1/*
2 *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.
3 *
4 *  Use of this source code is governed by a BSD-style license
5 *  that can be found in the LICENSE file in the root of the source
6 *  tree. An additional intellectual property rights grant can be found
7 *  in the file PATENTS.  All contributing project authors may
8 *  be found in the AUTHORS file in the root of the source tree.
9 */
10
11
12#include "vpx_ports/config.h"
13#include "vpx_ports/arm.h"
14#include "vp8/encoder/variance.h"
15#include "vp8/encoder/onyx_int.h"
16
17extern void (*vp8_yv12_copy_partial_frame_ptr)(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
18extern void vp8_yv12_copy_partial_frame(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
19extern void vpxyv12_copy_partial_frame_neon(YV12_BUFFER_CONFIG *src_ybc, YV12_BUFFER_CONFIG *dst_ybc, int Fraction);
20
21void vp8_arch_arm_encoder_init(VP8_COMP *cpi)
22{
23#if CONFIG_RUNTIME_CPU_DETECT
24    int flags = cpi->common.rtcd.flags;
25    int has_edsp = flags & HAS_EDSP;
26    int has_media = flags & HAS_MEDIA;
27    int has_neon = flags & HAS_NEON;
28
29#if HAVE_ARMV6
30    if (has_media)
31    {
32        cpi->rtcd.variance.sad16x16              = vp8_sad16x16_armv6;
33        /*cpi->rtcd.variance.sad16x8               = vp8_sad16x8_c;
34        cpi->rtcd.variance.sad8x16               = vp8_sad8x16_c;
35        cpi->rtcd.variance.sad8x8                = vp8_sad8x8_c;
36        cpi->rtcd.variance.sad4x4                = vp8_sad4x4_c;*/
37
38        /*cpi->rtcd.variance.var4x4                = vp8_variance4x4_c;*/
39        cpi->rtcd.variance.var8x8                = vp8_variance8x8_armv6;
40        /*cpi->rtcd.variance.var8x16               = vp8_variance8x16_c;
41        cpi->rtcd.variance.var16x8               = vp8_variance16x8_c;*/
42        cpi->rtcd.variance.var16x16              = vp8_variance16x16_armv6;
43
44        /*cpi->rtcd.variance.subpixvar4x4          = vp8_sub_pixel_variance4x4_c;*/
45        cpi->rtcd.variance.subpixvar8x8          = vp8_sub_pixel_variance8x8_armv6;
46        /*cpi->rtcd.variance.subpixvar8x16         = vp8_sub_pixel_variance8x16_c;
47        cpi->rtcd.variance.subpixvar16x8         = vp8_sub_pixel_variance16x8_c;*/
48        cpi->rtcd.variance.subpixvar16x16        = vp8_sub_pixel_variance16x16_armv6;
49        cpi->rtcd.variance.halfpixvar16x16_h     = vp8_variance_halfpixvar16x16_h_armv6;
50        cpi->rtcd.variance.halfpixvar16x16_v     = vp8_variance_halfpixvar16x16_v_armv6;
51        cpi->rtcd.variance.halfpixvar16x16_hv    = vp8_variance_halfpixvar16x16_hv_armv6;
52
53        cpi->rtcd.variance.mse16x16              = vp8_mse16x16_armv6;
54        /*cpi->rtcd.variance.getmbss               = vp8_get_mb_ss_c;*/
55
56        /*cpi->rtcd.variance.get16x16prederror     = vp8_get16x16pred_error_c;
57        cpi->rtcd.variance.get8x8var             = vp8_get8x8var_c;
58        cpi->rtcd.variance.get16x16var           = vp8_get16x16var_c;;
59        cpi->rtcd.variance.get4x4sse_cs          = vp8_get4x4sse_cs_c;*/
60
61        /*cpi->rtcd.fdct.short4x4                  = vp8_short_fdct4x4_c;
62        cpi->rtcd.fdct.short8x4                  = vp8_short_fdct8x4_c;*/
63        cpi->rtcd.fdct.fast4x4                   = vp8_fast_fdct4x4_armv6;
64        cpi->rtcd.fdct.fast8x4                   = vp8_fast_fdct8x4_armv6;
65        cpi->rtcd.fdct.walsh_short4x4            = vp8_short_walsh4x4_armv6;
66
67        /*cpi->rtcd.encodemb.berr                  = vp8_block_error_c;
68        cpi->rtcd.encodemb.mberr                 = vp8_mbblock_error_c;
69        cpi->rtcd.encodemb.mbuverr               = vp8_mbuverror_c;*/
70        cpi->rtcd.encodemb.subb                  = vp8_subtract_b_armv6;
71        cpi->rtcd.encodemb.submby                = vp8_subtract_mby_armv6;
72        cpi->rtcd.encodemb.submbuv               = vp8_subtract_mbuv_armv6;
73
74        /*cpi->rtcd.quantize.quantb                = vp8_regular_quantize_b;*/
75        cpi->rtcd.quantize.fastquantb            = vp8_fast_quantize_b_armv6;
76    }
77#endif
78
79#if HAVE_ARMV7
80    if (has_neon)
81    {
82        cpi->rtcd.variance.sad16x16              = vp8_sad16x16_neon;
83        cpi->rtcd.variance.sad16x8               = vp8_sad16x8_neon;
84        cpi->rtcd.variance.sad8x16               = vp8_sad8x16_neon;
85        cpi->rtcd.variance.sad8x8                = vp8_sad8x8_neon;
86        cpi->rtcd.variance.sad4x4                = vp8_sad4x4_neon;
87
88        /*cpi->rtcd.variance.var4x4                = vp8_variance4x4_c;*/
89        cpi->rtcd.variance.var8x8                = vp8_variance8x8_neon;
90        cpi->rtcd.variance.var8x16               = vp8_variance8x16_neon;
91        cpi->rtcd.variance.var16x8               = vp8_variance16x8_neon;
92        cpi->rtcd.variance.var16x16              = vp8_variance16x16_neon;
93
94        /*cpi->rtcd.variance.subpixvar4x4          = vp8_sub_pixel_variance4x4_c;*/
95        cpi->rtcd.variance.subpixvar8x8          = vp8_sub_pixel_variance8x8_neon;
96        /*cpi->rtcd.variance.subpixvar8x16         = vp8_sub_pixel_variance8x16_c;
97        cpi->rtcd.variance.subpixvar16x8         = vp8_sub_pixel_variance16x8_c;*/
98        cpi->rtcd.variance.subpixvar16x16        = vp8_sub_pixel_variance16x16_neon;
99        cpi->rtcd.variance.halfpixvar16x16_h     = vp8_variance_halfpixvar16x16_h_neon;
100        cpi->rtcd.variance.halfpixvar16x16_v     = vp8_variance_halfpixvar16x16_v_neon;
101        cpi->rtcd.variance.halfpixvar16x16_hv    = vp8_variance_halfpixvar16x16_hv_neon;
102
103        cpi->rtcd.variance.mse16x16              = vp8_mse16x16_neon;
104        /*cpi->rtcd.variance.getmbss               = vp8_get_mb_ss_c;*/
105
106        cpi->rtcd.variance.get16x16prederror     = vp8_get16x16pred_error_neon;
107        /*cpi->rtcd.variance.get8x8var             = vp8_get8x8var_c;
108        cpi->rtcd.variance.get16x16var           = vp8_get16x16var_c;*/
109        cpi->rtcd.variance.get4x4sse_cs          = vp8_get4x4sse_cs_neon;
110
111        cpi->rtcd.fdct.short4x4                  = vp8_short_fdct4x4_neon;
112        cpi->rtcd.fdct.short8x4                  = vp8_short_fdct8x4_neon;
113        cpi->rtcd.fdct.fast4x4                   = vp8_fast_fdct4x4_neon;
114        cpi->rtcd.fdct.fast8x4                   = vp8_fast_fdct8x4_neon;
115        cpi->rtcd.fdct.walsh_short4x4            = vp8_short_walsh4x4_neon;
116
117        /*cpi->rtcd.encodemb.berr                  = vp8_block_error_c;
118        cpi->rtcd.encodemb.mberr                 = vp8_mbblock_error_c;
119        cpi->rtcd.encodemb.mbuverr               = vp8_mbuverror_c;*/
120        cpi->rtcd.encodemb.subb                  = vp8_subtract_b_neon;
121        cpi->rtcd.encodemb.submby                = vp8_subtract_mby_neon;
122        cpi->rtcd.encodemb.submbuv               = vp8_subtract_mbuv_neon;
123
124        /*cpi->rtcd.quantize.quantb                = vp8_regular_quantize_b;
125        cpi->rtcd.quantize.fastquantb            = vp8_fast_quantize_b_c;*/
126        /* The neon quantizer has not been updated to match the new exact
127         * quantizer introduced in commit e04e2935
128         */
129        /*cpi->rtcd.quantize.fastquantb            = vp8_fast_quantize_b_neon;*/
130    }
131#endif
132
133#if HAVE_ARMV7
134#if CONFIG_RUNTIME_CPU_DETECT
135    if (has_neon)
136#endif
137    {
138        vp8_yv12_copy_partial_frame_ptr = vpxyv12_copy_partial_frame_neon;
139    }
140#endif
141#endif
142}
143