Lines Matching refs:xx

143 static inline Int qadd32S ( Int xx, Int yy ) 
145 Long t = ((Long)xx) + ((Long)yy);
153 static inline Short qadd16S ( Short xx, Short yy )
155 Int t = ((Int)xx) + ((Int)yy);
161 static inline Char qadd8S ( Char xx, Char yy )
163 Int t = ((Int)xx) + ((Int)yy);
169 static inline UShort qadd16U ( UShort xx, UShort yy )
171 UInt t = ((UInt)xx) + ((UInt)yy);
176 static inline UChar qadd8U ( UChar xx, UChar yy )
178 UInt t = ((UInt)xx) + ((UInt)yy);
183 static inline Int qsub32S ( Int xx, Int yy )
185 Long t = ((Long)xx) - ((Long)yy);
193 static inline Short qsub16S ( Short xx, Short yy )
195 Int t = ((Int)xx) - ((Int)yy);
201 static inline Char qsub8S ( Char xx, Char yy )
203 Int t = ((Int)xx) - ((Int)yy);
209 static inline UShort qsub16U ( UShort xx, UShort yy )
211 Int t = ((Int)xx) - ((Int)yy);
217 static inline UChar qsub8U ( UChar xx, UChar yy )
219 Int t = ((Int)xx) - ((Int)yy);
225 static inline Short mul16 ( Short xx, Short yy )
227 Int t = ((Int)xx) * ((Int)yy);
231 static inline Int mul32 ( Int xx, Int yy )
233 Int t = ((Int)xx) * ((Int)yy);
237 static inline Short mulhi16S ( Short xx, Short yy )
239 Int t = ((Int)xx) * ((Int)yy);
244 static inline UShort mulhi16U ( UShort xx, UShort yy )
246 UInt t = ((UInt)xx) * ((UInt)yy);
251 static inline UInt cmpeq32 ( UInt xx, UInt yy )
253 return xx==yy ? 0xFFFFFFFF : 0;
256 static inline UShort cmpeq16 ( UShort xx, UShort yy )
258 return toUShort(xx==yy ? 0xFFFF : 0);
261 static inline UChar cmpeq8 ( UChar xx, UChar yy )
263 return toUChar(xx==yy ? 0xFF : 0);
266 static inline UInt cmpgt32S ( Int xx, Int yy )
268 return xx>yy ? 0xFFFFFFFF : 0;
271 static inline UShort cmpgt16S ( Short xx, Short yy )
273 return toUShort(xx>yy ? 0xFFFF : 0);
276 static inline UChar cmpgt8S ( Char xx, Char yy )
278 return toUChar(xx>yy ? 0xFF : 0);
281 static inline UInt cmpnez32 ( UInt xx )
283 return xx==0 ? 0 : 0xFFFFFFFF;
286 static inline UShort cmpnez16 ( UShort xx )
288 return toUShort(xx==0 ? 0 : 0xFFFF);
291 static inline UChar cmpnez8 ( UChar xx )
293 return toUChar(xx==0 ? 0 : 0xFF);
298 Int xx = (Int)xx0;
299 if (xx < -32768) xx = -32768;
300 if (xx > 32767) xx = 32767;
301 return (Short)xx;
306 Short xx = (Short)xx0;
307 if (xx < -128) xx = -128;
308 if (xx > 127) xx = 127;
309 return (Char)xx;
314 Short xx = (Short)xx0;
315 if (xx < 0) xx = 0;
316 if (xx > 255) xx = 255;
317 return (UChar)xx;
320 static inline UShort narrow32to16 ( UInt xx )
322 return (UShort)xx;
325 static inline UChar narrow16to8 ( UShort xx )
327 return (UChar)xx;
373 static inline UChar avg8U ( UChar xx, UChar yy )
375 UInt xxi = (UInt)xx;
381 static inline UShort avg16U ( UShort xx, UShort yy )
383 UInt xxi = (UInt)xx;
389 static inline Short max16S ( Short xx, Short yy )
391 return toUShort((xx > yy) ? xx : yy);
394 static inline UChar max8U ( UChar xx, UChar yy )
396 return toUChar((xx > yy) ? xx : yy);
399 static inline Short min16S ( Short xx, Short yy )
401 return toUShort((xx < yy) ? xx : yy);
404 static inline UChar min8U ( UChar xx, UChar yy )
406 return toUChar((xx < yy) ? xx : yy);
409 static inline UShort hadd16U ( UShort xx, UShort yy )
411 UInt xxi = (UInt)xx;
417 static inline Short hadd16S ( Short xx, Short yy )
419 Int xxi = (Int)xx;
425 static inline UShort hsub16U ( UShort xx, UShort yy )
427 UInt xxi = (UInt)xx;
433 static inline Short hsub16S ( Short xx, Short yy )
435 Int xxi = (Int)xx;
441 static inline UChar hadd8U ( UChar xx, UChar yy )
443 UInt xxi = (UInt)xx;
449 static inline Char hadd8S ( Char xx, Char yy )
451 Int xxi = (Int)xx;
457 static inline UChar hsub8U ( UChar xx, UChar yy )
459 UInt xxi = (UInt)xx;
465 static inline Char hsub8S ( Char xx, Char yy )
467 Int xxi = (Int)xx;
473 static inline UInt absdiff8U ( UChar xx, UChar yy )
475 UInt xxu = (UChar)xx;
487 ULong h_generic_calc_Add32x2 ( ULong xx, ULong yy )
490 sel32x2_1(xx) + sel32x2_1(yy),
491 sel32x2_0(xx) + sel32x2_0(yy)
495 ULong h_generic_calc_Add16x4 ( ULong xx, ULong yy )
498 toUShort( sel16x4_3(xx) + sel16x4_3(yy) ),
499 toUShort( sel16x4_2(xx) + sel16x4_2(yy) ),
500 toUShort( sel16x4_1(xx) + sel16x4_1(yy) ),
501 toUShort( sel16x4_0(xx) + sel16x4_0(yy) )
505 ULong h_generic_calc_Add8x8 ( ULong xx, ULong yy )
508 toUChar( sel8x8_7(xx) + sel8x8_7(yy) ),
509 toUChar( sel8x8_6(xx) + sel8x8_6(yy) ),
510 toUChar( sel8x8_5(xx) + sel8x8_5(yy) ),
511 toUChar( sel8x8_4(xx) + sel8x8_4(yy) ),
512 toUChar( sel8x8_3(xx) + sel8x8_3(yy) ),
513 toUChar( sel8x8_2(xx) + sel8x8_2(yy) ),
514 toUChar( sel8x8_1(xx) + sel8x8_1(yy) ),
515 toUChar( sel8x8_0(xx) + sel8x8_0(yy) )
521 ULong h_generic_calc_QAdd16Sx4 ( ULong xx, ULong yy )
524 qadd16S( sel16x4_3(xx), sel16x4_3(yy) ),
525 qadd16S( sel16x4_2(xx), sel16x4_2(yy) ),
526 qadd16S( sel16x4_1(xx), sel16x4_1(yy) ),
527 qadd16S( sel16x4_0(xx), sel16x4_0(yy) )
531 ULong h_generic_calc_QAdd8Sx8 ( ULong xx, ULong yy )
534 qadd8S( sel8x8_7(xx), sel8x8_7(yy) ),
535 qadd8S( sel8x8_6(xx), sel8x8_6(yy) ),
536 qadd8S( sel8x8_5(xx), sel8x8_5(yy) ),
537 qadd8S( sel8x8_4(xx), sel8x8_4(yy) ),
538 qadd8S( sel8x8_3(xx), sel8x8_3(yy) ),
539 qadd8S( sel8x8_2(xx), sel8x8_2(yy) ),
540 qadd8S( sel8x8_1(xx), sel8x8_1(yy) ),
541 qadd8S( sel8x8_0(xx), sel8x8_0(yy) )
545 ULong h_generic_calc_QAdd16Ux4 ( ULong xx, ULong yy )
548 qadd16U( sel16x4_3(xx), sel16x4_3(yy) ),
549 qadd16U( sel16x4_2(xx), sel16x4_2(yy) ),
550 qadd16U( sel16x4_1(xx), sel16x4_1(yy) ),
551 qadd16U( sel16x4_0(xx), sel16x4_0(yy) )
555 ULong h_generic_calc_QAdd8Ux8 ( ULong xx, ULong yy )
558 qadd8U( sel8x8_7(xx), sel8x8_7(yy) ),
559 qadd8U( sel8x8_6(xx), sel8x8_6(yy) ),
560 qadd8U( sel8x8_5(xx), sel8x8_5(yy) ),
561 qadd8U( sel8x8_4(xx), sel8x8_4(yy) ),
562 qadd8U( sel8x8_3(xx), sel8x8_3(yy) ),
563 qadd8U( sel8x8_2(xx), sel8x8_2(yy) ),
564 qadd8U( sel8x8_1(xx), sel8x8_1(yy) ),
565 qadd8U( sel8x8_0(xx), sel8x8_0(yy) )
571 ULong h_generic_calc_Sub32x2 ( ULong xx, ULong yy )
574 sel32x2_1(xx) - sel32x2_1(yy),
575 sel32x2_0(xx) - sel32x2_0(yy)
579 ULong h_generic_calc_Sub16x4 ( ULong xx, ULong yy )
582 toUShort( sel16x4_3(xx) - sel16x4_3(yy) ),
583 toUShort( sel16x4_2(xx) - sel16x4_2(yy) ),
584 toUShort( sel16x4_1(xx) - sel16x4_1(yy) ),
585 toUShort( sel16x4_0(xx) - sel16x4_0(yy) )
589 ULong h_generic_calc_Sub8x8 ( ULong xx, ULong yy )
592 toUChar( sel8x8_7(xx) - sel8x8_7(yy) ),
593 toUChar( sel8x8_6(xx) - sel8x8_6(yy) ),
594 toUChar( sel8x8_5(xx) - sel8x8_5(yy) ),
595 toUChar( sel8x8_4(xx) - sel8x8_4(yy) ),
596 toUChar( sel8x8_3(xx) - sel8x8_3(yy) ),
597 toUChar( sel8x8_2(xx) - sel8x8_2(yy) ),
598 toUChar( sel8x8_1(xx) - sel8x8_1(yy) ),
599 toUChar( sel8x8_0(xx) - sel8x8_0(yy) )
605 ULong h_generic_calc_QSub16Sx4 ( ULong xx, ULong yy )
608 qsub16S( sel16x4_3(xx), sel16x4_3(yy) ),
609 qsub16S( sel16x4_2(xx), sel16x4_2(yy) ),
610 qsub16S( sel16x4_1(xx), sel16x4_1(yy) ),
611 qsub16S( sel16x4_0(xx), sel16x4_0(yy) )
615 ULong h_generic_calc_QSub8Sx8 ( ULong xx, ULong yy )
618 qsub8S( sel8x8_7(xx), sel8x8_7(yy) ),
619 qsub8S( sel8x8_6(xx), sel8x8_6(yy) ),
620 qsub8S( sel8x8_5(xx), sel8x8_5(yy) ),
621 qsub8S( sel8x8_4(xx), sel8x8_4(yy) ),
622 qsub8S( sel8x8_3(xx), sel8x8_3(yy) ),
623 qsub8S( sel8x8_2(xx), sel8x8_2(yy) ),
624 qsub8S( sel8x8_1(xx), sel8x8_1(yy) ),
625 qsub8S( sel8x8_0(xx), sel8x8_0(yy) )
629 ULong h_generic_calc_QSub16Ux4 ( ULong xx, ULong yy )
632 qsub16U( sel16x4_3(xx), sel16x4_3(yy) ),
633 qsub16U( sel16x4_2(xx), sel16x4_2(yy) ),
634 qsub16U( sel16x4_1(xx), sel16x4_1(yy) ),
635 qsub16U( sel16x4_0(xx), sel16x4_0(yy) )
639 ULong h_generic_calc_QSub8Ux8 ( ULong xx, ULong yy )
642 qsub8U( sel8x8_7(xx), sel8x8_7(yy) ),
643 qsub8U( sel8x8_6(xx), sel8x8_6(yy) ),
644 qsub8U( sel8x8_5(xx), sel8x8_5(yy) ),
645 qsub8U( sel8x8_4(xx), sel8x8_4(yy) ),
646 qsub8U( sel8x8_3(xx), sel8x8_3(yy) ),
647 qsub8U( sel8x8_2(xx), sel8x8_2(yy) ),
648 qsub8U( sel8x8_1(xx), sel8x8_1(yy) ),
649 qsub8U( sel8x8_0(xx), sel8x8_0(yy) )
655 ULong h_generic_calc_Mul16x4 ( ULong xx, ULong yy )
658 mul16( sel16x4_3(xx), sel16x4_3(yy) ),
659 mul16( sel16x4_2(xx), sel16x4_2(yy) ),
660 mul16( sel16x4_1(xx), sel16x4_1(yy) ),
661 mul16( sel16x4_0(xx), sel16x4_0(yy) )
665 ULong h_generic_calc_Mul32x2 ( ULong xx, ULong yy )
668 mul32( sel32x2_1(xx), sel32x2_1(yy) ),
669 mul32( sel32x2_0(xx), sel32x2_0(yy) )
673 ULong h_generic_calc_MulHi16Sx4 ( ULong xx, ULong yy )
676 mulhi16S( sel16x4_3(xx), sel16x4_3(yy) ),
677 mulhi16S( sel16x4_2(xx), sel16x4_2(yy) ),
678 mulhi16S( sel16x4_1(xx), sel16x4_1(yy) ),
679 mulhi16S( sel16x4_0(xx), sel16x4_0(yy) )
683 ULong h_generic_calc_MulHi16Ux4 ( ULong xx, ULong yy )
686 mulhi16U( sel16x4_3(xx), sel16x4_3(yy) ),
687 mulhi16U( sel16x4_2(xx), sel16x4_2(yy) ),
688 mulhi16U( sel16x4_1(xx), sel16x4_1(yy) ),
689 mulhi16U( sel16x4_0(xx), sel16x4_0(yy) )
695 ULong h_generic_calc_CmpEQ32x2 ( ULong xx, ULong yy )
698 cmpeq32( sel32x2_1(xx), sel32x2_1(yy) ),
699 cmpeq32( sel32x2_0(xx), sel32x2_0(yy) )
703 ULong h_generic_calc_CmpEQ16x4 ( ULong xx, ULong yy )
706 cmpeq16( sel16x4_3(xx), sel16x4_3(yy) ),
707 cmpeq16( sel16x4_2(xx), sel16x4_2(yy) ),
708 cmpeq16( sel16x4_1(xx), sel16x4_1(yy) ),
709 cmpeq16( sel16x4_0(xx), sel16x4_0(yy) )
713 ULong h_generic_calc_CmpEQ8x8 ( ULong xx, ULong yy )
716 cmpeq8( sel8x8_7(xx), sel8x8_7(yy) ),
717 cmpeq8( sel8x8_6(xx), sel8x8_6(yy) ),
718 cmpeq8( sel8x8_5(xx), sel8x8_5(yy) ),
719 cmpeq8( sel8x8_4(xx), sel8x8_4(yy) ),
720 cmpeq8( sel8x8_3(xx), sel8x8_3(yy) ),
721 cmpeq8( sel8x8_2(xx), sel8x8_2(yy) ),
722 cmpeq8( sel8x8_1(xx), sel8x8_1(yy) ),
723 cmpeq8( sel8x8_0(xx), sel8x8_0(yy) )
727 ULong h_generic_calc_CmpGT32Sx2 ( ULong xx, ULong yy )
730 cmpgt32S( sel32x2_1(xx), sel32x2_1(yy) ),
731 cmpgt32S( sel32x2_0(xx), sel32x2_0(yy) )
735 ULong h_generic_calc_CmpGT16Sx4 ( ULong xx, ULong yy )
738 cmpgt16S( sel16x4_3(xx), sel16x4_3(yy) ),
739 cmpgt16S( sel16x4_2(xx), sel16x4_2(yy) ),
740 cmpgt16S( sel16x4_1(xx), sel16x4_1(yy) ),
741 cmpgt16S( sel16x4_0(xx), sel16x4_0(yy) )
745 ULong h_generic_calc_CmpGT8Sx8 ( ULong xx, ULong yy )
748 cmpgt8S( sel8x8_7(xx), sel8x8_7(yy) ),
749 cmpgt8S( sel8x8_6(xx), sel8x8_6(yy) ),
750 cmpgt8S( sel8x8_5(xx), sel8x8_5(yy) ),
751 cmpgt8S( sel8x8_4(xx), sel8x8_4(yy) ),
752 cmpgt8S( sel8x8_3(xx), sel8x8_3(yy) ),
753 cmpgt8S( sel8x8_2(xx), sel8x8_2(yy) ),
754 cmpgt8S( sel8x8_1(xx), sel8x8_1(yy) ),
755 cmpgt8S( sel8x8_0(xx), sel8x8_0(yy) )
759 ULong h_generic_calc_CmpNEZ32x2 ( ULong xx )
762 cmpnez32( sel32x2_1(xx) ),
763 cmpnez32( sel32x2_0(xx) )
767 ULong h_generic_calc_CmpNEZ16x4 ( ULong xx )
770 cmpnez16( sel16x4_3(xx) ),
771 cmpnez16( sel16x4_2(xx) ),
772 cmpnez16( sel16x4_1(xx) ),
773 cmpnez16( sel16x4_0(xx) )
777 ULong h_generic_calc_CmpNEZ8x8 ( ULong xx )
780 cmpnez8( sel8x8_7(xx) ),
781 cmpnez8( sel8x8_6(xx) ),
782 cmpnez8( sel8x8_5(xx) ),
783 cmpnez8( sel8x8_4(xx) ),
784 cmpnez8( sel8x8_3(xx) ),
785 cmpnez8( sel8x8_2(xx) ),
786 cmpnez8( sel8x8_1(xx) ),
787 cmpnez8( sel8x8_0(xx) )
999 ULong h_generic_calc_ShlN32x2 ( ULong xx, UInt nn )
1004 shl32( sel32x2_1(xx), nn ),
1005 shl32( sel32x2_0(xx), nn )
1009 ULong h_generic_calc_ShlN16x4 ( ULong xx, UInt nn )
1014 shl16( sel16x4_3(xx), nn ),
1015 shl16( sel16x4_2(xx), nn ),
1016 shl16( sel16x4_1(xx), nn ),
1017 shl16( sel16x4_0(xx), nn )
1021 ULong h_generic_calc_ShlN8x8 ( ULong xx, UInt nn )
1026 shl8( sel8x8_7(xx), nn ),
1027 shl8( sel8x8_6(xx), nn ),
1028 shl8( sel8x8_5(xx), nn ),
1029 shl8( sel8x8_4(xx), nn ),
1030 shl8( sel8x8_3(xx), nn ),
1031 shl8( sel8x8_2(xx), nn ),
1032 shl8( sel8x8_1(xx), nn ),
1033 shl8( sel8x8_0(xx), nn )
1037 ULong h_generic_calc_ShrN32x2 ( ULong xx, UInt nn )
1042 shr32( sel32x2_1(xx), nn ),
1043 shr32( sel32x2_0(xx), nn )
1047 ULong h_generic_calc_ShrN16x4 ( ULong xx, UInt nn )
1052 shr16( sel16x4_3(xx), nn ),
1053 shr16( sel16x4_2(xx), nn ),
1054 shr16( sel16x4_1(xx), nn ),
1055 shr16( sel16x4_0(xx), nn )
1059 ULong h_generic_calc_SarN32x2 ( ULong xx, UInt nn )
1064 sar32( sel32x2_1(xx), nn ),
1065 sar32( sel32x2_0(xx), nn )
1069 ULong h_generic_calc_SarN16x4 ( ULong xx, UInt nn )
1074 sar16( sel16x4_3(xx), nn ),
1075 sar16( sel16x4_2(xx), nn ),
1076 sar16( sel16x4_1(xx), nn ),
1077 sar16( sel16x4_0(xx), nn )
1081 ULong h_generic_calc_SarN8x8 ( ULong xx, UInt nn )
1086 sar8( sel8x8_7(xx), nn ),
1087 sar8( sel8x8_6(xx), nn ),
1088 sar8( sel8x8_5(xx), nn ),
1089 sar8( sel8x8_4(xx), nn ),
1090 sar8( sel8x8_3(xx), nn ),
1091 sar8( sel8x8_2(xx), nn ),
1092 sar8( sel8x8_1(xx), nn ),
1093 sar8( sel8x8_0(xx), nn )
1099 ULong h_generic_calc_Avg8Ux8 ( ULong xx, ULong yy )
1102 avg8U( sel8x8_7(xx), sel8x8_7(yy) ),
1103 avg8U( sel8x8_6(xx), sel8x8_6(yy) ),
1104 avg8U( sel8x8_5(xx), sel8x8_5(yy) ),
1105 avg8U( sel8x8_4(xx), sel8x8_4(yy) ),
1106 avg8U( sel8x8_3(xx), sel8x8_3(yy) ),
1107 avg8U( sel8x8_2(xx), sel8x8_2(yy) ),
1108 avg8U( sel8x8_1(xx), sel8x8_1(yy) ),
1109 avg8U( sel8x8_0(xx), sel8x8_0(yy) )
1113 ULong h_generic_calc_Avg16Ux4 ( ULong xx, ULong yy )
1116 avg16U( sel16x4_3(xx), sel16x4_3(yy) ),
1117 avg16U( sel16x4_2(xx), sel16x4_2(yy) ),
1118 avg16U( sel16x4_1(xx), sel16x4_1(yy) ),
1119 avg16U( sel16x4_0(xx), sel16x4_0(yy) )
1125 ULong h_generic_calc_Max16Sx4 ( ULong xx, ULong yy )
1128 max16S( sel16x4_3(xx), sel16x4_3(yy) ),
1129 max16S( sel16x4_2(xx), sel16x4_2(yy) ),
1130 max16S( sel16x4_1(xx), sel16x4_1(yy) ),
1131 max16S( sel16x4_0(xx), sel16x4_0(yy) )
1135 ULong h_generic_calc_Max8Ux8 ( ULong xx, ULong yy )
1138 max8U( sel8x8_7(xx), sel8x8_7(yy) ),
1139 max8U( sel8x8_6(xx), sel8x8_6(yy) ),
1140 max8U( sel8x8_5(xx), sel8x8_5(yy) ),
1141 max8U( sel8x8_4(xx), sel8x8_4(yy) ),
1142 max8U( sel8x8_3(xx), sel8x8_3(yy) ),
1143 max8U( sel8x8_2(xx), sel8x8_2(yy) ),
1144 max8U( sel8x8_1(xx), sel8x8_1(yy) ),
1145 max8U( sel8x8_0(xx), sel8x8_0(yy) )
1149 ULong h_generic_calc_Min16Sx4 ( ULong xx, ULong yy )
1152 min16S( sel16x4_3(xx), sel16x4_3(yy) ),
1153 min16S( sel16x4_2(xx), sel16x4_2(yy) ),
1154 min16S( sel16x4_1(xx), sel16x4_1(yy) ),
1155 min16S( sel16x4_0(xx), sel16x4_0(yy) )
1159 ULong h_generic_calc_Min8Ux8 ( ULong xx, ULong yy )
1162 min8U( sel8x8_7(xx), sel8x8_7(yy) ),
1163 min8U( sel8x8_6(xx), sel8x8_6(yy) ),
1164 min8U( sel8x8_5(xx), sel8x8_5(yy) ),
1165 min8U( sel8x8_4(xx), sel8x8_4(yy) ),
1166 min8U( sel8x8_3(xx), sel8x8_3(yy) ),
1167 min8U( sel8x8_2(xx), sel8x8_2(yy) ),
1168 min8U( sel8x8_1(xx), sel8x8_1(yy) ),
1169 min8U( sel8x8_0(xx), sel8x8_0(yy) )
1173 UInt h_generic_calc_GetMSBs8x8 ( ULong xx )
1176 if (xx & (1ULL << (64-1))) r |= (1<<7);
1177 if (xx & (1ULL << (56-1))) r |= (1<<6);
1178 if (xx & (1ULL << (48-1))) r |= (1<<5);
1179 if (xx & (1ULL << (40-1))) r |= (1<<4);
1180 if (xx & (1ULL << (32-1))) r |= (1<<3);
1181 if (xx & (1ULL << (24-1))) r |= (1<<2);
1182 if (xx & (1ULL << (16-1))) r |= (1<<1);
1183 if (xx & (1ULL << ( 8-1))) r |= (1<<0);
1229 UInt h_generic_calc_Add16x2 ( UInt xx, UInt yy )
1231 return mk16x2( sel16x2_1(xx) + sel16x2_1(yy),
1232 sel16x2_0(xx) + sel16x2_0(yy) );
1235 UInt h_generic_calc_Sub16x2 ( UInt xx, UInt yy )
1237 return mk16x2( sel16x2_1(xx) - sel16x2_1(yy),
1238 sel16x2_0(xx) - sel16x2_0(yy) );
1241 UInt h_generic_calc_HAdd16Ux2 ( UInt xx, UInt yy )
1243 return mk16x2( hadd16U( sel16x2_1(xx), sel16x2_1(yy) ),
1244 hadd16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1247 UInt h_generic_calc_HAdd16Sx2 ( UInt xx, UInt yy )
1249 return mk16x2( hadd16S( sel16x2_1(xx), sel16x2_1(yy) ),
1250 hadd16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1253 UInt h_generic_calc_HSub16Ux2 ( UInt xx, UInt yy )
1255 return mk16x2( hsub16U( sel16x2_1(xx), sel16x2_1(yy) ),
1256 hsub16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1259 UInt h_generic_calc_HSub16Sx2 ( UInt xx, UInt yy )
1261 return mk16x2( hsub16S( sel16x2_1(xx), sel16x2_1(yy) ),
1262 hsub16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1265 UInt h_generic_calc_QAdd16Ux2 ( UInt xx, UInt yy )
1267 return mk16x2( qadd16U( sel16x2_1(xx), sel16x2_1(yy) ),
1268 qadd16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1271 UInt h_generic_calc_QAdd16Sx2 ( UInt xx, UInt yy )
1273 return mk16x2( qadd16S( sel16x2_1(xx), sel16x2_1(yy) ),
1274 qadd16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1277 UInt h_generic_calc_QSub16Ux2 ( UInt xx, UInt yy )
1279 return mk16x2( qsub16U( sel16x2_1(xx), sel16x2_1(yy) ),
1280 qsub16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1283 UInt h_generic_calc_QSub16Sx2 ( UInt xx, UInt yy )
1285 return mk16x2( qsub16S( sel16x2_1(xx), sel16x2_1(yy) ),
1286 qsub16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1291 UInt h_generic_calc_Add8x4 ( UInt xx, UInt yy )
1294 sel8x4_3(xx) + sel8x4_3(yy),
1295 sel8x4_2(xx) + sel8x4_2(yy),
1296 sel8x4_1(xx) + sel8x4_1(yy),
1297 sel8x4_0(xx) + sel8x4_0(yy)
1301 UInt h_generic_calc_Sub8x4 ( UInt xx, UInt yy )
1304 sel8x4_3(xx) - sel8x4_3(yy),
1305 sel8x4_2(xx) - sel8x4_2(yy),
1306 sel8x4_1(xx) - sel8x4_1(yy),
1307 sel8x4_0(xx) - sel8x4_0(yy)
1311 UInt h_generic_calc_HAdd8Ux4 ( UInt xx, UInt yy )
1314 hadd8U( sel8x4_3(xx), sel8x4_3(yy) ),
1315 hadd8U( sel8x4_2(xx), sel8x4_2(yy) ),
1316 hadd8U( sel8x4_1(xx), sel8x4_1(yy) ),
1317 hadd8U( sel8x4_0(xx), sel8x4_0(yy) )
1321 UInt h_generic_calc_HAdd8Sx4 ( UInt xx, UInt yy )
1324 hadd8S( sel8x4_3(xx), sel8x4_3(yy) ),
1325 hadd8S( sel8x4_2(xx), sel8x4_2(yy) ),
1326 hadd8S( sel8x4_1(xx), sel8x4_1(yy) ),
1327 hadd8S( sel8x4_0(xx), sel8x4_0(yy) )
1331 UInt h_generic_calc_HSub8Ux4 ( UInt xx, UInt yy )
1334 hsub8U( sel8x4_3(xx), sel8x4_3(yy) ),
1335 hsub8U( sel8x4_2(xx), sel8x4_2(yy) ),
1336 hsub8U( sel8x4_1(xx), sel8x4_1(yy) ),
1337 hsub8U( sel8x4_0(xx), sel8x4_0(yy) )
1341 UInt h_generic_calc_HSub8Sx4 ( UInt xx, UInt yy )
1344 hsub8S( sel8x4_3(xx), sel8x4_3(yy) ),
1345 hsub8S( sel8x4_2(xx), sel8x4_2(yy) ),
1346 hsub8S( sel8x4_1(xx), sel8x4_1(yy) ),
1347 hsub8S( sel8x4_0(xx), sel8x4_0(yy) )
1351 UInt h_generic_calc_QAdd8Ux4 ( UInt xx, UInt yy )
1354 qadd8U( sel8x4_3(xx), sel8x4_3(yy) ),
1355 qadd8U( sel8x4_2(xx), sel8x4_2(yy) ),
1356 qadd8U( sel8x4_1(xx), sel8x4_1(yy) ),
1357 qadd8U( sel8x4_0(xx), sel8x4_0(yy) )
1361 UInt h_generic_calc_QAdd8Sx4 ( UInt xx, UInt yy )
1364 qadd8S( sel8x4_3(xx), sel8x4_3(yy) ),
1365 qadd8S( sel8x4_2(xx), sel8x4_2(yy) ),
1366 qadd8S( sel8x4_1(xx), sel8x4_1(yy) ),
1367 qadd8S( sel8x4_0(xx), sel8x4_0(yy) )
1371 UInt h_generic_calc_QSub8Ux4 ( UInt xx, UInt yy )
1374 qsub8U( sel8x4_3(xx), sel8x4_3(yy) ),
1375 qsub8U( sel8x4_2(xx), sel8x4_2(yy) ),
1376 qsub8U( sel8x4_1(xx), sel8x4_1(yy) ),
1377 qsub8U( sel8x4_0(xx), sel8x4_0(yy) )
1381 UInt h_generic_calc_QSub8Sx4 ( UInt xx, UInt yy )
1384 qsub8S( sel8x4_3(xx), sel8x4_3(yy) ),
1385 qsub8S( sel8x4_2(xx), sel8x4_2(yy) ),
1386 qsub8S( sel8x4_1(xx), sel8x4_1(yy) ),
1387 qsub8S( sel8x4_0(xx), sel8x4_0(yy) )
1391 UInt h_generic_calc_CmpNEZ16x2 ( UInt xx )
1394 cmpnez16( sel16x2_1(xx) ),
1395 cmpnez16( sel16x2_0(xx) )
1399 UInt h_generic_calc_CmpNEZ8x4 ( UInt xx )
1402 cmpnez8( sel8x4_3(xx) ),
1403 cmpnez8( sel8x4_2(xx) ),
1404 cmpnez8( sel8x4_1(xx) ),
1405 cmpnez8( sel8x4_0(xx) )
1409 UInt h_generic_calc_Sad8Ux4 ( UInt xx, UInt yy )
1411 return absdiff8U( sel8x4_3(xx), sel8x4_3(yy) )
1412 + absdiff8U( sel8x4_2(xx), sel8x4_2(yy) )
1413 + absdiff8U( sel8x4_1(xx), sel8x4_1(yy) )
1414 + absdiff8U( sel8x4_0(xx), sel8x4_0(yy) );
1417 UInt h_generic_calc_QAdd32S ( UInt xx, UInt yy )
1419 return qadd32S( xx, yy );
1422 UInt h_generic_calc_QSub32S ( UInt xx, UInt yy )
1424 return qsub32S( xx, yy );