return score;
}
-static int ssd_int8_vs_int16_c(int8_t *pix1, int16_t *pix2, int size){
+static int ssd_int8_vs_int16_c(const int8_t *pix1, const int16_t *pix2,
+ int size){
int score=0;
int i;
for(i=0; i<size; i++)
me_cmp_func ildct_cmp[5]; //only width 16 used
me_cmp_func frame_skip_cmp[5]; //only width 8 used
- int (*ssd_int8_vs_int16)(int8_t *pix1, int16_t *pix2, int size);
+ int (*ssd_int8_vs_int16)(const int8_t *pix1, const int16_t *pix2,
+ int size);
/**
* Halfpel motion compensation with rounding (a+b+1)>>1.
#undef HSUM
#undef DCT_SAD
-static int ssd_int8_vs_int16_mmx(int8_t *pix1, int16_t *pix2, int size){
+static int ssd_int8_vs_int16_mmx(const int8_t *pix1, const int16_t *pix2, int size){
int sum;
long i=size;
asm volatile(
#include "dsputil_altivec.h"
-static int ssd_int8_vs_int16_altivec(int8_t *pix1, int16_t *pix2, int size) {
+static int ssd_int8_vs_int16_altivec(const int8_t *pix1, const int16_t *pix2,
+ int size) {
int i, size16;
vector signed char vpix1;
vector signed short vpix2, vdiff, vpix1l,vpix1h;