1335 строки
60 KiB
C++
1335 строки
60 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebM project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include <cstdlib>
|
|
#include <new>
|
|
|
|
#include "third_party/googletest/src/include/gtest/gtest.h"
|
|
|
|
#include "./vpx_config.h"
|
|
#include "./vpx_dsp_rtcd.h"
|
|
#include "test/acm_random.h"
|
|
#include "test/clear_system_state.h"
|
|
#include "test/register_state_check.h"
|
|
#include "aom/vpx_codec.h"
|
|
#include "aom/vpx_integer.h"
|
|
#include "aom_mem/vpx_mem.h"
|
|
#include "aom_ports/mem.h"
|
|
|
|
namespace {
|
|
|
|
typedef unsigned int (*VarianceMxNFunc)(const uint8_t *a, int a_stride,
|
|
const uint8_t *b, int b_stride,
|
|
unsigned int *sse);
|
|
typedef unsigned int (*SubpixVarMxNFunc)(const uint8_t *a, int a_stride,
|
|
int xoffset, int yoffset,
|
|
const uint8_t *b, int b_stride,
|
|
unsigned int *sse);
|
|
typedef unsigned int (*SubpixAvgVarMxNFunc)(const uint8_t *a, int a_stride,
|
|
int xoffset, int yoffset,
|
|
const uint8_t *b, int b_stride,
|
|
uint32_t *sse,
|
|
const uint8_t *second_pred);
|
|
typedef unsigned int (*Get4x4SseFunc)(const uint8_t *a, int a_stride,
|
|
const uint8_t *b, int b_stride);
|
|
typedef unsigned int (*SumOfSquaresFunction)(const int16_t *src);
|
|
|
|
using ::std::tr1::get;
|
|
using ::std::tr1::make_tuple;
|
|
using ::std::tr1::tuple;
|
|
using libvpx_test::ACMRandom;
|
|
|
|
// Truncate high bit depth results by downshifting (with rounding) by:
|
|
// 2 * (bit_depth - 8) for sse
|
|
// (bit_depth - 8) for se
|
|
static void RoundHighBitDepth(int bit_depth, int64_t *se, uint64_t *sse) {
|
|
switch (bit_depth) {
|
|
case VPX_BITS_12:
|
|
*sse = (*sse + 128) >> 8;
|
|
*se = (*se + 8) >> 4;
|
|
break;
|
|
case VPX_BITS_10:
|
|
*sse = (*sse + 8) >> 4;
|
|
*se = (*se + 2) >> 2;
|
|
break;
|
|
case VPX_BITS_8:
|
|
default: break;
|
|
}
|
|
}
|
|
|
|
static unsigned int mb_ss_ref(const int16_t *src) {
|
|
unsigned int res = 0;
|
|
for (int i = 0; i < 256; ++i) {
|
|
res += src[i] * src[i];
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static uint32_t variance_ref(const uint8_t *src, const uint8_t *ref, int l2w,
|
|
int l2h, int src_stride_coeff,
|
|
int ref_stride_coeff, uint32_t *sse_ptr,
|
|
bool use_high_bit_depth_,
|
|
vpx_bit_depth_t bit_depth) {
|
|
int64_t se = 0;
|
|
uint64_t sse = 0;
|
|
const int w = 1 << l2w;
|
|
const int h = 1 << l2h;
|
|
for (int y = 0; y < h; y++) {
|
|
for (int x = 0; x < w; x++) {
|
|
int diff;
|
|
if (!use_high_bit_depth_) {
|
|
diff = ref[w * y * ref_stride_coeff + x] -
|
|
src[w * y * src_stride_coeff + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
diff = CONVERT_TO_SHORTPTR(ref)[w * y * ref_stride_coeff + x] -
|
|
CONVERT_TO_SHORTPTR(src)[w * y * src_stride_coeff + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
}
|
|
}
|
|
RoundHighBitDepth(bit_depth, &se, &sse);
|
|
*sse_ptr = static_cast<uint32_t>(sse);
|
|
return static_cast<uint32_t>(
|
|
sse - ((static_cast<int64_t>(se) * se) >> (l2w + l2h)));
|
|
}
|
|
|
|
/* The subpel reference functions differ from the codec version in one aspect:
|
|
* they calculate the bilinear factors directly instead of using a lookup table
|
|
* and therefore upshift xoff and yoff by 1. Only every other calculated value
|
|
* is used so the codec version shrinks the table to save space and maintain
|
|
* compatibility with vp8.
|
|
*/
|
|
static uint32_t subpel_variance_ref(const uint8_t *ref, const uint8_t *src,
|
|
int l2w, int l2h, int xoff, int yoff,
|
|
uint32_t *sse_ptr, bool use_high_bit_depth_,
|
|
vpx_bit_depth_t bit_depth) {
|
|
int64_t se = 0;
|
|
uint64_t sse = 0;
|
|
const int w = 1 << l2w;
|
|
const int h = 1 << l2h;
|
|
|
|
xoff <<= 1;
|
|
yoff <<= 1;
|
|
|
|
for (int y = 0; y < h; y++) {
|
|
for (int x = 0; x < w; x++) {
|
|
// Bilinear interpolation at a 16th pel step.
|
|
if (!use_high_bit_depth_) {
|
|
const int a1 = ref[(w + 1) * (y + 0) + x + 0];
|
|
const int a2 = ref[(w + 1) * (y + 0) + x + 1];
|
|
const int b1 = ref[(w + 1) * (y + 1) + x + 0];
|
|
const int b2 = ref[(w + 1) * (y + 1) + x + 1];
|
|
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4);
|
|
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4);
|
|
const int r = a + (((b - a) * yoff + 8) >> 4);
|
|
const int diff = r - src[w * y + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
uint16_t *ref16 = CONVERT_TO_SHORTPTR(ref);
|
|
uint16_t *src16 = CONVERT_TO_SHORTPTR(src);
|
|
const int a1 = ref16[(w + 1) * (y + 0) + x + 0];
|
|
const int a2 = ref16[(w + 1) * (y + 0) + x + 1];
|
|
const int b1 = ref16[(w + 1) * (y + 1) + x + 0];
|
|
const int b2 = ref16[(w + 1) * (y + 1) + x + 1];
|
|
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4);
|
|
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4);
|
|
const int r = a + (((b - a) * yoff + 8) >> 4);
|
|
const int diff = r - src16[w * y + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
}
|
|
}
|
|
RoundHighBitDepth(bit_depth, &se, &sse);
|
|
*sse_ptr = static_cast<uint32_t>(sse);
|
|
return static_cast<uint32_t>(
|
|
sse - ((static_cast<int64_t>(se) * se) >> (l2w + l2h)));
|
|
}
|
|
|
|
class SumOfSquaresTest : public ::testing::TestWithParam<SumOfSquaresFunction> {
|
|
public:
|
|
SumOfSquaresTest() : func_(GetParam()) {}
|
|
|
|
virtual ~SumOfSquaresTest() { libvpx_test::ClearSystemState(); }
|
|
|
|
protected:
|
|
void ConstTest();
|
|
void RefTest();
|
|
|
|
SumOfSquaresFunction func_;
|
|
ACMRandom rnd_;
|
|
};
|
|
|
|
void SumOfSquaresTest::ConstTest() {
|
|
int16_t mem[256];
|
|
unsigned int res;
|
|
for (int v = 0; v < 256; ++v) {
|
|
for (int i = 0; i < 256; ++i) {
|
|
mem[i] = v;
|
|
}
|
|
ASM_REGISTER_STATE_CHECK(res = func_(mem));
|
|
EXPECT_EQ(256u * (v * v), res);
|
|
}
|
|
}
|
|
|
|
void SumOfSquaresTest::RefTest() {
|
|
int16_t mem[256];
|
|
for (int i = 0; i < 100; ++i) {
|
|
for (int j = 0; j < 256; ++j) {
|
|
mem[j] = rnd_.Rand8() - rnd_.Rand8();
|
|
}
|
|
|
|
const unsigned int expected = mb_ss_ref(mem);
|
|
unsigned int res;
|
|
ASM_REGISTER_STATE_CHECK(res = func_(mem));
|
|
EXPECT_EQ(expected, res);
|
|
}
|
|
}
|
|
|
|
template <typename VarianceFunctionType>
|
|
class VarianceTest : public ::testing::TestWithParam<
|
|
tuple<int, int, VarianceFunctionType, int> > {
|
|
public:
|
|
virtual void SetUp() {
|
|
const tuple<int, int, VarianceFunctionType, int> ¶ms = this->GetParam();
|
|
log2width_ = get<0>(params);
|
|
width_ = 1 << log2width_;
|
|
log2height_ = get<1>(params);
|
|
height_ = 1 << log2height_;
|
|
variance_ = get<2>(params);
|
|
if (get<3>(params)) {
|
|
bit_depth_ = static_cast<vpx_bit_depth_t>(get<3>(params));
|
|
use_high_bit_depth_ = true;
|
|
} else {
|
|
bit_depth_ = VPX_BITS_8;
|
|
use_high_bit_depth_ = false;
|
|
}
|
|
mask_ = (1 << bit_depth_) - 1;
|
|
|
|
rnd_.Reset(ACMRandom::DeterministicSeed());
|
|
block_size_ = width_ * height_;
|
|
if (!use_high_bit_depth_) {
|
|
src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_ * 2));
|
|
ref_ = new uint8_t[block_size_ * 2];
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
src_ = CONVERT_TO_BYTEPTR(reinterpret_cast<uint16_t *>(
|
|
vpx_memalign(16, block_size_ * 2 * sizeof(uint16_t))));
|
|
ref_ = CONVERT_TO_BYTEPTR(new uint16_t[block_size_ * 2]);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
ASSERT_TRUE(src_ != NULL);
|
|
ASSERT_TRUE(ref_ != NULL);
|
|
}
|
|
|
|
virtual void TearDown() {
|
|
if (!use_high_bit_depth_) {
|
|
vpx_free(src_);
|
|
delete[] ref_;
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_free(CONVERT_TO_SHORTPTR(src_));
|
|
delete[] CONVERT_TO_SHORTPTR(ref_);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
libvpx_test::ClearSystemState();
|
|
}
|
|
|
|
protected:
|
|
void ZeroTest();
|
|
void RefTest();
|
|
void RefStrideTest();
|
|
void OneQuarterTest();
|
|
|
|
ACMRandom rnd_;
|
|
uint8_t *src_;
|
|
uint8_t *ref_;
|
|
int width_, log2width_;
|
|
int height_, log2height_;
|
|
vpx_bit_depth_t bit_depth_;
|
|
int mask_;
|
|
bool use_high_bit_depth_;
|
|
int block_size_;
|
|
VarianceFunctionType variance_;
|
|
};
|
|
|
|
template <typename VarianceFunctionType>
|
|
void VarianceTest<VarianceFunctionType>::ZeroTest() {
|
|
for (int i = 0; i <= 255; ++i) {
|
|
if (!use_high_bit_depth_) {
|
|
memset(src_, i, block_size_);
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(src_), i << (bit_depth_ - 8),
|
|
block_size_);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
for (int j = 0; j <= 255; ++j) {
|
|
if (!use_high_bit_depth_) {
|
|
memset(ref_, j, block_size_);
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(ref_), j << (bit_depth_ - 8),
|
|
block_size_);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
unsigned int sse;
|
|
unsigned int var;
|
|
ASM_REGISTER_STATE_CHECK(var =
|
|
variance_(src_, width_, ref_, width_, &sse));
|
|
EXPECT_EQ(0u, var) << "src values: " << i << " ref values: " << j;
|
|
}
|
|
}
|
|
}
|
|
|
|
template <typename VarianceFunctionType>
|
|
void VarianceTest<VarianceFunctionType>::RefTest() {
|
|
for (int i = 0; i < 10; ++i) {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
if (!use_high_bit_depth_) {
|
|
src_[j] = rnd_.Rand8();
|
|
ref_[j] = rnd_.Rand8();
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() && mask_;
|
|
CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() && mask_;
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
}
|
|
unsigned int sse1, sse2;
|
|
unsigned int var1;
|
|
const int stride_coeff = 1;
|
|
ASM_REGISTER_STATE_CHECK(var1 =
|
|
variance_(src_, width_, ref_, width_, &sse1));
|
|
const unsigned int var2 =
|
|
variance_ref(src_, ref_, log2width_, log2height_, stride_coeff,
|
|
stride_coeff, &sse2, use_high_bit_depth_, bit_depth_);
|
|
EXPECT_EQ(sse1, sse2);
|
|
EXPECT_EQ(var1, var2);
|
|
}
|
|
}
|
|
|
|
template <typename VarianceFunctionType>
|
|
void VarianceTest<VarianceFunctionType>::RefStrideTest() {
|
|
for (int i = 0; i < 10; ++i) {
|
|
int ref_stride_coeff = i % 2;
|
|
int src_stride_coeff = (i >> 1) % 2;
|
|
for (int j = 0; j < block_size_; j++) {
|
|
int ref_ind = (j / width_) * ref_stride_coeff * width_ + j % width_;
|
|
int src_ind = (j / width_) * src_stride_coeff * width_ + j % width_;
|
|
if (!use_high_bit_depth_) {
|
|
src_[src_ind] = rnd_.Rand8();
|
|
ref_[ref_ind] = rnd_.Rand8();
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
CONVERT_TO_SHORTPTR(src_)[src_ind] = rnd_.Rand16() && mask_;
|
|
CONVERT_TO_SHORTPTR(ref_)[ref_ind] = rnd_.Rand16() && mask_;
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
}
|
|
unsigned int sse1, sse2;
|
|
unsigned int var1;
|
|
|
|
ASM_REGISTER_STATE_CHECK(var1 = variance_(src_, width_ * src_stride_coeff,
|
|
ref_, width_ * ref_stride_coeff,
|
|
&sse1));
|
|
const unsigned int var2 =
|
|
variance_ref(src_, ref_, log2width_, log2height_, src_stride_coeff,
|
|
ref_stride_coeff, &sse2, use_high_bit_depth_, bit_depth_);
|
|
EXPECT_EQ(sse1, sse2);
|
|
EXPECT_EQ(var1, var2);
|
|
}
|
|
}
|
|
|
|
template <typename VarianceFunctionType>
|
|
void VarianceTest<VarianceFunctionType>::OneQuarterTest() {
|
|
const int half = block_size_ / 2;
|
|
if (!use_high_bit_depth_) {
|
|
memset(src_, 255, block_size_);
|
|
memset(ref_, 255, half);
|
|
memset(ref_ + half, 0, half);
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(src_), 255 << (bit_depth_ - 8),
|
|
block_size_);
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(ref_), 255 << (bit_depth_ - 8), half);
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(ref_) + half, 0, half);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
unsigned int sse;
|
|
unsigned int var;
|
|
ASM_REGISTER_STATE_CHECK(var = variance_(src_, width_, ref_, width_, &sse));
|
|
const unsigned int expected = block_size_ * 255 * 255 / 4;
|
|
EXPECT_EQ(expected, var);
|
|
}
|
|
|
|
template <typename MseFunctionType>
|
|
class MseTest
|
|
: public ::testing::TestWithParam<tuple<int, int, MseFunctionType> > {
|
|
public:
|
|
virtual void SetUp() {
|
|
const tuple<int, int, MseFunctionType> ¶ms = this->GetParam();
|
|
log2width_ = get<0>(params);
|
|
width_ = 1 << log2width_;
|
|
log2height_ = get<1>(params);
|
|
height_ = 1 << log2height_;
|
|
mse_ = get<2>(params);
|
|
|
|
rnd(ACMRandom::DeterministicSeed());
|
|
block_size_ = width_ * height_;
|
|
src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_));
|
|
ref_ = new uint8_t[block_size_];
|
|
ASSERT_TRUE(src_ != NULL);
|
|
ASSERT_TRUE(ref_ != NULL);
|
|
}
|
|
|
|
virtual void TearDown() {
|
|
vpx_free(src_);
|
|
delete[] ref_;
|
|
libvpx_test::ClearSystemState();
|
|
}
|
|
|
|
protected:
|
|
void RefTest_mse();
|
|
void RefTest_sse();
|
|
void MaxTest_mse();
|
|
void MaxTest_sse();
|
|
|
|
ACMRandom rnd;
|
|
uint8_t *src_;
|
|
uint8_t *ref_;
|
|
int width_, log2width_;
|
|
int height_, log2height_;
|
|
int block_size_;
|
|
MseFunctionType mse_;
|
|
};
|
|
|
|
template <typename MseFunctionType>
|
|
void MseTest<MseFunctionType>::RefTest_mse() {
|
|
for (int i = 0; i < 10; ++i) {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
src_[j] = rnd.Rand8();
|
|
ref_[j] = rnd.Rand8();
|
|
}
|
|
unsigned int sse1, sse2;
|
|
const int stride_coeff = 1;
|
|
ASM_REGISTER_STATE_CHECK(mse_(src_, width_, ref_, width_, &sse1));
|
|
variance_ref(src_, ref_, log2width_, log2height_, stride_coeff,
|
|
stride_coeff, &sse2, false, VPX_BITS_8);
|
|
EXPECT_EQ(sse1, sse2);
|
|
}
|
|
}
|
|
|
|
template <typename MseFunctionType>
|
|
void MseTest<MseFunctionType>::RefTest_sse() {
|
|
for (int i = 0; i < 10; ++i) {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
src_[j] = rnd.Rand8();
|
|
ref_[j] = rnd.Rand8();
|
|
}
|
|
unsigned int sse2;
|
|
unsigned int var1;
|
|
const int stride_coeff = 1;
|
|
ASM_REGISTER_STATE_CHECK(var1 = mse_(src_, width_, ref_, width_));
|
|
variance_ref(src_, ref_, log2width_, log2height_, stride_coeff,
|
|
stride_coeff, &sse2, false, VPX_BITS_8);
|
|
EXPECT_EQ(var1, sse2);
|
|
}
|
|
}
|
|
|
|
template <typename MseFunctionType>
|
|
void MseTest<MseFunctionType>::MaxTest_mse() {
|
|
memset(src_, 255, block_size_);
|
|
memset(ref_, 0, block_size_);
|
|
unsigned int sse;
|
|
ASM_REGISTER_STATE_CHECK(mse_(src_, width_, ref_, width_, &sse));
|
|
const unsigned int expected = block_size_ * 255 * 255;
|
|
EXPECT_EQ(expected, sse);
|
|
}
|
|
|
|
template <typename MseFunctionType>
|
|
void MseTest<MseFunctionType>::MaxTest_sse() {
|
|
memset(src_, 255, block_size_);
|
|
memset(ref_, 0, block_size_);
|
|
unsigned int var;
|
|
ASM_REGISTER_STATE_CHECK(var = mse_(src_, width_, ref_, width_));
|
|
const unsigned int expected = block_size_ * 255 * 255;
|
|
EXPECT_EQ(expected, var);
|
|
}
|
|
|
|
static uint32_t subpel_avg_variance_ref(const uint8_t *ref, const uint8_t *src,
|
|
const uint8_t *second_pred, int l2w,
|
|
int l2h, int xoff, int yoff,
|
|
uint32_t *sse_ptr,
|
|
bool use_high_bit_depth,
|
|
vpx_bit_depth_t bit_depth) {
|
|
int64_t se = 0;
|
|
uint64_t sse = 0;
|
|
const int w = 1 << l2w;
|
|
const int h = 1 << l2h;
|
|
|
|
xoff <<= 1;
|
|
yoff <<= 1;
|
|
|
|
for (int y = 0; y < h; y++) {
|
|
for (int x = 0; x < w; x++) {
|
|
// bilinear interpolation at a 16th pel step
|
|
if (!use_high_bit_depth) {
|
|
const int a1 = ref[(w + 1) * (y + 0) + x + 0];
|
|
const int a2 = ref[(w + 1) * (y + 0) + x + 1];
|
|
const int b1 = ref[(w + 1) * (y + 1) + x + 0];
|
|
const int b2 = ref[(w + 1) * (y + 1) + x + 1];
|
|
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4);
|
|
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4);
|
|
const int r = a + (((b - a) * yoff + 8) >> 4);
|
|
const int diff =
|
|
((r + second_pred[w * y + x] + 1) >> 1) - src[w * y + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
uint16_t *ref16 = CONVERT_TO_SHORTPTR(ref);
|
|
uint16_t *src16 = CONVERT_TO_SHORTPTR(src);
|
|
uint16_t *sec16 = CONVERT_TO_SHORTPTR(second_pred);
|
|
const int a1 = ref16[(w + 1) * (y + 0) + x + 0];
|
|
const int a2 = ref16[(w + 1) * (y + 0) + x + 1];
|
|
const int b1 = ref16[(w + 1) * (y + 1) + x + 0];
|
|
const int b2 = ref16[(w + 1) * (y + 1) + x + 1];
|
|
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4);
|
|
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4);
|
|
const int r = a + (((b - a) * yoff + 8) >> 4);
|
|
const int diff = ((r + sec16[w * y + x] + 1) >> 1) - src16[w * y + x];
|
|
se += diff;
|
|
sse += diff * diff;
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
}
|
|
}
|
|
RoundHighBitDepth(bit_depth, &se, &sse);
|
|
*sse_ptr = static_cast<uint32_t>(sse);
|
|
return static_cast<uint32_t>(
|
|
sse - ((static_cast<int64_t>(se) * se) >> (l2w + l2h)));
|
|
}
|
|
|
|
template <typename SubpelVarianceFunctionType>
|
|
class SubpelVarianceTest
|
|
: public ::testing::TestWithParam<
|
|
tuple<int, int, SubpelVarianceFunctionType, int> > {
|
|
public:
|
|
virtual void SetUp() {
|
|
const tuple<int, int, SubpelVarianceFunctionType, int> ¶ms =
|
|
this->GetParam();
|
|
log2width_ = get<0>(params);
|
|
width_ = 1 << log2width_;
|
|
log2height_ = get<1>(params);
|
|
height_ = 1 << log2height_;
|
|
subpel_variance_ = get<2>(params);
|
|
if (get<3>(params)) {
|
|
bit_depth_ = (vpx_bit_depth_t)get<3>(params);
|
|
use_high_bit_depth_ = true;
|
|
} else {
|
|
bit_depth_ = VPX_BITS_8;
|
|
use_high_bit_depth_ = false;
|
|
}
|
|
mask_ = (1 << bit_depth_) - 1;
|
|
|
|
rnd_.Reset(ACMRandom::DeterministicSeed());
|
|
block_size_ = width_ * height_;
|
|
if (!use_high_bit_depth_) {
|
|
src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_));
|
|
sec_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_));
|
|
ref_ = new uint8_t[block_size_ + width_ + height_ + 1];
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
src_ = CONVERT_TO_BYTEPTR(reinterpret_cast<uint16_t *>(
|
|
vpx_memalign(16, block_size_ * sizeof(uint16_t))));
|
|
sec_ = CONVERT_TO_BYTEPTR(reinterpret_cast<uint16_t *>(
|
|
vpx_memalign(16, block_size_ * sizeof(uint16_t))));
|
|
ref_ =
|
|
CONVERT_TO_BYTEPTR(new uint16_t[block_size_ + width_ + height_ + 1]);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
ASSERT_TRUE(src_ != NULL);
|
|
ASSERT_TRUE(sec_ != NULL);
|
|
ASSERT_TRUE(ref_ != NULL);
|
|
}
|
|
|
|
virtual void TearDown() {
|
|
if (!use_high_bit_depth_) {
|
|
vpx_free(src_);
|
|
delete[] ref_;
|
|
vpx_free(sec_);
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_free(CONVERT_TO_SHORTPTR(src_));
|
|
delete[] CONVERT_TO_SHORTPTR(ref_);
|
|
vpx_free(CONVERT_TO_SHORTPTR(sec_));
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
libvpx_test::ClearSystemState();
|
|
}
|
|
|
|
protected:
|
|
void RefTest();
|
|
void ExtremeRefTest();
|
|
|
|
ACMRandom rnd_;
|
|
uint8_t *src_;
|
|
uint8_t *ref_;
|
|
uint8_t *sec_;
|
|
bool use_high_bit_depth_;
|
|
vpx_bit_depth_t bit_depth_;
|
|
int width_, log2width_;
|
|
int height_, log2height_;
|
|
int block_size_, mask_;
|
|
SubpelVarianceFunctionType subpel_variance_;
|
|
};
|
|
|
|
template <typename SubpelVarianceFunctionType>
|
|
void SubpelVarianceTest<SubpelVarianceFunctionType>::RefTest() {
|
|
for (int x = 0; x < 8; ++x) {
|
|
for (int y = 0; y < 8; ++y) {
|
|
if (!use_high_bit_depth_) {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
src_[j] = rnd_.Rand8();
|
|
}
|
|
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) {
|
|
ref_[j] = rnd_.Rand8();
|
|
}
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() & mask_;
|
|
}
|
|
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) {
|
|
CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() & mask_;
|
|
}
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
unsigned int sse1, sse2;
|
|
unsigned int var1;
|
|
ASM_REGISTER_STATE_CHECK(
|
|
var1 = subpel_variance_(ref_, width_ + 1, x, y, src_, width_, &sse1));
|
|
const unsigned int var2 =
|
|
subpel_variance_ref(ref_, src_, log2width_, log2height_, x, y, &sse2,
|
|
use_high_bit_depth_, bit_depth_);
|
|
EXPECT_EQ(sse1, sse2) << "at position " << x << ", " << y;
|
|
EXPECT_EQ(var1, var2) << "at position " << x << ", " << y;
|
|
}
|
|
}
|
|
}
|
|
|
|
template <typename SubpelVarianceFunctionType>
|
|
void SubpelVarianceTest<SubpelVarianceFunctionType>::ExtremeRefTest() {
|
|
// Compare against reference.
|
|
// Src: Set the first half of values to 0, the second half to the maximum.
|
|
// Ref: Set the first half of values to the maximum, the second half to 0.
|
|
for (int x = 0; x < 8; ++x) {
|
|
for (int y = 0; y < 8; ++y) {
|
|
const int half = block_size_ / 2;
|
|
if (!use_high_bit_depth_) {
|
|
memset(src_, 0, half);
|
|
memset(src_ + half, 255, half);
|
|
memset(ref_, 255, half);
|
|
memset(ref_ + half, 0, half + width_ + height_ + 1);
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(src_), mask_, half);
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(src_) + half, 0, half);
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(ref_), 0, half);
|
|
vpx_memset16(CONVERT_TO_SHORTPTR(ref_) + half, mask_,
|
|
half + width_ + height_ + 1);
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
unsigned int sse1, sse2;
|
|
unsigned int var1;
|
|
ASM_REGISTER_STATE_CHECK(
|
|
var1 = subpel_variance_(ref_, width_ + 1, x, y, src_, width_, &sse1));
|
|
const unsigned int var2 =
|
|
subpel_variance_ref(ref_, src_, log2width_, log2height_, x, y, &sse2,
|
|
use_high_bit_depth_, bit_depth_);
|
|
EXPECT_EQ(sse1, sse2) << "for xoffset " << x << " and yoffset " << y;
|
|
EXPECT_EQ(var1, var2) << "for xoffset " << x << " and yoffset " << y;
|
|
}
|
|
}
|
|
}
|
|
|
|
template <>
|
|
void SubpelVarianceTest<SubpixAvgVarMxNFunc>::RefTest() {
|
|
for (int x = 0; x < 8; ++x) {
|
|
for (int y = 0; y < 8; ++y) {
|
|
if (!use_high_bit_depth_) {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
src_[j] = rnd_.Rand8();
|
|
sec_[j] = rnd_.Rand8();
|
|
}
|
|
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) {
|
|
ref_[j] = rnd_.Rand8();
|
|
}
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
} else {
|
|
for (int j = 0; j < block_size_; j++) {
|
|
CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() & mask_;
|
|
CONVERT_TO_SHORTPTR(sec_)[j] = rnd_.Rand16() & mask_;
|
|
}
|
|
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) {
|
|
CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() & mask_;
|
|
}
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
}
|
|
unsigned int sse1, sse2;
|
|
unsigned int var1;
|
|
ASM_REGISTER_STATE_CHECK(var1 =
|
|
subpel_variance_(ref_, width_ + 1, x, y,
|
|
src_, width_, &sse1, sec_));
|
|
const unsigned int var2 =
|
|
subpel_avg_variance_ref(ref_, src_, sec_, log2width_, log2height_, x,
|
|
y, &sse2, use_high_bit_depth_, bit_depth_);
|
|
EXPECT_EQ(sse1, sse2) << "at position " << x << ", " << y;
|
|
EXPECT_EQ(var1, var2) << "at position " << x << ", " << y;
|
|
}
|
|
}
|
|
}
|
|
|
|
typedef MseTest<Get4x4SseFunc> VpxSseTest;
|
|
typedef MseTest<VarianceMxNFunc> VpxMseTest;
|
|
typedef VarianceTest<VarianceMxNFunc> VpxVarianceTest;
|
|
typedef SubpelVarianceTest<SubpixVarMxNFunc> VpxSubpelVarianceTest;
|
|
typedef SubpelVarianceTest<SubpixAvgVarMxNFunc> VpxSubpelAvgVarianceTest;
|
|
|
|
TEST_P(VpxSseTest, Ref_sse) { RefTest_sse(); }
|
|
TEST_P(VpxSseTest, Max_sse) { MaxTest_sse(); }
|
|
TEST_P(VpxMseTest, Ref_mse) { RefTest_mse(); }
|
|
TEST_P(VpxMseTest, Max_mse) { MaxTest_mse(); }
|
|
TEST_P(VpxVarianceTest, Zero) { ZeroTest(); }
|
|
TEST_P(VpxVarianceTest, Ref) { RefTest(); }
|
|
TEST_P(VpxVarianceTest, RefStride) { RefStrideTest(); }
|
|
TEST_P(VpxVarianceTest, OneQuarter) { OneQuarterTest(); }
|
|
TEST_P(SumOfSquaresTest, Const) { ConstTest(); }
|
|
TEST_P(SumOfSquaresTest, Ref) { RefTest(); }
|
|
TEST_P(VpxSubpelVarianceTest, Ref) { RefTest(); }
|
|
TEST_P(VpxSubpelVarianceTest, ExtremeRef) { ExtremeRefTest(); }
|
|
TEST_P(VpxSubpelAvgVarianceTest, Ref) { RefTest(); }
|
|
|
|
INSTANTIATE_TEST_CASE_P(C, SumOfSquaresTest,
|
|
::testing::Values(vpx_get_mb_ss_c));
|
|
|
|
INSTANTIATE_TEST_CASE_P(C, VpxSseTest,
|
|
::testing::Values(make_tuple(2, 2,
|
|
&vpx_get4x4sse_cs_c)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(C, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_mse16x16_c),
|
|
make_tuple(4, 3, &vpx_mse16x8_c),
|
|
make_tuple(3, 4, &vpx_mse8x16_c),
|
|
make_tuple(3, 3, &vpx_mse8x8_c)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_variance64x64_c, 0),
|
|
make_tuple(6, 5, &vpx_variance64x32_c, 0),
|
|
make_tuple(5, 6, &vpx_variance32x64_c, 0),
|
|
make_tuple(5, 5, &vpx_variance32x32_c, 0),
|
|
make_tuple(5, 4, &vpx_variance32x16_c, 0),
|
|
make_tuple(4, 5, &vpx_variance16x32_c, 0),
|
|
make_tuple(4, 4, &vpx_variance16x16_c, 0),
|
|
make_tuple(4, 3, &vpx_variance16x8_c, 0),
|
|
make_tuple(3, 4, &vpx_variance8x16_c, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_c, 0),
|
|
make_tuple(3, 2, &vpx_variance8x4_c, 0),
|
|
make_tuple(2, 3, &vpx_variance4x8_c, 0),
|
|
make_tuple(2, 2, &vpx_variance4x4_c, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_variance64x64_c, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_variance64x32_c, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_variance32x64_c, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_c, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_variance32x16_c, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_variance16x32_c, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_variance16x16_c, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_variance16x8_c, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_variance8x16_c, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_c, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_variance8x4_c, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_variance4x8_c, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_variance4x4_c, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxSubpelAvgVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_avg_variance64x64_c, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_avg_variance64x32_c, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_avg_variance32x64_c, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_avg_variance32x32_c, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_avg_variance32x16_c, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_avg_variance16x32_c, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_avg_variance16x16_c, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_avg_variance16x8_c, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_avg_variance8x16_c, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_avg_variance8x8_c, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_avg_variance8x4_c, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_avg_variance4x8_c, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_avg_variance4x4_c, 0)));
|
|
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
typedef MseTest<VarianceMxNFunc> VpxHBDMseTest;
|
|
typedef VarianceTest<VarianceMxNFunc> VpxHBDVarianceTest;
|
|
typedef SubpelVarianceTest<SubpixVarMxNFunc> VpxHBDSubpelVarianceTest;
|
|
typedef SubpelVarianceTest<SubpixAvgVarMxNFunc> VpxHBDSubpelAvgVarianceTest;
|
|
|
|
TEST_P(VpxHBDMseTest, Ref_mse) { RefTest_mse(); }
|
|
TEST_P(VpxHBDMseTest, Max_mse) { MaxTest_mse(); }
|
|
TEST_P(VpxHBDVarianceTest, Zero) { ZeroTest(); }
|
|
TEST_P(VpxHBDVarianceTest, Ref) { RefTest(); }
|
|
TEST_P(VpxHBDVarianceTest, RefStride) { RefStrideTest(); }
|
|
TEST_P(VpxHBDVarianceTest, OneQuarter) { OneQuarterTest(); }
|
|
TEST_P(VpxHBDSubpelVarianceTest, Ref) { RefTest(); }
|
|
TEST_P(VpxHBDSubpelVarianceTest, ExtremeRef) { ExtremeRefTest(); }
|
|
TEST_P(VpxHBDSubpelAvgVarianceTest, Ref) { RefTest(); }
|
|
|
|
/* TODO(debargha): This test does not support the highbd version
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxHBDMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_highbd_12_mse16x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_12_mse16x8_c),
|
|
make_tuple(4, 4, &vpx_highbd_12_mse8x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_12_mse8x8_c),
|
|
make_tuple(4, 4, &vpx_highbd_10_mse16x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_10_mse16x8_c),
|
|
make_tuple(4, 4, &vpx_highbd_10_mse8x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_10_mse8x8_c),
|
|
make_tuple(4, 4, &vpx_highbd_8_mse16x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_8_mse16x8_c),
|
|
make_tuple(4, 4, &vpx_highbd_8_mse8x16_c),
|
|
make_tuple(4, 4, &vpx_highbd_8_mse8x8_c)));
|
|
*/
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxHBDVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_highbd_12_variance64x64_c, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_variance64x32_c, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_variance32x64_c, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_variance32x32_c, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_variance32x16_c, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_variance16x32_c, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_variance16x16_c, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_variance16x8_c, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_variance8x16_c, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_variance8x8_c, 12),
|
|
make_tuple(3, 2, &vpx_highbd_12_variance8x4_c, 12),
|
|
make_tuple(2, 3, &vpx_highbd_12_variance4x8_c, 12),
|
|
make_tuple(2, 2, &vpx_highbd_12_variance4x4_c, 12),
|
|
make_tuple(6, 6, &vpx_highbd_10_variance64x64_c, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_variance64x32_c, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_variance32x64_c, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_variance32x32_c, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_variance32x16_c, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_variance16x32_c, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_variance16x16_c, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_variance16x8_c, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_variance8x16_c, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_variance8x8_c, 10),
|
|
make_tuple(3, 2, &vpx_highbd_10_variance8x4_c, 10),
|
|
make_tuple(2, 3, &vpx_highbd_10_variance4x8_c, 10),
|
|
make_tuple(2, 2, &vpx_highbd_10_variance4x4_c, 10),
|
|
make_tuple(6, 6, &vpx_highbd_8_variance64x64_c, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_variance64x32_c, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_variance32x64_c, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_variance32x32_c, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_variance32x16_c, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_variance16x32_c, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_variance16x16_c, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_variance16x8_c, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_variance8x16_c, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_variance8x8_c, 8),
|
|
make_tuple(3, 2, &vpx_highbd_8_variance8x4_c, 8),
|
|
make_tuple(2, 3, &vpx_highbd_8_variance4x8_c, 8),
|
|
make_tuple(2, 2, &vpx_highbd_8_variance4x4_c, 8)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxHBDSubpelVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_highbd_8_sub_pixel_variance64x64_c, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_sub_pixel_variance64x32_c, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_sub_pixel_variance32x64_c, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_sub_pixel_variance32x32_c, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_sub_pixel_variance32x16_c, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_sub_pixel_variance16x32_c, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_sub_pixel_variance16x16_c, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_sub_pixel_variance16x8_c, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_sub_pixel_variance8x16_c, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_sub_pixel_variance8x8_c, 8),
|
|
make_tuple(3, 2, &vpx_highbd_8_sub_pixel_variance8x4_c, 8),
|
|
make_tuple(2, 3, &vpx_highbd_8_sub_pixel_variance4x8_c, 8),
|
|
make_tuple(2, 2, &vpx_highbd_8_sub_pixel_variance4x4_c, 8),
|
|
make_tuple(6, 6, &vpx_highbd_10_sub_pixel_variance64x64_c, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_sub_pixel_variance64x32_c, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_sub_pixel_variance32x64_c, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_sub_pixel_variance32x32_c, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_sub_pixel_variance32x16_c, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_sub_pixel_variance16x32_c, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_sub_pixel_variance16x16_c, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_sub_pixel_variance16x8_c, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_sub_pixel_variance8x16_c, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_sub_pixel_variance8x8_c, 10),
|
|
make_tuple(3, 2, &vpx_highbd_10_sub_pixel_variance8x4_c, 10),
|
|
make_tuple(2, 3, &vpx_highbd_10_sub_pixel_variance4x8_c, 10),
|
|
make_tuple(2, 2, &vpx_highbd_10_sub_pixel_variance4x4_c, 10),
|
|
make_tuple(6, 6, &vpx_highbd_12_sub_pixel_variance64x64_c, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_sub_pixel_variance64x32_c, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_sub_pixel_variance32x64_c, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_sub_pixel_variance32x32_c, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_sub_pixel_variance32x16_c, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_sub_pixel_variance16x32_c, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_sub_pixel_variance16x16_c, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_sub_pixel_variance16x8_c, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_sub_pixel_variance8x16_c, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_sub_pixel_variance8x8_c, 12),
|
|
make_tuple(3, 2, &vpx_highbd_12_sub_pixel_variance8x4_c, 12),
|
|
make_tuple(2, 3, &vpx_highbd_12_sub_pixel_variance4x8_c, 12),
|
|
make_tuple(2, 2, &vpx_highbd_12_sub_pixel_variance4x4_c, 12)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
C, VpxHBDSubpelAvgVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_highbd_8_sub_pixel_avg_variance64x64_c, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_sub_pixel_avg_variance64x32_c, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_sub_pixel_avg_variance32x64_c, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_sub_pixel_avg_variance32x32_c, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_sub_pixel_avg_variance32x16_c, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_sub_pixel_avg_variance16x32_c, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_sub_pixel_avg_variance16x16_c, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_sub_pixel_avg_variance16x8_c, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_sub_pixel_avg_variance8x16_c, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_sub_pixel_avg_variance8x8_c, 8),
|
|
make_tuple(3, 2, &vpx_highbd_8_sub_pixel_avg_variance8x4_c, 8),
|
|
make_tuple(2, 3, &vpx_highbd_8_sub_pixel_avg_variance4x8_c, 8),
|
|
make_tuple(2, 2, &vpx_highbd_8_sub_pixel_avg_variance4x4_c, 8),
|
|
make_tuple(6, 6, &vpx_highbd_10_sub_pixel_avg_variance64x64_c, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_sub_pixel_avg_variance64x32_c, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_sub_pixel_avg_variance32x64_c, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_sub_pixel_avg_variance32x32_c, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_sub_pixel_avg_variance32x16_c, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_sub_pixel_avg_variance16x32_c, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_sub_pixel_avg_variance16x16_c, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_sub_pixel_avg_variance16x8_c, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_sub_pixel_avg_variance8x16_c, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_sub_pixel_avg_variance8x8_c, 10),
|
|
make_tuple(3, 2, &vpx_highbd_10_sub_pixel_avg_variance8x4_c, 10),
|
|
make_tuple(2, 3, &vpx_highbd_10_sub_pixel_avg_variance4x8_c, 10),
|
|
make_tuple(2, 2, &vpx_highbd_10_sub_pixel_avg_variance4x4_c, 10),
|
|
make_tuple(6, 6, &vpx_highbd_12_sub_pixel_avg_variance64x64_c, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_sub_pixel_avg_variance64x32_c, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_sub_pixel_avg_variance32x64_c, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_sub_pixel_avg_variance32x32_c, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_sub_pixel_avg_variance32x16_c, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_sub_pixel_avg_variance16x32_c, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_sub_pixel_avg_variance16x16_c, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_sub_pixel_avg_variance16x8_c, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_sub_pixel_avg_variance8x16_c, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_sub_pixel_avg_variance8x8_c, 12),
|
|
make_tuple(3, 2, &vpx_highbd_12_sub_pixel_avg_variance8x4_c, 12),
|
|
make_tuple(2, 3, &vpx_highbd_12_sub_pixel_avg_variance4x8_c, 12),
|
|
make_tuple(2, 2, &vpx_highbd_12_sub_pixel_avg_variance4x4_c, 12)));
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
|
|
#if HAVE_MMX
|
|
INSTANTIATE_TEST_CASE_P(MMX, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_mse16x16_mmx)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(MMX, SumOfSquaresTest,
|
|
::testing::Values(vpx_get_mb_ss_mmx));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MMX, VpxVarianceTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_variance16x16_mmx, 0),
|
|
make_tuple(4, 3, &vpx_variance16x8_mmx, 0),
|
|
make_tuple(3, 4, &vpx_variance8x16_mmx, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_mmx, 0),
|
|
make_tuple(2, 2, &vpx_variance4x4_mmx, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MMX, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_sub_pixel_variance16x16_mmx, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_variance16x8_mmx, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_variance8x16_mmx, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_mmx, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_variance4x4_mmx, 0)));
|
|
#endif // HAVE_MMX
|
|
|
|
#if HAVE_SSE2
|
|
INSTANTIATE_TEST_CASE_P(SSE2, SumOfSquaresTest,
|
|
::testing::Values(vpx_get_mb_ss_sse2));
|
|
|
|
INSTANTIATE_TEST_CASE_P(SSE2, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_mse16x16_sse2),
|
|
make_tuple(4, 3, &vpx_mse16x8_sse2),
|
|
make_tuple(3, 4, &vpx_mse8x16_sse2),
|
|
make_tuple(3, 3, &vpx_mse8x8_sse2)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_variance64x64_sse2, 0),
|
|
make_tuple(6, 5, &vpx_variance64x32_sse2, 0),
|
|
make_tuple(5, 6, &vpx_variance32x64_sse2, 0),
|
|
make_tuple(5, 5, &vpx_variance32x32_sse2, 0),
|
|
make_tuple(5, 4, &vpx_variance32x16_sse2, 0),
|
|
make_tuple(4, 5, &vpx_variance16x32_sse2, 0),
|
|
make_tuple(4, 4, &vpx_variance16x16_sse2, 0),
|
|
make_tuple(4, 3, &vpx_variance16x8_sse2, 0),
|
|
make_tuple(3, 4, &vpx_variance8x16_sse2, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_sse2, 0),
|
|
make_tuple(3, 2, &vpx_variance8x4_sse2, 0),
|
|
make_tuple(2, 3, &vpx_variance4x8_sse2, 0),
|
|
make_tuple(2, 2, &vpx_variance4x4_sse2, 0)));
|
|
|
|
#if CONFIG_USE_X86INC
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_variance64x64_sse2, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_variance64x32_sse2, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_variance32x64_sse2, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_sse2, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_variance32x16_sse2, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_variance16x32_sse2, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_variance16x16_sse2, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_variance16x8_sse2, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_variance8x16_sse2, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_sse2, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_variance8x4_sse2, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_variance4x8_sse, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_variance4x4_sse, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxSubpelAvgVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_sub_pixel_avg_variance64x64_sse2, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_avg_variance64x32_sse2, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_avg_variance32x64_sse2, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_avg_variance32x32_sse2, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_avg_variance32x16_sse2, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_avg_variance16x32_sse2, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_avg_variance16x16_sse2, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_avg_variance16x8_sse2, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_avg_variance8x16_sse2, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_avg_variance8x8_sse2, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_avg_variance8x4_sse2, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_avg_variance4x8_sse, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_avg_variance4x4_sse, 0)));
|
|
#endif // CONFIG_USE_X86INC
|
|
|
|
#if CONFIG_VPX_HIGHBITDEPTH
|
|
/* TODO(debargha): This test does not support the highbd version
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxHBDMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_highbd_12_mse16x16_sse2),
|
|
make_tuple(4, 3, &vpx_highbd_12_mse16x8_sse2),
|
|
make_tuple(3, 4, &vpx_highbd_12_mse8x16_sse2),
|
|
make_tuple(3, 3, &vpx_highbd_12_mse8x8_sse2),
|
|
make_tuple(4, 4, &vpx_highbd_10_mse16x16_sse2),
|
|
make_tuple(4, 3, &vpx_highbd_10_mse16x8_sse2),
|
|
make_tuple(3, 4, &vpx_highbd_10_mse8x16_sse2),
|
|
make_tuple(3, 3, &vpx_highbd_10_mse8x8_sse2),
|
|
make_tuple(4, 4, &vpx_highbd_8_mse16x16_sse2),
|
|
make_tuple(4, 3, &vpx_highbd_8_mse16x8_sse2),
|
|
make_tuple(3, 4, &vpx_highbd_8_mse8x16_sse2),
|
|
make_tuple(3, 3, &vpx_highbd_8_mse8x8_sse2)));
|
|
*/
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxHBDVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_highbd_12_variance64x64_sse2, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_variance64x32_sse2, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_variance32x64_sse2, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_variance32x32_sse2, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_variance32x16_sse2, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_variance16x32_sse2, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_variance16x16_sse2, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_variance16x8_sse2, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_variance8x16_sse2, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_variance8x8_sse2, 12),
|
|
make_tuple(6, 6, &vpx_highbd_10_variance64x64_sse2, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_variance64x32_sse2, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_variance32x64_sse2, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_variance32x32_sse2, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_variance32x16_sse2, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_variance16x32_sse2, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_variance16x16_sse2, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_variance16x8_sse2, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_variance8x16_sse2, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_variance8x8_sse2, 10),
|
|
make_tuple(6, 6, &vpx_highbd_8_variance64x64_sse2, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_variance64x32_sse2, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_variance32x64_sse2, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_variance32x32_sse2, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_variance32x16_sse2, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_variance16x32_sse2, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_variance16x16_sse2, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_variance16x8_sse2, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_variance8x16_sse2, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_variance8x8_sse2, 8)));
|
|
|
|
#if CONFIG_USE_X86INC
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxHBDSubpelVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_highbd_12_sub_pixel_variance64x64_sse2, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_sub_pixel_variance64x32_sse2, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_sub_pixel_variance32x64_sse2, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_sub_pixel_variance32x32_sse2, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_sub_pixel_variance32x16_sse2, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_sub_pixel_variance16x32_sse2, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_sub_pixel_variance16x16_sse2, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_sub_pixel_variance16x8_sse2, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_sub_pixel_variance8x16_sse2, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_sub_pixel_variance8x8_sse2, 12),
|
|
make_tuple(3, 2, &vpx_highbd_12_sub_pixel_variance8x4_sse2, 12),
|
|
make_tuple(6, 6, &vpx_highbd_10_sub_pixel_variance64x64_sse2, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_sub_pixel_variance64x32_sse2, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_sub_pixel_variance32x64_sse2, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_sub_pixel_variance32x32_sse2, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_sub_pixel_variance32x16_sse2, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_sub_pixel_variance16x32_sse2, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_sub_pixel_variance16x16_sse2, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_sub_pixel_variance16x8_sse2, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_sub_pixel_variance8x16_sse2, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_sub_pixel_variance8x8_sse2, 10),
|
|
make_tuple(3, 2, &vpx_highbd_10_sub_pixel_variance8x4_sse2, 10),
|
|
make_tuple(6, 6, &vpx_highbd_8_sub_pixel_variance64x64_sse2, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_sub_pixel_variance64x32_sse2, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_sub_pixel_variance32x64_sse2, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_sub_pixel_variance32x32_sse2, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_sub_pixel_variance32x16_sse2, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_sub_pixel_variance16x32_sse2, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_sub_pixel_variance16x16_sse2, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_sub_pixel_variance16x8_sse2, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_sub_pixel_variance8x16_sse2, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_sub_pixel_variance8x8_sse2, 8),
|
|
make_tuple(3, 2, &vpx_highbd_8_sub_pixel_variance8x4_sse2, 8)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSE2, VpxHBDSubpelAvgVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_highbd_12_sub_pixel_avg_variance64x64_sse2, 12),
|
|
make_tuple(6, 5, &vpx_highbd_12_sub_pixel_avg_variance64x32_sse2, 12),
|
|
make_tuple(5, 6, &vpx_highbd_12_sub_pixel_avg_variance32x64_sse2, 12),
|
|
make_tuple(5, 5, &vpx_highbd_12_sub_pixel_avg_variance32x32_sse2, 12),
|
|
make_tuple(5, 4, &vpx_highbd_12_sub_pixel_avg_variance32x16_sse2, 12),
|
|
make_tuple(4, 5, &vpx_highbd_12_sub_pixel_avg_variance16x32_sse2, 12),
|
|
make_tuple(4, 4, &vpx_highbd_12_sub_pixel_avg_variance16x16_sse2, 12),
|
|
make_tuple(4, 3, &vpx_highbd_12_sub_pixel_avg_variance16x8_sse2, 12),
|
|
make_tuple(3, 4, &vpx_highbd_12_sub_pixel_avg_variance8x16_sse2, 12),
|
|
make_tuple(3, 3, &vpx_highbd_12_sub_pixel_avg_variance8x8_sse2, 12),
|
|
make_tuple(3, 2, &vpx_highbd_12_sub_pixel_avg_variance8x4_sse2, 12),
|
|
make_tuple(6, 6, &vpx_highbd_10_sub_pixel_avg_variance64x64_sse2, 10),
|
|
make_tuple(6, 5, &vpx_highbd_10_sub_pixel_avg_variance64x32_sse2, 10),
|
|
make_tuple(5, 6, &vpx_highbd_10_sub_pixel_avg_variance32x64_sse2, 10),
|
|
make_tuple(5, 5, &vpx_highbd_10_sub_pixel_avg_variance32x32_sse2, 10),
|
|
make_tuple(5, 4, &vpx_highbd_10_sub_pixel_avg_variance32x16_sse2, 10),
|
|
make_tuple(4, 5, &vpx_highbd_10_sub_pixel_avg_variance16x32_sse2, 10),
|
|
make_tuple(4, 4, &vpx_highbd_10_sub_pixel_avg_variance16x16_sse2, 10),
|
|
make_tuple(4, 3, &vpx_highbd_10_sub_pixel_avg_variance16x8_sse2, 10),
|
|
make_tuple(3, 4, &vpx_highbd_10_sub_pixel_avg_variance8x16_sse2, 10),
|
|
make_tuple(3, 3, &vpx_highbd_10_sub_pixel_avg_variance8x8_sse2, 10),
|
|
make_tuple(3, 2, &vpx_highbd_10_sub_pixel_avg_variance8x4_sse2, 10),
|
|
make_tuple(6, 6, &vpx_highbd_8_sub_pixel_avg_variance64x64_sse2, 8),
|
|
make_tuple(6, 5, &vpx_highbd_8_sub_pixel_avg_variance64x32_sse2, 8),
|
|
make_tuple(5, 6, &vpx_highbd_8_sub_pixel_avg_variance32x64_sse2, 8),
|
|
make_tuple(5, 5, &vpx_highbd_8_sub_pixel_avg_variance32x32_sse2, 8),
|
|
make_tuple(5, 4, &vpx_highbd_8_sub_pixel_avg_variance32x16_sse2, 8),
|
|
make_tuple(4, 5, &vpx_highbd_8_sub_pixel_avg_variance16x32_sse2, 8),
|
|
make_tuple(4, 4, &vpx_highbd_8_sub_pixel_avg_variance16x16_sse2, 8),
|
|
make_tuple(4, 3, &vpx_highbd_8_sub_pixel_avg_variance16x8_sse2, 8),
|
|
make_tuple(3, 4, &vpx_highbd_8_sub_pixel_avg_variance8x16_sse2, 8),
|
|
make_tuple(3, 3, &vpx_highbd_8_sub_pixel_avg_variance8x8_sse2, 8),
|
|
make_tuple(3, 2, &vpx_highbd_8_sub_pixel_avg_variance8x4_sse2, 8)));
|
|
#endif // CONFIG_USE_X86INC
|
|
#endif // CONFIG_VPX_HIGHBITDEPTH
|
|
#endif // HAVE_SSE2
|
|
|
|
#if HAVE_SSSE3
|
|
#if CONFIG_USE_X86INC
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSSE3, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_variance64x64_ssse3, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_variance64x32_ssse3, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_variance32x64_ssse3, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_ssse3, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_variance32x16_ssse3, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_variance16x32_ssse3, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_variance16x16_ssse3, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_variance16x8_ssse3, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_variance8x16_ssse3, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_ssse3, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_variance8x4_ssse3, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_variance4x8_ssse3, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_variance4x4_ssse3, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
SSSE3, VpxSubpelAvgVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_sub_pixel_avg_variance64x64_ssse3, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_avg_variance64x32_ssse3, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_avg_variance32x64_ssse3, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_avg_variance32x32_ssse3, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_avg_variance32x16_ssse3, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_avg_variance16x32_ssse3, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_avg_variance16x16_ssse3, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_avg_variance16x8_ssse3, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_avg_variance8x16_ssse3, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_avg_variance8x8_ssse3, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_avg_variance8x4_ssse3, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_avg_variance4x8_ssse3, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_avg_variance4x4_ssse3, 0)));
|
|
#endif // CONFIG_USE_X86INC
|
|
#endif // HAVE_SSSE3
|
|
|
|
#if HAVE_AVX2
|
|
INSTANTIATE_TEST_CASE_P(AVX2, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4,
|
|
&vpx_mse16x16_avx2)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
AVX2, VpxVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_variance64x64_avx2, 0),
|
|
make_tuple(6, 5, &vpx_variance64x32_avx2, 0),
|
|
make_tuple(5, 5, &vpx_variance32x32_avx2, 0),
|
|
make_tuple(5, 4, &vpx_variance32x16_avx2, 0),
|
|
make_tuple(4, 4, &vpx_variance16x16_avx2, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
AVX2, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_variance64x64_avx2, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_avx2, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
AVX2, VpxSubpelAvgVarianceTest,
|
|
::testing::Values(
|
|
make_tuple(6, 6, &vpx_sub_pixel_avg_variance64x64_avx2, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_avg_variance32x32_avx2, 0)));
|
|
#endif // HAVE_AVX2
|
|
|
|
#if HAVE_MEDIA
|
|
INSTANTIATE_TEST_CASE_P(MEDIA, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4,
|
|
&vpx_mse16x16_media)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MEDIA, VpxVarianceTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_variance16x16_media, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_media, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MEDIA, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_sub_pixel_variance16x16_media, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_media, 0)));
|
|
#endif // HAVE_MEDIA
|
|
|
|
#if HAVE_NEON
|
|
INSTANTIATE_TEST_CASE_P(NEON, VpxSseTest,
|
|
::testing::Values(make_tuple(2, 2,
|
|
&vpx_get4x4sse_cs_neon)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(NEON, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4,
|
|
&vpx_mse16x16_neon)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
NEON, VpxVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_variance64x64_neon, 0),
|
|
make_tuple(6, 5, &vpx_variance64x32_neon, 0),
|
|
make_tuple(5, 6, &vpx_variance32x64_neon, 0),
|
|
make_tuple(5, 5, &vpx_variance32x32_neon, 0),
|
|
make_tuple(4, 4, &vpx_variance16x16_neon, 0),
|
|
make_tuple(4, 3, &vpx_variance16x8_neon, 0),
|
|
make_tuple(3, 4, &vpx_variance8x16_neon, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_neon, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
NEON, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_variance64x64_neon, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_neon, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_variance16x16_neon, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_neon, 0)));
|
|
#endif // HAVE_NEON
|
|
|
|
#if HAVE_MSA
|
|
INSTANTIATE_TEST_CASE_P(MSA, SumOfSquaresTest,
|
|
::testing::Values(vpx_get_mb_ss_msa));
|
|
|
|
INSTANTIATE_TEST_CASE_P(MSA, VpxSseTest,
|
|
::testing::Values(make_tuple(2, 2,
|
|
&vpx_get4x4sse_cs_msa)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(MSA, VpxMseTest,
|
|
::testing::Values(make_tuple(4, 4, &vpx_mse16x16_msa),
|
|
make_tuple(4, 3, &vpx_mse16x8_msa),
|
|
make_tuple(3, 4, &vpx_mse8x16_msa),
|
|
make_tuple(3, 3, &vpx_mse8x8_msa)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MSA, VpxVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_variance64x64_msa, 0),
|
|
make_tuple(6, 5, &vpx_variance64x32_msa, 0),
|
|
make_tuple(5, 6, &vpx_variance32x64_msa, 0),
|
|
make_tuple(5, 5, &vpx_variance32x32_msa, 0),
|
|
make_tuple(5, 4, &vpx_variance32x16_msa, 0),
|
|
make_tuple(4, 5, &vpx_variance16x32_msa, 0),
|
|
make_tuple(4, 4, &vpx_variance16x16_msa, 0),
|
|
make_tuple(4, 3, &vpx_variance16x8_msa, 0),
|
|
make_tuple(3, 4, &vpx_variance8x16_msa, 0),
|
|
make_tuple(3, 3, &vpx_variance8x8_msa, 0),
|
|
make_tuple(3, 2, &vpx_variance8x4_msa, 0),
|
|
make_tuple(2, 3, &vpx_variance4x8_msa, 0),
|
|
make_tuple(2, 2, &vpx_variance4x4_msa, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MSA, VpxSubpelVarianceTest,
|
|
::testing::Values(make_tuple(2, 2, &vpx_sub_pixel_variance4x4_msa, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_variance4x8_msa, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_variance8x4_msa, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_variance8x8_msa, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_variance8x16_msa, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_variance16x8_msa, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_variance16x16_msa, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_variance16x32_msa, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_variance32x16_msa, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_variance32x32_msa, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_variance32x64_msa, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_variance64x32_msa, 0),
|
|
make_tuple(6, 6, &vpx_sub_pixel_variance64x64_msa, 0)));
|
|
|
|
INSTANTIATE_TEST_CASE_P(
|
|
MSA, VpxSubpelAvgVarianceTest,
|
|
::testing::Values(make_tuple(6, 6, &vpx_sub_pixel_avg_variance64x64_msa, 0),
|
|
make_tuple(6, 5, &vpx_sub_pixel_avg_variance64x32_msa, 0),
|
|
make_tuple(5, 6, &vpx_sub_pixel_avg_variance32x64_msa, 0),
|
|
make_tuple(5, 5, &vpx_sub_pixel_avg_variance32x32_msa, 0),
|
|
make_tuple(5, 4, &vpx_sub_pixel_avg_variance32x16_msa, 0),
|
|
make_tuple(4, 5, &vpx_sub_pixel_avg_variance16x32_msa, 0),
|
|
make_tuple(4, 4, &vpx_sub_pixel_avg_variance16x16_msa, 0),
|
|
make_tuple(4, 3, &vpx_sub_pixel_avg_variance16x8_msa, 0),
|
|
make_tuple(3, 4, &vpx_sub_pixel_avg_variance8x16_msa, 0),
|
|
make_tuple(3, 3, &vpx_sub_pixel_avg_variance8x8_msa, 0),
|
|
make_tuple(3, 2, &vpx_sub_pixel_avg_variance8x4_msa, 0),
|
|
make_tuple(2, 3, &vpx_sub_pixel_avg_variance4x8_msa, 0),
|
|
make_tuple(2, 2, &vpx_sub_pixel_avg_variance4x4_msa, 0)));
|
|
#endif // HAVE_MSA
|
|
} // namespace
|