2019-02-18 15:52:47 +00:00
|
|
|
/*****************************************************************************
|
|
|
|
* This file is part of Kvazaar HEVC encoder.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2013-2015 Tampere University of Technology and others (see
|
|
|
|
* COPYING file).
|
|
|
|
*
|
|
|
|
* Kvazaar is free software: you can redistribute it and/or modify it under
|
|
|
|
* the terms of the GNU Lesser General Public License as published by the
|
|
|
|
* Free Software Foundation; either version 2.1 of the License, or (at your
|
|
|
|
* option) any later version.
|
|
|
|
*
|
|
|
|
* Kvazaar is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
|
|
* WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
|
|
* FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for
|
|
|
|
* more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License along
|
|
|
|
* with Kvazaar. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
****************************************************************************/
|
|
|
|
|
2019-01-15 17:37:53 +00:00
|
|
|
#ifndef REG_SAD_POW2_WIDTHS_SSE41_H_
|
|
|
|
#define REG_SAD_POW2_WIDTHS_SSE41_H_
|
|
|
|
|
|
|
|
#include <immintrin.h>
|
|
|
|
#include "kvazaar.h"
|
|
|
|
|
2019-01-18 15:32:52 +00:00
|
|
|
static INLINE uint32_t reg_sad_w0(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_w4(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 19:09:18 +00:00
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
2019-01-16 19:48:18 +00:00
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
2019-01-15 19:09:18 +00:00
|
|
|
|
2019-01-16 19:48:18 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
2019-01-15 19:09:18 +00:00
|
|
|
__m128i a = _mm_cvtsi32_si128(*(uint32_t *)(data1 + y * stride1));
|
|
|
|
__m128i b = _mm_cvtsi32_si128(*(uint32_t *)(data2 + y * stride2));
|
|
|
|
|
2019-01-16 18:39:51 +00:00
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(data1 + (y + 1) * stride1), 1);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(data2 + (y + 1) * stride2), 1);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(data1 + (y + 2) * stride1), 2);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(data2 + (y + 2) * stride2), 2);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(data1 + (y + 3) * stride1), 3);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(data2 + (y + 3) * stride2), 3);
|
2019-01-15 19:09:18 +00:00
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, b);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
2019-01-16 19:48:18 +00:00
|
|
|
if (height_residual_lines) {
|
2019-01-15 19:09:18 +00:00
|
|
|
for (; y < height; y++) {
|
2019-01-16 18:39:51 +00:00
|
|
|
__m128i a = _mm_cvtsi32_si128(*(const uint32_t *)(data1 + y * stride1));
|
|
|
|
__m128i b = _mm_cvtsi32_si128(*(const uint32_t *)(data2 + y * stride2));
|
2019-01-15 19:09:18 +00:00
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, b);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_w8(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 17:37:53 +00:00
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
2019-01-16 19:41:42 +00:00
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
2019-01-15 17:37:53 +00:00
|
|
|
|
2019-01-16 19:41:42 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
2019-01-15 17:37:53 +00:00
|
|
|
__m128d a_d = _mm_setzero_pd();
|
|
|
|
__m128d b_d = _mm_setzero_pd();
|
2019-01-16 19:41:42 +00:00
|
|
|
__m128d c_d = _mm_setzero_pd();
|
|
|
|
__m128d d_d = _mm_setzero_pd();
|
2019-01-15 17:37:53 +00:00
|
|
|
|
|
|
|
a_d = _mm_loadl_pd(a_d, (const double *)(data1 + (y + 0) * stride1));
|
|
|
|
b_d = _mm_loadl_pd(b_d, (const double *)(data2 + (y + 0) * stride2));
|
|
|
|
a_d = _mm_loadh_pd(a_d, (const double *)(data1 + (y + 1) * stride1));
|
|
|
|
b_d = _mm_loadh_pd(b_d, (const double *)(data2 + (y + 1) * stride2));
|
|
|
|
|
2019-01-16 19:41:42 +00:00
|
|
|
c_d = _mm_loadl_pd(c_d, (const double *)(data1 + (y + 2) * stride1));
|
|
|
|
d_d = _mm_loadl_pd(d_d, (const double *)(data2 + (y + 2) * stride2));
|
|
|
|
c_d = _mm_loadh_pd(c_d, (const double *)(data1 + (y + 3) * stride1));
|
|
|
|
d_d = _mm_loadh_pd(d_d, (const double *)(data2 + (y + 3) * stride2));
|
|
|
|
|
2019-01-15 17:37:53 +00:00
|
|
|
__m128i a = _mm_castpd_si128(a_d);
|
|
|
|
__m128i b = _mm_castpd_si128(b_d);
|
2019-01-16 19:41:42 +00:00
|
|
|
__m128i c = _mm_castpd_si128(c_d);
|
|
|
|
__m128i d = _mm_castpd_si128(d_d);
|
2019-01-15 17:37:53 +00:00
|
|
|
|
2019-01-16 19:41:42 +00:00
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
2019-01-16 19:41:42 +00:00
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
2019-02-18 18:13:33 +00:00
|
|
|
__m128i a = _mm_cvtsi64_si128(*(int64_t *)(data1 + y * stride1));
|
|
|
|
__m128i b = _mm_cvtsi64_si128(*(int64_t *)(data2 + y * stride2));
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
2019-01-16 19:41:42 +00:00
|
|
|
}
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
2019-02-18 18:13:33 +00:00
|
|
|
return _mm_cvtsi128_si32(sad);
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_w12(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 19:09:18 +00:00
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
for (y = 0; y < height; y++) {
|
2019-01-16 18:39:51 +00:00
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + y * stride1));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + y * stride2));
|
2019-01-15 19:09:18 +00:00
|
|
|
|
|
|
|
__m128i b_masked = _mm_blend_epi16(a, b, 0x3f);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (a, b_masked);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_w16(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 17:37:53 +00:00
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
2019-01-16 18:54:34 +00:00
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + (y + 0) * stride1));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + (y + 0) * stride2));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(data1 + (y + 1) * stride1));
|
|
|
|
__m128i d = _mm_loadu_si128((const __m128i *)(data2 + (y + 1) * stride2));
|
|
|
|
__m128i e = _mm_loadu_si128((const __m128i *)(data1 + (y + 2) * stride1));
|
|
|
|
__m128i f = _mm_loadu_si128((const __m128i *)(data2 + (y + 2) * stride2));
|
|
|
|
__m128i g = _mm_loadu_si128((const __m128i *)(data1 + (y + 3) * stride1));
|
|
|
|
__m128i h = _mm_loadu_si128((const __m128i *)(data2 + (y + 3) * stride2));
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(e, f);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(g, h);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + (y + 0) * stride1));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + (y + 0) * stride2));
|
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, b);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
2019-01-16 18:54:34 +00:00
|
|
|
|
2019-01-15 17:37:53 +00:00
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_w24(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 20:08:50 +00:00
|
|
|
const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
|
|
|
const int32_t height_doublelines = height & ~1;
|
|
|
|
const int32_t height_parity = height & 1;
|
|
|
|
|
|
|
|
for (y = 0; y < height_doublelines; y += 2) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + (y + 0) * stride1));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + (y + 0) * stride2));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(data1 + (y + 1) * stride1));
|
|
|
|
__m128i d = _mm_loadu_si128((const __m128i *)(data2 + (y + 1) * stride2));
|
|
|
|
|
|
|
|
__m128d e_d = _mm_setzero_pd();
|
|
|
|
__m128d f_d = _mm_setzero_pd();
|
|
|
|
|
|
|
|
e_d = _mm_loadl_pd(e_d, (const double *)(data1 + (y + 0) * stride1 + 16));
|
|
|
|
f_d = _mm_loadl_pd(f_d, (const double *)(data2 + (y + 0) * stride2 + 16));
|
|
|
|
e_d = _mm_loadh_pd(e_d, (const double *)(data1 + (y + 1) * stride1 + 16));
|
|
|
|
f_d = _mm_loadh_pd(f_d, (const double *)(data2 + (y + 1) * stride2 + 16));
|
|
|
|
|
|
|
|
__m128i e = _mm_castpd_si128(e_d);
|
|
|
|
__m128i f = _mm_castpd_si128(f_d);
|
|
|
|
|
|
|
|
__m128i curr_sads_1 = _mm_sad_epu8(a, b);
|
|
|
|
__m128i curr_sads_2 = _mm_sad_epu8(c, d);
|
|
|
|
__m128i curr_sads_3 = _mm_sad_epu8(e, f);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_1);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_2);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_3);
|
|
|
|
}
|
|
|
|
if (height_parity) {
|
2019-01-16 18:39:51 +00:00
|
|
|
__m128i a = _mm_loadu_si128 ((const __m128i *) (data1 + y * stride1));
|
|
|
|
__m128i b = _mm_loadu_si128 ((const __m128i *) (data2 + y * stride2));
|
2019-01-24 13:25:45 +00:00
|
|
|
__m128i c = _mm_cvtsi64_si128(*(const uint64_t *)(data1 + y * stride1 + 16));
|
|
|
|
__m128i d = _mm_cvtsi64_si128(*(const uint64_t *)(data2 + y * stride2 + 16));
|
2019-01-15 20:08:50 +00:00
|
|
|
|
|
|
|
__m128i curr_sads_1 = _mm_sad_epu8(a, b);
|
|
|
|
__m128i curr_sads_2 = _mm_sad_epu8(c, d);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_1);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_2);
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-16 17:57:53 +00:00
|
|
|
static INLINE uint32_t reg_sad_arbitrary(const kvz_pixel * const data1, const kvz_pixel * const data2,
|
2019-01-15 17:37:53 +00:00
|
|
|
const int32_t width, const int32_t height, const uint32_t stride1,
|
|
|
|
const uint32_t stride2)
|
|
|
|
{
|
|
|
|
int32_t y, x;
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
|
|
|
|
// Bytes in block in 128-bit blocks per each scanline, and remainder
|
2019-01-16 19:48:18 +00:00
|
|
|
const int32_t width_xmms = width & ~15;
|
|
|
|
const int32_t width_residual_pixels = width & 15;
|
2019-01-16 19:35:33 +00:00
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
2019-01-15 17:37:53 +00:00
|
|
|
|
2019-01-16 19:48:18 +00:00
|
|
|
const __m128i rds = _mm_set1_epi8 (width_residual_pixels);
|
2019-01-15 17:37:53 +00:00
|
|
|
const __m128i ns = _mm_setr_epi8 (0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
|
|
const __m128i rdmask = _mm_cmpgt_epi8(rds, ns);
|
|
|
|
|
2019-01-16 19:48:18 +00:00
|
|
|
for (x = 0; x < width_xmms; x += 16) {
|
2019-01-16 19:35:33 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + (y + 0) * stride1 + x));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + (y + 0) * stride2 + x));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(data1 + (y + 1) * stride1 + x));
|
|
|
|
__m128i d = _mm_loadu_si128((const __m128i *)(data2 + (y + 1) * stride2 + x));
|
|
|
|
__m128i e = _mm_loadu_si128((const __m128i *)(data1 + (y + 2) * stride1 + x));
|
|
|
|
__m128i f = _mm_loadu_si128((const __m128i *)(data2 + (y + 2) * stride2 + x));
|
|
|
|
__m128i g = _mm_loadu_si128((const __m128i *)(data1 + (y + 3) * stride1 + x));
|
|
|
|
__m128i h = _mm_loadu_si128((const __m128i *)(data2 + (y + 3) * stride2 + x));
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(e, f);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(g, h);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + y * stride1 + x));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + y * stride2 + x));
|
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, b);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-16 19:48:18 +00:00
|
|
|
if (width_residual_pixels) {
|
2019-01-16 19:35:33 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + (y + 0) * stride1 + x));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + (y + 0) * stride2 + x));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(data1 + (y + 1) * stride1 + x));
|
|
|
|
__m128i d = _mm_loadu_si128((const __m128i *)(data2 + (y + 1) * stride2 + x));
|
|
|
|
__m128i e = _mm_loadu_si128((const __m128i *)(data1 + (y + 2) * stride1 + x));
|
|
|
|
__m128i f = _mm_loadu_si128((const __m128i *)(data2 + (y + 2) * stride2 + x));
|
|
|
|
__m128i g = _mm_loadu_si128((const __m128i *)(data1 + (y + 3) * stride1 + x));
|
|
|
|
__m128i h = _mm_loadu_si128((const __m128i *)(data2 + (y + 3) * stride2 + x));
|
|
|
|
|
|
|
|
__m128i b_masked = _mm_blendv_epi8(a, b, rdmask);
|
|
|
|
__m128i d_masked = _mm_blendv_epi8(c, d, rdmask);
|
|
|
|
__m128i f_masked = _mm_blendv_epi8(e, f, rdmask);
|
|
|
|
__m128i h_masked = _mm_blendv_epi8(g, h, rdmask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8 (a, b_masked);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8 (c, d_masked);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8 (e, f_masked);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8 (g, h_masked);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
2019-01-16 19:35:33 +00:00
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(data1 + y * stride1 + x));
|
|
|
|
__m128i b = _mm_loadu_si128((const __m128i *)(data2 + y * stride2 + x));
|
|
|
|
|
|
|
|
__m128i b_masked = _mm_blendv_epi8(a, b, rdmask);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (a, b_masked);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
2019-01-15 17:37:53 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-22 13:57:16 +00:00
|
|
|
static uint32_t ver_sad_w4(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t height, uint32_t stride)
|
|
|
|
{
|
|
|
|
__m128i ref_row = _mm_set1_epi32(*(const uint32_t *)ref_data);
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_cvtsi32_si128(*(uint32_t *)(pic_data + y * stride));
|
|
|
|
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 1) * stride), 1);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 2) * stride), 2);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 3) * stride), 3);
|
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, ref_row);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
// Only pick the last dword, because we're comparing single dwords (lines)
|
|
|
|
ref_row = _mm_bsrli_si128(ref_row, 12);
|
|
|
|
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_cvtsi32_si128(*(const uint32_t *)(pic_data + y * stride));
|
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, ref_row);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
|
|
|
static uint32_t ver_sad_w8(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t height, uint32_t stride)
|
|
|
|
{
|
|
|
|
const __m128i ref_row = _mm_set1_epi64x(*(const uint64_t *)ref_data);
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128d a_d = _mm_setzero_pd();
|
|
|
|
__m128d c_d = _mm_setzero_pd();
|
|
|
|
|
|
|
|
a_d = _mm_loadl_pd(a_d, (const double *)(pic_data + (y + 0) * stride));
|
|
|
|
a_d = _mm_loadh_pd(a_d, (const double *)(pic_data + (y + 1) * stride));
|
|
|
|
|
|
|
|
c_d = _mm_loadl_pd(c_d, (const double *)(pic_data + (y + 2) * stride));
|
|
|
|
c_d = _mm_loadh_pd(c_d, (const double *)(pic_data + (y + 3) * stride));
|
|
|
|
|
|
|
|
__m128i a = _mm_castpd_si128(a_d);
|
|
|
|
__m128i c = _mm_castpd_si128(c_d);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, ref_row);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, ref_row);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
2019-02-18 18:13:33 +00:00
|
|
|
__m128i b = _mm_move_epi64(ref_row);
|
2019-01-22 13:57:16 +00:00
|
|
|
|
|
|
|
for (; y < height; y++) {
|
2019-02-18 18:13:33 +00:00
|
|
|
__m128i a = _mm_cvtsi64_si128(*(int64_t *)(pic_data + y * stride));
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
2019-01-22 13:57:16 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
2019-02-18 18:13:33 +00:00
|
|
|
return _mm_cvtsi128_si32(sad);
|
2019-01-22 13:57:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static uint32_t ver_sad_w12(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t height, uint32_t stride)
|
|
|
|
{
|
|
|
|
const __m128i ref_row = _mm_loadu_si128((__m128i *)ref_data);
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
|
|
|
for (y = 0; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(pic_data + y * stride));
|
|
|
|
|
|
|
|
__m128i a_masked = _mm_blend_epi16(ref_row, a, 0x3f);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (ref_row, a_masked);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
|
|
|
static uint32_t ver_sad_w16(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t height, uint32_t stride)
|
|
|
|
{
|
|
|
|
const __m128i ref_row = _mm_loadu_si128((__m128i *)ref_data);
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i pic_row_1 = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * stride));
|
|
|
|
__m128i pic_row_2 = _mm_loadu_si128((__m128i *)(pic_data + (y + 1) * stride));
|
|
|
|
__m128i pic_row_3 = _mm_loadu_si128((__m128i *)(pic_data + (y + 2) * stride));
|
|
|
|
__m128i pic_row_4 = _mm_loadu_si128((__m128i *)(pic_data + (y + 3) * stride));
|
|
|
|
|
|
|
|
__m128i curr_sads_1 = _mm_sad_epu8 (pic_row_1, ref_row);
|
|
|
|
__m128i curr_sads_2 = _mm_sad_epu8 (pic_row_2, ref_row);
|
|
|
|
__m128i curr_sads_3 = _mm_sad_epu8 (pic_row_3, ref_row);
|
|
|
|
__m128i curr_sads_4 = _mm_sad_epu8 (pic_row_4, ref_row);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_1);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_2);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_3);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_4);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i pic_row = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * stride));
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (pic_row, ref_row);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
|
|
|
static uint32_t ver_sad_arbitrary(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t width, int32_t height, uint32_t stride)
|
|
|
|
{
|
|
|
|
int32_t y, x;
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
|
|
|
|
// Bytes in block in 128-bit blocks per each scanline, and remainder
|
|
|
|
const int32_t width_xmms = width & ~15;
|
|
|
|
const int32_t width_residual_pixels = width & 15;
|
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
const __m128i rds = _mm_set1_epi8 (width_residual_pixels);
|
|
|
|
const __m128i ns = _mm_setr_epi8 (0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
|
|
const __m128i rdmask = _mm_cmpgt_epi8(rds, ns);
|
|
|
|
|
|
|
|
for (x = 0; x < width_xmms; x += 16) {
|
|
|
|
const __m128i ref_row = _mm_loadu_si128((__m128i *)(ref_data + x));
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(pic_data + (y + 0) * stride + x));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(pic_data + (y + 1) * stride + x));
|
|
|
|
__m128i e = _mm_loadu_si128((const __m128i *)(pic_data + (y + 2) * stride + x));
|
|
|
|
__m128i g = _mm_loadu_si128((const __m128i *)(pic_data + (y + 3) * stride + x));
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(ref_row, a);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(ref_row, c);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(ref_row, e);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(ref_row, g);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(pic_data + y * stride + x));
|
|
|
|
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, ref_row);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (width_residual_pixels) {
|
|
|
|
const __m128i ref_row = _mm_loadu_si128((__m128i *)(ref_data + x));
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(pic_data + (y + 0) * stride + x));
|
|
|
|
__m128i c = _mm_loadu_si128((const __m128i *)(pic_data + (y + 1) * stride + x));
|
|
|
|
__m128i e = _mm_loadu_si128((const __m128i *)(pic_data + (y + 2) * stride + x));
|
|
|
|
__m128i g = _mm_loadu_si128((const __m128i *)(pic_data + (y + 3) * stride + x));
|
|
|
|
|
|
|
|
__m128i a_masked = _mm_blendv_epi8(ref_row, a, rdmask);
|
|
|
|
__m128i c_masked = _mm_blendv_epi8(ref_row, c, rdmask);
|
|
|
|
__m128i e_masked = _mm_blendv_epi8(ref_row, e, rdmask);
|
|
|
|
__m128i g_masked = _mm_blendv_epi8(ref_row, g, rdmask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8 (ref_row, a_masked);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8 (ref_row, c_masked);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8 (ref_row, e_masked);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8 (ref_row, g_masked);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((const __m128i *)(pic_data + y * stride + x));
|
|
|
|
|
|
|
|
__m128i a_masked = _mm_blendv_epi8(ref_row, a, rdmask);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (ref_row, a_masked);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-31 18:33:41 +00:00
|
|
|
static uint32_t hor_sad_sse41_w4(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
2019-02-01 12:11:30 +00:00
|
|
|
int32_t height, uint32_t pic_stride, uint32_t ref_stride,
|
|
|
|
uint32_t left, uint32_t right)
|
2019-01-22 19:27:21 +00:00
|
|
|
{
|
2019-02-06 17:26:08 +00:00
|
|
|
const int32_t right_border_idx = 3 - right;
|
|
|
|
const int32_t border_idx = left ? left : right_border_idx;
|
|
|
|
|
|
|
|
const __m128i ns = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
|
|
|
|
|
|
const int32_t border_idx_negative = border_idx >> 31;
|
|
|
|
const int32_t leftoff = border_idx_negative | left;
|
2019-01-31 18:33:41 +00:00
|
|
|
|
2019-01-22 19:27:21 +00:00
|
|
|
// Dualword (ie. line) base indexes, ie. the edges the lines read will be
|
|
|
|
// clamped towards
|
2019-01-31 18:33:41 +00:00
|
|
|
const __m128i dwbaseids = _mm_setr_epi8(0, 0, 0, 0, 4, 4, 4, 4,
|
|
|
|
8, 8, 8, 8, 12, 12, 12, 12);
|
|
|
|
|
2019-02-06 17:26:08 +00:00
|
|
|
__m128i right_border_idxs = _mm_set1_epi8((int8_t)right_border_idx);
|
|
|
|
__m128i left_128 = _mm_set1_epi8((int8_t)left);
|
|
|
|
|
|
|
|
right_border_idxs = _mm_add_epi8 (right_border_idxs, dwbaseids);
|
|
|
|
|
|
|
|
__m128i mask_right = _mm_min_epi8 (ns, right_border_idxs);
|
|
|
|
__m128i mask1 = _mm_sub_epi8 (mask_right, left_128);
|
|
|
|
|
|
|
|
const __m128i epol_mask = _mm_max_epi8(mask1, dwbaseids);
|
|
|
|
|
2019-01-22 19:27:21 +00:00
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_cvtsi32_si128(*(const uint32_t *)(pic_data + y * pic_stride));
|
2019-01-31 18:33:41 +00:00
|
|
|
__m128i b = _mm_cvtsi32_si128(*(const uint32_t *)(ref_data + y * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
|
2019-01-31 18:33:41 +00:00
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 1) * pic_stride), 1);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(ref_data + (y + 1) * ref_stride + leftoff), 1);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 2) * pic_stride), 2);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(ref_data + (y + 2) * ref_stride + leftoff), 2);
|
|
|
|
a = _mm_insert_epi32(a, *(const uint32_t *)(pic_data + (y + 3) * pic_stride), 3);
|
|
|
|
b = _mm_insert_epi32(b, *(const uint32_t *)(ref_data + (y + 3) * ref_stride + leftoff), 3);
|
2019-01-22 19:27:21 +00:00
|
|
|
|
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (a, b_epol);
|
|
|
|
sse_inc = _mm_add_epi64 (sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_cvtsi32_si128(*(const uint32_t *)(pic_data + y * pic_stride));
|
2019-01-31 18:33:41 +00:00
|
|
|
__m128i b = _mm_cvtsi32_si128(*(const uint32_t *)(ref_data + y * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
|
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8 (a, b_epol);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-31 18:08:31 +00:00
|
|
|
static uint32_t hor_sad_sse41_w8(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
2019-02-01 12:11:30 +00:00
|
|
|
int32_t height, uint32_t pic_stride, uint32_t ref_stride,
|
|
|
|
uint32_t left, uint32_t right)
|
2019-01-22 19:27:21 +00:00
|
|
|
{
|
2019-02-06 17:26:08 +00:00
|
|
|
// right is the number of overhanging pixels in the vector, so it has to be
|
|
|
|
// handled this way to produce the index of last valid (border) pixel
|
|
|
|
const int32_t right_border_idx = 7 - right;
|
|
|
|
const int32_t border_idx = left ? left : right_border_idx;
|
|
|
|
|
|
|
|
const __m128i ns = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
2019-01-31 18:08:31 +00:00
|
|
|
|
2019-01-22 19:27:21 +00:00
|
|
|
// Quadword (ie. line) base indexes, ie. the edges the lines read will be
|
|
|
|
// clamped towards; higher qword (lower line) bytes tend towards 8 and lower
|
|
|
|
// qword (higher line) bytes towards 0
|
2019-01-31 18:08:31 +00:00
|
|
|
const __m128i qwbaseids = _mm_setr_epi8(0, 0, 0, 0, 0, 0, 0, 0,
|
|
|
|
8, 8, 8, 8, 8, 8, 8, 8);
|
|
|
|
|
2019-02-06 17:26:08 +00:00
|
|
|
// Dirty hack alert! If right == block_width (ie. the entire vector is
|
|
|
|
// outside the frame), move the block offset one pixel to the left (so
|
|
|
|
// that the leftmost pixel in vector is actually the valid border pixel
|
|
|
|
// from which we want to extrapolate), and use an epol mask that will
|
|
|
|
// simply stretch the pixel all over the vector.
|
|
|
|
//
|
|
|
|
// To avoid a branch here:
|
|
|
|
// The mask will be -1 (0xffffffff) for border_idx -1 and 0 for >= 0
|
|
|
|
const int32_t border_idx_negative = border_idx >> 31;
|
|
|
|
const int32_t leftoff = border_idx_negative | left;
|
|
|
|
|
|
|
|
__m128i right_border_idxs = _mm_set1_epi8((int8_t)right_border_idx);
|
|
|
|
__m128i left_128 = _mm_set1_epi8((int8_t)left);
|
|
|
|
|
|
|
|
right_border_idxs = _mm_add_epi8 (right_border_idxs, qwbaseids);
|
|
|
|
|
|
|
|
// If we're straddling the left border, right_border_idx is 7 and the first
|
|
|
|
// operation does nothing. If right border, left is 0 and the second
|
|
|
|
// operation does nothing.
|
|
|
|
__m128i mask_right = _mm_min_epi8 (ns, right_border_idxs);
|
|
|
|
__m128i mask1 = _mm_sub_epi8 (mask_right, left_128);
|
|
|
|
|
|
|
|
// If right == 8 (we're completely outside the frame), right_border_idx is
|
|
|
|
// -1 and so is mask1. Clamp negative values to qwbaseid and as discussed
|
|
|
|
// earlier, adjust the load offset instead to load the "-1'st" pixels and
|
|
|
|
// using qwbaseids as the shuffle mask, broadcast it all over the rows.
|
|
|
|
const __m128i epol_mask = _mm_max_epi8(mask1, qwbaseids);
|
|
|
|
|
2019-01-22 19:27:21 +00:00
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128d a_d = _mm_setzero_pd();
|
|
|
|
__m128d b_d = _mm_setzero_pd();
|
|
|
|
__m128d c_d = _mm_setzero_pd();
|
|
|
|
__m128d d_d = _mm_setzero_pd();
|
|
|
|
|
|
|
|
a_d = _mm_loadl_pd(a_d, (const double *)(pic_data + (y + 0) * pic_stride));
|
2019-01-31 18:08:31 +00:00
|
|
|
b_d = _mm_loadl_pd(b_d, (const double *)(ref_data + (y + 0) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
a_d = _mm_loadh_pd(a_d, (const double *)(pic_data + (y + 1) * pic_stride));
|
2019-01-31 18:08:31 +00:00
|
|
|
b_d = _mm_loadh_pd(b_d, (const double *)(ref_data + (y + 1) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
|
|
|
|
c_d = _mm_loadl_pd(c_d, (const double *)(pic_data + (y + 2) * pic_stride));
|
2019-01-31 18:08:31 +00:00
|
|
|
d_d = _mm_loadl_pd(d_d, (const double *)(ref_data + (y + 2) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
c_d = _mm_loadh_pd(c_d, (const double *)(pic_data + (y + 3) * pic_stride));
|
2019-01-31 18:08:31 +00:00
|
|
|
d_d = _mm_loadh_pd(d_d, (const double *)(ref_data + (y + 3) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
|
|
|
|
__m128i a = _mm_castpd_si128(a_d);
|
|
|
|
__m128i b = _mm_castpd_si128(b_d);
|
|
|
|
__m128i c = _mm_castpd_si128(c_d);
|
|
|
|
__m128i d = _mm_castpd_si128(d_d);
|
|
|
|
|
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
|
|
|
__m128i d_epol = _mm_shuffle_epi8(d, epol_mask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d_epol);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
2019-02-18 18:13:33 +00:00
|
|
|
__m128i a = _mm_cvtsi64_si128(*(int64_t *)(pic_data + y * pic_stride));
|
|
|
|
__m128i b = _mm_cvtsi64_si128(*(int64_t *)(ref_data + y * ref_stride + leftoff));
|
|
|
|
|
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
2019-01-22 19:27:21 +00:00
|
|
|
|
2019-02-18 18:13:33 +00:00
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
2019-01-22 19:27:21 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
2019-02-18 18:13:33 +00:00
|
|
|
return _mm_cvtsi128_si32(sad);
|
2019-01-22 19:27:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2019-01-31 13:40:02 +00:00
|
|
|
* left and right measure how many pixels of one horizontal scanline will be
|
|
|
|
* outside either the left or the right screen border. For blocks straddling
|
|
|
|
* the left border, read the scanlines starting from the left border instead,
|
|
|
|
* and use the extrapolation mask to essentially move the pixels right while
|
|
|
|
* copying the left border pixel to the vector positions that logically point
|
|
|
|
* outside of the buffer.
|
|
|
|
*
|
|
|
|
* For blocks straddling the right border, just read over the right border,
|
|
|
|
* and extrapolate all pixels beyond the border idx to copy the value of the
|
|
|
|
* border pixel. An exception is right == width (leftmost reference pixel is
|
|
|
|
* one place right from the right border, it's ugly because the pixel to
|
|
|
|
* extrapolate from is located at relative X offset -1), abuse the left border
|
|
|
|
* aligning functionality instead to actually read starting from the valid
|
|
|
|
* border pixel, and use a suitable mask to fill all the other pixels with
|
|
|
|
* that value.
|
2019-01-22 19:27:21 +00:00
|
|
|
*/
|
2019-01-31 13:40:02 +00:00
|
|
|
static uint32_t hor_sad_sse41_w16(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
2019-02-01 12:11:30 +00:00
|
|
|
int32_t height, uint32_t pic_stride, uint32_t ref_stride,
|
2019-02-04 19:27:53 +00:00
|
|
|
const uint32_t left, const uint32_t right)
|
2019-01-22 19:27:21 +00:00
|
|
|
{
|
2019-02-04 19:27:53 +00:00
|
|
|
// right is the number of overhanging pixels in the vector, so it has to be
|
|
|
|
// handled this way to produce the index of last valid (border) pixel
|
|
|
|
const int32_t right_border_idx = 15 - right;
|
|
|
|
const int32_t border_idx = left ? left : right_border_idx;
|
|
|
|
|
|
|
|
const __m128i ns = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
|
|
const __m128i zero = _mm_setzero_si128();
|
|
|
|
|
|
|
|
// Dirty hack alert! If right == block_width (ie. the entire vector is
|
|
|
|
// outside the frame), move the block offset one pixel to the left (so
|
|
|
|
// that the leftmost pixel in vector is actually the valid border pixel
|
|
|
|
// from which we want to extrapolate), and use an epol mask that will
|
|
|
|
// simply stretch the pixel all over the vector.
|
|
|
|
//
|
|
|
|
// To avoid a branch here:
|
|
|
|
// The mask will be -1 (0xffffffff) for border_idx -1 and 0 for >= 0
|
|
|
|
const int32_t border_idx_negative = border_idx >> 31;
|
|
|
|
const int32_t leftoff = border_idx_negative | left;
|
|
|
|
|
|
|
|
__m128i right_border_idxs = _mm_set1_epi8((int8_t)right_border_idx);
|
|
|
|
__m128i left_128 = _mm_set1_epi8((int8_t)left);
|
|
|
|
|
|
|
|
// If we're straddling the left border, right_border_idx is 15 and the first
|
|
|
|
// operation does nothing. If right border, left is 0 and the second
|
|
|
|
// operation does nothing.
|
|
|
|
__m128i mask_right = _mm_min_epi8 (ns, right_border_idxs);
|
|
|
|
__m128i mask1 = _mm_sub_epi8 (mask_right, left_128);
|
|
|
|
|
|
|
|
// If right == 16 (we're completely outside the frame), right_border_idx is
|
|
|
|
// -1 and so is mask1. Clamp negative values to zero and as discussed
|
|
|
|
// earlier, adjust the load offset instead to load the "-1'st" pixel and
|
|
|
|
// using an all-zero shuffle mask, broadcast it all over the vector.
|
|
|
|
const __m128i epol_mask = _mm_max_epi8(mask1, zero);
|
2019-01-31 13:40:02 +00:00
|
|
|
|
2019-01-22 19:27:21 +00:00
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
int32_t y;
|
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride));
|
2019-01-31 13:40:02 +00:00
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
__m128i c = _mm_loadu_si128((__m128i *)(pic_data + (y + 1) * pic_stride));
|
2019-01-31 13:40:02 +00:00
|
|
|
__m128i d = _mm_loadu_si128((__m128i *)(ref_data + (y + 1) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
__m128i e = _mm_loadu_si128((__m128i *)(pic_data + (y + 2) * pic_stride));
|
2019-01-31 13:40:02 +00:00
|
|
|
__m128i f = _mm_loadu_si128((__m128i *)(ref_data + (y + 2) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
__m128i g = _mm_loadu_si128((__m128i *)(pic_data + (y + 3) * pic_stride));
|
2019-01-31 13:40:02 +00:00
|
|
|
__m128i h = _mm_loadu_si128((__m128i *)(ref_data + (y + 3) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
|
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
|
|
|
__m128i d_epol = _mm_shuffle_epi8(d, epol_mask);
|
|
|
|
__m128i f_epol = _mm_shuffle_epi8(f, epol_mask);
|
|
|
|
__m128i h_epol = _mm_shuffle_epi8(h, epol_mask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol);
|
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d_epol);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(e, f_epol);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(g, h_epol);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride));
|
2019-01-31 13:40:02 +00:00
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + leftoff));
|
2019-01-22 19:27:21 +00:00
|
|
|
__m128i b_epol = _mm_shuffle_epi8(b, epol_mask);
|
|
|
|
__m128i curr_sads = _mm_sad_epu8(a, b_epol);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-30 20:57:06 +00:00
|
|
|
static uint32_t hor_sad_sse41_arbitrary(const kvz_pixel *pic_data, const kvz_pixel *ref_data,
|
|
|
|
int32_t width, int32_t height, uint32_t pic_stride,
|
|
|
|
uint32_t ref_stride, uint32_t left, uint32_t right)
|
|
|
|
{
|
|
|
|
const size_t xmm_width = 16;
|
|
|
|
const __m128i xmm_widths = _mm_set1_epi8(xmm_width);
|
|
|
|
|
|
|
|
// Bytes in block in 128-bit blocks per each scanline, and remainder
|
|
|
|
const int32_t width_xmms = width & ~(xmm_width - 1);
|
|
|
|
const int32_t width_residual_pixels = width & (xmm_width - 1);
|
|
|
|
|
|
|
|
const int32_t height_fourline_groups = height & ~3;
|
|
|
|
const int32_t height_residual_lines = height & 3;
|
|
|
|
|
|
|
|
__m128i ns = _mm_setr_epi8(0, 1, 2, 3, 4, 5, 6, 7,
|
|
|
|
8, 9, 10, 11, 12, 13, 14, 15);
|
|
|
|
|
|
|
|
const __m128i rds = _mm_set1_epi8 (width_residual_pixels);
|
|
|
|
const __m128i rdmask = _mm_cmpgt_epi8(rds, ns);
|
|
|
|
|
|
|
|
int32_t border_idx;
|
|
|
|
__m128i is_right_border = _mm_setzero_si128();
|
|
|
|
if (left) {
|
|
|
|
border_idx = left;
|
|
|
|
} else {
|
|
|
|
border_idx = width - (right + 1);
|
|
|
|
is_right_border = _mm_cmpeq_epi8(is_right_border, is_right_border);
|
|
|
|
}
|
|
|
|
const __m128i epol_src_idx = _mm_set1_epi8(border_idx);
|
|
|
|
|
|
|
|
int32_t x, y;
|
|
|
|
__m128i sse_inc = _mm_setzero_si128();
|
|
|
|
__m128i epol_mask;
|
|
|
|
for (x = 0; x < width_xmms; x += xmm_width) {
|
|
|
|
|
|
|
|
// This is a dirty hack, but it saves us an easily predicted branch! It
|
|
|
|
// also marks the first or last valid pixel (the border one) for
|
|
|
|
// extrapolating, but that makes no difference since the pixels marked
|
|
|
|
// for extrapolation will always be written over with that exact pixel's
|
|
|
|
// value.
|
|
|
|
epol_mask = _mm_cmpgt_epi8(epol_src_idx, ns);
|
|
|
|
epol_mask = _mm_xor_si128 (epol_mask, is_right_border);
|
|
|
|
|
2019-01-31 12:21:15 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
2019-01-30 20:57:06 +00:00
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride + x));
|
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + x));
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i c = _mm_loadu_si128((__m128i *)(pic_data + (y + 1) * pic_stride + x));
|
|
|
|
__m128i d = _mm_loadu_si128((__m128i *)(ref_data + (y + 1) * ref_stride + x));
|
|
|
|
__m128i e = _mm_loadu_si128((__m128i *)(pic_data + (y + 2) * pic_stride + x));
|
|
|
|
__m128i f = _mm_loadu_si128((__m128i *)(ref_data + (y + 2) * ref_stride + x));
|
|
|
|
__m128i g = _mm_loadu_si128((__m128i *)(pic_data + (y + 3) * pic_stride + x));
|
|
|
|
__m128i h = _mm_loadu_si128((__m128i *)(ref_data + (y + 3) * ref_stride + x));
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
__m128i border_px_b = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 0) * ref_stride + border_idx));
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i border_px_d = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 1) * ref_stride + border_idx));
|
|
|
|
__m128i border_px_f = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 2) * ref_stride + border_idx));
|
|
|
|
__m128i border_px_h = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 3) * ref_stride + border_idx));
|
2019-01-30 20:57:06 +00:00
|
|
|
__m128i b_epol = _mm_blendv_epi8(b, border_px_b, epol_mask);
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i d_epol = _mm_blendv_epi8(d, border_px_d, epol_mask);
|
|
|
|
__m128i f_epol = _mm_blendv_epi8(f, border_px_f, epol_mask);
|
|
|
|
__m128i h_epol = _mm_blendv_epi8(h, border_px_h, epol_mask);
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol);
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d_epol);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(e, f_epol);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(g, h_epol);
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
2019-01-31 12:21:15 +00:00
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride + x));
|
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + x));
|
|
|
|
|
|
|
|
__m128i border_px_b = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 0) * ref_stride + border_idx));
|
|
|
|
__m128i b_epol = _mm_blendv_epi8(b, border_px_b, epol_mask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
}
|
2019-01-30 20:57:06 +00:00
|
|
|
}
|
|
|
|
ns = _mm_add_epi8(ns, xmm_widths);
|
|
|
|
}
|
|
|
|
if (width_residual_pixels) {
|
|
|
|
epol_mask = _mm_cmpgt_epi8(epol_src_idx, ns);
|
|
|
|
epol_mask = _mm_xor_si128 (epol_mask, is_right_border);
|
|
|
|
|
2019-01-31 12:21:15 +00:00
|
|
|
for (y = 0; y < height_fourline_groups; y += 4) {
|
2019-01-30 20:57:06 +00:00
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride + x));
|
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + x));
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i c = _mm_loadu_si128((__m128i *)(pic_data + (y + 1) * pic_stride + x));
|
|
|
|
__m128i d = _mm_loadu_si128((__m128i *)(ref_data + (y + 1) * ref_stride + x));
|
|
|
|
__m128i e = _mm_loadu_si128((__m128i *)(pic_data + (y + 2) * pic_stride + x));
|
|
|
|
__m128i f = _mm_loadu_si128((__m128i *)(ref_data + (y + 2) * ref_stride + x));
|
|
|
|
__m128i g = _mm_loadu_si128((__m128i *)(pic_data + (y + 3) * pic_stride + x));
|
|
|
|
__m128i h = _mm_loadu_si128((__m128i *)(ref_data + (y + 3) * ref_stride + x));
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
__m128i border_px_b = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 0) * ref_stride + border_idx));
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i border_px_d = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 1) * ref_stride + border_idx));
|
|
|
|
__m128i border_px_f = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 2) * ref_stride + border_idx));
|
|
|
|
__m128i border_px_h = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 3) * ref_stride + border_idx));
|
|
|
|
|
2019-01-30 20:57:06 +00:00
|
|
|
__m128i b_epol_1 = _mm_blendv_epi8(b, border_px_b, epol_mask);
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i d_epol_1 = _mm_blendv_epi8(d, border_px_d, epol_mask);
|
|
|
|
__m128i f_epol_1 = _mm_blendv_epi8(f, border_px_f, epol_mask);
|
|
|
|
__m128i h_epol_1 = _mm_blendv_epi8(h, border_px_h, epol_mask);
|
|
|
|
|
2019-01-30 20:57:06 +00:00
|
|
|
__m128i b_epol_2 = _mm_blendv_epi8(a, b_epol_1, rdmask);
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i d_epol_2 = _mm_blendv_epi8(c, d_epol_1, rdmask);
|
|
|
|
__m128i f_epol_2 = _mm_blendv_epi8(e, f_epol_1, rdmask);
|
|
|
|
__m128i h_epol_2 = _mm_blendv_epi8(g, h_epol_1, rdmask);
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol_2);
|
2019-01-31 12:21:15 +00:00
|
|
|
__m128i curr_sads_cd = _mm_sad_epu8(c, d_epol_2);
|
|
|
|
__m128i curr_sads_ef = _mm_sad_epu8(e, f_epol_2);
|
|
|
|
__m128i curr_sads_gh = _mm_sad_epu8(g, h_epol_2);
|
2019-01-30 20:57:06 +00:00
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
2019-01-31 12:21:15 +00:00
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_cd);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ef);
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_gh);
|
|
|
|
}
|
|
|
|
if (height_residual_lines) {
|
|
|
|
for (; y < height; y++) {
|
|
|
|
__m128i a = _mm_loadu_si128((__m128i *)(pic_data + (y + 0) * pic_stride + x));
|
|
|
|
__m128i b = _mm_loadu_si128((__m128i *)(ref_data + (y + 0) * ref_stride + x));
|
|
|
|
|
|
|
|
__m128i border_px_b = _mm_set1_epi8 (*(uint8_t *)(ref_data + (y + 0) * ref_stride + border_idx));
|
|
|
|
__m128i b_epol_1 = _mm_blendv_epi8(b, border_px_b, epol_mask);
|
|
|
|
__m128i b_epol_2 = _mm_blendv_epi8(a, b_epol_1, rdmask);
|
|
|
|
|
|
|
|
__m128i curr_sads_ab = _mm_sad_epu8(a, b_epol_2);
|
|
|
|
|
|
|
|
sse_inc = _mm_add_epi64(sse_inc, curr_sads_ab);
|
|
|
|
}
|
2019-01-30 20:57:06 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
__m128i sse_inc_2 = _mm_shuffle_epi32(sse_inc, _MM_SHUFFLE(1, 0, 3, 2));
|
|
|
|
__m128i sad = _mm_add_epi64 (sse_inc, sse_inc_2);
|
|
|
|
return _mm_cvtsi128_si32(sad);
|
|
|
|
}
|
|
|
|
|
2019-01-15 17:37:53 +00:00
|
|
|
#endif
|