mirror of
https://github.com/ultravideo/uvg266.git
synced 2024-11-28 11:44:08 +00:00
bdc16d2612
- It works just like the old structure except that the flags are checked with bitmasks instead of having the flag value be propagated upwards. There isn't really any benefit to this because the flags still have to be propagated to parent CUs. - Wrapped them inside a struct to make copying them easier. (Just need to copy the struct instead of making individual copies)
1146 lines
40 KiB
C
1146 lines
40 KiB
C
/*****************************************************************************
|
|
* This file is part of Kvazaar HEVC encoder.
|
|
*
|
|
* Copyright (C) 2013-2014 Tampere University of Technology and others (see
|
|
* COPYING file).
|
|
*
|
|
* Kvazaar is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as published
|
|
* by the Free Software Foundation.
|
|
*
|
|
* Kvazaar is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with Kvazaar. If not, see <http://www.gnu.org/licenses/>.
|
|
****************************************************************************/
|
|
|
|
/*
|
|
* \file
|
|
*/
|
|
|
|
#include "search.h"
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
|
|
#include "config.h"
|
|
#include "bitstream.h"
|
|
#include "picture.h"
|
|
#include "intra.h"
|
|
#include "inter.h"
|
|
#include "filter.h"
|
|
#include "rdo.h"
|
|
|
|
// Temporarily for debugging.
|
|
#define SEARCH_MV_FULL_RADIUS 0
|
|
|
|
#define IN_FRAME(x, y, width, height, block_width, block_height) \
|
|
((x) >= 0 && (y) >= 0 \
|
|
&& (x) + (block_width) <= (width) \
|
|
&& (y) + (block_height) <= (height))
|
|
|
|
/**
|
|
* This is used in the hexagon_search to select 3 points to search.
|
|
*
|
|
* The start of the hexagonal pattern has been repeated at the end so that
|
|
* the indices between 1-6 can be used as the start of a 3-point list of new
|
|
* points to search.
|
|
*
|
|
* 6 o-o 1 / 7
|
|
* / \
|
|
* 5 o 0 o 2 / 8
|
|
* \ /
|
|
* 4 o-o 3
|
|
*/
|
|
const vector2d large_hexbs[10] = {
|
|
{ 0, 0 },
|
|
{ 1, -2 }, { 2, 0 }, { 1, 2 }, { -1, 2 }, { -2, 0 }, { -1, -2 },
|
|
{ 1, -2 }, { 2, 0 }
|
|
};
|
|
|
|
/**
|
|
* This is used as the last step of the hexagon search.
|
|
*/
|
|
const vector2d small_hexbs[5] = {
|
|
{ 0, 0 },
|
|
{ -1, -1 }, { -1, 0 }, { 1, 0 }, { 1, 1 }
|
|
};
|
|
|
|
|
|
static uint32_t get_ep_ex_golomb_bitcost(uint32_t symbol, uint32_t count)
|
|
{
|
|
int32_t num_bins = 0;
|
|
while (symbol >= (uint32_t)(1 << count)) {
|
|
++num_bins;
|
|
symbol -= 1 << count;
|
|
++count;
|
|
}
|
|
num_bins ++;
|
|
|
|
return num_bins;
|
|
}
|
|
|
|
static uint32_t get_mvd_coding_cost(vector2d *mvd)
|
|
{
|
|
uint32_t bitcost = 0;
|
|
const int32_t mvd_hor = mvd->x;
|
|
const int32_t mvd_ver = mvd->y;
|
|
const int8_t hor_abs_gr0 = mvd_hor != 0;
|
|
const int8_t ver_abs_gr0 = mvd_ver != 0;
|
|
const uint32_t mvd_hor_abs = abs(mvd_hor);
|
|
const uint32_t mvd_ver_abs = abs(mvd_ver);
|
|
|
|
// Greater than 0 for x/y
|
|
bitcost += 2;
|
|
|
|
if (hor_abs_gr0) {
|
|
if (mvd_hor_abs > 1) {
|
|
bitcost += get_ep_ex_golomb_bitcost(mvd_hor_abs-2, 1) - 2; // TODO: tune the costs
|
|
}
|
|
// Greater than 1 + sign
|
|
bitcost += 2;
|
|
}
|
|
|
|
if (ver_abs_gr0) {
|
|
if (mvd_ver_abs > 1) {
|
|
bitcost += get_ep_ex_golomb_bitcost(mvd_ver_abs-2, 1) - 2; // TODO: tune the costs
|
|
}
|
|
// Greater than 1 + sign
|
|
bitcost += 2;
|
|
}
|
|
|
|
return bitcost;
|
|
}
|
|
|
|
static int calc_mvd_cost(const encoder_state * const encoder_state, int x, int y,
|
|
int16_t mv_cand[2][2], int16_t merge_cand[MRG_MAX_NUM_CANDS][3],
|
|
int16_t num_cand,int32_t ref_idx, uint32_t *bitcost)
|
|
{
|
|
uint32_t temp_bitcost = 0;
|
|
uint32_t merge_idx;
|
|
int cand1_cost,cand2_cost;
|
|
vector2d mvd_temp1, mvd_temp2;
|
|
int8_t merged = 0;
|
|
int8_t cur_mv_cand = 0;
|
|
|
|
x <<= 2;
|
|
y <<= 2;
|
|
|
|
// Check every candidate to find a match
|
|
for(merge_idx = 0; merge_idx < (uint32_t)num_cand; merge_idx++) {
|
|
if (merge_cand[merge_idx][0] == x &&
|
|
merge_cand[merge_idx][1] == y &&
|
|
merge_cand[merge_idx][2] == ref_idx) {
|
|
temp_bitcost += merge_idx;
|
|
merged = 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Check mvd cost only if mv is not merged
|
|
if(!merged) {
|
|
mvd_temp1.x = x - mv_cand[0][0];
|
|
mvd_temp1.y = y - mv_cand[0][1];
|
|
cand1_cost = get_mvd_coding_cost(&mvd_temp1);
|
|
|
|
mvd_temp2.x = x - mv_cand[1][0];
|
|
mvd_temp2.y = y - mv_cand[1][1];
|
|
cand2_cost = get_mvd_coding_cost(&mvd_temp2);
|
|
|
|
// Select candidate 1 if it has lower cost
|
|
if (cand2_cost < cand1_cost) {
|
|
cur_mv_cand = 1;
|
|
}
|
|
temp_bitcost += cur_mv_cand ? cand2_cost : cand1_cost;
|
|
}
|
|
*bitcost = temp_bitcost;
|
|
return temp_bitcost*(int32_t)(encoder_state->cur_lambda_cost+0.5);
|
|
}
|
|
|
|
|
|
/**
|
|
* \brief Do motion search using the HEXBS algorithm.
|
|
*
|
|
* \param depth log2 depth of the search
|
|
* \param pic Picture motion vector is searched for.
|
|
* \param ref Picture motion vector is searched from.
|
|
* \param orig Top left corner of the searched for block.
|
|
* \param mv_in_out Predicted mv in and best out. Quarter pixel precision.
|
|
*
|
|
* \returns Cost of the motion vector.
|
|
*
|
|
* Motion vector is searched by first searching iteratively with the large
|
|
* hexagon pattern until the best match is at the center of the hexagon.
|
|
* As a final step a smaller hexagon is used to check the adjacent pixels.
|
|
*
|
|
* If a non 0,0 predicted motion vector predictor is given as mv_in_out,
|
|
* the 0,0 vector is also tried. This is hoped to help in the case where
|
|
* the predicted motion vector is way off. In the future even more additional
|
|
* points like 0,0 might be used, such as vectors from top or left.
|
|
*/
|
|
static unsigned hexagon_search(const encoder_state * const encoder_state, unsigned depth,
|
|
const picture *pic, const picture *ref,
|
|
const vector2d *orig, vector2d *mv_in_out,
|
|
int16_t mv_cand[2][2], int16_t merge_cand[MRG_MAX_NUM_CANDS][3],
|
|
int16_t num_cand, int32_t ref_idx, uint32_t *bitcost_out)
|
|
{
|
|
vector2d mv = { mv_in_out->x >> 2, mv_in_out->y >> 2 };
|
|
int block_width = CU_WIDTH_FROM_DEPTH(depth);
|
|
unsigned best_cost = UINT32_MAX;
|
|
uint32_t best_bitcost = 0, bitcost;
|
|
unsigned i;
|
|
unsigned best_index = 0; // Index of large_hexbs or finally small_hexbs.
|
|
|
|
|
|
// Search the initial 7 points of the hexagon.
|
|
for (i = 0; i < 7; ++i) {
|
|
const vector2d *pattern = &large_hexbs[i];
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
(encoder_state->lcu_offset_x * LCU_WIDTH) + orig->x + mv.x + pattern->x,
|
|
(encoder_state->lcu_offset_y * LCU_WIDTH) + orig->y + mv.y + pattern->y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(encoder_state, mv.x + pattern->x, mv.y + pattern->y, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
|
|
if (cost < best_cost) {
|
|
best_cost = cost;
|
|
best_index = i;
|
|
best_bitcost = bitcost;
|
|
}
|
|
}
|
|
|
|
// Try the 0,0 vector.
|
|
if (!(mv.x == 0 && mv.y == 0)) {
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
(encoder_state->lcu_offset_x * LCU_WIDTH) + orig->x,
|
|
(encoder_state->lcu_offset_y * LCU_WIDTH) + orig->y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(encoder_state, 0, 0, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
|
|
// If the 0,0 is better, redo the hexagon around that point.
|
|
if (cost < best_cost) {
|
|
best_cost = cost;
|
|
best_bitcost = bitcost;
|
|
best_index = 0;
|
|
mv.x = 0;
|
|
mv.y = 0;
|
|
|
|
for (i = 1; i < 7; ++i) {
|
|
const vector2d *pattern = &large_hexbs[i];
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
(encoder_state->lcu_offset_x * LCU_WIDTH) + orig->x + pattern->x,
|
|
(encoder_state->lcu_offset_y * LCU_WIDTH) + orig->y + pattern->y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(encoder_state, pattern->x, pattern->y, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
|
|
if (cost < best_cost) {
|
|
best_cost = cost;
|
|
best_index = i;
|
|
best_bitcost = bitcost;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Iteratively search the 3 new points around the best match, until the best
|
|
// match is in the center.
|
|
while (best_index != 0) {
|
|
unsigned start; // Starting point of the 3 offsets to be searched.
|
|
if (best_index == 1) {
|
|
start = 6;
|
|
} else if (best_index == 8) {
|
|
start = 1;
|
|
} else {
|
|
start = best_index - 1;
|
|
}
|
|
|
|
// Move the center to the best match.
|
|
mv.x += large_hexbs[best_index].x;
|
|
mv.y += large_hexbs[best_index].y;
|
|
best_index = 0;
|
|
|
|
// Iterate through the next 3 points.
|
|
for (i = 0; i < 3; ++i) {
|
|
const vector2d *offset = &large_hexbs[start + i];
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
(encoder_state->lcu_offset_x * LCU_WIDTH) + orig->x + mv.x + offset->x,
|
|
(encoder_state->lcu_offset_y * LCU_WIDTH) + orig->y + mv.y + offset->y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(encoder_state, mv.x + offset->x, mv.y + offset->y, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
|
|
if (cost < best_cost) {
|
|
best_cost = cost;
|
|
best_index = start + i;
|
|
best_bitcost = bitcost;
|
|
}
|
|
++offset;
|
|
}
|
|
}
|
|
|
|
// Move the center to the best match.
|
|
mv.x += large_hexbs[best_index].x;
|
|
mv.y += large_hexbs[best_index].y;
|
|
best_index = 0;
|
|
|
|
// Do the final step of the search with a small pattern.
|
|
for (i = 1; i < 5; ++i) {
|
|
const vector2d *offset = &small_hexbs[i];
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
(encoder_state->lcu_offset_x * LCU_WIDTH) + orig->x + mv.x + offset->x,
|
|
(encoder_state->lcu_offset_y * LCU_WIDTH) + orig->y + mv.y + offset->y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(encoder_state, mv.x + offset->x, mv.y + offset->y, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
|
|
if (cost > 0 && cost < best_cost) {
|
|
best_cost = cost;
|
|
best_index = i;
|
|
best_bitcost = bitcost;
|
|
}
|
|
}
|
|
|
|
// Adjust the movement vector according to the final best match.
|
|
mv.x += small_hexbs[best_index].x;
|
|
mv.y += small_hexbs[best_index].y;
|
|
|
|
// Return final movement vector in quarter-pixel precision.
|
|
mv_in_out->x = mv.x << 2;
|
|
mv_in_out->y = mv.y << 2;
|
|
|
|
*bitcost_out = best_bitcost;
|
|
|
|
return best_cost;
|
|
}
|
|
|
|
|
|
#if SEARCH_MV_FULL_RADIUS
|
|
static unsigned search_mv_full(unsigned depth,
|
|
const picture *pic, const picture *ref,
|
|
const vector2d *orig, vector2d *mv_in_out,
|
|
int16_t mv_cand[2][2], int16_t merge_cand[MRG_MAX_NUM_CANDS][3],
|
|
int16_t num_cand, int32_t ref_idx, uint32_t *bitcost_out)
|
|
{
|
|
vector2d mv = { mv_in_out->x >> 2, mv_in_out->y >> 2 };
|
|
int block_width = CU_WIDTH_FROM_DEPTH(depth);
|
|
unsigned best_cost = UINT32_MAX;
|
|
int x, y;
|
|
uint32_t best_bitcost = 0, bitcost;
|
|
vector2d min_mv, max_mv;
|
|
|
|
/*if (abs(mv.x) > SEARCH_MV_FULL_RADIUS || abs(mv.y) > SEARCH_MV_FULL_RADIUS) {
|
|
best_cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
orig->x, orig->y,
|
|
block_width, block_width);
|
|
mv.x = 0;
|
|
mv.y = 0;
|
|
}*/
|
|
|
|
min_mv.x = mv.x - SEARCH_MV_FULL_RADIUS;
|
|
min_mv.y = mv.y - SEARCH_MV_FULL_RADIUS;
|
|
max_mv.x = mv.x + SEARCH_MV_FULL_RADIUS;
|
|
max_mv.y = mv.y + SEARCH_MV_FULL_RADIUS;
|
|
|
|
for (y = min_mv.y; y < max_mv.y; ++y) {
|
|
for (x = min_mv.x; x < max_mv.x; ++x) {
|
|
unsigned cost = calc_sad(pic, ref, orig->x, orig->y,
|
|
orig->x + x,
|
|
orig->y + y,
|
|
block_width, block_width);
|
|
cost += calc_mvd_cost(x, y, mv_cand,merge_cand,num_cand,ref_idx, &bitcost);
|
|
if (cost < best_cost) {
|
|
best_cost = cost;
|
|
best_bitcost = bitcost;
|
|
mv.x = x;
|
|
mv.y = y;
|
|
}
|
|
}
|
|
}
|
|
|
|
mv_in_out->x = mv.x << 2;
|
|
mv_in_out->y = mv.y << 2;
|
|
|
|
*bitcost_out = best_bitcost;
|
|
|
|
return best_cost;
|
|
}
|
|
#endif
|
|
|
|
/**
|
|
* Update lcu to have best modes at this depth.
|
|
* \return Cost of best mode.
|
|
*/
|
|
static int search_cu_inter(const encoder_state * const encoder_state, int x, int y, int depth, lcu_t *lcu)
|
|
{
|
|
const picture * const cur_pic = encoder_state->cur_pic;
|
|
uint32_t ref_idx = 0;
|
|
int x_local = (x&0x3f), y_local = (y&0x3f);
|
|
int x_cu = x>>3;
|
|
int y_cu = y>>3;
|
|
int cu_pos = LCU_CU_OFFSET+(x_local>>3) + (y_local>>3)*LCU_T_CU_WIDTH;
|
|
|
|
cu_info *cur_cu = &lcu->cu[cu_pos];
|
|
|
|
int16_t mv_cand[2][2];
|
|
// Search for merge mode candidate
|
|
int16_t merge_cand[MRG_MAX_NUM_CANDS][3];
|
|
// Get list of candidates
|
|
int16_t num_cand = inter_get_merge_cand(x, y, depth, merge_cand, lcu);
|
|
|
|
// Select better candidate
|
|
cur_cu->inter.mv_cand = 0; // Default to candidate 0
|
|
|
|
cur_cu->inter.cost = UINT_MAX;
|
|
|
|
for (ref_idx = 0; ref_idx < encoder_state->ref->used_size; ref_idx++) {
|
|
picture *ref_pic = encoder_state->ref->pics[ref_idx];
|
|
unsigned width_in_scu = NO_SCU_IN_LCU(ref_pic->width_in_lcu);
|
|
cu_info *ref_cu = &ref_pic->cu_array[y_cu * width_in_scu + x_cu];
|
|
uint32_t temp_bitcost = 0;
|
|
uint32_t temp_cost = 0;
|
|
vector2d orig, mv, mvd;
|
|
int32_t merged = 0;
|
|
uint8_t cu_mv_cand = 0;
|
|
int8_t merge_idx = 0;
|
|
int8_t temp_ref_idx = cur_cu->inter.mv_ref;
|
|
orig.x = x_cu * CU_MIN_SIZE_PIXELS;
|
|
orig.y = y_cu * CU_MIN_SIZE_PIXELS;
|
|
mv.x = 0;
|
|
mv.y = 0;
|
|
if (ref_cu->type == CU_INTER) {
|
|
mv.x = ref_cu->inter.mv[0];
|
|
mv.y = ref_cu->inter.mv[1];
|
|
}
|
|
// Get MV candidates
|
|
cur_cu->inter.mv_ref = ref_idx;
|
|
inter_get_mv_cand(encoder_state, x, y, depth, mv_cand, cur_cu, lcu);
|
|
cur_cu->inter.mv_ref = temp_ref_idx;
|
|
|
|
#if SEARCH_MV_FULL_RADIUS
|
|
temp_cost += search_mv_full(depth, cur_pic, ref_pic, &orig, &mv, mv_cand, merge_cand, num_cand, ref_idx, &temp_bitcost);
|
|
#else
|
|
temp_cost += hexagon_search(encoder_state, depth, cur_pic, ref_pic, &orig, &mv, mv_cand, merge_cand, num_cand, ref_idx, &temp_bitcost);
|
|
#endif
|
|
|
|
merged = 0;
|
|
// Check every candidate to find a match
|
|
for(merge_idx = 0; merge_idx < num_cand; merge_idx++) {
|
|
if (merge_cand[merge_idx][0] == mv.x &&
|
|
merge_cand[merge_idx][1] == mv.y &&
|
|
(uint32_t)merge_cand[merge_idx][2] == ref_idx) {
|
|
merged = 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Only check when candidates are different
|
|
if (!merged && (mv_cand[0][0] != mv_cand[1][0] || mv_cand[0][1] != mv_cand[1][1])) {
|
|
vector2d mvd_temp1, mvd_temp2;
|
|
int cand1_cost,cand2_cost;
|
|
|
|
mvd_temp1.x = mv.x - mv_cand[0][0];
|
|
mvd_temp1.y = mv.y - mv_cand[0][1];
|
|
cand1_cost = get_mvd_coding_cost(&mvd_temp1);
|
|
|
|
mvd_temp2.x = mv.x - mv_cand[1][0];
|
|
mvd_temp2.y = mv.y - mv_cand[1][1];
|
|
cand2_cost = get_mvd_coding_cost(&mvd_temp2);
|
|
|
|
// Select candidate 1 if it has lower cost
|
|
if (cand2_cost < cand1_cost) {
|
|
cu_mv_cand = 1;
|
|
}
|
|
}
|
|
mvd.x = mv.x - mv_cand[cu_mv_cand][0];
|
|
mvd.y = mv.y - mv_cand[cu_mv_cand][1];
|
|
|
|
if(temp_cost < cur_cu->inter.cost) {
|
|
cur_cu->merged = merged;
|
|
cur_cu->merge_idx = merge_idx;
|
|
cur_cu->inter.mv_ref = ref_idx;
|
|
cur_cu->inter.mv_dir = 1;
|
|
cur_cu->inter.mv[0] = (int16_t)mv.x;
|
|
cur_cu->inter.mv[1] = (int16_t)mv.y;
|
|
cur_cu->inter.mvd[0] = (int16_t)mvd.x;
|
|
cur_cu->inter.mvd[1] = (int16_t)mvd.y;
|
|
cur_cu->inter.cost = temp_cost;
|
|
cur_cu->inter.bitcost = temp_bitcost + ref_idx;
|
|
cur_cu->inter.mv_cand = cu_mv_cand;
|
|
}
|
|
}
|
|
|
|
return cur_cu->inter.cost;
|
|
}
|
|
|
|
|
|
/**
|
|
* Copy all non-reference CU data from depth+1 to depth.
|
|
*/
|
|
static void work_tree_copy_up(int x_px, int y_px, int depth, lcu_t work_tree[MAX_PU_DEPTH + 1])
|
|
{
|
|
// Copy non-reference CUs.
|
|
{
|
|
const int x_cu = SUB_SCU(x_px) >> MAX_DEPTH;
|
|
const int y_cu = SUB_SCU(y_px) >> MAX_DEPTH;
|
|
const int width_cu = LCU_WIDTH >> MAX_DEPTH >> depth;
|
|
int x, y;
|
|
for (y = y_cu; y < y_cu + width_cu; ++y) {
|
|
for (x = x_cu; x < x_cu + width_cu; ++x) {
|
|
const cu_info *from_cu = &work_tree[depth + 1].cu[LCU_CU_OFFSET + x + y * LCU_T_CU_WIDTH];
|
|
cu_info *to_cu = &work_tree[depth].cu[LCU_CU_OFFSET + x + y * LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
}
|
|
|
|
// Copy reconstructed pixels.
|
|
{
|
|
const int x = SUB_SCU(x_px);
|
|
const int y = SUB_SCU(y_px);
|
|
const int width_px = LCU_WIDTH >> depth;
|
|
const int luma_index = x + y * LCU_WIDTH;
|
|
const int chroma_index = (x / 2) + (y / 2) * (LCU_WIDTH / 2);
|
|
|
|
const lcu_yuv_t *from = &work_tree[depth + 1].rec;
|
|
lcu_yuv_t *to = &work_tree[depth].rec;
|
|
|
|
const lcu_coeff_t *from_coeff = &work_tree[depth + 1].coeff;
|
|
lcu_coeff_t *to_coeff = &work_tree[depth].coeff;
|
|
|
|
picture_blit_pixels(&from->y[luma_index], &to->y[luma_index],
|
|
width_px, width_px, LCU_WIDTH, LCU_WIDTH);
|
|
picture_blit_pixels(&from->u[chroma_index], &to->u[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
picture_blit_pixels(&from->v[chroma_index], &to->v[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
|
|
// Copy coefficients up. They do not have to be copied down because they
|
|
// are not used for the search.
|
|
picture_blit_coeffs(&from_coeff->y[luma_index], &to_coeff->y[luma_index],
|
|
width_px, width_px, LCU_WIDTH, LCU_WIDTH);
|
|
picture_blit_coeffs(&from_coeff->u[chroma_index], &to_coeff->u[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
picture_blit_coeffs(&from_coeff->v[chroma_index], &to_coeff->v[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* Copy all non-reference CU data from depth to depth+1..MAX_PU_DEPTH.
|
|
*/
|
|
static void work_tree_copy_down(int x_px, int y_px, int depth, lcu_t work_tree[MAX_PU_DEPTH + 1])
|
|
{
|
|
// TODO: clean up to remove the copy pasta
|
|
const int width_px = LCU_WIDTH >> depth;
|
|
|
|
int d;
|
|
|
|
for (d = depth + 1; d < MAX_PU_DEPTH + 1; ++d) {
|
|
const int x_cu = SUB_SCU(x_px) >> MAX_DEPTH;
|
|
const int y_cu = SUB_SCU(y_px) >> MAX_DEPTH;
|
|
const int width_cu = width_px >> MAX_DEPTH;
|
|
|
|
int x, y;
|
|
for (y = y_cu; y < y_cu + width_cu; ++y) {
|
|
for (x = x_cu; x < x_cu + width_cu; ++x) {
|
|
const cu_info *from_cu = &work_tree[depth].cu[LCU_CU_OFFSET + x + y * LCU_T_CU_WIDTH];
|
|
cu_info *to_cu = &work_tree[d].cu[LCU_CU_OFFSET + x + y * LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
}
|
|
|
|
// Copy reconstructed pixels.
|
|
for (d = depth + 1; d < MAX_PU_DEPTH + 1; ++d) {
|
|
const int x = SUB_SCU(x_px);
|
|
const int y = SUB_SCU(y_px);
|
|
|
|
const int luma_index = x + y * LCU_WIDTH;
|
|
const int chroma_index = (x / 2) + (y / 2) * (LCU_WIDTH / 2);
|
|
|
|
lcu_yuv_t *from = &work_tree[depth].rec;
|
|
lcu_yuv_t *to = &work_tree[d].rec;
|
|
|
|
picture_blit_pixels(&from->y[luma_index], &to->y[luma_index],
|
|
width_px, width_px, LCU_WIDTH, LCU_WIDTH);
|
|
picture_blit_pixels(&from->u[chroma_index], &to->u[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
picture_blit_pixels(&from->v[chroma_index], &to->v[chroma_index],
|
|
width_px / 2, width_px / 2, LCU_WIDTH / 2, LCU_WIDTH / 2);
|
|
}
|
|
}
|
|
|
|
|
|
static void lcu_set_intra_mode(lcu_t *lcu, int x_px, int y_px, int depth, int pred_mode, int part_mode)
|
|
{
|
|
const int width_cu = LCU_CU_WIDTH >> depth;
|
|
const int x_cu = SUB_SCU(x_px) >> MAX_DEPTH;
|
|
const int y_cu = SUB_SCU(y_px) >> MAX_DEPTH;
|
|
cu_info *const lcu_cu = &lcu->cu[LCU_CU_OFFSET];
|
|
int x, y;
|
|
|
|
// NxN can only be applied to a single CU at a time.
|
|
if (part_mode == SIZE_NxN) {
|
|
cu_info *cu = &lcu_cu[x_cu + y_cu * LCU_T_CU_WIDTH];
|
|
cu->depth = MAX_DEPTH;
|
|
cu->type = CU_INTRA;
|
|
// It is assumed that cu->intra[].mode's are already set.
|
|
cu->part_size = part_mode;
|
|
cu->tr_depth = depth;
|
|
return;
|
|
}
|
|
|
|
// Set mode in every CU covered by part_mode in this depth.
|
|
for (y = y_cu; y < y_cu + width_cu; ++y) {
|
|
for (x = x_cu; x < x_cu + width_cu; ++x) {
|
|
cu_info *cu = &lcu_cu[x + y * LCU_T_CU_WIDTH];
|
|
cu->depth = depth;
|
|
cu->type = CU_INTRA;
|
|
cu->intra[0].mode = pred_mode;
|
|
cu->intra[1].mode = pred_mode;
|
|
cu->intra[2].mode = pred_mode;
|
|
cu->intra[3].mode = pred_mode;
|
|
cu->part_size = part_mode;
|
|
cu->tr_depth = depth;
|
|
cu->coded = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static void lcu_set_inter(lcu_t *lcu, int x_px, int y_px, int depth, cu_info *cur_cu)
|
|
{
|
|
const int width_cu = LCU_CU_WIDTH >> depth;
|
|
const int x_cu = SUB_SCU(x_px) >> MAX_DEPTH;
|
|
const int y_cu = SUB_SCU(y_px) >> MAX_DEPTH;
|
|
cu_info *const lcu_cu = &lcu->cu[LCU_CU_OFFSET];
|
|
int x, y;
|
|
// Set mode in every CU covered by part_mode in this depth.
|
|
for (y = y_cu; y < y_cu + width_cu; ++y) {
|
|
for (x = x_cu; x < x_cu + width_cu; ++x) {
|
|
cu_info *cu = &lcu_cu[x + y * LCU_T_CU_WIDTH];
|
|
//Check if this could be moved inside the if
|
|
cu->coded = 1;
|
|
if (cu != cur_cu) {
|
|
cu->depth = cur_cu->depth;
|
|
cu->type = CU_INTER;
|
|
cu->tr_depth = cur_cu->tr_depth;
|
|
cu->merged = cur_cu->merged;
|
|
cu->skipped = cur_cu->skipped;
|
|
memcpy(&cu->inter, &cur_cu->inter, sizeof(cu_info_inter));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static void lcu_set_coeff(lcu_t *lcu, int x_px, int y_px, int depth, cu_info *cur_cu)
|
|
{
|
|
const int width_cu = LCU_CU_WIDTH >> depth;
|
|
const int x_cu = SUB_SCU(x_px) >> MAX_DEPTH;
|
|
const int y_cu = SUB_SCU(y_px) >> MAX_DEPTH;
|
|
cu_info *const lcu_cu = &lcu->cu[LCU_CU_OFFSET];
|
|
int x, y;
|
|
int tr_split = cur_cu->tr_depth-cur_cu->depth;
|
|
|
|
// Set coeff flags in every CU covered by part_mode in this depth.
|
|
for (y = y_cu; y < y_cu + width_cu; ++y) {
|
|
for (x = x_cu; x < x_cu + width_cu; ++x) {
|
|
cu_info *cu = &lcu_cu[x + y * LCU_T_CU_WIDTH];
|
|
// Use TU top-left CU to propagate coeff flags
|
|
uint32_t mask = ~((width_cu>>tr_split)-1);
|
|
cu_info *cu_from = &lcu_cu[(x & mask) + (y & mask) * LCU_T_CU_WIDTH];
|
|
if (cu != cu_from) {
|
|
// Chroma coeff data is not used, luma is needed for deblocking
|
|
cu->cbf.y = cu_from->cbf.y;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* Update lcu to have best modes at this depth.
|
|
* \return Cost of best mode.
|
|
*/
|
|
static int search_cu_intra(encoder_state * const encoder_state,
|
|
const int x_px, const int y_px,
|
|
const int depth, lcu_t *lcu)
|
|
{
|
|
const picture * const cur_pic = encoder_state->cur_pic;
|
|
const vector2d lcu_px = { x_px & 0x3f, y_px & 0x3f };
|
|
const vector2d lcu_cu = { lcu_px.x >> 3, lcu_px.y >> 3 };
|
|
const int8_t cu_width = (LCU_WIDTH >> (depth));
|
|
const int cu_index = LCU_CU_OFFSET + lcu_cu.x + lcu_cu.y * LCU_T_CU_WIDTH;
|
|
|
|
cu_info *cur_cu = &lcu->cu[cu_index];
|
|
|
|
pixel rec_buffer[(LCU_WIDTH * 2 + 1) * (LCU_WIDTH * 2 + 1)];
|
|
pixel *cu_in_rec_buffer = &rec_buffer[cu_width * 2 + 8 + 1];
|
|
|
|
int8_t candidate_modes[3];
|
|
|
|
cu_info *left_cu = 0;
|
|
cu_info *above_cu = 0;
|
|
|
|
if ((x_px >> 3) > 0) {
|
|
left_cu = &lcu->cu[cu_index - 1];
|
|
}
|
|
// Don't take the above CU across the LCU boundary.
|
|
if ((y_px >> 3) > 0 && lcu_cu.y != 0) {
|
|
above_cu = &lcu->cu[cu_index - LCU_T_CU_WIDTH];
|
|
}
|
|
|
|
// Get intra predictors
|
|
intra_get_dir_luma_predictor(x_px, y_px, candidate_modes, cur_cu, left_cu, above_cu);
|
|
|
|
// Build reconstructed block to use in prediction with extrapolated borders
|
|
intra_build_reference_border(encoder_state->encoder_control, x_px, y_px, cu_width * 2 + 8,
|
|
rec_buffer, cu_width * 2 + 8, 0,
|
|
cur_pic->width,
|
|
cur_pic->height,
|
|
lcu);
|
|
|
|
// Find best intra mode for 2Nx2N.
|
|
{
|
|
uint32_t cost = UINT32_MAX;
|
|
int16_t mode = -1;
|
|
uint32_t bitcost = UINT32_MAX;
|
|
pixel *ref_pixels = &lcu->ref.y[lcu_px.x + lcu_px.y * LCU_WIDTH];
|
|
unsigned pu_index = PU_INDEX(x_px >> 2, y_px >> 2);
|
|
mode = intra_prediction(encoder_state,ref_pixels, LCU_WIDTH,
|
|
cu_in_rec_buffer, cu_width * 2 + 8, cu_width,
|
|
&cost, candidate_modes, &bitcost);
|
|
cur_cu->intra[pu_index].mode = (int8_t)mode;
|
|
cur_cu->intra[pu_index].cost = cost;
|
|
cur_cu->intra[pu_index].bitcost = bitcost;
|
|
}
|
|
|
|
return cur_cu->intra[PU_INDEX(x_px >> 2, y_px >> 2)].cost;
|
|
}
|
|
|
|
/**
|
|
* Calculate "final cost" for the block
|
|
* \return Cost of block
|
|
*
|
|
* Take SSD between reconstruction and original and add cost from
|
|
* coding (bitcost * lambda) and cost for coding coefficients (estimated
|
|
* here as (coefficient_sum * 1.5) * lambda)
|
|
*/
|
|
static int lcu_get_final_cost(const encoder_state * const encoder_state,
|
|
const int x_px, const int y_px,
|
|
const int depth, lcu_t *lcu)
|
|
{
|
|
cu_info *cur_cu;
|
|
int x_local = (x_px&0x3f), y_local = (y_px&0x3f);
|
|
int cost = 0;
|
|
int coeff_cost = 0;
|
|
const int rdo = encoder_state->encoder_control->rdo;
|
|
|
|
int width = LCU_WIDTH>>depth;
|
|
int x,y;
|
|
cur_cu = &lcu->cu[LCU_CU_OFFSET+(x_local>>3) + (y_local>>3)*LCU_T_CU_WIDTH];
|
|
|
|
// SSD between reconstruction and original
|
|
for (y = y_local; y < y_local+width; ++y) {
|
|
for (x = x_local; x < x_local+width; ++x) {
|
|
int diff = (int)lcu->rec.y[y * LCU_WIDTH + x] - (int)lcu->ref.y[y * LCU_WIDTH + x];
|
|
cost += diff*diff;
|
|
}
|
|
}
|
|
// Chroma SSD
|
|
for (y = y_local>>1; y < (y_local+width)>>1; ++y) {
|
|
for (x = x_local>>1; x < (x_local+width)>>1; ++x) {
|
|
int diff = (int)lcu->rec.u[y * (LCU_WIDTH>>1) + x] - (int)lcu->ref.u[y * (LCU_WIDTH>>1) + x];
|
|
cost += diff*diff;
|
|
diff = (int)lcu->rec.v[y * (LCU_WIDTH>>1) + x] - (int)lcu->ref.v[y * (LCU_WIDTH>>1) + x];
|
|
cost += diff*diff;
|
|
}
|
|
}
|
|
|
|
if(rdo == 1) {
|
|
// sum of coeffs
|
|
for (y = y_local; y < y_local+width; ++y) {
|
|
for (x = x_local; x < x_local+width; ++x) {
|
|
coeff_cost += abs((int)lcu->coeff.y[y * LCU_WIDTH + x]);
|
|
}
|
|
}
|
|
// Chroma sum of coeffs
|
|
for (y = y_local>>1; y < (y_local+width)>>1; ++y) {
|
|
for (x = x_local>>1; x < (x_local+width)>>1; ++x) {
|
|
coeff_cost += abs((int)lcu->coeff.u[y * (LCU_WIDTH>>1) + x]);
|
|
coeff_cost += abs((int)lcu->coeff.v[y * (LCU_WIDTH>>1) + x]);
|
|
}
|
|
}
|
|
// Coefficient costs
|
|
cost += (coeff_cost + (coeff_cost>>1)) * (int32_t)(encoder_state->cur_lambda_cost+0.5);
|
|
|
|
// Calculate actual bit costs for coding the coeffs
|
|
// RDO
|
|
} else if (rdo == 2) {
|
|
coefficient coeff_temp[32*32];
|
|
coefficient coeff_temp_u[16*16];
|
|
coefficient coeff_temp_v[16*16];
|
|
int i;
|
|
int blocks = (width == 64)?4:1;
|
|
int8_t luma_scan_mode = SCAN_DIAG;
|
|
int8_t chroma_scan_mode = SCAN_DIAG;
|
|
|
|
for(i = 0; i < blocks; i++) {
|
|
// For 64x64 blocks we need to do transform split to 32x32
|
|
int blk_y = i&2 ? 32:0 + y_local;
|
|
int blk_x = i&1 ? 32:0 + x_local;
|
|
int blockwidth = (width == 64)?32:width;
|
|
|
|
if (cur_cu->type == CU_INTRA) {
|
|
// Scan mode is diagonal, except for 4x4 and 8x8, where:
|
|
// - angular 6-14 = vertical
|
|
// - angular 22-30 = horizontal
|
|
int luma_mode = cur_cu->intra[i].mode;
|
|
int chroma_mode = cur_cu->intra[0].mode_chroma;
|
|
|
|
if (width <= 8) {
|
|
if (luma_mode >= 6 && luma_mode <= 14) {
|
|
luma_scan_mode = SCAN_VER;
|
|
} else if (luma_mode >= 22 && luma_mode <= 30) {
|
|
luma_scan_mode = SCAN_HOR;
|
|
}
|
|
|
|
if (chroma_mode >= 6 && chroma_mode <= 14) {
|
|
chroma_scan_mode = SCAN_VER;
|
|
} else if (chroma_mode >= 22 && chroma_mode <= 30) {
|
|
chroma_scan_mode = SCAN_HOR;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Calculate luma coeff bit count
|
|
picture_blit_coeffs(&lcu->coeff.y[(blk_y*LCU_WIDTH)+blk_x],coeff_temp,blockwidth,blockwidth,LCU_WIDTH,blockwidth);
|
|
coeff_cost += get_coeff_cost(encoder_state, coeff_temp, blockwidth, 0, luma_scan_mode);
|
|
|
|
blk_y >>= 1;
|
|
blk_x >>= 1;
|
|
if (blockwidth > 4) {
|
|
// Chroma is 1/4th of luma unless luma is 4x4.
|
|
blockwidth >>= 1;
|
|
} else if (x_px % 8 != 0 || y_px % 8 != 0) {
|
|
// Only add chroma cost for 4x4 blocks for the one on the 8x8 grid.
|
|
break;
|
|
}
|
|
|
|
picture_blit_coeffs(&lcu->coeff.u[(blk_y*(LCU_WIDTH>>1))+blk_x],coeff_temp_u,blockwidth,blockwidth,LCU_WIDTH>>1,blockwidth);
|
|
picture_blit_coeffs(&lcu->coeff.v[(blk_y*(LCU_WIDTH>>1))+blk_x],coeff_temp_v,blockwidth,blockwidth,LCU_WIDTH>>1,blockwidth);
|
|
|
|
coeff_cost += get_coeff_cost(encoder_state, coeff_temp_u, blockwidth, 2, chroma_scan_mode);
|
|
coeff_cost += get_coeff_cost(encoder_state, coeff_temp_v, blockwidth, 2, chroma_scan_mode);
|
|
}
|
|
// Multiply bit count with lambda to get RD-cost
|
|
cost += coeff_cost * (int32_t)(encoder_state->cur_lambda_cost+0.5);
|
|
}
|
|
|
|
// Bitcost
|
|
cost += (cur_cu->type == CU_INTER ? cur_cu->inter.bitcost : cur_cu->intra[PU_INDEX(x_px >> 2, y_px >> 2)].bitcost)*(int32_t)(encoder_state->cur_lambda_cost+0.5);
|
|
|
|
return cost;
|
|
}
|
|
|
|
/**
|
|
* Search every mode from 0 to MAX_PU_DEPTH and return cost of best mode.
|
|
* - The recursion is started at depth 0 and goes in Z-order to MAX_PU_DEPTH.
|
|
* - Data structure work_tree is maintained such that the neighbouring SCUs
|
|
* and pixels to the left and up of current CU are the final CUs decided
|
|
* via the search. This is done by copying the relevant data to all
|
|
* relevant levels whenever a decision is made whether to split or not.
|
|
* - All the final data for the LCU gets eventually copied to depth 0, which
|
|
* will be the final output of the recursion.
|
|
*/
|
|
static int search_cu(encoder_state * const encoder_state, int x, int y, int depth, lcu_t work_tree[MAX_PU_DEPTH])
|
|
{
|
|
const picture * const cur_pic = encoder_state->cur_pic;
|
|
int cu_width = LCU_WIDTH >> depth;
|
|
int cost = MAX_INT;
|
|
cu_info *cur_cu;
|
|
int x_local = (x&0x3f), y_local = (y&0x3f);
|
|
|
|
// Stop recursion if the CU is completely outside the frame.
|
|
if (x >= cur_pic->width || y >= cur_pic->height) {
|
|
// Return zero cost because this CU does not have to be coded.
|
|
return 0;
|
|
}
|
|
|
|
cur_cu = &(&work_tree[depth])->cu[LCU_CU_OFFSET+(x_local>>3) + (y_local>>3)*LCU_T_CU_WIDTH];
|
|
// Assign correct depth
|
|
cur_cu->depth = depth > MAX_DEPTH ? MAX_DEPTH : depth;
|
|
cur_cu->tr_depth = depth > 0 ? depth : 1;
|
|
cur_cu->type = CU_NOTSET;
|
|
cur_cu->part_size = depth > MAX_DEPTH ? SIZE_NxN : SIZE_2Nx2N;
|
|
// If the CU is completely inside the frame at this depth, search for
|
|
// prediction modes at this depth.
|
|
if (x + cu_width <= cur_pic->width &&
|
|
y + cu_width <= cur_pic->height)
|
|
{
|
|
|
|
if (cur_pic->slicetype != SLICE_I &&
|
|
depth >= MIN_INTER_SEARCH_DEPTH &&
|
|
depth <= MAX_INTER_SEARCH_DEPTH)
|
|
{
|
|
int mode_cost = search_cu_inter(encoder_state, x, y, depth, &work_tree[depth]);
|
|
if (mode_cost < cost) {
|
|
cost = mode_cost;
|
|
cur_cu->type = CU_INTER;
|
|
}
|
|
}
|
|
|
|
if (depth >= MIN_INTRA_SEARCH_DEPTH &&
|
|
depth <= MAX_INTRA_SEARCH_DEPTH)
|
|
{
|
|
int mode_cost = search_cu_intra(encoder_state, x, y, depth, &work_tree[depth]);
|
|
if (mode_cost < cost) {
|
|
cost = mode_cost;
|
|
cur_cu->type = CU_INTRA;
|
|
}
|
|
}
|
|
|
|
// Reconstruct best mode because we need the reconstructed pixels for
|
|
// mode search of adjacent CUs.
|
|
if (cur_cu->type == CU_INTRA) {
|
|
lcu_set_intra_mode(&work_tree[depth], x, y, depth, cur_cu->intra[PU_INDEX(x >> 2, y >> 2)].mode, cur_cu->part_size);
|
|
intra_recon_lcu(encoder_state, x, y, depth,&work_tree[depth], cur_pic->width, cur_pic->height);
|
|
} else if (cur_cu->type == CU_INTER) {
|
|
int cbf;
|
|
inter_recon_lcu(encoder_state, encoder_state->ref->pics[cur_cu->inter.mv_ref], x, y, LCU_WIDTH>>depth, cur_cu->inter.mv, &work_tree[depth]);
|
|
encode_transform_tree(encoder_state, x, y, depth, &work_tree[depth]);
|
|
|
|
cbf = cbf_is_set(cur_cu->cbf.y, depth) || cbf_is_set(cur_cu->cbf.u, depth) || cbf_is_set(cur_cu->cbf.v, depth);
|
|
|
|
if(cur_cu->merged && !cbf) {
|
|
cur_cu->merged = 0;
|
|
cur_cu->skipped = 1;
|
|
// Selecting skip reduces bits needed to code the CU
|
|
cur_cu->inter.bitcost--;
|
|
}
|
|
lcu_set_inter(&work_tree[depth], x, y, depth, cur_cu);
|
|
lcu_set_coeff(&work_tree[depth], x, y, depth, cur_cu);
|
|
}
|
|
}
|
|
if (cur_cu->type == CU_INTRA || cur_cu->type == CU_INTER) {
|
|
cost = lcu_get_final_cost(encoder_state, x, y, depth, &work_tree[depth]);
|
|
}
|
|
|
|
// Recursively split all the way to max search depth.
|
|
if (depth < MAX_INTRA_SEARCH_DEPTH || depth < MAX_INTER_SEARCH_DEPTH) {
|
|
int half_cu = cu_width / 2;
|
|
int split_cost = (int)(4.5 * encoder_state->cur_lambda_cost);
|
|
int cbf = cbf_is_set(cur_cu->cbf.y, depth) || cbf_is_set(cur_cu->cbf.u, depth) || cbf_is_set(cur_cu->cbf.v, depth);
|
|
|
|
// If skip mode was selected for the block, skip further search.
|
|
// Skip mode means there's no coefficients in the block, so splitting
|
|
// might not give any better results but takes more time to do.
|
|
if(cur_cu->type == CU_NOTSET || cbf) {
|
|
split_cost += search_cu(encoder_state, x, y, depth + 1, work_tree);
|
|
split_cost += search_cu(encoder_state, x + half_cu, y, depth + 1, work_tree);
|
|
split_cost += search_cu(encoder_state, x, y + half_cu, depth + 1, work_tree);
|
|
split_cost += search_cu(encoder_state, x + half_cu, y + half_cu, depth + 1, work_tree);
|
|
} else {
|
|
split_cost = INT_MAX;
|
|
}
|
|
if (split_cost < cost) {
|
|
// Copy split modes to this depth.
|
|
cost = split_cost;
|
|
work_tree_copy_up(x, y, depth, work_tree);
|
|
} else {
|
|
// Copy this CU's mode all the way down for use in adjacent CUs mode
|
|
// search.
|
|
work_tree_copy_down(x, y, depth, work_tree);
|
|
}
|
|
}
|
|
|
|
return cost;
|
|
}
|
|
|
|
|
|
/**
|
|
* Initialize lcu_t for search.
|
|
* - Copy reference CUs.
|
|
* - Copy reference pixels from neighbouring LCUs.
|
|
* - Copy reference pixels from this LCU.
|
|
*/
|
|
static void init_lcu_t(const encoder_state * const encoder_state, const int x, const int y, lcu_t *lcu, const yuv_t *hor_buf, const yuv_t *ver_buf)
|
|
{
|
|
const picture * const cur_pic = encoder_state->cur_pic;
|
|
|
|
// Copy reference cu_info structs from neighbouring LCUs.
|
|
{
|
|
const int x_cu = x >> MAX_DEPTH;
|
|
const int y_cu = y >> MAX_DEPTH;
|
|
const int cu_array_width = cur_pic->width_in_lcu << MAX_DEPTH;
|
|
cu_info *const cu_array = cur_pic->cu_array;
|
|
|
|
// Use top-left sub-cu of LCU as pointer to lcu->cu array to make things
|
|
// simpler.
|
|
cu_info *lcu_cu = &lcu->cu[LCU_CU_OFFSET];
|
|
|
|
// Copy top CU row.
|
|
if (y_cu > 0) {
|
|
int i;
|
|
for (i = 0; i < LCU_CU_WIDTH; ++i) {
|
|
const cu_info *from_cu = &cu_array[(x_cu + i) + (y_cu - 1) * cu_array_width];
|
|
cu_info *to_cu = &lcu_cu[i - LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
// Copy left CU column.
|
|
if (x_cu > 0) {
|
|
int i;
|
|
for (i = 0; i < LCU_CU_WIDTH; ++i) {
|
|
const cu_info *from_cu = &cu_array[(x_cu - 1) + (y_cu + i) * cu_array_width];
|
|
cu_info *to_cu = &lcu_cu[-1 + i * LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
// Copy top-left CU.
|
|
if (x_cu > 0 && y_cu > 0) {
|
|
const cu_info *from_cu = &cu_array[(x_cu - 1) + (y_cu - 1) * cu_array_width];
|
|
cu_info *to_cu = &lcu_cu[-1 - LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
|
|
// Copy top-right CU.
|
|
if (y_cu > 0 && x + LCU_WIDTH < cur_pic->width) {
|
|
const cu_info *from_cu = &cu_array[(x_cu + LCU_CU_WIDTH) + (y_cu - 1) * cu_array_width];
|
|
cu_info *to_cu = &lcu->cu[LCU_T_CU_WIDTH*LCU_T_CU_WIDTH];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
|
|
// Copy reference pixels.
|
|
{
|
|
const int pic_width = cur_pic->width;
|
|
|
|
// Copy top reference pixels.
|
|
if (y > 0) {
|
|
// hor_buf is of size pic_width so there might not be LCU_REF_PX_WIDTH
|
|
// number of allocated pixels left.
|
|
int x_max = MIN(LCU_REF_PX_WIDTH, pic_width - x);
|
|
memcpy(&lcu->top_ref.y[1], &hor_buf->y[x], x_max);
|
|
memcpy(&lcu->top_ref.u[1], &hor_buf->u[x / 2], x_max / 2);
|
|
memcpy(&lcu->top_ref.v[1], &hor_buf->v[x / 2], x_max / 2);
|
|
}
|
|
// Copy left reference pixels.
|
|
if (x > 0) {
|
|
memcpy(&lcu->left_ref.y[1], &ver_buf->y[1], LCU_WIDTH);
|
|
memcpy(&lcu->left_ref.u[1], &ver_buf->u[1], LCU_WIDTH / 2);
|
|
memcpy(&lcu->left_ref.v[1], &ver_buf->v[1], LCU_WIDTH / 2);
|
|
}
|
|
// Copy top-left reference pixel.
|
|
if (x > 0 && y > 0) {
|
|
lcu->top_ref.y[0] = ver_buf->y[0];
|
|
lcu->left_ref.y[0] = ver_buf->y[0];
|
|
|
|
lcu->top_ref.u[0] = ver_buf->u[0];
|
|
lcu->left_ref.u[0] = ver_buf->u[0];
|
|
|
|
lcu->top_ref.v[0] = ver_buf->v[0];
|
|
lcu->left_ref.v[0] = ver_buf->v[0];
|
|
}
|
|
}
|
|
|
|
// Copy LCU pixels.
|
|
{
|
|
const picture * const pic = encoder_state->cur_pic;
|
|
int pic_width = cur_pic->width;
|
|
int x_max = MIN(x + LCU_WIDTH, pic_width) - x;
|
|
int y_max = MIN(y + LCU_WIDTH, cur_pic->height) - y;
|
|
|
|
int x_c = x / 2;
|
|
int y_c = y / 2;
|
|
int pic_width_c = pic_width / 2;
|
|
int x_max_c = x_max / 2;
|
|
int y_max_c = y_max / 2;
|
|
|
|
picture_blit_pixels(&pic->y_data[x + y * pic_width], lcu->ref.y,
|
|
x_max, y_max, pic_width, LCU_WIDTH);
|
|
picture_blit_pixels(&pic->u_data[x_c + y_c * pic_width_c], lcu->ref.u,
|
|
x_max_c, y_max_c, pic_width_c, LCU_WIDTH / 2);
|
|
picture_blit_pixels(&pic->v_data[x_c + y_c * pic_width_c], lcu->ref.v,
|
|
x_max_c, y_max_c, pic_width_c, LCU_WIDTH / 2);
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* Copy CU and pixel data to it's place in picture datastructure.
|
|
*/
|
|
static void copy_lcu_to_cu_data(const encoder_state * const encoder_state, int x_px, int y_px, const lcu_t *lcu)
|
|
{
|
|
// Copy non-reference CUs to picture.
|
|
{
|
|
const int x_cu = x_px >> MAX_DEPTH;
|
|
const int y_cu = y_px >> MAX_DEPTH;
|
|
const picture * const cur_pic = encoder_state->cur_pic;
|
|
const int cu_array_width = cur_pic->width_in_lcu << MAX_DEPTH;
|
|
cu_info *const cu_array = cur_pic->cu_array;
|
|
|
|
// Use top-left sub-cu of LCU as pointer to lcu->cu array to make things
|
|
// simpler.
|
|
const cu_info *const lcu_cu = &lcu->cu[LCU_CU_OFFSET];
|
|
|
|
int x, y;
|
|
for (y = 0; y < LCU_CU_WIDTH; ++y) {
|
|
for (x = 0; x < LCU_CU_WIDTH; ++x) {
|
|
const cu_info *from_cu = &lcu_cu[x + y * LCU_T_CU_WIDTH];
|
|
cu_info *to_cu = &cu_array[(x_cu + x) + (y_cu + y) * cu_array_width];
|
|
memcpy(to_cu, from_cu, sizeof(*to_cu));
|
|
}
|
|
}
|
|
}
|
|
|
|
// Copy pixels to picture.
|
|
{
|
|
picture * const pic = encoder_state->cur_pic;
|
|
const int pic_width = pic->width;
|
|
const int x_max = MIN(x_px + LCU_WIDTH, pic_width) - x_px;
|
|
const int y_max = MIN(y_px + LCU_WIDTH, pic->height) - y_px;
|
|
const int luma_index = x_px + y_px * pic_width;
|
|
const int chroma_index = (x_px / 2) + (y_px / 2) * (pic_width / 2);
|
|
|
|
picture_blit_pixels(lcu->rec.y, &pic->y_recdata[luma_index],
|
|
x_max, y_max, LCU_WIDTH, pic_width);
|
|
picture_blit_coeffs(lcu->coeff.y, &pic->coeff_y[luma_index],
|
|
x_max, y_max, LCU_WIDTH, pic_width);
|
|
|
|
picture_blit_pixels(lcu->rec.u, &pic->u_recdata[chroma_index],
|
|
x_max / 2, y_max / 2, LCU_WIDTH / 2, pic_width / 2);
|
|
picture_blit_pixels(lcu->rec.v, &pic->v_recdata[chroma_index],
|
|
x_max / 2, y_max / 2, LCU_WIDTH / 2, pic_width / 2);
|
|
picture_blit_coeffs(lcu->coeff.u, &pic->coeff_u[chroma_index],
|
|
x_max / 2, y_max / 2, LCU_WIDTH / 2, pic_width / 2);
|
|
picture_blit_coeffs(lcu->coeff.v, &pic->coeff_v[chroma_index],
|
|
x_max / 2, y_max / 2, LCU_WIDTH / 2, pic_width / 2);
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
* Search LCU for modes.
|
|
* - Best mode gets copied to current picture.
|
|
*/
|
|
void search_lcu(encoder_state * const encoder_state, int x, int y, yuv_t* hor_buf, yuv_t* ver_buf)
|
|
{
|
|
lcu_t work_tree[MAX_PU_DEPTH + 1];
|
|
int depth;
|
|
// Initialize work tree.
|
|
for (depth = 0; depth <= MAX_PU_DEPTH; ++depth) {
|
|
memset(&work_tree[depth], 0, sizeof(work_tree[depth]));
|
|
init_lcu_t(encoder_state, x, y, &work_tree[depth], hor_buf, ver_buf);
|
|
}
|
|
|
|
// Start search from depth 0.
|
|
search_cu(encoder_state, x, y, 0, work_tree);
|
|
|
|
copy_lcu_to_cu_data(encoder_state, x, y, &work_tree[0]);
|
|
}
|