/*
|
* Copyright 2020 Rockchip Electronics Co. LTD
|
*
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
* you may not use this file except in compliance with the License.
|
* You may obtain a copy of the License at
|
*
|
* http://www.apache.org/licenses/LICENSE-2.0
|
*
|
* Unless required by applicable law or agreed to in writing, software
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
* See the License for the specific language governing permissions and
|
* limitations under the License.
|
*/
|
|
#define MODULE_TAG "hal_h265d_rkv"
|
|
#include <stdio.h>
|
#include <string.h>
|
|
#include "mpp_mem.h"
|
#include "mpp_bitread.h"
|
#include "mpp_bitput.h"
|
|
#include "hal_h265d_debug.h"
|
#include "hal_h265d_ctx.h"
|
#include "hal_h265d_com.h"
|
#include "hal_h265d_rkv.h"
|
#include "hal_h265d_rkv_reg.h"
|
#include "h265d_syntax.h"
|
|
/* #define dump */
|
#ifdef dump
|
static FILE *fp = NULL;
|
#endif
|
#define HW_RPS
|
|
#define PPS_SIZE (96 * 64)
|
|
static MPP_RET hal_h265d_alloc_res(void *hal)
|
{
|
RK_S32 i = 0;
|
RK_S32 ret = 0;
|
HalH265dCtx *reg_cxt = (HalH265dCtx *)hal;
|
if (reg_cxt->fast_mode) {
|
for (i = 0; i < MAX_GEN_REG; i++) {
|
reg_cxt->g_buf[i].hw_regs =
|
mpp_calloc_size(void, sizeof(H265d_REGS_t));
|
ret = mpp_buffer_get(reg_cxt->group,
|
®_cxt->g_buf[i].scaling_list_data,
|
SCALING_LIST_SIZE);
|
if (ret) {
|
mpp_err("h265d scaling_list_data get buffer failed\n");
|
return ret;
|
}
|
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->g_buf[i].pps_data,
|
PPS_SIZE);
|
if (ret) {
|
mpp_err("h265d pps_data get buffer failed\n");
|
return ret;
|
}
|
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->g_buf[i].rps_data,
|
RPS_SIZE);
|
if (ret) {
|
mpp_err("h265d rps_data get buffer failed\n");
|
return ret;
|
}
|
}
|
} else {
|
reg_cxt->hw_regs = mpp_calloc_size(void, sizeof(H265d_REGS_t));
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->scaling_list_data,
|
SCALING_LIST_SIZE);
|
if (ret) {
|
mpp_err("h265d scaling_list_data get buffer failed\n");
|
return ret;
|
}
|
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->pps_data, PPS_SIZE);
|
if (ret) {
|
mpp_err("h265d pps_data get buffer failed\n");
|
return ret;
|
}
|
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->rps_data, RPS_SIZE);
|
if (ret) {
|
mpp_err("h265d rps_data get buffer failed\n");
|
return ret;
|
}
|
|
}
|
return MPP_OK;
|
}
|
|
static MPP_RET hal_h265d_release_res(void *hal)
|
{
|
RK_S32 ret = 0;
|
HalH265dCtx *reg_cxt = ( HalH265dCtx *)hal;
|
RK_S32 i = 0;
|
if (reg_cxt->fast_mode) {
|
for (i = 0; i < MAX_GEN_REG; i++) {
|
if (reg_cxt->g_buf[i].scaling_list_data) {
|
ret = mpp_buffer_put(reg_cxt->g_buf[i].scaling_list_data);
|
if (ret) {
|
mpp_err("h265d scaling_list_data free buffer failed\n");
|
return ret;
|
}
|
}
|
if (reg_cxt->g_buf[i].pps_data) {
|
ret = mpp_buffer_put(reg_cxt->g_buf[i].pps_data);
|
if (ret) {
|
mpp_err("h265d pps_data free buffer failed\n");
|
return ret;
|
}
|
}
|
|
if (reg_cxt->g_buf[i].rps_data) {
|
ret = mpp_buffer_put(reg_cxt->g_buf[i].rps_data);
|
if (ret) {
|
mpp_err("h265d rps_data free buffer failed\n");
|
return ret;
|
}
|
}
|
|
if (reg_cxt->g_buf[i].hw_regs) {
|
mpp_free(reg_cxt->g_buf[i].hw_regs);
|
reg_cxt->g_buf[i].hw_regs = NULL;
|
}
|
}
|
} else {
|
if (reg_cxt->scaling_list_data) {
|
ret = mpp_buffer_put(reg_cxt->scaling_list_data);
|
if (ret) {
|
mpp_err("h265d scaling_list_data free buffer failed\n");
|
return ret;
|
}
|
}
|
if (reg_cxt->pps_data) {
|
ret = mpp_buffer_put(reg_cxt->pps_data);
|
if (ret) {
|
mpp_err("h265d pps_data free buffer failed\n");
|
return ret;
|
}
|
}
|
|
if (reg_cxt->rps_data) {
|
ret = mpp_buffer_put(reg_cxt->rps_data);
|
if (ret) {
|
mpp_err("h265d rps_data free buffer failed\n");
|
return ret;
|
}
|
}
|
|
if (reg_cxt->hw_regs) {
|
mpp_free(reg_cxt->hw_regs);
|
reg_cxt->hw_regs = NULL;
|
}
|
}
|
return MPP_OK;
|
}
|
|
MPP_RET hal_h265d_rkv_init(void *hal, MppHalCfg *cfg)
|
{
|
MPP_RET ret = MPP_NOK;
|
HalH265dCtx *reg_cxt = (HalH265dCtx *)hal;
|
|
mpp_slots_set_prop(reg_cxt->slots, SLOTS_HOR_ALIGN, hevc_hor_align);
|
mpp_slots_set_prop(reg_cxt->slots, SLOTS_VER_ALIGN, hevc_ver_align);
|
|
reg_cxt->scaling_qm = mpp_calloc(DXVA_Qmatrix_HEVC, 1);
|
reg_cxt->sw_rps_buf = mpp_calloc(RK_U64, 400);
|
|
if (reg_cxt->scaling_qm == NULL) {
|
mpp_err("scaling_org alloc fail");
|
return MPP_ERR_MALLOC;
|
}
|
|
reg_cxt->scaling_rk = mpp_calloc(scalingFactor_t, 1);
|
if (reg_cxt->scaling_rk == NULL) {
|
mpp_err("scaling_rk alloc fail");
|
return MPP_ERR_MALLOC;
|
}
|
|
if (reg_cxt->group == NULL) {
|
ret = mpp_buffer_group_get_internal(®_cxt->group, MPP_BUFFER_TYPE_ION);
|
if (ret) {
|
mpp_err("h265d mpp_buffer_group_get failed\n");
|
return ret;
|
}
|
}
|
|
ret = mpp_buffer_get(reg_cxt->group, ®_cxt->cabac_table_data, sizeof(cabac_table));
|
if (ret) {
|
mpp_err("h265d cabac_table get buffer failed\n");
|
return ret;
|
}
|
|
ret = mpp_buffer_write(reg_cxt->cabac_table_data, 0, (void*)cabac_table, sizeof(cabac_table));
|
if (ret) {
|
mpp_err("h265d write cabac_table data failed\n");
|
return ret;
|
}
|
|
ret = hal_h265d_alloc_res(hal);
|
if (ret) {
|
mpp_err("hal_h265d_alloc_res failed\n");
|
return ret;
|
}
|
|
|
{
|
// report hw_info to parser
|
const MppSocInfo *info = mpp_get_soc_info();
|
const void *hw_info = NULL;
|
RK_U32 i = 0;
|
|
for (i = 0; i < MPP_ARRAY_ELEMS(info->dec_caps); i++) {
|
if (info->dec_caps[i] && ( info->dec_caps[i]->type == VPU_CLIENT_RKVDEC ||
|
info->dec_caps[i]->type == VPU_CLIENT_HEVC_DEC)) {
|
hw_info = info->dec_caps[i];
|
break;
|
}
|
}
|
|
mpp_assert(hw_info);
|
cfg->hw_info = hw_info;
|
}
|
|
#ifdef dump
|
fp = fopen("/data/hal.bin", "wb");
|
#endif
|
(void) cfg;
|
return MPP_OK;
|
}
|
|
MPP_RET hal_h265d_rkv_deinit(void *hal)
|
{
|
RK_S32 ret = 0;
|
HalH265dCtx *reg_cxt = (HalH265dCtx *)hal;
|
|
ret = mpp_buffer_put(reg_cxt->cabac_table_data);
|
if (ret) {
|
mpp_err("h265d cabac_table free buffer failed\n");
|
return ret;
|
}
|
|
if (reg_cxt->scaling_qm) {
|
mpp_free(reg_cxt->scaling_qm);
|
}
|
|
if (reg_cxt->sw_rps_buf) {
|
mpp_free(reg_cxt->sw_rps_buf);
|
}
|
|
if (reg_cxt->scaling_rk) {
|
mpp_free(reg_cxt->scaling_rk);
|
}
|
|
hal_h265d_release_res(hal);
|
|
if (reg_cxt->group) {
|
ret = mpp_buffer_group_put(reg_cxt->group);
|
if (ret) {
|
mpp_err("h265d group free buffer failed\n");
|
return ret;
|
}
|
}
|
return MPP_OK;
|
}
|
|
static RK_S32 hal_h265d_v345_output_pps_packet(void *hal, void *dxva)
|
{
|
RK_S32 fifo_len = 12;
|
RK_S32 i, j;
|
RK_U32 log2_min_cb_size;
|
RK_S32 width, height;
|
HalH265dCtx *reg_cxt = ( HalH265dCtx *)hal;
|
h265d_dxva2_picture_context_t *dxva_cxt = (h265d_dxva2_picture_context_t*)dxva;
|
BitputCtx_t bp;
|
RK_U64 *pps_packet = mpp_calloc(RK_U64, fifo_len + 1);
|
|
if (NULL == reg_cxt || dxva_cxt == NULL) {
|
mpp_err("%s:%s:%d reg_cxt or dxva_cxt is NULL",
|
__FILE__, __FUNCTION__, __LINE__);
|
MPP_FREE(pps_packet);
|
return MPP_ERR_NULL_PTR;
|
}
|
|
void *pps_ptr = mpp_buffer_get_ptr(reg_cxt->pps_data);
|
if (NULL == pps_ptr) {
|
mpp_err("pps_data get ptr error");
|
return MPP_ERR_NOMEM;
|
}
|
memset(pps_ptr, 0, 96 * 64);
|
// pps_packet = (RK_U64 *)(pps_ptr + dxva_cxt->pp.pps_id * 80);
|
|
for (i = 0; i < 12; i++) pps_packet[i] = 0;
|
|
mpp_set_bitput_ctx(&bp, pps_packet, fifo_len);
|
|
// SPS
|
mpp_put_bits(&bp, dxva_cxt->pp.vps_id , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_id , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.chroma_format_idc , 2);
|
|
log2_min_cb_size = dxva_cxt->pp.log2_min_luma_coding_block_size_minus3 + 3;
|
width = (dxva_cxt->pp.PicWidthInMinCbsY << log2_min_cb_size);
|
height = (dxva_cxt->pp.PicHeightInMinCbsY << log2_min_cb_size);
|
|
mpp_put_bits(&bp, width , 13);
|
mpp_put_bits(&bp, height , 13);
|
mpp_put_bits(&bp, dxva_cxt->pp.bit_depth_luma_minus8 + 8 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.bit_depth_chroma_minus8 + 8 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_max_pic_order_cnt_lsb_minus4 + 4 , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size , 2); //log2_maxa_coding_block_depth
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_min_luma_coding_block_size_minus3 + 3 , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_min_transform_block_size_minus2 + 2 , 3);
|
///<-zrh comment ^ 57 bit above
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_transform_block_size , 2);
|
mpp_put_bits(&bp, dxva_cxt->pp.max_transform_hierarchy_depth_inter , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.max_transform_hierarchy_depth_intra , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.scaling_list_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.amp_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.sample_adaptive_offset_enabled_flag , 1);
|
///<-zrh comment ^ 68 bit above
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.pcm_sample_bit_depth_luma_minus1 + 1) : 0 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.pcm_sample_bit_depth_chroma_minus1 + 1) : 0 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_loop_filter_disabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_pcm_luma_coding_block_size , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.log2_min_pcm_luma_coding_block_size_minus3 + 3) : 0, 3);
|
|
mpp_put_bits(&bp, dxva_cxt->pp.num_short_term_ref_pic_sets , 7);
|
mpp_put_bits(&bp, dxva_cxt->pp.long_term_ref_pics_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_long_term_ref_pics_sps , 6);
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_temporal_mvp_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.strong_intra_smoothing_enabled_flag , 1);
|
///<-zrh comment ^ 100 bit above
|
|
mpp_put_bits(&bp, 0 , 7 );
|
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_max_dec_pic_buffering_minus1, 4);
|
mpp_put_bits(&bp, 0, 3);
|
mpp_put_align(&bp , 32, 0xf); //128
|
|
// PPS
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_id , 6 );
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_id , 4 );
|
mpp_put_bits(&bp, dxva_cxt->pp.dependent_slice_segments_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.output_flag_present_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.num_extra_slice_header_bits , 13);
|
mpp_put_bits(&bp, dxva_cxt->pp.sign_data_hiding_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.cabac_init_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_ref_idx_l0_default_active_minus1 + 1 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_ref_idx_l1_default_active_minus1 + 1 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.init_qp_minus26 , 7);
|
mpp_put_bits(&bp, dxva_cxt->pp.constrained_intra_pred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.transform_skip_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.cu_qp_delta_enabled_flag , 1); //164
|
|
mpp_put_bits(&bp, log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size -
|
dxva_cxt->pp.diff_cu_qp_delta_depth , 3);
|
|
h265h_dbg(H265H_DBG_PPS, "log2_min_cb_size %d %d %d \n", log2_min_cb_size,
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size, dxva_cxt->pp.diff_cu_qp_delta_depth );
|
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_cb_qp_offset , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_cr_qp_offset , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_slice_chroma_qp_offsets_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.weighted_pred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.weighted_bipred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.transquant_bypass_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.tiles_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.entropy_coding_sync_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_loop_filter_across_slices_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.loop_filter_across_tiles_enabled_flag , 1); //185
|
mpp_put_bits(&bp, dxva_cxt->pp.deblocking_filter_override_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_deblocking_filter_disabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_beta_offset_div2 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_tc_offset_div2 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.lists_modification_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_parallel_merge_level_minus2 + 2 , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.slice_segment_header_extension_present_flag , 1);
|
mpp_put_bits(&bp, 0 , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.tiles_enabled_flag ? dxva_cxt->pp.num_tile_columns_minus1 + 1 : 0, 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.tiles_enabled_flag ? dxva_cxt->pp.num_tile_rows_minus1 + 1 : 0 , 5 );
|
mpp_put_bits(&bp, 0, 4); //mSps_Pps[i]->mMode
|
mpp_put_align(&bp, 64, 0xf);
|
{
|
/// tiles info begin
|
RK_U16 column_width[20];
|
RK_U16 row_height[22];
|
|
memset(column_width, 0, sizeof(column_width));
|
memset(row_height, 0, sizeof(row_height));
|
|
if (dxva_cxt->pp.tiles_enabled_flag) {
|
|
if (dxva_cxt->pp.uniform_spacing_flag == 0) {
|
RK_S32 maxcuwidth = dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size + log2_min_cb_size;
|
RK_S32 ctu_width_in_pic = (width +
|
(1 << maxcuwidth) - 1) / (1 << maxcuwidth) ;
|
RK_S32 ctu_height_in_pic = (height +
|
(1 << maxcuwidth) - 1) / (1 << maxcuwidth) ;
|
RK_S32 sum = 0;
|
for (i = 0; i < dxva_cxt->pp.num_tile_columns_minus1; i++) {
|
column_width[i] = dxva_cxt->pp.column_width_minus1[i] + 1;
|
sum += column_width[i] ;
|
}
|
column_width[i] = ctu_width_in_pic - sum;
|
|
sum = 0;
|
for (i = 0; i < dxva_cxt->pp.num_tile_rows_minus1; i++) {
|
row_height[i] = dxva_cxt->pp.row_height_minus1[i] + 1;
|
sum += row_height[i];
|
}
|
row_height[i] = ctu_height_in_pic - sum;
|
} // end of (pps->uniform_spacing_flag == 0)
|
else {
|
|
RK_S32 pic_in_cts_width = (width +
|
(1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size)) - 1)
|
/ (1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size));
|
RK_S32 pic_in_cts_height = (height +
|
(1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size)) - 1)
|
/ (1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size));
|
|
for (i = 0; i < dxva_cxt->pp.num_tile_columns_minus1 + 1; i++)
|
column_width[i] = ((i + 1) * pic_in_cts_width) / (dxva_cxt->pp.num_tile_columns_minus1 + 1) -
|
(i * pic_in_cts_width) / (dxva_cxt->pp.num_tile_columns_minus1 + 1);
|
|
for (i = 0; i < dxva_cxt->pp.num_tile_rows_minus1 + 1; i++)
|
row_height[i] = ((i + 1) * pic_in_cts_height) / (dxva_cxt->pp.num_tile_rows_minus1 + 1) -
|
(i * pic_in_cts_height) / (dxva_cxt->pp.num_tile_rows_minus1 + 1);
|
}
|
} // pps->tiles_enabled_flag
|
else {
|
RK_S32 MaxCUWidth = (1 << (dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size + log2_min_cb_size));
|
column_width[0] = (width + MaxCUWidth - 1) / MaxCUWidth;
|
row_height[0] = (height + MaxCUWidth - 1) / MaxCUWidth;
|
}
|
|
for (j = 0; j < 20; j++) {
|
if (column_width[j] > 0)
|
column_width[j]--;
|
mpp_put_bits(&bp, column_width[j], 8);
|
}
|
|
for (j = 0; j < 22; j++) {
|
if (row_height[j] > 0)
|
row_height[j]--;
|
mpp_put_bits(&bp, row_height[j], 8);
|
}
|
}
|
|
{
|
RK_U8 *ptr_scaling = (RK_U8 *)mpp_buffer_get_ptr(reg_cxt->scaling_list_data);
|
RK_U32 fd = mpp_buffer_get_fd(reg_cxt->scaling_list_data);
|
|
hal_h265d_output_scalinglist_packet(hal, ptr_scaling, dxva);
|
mpp_put_bits(&bp, fd, 32);
|
mpp_put_bits(&bp, 0, 60);
|
mpp_put_align(&bp, 128, 0xf);
|
}
|
|
for (i = 0; i < 64; i++)
|
memcpy(pps_ptr + i * 96, pps_packet, 96);
|
|
#ifdef dump
|
fwrite(pps_ptr, 1, 80 * 64, fp);
|
RK_U32 *tmp = (RK_U32 *)pps_ptr;
|
for (i = 0; i < 96 / 4; i++) {
|
h265h_dbg(H265H_DBG_PPS, "pps[%3d] = 0x%08x\n", i, tmp[i]);
|
}
|
#endif
|
MPP_FREE(pps_packet);
|
return 0;
|
}
|
|
static RK_S32 hal_h265d_output_pps_packet(void *hal, void *dxva)
|
{
|
RK_S32 fifo_len = 10;
|
RK_S32 i, j;
|
RK_U32 log2_min_cb_size;
|
RK_S32 width, height;
|
HalH265dCtx *reg_cxt = ( HalH265dCtx *)hal;
|
h265d_dxva2_picture_context_t *dxva_cxt = (h265d_dxva2_picture_context_t*)dxva;
|
BitputCtx_t bp;
|
RK_U64 *pps_packet = mpp_calloc(RK_U64, fifo_len + 1);
|
|
if (NULL == reg_cxt || dxva_cxt == NULL) {
|
mpp_err("%s:%s:%d reg_cxt or dxva_cxt is NULL",
|
__FILE__, __FUNCTION__, __LINE__);
|
MPP_FREE(pps_packet);
|
return MPP_ERR_NULL_PTR;
|
}
|
|
void *pps_ptr = mpp_buffer_get_ptr(reg_cxt->pps_data);
|
if (NULL == pps_ptr) {
|
mpp_err("pps_data get ptr error");
|
return MPP_ERR_NOMEM;
|
}
|
memset(pps_ptr, 0, 80 * 64);
|
// pps_packet = (RK_U64 *)(pps_ptr + dxva_cxt->pp.pps_id * 80);
|
|
for (i = 0; i < 10; i++) pps_packet[i] = 0;
|
|
mpp_set_bitput_ctx(&bp, pps_packet, fifo_len);
|
|
// SPS
|
mpp_put_bits(&bp, dxva_cxt->pp.vps_id , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_id , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.chroma_format_idc , 2);
|
|
log2_min_cb_size = dxva_cxt->pp.log2_min_luma_coding_block_size_minus3 + 3;
|
width = (dxva_cxt->pp.PicWidthInMinCbsY << log2_min_cb_size);
|
height = (dxva_cxt->pp.PicHeightInMinCbsY << log2_min_cb_size);
|
|
mpp_put_bits(&bp, width , 13);
|
mpp_put_bits(&bp, height , 13);
|
mpp_put_bits(&bp, dxva_cxt->pp.bit_depth_luma_minus8 + 8 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.bit_depth_chroma_minus8 + 8 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_max_pic_order_cnt_lsb_minus4 + 4 , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size , 2); //log2_maxa_coding_block_depth
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_min_luma_coding_block_size_minus3 + 3 , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_min_transform_block_size_minus2 + 2 , 3);
|
///<-zrh comment ^ 57 bit above
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_transform_block_size , 2);
|
mpp_put_bits(&bp, dxva_cxt->pp.max_transform_hierarchy_depth_inter , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.max_transform_hierarchy_depth_intra , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.scaling_list_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.amp_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.sample_adaptive_offset_enabled_flag , 1);
|
///<-zrh comment ^ 68 bit above
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.pcm_sample_bit_depth_luma_minus1 + 1) : 0 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.pcm_sample_bit_depth_chroma_minus1 + 1) : 0 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_loop_filter_disabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_diff_max_min_pcm_luma_coding_block_size , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.pcm_enabled_flag ? (dxva_cxt->pp.log2_min_pcm_luma_coding_block_size_minus3 + 3) : 0, 3);
|
|
mpp_put_bits(&bp, dxva_cxt->pp.num_short_term_ref_pic_sets , 7);
|
mpp_put_bits(&bp, dxva_cxt->pp.long_term_ref_pics_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_long_term_ref_pics_sps , 6);
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_temporal_mvp_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.strong_intra_smoothing_enabled_flag , 1);
|
///<-zrh comment ^ 100 bit above
|
|
mpp_put_bits(&bp, 0 , 7 );
|
mpp_put_align(&bp , 32, 0xf);
|
|
// PPS
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_id , 6 );
|
mpp_put_bits(&bp, dxva_cxt->pp.sps_id , 4 );
|
mpp_put_bits(&bp, dxva_cxt->pp.dependent_slice_segments_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.output_flag_present_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.num_extra_slice_header_bits , 13);
|
mpp_put_bits(&bp, dxva_cxt->pp.sign_data_hiding_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.cabac_init_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_ref_idx_l0_default_active_minus1 + 1 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_ref_idx_l1_default_active_minus1 + 1 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.init_qp_minus26 , 7);
|
mpp_put_bits(&bp, dxva_cxt->pp.constrained_intra_pred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.transform_skip_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.cu_qp_delta_enabled_flag , 1);
|
|
mpp_put_bits(&bp, log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size -
|
dxva_cxt->pp.diff_cu_qp_delta_depth , 3);
|
|
h265h_dbg(H265H_DBG_PPS, "log2_min_cb_size %d %d %d \n", log2_min_cb_size,
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size, dxva_cxt->pp.diff_cu_qp_delta_depth );
|
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_cb_qp_offset , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_cr_qp_offset , 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_slice_chroma_qp_offsets_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.weighted_pred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.weighted_bipred_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.transquant_bypass_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.tiles_enabled_flag , 1 );
|
mpp_put_bits(&bp, dxva_cxt->pp.entropy_coding_sync_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_loop_filter_across_slices_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.loop_filter_across_tiles_enabled_flag , 1);
|
|
mpp_put_bits(&bp, dxva_cxt->pp.deblocking_filter_override_enabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_deblocking_filter_disabled_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_beta_offset_div2 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.pps_tc_offset_div2 , 4);
|
mpp_put_bits(&bp, dxva_cxt->pp.lists_modification_present_flag , 1);
|
mpp_put_bits(&bp, dxva_cxt->pp.log2_parallel_merge_level_minus2 + 2 , 3);
|
/*slice_segment_header_extension_present_flag need set 0 */
|
mpp_put_bits(&bp, 0 , 1);
|
mpp_put_bits(&bp, 0 , 3);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_tile_columns_minus1 + 1, 5);
|
mpp_put_bits(&bp, dxva_cxt->pp.num_tile_rows_minus1 + 1 , 5 );
|
mpp_put_bits(&bp, 3, 2); //mSps_Pps[i]->mMode
|
mpp_put_align(&bp, 64, 0xf);
|
|
{
|
/// tiles info begin
|
RK_U16 column_width[20];
|
RK_U16 row_height[22];
|
|
memset(column_width, 0, sizeof(column_width));
|
memset(row_height, 0, sizeof(row_height));
|
|
if (dxva_cxt->pp.tiles_enabled_flag) {
|
if (dxva_cxt->pp.uniform_spacing_flag == 0) {
|
RK_S32 maxcuwidth = dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size + log2_min_cb_size;
|
RK_S32 ctu_width_in_pic = (width +
|
(1 << maxcuwidth) - 1) / (1 << maxcuwidth) ;
|
RK_S32 ctu_height_in_pic = (height +
|
(1 << maxcuwidth) - 1) / (1 << maxcuwidth) ;
|
RK_S32 sum = 0;
|
for (i = 0; i < dxva_cxt->pp.num_tile_columns_minus1; i++) {
|
column_width[i] = dxva_cxt->pp.column_width_minus1[i] + 1;
|
sum += column_width[i] ;
|
}
|
column_width[i] = ctu_width_in_pic - sum;
|
|
sum = 0;
|
for (i = 0; i < dxva_cxt->pp.num_tile_rows_minus1; i++) {
|
row_height[i] = dxva_cxt->pp.row_height_minus1[i] + 1;
|
sum += row_height[i];
|
}
|
row_height[i] = ctu_height_in_pic - sum;
|
} // end of (pps->uniform_spacing_flag == 0)
|
else {
|
|
RK_S32 pic_in_cts_width = (width +
|
(1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size)) - 1)
|
/ (1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size));
|
RK_S32 pic_in_cts_height = (height +
|
(1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size)) - 1)
|
/ (1 << (log2_min_cb_size +
|
dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size));
|
|
for (i = 0; i < dxva_cxt->pp.num_tile_columns_minus1 + 1; i++)
|
column_width[i] = ((i + 1) * pic_in_cts_width) / (dxva_cxt->pp.num_tile_columns_minus1 + 1) -
|
(i * pic_in_cts_width) / (dxva_cxt->pp.num_tile_columns_minus1 + 1);
|
|
for (i = 0; i < dxva_cxt->pp.num_tile_rows_minus1 + 1; i++)
|
row_height[i] = ((i + 1) * pic_in_cts_height) / (dxva_cxt->pp.num_tile_rows_minus1 + 1) -
|
(i * pic_in_cts_height) / (dxva_cxt->pp.num_tile_rows_minus1 + 1);
|
}
|
} // pps->tiles_enabled_flag
|
else {
|
RK_S32 MaxCUWidth = (1 << (dxva_cxt->pp.log2_diff_max_min_luma_coding_block_size + log2_min_cb_size));
|
column_width[0] = (width + MaxCUWidth - 1) / MaxCUWidth;
|
row_height[0] = (height + MaxCUWidth - 1) / MaxCUWidth;
|
}
|
|
for (j = 0; j < 20; j++) {
|
if (column_width[j] > 0)
|
column_width[j]--;
|
mpp_put_bits(&bp, column_width[j], 8);
|
}
|
|
for (j = 0; j < 22; j++) {
|
if (row_height[j] > 0)
|
row_height[j]--;
|
mpp_put_bits(&bp, row_height[j], 8);
|
}
|
}
|
|
{
|
RK_U8 *ptr_scaling = (RK_U8 *)mpp_buffer_get_ptr(reg_cxt->scaling_list_data);
|
RK_U32 fd = mpp_buffer_get_fd(reg_cxt->scaling_list_data);
|
|
hal_h265d_output_scalinglist_packet(hal, ptr_scaling, dxva);
|
mpp_put_bits(&bp, fd, 32);
|
mpp_put_align(&bp, 64, 0xf);
|
}
|
|
for (i = 0; i < 64; i++)
|
memcpy(pps_ptr + i * 80, pps_packet, 80);
|
|
#ifdef dump
|
fwrite(pps_ptr, 1, 80 * 64, fp);
|
fflush(fp);
|
#endif
|
|
MPP_FREE(pps_packet);
|
return 0;
|
}
|
|
static void update_stream_buffer(MppBuffer streambuf, HalTaskInfo *syn)
|
{
|
h265d_dxva2_picture_context_t *dxva_cxt =
|
(h265d_dxva2_picture_context_t *)syn->dec.syntax.data;
|
RK_U8 *ptr = (RK_U8*)mpp_buffer_get_ptr(streambuf);
|
RK_U8 bit_left = 0;
|
RK_U16 start_byte, end_byte, i = 0;
|
RK_U32 stream_size = dxva_cxt->bitstream_size;
|
RK_U8 *buf = NULL;
|
RK_U8 *temp = NULL;
|
RK_U32 cut_byte = 0, cut_byte_acc = 0;
|
|
for (i = 0; i < dxva_cxt->slice_count; i++) {
|
if (dxva_cxt->slice_cut_param[i].is_enable) {
|
|
bit_left = 8 - (dxva_cxt->slice_cut_param[i].start_bit & 0x7);
|
start_byte = dxva_cxt->slice_cut_param[i].start_bit >> 3;
|
end_byte = (dxva_cxt->slice_cut_param[i].end_bit + 7) >> 3;
|
buf = ptr + (dxva_cxt->slice_short[i].BSNALunitDataLocation - cut_byte_acc);
|
temp = buf + start_byte;
|
|
h265h_dbg(H265H_DBG_FUNCTION, "start bit %d start byte[%d] 0x%x end bit %d end byte[%d] 0x%x\n",
|
dxva_cxt->slice_cut_param[i].start_bit, start_byte, buf[start_byte],
|
dxva_cxt->slice_cut_param[i].end_bit, end_byte, buf[end_byte]);
|
if (bit_left < 8) {
|
*temp = (*temp >> bit_left) << bit_left;
|
*temp |= 1 << (bit_left - 1);
|
} else {
|
*temp = 0x80;
|
}
|
if ((dxva_cxt->slice_cut_param[i].end_bit & 0x7) == 0 && buf[end_byte] == 0x80)
|
end_byte += 1;
|
|
h265h_dbg(H265H_DBG_FUNCTION, "i %d location %d count %d SliceBytesInBuffer %d bitstream_size %d\n",
|
i, dxva_cxt->slice_short[i].BSNALunitDataLocation, dxva_cxt->slice_count,
|
dxva_cxt->slice_short[i].SliceBytesInBuffer, dxva_cxt->bitstream_size);
|
|
memcpy(buf + start_byte + 1, buf + end_byte,
|
stream_size - dxva_cxt->slice_short[i].BSNALunitDataLocation - end_byte);
|
|
cut_byte = end_byte - start_byte - 1;
|
dxva_cxt->slice_short[i].SliceBytesInBuffer -= cut_byte;
|
dxva_cxt->bitstream_size -= cut_byte;
|
cut_byte_acc += cut_byte;
|
}
|
}
|
}
|
|
MPP_RET hal_h265d_rkv_gen_regs(void *hal, HalTaskInfo *syn)
|
{
|
RK_S32 i = 0;
|
RK_S32 log2_min_cb_size;
|
RK_S32 width, height;
|
RK_S32 stride_y, stride_uv, virstrid_y, virstrid_yuv;
|
H265d_REGS_t *hw_regs;
|
RK_S32 ret = MPP_SUCCESS;
|
MppBuffer streambuf = NULL;
|
RK_S32 aglin_offset = 0;
|
RK_S32 valid_ref = -1;
|
MppBuffer framebuf = NULL;
|
RK_U32 sw_ref_valid = 0;
|
|
if (syn->dec.flags.parse_err ||
|
syn->dec.flags.ref_err) {
|
h265h_dbg(H265H_DBG_TASK_ERR, "%s found task error\n", __FUNCTION__);
|
return MPP_OK;
|
}
|
|
h265d_dxva2_picture_context_t *dxva_cxt =
|
(h265d_dxva2_picture_context_t *)syn->dec.syntax.data;
|
HalH265dCtx *reg_cxt = ( HalH265dCtx *)hal;
|
|
void *rps_ptr = NULL;
|
if (reg_cxt ->fast_mode) {
|
for (i = 0; i < MAX_GEN_REG; i++) {
|
if (!reg_cxt->g_buf[i].use_flag) {
|
syn->dec.reg_index = i;
|
reg_cxt->rps_data = reg_cxt->g_buf[i].rps_data;
|
reg_cxt->scaling_list_data =
|
reg_cxt->g_buf[i].scaling_list_data;
|
reg_cxt->pps_data = reg_cxt->g_buf[i].pps_data;
|
reg_cxt->hw_regs = reg_cxt->g_buf[i].hw_regs;
|
reg_cxt->g_buf[i].use_flag = 1;
|
break;
|
}
|
}
|
if (i == MAX_GEN_REG) {
|
mpp_err("hevc rps buf all used");
|
return MPP_ERR_NOMEM;
|
}
|
}
|
rps_ptr = mpp_buffer_get_ptr(reg_cxt->rps_data);
|
if (NULL == rps_ptr) {
|
|
mpp_err("rps_data get ptr error");
|
return MPP_ERR_NOMEM;
|
}
|
|
|
if (syn->dec.syntax.data == NULL) {
|
mpp_err("%s:%s:%d dxva is NULL", __FILE__, __FUNCTION__, __LINE__);
|
return MPP_ERR_NULL_PTR;
|
}
|
|
/* output pps */
|
if (reg_cxt->is_v345) {
|
hal_h265d_v345_output_pps_packet(hal, syn->dec.syntax.data);
|
} else {
|
hal_h265d_output_pps_packet(hal, syn->dec.syntax.data);
|
}
|
|
if (NULL == reg_cxt->hw_regs) {
|
return MPP_ERR_NULL_PTR;
|
}
|
|
hw_regs = (H265d_REGS_t*)reg_cxt->hw_regs;
|
memset(hw_regs, 0, sizeof(H265d_REGS_t));
|
|
log2_min_cb_size = dxva_cxt->pp.log2_min_luma_coding_block_size_minus3 + 3;
|
|
width = (dxva_cxt->pp.PicWidthInMinCbsY << log2_min_cb_size);
|
height = (dxva_cxt->pp.PicHeightInMinCbsY << log2_min_cb_size);
|
|
stride_y = ((MPP_ALIGN(width, 64)
|
* (dxva_cxt->pp.bit_depth_luma_minus8 + 8)) >> 3);
|
stride_uv = ((MPP_ALIGN(width, 64)
|
* (dxva_cxt->pp.bit_depth_chroma_minus8 + 8)) >> 3);
|
|
stride_y = hevc_hor_align(stride_y);
|
stride_uv = hevc_hor_align(stride_uv);
|
virstrid_y = hevc_ver_align(height) * stride_y;
|
virstrid_yuv = virstrid_y + stride_uv * hevc_ver_align(height) / 2;
|
|
hw_regs->sw_picparameter.sw_slice_num = dxva_cxt->slice_count;
|
hw_regs->sw_picparameter.sw_y_hor_virstride = stride_y >> 4;
|
hw_regs->sw_picparameter.sw_uv_hor_virstride = stride_uv >> 4;
|
hw_regs->sw_y_virstride = virstrid_y >> 4;
|
hw_regs->sw_yuv_virstride = virstrid_yuv >> 4;
|
hw_regs->sw_sysctrl.sw_h26x_rps_mode = 0;
|
mpp_buf_slot_get_prop(reg_cxt->slots, dxva_cxt->pp.CurrPic.Index7Bits,
|
SLOT_BUFFER, &framebuf);
|
hw_regs->sw_decout_base = mpp_buffer_get_fd(framebuf); //just index need map
|
|
/*if out_base is equal to zero it means this frame may error
|
we return directly add by csy*/
|
|
if (hw_regs->sw_decout_base == 0) {
|
return 0;
|
}
|
|
hw_regs->sw_cur_poc = dxva_cxt->pp.CurrPicOrderCntVal;
|
|
mpp_buf_slot_get_prop(reg_cxt->packet_slots, syn->dec.input, SLOT_BUFFER,
|
&streambuf);
|
|
if ( dxva_cxt->bitstream == NULL) {
|
dxva_cxt->bitstream = mpp_buffer_get_ptr(streambuf);
|
}
|
if (reg_cxt->is_v345) {
|
#ifdef HW_RPS
|
hw_regs->sw_sysctrl.sw_wait_reset_en = 1;
|
hw_regs->v345_reg_ends.reg064_mvc0.refp_layer_same_with_cur = 0xffff;
|
hal_h265d_slice_hw_rps(syn->dec.syntax.data, rps_ptr, reg_cxt->sw_rps_buf, reg_cxt->fast_mode);
|
#else
|
hw_regs->sw_sysctrl.sw_h26x_rps_mode = 1;
|
hal_h265d_slice_output_rps(syn->dec.syntax.data, rps_ptr);
|
#endif
|
} else {
|
hal_h265d_slice_output_rps(syn->dec.syntax.data, rps_ptr);
|
}
|
|
if (dxva_cxt->pp.slice_segment_header_extension_present_flag && !reg_cxt->is_v345) {
|
update_stream_buffer(streambuf, syn);
|
}
|
|
hw_regs->sw_cabactbl_base = mpp_buffer_get_fd(reg_cxt->cabac_table_data);
|
hw_regs->sw_pps_base = mpp_buffer_get_fd(reg_cxt->pps_data);
|
hw_regs->sw_rps_base = mpp_buffer_get_fd(reg_cxt->rps_data);
|
hw_regs->sw_strm_rlc_base = mpp_buffer_get_fd(streambuf);
|
hw_regs->sw_stream_len = ((dxva_cxt->bitstream_size + 15)
|
& (~15)) + 64;
|
aglin_offset = hw_regs->sw_stream_len - dxva_cxt->bitstream_size;
|
if (aglin_offset > 0) {
|
memset((void *)(dxva_cxt->bitstream + dxva_cxt->bitstream_size), 0,
|
aglin_offset);
|
}
|
hw_regs->sw_interrupt.sw_dec_e = 1;
|
hw_regs->sw_interrupt.sw_dec_timeout_e = 1;
|
hw_regs->sw_interrupt.sw_wr_ddr_align_en = dxva_cxt->pp.tiles_enabled_flag
|
? 0 : 1;
|
|
///find s->rps_model[i] position, and set register
|
hw_regs->cabac_error_en = 0xfdfffffd;
|
hw_regs->rkv_reg_ends.extern_error_en = 0x30000000;
|
|
valid_ref = hw_regs->sw_decout_base;
|
for (i = 0; i < (RK_S32)MPP_ARRAY_ELEMS(dxva_cxt->pp.RefPicList); i++) {
|
if (dxva_cxt->pp.RefPicList[i].bPicEntry != 0xff &&
|
dxva_cxt->pp.RefPicList[i].bPicEntry != 0x7f) {
|
hw_regs->sw_refer_poc[i] = dxva_cxt->pp.PicOrderCntValList[i];
|
mpp_buf_slot_get_prop(reg_cxt->slots,
|
dxva_cxt->pp.RefPicList[i].Index7Bits,
|
SLOT_BUFFER, &framebuf);
|
if (framebuf != NULL) {
|
hw_regs->sw_refer_base[i] = mpp_buffer_get_fd(framebuf);
|
valid_ref = hw_regs->sw_refer_base[i];
|
} else {
|
hw_regs->sw_refer_base[i] = valid_ref;
|
}
|
sw_ref_valid |= (1 << i);
|
} else {
|
hw_regs->sw_refer_base[i] = hw_regs->sw_decout_base;
|
}
|
}
|
|
if (sw_ref_valid) {
|
mpp_dev_set_reg_offset(reg_cxt->dev, 10, sw_ref_valid & 0xf);
|
mpp_dev_set_reg_offset(reg_cxt->dev, 11, ((sw_ref_valid >> 4) & 0xf));
|
mpp_dev_set_reg_offset(reg_cxt->dev, 12, ((sw_ref_valid >> 8) & 0xf));
|
mpp_dev_set_reg_offset(reg_cxt->dev, 13, ((sw_ref_valid >> 12) & 0xf));
|
}
|
|
return ret;
|
}
|
|
MPP_RET hal_h265d_rkv_start(void *hal, HalTaskInfo *task)
|
{
|
MPP_RET ret = MPP_OK;
|
H265d_REGS_t *hw_regs = NULL;
|
HalH265dCtx *reg_cxt = (HalH265dCtx *)hal;
|
RK_S32 index = task->dec.reg_index;
|
|
RK_U32 i;
|
|
if (task->dec.flags.parse_err ||
|
task->dec.flags.ref_err) {
|
h265h_dbg(H265H_DBG_TASK_ERR, "%s found task error\n", __FUNCTION__);
|
return MPP_OK;
|
}
|
|
if (reg_cxt->fast_mode) {
|
hw_regs = ( H265d_REGS_t *)reg_cxt->g_buf[index].hw_regs;
|
} else {
|
hw_regs = ( H265d_REGS_t *)reg_cxt->hw_regs;
|
}
|
|
if (hw_regs == NULL) {
|
mpp_err("hal_h265d_start hw_regs is NULL");
|
return MPP_ERR_NULL_PTR;
|
}
|
|
do {
|
MppDevRegWrCfg wr_cfg;
|
MppDevRegRdCfg rd_cfg;
|
RK_U32 reg_size = (reg_cxt->is_v345) ? V345_HEVC_REGISTERS :
|
(reg_cxt->client_type == VPU_CLIENT_RKVDEC) ?
|
RKVDEC_V1_REGISTERS : RKVDEC_HEVC_REGISTERS;
|
|
reg_size *= sizeof(RK_U32);
|
|
wr_cfg.reg = hw_regs;
|
wr_cfg.size = reg_size;
|
wr_cfg.offset = 0;
|
|
if (hal_h265d_debug & H265H_DBG_REG) {
|
for (i = 0; i < reg_size / sizeof(RK_U32); i++)
|
h265h_dbg(H265H_DBG_REG, "RK_HEVC_DEC: regs[%02d]=%08X\n", i, ((RK_U32 *)hw_regs)[i]);
|
}
|
|
ret = mpp_dev_ioctl(reg_cxt->dev, MPP_DEV_REG_WR, &wr_cfg);
|
if (ret) {
|
mpp_err_f("set register write failed %d\n", ret);
|
break;
|
}
|
|
rd_cfg.reg = hw_regs;
|
rd_cfg.size = reg_size;
|
rd_cfg.offset = 0;
|
|
ret = mpp_dev_ioctl(reg_cxt->dev, MPP_DEV_REG_RD, &rd_cfg);
|
if (ret) {
|
mpp_err_f("set register read failed %d\n", ret);
|
break;
|
}
|
|
ret = mpp_dev_ioctl(reg_cxt->dev, MPP_DEV_CMD_SEND, NULL);
|
if (ret) {
|
mpp_err_f("send cmd failed %d\n", ret);
|
break;
|
}
|
} while (0);
|
|
return ret;
|
}
|
|
MPP_RET hal_h265d_rkv_wait(void *hal, HalTaskInfo *task)
|
{
|
MPP_RET ret = MPP_OK;
|
RK_S32 index = task->dec.reg_index;
|
HalH265dCtx *reg_cxt = (HalH265dCtx *)hal;
|
H265d_REGS_t *hw_regs = NULL;
|
RK_S32 i;
|
|
if (task->dec.flags.parse_err ||
|
task->dec.flags.ref_err) {
|
h265h_dbg(H265H_DBG_TASK_ERR, "%s found task error\n", __FUNCTION__);
|
goto ERR_PROC;
|
}
|
|
if (reg_cxt->fast_mode) {
|
hw_regs = ( H265d_REGS_t *)reg_cxt->g_buf[index].hw_regs;
|
} else {
|
hw_regs = ( H265d_REGS_t *)reg_cxt->hw_regs;
|
}
|
|
ret = mpp_dev_ioctl(reg_cxt->dev, MPP_DEV_CMD_POLL, NULL);
|
if (ret)
|
mpp_err_f("poll cmd failed %d\n", ret);
|
|
ERR_PROC:
|
if (task->dec.flags.parse_err ||
|
task->dec.flags.ref_err ||
|
hw_regs->sw_interrupt.sw_dec_error_sta ||
|
hw_regs->sw_interrupt.sw_dec_timeout_sta ||
|
hw_regs->sw_interrupt.sw_dec_empty_sta) {
|
if (!reg_cxt->fast_mode) {
|
if (reg_cxt->dec_cb)
|
mpp_callback(reg_cxt->dec_cb, &task->dec);
|
} else {
|
MppFrame mframe = NULL;
|
mpp_buf_slot_get_prop(reg_cxt->slots, task->dec.output,
|
SLOT_FRAME_PTR, &mframe);
|
if (mframe) {
|
reg_cxt->fast_mode_err_found = 1;
|
mpp_frame_set_errinfo(mframe, 1);
|
}
|
}
|
} else {
|
if (reg_cxt->fast_mode && reg_cxt->fast_mode_err_found) {
|
for (i = 0; i < (RK_S32)MPP_ARRAY_ELEMS(task->dec.refer); i++) {
|
if (task->dec.refer[i] >= 0) {
|
MppFrame frame_ref = NULL;
|
|
mpp_buf_slot_get_prop(reg_cxt->slots, task->dec.refer[i],
|
SLOT_FRAME_PTR, &frame_ref);
|
h265h_dbg(H265H_DBG_FAST_ERR, "refer[%d] %d frame %p\n",
|
i, task->dec.refer[i], frame_ref);
|
if (frame_ref && mpp_frame_get_errinfo(frame_ref)) {
|
MppFrame frame_out = NULL;
|
mpp_buf_slot_get_prop(reg_cxt->slots, task->dec.output,
|
SLOT_FRAME_PTR, &frame_out);
|
mpp_frame_set_errinfo(frame_out, 1);
|
break;
|
}
|
}
|
}
|
}
|
}
|
|
if (hal_h265d_debug & H265H_DBG_REG) {
|
h265h_dbg(H265H_DBG_REG, "RK_HEVC_DEC: regs[1]=0x%08X, regs[45]=0x%08x\n", ((RK_U32 *)hw_regs)[1], ((RK_U32 *)hw_regs)[45]);
|
}
|
|
if (reg_cxt->fast_mode) {
|
reg_cxt->g_buf[index].use_flag = 0;
|
}
|
|
return ret;
|
}
|
|
MPP_RET hal_h265d_rkv_reset(void *hal)
|
{
|
MPP_RET ret = MPP_OK;
|
HalH265dCtx *p_hal = (HalH265dCtx *)hal;
|
|
p_hal->fast_mode_err_found = 0;
|
|
return ret;
|
}
|
|
const MppHalApi hal_h265d_rkv = {
|
.name = "h265d_rkdec",
|
.type = MPP_CTX_DEC,
|
.coding = MPP_VIDEO_CodingHEVC,
|
.ctx_size = sizeof(HalH265dCtx),
|
.flag = 0,
|
.init = hal_h265d_rkv_init,
|
.deinit = hal_h265d_rkv_deinit,
|
.reg_gen = hal_h265d_rkv_gen_regs,
|
.start = hal_h265d_rkv_start,
|
.wait = hal_h265d_rkv_wait,
|
.reset = hal_h265d_rkv_reset,
|
.flush = NULL,
|
.control = NULL,
|
};
|