From 271195f85bbce284ac80ed31c62fba9b7e74e99d Mon Sep 17 00:00:00 2001 From: Shivraj Patil Date: Tue, 9 Jun 2015 20:50:26 +0530 Subject: [PATCH] avcodec/mips: MSA (MIPS-SIMD-Arch) optimizations for HEVC loop filter and sao functions This patch adds MSA (MIPS-SIMD-Arch) optimizations for HEVC loop filter and sao functions in new file hevc_lpf_sao_msa.c Adds new generic macros (needed for this patch) in libavutil/mips/generic_macros_msa.h In this patch, in comparision with previous patch, duplicated c functions are removed. Signed-off-by: Shivraj Patil Signed-off-by: Michael Niedermayer --- libavcodec/mips/Makefile | 3 +- libavcodec/mips/hevc_lpf_sao_msa.c | 2088 +++++++++++++++++++++++++++ libavcodec/mips/hevcdsp_init_mips.c | 26 + libavcodec/mips/hevcdsp_mips.h | 30 + libavutil/mips/generic_macros_msa.h | 111 +- 5 files changed, 2256 insertions(+), 2 deletions(-) create mode 100644 libavcodec/mips/hevc_lpf_sao_msa.c diff --git a/libavcodec/mips/Makefile b/libavcodec/mips/Makefile index 463072a76d..b8bb1fef2e 100644 --- a/libavcodec/mips/Makefile +++ b/libavcodec/mips/Makefile @@ -26,7 +26,8 @@ MSA-OBJS-$(CONFIG_HEVC_DECODER) += mips/hevcdsp_msa.o \ mips/hevc_mc_uniw_msa.o \ mips/hevc_mc_bi_msa.o \ mips/hevc_mc_biw_msa.o \ - mips/hevc_idct_msa.o + mips/hevc_idct_msa.o \ + mips/hevc_lpf_sao_msa.o MSA-OBJS-$(CONFIG_H264DSP) += mips/h264dsp_msa.o LOONGSON3-OBJS-$(CONFIG_H264DSP) += mips/h264dsp_mmi.o LOONGSON3-OBJS-$(CONFIG_H264CHROMA) += mips/h264chroma_mmi.o diff --git a/libavcodec/mips/hevc_lpf_sao_msa.c b/libavcodec/mips/hevc_lpf_sao_msa.c new file mode 100644 index 0000000000..da1db51ef5 --- /dev/null +++ b/libavcodec/mips/hevc_lpf_sao_msa.c @@ -0,0 +1,2088 @@ +/* + * Copyright (c) 2015 Manojkumar Bhosale (Manojkumar.Bhosale@imgtec.com) + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "libavutil/mips/generic_macros_msa.h" +#include "libavcodec/mips/hevcdsp_mips.h" + +static void hevc_loopfilter_luma_hor_msa(uint8_t *src, int32_t stride, + int32_t beta, int32_t *tc, + uint8_t *p_is_pcm, uint8_t *q_is_pcm) +{ + uint8_t *p3 = src - (stride << 2); + uint8_t *p2 = src - ((stride << 1) + stride); + uint8_t *p1 = src - (stride << 1); + uint8_t *p0 = src - stride; + uint8_t *q0 = src; + uint8_t *q1 = src + stride; + uint8_t *q2 = src + (stride << 1); + uint8_t *q3 = src + (stride << 1) + stride; + uint8_t flag0, flag1; + int32_t dp00, dq00, dp30, dq30, d00, d30; + int32_t dp04, dq04, dp34, dq34, d04, d34; + int32_t tc0, p_is_pcm0, q_is_pcm0, beta30, beta20, tc250; + int32_t tc4, p_is_pcm4, q_is_pcm4, tc254, tmp; + uint64_t dst_val0, dst_val1; + v16u8 dst0, dst1, dst2, dst3, dst4, dst5; + v2i64 cmp0, cmp1, cmp2, p_is_pcm_vec, q_is_pcm_vec; + v8u16 temp0, temp1; + v8i16 temp2; + v8i16 tc_pos, tc_neg; + v8i16 diff0, diff1, delta0, delta1, delta2, abs_delta0; + v16i8 zero = { 0 }; + v8u16 p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, q2_src, q3_src; + + dp00 = abs(p2[0] - (p1[0] << 1) + p0[0]); + dq00 = abs(q2[0] - (q1[0] << 1) + q0[0]); + dp30 = abs(p2[3] - (p1[3] << 1) + p0[3]); + dq30 = abs(q2[3] - (q1[3] << 1) + q0[3]); + d00 = dp00 + dq00; + d30 = dp30 + dq30; + p_is_pcm0 = p_is_pcm[0]; + q_is_pcm0 = q_is_pcm[0]; + dp04 = abs(p2[4] - (p1[4] << 1) + p0[4]); + dq04 = abs(q2[4] - (q1[4] << 1) + q0[4]); + dp34 = abs(p2[7] - (p1[7] << 1) + p0[7]); + dq34 = abs(q2[7] - (q1[7] << 1) + q0[7]); + d04 = dp04 + dq04; + d34 = dp34 + dq34; + p_is_pcm4 = p_is_pcm[1]; + q_is_pcm4 = q_is_pcm[1]; + + if (!p_is_pcm0 || !p_is_pcm4 || !q_is_pcm0 || !q_is_pcm4) { + if (!(d00 + d30 >= beta) || !(d04 + d34 >= beta)) { + p3_src = LD_UH(p3); + p2_src = LD_UH(p2); + p1_src = LD_UH(p1); + p0_src = LD_UH(p0); + q0_src = LD_UH(q0); + q1_src = LD_UH(q1); + q2_src = LD_UH(q2); + q3_src = LD_UH(q3); + + tc0 = tc[0]; + beta30 = beta >> 3; + beta20 = beta >> 2; + tc250 = ((tc0 * 5 + 1) >> 1); + tc4 = tc[1]; + tc254 = ((tc4 * 5 + 1) >> 1); + + flag0 = (abs(p3[0] - p0[0]) + abs(q3[0] - q0[0]) < beta30 && + abs(p0[0] - q0[0]) < tc250 && + abs(p3[3] - p0[3]) + abs(q3[3] - q0[3]) < beta30 && + abs(p0[3] - q0[3]) < tc250 && + (d00 << 1) < beta20 && (d30 << 1) < beta20); + cmp0 = __msa_fill_d(flag0); + + flag1 = (abs(p3[4] - p0[4]) + abs(q3[4] - q0[4]) < beta30 && + abs(p0[4] - q0[4]) < tc254 && + abs(p3[7] - p0[7]) + abs(q3[7] - q0[7]) < beta30 && + abs(p0[7] - q0[7]) < tc254 && + (d04 << 1) < beta20 && (d34 << 1) < beta20); + cmp1 = __msa_fill_d(flag1); + cmp2 = __msa_ilvev_d(cmp1, cmp0); + cmp2 = __msa_ceqi_d(cmp2, 0); + + ILVR_B8_UH(zero, p3_src, zero, p2_src, zero, p1_src, zero, p0_src, + zero, q0_src, zero, q1_src, zero, q2_src, zero, q3_src, + p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, q2_src, + q3_src); + + cmp0 = (v2i64) __msa_fill_h(tc0); + cmp1 = (v2i64) __msa_fill_h(tc4); + tc_pos = (v8i16) __msa_ilvev_d(cmp1, cmp0); + tc_pos <<= 1; + tc_neg = -tc_pos; + + temp0 = (p1_src + p0_src + q0_src); + temp1 = ((p3_src + p2_src) << 1) + p2_src + temp0; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - p2_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst0 = (v16u8) (temp2 + (v8i16) p2_src); + + temp1 = temp0 + p2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 2); + temp2 = (v8i16) (temp1 - p1_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst1 = (v16u8) (temp2 + (v8i16) p1_src); + + temp1 = (temp0 << 1) + p2_src + q1_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - p0_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst2 = (v16u8) (temp2 + (v8i16) p0_src); + + cmp0 = __msa_fill_d(p_is_pcm0); + cmp1 = __msa_fill_d(p_is_pcm4); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + dst0 = __msa_bmz_v(dst0, (v16u8) p2_src, (v16u8) p_is_pcm_vec); + dst1 = __msa_bmz_v(dst1, (v16u8) p1_src, (v16u8) p_is_pcm_vec); + dst2 = __msa_bmz_v(dst2, (v16u8) p0_src, (v16u8) p_is_pcm_vec); + + temp0 = (q1_src + p0_src + q0_src); + + temp1 = ((q3_src + q2_src) << 1) + q2_src + temp0; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - q2_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst5 = (v16u8) (temp2 + (v8i16) q2_src); + + temp1 = temp0 + q2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 2); + temp2 = (v8i16) (temp1 - q1_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst4 = (v16u8) (temp2 + (v8i16) q1_src); + + temp1 = (temp0 << 1) + p1_src + q2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - q0_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst3 = (v16u8) (temp2 + (v8i16) q0_src); + + cmp0 = __msa_fill_d(q_is_pcm0); + cmp1 = __msa_fill_d(q_is_pcm4); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + dst3 = __msa_bmz_v(dst3, (v16u8) q0_src, (v16u8) q_is_pcm_vec); + dst4 = __msa_bmz_v(dst4, (v16u8) q1_src, (v16u8) q_is_pcm_vec); + dst5 = __msa_bmz_v(dst5, (v16u8) q2_src, (v16u8) q_is_pcm_vec); + + tc_pos >>= 1; + tc_neg = -tc_pos; + + diff0 = (v8i16) (q0_src - p0_src); + diff1 = (v8i16) (q1_src - p1_src); + diff0 = (diff0 << 3) + diff0; + diff1 = (diff1 << 1) + diff1; + delta0 = diff0 - diff1; + delta0 = __msa_srari_h(delta0, 4); + + temp1 = (v8u16) ((tc_pos << 3) + (tc_pos << 1)); + abs_delta0 = __msa_add_a_h(delta0, (v8i16) zero); + abs_delta0 = (v8u16) abs_delta0 < temp1; + + delta0 = CLIP_SH(delta0, tc_neg, tc_pos); + + temp0 = (v8u16) (delta0 + p0_src); + temp0 = (v8u16) CLIP_SH_0_255(temp0); + temp0 = (v8u16) __msa_bmz_v((v16u8) temp0, (v16u8) p0_src, + (v16u8) p_is_pcm_vec); + + temp2 = (v8i16) (q0_src - delta0); + temp2 = CLIP_SH_0_255(temp2); + temp2 = (v8i16) __msa_bmz_v((v16u8) temp2, (v16u8) q0_src, + (v16u8) q_is_pcm_vec); + + tmp = (beta + (beta >> 1)) >> 3; + cmp0 = __msa_fill_d(!p_is_pcm0 && ((dp00 + dp30) < tmp)); + cmp1 = __msa_fill_d(!p_is_pcm4 && ((dp04 + dp34) < tmp)); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + cmp0 = (v2i64) __msa_fill_h((!q_is_pcm0) && (dq00 + dq30 < tmp)); + cmp1 = (v2i64) __msa_fill_h((!q_is_pcm4) && (dq04 + dq34 < tmp)); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + tc_pos >>= 1; + tc_neg = -tc_pos; + + delta1 = (v8i16) __msa_aver_u_h(p2_src, p0_src); + delta1 -= (v8i16) p1_src; + delta1 += delta0; + delta1 >>= 1; + delta1 = CLIP_SH(delta1, tc_neg, tc_pos); + delta1 = (v8i16) p1_src + (v8i16) delta1; + delta1 = CLIP_SH_0_255(delta1); + delta1 = (v8i16) __msa_bmnz_v((v16u8) delta1, (v16u8) p1_src, + (v16u8) p_is_pcm_vec); + + delta2 = (v8i16) __msa_aver_u_h(q0_src, q2_src); + delta2 = delta2 - (v8i16) q1_src; + delta2 = delta2 - delta0; + delta2 = delta2 >> 1; + delta2 = CLIP_SH(delta2, tc_neg, tc_pos); + delta2 = (v8i16) q1_src + (v8i16) delta2; + delta2 = CLIP_SH_0_255(delta2); + delta2 = (v8i16) __msa_bmnz_v((v16u8) delta2, (v16u8) q1_src, + (v16u8) q_is_pcm_vec); + + delta1 = (v8i16) __msa_bmz_v((v16u8) delta1, (v16u8) p1_src, + (v16u8) abs_delta0); + temp0 = (v8u16) __msa_bmz_v((v16u8) temp0, (v16u8) p0_src, + (v16u8) abs_delta0); + temp2 = (v8i16) __msa_bmz_v((v16u8) temp2, (v16u8) q0_src, + (v16u8) abs_delta0); + delta2 = (v8i16) __msa_bmz_v((v16u8) delta2, (v16u8) q1_src, + (v16u8) abs_delta0); + + dst2 = __msa_bmnz_v(dst2, (v16u8) temp0, (v16u8) cmp2); + dst3 = __msa_bmnz_v(dst3, (v16u8) temp2, (v16u8) cmp2); + dst1 = __msa_bmnz_v(dst1, (v16u8) delta1, (v16u8) cmp2); + dst4 = __msa_bmnz_v(dst4, (v16u8) delta2, (v16u8) cmp2); + dst0 = __msa_bmnz_v(dst0, (v16u8) p2_src, (v16u8) cmp2); + dst5 = __msa_bmnz_v(dst5, (v16u8) q2_src, (v16u8) cmp2); + + cmp0 = __msa_fill_d(d00 + d30 >= beta); + cmp1 = __msa_fill_d(d04 + d34 >= beta); + cmp0 = __msa_ilvev_d(cmp1, cmp0); + cmp0 = __msa_ceqi_d(cmp0, 0); + + dst0 = __msa_bmz_v(dst0, (v16u8) p2_src, (v16u8) cmp0); + dst1 = __msa_bmz_v(dst1, (v16u8) p1_src, (v16u8) cmp0); + dst2 = __msa_bmz_v(dst2, (v16u8) p0_src, (v16u8) cmp0); + dst3 = __msa_bmz_v(dst3, (v16u8) q0_src, (v16u8) cmp0); + dst4 = __msa_bmz_v(dst4, (v16u8) q1_src, (v16u8) cmp0); + dst5 = __msa_bmz_v(dst5, (v16u8) q2_src, (v16u8) cmp0); + + PCKEV_B2_UB(dst1, dst0, dst3, dst2, dst0, dst1); + dst2 = (v16u8) __msa_pckev_b((v16i8) dst5, (v16i8) dst4); + + dst_val0 = __msa_copy_u_d((v2i64) dst2, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst2, 1); + + ST8x4_UB(dst0, dst1, p2, stride); + p2 += (4 * stride); + SD(dst_val0, p2); + p2 += stride; + SD(dst_val1, p2); + } + } +} + +static void hevc_loopfilter_luma_ver_msa(uint8_t *src, int32_t stride, + int32_t beta, int32_t *tc, + uint8_t *p_is_pcm, uint8_t *q_is_pcm) +{ + uint8_t *p3 = src; + uint8_t *p2 = src + 3 * stride; + uint8_t *p1 = src + (stride << 2); + uint8_t *p0 = src + 7 * stride; + uint8_t flag0, flag1; + uint16_t tmp0, tmp1; + uint32_t tmp2, tmp3; + int32_t dp00, dq00, dp30, dq30, d00, d30; + int32_t dp04, dq04, dp34, dq34, d04, d34; + int32_t tc0, p_is_pcm0, q_is_pcm0, beta30, beta20, tc250; + int32_t tc4, p_is_pcm4, q_is_pcm4, tc254, tmp; + v16u8 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; + v2i64 cmp0, cmp1, cmp2, p_is_pcm_vec, q_is_pcm_vec; + v8u16 temp0, temp1; + v8i16 temp2; + v8i16 tc_pos, tc_neg; + v8i16 diff0, diff1, delta0, delta1, delta2, abs_delta0; + v16i8 zero = { 0 }; + v8u16 p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, q2_src, q3_src; + + dp00 = abs(p3[-3] - (p3[-2] << 1) + p3[-1]); + dq00 = abs(p3[2] - (p3[1] << 1) + p3[0]); + dp30 = abs(p2[-3] - (p2[-2] << 1) + p2[-1]); + dq30 = abs(p2[2] - (p2[1] << 1) + p2[0]); + d00 = dp00 + dq00; + d30 = dp30 + dq30; + p_is_pcm0 = p_is_pcm[0]; + q_is_pcm0 = q_is_pcm[0]; + + dp04 = abs(p1[-3] - (p1[-2] << 1) + p1[-1]); + dq04 = abs(p1[2] - (p1[1] << 1) + p1[0]); + dp34 = abs(p0[-3] - (p0[-2] << 1) + p0[-1]); + dq34 = abs(p0[2] - (p0[1] << 1) + p0[0]); + d04 = dp04 + dq04; + d34 = dp34 + dq34; + p_is_pcm4 = p_is_pcm[1]; + q_is_pcm4 = q_is_pcm[1]; + + if (!p_is_pcm0 || !p_is_pcm4 || !q_is_pcm0 || !q_is_pcm4) { + if (!(d00 + d30 >= beta) || !(d04 + d34 >= beta)) { + src -= 4; + LD_UH8(src, stride, + p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, q2_src, + q3_src); + + tc0 = tc[0]; + beta30 = beta >> 3; + beta20 = beta >> 2; + tc250 = ((tc0 * 5 + 1) >> 1); + + tc4 = tc[1]; + tc254 = ((tc4 * 5 + 1) >> 1); + + TRANSPOSE8x8_UB_UH(p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, + q2_src, q3_src, p3_src, p2_src, p1_src, p0_src, + q0_src, q1_src, q2_src, q3_src); + + flag0 = (abs(p3[-4] - p3[-1]) + abs(p3[3] - p3[0]) < beta30 && + abs(p3[-1] - p3[0]) < tc250 && + abs(p2[-4] - p2[-1]) + abs(p2[3] - p2[0]) < beta30 && + abs(p2[-1] - p2[0]) < tc250 && + (d00 << 1) < beta20 && (d30 << 1) < beta20); + cmp0 = __msa_fill_d(flag0); + + flag1 = (abs(p1[-4] - p1[-1]) + abs(p1[3] - p1[0]) < beta30 && + abs(p1[-1] - p1[0]) < tc254 && + abs(p0[-4] - p0[-1]) + abs(p0[3] - p0[0]) < beta30 && + abs(p0[-1] - p0[0]) < tc254 && + (d04 << 1) < beta20 && (d34 << 1) < beta20); + cmp1 = __msa_fill_d(flag1); + cmp2 = __msa_ilvev_d(cmp1, cmp0); + cmp2 = __msa_ceqi_d(cmp2, 0); + + ILVR_B8_UH(zero, p3_src, zero, p2_src, zero, p1_src, zero, p0_src, + zero, q0_src, zero, q1_src, zero, q2_src, zero, q3_src, + p3_src, p2_src, p1_src, p0_src, q0_src, q1_src, q2_src, + q3_src); + + cmp0 = (v2i64) __msa_fill_h(tc0 << 1); + cmp1 = (v2i64) __msa_fill_h(tc4 << 1); + tc_pos = (v8i16) __msa_ilvev_d(cmp1, cmp0); + tc_neg = -tc_pos; + + temp0 = (p1_src + p0_src + q0_src); + + temp1 = ((p3_src + p2_src) << 1) + p2_src + temp0; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - p2_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst0 = (v16u8) (temp2 + (v8i16) p2_src); + + temp1 = temp0 + p2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 2); + temp2 = (v8i16) (temp1 - p1_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst1 = (v16u8) (temp2 + (v8i16) p1_src); + + temp1 = (temp0 << 1) + p2_src + q1_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - p0_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst2 = (v16u8) (temp2 + (v8i16) p0_src); + + cmp0 = __msa_fill_d(p_is_pcm0); + cmp1 = __msa_fill_d(p_is_pcm4); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + dst0 = __msa_bmz_v(dst0, (v16u8) p2_src, (v16u8) p_is_pcm_vec); + dst1 = __msa_bmz_v(dst1, (v16u8) p1_src, (v16u8) p_is_pcm_vec); + dst2 = __msa_bmz_v(dst2, (v16u8) p0_src, (v16u8) p_is_pcm_vec); + + temp0 = (q1_src + p0_src + q0_src); + temp1 = ((q3_src + q2_src) << 1) + q2_src + temp0; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - q2_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst5 = (v16u8) (temp2 + (v8i16) q2_src); + + temp1 = temp0 + q2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 2); + temp2 = (v8i16) (temp1 - q1_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst4 = (v16u8) (temp2 + (v8i16) q1_src); + + temp1 = (temp0 << 1) + p1_src + q2_src; + temp1 = (v8u16) __msa_srari_h((v8i16) temp1, 3); + temp2 = (v8i16) (temp1 - q0_src); + temp2 = CLIP_SH(temp2, tc_neg, tc_pos); + dst3 = (v16u8) (temp2 + (v8i16) q0_src); + + cmp0 = __msa_fill_d(q_is_pcm0); + cmp1 = __msa_fill_d(q_is_pcm4); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + dst3 = __msa_bmz_v(dst3, (v16u8) q0_src, (v16u8) q_is_pcm_vec); + dst4 = __msa_bmz_v(dst4, (v16u8) q1_src, (v16u8) q_is_pcm_vec); + dst5 = __msa_bmz_v(dst5, (v16u8) q2_src, (v16u8) q_is_pcm_vec); + + tc_pos >>= 1; + tc_neg = -tc_pos; + + diff0 = (v8i16) (q0_src - p0_src); + diff1 = (v8i16) (q1_src - p1_src); + diff0 = (v8i16) (diff0 << 3) + diff0; + diff1 = (v8i16) (diff1 << 1) + diff1; + delta0 = diff0 - diff1; + delta0 = __msa_srari_h(delta0, 4); + + temp1 = (v8u16) ((tc_pos << 3) + (tc_pos << 1)); + abs_delta0 = __msa_add_a_h(delta0, (v8i16) zero); + abs_delta0 = (v8u16) abs_delta0 < temp1; + + delta0 = CLIP_SH(delta0, tc_neg, tc_pos); + temp0 = (v8u16) delta0 + p0_src; + temp0 = (v8u16) CLIP_SH_0_255(temp0); + temp0 = (v8u16) __msa_bmz_v((v16u8) temp0, (v16u8) p0_src, + (v16u8) p_is_pcm_vec); + + temp2 = (v8i16) q0_src - delta0; + temp2 = CLIP_SH_0_255(temp2); + temp2 = (v8i16) __msa_bmz_v((v16u8) temp2, (v16u8) q0_src, + (v16u8) q_is_pcm_vec); + + tmp = ((beta + (beta >> 1)) >> 3); + cmp0 = __msa_fill_d(!p_is_pcm0 && (dp00 + dp30 < tmp)); + cmp1 = __msa_fill_d(!p_is_pcm4 && (dp04 + dp34 < tmp)); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + cmp0 = (v2i64) __msa_fill_h((!q_is_pcm0) && (dq00 + dq30 < tmp)); + cmp1 = (v2i64) __msa_fill_h((!q_is_pcm4) && (dq04 + dq34 < tmp)); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + tc_pos >>= 1; + tc_neg = -tc_pos; + + delta1 = (v8i16) __msa_aver_u_h(p2_src, p0_src); + delta1 -= (v8i16) p1_src; + delta1 += delta0; + delta1 >>= 1; + delta1 = CLIP_SH(delta1, tc_neg, tc_pos); + delta1 = (v8i16) p1_src + (v8i16) delta1; + delta1 = CLIP_SH_0_255(delta1); + delta1 = (v8i16) __msa_bmnz_v((v16u8) delta1, (v16u8) p1_src, + (v16u8) p_is_pcm_vec); + + delta2 = (v8i16) __msa_aver_u_h(q0_src, q2_src); + delta2 = delta2 - (v8i16) q1_src; + delta2 = delta2 - delta0; + delta2 = delta2 >> 1; + delta2 = CLIP_SH(delta2, tc_neg, tc_pos); + delta2 = (v8i16) q1_src + (v8i16) delta2; + delta2 = CLIP_SH_0_255(delta2); + delta2 = (v8i16) __msa_bmnz_v((v16u8) delta2, (v16u8) q1_src, + (v16u8) q_is_pcm_vec); + delta1 = (v8i16) __msa_bmz_v((v16u8) delta1, (v16u8) p1_src, + (v16u8) abs_delta0); + temp0 = (v8u16) __msa_bmz_v((v16u8) temp0, (v16u8) p0_src, + (v16u8) abs_delta0); + temp2 = (v8i16) __msa_bmz_v((v16u8) temp2, (v16u8) q0_src, + (v16u8) abs_delta0); + delta2 = (v8i16) __msa_bmz_v((v16u8) delta2, (v16u8) q1_src, + (v16u8) abs_delta0); + + dst2 = __msa_bmnz_v(dst2, (v16u8) temp0, (v16u8) cmp2); + dst3 = __msa_bmnz_v(dst3, (v16u8) temp2, (v16u8) cmp2); + dst1 = __msa_bmnz_v(dst1, (v16u8) delta1, (v16u8) cmp2); + dst4 = __msa_bmnz_v(dst4, (v16u8) delta2, (v16u8) cmp2); + dst0 = __msa_bmnz_v(dst0, (v16u8) p2_src, (v16u8) cmp2); + dst5 = __msa_bmnz_v(dst5, (v16u8) q2_src, (v16u8) cmp2); + + cmp0 = __msa_fill_d(d00 + d30 >= beta); + dst7 = (v16u8) __msa_fill_d(d04 + d34 >= beta); + cmp0 = __msa_ilvev_d((v2i64) dst7, cmp0); + dst6 = (v16u8) __msa_ceqi_d(cmp0, 0); + + dst0 = __msa_bmz_v(dst0, (v16u8) p2_src, dst6); + dst1 = __msa_bmz_v(dst1, (v16u8) p1_src, dst6); + dst2 = __msa_bmz_v(dst2, (v16u8) p0_src, dst6); + dst3 = __msa_bmz_v(dst3, (v16u8) q0_src, dst6); + dst4 = __msa_bmz_v(dst4, (v16u8) q1_src, dst6); + dst5 = __msa_bmz_v(dst5, (v16u8) q2_src, dst6); + + PCKEV_B4_UB(dst0, dst0, dst1, dst1, dst2, dst2, dst3, dst3, + dst0, dst1, dst2, dst3); + PCKEV_B2_UB(dst4, dst4, dst5, dst5, dst4, dst5); + + TRANSPOSE8x8_UB_UB(dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7, + dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7); + + src += 1; + + tmp2 = __msa_copy_u_w((v4i32) dst0, 0); + tmp0 = __msa_copy_u_h((v8i16) dst0, 2); + tmp3 = __msa_copy_u_w((v4i32) dst1, 0); + tmp1 = __msa_copy_u_h((v8i16) dst1, 2); + SW(tmp2, src); + SH(tmp0, src + 4); + src += stride; + SW(tmp3, src); + SH(tmp1, src + 4); + src += stride; + + tmp2 = __msa_copy_u_w((v4i32) dst2, 0); + tmp0 = __msa_copy_u_h((v8i16) dst2, 2); + tmp3 = __msa_copy_u_w((v4i32) dst3, 0); + tmp1 = __msa_copy_u_h((v8i16) dst3, 2); + SW(tmp2, src); + SH(tmp0, src + 4); + src += stride; + SW(tmp3, src); + SH(tmp1, src + 4); + src += stride; + + tmp2 = __msa_copy_u_w((v4i32) dst4, 0); + tmp0 = __msa_copy_u_h((v8i16) dst4, 2); + tmp3 = __msa_copy_u_w((v4i32) dst5, 0); + tmp1 = __msa_copy_u_h((v8i16) dst5, 2); + SW(tmp2, src); + SH(tmp0, src + 4); + src += stride; + SW(tmp3, src); + SH(tmp1, src + 4); + src += stride; + + tmp2 = __msa_copy_u_w((v4i32) dst6, 0); + tmp0 = __msa_copy_u_h((v8i16) dst6, 2); + tmp3 = __msa_copy_u_w((v4i32) dst7, 0); + tmp1 = __msa_copy_u_h((v8i16) dst7, 2); + SW(tmp2, src); + SH(tmp0, src + 4); + src += stride; + SW(tmp3, src); + SH(tmp1, src + 4); + } + } +} + +static void hevc_loopfilter_chroma_hor_msa(uint8_t *src, int32_t stride, + int32_t *tc, uint8_t *p_is_pcm, + uint8_t *q_is_pcm) +{ + uint8_t *p1_ptr = src - (stride << 1); + uint8_t *p0_ptr = src - stride; + uint8_t *q0_ptr = src; + uint8_t *q1_ptr = src + stride; + v2i64 cmp0, cmp1, p_is_pcm_vec, q_is_pcm_vec; + v8u16 p1, p0, q0, q1; + v8i16 tc_pos, tc_neg; + v16i8 zero = { 0 }; + v8i16 temp0, temp1, delta; + + if (!(tc[0] <= 0) || !(tc[1] <= 0)) { + cmp0 = (v2i64) __msa_fill_h(tc[0]); + cmp1 = (v2i64) __msa_fill_h(tc[1]); + tc_pos = (v8i16) __msa_ilvev_d(cmp1, cmp0); + tc_neg = -tc_pos; + + cmp0 = __msa_fill_d(p_is_pcm[0]); + cmp1 = __msa_fill_d(p_is_pcm[1]); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + cmp0 = __msa_fill_d(q_is_pcm[0]); + cmp1 = __msa_fill_d(q_is_pcm[1]); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + p1 = LD_UH(p1_ptr); + p0 = LD_UH(p0_ptr); + q0 = LD_UH(q0_ptr); + q1 = LD_UH(q1_ptr); + + ILVR_B4_UH(zero, p1, zero, p0, zero, q0, zero, q1, p1, p0, q0, q1); + + temp0 = (v8i16) (q0 - p0); + temp1 = (v8i16) (p1 - q1); + temp0 <<= 2; + temp0 += temp1; + delta = __msa_srari_h((v8i16) temp0, 3); + delta = CLIP_SH(delta, tc_neg, tc_pos); + + temp0 = (v8i16) ((v8i16) p0 + delta); + temp0 = CLIP_SH_0_255(temp0); + temp0 = (v8i16) __msa_bmz_v((v16u8) temp0, (v16u8) p0, + (v16u8) p_is_pcm_vec); + + temp1 = (v8i16) ((v8i16) q0 - delta); + temp1 = CLIP_SH_0_255(temp1); + temp1 = (v8i16) __msa_bmz_v((v16u8) temp1, (v16u8) q0, + (v16u8) q_is_pcm_vec); + + tc_pos = (v8i16) __msa_clei_s_d((v2i64) tc_pos, 0); + temp0 = (v8i16) __msa_bmnz_v((v16u8) temp0, (v16u8) p0, (v16u8) tc_pos); + temp1 = (v8i16) __msa_bmnz_v((v16u8) temp1, (v16u8) q0, (v16u8) tc_pos); + + temp0 = (v8i16) __msa_pckev_b((v16i8) temp1, (v16i8) temp0); + ST8x2_UB(temp0, p0_ptr, stride); + } +} + +static void hevc_loopfilter_chroma_ver_msa(uint8_t *src, int32_t stride, + int32_t *tc, uint8_t *p_is_pcm, + uint8_t *q_is_pcm) +{ + v2i64 cmp0, cmp1, p_is_pcm_vec, q_is_pcm_vec; + v16u8 src0, src1, src2, src3, src4, src5, src6, src7; + v8u16 p1, p0, q0, q1; + v8i16 tc_pos, tc_neg; + v16i8 zero = { 0 }; + v8i16 temp0, temp1, delta; + + if (!(tc[0] <= 0) || !(tc[1] <= 0)) { + cmp0 = (v2i64) __msa_fill_h(tc[0]); + cmp1 = (v2i64) __msa_fill_h(tc[1]); + tc_pos = (v8i16) __msa_ilvev_d(cmp1, cmp0); + tc_neg = -tc_pos; + + cmp0 = __msa_fill_d(p_is_pcm[0]); + cmp1 = __msa_fill_d(p_is_pcm[1]); + p_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + p_is_pcm_vec = __msa_ceqi_d(p_is_pcm_vec, 0); + + cmp0 = __msa_fill_d(q_is_pcm[0]); + cmp1 = __msa_fill_d(q_is_pcm[1]); + q_is_pcm_vec = __msa_ilvev_d(cmp1, cmp0); + q_is_pcm_vec = __msa_ceqi_d(q_is_pcm_vec, 0); + + src -= 2; + LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); + TRANSPOSE8x4_UB_UH(src0, src1, src2, src3, src4, src5, src6, src7, + p1, p0, q0, q1); + ILVR_B4_UH(zero, p1, zero, p0, zero, q0, zero, q1, p1, p0, q0, q1); + + temp0 = (v8i16) (q0 - p0); + temp1 = (v8i16) (p1 - q1); + temp0 <<= 2; + temp0 += temp1; + delta = __msa_srari_h((v8i16) temp0, 3); + delta = CLIP_SH(delta, tc_neg, tc_pos); + + temp0 = (v8i16) ((v8i16) p0 + delta); + temp0 = CLIP_SH_0_255(temp0); + temp0 = (v8i16) __msa_bmz_v((v16u8) temp0, (v16u8) p0, + (v16u8) p_is_pcm_vec); + + temp1 = (v8i16) ((v8i16) q0 - delta); + temp1 = CLIP_SH_0_255(temp1); + temp1 = (v8i16) __msa_bmz_v((v16u8) temp1, (v16u8) q0, + (v16u8) q_is_pcm_vec); + + tc_pos = (v8i16) __msa_clei_s_d((v2i64) tc_pos, 0); + temp0 = (v8i16) __msa_bmnz_v((v16u8) temp0, (v16u8) p0, (v16u8) tc_pos); + temp1 = (v8i16) __msa_bmnz_v((v16u8) temp1, (v16u8) q0, (v16u8) tc_pos); + + temp0 = (v8i16) __msa_ilvev_b((v16i8) temp1, (v16i8) temp0); + + src += 1; + ST2x4_UB(temp0, 0, src, stride); + src += (4 * stride); + ST2x4_UB(temp0, 4, src, stride); + } +} + +static void hevc_sao_band_filter_4width_msa(uint8_t *dst, int32_t dst_stride, + uint8_t *src, int32_t src_stride, + int32_t sao_left_class, + int16_t *sao_offset_val, + int32_t height) +{ + int32_t h_cnt; + v16u8 src0, src1, src2, src3; + v16i8 src0_r, src1_r; + v16i8 offset, offset_val, mask; + v16i8 offset0 = { 0 }; + v16i8 offset1 = { 0 }; + v16i8 zero = { 0 }; + v8i16 temp0, temp1, dst0, dst1; + + offset_val = LD_SB(sao_offset_val + 1); + offset_val = (v16i8) __msa_pckev_d((v2i64) offset_val, (v2i64) offset_val); + + offset_val = __msa_pckev_b(offset_val, offset_val); + offset1 = (v16i8) __msa_insve_w((v4i32) offset1, 3, (v4i32) offset_val); + offset0 = __msa_sld_b(offset1, offset0, 28 - ((sao_left_class) & 31)); + offset1 = __msa_sld_b(zero, offset1, 28 - ((sao_left_class) & 31)); + + if (!((sao_left_class > 12) & (sao_left_class < 29))) { + SWAP(offset0, offset1); + } + + for (h_cnt = height >> 2; h_cnt--;) { + LD_UB4(src, src_stride, src0, src1, src2, src3); + src += (4 * src_stride); + + ILVEV_D2_SB(src0, src1, src2, src3, src0_r, src1_r); + + src0_r = (v16i8) __msa_pckev_w((v4i32) src1_r, (v4i32) src0_r); + mask = __msa_srli_b(src0_r, 3); + offset = __msa_vshf_b(mask, offset1, offset0); + + UNPCK_SB_SH(offset, temp0, temp1); + ILVRL_B2_SH(zero, src0_r, dst0, dst1); + ADD2(dst0, temp0, dst1, temp1, dst0, dst1); + CLIP_SH2_0_255(dst0, dst1); + dst0 = (v8i16) __msa_pckev_b((v16i8) dst1, (v16i8) dst0); + ST4x4_UB(dst0, dst0, 0, 1, 2, 3, dst, dst_stride); + dst += (4 * dst_stride); + } +} + +static void hevc_sao_band_filter_8width_msa(uint8_t *dst, int32_t dst_stride, + uint8_t *src, int32_t src_stride, + int32_t sao_left_class, + int16_t *sao_offset_val, + int32_t height) +{ + int32_t h_cnt; + v16u8 src0, src1, src2, src3; + v16i8 src0_r, src1_r, mask0, mask1; + v16i8 offset, offset_val; + v16i8 offset0 = { 0 }; + v16i8 offset1 = { 0 }; + v16i8 zero = { 0 }; + v8i16 dst0, dst1, dst2, dst3; + v8i16 temp0, temp1, temp2, temp3; + + offset_val = LD_SB(sao_offset_val + 1); + offset_val = (v16i8) __msa_pckev_d((v2i64) offset_val, (v2i64) offset_val); + offset_val = __msa_pckev_b(offset_val, offset_val); + offset1 = (v16i8) __msa_insve_w((v4i32) offset1, 3, (v4i32) offset_val); + offset0 = __msa_sld_b(offset1, offset0, 28 - ((sao_left_class) & 31)); + offset1 = __msa_sld_b(zero, offset1, 28 - ((sao_left_class) & 31)); + + if (!((sao_left_class > 12) & (sao_left_class < 29))) { + SWAP(offset0, offset1); + } + + for (h_cnt = height >> 2; h_cnt--;) { + LD_UB4(src, src_stride, src0, src1, src2, src3); + src += (4 * src_stride); + + ILVR_D2_SB(src1, src0, src3, src2, src0_r, src1_r); + + mask0 = __msa_srli_b(src0_r, 3); + mask1 = __msa_srli_b(src1_r, 3); + + offset = __msa_vshf_b(mask0, offset1, offset0); + UNPCK_SB_SH(offset, temp0, temp1); + + offset = __msa_vshf_b(mask1, offset1, offset0); + UNPCK_SB_SH(offset, temp2, temp3); + + UNPCK_UB_SH(src0_r, dst0, dst1); + UNPCK_UB_SH(src1_r, dst2, dst3); + ADD4(dst0, temp0, dst1, temp1, dst2, temp2, dst3, temp3, + dst0, dst1, dst2, dst3); + CLIP_SH4_0_255(dst0, dst1, dst2, dst3); + PCKEV_B2_SH(dst1, dst0, dst3, dst2, dst0, dst2); + ST8x4_UB(dst0, dst2, dst, dst_stride); + dst += (4 * dst_stride); + } +} + +static void hevc_sao_band_filter_16multiple_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int32_t sao_left_class, + int16_t *sao_offset_val, + int32_t width, int32_t height) +{ + int32_t h_cnt, w_cnt; + v16u8 src0, src1, src2, src3; + v8i16 dst0, dst1, dst2, dst3, dst4, dst5, dst6, dst7; + v16i8 out0, out1, out2, out3; + v16i8 mask0, mask1, mask2, mask3; + v16i8 tmp0, tmp1, tmp2, tmp3, offset_val; + v16i8 offset0 = { 0 }; + v16i8 offset1 = { 0 }; + v16i8 zero = { 0 }; + v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; + + offset_val = LD_SB(sao_offset_val + 1); + offset_val = (v16i8) __msa_pckev_d((v2i64) offset_val, (v2i64) offset_val); + offset_val = __msa_pckev_b(offset_val, offset_val); + offset1 = (v16i8) __msa_insve_w((v4i32) offset1, 3, (v4i32) offset_val); + offset0 = __msa_sld_b(offset1, offset0, 28 - ((sao_left_class) & 31)); + offset1 = __msa_sld_b(zero, offset1, 28 - ((sao_left_class) & 31)); + + if (!((sao_left_class > 12) & (sao_left_class < 29))) { + SWAP(offset0, offset1); + } + + for (h_cnt = height >> 2; h_cnt--;) { + for (w_cnt = 0; w_cnt < (width >> 4); w_cnt++) { + LD_UB4(src + w_cnt * 16, src_stride, src0, src1, src2, src3); + + mask0 = __msa_srli_b((v16i8) src0, 3); + mask1 = __msa_srli_b((v16i8) src1, 3); + mask2 = __msa_srli_b((v16i8) src2, 3); + mask3 = __msa_srli_b((v16i8) src3, 3); + + VSHF_B2_SB(offset0, offset1, offset0, offset1, mask0, mask1, + tmp0, tmp1); + VSHF_B2_SB(offset0, offset1, offset0, offset1, mask2, mask3, + tmp2, tmp3); + UNPCK_SB_SH(tmp0, temp0, temp1); + UNPCK_SB_SH(tmp1, temp2, temp3); + UNPCK_SB_SH(tmp2, temp4, temp5); + UNPCK_SB_SH(tmp3, temp6, temp7); + ILVRL_B2_SH(zero, src0, dst0, dst1); + ILVRL_B2_SH(zero, src1, dst2, dst3); + ILVRL_B2_SH(zero, src2, dst4, dst5); + ILVRL_B2_SH(zero, src3, dst6, dst7); + ADD4(dst0, temp0, dst1, temp1, dst2, temp2, dst3, temp3, + dst0, dst1, dst2, dst3); + ADD4(dst4, temp4, dst5, temp5, dst6, temp6, dst7, temp7, + dst4, dst5, dst6, dst7); + CLIP_SH4_0_255(dst0, dst1, dst2, dst3); + CLIP_SH4_0_255(dst4, dst5, dst6, dst7); + PCKEV_B4_SB(dst1, dst0, dst3, dst2, dst5, dst4, dst7, dst6, + out0, out1, out2, out3); + ST_SB4(out0, out1, out2, out3, dst + w_cnt * 16, dst_stride); + } + + src += src_stride << 2; + dst += dst_stride << 2; + } +} + +static void hevc_sao_edge_filter_0degree_4width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + int32_t h_cnt; + uint32_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src_minus11; + v16i8 zero = { 0 }; + v16i8 src_zero0, src_zero1, src_plus10, src_plus11, dst0; + v8i16 offset_mask0, offset_mask1; + v8i16 sao_offset, src00, src01; + + sao_offset = LD_SH(sao_offset_val); + src -= 1; + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src, src_stride, src_minus10, src_minus11); + src += (2 * src_stride); + + SLDI_B2_0_SB(src_minus10, src_minus11, src_zero0, src_zero1, 1); + SLDI_B2_0_SB(src_minus10, src_minus11, src_plus10, src_plus11, 2); + ILVR_B2_UB(src_plus10, src_minus10, src_plus11, src_minus11, + src_minus10, src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, src_zero0, + src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + + dst0 = __msa_pckev_b((v16i8) offset_mask1, (v16i8) offset_mask0); + dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); + dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); + SW(dst_val0, dst); + dst += dst_stride; + SW(dst_val1, dst); + dst += dst_stride; + } +} + +static void hevc_sao_edge_filter_0degree_8width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + uint8_t *src_minus1; + int32_t h_cnt; + uint64_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 dst0, dst1; + v16i8 zero = { 0 }; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src_minus11; + v16i8 src_zero0, src_plus10, src_zero1, src_plus11; + v8i16 sao_offset, src00, offset_mask0, src01, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + + for (h_cnt = (height >> 1); h_cnt--;) { + src_minus1 = src - 1; + LD_UB2(src_minus1, src_stride, src_minus10, src_minus11); + + SLDI_B2_0_SB(src_minus10, src_minus11, src_zero0, src_zero1, 1); + SLDI_B2_0_SB(src_minus10, src_minus11, src_plus10, src_plus11, 2); + ILVR_B2_UB(src_plus10, src_minus10, src_plus11, src_minus11, + src_minus10, src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, + src_zero0, src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + PCKEV_B2_SB(offset_mask0, offset_mask0, offset_mask1, offset_mask1, + dst0, dst1); + + dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst1, 0); + SD(dst_val0, dst); + dst += dst_stride; + SD(dst_val1, dst); + dst += dst_stride; + src += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_0degree_16multiple_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t width, + int32_t height) +{ + uint8_t *dst_ptr, *src_minus1; + int32_t h_cnt, v_cnt; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 sao_offset; + v16u8 cmp_minus10, cmp_plus10, diff_minus10, diff_plus10, cmp_minus11; + v16u8 cmp_plus11, diff_minus11, diff_plus11, cmp_minus12, cmp_plus12; + v16u8 diff_minus12, diff_plus12, cmp_minus13, cmp_plus13, diff_minus13; + v16u8 diff_plus13; + v16u8 src10, src11, src12, src13, dst0, dst1, dst2, dst3; + v16u8 src_minus10, src_minus11, src_minus12, src_minus13; + v16i8 offset_mask0, offset_mask1, offset_mask2, offset_mask3; + v16i8 src_zero0, src_zero1, src_zero2, src_zero3; + v16i8 src_plus10, src_plus11, src_plus12, src_plus13; + v8i16 src0, src1, src2, src3, src4, src5, src6, src7; + v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; + + sao_offset = LD_SB(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); + + for (h_cnt = (height >> 2); h_cnt--;) { + src_minus1 = src - 1; + LD_UB4(src_minus1, src_stride, + src_minus10, src_minus11, src_minus12, src_minus13); + + for (v_cnt = 0; v_cnt < (width >> 4); v_cnt++) { + src_minus1 += 16; + dst_ptr = dst + (v_cnt << 4); + LD_UB4(src_minus1, src_stride, src10, src11, src12, src13); + + SLDI_B2_SB(src10, src11, src_minus10, src_minus11, src_zero0, + src_zero1, 1); + SLDI_B2_SB(src12, src13, src_minus12, src_minus13, src_zero2, + src_zero3, 1); + SLDI_B2_SB(src10, src11, src_minus10, src_minus11, src_plus10, + src_plus11, 2); + SLDI_B2_SB(src12, src13, src_minus12, src_minus13, src_plus12, + src_plus13, 2); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + cmp_plus10 = ((v16u8) src_zero0 == (v16u8) src_plus10); + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + cmp_plus11 = ((v16u8) src_zero1 == (v16u8) src_plus11); + cmp_minus12 = ((v16u8) src_zero2 == src_minus12); + cmp_plus12 = ((v16u8) src_zero2 == (v16u8) src_plus12); + cmp_minus13 = ((v16u8) src_zero3 == src_minus13); + cmp_plus13 = ((v16u8) src_zero3 == (v16u8) src_plus13); + + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + diff_plus10 = __msa_nor_v(cmp_plus10, cmp_plus10); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + diff_plus11 = __msa_nor_v(cmp_plus11, cmp_plus11); + diff_minus12 = __msa_nor_v(cmp_minus12, cmp_minus12); + diff_plus12 = __msa_nor_v(cmp_plus12, cmp_plus12); + diff_minus13 = __msa_nor_v(cmp_minus13, cmp_minus13); + diff_plus13 = __msa_nor_v(cmp_plus13, cmp_plus13); + + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + cmp_plus10 = ((v16u8) src_plus10 < (v16u8) src_zero0); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + cmp_plus11 = ((v16u8) src_plus11 < (v16u8) src_zero1); + cmp_minus12 = (src_minus12 < (v16u8) src_zero2); + cmp_plus12 = ((v16u8) src_plus12 < (v16u8) src_zero2); + cmp_minus13 = (src_minus13 < (v16u8) src_zero3); + cmp_plus13 = ((v16u8) src_plus13 < (v16u8) src_zero3); + + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + diff_plus10 = __msa_bmnz_v(diff_plus10, const1, cmp_plus10); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + diff_plus11 = __msa_bmnz_v(diff_plus11, const1, cmp_plus11); + diff_minus12 = __msa_bmnz_v(diff_minus12, const1, cmp_minus12); + diff_plus12 = __msa_bmnz_v(diff_plus12, const1, cmp_plus12); + diff_minus13 = __msa_bmnz_v(diff_minus13, const1, cmp_minus13); + diff_plus13 = __msa_bmnz_v(diff_plus13, const1, cmp_plus13); + + offset_mask0 = 2 + (v16i8) diff_minus10 + (v16i8) diff_plus10; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + offset_mask1 = 2 + (v16i8) diff_minus11 + (v16i8) diff_plus11; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + offset_mask2 = 2 + (v16i8) diff_minus12 + (v16i8) diff_plus12; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask2, + offset_mask2, offset_mask2, offset_mask2); + offset_mask3 = 2 + (v16i8) diff_minus13 + (v16i8) diff_plus13; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask3, + offset_mask3, offset_mask3, offset_mask3); + + UNPCK_UB_SH(src_zero0, src0, src1); + UNPCK_SB_SH(offset_mask0, temp0, temp1); + UNPCK_UB_SH(src_zero1, src2, src3); + UNPCK_SB_SH(offset_mask1, temp2, temp3); + UNPCK_UB_SH(src_zero2, src4, src5); + UNPCK_SB_SH(offset_mask2, temp4, temp5); + UNPCK_UB_SH(src_zero3, src6, src7); + UNPCK_SB_SH(offset_mask3, temp6, temp7); + ADD4(temp0, src0, temp1, src1, temp2, src2, temp3, src3, temp0, + temp1, temp2, temp3); + ADD4(temp4, src4, temp5, src5, temp6, src6, temp7, src7, temp4, + temp5, temp6, temp7); + CLIP_SH4_0_255(temp0, temp1, temp2, temp3); + CLIP_SH4_0_255(temp4, temp5, temp6, temp7); + PCKEV_B4_UB(temp1, temp0, temp3, temp2, temp5, temp4, temp7, temp6, + dst0, dst1, dst2, dst3); + + src_minus10 = src10; + ST_UB(dst0, dst_ptr); + src_minus11 = src11; + ST_UB(dst1, dst_ptr + dst_stride); + src_minus12 = src12; + ST_UB(dst2, dst_ptr + (dst_stride << 1)); + src_minus13 = src13; + ST_UB(dst3, dst_ptr + (dst_stride * 3)); + } + + src += (src_stride << 2); + dst += (dst_stride << 2); + } +} + +static void hevc_sao_edge_filter_90degree_4width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + int32_t h_cnt; + uint32_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 dst0; + v16i8 zero = { 0 }; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src_minus11, src10, src11; + v16i8 src_zero0, src_zero1; + v8i16 sao_offset, src00, src01, offset_mask0, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + + LD_UB2(src - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src + src_stride, src_stride, src10, src11); + + src_minus10 = (v16u8) __msa_ilvr_b((v16i8) src10, (v16i8) src_minus10); + src_zero0 = __msa_ilvr_b((v16i8) src_minus11, (v16i8) src_minus11); + src_minus11 = (v16u8) __msa_ilvr_b((v16i8) src11, (v16i8) src_minus11); + src_zero1 = __msa_ilvr_b((v16i8) src10, (v16i8) src10); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + dst0 = __msa_pckev_b((v16i8) offset_mask1, (v16i8) offset_mask0); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); + dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); + SW(dst_val0, dst); + dst += dst_stride; + SW(dst_val1, dst); + + dst += dst_stride; + src += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_90degree_8width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + int32_t h_cnt; + uint64_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 zero = { 0 }; + v16i8 src_zero0, src_zero1, dst0, dst1; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src_minus11, src10, src11; + v8i16 sao_offset, src00, offset_mask0, src01, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + + LD_UB2(src - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src + src_stride, src_stride, src10, src11); + + src_minus10 = (v16u8) __msa_ilvr_b((v16i8) src10, (v16i8) src_minus10); + src_zero0 = __msa_ilvr_b((v16i8) src_minus11, (v16i8) src_minus11); + src_minus11 = (v16u8) __msa_ilvr_b((v16i8) src11, (v16i8) src_minus11); + src_zero1 = __msa_ilvr_b((v16i8) src10, (v16i8) src10); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + PCKEV_B2_SB(offset_mask0, offset_mask0, offset_mask1, offset_mask1, + dst0, dst1); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst1, 0); + SD(dst_val0, dst); + dst += dst_stride; + SD(dst_val1, dst); + dst += dst_stride; + src += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_90degree_16multiple_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t * + sao_offset_val, + int32_t width, + int32_t height) +{ + uint8_t *src_orig = src; + uint8_t *dst_orig = dst; + int32_t h_cnt, v_cnt; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16u8 cmp_minus10, cmp_plus10, diff_minus10, diff_plus10, cmp_minus11; + v16u8 cmp_plus11, diff_minus11, diff_plus11, cmp_minus12, cmp_plus12; + v16u8 diff_minus12, diff_plus12, cmp_minus13, cmp_plus13, diff_minus13; + v16u8 diff_plus13; + v16u8 src10, src_minus10, dst0, src11, src_minus11, dst1; + v16u8 src12, dst2, src13, dst3; + v16i8 offset_mask0, offset_mask1, offset_mask2, offset_mask3, sao_offset; + v8i16 src0, src1, src2, src3, src4, src5, src6, src7; + v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; + + sao_offset = LD_SB(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); + + for (v_cnt = 0; v_cnt < (width >> 4); v_cnt++) { + src = src_orig + (v_cnt << 4); + dst = dst_orig + (v_cnt << 4); + + LD_UB2(src - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 2); h_cnt--;) { + LD_UB4(src + src_stride, src_stride, src10, src11, src12, src13); + + cmp_minus10 = (src_minus11 == src_minus10); + cmp_plus10 = (src_minus11 == src10); + cmp_minus11 = (src10 == src_minus11); + cmp_plus11 = (src10 == src11); + cmp_minus12 = (src11 == src10); + cmp_plus12 = (src11 == src12); + cmp_minus13 = (src12 == src11); + cmp_plus13 = (src12 == src13); + + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + diff_plus10 = __msa_nor_v(cmp_plus10, cmp_plus10); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + diff_plus11 = __msa_nor_v(cmp_plus11, cmp_plus11); + diff_minus12 = __msa_nor_v(cmp_minus12, cmp_minus12); + diff_plus12 = __msa_nor_v(cmp_plus12, cmp_plus12); + diff_minus13 = __msa_nor_v(cmp_minus13, cmp_minus13); + diff_plus13 = __msa_nor_v(cmp_plus13, cmp_plus13); + + cmp_minus10 = (src_minus10 < src_minus11); + cmp_plus10 = (src10 < src_minus11); + cmp_minus11 = (src_minus11 < src10); + cmp_plus11 = (src11 < src10); + cmp_minus12 = (src10 < src11); + cmp_plus12 = (src12 < src11); + cmp_minus13 = (src11 < src12); + cmp_plus13 = (src13 < src12); + + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + diff_plus10 = __msa_bmnz_v(diff_plus10, const1, cmp_plus10); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + diff_plus11 = __msa_bmnz_v(diff_plus11, const1, cmp_plus11); + diff_minus12 = __msa_bmnz_v(diff_minus12, const1, cmp_minus12); + diff_plus12 = __msa_bmnz_v(diff_plus12, const1, cmp_plus12); + diff_minus13 = __msa_bmnz_v(diff_minus13, const1, cmp_minus13); + diff_plus13 = __msa_bmnz_v(diff_plus13, const1, cmp_plus13); + + offset_mask0 = 2 + (v16i8) diff_minus10 + (v16i8) diff_plus10; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask0, offset_mask0, offset_mask0, offset_mask0); + offset_mask1 = 2 + (v16i8) diff_minus11 + (v16i8) diff_plus11; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask1, offset_mask1, offset_mask1, offset_mask1); + offset_mask2 = 2 + (v16i8) diff_minus12 + (v16i8) diff_plus12; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask2, offset_mask2, offset_mask2, offset_mask2); + offset_mask3 = 2 + (v16i8) diff_minus13 + (v16i8) diff_plus13; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask3, offset_mask3, offset_mask3, offset_mask3); + + UNPCK_UB_SH(src_minus11, src0, src1); + UNPCK_SB_SH(offset_mask0, temp0, temp1); + UNPCK_UB_SH(src10, src2, src3); + UNPCK_SB_SH(offset_mask1, temp2, temp3); + UNPCK_UB_SH(src11, src4, src5); + UNPCK_SB_SH(offset_mask2, temp4, temp5); + UNPCK_UB_SH(src12, src6, src7); + UNPCK_SB_SH(offset_mask3, temp6, temp7); + ADD4(temp0, src0, temp1, src1, temp2, src2, temp3, src3, temp0, + temp1, temp2, temp3); + ADD4(temp4, src4, temp5, src5, temp6, src6, temp7, src7, temp4, + temp5, temp6, temp7); + CLIP_SH4_0_255(temp0, temp1, temp2, temp3); + CLIP_SH4_0_255(temp4, temp5, temp6, temp7); + PCKEV_B4_UB(temp1, temp0, temp3, temp2, temp5, temp4, temp7, temp6, + dst0, dst1, dst2, dst3); + + src_minus10 = src12; + src_minus11 = src13; + + ST_UB4(dst0, dst1, dst2, dst3, dst, dst_stride); + + src += (src_stride << 2); + dst += (dst_stride << 2); + } + } +} + +static void hevc_sao_edge_filter_45degree_4width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + uint8_t *src_orig; + int32_t h_cnt; + uint32_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 zero = { 0 }; + v16u8 cmp_minus10, diff_minus10, src_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus11, src10, src11; + v16i8 src_plus0, src_zero0, src_plus1, src_zero1, dst0; + v8i16 sao_offset, src00, src01, offset_mask0, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + + src_orig = src - 1; + LD_UB2(src_orig - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src_orig + src_stride, src_stride, src10, src11); + + SLDI_B2_0_SB(src_minus11, src10, src_zero0, src_zero1, 1); + SLDI_B2_0_SB(src10, src11, src_plus0, src_plus1, 2); + + ILVR_B2_UB(src_plus0, src_minus10, src_plus1, src_minus11, src_minus10, + src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, src_zero0, + src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + + dst0 = __msa_pckev_b((v16i8) offset_mask1, (v16i8) offset_mask0); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); + dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); + SW(dst_val0, dst); + dst += dst_stride; + SW(dst_val1, dst); + + dst += dst_stride; + src_orig += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_45degree_8width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + uint8_t *src_orig; + int32_t h_cnt; + uint64_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 zero = { 0 }; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src10, src_minus11, src11; + v16i8 src_zero0, src_plus10, src_zero1, src_plus11, dst0, dst1; + v8i16 sao_offset, src00, offset_mask0, src01, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + src_orig = src - 1; + + LD_UB2(src_orig - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src_orig + src_stride, src_stride, src10, src11); + + SLDI_B2_0_SB(src_minus11, src10, src_zero0, src_zero1, 1); + SLDI_B2_0_SB(src10, src11, src_plus10, src_plus11, 2); + + ILVR_B2_UB(src_plus10, src_minus10, src_plus11, src_minus11, + src_minus10, src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, + src_zero0, src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + PCKEV_B2_SB(offset_mask0, offset_mask0, offset_mask1, offset_mask1, + dst0, dst1); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst1, 0); + SD(dst_val0, dst); + dst += dst_stride; + SD(dst_val1, dst); + + dst += dst_stride; + src_orig += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_45degree_16multiple_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t * + sao_offset_val, + int32_t width, + int32_t height) +{ + uint8_t *src_orig = src; + uint8_t *dst_orig = dst; + int32_t h_cnt, v_cnt; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16u8 cmp_minus10, cmp_plus10, diff_minus10, diff_plus10, cmp_minus11; + v16u8 cmp_plus11, diff_minus11, diff_plus11, cmp_minus12, cmp_plus12; + v16u8 diff_minus12, diff_plus12, cmp_minus13, cmp_plus13, diff_minus13; + v16u8 diff_plus13, src_minus14, src_plus13; + v16i8 offset_mask0, offset_mask1, offset_mask2, offset_mask3; + v16u8 src10, src_minus10, dst0, src11, src_minus11, dst1; + v16u8 src12, src_minus12, dst2, src13, src_minus13, dst3; + v16i8 src_zero0, src_plus10, src_zero1, src_plus11, src_zero2, src_plus12; + v16i8 src_zero3, sao_offset; + v8i16 src0, src1, src2, src3, src4, src5, src6, src7; + v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; + + sao_offset = LD_SB(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); + + for (h_cnt = (height >> 2); h_cnt--;) { + src_orig = src - 1; + dst_orig = dst; + LD_UB4(src_orig, src_stride, + src_minus11, src_minus12, src_minus13, src_minus14); + + for (v_cnt = 0; v_cnt < (width >> 4); v_cnt++) { + src_minus10 = LD_UB(src_orig - src_stride); + LD_UB4(src_orig + 16, src_stride, src10, src11, src12, src13); + src_plus13 = LD_UB(src + 1 + (v_cnt << 4) + (src_stride << 2)); + src_orig += 16; + + SLDI_B2_SB(src10, src11, src_minus11, src_minus12, src_zero0, + src_zero1, 1); + SLDI_B2_SB(src12, src13, src_minus13, src_minus14, src_zero2, + src_zero3, 1); + SLDI_B2_SB(src11, src12, src_minus12, src_minus13, src_plus10, + src_plus11, 2); + + src_plus12 = __msa_sldi_b((v16i8) src13, (v16i8) src_minus14, 2); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + cmp_plus10 = ((v16u8) src_zero0 == (v16u8) src_plus10); + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + cmp_plus11 = ((v16u8) src_zero1 == (v16u8) src_plus11); + cmp_minus12 = ((v16u8) src_zero2 == src_minus12); + cmp_plus12 = ((v16u8) src_zero2 == (v16u8) src_plus12); + cmp_minus13 = ((v16u8) src_zero3 == src_minus13); + cmp_plus13 = ((v16u8) src_zero3 == src_plus13); + + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + diff_plus10 = __msa_nor_v(cmp_plus10, cmp_plus10); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + diff_plus11 = __msa_nor_v(cmp_plus11, cmp_plus11); + diff_minus12 = __msa_nor_v(cmp_minus12, cmp_minus12); + diff_plus12 = __msa_nor_v(cmp_plus12, cmp_plus12); + diff_minus13 = __msa_nor_v(cmp_minus13, cmp_minus13); + diff_plus13 = __msa_nor_v(cmp_plus13, cmp_plus13); + + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + cmp_plus10 = ((v16u8) src_plus10 < (v16u8) src_zero0); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + cmp_plus11 = ((v16u8) src_plus11 < (v16u8) src_zero1); + cmp_minus12 = (src_minus12 < (v16u8) src_zero2); + cmp_plus12 = ((v16u8) src_plus12 < (v16u8) src_zero2); + cmp_minus13 = (src_minus13 < (v16u8) src_zero3); + cmp_plus13 = (src_plus13 < (v16u8) src_zero3); + + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + diff_plus10 = __msa_bmnz_v(diff_plus10, const1, cmp_plus10); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + diff_plus11 = __msa_bmnz_v(diff_plus11, const1, cmp_plus11); + diff_minus12 = __msa_bmnz_v(diff_minus12, const1, cmp_minus12); + diff_plus12 = __msa_bmnz_v(diff_plus12, const1, cmp_plus12); + diff_minus13 = __msa_bmnz_v(diff_minus13, const1, cmp_minus13); + diff_plus13 = __msa_bmnz_v(diff_plus13, const1, cmp_plus13); + + offset_mask0 = 2 + (v16i8) diff_minus10 + (v16i8) diff_plus10; + offset_mask1 = 2 + (v16i8) diff_minus11 + (v16i8) diff_plus11; + offset_mask2 = 2 + (v16i8) diff_minus12 + (v16i8) diff_plus12; + offset_mask3 = 2 + (v16i8) diff_minus13 + (v16i8) diff_plus13; + + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask0, offset_mask0, offset_mask0, offset_mask0); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask1, offset_mask1, offset_mask1, offset_mask1); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask2, offset_mask2, offset_mask2, offset_mask2); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask3, offset_mask3, offset_mask3, offset_mask3); + + UNPCK_UB_SH(src_zero0, src0, src1); + UNPCK_SB_SH(offset_mask0, temp0, temp1); + UNPCK_UB_SH(src_zero1, src2, src3); + UNPCK_SB_SH(offset_mask1, temp2, temp3); + UNPCK_UB_SH(src_zero2, src4, src5); + UNPCK_SB_SH(offset_mask2, temp4, temp5); + UNPCK_UB_SH(src_zero3, src6, src7); + UNPCK_SB_SH(offset_mask3, temp6, temp7); + ADD4(temp0, src0, temp1, src1, temp2, src2, temp3, src3, temp0, + temp1, temp2, temp3); + ADD4(temp4, src4, temp5, src5, temp6, src6, temp7, src7, temp4, + temp5, temp6, temp7); + CLIP_SH4_0_255(temp0, temp1, temp2, temp3); + CLIP_SH4_0_255(temp4, temp5, temp6, temp7); + PCKEV_B4_UB(temp1, temp0, temp3, temp2, temp5, temp4, + temp7, temp6, dst0, dst1, dst2, dst3); + + src_minus11 = src10; + src_minus12 = src11; + src_minus13 = src12; + src_minus14 = src13; + + ST_UB4(dst0, dst1, dst2, dst3, dst_orig, dst_stride); + dst_orig += 16; + } + + src += (src_stride << 2); + dst += (dst_stride << 2); + } +} + +static void hevc_sao_edge_filter_135degree_4width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + uint8_t *src_orig; + int32_t h_cnt; + uint32_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 zero = { 0 }; + v16i8 src_zero0, src_zero1, dst0; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src10, src_minus11, src11; + v8i16 offset_mask0, offset_mask1, sao_offset, src00, src01; + + sao_offset = LD_SH(sao_offset_val); + src_orig = src - 1; + + LD_UB2(src_orig - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src_orig + src_stride, src_stride, src10, src11); + + SLDI_B2_0_SB(src_minus11, src10, src_zero0, src_zero1, 1); + SLDI_B2_0_UB(src_minus10, src_minus11, src_minus10, src_minus11, 2); + + ILVR_B2_UB(src10, src_minus10, src11, src_minus11, src_minus10, + src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, src_zero0, + src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + dst0 = __msa_pckev_b((v16i8) offset_mask1, (v16i8) offset_mask0); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); + dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); + + SW(dst_val0, dst); + dst += dst_stride; + SW(dst_val1, dst); + + dst += dst_stride; + src_orig += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_135degree_8width_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t *sao_offset_val, + int32_t height) +{ + uint8_t *src_orig; + int32_t h_cnt; + uint64_t dst_val0, dst_val1; + v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16i8 zero = { 0 }; + v16i8 src_zero0, src_zero1, dst0, dst1; + v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; + v16u8 src_minus10, src10, src_minus11, src11; + v8i16 sao_offset, src00, offset_mask0, src01, offset_mask1; + + sao_offset = LD_SH(sao_offset_val); + src_orig = src - 1; + + LD_UB2(src_orig - src_stride, src_stride, src_minus10, src_minus11); + + for (h_cnt = (height >> 1); h_cnt--;) { + LD_UB2(src_orig + src_stride, src_stride, src10, src11); + + SLDI_B2_0_SB(src_minus11, src10, src_zero0, src_zero1, 1); + SLDI_B2_0_UB(src_minus10, src_minus11, src_minus10, src_minus11, 2); + ILVR_B2_UB(src10, src_minus10, src11, src_minus11, src_minus10, + src_minus11); + ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, src_zero0, + src_zero1); + + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + + offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); + offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, + offset_mask0, offset_mask0, offset_mask0); + VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, + offset_mask1, offset_mask1, offset_mask1); + ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, + offset_mask1); + CLIP_SH2_0_255(offset_mask0, offset_mask1); + PCKEV_B2_SB(offset_mask0, offset_mask0, offset_mask1, offset_mask1, + dst0, dst1); + + src_minus10 = src10; + src_minus11 = src11; + + dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst1, 0); + + SD(dst_val0, dst); + dst += dst_stride; + SD(dst_val1, dst); + dst += dst_stride; + + src_orig += (src_stride << 1); + } +} + +static void hevc_sao_edge_filter_135degree_16multiple_msa(uint8_t *dst, + int32_t dst_stride, + uint8_t *src, + int32_t src_stride, + int16_t * + sao_offset_val, + int32_t width, + int32_t height) +{ + uint8_t *src_orig, *dst_orig; + int32_t h_cnt, v_cnt; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; + v16u8 const1 = (v16u8) __msa_ldi_b(1); + v16u8 dst0, dst1, dst2, dst3; + v16u8 cmp_minus10, cmp_minus11, cmp_minus12, cmp_minus13, cmp_plus10; + v16u8 cmp_plus11, cmp_plus12, cmp_plus13, diff_minus10, diff_minus11; + v16u8 diff_minus12, diff_minus13, diff_plus10, diff_plus11, diff_plus12; + v16u8 diff_plus13, src10, src11, src12, src13, src_minus10, src_minus11; + v16u8 src_plus10, src_plus11, src_plus12, src_plus13; + v16i8 src_minus12, src_minus13, src_zero0, src_zero1, src_zero2, src_zero3; + v16i8 offset_mask0, offset_mask1, offset_mask2, offset_mask3, sao_offset; + v8i16 src0, src1, src2, src3, src4, src5, src6, src7; + v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; + + sao_offset = LD_SB(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); + + for (h_cnt = (height >> 2); h_cnt--;) { + src_orig = src - 1; + dst_orig = dst; + + LD_UB4(src_orig, src_stride, + src_minus11, src_plus10, src_plus11, src_plus12); + + for (v_cnt = 0; v_cnt < (width >> 4); v_cnt++) { + src_minus10 = LD_UB(src_orig + 2 - src_stride); + LD_UB4(src_orig + 16, src_stride, src10, src11, src12, src13); + src_plus13 = LD_UB(src_orig + (src_stride << 2)); + src_orig += 16; + + src_zero0 = __msa_sldi_b((v16i8) src10, (v16i8) src_minus11, 1); + cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + cmp_plus10 = ((v16u8) src_zero0 == src_plus10); + + src_zero1 = __msa_sldi_b((v16i8) src11, (v16i8) src_plus10, 1); + src_minus11 = (v16u8) __msa_sldi_b((v16i8) src10, + (v16i8) src_minus11, 2); + cmp_minus11 = ((v16u8) src_zero1 == src_minus11); + cmp_plus11 = ((v16u8) src_zero1 == src_plus11); + + src_zero2 = __msa_sldi_b((v16i8) src12, (v16i8) src_plus11, 1); + src_minus12 = __msa_sldi_b((v16i8) src11, (v16i8) src_plus10, 2); + cmp_minus12 = ((v16u8) src_zero2 == (v16u8) src_minus12); + cmp_plus12 = ((v16u8) src_zero2 == src_plus12); + + src_zero3 = __msa_sldi_b((v16i8) src13, (v16i8) src_plus12, 1); + src_minus13 = __msa_sldi_b((v16i8) src12, (v16i8) src_plus11, 2); + cmp_minus13 = ((v16u8) src_zero3 == (v16u8) src_minus13); + cmp_plus13 = ((v16u8) src_zero3 == src_plus13); + + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + diff_plus10 = __msa_nor_v(cmp_plus10, cmp_plus10); + diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); + diff_plus11 = __msa_nor_v(cmp_plus11, cmp_plus11); + diff_minus12 = __msa_nor_v(cmp_minus12, cmp_minus12); + diff_plus12 = __msa_nor_v(cmp_plus12, cmp_plus12); + diff_minus13 = __msa_nor_v(cmp_minus13, cmp_minus13); + diff_plus13 = __msa_nor_v(cmp_plus13, cmp_plus13); + + cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + cmp_plus10 = (src_plus10 < (v16u8) src_zero0); + cmp_minus11 = (src_minus11 < (v16u8) src_zero1); + cmp_plus11 = (src_plus11 < (v16u8) src_zero1); + cmp_minus12 = ((v16u8) src_minus12 < (v16u8) src_zero2); + cmp_plus12 = (src_plus12 < (v16u8) src_zero2); + cmp_minus13 = ((v16u8) src_minus13 < (v16u8) src_zero3); + cmp_plus13 = (src_plus13 < (v16u8) src_zero3); + + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + diff_plus10 = __msa_bmnz_v(diff_plus10, const1, cmp_plus10); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + diff_plus11 = __msa_bmnz_v(diff_plus11, const1, cmp_plus11); + diff_minus12 = __msa_bmnz_v(diff_minus12, const1, cmp_minus12); + diff_plus12 = __msa_bmnz_v(diff_plus12, const1, cmp_plus12); + diff_minus13 = __msa_bmnz_v(diff_minus13, const1, cmp_minus13); + diff_plus13 = __msa_bmnz_v(diff_plus13, const1, cmp_plus13); + + offset_mask0 = 2 + (v16i8) diff_minus10 + (v16i8) diff_plus10; + offset_mask1 = 2 + (v16i8) diff_minus11 + (v16i8) diff_plus11; + offset_mask2 = 2 + (v16i8) diff_minus12 + (v16i8) diff_plus12; + offset_mask3 = 2 + (v16i8) diff_minus13 + (v16i8) diff_plus13; + + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask0, offset_mask0, offset_mask0, offset_mask0); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask1, offset_mask1, offset_mask1, offset_mask1); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask2, offset_mask2, offset_mask2, offset_mask2); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, + offset_mask3, offset_mask3, offset_mask3, offset_mask3); + + UNPCK_UB_SH(src_zero0, src0, src1); + UNPCK_SB_SH(offset_mask0, temp0, temp1); + UNPCK_UB_SH(src_zero1, src2, src3); + UNPCK_SB_SH(offset_mask1, temp2, temp3); + UNPCK_UB_SH(src_zero2, src4, src5); + UNPCK_SB_SH(offset_mask2, temp4, temp5); + UNPCK_UB_SH(src_zero3, src6, src7); + UNPCK_SB_SH(offset_mask3, temp6, temp7); + + ADD4(temp0, src0, temp1, src1, temp2, src2, temp3, src3, temp0, + temp1, temp2, temp3); + ADD4(temp4, src4, temp5, src5, temp6, src6, temp7, src7, temp4, + temp5, temp6, temp7); + CLIP_SH4_0_255(temp0, temp1, temp2, temp3); + CLIP_SH4_0_255(temp4, temp5, temp6, temp7); + PCKEV_B4_UB(temp1, temp0, temp3, temp2, temp5, temp4, temp7, temp6, + dst0, dst1, dst2, dst3); + + src_minus11 = src10; + src_plus10 = src11; + src_plus11 = src12; + src_plus12 = src13; + + ST_UB4(dst0, dst1, dst2, dst3, dst_orig, dst_stride); + dst_orig += 16; + } + + src += (src_stride << 2); + dst += (dst_stride << 2); + } +} + +void ff_hevc_loop_filter_luma_h_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t beta, int32_t *tc, + uint8_t *no_p, uint8_t *no_q) +{ + hevc_loopfilter_luma_hor_msa(src, src_stride, beta, tc, no_p, no_q); +} + +void ff_hevc_loop_filter_luma_v_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t beta, int32_t *tc, + uint8_t *no_p, uint8_t *no_q) +{ + hevc_loopfilter_luma_ver_msa(src, src_stride, beta, tc, no_p, no_q); +} + +void ff_hevc_loop_filter_chroma_h_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t *tc, uint8_t *no_p, + uint8_t *no_q) +{ + hevc_loopfilter_chroma_hor_msa(src, src_stride, tc, no_p, no_q); +} + +void ff_hevc_loop_filter_chroma_v_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t *tc, uint8_t *no_p, + uint8_t *no_q) +{ + hevc_loopfilter_chroma_ver_msa(src, src_stride, tc, no_p, no_q); +} + +void ff_hevc_sao_band_filter_0_8_msa(uint8_t *dst, uint8_t *src, + ptrdiff_t stride_dst, ptrdiff_t stride_src, + int16_t *sao_offset_val, int sao_left_class, + int width, int height) +{ + if (width >> 4) { + hevc_sao_band_filter_16multiple_msa(dst, stride_dst, src, stride_src, + sao_left_class, sao_offset_val, + width - (width % 16), height); + dst += width - (width % 16); + src += width - (width % 16); + width %= 16; + } + + if (width >> 3) { + hevc_sao_band_filter_8width_msa(dst, stride_dst, src, stride_src, + sao_left_class, sao_offset_val, height); + dst += 8; + src += 8; + width %= 8; + } + + if (width) { + hevc_sao_band_filter_4width_msa(dst, stride_dst, src, stride_src, + sao_left_class, sao_offset_val, height); + } +} + +void ff_hevc_sao_edge_filter_8_msa(uint8_t *dst, uint8_t *src, + ptrdiff_t stride_dst, + int16_t *sao_offset_val, + int eo, int width, int height) +{ + ptrdiff_t stride_src = (2 * 64 + 32) / sizeof(uint8_t); + + switch (eo) { + case 0: + if (width >> 4) { + hevc_sao_edge_filter_0degree_16multiple_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, + width - (width % 16), + height); + dst += width - (width % 16); + src += width - (width % 16); + width %= 16; + } + + if (width >> 3) { + hevc_sao_edge_filter_0degree_8width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + dst += 8; + src += 8; + width %= 8; + } + + if (width) { + hevc_sao_edge_filter_0degree_4width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + } + break; + + case 1: + if (width >> 4) { + hevc_sao_edge_filter_90degree_16multiple_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, + width - (width % 16), + height); + dst += width - (width % 16); + src += width - (width % 16); + width %= 16; + } + + if (width >> 3) { + hevc_sao_edge_filter_90degree_8width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + dst += 8; + src += 8; + width %= 8; + } + + if (width) { + hevc_sao_edge_filter_90degree_4width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + } + break; + + case 2: + if (width >> 4) { + hevc_sao_edge_filter_45degree_16multiple_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, + width - (width % 16), + height); + dst += width - (width % 16); + src += width - (width % 16); + width %= 16; + } + + if (width >> 3) { + hevc_sao_edge_filter_45degree_8width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + dst += 8; + src += 8; + width %= 8; + } + + if (width) { + hevc_sao_edge_filter_45degree_4width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + } + break; + + case 3: + if (width >> 4) { + hevc_sao_edge_filter_135degree_16multiple_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, + width - (width % 16), + height); + dst += width - (width % 16); + src += width - (width % 16); + width %= 16; + } + + if (width >> 3) { + hevc_sao_edge_filter_135degree_8width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + dst += 8; + src += 8; + width %= 8; + } + + if (width) { + hevc_sao_edge_filter_135degree_4width_msa(dst, stride_dst, + src, stride_src, + sao_offset_val, height); + } + break; + } +} diff --git a/libavcodec/mips/hevcdsp_init_mips.c b/libavcodec/mips/hevcdsp_init_mips.c index e3d98b1c1f..3675b93155 100644 --- a/libavcodec/mips/hevcdsp_init_mips.c +++ b/libavcodec/mips/hevcdsp_init_mips.c @@ -403,6 +403,32 @@ static av_cold void hevc_dsp_init_msa(HEVCDSPContext *c, c->put_hevc_epel_bi_w[6][1][1] = ff_hevc_put_hevc_bi_w_epel_hv24_8_msa; c->put_hevc_epel_bi_w[7][1][1] = ff_hevc_put_hevc_bi_w_epel_hv32_8_msa; + c->sao_band_filter[0] = + c->sao_band_filter[1] = + c->sao_band_filter[2] = + c->sao_band_filter[3] = + c->sao_band_filter[4] = ff_hevc_sao_band_filter_0_8_msa; + + c->sao_edge_filter[0] = + c->sao_edge_filter[1] = + c->sao_edge_filter[2] = + c->sao_edge_filter[3] = + c->sao_edge_filter[4] = ff_hevc_sao_edge_filter_8_msa; + + c->hevc_h_loop_filter_luma = ff_hevc_loop_filter_luma_h_8_msa; + c->hevc_v_loop_filter_luma = ff_hevc_loop_filter_luma_v_8_msa; + + c->hevc_h_loop_filter_chroma = ff_hevc_loop_filter_chroma_h_8_msa; + c->hevc_v_loop_filter_chroma = ff_hevc_loop_filter_chroma_v_8_msa; + + c->hevc_h_loop_filter_luma_c = ff_hevc_loop_filter_luma_h_8_msa; + c->hevc_v_loop_filter_luma_c = ff_hevc_loop_filter_luma_v_8_msa; + + c->hevc_h_loop_filter_chroma_c = + ff_hevc_loop_filter_chroma_h_8_msa; + c->hevc_v_loop_filter_chroma_c = + ff_hevc_loop_filter_chroma_v_8_msa; + c->idct[0] = ff_hevc_idct_4x4_msa; c->idct[1] = ff_hevc_idct_8x8_msa; c->idct[2] = ff_hevc_idct_16x16_msa; diff --git a/libavcodec/mips/hevcdsp_mips.h b/libavcodec/mips/hevcdsp_mips.h index b7c903e80e..f26d7d9a0b 100644 --- a/libavcodec/mips/hevcdsp_mips.h +++ b/libavcodec/mips/hevcdsp_mips.h @@ -431,6 +431,36 @@ BI_W_MC(epel, hv, 64); #undef BI_W_MC +void ff_hevc_loop_filter_luma_h_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t beta, int32_t *tc, + uint8_t *no_p, uint8_t *no_q); + +void ff_hevc_loop_filter_luma_v_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t beta, int32_t *tc, + uint8_t *no_p, uint8_t *no_q); + +void ff_hevc_loop_filter_chroma_h_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t *tc, uint8_t *no_p, + uint8_t *no_q); + +void ff_hevc_loop_filter_chroma_v_8_msa(uint8_t *src, + ptrdiff_t src_stride, + int32_t *tc, uint8_t *no_p, + uint8_t *no_q); + +void ff_hevc_sao_band_filter_0_8_msa(uint8_t *dst, uint8_t *src, + ptrdiff_t stride_dst, ptrdiff_t stride_src, + int16_t *sao_offset_val, int sao_left_class, + int width, int height); + +void ff_hevc_sao_edge_filter_8_msa(uint8_t *dst, uint8_t *src, + ptrdiff_t stride_dst, + int16_t *sao_offset_val, + int eo, int width, int height); + void ff_hevc_idct_4x4_msa(int16_t *coeffs, int col_limit); void ff_hevc_idct_8x8_msa(int16_t *coeffs, int col_limit); void ff_hevc_idct_16x16_msa(int16_t *coeffs, int col_limit); diff --git a/libavutil/mips/generic_macros_msa.h b/libavutil/mips/generic_macros_msa.h index 234aead065..f43798d6bb 100644 --- a/libavutil/mips/generic_macros_msa.h +++ b/libavutil/mips/generic_macros_msa.h @@ -761,6 +761,8 @@ out1 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in1, slide_val); \ } #define SLDI_B2_0_UB(...) SLDI_B2_0(v16u8, __VA_ARGS__) +#define SLDI_B2_0_SB(...) SLDI_B2_0(v16i8, __VA_ARGS__) +#define SLDI_B2_0_SW(...) SLDI_B2_0(v4i32, __VA_ARGS__) #define SLDI_B4_0(RTYPE, in0, in1, in2, in3, \ out0, out1, out2, out3, slide_val) \ @@ -821,6 +823,23 @@ VSHF_B2(RTYPE, in0, in1, in0, in1, mask2, mask3, out2, out3); \ } #define VSHF_B4_SB(...) VSHF_B4(v16i8, __VA_ARGS__) +#define VSHF_B4_SH(...) VSHF_B4(v8i16, __VA_ARGS__) + +/* Description : Shuffle halfword vector elements as per mask vector + Arguments : Inputs - in0, in1, in2, in3, mask0, mask1 + Outputs - out0, out1 + Return Type - as per RTYPE + Details : Selective halfword elements from in0 & in1 are copied to out0 + as per control vector mask0 + Selective halfword elements from in2 & in3 are copied to out1 + as per control vector mask1 +*/ +#define VSHF_H2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ +{ \ + out0 = (RTYPE) __msa_vshf_h((v8i16) mask0, (v8i16) in1, (v8i16) in0); \ + out1 = (RTYPE) __msa_vshf_h((v8i16) mask1, (v8i16) in3, (v8i16) in2); \ +} +#define VSHF_H2_SH(...) VSHF_H2(v8i16, __VA_ARGS__) /* Description : Shuffle byte vector elements as per mask vector Arguments : Inputs - in0, in1, in2, in3, mask0, mask1 @@ -1060,6 +1079,25 @@ #define INSERT_D2_SB(...) INSERT_D2(v16i8, __VA_ARGS__) #define INSERT_D2_SD(...) INSERT_D2(v2i64, __VA_ARGS__) +/* Description : Interleave even byte elements from vectors + Arguments : Inputs - in0, in1, in2, in3 + Outputs - out0, out1 + Return Type - as per RTYPE + Details : Even byte elements of 'in0' and even byte + elements of 'in1' are interleaved and copied to 'out0' + Even byte elements of 'in2' and even byte + elements of 'in3' are interleaved and copied to 'out1' +*/ +#define ILVEV_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ +{ \ + out0 = (RTYPE) __msa_ilvev_b((v16i8) in1, (v16i8) in0); \ + out1 = (RTYPE) __msa_ilvev_b((v16i8) in3, (v16i8) in2); \ +} +#define ILVEV_B2_UB(...) ILVEV_B2(v16u8, __VA_ARGS__) +#define ILVEV_B2_SB(...) ILVEV_B2(v16i8, __VA_ARGS__) +#define ILVEV_B2_SH(...) ILVEV_B2(v8i16, __VA_ARGS__) +#define ILVEV_B2_SD(...) ILVEV_B2(v2i64, __VA_ARGS__) + /* Description : Interleave even halfword elements from vectors Arguments : Inputs - in0, in1, in2, in3 Outputs - out0, out1 @@ -1107,6 +1145,8 @@ out1 = (RTYPE) __msa_ilvev_d((v2i64) in3, (v2i64) in2); \ } #define ILVEV_D2_UB(...) ILVEV_D2(v16u8, __VA_ARGS__) +#define ILVEV_D2_SB(...) ILVEV_D2(v16i8, __VA_ARGS__) +#define ILVEV_D2_SW(...) ILVEV_D2(v4i32, __VA_ARGS__) /* Description : Interleave left half of byte elements from vectors Arguments : Inputs - in0, in1, in2, in3 @@ -1212,9 +1252,22 @@ ILVR_B2(RTYPE, in0, in1, in2, in3, out0, out1); \ ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \ } +#define ILVR_B4_UB(...) ILVR_B4(v16u8, __VA_ARGS__) #define ILVR_B4_SB(...) ILVR_B4(v16i8, __VA_ARGS__) #define ILVR_B4_UH(...) ILVR_B4(v8u16, __VA_ARGS__) #define ILVR_B4_SH(...) ILVR_B4(v8i16, __VA_ARGS__) +#define ILVR_B4_SW(...) ILVR_B4(v4i32, __VA_ARGS__) + +#define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ + in8, in9, in10, in11, in12, in13, in14, in15, \ + out0, out1, out2, out3, out4, out5, out6, out7) \ +{ \ + ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ + out0, out1, out2, out3); \ + ILVR_B4(RTYPE, in8, in9, in10, in11, in12, in13, in14, in15, \ + out4, out5, out6, out7); \ +} +#define ILVR_B8_UH(...) ILVR_B8(v8u16, __VA_ARGS__) /* Description : Interleave right half of halfword elements from vectors Arguments : Inputs - in0, in1, in2, in3, in4, in5, in6, in7 @@ -1870,6 +1923,25 @@ ADD2(in4, in5, in6, in7, out2, out3); \ } +/* Description : Sign extend byte elements from input vector and return + halfword results in pair of vectors + Arguments : Inputs - in (1 input byte vector) + Outputs - out0, out1 (sign extended 2 halfword vectors) + Return Type - signed halfword + Details : Sign bit of byte elements from input vector 'in' is + extracted and interleaved right with same vector 'in0' to + generate 8 signed halfword elements in 'out0' + Then interleaved left with same vector 'in0' to + generate 8 signed halfword elements in 'out1' +*/ +#define UNPCK_SB_SH(in, out0, out1) \ +{ \ + v16i8 tmp_m; \ + \ + tmp_m = __msa_clti_s_b((v16i8) in, 0); \ + ILVRL_B2_SH(tmp_m, in, out0, out1); \ +} + /* Description : Zero extend unsigned byte elements to halfword elements Arguments : Inputs - in (1 input unsigned byte vector) Outputs - out0, out1 (unsigned 2 halfword vectors) @@ -1903,6 +1975,18 @@ ILVRL_H2_SW(tmp_m, in, out0, out1); \ } +/* Description : Swap two variables + Arguments : Inputs - in0, in1 + Outputs - in0, in1 (in-place) + Details : Swapping of two input variables using xor +*/ +#define SWAP(in0, in1) \ +{ \ + in0 = in0 ^ in1; \ + in1 = in0 ^ in1; \ + in0 = in0 ^ in1; \ +} + /* Description : Butterfly of 4 input vectors Arguments : Inputs - in0, in1, in2, in3 Outputs - out0, out1, out2, out3 @@ -1959,9 +2043,34 @@ out1 = (RTYPE) __msa_ilvl_d((v2i64) out2, (v2i64) out0); \ out3 = (RTYPE) __msa_ilvl_d((v2i64) out0, (v2i64) out2); \ } - #define TRANSPOSE8x4_UB_UB(...) TRANSPOSE8x4_UB(v16u8, __VA_ARGS__) +#define TRANSPOSE8x4_UB_UH(...) TRANSPOSE8x4_UB(v8u16, __VA_ARGS__) +/* Description : Transposes input 8x8 byte block + Arguments : Inputs - in0, in1, in2, in3, in4, in5, in6, in7 + (input 8x8 byte block) + Outputs - out0, out1, out2, out3, out4, out5, out6, out7 + (output 8x8 byte block) + Return Type - unsigned byte + Details : +*/ +#define TRANSPOSE8x8_UB(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ + out0, out1, out2, out3, out4, out5, out6, out7) \ +{ \ + v16i8 tmp0_m, tmp1_m, tmp2_m, tmp3_m; \ + v16i8 tmp4_m, tmp5_m, tmp6_m, tmp7_m; \ + \ + ILVR_B4_SB(in2, in0, in3, in1, in6, in4, in7, in5, \ + tmp0_m, tmp1_m, tmp2_m, tmp3_m); \ + ILVRL_B2_SB(tmp1_m, tmp0_m, tmp4_m, tmp5_m); \ + ILVRL_B2_SB(tmp3_m, tmp2_m, tmp6_m, tmp7_m); \ + ILVRL_W2(RTYPE, tmp6_m, tmp4_m, out0, out2); \ + ILVRL_W2(RTYPE, tmp7_m, tmp5_m, out4, out6); \ + SLDI_B2_0(RTYPE, out0, out2, out1, out3, 8); \ + SLDI_B2_0(RTYPE, out4, out6, out5, out7, 8); \ +} +#define TRANSPOSE8x8_UB_UB(...) TRANSPOSE8x8_UB(v16u8, __VA_ARGS__) +#define TRANSPOSE8x8_UB_UH(...) TRANSPOSE8x8_UB(v8u16, __VA_ARGS__) /* Description : Transposes 16x8 block into 8x16 with byte elements in vectors Arguments : Inputs - in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, in12, in13, in14, in15