Source code

Revision control

Copy as Markdown

Other Tools

/*
* Copyright 2022 The LibYuv Project Authors. All rights reserved.
*
* Copyright (c) 2022 Loongson Technology Corporation Limited
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "libyuv/row.h"
#if !defined(LIBYUV_DISABLE_LSX) && defined(__loongarch_sx)
#include "libyuv/loongson_intrinsics.h"
#ifdef __cplusplus
namespace libyuv {
extern "C" {
#endif
// Fill YUV -> RGB conversion constants into vectors
#define YUVTORGB_SETUP(yuvconst, vr, ub, vg, ug, yg, yb) \
{ \
ub = __lsx_vreplgr2vr_h(yuvconst->kUVToB[0]); \
vr = __lsx_vreplgr2vr_h(yuvconst->kUVToR[1]); \
ug = __lsx_vreplgr2vr_h(yuvconst->kUVToG[0]); \
vg = __lsx_vreplgr2vr_h(yuvconst->kUVToG[1]); \
yg = __lsx_vreplgr2vr_h(yuvconst->kYToRgb[0]); \
yb = __lsx_vreplgr2vr_w(yuvconst->kYBiasToRgb[0]); \
}
// Convert 8 pixels of YUV420 to RGB.
#define YUVTORGB(in_y, in_vu, vrub, vgug, yg, yb, out_b, out_g, out_r) \
{ \
__m128i y_ev, y_od, u_l, v_l; \
__m128i tmp0, tmp1, tmp2, tmp3; \
\
tmp0 = __lsx_vilvl_b(in_y, in_y); \
y_ev = __lsx_vmulwev_w_hu_h(tmp0, yg); \
y_od = __lsx_vmulwod_w_hu_h(tmp0, yg); \
y_ev = __lsx_vsrai_w(y_ev, 16); \
y_od = __lsx_vsrai_w(y_od, 16); \
y_ev = __lsx_vadd_w(y_ev, yb); \
y_od = __lsx_vadd_w(y_od, yb); \
in_vu = __lsx_vilvl_b(zero, in_vu); \
in_vu = __lsx_vsub_h(in_vu, const_80); \
u_l = __lsx_vmulwev_w_h(in_vu, vrub); \
v_l = __lsx_vmulwod_w_h(in_vu, vrub); \
tmp0 = __lsx_vadd_w(y_ev, u_l); \
tmp1 = __lsx_vadd_w(y_od, u_l); \
tmp2 = __lsx_vadd_w(y_ev, v_l); \
tmp3 = __lsx_vadd_w(y_od, v_l); \
tmp0 = __lsx_vsrai_w(tmp0, 6); \
tmp1 = __lsx_vsrai_w(tmp1, 6); \
tmp2 = __lsx_vsrai_w(tmp2, 6); \
tmp3 = __lsx_vsrai_w(tmp3, 6); \
tmp0 = __lsx_vclip255_w(tmp0); \
tmp1 = __lsx_vclip255_w(tmp1); \
tmp2 = __lsx_vclip255_w(tmp2); \
tmp3 = __lsx_vclip255_w(tmp3); \
out_b = __lsx_vpackev_h(tmp1, tmp0); \
out_r = __lsx_vpackev_h(tmp3, tmp2); \
tmp0 = __lsx_vdp2_w_h(in_vu, vgug); \
tmp1 = __lsx_vsub_w(y_ev, tmp0); \
tmp2 = __lsx_vsub_w(y_od, tmp0); \
tmp1 = __lsx_vsrai_w(tmp1, 6); \
tmp2 = __lsx_vsrai_w(tmp2, 6); \
tmp1 = __lsx_vclip255_w(tmp1); \
tmp2 = __lsx_vclip255_w(tmp2); \
out_g = __lsx_vpackev_h(tmp2, tmp1); \
}
// Convert I444 pixels of YUV420 to RGB.
#define I444TORGB(in_yy, in_u, in_v, ub, vr, ugvg, yg, yb, out_b, out_g, \
out_r) \
{ \
__m128i y_ev, y_od, u_ev, v_ev, u_od, v_od; \
__m128i tmp0, tmp1, tmp2, tmp3; \
\
y_ev = __lsx_vmulwev_w_hu_h(in_yy, yg); \
y_od = __lsx_vmulwod_w_hu_h(in_yy, yg); \
y_ev = __lsx_vsrai_w(y_ev, 16); \
y_od = __lsx_vsrai_w(y_od, 16); \
y_ev = __lsx_vadd_w(y_ev, yb); \
y_od = __lsx_vadd_w(y_od, yb); \
in_u = __lsx_vsub_h(in_u, const_80); \
in_v = __lsx_vsub_h(in_v, const_80); \
u_ev = __lsx_vmulwev_w_h(in_u, ub); \
u_od = __lsx_vmulwod_w_h(in_u, ub); \
v_ev = __lsx_vmulwev_w_h(in_v, vr); \
v_od = __lsx_vmulwod_w_h(in_v, vr); \
tmp0 = __lsx_vadd_w(y_ev, u_ev); \
tmp1 = __lsx_vadd_w(y_od, u_od); \
tmp2 = __lsx_vadd_w(y_ev, v_ev); \
tmp3 = __lsx_vadd_w(y_od, v_od); \
tmp0 = __lsx_vsrai_w(tmp0, 6); \
tmp1 = __lsx_vsrai_w(tmp1, 6); \
tmp2 = __lsx_vsrai_w(tmp2, 6); \
tmp3 = __lsx_vsrai_w(tmp3, 6); \
tmp0 = __lsx_vclip255_w(tmp0); \
tmp1 = __lsx_vclip255_w(tmp1); \
tmp2 = __lsx_vclip255_w(tmp2); \
tmp3 = __lsx_vclip255_w(tmp3); \
out_b = __lsx_vpackev_h(tmp1, tmp0); \
out_r = __lsx_vpackev_h(tmp3, tmp2); \
u_ev = __lsx_vpackev_h(in_u, in_v); \
u_od = __lsx_vpackod_h(in_u, in_v); \
v_ev = __lsx_vdp2_w_h(u_ev, ugvg); \
v_od = __lsx_vdp2_w_h(u_od, ugvg); \
tmp0 = __lsx_vsub_w(y_ev, v_ev); \
tmp1 = __lsx_vsub_w(y_od, v_od); \
tmp0 = __lsx_vsrai_w(tmp0, 6); \
tmp1 = __lsx_vsrai_w(tmp1, 6); \
tmp0 = __lsx_vclip255_w(tmp0); \
tmp1 = __lsx_vclip255_w(tmp1); \
out_g = __lsx_vpackev_h(tmp1, tmp0); \
}
// Pack and Store 8 ARGB values.
#define STOREARGB(in_a, in_r, in_g, in_b, pdst_argb) \
{ \
__m128i temp0, temp1; \
__m128i dst0, dst1; \
\
temp0 = __lsx_vpackev_b(in_g, in_b); \
temp1 = __lsx_vpackev_b(in_a, in_r); \
dst0 = __lsx_vilvl_h(temp1, temp0); \
dst1 = __lsx_vilvh_h(temp1, temp0); \
__lsx_vst(dst0, pdst_argb, 0); \
__lsx_vst(dst1, pdst_argb, 16); \
pdst_argb += 32; \
}
#define RGBTOUV(_tmpb, _tmpg, _tmpr, _nexb, _nexg, _nexr, _dst0) \
{ \
__m128i _tmp0, _tmp1, _tmp2, _tmp3; \
__m128i _reg0, _reg1; \
_tmp0 = __lsx_vaddwev_h_bu(_tmpb, _nexb); \
_tmp1 = __lsx_vaddwod_h_bu(_tmpb, _nexb); \
_tmp2 = __lsx_vaddwev_h_bu(_tmpg, _nexg); \
_tmp3 = __lsx_vaddwod_h_bu(_tmpg, _nexg); \
_reg0 = __lsx_vaddwev_h_bu(_tmpr, _nexr); \
_reg1 = __lsx_vaddwod_h_bu(_tmpr, _nexr); \
_tmpb = __lsx_vavgr_hu(_tmp0, _tmp1); \
_tmpg = __lsx_vavgr_hu(_tmp2, _tmp3); \
_tmpr = __lsx_vavgr_hu(_reg0, _reg1); \
_reg0 = __lsx_vmadd_h(const_8080, const_112, _tmpb); \
_reg1 = __lsx_vmadd_h(const_8080, const_112, _tmpr); \
_reg0 = __lsx_vmsub_h(_reg0, const_74, _tmpg); \
_reg1 = __lsx_vmsub_h(_reg1, const_94, _tmpg); \
_reg0 = __lsx_vmsub_h(_reg0, const_38, _tmpr); \
_reg1 = __lsx_vmsub_h(_reg1, const_18, _tmpb); \
_dst0 = __lsx_vpickod_b(_reg1, _reg0); \
}
void ARGB4444ToARGBRow_LSX(const uint8_t* src_argb4444,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src0, src1;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i reg0, reg1, reg2, reg3;
__m128i dst0, dst1, dst2, dst3;
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_argb4444, 0);
src1 = __lsx_vld(src_argb4444, 16);
tmp0 = __lsx_vandi_b(src0, 0x0F);
tmp1 = __lsx_vandi_b(src0, 0xF0);
tmp2 = __lsx_vandi_b(src1, 0x0F);
tmp3 = __lsx_vandi_b(src1, 0xF0);
reg0 = __lsx_vslli_b(tmp0, 4);
reg2 = __lsx_vslli_b(tmp2, 4);
reg1 = __lsx_vsrli_b(tmp1, 4);
reg3 = __lsx_vsrli_b(tmp3, 4);
DUP4_ARG2(__lsx_vor_v, tmp0, reg0, tmp1, reg1, tmp2, reg2, tmp3, reg3, tmp0,
tmp1, tmp2, tmp3);
dst0 = __lsx_vilvl_b(tmp1, tmp0);
dst2 = __lsx_vilvl_b(tmp3, tmp2);
dst1 = __lsx_vilvh_b(tmp1, tmp0);
dst3 = __lsx_vilvh_b(tmp3, tmp2);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_argb4444 += 32;
}
}
void ARGB1555ToARGBRow_LSX(const uint8_t* src_argb1555,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src0, src1;
__m128i tmp0, tmp1, tmpb, tmpg, tmpr, tmpa;
__m128i reg0, reg1, reg2;
__m128i dst0, dst1, dst2, dst3;
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_argb1555, 0);
src1 = __lsx_vld(src_argb1555, 16);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
tmpg = __lsx_vsrli_b(tmp0, 5);
reg0 = __lsx_vandi_b(tmp1, 0x03);
reg0 = __lsx_vslli_b(reg0, 3);
tmpg = __lsx_vor_v(tmpg, reg0);
reg1 = __lsx_vandi_b(tmp1, 0x7C);
tmpr = __lsx_vsrli_b(reg1, 2);
tmpa = __lsx_vsrli_b(tmp1, 7);
tmpa = __lsx_vneg_b(tmpa);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vslli_b(tmpg, 3);
reg2 = __lsx_vslli_b(tmpr, 3);
tmpb = __lsx_vsrli_b(tmpb, 2);
tmpg = __lsx_vsrli_b(tmpg, 2);
tmpr = __lsx_vsrli_b(tmpr, 2);
tmpb = __lsx_vor_v(reg0, tmpb);
tmpg = __lsx_vor_v(reg1, tmpg);
tmpr = __lsx_vor_v(reg2, tmpr);
DUP2_ARG2(__lsx_vilvl_b, tmpg, tmpb, tmpa, tmpr, reg0, reg1);
dst0 = __lsx_vilvl_h(reg1, reg0);
dst1 = __lsx_vilvh_h(reg1, reg0);
DUP2_ARG2(__lsx_vilvh_b, tmpg, tmpb, tmpa, tmpr, reg0, reg1);
dst2 = __lsx_vilvl_h(reg1, reg0);
dst3 = __lsx_vilvh_h(reg1, reg0);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_argb1555 += 32;
}
}
void RGB565ToARGBRow_LSX(const uint8_t* src_rgb565,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src0, src1;
__m128i tmp0, tmp1, tmpb, tmpg, tmpr;
__m128i reg0, reg1, dst0, dst1, dst2, dst3;
__m128i alpha = __lsx_vldi(0xFF);
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_rgb565, 0);
src1 = __lsx_vld(src_rgb565, 16);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
tmpr = __lsx_vandi_b(tmp1, 0xF8);
reg1 = __lsx_vandi_b(tmp1, 0x07);
reg0 = __lsx_vsrli_b(tmp0, 5);
reg1 = __lsx_vslli_b(reg1, 3);
tmpg = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vsrli_b(tmpb, 2);
tmpb = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vslli_b(tmpg, 2);
reg1 = __lsx_vsrli_b(tmpg, 4);
tmpg = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vsrli_b(tmpr, 5);
tmpr = __lsx_vor_v(tmpr, reg0);
DUP2_ARG2(__lsx_vilvl_b, tmpg, tmpb, alpha, tmpr, reg0, reg1);
dst0 = __lsx_vilvl_h(reg1, reg0);
dst1 = __lsx_vilvh_h(reg1, reg0);
DUP2_ARG2(__lsx_vilvh_b, tmpg, tmpb, alpha, tmpr, reg0, reg1);
dst2 = __lsx_vilvl_h(reg1, reg0);
dst3 = __lsx_vilvh_h(reg1, reg0);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_rgb565 += 32;
}
}
void RGB24ToARGBRow_LSX(const uint8_t* src_rgb24,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src0, src1, src2;
__m128i tmp0, tmp1, tmp2;
__m128i dst0, dst1, dst2, dst3;
__m128i alpha = __lsx_vldi(0xFF);
__m128i shuf0 = {0x131211100F0E0D0C, 0x1B1A191817161514};
__m128i shuf1 = {0x1F1E1D1C1B1A1918, 0x0706050403020100};
__m128i shuf2 = {0x0B0A090807060504, 0x131211100F0E0D0C};
__m128i shuf3 = {0x1005040310020100, 0x100B0A0910080706};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_rgb24, 0);
src1 = __lsx_vld(src_rgb24, 16);
src2 = __lsx_vld(src_rgb24, 32);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuf0, src1, src2, shuf1, tmp0, tmp1);
tmp2 = __lsx_vshuf_b(src1, src2, shuf2);
DUP4_ARG3(__lsx_vshuf_b, alpha, src0, shuf3, alpha, tmp0, shuf3, alpha,
tmp1, shuf3, alpha, tmp2, shuf3, dst0, dst1, dst2, dst3);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_rgb24 += 48;
}
}
void RAWToARGBRow_LSX(const uint8_t* src_raw, uint8_t* dst_argb, int width) {
int x;
int len = width / 16;
__m128i src0, src1, src2;
__m128i tmp0, tmp1, tmp2;
__m128i dst0, dst1, dst2, dst3;
__m128i alpha = __lsx_vldi(0xFF);
__m128i shuf0 = {0x131211100F0E0D0C, 0x1B1A191817161514};
__m128i shuf1 = {0x1F1E1D1C1B1A1918, 0x0706050403020100};
__m128i shuf2 = {0x0B0A090807060504, 0x131211100F0E0D0C};
__m128i shuf3 = {0x1003040510000102, 0x10090A0B10060708};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_raw, 0);
src1 = __lsx_vld(src_raw, 16);
src2 = __lsx_vld(src_raw, 32);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuf0, src1, src2, shuf1, tmp0, tmp1);
tmp2 = __lsx_vshuf_b(src1, src2, shuf2);
DUP4_ARG3(__lsx_vshuf_b, alpha, src0, shuf3, alpha, tmp0, shuf3, alpha,
tmp1, shuf3, alpha, tmp2, shuf3, dst0, dst1, dst2, dst3);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_raw += 48;
}
}
void ARGB1555ToYRow_LSX(const uint8_t* src_argb1555,
uint8_t* dst_y,
int width) {
int x;
int len = width / 16;
__m128i src0, src1;
__m128i tmp0, tmp1, tmpb, tmpg, tmpr;
__m128i reg0, reg1, reg2, dst0;
__m128i const_66 = __lsx_vldi(66);
__m128i const_129 = __lsx_vldi(129);
__m128i const_25 = __lsx_vldi(25);
__m128i const_1080 = {0x1080108010801080, 0x1080108010801080};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_argb1555, 0);
src1 = __lsx_vld(src_argb1555, 16);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
tmpg = __lsx_vsrli_b(tmp0, 5);
reg0 = __lsx_vandi_b(tmp1, 0x03);
reg0 = __lsx_vslli_b(reg0, 3);
tmpg = __lsx_vor_v(tmpg, reg0);
reg1 = __lsx_vandi_b(tmp1, 0x7C);
tmpr = __lsx_vsrli_b(reg1, 2);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vslli_b(tmpg, 3);
reg2 = __lsx_vslli_b(tmpr, 3);
tmpb = __lsx_vsrli_b(tmpb, 2);
tmpg = __lsx_vsrli_b(tmpg, 2);
tmpr = __lsx_vsrli_b(tmpr, 2);
tmpb = __lsx_vor_v(reg0, tmpb);
tmpg = __lsx_vor_v(reg1, tmpg);
tmpr = __lsx_vor_v(reg2, tmpr);
reg0 = __lsx_vmaddwev_h_bu(const_1080, tmpb, const_25);
reg1 = __lsx_vmaddwod_h_bu(const_1080, tmpb, const_25);
reg0 = __lsx_vmaddwev_h_bu(reg0, tmpg, const_129);
reg1 = __lsx_vmaddwod_h_bu(reg1, tmpg, const_129);
reg0 = __lsx_vmaddwev_h_bu(reg0, tmpr, const_66);
reg1 = __lsx_vmaddwod_h_bu(reg1, tmpr, const_66);
dst0 = __lsx_vpackod_b(reg1, reg0);
__lsx_vst(dst0, dst_y, 0);
dst_y += 16;
src_argb1555 += 32;
}
}
void ARGB1555ToUVRow_LSX(const uint8_t* src_argb1555,
int src_stride_argb1555,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
int len = width / 16;
const uint8_t* next_argb1555 = src_argb1555 + src_stride_argb1555;
__m128i src0, src1, src2, src3;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i reg0, reg1, reg2, reg3, dst0;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_argb1555, 0, src_argb1555, 16, next_argb1555, 0,
next_argb1555, 16, src0, src1, src2, src3);
DUP2_ARG2(__lsx_vpickev_b, src1, src0, src3, src2, tmp0, tmp2);
DUP2_ARG2(__lsx_vpickod_b, src1, src0, src3, src2, tmp1, tmp3);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
nexb = __lsx_vandi_b(tmp2, 0x1F);
tmpg = __lsx_vsrli_b(tmp0, 5);
nexg = __lsx_vsrli_b(tmp2, 5);
reg0 = __lsx_vandi_b(tmp1, 0x03);
reg2 = __lsx_vandi_b(tmp3, 0x03);
reg0 = __lsx_vslli_b(reg0, 3);
reg2 = __lsx_vslli_b(reg2, 3);
tmpg = __lsx_vor_v(tmpg, reg0);
nexg = __lsx_vor_v(nexg, reg2);
reg1 = __lsx_vandi_b(tmp1, 0x7C);
reg3 = __lsx_vandi_b(tmp3, 0x7C);
tmpr = __lsx_vsrli_b(reg1, 2);
nexr = __lsx_vsrli_b(reg3, 2);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vslli_b(tmpg, 3);
reg2 = __lsx_vslli_b(tmpr, 3);
tmpb = __lsx_vsrli_b(tmpb, 2);
tmpg = __lsx_vsrli_b(tmpg, 2);
tmpr = __lsx_vsrli_b(tmpr, 2);
tmpb = __lsx_vor_v(reg0, tmpb);
tmpg = __lsx_vor_v(reg1, tmpg);
tmpr = __lsx_vor_v(reg2, tmpr);
reg0 = __lsx_vslli_b(nexb, 3);
reg1 = __lsx_vslli_b(nexg, 3);
reg2 = __lsx_vslli_b(nexr, 3);
nexb = __lsx_vsrli_b(nexb, 2);
nexg = __lsx_vsrli_b(nexg, 2);
nexr = __lsx_vsrli_b(nexr, 2);
nexb = __lsx_vor_v(reg0, nexb);
nexg = __lsx_vor_v(reg1, nexg);
nexr = __lsx_vor_v(reg2, nexr);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_argb1555 += 32;
next_argb1555 += 32;
}
}
void RGB565ToYRow_LSX(const uint8_t* src_rgb565, uint8_t* dst_y, int width) {
int x;
int len = width / 16;
__m128i src0, src1;
__m128i tmp0, tmp1, tmpb, tmpg, tmpr;
__m128i reg0, reg1, dst0;
__m128i const_66 = __lsx_vldi(66);
__m128i const_129 = __lsx_vldi(129);
__m128i const_25 = __lsx_vldi(25);
__m128i const_1080 = {0x1080108010801080, 0x1080108010801080};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_rgb565, 0);
src1 = __lsx_vld(src_rgb565, 16);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
tmpr = __lsx_vandi_b(tmp1, 0xF8);
reg1 = __lsx_vandi_b(tmp1, 0x07);
reg0 = __lsx_vsrli_b(tmp0, 5);
reg1 = __lsx_vslli_b(reg1, 3);
tmpg = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vsrli_b(tmpb, 2);
tmpb = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vslli_b(tmpg, 2);
reg1 = __lsx_vsrli_b(tmpg, 4);
tmpg = __lsx_vor_v(reg1, reg0);
reg0 = __lsx_vsrli_b(tmpr, 5);
tmpr = __lsx_vor_v(tmpr, reg0);
reg0 = __lsx_vmaddwev_h_bu(const_1080, tmpb, const_25);
reg1 = __lsx_vmaddwod_h_bu(const_1080, tmpb, const_25);
reg0 = __lsx_vmaddwev_h_bu(reg0, tmpg, const_129);
reg1 = __lsx_vmaddwod_h_bu(reg1, tmpg, const_129);
reg0 = __lsx_vmaddwev_h_bu(reg0, tmpr, const_66);
reg1 = __lsx_vmaddwod_h_bu(reg1, tmpr, const_66);
dst0 = __lsx_vpackod_b(reg1, reg0);
__lsx_vst(dst0, dst_y, 0);
dst_y += 16;
src_rgb565 += 32;
}
}
void RGB565ToUVRow_LSX(const uint8_t* src_rgb565,
int src_stride_rgb565,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
int len = width / 16;
const uint8_t* next_rgb565 = src_rgb565 + src_stride_rgb565;
__m128i src0, src1, src2, src3;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i reg0, reg1, reg2, reg3, dst0;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_rgb565, 0, src_rgb565, 16, next_rgb565, 0,
next_rgb565, 16, src0, src1, src2, src3);
DUP2_ARG2(__lsx_vpickev_b, src1, src0, src3, src2, tmp0, tmp2);
DUP2_ARG2(__lsx_vpickod_b, src1, src0, src3, src2, tmp1, tmp3);
tmpb = __lsx_vandi_b(tmp0, 0x1F);
tmpr = __lsx_vandi_b(tmp1, 0xF8);
nexb = __lsx_vandi_b(tmp2, 0x1F);
nexr = __lsx_vandi_b(tmp3, 0xF8);
reg1 = __lsx_vandi_b(tmp1, 0x07);
reg3 = __lsx_vandi_b(tmp3, 0x07);
reg0 = __lsx_vsrli_b(tmp0, 5);
reg1 = __lsx_vslli_b(reg1, 3);
reg2 = __lsx_vsrli_b(tmp2, 5);
reg3 = __lsx_vslli_b(reg3, 3);
tmpg = __lsx_vor_v(reg1, reg0);
nexg = __lsx_vor_v(reg2, reg3);
reg0 = __lsx_vslli_b(tmpb, 3);
reg1 = __lsx_vsrli_b(tmpb, 2);
reg2 = __lsx_vslli_b(nexb, 3);
reg3 = __lsx_vsrli_b(nexb, 2);
tmpb = __lsx_vor_v(reg1, reg0);
nexb = __lsx_vor_v(reg2, reg3);
reg0 = __lsx_vslli_b(tmpg, 2);
reg1 = __lsx_vsrli_b(tmpg, 4);
reg2 = __lsx_vslli_b(nexg, 2);
reg3 = __lsx_vsrli_b(nexg, 4);
tmpg = __lsx_vor_v(reg1, reg0);
nexg = __lsx_vor_v(reg2, reg3);
reg0 = __lsx_vsrli_b(tmpr, 5);
reg2 = __lsx_vsrli_b(nexr, 5);
tmpr = __lsx_vor_v(tmpr, reg0);
nexr = __lsx_vor_v(nexr, reg2);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_rgb565 += 32;
next_rgb565 += 32;
}
}
void RGB24ToUVRow_LSX(const uint8_t* src_rgb24,
int src_stride_rgb24,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_rgb24 = src_rgb24 + src_stride_rgb24;
int len = width / 16;
__m128i src0, src1, src2;
__m128i nex0, nex1, nex2, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
__m128i shuff0_b = {0x15120F0C09060300, 0x00000000001E1B18};
__m128i shuff1_b = {0x0706050403020100, 0x1D1A1714110A0908};
__m128i shuff0_g = {0x1613100D0A070401, 0x00000000001F1C19};
__m128i shuff1_g = {0x0706050403020100, 0x1E1B1815120A0908};
__m128i shuff0_r = {0x1714110E0B080502, 0x0000000000001D1A};
__m128i shuff1_r = {0x0706050403020100, 0x1F1C191613100908};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_rgb24, 0);
src1 = __lsx_vld(src_rgb24, 16);
src2 = __lsx_vld(src_rgb24, 32);
nex0 = __lsx_vld(next_rgb24, 0);
nex1 = __lsx_vld(next_rgb24, 16);
nex2 = __lsx_vld(next_rgb24, 32);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_b, nex1, nex0, shuff0_b, tmpb,
nexb);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_g, nex1, nex0, shuff0_g, tmpg,
nexg);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_r, nex1, nex0, shuff0_r, tmpr,
nexr);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpb, shuff1_b, nex2, nexb, shuff1_b, tmpb,
nexb);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpg, shuff1_g, nex2, nexg, shuff1_g, tmpg,
nexg);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpr, shuff1_r, nex2, nexr, shuff1_r, tmpr,
nexr);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_rgb24 += 48;
next_rgb24 += 48;
}
}
void RAWToUVRow_LSX(const uint8_t* src_raw,
int src_stride_raw,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_raw = src_raw + src_stride_raw;
int len = width / 16;
__m128i src0, src1, src2;
__m128i nex0, nex1, nex2, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
__m128i shuff0_r = {0x15120F0C09060300, 0x00000000001E1B18};
__m128i shuff1_r = {0x0706050403020100, 0x1D1A1714110A0908};
__m128i shuff0_g = {0x1613100D0A070401, 0x00000000001F1C19};
__m128i shuff1_g = {0x0706050403020100, 0x1E1B1815120A0908};
__m128i shuff0_b = {0x1714110E0B080502, 0x0000000000001D1A};
__m128i shuff1_b = {0x0706050403020100, 0x1F1C191613100908};
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_raw, 0);
src1 = __lsx_vld(src_raw, 16);
src2 = __lsx_vld(src_raw, 32);
nex0 = __lsx_vld(next_raw, 0);
nex1 = __lsx_vld(next_raw, 16);
nex2 = __lsx_vld(next_raw, 32);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_b, nex1, nex0, shuff0_b, tmpb,
nexb);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_g, nex1, nex0, shuff0_g, tmpg,
nexg);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuff0_r, nex1, nex0, shuff0_r, tmpr,
nexr);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpb, shuff1_b, nex2, nexb, shuff1_b, tmpb,
nexb);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpg, shuff1_g, nex2, nexg, shuff1_g, tmpg,
nexg);
DUP2_ARG3(__lsx_vshuf_b, src2, tmpr, shuff1_r, nex2, nexr, shuff1_r, tmpr,
nexr);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_raw += 48;
next_raw += 48;
}
}
void NV12ToARGBRow_LSX(const uint8_t* src_y,
const uint8_t* src_uv,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 8;
__m128i vec_y, vec_vu;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb;
__m128i vec_vrub, vec_vgug;
__m128i out_b, out_g, out_r;
__m128i const_80 = __lsx_vldi(0x480);
__m128i alpha = __lsx_vldi(0xFF);
__m128i zero = __lsx_vldi(0);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_vrub = __lsx_vilvl_h(vec_vr, vec_ub);
vec_vgug = __lsx_vilvl_h(vec_vg, vec_ug);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
vec_vu = __lsx_vld(src_uv, 0);
YUVTORGB(vec_y, vec_vu, vec_vrub, vec_vgug, vec_yg, vec_yb, out_b, out_g,
out_r);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
src_y += 8;
src_uv += 8;
}
}
void NV12ToRGB565Row_LSX(const uint8_t* src_y,
const uint8_t* src_uv,
uint8_t* dst_rgb565,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 8;
__m128i vec_y, vec_vu;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb;
__m128i vec_vrub, vec_vgug;
__m128i out_b, out_g, out_r;
__m128i const_80 = __lsx_vldi(0x480);
__m128i zero = __lsx_vldi(0);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_vrub = __lsx_vilvl_h(vec_vr, vec_ub);
vec_vgug = __lsx_vilvl_h(vec_vg, vec_ug);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
vec_vu = __lsx_vld(src_uv, 0);
YUVTORGB(vec_y, vec_vu, vec_vrub, vec_vgug, vec_yg, vec_yb, out_b, out_g,
out_r);
out_b = __lsx_vsrli_h(out_b, 3);
out_g = __lsx_vsrli_h(out_g, 2);
out_r = __lsx_vsrli_h(out_r, 3);
out_g = __lsx_vslli_h(out_g, 5);
out_r = __lsx_vslli_h(out_r, 11);
out_r = __lsx_vor_v(out_r, out_g);
out_r = __lsx_vor_v(out_r, out_b);
__lsx_vst(out_r, dst_rgb565, 0);
src_y += 8;
src_uv += 8;
dst_rgb565 += 16;
}
}
void NV21ToARGBRow_LSX(const uint8_t* src_y,
const uint8_t* src_vu,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 8;
__m128i vec_y, vec_uv;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb;
__m128i vec_ubvr, vec_ugvg;
__m128i out_b, out_g, out_r;
__m128i const_80 = __lsx_vldi(0x480);
__m128i alpha = __lsx_vldi(0xFF);
__m128i zero = __lsx_vldi(0);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_ubvr = __lsx_vilvl_h(vec_ub, vec_vr);
vec_ugvg = __lsx_vilvl_h(vec_ug, vec_vg);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
vec_uv = __lsx_vld(src_vu, 0);
YUVTORGB(vec_y, vec_uv, vec_ubvr, vec_ugvg, vec_yg, vec_yb, out_r, out_g,
out_b);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
src_y += 8;
src_vu += 8;
}
}
void SobelRow_LSX(const uint8_t* src_sobelx,
const uint8_t* src_sobely,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src0, src1, tmp0;
__m128i out0, out1, out2, out3;
__m128i alpha = __lsx_vldi(0xFF);
__m128i shuff0 = {0x1001010110000000, 0x1003030310020202};
__m128i shuff1 = __lsx_vaddi_bu(shuff0, 0x04);
__m128i shuff2 = __lsx_vaddi_bu(shuff1, 0x04);
__m128i shuff3 = __lsx_vaddi_bu(shuff2, 0x04);
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_sobelx, 0);
src1 = __lsx_vld(src_sobely, 0);
tmp0 = __lsx_vsadd_bu(src0, src1);
DUP4_ARG3(__lsx_vshuf_b, alpha, tmp0, shuff0, alpha, tmp0, shuff1, alpha,
tmp0, shuff2, alpha, tmp0, shuff3, out0, out1, out2, out3);
__lsx_vst(out0, dst_argb, 0);
__lsx_vst(out1, dst_argb, 16);
__lsx_vst(out2, dst_argb, 32);
__lsx_vst(out3, dst_argb, 48);
src_sobelx += 16;
src_sobely += 16;
dst_argb += 64;
}
}
void SobelToPlaneRow_LSX(const uint8_t* src_sobelx,
const uint8_t* src_sobely,
uint8_t* dst_y,
int width) {
int x;
int len = width / 32;
__m128i src0, src1, src2, src3, dst0, dst1;
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_sobelx, 0, src_sobelx, 16, src0, src1);
DUP2_ARG2(__lsx_vld, src_sobely, 0, src_sobely, 16, src2, src3);
dst0 = __lsx_vsadd_bu(src0, src2);
dst1 = __lsx_vsadd_bu(src1, src3);
__lsx_vst(dst0, dst_y, 0);
__lsx_vst(dst1, dst_y, 16);
src_sobelx += 32;
src_sobely += 32;
dst_y += 32;
}
}
void SobelXYRow_LSX(const uint8_t* src_sobelx,
const uint8_t* src_sobely,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 16;
__m128i src_r, src_b, src_g;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i dst0, dst1, dst2, dst3;
__m128i alpha = __lsx_vldi(0xFF);
for (x = 0; x < len; x++) {
src_r = __lsx_vld(src_sobelx, 0);
src_b = __lsx_vld(src_sobely, 0);
src_g = __lsx_vsadd_bu(src_r, src_b);
tmp0 = __lsx_vilvl_b(src_g, src_b);
tmp1 = __lsx_vilvh_b(src_g, src_b);
tmp2 = __lsx_vilvl_b(alpha, src_r);
tmp3 = __lsx_vilvh_b(alpha, src_r);
dst0 = __lsx_vilvl_h(tmp2, tmp0);
dst1 = __lsx_vilvh_h(tmp2, tmp0);
dst2 = __lsx_vilvl_h(tmp3, tmp1);
dst3 = __lsx_vilvh_h(tmp3, tmp1);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
src_sobelx += 16;
src_sobely += 16;
dst_argb += 64;
}
}
void BGRAToUVRow_LSX(const uint8_t* src_bgra,
int src_stride_bgra,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_bgra = src_bgra + src_stride_bgra;
int len = width / 16;
__m128i src0, src1, src2, src3;
__m128i nex0, nex1, nex2, nex3;
__m128i tmp0, tmp1, tmp2, tmp3, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_bgra, 0, src_bgra, 16, src_bgra, 32, src_bgra, 48,
src0, src1, src2, src3);
DUP4_ARG2(__lsx_vld, next_bgra, 0, next_bgra, 16, next_bgra, 32, next_bgra,
48, nex0, nex1, nex2, nex3);
tmp0 = __lsx_vpickod_b(src1, src0);
tmp1 = __lsx_vpickev_b(src1, src0);
tmp2 = __lsx_vpickod_b(src3, src2);
tmp3 = __lsx_vpickev_b(src3, src2);
tmpb = __lsx_vpickod_b(tmp2, tmp0);
tmpr = __lsx_vpickev_b(tmp2, tmp0);
tmpg = __lsx_vpickod_b(tmp3, tmp1);
tmp0 = __lsx_vpickod_b(nex1, nex0);
tmp1 = __lsx_vpickev_b(nex1, nex0);
tmp2 = __lsx_vpickod_b(nex3, nex2);
tmp3 = __lsx_vpickev_b(nex3, nex2);
nexb = __lsx_vpickod_b(tmp2, tmp0);
nexr = __lsx_vpickev_b(tmp2, tmp0);
nexg = __lsx_vpickod_b(tmp3, tmp1);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_bgra += 64;
next_bgra += 64;
}
}
void ABGRToUVRow_LSX(const uint8_t* src_abgr,
int src_stride_abgr,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_abgr = src_abgr + src_stride_abgr;
int len = width / 16;
__m128i src0, src1, src2, src3;
__m128i nex0, nex1, nex2, nex3;
__m128i tmp0, tmp1, tmp2, tmp3, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_abgr, 0, src_abgr, 16, src_abgr, 32, src_abgr, 48,
src0, src1, src2, src3);
DUP4_ARG2(__lsx_vld, next_abgr, 0, next_abgr, 16, next_abgr, 32, next_abgr,
48, nex0, nex1, nex2, nex3);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmp2 = __lsx_vpickev_b(src3, src2);
tmp3 = __lsx_vpickod_b(src3, src2);
tmpb = __lsx_vpickod_b(tmp2, tmp0);
tmpr = __lsx_vpickev_b(tmp2, tmp0);
tmpg = __lsx_vpickev_b(tmp3, tmp1);
tmp0 = __lsx_vpickev_b(nex1, nex0);
tmp1 = __lsx_vpickod_b(nex1, nex0);
tmp2 = __lsx_vpickev_b(nex3, nex2);
tmp3 = __lsx_vpickod_b(nex3, nex2);
nexb = __lsx_vpickod_b(tmp2, tmp0);
nexr = __lsx_vpickev_b(tmp2, tmp0);
nexg = __lsx_vpickev_b(tmp3, tmp1);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_abgr += 64;
next_abgr += 64;
}
}
void RGBAToUVRow_LSX(const uint8_t* src_rgba,
int src_stride_rgba,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_rgba = src_rgba + src_stride_rgba;
int len = width / 16;
__m128i src0, src1, src2, src3;
__m128i nex0, nex1, nex2, nex3;
__m128i tmp0, tmp1, tmp2, tmp3, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_112 = __lsx_vldi(0x438);
__m128i const_74 = __lsx_vldi(0x425);
__m128i const_38 = __lsx_vldi(0x413);
__m128i const_94 = __lsx_vldi(0x42F);
__m128i const_18 = __lsx_vldi(0x409);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_rgba, 0, src_rgba, 16, src_rgba, 32, src_rgba, 48,
src0, src1, src2, src3);
DUP4_ARG2(__lsx_vld, next_rgba, 0, next_rgba, 16, next_rgba, 32, next_rgba,
48, nex0, nex1, nex2, nex3);
tmp0 = __lsx_vpickod_b(src1, src0);
tmp1 = __lsx_vpickev_b(src1, src0);
tmp2 = __lsx_vpickod_b(src3, src2);
tmp3 = __lsx_vpickev_b(src3, src2);
tmpr = __lsx_vpickod_b(tmp2, tmp0);
tmpb = __lsx_vpickev_b(tmp2, tmp0);
tmpg = __lsx_vpickod_b(tmp3, tmp1);
tmp0 = __lsx_vpickod_b(nex1, nex0);
tmp1 = __lsx_vpickev_b(nex1, nex0);
tmp2 = __lsx_vpickod_b(nex3, nex2);
tmp3 = __lsx_vpickev_b(nex3, nex2);
nexr = __lsx_vpickod_b(tmp2, tmp0);
nexb = __lsx_vpickev_b(tmp2, tmp0);
nexg = __lsx_vpickod_b(tmp3, tmp1);
RGBTOUV(tmpb, tmpg, tmpr, nexb, nexg, nexr, dst0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_rgba += 64;
next_rgba += 64;
}
}
void ARGBToUVJRow_LSX(const uint8_t* src_argb,
int src_stride_argb,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
const uint8_t* next_argb = src_argb + src_stride_argb;
int len = width / 16;
__m128i src0, src1, src2, src3;
__m128i nex0, nex1, nex2, nex3;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i reg0, reg1, dst0;
__m128i tmpb, tmpg, tmpr, nexb, nexg, nexr;
__m128i const_63 = __lsx_vldi(0x43F);
__m128i const_42 = __lsx_vldi(0x42A);
__m128i const_21 = __lsx_vldi(0x415);
__m128i const_53 = __lsx_vldi(0x435);
__m128i const_10 = __lsx_vldi(0x40A);
__m128i const_8080 = {0x8080808080808080, 0x8080808080808080};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src_argb, 32, src_argb, 48,
src0, src1, src2, src3);
DUP4_ARG2(__lsx_vld, next_argb, 0, next_argb, 16, next_argb, 32, next_argb,
48, nex0, nex1, nex2, nex3);
tmp0 = __lsx_vpickev_b(src1, src0);
tmp1 = __lsx_vpickod_b(src1, src0);
tmp2 = __lsx_vpickev_b(src3, src2);
tmp3 = __lsx_vpickod_b(src3, src2);
tmpr = __lsx_vpickod_b(tmp2, tmp0);
tmpb = __lsx_vpickev_b(tmp2, tmp0);
tmpg = __lsx_vpickev_b(tmp3, tmp1);
tmp0 = __lsx_vpickev_b(nex1, nex0);
tmp1 = __lsx_vpickod_b(nex1, nex0);
tmp2 = __lsx_vpickev_b(nex3, nex2);
tmp3 = __lsx_vpickod_b(nex3, nex2);
nexr = __lsx_vpickod_b(tmp2, tmp0);
nexb = __lsx_vpickev_b(tmp2, tmp0);
nexg = __lsx_vpickev_b(tmp3, tmp1);
tmp0 = __lsx_vaddwev_h_bu(tmpb, nexb);
tmp1 = __lsx_vaddwod_h_bu(tmpb, nexb);
tmp2 = __lsx_vaddwev_h_bu(tmpg, nexg);
tmp3 = __lsx_vaddwod_h_bu(tmpg, nexg);
reg0 = __lsx_vaddwev_h_bu(tmpr, nexr);
reg1 = __lsx_vaddwod_h_bu(tmpr, nexr);
tmpb = __lsx_vavgr_hu(tmp0, tmp1);
tmpg = __lsx_vavgr_hu(tmp2, tmp3);
tmpr = __lsx_vavgr_hu(reg0, reg1);
reg0 = __lsx_vmadd_h(const_8080, const_63, tmpb);
reg1 = __lsx_vmadd_h(const_8080, const_63, tmpr);
reg0 = __lsx_vmsub_h(reg0, const_42, tmpg);
reg1 = __lsx_vmsub_h(reg1, const_53, tmpg);
reg0 = __lsx_vmsub_h(reg0, const_21, tmpr);
reg1 = __lsx_vmsub_h(reg1, const_10, tmpb);
dst0 = __lsx_vpickod_b(reg1, reg0);
__lsx_vstelm_d(dst0, dst_u, 0, 0);
__lsx_vstelm_d(dst0, dst_v, 0, 1);
dst_u += 8;
dst_v += 8;
src_argb += 64;
next_argb += 64;
}
}
void I444ToARGBRow_LSX(const uint8_t* src_y,
const uint8_t* src_u,
const uint8_t* src_v,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 16;
__m128i vec_y, vec_u, vec_v, out_b, out_g, out_r;
__m128i vec_yl, vec_yh, vec_ul, vec_vl, vec_uh, vec_vh;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb, vec_ugvg;
__m128i const_80 = __lsx_vldi(0x480);
__m128i alpha = __lsx_vldi(0xFF);
__m128i zero = __lsx_vldi(0);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_ugvg = __lsx_vilvl_h(vec_ug, vec_vg);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
vec_u = __lsx_vld(src_u, 0);
vec_v = __lsx_vld(src_v, 0);
vec_yl = __lsx_vilvl_b(vec_y, vec_y);
vec_ul = __lsx_vilvl_b(zero, vec_u);
vec_vl = __lsx_vilvl_b(zero, vec_v);
I444TORGB(vec_yl, vec_ul, vec_vl, vec_ub, vec_vr, vec_ugvg, vec_yg, vec_yb,
out_b, out_g, out_r);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
vec_yh = __lsx_vilvh_b(vec_y, vec_y);
vec_uh = __lsx_vilvh_b(zero, vec_u);
vec_vh = __lsx_vilvh_b(zero, vec_v);
I444TORGB(vec_yh, vec_uh, vec_vh, vec_ub, vec_vr, vec_ugvg, vec_yg, vec_yb,
out_b, out_g, out_r);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
src_y += 16;
src_u += 16;
src_v += 16;
}
}
void I400ToARGBRow_LSX(const uint8_t* src_y,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 16;
__m128i vec_y, vec_yl, vec_yh, out0;
__m128i y_ev, y_od, dst0, dst1, dst2, dst3;
__m128i temp0, temp1;
__m128i alpha = __lsx_vldi(0xFF);
__m128i vec_yg = __lsx_vreplgr2vr_h(yuvconstants->kYToRgb[0]);
__m128i vec_yb = __lsx_vreplgr2vr_w(yuvconstants->kYBiasToRgb[0]);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
vec_yl = __lsx_vilvl_b(vec_y, vec_y);
y_ev = __lsx_vmulwev_w_hu_h(vec_yl, vec_yg);
y_od = __lsx_vmulwod_w_hu_h(vec_yl, vec_yg);
y_ev = __lsx_vsrai_w(y_ev, 16);
y_od = __lsx_vsrai_w(y_od, 16);
y_ev = __lsx_vadd_w(y_ev, vec_yb);
y_od = __lsx_vadd_w(y_od, vec_yb);
y_ev = __lsx_vsrai_w(y_ev, 6);
y_od = __lsx_vsrai_w(y_od, 6);
y_ev = __lsx_vclip255_w(y_ev);
y_od = __lsx_vclip255_w(y_od);
out0 = __lsx_vpackev_h(y_od, y_ev);
temp0 = __lsx_vpackev_b(out0, out0);
temp1 = __lsx_vpackev_b(alpha, out0);
dst0 = __lsx_vilvl_h(temp1, temp0);
dst1 = __lsx_vilvh_h(temp1, temp0);
vec_yh = __lsx_vilvh_b(vec_y, vec_y);
y_ev = __lsx_vmulwev_w_hu_h(vec_yh, vec_yg);
y_od = __lsx_vmulwod_w_hu_h(vec_yh, vec_yg);
y_ev = __lsx_vsrai_w(y_ev, 16);
y_od = __lsx_vsrai_w(y_od, 16);
y_ev = __lsx_vadd_w(y_ev, vec_yb);
y_od = __lsx_vadd_w(y_od, vec_yb);
y_ev = __lsx_vsrai_w(y_ev, 6);
y_od = __lsx_vsrai_w(y_od, 6);
y_ev = __lsx_vclip255_w(y_ev);
y_od = __lsx_vclip255_w(y_od);
out0 = __lsx_vpackev_h(y_od, y_ev);
temp0 = __lsx_vpackev_b(out0, out0);
temp1 = __lsx_vpackev_b(alpha, out0);
dst2 = __lsx_vilvl_h(temp1, temp0);
dst3 = __lsx_vilvh_h(temp1, temp0);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_y += 16;
}
}
void J400ToARGBRow_LSX(const uint8_t* src_y, uint8_t* dst_argb, int width) {
int x;
int len = width / 16;
__m128i vec_y, dst0, dst1, dst2, dst3;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i alpha = __lsx_vldi(0xFF);
for (x = 0; x < len; x++) {
vec_y = __lsx_vld(src_y, 0);
tmp0 = __lsx_vilvl_b(vec_y, vec_y);
tmp1 = __lsx_vilvh_b(vec_y, vec_y);
tmp2 = __lsx_vilvl_b(alpha, vec_y);
tmp3 = __lsx_vilvh_b(alpha, vec_y);
dst0 = __lsx_vilvl_h(tmp2, tmp0);
dst1 = __lsx_vilvh_h(tmp2, tmp0);
dst2 = __lsx_vilvl_h(tmp3, tmp1);
dst3 = __lsx_vilvh_h(tmp3, tmp1);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
src_y += 16;
}
}
void YUY2ToARGBRow_LSX(const uint8_t* src_yuy2,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 8;
__m128i src0, vec_y, vec_vu;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb;
__m128i vec_vrub, vec_vgug;
__m128i out_b, out_g, out_r;
__m128i const_80 = __lsx_vldi(0x480);
__m128i zero = __lsx_vldi(0);
__m128i alpha = __lsx_vldi(0xFF);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_vrub = __lsx_vilvl_h(vec_vr, vec_ub);
vec_vgug = __lsx_vilvl_h(vec_vg, vec_ug);
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_yuy2, 0);
vec_y = __lsx_vpickev_b(src0, src0);
vec_vu = __lsx_vpickod_b(src0, src0);
YUVTORGB(vec_y, vec_vu, vec_vrub, vec_vgug, vec_yg, vec_yb, out_b, out_g,
out_r);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
src_yuy2 += 16;
}
}
void UYVYToARGBRow_LSX(const uint8_t* src_uyvy,
uint8_t* dst_argb,
const struct YuvConstants* yuvconstants,
int width) {
int x;
int len = width / 8;
__m128i src0, vec_y, vec_vu;
__m128i vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb;
__m128i vec_vrub, vec_vgug;
__m128i out_b, out_g, out_r;
__m128i const_80 = __lsx_vldi(0x480);
__m128i zero = __lsx_vldi(0);
__m128i alpha = __lsx_vldi(0xFF);
YUVTORGB_SETUP(yuvconstants, vec_vr, vec_ub, vec_vg, vec_ug, vec_yg, vec_yb);
vec_vrub = __lsx_vilvl_h(vec_vr, vec_ub);
vec_vgug = __lsx_vilvl_h(vec_vg, vec_ug);
for (x = 0; x < len; x++) {
src0 = __lsx_vld(src_uyvy, 0);
vec_y = __lsx_vpickod_b(src0, src0);
vec_vu = __lsx_vpickev_b(src0, src0);
YUVTORGB(vec_y, vec_vu, vec_vrub, vec_vgug, vec_yg, vec_yb, out_b, out_g,
out_r);
STOREARGB(alpha, out_r, out_g, out_b, dst_argb);
src_uyvy += 16;
}
}
void InterpolateRow_LSX(uint8_t* dst_ptr,
const uint8_t* src_ptr,
ptrdiff_t src_stride,
int width,
int32_t source_y_fraction) {
int x;
int y1_fraction = source_y_fraction;
int y0_fraction = 256 - y1_fraction;
const uint8_t* nex_ptr = src_ptr + src_stride;
uint16_t y_fractions;
int len = width / 32;
__m128i src0, src1, nex0, nex1;
__m128i dst0, dst1, y_frac;
__m128i tmp0, tmp1, tmp2, tmp3;
__m128i const_128 = __lsx_vldi(0x480);
if (y1_fraction == 0) {
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_ptr, 0, src_ptr, 16, src0, src1);
__lsx_vst(src0, dst_ptr, 0);
__lsx_vst(src1, dst_ptr, 16);
src_ptr += 32;
dst_ptr += 32;
}
return;
}
if (y1_fraction == 128) {
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_ptr, 0, src_ptr, 16, src0, src1);
DUP2_ARG2(__lsx_vld, nex_ptr, 0, nex_ptr, 16, nex0, nex1);
dst0 = __lsx_vavgr_bu(src0, nex0);
dst1 = __lsx_vavgr_bu(src1, nex1);
__lsx_vst(dst0, dst_ptr, 0);
__lsx_vst(dst1, dst_ptr, 16);
src_ptr += 32;
nex_ptr += 32;
dst_ptr += 32;
}
return;
}
y_fractions = (uint16_t)(y0_fraction + (y1_fraction << 8));
y_frac = __lsx_vreplgr2vr_h(y_fractions);
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_ptr, 0, src_ptr, 16, src0, src1);
DUP2_ARG2(__lsx_vld, nex_ptr, 0, nex_ptr, 16, nex0, nex1);
tmp0 = __lsx_vilvl_b(nex0, src0);
tmp1 = __lsx_vilvh_b(nex0, src0);
tmp2 = __lsx_vilvl_b(nex1, src1);
tmp3 = __lsx_vilvh_b(nex1, src1);
tmp0 = __lsx_vdp2add_h_bu(const_128, tmp0, y_frac);
tmp1 = __lsx_vdp2add_h_bu(const_128, tmp1, y_frac);
tmp2 = __lsx_vdp2add_h_bu(const_128, tmp2, y_frac);
tmp3 = __lsx_vdp2add_h_bu(const_128, tmp3, y_frac);
dst0 = __lsx_vsrlni_b_h(tmp1, tmp0, 8);
dst1 = __lsx_vsrlni_b_h(tmp3, tmp2, 8);
__lsx_vst(dst0, dst_ptr, 0);
__lsx_vst(dst1, dst_ptr, 16);
src_ptr += 32;
nex_ptr += 32;
dst_ptr += 32;
}
}
void ARGBSetRow_LSX(uint8_t* dst_argb, uint32_t v32, int width) {
int x;
int len = width / 4;
__m128i dst0 = __lsx_vreplgr2vr_w(v32);
for (x = 0; x < len; x++) {
__lsx_vst(dst0, dst_argb, 0);
dst_argb += 16;
}
}
void RAWToRGB24Row_LSX(const uint8_t* src_raw, uint8_t* dst_rgb24, int width) {
int x;
int len = width / 16;
__m128i src0, src1, src2;
__m128i dst0, dst1, dst2;
__m128i shuf0 = {0x0708030405000102, 0x110C0D0E090A0B06};
__m128i shuf1 = {0x1516171213140F10, 0x1F1E1B1C1D18191A};
__m128i shuf2 = {0x090405060102031E, 0x0D0E0F0A0B0C0708};
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_raw, 0, src_raw, 16, src0, src1);
src2 = __lsx_vld(src_raw, 32);
DUP2_ARG3(__lsx_vshuf_b, src1, src0, shuf0, src1, src0, shuf1, dst0, dst1);
dst2 = __lsx_vshuf_b(src1, src2, shuf2);
dst1 = __lsx_vinsgr2vr_b(dst1, src_raw[32], 0x0E);
__lsx_vst(dst0, dst_rgb24, 0);
__lsx_vst(dst1, dst_rgb24, 16);
__lsx_vst(dst2, dst_rgb24, 32);
dst_rgb24 += 48;
src_raw += 48;
}
}
void MergeUVRow_LSX(const uint8_t* src_u,
const uint8_t* src_v,
uint8_t* dst_uv,
int width) {
int x;
int len = width / 16;
__m128i src0, src1, dst0, dst1;
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_u, 0, src_v, 0, src0, src1);
dst0 = __lsx_vilvl_b(src1, src0);
dst1 = __lsx_vilvh_b(src1, src0);
__lsx_vst(dst0, dst_uv, 0);
__lsx_vst(dst1, dst_uv, 16);
src_u += 16;
src_v += 16;
dst_uv += 32;
}
}
void ARGBExtractAlphaRow_LSX(const uint8_t* src_argb,
uint8_t* dst_a,
int width) {
int x;
int len = width / 16;
__m128i src0, src1, src2, src3, tmp0, tmp1, dst0;
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src_argb, 32, src_argb, 48,
src0, src1, src2, src3);
tmp0 = __lsx_vpickod_b(src1, src0);
tmp1 = __lsx_vpickod_b(src3, src2);
dst0 = __lsx_vpickod_b(tmp1, tmp0);
__lsx_vst(dst0, dst_a, 0);
src_argb += 64;
dst_a += 16;
}
}
void ARGBBlendRow_LSX(const uint8_t* src_argb,
const uint8_t* src_argb1,
uint8_t* dst_argb,
int width) {
int x;
int len = width / 8;
__m128i src0, src1, src2, src3;
__m128i tmp0, tmp1, dst0, dst1;
__m128i reg0, reg1, reg2, reg3;
__m128i a0, a1, a2, a3;
__m128i const_256 = __lsx_vldi(0x500);
__m128i zero = __lsx_vldi(0);
__m128i alpha = __lsx_vldi(0xFF);
__m128i control = {0xFF000000FF000000, 0xFF000000FF000000};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src_argb1, 0, src_argb1, 16,
src0, src1, src2, src3);
tmp0 = __lsx_vshuf4i_b(src0, 0xFF);
tmp1 = __lsx_vshuf4i_b(src1, 0xFF);
a0 = __lsx_vilvl_b(zero, tmp0);
a1 = __lsx_vilvh_b(zero, tmp0);
a2 = __lsx_vilvl_b(zero, tmp1);
a3 = __lsx_vilvh_b(zero, tmp1);
reg0 = __lsx_vilvl_b(zero, src2);
reg1 = __lsx_vilvh_b(zero, src2);
reg2 = __lsx_vilvl_b(zero, src3);
reg3 = __lsx_vilvh_b(zero, src3);
DUP4_ARG2(__lsx_vsub_h, const_256, a0, const_256, a1, const_256, a2,
const_256, a3, a0, a1, a2, a3);
DUP4_ARG2(__lsx_vmul_h, a0, reg0, a1, reg1, a2, reg2, a3, reg3, reg0, reg1,
reg2, reg3);
DUP2_ARG3(__lsx_vsrani_b_h, reg1, reg0, 8, reg3, reg2, 8, dst0, dst1);
dst0 = __lsx_vsadd_bu(dst0, src0);
dst1 = __lsx_vsadd_bu(dst1, src1);
dst0 = __lsx_vbitsel_v(dst0, alpha, control);
dst1 = __lsx_vbitsel_v(dst1, alpha, control);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
src_argb += 32;
src_argb1 += 32;
dst_argb += 32;
}
}
void ARGBQuantizeRow_LSX(uint8_t* dst_argb,
int scale,
int interval_size,
int interval_offset,
int width) {
int x;
int len = width / 16;
__m128i src0, src1, src2, src3, dst0, dst1, dst2, dst3;
__m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7;
__m128i reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
__m128i vec_size = __lsx_vreplgr2vr_b(interval_size);
__m128i vec_offset = __lsx_vreplgr2vr_b(interval_offset);
__m128i vec_scale = __lsx_vreplgr2vr_w(scale);
__m128i zero = __lsx_vldi(0);
__m128i control = {0xFF000000FF000000, 0xFF000000FF000000};
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, dst_argb, 0, dst_argb, 16, dst_argb, 32, dst_argb, 48,
src0, src1, src2, src3);
reg0 = __lsx_vilvl_b(zero, src0);
reg1 = __lsx_vilvh_b(zero, src0);
reg2 = __lsx_vilvl_b(zero, src1);
reg3 = __lsx_vilvh_b(zero, src1);
reg4 = __lsx_vilvl_b(zero, src2);
reg5 = __lsx_vilvh_b(zero, src2);
reg6 = __lsx_vilvl_b(zero, src3);
reg7 = __lsx_vilvh_b(zero, src3);
tmp0 = __lsx_vilvl_h(zero, reg0);
tmp1 = __lsx_vilvh_h(zero, reg0);
tmp2 = __lsx_vilvl_h(zero, reg1);
tmp3 = __lsx_vilvh_h(zero, reg1);
tmp4 = __lsx_vilvl_h(zero, reg2);
tmp5 = __lsx_vilvh_h(zero, reg2);
tmp6 = __lsx_vilvl_h(zero, reg3);
tmp7 = __lsx_vilvh_h(zero, reg3);
DUP4_ARG2(__lsx_vmul_w, tmp0, vec_scale, tmp1, vec_scale, tmp2, vec_scale,
tmp3, vec_scale, tmp0, tmp1, tmp2, tmp3);
DUP4_ARG2(__lsx_vmul_w, tmp4, vec_scale, tmp5, vec_scale, tmp6, vec_scale,
tmp7, vec_scale, tmp4, tmp5, tmp6, tmp7);
DUP4_ARG3(__lsx_vsrani_h_w, tmp1, tmp0, 16, tmp3, tmp2, 16, tmp5, tmp4, 16,
tmp7, tmp6, 16, reg0, reg1, reg2, reg3);
dst0 = __lsx_vpickev_b(reg1, reg0);
dst1 = __lsx_vpickev_b(reg3, reg2);
tmp0 = __lsx_vilvl_h(zero, reg4);
tmp1 = __lsx_vilvh_h(zero, reg4);
tmp2 = __lsx_vilvl_h(zero, reg5);
tmp3 = __lsx_vilvh_h(zero, reg5);
tmp4 = __lsx_vilvl_h(zero, reg6);
tmp5 = __lsx_vilvh_h(zero, reg6);
tmp6 = __lsx_vilvl_h(zero, reg7);
tmp7 = __lsx_vilvh_h(zero, reg7);
DUP4_ARG2(__lsx_vmul_w, tmp0, vec_scale, tmp1, vec_scale, tmp2, vec_scale,
tmp3, vec_scale, tmp0, tmp1, tmp2, tmp3);
DUP4_ARG2(__lsx_vmul_w, tmp4, vec_scale, tmp5, vec_scale, tmp6, vec_scale,
tmp7, vec_scale, tmp4, tmp5, tmp6, tmp7);
DUP4_ARG3(__lsx_vsrani_h_w, tmp1, tmp0, 16, tmp3, tmp2, 16, tmp5, tmp4, 16,
tmp7, tmp6, 16, reg0, reg1, reg2, reg3);
dst2 = __lsx_vpickev_b(reg1, reg0);
dst3 = __lsx_vpickev_b(reg3, reg2);
DUP4_ARG2(__lsx_vmul_b, dst0, vec_size, dst1, vec_size, dst2, vec_size,
dst3, vec_size, dst0, dst1, dst2, dst3);
DUP4_ARG2(__lsx_vadd_b, dst0, vec_offset, dst1, vec_offset, dst2,
vec_offset, dst3, vec_offset, dst0, dst1, dst2, dst3);
DUP4_ARG3(__lsx_vbitsel_v, dst0, src0, control, dst1, src1, control, dst2,
src2, control, dst3, src3, control, dst0, dst1, dst2, dst3);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
__lsx_vst(dst2, dst_argb, 32);
__lsx_vst(dst3, dst_argb, 48);
dst_argb += 64;
}
}
void ARGBColorMatrixRow_LSX(const uint8_t* src_argb,
uint8_t* dst_argb,
const int8_t* matrix_argb,
int width) {
int x;
int len = width / 8;
__m128i src0, src1, tmp0, tmp1, dst0, dst1;
__m128i tmp_b, tmp_g, tmp_r, tmp_a;
__m128i reg_b, reg_g, reg_r, reg_a;
__m128i matrix_b = __lsx_vldrepl_w(matrix_argb, 0);
__m128i matrix_g = __lsx_vldrepl_w(matrix_argb, 4);
__m128i matrix_r = __lsx_vldrepl_w(matrix_argb, 8);
__m128i matrix_a = __lsx_vldrepl_w(matrix_argb, 12);
for (x = 0; x < len; x++) {
DUP2_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src0, src1);
DUP4_ARG2(__lsx_vdp2_h_bu_b, src0, matrix_b, src0, matrix_g, src0, matrix_r,
src0, matrix_a, tmp_b, tmp_g, tmp_r, tmp_a);
DUP4_ARG2(__lsx_vdp2_h_bu_b, src1, matrix_b, src1, matrix_g, src1, matrix_r,
src1, matrix_a, reg_b, reg_g, reg_r, reg_a);
DUP4_ARG2(__lsx_vhaddw_w_h, tmp_b, tmp_b, tmp_g, tmp_g, tmp_r, tmp_r, tmp_a,
tmp_a, tmp_b, tmp_g, tmp_r, tmp_a);
DUP4_ARG2(__lsx_vhaddw_w_h, reg_b, reg_b, reg_g, reg_g, reg_r, reg_r, reg_a,
reg_a, reg_b, reg_g, reg_r, reg_a);
DUP4_ARG2(__lsx_vsrai_w, tmp_b, 6, tmp_g, 6, tmp_r, 6, tmp_a, 6, tmp_b,
tmp_g, tmp_r, tmp_a);
DUP4_ARG2(__lsx_vsrai_w, reg_b, 6, reg_g, 6, reg_r, 6, reg_a, 6, reg_b,
reg_g, reg_r, reg_a);
DUP4_ARG1(__lsx_vclip255_w, tmp_b, tmp_g, tmp_r, tmp_a, tmp_b, tmp_g, tmp_r,
tmp_a)
DUP4_ARG1(__lsx_vclip255_w, reg_b, reg_g, reg_r, reg_a, reg_b, reg_g, reg_r,
reg_a)
DUP4_ARG2(__lsx_vpickev_h, reg_b, tmp_b, reg_g, tmp_g, reg_r, tmp_r, reg_a,
tmp_a, tmp_b, tmp_g, tmp_r, tmp_a);
tmp0 = __lsx_vpackev_b(tmp_g, tmp_b);
tmp1 = __lsx_vpackev_b(tmp_a, tmp_r);
dst0 = __lsx_vilvl_h(tmp1, tmp0);
dst1 = __lsx_vilvh_h(tmp1, tmp0);
__lsx_vst(dst0, dst_argb, 0);
__lsx_vst(dst1, dst_argb, 16);
src_argb += 32;
dst_argb += 32;
}
}
void SplitUVRow_LSX(const uint8_t* src_uv,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
int len = width / 32;
__m128i src0, src1, src2, src3;
__m128i dst0, dst1, dst2, dst3;
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src_uv, 0, src_uv, 16, src_uv, 32, src_uv, 48, src0,
src1, src2, src3);
DUP2_ARG2(__lsx_vpickev_b, src1, src0, src3, src2, dst0, dst1);
DUP2_ARG2(__lsx_vpickod_b, src1, src0, src3, src2, dst2, dst3);
__lsx_vst(dst0, dst_u, 0);
__lsx_vst(dst1, dst_u, 16);
__lsx_vst(dst2, dst_v, 0);
__lsx_vst(dst3, dst_v, 16);
src_uv += 64;
dst_u += 32;
dst_v += 32;
}
}
void SetRow_LSX(uint8_t* dst, uint8_t v8, int width) {
int x;
int len = width / 16;
__m128i dst0 = __lsx_vreplgr2vr_b(v8);
for (x = 0; x < len; x++) {
__lsx_vst(dst0, dst, 0);
dst += 16;
}
}
void MirrorSplitUVRow_LSX(const uint8_t* src_uv,
uint8_t* dst_u,
uint8_t* dst_v,
int width) {
int x;
int len = width / 32;
__m128i src0, src1, src2, src3;
__m128i dst0, dst1, dst2, dst3;
__m128i shuff0 = {0x10121416181A1C1E, 0x00020406080A0C0E};
__m128i shuff1 = {0x11131517191B1D1F, 0x01030507090B0D0F};
src_uv += (width << 1);
for (x = 0; x < len; x++) {
src_uv -= 64;
DUP4_ARG2(__lsx_vld, src_uv, 0, src_uv, 16, src_uv, 32, src_uv, 48, src2,
src3, src0, src1);
DUP4_ARG3(__lsx_vshuf_b, src1, src0, shuff1, src3, src2, shuff1, src1, src0,
shuff0, src3, src2, shuff0, dst0, dst1, dst2, dst3);
__lsx_vst(dst0, dst_v, 0);
__lsx_vst(dst1, dst_v, 16);
__lsx_vst(dst2, dst_u, 0);
__lsx_vst(dst3, dst_u, 16);
dst_u += 32;
dst_v += 32;
}
}
void HalfFloatRow_LSX(const uint16_t* src,
uint16_t* dst,
float scale,
int width) {
int x;
int len = width / 32;
float mult = 1.9259299444e-34f * scale;
__m128i src0, src1, src2, src3, dst0, dst1, dst2, dst3;
__m128i tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7;
__m128 reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7;
__m128 vec_mult = (__m128)__lsx_vldrepl_w(&mult, 0);
__m128i zero = __lsx_vldi(0);
for (x = 0; x < len; x++) {
DUP4_ARG2(__lsx_vld, src, 0, src, 16, src, 32, src, 48, src0, src1, src2,
src3);
DUP4_ARG2(__lsx_vilvl_h, zero, src0, zero, src1, zero, src2, zero, src3,
tmp0, tmp2, tmp4, tmp6);
DUP4_ARG2(__lsx_vilvh_h, zero, src0, zero, src1, zero, src2, zero, src3,
tmp1, tmp3, tmp5, tmp7);
DUP4_ARG1(__lsx_vffint_s_wu, tmp0, tmp2, tmp4, tmp6, reg0, reg2, reg4,
reg6);
DUP4_ARG1(__lsx_vffint_s_wu, tmp1, tmp3, tmp5, tmp7, reg1, reg3, reg5,
reg7);
DUP4_ARG2(__lsx_vfmul_s, reg0, vec_mult, reg1, vec_mult, reg2, vec_mult,
reg3, vec_mult, reg0, reg1, reg2, reg3);
DUP4_ARG2(__lsx_vfmul_s, reg4, vec_mult, reg5, vec_mult, reg6, vec_mult,
reg7, vec_mult, reg4, reg5, reg6, reg7);
DUP4_ARG2(__lsx_vsrli_w, (v4u32)reg0, 13, (v4u32)reg1, 13, (v4u32)reg2, 13,
(v4u32)reg3, 13, tmp0, tmp1, tmp2, tmp3);
DUP4_ARG2(__lsx_vsrli_w, (v4u32)reg4, 13, (v4u32)reg5, 13, (v4u32)reg6, 13,
(v4u32)reg7, 13, tmp4, tmp5, tmp6, tmp7);
DUP4_ARG2(__lsx_vpickev_h, tmp1, tmp0, tmp3, tmp2, tmp5, tmp4, tmp7, tmp6,
dst0, dst1, dst2, dst3);
__lsx_vst(dst0, dst, 0);
__lsx_vst(dst1, dst, 16);
__lsx_vst(dst2, dst, 32);
__lsx_vst(dst3, dst, 48);
src += 32;
dst += 32;
}
}
struct RgbConstants {
uint8_t kRGBToY[4];
uint16_t kAddY;
uint16_t pad;
};
// RGB to JPeg coefficients
// B * 0.1140 coefficient = 29
// G * 0.5870 coefficient = 150
// R * 0.2990 coefficient = 77
// Add 0.5 = 0x80
static const struct RgbConstants kRgb24JPEGConstants = {{29, 150, 77, 0},
128,
0};
static const struct RgbConstants kRawJPEGConstants = {{77, 150, 29, 0}, 128, 0};
// RGB to BT.601 coefficients
// B * 0.1016 coefficient = 25
// G * 0.5078 coefficient = 129
// R * 0.2578 coefficient = 66
// Add 16.5 = 0x1080
static const struct RgbConstants kRgb24I601Constants = {{25, 129, 66, 0},
0x1080,
0};
static const struct RgbConstants kRawI601Constants = {{66, 129, 25, 0},
0x1080,
0};
// ARGB expects first 3 values to contain RGB and 4th value is ignored.
static void ARGBToYMatrixRow_LSX(const uint8_t* src_argb,
uint8_t* dst_y,
int width,
const struct RgbConstants* rgbconstants) {
asm volatile(
"vldrepl.b $vr0, %3, 0 \n\t" // load rgbconstants
"vldrepl.b $vr1, %3, 1 \n\t" // load rgbconstants
"vldrepl.b $vr2, %3, 2 \n\t" // load rgbconstants
"vldrepl.h $vr3, %3, 4 \n\t" // load rgbconstants
"1: \n\t"
"vld $vr4, %0, 0 \n\t"
"vld $vr5, %0, 16 \n\t"
"vld $vr6, %0, 32 \n\t"
"vld $vr7, %0, 48 \n\t" // load 16 pixels of
// ARGB
"vor.v $vr12, $vr3, $vr3 \n\t"
"vor.v $vr13, $vr3, $vr3 \n\t"
"addi.d %2, %2, -16 \n\t" // 16 processed per
// loop.
"vpickev.b $vr8, $vr5, $vr4 \n\t" // BR
"vpickev.b $vr10, $vr7, $vr6 \n\t"
"vpickod.b $vr9, $vr5, $vr4 \n\t" // GA
"vpickod.b $vr11, $vr7, $vr6 \n\t"
"vmaddwev.h.bu $vr12, $vr8, $vr0 \n\t" // B
"vmaddwev.h.bu $vr13, $vr10, $vr0 \n\t"
"vmaddwev.h.bu $vr12, $vr9, $vr1 \n\t" // G
"vmaddwev.h.bu $vr13, $vr11, $vr1 \n\t"
"vmaddwod.h.bu $vr12, $vr8, $vr2 \n\t" // R
"vmaddwod.h.bu $vr13, $vr10, $vr2 \n\t"
"addi.d %0, %0, 64 \n\t"
"vpickod.b $vr10, $vr13, $vr12 \n\t"
"vst $vr10, %1, 0 \n\t"
"addi.d %1, %1, 16 \n\t"
"bnez %2, 1b \n\t"
: "+&r"(src_argb), // %0
"+&r"(dst_y), // %1
"+&r"(width) // %2
: "r"(rgbconstants)
: "memory");
}
void ARGBToYRow_LSX(const uint8_t* src_argb, uint8_t* dst_y, int width) {
ARGBToYMatrixRow_LSX(src_argb, dst_y, width, &kRgb24I601Constants);
}
void ARGBToYJRow_LSX(const uint8_t* src_argb, uint8_t* dst_yj, int width) {
ARGBToYMatrixRow_LSX(src_argb, dst_yj, width, &kRgb24JPEGConstants);
}
void ABGRToYRow_LSX(const uint8_t* src_abgr, uint8_t* dst_y, int width) {
ARGBToYMatrixRow_LSX(src_abgr, dst_y, width, &kRawI601Constants);
}
void ABGRToYJRow_LSX(const uint8_t* src_abgr, uint8_t* dst_yj, int width) {
ARGBToYMatrixRow_LSX(src_abgr, dst_yj, width, &kRawJPEGConstants);
}
// RGBA expects first value to be A and ignored, then 3 values to contain RGB.
// Same code as ARGB, except the LD4
static void RGBAToYMatrixRow_LSX(const uint8_t* src_rgba,
uint8_t* dst_y,
int width,
const struct RgbConstants* rgbconstants) {
asm volatile(
"vldrepl.b $vr0, %3, 0 \n\t" // load rgbconstants
"vldrepl.b $vr1, %3, 1 \n\t" // load rgbconstants
"vldrepl.b $vr2, %3, 2 \n\t" // load rgbconstants
"vldrepl.h $vr3, %3, 4 \n\t" // load rgbconstants
"1: \n\t"
"vld $vr4, %0, 0 \n\t"
"vld $vr5, %0, 16 \n\t"
"vld $vr6, %0, 32 \n\t"
"vld $vr7, %0, 48 \n\t" // load 16 pixels of
// RGBA
"vor.v $vr12, $vr3, $vr3 \n\t"
"vor.v $vr13, $vr3, $vr3 \n\t"
"addi.d %2, %2, -16 \n\t" // 16 processed per
// loop.
"vpickev.b $vr8, $vr5, $vr4 \n\t" // AG
"vpickev.b $vr10, $vr7, $vr6 \n\t"
"vpickod.b $vr9, $vr5, $vr4 \n\t" // BR
"vpickod.b $vr11, $vr7, $vr6 \n\t"
"vmaddwev.h.bu $vr12, $vr9, $vr0 \n\t" // B
"vmaddwev.h.bu $vr13, $vr11, $vr0 \n\t"
"vmaddwod.h.bu $vr12, $vr8, $vr1 \n\t" // G
"vmaddwod.h.bu $vr13, $vr10, $vr1 \n\t"
"vmaddwod.h.bu $vr12, $vr9, $vr2 \n\t" // R
"vmaddwod.h.bu $vr13, $vr11, $vr2 \n\t"
"addi.d %0, %0, 64 \n\t"
"vpickod.b $vr10, $vr13, $vr12 \n\t"
"vst $vr10, %1, 0 \n\t"
"addi.d %1, %1, 16 \n\t"
"bnez %2, 1b \n\t"
: "+&r"(src_rgba), // %0
"+&r"(dst_y), // %1
"+&r"(width) // %2
: "r"(rgbconstants)
: "memory");
}
void RGBAToYRow_LSX(const uint8_t* src_rgba, uint8_t* dst_y, int width) {
RGBAToYMatrixRow_LSX(src_rgba, dst_y, width, &kRgb24I601Constants);
}
void RGBAToYJRow_LSX(const uint8_t* src_rgba, uint8_t* dst_yj, int width) {
RGBAToYMatrixRow_LSX(src_rgba, dst_yj, width, &kRgb24JPEGConstants);
}
void BGRAToYRow_LSX(const uint8_t* src_bgra, uint8_t* dst_y, int width) {
RGBAToYMatrixRow_LSX(src_bgra, dst_y, width, &kRawI601Constants);
}
static void RGBToYMatrixRow_LSX(const uint8_t* src_rgba,
uint8_t* dst_y,
int width,
const struct RgbConstants* rgbconstants) {
int8_t shuff[64] = {0, 2, 3, 5, 6, 8, 9, 11, 12, 14, 15, 17, 18,
20, 21, 23, 24, 26, 27, 29, 30, 0, 1, 3, 4, 6,
7, 9, 10, 12, 13, 15, 1, 0, 4, 0, 7, 0, 10,
0, 13, 0, 16, 0, 19, 0, 22, 0, 25, 0, 28, 0,
31, 0, 2, 0, 5, 0, 8, 0, 11, 0, 14, 0};
asm volatile(
"vldrepl.b $vr0, %3, 0 \n\t" // load rgbconstants
"vldrepl.b $vr1, %3, 1 \n\t" // load rgbconstants
"vldrepl.b $vr2, %3, 2 \n\t" // load rgbconstants
"vldrepl.h $vr3, %3, 4 \n\t" // load rgbconstants
"vld $vr4, %4, 0 \n\t" // load shuff
"vld $vr5, %4, 16 \n\t"
"vld $vr6, %4, 32 \n\t"
"vld $vr7, %4, 48 \n\t"
"1: \n\t"
"vld $vr8, %0, 0 \n\t"
"vld $vr9, %0, 16 \n\t"
"vld $vr10, %0, 32 \n\t" // load 16 pixels of
// RGB
"vor.v $vr12, $vr3, $vr3 \n\t"
"vor.v $vr13, $vr3, $vr3 \n\t"
"addi.d %2, %2, -16 \n\t" // 16 processed per
// loop.
"vshuf.b $vr14, $vr9, $vr8, $vr4 \n\t"
"vshuf.b $vr15, $vr9, $vr10, $vr5 \n\t"
"vshuf.b $vr16, $vr9, $vr8, $vr6 \n\t"
"vshuf.b $vr17, $vr9, $vr10, $vr7 \n\t"
"vmaddwev.h.bu $vr12, $vr16, $vr1 \n\t" // G
"vmaddwev.h.bu $vr13, $vr17, $vr1 \n\t"
"vmaddwev.h.bu $vr12, $vr14, $vr0 \n\t" // B
"vmaddwev.h.bu $vr13, $vr15, $vr0 \n\t"
"vmaddwod.h.bu $vr12, $vr14, $vr2 \n\t" // R
"vmaddwod.h.bu $vr13, $vr15, $vr2 \n\t"
"addi.d %0, %0, 48 \n\t"
"vpickod.b $vr10, $vr13, $vr12 \n\t"
"vst $vr10, %1, 0 \n\t"
"addi.d %1, %1, 16 \n\t"
"bnez %2, 1b \n\t"
: "+&r"(src_rgba), // %0
"+&r"(dst_y), // %1
"+&r"(width) // %2
: "r"(rgbconstants), // %3
"r"(shuff) // %4
: "memory");
}
void RGB24ToYJRow_LSX(const uint8_t* src_rgb24, uint8_t* dst_yj, int width) {
RGBToYMatrixRow_LSX(src_rgb24, dst_yj, width, &kRgb24JPEGConstants);
}
void RAWToYJRow_LSX(const uint8_t* src_raw, uint8_t* dst_yj, int width) {
RGBToYMatrixRow_LSX(src_raw, dst_yj, width, &kRawJPEGConstants);
}
void RGB24ToYRow_LSX(const uint8_t* src_rgb24, uint8_t* dst_y, int width) {
RGBToYMatrixRow_LSX(src_rgb24, dst_y, width, &kRgb24I601Constants);
}
void RAWToYRow_LSX(const uint8_t* src_raw, uint8_t* dst_y, int width) {
RGBToYMatrixRow_LSX(src_raw, dst_y, width, &kRawI601Constants);
}
#ifdef __cplusplus
} // extern "C"
} // namespace libyuv
#endif
#endif // !defined(LIBYUV_DISABLE_LSX) && defined(__loongarch_sx)