diff options
Diffstat (limited to 'common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s')
-rw-r--r-- | common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s | 483 |
1 files changed, 483 insertions, 0 deletions
diff --git a/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s b/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s new file mode 100644 index 0000000..da316ae --- /dev/null +++ b/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s @@ -0,0 +1,483 @@ +///***************************************************************************** +//* +//* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore +//* +//* Licensed under the Apache License, Version 2.0 (the "License"); +//* you may not use this file except in compliance with the License. +//* You may obtain a copy of the License at: +//* +//* http://www.apache.org/licenses/LICENSE-2.0 +//* +//* Unless required by applicable law or agreed to in writing, software +//* distributed under the License is distributed on an "AS IS" BASIS, +//* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +//* See the License for the specific language governing permissions and +//* limitations under the License. +//* +//*****************************************************************************/ +///** +//****************************************************************************** +//* //file +//* ihevc_inter_pred_filters_luma_vert.s +//* +//* //brief +//* contains function definitions for inter prediction interpolation. +//* functions are coded using neon intrinsics and can be compiled using + +//* rvct +//* +//* //author +//* parthiban v +//* +//* //par list of functions: +//* +//* - ihevc_inter_pred_luma_vert() +//* +//* //remarks +//* none +//* +//******************************************************************************* +//*/ + +///* all the functions here are replicated from ihevc_inter_pred_filters.c and modified to */ +///* include reconstruction */ + + +//void ihevc_inter_pred_luma_vert_w16out(uword8 *pu1_src, +// word16 *pi2_dst, +// word32 src_strd, +// word32 dst_strd, +// word8 *pi1_coeff, +// word32 ht, +// word32 wd ) + +//**************variables vs registers***************************************** +// x0 => *pu1_src +// x1 => *pu1_dst +// x2 => src_strd +// x6 => dst_strd +// x12 => *pi1_coeff +// x5 => ht +// x3 => wd + + +.include "ihevc_neon_macros.s" + +.globl ihevc_inter_pred_luma_vert_w16out_av8 + +.type ihevc_inter_pred_luma_vert_w16out_av8, %function + +ihevc_inter_pred_luma_vert_w16out_av8: + + // stmfd sp!, {x4-x12, x14} //stack stores the values of the arguments + push_v_regs + stp x19, x20,[sp,#-16]! + + mov x15,x4 // pi1_coeff + mov x16,x5 // ht + mov x17,x6 // wd + + mov x12,x15 //load pi1_coeff + mov x6,x3 + mov x5,x17 //load wd + ld1 {v0.8b},[x12] //coeff = vld1_s8(pi1_coeff) + sub x12,x2,x2,lsl #2 //src_ctrd & pi1_coeff + abs v0.8b, v0.8b //vabs_s8(coeff) + add x0,x0,x12 //x0->pu1_src x12->pi1_coeff + mov x3,x16 //load ht + subs x7,x3,#0 //x3->ht + //ble end_loops_16out //end loop jump + dup v22.8b, v0.8b[0] //coeffabs_0 = vdup_lane_u8(coeffabs, 0)// + cmp x5,#8 + dup v23.8b, v0.8b[1] //coeffabs_1 = vdup_lane_u8(coeffabs, 1)// + dup v24.8b, v0.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2)// + dup v25.8b, v0.8b[3] //coeffabs_3 = vdup_lane_u8(coeffabs, 3)// + dup v26.8b, v0.8b[4] //coeffabs_4 = vdup_lane_u8(coeffabs, 4)// + dup v27.8b, v0.8b[5] //coeffabs_5 = vdup_lane_u8(coeffabs, 5)// + dup v28.8b, v0.8b[6] //coeffabs_6 = vdup_lane_u8(coeffabs, 6)// + dup v29.8b, v0.8b[7] //coeffabs_7 = vdup_lane_u8(coeffabs, 7)// + blt core_loop_wd_4_16out //core loop wd 4 jump + stp x0,x1, [sp, #-16]! + + bic x4,x5,#7 //x5 ->wd + sub x20,x4,x6,lsl #2 //x6->dst_strd x5 ->wd + neg x9, x20 + sub x20,x4,x2,lsl #2 //x2->src_strd + neg x8, x20 + lsl x6, x6, #1 + lsr x3, x5, #3 //divide by 8 + mul x7, x7, x3 //multiply height by width + sub x7, x7,#4 //subtract by one for epilog + +prolog_16out: + + and x10, x0, #31 + add x3,x0,x2 //pu1_src_tmp += src_strd// + + ld1 {v1.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)// + subs x4,x4,#8 + ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// + ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)// + ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)// + ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)// + ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)// + ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)// + ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)// + ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)// + + + add x20,x0,x8 + csel x0, x20, x0,le + umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// + + bic x20,x5,#7 //x5 ->wd + csel x4, x20, x4,le + umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// + + ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// + + add x20,x20,x3 + prfm PLDL1KEEP,[x20] + umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// + add x20,x3, x2 + prfm PLDL1KEEP,[x20] + umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// + add x20,x3, x2, lsl #1 + prfm PLDL1KEEP,[x20] + umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// + add x3, x3, x2 + umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// + add x20,x3, x2, lsl #1 + prfm PLDL1KEEP,[x20] + umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// + + add x3,x0,x2 //pu1_src_tmp += src_strd// + umull v12.8h, v3.8b, v23.8b + ld1 {v1.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlsl v12.8h, v2.8b, v22.8b + ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlsl v12.8h, v4.8b, v24.8b + ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlal v12.8h, v5.8b, v25.8b + umlal v12.8h, v6.8b, v26.8b + umlsl v12.8h, v7.8b, v27.8b + umlal v12.8h, v16.8b, v28.8b + umlsl v12.8h, v17.8b, v29.8b + add x14,x1,x6 + st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)// + //vqrshrun.s16 d10,q5,#6 //sto_res = vqmovun_s16(sto_res_tmp)// + add x20,x1,x9,lsl #1 + csel x1, x20, x1,le + + umull v14.8h, v4.8b, v23.8b + subs x7,x7,#4 + umlsl v14.8h, v3.8b, v22.8b + umlsl v14.8h, v5.8b, v24.8b + umlal v14.8h, v6.8b, v25.8b + ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + umlal v14.8h, v7.8b, v26.8b + ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlsl v14.8h, v16.8b, v27.8b + ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlal v14.8h, v17.8b, v28.8b + ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlsl v14.8h, v18.8b, v29.8b + ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + + st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// + //vqrshrun.s16 d12,q6,#6 + + + blt epilog_end_16out + beq epilog_16out //jumps to epilog + +kernel_8_16out: + + subs x4,x4,#8 + umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// + + add x20,x0,x8 + csel x0, x20, x0,le + umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)// + + ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)// + + ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)// + + bic x20,x5,#7 //x5 ->wd + csel x4, x20, x4,le + umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)// + + ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)// + + st1 {v12.16b},[x14],x6 + umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)// + + add x3,x0,x2 //pu1_src_tmp += src_strd// + umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)// + + +// and x11, x0, #31 + umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// + + st1 {v14.16b},[x14],x6 + umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// + + add x14,x1,x6 + umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// + + ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// + + ld1 {v1.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// + + st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)// + umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// + + add x20,x1,x9,lsl #1 + csel x1, x20, x1,le + umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// + +// cmp x11, x10 + umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// + + add x10, x3, x2, lsl #3 // 10*strd - 8+2 + umull v12.8h, v3.8b, v23.8b + + add x10, x10, x2 // 11*strd + umlsl v12.8h, v2.8b, v22.8b + + add x20,x20,x10 + prfm PLDL1KEEP,[x20] //11+ 0 + umlsl v12.8h, v4.8b, v24.8b + + add x20,x10, x2 + prfm PLDL1KEEP,[x20] //11+ 1*strd + umlal v12.8h, v5.8b, v25.8b + + add x20,x10, x2, lsl #1 + prfm PLDL1KEEP,[x20] //11+ 2*strd + umlal v12.8h, v6.8b, v26.8b + + add x10, x10, x2 //12*strd + umlsl v12.8h, v7.8b, v27.8b + + add x20,x10, x2, lsl #1 + prfm PLDL1KEEP,[x20] //11+ 3*strd + umlal v12.8h, v16.8b, v28.8b + +// mov x10, x11 + umlsl v12.8h, v17.8b, v29.8b + + ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umull v14.8h, v4.8b, v23.8b + + subs x7,x7,#4 + umlsl v14.8h, v3.8b, v22.8b + + st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// + umlsl v14.8h, v5.8b, v24.8b + + ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + umlal v14.8h, v6.8b, v25.8b + + ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umlal v14.8h, v7.8b, v26.8b + + ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umlsl v14.8h, v16.8b, v27.8b + + ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umlal v14.8h, v17.8b, v28.8b + + ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)// + umlsl v14.8h, v18.8b, v29.8b + + + bgt kernel_8_16out //jumps to kernel_8 + +epilog_16out: + + umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// + umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)// + umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)// + umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)// + umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)// + umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)// + umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)// + umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)// + st1 {v12.16b},[x14],x6 + + //vqrshrun.s16 d14,q7,#6 + + ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)// + umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// + umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// + umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// + umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// + umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// + umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// + umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// + umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// + st1 {v14.16b},[x14],x6 + + //vqrshrun.s16 d8,q4,#6 //sto_res = vqmovun_s16(sto_res_tmp)// + + ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)// + umull v12.8h, v3.8b, v23.8b + umlsl v12.8h, v2.8b, v22.8b + umlsl v12.8h, v4.8b, v24.8b + umlal v12.8h, v5.8b, v25.8b + umlal v12.8h, v6.8b, v26.8b + umlsl v12.8h, v7.8b, v27.8b + umlal v12.8h, v16.8b, v28.8b + umlsl v12.8h, v17.8b, v29.8b + add x14,x1,x6 + st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)// + //vqrshrun.s16 d10,q5,#6 //sto_res = vqmovun_s16(sto_res_tmp)// + + ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)// + umull v14.8h, v4.8b, v23.8b + umlsl v14.8h, v3.8b, v22.8b + umlsl v14.8h, v5.8b, v24.8b + umlal v14.8h, v6.8b, v25.8b + umlal v14.8h, v7.8b, v26.8b + umlsl v14.8h, v16.8b, v27.8b + umlal v14.8h, v17.8b, v28.8b + umlsl v14.8h, v18.8b, v29.8b + + st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// + //vqrshrun.s16 d12,q6,#6 + +epilog_end_16out: + st1 {v12.16b},[x14],x6 + //vqrshrun.s16 d14,q7,#6 + + st1 {v14.16b},[x14],x6 + + +end_loops_16out: + tst x5,#7 + ldp x0,x1, [sp], #16 + + // ldmeqfd sp!,{x4-x12,x15} //reload the registers from sp + bne lbl355 + ldp x19, x20,[sp], #16 + pop_v_regs + ret +lbl355: + mov x5, #4 + add x0, x0, #8 + add x1, x1, #16 + mov x7, #16 + lsr x6, x6, #1 + + // + +core_loop_wd_4_16out: + sub x20,x5,x6,lsl #2 //x6->dst_strd x5 ->wd + neg x9, x20 + sub x20,x5,x2,lsl #2 //x2->src_strd + neg x8, x20 + movi v4.8b, #0 + lsl x6, x6, #1 + +outer_loop_wd_4_16out: + subs x12,x5,#0 + ble end_inner_loop_wd_4_16out //outer loop jump + +inner_loop_wd_4_16out: + add x3,x0,x2 + ld1 {v4.s}[1],[x3],x2 //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 1)// + subs x12,x12,#4 + dup v5.2s, v4.2s[1] //src_tmp2 = vdup_lane_u32(src_tmp1, 1)// + ld1 {v5.s}[1],[x3],x2 //src_tmp2 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp2, 1)// + ld1 {v4.s}[0],[x0] //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 0)// + umull v0.8h, v5.8b, v23.8b //mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1)// + + dup v6.2s, v5.2s[1] //src_tmp3 = vdup_lane_u32(src_tmp2, 1)// + add x0,x0,#4 + ld1 {v6.s}[1],[x3],x2 //src_tmp3 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp3, 1)// + umlsl v0.8h, v4.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_0)// + + dup v7.2s, v6.2s[1] //src_tmp4 = vdup_lane_u32(src_tmp3, 1)// + ld1 {v7.s}[1],[x3],x2 //src_tmp4 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp4, 1)// + umlsl v0.8h, v6.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_2)// + + umull v8.8h, v7.8b, v23.8b + dup v4.2s, v7.2s[1] //src_tmp1 = vdup_lane_u32(src_tmp4, 1)// + umull v2.8h, v7.8b, v25.8b //mul_res2 = vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3)// + ld1 {v4.s}[1],[x3],x2 //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 1)// + umlsl v8.8h, v6.8b, v22.8b + umlal v0.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_4)// + + dup v5.2s, v4.2s[1] //src_tmp2 = vdup_lane_u32(src_tmp1, 1)// + umlsl v8.8h, v4.8b, v24.8b + ld1 {v5.s}[1],[x3],x2 //src_tmp2 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp2, 1)// + umlsl v2.8h, v5.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp2), coeffabs_5)// + + dup v6.2s, v5.2s[1] //src_tmp3 = vdup_lane_u32(src_tmp2, 1)// + umlal v8.8h, v5.8b, v25.8b + ld1 {v6.s}[1],[x3],x2 //src_tmp3 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp3, 1)// + umlal v0.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_6)// + + dup v7.2s, v6.2s[1] //src_tmp4 = vdup_lane_u32(src_tmp3, 1)// + umlal v8.8h, v6.8b, v26.8b + ld1 {v7.s}[1],[x3],x2 //src_tmp4 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp4, 1)// + umlsl v2.8h, v7.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp4), coeffabs_7)// + + dup v4.2s, v7.2s[1] + add v0.8h, v0.8h , v2.8h //mul_res1 = vaddq_u16(mul_res1, mul_res2)// + + umlsl v8.8h, v7.8b, v27.8b + ld1 {v4.s}[1],[x3],x2 + umlal v8.8h, v4.8b, v28.8b + dup v5.2s, v4.2s[1] + //vqrshrun.s16 d0,q0,#6 //sto_res = vqmovun_s16(sto_res_tmp)// + + ld1 {v5.s}[1],[x3] + add x3,x1,x6 + st1 {v0.d}[0],[x1],#8 //vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0)// + + umlsl v8.8h, v5.8b, v29.8b + st1 {v0.d}[1],[x3],x6 //vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 1)// + //vqrshrun.s16 d8,q4,#6 + + st1 {v8.d}[0],[x3],x6 + //add x1,x1,#4 + st1 {v8.d}[1],[x3] + bgt inner_loop_wd_4_16out + +end_inner_loop_wd_4_16out: + subs x7,x7,#4 + add x1,x1,x9,lsl #1 + add x0,x0,x8 + bgt outer_loop_wd_4_16out + + // ldmfd sp!, {x4-x12, x15} //reload the registers from sp + ldp x19, x20,[sp], #16 + pop_v_regs + ret + + + + + + + + + |