summaryrefslogtreecommitdiffstats
path: root/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s
diff options
context:
space:
mode:
Diffstat (limited to 'common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s')
-rw-r--r--common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s483
1 files changed, 483 insertions, 0 deletions
diff --git a/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s b/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s
new file mode 100644
index 0000000..da316ae
--- /dev/null
+++ b/common/arm64/ihevc_inter_pred_filters_luma_vert_w16out.s
@@ -0,0 +1,483 @@
+///*****************************************************************************
+//*
+//* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore
+//*
+//* Licensed under the Apache License, Version 2.0 (the "License");
+//* you may not use this file except in compliance with the License.
+//* You may obtain a copy of the License at:
+//*
+//* http://www.apache.org/licenses/LICENSE-2.0
+//*
+//* Unless required by applicable law or agreed to in writing, software
+//* distributed under the License is distributed on an "AS IS" BASIS,
+//* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+//* See the License for the specific language governing permissions and
+//* limitations under the License.
+//*
+//*****************************************************************************/
+///**
+//******************************************************************************
+//* //file
+//* ihevc_inter_pred_filters_luma_vert.s
+//*
+//* //brief
+//* contains function definitions for inter prediction interpolation.
+//* functions are coded using neon intrinsics and can be compiled using
+
+//* rvct
+//*
+//* //author
+//* parthiban v
+//*
+//* //par list of functions:
+//*
+//* - ihevc_inter_pred_luma_vert()
+//*
+//* //remarks
+//* none
+//*
+//*******************************************************************************
+//*/
+
+///* all the functions here are replicated from ihevc_inter_pred_filters.c and modified to */
+///* include reconstruction */
+
+
+//void ihevc_inter_pred_luma_vert_w16out(uword8 *pu1_src,
+// word16 *pi2_dst,
+// word32 src_strd,
+// word32 dst_strd,
+// word8 *pi1_coeff,
+// word32 ht,
+// word32 wd )
+
+//**************variables vs registers*****************************************
+// x0 => *pu1_src
+// x1 => *pu1_dst
+// x2 => src_strd
+// x6 => dst_strd
+// x12 => *pi1_coeff
+// x5 => ht
+// x3 => wd
+
+
+.include "ihevc_neon_macros.s"
+
+.globl ihevc_inter_pred_luma_vert_w16out_av8
+
+.type ihevc_inter_pred_luma_vert_w16out_av8, %function
+
+ihevc_inter_pred_luma_vert_w16out_av8:
+
+ // stmfd sp!, {x4-x12, x14} //stack stores the values of the arguments
+ push_v_regs
+ stp x19, x20,[sp,#-16]!
+
+ mov x15,x4 // pi1_coeff
+ mov x16,x5 // ht
+ mov x17,x6 // wd
+
+ mov x12,x15 //load pi1_coeff
+ mov x6,x3
+ mov x5,x17 //load wd
+ ld1 {v0.8b},[x12] //coeff = vld1_s8(pi1_coeff)
+ sub x12,x2,x2,lsl #2 //src_ctrd & pi1_coeff
+ abs v0.8b, v0.8b //vabs_s8(coeff)
+ add x0,x0,x12 //x0->pu1_src x12->pi1_coeff
+ mov x3,x16 //load ht
+ subs x7,x3,#0 //x3->ht
+ //ble end_loops_16out //end loop jump
+ dup v22.8b, v0.8b[0] //coeffabs_0 = vdup_lane_u8(coeffabs, 0)//
+ cmp x5,#8
+ dup v23.8b, v0.8b[1] //coeffabs_1 = vdup_lane_u8(coeffabs, 1)//
+ dup v24.8b, v0.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2)//
+ dup v25.8b, v0.8b[3] //coeffabs_3 = vdup_lane_u8(coeffabs, 3)//
+ dup v26.8b, v0.8b[4] //coeffabs_4 = vdup_lane_u8(coeffabs, 4)//
+ dup v27.8b, v0.8b[5] //coeffabs_5 = vdup_lane_u8(coeffabs, 5)//
+ dup v28.8b, v0.8b[6] //coeffabs_6 = vdup_lane_u8(coeffabs, 6)//
+ dup v29.8b, v0.8b[7] //coeffabs_7 = vdup_lane_u8(coeffabs, 7)//
+ blt core_loop_wd_4_16out //core loop wd 4 jump
+ stp x0,x1, [sp, #-16]!
+
+ bic x4,x5,#7 //x5 ->wd
+ sub x20,x4,x6,lsl #2 //x6->dst_strd x5 ->wd
+ neg x9, x20
+ sub x20,x4,x2,lsl #2 //x2->src_strd
+ neg x8, x20
+ lsl x6, x6, #1
+ lsr x3, x5, #3 //divide by 8
+ mul x7, x7, x3 //multiply height by width
+ sub x7, x7,#4 //subtract by one for epilog
+
+prolog_16out:
+
+ and x10, x0, #31
+ add x3,x0,x2 //pu1_src_tmp += src_strd//
+
+ ld1 {v1.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ subs x4,x4,#8
+ ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)//
+ ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)//
+ ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)//
+ ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)//
+ ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)//
+ ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)//
+ ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)//
+ ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)//
+
+
+ add x20,x0,x8
+ csel x0, x20, x0,le
+ umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)//
+
+ bic x20,x5,#7 //x5 ->wd
+ csel x4, x20, x4,le
+ umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)//
+
+ ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)//
+
+ add x20,x20,x3
+ prfm PLDL1KEEP,[x20]
+ umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)//
+ add x20,x3, x2
+ prfm PLDL1KEEP,[x20]
+ umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)//
+ add x20,x3, x2, lsl #1
+ prfm PLDL1KEEP,[x20]
+ umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)//
+ add x3, x3, x2
+ umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)//
+ add x20,x3, x2, lsl #1
+ prfm PLDL1KEEP,[x20]
+ umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)//
+
+ add x3,x0,x2 //pu1_src_tmp += src_strd//
+ umull v12.8h, v3.8b, v23.8b
+ ld1 {v1.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlsl v12.8h, v2.8b, v22.8b
+ ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlsl v12.8h, v4.8b, v24.8b
+ ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlal v12.8h, v5.8b, v25.8b
+ umlal v12.8h, v6.8b, v26.8b
+ umlsl v12.8h, v7.8b, v27.8b
+ umlal v12.8h, v16.8b, v28.8b
+ umlsl v12.8h, v17.8b, v29.8b
+ add x14,x1,x6
+ st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)//
+ //vqrshrun.s16 d10,q5,#6 //sto_res = vqmovun_s16(sto_res_tmp)//
+ add x20,x1,x9,lsl #1
+ csel x1, x20, x1,le
+
+ umull v14.8h, v4.8b, v23.8b
+ subs x7,x7,#4
+ umlsl v14.8h, v3.8b, v22.8b
+ umlsl v14.8h, v5.8b, v24.8b
+ umlal v14.8h, v6.8b, v25.8b
+ ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+ umlal v14.8h, v7.8b, v26.8b
+ ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlsl v14.8h, v16.8b, v27.8b
+ ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlal v14.8h, v17.8b, v28.8b
+ ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlsl v14.8h, v18.8b, v29.8b
+ ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+
+ st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)//
+ //vqrshrun.s16 d12,q6,#6
+
+
+ blt epilog_end_16out
+ beq epilog_16out //jumps to epilog
+
+kernel_8_16out:
+
+ subs x4,x4,#8
+ umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)//
+
+ add x20,x0,x8
+ csel x0, x20, x0,le
+ umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)//
+
+ ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)//
+
+ ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)//
+
+ bic x20,x5,#7 //x5 ->wd
+ csel x4, x20, x4,le
+ umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)//
+
+ ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)//
+
+ st1 {v12.16b},[x14],x6
+ umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)//
+
+ add x3,x0,x2 //pu1_src_tmp += src_strd//
+ umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)//
+
+
+// and x11, x0, #31
+ umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)//
+
+ st1 {v14.16b},[x14],x6
+ umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)//
+
+ add x14,x1,x6
+ umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)//
+
+ ld1 {v0.8b},[x0],#8 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)//
+
+ ld1 {v1.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)//
+
+ st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)//
+ umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)//
+
+ add x20,x1,x9,lsl #1
+ csel x1, x20, x1,le
+ umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)//
+
+// cmp x11, x10
+ umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)//
+
+ add x10, x3, x2, lsl #3 // 10*strd - 8+2
+ umull v12.8h, v3.8b, v23.8b
+
+ add x10, x10, x2 // 11*strd
+ umlsl v12.8h, v2.8b, v22.8b
+
+ add x20,x20,x10
+ prfm PLDL1KEEP,[x20] //11+ 0
+ umlsl v12.8h, v4.8b, v24.8b
+
+ add x20,x10, x2
+ prfm PLDL1KEEP,[x20] //11+ 1*strd
+ umlal v12.8h, v5.8b, v25.8b
+
+ add x20,x10, x2, lsl #1
+ prfm PLDL1KEEP,[x20] //11+ 2*strd
+ umlal v12.8h, v6.8b, v26.8b
+
+ add x10, x10, x2 //12*strd
+ umlsl v12.8h, v7.8b, v27.8b
+
+ add x20,x10, x2, lsl #1
+ prfm PLDL1KEEP,[x20] //11+ 3*strd
+ umlal v12.8h, v16.8b, v28.8b
+
+// mov x10, x11
+ umlsl v12.8h, v17.8b, v29.8b
+
+ ld1 {v2.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umull v14.8h, v4.8b, v23.8b
+
+ subs x7,x7,#4
+ umlsl v14.8h, v3.8b, v22.8b
+
+ st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)//
+ umlsl v14.8h, v5.8b, v24.8b
+
+ ld1 {v3.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+ umlal v14.8h, v6.8b, v25.8b
+
+ ld1 {v4.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umlal v14.8h, v7.8b, v26.8b
+
+ ld1 {v5.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umlsl v14.8h, v16.8b, v27.8b
+
+ ld1 {v6.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umlal v14.8h, v17.8b, v28.8b
+
+ ld1 {v7.8b},[x3],x2 //src_tmp4 = vld1_u8(pu1_src_tmp)//
+ umlsl v14.8h, v18.8b, v29.8b
+
+
+ bgt kernel_8_16out //jumps to kernel_8
+
+epilog_16out:
+
+ umull v8.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)//
+ umlsl v8.8h, v0.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp1, coeffabs_0)//
+ umlsl v8.8h, v2.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp3, coeffabs_2)//
+ umlal v8.8h, v3.8b, v25.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp4, coeffabs_3)//
+ umlal v8.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp1, coeffabs_4)//
+ umlsl v8.8h, v5.8b, v27.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp2, coeffabs_5)//
+ umlal v8.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, src_tmp3, coeffabs_6)//
+ umlsl v8.8h, v7.8b, v29.8b //mul_res1 = vmlsl_u8(mul_res1, src_tmp4, coeffabs_7)//
+ st1 {v12.16b},[x14],x6
+
+ //vqrshrun.s16 d14,q7,#6
+
+ ld1 {v16.8b},[x3],x2 //src_tmp1 = vld1_u8(pu1_src_tmp)//
+ umull v10.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)//
+ umlsl v10.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)//
+ umlsl v10.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)//
+ umlal v10.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)//
+ umlal v10.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)//
+ umlsl v10.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)//
+ umlal v10.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)//
+ umlsl v10.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)//
+ st1 {v14.16b},[x14],x6
+
+ //vqrshrun.s16 d8,q4,#6 //sto_res = vqmovun_s16(sto_res_tmp)//
+
+ ld1 {v17.8b},[x3],x2 //src_tmp2 = vld1_u8(pu1_src_tmp)//
+ umull v12.8h, v3.8b, v23.8b
+ umlsl v12.8h, v2.8b, v22.8b
+ umlsl v12.8h, v4.8b, v24.8b
+ umlal v12.8h, v5.8b, v25.8b
+ umlal v12.8h, v6.8b, v26.8b
+ umlsl v12.8h, v7.8b, v27.8b
+ umlal v12.8h, v16.8b, v28.8b
+ umlsl v12.8h, v17.8b, v29.8b
+ add x14,x1,x6
+ st1 {v8.16b},[x1],#16 //vst1_u8(pu1_dst,sto_res)//
+ //vqrshrun.s16 d10,q5,#6 //sto_res = vqmovun_s16(sto_res_tmp)//
+
+ ld1 {v18.8b},[x3],x2 //src_tmp3 = vld1_u8(pu1_src_tmp)//
+ umull v14.8h, v4.8b, v23.8b
+ umlsl v14.8h, v3.8b, v22.8b
+ umlsl v14.8h, v5.8b, v24.8b
+ umlal v14.8h, v6.8b, v25.8b
+ umlal v14.8h, v7.8b, v26.8b
+ umlsl v14.8h, v16.8b, v27.8b
+ umlal v14.8h, v17.8b, v28.8b
+ umlsl v14.8h, v18.8b, v29.8b
+
+ st1 {v10.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)//
+ //vqrshrun.s16 d12,q6,#6
+
+epilog_end_16out:
+ st1 {v12.16b},[x14],x6
+ //vqrshrun.s16 d14,q7,#6
+
+ st1 {v14.16b},[x14],x6
+
+
+end_loops_16out:
+ tst x5,#7
+ ldp x0,x1, [sp], #16
+
+ // ldmeqfd sp!,{x4-x12,x15} //reload the registers from sp
+ bne lbl355
+ ldp x19, x20,[sp], #16
+ pop_v_regs
+ ret
+lbl355:
+ mov x5, #4
+ add x0, x0, #8
+ add x1, x1, #16
+ mov x7, #16
+ lsr x6, x6, #1
+
+ //
+
+core_loop_wd_4_16out:
+ sub x20,x5,x6,lsl #2 //x6->dst_strd x5 ->wd
+ neg x9, x20
+ sub x20,x5,x2,lsl #2 //x2->src_strd
+ neg x8, x20
+ movi v4.8b, #0
+ lsl x6, x6, #1
+
+outer_loop_wd_4_16out:
+ subs x12,x5,#0
+ ble end_inner_loop_wd_4_16out //outer loop jump
+
+inner_loop_wd_4_16out:
+ add x3,x0,x2
+ ld1 {v4.s}[1],[x3],x2 //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 1)//
+ subs x12,x12,#4
+ dup v5.2s, v4.2s[1] //src_tmp2 = vdup_lane_u32(src_tmp1, 1)//
+ ld1 {v5.s}[1],[x3],x2 //src_tmp2 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp2, 1)//
+ ld1 {v4.s}[0],[x0] //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 0)//
+ umull v0.8h, v5.8b, v23.8b //mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1)//
+
+ dup v6.2s, v5.2s[1] //src_tmp3 = vdup_lane_u32(src_tmp2, 1)//
+ add x0,x0,#4
+ ld1 {v6.s}[1],[x3],x2 //src_tmp3 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp3, 1)//
+ umlsl v0.8h, v4.8b, v22.8b //mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_0)//
+
+ dup v7.2s, v6.2s[1] //src_tmp4 = vdup_lane_u32(src_tmp3, 1)//
+ ld1 {v7.s}[1],[x3],x2 //src_tmp4 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp4, 1)//
+ umlsl v0.8h, v6.8b, v24.8b //mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_2)//
+
+ umull v8.8h, v7.8b, v23.8b
+ dup v4.2s, v7.2s[1] //src_tmp1 = vdup_lane_u32(src_tmp4, 1)//
+ umull v2.8h, v7.8b, v25.8b //mul_res2 = vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3)//
+ ld1 {v4.s}[1],[x3],x2 //src_tmp1 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp1, 1)//
+ umlsl v8.8h, v6.8b, v22.8b
+ umlal v0.8h, v4.8b, v26.8b //mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_4)//
+
+ dup v5.2s, v4.2s[1] //src_tmp2 = vdup_lane_u32(src_tmp1, 1)//
+ umlsl v8.8h, v4.8b, v24.8b
+ ld1 {v5.s}[1],[x3],x2 //src_tmp2 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp2, 1)//
+ umlsl v2.8h, v5.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp2), coeffabs_5)//
+
+ dup v6.2s, v5.2s[1] //src_tmp3 = vdup_lane_u32(src_tmp2, 1)//
+ umlal v8.8h, v5.8b, v25.8b
+ ld1 {v6.s}[1],[x3],x2 //src_tmp3 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp3, 1)//
+ umlal v0.8h, v6.8b, v28.8b //mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_6)//
+
+ dup v7.2s, v6.2s[1] //src_tmp4 = vdup_lane_u32(src_tmp3, 1)//
+ umlal v8.8h, v6.8b, v26.8b
+ ld1 {v7.s}[1],[x3],x2 //src_tmp4 = vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp4, 1)//
+ umlsl v2.8h, v7.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp4), coeffabs_7)//
+
+ dup v4.2s, v7.2s[1]
+ add v0.8h, v0.8h , v2.8h //mul_res1 = vaddq_u16(mul_res1, mul_res2)//
+
+ umlsl v8.8h, v7.8b, v27.8b
+ ld1 {v4.s}[1],[x3],x2
+ umlal v8.8h, v4.8b, v28.8b
+ dup v5.2s, v4.2s[1]
+ //vqrshrun.s16 d0,q0,#6 //sto_res = vqmovun_s16(sto_res_tmp)//
+
+ ld1 {v5.s}[1],[x3]
+ add x3,x1,x6
+ st1 {v0.d}[0],[x1],#8 //vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0)//
+
+ umlsl v8.8h, v5.8b, v29.8b
+ st1 {v0.d}[1],[x3],x6 //vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 1)//
+ //vqrshrun.s16 d8,q4,#6
+
+ st1 {v8.d}[0],[x3],x6
+ //add x1,x1,#4
+ st1 {v8.d}[1],[x3]
+ bgt inner_loop_wd_4_16out
+
+end_inner_loop_wd_4_16out:
+ subs x7,x7,#4
+ add x1,x1,x9,lsl #1
+ add x0,x0,x8
+ bgt outer_loop_wd_4_16out
+
+ // ldmfd sp!, {x4-x12, x15} //reload the registers from sp
+ ldp x19, x20,[sp], #16
+ pop_v_regs
+ ret
+
+
+
+
+
+
+
+
+