diff options
Diffstat (limited to 'common/arm/ihevc_itrans_recon_4x4_ttype1.s')
-rw-r--r-- | common/arm/ihevc_itrans_recon_4x4_ttype1.s | 236 |
1 files changed, 236 insertions, 0 deletions
diff --git a/common/arm/ihevc_itrans_recon_4x4_ttype1.s b/common/arm/ihevc_itrans_recon_4x4_ttype1.s new file mode 100644 index 0000000..ab65dae --- /dev/null +++ b/common/arm/ihevc_itrans_recon_4x4_ttype1.s @@ -0,0 +1,236 @@ +@/***************************************************************************** +@* +@* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore +@* +@* Licensed under the Apache License, Version 2.0 (the "License"); +@* you may not use this file except in compliance with the License. +@* You may obtain a copy of the License at: +@* +@* http://www.apache.org/licenses/LICENSE-2.0 +@* +@* Unless required by applicable law or agreed to in writing, software +@* distributed under the License is distributed on an "AS IS" BASIS, +@* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +@* See the License for the specific language governing permissions and +@* limitations under the License. +@* +@*****************************************************************************/ +@/** +@ ******************************************************************************* +@ * @file +@ * ihevc_itrans_recon_4x4_ttype1.s +@ * +@ * @brief +@ * contains function definitions for inverse transform and reconstruction +@ * +@ * +@ * @author +@ * naveen sr +@ * +@ * @par list of functions: +@ * - ihevc_itrans_recon_4x4_ttype1() +@ * +@ * @remarks +@ * none +@ * +@ ******************************************************************************* +@ */ + +@/* all the functions here are replicated from ihevc_itrans.c and modified to */ +@/* include reconstruction */ +@ +@/** +@ ******************************************************************************* +@ * +@ * @brief +@ * this function performs inverse transform type 1 (dst) and reconstruction +@ * for 4x4 input block +@ * +@ * @par description: +@ * performs inverse transform and adds the prediction data and clips output +@ * to 8 bit +@ * +@ * @param[in] pi2_src +@ * input 4x4 coefficients +@ * +@ * @param[in] pi2_tmp +@ * temporary 4x4 buffer for storing inverse +@ * +@ * transform +@ * 1st stage output +@ * +@ * @param[in] pu1_pred +@ * prediction 4x4 block +@ * +@ * @param[out] pu1_dst +@ * output 4x4 block +@ * +@ * @param[in] src_strd +@ * input stride +@ * +@ * @param[in] pred_strd +@ * prediction stride +@ * +@ * @param[in] dst_strd +@ * output stride +@ * +@ * @param[in] zero_cols +@ * zero columns in pi2_src +@ * +@ * @returns void +@ * +@ * @remarks +@ * none +@ * +@ ******************************************************************************* +@ */ +@void ihevc_itrans_recon_4x4_ttype1(word16 *pi2_src, +@ word16 *pi2_tmp, +@ uword8 *pu1_pred, +@ uword8 *pu1_dst, +@ word32 src_strd, +@ word32 pred_strd, +@ word32 dst_strd, +@ word32 zero_cols) + +@**************variables vs registers************************* +@ r0 => *pi2_src +@ r1 => *pi2_tmp +@ r2 => *pu1_pred +@ r3 => *pu1_dst +@ r4 => src_strd +@ r5 => pred_strd +@ r6 => dst_strd +@ r7 => zero_cols + +.text +.align 4 + + + + +.set shift_stage1_idct , 7 +.set shift_stage2_idct , 12 + +.globl ihevc_itrans_recon_4x4_ttype1_a9q + +.type ihevc_itrans_recon_4x4_ttype1_a9q, %function + +ihevc_itrans_recon_4x4_ttype1_a9q: + + stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments + ldr r4,[sp,#40] @loading src_strd + ldr r5,[sp,#44] @loading pred_strd + ldr r6,[sp,#48] @loading dst_strd + ldr r7,[sp,#52] @loading zero_cols + + add r4,r4,r4 @ src_strd in terms of word16 + + mov r8,#29 + mov r9,#55 + mov r10,#74 + mov r11,#84 + vmov.i16 d4[0],r8 + vld1.16 d0,[r0],r4 @loading pi2_src 1st row + vmov.i16 d4[1],r9 + vld1.16 d1,[r0],r4 @loading pi2_src 2nd row + vmov.i16 d4[2],r10 + vld1.16 d2,[r0],r4 @loading pi2_src 3rd row + vmov.i16 d4[3],r11 + vld1.16 d3,[r0],r4 @loading pi2_src 4th row + + @ first stage computation starts + vmull.s16 q3,d1,d4[2] @74 * pi2_src[1] + vmlal.s16 q3,d0,d4[0] @74 * pi2_src[1] + 29 * pi2_src[0] + vmlal.s16 q3,d3,d4[1] @74 * pi2_src[1] + 29 * pi2_src[0] + 55 * pi2_src[3] + vmlal.s16 q3,d2,d4[3] @pi2_out[0] = 29* pi2_src[0] + 74 * pi2_src[1] + 84* pi2_src[2] + 55 * pi2_src[3] + + vmull.s16 q4,d1,d4[2] @74 * pi2_src[1] + vmlal.s16 q4,d0,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] + vmlsl.s16 q4,d2,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] + vmlsl.s16 q4,d3,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3]) + + vmull.s16 q5,d0,d4[2] @ 74 * pi2_src[0] + vmlsl.s16 q5,d2,d4[2] @ 74 * pi2_src[0] - 74 * pi2_src[2] + vmlal.s16 q5,d3,d4[2] @pi2_out[2] = 74 * pi2_src[0] - 74 * pi2_src[2] + 74 * pi2_src[3] + + vmull.s16 q6,d2,d4[1] @ 55 * pi2_src[2] + vmlsl.s16 q6,d1,d4[2] @ 55 * pi2_src[2] - 74 * pi2_src[1] + vmlsl.s16 q6,d3,d4[0] @ - 74 * pi2_src[1] + 55 * pi2_src[2] - 29 * pi2_src[3] + vmlal.s16 q6,d0,d4[3] @pi2_out[3] = 84 * pi2_src[0] - 74 * pi2_src[1] + 55 * pi2_src[2] - 29 * pi2_src[3] + + vqrshrn.s32 d14,q3,#shift_stage1_idct @ (pi2_out[0] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d15,q4,#shift_stage1_idct @ (pi2_out[1] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d16,q5,#shift_stage1_idct @ (pi2_out[2] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d17,q6,#shift_stage1_idct @ (pi2_out[3] + rounding ) >> shift_stage1_idct + vld1.32 d18[0], [r2],r5 + + vtrn.16 d14,d15 + vtrn.16 d16,d17 + vtrn.32 d14,d16 + vtrn.32 d15,d17 + @ output in d14,d15,d16,d17 + @ first stage computation ends + + @ second stage computation starts : copy pasting 1st stage + @ register changes + @ d14 - d0 + @ d15 - d1 + @ d16 - d2 + @ d17 - d3 + vld1.32 d18[1], [r2],r5 + vmull.s16 q3,d15,d4[2] @74 * pi2_src[1] + vmlal.s16 q3,d14,d4[0] @74 * pi2_src[1] + 29 * pi2_src[0] + vmlal.s16 q3,d17,d4[1] @74 * pi2_src[1] + 29 * pi2_src[0] + 55 * pi2_src[3] + vmlal.s16 q3,d16,d4[3] @pi2_out[0] = 29* pi2_src[0] + 74 * pi2_src[1] + 84* pi2_src[2] + 55 * pi2_src[3] + + vmull.s16 q4,d15,d4[2] @74 * pi2_src[1] + vmlal.s16 q4,d14,d4[1] @74 * pi2_src[1] + 55 * pi2_src[0] + vmlsl.s16 q4,d16,d4[0] @74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] + vmlsl.s16 q4,d17,d4[3] @pi2_out[1] = 74 * pi2_src[1] + 55 * pi2_src[0] - 29 * pi2_src[2] - 84 * pi2_src[3]) + + vmull.s16 q5,d14,d4[2] @ 74 * pi2_src[0] + vmlsl.s16 q5,d16,d4[2] @ 74 * pi2_src[0] - 74 * pi2_src[2] + vmlal.s16 q5,d17,d4[2] @pi2_out[2] = 74 * pi2_src[0] - 74 * pi2_src[2] + 74 * pi2_src[3] + vld1.32 d19[0], [r2],r5 + + vmull.s16 q6,d16,d4[1] @ 55 * pi2_src[2] + vmlsl.s16 q6,d15,d4[2] @ - 74 * pi2_src[1] + 55 * pi2_src[2] + vmlsl.s16 q6,d17,d4[0] @ - 74 * pi2_src[1] + 55 * pi2_src[2] - 29 * pi2_src[3] + vmlal.s16 q6,d14,d4[3] @pi2_out[3] = 84 * pi2_src[0] - 74 * pi2_src[1] + 55 * pi2_src[2] - 29 * pi2_src[3] + + vqrshrn.s32 d0,q3,#shift_stage2_idct @ (pi2_out[0] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d1,q4,#shift_stage2_idct @ (pi2_out[1] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d2,q5,#shift_stage2_idct @ (pi2_out[2] + rounding ) >> shift_stage1_idct + vqrshrn.s32 d3,q6,#shift_stage2_idct @ (pi2_out[3] + rounding ) >> shift_stage1_idct + vld1.32 d19[1], [r2],r5 + vtrn.16 d0,d1 + vtrn.16 d2,d3 + vtrn.32 d0,d2 + vtrn.32 d1,d3 + @ output in d0,d1,d2,d3 + @ second stage computation ends + + @ loading pred + + vaddw.u8 q0,q0,d18 @ pi2_out(16bit) + pu1_pred(8bit) + vqmovun.s16 d0,q0 @ clip_u8(pi2_out(16bit) + pu1_pred(8bit)) + vaddw.u8 q1,q1,d19 @ pi2_out(16bit) + pu1_pred(8bit) + vqmovun.s16 d1,q1 @ clip_u8(pi2_out(16bit) + pu1_pred(8bit)) + + @ storing destination + vst1.32 {d0[0]},[r3],r6 + vst1.32 {d0[1]},[r3],r6 + vst1.32 {d1[0]},[r3],r6 + vst1.32 {d1[1]},[r3],r6 + + ldmfd sp!,{r4-r12,r15} @reload the registers from sp + + + + + + + + + |