diff options
Diffstat (limited to 'common/arm/ihevc_intra_pred_chroma_mode_27_to_33.s')
-rw-r--r-- | common/arm/ihevc_intra_pred_chroma_mode_27_to_33.s | 542 |
1 files changed, 542 insertions, 0 deletions
diff --git a/common/arm/ihevc_intra_pred_chroma_mode_27_to_33.s b/common/arm/ihevc_intra_pred_chroma_mode_27_to_33.s new file mode 100644 index 0000000..f2431e1 --- /dev/null +++ b/common/arm/ihevc_intra_pred_chroma_mode_27_to_33.s @@ -0,0 +1,542 @@ +@/***************************************************************************** +@* +@* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore +@* +@* Licensed under the Apache License, Version 2.0 (the "License"); +@* you may not use this file except in compliance with the License. +@* You may obtain a copy of the License at: +@* +@* http://www.apache.org/licenses/LICENSE-2.0 +@* +@* Unless required by applicable law or agreed to in writing, software +@* distributed under the License is distributed on an "AS IS" BASIS, +@* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +@* See the License for the specific language governing permissions and +@* limitations under the License. +@* +@*****************************************************************************/ +@/** +@******************************************************************************* +@* @file +@* ihevc_intra_pred_chroma_mode_27_to_33.s +@* +@* @brief +@* contains function definition for intra prediction interpolation filters +@* +@* +@* @author +@* parthiban v +@* +@* @par list of functions: +@* - ihevc_intra_pred_chroma_mode_27_to_33() +@* +@* @remarksll +@* none +@* +@******************************************************************************* +@*/ + +@/** +@******************************************************************************* +@* +@* @brief +@* intraprediction for mode 27 to 33 (positive angle, vertical mode ) with +@*.extern neighboring samples location pointed by 'pu1_ref' to the tu +@* block location pointed by 'pu1_dst' +@* +@* @par description: +@* +@* +@* @param[in] pu1_src +@* uword8 pointer to the source +@* +@* @param[in] pu1_dst +@* uword8 pointer to the destination +@* +@* @param[in] src_strd +@* integer source stride +@* +@* @param[in] dst_strd +@* integer destination stride +@* +@* @param[in] nt +@* integer transform block size +@* +@* @param[in] mode +@* integer intraprediction mode +@* +@* @returns +@* +@* @remarks +@* none +@* +@******************************************************************************* +@*/ + +@.if intra_pred_chroma_27_t0_33 == c +@void ihevc_intra_pred_chroma_mode_27_to_33(uword8 *pu1_ref, +@ word32 src_strd, +@ uword8 *pu1_dst, +@ word32 dst_strd, +@ word32 nt, +@ word32 mode) + +.text +.align 4 + + + + +.globl ihevc_intra_pred_chroma_mode_27_to_33_a9q +.extern gai4_ihevc_ang_table +.extern gau1_ihevc_planar_factor + +gai4_ihevc_ang_table_addr: +.long gai4_ihevc_ang_table - ulbl1 - 8 + +gau1_ihevc_planar_factor_addr: +.long gau1_ihevc_planar_factor - ulbl2 - 8 + + +.type ihevc_intra_pred_chroma_mode_27_to_33_a9q, %function + +ihevc_intra_pred_chroma_mode_27_to_33_a9q: + + stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments + + ldr r4,[sp,#40] @loads nt + ldr r5,[sp,#44] @loads mode + ldr r6,gai4_ihevc_ang_table_addr @loads word32 gai4_ihevc_ang_table[35] +ulbl1: + add r6,r6,pc + + lsl r7,r4,#2 @four_nt + + add r8,r6,r5,lsl #2 @*gai4_ihevc_ang_table[mode] + ldr r9,[r8] @intra_pred_ang = gai4_ihevc_ang_table[mode] + ldr r1,gau1_ihevc_planar_factor_addr @used for ((row + 1) * intra_pred_ang) row values +ulbl2: + add r1,r1,pc + add r6,r1,#1 + + tst r4,#7 + add r8,r0,r7 @pu1_ref + four_nt + mov lr,#0 @row + mov r12,r4 + bne core_loop_4 + lsl r4,r4,#1 + b core_loop_8 + +core_loop_8: + add r8,r8,#2 @pu1_ref_main_idx += (four_nt + 1) + vdup.8 d0,r9 @intra_pred_ang + mov r12,r4,lsr #4 @divide by 8 + + vmov.i8 d1,#32 + mul r7,r4,r12 + + vmov.i16 q3,#31 + + mov r1,r8 + mov r5,r4 + mov r11,#2 + +prologue: + vld1.8 {d3},[r6] @loads the row value + vmull.u8 q1,d3,d0 @pos = ((row + 1) * intra_pred_ang) + vand q2,q1,q3 @dup_const_fract(fract = pos & (31)) + vmovn.i16 d4,q2 + vshrn.u16 d5,q1,#5 @idx = pos >> 5 + + vdup.8 d31,d4[0] + add r0,r2,r3 + + vmov.u32 lr,d5[0] @(i row)extract idx to the r register + lsl lr,lr,#1 + + vdup.8 d29,d4[1] @(ii) + and r9,lr,#0xff @(i row) get the last byte + + add r10,r8,r9 @(i row)*pu1_ref[ref_main_idx] + + asr lr,lr,#8 @(ii)shift by 8 + vld1.8 {d8},[r10],r11 @(i row)ref_main_idx + and r9,lr,#0xff @(ii)get the last byte + + asr lr,lr,#8 @(iii) + vld1.8 {d9},[r10] @(i row)ref_main_idx_1 + add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx] + + and r9,lr,#0xff @(iii) + vsub.u8 d30,d1,d31 @32-fract(dup_const_32_fract) + add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx] + + vld1.8 {d12},[r12],r11 @(ii)ref_main_idx + vmull.u8 q5,d8,d30 @(i row)vmull_u8(ref_main_idx, dup_const_32_fract) + + vld1.8 {d13},[r12] @(ii)ref_main_idx_1 + vmlal.u8 q5,d9,d31 @(i row)vmull_u8(ref_main_idx_1, dup_const_fract) + asr lr,lr,#8 @(iv) + + vdup.8 d27,d4[2] @(iii) + vsub.u8 d28,d1,d29 @(ii)32-fract(dup_const_32_fract) + and r9,lr,#0xff @(iv) + + vdup.8 d25,d4[3] @(iv) + vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract) + add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx] + + vld1.8 {d16},[r10],r11 @(iii)ref_main_idx + vmlal.u8 q7,d13,d29 @(ii)vmull_u8(ref_main_idx_1, dup_const_fract) + + vld1.8 {d17},[r10] @(iii)ref_main_idx_1 + vrshrn.i16 d10,q5,#5 @(i row)shift_res = vrshrn_n_u16(add_res, 5) + + vld1.8 {d20},[r12],r11 @(iv)ref_main_idx + vsub.u8 d26,d1,d27 @(iii)32-fract(dup_const_32_fract) + + vld1.8 {d21},[r12] @(iv)ref_main_idx_1 + + vdup.8 d31,d4[4] @(v) + vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) + + vmov.u32 lr,d5[1] @extract idx to the r register + vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) + lsl lr,lr,#1 + + vst1.8 {d10},[r2]! @(i row) + vrshrn.i16 d14,q7,#5 @(ii)shift_res = vrshrn_n_u16(add_res, 5) + + and r9,lr,#0xff @(v) + vdup.8 d29,d4[5] @(vi) + add r10,r8,r9 @(v)*pu1_ref[ref_main_idx] + + vld1.8 {d8},[r10],r11 @(v)ref_main_idx + vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) + + asr lr,lr,#8 @(vi) + vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) + and r9,lr,#0xff @(vi) + + vld1.8 {d9},[r10] @(v)ref_main_idx_1 + vmlal.u8 q11,d21,d25 @(iv)vmull_u8(ref_main_idx_1, dup_const_fract) + + vst1.8 {d14},[r0],r3 @(ii) + vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) + + add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx] + vdup.8 d27,d4[6] @(vii) + asr lr,lr,#8 @(vii) + + and r9,lr,#0xff @(vii) + vsub.u8 d30,d1,d31 @(v)32-fract(dup_const_32_fract) + add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx] + + vld1.8 {d12},[r12],r11 @(vi)ref_main_idx + vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract) + + vld1.8 {d13},[r12] @(vi)ref_main_idx_1 + vmlal.u8 q5,d9,d31 @(v)vmull_u8(ref_main_idx_1, dup_const_fract) + + vst1.8 {d18},[r0],r3 @(iii) + vrshrn.i16 d22,q11,#5 @(iv)shift_res = vrshrn_n_u16(add_res, 5) + + asr lr,lr,#8 @(viii) + vdup.8 d25,d4[7] @(viii) + and r9,lr,#0xff @(viii) + + vld1.8 {d16},[r10],r11 @(vii)ref_main_idx + vsub.u8 d28,d1,d29 @(vi)32-fract(dup_const_32_fract) + + vld1.8 {d17},[r10] @(vii)ref_main_idx_1 + vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract) + + add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx] + vmlal.u8 q7,d13,d29 @(vi)vmull_u8(ref_main_idx_1, dup_const_fract) + subs r7,r7,#8 + + vst1.8 {d22},[r0],r3 @(iv) + vrshrn.i16 d10,q5,#5 @(v)shift_res = vrshrn_n_u16(add_res, 5) + + vld1.8 {d20},[r12],r11 @(viii)ref_main_idx + vsub.u8 d26,d1,d27 @(vii)32-fract(dup_const_32_fract) + + vld1.8 {d21},[r12] @(viii)ref_main_idx_1 + vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) + + addgt r8,r8,#8 + vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) + subgt r4,r4,#8 + + vst1.8 {d10},[r0],r3 @(v) + vrshrn.i16 d14,q7,#5 @(vi)shift_res = vrshrn_n_u16(add_res, 5) + + beq epilogue + + vld1.8 {d5},[r6] @loads the row value + vmull.u8 q1,d5,d0 @pos = ((row + 1) * intra_pred_ang) + vand q2,q1,q3 @dup_const_fract(fract = pos & (31)) + vmovn.i16 d4,q2 + vshrn.u16 d3,q1,#5 @idx = pos >> 5 + vmov.u32 lr,d3[0] @(i)extract idx to the r register + lsl lr,lr,#1 + and r9,lr,#0xff @(i) + add r10,r8,r9 @(i)*pu1_ref[ref_main_idx] + +kernel_8_rows: + asr lr,lr,#8 @(ii) + vdup.8 d31,d4[0] + subs r4,r4,#8 + + vld1.8 {d8},[r10],r11 @(i)ref_main_idx + vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) + and r9,lr,#0xff @(ii) + addle r6,r6,#8 @increment the row value + + vld1.8 {d9},[r10] @(i)ref_main_idx_1 + vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract) + add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx] + + vld1.8 {d5},[r6] @loads the row value + vmlal.u8 q11,d21,d25 @(viii)vmull_u8(ref_main_idx_1, dup_const_fract) + asr lr,lr,#8 @(iii) + + vdup.8 d29,d4[1] @(ii) + vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) + and r9,lr,#0xff @(iii) + + vst1.8 {d14},[r0],r3 @(vi) + vsub.u8 d30,d1,d31 @(i)32-fract(dup_const_32_fract) + add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx] + + vld1.8 {d12},[r12],r11 @(ii)ref_main_idx + vmull.u8 q5,d8,d30 @(i)vmull_u8(ref_main_idx, dup_const_32_fract) + asr lr,lr,#8 @(iv) + + vld1.8 {d13},[r12] @(ii)ref_main_idx_1 + vmlal.u8 q5,d9,d31 @(i)vmull_u8(ref_main_idx_1, dup_const_fract) + and r9,lr,#0xff @(iv) + + vmov.u32 lr,d3[1] @extract idx to the r register + vrshrn.i16 d22,q11,#5 @(viii)shift_res = vrshrn_n_u16(add_res, 5) + + vdup.8 d27,d4[2] @(iii) + vsub.u8 d28,d1,d29 @(ii)32-fract(dup_const_32_fract) + movle r4,r5 @reload nt + + vld1.8 {d16},[r10],r11 @(iii)ref_main_idx + vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract) + add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx] + + vst1.8 {d18},[r0],r3 @(vii) + vmlal.u8 q7,d13,d29 @(ii)vmull_u8(ref_main_idx_1, dup_const_fract) + + vld1.8 {d17},[r10] @(iii)ref_main_idx_1 + vrshrn.i16 d10,q5,#5 @(i)shift_res = vrshrn_n_u16(add_res, 5) + + vdup.8 d25,d4[3] @(iv) + vmull.u8 q1,d5,d0 @pos = ((row + 1) * intra_pred_ang) + + vst1.8 {d22},[r0] @(viii) + vsub.u8 d26,d1,d27 @(iii)32-fract(dup_const_32_fract) + + vld1.8 {d20},[r12],r11 @(iv)ref_main_idx + vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract) + lsl lr,lr,#1 + + vld1.8 {d21},[r12] @(iv)ref_main_idx_1 + vmlal.u8 q9,d17,d27 @(iii)vmull_u8(ref_main_idx_1, dup_const_fract) + add r0,r2,r3 + + vdup.8 d31,d4[4] @(v) + vrshrn.i16 d14,q7,#5 @(ii)shift_res = vrshrn_n_u16(add_res, 5) + and r9,lr,#0xff @(v) + + vst1.8 {d10},[r2]! @(i) + vsub.u8 d24,d1,d25 @(iv)32-fract(dup_const_32_fract) + add r10,r8,r9 @(v)*pu1_ref[ref_main_idx] + + vdup.8 d29,d4[5] @(vi) + vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract) + asr lr,lr,#8 @(vi) + + vdup.8 d27,d4[6] @(vii) + vmlal.u8 q11,d21,d25 @(iv)vmull_u8(ref_main_idx_1, dup_const_fract) + and r9,lr,#0xff @(vi) + + vdup.8 d25,d4[7] @(viii) + vrshrn.i16 d18,q9,#5 @(iii)shift_res = vrshrn_n_u16(add_res, 5) + add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx] + + vld1.8 {d8},[r10],r11 @(v)ref_main_idx + vand q2,q1,q3 @dup_const_fract(fract = pos & (31)) + asr lr,lr,#8 @(vii) + + vld1.8 {d9},[r10] @(v)ref_main_idx_1 + vshrn.u16 d3,q1,#5 @idx = pos >> 5 + and r9,lr,#0xff @(vii) + + vst1.8 {d14},[r0],r3 @(ii) + vrshrn.i16 d22,q11,#5 @(iv)shift_res = vrshrn_n_u16(add_res, 5) + asr lr,lr,#8 @(viii) + + vld1.8 {d12},[r12],r11 @(vi)ref_main_idx + vsub.u8 d30,d1,d31 @(v)32-fract(dup_const_32_fract) + add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx] + + vld1.8 {d13},[r12] @(vi)ref_main_idx_1 + vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract) + and r9,lr,#0xff @(viii) + + vmov.u32 lr,d3[0] @(i)extract idx to the r register + vmlal.u8 q5,d9,d31 @(v)vmull_u8(ref_main_idx_1, dup_const_fract) + add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx] + + vld1.8 {d16},[r10],r11 @(vii)ref_main_idx + vsub.u8 d28,d1,d29 @(vi)32-fract(dup_const_32_fract) + + vst1.8 {d18},[r0],r3 @(iii) + vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract) + movle r8,r1 @reload the source to pu1_src+2nt + + vld1.8 {d17},[r10] @(vii)ref_main_idx_1 + vmlal.u8 q7,d13,d29 @(vi)vmull_u8(ref_main_idx_1, dup_const_fract) + addgt r8,r8,#8 @increment the source next set 8 columns in same row + + vld1.8 {d20},[r12],r11 @(viii)ref_main_idx + vrshrn.i16 d10,q5,#5 @(v)shift_res = vrshrn_n_u16(add_res, 5) + + vld1.8 {d21},[r12] @(viii)ref_main_idx_1 + vsub.u8 d26,d1,d27 @(vii)32-fract(dup_const_32_fract) + lslle r12,r3,#3 + + vst1.8 {d22},[r0],r3 @(iv) + vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract) + suble r12,r12,r5 + + vst1.8 {d10},[r0],r3 @(v) + vmlal.u8 q9,d17,d27 @(vii)vmull_u8(ref_main_idx_1, dup_const_fract) + addle r2,r2,r12 @increment the dst pointer to 8*dst_strd - nt + + vmovn.i16 d4,q2 + vrshrn.i16 d14,q7,#5 @(vi)shift_res = vrshrn_n_u16(add_res, 5) + lsl lr,lr,#1 + + and r9,lr,#0xff @(i) + subs r7,r7,#8 + add r10,r8,r9 @(i)*pu1_ref[ref_main_idx] + + bne kernel_8_rows + +epilogue: + vst1.8 {d14},[r0],r3 @(vi) + vrshrn.i16 d18,q9,#5 @(vii)shift_res = vrshrn_n_u16(add_res, 5) + + vsub.u8 d24,d1,d25 @(viii)32-fract(dup_const_32_fract) + vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract) + vmlal.u8 q11,d21,d25 @(viii)vmull_u8(ref_main_idx_1, dup_const_fract) + + vst1.8 {d18},[r0],r3 @(vii) + vrshrn.i16 d22,q11,#5 @(viii)shift_res = vrshrn_n_u16(add_res, 5) + + vst1.8 {d22},[r0],r3 @(viii) + b end_loops + +core_loop_4: + add r10,r8,#2 @pu1_ref_main_idx += (four_nt + 1) + add r11,r8,#4 @pu1_ref_main_idx_1 += (four_nt + 2) + mov r8,#0 + + add r5,r8,#1 @row + 1 + mul r5,r5,r9 @pos = ((row + 1) * intra_pred_ang) + and r5,r5,#31 @fract = pos & (31) + cmp lr,r5 @if(fract_prev > fract) + addgt r10,r10,#2 @pu1_ref_main_idx += 2 + add r11,r10,#2 @pu1_ref_main_idx_1 += 2 + vdup.8 d0,r5 @dup_const_fract + rsb r4,r5,#32 + vdup.8 d1,r4 @dup_const_32_fract + +@inner_loop_4 + vld1.8 {d2},[r10] @ref_main_idx + add r8,r8,#1 + mov lr,r5 @fract_prev = fract + + vld1.8 {d3},[r11] @ref_main_idx_1 + add r5,r8,#1 @row + 1 + mul r5,r5,r9 @pos = ((row + 1) * intra_pred_ang) + and r5,r5,#31 @fract = pos & (31) + cmp lr,r5 @if(fract_prev > fract) + addgt r10,r10,#2 @pu1_ref_main_idx += 1 + add r11,r10,#2 @pu1_ref_main_idx_1 += 1 + + vdup.8 d6,r5 @dup_const_fract + vmull.u8 q2,d2,d1 @vmull_u8(ref_main_idx, dup_const_32_fract) + + rsb r4,r5,#32 + vdup.8 d7,r4 @dup_const_32_fract + vmlal.u8 q2,d3,d0 @vmull_u8(ref_main_idx_1, dup_const_fract) + + vld1.8 {d8},[r10] @ref_main_idx + add r8,r8,#1 + + vld1.8 {d9},[r11] @ref_main_idx_1 + vrshrn.i16 d4,q2,#5 @shift_res = vrshrn_n_u16(add_res, 5) + + mov lr,r5 @fract_prev = fract + add r5,r8,#1 @row + 1 + mul r5,r5,r9 @pos = ((row + 1) * intra_pred_ang) + and r5,r5,#31 @fract = pos & (31) + cmp lr,r5 @if(fract_prev > fract) + addgt r10,r10,#2 @pu1_ref_main_idx += 1 + add r11,r10,#2 @pu1_ref_main_idx_1 += 1 + + vdup.8 d12,r5 @dup_const_fract + vmull.u8 q5,d8,d7 @vmull_u8(ref_main_idx, dup_const_32_fract) + + rsb r4,r5,#32 + vdup.8 d13,r4 @dup_const_32_fract + vmlal.u8 q5,d9,d6 @vmull_u8(ref_main_idx_1, dup_const_fract) + + vld1.8 {d14},[r10] @ref_main_idx + add r8,r8,#1 + + vst1.8 {d4},[r2],r3 + vrshrn.i16 d10,q5,#5 @shift_res = vrshrn_n_u16(add_res, 5) + + vld1.8 {d15},[r11] @ref_main_idx_1 + mov lr,r5 @fract_prev = fract + add r5,r8,#1 @row + 1 + mul r5,r5,r9 @pos = ((row + 1) * intra_pred_ang) + and r5,r5,#31 @fract = pos & (31) + cmp lr,r5 @if(fract_prev > fract) + addgt r10,r10,#2 @pu1_ref_main_idx += 1 + add r11,r10,#2 @pu1_ref_main_idx_1 += 1 + + vdup.8 d18,r5 @dup_const_fract + vmull.u8 q8,d14,d13 @vmull_u8(ref_main_idx, dup_const_32_fract) + + rsb r4,r5,#32 + vdup.8 d19,r4 @dup_const_32_fract + vmlal.u8 q8,d15,d12 @vmull_u8(ref_main_idx_1, dup_const_fract) + + vld1.8 {d20},[r10] @ref_main_idx + + vst1.8 {d10},[r2],r3 + vrshrn.i16 d16,q8,#5 @shift_res = vrshrn_n_u16(add_res, 5) + vld1.8 {d21},[r11] @ref_main_idx_1 + + vmull.u8 q11,d20,d19 @vmull_u8(ref_main_idx, dup_const_32_fract) + vmlal.u8 q11,d21,d18 @vmull_u8(ref_main_idx_1, dup_const_fract) + + vst1.8 {d16},[r2],r3 + vrshrn.i16 d22,q11,#5 @shift_res = vrshrn_n_u16(add_res, 5) + + vst1.8 {d22},[r2],r3 + +end_loops: + ldmfd sp!,{r4-r12,r15} @reload the registers from sp + + + + |