diff options
author | Måns Rullgård <mans@mansr.com> | 2007-02-05 21:18:11 +0000 |
---|---|---|
committer | Måns Rullgård <mans@mansr.com> | 2007-02-05 21:18:11 +0000 |
commit | 7ee829922b1053a2496e5f0849265c4cb9060baf (patch) | |
tree | 0dcbfeea4da20e26b226ed5c875ca549aee2080e /libavcodec/armv4l/simple_idct_armv6.S | |
parent | 271593f123caf46d4703f278bc59134c07cd47d1 (diff) | |
download | ffmpeg-7ee829922b1053a2496e5f0849265c4cb9060baf.tar.gz |
oops, revert accidental checkin
Originally committed as revision 7840 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diffstat (limited to 'libavcodec/armv4l/simple_idct_armv6.S')
-rw-r--r-- | libavcodec/armv4l/simple_idct_armv6.S | 144 |
1 files changed, 40 insertions, 104 deletions
diff --git a/libavcodec/armv4l/simple_idct_armv6.S b/libavcodec/armv4l/simple_idct_armv6.S index 328e0a9147..401e1910df 100644 --- a/libavcodec/armv4l/simple_idct_armv6.S +++ b/libavcodec/armv4l/simple_idct_armv6.S @@ -47,19 +47,6 @@ w42n: .long W42n w46: .long W46 w57: .long W57 - .macro idct_row_start shift - ldr ip, [pc, #(w42-.-8)] /* ip = W4 | (W2 << 16) */ - ldr lr, [pc, #(w46-.-8)] /* lr = W4 | (W6 << 16) */ - ldr v7, [pc, #(w57-.-8)] /* v7 = W5 | (W7 << 16) */ - mov a2, #(1<<(\shift-1)) - smlad v1, a3, ip, a2 - smlsd v4, a3, ip, a2 - ldr ip, [pc, #(w13-.-8)] /* ip = W1 | (W3 << 16) */ - smlad v2, a3, lr, a2 - smlsd v3, a3, lr, a2 - smusdx fp, a4, v7 /* fp = B3 = W7*row[1] - W5*row[3] */ - smuad v5, a4, ip /* v5 = B0 = W1*row[1] + W3*row[3] */ - .endm /* Compute partial IDCT of single row. shift = left-shift amount @@ -71,6 +58,17 @@ w57: .long W57 Output in registers v1--v8 */ .macro idct_row shift + ldr lr, [pc, #(w46-.-8)] /* lr = W4 | (W6 << 16) */ + mov a2, #(1<<(\shift-1)) + smlad v1, a3, ip, a2 + smlsd v4, a3, ip, a2 + ldr ip, [pc, #(w13-.-8)] /* ip = W1 | (W3 << 16) */ + ldr v7, [pc, #(w57-.-8)] /* v7 = W5 | (W7 << 16) */ + smlad v2, a3, lr, a2 + smlsd v3, a3, lr, a2 + + smuad v5, a4, ip /* v5 = B0 = W1*row[1] + W3*row[3] */ + smusdx fp, a4, v7 /* fp = B3 = W7*row[1] - W5*row[3] */ ldr lr, [a1, #12] /* lr = row[7,5] */ pkhtb a3, ip, v7, asr #16 /* a4 = W7 | (W3 << 16) */ pkhbt a2, ip, v7, lsl #16 /* a2 = W1 | (W5 << 16) */ @@ -101,6 +99,16 @@ w57: .long W57 Output in registers v1--v8 */ .macro idct_row4 shift + ldr lr, [pc, #(w46-.-8)] /* lr = W4 | (W6 << 16) */ + ldr v7, [pc, #(w57-.-8)] /* v7 = W5 | (W7 << 16) */ + mov a2, #(1<<(\shift-1)) + smlad v1, a3, ip, a2 + smlsd v4, a3, ip, a2 + ldr ip, [pc, #(w13-.-8)] /* ip = W1 | (W3 << 16) */ + smlad v2, a3, lr, a2 + smlsd v3, a3, lr, a2 + smusdx fp, a4, v7 /* fp = B3 = W7*row[1] - W5*row[3] */ + smuad v5, a4, ip /* v5 = B0 = W1*row[1] + W3*row[3] */ pkhtb a3, ip, v7, asr #16 /* a4 = W7 | (W3 << 16) */ pkhbt a2, ip, v7, lsl #16 /* a2 = W1 | (W5 << 16) */ smusdx v6, a3, a4 /* v6 = -B1 = W7*row[3] - W3*row[1] */ @@ -185,100 +193,28 @@ w57: .long W57 .align .func idct_row_armv6 idct_row_armv6: - ldr fp, [a1, #12] /* fp = row[7,5] */ - ldr v7, [a1, #4] /* v7 = row[6,4] */ + str lr, [sp, #-4]! + + ldr lr, [a1, #12] /* lr = row[7,5] */ + ldr ip, [a1, #4] /* ip = row[6,4] */ ldr a4, [a1, #8] /* a4 = row[3,1] */ ldr a3, [a1] /* a3 = row[2,0] */ - mov ip, #(1<<(ROW_SHIFT-1)) - orrs v5, fp, v7 - cmpeq v5, a4 - cmpeq v5, a3, lsr #16 + orrs lr, lr, ip + cmpeq lr, a4 + cmpeq lr, a3, lsr #16 beq 1f - cmp v5, #0 - stmfd sp!, {a2, lr} - ldr v5, [pc, #(w42-.-8)] /* v5 = W4 | (W2 << 16) */ - ldr v6, [pc, #(w46-.-8)] /* v6 = W4 | (W6 << 16) */ - ldr v7, [pc, #(w57-.-8)] /* v7 = W5 | (W7 << 16) */ - - smlad v1, a3, v5, ip - smlsd v4, a3, v5, ip - ldr a2, [pc, #(w13-.-8)] /* a2 = W1 | (W3 << 16) */ - smlad v2, a3, v6, ip - smlsd v3, a3, v6, ip - smusdx lr, a4, v7 /* lr = B3 = W7*row[1] - W5*row[3] */ - smuad v5, a4, a2 /* v5 = B0 = W1*row[1] + W3*row[3] */ - - pkhtb a3, a2, v7, asr #16 /* a3 = W7 | (W3 << 16) */ - pkhbt ip, a2, v7, lsl #16 /* ip = W1 | (W5 << 16) */ - smusdx v6, a3, a4 /* v6 = -B1 = W7*row[3] - W3*row[1] */ - smusdx a4, a4, ip /* v7 = B2 = W5*row[1] - W1*row[3] */ - beq 3f - - smlad v5, fp, v7, v5 /* B0 += W5*row[5] + W7*row[7] */ - smlad v7, fp, a3, a4 /* B2 += W7*row[5] + W3*row[7] */ - ldr a4, [pc, #(w42n-.-8)] /* a4 = -W4 | (-W2 << 16) */ - ldr a3, [a1, #4] /* a3 = row[6,4] */ - smlsdx lr, fp, a2, lr /* B3 += W3*row[5] - W1*row[7] */ - ldr a2, [pc, #(w46-.-8)] /* a2 = W4 | (W6 << 16) */ - smlad v6, fp, ip, v6 /* B1 -= W1*row[5] + W5*row[7] */ - - smlad v2, a3, a4, v2 /* A1 += -W4*row[4] - W2*row[6] */ - smlsd v3, a3, a4, v3 /* A2 += -W4*row[4] + W2*row[6] */ - smlad v1, a3, a2, v1 /* A0 += W4*row[4] + W6*row[6] */ - smlsd v4, a3, a2, v4 /* A3 += W4*row[4] - W6*row[6] */ - - ldr a2, [sp], #4 - add a4, v1, v5 /* a4 = A0 + B0 */ - sub a3, v1, v5 /* a3 = A0 - B0 */ - mov v1, a4, asr #ROW_SHIFT - mov v5, a3, asr #ROW_SHIFT - - sub a4, v2, v6 /* a4 = A1 + B1 */ - add a3, v2, v6 /* a3 = A1 - B1 */ - mov v2, a4, asr #ROW_SHIFT - mov v6, a3, asr #ROW_SHIFT - - add a4, v3, v7 /* a4 = A2 + B2 */ - sub a3, v3, v7 /* a3 = A2 - B2 */ - mov v3, a4, asr #ROW_SHIFT - mov v7, a3, asr #ROW_SHIFT - - add a4, v4, lr /* a4 = A3 + B3 */ - sub a3, v4, lr /* a3 = A3 - B3 */ - mov v4, a4, asr #ROW_SHIFT - mov fp, a3, asr #ROW_SHIFT + str a2, [sp, #-4]! + ldr ip, [pc, #(w42-.-8)] /* ip = W4 | (W2 << 16) */ + cmp lr, #0 + beq 2f - strh v1, [a2] - strh v2, [a2, #(16*2)] - strh v3, [a2, #(16*4)] - strh v4, [a2, #(16*6)] - strh fp, [a2, #(16*1)] - strh v7, [a2, #(16*3)] - strh v6, [a2, #(16*5)] - strh v5, [a2, #(16*7)] + idct_row ROW_SHIFT + b 3f - ldr pc, [sp], #4 +2: idct_row4 ROW_SHIFT 3: ldr a2, [sp], #4 - add v7, v1, v5 /* v7 = A0 + B0 */ - sub a3, v1, v5 /* a3 = A0 - B0 */ - mov v1, v7, asr #ROW_SHIFT - mov v5, a3, asr #ROW_SHIFT - - sub v7, v2, v6 /* v7 = A1 + B1 */ - add a3, v2, v6 /* a3 = A1 - B1 */ - mov v2, v7, asr #ROW_SHIFT - mov v6, a3, asr #ROW_SHIFT - - add v7, v3, a4 /* v7 = A2 + B2 */ - sub a3, v3, a4 /* a3 = A2 - B2 */ - mov v3, v7, asr #ROW_SHIFT - mov v7, a3, asr #ROW_SHIFT - - add a4, v4, lr /* xx = A3 + B3 */ - sub a3, v4, lr /* a3 = A3 - B3 */ - mov v4, a4, asr #ROW_SHIFT - mov fp, a3, asr #ROW_SHIFT + idct_finish_shift ROW_SHIFT strh v1, [a2] strh v2, [a2, #(16*2)] @@ -300,7 +236,7 @@ idct_row_armv6: strh a3, [a2, #(16*3)] strh a3, [a2, #(16*5)] strh a3, [a2, #(16*7)] - mov pc, lr + ldr pc, [sp], #4 .endfunc /* @@ -314,8 +250,8 @@ idct_col_armv6: stmfd sp!, {a2, lr} ldr a3, [a1] /* a3 = row[2,0] */ + ldr ip, [pc, #(w42-.-8)] /* ip = W4 | (W2 << 16) */ ldr a4, [a1, #8] /* a4 = row[3,1] */ - idct_row_start COL_SHIFT idct_row COL_SHIFT ldr a2, [sp], #4 idct_finish_shift COL_SHIFT @@ -344,8 +280,8 @@ idct_col_put_armv6: stmfd sp!, {a2, a3, lr} ldr a3, [a1] /* a3 = row[2,0] */ + ldr ip, [pc, #(w42-.-8)] /* ip = W4 | (W2 << 16) */ ldr a4, [a1, #8] /* a4 = row[3,1] */ - idct_row_start COL_SHIFT idct_row COL_SHIFT ldmfd sp!, {a2, a3} idct_finish_shift_sat COL_SHIFT @@ -376,8 +312,8 @@ idct_col_add_armv6: stmfd sp!, {a2, a3, lr} ldr a3, [a1] /* a3 = row[2,0] */ + ldr ip, [pc, #(w42-.-8)] /* ip = W4 | (W2 << 16) */ ldr a4, [a1, #8] /* a4 = row[3,1] */ - idct_row_start COL_SHIFT idct_row COL_SHIFT ldmfd sp!, {a2, a3} idct_finish |