early-access version 1432
This commit is contained in:
358
externals/ffmpeg/libavresample/arm/resample_neon.S
vendored
Executable file
358
externals/ffmpeg/libavresample/arm/resample_neon.S
vendored
Executable file
@@ -0,0 +1,358 @@
|
||||
/*
|
||||
* Copyright (c) 2014 Peter Meerwald <pmeerw@pmeerw.net>
|
||||
*
|
||||
* This file is part of FFmpeg.
|
||||
*
|
||||
* FFmpeg is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU Lesser General Public
|
||||
* License as published by the Free Software Foundation; either
|
||||
* version 2.1 of the License, or (at your option) any later version.
|
||||
*
|
||||
* FFmpeg is distributed in the hope that it will be useful,
|
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
* Lesser General Public License for more details.
|
||||
*
|
||||
* You should have received a copy of the GNU Lesser General Public
|
||||
* License along with FFmpeg; if not, write to the Free Software
|
||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
||||
*/
|
||||
|
||||
#include "libavutil/arm/asm.S"
|
||||
|
||||
#include "asm-offsets.h"
|
||||
|
||||
.macro resample_one fmt, es=2
|
||||
function ff_resample_one_\fmt\()_neon, export=1
|
||||
push {r4, r5}
|
||||
add r1, r1, r2, lsl #\es
|
||||
|
||||
ldr r2, [r0, #PHASE_SHIFT+4] /* phase_mask */
|
||||
ldr ip, [sp, #8] /* index */
|
||||
ldr r5, [r0, #FILTER_LENGTH]
|
||||
and r2, ip, r2 /* (index & phase_mask) */
|
||||
ldr r4, [r0, #PHASE_SHIFT]
|
||||
lsr r4, ip, r4 /* compute sample_index */
|
||||
mul r2, r2, r5
|
||||
|
||||
ldr ip, [r0, #FILTER_BANK]
|
||||
add r3, r3, r4, lsl #\es /* &src[sample_index] */
|
||||
|
||||
cmp r5, #8
|
||||
add r0, ip, r2, lsl #\es /* filter = &filter_bank[...] */
|
||||
|
||||
blt 5f
|
||||
8:
|
||||
subs r5, r5, #8
|
||||
LOAD4
|
||||
MUL4
|
||||
7:
|
||||
LOAD4
|
||||
beq 6f
|
||||
cmp r5, #8
|
||||
MLA4
|
||||
blt 4f
|
||||
subs r5, r5, #8
|
||||
LOAD4
|
||||
MLA4
|
||||
b 7b
|
||||
6:
|
||||
MLA4
|
||||
STORE
|
||||
pop {r4, r5}
|
||||
bx lr
|
||||
5:
|
||||
INIT4
|
||||
4: /* remaining filter_length 1 to 7 */
|
||||
cmp r5, #4
|
||||
blt 2f
|
||||
subs r5, r5, #4
|
||||
LOAD4
|
||||
MLA4
|
||||
beq 0f
|
||||
2: /* remaining filter_length 1 to 3 */
|
||||
cmp r5, #2
|
||||
blt 1f
|
||||
subs r5, r5, #2
|
||||
LOAD2
|
||||
MLA2
|
||||
beq 0f
|
||||
1: /* remaining filter_length 1 */
|
||||
LOAD1
|
||||
MLA1
|
||||
0:
|
||||
STORE
|
||||
pop {r4, r5}
|
||||
bx lr
|
||||
endfunc
|
||||
|
||||
.purgem LOAD1
|
||||
.purgem LOAD2
|
||||
.purgem LOAD4
|
||||
.purgem MLA1
|
||||
.purgem MLA2
|
||||
.purgem MLA4
|
||||
.purgem MUL4
|
||||
.purgem INIT4
|
||||
.purgem STORE
|
||||
.endm
|
||||
|
||||
|
||||
/* float32 */
|
||||
.macro LOAD1
|
||||
veor.32 d0, d0
|
||||
vld1.32 {d0[0]}, [r0]! /* load filter */
|
||||
vld1.32 {d4[0]}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD2
|
||||
vld1.32 {d0}, [r0]! /* load filter */
|
||||
vld1.32 {d4}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD4
|
||||
vld1.32 {d0,d1}, [r0]! /* load filter */
|
||||
vld1.32 {d4,d5}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro MLA1
|
||||
vmla.f32 d16, d0, d4[0]
|
||||
.endm
|
||||
.macro MLA2
|
||||
vmla.f32 d16, d0, d4
|
||||
.endm
|
||||
.macro MLA4
|
||||
vmla.f32 d16, d0, d4
|
||||
vmla.f32 d17, d1, d5
|
||||
.endm
|
||||
.macro MUL4
|
||||
vmul.f32 d16, d0, d4
|
||||
vmul.f32 d17, d1, d5
|
||||
.endm
|
||||
.macro INIT4
|
||||
veor.f32 q8, q8
|
||||
.endm
|
||||
.macro STORE
|
||||
vpadd.f32 d16, d16, d17
|
||||
vpadd.f32 d16, d16, d16
|
||||
vst1.32 d16[0], [r1]
|
||||
.endm
|
||||
|
||||
resample_one flt, 2
|
||||
|
||||
|
||||
/* s32 */
|
||||
.macro LOAD1
|
||||
veor.32 d0, d0
|
||||
vld1.32 {d0[0]}, [r0]! /* load filter */
|
||||
vld1.32 {d4[0]}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD2
|
||||
vld1.32 {d0}, [r0]! /* load filter */
|
||||
vld1.32 {d4}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD4
|
||||
vld1.32 {d0,d1}, [r0]! /* load filter */
|
||||
vld1.32 {d4,d5}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro MLA1
|
||||
vmlal.s32 q8, d0, d4[0]
|
||||
.endm
|
||||
.macro MLA2
|
||||
vmlal.s32 q8, d0, d4
|
||||
.endm
|
||||
.macro MLA4
|
||||
vmlal.s32 q8, d0, d4
|
||||
vmlal.s32 q9, d1, d5
|
||||
.endm
|
||||
.macro MUL4
|
||||
vmull.s32 q8, d0, d4
|
||||
vmull.s32 q9, d1, d5
|
||||
.endm
|
||||
.macro INIT4
|
||||
veor.s64 q8, q8
|
||||
veor.s64 q9, q9
|
||||
.endm
|
||||
.macro STORE
|
||||
vadd.s64 q8, q8, q9
|
||||
vadd.s64 d16, d16, d17
|
||||
vqrshrn.s64 d16, q8, #30
|
||||
vst1.32 d16[0], [r1]
|
||||
.endm
|
||||
|
||||
resample_one s32, 2
|
||||
|
||||
|
||||
/* s16 */
|
||||
.macro LOAD1
|
||||
veor.16 d0, d0
|
||||
vld1.16 {d0[0]}, [r0]! /* load filter */
|
||||
vld1.16 {d4[0]}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD2
|
||||
veor.16 d0, d0
|
||||
vld1.32 {d0[0]}, [r0]! /* load filter */
|
||||
veor.16 d4, d4
|
||||
vld1.32 {d4[0]}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD4
|
||||
vld1.16 {d0}, [r0]! /* load filter */
|
||||
vld1.16 {d4}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro MLA1
|
||||
vmlal.s16 q8, d0, d4[0]
|
||||
.endm
|
||||
.macro MLA2
|
||||
vmlal.s16 q8, d0, d4
|
||||
.endm
|
||||
.macro MLA4
|
||||
vmlal.s16 q8, d0, d4
|
||||
.endm
|
||||
.macro MUL4
|
||||
vmull.s16 q8, d0, d4
|
||||
.endm
|
||||
.macro INIT4
|
||||
veor.s32 q8, q8
|
||||
.endm
|
||||
.macro STORE
|
||||
vpadd.s32 d16, d16, d17
|
||||
vpadd.s32 d16, d16, d16
|
||||
vqrshrn.s32 d16, q8, #15
|
||||
vst1.16 d16[0], [r1]
|
||||
.endm
|
||||
|
||||
resample_one s16, 1
|
||||
|
||||
|
||||
.macro resample_linear fmt, es=2
|
||||
function ff_resample_linear_\fmt\()_neon, export=1
|
||||
push {r4, r5}
|
||||
add r1, r1, r2, lsl #\es
|
||||
|
||||
ldr r2, [r0, #PHASE_SHIFT+4] /* phase_mask */
|
||||
ldr ip, [sp, #8] /* index */
|
||||
ldr r5, [r0, #FILTER_LENGTH]
|
||||
and r2, ip, r2 /* (index & phase_mask) */
|
||||
ldr r4, [r0, #PHASE_SHIFT]
|
||||
lsr r4, ip, r4 /* compute sample_index */
|
||||
mul r2, r2, r5
|
||||
|
||||
ldr ip, [r0, #FILTER_BANK]
|
||||
add r3, r3, r4, lsl #\es /* &src[sample_index] */
|
||||
|
||||
cmp r5, #8
|
||||
ldr r4, [r0, #SRC_INCR]
|
||||
add r0, ip, r2, lsl #\es /* filter = &filter_bank[...] */
|
||||
add r2, r0, r5, lsl #\es /* filter[... + c->filter_length] */
|
||||
|
||||
blt 5f
|
||||
8:
|
||||
subs r5, r5, #8
|
||||
LOAD4
|
||||
MUL4
|
||||
7:
|
||||
LOAD4
|
||||
beq 6f
|
||||
cmp r5, #8
|
||||
MLA4
|
||||
blt 4f
|
||||
subs r5, r5, #8
|
||||
LOAD4
|
||||
MLA4
|
||||
b 7b
|
||||
6:
|
||||
MLA4
|
||||
STORE
|
||||
pop {r4, r5}
|
||||
bx lr
|
||||
5:
|
||||
INIT4
|
||||
4: /* remaining filter_length 1 to 7 */
|
||||
cmp r5, #4
|
||||
blt 2f
|
||||
subs r5, r5, #4
|
||||
LOAD4
|
||||
MLA4
|
||||
beq 0f
|
||||
2: /* remaining filter_length 1 to 3 */
|
||||
cmp r5, #2
|
||||
blt 1f
|
||||
subs r5, r5, #2
|
||||
LOAD2
|
||||
MLA2
|
||||
beq 0f
|
||||
1: /* remaining filter_length 1 */
|
||||
LOAD1
|
||||
MLA1
|
||||
0:
|
||||
STORE
|
||||
pop {r4, r5}
|
||||
bx lr
|
||||
endfunc
|
||||
|
||||
.purgem LOAD1
|
||||
.purgem LOAD2
|
||||
.purgem LOAD4
|
||||
.purgem MLA1
|
||||
.purgem MLA2
|
||||
.purgem MLA4
|
||||
.purgem MUL4
|
||||
.purgem INIT4
|
||||
.purgem STORE
|
||||
.endm
|
||||
|
||||
|
||||
/* float32 linear */
|
||||
.macro LOAD1
|
||||
veor.32 d0, d0
|
||||
veor.32 d2, d2
|
||||
vld1.32 {d0[0]}, [r0]! /* load filter */
|
||||
vld1.32 {d2[0]}, [r2]! /* load filter */
|
||||
vld1.32 {d4[0]}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD2
|
||||
vld1.32 {d0}, [r0]! /* load filter */
|
||||
vld1.32 {d2}, [r2]! /* load filter */
|
||||
vld1.32 {d4}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro LOAD4
|
||||
vld1.32 {d0,d1}, [r0]! /* load filter */
|
||||
vld1.32 {d2,d3}, [r2]! /* load filter */
|
||||
vld1.32 {d4,d5}, [r3]! /* load src */
|
||||
.endm
|
||||
.macro MLA1
|
||||
vmla.f32 d18, d0, d4[0]
|
||||
vmla.f32 d16, d2, d4[0]
|
||||
.endm
|
||||
.macro MLA2
|
||||
vmla.f32 d18, d0, d4
|
||||
vmla.f32 d16, d2, d4
|
||||
.endm
|
||||
.macro MLA4
|
||||
vmla.f32 q9, q0, q2
|
||||
vmla.f32 q8, q1, q2
|
||||
.endm
|
||||
.macro MUL4
|
||||
vmul.f32 q9, q0, q2
|
||||
vmul.f32 q8, q1, q2
|
||||
.endm
|
||||
.macro INIT4
|
||||
veor.f32 q9, q9
|
||||
veor.f32 q8, q8
|
||||
.endm
|
||||
.macro STORE
|
||||
vldr s0, [sp, #12] /* frac */
|
||||
vmov s1, r4
|
||||
vcvt.f32.s32 d0, d0
|
||||
|
||||
vsub.f32 q8, q8, q9 /* v2 - val */
|
||||
vpadd.f32 d18, d18, d19
|
||||
vpadd.f32 d16, d16, d17
|
||||
vpadd.f32 d2, d18, d18
|
||||
vpadd.f32 d1, d16, d16
|
||||
|
||||
vmul.f32 s2, s2, s0 /* (v2 - val) * frac */
|
||||
vdiv.f32 s2, s2, s1 /* / c->src_incr */
|
||||
vadd.f32 s4, s4, s2
|
||||
|
||||
vstr s4, [r1]
|
||||
.endm
|
||||
|
||||
resample_linear flt, 2
|
Reference in New Issue
Block a user