301 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			NASM
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			301 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			NASM
		
	
	
		
			Executable File
		
	
	
	
	
| ;******************************************************************************
 | |
| ;* Copyright Nick Kurshev
 | |
| ;* Copyright Michael (michaelni@gmx.at)
 | |
| ;* Copyright 2018 Jokyo Images
 | |
| ;* Copyright Ivo van Poorten
 | |
| ;*
 | |
| ;* This file is part of FFmpeg.
 | |
| ;*
 | |
| ;* FFmpeg is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* FFmpeg is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with FFmpeg; if not, write to the Free Software
 | |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| 
 | |
| SECTION_RODATA
 | |
| 
 | |
| pb_mask_shuffle2103_mmx times 8 dw 255
 | |
| pb_shuffle2103: db 2, 1, 0, 3, 6, 5, 4, 7, 10, 9, 8, 11, 14, 13, 12, 15
 | |
| pb_shuffle0321: db 0, 3, 2, 1, 4, 7, 6, 5, 8, 11, 10, 9, 12, 15, 14, 13
 | |
| pb_shuffle1230: db 1, 2, 3, 0, 5, 6, 7, 4, 9, 10, 11, 8, 13, 14, 15, 12
 | |
| pb_shuffle3012: db 3, 0, 1, 2, 7, 4, 5, 6, 11, 8, 9, 10, 15, 12, 13, 14
 | |
| pb_shuffle3210: db 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12
 | |
| 
 | |
| SECTION .text
 | |
| 
 | |
| %macro RSHIFT_COPY 3
 | |
| ; %1 dst ; %2 src ; %3 shift
 | |
| %if cpuflag(avx)
 | |
|     psrldq  %1, %2, %3
 | |
| %else
 | |
|     mova           %1, %2
 | |
|     RSHIFT         %1, %3
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; shuffle_bytes_2103_mmext (const uint8_t *src, uint8_t *dst, int src_size)
 | |
| ;------------------------------------------------------------------------------
 | |
| INIT_MMX mmxext
 | |
| cglobal shuffle_bytes_2103, 3, 5, 8, src, dst, w, tmp, x
 | |
|     mova   m6, [pb_mask_shuffle2103_mmx]
 | |
|     mova   m7, m6
 | |
|     psllq  m7, 8
 | |
| 
 | |
|     movsxdifnidn wq, wd
 | |
|     mov xq, wq
 | |
| 
 | |
|     add        srcq, wq
 | |
|     add        dstq, wq
 | |
|     neg          wq
 | |
| 
 | |
| ;calc scalar loop
 | |
|     and xq, mmsize*2 -4
 | |
|     je .loop_simd
 | |
| 
 | |
| .loop_scalar:
 | |
|    mov          tmpb, [srcq + wq + 2]
 | |
|    mov [dstq+wq + 0], tmpb
 | |
|    mov          tmpb, [srcq + wq + 1]
 | |
|    mov [dstq+wq + 1], tmpb
 | |
|    mov          tmpb, [srcq + wq + 0]
 | |
|    mov [dstq+wq + 2], tmpb
 | |
|    mov          tmpb, [srcq + wq + 3]
 | |
|    mov [dstq+wq + 3], tmpb
 | |
|    add            wq, 4
 | |
|    sub            xq, 4
 | |
|    jg .loop_scalar
 | |
| 
 | |
| ;check if src_size < mmsize * 2
 | |
| cmp wq, 0
 | |
| jge .end
 | |
| 
 | |
| .loop_simd:
 | |
|     movu     m0, [srcq+wq]
 | |
|     movu     m1, [srcq+wq+8]
 | |
| 
 | |
|     pshufw   m3, m0, 177
 | |
|     pshufw   m5, m1, 177
 | |
| 
 | |
|     pand     m0, m7
 | |
|     pand     m3, m6
 | |
| 
 | |
|     pand     m1, m7
 | |
|     pand     m5, m6
 | |
| 
 | |
|     por      m0, m3
 | |
|     por      m1, m5
 | |
| 
 | |
|     movu      [dstq+wq], m0
 | |
|     movu  [dstq+wq + 8], m1
 | |
| 
 | |
|     add              wq, mmsize*2
 | |
|     jl .loop_simd
 | |
| 
 | |
| .end:
 | |
|     RET
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; shuffle_bytes_## (const uint8_t *src, uint8_t *dst, int src_size)
 | |
| ;------------------------------------------------------------------------------
 | |
| ; %1-4 index shuffle
 | |
| %macro SHUFFLE_BYTES 4
 | |
| cglobal shuffle_bytes_%1%2%3%4, 3, 5, 2, src, dst, w, tmp, x
 | |
|     VBROADCASTI128    m0, [pb_shuffle%1%2%3%4]
 | |
|     movsxdifnidn wq, wd
 | |
|     mov xq, wq
 | |
| 
 | |
|     add        srcq, wq
 | |
|     add        dstq, wq
 | |
|     neg          wq
 | |
| 
 | |
| ;calc scalar loop
 | |
|     and xq, mmsize-4
 | |
|     je .loop_simd
 | |
| 
 | |
| .loop_scalar:
 | |
|    mov          tmpb, [srcq + wq + %1]
 | |
|    mov [dstq+wq + 0], tmpb
 | |
|    mov          tmpb, [srcq + wq + %2]
 | |
|    mov [dstq+wq + 1], tmpb
 | |
|    mov          tmpb, [srcq + wq + %3]
 | |
|    mov [dstq+wq + 2], tmpb
 | |
|    mov          tmpb, [srcq + wq + %4]
 | |
|    mov [dstq+wq + 3], tmpb
 | |
|    add            wq, 4
 | |
|    sub            xq, 4
 | |
|    jg .loop_scalar
 | |
| 
 | |
| ;check if src_size < mmsize
 | |
| cmp wq, 0
 | |
| jge .end
 | |
| 
 | |
| .loop_simd:
 | |
|     movu           m1, [srcq+wq]
 | |
|     pshufb         m1, m0
 | |
|     movu    [dstq+wq], m1
 | |
|     add            wq, mmsize
 | |
|     jl .loop_simd
 | |
| 
 | |
| .end:
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM ssse3
 | |
| SHUFFLE_BYTES 2, 1, 0, 3
 | |
| SHUFFLE_BYTES 0, 3, 2, 1
 | |
| SHUFFLE_BYTES 1, 2, 3, 0
 | |
| SHUFFLE_BYTES 3, 0, 1, 2
 | |
| SHUFFLE_BYTES 3, 2, 1, 0
 | |
| 
 | |
| ;-----------------------------------------------------------------------------------------------
 | |
| ; uyvytoyuv422(uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 | |
| ;              const uint8_t *src, int width, int height,
 | |
| ;              int lumStride, int chromStride, int srcStride)
 | |
| ;-----------------------------------------------------------------------------------------------
 | |
| %macro UYVY_TO_YUV422 0
 | |
| cglobal uyvytoyuv422, 9, 14, 8, ydst, udst, vdst, src, w, h, lum_stride, chrom_stride, src_stride, wtwo, whalf, tmp, x, back_w
 | |
|     pxor         m0, m0
 | |
|     pcmpeqw      m1, m1
 | |
|     psrlw        m1, 8
 | |
| 
 | |
|     movsxdifnidn            wq, wd
 | |
|     movsxdifnidn   lum_strideq, lum_strided
 | |
|     movsxdifnidn chrom_strideq, chrom_strided
 | |
|     movsxdifnidn   src_strideq, src_strided
 | |
| 
 | |
|     mov     back_wq, wq
 | |
|     mov      whalfq, wq
 | |
|     shr      whalfq, 1     ; whalf = width / 2
 | |
| 
 | |
|     lea srcq, [srcq + wq * 2]
 | |
|     add    ydstq, wq
 | |
|     add    udstq, whalfq
 | |
|     add    vdstq, whalfq
 | |
| 
 | |
| .loop_line:
 | |
|     mov          xq, wq
 | |
|     mov       wtwoq, wq
 | |
|     add       wtwoq, wtwoq ; wtwo = width * 2
 | |
| 
 | |
|     neg       wq
 | |
|     neg    wtwoq
 | |
|     neg   whalfq
 | |
| 
 | |
|     ;calc scalar loop count
 | |
|     and       xq, mmsize * 2 - 1
 | |
|     je .loop_simd
 | |
| 
 | |
|     .loop_scalar:
 | |
|         mov             tmpb, [srcq + wtwoq + 0]
 | |
|         mov [udstq + whalfq], tmpb
 | |
| 
 | |
|         mov             tmpb, [srcq + wtwoq + 1]
 | |
|         mov     [ydstq + wq], tmpb
 | |
| 
 | |
|         mov             tmpb, [srcq + wtwoq + 2]
 | |
|         mov [vdstq + whalfq], tmpb
 | |
| 
 | |
|         mov             tmpb, [srcq + wtwoq + 3]
 | |
|         mov [ydstq + wq + 1], tmpb
 | |
| 
 | |
|         add      wq, 2
 | |
|         add   wtwoq, 4
 | |
|         add  whalfq, 1
 | |
|         sub      xq, 2
 | |
|         jg .loop_scalar
 | |
| 
 | |
|     ; check if simd loop is need
 | |
|     cmp      wq, 0
 | |
|     jge .end_line
 | |
| 
 | |
|     .loop_simd:
 | |
|         movu    m2, [srcq + wtwoq             ]
 | |
|         movu    m3, [srcq + wtwoq + mmsize    ]
 | |
|         movu    m4, [srcq + wtwoq + mmsize * 2]
 | |
|         movu    m5, [srcq + wtwoq + mmsize * 3]
 | |
| 
 | |
|         ; extract y part 1
 | |
|         RSHIFT_COPY    m6, m2, 1 ; UYVY UYVY -> YVYU YVY...
 | |
|         pand           m6, m1; YxYx YxYx...
 | |
| 
 | |
|         RSHIFT_COPY    m7, m3, 1 ; UYVY UYVY -> YVYU YVY...
 | |
|         pand           m7, m1 ; YxYx YxYx...
 | |
| 
 | |
|         packuswb       m6, m7 ; YYYY YYYY...
 | |
|         movu [ydstq + wq], m6
 | |
| 
 | |
|         ; extract y part 2
 | |
|         RSHIFT_COPY    m6, m4, 1 ; UYVY UYVY -> YVYU YVY...
 | |
|         pand           m6, m1; YxYx YxYx...
 | |
| 
 | |
|         RSHIFT_COPY    m7, m5, 1 ; UYVY UYVY -> YVYU YVY...
 | |
|         pand           m7, m1 ; YxYx YxYx...
 | |
| 
 | |
|         packuswb                m6, m7 ; YYYY YYYY...
 | |
|         movu [ydstq + wq + mmsize], m6
 | |
| 
 | |
|         ; extract uv
 | |
|         pand       m2, m1   ; UxVx...
 | |
|         pand       m3, m1   ; UxVx...
 | |
|         pand       m4, m1   ; UxVx...
 | |
|         pand       m5, m1   ; UxVx...
 | |
| 
 | |
|         packuswb   m2, m3   ; UVUV...
 | |
|         packuswb   m4, m5   ; UVUV...
 | |
| 
 | |
|         ; U
 | |
|         pand       m6, m2, m1 ; UxUx...
 | |
|         pand       m7, m4, m1 ; UxUx...
 | |
| 
 | |
|         packuswb m6, m7 ; UUUU
 | |
|         movu   [udstq + whalfq], m6
 | |
| 
 | |
| 
 | |
|         ; V
 | |
|         psrlw      m2, 8  ; VxVx...
 | |
|         psrlw      m4, 8  ; VxVx...
 | |
|         packuswb   m2, m4 ; VVVV
 | |
|         movu   [vdstq + whalfq], m2
 | |
| 
 | |
|         add   whalfq, mmsize
 | |
|         add    wtwoq, mmsize * 4
 | |
|         add       wq, mmsize * 2
 | |
|         jl .loop_simd
 | |
| 
 | |
|     .end_line:
 | |
|         add        srcq, src_strideq
 | |
|         add        ydstq, lum_strideq
 | |
|         add        udstq, chrom_strideq
 | |
|         add        vdstq, chrom_strideq
 | |
| 
 | |
|         ;restore initial state of line variable
 | |
|         mov           wq, back_wq
 | |
|         mov          xq, wq
 | |
|         mov      whalfq, wq
 | |
|         shr      whalfq, 1     ; whalf = width / 2
 | |
|         sub          hd, 1
 | |
|         jg .loop_line
 | |
| 
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| %if ARCH_X86_64
 | |
| INIT_XMM sse2
 | |
| UYVY_TO_YUV422
 | |
| 
 | |
| INIT_XMM avx
 | |
| UYVY_TO_YUV422
 | |
| %endif
 |