diff options
| author | Tim Redfern <tim@eclectronics.org> | 2014-02-17 13:36:38 +0000 |
|---|---|---|
| committer | Tim Redfern <tim@eclectronics.org> | 2014-02-17 13:36:38 +0000 |
| commit | 22e28216336da876e1fd17f380ce42eaf1446769 (patch) | |
| tree | 444dad3dc7e2656992d29f34f7bce31970c122a5 /ffmpeg/libavcodec/x86/diracdsp_yasm.asm | |
| parent | ae5e8541f6e06e64c28719467cdf366ac57aff31 (diff) | |
chasing indexing error
Diffstat (limited to 'ffmpeg/libavcodec/x86/diracdsp_yasm.asm')
| -rw-r--r-- | ffmpeg/libavcodec/x86/diracdsp_yasm.asm | 264 |
1 files changed, 0 insertions, 264 deletions
diff --git a/ffmpeg/libavcodec/x86/diracdsp_yasm.asm b/ffmpeg/libavcodec/x86/diracdsp_yasm.asm deleted file mode 100644 index 3e9765b..0000000 --- a/ffmpeg/libavcodec/x86/diracdsp_yasm.asm +++ /dev/null @@ -1,264 +0,0 @@ -;****************************************************************************** -;* Copyright (c) 2010 David Conrad -;* -;* This file is part of FFmpeg. -;* -;* FFmpeg is free software; you can redistribute it and/or -;* modify it under the terms of the GNU Lesser General Public -;* License as published by the Free Software Foundation; either -;* version 2.1 of the License, or (at your option) any later version. -;* -;* FFmpeg is distributed in the hope that it will be useful, -;* but WITHOUT ANY WARRANTY; without even the implied warranty of -;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -;* Lesser General Public License for more details. -;* -;* You should have received a copy of the GNU Lesser General Public -;* License along with FFmpeg; if not, write to the Free Software -;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA -;****************************************************************************** - -%include "libavutil/x86/x86util.asm" - -SECTION_RODATA -pw_3: times 8 dw 3 -pw_7: times 8 dw 7 -pw_16: times 8 dw 16 -pw_32: times 8 dw 32 -pb_128: times 16 db 128 - -section .text - -%macro UNPACK_ADD 6 - mov%5 %1, %3 - mov%6 m5, %4 - mova m4, %1 - mova %2, m5 - punpcklbw %1, m7 - punpcklbw m5, m7 - punpckhbw m4, m7 - punpckhbw %2, m7 - paddw %1, m5 - paddw %2, m4 -%endmacro - -%macro HPEL_FILTER 1 -; dirac_hpel_filter_v_sse2(uint8_t *dst, uint8_t *src, int stride, int width); -cglobal dirac_hpel_filter_v_%1, 4,6,8, dst, src, stride, width, src0, stridex3 - mov src0q, srcq - lea stridex3q, [3*strideq] - sub src0q, stridex3q - pxor m7, m7 -.loop: - ; 7*(src[0] + src[1]) - UNPACK_ADD m0, m1, [srcq], [srcq + strideq], a,a - pmullw m0, [pw_7] - pmullw m1, [pw_7] - - ; 3*( ... + src[-2] + src[3]) - UNPACK_ADD m2, m3, [src0q + strideq], [srcq + stridex3q], a,a - paddw m0, m2 - paddw m1, m3 - pmullw m0, [pw_3] - pmullw m1, [pw_3] - - ; ... - 7*(src[-1] + src[2]) - UNPACK_ADD m2, m3, [src0q + strideq*2], [srcq + strideq*2], a,a - pmullw m2, [pw_7] - pmullw m3, [pw_7] - psubw m0, m2 - psubw m1, m3 - - ; ... - (src[-3] + src[4]) - UNPACK_ADD m2, m3, [src0q], [srcq + strideq*4], a,a - psubw m0, m2 - psubw m1, m3 - - paddw m0, [pw_16] - paddw m1, [pw_16] - psraw m0, 5 - psraw m1, 5 - packuswb m0, m1 - mova [dstq], m0 - add dstq, mmsize - add srcq, mmsize - add src0q, mmsize - sub widthd, mmsize - jg .loop - RET - -; dirac_hpel_filter_h_sse2(uint8_t *dst, uint8_t *src, int width); -cglobal dirac_hpel_filter_h_%1, 3,3,8, dst, src, width - dec widthd - pxor m7, m7 - and widthd, ~(mmsize-1) -.loop: - ; 7*(src[0] + src[1]) - UNPACK_ADD m0, m1, [srcq + widthq], [srcq + widthq + 1], u,u - pmullw m0, [pw_7] - pmullw m1, [pw_7] - - ; 3*( ... + src[-2] + src[3]) - UNPACK_ADD m2, m3, [srcq + widthq - 2], [srcq + widthq + 3], u,u - paddw m0, m2 - paddw m1, m3 - pmullw m0, [pw_3] - pmullw m1, [pw_3] - - ; ... - 7*(src[-1] + src[2]) - UNPACK_ADD m2, m3, [srcq + widthq - 1], [srcq + widthq + 2], u,u - pmullw m2, [pw_7] - pmullw m3, [pw_7] - psubw m0, m2 - psubw m1, m3 - - ; ... - (src[-3] + src[4]) - UNPACK_ADD m2, m3, [srcq + widthq - 3], [srcq + widthq + 4], u,u - psubw m0, m2 - psubw m1, m3 - - paddw m0, [pw_16] - paddw m1, [pw_16] - psraw m0, 5 - psraw m1, 5 - packuswb m0, m1 - mova [dstq + widthq], m0 - sub widthd, mmsize - jge .loop - RET -%endmacro - -%macro PUT_RECT 1 -; void put_rect_clamped(uint8_t *dst, int dst_stride, int16_t *src, int src_stride, int width, int height) -cglobal put_signed_rect_clamped_%1, 5,9,3, dst, dst_stride, src, src_stride, w, dst2, src2 - mova m0, [pb_128] - add wd, (mmsize-1) - and wd, ~(mmsize-1) - -%if ARCH_X86_64 - movsxd dst_strideq, dst_strided - movsxd src_strideq, src_strided - mov r7d, r5m - mov r8d, wd - %define wspill r8d - %define hd r7d -%else - mov r4m, wd - %define wspill r4m - %define hd r5mp -%endif - -.loopy - lea src2q, [srcq+src_strideq*2] - lea dst2q, [dstq+dst_strideq] -.loopx: - sub wd, mmsize - mova m1, [srcq +2*wq] - mova m2, [src2q+2*wq] - packsswb m1, [srcq +2*wq+mmsize] - packsswb m2, [src2q+2*wq+mmsize] - paddb m1, m0 - paddb m2, m0 - mova [dstq +wq], m1 - mova [dst2q+wq], m2 - jg .loopx - - lea srcq, [srcq+src_strideq*4] - lea dstq, [dstq+dst_strideq*2] - sub hd, 2 - mov wd, wspill - jg .loopy - RET -%endm - -%macro ADD_RECT 1 -; void add_rect_clamped(uint8_t *dst, uint16_t *src, int stride, int16_t *idwt, int idwt_stride, int width, int height) -cglobal add_rect_clamped_%1, 7,9,3, dst, src, stride, idwt, idwt_stride, w, h - mova m0, [pw_32] - add wd, (mmsize-1) - and wd, ~(mmsize-1) - -%if ARCH_X86_64 - movsxd strideq, strided - movsxd idwt_strideq, idwt_strided - mov r8d, wd - %define wspill r8d -%else - mov r5m, wd - %define wspill r5m -%endif - -.loop: - sub wd, mmsize - movu m1, [srcq +2*wq] ; FIXME: ensure alignment - paddw m1, m0 - psraw m1, 6 - movu m2, [srcq +2*wq+mmsize] ; FIXME: ensure alignment - paddw m2, m0 - psraw m2, 6 - paddw m1, [idwtq+2*wq] - paddw m2, [idwtq+2*wq+mmsize] - packuswb m1, m2 - mova [dstq +wq], m1 - jg .loop - - lea srcq, [srcq + 2*strideq] - add dstq, strideq - lea idwtq, [idwtq+ 2*idwt_strideq] - sub hd, 1 - mov wd, wspill - jg .loop - RET -%endm - -%macro ADD_OBMC 2 -; void add_obmc(uint16_t *dst, uint8_t *src, int stride, uint8_t *obmc_weight, int yblen) -cglobal add_dirac_obmc%1_%2, 6,6,5, dst, src, stride, obmc, yblen - pxor m4, m4 -.loop: -%assign i 0 -%rep %1 / mmsize - mova m0, [srcq+i] - mova m1, m0 - punpcklbw m0, m4 - punpckhbw m1, m4 - mova m2, [obmcq+i] - mova m3, m2 - punpcklbw m2, m4 - punpckhbw m3, m4 - pmullw m0, m2 - pmullw m1, m3 - movu m2, [dstq+2*i] - movu m3, [dstq+2*i+mmsize] - paddw m0, m2 - paddw m1, m3 - movu [dstq+2*i], m0 - movu [dstq+2*i+mmsize], m1 -%assign i i+mmsize -%endrep - lea srcq, [srcq+strideq] - lea dstq, [dstq+2*strideq] - add obmcq, 32 - sub yblend, 1 - jg .loop - RET -%endm - -INIT_MMX -%if ARCH_X86_64 == 0 -PUT_RECT mmx -ADD_RECT mmx - -HPEL_FILTER mmx -ADD_OBMC 32, mmx -ADD_OBMC 16, mmx -%endif -ADD_OBMC 8, mmx - -INIT_XMM -PUT_RECT sse2 -ADD_RECT sse2 - -HPEL_FILTER sse2 -ADD_OBMC 32, sse2 -ADD_OBMC 16, sse2 |
