vc1dsp.asm 7.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320
  1. ;******************************************************************************
  2. ;* VC1 deblocking optimizations
  3. ;* Copyright (c) 2009 David Conrad
  4. ;*
  5. ;* This file is part of FFmpeg.
  6. ;*
  7. ;* FFmpeg is free software; you can redistribute it and/or
  8. ;* modify it under the terms of the GNU Lesser General Public
  9. ;* License as published by the Free Software Foundation; either
  10. ;* version 2.1 of the License, or (at your option) any later version.
  11. ;*
  12. ;* FFmpeg is distributed in the hope that it will be useful,
  13. ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. ;* Lesser General Public License for more details.
  16. ;*
  17. ;* You should have received a copy of the GNU Lesser General Public
  18. ;* License along with FFmpeg; if not, write to the Free Software
  19. ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. ;******************************************************************************
  21. %include "libavutil/x86/x86inc.asm"
  22. %include "libavutil/x86/x86util.asm"
  23. cextern pw_4
  24. cextern pw_5
  25. section .text
  26. ; dst_low, dst_high (src), zero
  27. ; zero-extends one vector from 8 to 16 bits
  28. %macro UNPACK_8TO16 4
  29. mova m%2, m%3
  30. punpckh%1 m%3, m%4
  31. punpckl%1 m%2, m%4
  32. %endmacro
  33. %macro STORE_4_WORDS_MMX 6
  34. movd %6d, %5
  35. %if mmsize==16
  36. psrldq %5, 4
  37. %else
  38. psrlq %5, 32
  39. %endif
  40. mov %1, %6w
  41. shr %6, 16
  42. mov %2, %6w
  43. movd %6d, %5
  44. mov %3, %6w
  45. shr %6, 16
  46. mov %4, %6w
  47. %endmacro
  48. %macro STORE_4_WORDS_SSE4 6
  49. pextrw %1, %5, %6+0
  50. pextrw %2, %5, %6+1
  51. pextrw %3, %5, %6+2
  52. pextrw %4, %5, %6+3
  53. %endmacro
  54. ; in: p1 p0 q0 q1, clobbers p0
  55. ; out: p1 = (2*(p1 - q1) - 5*(p0 - q0) + 4) >> 3
  56. %macro VC1_LOOP_FILTER_A0 4
  57. psubw %1, %4
  58. psubw %2, %3
  59. paddw %1, %1
  60. pmullw %2, [pw_5]
  61. psubw %1, %2
  62. paddw %1, [pw_4]
  63. psraw %1, 3
  64. %endmacro
  65. ; in: p0 q0 a0 a1 a2
  66. ; m0 m1 m7 m6 m5
  67. ; %1: size
  68. ; out: m0=p0' m1=q0'
  69. %macro VC1_FILTER 1
  70. PABSW m4, m7
  71. PABSW m3, m6
  72. PABSW m2, m5
  73. mova m6, m4
  74. pminsw m3, m2
  75. pcmpgtw m6, m3 ; if (a2 < a0 || a1 < a0)
  76. psubw m3, m4
  77. pmullw m3, [pw_5] ; 5*(a3 - a0)
  78. PABSW m2, m3
  79. psraw m2, 3 ; abs(d/8)
  80. pxor m7, m3 ; d_sign ^= a0_sign
  81. pxor m5, m5
  82. movd m3, r2d
  83. %if %1 > 4
  84. punpcklbw m3, m3
  85. %endif
  86. punpcklbw m3, m5
  87. pcmpgtw m3, m4 ; if (a0 < pq)
  88. pand m6, m3
  89. mova m3, m0
  90. psubw m3, m1
  91. PABSW m4, m3
  92. psraw m4, 1
  93. pxor m3, m7 ; d_sign ^ clip_sign
  94. psraw m3, 15
  95. pminsw m2, m4 ; min(d, clip)
  96. pcmpgtw m4, m5
  97. pand m6, m4 ; filt3 (C return value)
  98. ; each set of 4 pixels is not filtered if the 3rd is not
  99. %if mmsize==16
  100. pshuflw m4, m6, 0xaa
  101. %if %1 > 4
  102. pshufhw m4, m4, 0xaa
  103. %endif
  104. %else
  105. pshufw m4, m6, 0xaa
  106. %endif
  107. pandn m3, m4
  108. pand m2, m6
  109. pand m3, m2 ; d final
  110. psraw m7, 15
  111. pxor m3, m7
  112. psubw m3, m7
  113. psubw m0, m3
  114. paddw m1, m3
  115. packuswb m0, m0
  116. packuswb m1, m1
  117. %endmacro
  118. ; 1st param: size of filter
  119. ; 2nd param: mov suffix equivalent to the filter size
  120. %macro VC1_V_LOOP_FILTER 2
  121. pxor m5, m5
  122. mov%2 m6, [r4]
  123. mov%2 m4, [r4+r1]
  124. mov%2 m7, [r4+2*r1]
  125. mov%2 m0, [r4+r3]
  126. punpcklbw m6, m5
  127. punpcklbw m4, m5
  128. punpcklbw m7, m5
  129. punpcklbw m0, m5
  130. VC1_LOOP_FILTER_A0 m6, m4, m7, m0
  131. mov%2 m1, [r0]
  132. mov%2 m2, [r0+r1]
  133. punpcklbw m1, m5
  134. punpcklbw m2, m5
  135. mova m4, m0
  136. VC1_LOOP_FILTER_A0 m7, m4, m1, m2
  137. mov%2 m3, [r0+2*r1]
  138. mov%2 m4, [r0+r3]
  139. punpcklbw m3, m5
  140. punpcklbw m4, m5
  141. mova m5, m1
  142. VC1_LOOP_FILTER_A0 m5, m2, m3, m4
  143. VC1_FILTER %1
  144. mov%2 [r4+r3], m0
  145. mov%2 [r0], m1
  146. %endmacro
  147. ; 1st param: size of filter
  148. ; NOTE: UNPACK_8TO16 this number of 8 bit numbers are in half a register
  149. ; 2nd (optional) param: temp register to use for storing words
  150. %macro VC1_H_LOOP_FILTER 1-2
  151. %if %1 == 4
  152. movq m0, [r0 -4]
  153. movq m1, [r0+ r1-4]
  154. movq m2, [r0+2*r1-4]
  155. movq m3, [r0+ r3-4]
  156. TRANSPOSE4x4B 0, 1, 2, 3, 4
  157. %else
  158. movq m0, [r0 -4]
  159. movq m4, [r0+ r1-4]
  160. movq m1, [r0+2*r1-4]
  161. movq m5, [r0+ r3-4]
  162. movq m2, [r4 -4]
  163. movq m6, [r4+ r1-4]
  164. movq m3, [r4+2*r1-4]
  165. movq m7, [r4+ r3-4]
  166. punpcklbw m0, m4
  167. punpcklbw m1, m5
  168. punpcklbw m2, m6
  169. punpcklbw m3, m7
  170. TRANSPOSE4x4W 0, 1, 2, 3, 4
  171. %endif
  172. pxor m5, m5
  173. UNPACK_8TO16 bw, 6, 0, 5
  174. UNPACK_8TO16 bw, 7, 1, 5
  175. VC1_LOOP_FILTER_A0 m6, m0, m7, m1
  176. UNPACK_8TO16 bw, 4, 2, 5
  177. mova m0, m1 ; m0 = p0
  178. VC1_LOOP_FILTER_A0 m7, m1, m4, m2
  179. UNPACK_8TO16 bw, 1, 3, 5
  180. mova m5, m4
  181. VC1_LOOP_FILTER_A0 m5, m2, m1, m3
  182. SWAP 1, 4 ; m1 = q0
  183. VC1_FILTER %1
  184. punpcklbw m0, m1
  185. %if %0 > 1
  186. STORE_4_WORDS_MMX [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, %2
  187. %if %1 > 4
  188. psrldq m0, 4
  189. STORE_4_WORDS_MMX [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, %2
  190. %endif
  191. %else
  192. STORE_4_WORDS_SSE4 [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, 0
  193. STORE_4_WORDS_SSE4 [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, 4
  194. %endif
  195. %endmacro
  196. %macro START_V_FILTER 0
  197. mov r4, r0
  198. lea r3, [4*r1]
  199. sub r4, r3
  200. lea r3, [r1+2*r1]
  201. imul r2, 0x01010101
  202. %endmacro
  203. %macro START_H_FILTER 1
  204. lea r3, [r1+2*r1]
  205. %if %1 > 4
  206. lea r4, [r0+4*r1]
  207. %endif
  208. imul r2, 0x01010101
  209. %endmacro
  210. %macro VC1_LF_MMX 1
  211. INIT_MMX
  212. cglobal vc1_v_loop_filter_internal_%1
  213. VC1_V_LOOP_FILTER 4, d
  214. ret
  215. cglobal vc1_h_loop_filter_internal_%1
  216. VC1_H_LOOP_FILTER 4, r4
  217. ret
  218. ; void ff_vc1_v_loop_filter4_mmx2(uint8_t *src, int stride, int pq)
  219. cglobal vc1_v_loop_filter4_%1, 3,5,0
  220. START_V_FILTER
  221. call vc1_v_loop_filter_internal_%1
  222. RET
  223. ; void ff_vc1_h_loop_filter4_mmx2(uint8_t *src, int stride, int pq)
  224. cglobal vc1_h_loop_filter4_%1, 3,5,0
  225. START_H_FILTER 4
  226. call vc1_h_loop_filter_internal_%1
  227. RET
  228. ; void ff_vc1_v_loop_filter8_mmx2(uint8_t *src, int stride, int pq)
  229. cglobal vc1_v_loop_filter8_%1, 3,5,0
  230. START_V_FILTER
  231. call vc1_v_loop_filter_internal_%1
  232. add r4, 4
  233. add r0, 4
  234. call vc1_v_loop_filter_internal_%1
  235. RET
  236. ; void ff_vc1_h_loop_filter8_mmx2(uint8_t *src, int stride, int pq)
  237. cglobal vc1_h_loop_filter8_%1, 3,5,0
  238. START_H_FILTER 4
  239. call vc1_h_loop_filter_internal_%1
  240. lea r0, [r0+4*r1]
  241. call vc1_h_loop_filter_internal_%1
  242. RET
  243. %endmacro
  244. %define PABSW PABSW_MMX2
  245. VC1_LF_MMX mmx2
  246. INIT_XMM
  247. ; void ff_vc1_v_loop_filter8_sse2(uint8_t *src, int stride, int pq)
  248. cglobal vc1_v_loop_filter8_sse2, 3,5,8
  249. START_V_FILTER
  250. VC1_V_LOOP_FILTER 8, q
  251. RET
  252. ; void ff_vc1_h_loop_filter8_sse2(uint8_t *src, int stride, int pq)
  253. cglobal vc1_h_loop_filter8_sse2, 3,6,8
  254. START_H_FILTER 8
  255. VC1_H_LOOP_FILTER 8, r5
  256. RET
  257. %define PABSW PABSW_SSSE3
  258. INIT_MMX
  259. ; void ff_vc1_v_loop_filter4_ssse3(uint8_t *src, int stride, int pq)
  260. cglobal vc1_v_loop_filter4_ssse3, 3,5,0
  261. START_V_FILTER
  262. VC1_V_LOOP_FILTER 4, d
  263. RET
  264. ; void ff_vc1_h_loop_filter4_ssse3(uint8_t *src, int stride, int pq)
  265. cglobal vc1_h_loop_filter4_ssse3, 3,5,0
  266. START_H_FILTER 4
  267. VC1_H_LOOP_FILTER 4, r4
  268. RET
  269. INIT_XMM
  270. ; void ff_vc1_v_loop_filter8_ssse3(uint8_t *src, int stride, int pq)
  271. cglobal vc1_v_loop_filter8_ssse3, 3,5,8
  272. START_V_FILTER
  273. VC1_V_LOOP_FILTER 8, q
  274. RET
  275. ; void ff_vc1_h_loop_filter8_ssse3(uint8_t *src, int stride, int pq)
  276. cglobal vc1_h_loop_filter8_ssse3, 3,6,8
  277. START_H_FILTER 8
  278. VC1_H_LOOP_FILTER 8, r5
  279. RET
  280. ; void ff_vc1_h_loop_filter8_sse4(uint8_t *src, int stride, int pq)
  281. cglobal vc1_h_loop_filter8_sse4, 3,5,8
  282. START_H_FILTER 8
  283. VC1_H_LOOP_FILTER 8
  284. RET