jcgryext-avx2.asm 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438
  1. ;
  2. ; jcgryext.asm - grayscale colorspace conversion (64-bit AVX2)
  3. ;
  4. ; Copyright (C) 2011, 2016, D. R. Commander.
  5. ; Copyright (C) 2015, Intel Corporation.
  6. ; Copyright (C) 2018, Matthias Räncker.
  7. ;
  8. ; Based on the x86 SIMD extension for IJG JPEG library
  9. ; Copyright (C) 1999-2006, MIYASAKA Masaru.
  10. ; For conditions of distribution and use, see copyright notice in jsimdext.inc
  11. ;
  12. ; This file should be assembled with NASM (Netwide Assembler),
  13. ; can *not* be assembled with Microsoft's MASM or any compatible
  14. ; assembler (including Borland's Turbo Assembler).
  15. ; NASM is available from http://nasm.sourceforge.net/ or
  16. ; http://sourceforge.net/project/showfiles.php?group_id=6208
  17. %include "jcolsamp.inc"
  18. ; --------------------------------------------------------------------------
  19. ;
  20. ; Convert some rows of samples to the output colorspace.
  21. ;
  22. ; GLOBAL(void)
  23. ; jsimd_rgb_gray_convert_avx2(JDIMENSION img_width, JSAMPARRAY input_buf,
  24. ; JSAMPIMAGE output_buf, JDIMENSION output_row,
  25. ; int num_rows);
  26. ;
  27. ; r10d = JDIMENSION img_width
  28. ; r11 = JSAMPARRAY input_buf
  29. ; r12 = JSAMPIMAGE output_buf
  30. ; r13d = JDIMENSION output_row
  31. ; r14d = int num_rows
  32. %define wk(i) rbp - (WK_NUM - (i)) * SIZEOF_YMMWORD ; ymmword wk[WK_NUM]
  33. %define WK_NUM 2
  34. align 32
  35. GLOBAL_FUNCTION(jsimd_rgb_gray_convert_avx2)
  36. EXTN(jsimd_rgb_gray_convert_avx2):
  37. push rbp
  38. mov rax, rsp ; rax = original rbp
  39. sub rsp, byte 4
  40. and rsp, byte (-SIZEOF_YMMWORD) ; align to 256 bits
  41. mov [rsp], rax
  42. mov rbp, rsp ; rbp = aligned rbp
  43. lea rsp, [wk(0)]
  44. collect_args 5
  45. push rbx
  46. mov ecx, r10d
  47. test rcx, rcx
  48. jz near .return
  49. push rcx
  50. mov rsi, r12
  51. mov ecx, r13d
  52. mov rdip, JSAMPARRAY [rsi+0*SIZEOF_JSAMPARRAY]
  53. lea rdi, [rdi+rcx*SIZEOF_JSAMPROW]
  54. pop rcx
  55. mov rsi, r11
  56. mov eax, r14d
  57. test rax, rax
  58. jle near .return
  59. .rowloop:
  60. push rdi
  61. push rsi
  62. push rcx ; col
  63. mov rsip, JSAMPROW [rsi] ; inptr
  64. mov rdip, JSAMPROW [rdi] ; outptr0
  65. cmp rcx, byte SIZEOF_YMMWORD
  66. jae near .columnloop
  67. %if RGB_PIXELSIZE == 3 ; ---------------
  68. .column_ld1:
  69. push rax
  70. push rdx
  71. lea rcx, [rcx+rcx*2] ; imul ecx,RGB_PIXELSIZE
  72. test cl, SIZEOF_BYTE
  73. jz short .column_ld2
  74. sub rcx, byte SIZEOF_BYTE
  75. movzx rax, byte [rsi+rcx]
  76. .column_ld2:
  77. test cl, SIZEOF_WORD
  78. jz short .column_ld4
  79. sub rcx, byte SIZEOF_WORD
  80. movzx rdx, word [rsi+rcx]
  81. shl rax, WORD_BIT
  82. or rax, rdx
  83. .column_ld4:
  84. vmovd xmmA, eax
  85. pop rdx
  86. pop rax
  87. test cl, SIZEOF_DWORD
  88. jz short .column_ld8
  89. sub rcx, byte SIZEOF_DWORD
  90. vmovd xmmF, XMM_DWORD [rsi+rcx]
  91. vpslldq xmmA, xmmA, SIZEOF_DWORD
  92. vpor xmmA, xmmA, xmmF
  93. .column_ld8:
  94. test cl, SIZEOF_MMWORD
  95. jz short .column_ld16
  96. sub rcx, byte SIZEOF_MMWORD
  97. vmovq xmmB, XMM_MMWORD [rsi+rcx]
  98. vpslldq xmmA, xmmA, SIZEOF_MMWORD
  99. vpor xmmA, xmmA, xmmB
  100. .column_ld16:
  101. test cl, SIZEOF_XMMWORD
  102. jz short .column_ld32
  103. sub rcx, byte SIZEOF_XMMWORD
  104. vmovdqu xmmB, XMM_MMWORD [rsi+rcx]
  105. vperm2i128 ymmA, ymmA, ymmA, 1
  106. vpor ymmA, ymmB
  107. .column_ld32:
  108. test cl, SIZEOF_YMMWORD
  109. jz short .column_ld64
  110. sub rcx, byte SIZEOF_YMMWORD
  111. vmovdqa ymmF, ymmA
  112. vmovdqu ymmA, YMMWORD [rsi+0*SIZEOF_YMMWORD]
  113. .column_ld64:
  114. test cl, 2*SIZEOF_YMMWORD
  115. mov rcx, SIZEOF_YMMWORD
  116. jz short .rgb_gray_cnv
  117. vmovdqa ymmB, ymmA
  118. vmovdqu ymmA, YMMWORD [rsi+0*SIZEOF_YMMWORD]
  119. vmovdqu ymmF, YMMWORD [rsi+1*SIZEOF_YMMWORD]
  120. jmp short .rgb_gray_cnv
  121. .columnloop:
  122. vmovdqu ymmA, YMMWORD [rsi+0*SIZEOF_YMMWORD]
  123. vmovdqu ymmF, YMMWORD [rsi+1*SIZEOF_YMMWORD]
  124. vmovdqu ymmB, YMMWORD [rsi+2*SIZEOF_YMMWORD]
  125. .rgb_gray_cnv:
  126. ; ymmA=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
  127. ; 15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A)
  128. ; ymmF=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
  129. ; 0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
  130. ; ymmB=(1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q
  131. ; 2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
  132. vmovdqu ymmC, ymmA
  133. vinserti128 ymmA, ymmF, xmmA, 0 ; ymmA=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
  134. ; 0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
  135. vinserti128 ymmC, ymmC, xmmB, 0 ; ymmC=(1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q
  136. ; 15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A)
  137. vinserti128 ymmB, ymmB, xmmF, 0 ; ymmB=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
  138. ; 2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
  139. vperm2i128 ymmF, ymmC, ymmC, 1 ; ymmF=(15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A
  140. ; 1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q)
  141. vmovdqa ymmG, ymmA
  142. vpslldq ymmA, ymmA, 8 ; ymmA=(-- -- -- -- -- -- -- -- 00 10 20 01 11 21 02 12
  143. ; 22 03 13 23 04 14 24 05 0G 1G 2G 0H 1H 2H 0I 1I)
  144. vpsrldq ymmG, ymmG, 8 ; ymmG=(22 03 13 23 04 14 24 05 0G 1G 2G 0H 1H 2H 0I 1I
  145. ; 2I 0J 1J 2J 0K 1K 2K 0L -- -- -- -- -- -- -- --)
  146. vpunpckhbw ymmA, ymmA, ymmF ; ymmA=(00 08 10 18 20 28 01 09 11 19 21 29 02 0A 12 1A
  147. ; 0G 0O 1G 1O 2G 2O 0H 0P 1H 1P 2H 2P 0I 0Q 1I 1Q)
  148. vpslldq ymmF, ymmF, 8 ; ymmF=(-- -- -- -- -- -- -- -- 15 25 06 16 26 07 17 27
  149. ; 08 18 28 09 19 29 0A 1A 1L 2L 0M 1M 2M 0N 1N 2N)
  150. vpunpcklbw ymmG, ymmG, ymmB ; ymmG=(22 2A 03 0B 13 1B 23 2B 04 0C 14 1C 24 2C 05 0D
  151. ; 2I 2Q 0J 0R 1J 1R 2J 2R 0K 0S 1K 1S 2K 2S 0L 0T)
  152. vpunpckhbw ymmF, ymmF, ymmB ; ymmF=(15 1D 25 2D 06 0E 16 1E 26 2E 07 0F 17 1F 27 2F
  153. ; 1L 1T 2L 2T 0M 0U 1M 1U 2M 2U 0N 0V 1N 1V 2N 2V)
  154. vmovdqa ymmD, ymmA
  155. vpslldq ymmA, ymmA, 8 ; ymmA=(-- -- -- -- -- -- -- -- 00 08 10 18 20 28 01 09
  156. ; 11 19 21 29 02 0A 12 1A 0G 0O 1G 1O 2G 2O 0H 0P)
  157. vpsrldq ymmD, ymmD, 8 ; ymmD=(11 19 21 29 02 0A 12 1A 0G 0O 1G 1O 2G 2O 0H 0P
  158. ; 1H 1P 2H 2P 0I 0Q 1I 1Q -- -- -- -- -- -- -- --)
  159. vpunpckhbw ymmA, ymmA, ymmG ; ymmA=(00 04 08 0C 10 14 18 1C 20 24 28 2C 01 05 09 0D
  160. ; 0G 0K 0O 0S 1G 1K 1O 1S 2G 2K 2O 2S 0H 0L 0P 0T)
  161. vpslldq ymmG, ymmG, 8 ; ymmG=(-- -- -- -- -- -- -- -- 22 2A 03 0B 13 1B 23 2B
  162. ; 04 0C 14 1C 24 2C 05 0D 2I 2Q 0J 0R 1J 1R 2J 2R)
  163. vpunpcklbw ymmD, ymmD, ymmF ; ymmD=(11 15 19 1D 21 25 29 2D 02 06 0A 0E 12 16 1A 1E
  164. ; 1H 1L 1P 1T 2H 2L 2P 2T 0I 0M 0Q 0U 1I 1M 1Q 1U)
  165. vpunpckhbw ymmG, ymmG, ymmF ; ymmG=(22 26 2A 2E 03 07 0B 0F 13 17 1B 1F 23 27 2B 2F
  166. ; 2I 2M 2Q 2U 0J 0N 0R 0V 1J 1N 1R 1V 2J 2N 2R 2V)
  167. vmovdqa ymmE, ymmA
  168. vpslldq ymmA, ymmA, 8 ; ymmA=(-- -- -- -- -- -- -- -- 00 04 08 0C 10 14 18 1C
  169. ; 20 24 28 2C 01 05 09 0D 0G 0K 0O 0S 1G 1K 1O 1S)
  170. vpsrldq ymmE, ymmE, 8 ; ymmE=(20 24 28 2C 01 05 09 0D 0G 0K 0O 0S 1G 1K 1O 1S
  171. ; 2G 2K 2O 2S 0H 0L 0P 0T -- -- -- -- -- -- -- --)
  172. vpunpckhbw ymmA, ymmA, ymmD ; ymmA=(00 02 04 06 08 0A 0C 0E 10 12 14 16 18 1A 1C 1E
  173. ; 0G 0I 0K 0M 0O 0Q 0S 0U 1G 1I 1K 1M 1O 1Q 1S 1U)
  174. vpslldq ymmD, ymmD, 8 ; ymmD=(-- -- -- -- -- -- -- -- 11 15 19 1D 21 25 29 2D
  175. ; 02 06 0A 0E 12 16 1A 1E 1H 1L 1P 1T 2H 2L 2P 2T)
  176. vpunpcklbw ymmE, ymmE, ymmG ; ymmE=(20 22 24 26 28 2A 2C 2E 01 03 05 07 09 0B 0D 0F
  177. ; 2G 2I 2K 2M 2O 2Q 2S 2U 0H 0J 0L 0N 0P 0R 0T 0V)
  178. vpunpckhbw ymmD, ymmD, ymmG ; ymmD=(11 13 15 17 19 1B 1D 1F 21 23 25 27 29 2B 2D 2F
  179. ; 1H 1J 1L 1N 1P 1R 1T 1V 2H 2J 2L 2N 2P 2R 2T 2V)
  180. vpxor ymmH, ymmH, ymmH
  181. vmovdqa ymmC, ymmA
  182. vpunpcklbw ymmA, ymmA, ymmH ; ymmA=(00 02 04 06 08 0A 0C 0E 0G 0I 0K 0M 0O 0Q 0S 0U)
  183. vpunpckhbw ymmC, ymmC, ymmH ; ymmC=(10 12 14 16 18 1A 1C 1E 1G 1I 1K 1M 1O 1Q 1S 1U)
  184. vmovdqa ymmB, ymmE
  185. vpunpcklbw ymmE, ymmE, ymmH ; ymmE=(20 22 24 26 28 2A 2C 2E 2G 2I 2K 2M 2O 2Q 2S 2U)
  186. vpunpckhbw ymmB, ymmB, ymmH ; ymmB=(01 03 05 07 09 0B 0D 0F 0H 0J 0L 0N 0P 0R 0T 0V)
  187. vmovdqa ymmF, ymmD
  188. vpunpcklbw ymmD, ymmD, ymmH ; ymmD=(11 13 15 17 19 1B 1D 1F 1H 1J 1L 1N 1P 1R 1T 1V)
  189. vpunpckhbw ymmF, ymmF, ymmH ; ymmF=(21 23 25 27 29 2B 2D 2F 2H 2J 2L 2N 2P 2R 2T 2V)
  190. %else ; RGB_PIXELSIZE == 4 ; -----------
  191. .column_ld1:
  192. test cl, SIZEOF_XMMWORD/16
  193. jz short .column_ld2
  194. sub rcx, byte SIZEOF_XMMWORD/16
  195. vmovd xmmA, XMM_DWORD [rsi+rcx*RGB_PIXELSIZE]
  196. .column_ld2:
  197. test cl, SIZEOF_XMMWORD/8
  198. jz short .column_ld4
  199. sub rcx, byte SIZEOF_XMMWORD/8
  200. vmovq xmmF, XMM_MMWORD [rsi+rcx*RGB_PIXELSIZE]
  201. vpslldq xmmA, xmmA, SIZEOF_MMWORD
  202. vpor xmmA, xmmA, xmmF
  203. .column_ld4:
  204. test cl, SIZEOF_XMMWORD/4
  205. jz short .column_ld8
  206. sub rcx, byte SIZEOF_XMMWORD/4
  207. vmovdqa xmmF, xmmA
  208. vperm2i128 ymmF, ymmF, ymmF, 1
  209. vmovdqu xmmA, XMMWORD [rsi+rcx*RGB_PIXELSIZE]
  210. vpor ymmA, ymmA, ymmF
  211. .column_ld8:
  212. test cl, SIZEOF_XMMWORD/2
  213. jz short .column_ld16
  214. sub rcx, byte SIZEOF_XMMWORD/2
  215. vmovdqa ymmF, ymmA
  216. vmovdqu ymmA, YMMWORD [rsi+rcx*RGB_PIXELSIZE]
  217. .column_ld16:
  218. test cl, SIZEOF_XMMWORD
  219. mov rcx, SIZEOF_YMMWORD
  220. jz short .rgb_gray_cnv
  221. vmovdqa ymmE, ymmA
  222. vmovdqa ymmH, ymmF
  223. vmovdqu ymmA, YMMWORD [rsi+0*SIZEOF_YMMWORD]
  224. vmovdqu ymmF, YMMWORD [rsi+1*SIZEOF_YMMWORD]
  225. jmp short .rgb_gray_cnv
  226. .columnloop:
  227. vmovdqu ymmA, YMMWORD [rsi+0*SIZEOF_YMMWORD]
  228. vmovdqu ymmF, YMMWORD [rsi+1*SIZEOF_YMMWORD]
  229. vmovdqu ymmE, YMMWORD [rsi+2*SIZEOF_YMMWORD]
  230. vmovdqu ymmH, YMMWORD [rsi+3*SIZEOF_YMMWORD]
  231. .rgb_gray_cnv:
  232. ; ymmA=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
  233. ; 04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37)
  234. ; ymmF=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
  235. ; 0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F)
  236. ; ymmE=(0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J
  237. ; 0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
  238. ; ymmH=(0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R
  239. ; 0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
  240. vmovdqa ymmB, ymmA
  241. vinserti128 ymmA, ymmA, xmmE, 1 ; ymmA=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
  242. ; 0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J)
  243. vperm2i128 ymmE, ymmB, ymmE, 0x31 ; ymmE=(04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37
  244. ; 0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
  245. vmovdqa ymmB, ymmF
  246. vinserti128 ymmF, ymmF, xmmH, 1 ; ymmF=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
  247. ; 0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R)
  248. vperm2i128 ymmH, ymmB, ymmH, 0x31 ; ymmH=(0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F
  249. ; 0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
  250. vmovdqa ymmD, ymmA
  251. vpunpcklbw ymmA, ymmA, ymmE ; ymmA=(00 04 10 14 20 24 30 34 01 05 11 15 21 25 31 35
  252. ; 0G 0K 1G 1K 2G 2K 3G 3K 0H 0L 1H 1L 2H 2L 3H 3L)
  253. vpunpckhbw ymmD, ymmD, ymmE ; ymmD=(02 06 12 16 22 26 32 36 03 07 13 17 23 27 33 37
  254. ; 0I 0M 1I 1M 2I 2M 3I 3M 0J 0N 1J 1N 2J 2N 3J 3N)
  255. vmovdqa ymmC, ymmF
  256. vpunpcklbw ymmF, ymmF, ymmH ; ymmF=(08 0C 18 1C 28 2C 38 3C 09 0D 19 1D 29 2D 39 3D
  257. ; 0O 0S 1O 1S 2O 2S 3O 3S 0P 0T 1P 1T 2P 2T 3P 3T)
  258. vpunpckhbw ymmC, ymmC, ymmH ; ymmC=(0A 0E 1A 1E 2A 2E 3A 3E 0B 0F 1B 1F 2B 2F 3B 3F
  259. ; 0Q 0U 1Q 1U 2Q 2U 3Q 3U 0R 0V 1R 1V 2R 2V 3R 3V)
  260. vmovdqa ymmB, ymmA
  261. vpunpcklwd ymmA, ymmA, ymmF ; ymmA=(00 04 08 0C 10 14 18 1C 20 24 28 2C 30 34 38 3C
  262. ; 0G 0K 0O 0S 1G 1K 1O 1S 2G 2K 2O 2S 3G 3K 3O 3S)
  263. vpunpckhwd ymmB, ymmB, ymmF ; ymmB=(01 05 09 0D 11 15 19 1D 21 25 29 2D 31 35 39 3D
  264. ; 0H 0L 0P 0T 1H 1L 1P 1T 2H 2L 2P 2T 3H 3L 3P 3T)
  265. vmovdqa ymmG, ymmD
  266. vpunpcklwd ymmD, ymmD, ymmC ; ymmD=(02 06 0A 0E 12 16 1A 1E 22 26 2A 2E 32 36 3A 3E
  267. ; 0I 0M 0Q 0U 1I 1M 1Q 1U 2I 2M 2Q 2U 3I 3M 3Q 3U)
  268. vpunpckhwd ymmG, ymmG, ymmC ; ymmG=(03 07 0B 0F 13 17 1B 1F 23 27 2B 2F 33 37 3B 3F
  269. ; 0J 0N 0R 0V 1J 1N 1R 1V 2J 2N 2R 2V 3J 3N 3R 3V)
  270. vmovdqa ymmE, ymmA
  271. vpunpcklbw ymmA, ymmA, ymmD ; ymmA=(00 02 04 06 08 0A 0C 0E 10 12 14 16 18 1A 1C 1E
  272. ; 0G 0I 0K 0M 0O 0Q 0S 0U 1G 1I 1K 1M 1O 1Q 1S 1U)
  273. vpunpckhbw ymmE, ymmE, ymmD ; ymmE=(20 22 24 26 28 2A 2C 2E 30 32 34 36 38 3A 3C 3E
  274. ; 2G 2I 2K 2M 2O 2Q 2S 2U 3G 3I 3K 3M 3O 3Q 3S 3U)
  275. vmovdqa ymmH, ymmB
  276. vpunpcklbw ymmB, ymmB, ymmG ; ymmB=(01 03 05 07 09 0B 0D 0F 11 13 15 17 19 1B 1D 1F
  277. ; 0H 0J 0L 0N 0P 0R 0T 0V 1H 1J 1L 1N 1P 1R 1T 1V)
  278. vpunpckhbw ymmH, ymmH, ymmG ; ymmH=(21 23 25 27 29 2B 2D 2F 31 33 35 37 39 3B 3D 3F
  279. ; 2H 2J 2L 2N 2P 2R 2T 2V 3H 3J 3L 3N 3P 3R 3T 3V)
  280. vpxor ymmF, ymmF, ymmF
  281. vmovdqa ymmC, ymmA
  282. vpunpcklbw ymmA, ymmA, ymmF ; ymmA=(00 02 04 06 08 0A 0C 0E 0G 0I 0K 0M 0O 0Q 0S 0U)
  283. vpunpckhbw ymmC, ymmC, ymmF ; ymmC=(10 12 14 16 18 1A 1C 1E 1G 1I 1K 1M 1O 1Q 1S 1U)
  284. vmovdqa ymmD, ymmB
  285. vpunpcklbw ymmB, ymmB, ymmF ; ymmB=(01 03 05 07 09 0B 0D 0F 0H 0J 0L 0N 0P 0R 0T 0V)
  286. vpunpckhbw ymmD, ymmD, ymmF ; ymmD=(11 13 15 17 19 1B 1D 1F 1H 1J 1L 1N 1P 1R 1T 1V)
  287. vmovdqa ymmG, ymmE
  288. vpunpcklbw ymmE, ymmE, ymmF ; ymmE=(20 22 24 26 28 2A 2C 2E 2G 2I 2K 2M 2O 2Q 2S 2U)
  289. vpunpckhbw ymmG, ymmG, ymmF ; ymmG=(30 32 34 36 38 3A 3C 3E 3G 3I 3K 3M 3O 3Q 3S 3U)
  290. vpunpcklbw ymmF, ymmF, ymmH
  291. vpunpckhbw ymmH, ymmH, ymmH
  292. vpsrlw ymmF, ymmF, BYTE_BIT ; ymmF=(21 23 25 27 29 2B 2D 2F 2H 2J 2L 2N 2P 2R 2T 2V)
  293. vpsrlw ymmH, ymmH, BYTE_BIT ; ymmH=(31 33 35 37 39 3B 3D 3F 3H 3J 3L 3N 3P 3R 3T 3V)
  294. %endif ; RGB_PIXELSIZE ; ---------------
  295. ; ymm0=R(02468ACEGIKMOQSU)=RE, ymm2=G(02468ACEGIKMOQSU)=GE, ymm4=B(02468ACEGIKMOQSU)=BE
  296. ; ymm1=R(13579BDFHJLNPRTV)=RO, ymm3=G(13579BDFHJLNPRTV)=GO, ymm5=B(13579BDFHJLNPRTV)=BO
  297. ; (Original)
  298. ; Y = 0.29900 * R + 0.58700 * G + 0.11400 * B
  299. ;
  300. ; (This implementation)
  301. ; Y = 0.29900 * R + 0.33700 * G + 0.11400 * B + 0.25000 * G
  302. vmovdqa ymm6, ymm1
  303. vpunpcklwd ymm1, ymm1, ymm3
  304. vpunpckhwd ymm6, ymm6, ymm3
  305. vpmaddwd ymm1, ymm1, [rel PW_F0299_F0337] ; ymm1=ROL*FIX(0.299)+GOL*FIX(0.337)
  306. vpmaddwd ymm6, ymm6, [rel PW_F0299_F0337] ; ymm6=ROH*FIX(0.299)+GOH*FIX(0.337)
  307. vmovdqa ymm7, ymm6 ; ymm7=ROH*FIX(0.299)+GOH*FIX(0.337)
  308. vmovdqa ymm6, ymm0
  309. vpunpcklwd ymm0, ymm0, ymm2
  310. vpunpckhwd ymm6, ymm6, ymm2
  311. vpmaddwd ymm0, ymm0, [rel PW_F0299_F0337] ; ymm0=REL*FIX(0.299)+GEL*FIX(0.337)
  312. vpmaddwd ymm6, ymm6, [rel PW_F0299_F0337] ; ymm6=REH*FIX(0.299)+GEH*FIX(0.337)
  313. vmovdqa YMMWORD [wk(0)], ymm0 ; wk(0)=REL*FIX(0.299)+GEL*FIX(0.337)
  314. vmovdqa YMMWORD [wk(1)], ymm6 ; wk(1)=REH*FIX(0.299)+GEH*FIX(0.337)
  315. vmovdqa ymm0, ymm5 ; ymm0=BO
  316. vmovdqa ymm6, ymm4 ; ymm6=BE
  317. vmovdqa ymm4, ymm0
  318. vpunpcklwd ymm0, ymm0, ymm3
  319. vpunpckhwd ymm4, ymm4, ymm3
  320. vpmaddwd ymm0, ymm0, [rel PW_F0114_F0250] ; ymm0=BOL*FIX(0.114)+GOL*FIX(0.250)
  321. vpmaddwd ymm4, ymm4, [rel PW_F0114_F0250] ; ymm4=BOH*FIX(0.114)+GOH*FIX(0.250)
  322. vmovdqa ymm3, [rel PD_ONEHALF] ; ymm3=[PD_ONEHALF]
  323. vpaddd ymm0, ymm0, ymm1
  324. vpaddd ymm4, ymm4, ymm7
  325. vpaddd ymm0, ymm0, ymm3
  326. vpaddd ymm4, ymm4, ymm3
  327. vpsrld ymm0, ymm0, SCALEBITS ; ymm0=YOL
  328. vpsrld ymm4, ymm4, SCALEBITS ; ymm4=YOH
  329. vpackssdw ymm0, ymm0, ymm4 ; ymm0=YO
  330. vmovdqa ymm4, ymm6
  331. vpunpcklwd ymm6, ymm6, ymm2
  332. vpunpckhwd ymm4, ymm4, ymm2
  333. vpmaddwd ymm6, ymm6, [rel PW_F0114_F0250] ; ymm6=BEL*FIX(0.114)+GEL*FIX(0.250)
  334. vpmaddwd ymm4, ymm4, [rel PW_F0114_F0250] ; ymm4=BEH*FIX(0.114)+GEH*FIX(0.250)
  335. vmovdqa ymm2, [rel PD_ONEHALF] ; ymm2=[PD_ONEHALF]
  336. vpaddd ymm6, ymm6, YMMWORD [wk(0)]
  337. vpaddd ymm4, ymm4, YMMWORD [wk(1)]
  338. vpaddd ymm6, ymm6, ymm2
  339. vpaddd ymm4, ymm4, ymm2
  340. vpsrld ymm6, ymm6, SCALEBITS ; ymm6=YEL
  341. vpsrld ymm4, ymm4, SCALEBITS ; ymm4=YEH
  342. vpackssdw ymm6, ymm6, ymm4 ; ymm6=YE
  343. vpsllw ymm0, ymm0, BYTE_BIT
  344. vpor ymm6, ymm6, ymm0 ; ymm6=Y
  345. vmovdqu YMMWORD [rdi], ymm6 ; Save Y
  346. sub rcx, byte SIZEOF_YMMWORD
  347. add rsi, RGB_PIXELSIZE*SIZEOF_YMMWORD ; inptr
  348. add rdi, byte SIZEOF_YMMWORD ; outptr0
  349. cmp rcx, byte SIZEOF_YMMWORD
  350. jae near .columnloop
  351. test rcx, rcx
  352. jnz near .column_ld1
  353. pop rcx ; col
  354. pop rsi
  355. pop rdi
  356. add rsi, byte SIZEOF_JSAMPROW ; input_buf
  357. add rdi, byte SIZEOF_JSAMPROW
  358. dec rax ; num_rows
  359. jg near .rowloop
  360. .return:
  361. pop rbx
  362. vzeroupper
  363. uncollect_args 5
  364. mov rsp, rbp ; rsp <- aligned rbp
  365. pop rsp ; rsp <- original rbp
  366. pop rbp
  367. ret
  368. ; For some reason, the OS X linker does not honor the request to align the
  369. ; segment unless we do this.
  370. align 32