jdmrgext-avx2.asm 27 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575
  1. ;
  2. ; jdmrgext.asm - merged upsampling/color conversion (AVX2)
  3. ;
  4. ; Copyright 2009, 2012 Pierre Ossman <ossman@cendio.se> for Cendio AB
  5. ; Copyright (C) 2012, 2016, D. R. Commander.
  6. ; Copyright (C) 2015, Intel Corporation.
  7. ;
  8. ; Based on the x86 SIMD extension for IJG JPEG library
  9. ; Copyright (C) 1999-2006, MIYASAKA Masaru.
  10. ; For conditions of distribution and use, see copyright notice in jsimdext.inc
  11. ;
  12. ; This file should be assembled with NASM (Netwide Assembler),
  13. ; can *not* be assembled with Microsoft's MASM or any compatible
  14. ; assembler (including Borland's Turbo Assembler).
  15. ; NASM is available from http://nasm.sourceforge.net/ or
  16. ; http://sourceforge.net/project/showfiles.php?group_id=6208
  17. %include "jcolsamp.inc"
  18. ; --------------------------------------------------------------------------
  19. ;
  20. ; Upsample and color convert for the case of 2:1 horizontal and 1:1 vertical.
  21. ;
  22. ; GLOBAL(void)
  23. ; jsimd_h2v1_merged_upsample_avx2(JDIMENSION output_width,
  24. ; JSAMPIMAGE input_buf,
  25. ; JDIMENSION in_row_group_ctr,
  26. ; JSAMPARRAY output_buf);
  27. ;
  28. %define output_width(b) (b) + 8 ; JDIMENSION output_width
  29. %define input_buf(b) (b) + 12 ; JSAMPIMAGE input_buf
  30. %define in_row_group_ctr(b) (b) + 16 ; JDIMENSION in_row_group_ctr
  31. %define output_buf(b) (b) + 20 ; JSAMPARRAY output_buf
  32. %define original_ebp ebp + 0
  33. %define wk(i) ebp - (WK_NUM - (i)) * SIZEOF_YMMWORD
  34. ; ymmword wk[WK_NUM]
  35. %define WK_NUM 3
  36. %define gotptr wk(0) - SIZEOF_POINTER ; void * gotptr
  37. align 32
  38. GLOBAL_FUNCTION(jsimd_h2v1_merged_upsample_avx2)
  39. EXTN(jsimd_h2v1_merged_upsample_avx2):
  40. push ebp
  41. mov eax, esp ; eax = original ebp
  42. sub esp, byte 4
  43. and esp, byte (-SIZEOF_YMMWORD) ; align to 256 bits
  44. mov [esp], eax
  45. mov ebp, esp ; ebp = aligned ebp
  46. lea esp, [wk(0)]
  47. pushpic eax ; make a room for GOT address
  48. push ebx
  49. ; push ecx ; need not be preserved
  50. ; push edx ; need not be preserved
  51. push esi
  52. push edi
  53. get_GOT ebx ; get GOT address
  54. movpic POINTER [gotptr], ebx ; save GOT address
  55. mov ecx, JDIMENSION [output_width(eax)] ; col
  56. test ecx, ecx
  57. jz near .return
  58. push ecx
  59. mov edi, JSAMPIMAGE [input_buf(eax)]
  60. mov ecx, JDIMENSION [in_row_group_ctr(eax)]
  61. mov esi, JSAMPARRAY [edi+0*SIZEOF_JSAMPARRAY]
  62. mov ebx, JSAMPARRAY [edi+1*SIZEOF_JSAMPARRAY]
  63. mov edx, JSAMPARRAY [edi+2*SIZEOF_JSAMPARRAY]
  64. mov edi, JSAMPARRAY [output_buf(eax)]
  65. mov esi, JSAMPROW [esi+ecx*SIZEOF_JSAMPROW] ; inptr0
  66. mov ebx, JSAMPROW [ebx+ecx*SIZEOF_JSAMPROW] ; inptr1
  67. mov edx, JSAMPROW [edx+ecx*SIZEOF_JSAMPROW] ; inptr2
  68. mov edi, JSAMPROW [edi] ; outptr
  69. pop ecx ; col
  70. alignx 16, 7
  71. .columnloop:
  72. movpic eax, POINTER [gotptr] ; load GOT address (eax)
  73. vmovdqu ymm6, YMMWORD [ebx] ; ymm6=Cb(0123456789ABCDEFGHIJKLMNOPQRSTUV)
  74. vmovdqu ymm7, YMMWORD [edx] ; ymm7=Cr(0123456789ABCDEFGHIJKLMNOPQRSTUV)
  75. vpxor ymm1, ymm1, ymm1 ; ymm1=(all 0's)
  76. vpcmpeqw ymm3, ymm3, ymm3
  77. vpsllw ymm3, ymm3, 7 ; ymm3={0xFF80 0xFF80 0xFF80 0xFF80 ..}
  78. vpermq ymm6, ymm6, 0xd8 ; ymm6=Cb(01234567GHIJKLMN89ABCDEFOPQRSTUV)
  79. vpermq ymm7, ymm7, 0xd8 ; ymm7=Cr(01234567GHIJKLMN89ABCDEFOPQRSTUV)
  80. vpunpcklbw ymm4, ymm6, ymm1 ; ymm4=Cb(0123456789ABCDEF)=CbL
  81. vpunpckhbw ymm6, ymm6, ymm1 ; ymm6=Cb(GHIJKLMNOPQRSTUV)=CbH
  82. vpunpcklbw ymm0, ymm7, ymm1 ; ymm0=Cr(0123456789ABCDEF)=CrL
  83. vpunpckhbw ymm7, ymm7, ymm1 ; ymm7=Cr(GHIJKLMNOPQRSTUV)=CrH
  84. vpaddw ymm5, ymm6, ymm3
  85. vpaddw ymm2, ymm4, ymm3
  86. vpaddw ymm1, ymm7, ymm3
  87. vpaddw ymm3, ymm0, ymm3
  88. ; (Original)
  89. ; R = Y + 1.40200 * Cr
  90. ; G = Y - 0.34414 * Cb - 0.71414 * Cr
  91. ; B = Y + 1.77200 * Cb
  92. ;
  93. ; (This implementation)
  94. ; R = Y + 0.40200 * Cr + Cr
  95. ; G = Y - 0.34414 * Cb + 0.28586 * Cr - Cr
  96. ; B = Y - 0.22800 * Cb + Cb + Cb
  97. vpaddw ymm6, ymm5, ymm5 ; ymm6=2*CbH
  98. vpaddw ymm4, ymm2, ymm2 ; ymm4=2*CbL
  99. vpaddw ymm7, ymm1, ymm1 ; ymm7=2*CrH
  100. vpaddw ymm0, ymm3, ymm3 ; ymm0=2*CrL
  101. vpmulhw ymm6, ymm6, [GOTOFF(eax,PW_MF0228)] ; ymm6=(2*CbH * -FIX(0.22800))
  102. vpmulhw ymm4, ymm4, [GOTOFF(eax,PW_MF0228)] ; ymm4=(2*CbL * -FIX(0.22800))
  103. vpmulhw ymm7, ymm7, [GOTOFF(eax,PW_F0402)] ; ymm7=(2*CrH * FIX(0.40200))
  104. vpmulhw ymm0, ymm0, [GOTOFF(eax,PW_F0402)] ; ymm0=(2*CrL * FIX(0.40200))
  105. vpaddw ymm6, ymm6, [GOTOFF(eax,PW_ONE)]
  106. vpaddw ymm4, ymm4, [GOTOFF(eax,PW_ONE)]
  107. vpsraw ymm6, ymm6, 1 ; ymm6=(CbH * -FIX(0.22800))
  108. vpsraw ymm4, ymm4, 1 ; ymm4=(CbL * -FIX(0.22800))
  109. vpaddw ymm7, ymm7, [GOTOFF(eax,PW_ONE)]
  110. vpaddw ymm0, ymm0, [GOTOFF(eax,PW_ONE)]
  111. vpsraw ymm7, ymm7, 1 ; ymm7=(CrH * FIX(0.40200))
  112. vpsraw ymm0, ymm0, 1 ; ymm0=(CrL * FIX(0.40200))
  113. vpaddw ymm6, ymm6, ymm5
  114. vpaddw ymm4, ymm4, ymm2
  115. vpaddw ymm6, ymm6, ymm5 ; ymm6=(CbH * FIX(1.77200))=(B-Y)H
  116. vpaddw ymm4, ymm4, ymm2 ; ymm4=(CbL * FIX(1.77200))=(B-Y)L
  117. vpaddw ymm7, ymm7, ymm1 ; ymm7=(CrH * FIX(1.40200))=(R-Y)H
  118. vpaddw ymm0, ymm0, ymm3 ; ymm0=(CrL * FIX(1.40200))=(R-Y)L
  119. vmovdqa YMMWORD [wk(0)], ymm6 ; wk(0)=(B-Y)H
  120. vmovdqa YMMWORD [wk(1)], ymm7 ; wk(1)=(R-Y)H
  121. vpunpckhwd ymm6, ymm5, ymm1
  122. vpunpcklwd ymm5, ymm5, ymm1
  123. vpmaddwd ymm5, ymm5, [GOTOFF(eax,PW_MF0344_F0285)]
  124. vpmaddwd ymm6, ymm6, [GOTOFF(eax,PW_MF0344_F0285)]
  125. vpunpckhwd ymm7, ymm2, ymm3
  126. vpunpcklwd ymm2, ymm2, ymm3
  127. vpmaddwd ymm2, ymm2, [GOTOFF(eax,PW_MF0344_F0285)]
  128. vpmaddwd ymm7, ymm7, [GOTOFF(eax,PW_MF0344_F0285)]
  129. vpaddd ymm5, ymm5, [GOTOFF(eax,PD_ONEHALF)]
  130. vpaddd ymm6, ymm6, [GOTOFF(eax,PD_ONEHALF)]
  131. vpsrad ymm5, ymm5, SCALEBITS
  132. vpsrad ymm6, ymm6, SCALEBITS
  133. vpaddd ymm2, ymm2, [GOTOFF(eax,PD_ONEHALF)]
  134. vpaddd ymm7, ymm7, [GOTOFF(eax,PD_ONEHALF)]
  135. vpsrad ymm2, ymm2, SCALEBITS
  136. vpsrad ymm7, ymm7, SCALEBITS
  137. vpackssdw ymm5, ymm5, ymm6 ; ymm5=CbH*-FIX(0.344)+CrH*FIX(0.285)
  138. vpackssdw ymm2, ymm2, ymm7 ; ymm2=CbL*-FIX(0.344)+CrL*FIX(0.285)
  139. vpsubw ymm5, ymm5, ymm1 ; ymm5=CbH*-FIX(0.344)+CrH*-FIX(0.714)=(G-Y)H
  140. vpsubw ymm2, ymm2, ymm3 ; ymm2=CbL*-FIX(0.344)+CrL*-FIX(0.714)=(G-Y)L
  141. vmovdqa YMMWORD [wk(2)], ymm5 ; wk(2)=(G-Y)H
  142. mov al, 2 ; Yctr
  143. jmp short .Yloop_1st
  144. alignx 16, 7
  145. .Yloop_2nd:
  146. vmovdqa ymm0, YMMWORD [wk(1)] ; ymm0=(R-Y)H
  147. vmovdqa ymm2, YMMWORD [wk(2)] ; ymm2=(G-Y)H
  148. vmovdqa ymm4, YMMWORD [wk(0)] ; ymm4=(B-Y)H
  149. alignx 16, 7
  150. .Yloop_1st:
  151. vmovdqu ymm7, YMMWORD [esi] ; ymm7=Y(0123456789ABCDEFGHIJKLMNOPQRSTUV)
  152. vpcmpeqw ymm6, ymm6, ymm6
  153. vpsrlw ymm6, ymm6, BYTE_BIT ; ymm6={0xFF 0x00 0xFF 0x00 ..}
  154. vpand ymm6, ymm6, ymm7 ; ymm6=Y(02468ACEGIKMOQSU)=YE
  155. vpsrlw ymm7, ymm7, BYTE_BIT ; ymm7=Y(13579BDFHJLNPRTV)=YO
  156. vmovdqa ymm1, ymm0 ; ymm1=ymm0=(R-Y)(L/H)
  157. vmovdqa ymm3, ymm2 ; ymm3=ymm2=(G-Y)(L/H)
  158. vmovdqa ymm5, ymm4 ; ymm5=ymm4=(B-Y)(L/H)
  159. vpaddw ymm0, ymm0, ymm6 ; ymm0=((R-Y)+YE)=RE=R(02468ACEGIKMOQSU)
  160. vpaddw ymm1, ymm1, ymm7 ; ymm1=((R-Y)+YO)=RO=R(13579BDFHJLNPRTV)
  161. vpackuswb ymm0, ymm0, ymm0 ; ymm0=R(02468ACE********GIKMOQSU********)
  162. vpackuswb ymm1, ymm1, ymm1 ; ymm1=R(13579BDF********HJLNPRTV********)
  163. vpaddw ymm2, ymm2, ymm6 ; ymm2=((G-Y)+YE)=GE=G(02468ACEGIKMOQSU)
  164. vpaddw ymm3, ymm3, ymm7 ; ymm3=((G-Y)+YO)=GO=G(13579BDFHJLNPRTV)
  165. vpackuswb ymm2, ymm2, ymm2 ; ymm2=G(02468ACE********GIKMOQSU********)
  166. vpackuswb ymm3, ymm3, ymm3 ; ymm3=G(13579BDF********HJLNPRTV********)
  167. vpaddw ymm4, ymm4, ymm6 ; ymm4=((B-Y)+YE)=BE=B(02468ACEGIKMOQSU)
  168. vpaddw ymm5, ymm5, ymm7 ; ymm5=((B-Y)+YO)=BO=B(13579BDFHJLNPRTV)
  169. vpackuswb ymm4, ymm4, ymm4 ; ymm4=B(02468ACE********GIKMOQSU********)
  170. vpackuswb ymm5, ymm5, ymm5 ; ymm5=B(13579BDF********HJLNPRTV********)
  171. %if RGB_PIXELSIZE == 3 ; ---------------
  172. ; ymmA=(00 02 04 06 08 0A 0C 0E ** 0G 0I 0K 0M 0O 0Q 0S 0U **)
  173. ; ymmB=(01 03 05 07 09 0B 0D 0F ** 0H 0J 0L 0N 0P 0R 0T 0V **)
  174. ; ymmC=(10 12 14 16 18 1A 1C 1E ** 1G 1I 1K 1M 1O 1Q 1S 1U **)
  175. ; ymmD=(11 13 15 17 19 1B 1D 1F ** 1H 1J 1L 1N 1P 1R 1T 1V **)
  176. ; ymmE=(20 22 24 26 28 2A 2C 2E ** 2G 2I 2K 2M 2O 2Q 2S 2U **)
  177. ; ymmF=(21 23 25 27 29 2B 2D 2F ** 2H 2J 2L 2N 2P 2R 2T 2V **)
  178. ; ymmG=(** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** **)
  179. ; ymmH=(** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** ** **)
  180. vpunpcklbw ymmA, ymmA, ymmC ; ymmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E
  181. ; 0G 1G 0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U)
  182. vpunpcklbw ymmE, ymmE, ymmB ; ymmE=(20 01 22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F
  183. ; 2G 0H 2I 0J 2K 0L 2M 0N 2O 0P 2Q 0R 2S 0T 2U 0V)
  184. vpunpcklbw ymmD, ymmD, ymmF ; ymmD=(11 21 13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F
  185. ; 1H 2H 1J 2J 1L 2L 1N 2N 1P 2P 1R 2R 1T 2T 1V 2V)
  186. vpsrldq ymmH, ymmA, 2 ; ymmH=(02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E 0G 1G
  187. ; 0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U -- --)
  188. vpunpckhwd ymmG, ymmA, ymmE ; ymmG=(08 18 28 09 0A 1A 2A 0B 0C 1C 2C 0D 0E 1E 2E 0F
  189. ; 0O 1O 2O 0P 0Q 1Q 2Q 0R 0S 1S 2S 0T 0U 1U 2U 0V)
  190. vpunpcklwd ymmA, ymmA, ymmE ; ymmA=(00 10 20 01 02 12 22 03 04 14 24 05 06 16 26 07
  191. ; 0G 1G 2G 0H 0I 1I 2I 0J 0K 1K 2K 0L 0M 1M 2M 0N)
  192. vpsrldq ymmE, ymmE, 2 ; ymmE=(22 03 24 05 26 07 28 09 2A 0B 2C 0D 2E 0F 2G 0H
  193. ; 2I 0J 2K 0L 2M 0N 2O 0P 2Q 0R 2S 0T 2U 0V -- --)
  194. vpsrldq ymmB, ymmD, 2 ; ymmB=(13 23 15 25 17 27 19 29 1B 2B 1D 2D 1F 2F 1H 2H
  195. ; 1J 2J 1L 2L 1N 2N 1P 2P 1R 2R 1T 2T 1V 2V -- --)
  196. vpunpckhwd ymmC, ymmD, ymmH ; ymmC=(19 29 0A 1A 1B 2B 0C 1C 1D 2D 0E 1E 1F 2F 0G 1G
  197. ; 1P 2P 0Q 1Q 1R 2R 0S 1S 1T 2T 0U 1U 1V 2V -- --)
  198. vpunpcklwd ymmD, ymmD, ymmH ; ymmD=(11 21 02 12 13 23 04 14 15 25 06 16 17 27 08 18
  199. ; 1H 2H 0I 1I 1J 2J 0K 1K 1L 2L 0M 1M 1N 2N 0O 1O)
  200. vpunpckhwd ymmF, ymmE, ymmB ; ymmF=(2A 0B 1B 2B 2C 0D 1D 2D 2E 0F 1F 2F 2G 0H 1H 2H
  201. ; 2Q 0R 1R 2R 2S 0T 1T 2T 2U 0V 1V 2V -- -- -- --)
  202. vpunpcklwd ymmE, ymmE, ymmB ; ymmE=(22 03 13 23 24 05 15 25 26 07 17 27 28 09 19 29
  203. ; 2I 0J 1J 2J 2K 0L 1L 2L 2M 0N 1N 2N 2O 0P 1P 2P)
  204. vpshufd ymmH, ymmA, 0x4E ; ymmH=(04 14 24 05 06 16 26 07 00 10 20 01 02 12 22 03
  205. ; 0K 1K 2K 0L 0M 1M 2M 0N 0G 1G 2G 0H 0I 1I 2I 0J)
  206. vpunpckldq ymmA, ymmA, ymmD ; ymmA=(00 10 20 01 11 21 02 12 02 12 22 03 13 23 04 14
  207. ; 0G 1G 2G 0H 1H 2H 0I 1I 0I 1I 2I 0J 1J 2J 0K 1K)
  208. vpunpckhdq ymmD, ymmD, ymmE ; ymmD=(15 25 06 16 26 07 17 27 17 27 08 18 28 09 19 29
  209. ; 1L 2L 0M 1M 2M 0N 1N 2N 1N 2N 0O 1O 2O 0P 1P 2P)
  210. vpunpckldq ymmE, ymmE, ymmH ; ymmE=(22 03 13 23 04 14 24 05 24 05 15 25 06 16 26 07
  211. ; 2I 0J 1J 2J 0K 1K 2K 0L 2K 0L 1L 2L 0M 1M 2M 0N)
  212. vpshufd ymmH, ymmG, 0x4E ; ymmH=(0C 1C 2C 0D 0E 1E 2E 0F 08 18 28 09 0A 1A 2A 0B
  213. ; 0S 1S 2S 0T 0U 1U 2U 0V 0O 1O 2O 0P 0Q 1Q 2Q 0R)
  214. vpunpckldq ymmG, ymmG, ymmC ; ymmG=(08 18 28 09 19 29 0A 1A 0A 1A 2A 0B 1B 2B 0C 1C
  215. ; 0O 1O 2O 0P 1P 2P 0Q 1Q 0Q 1Q 2Q 0R 1R 2R 0S 1S)
  216. vpunpckhdq ymmC, ymmC, ymmF ; ymmC=(1D 2D 0E 1E 2E 0F 1F 2F 1F 2F 0G 1G 2G 0H 1H 2H
  217. ; 1T 2T 0U 1U 2U 0V 1V 2V 1V 2V -- -- -- -- -- --)
  218. vpunpckldq ymmF, ymmF, ymmH ; ymmF=(2A 0B 1B 2B 0C 1C 2C 0D 2C 0D 1D 2D 0E 1E 2E 0F
  219. ; 2Q 0R 1R 2R 0S 1S 2S 0T 2S 0T 1T 2T 0U 1U 2U 0V)
  220. vpunpcklqdq ymmH, ymmA, ymmE ; ymmH=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
  221. ; 0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
  222. vpunpcklqdq ymmG, ymmD, ymmG ; ymmG=(15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A
  223. ; 1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q)
  224. vpunpcklqdq ymmC, ymmF, ymmC ; ymmC=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
  225. ; 2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
  226. vperm2i128 ymmA, ymmH, ymmG, 0x20 ; ymmA=(00 10 20 01 11 21 02 12 22 03 13 23 04 14 24 05
  227. ; 15 25 06 16 26 07 17 27 08 18 28 09 19 29 0A 1A)
  228. vperm2i128 ymmD, ymmC, ymmH, 0x30 ; ymmD=(2A 0B 1B 2B 0C 1C 2C 0D 1D 2D 0E 1E 2E 0F 1F 2F
  229. ; 0G 1G 2G 0H 1H 2H 0I 1I 2I 0J 1J 2J 0K 1K 2K 0L)
  230. vperm2i128 ymmF, ymmG, ymmC, 0x31 ; ymmF=(1L 2L 0M 1M 2M 0N 1N 2N 0O 1O 2O 0P 1P 2P 0Q 1Q
  231. ; 2Q 0R 1R 2R 0S 1S 2S 0T 1T 2T 0U 1U 2U 0V 1V 2V)
  232. cmp ecx, byte SIZEOF_YMMWORD
  233. jb short .column_st64
  234. test edi, SIZEOF_YMMWORD-1
  235. jnz short .out1
  236. ; --(aligned)-------------------
  237. vmovntdq YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  238. vmovntdq YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  239. vmovntdq YMMWORD [edi+2*SIZEOF_YMMWORD], ymmF
  240. jmp short .out0
  241. .out1: ; --(unaligned)-----------------
  242. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  243. vmovdqu YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  244. vmovdqu YMMWORD [edi+2*SIZEOF_YMMWORD], ymmF
  245. .out0:
  246. add edi, byte RGB_PIXELSIZE*SIZEOF_YMMWORD ; outptr
  247. sub ecx, byte SIZEOF_YMMWORD
  248. jz near .endcolumn
  249. add esi, byte SIZEOF_YMMWORD ; inptr0
  250. dec al ; Yctr
  251. jnz near .Yloop_2nd
  252. add ebx, byte SIZEOF_YMMWORD ; inptr1
  253. add edx, byte SIZEOF_YMMWORD ; inptr2
  254. jmp near .columnloop
  255. alignx 16, 7
  256. .column_st64:
  257. lea ecx, [ecx+ecx*2] ; imul ecx, RGB_PIXELSIZE
  258. cmp ecx, byte 2*SIZEOF_YMMWORD
  259. jb short .column_st32
  260. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  261. vmovdqu YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  262. add edi, byte 2*SIZEOF_YMMWORD ; outptr
  263. vmovdqa ymmA, ymmF
  264. sub ecx, byte 2*SIZEOF_YMMWORD
  265. jmp short .column_st31
  266. .column_st32:
  267. cmp ecx, byte SIZEOF_YMMWORD
  268. jb short .column_st31
  269. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  270. add edi, byte SIZEOF_YMMWORD ; outptr
  271. vmovdqa ymmA, ymmD
  272. sub ecx, byte SIZEOF_YMMWORD
  273. jmp short .column_st31
  274. .column_st31:
  275. cmp ecx, byte SIZEOF_XMMWORD
  276. jb short .column_st15
  277. vmovdqu XMMWORD [edi+0*SIZEOF_XMMWORD], xmmA
  278. add edi, byte SIZEOF_XMMWORD ; outptr
  279. vperm2i128 ymmA, ymmA, ymmA, 1
  280. sub ecx, byte SIZEOF_XMMWORD
  281. .column_st15:
  282. ; Store the lower 8 bytes of xmmA to the output when it has enough
  283. ; space.
  284. cmp ecx, byte SIZEOF_MMWORD
  285. jb short .column_st7
  286. vmovq XMM_MMWORD [edi], xmmA
  287. add edi, byte SIZEOF_MMWORD
  288. sub ecx, byte SIZEOF_MMWORD
  289. vpsrldq xmmA, xmmA, SIZEOF_MMWORD
  290. .column_st7:
  291. ; Store the lower 4 bytes of xmmA to the output when it has enough
  292. ; space.
  293. cmp ecx, byte SIZEOF_DWORD
  294. jb short .column_st3
  295. vmovd XMM_DWORD [edi], xmmA
  296. add edi, byte SIZEOF_DWORD
  297. sub ecx, byte SIZEOF_DWORD
  298. vpsrldq xmmA, xmmA, SIZEOF_DWORD
  299. .column_st3:
  300. ; Store the lower 2 bytes of eax to the output when it has enough
  301. ; space.
  302. vmovd eax, xmmA
  303. cmp ecx, byte SIZEOF_WORD
  304. jb short .column_st1
  305. mov word [edi], ax
  306. add edi, byte SIZEOF_WORD
  307. sub ecx, byte SIZEOF_WORD
  308. shr eax, 16
  309. .column_st1:
  310. ; Store the lower 1 byte of eax to the output when it has enough
  311. ; space.
  312. test ecx, ecx
  313. jz short .endcolumn
  314. mov byte [edi], al
  315. %else ; RGB_PIXELSIZE == 4 ; -----------
  316. %ifdef RGBX_FILLER_0XFF
  317. vpcmpeqb ymm6, ymm6, ymm6 ; ymm6=XE=X(02468ACE********GIKMOQSU********)
  318. vpcmpeqb ymm7, ymm7, ymm7 ; ymm7=XO=X(13579BDF********HJLNPRTV********)
  319. %else
  320. vpxor ymm6, ymm6, ymm6 ; ymm6=XE=X(02468ACE********GIKMOQSU********)
  321. vpxor ymm7, ymm7, ymm7 ; ymm7=XO=X(13579BDF********HJLNPRTV********)
  322. %endif
  323. ; ymmA=(00 02 04 06 08 0A 0C 0E ** 0G 0I 0K 0M 0O 0Q 0S 0U **)
  324. ; ymmB=(01 03 05 07 09 0B 0D 0F ** 0H 0J 0L 0N 0P 0R 0T 0V **)
  325. ; ymmC=(10 12 14 16 18 1A 1C 1E ** 1G 1I 1K 1M 1O 1Q 1S 1U **)
  326. ; ymmD=(11 13 15 17 19 1B 1D 1F ** 1H 1J 1L 1N 1P 1R 1T 1V **)
  327. ; ymmE=(20 22 24 26 28 2A 2C 2E ** 2G 2I 2K 2M 2O 2Q 2S 2U **)
  328. ; ymmF=(21 23 25 27 29 2B 2D 2F ** 2H 2J 2L 2N 2P 2R 2T 2V **)
  329. ; ymmG=(30 32 34 36 38 3A 3C 3E ** 3G 3I 3K 3M 3O 3Q 3S 3U **)
  330. ; ymmH=(31 33 35 37 39 3B 3D 3F ** 3H 3J 3L 3N 3P 3R 3T 3V **)
  331. vpunpcklbw ymmA, ymmA, ymmC ; ymmA=(00 10 02 12 04 14 06 16 08 18 0A 1A 0C 1C 0E 1E
  332. ; 0G 1G 0I 1I 0K 1K 0M 1M 0O 1O 0Q 1Q 0S 1S 0U 1U)
  333. vpunpcklbw ymmE, ymmE, ymmG ; ymmE=(20 30 22 32 24 34 26 36 28 38 2A 3A 2C 3C 2E 3E
  334. ; 2G 3G 2I 3I 2K 3K 2M 3M 2O 3O 2Q 3Q 2S 3S 2U 3U)
  335. vpunpcklbw ymmB, ymmB, ymmD ; ymmB=(01 11 03 13 05 15 07 17 09 19 0B 1B 0D 1D 0F 1F
  336. ; 0H 1H 0J 1J 0L 1L 0N 1N 0P 1P 0R 1R 0T 1T 0V 1V)
  337. vpunpcklbw ymmF, ymmF, ymmH ; ymmF=(21 31 23 33 25 35 27 37 29 39 2B 3B 2D 3D 2F 3F
  338. ; 2H 3H 2J 3J 2L 3L 2N 3N 2P 3P 2R 3R 2T 3T 2V 3V)
  339. vpunpckhwd ymmC, ymmA, ymmE ; ymmC=(08 18 28 38 0A 1A 2A 3A 0C 1C 2C 3C 0E 1E 2E 3E
  340. ; 0O 1O 2O 3O 0Q 1Q 2Q 3Q 0S 1S 2S 3S 0U 1U 2U 3U)
  341. vpunpcklwd ymmA, ymmA, ymmE ; ymmA=(00 10 20 30 02 12 22 32 04 14 24 34 06 16 26 36
  342. ; 0G 1G 2G 3G 0I 1I 2I 3I 0K 1K 2K 3K 0M 1M 2M 3M)
  343. vpunpckhwd ymmG, ymmB, ymmF ; ymmG=(09 19 29 39 0B 1B 2B 3B 0D 1D 2D 3D 0F 1F 2F 3F
  344. ; 0P 1P 2P 3P 0R 1R 2R 3R 0T 1T 2T 3T 0V 1V 2V 3V)
  345. vpunpcklwd ymmB, ymmB, ymmF ; ymmB=(01 11 21 31 03 13 23 33 05 15 25 35 07 17 27 37
  346. ; 0H 1H 2H 3H 0J 1J 2J 3J 0L 1L 2L 3L 0N 1N 2N 3N)
  347. vpunpckhdq ymmE, ymmA, ymmB ; ymmE=(04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37
  348. ; 0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
  349. vpunpckldq ymmB, ymmA, ymmB ; ymmB=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
  350. ; 0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J)
  351. vpunpckhdq ymmF, ymmC, ymmG ; ymmF=(0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F
  352. ; 0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
  353. vpunpckldq ymmG, ymmC, ymmG ; ymmG=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
  354. ; 0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R)
  355. vperm2i128 ymmA, ymmB, ymmE, 0x20 ; ymmA=(00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
  356. ; 04 14 24 34 05 15 25 35 06 16 26 36 07 17 27 37)
  357. vperm2i128 ymmD, ymmG, ymmF, 0x20 ; ymmD=(08 18 28 38 09 19 29 39 0A 1A 2A 3A 0B 1B 2B 3B
  358. ; 0C 1C 2C 3C 0D 1D 2D 3D 0E 1E 2E 3E 0F 1F 2F 3F)
  359. vperm2i128 ymmC, ymmB, ymmE, 0x31 ; ymmC=(0G 1G 2G 3G 0H 1H 2H 3H 0I 1I 2I 3I 0J 1J 2J 3J
  360. ; 0K 1K 2K 3K 0L 1L 2L 3L 0M 1M 2M 3M 0N 1N 2N 3N)
  361. vperm2i128 ymmH, ymmG, ymmF, 0x31 ; ymmH=(0O 1O 2O 3O 0P 1P 2P 3P 0Q 1Q 2Q 3Q 0R 1R 2R 3R
  362. ; 0S 1S 2S 3S 0T 1T 2T 3T 0U 1U 2U 3U 0V 1V 2V 3V)
  363. cmp ecx, byte SIZEOF_YMMWORD
  364. jb short .column_st64
  365. test edi, SIZEOF_YMMWORD-1
  366. jnz short .out1
  367. ; --(aligned)-------------------
  368. vmovntdq YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  369. vmovntdq YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  370. vmovntdq YMMWORD [edi+2*SIZEOF_YMMWORD], ymmC
  371. vmovntdq YMMWORD [edi+3*SIZEOF_YMMWORD], ymmH
  372. jmp short .out0
  373. .out1: ; --(unaligned)-----------------
  374. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  375. vmovdqu YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  376. vmovdqu YMMWORD [edi+2*SIZEOF_YMMWORD], ymmC
  377. vmovdqu YMMWORD [edi+3*SIZEOF_YMMWORD], ymmH
  378. .out0:
  379. add edi, RGB_PIXELSIZE*SIZEOF_YMMWORD ; outptr
  380. sub ecx, byte SIZEOF_YMMWORD
  381. jz near .endcolumn
  382. add esi, byte SIZEOF_YMMWORD ; inptr0
  383. dec al
  384. jnz near .Yloop_2nd
  385. add ebx, byte SIZEOF_YMMWORD ; inptr1
  386. add edx, byte SIZEOF_YMMWORD ; inptr2
  387. jmp near .columnloop
  388. alignx 16, 7
  389. .column_st64:
  390. cmp ecx, byte SIZEOF_YMMWORD/2
  391. jb short .column_st32
  392. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  393. vmovdqu YMMWORD [edi+1*SIZEOF_YMMWORD], ymmD
  394. add edi, byte 2*SIZEOF_YMMWORD ; outptr
  395. vmovdqa ymmA, ymmC
  396. vmovdqa ymmD, ymmH
  397. sub ecx, byte SIZEOF_YMMWORD/2
  398. .column_st32:
  399. cmp ecx, byte SIZEOF_YMMWORD/4
  400. jb short .column_st16
  401. vmovdqu YMMWORD [edi+0*SIZEOF_YMMWORD], ymmA
  402. add edi, byte SIZEOF_YMMWORD ; outptr
  403. vmovdqa ymmA, ymmD
  404. sub ecx, byte SIZEOF_YMMWORD/4
  405. .column_st16:
  406. cmp ecx, byte SIZEOF_YMMWORD/8
  407. jb short .column_st15
  408. vmovdqu XMMWORD [edi+0*SIZEOF_XMMWORD], xmmA
  409. add edi, byte SIZEOF_XMMWORD ; outptr
  410. vperm2i128 ymmA, ymmA, ymmA, 1
  411. sub ecx, byte SIZEOF_YMMWORD/8
  412. .column_st15:
  413. ; Store two pixels (8 bytes) of ymmA to the output when it has enough
  414. ; space.
  415. cmp ecx, byte SIZEOF_YMMWORD/16
  416. jb short .column_st7
  417. vmovq MMWORD [edi], xmmA
  418. add edi, byte SIZEOF_YMMWORD/16*4
  419. sub ecx, byte SIZEOF_YMMWORD/16
  420. vpsrldq xmmA, SIZEOF_YMMWORD/16*4
  421. .column_st7:
  422. ; Store one pixel (4 bytes) of ymmA to the output when it has enough
  423. ; space.
  424. test ecx, ecx
  425. jz short .endcolumn
  426. vmovd XMM_DWORD [edi], xmmA
  427. %endif ; RGB_PIXELSIZE ; ---------------
  428. .endcolumn:
  429. sfence ; flush the write buffer
  430. .return:
  431. vzeroupper
  432. pop edi
  433. pop esi
  434. ; pop edx ; need not be preserved
  435. ; pop ecx ; need not be preserved
  436. pop ebx
  437. mov esp, ebp ; esp <- aligned ebp
  438. pop esp ; esp <- original ebp
  439. pop ebp
  440. ret
  441. ; --------------------------------------------------------------------------
  442. ;
  443. ; Upsample and color convert for the case of 2:1 horizontal and 2:1 vertical.
  444. ;
  445. ; GLOBAL(void)
  446. ; jsimd_h2v2_merged_upsample_avx2(JDIMENSION output_width,
  447. ; JSAMPIMAGE input_buf,
  448. ; JDIMENSION in_row_group_ctr,
  449. ; JSAMPARRAY output_buf);
  450. ;
  451. %define output_width(b) (b) + 8 ; JDIMENSION output_width
  452. %define input_buf(b) (b) + 12 ; JSAMPIMAGE input_buf
  453. %define in_row_group_ctr(b) (b) + 16 ; JDIMENSION in_row_group_ctr
  454. %define output_buf(b) (b) + 20 ; JSAMPARRAY output_buf
  455. align 32
  456. GLOBAL_FUNCTION(jsimd_h2v2_merged_upsample_avx2)
  457. EXTN(jsimd_h2v2_merged_upsample_avx2):
  458. push ebp
  459. mov ebp, esp
  460. push ebx
  461. ; push ecx ; need not be preserved
  462. ; push edx ; need not be preserved
  463. push esi
  464. push edi
  465. mov eax, POINTER [output_width(ebp)]
  466. mov edi, JSAMPIMAGE [input_buf(ebp)]
  467. mov ecx, JDIMENSION [in_row_group_ctr(ebp)]
  468. mov esi, JSAMPARRAY [edi+0*SIZEOF_JSAMPARRAY]
  469. mov ebx, JSAMPARRAY [edi+1*SIZEOF_JSAMPARRAY]
  470. mov edx, JSAMPARRAY [edi+2*SIZEOF_JSAMPARRAY]
  471. mov edi, JSAMPARRAY [output_buf(ebp)]
  472. lea esi, [esi+ecx*SIZEOF_JSAMPROW]
  473. push edx ; inptr2
  474. push ebx ; inptr1
  475. push esi ; inptr00
  476. mov ebx, esp
  477. push edi ; output_buf (outptr0)
  478. push ecx ; in_row_group_ctr
  479. push ebx ; input_buf
  480. push eax ; output_width
  481. call near EXTN(jsimd_h2v1_merged_upsample_avx2)
  482. add esi, byte SIZEOF_JSAMPROW ; inptr01
  483. add edi, byte SIZEOF_JSAMPROW ; outptr1
  484. mov POINTER [ebx+0*SIZEOF_POINTER], esi
  485. mov POINTER [ebx-1*SIZEOF_POINTER], edi
  486. call near EXTN(jsimd_h2v1_merged_upsample_avx2)
  487. add esp, byte 7*SIZEOF_DWORD
  488. pop edi
  489. pop esi
  490. ; pop edx ; need not be preserved
  491. ; pop ecx ; need not be preserved
  492. pop ebx
  493. pop ebp
  494. ret
  495. ; For some reason, the OS X linker does not honor the request to align the
  496. ; segment unless we do this.
  497. align 32