1; 2; jidctint.asm - accurate integer IDCT (SSE2) 3; 4; Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB 5; 6; Based on 7; x86 SIMD extension for IJG JPEG library 8; Copyright (C) 1999-2006, MIYASAKA Masaru. 9; For conditions of distribution and use, see copyright notice in jsimdext.inc 10; 11; This file should be assembled with NASM (Netwide Assembler), 12; can *not* be assembled with Microsoft's MASM or any compatible 13; assembler (including Borland's Turbo Assembler). 14; NASM is available from http://nasm.sourceforge.net/ or 15; http://sourceforge.net/project/showfiles.php?group_id=6208 16; 17; This file contains a slow-but-accurate integer implementation of the 18; inverse DCT (Discrete Cosine Transform). The following code is based 19; directly on the IJG's original jidctint.c; see the jidctint.c for 20; more details. 21; 22; [TAB8] 23 24%include "jsimdext.inc" 25%include "jdct.inc" 26 27; -------------------------------------------------------------------------- 28 29%define CONST_BITS 13 30%define PASS1_BITS 2 31 32%define DESCALE_P1 (CONST_BITS-PASS1_BITS) 33%define DESCALE_P2 (CONST_BITS+PASS1_BITS+3) 34 35%if CONST_BITS == 13 36F_0_298 equ 2446 ; FIX(0.298631336) 37F_0_390 equ 3196 ; FIX(0.390180644) 38F_0_541 equ 4433 ; FIX(0.541196100) 39F_0_765 equ 6270 ; FIX(0.765366865) 40F_0_899 equ 7373 ; FIX(0.899976223) 41F_1_175 equ 9633 ; FIX(1.175875602) 42F_1_501 equ 12299 ; FIX(1.501321110) 43F_1_847 equ 15137 ; FIX(1.847759065) 44F_1_961 equ 16069 ; FIX(1.961570560) 45F_2_053 equ 16819 ; FIX(2.053119869) 46F_2_562 equ 20995 ; FIX(2.562915447) 47F_3_072 equ 25172 ; FIX(3.072711026) 48%else 49; NASM cannot do compile-time arithmetic on floating-point constants. 50%define DESCALE(x,n) (((x)+(1<<((n)-1)))>>(n)) 51F_0_298 equ DESCALE( 320652955,30-CONST_BITS) ; FIX(0.298631336) 52F_0_390 equ DESCALE( 418953276,30-CONST_BITS) ; FIX(0.390180644) 53F_0_541 equ DESCALE( 581104887,30-CONST_BITS) ; FIX(0.541196100) 54F_0_765 equ DESCALE( 821806413,30-CONST_BITS) ; FIX(0.765366865) 55F_0_899 equ DESCALE( 966342111,30-CONST_BITS) ; FIX(0.899976223) 56F_1_175 equ DESCALE(1262586813,30-CONST_BITS) ; FIX(1.175875602) 57F_1_501 equ DESCALE(1612031267,30-CONST_BITS) ; FIX(1.501321110) 58F_1_847 equ DESCALE(1984016188,30-CONST_BITS) ; FIX(1.847759065) 59F_1_961 equ DESCALE(2106220350,30-CONST_BITS) ; FIX(1.961570560) 60F_2_053 equ DESCALE(2204520673,30-CONST_BITS) ; FIX(2.053119869) 61F_2_562 equ DESCALE(2751909506,30-CONST_BITS) ; FIX(2.562915447) 62F_3_072 equ DESCALE(3299298341,30-CONST_BITS) ; FIX(3.072711026) 63%endif 64 65; -------------------------------------------------------------------------- 66 SECTION SEG_CONST 67 68 alignz 16 69 global EXTN(jconst_idct_islow_sse2) 70 71EXTN(jconst_idct_islow_sse2): 72 73PW_F130_F054 times 4 dw (F_0_541+F_0_765), F_0_541 74PW_F054_MF130 times 4 dw F_0_541, (F_0_541-F_1_847) 75PW_MF078_F117 times 4 dw (F_1_175-F_1_961), F_1_175 76PW_F117_F078 times 4 dw F_1_175, (F_1_175-F_0_390) 77PW_MF060_MF089 times 4 dw (F_0_298-F_0_899),-F_0_899 78PW_MF089_F060 times 4 dw -F_0_899, (F_1_501-F_0_899) 79PW_MF050_MF256 times 4 dw (F_2_053-F_2_562),-F_2_562 80PW_MF256_F050 times 4 dw -F_2_562, (F_3_072-F_2_562) 81PD_DESCALE_P1 times 4 dd 1 << (DESCALE_P1-1) 82PD_DESCALE_P2 times 4 dd 1 << (DESCALE_P2-1) 83PB_CENTERJSAMP times 16 db CENTERJSAMPLE 84 85 alignz 16 86 87; -------------------------------------------------------------------------- 88 SECTION SEG_TEXT 89 BITS 32 90; 91; Perform dequantization and inverse DCT on one block of coefficients. 92; 93; GLOBAL(void) 94; jsimd_idct_islow_sse2 (void * dct_table, JCOEFPTR coef_block, 95; JSAMPARRAY output_buf, JDIMENSION output_col) 96; 97 98%define dct_table(b) (b)+8 ; jpeg_component_info * compptr 99%define coef_block(b) (b)+12 ; JCOEFPTR coef_block 100%define output_buf(b) (b)+16 ; JSAMPARRAY output_buf 101%define output_col(b) (b)+20 ; JDIMENSION output_col 102 103%define original_ebp ebp+0 104%define wk(i) ebp-(WK_NUM-(i))*SIZEOF_XMMWORD ; xmmword wk[WK_NUM] 105%define WK_NUM 12 106 107 align 16 108 global EXTN(jsimd_idct_islow_sse2) 109 110EXTN(jsimd_idct_islow_sse2): 111 push ebp 112 mov eax,esp ; eax = original ebp 113 sub esp, byte 4 114 and esp, byte (-SIZEOF_XMMWORD) ; align to 128 bits 115 mov [esp],eax 116 mov ebp,esp ; ebp = aligned ebp 117 lea esp, [wk(0)] 118 pushpic ebx 119; push ecx ; unused 120; push edx ; need not be preserved 121 push esi 122 push edi 123 124 get_GOT ebx ; get GOT address 125 126 ; ---- Pass 1: process columns from input. 127 128; mov eax, [original_ebp] 129 mov edx, POINTER [dct_table(eax)] ; quantptr 130 mov esi, JCOEFPTR [coef_block(eax)] ; inptr 131 132%ifndef NO_ZERO_COLUMN_TEST_ISLOW_SSE2 133 mov eax, DWORD [DWBLOCK(1,0,esi,SIZEOF_JCOEF)] 134 or eax, DWORD [DWBLOCK(2,0,esi,SIZEOF_JCOEF)] 135 jnz near .columnDCT 136 137 movdqa xmm0, XMMWORD [XMMBLOCK(1,0,esi,SIZEOF_JCOEF)] 138 movdqa xmm1, XMMWORD [XMMBLOCK(2,0,esi,SIZEOF_JCOEF)] 139 por xmm0, XMMWORD [XMMBLOCK(3,0,esi,SIZEOF_JCOEF)] 140 por xmm1, XMMWORD [XMMBLOCK(4,0,esi,SIZEOF_JCOEF)] 141 por xmm0, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] 142 por xmm1, XMMWORD [XMMBLOCK(6,0,esi,SIZEOF_JCOEF)] 143 por xmm0, XMMWORD [XMMBLOCK(7,0,esi,SIZEOF_JCOEF)] 144 por xmm1,xmm0 145 packsswb xmm1,xmm1 146 packsswb xmm1,xmm1 147 movd eax,xmm1 148 test eax,eax 149 jnz short .columnDCT 150 151 ; -- AC terms all zero 152 153 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] 154 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 155 156 psllw xmm5,PASS1_BITS 157 158 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) 159 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) 160 punpckhwd xmm4,xmm4 ; xmm4=(04 04 05 05 06 06 07 07) 161 162 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) 163 pshufd xmm6,xmm5,0x55 ; xmm6=col1=(01 01 01 01 01 01 01 01) 164 pshufd xmm1,xmm5,0xAA ; xmm1=col2=(02 02 02 02 02 02 02 02) 165 pshufd xmm5,xmm5,0xFF ; xmm5=col3=(03 03 03 03 03 03 03 03) 166 pshufd xmm0,xmm4,0x00 ; xmm0=col4=(04 04 04 04 04 04 04 04) 167 pshufd xmm3,xmm4,0x55 ; xmm3=col5=(05 05 05 05 05 05 05 05) 168 pshufd xmm2,xmm4,0xAA ; xmm2=col6=(06 06 06 06 06 06 06 06) 169 pshufd xmm4,xmm4,0xFF ; xmm4=col7=(07 07 07 07 07 07 07 07) 170 171 movdqa XMMWORD [wk(8)], xmm6 ; wk(8)=col1 172 movdqa XMMWORD [wk(9)], xmm5 ; wk(9)=col3 173 movdqa XMMWORD [wk(10)], xmm3 ; wk(10)=col5 174 movdqa XMMWORD [wk(11)], xmm4 ; wk(11)=col7 175 jmp near .column_end 176 alignx 16,7 177%endif 178.columnDCT: 179 180 ; -- Even part 181 182 movdqa xmm0, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] 183 movdqa xmm1, XMMWORD [XMMBLOCK(2,0,esi,SIZEOF_JCOEF)] 184 pmullw xmm0, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 185 pmullw xmm1, XMMWORD [XMMBLOCK(2,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 186 movdqa xmm2, XMMWORD [XMMBLOCK(4,0,esi,SIZEOF_JCOEF)] 187 movdqa xmm3, XMMWORD [XMMBLOCK(6,0,esi,SIZEOF_JCOEF)] 188 pmullw xmm2, XMMWORD [XMMBLOCK(4,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 189 pmullw xmm3, XMMWORD [XMMBLOCK(6,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 190 191 ; (Original) 192 ; z1 = (z2 + z3) * 0.541196100; 193 ; tmp2 = z1 + z3 * -1.847759065; 194 ; tmp3 = z1 + z2 * 0.765366865; 195 ; 196 ; (This implementation) 197 ; tmp2 = z2 * 0.541196100 + z3 * (0.541196100 - 1.847759065); 198 ; tmp3 = z2 * (0.541196100 + 0.765366865) + z3 * 0.541196100; 199 200 movdqa xmm4,xmm1 ; xmm1=in2=z2 201 movdqa xmm5,xmm1 202 punpcklwd xmm4,xmm3 ; xmm3=in6=z3 203 punpckhwd xmm5,xmm3 204 movdqa xmm1,xmm4 205 movdqa xmm3,xmm5 206 pmaddwd xmm4,[GOTOFF(ebx,PW_F130_F054)] ; xmm4=tmp3L 207 pmaddwd xmm5,[GOTOFF(ebx,PW_F130_F054)] ; xmm5=tmp3H 208 pmaddwd xmm1,[GOTOFF(ebx,PW_F054_MF130)] ; xmm1=tmp2L 209 pmaddwd xmm3,[GOTOFF(ebx,PW_F054_MF130)] ; xmm3=tmp2H 210 211 movdqa xmm6,xmm0 212 paddw xmm0,xmm2 ; xmm0=in0+in4 213 psubw xmm6,xmm2 ; xmm6=in0-in4 214 215 pxor xmm7,xmm7 216 pxor xmm2,xmm2 217 punpcklwd xmm7,xmm0 ; xmm7=tmp0L 218 punpckhwd xmm2,xmm0 ; xmm2=tmp0H 219 psrad xmm7,(16-CONST_BITS) ; psrad xmm7,16 & pslld xmm7,CONST_BITS 220 psrad xmm2,(16-CONST_BITS) ; psrad xmm2,16 & pslld xmm2,CONST_BITS 221 222 movdqa xmm0,xmm7 223 paddd xmm7,xmm4 ; xmm7=tmp10L 224 psubd xmm0,xmm4 ; xmm0=tmp13L 225 movdqa xmm4,xmm2 226 paddd xmm2,xmm5 ; xmm2=tmp10H 227 psubd xmm4,xmm5 ; xmm4=tmp13H 228 229 movdqa XMMWORD [wk(0)], xmm7 ; wk(0)=tmp10L 230 movdqa XMMWORD [wk(1)], xmm2 ; wk(1)=tmp10H 231 movdqa XMMWORD [wk(2)], xmm0 ; wk(2)=tmp13L 232 movdqa XMMWORD [wk(3)], xmm4 ; wk(3)=tmp13H 233 234 pxor xmm5,xmm5 235 pxor xmm7,xmm7 236 punpcklwd xmm5,xmm6 ; xmm5=tmp1L 237 punpckhwd xmm7,xmm6 ; xmm7=tmp1H 238 psrad xmm5,(16-CONST_BITS) ; psrad xmm5,16 & pslld xmm5,CONST_BITS 239 psrad xmm7,(16-CONST_BITS) ; psrad xmm7,16 & pslld xmm7,CONST_BITS 240 241 movdqa xmm2,xmm5 242 paddd xmm5,xmm1 ; xmm5=tmp11L 243 psubd xmm2,xmm1 ; xmm2=tmp12L 244 movdqa xmm0,xmm7 245 paddd xmm7,xmm3 ; xmm7=tmp11H 246 psubd xmm0,xmm3 ; xmm0=tmp12H 247 248 movdqa XMMWORD [wk(4)], xmm5 ; wk(4)=tmp11L 249 movdqa XMMWORD [wk(5)], xmm7 ; wk(5)=tmp11H 250 movdqa XMMWORD [wk(6)], xmm2 ; wk(6)=tmp12L 251 movdqa XMMWORD [wk(7)], xmm0 ; wk(7)=tmp12H 252 253 ; -- Odd part 254 255 movdqa xmm4, XMMWORD [XMMBLOCK(1,0,esi,SIZEOF_JCOEF)] 256 movdqa xmm6, XMMWORD [XMMBLOCK(3,0,esi,SIZEOF_JCOEF)] 257 pmullw xmm4, XMMWORD [XMMBLOCK(1,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 258 pmullw xmm6, XMMWORD [XMMBLOCK(3,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 259 movdqa xmm1, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] 260 movdqa xmm3, XMMWORD [XMMBLOCK(7,0,esi,SIZEOF_JCOEF)] 261 pmullw xmm1, XMMWORD [XMMBLOCK(5,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 262 pmullw xmm3, XMMWORD [XMMBLOCK(7,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 263 264 movdqa xmm5,xmm6 265 movdqa xmm7,xmm4 266 paddw xmm5,xmm3 ; xmm5=z3 267 paddw xmm7,xmm1 ; xmm7=z4 268 269 ; (Original) 270 ; z5 = (z3 + z4) * 1.175875602; 271 ; z3 = z3 * -1.961570560; z4 = z4 * -0.390180644; 272 ; z3 += z5; z4 += z5; 273 ; 274 ; (This implementation) 275 ; z3 = z3 * (1.175875602 - 1.961570560) + z4 * 1.175875602; 276 ; z4 = z3 * 1.175875602 + z4 * (1.175875602 - 0.390180644); 277 278 movdqa xmm2,xmm5 279 movdqa xmm0,xmm5 280 punpcklwd xmm2,xmm7 281 punpckhwd xmm0,xmm7 282 movdqa xmm5,xmm2 283 movdqa xmm7,xmm0 284 pmaddwd xmm2,[GOTOFF(ebx,PW_MF078_F117)] ; xmm2=z3L 285 pmaddwd xmm0,[GOTOFF(ebx,PW_MF078_F117)] ; xmm0=z3H 286 pmaddwd xmm5,[GOTOFF(ebx,PW_F117_F078)] ; xmm5=z4L 287 pmaddwd xmm7,[GOTOFF(ebx,PW_F117_F078)] ; xmm7=z4H 288 289 movdqa XMMWORD [wk(10)], xmm2 ; wk(10)=z3L 290 movdqa XMMWORD [wk(11)], xmm0 ; wk(11)=z3H 291 292 ; (Original) 293 ; z1 = tmp0 + tmp3; z2 = tmp1 + tmp2; 294 ; tmp0 = tmp0 * 0.298631336; tmp1 = tmp1 * 2.053119869; 295 ; tmp2 = tmp2 * 3.072711026; tmp3 = tmp3 * 1.501321110; 296 ; z1 = z1 * -0.899976223; z2 = z2 * -2.562915447; 297 ; tmp0 += z1 + z3; tmp1 += z2 + z4; 298 ; tmp2 += z2 + z3; tmp3 += z1 + z4; 299 ; 300 ; (This implementation) 301 ; tmp0 = tmp0 * (0.298631336 - 0.899976223) + tmp3 * -0.899976223; 302 ; tmp1 = tmp1 * (2.053119869 - 2.562915447) + tmp2 * -2.562915447; 303 ; tmp2 = tmp1 * -2.562915447 + tmp2 * (3.072711026 - 2.562915447); 304 ; tmp3 = tmp0 * -0.899976223 + tmp3 * (1.501321110 - 0.899976223); 305 ; tmp0 += z3; tmp1 += z4; 306 ; tmp2 += z3; tmp3 += z4; 307 308 movdqa xmm2,xmm3 309 movdqa xmm0,xmm3 310 punpcklwd xmm2,xmm4 311 punpckhwd xmm0,xmm4 312 movdqa xmm3,xmm2 313 movdqa xmm4,xmm0 314 pmaddwd xmm2,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm2=tmp0L 315 pmaddwd xmm0,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm0=tmp0H 316 pmaddwd xmm3,[GOTOFF(ebx,PW_MF089_F060)] ; xmm3=tmp3L 317 pmaddwd xmm4,[GOTOFF(ebx,PW_MF089_F060)] ; xmm4=tmp3H 318 319 paddd xmm2, XMMWORD [wk(10)] ; xmm2=tmp0L 320 paddd xmm0, XMMWORD [wk(11)] ; xmm0=tmp0H 321 paddd xmm3,xmm5 ; xmm3=tmp3L 322 paddd xmm4,xmm7 ; xmm4=tmp3H 323 324 movdqa XMMWORD [wk(8)], xmm2 ; wk(8)=tmp0L 325 movdqa XMMWORD [wk(9)], xmm0 ; wk(9)=tmp0H 326 327 movdqa xmm2,xmm1 328 movdqa xmm0,xmm1 329 punpcklwd xmm2,xmm6 330 punpckhwd xmm0,xmm6 331 movdqa xmm1,xmm2 332 movdqa xmm6,xmm0 333 pmaddwd xmm2,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm2=tmp1L 334 pmaddwd xmm0,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm0=tmp1H 335 pmaddwd xmm1,[GOTOFF(ebx,PW_MF256_F050)] ; xmm1=tmp2L 336 pmaddwd xmm6,[GOTOFF(ebx,PW_MF256_F050)] ; xmm6=tmp2H 337 338 paddd xmm2,xmm5 ; xmm2=tmp1L 339 paddd xmm0,xmm7 ; xmm0=tmp1H 340 paddd xmm1, XMMWORD [wk(10)] ; xmm1=tmp2L 341 paddd xmm6, XMMWORD [wk(11)] ; xmm6=tmp2H 342 343 movdqa XMMWORD [wk(10)], xmm2 ; wk(10)=tmp1L 344 movdqa XMMWORD [wk(11)], xmm0 ; wk(11)=tmp1H 345 346 ; -- Final output stage 347 348 movdqa xmm5, XMMWORD [wk(0)] ; xmm5=tmp10L 349 movdqa xmm7, XMMWORD [wk(1)] ; xmm7=tmp10H 350 351 movdqa xmm2,xmm5 352 movdqa xmm0,xmm7 353 paddd xmm5,xmm3 ; xmm5=data0L 354 paddd xmm7,xmm4 ; xmm7=data0H 355 psubd xmm2,xmm3 ; xmm2=data7L 356 psubd xmm0,xmm4 ; xmm0=data7H 357 358 movdqa xmm3,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm3=[PD_DESCALE_P1] 359 360 paddd xmm5,xmm3 361 paddd xmm7,xmm3 362 psrad xmm5,DESCALE_P1 363 psrad xmm7,DESCALE_P1 364 paddd xmm2,xmm3 365 paddd xmm0,xmm3 366 psrad xmm2,DESCALE_P1 367 psrad xmm0,DESCALE_P1 368 369 packssdw xmm5,xmm7 ; xmm5=data0=(00 01 02 03 04 05 06 07) 370 packssdw xmm2,xmm0 ; xmm2=data7=(70 71 72 73 74 75 76 77) 371 372 movdqa xmm4, XMMWORD [wk(4)] ; xmm4=tmp11L 373 movdqa xmm3, XMMWORD [wk(5)] ; xmm3=tmp11H 374 375 movdqa xmm7,xmm4 376 movdqa xmm0,xmm3 377 paddd xmm4,xmm1 ; xmm4=data1L 378 paddd xmm3,xmm6 ; xmm3=data1H 379 psubd xmm7,xmm1 ; xmm7=data6L 380 psubd xmm0,xmm6 ; xmm0=data6H 381 382 movdqa xmm1,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm1=[PD_DESCALE_P1] 383 384 paddd xmm4,xmm1 385 paddd xmm3,xmm1 386 psrad xmm4,DESCALE_P1 387 psrad xmm3,DESCALE_P1 388 paddd xmm7,xmm1 389 paddd xmm0,xmm1 390 psrad xmm7,DESCALE_P1 391 psrad xmm0,DESCALE_P1 392 393 packssdw xmm4,xmm3 ; xmm4=data1=(10 11 12 13 14 15 16 17) 394 packssdw xmm7,xmm0 ; xmm7=data6=(60 61 62 63 64 65 66 67) 395 396 movdqa xmm6,xmm5 ; transpose coefficients(phase 1) 397 punpcklwd xmm5,xmm4 ; xmm5=(00 10 01 11 02 12 03 13) 398 punpckhwd xmm6,xmm4 ; xmm6=(04 14 05 15 06 16 07 17) 399 movdqa xmm1,xmm7 ; transpose coefficients(phase 1) 400 punpcklwd xmm7,xmm2 ; xmm7=(60 70 61 71 62 72 63 73) 401 punpckhwd xmm1,xmm2 ; xmm1=(64 74 65 75 66 76 67 77) 402 403 movdqa xmm3, XMMWORD [wk(6)] ; xmm3=tmp12L 404 movdqa xmm0, XMMWORD [wk(7)] ; xmm0=tmp12H 405 movdqa xmm4, XMMWORD [wk(10)] ; xmm4=tmp1L 406 movdqa xmm2, XMMWORD [wk(11)] ; xmm2=tmp1H 407 408 movdqa XMMWORD [wk(0)], xmm5 ; wk(0)=(00 10 01 11 02 12 03 13) 409 movdqa XMMWORD [wk(1)], xmm6 ; wk(1)=(04 14 05 15 06 16 07 17) 410 movdqa XMMWORD [wk(4)], xmm7 ; wk(4)=(60 70 61 71 62 72 63 73) 411 movdqa XMMWORD [wk(5)], xmm1 ; wk(5)=(64 74 65 75 66 76 67 77) 412 413 movdqa xmm5,xmm3 414 movdqa xmm6,xmm0 415 paddd xmm3,xmm4 ; xmm3=data2L 416 paddd xmm0,xmm2 ; xmm0=data2H 417 psubd xmm5,xmm4 ; xmm5=data5L 418 psubd xmm6,xmm2 ; xmm6=data5H 419 420 movdqa xmm7,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm7=[PD_DESCALE_P1] 421 422 paddd xmm3,xmm7 423 paddd xmm0,xmm7 424 psrad xmm3,DESCALE_P1 425 psrad xmm0,DESCALE_P1 426 paddd xmm5,xmm7 427 paddd xmm6,xmm7 428 psrad xmm5,DESCALE_P1 429 psrad xmm6,DESCALE_P1 430 431 packssdw xmm3,xmm0 ; xmm3=data2=(20 21 22 23 24 25 26 27) 432 packssdw xmm5,xmm6 ; xmm5=data5=(50 51 52 53 54 55 56 57) 433 434 movdqa xmm1, XMMWORD [wk(2)] ; xmm1=tmp13L 435 movdqa xmm4, XMMWORD [wk(3)] ; xmm4=tmp13H 436 movdqa xmm2, XMMWORD [wk(8)] ; xmm2=tmp0L 437 movdqa xmm7, XMMWORD [wk(9)] ; xmm7=tmp0H 438 439 movdqa xmm0,xmm1 440 movdqa xmm6,xmm4 441 paddd xmm1,xmm2 ; xmm1=data3L 442 paddd xmm4,xmm7 ; xmm4=data3H 443 psubd xmm0,xmm2 ; xmm0=data4L 444 psubd xmm6,xmm7 ; xmm6=data4H 445 446 movdqa xmm2,[GOTOFF(ebx,PD_DESCALE_P1)] ; xmm2=[PD_DESCALE_P1] 447 448 paddd xmm1,xmm2 449 paddd xmm4,xmm2 450 psrad xmm1,DESCALE_P1 451 psrad xmm4,DESCALE_P1 452 paddd xmm0,xmm2 453 paddd xmm6,xmm2 454 psrad xmm0,DESCALE_P1 455 psrad xmm6,DESCALE_P1 456 457 packssdw xmm1,xmm4 ; xmm1=data3=(30 31 32 33 34 35 36 37) 458 packssdw xmm0,xmm6 ; xmm0=data4=(40 41 42 43 44 45 46 47) 459 460 movdqa xmm7, XMMWORD [wk(0)] ; xmm7=(00 10 01 11 02 12 03 13) 461 movdqa xmm2, XMMWORD [wk(1)] ; xmm2=(04 14 05 15 06 16 07 17) 462 463 movdqa xmm4,xmm3 ; transpose coefficients(phase 1) 464 punpcklwd xmm3,xmm1 ; xmm3=(20 30 21 31 22 32 23 33) 465 punpckhwd xmm4,xmm1 ; xmm4=(24 34 25 35 26 36 27 37) 466 movdqa xmm6,xmm0 ; transpose coefficients(phase 1) 467 punpcklwd xmm0,xmm5 ; xmm0=(40 50 41 51 42 52 43 53) 468 punpckhwd xmm6,xmm5 ; xmm6=(44 54 45 55 46 56 47 57) 469 470 movdqa xmm1,xmm7 ; transpose coefficients(phase 2) 471 punpckldq xmm7,xmm3 ; xmm7=(00 10 20 30 01 11 21 31) 472 punpckhdq xmm1,xmm3 ; xmm1=(02 12 22 32 03 13 23 33) 473 movdqa xmm5,xmm2 ; transpose coefficients(phase 2) 474 punpckldq xmm2,xmm4 ; xmm2=(04 14 24 34 05 15 25 35) 475 punpckhdq xmm5,xmm4 ; xmm5=(06 16 26 36 07 17 27 37) 476 477 movdqa xmm3, XMMWORD [wk(4)] ; xmm3=(60 70 61 71 62 72 63 73) 478 movdqa xmm4, XMMWORD [wk(5)] ; xmm4=(64 74 65 75 66 76 67 77) 479 480 movdqa XMMWORD [wk(6)], xmm2 ; wk(6)=(04 14 24 34 05 15 25 35) 481 movdqa XMMWORD [wk(7)], xmm5 ; wk(7)=(06 16 26 36 07 17 27 37) 482 483 movdqa xmm2,xmm0 ; transpose coefficients(phase 2) 484 punpckldq xmm0,xmm3 ; xmm0=(40 50 60 70 41 51 61 71) 485 punpckhdq xmm2,xmm3 ; xmm2=(42 52 62 72 43 53 63 73) 486 movdqa xmm5,xmm6 ; transpose coefficients(phase 2) 487 punpckldq xmm6,xmm4 ; xmm6=(44 54 64 74 45 55 65 75) 488 punpckhdq xmm5,xmm4 ; xmm5=(46 56 66 76 47 57 67 77) 489 490 movdqa xmm3,xmm7 ; transpose coefficients(phase 3) 491 punpcklqdq xmm7,xmm0 ; xmm7=col0=(00 10 20 30 40 50 60 70) 492 punpckhqdq xmm3,xmm0 ; xmm3=col1=(01 11 21 31 41 51 61 71) 493 movdqa xmm4,xmm1 ; transpose coefficients(phase 3) 494 punpcklqdq xmm1,xmm2 ; xmm1=col2=(02 12 22 32 42 52 62 72) 495 punpckhqdq xmm4,xmm2 ; xmm4=col3=(03 13 23 33 43 53 63 73) 496 497 movdqa xmm0, XMMWORD [wk(6)] ; xmm0=(04 14 24 34 05 15 25 35) 498 movdqa xmm2, XMMWORD [wk(7)] ; xmm2=(06 16 26 36 07 17 27 37) 499 500 movdqa XMMWORD [wk(8)], xmm3 ; wk(8)=col1 501 movdqa XMMWORD [wk(9)], xmm4 ; wk(9)=col3 502 503 movdqa xmm3,xmm0 ; transpose coefficients(phase 3) 504 punpcklqdq xmm0,xmm6 ; xmm0=col4=(04 14 24 34 44 54 64 74) 505 punpckhqdq xmm3,xmm6 ; xmm3=col5=(05 15 25 35 45 55 65 75) 506 movdqa xmm4,xmm2 ; transpose coefficients(phase 3) 507 punpcklqdq xmm2,xmm5 ; xmm2=col6=(06 16 26 36 46 56 66 76) 508 punpckhqdq xmm4,xmm5 ; xmm4=col7=(07 17 27 37 47 57 67 77) 509 510 movdqa XMMWORD [wk(10)], xmm3 ; wk(10)=col5 511 movdqa XMMWORD [wk(11)], xmm4 ; wk(11)=col7 512.column_end: 513 514 ; -- Prefetch the next coefficient block 515 516 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 0*32] 517 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 1*32] 518 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 2*32] 519 prefetchnta [esi + DCTSIZE2*SIZEOF_JCOEF + 3*32] 520 521 ; ---- Pass 2: process rows from work array, store into output array. 522 523 mov eax, [original_ebp] 524 mov edi, JSAMPARRAY [output_buf(eax)] ; (JSAMPROW *) 525 mov eax, JDIMENSION [output_col(eax)] 526 527 ; -- Even part 528 529 ; xmm7=col0, xmm1=col2, xmm0=col4, xmm2=col6 530 531 ; (Original) 532 ; z1 = (z2 + z3) * 0.541196100; 533 ; tmp2 = z1 + z3 * -1.847759065; 534 ; tmp3 = z1 + z2 * 0.765366865; 535 ; 536 ; (This implementation) 537 ; tmp2 = z2 * 0.541196100 + z3 * (0.541196100 - 1.847759065); 538 ; tmp3 = z2 * (0.541196100 + 0.765366865) + z3 * 0.541196100; 539 540 movdqa xmm6,xmm1 ; xmm1=in2=z2 541 movdqa xmm5,xmm1 542 punpcklwd xmm6,xmm2 ; xmm2=in6=z3 543 punpckhwd xmm5,xmm2 544 movdqa xmm1,xmm6 545 movdqa xmm2,xmm5 546 pmaddwd xmm6,[GOTOFF(ebx,PW_F130_F054)] ; xmm6=tmp3L 547 pmaddwd xmm5,[GOTOFF(ebx,PW_F130_F054)] ; xmm5=tmp3H 548 pmaddwd xmm1,[GOTOFF(ebx,PW_F054_MF130)] ; xmm1=tmp2L 549 pmaddwd xmm2,[GOTOFF(ebx,PW_F054_MF130)] ; xmm2=tmp2H 550 551 movdqa xmm3,xmm7 552 paddw xmm7,xmm0 ; xmm7=in0+in4 553 psubw xmm3,xmm0 ; xmm3=in0-in4 554 555 pxor xmm4,xmm4 556 pxor xmm0,xmm0 557 punpcklwd xmm4,xmm7 ; xmm4=tmp0L 558 punpckhwd xmm0,xmm7 ; xmm0=tmp0H 559 psrad xmm4,(16-CONST_BITS) ; psrad xmm4,16 & pslld xmm4,CONST_BITS 560 psrad xmm0,(16-CONST_BITS) ; psrad xmm0,16 & pslld xmm0,CONST_BITS 561 562 movdqa xmm7,xmm4 563 paddd xmm4,xmm6 ; xmm4=tmp10L 564 psubd xmm7,xmm6 ; xmm7=tmp13L 565 movdqa xmm6,xmm0 566 paddd xmm0,xmm5 ; xmm0=tmp10H 567 psubd xmm6,xmm5 ; xmm6=tmp13H 568 569 movdqa XMMWORD [wk(0)], xmm4 ; wk(0)=tmp10L 570 movdqa XMMWORD [wk(1)], xmm0 ; wk(1)=tmp10H 571 movdqa XMMWORD [wk(2)], xmm7 ; wk(2)=tmp13L 572 movdqa XMMWORD [wk(3)], xmm6 ; wk(3)=tmp13H 573 574 pxor xmm5,xmm5 575 pxor xmm4,xmm4 576 punpcklwd xmm5,xmm3 ; xmm5=tmp1L 577 punpckhwd xmm4,xmm3 ; xmm4=tmp1H 578 psrad xmm5,(16-CONST_BITS) ; psrad xmm5,16 & pslld xmm5,CONST_BITS 579 psrad xmm4,(16-CONST_BITS) ; psrad xmm4,16 & pslld xmm4,CONST_BITS 580 581 movdqa xmm0,xmm5 582 paddd xmm5,xmm1 ; xmm5=tmp11L 583 psubd xmm0,xmm1 ; xmm0=tmp12L 584 movdqa xmm7,xmm4 585 paddd xmm4,xmm2 ; xmm4=tmp11H 586 psubd xmm7,xmm2 ; xmm7=tmp12H 587 588 movdqa XMMWORD [wk(4)], xmm5 ; wk(4)=tmp11L 589 movdqa XMMWORD [wk(5)], xmm4 ; wk(5)=tmp11H 590 movdqa XMMWORD [wk(6)], xmm0 ; wk(6)=tmp12L 591 movdqa XMMWORD [wk(7)], xmm7 ; wk(7)=tmp12H 592 593 ; -- Odd part 594 595 movdqa xmm6, XMMWORD [wk(9)] ; xmm6=col3 596 movdqa xmm3, XMMWORD [wk(8)] ; xmm3=col1 597 movdqa xmm1, XMMWORD [wk(11)] ; xmm1=col7 598 movdqa xmm2, XMMWORD [wk(10)] ; xmm2=col5 599 600 movdqa xmm5,xmm6 601 movdqa xmm4,xmm3 602 paddw xmm5,xmm1 ; xmm5=z3 603 paddw xmm4,xmm2 ; xmm4=z4 604 605 ; (Original) 606 ; z5 = (z3 + z4) * 1.175875602; 607 ; z3 = z3 * -1.961570560; z4 = z4 * -0.390180644; 608 ; z3 += z5; z4 += z5; 609 ; 610 ; (This implementation) 611 ; z3 = z3 * (1.175875602 - 1.961570560) + z4 * 1.175875602; 612 ; z4 = z3 * 1.175875602 + z4 * (1.175875602 - 0.390180644); 613 614 movdqa xmm0,xmm5 615 movdqa xmm7,xmm5 616 punpcklwd xmm0,xmm4 617 punpckhwd xmm7,xmm4 618 movdqa xmm5,xmm0 619 movdqa xmm4,xmm7 620 pmaddwd xmm0,[GOTOFF(ebx,PW_MF078_F117)] ; xmm0=z3L 621 pmaddwd xmm7,[GOTOFF(ebx,PW_MF078_F117)] ; xmm7=z3H 622 pmaddwd xmm5,[GOTOFF(ebx,PW_F117_F078)] ; xmm5=z4L 623 pmaddwd xmm4,[GOTOFF(ebx,PW_F117_F078)] ; xmm4=z4H 624 625 movdqa XMMWORD [wk(10)], xmm0 ; wk(10)=z3L 626 movdqa XMMWORD [wk(11)], xmm7 ; wk(11)=z3H 627 628 ; (Original) 629 ; z1 = tmp0 + tmp3; z2 = tmp1 + tmp2; 630 ; tmp0 = tmp0 * 0.298631336; tmp1 = tmp1 * 2.053119869; 631 ; tmp2 = tmp2 * 3.072711026; tmp3 = tmp3 * 1.501321110; 632 ; z1 = z1 * -0.899976223; z2 = z2 * -2.562915447; 633 ; tmp0 += z1 + z3; tmp1 += z2 + z4; 634 ; tmp2 += z2 + z3; tmp3 += z1 + z4; 635 ; 636 ; (This implementation) 637 ; tmp0 = tmp0 * (0.298631336 - 0.899976223) + tmp3 * -0.899976223; 638 ; tmp1 = tmp1 * (2.053119869 - 2.562915447) + tmp2 * -2.562915447; 639 ; tmp2 = tmp1 * -2.562915447 + tmp2 * (3.072711026 - 2.562915447); 640 ; tmp3 = tmp0 * -0.899976223 + tmp3 * (1.501321110 - 0.899976223); 641 ; tmp0 += z3; tmp1 += z4; 642 ; tmp2 += z3; tmp3 += z4; 643 644 movdqa xmm0,xmm1 645 movdqa xmm7,xmm1 646 punpcklwd xmm0,xmm3 647 punpckhwd xmm7,xmm3 648 movdqa xmm1,xmm0 649 movdqa xmm3,xmm7 650 pmaddwd xmm0,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm0=tmp0L 651 pmaddwd xmm7,[GOTOFF(ebx,PW_MF060_MF089)] ; xmm7=tmp0H 652 pmaddwd xmm1,[GOTOFF(ebx,PW_MF089_F060)] ; xmm1=tmp3L 653 pmaddwd xmm3,[GOTOFF(ebx,PW_MF089_F060)] ; xmm3=tmp3H 654 655 paddd xmm0, XMMWORD [wk(10)] ; xmm0=tmp0L 656 paddd xmm7, XMMWORD [wk(11)] ; xmm7=tmp0H 657 paddd xmm1,xmm5 ; xmm1=tmp3L 658 paddd xmm3,xmm4 ; xmm3=tmp3H 659 660 movdqa XMMWORD [wk(8)], xmm0 ; wk(8)=tmp0L 661 movdqa XMMWORD [wk(9)], xmm7 ; wk(9)=tmp0H 662 663 movdqa xmm0,xmm2 664 movdqa xmm7,xmm2 665 punpcklwd xmm0,xmm6 666 punpckhwd xmm7,xmm6 667 movdqa xmm2,xmm0 668 movdqa xmm6,xmm7 669 pmaddwd xmm0,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm0=tmp1L 670 pmaddwd xmm7,[GOTOFF(ebx,PW_MF050_MF256)] ; xmm7=tmp1H 671 pmaddwd xmm2,[GOTOFF(ebx,PW_MF256_F050)] ; xmm2=tmp2L 672 pmaddwd xmm6,[GOTOFF(ebx,PW_MF256_F050)] ; xmm6=tmp2H 673 674 paddd xmm0,xmm5 ; xmm0=tmp1L 675 paddd xmm7,xmm4 ; xmm7=tmp1H 676 paddd xmm2, XMMWORD [wk(10)] ; xmm2=tmp2L 677 paddd xmm6, XMMWORD [wk(11)] ; xmm6=tmp2H 678 679 movdqa XMMWORD [wk(10)], xmm0 ; wk(10)=tmp1L 680 movdqa XMMWORD [wk(11)], xmm7 ; wk(11)=tmp1H 681 682 ; -- Final output stage 683 684 movdqa xmm5, XMMWORD [wk(0)] ; xmm5=tmp10L 685 movdqa xmm4, XMMWORD [wk(1)] ; xmm4=tmp10H 686 687 movdqa xmm0,xmm5 688 movdqa xmm7,xmm4 689 paddd xmm5,xmm1 ; xmm5=data0L 690 paddd xmm4,xmm3 ; xmm4=data0H 691 psubd xmm0,xmm1 ; xmm0=data7L 692 psubd xmm7,xmm3 ; xmm7=data7H 693 694 movdqa xmm1,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm1=[PD_DESCALE_P2] 695 696 paddd xmm5,xmm1 697 paddd xmm4,xmm1 698 psrad xmm5,DESCALE_P2 699 psrad xmm4,DESCALE_P2 700 paddd xmm0,xmm1 701 paddd xmm7,xmm1 702 psrad xmm0,DESCALE_P2 703 psrad xmm7,DESCALE_P2 704 705 packssdw xmm5,xmm4 ; xmm5=data0=(00 10 20 30 40 50 60 70) 706 packssdw xmm0,xmm7 ; xmm0=data7=(07 17 27 37 47 57 67 77) 707 708 movdqa xmm3, XMMWORD [wk(4)] ; xmm3=tmp11L 709 movdqa xmm1, XMMWORD [wk(5)] ; xmm1=tmp11H 710 711 movdqa xmm4,xmm3 712 movdqa xmm7,xmm1 713 paddd xmm3,xmm2 ; xmm3=data1L 714 paddd xmm1,xmm6 ; xmm1=data1H 715 psubd xmm4,xmm2 ; xmm4=data6L 716 psubd xmm7,xmm6 ; xmm7=data6H 717 718 movdqa xmm2,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm2=[PD_DESCALE_P2] 719 720 paddd xmm3,xmm2 721 paddd xmm1,xmm2 722 psrad xmm3,DESCALE_P2 723 psrad xmm1,DESCALE_P2 724 paddd xmm4,xmm2 725 paddd xmm7,xmm2 726 psrad xmm4,DESCALE_P2 727 psrad xmm7,DESCALE_P2 728 729 packssdw xmm3,xmm1 ; xmm3=data1=(01 11 21 31 41 51 61 71) 730 packssdw xmm4,xmm7 ; xmm4=data6=(06 16 26 36 46 56 66 76) 731 732 packsswb xmm5,xmm4 ; xmm5=(00 10 20 30 40 50 60 70 06 16 26 36 46 56 66 76) 733 packsswb xmm3,xmm0 ; xmm3=(01 11 21 31 41 51 61 71 07 17 27 37 47 57 67 77) 734 735 movdqa xmm6, XMMWORD [wk(6)] ; xmm6=tmp12L 736 movdqa xmm2, XMMWORD [wk(7)] ; xmm2=tmp12H 737 movdqa xmm1, XMMWORD [wk(10)] ; xmm1=tmp1L 738 movdqa xmm7, XMMWORD [wk(11)] ; xmm7=tmp1H 739 740 movdqa XMMWORD [wk(0)], xmm5 ; wk(0)=(00 10 20 30 40 50 60 70 06 16 26 36 46 56 66 76) 741 movdqa XMMWORD [wk(1)], xmm3 ; wk(1)=(01 11 21 31 41 51 61 71 07 17 27 37 47 57 67 77) 742 743 movdqa xmm4,xmm6 744 movdqa xmm0,xmm2 745 paddd xmm6,xmm1 ; xmm6=data2L 746 paddd xmm2,xmm7 ; xmm2=data2H 747 psubd xmm4,xmm1 ; xmm4=data5L 748 psubd xmm0,xmm7 ; xmm0=data5H 749 750 movdqa xmm5,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm5=[PD_DESCALE_P2] 751 752 paddd xmm6,xmm5 753 paddd xmm2,xmm5 754 psrad xmm6,DESCALE_P2 755 psrad xmm2,DESCALE_P2 756 paddd xmm4,xmm5 757 paddd xmm0,xmm5 758 psrad xmm4,DESCALE_P2 759 psrad xmm0,DESCALE_P2 760 761 packssdw xmm6,xmm2 ; xmm6=data2=(02 12 22 32 42 52 62 72) 762 packssdw xmm4,xmm0 ; xmm4=data5=(05 15 25 35 45 55 65 75) 763 764 movdqa xmm3, XMMWORD [wk(2)] ; xmm3=tmp13L 765 movdqa xmm1, XMMWORD [wk(3)] ; xmm1=tmp13H 766 movdqa xmm7, XMMWORD [wk(8)] ; xmm7=tmp0L 767 movdqa xmm5, XMMWORD [wk(9)] ; xmm5=tmp0H 768 769 movdqa xmm2,xmm3 770 movdqa xmm0,xmm1 771 paddd xmm3,xmm7 ; xmm3=data3L 772 paddd xmm1,xmm5 ; xmm1=data3H 773 psubd xmm2,xmm7 ; xmm2=data4L 774 psubd xmm0,xmm5 ; xmm0=data4H 775 776 movdqa xmm7,[GOTOFF(ebx,PD_DESCALE_P2)] ; xmm7=[PD_DESCALE_P2] 777 778 paddd xmm3,xmm7 779 paddd xmm1,xmm7 780 psrad xmm3,DESCALE_P2 781 psrad xmm1,DESCALE_P2 782 paddd xmm2,xmm7 783 paddd xmm0,xmm7 784 psrad xmm2,DESCALE_P2 785 psrad xmm0,DESCALE_P2 786 787 movdqa xmm5,[GOTOFF(ebx,PB_CENTERJSAMP)] ; xmm5=[PB_CENTERJSAMP] 788 789 packssdw xmm3,xmm1 ; xmm3=data3=(03 13 23 33 43 53 63 73) 790 packssdw xmm2,xmm0 ; xmm2=data4=(04 14 24 34 44 54 64 74) 791 792 movdqa xmm7, XMMWORD [wk(0)] ; xmm7=(00 10 20 30 40 50 60 70 06 16 26 36 46 56 66 76) 793 movdqa xmm1, XMMWORD [wk(1)] ; xmm1=(01 11 21 31 41 51 61 71 07 17 27 37 47 57 67 77) 794 795 packsswb xmm6,xmm2 ; xmm6=(02 12 22 32 42 52 62 72 04 14 24 34 44 54 64 74) 796 packsswb xmm3,xmm4 ; xmm3=(03 13 23 33 43 53 63 73 05 15 25 35 45 55 65 75) 797 798 paddb xmm7,xmm5 799 paddb xmm1,xmm5 800 paddb xmm6,xmm5 801 paddb xmm3,xmm5 802 803 movdqa xmm0,xmm7 ; transpose coefficients(phase 1) 804 punpcklbw xmm7,xmm1 ; xmm7=(00 01 10 11 20 21 30 31 40 41 50 51 60 61 70 71) 805 punpckhbw xmm0,xmm1 ; xmm0=(06 07 16 17 26 27 36 37 46 47 56 57 66 67 76 77) 806 movdqa xmm2,xmm6 ; transpose coefficients(phase 1) 807 punpcklbw xmm6,xmm3 ; xmm6=(02 03 12 13 22 23 32 33 42 43 52 53 62 63 72 73) 808 punpckhbw xmm2,xmm3 ; xmm2=(04 05 14 15 24 25 34 35 44 45 54 55 64 65 74 75) 809 810 movdqa xmm4,xmm7 ; transpose coefficients(phase 2) 811 punpcklwd xmm7,xmm6 ; xmm7=(00 01 02 03 10 11 12 13 20 21 22 23 30 31 32 33) 812 punpckhwd xmm4,xmm6 ; xmm4=(40 41 42 43 50 51 52 53 60 61 62 63 70 71 72 73) 813 movdqa xmm5,xmm2 ; transpose coefficients(phase 2) 814 punpcklwd xmm2,xmm0 ; xmm2=(04 05 06 07 14 15 16 17 24 25 26 27 34 35 36 37) 815 punpckhwd xmm5,xmm0 ; xmm5=(44 45 46 47 54 55 56 57 64 65 66 67 74 75 76 77) 816 817 movdqa xmm1,xmm7 ; transpose coefficients(phase 3) 818 punpckldq xmm7,xmm2 ; xmm7=(00 01 02 03 04 05 06 07 10 11 12 13 14 15 16 17) 819 punpckhdq xmm1,xmm2 ; xmm1=(20 21 22 23 24 25 26 27 30 31 32 33 34 35 36 37) 820 movdqa xmm3,xmm4 ; transpose coefficients(phase 3) 821 punpckldq xmm4,xmm5 ; xmm4=(40 41 42 43 44 45 46 47 50 51 52 53 54 55 56 57) 822 punpckhdq xmm3,xmm5 ; xmm3=(60 61 62 63 64 65 66 67 70 71 72 73 74 75 76 77) 823 824 pshufd xmm6,xmm7,0x4E ; xmm6=(10 11 12 13 14 15 16 17 00 01 02 03 04 05 06 07) 825 pshufd xmm0,xmm1,0x4E ; xmm0=(30 31 32 33 34 35 36 37 20 21 22 23 24 25 26 27) 826 pshufd xmm2,xmm4,0x4E ; xmm2=(50 51 52 53 54 55 56 57 40 41 42 43 44 45 46 47) 827 pshufd xmm5,xmm3,0x4E ; xmm5=(70 71 72 73 74 75 76 77 60 61 62 63 64 65 66 67) 828 829 mov edx, JSAMPROW [edi+0*SIZEOF_JSAMPROW] 830 mov esi, JSAMPROW [edi+2*SIZEOF_JSAMPROW] 831 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm7 832 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm1 833 mov edx, JSAMPROW [edi+4*SIZEOF_JSAMPROW] 834 mov esi, JSAMPROW [edi+6*SIZEOF_JSAMPROW] 835 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm4 836 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm3 837 838 mov edx, JSAMPROW [edi+1*SIZEOF_JSAMPROW] 839 mov esi, JSAMPROW [edi+3*SIZEOF_JSAMPROW] 840 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm6 841 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm0 842 mov edx, JSAMPROW [edi+5*SIZEOF_JSAMPROW] 843 mov esi, JSAMPROW [edi+7*SIZEOF_JSAMPROW] 844 movq XMM_MMWORD [edx+eax*SIZEOF_JSAMPLE], xmm2 845 movq XMM_MMWORD [esi+eax*SIZEOF_JSAMPLE], xmm5 846 847 pop edi 848 pop esi 849; pop edx ; need not be preserved 850; pop ecx ; unused 851 poppic ebx 852 mov esp,ebp ; esp <- aligned ebp 853 pop esp ; esp <- original ebp 854 pop ebp 855 ret 856 857; For some reason, the OS X linker does not honor the request to align the 858; segment unless we do this. 859 align 16 860