aboutsummaryrefslogtreecommitdiffhomepage
path: root/src/jumper/SkJumper_generated.S
diff options
context:
space:
mode:
Diffstat (limited to 'src/jumper/SkJumper_generated.S')
-rw-r--r--src/jumper/SkJumper_generated.S6918
1 files changed, 3151 insertions, 3767 deletions
diff --git a/src/jumper/SkJumper_generated.S b/src/jumper/SkJumper_generated.S
index 9f3ec4a9d3..07f0b75c8d 100644
--- a/src/jumper/SkJumper_generated.S
+++ b/src/jumper/SkJumper_generated.S
@@ -9576,9 +9576,9 @@ _sk_seed_shader_skx:
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
.byte 98,242,125,40,124,201 // vpbroadcastd %ecx,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 98,241,116,56,88,13,50,150,3,0 // vaddps 0x39632(%rip){1to8},%ymm1,%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,241,116,56,88,13,174,136,3,0 // vaddps 0x388ae(%rip){1to8},%ymm1,%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,21,43,150,3,0 // vbroadcastss 0x3962b(%rip),%ymm2 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,21,167,136,3,0 // vbroadcastss 0x388a7(%rip),%ymm2 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
.byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
@@ -9592,16 +9592,16 @@ FUNCTION(_sk_dither_skx)
_sk_dither_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 98,114,125,40,124,194 // vpbroadcastd %edx,%ymm8
- .byte 197,61,254,5,173,151,3,0 // vpaddd 0x397ad(%rip),%ymm8,%ymm8 # 398e0 <_sk_srcover_bgra_8888_sse2_lowp+0x458>
+ .byte 197,61,254,5,45,138,3,0 // vpaddd 0x38a2d(%rip),%ymm8,%ymm8 # 38b60 <_sk_srcover_bgra_8888_sse2_lowp+0x45c>
.byte 98,114,125,40,124,201 // vpbroadcastd %ecx,%ymm9
.byte 196,65,53,239,200 // vpxor %ymm8,%ymm9,%ymm9
- .byte 196,98,125,88,21,245,149,3,0 // vpbroadcastd 0x395f5(%rip),%ymm10 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,98,125,88,21,113,136,3,0 // vpbroadcastd 0x38871(%rip),%ymm10 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 196,65,53,219,218 // vpand %ymm10,%ymm9,%ymm11
.byte 196,193,37,114,243,5 // vpslld $0x5,%ymm11,%ymm11
.byte 196,65,61,219,210 // vpand %ymm10,%ymm8,%ymm10
.byte 196,193,45,114,242,4 // vpslld $0x4,%ymm10,%ymm10
- .byte 196,98,125,88,37,218,149,3,0 // vpbroadcastd 0x395da(%rip),%ymm12 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
- .byte 196,98,125,88,45,213,149,3,0 // vpbroadcastd 0x395d5(%rip),%ymm13 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,125,88,37,86,136,3,0 // vpbroadcastd 0x38856(%rip),%ymm12 # 389bc <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
+ .byte 196,98,125,88,45,81,136,3,0 // vpbroadcastd 0x38851(%rip),%ymm13 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,65,53,219,245 // vpand %ymm13,%ymm9,%ymm14
.byte 196,193,13,114,246,2 // vpslld $0x2,%ymm14,%ymm14
.byte 196,65,37,235,222 // vpor %ymm14,%ymm11,%ymm11
@@ -9616,8 +9616,8 @@ _sk_dither_skx:
.byte 196,65,61,235,195 // vpor %ymm11,%ymm8,%ymm8
.byte 196,65,61,235,193 // vpor %ymm9,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,135,149,3,0 // vbroadcastss 0x39587(%rip),%ymm9 # 39748 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
- .byte 98,114,61,56,168,13,129,149,3,0 // vfmadd213ps 0x39581(%rip){1to8},%ymm8,%ymm9 # 3974c <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
+ .byte 196,98,125,24,13,3,136,3,0 // vbroadcastss 0x38803(%rip),%ymm9 # 389c4 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
+ .byte 98,114,61,56,168,13,253,135,3,0 // vfmadd213ps 0x387fd(%rip){1to8},%ymm8,%ymm9 # 389c8 <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
.byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
.byte 197,188,88,192 // vaddps %ymm0,%ymm8,%ymm0
@@ -9650,7 +9650,7 @@ HIDDEN _sk_black_color_skx
FUNCTION(_sk_black_color_skx)
_sk_black_color_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,14,149,3,0 // vbroadcastss 0x3950e(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,138,135,3,0 // vbroadcastss 0x3878a(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
@@ -9661,7 +9661,7 @@ HIDDEN _sk_white_color_skx
FUNCTION(_sk_white_color_skx)
_sk_white_color_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,245,148,3,0 // vbroadcastss 0x394f5(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,113,135,3,0 // vbroadcastss 0x38771(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 197,252,40,216 // vmovaps %ymm0,%ymm3
@@ -9707,7 +9707,7 @@ HIDDEN _sk_srcatop_skx
FUNCTION(_sk_srcatop_skx)
_sk_srcatop_skx:
.byte 197,252,89,199 // vmulps %ymm7,%ymm0,%ymm0
- .byte 196,98,125,24,5,148,148,3,0 // vbroadcastss 0x39494(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,16,135,3,0 // vbroadcastss 0x38710(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,226,61,184,196 // vfmadd231ps %ymm4,%ymm8,%ymm0
.byte 197,244,89,207 // vmulps %ymm7,%ymm1,%ymm1
@@ -9723,7 +9723,7 @@ HIDDEN _sk_dstatop_skx
.globl _sk_dstatop_skx
FUNCTION(_sk_dstatop_skx)
_sk_dstatop_skx:
- .byte 196,98,125,24,5,99,148,3,0 // vbroadcastss 0x39463(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,223,134,3,0 // vbroadcastss 0x386df(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,226,101,184,196 // vfmadd231ps %ymm4,%ymm3,%ymm0
@@ -9762,7 +9762,7 @@ HIDDEN _sk_srcout_skx
.globl _sk_srcout_skx
FUNCTION(_sk_srcout_skx)
_sk_srcout_skx:
- .byte 196,98,125,24,5,6,148,3,0 // vbroadcastss 0x39406(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,130,134,3,0 // vbroadcastss 0x38682(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
@@ -9775,7 +9775,7 @@ HIDDEN _sk_dstout_skx
.globl _sk_dstout_skx
FUNCTION(_sk_dstout_skx)
_sk_dstout_skx:
- .byte 196,226,125,24,5,229,147,3,0 // vbroadcastss 0x393e5(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,97,134,3,0 // vbroadcastss 0x38661(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,92,219 // vsubps %ymm3,%ymm0,%ymm3
.byte 197,228,89,196 // vmulps %ymm4,%ymm3,%ymm0
.byte 197,228,89,205 // vmulps %ymm5,%ymm3,%ymm1
@@ -9788,7 +9788,7 @@ HIDDEN _sk_srcover_skx
.globl _sk_srcover_skx
FUNCTION(_sk_srcover_skx)
_sk_srcover_skx:
- .byte 196,98,125,24,5,196,147,3,0 // vbroadcastss 0x393c4(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,64,134,3,0 // vbroadcastss 0x38640(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,194,85,184,200 // vfmadd231ps %ymm8,%ymm5,%ymm1
@@ -9801,7 +9801,7 @@ HIDDEN _sk_dstover_skx
.globl _sk_dstover_skx
FUNCTION(_sk_dstover_skx)
_sk_dstover_skx:
- .byte 196,98,125,24,5,159,147,3,0 // vbroadcastss 0x3939f(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,27,134,3,0 // vbroadcastss 0x3861b(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
.byte 196,226,61,168,205 // vfmadd213ps %ymm5,%ymm8,%ymm1
@@ -9825,7 +9825,7 @@ HIDDEN _sk_multiply_skx
.globl _sk_multiply_skx
FUNCTION(_sk_multiply_skx)
_sk_multiply_skx:
- .byte 196,98,125,24,5,102,147,3,0 // vbroadcastss 0x39366(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,226,133,3,0 // vbroadcastss 0x385e2(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,208 // vmulps %ymm0,%ymm9,%ymm10
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -9848,7 +9848,7 @@ HIDDEN _sk_plus__skx
FUNCTION(_sk_plus__skx)
_sk_plus__skx:
.byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
- .byte 196,98,125,24,5,21,147,3,0 // vbroadcastss 0x39315(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,145,133,3,0 // vbroadcastss 0x38591(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
@@ -9878,7 +9878,7 @@ HIDDEN _sk_xor__skx
.globl _sk_xor__skx
FUNCTION(_sk_xor__skx)
_sk_xor__skx:
- .byte 196,98,125,24,5,192,146,3,0 // vbroadcastss 0x392c0(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,60,133,3,0 // vbroadcastss 0x3853c(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -9912,7 +9912,7 @@ _sk_darken_skx:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,95,209 // vmaxps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,68,146,3,0 // vbroadcastss 0x39244(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,192,132,3,0 // vbroadcastss 0x384c0(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -9937,7 +9937,7 @@ _sk_lighten_skx:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,239,145,3,0 // vbroadcastss 0x391ef(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,107,132,3,0 // vbroadcastss 0x3846b(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -9965,7 +9965,7 @@ _sk_difference_skx:
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,142,145,3,0 // vbroadcastss 0x3918e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,10,132,3,0 // vbroadcastss 0x3840a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -9987,7 +9987,7 @@ _sk_exclusion_skx:
.byte 197,236,89,214 // vmulps %ymm6,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,72,145,3,0 // vbroadcastss 0x39148(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,196,131,3,0 // vbroadcastss 0x383c4(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -9999,7 +9999,7 @@ FUNCTION(_sk_colorburn_skx)
_sk_colorburn_skx:
.byte 98,241,92,40,194,199,0 // vcmpeqps %ymm7,%ymm4,%k0
.byte 98,114,126,40,56,192 // vpmovm2d %k0,%ymm8
- .byte 196,98,125,24,13,37,145,3,0 // vbroadcastss 0x39125(%rip),%ymm9 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,13,161,131,3,0 // vbroadcastss 0x383a1(%rip),%ymm9 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,52,92,215 // vsubps %ymm7,%ymm9,%ymm10
.byte 197,44,89,216 // vmulps %ymm0,%ymm10,%ymm11
.byte 197,36,88,228 // vaddps %ymm4,%ymm11,%ymm12
@@ -10063,7 +10063,7 @@ _sk_colordodge_skx:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 98,209,92,40,194,192,0 // vcmpeqps %ymm8,%ymm4,%k0
.byte 98,114,126,40,56,200 // vpmovm2d %k0,%ymm9
- .byte 196,98,125,24,21,246,143,3,0 // vbroadcastss 0x38ff6(%rip),%ymm10 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,21,114,130,3,0 // vbroadcastss 0x38272(%rip),%ymm10 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,44,92,223 // vsubps %ymm7,%ymm10,%ymm11
.byte 197,36,89,224 // vmulps %ymm0,%ymm11,%ymm12
.byte 98,241,124,40,194,195,0 // vcmpeqps %ymm3,%ymm0,%k0
@@ -10120,7 +10120,7 @@ HIDDEN _sk_hardlight_skx
.globl _sk_hardlight_skx
FUNCTION(_sk_hardlight_skx)
_sk_hardlight_skx:
- .byte 196,98,125,24,5,238,142,3,0 // vbroadcastss 0x38eee(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,106,129,3,0 // vbroadcastss 0x3816a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,208 // vmulps %ymm0,%ymm9,%ymm10
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -10174,7 +10174,7 @@ HIDDEN _sk_overlay_skx
.globl _sk_overlay_skx
FUNCTION(_sk_overlay_skx)
_sk_overlay_skx:
- .byte 196,98,125,24,5,10,142,3,0 // vbroadcastss 0x38e0a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,134,128,3,0 // vbroadcastss 0x38086(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,208 // vmulps %ymm0,%ymm9,%ymm10
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -10240,15 +10240,15 @@ _sk_softlight_skx:
.byte 196,65,52,88,217 // vaddps %ymm9,%ymm9,%ymm11
.byte 196,65,36,88,235 // vaddps %ymm11,%ymm11,%ymm13
.byte 197,4,92,243 // vsubps %ymm3,%ymm15,%ymm14
- .byte 196,98,125,24,29,232,140,3,0 // vbroadcastss 0x38ce8(%rip),%ymm11 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,29,100,127,3,0 // vbroadcastss 0x37f64(%rip),%ymm11 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,65,36,92,225 // vsubps %ymm9,%ymm11,%ymm12
.byte 196,98,13,168,227 // vfmadd213ps %ymm3,%ymm14,%ymm12
.byte 197,156,89,212 // vmulps %ymm4,%ymm12,%ymm2
.byte 196,66,21,168,237 // vfmadd213ps %ymm13,%ymm13,%ymm13
- .byte 98,226,125,40,24,21,227,140,3,0 // vbroadcastss 0x38ce3(%rip),%ymm18 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 98,226,125,40,24,21,95,127,3,0 // vbroadcastss 0x37f5f(%rip),%ymm18 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 98,161,52,40,88,202 // vaddps %ymm18,%ymm9,%ymm17
.byte 98,81,116,32,89,229 // vmulps %ymm13,%ymm17,%ymm12
- .byte 98,226,125,40,24,29,209,140,3,0 // vbroadcastss 0x38cd1(%rip),%ymm19 # 39754 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
+ .byte 98,226,125,40,24,29,77,127,3,0 // vbroadcastss 0x37f4d(%rip),%ymm19 # 389d0 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
.byte 98,50,53,40,184,227 // vfmadd231ps %ymm19,%ymm9,%ymm12
.byte 98,194,125,40,78,201 // vrsqrt14ps %ymm9,%ymm17
.byte 98,162,125,40,76,201 // vrcp14ps %ymm17,%ymm17
@@ -10370,11 +10370,11 @@ _sk_hue_skx:
.byte 196,65,28,89,219 // vmulps %ymm11,%ymm12,%ymm11
.byte 196,65,36,94,222 // vdivps %ymm14,%ymm11,%ymm11
.byte 196,99,37,74,225,240 // vblendvps %ymm15,%ymm1,%ymm11,%ymm12
- .byte 196,98,125,24,45,108,138,3,0 // vbroadcastss 0x38a6c(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,53,103,138,3,0 // vbroadcastss 0x38a67(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,45,232,124,3,0 // vbroadcastss 0x37ce8(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,227,124,3,0 // vbroadcastss 0x37ce3(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,254 // vmulps %ymm14,%ymm5,%ymm15
.byte 196,66,93,184,253 // vfmadd231ps %ymm13,%ymm4,%ymm15
- .byte 98,226,125,40,24,5,87,138,3,0 // vbroadcastss 0x38a57(%rip),%ymm16 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 98,226,125,40,24,5,211,124,3,0 // vbroadcastss 0x37cd3(%rip),%ymm16 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 98,50,77,40,184,248 // vfmadd231ps %ymm16,%ymm6,%ymm15
.byte 196,65,44,89,222 // vmulps %ymm14,%ymm10,%ymm11
.byte 196,66,53,184,221 // vfmadd231ps %ymm13,%ymm9,%ymm11
@@ -10431,7 +10431,7 @@ _sk_hue_skx:
.byte 196,65,12,88,219 // vaddps %ymm11,%ymm14,%ymm11
.byte 196,67,45,74,203,144 // vblendvps %ymm9,%ymm11,%ymm10,%ymm9
.byte 197,52,95,201 // vmaxps %ymm1,%ymm9,%ymm9
- .byte 196,226,125,24,13,243,136,3,0 // vbroadcastss 0x388f3(%rip),%ymm1 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,13,111,123,3,0 // vbroadcastss 0x37b6f(%rip),%ymm1 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,116,92,215 // vsubps %ymm7,%ymm1,%ymm10
.byte 197,172,89,192 // vmulps %ymm0,%ymm10,%ymm0
.byte 197,116,92,219 // vsubps %ymm3,%ymm1,%ymm11
@@ -10483,11 +10483,11 @@ _sk_saturation_skx:
.byte 196,65,28,89,219 // vmulps %ymm11,%ymm12,%ymm11
.byte 196,65,36,94,222 // vdivps %ymm14,%ymm11,%ymm11
.byte 196,99,37,74,225,240 // vblendvps %ymm15,%ymm1,%ymm11,%ymm12
- .byte 196,98,125,24,45,40,136,3,0 // vbroadcastss 0x38828(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,53,35,136,3,0 // vbroadcastss 0x38823(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,45,164,122,3,0 // vbroadcastss 0x37aa4(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,159,122,3,0 // vbroadcastss 0x37a9f(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,254 // vmulps %ymm14,%ymm5,%ymm15
.byte 196,66,93,184,253 // vfmadd231ps %ymm13,%ymm4,%ymm15
- .byte 98,226,125,40,24,5,19,136,3,0 // vbroadcastss 0x38813(%rip),%ymm16 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 98,226,125,40,24,5,143,122,3,0 // vbroadcastss 0x37a8f(%rip),%ymm16 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 98,50,77,40,184,248 // vfmadd231ps %ymm16,%ymm6,%ymm15
.byte 196,65,44,89,222 // vmulps %ymm14,%ymm10,%ymm11
.byte 196,66,53,184,221 // vfmadd231ps %ymm13,%ymm9,%ymm11
@@ -10544,7 +10544,7 @@ _sk_saturation_skx:
.byte 196,65,12,88,219 // vaddps %ymm11,%ymm14,%ymm11
.byte 196,67,45,74,203,144 // vblendvps %ymm9,%ymm11,%ymm10,%ymm9
.byte 197,52,95,201 // vmaxps %ymm1,%ymm9,%ymm9
- .byte 196,226,125,24,13,175,134,3,0 // vbroadcastss 0x386af(%rip),%ymm1 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,13,43,121,3,0 // vbroadcastss 0x3792b(%rip),%ymm1 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,116,92,215 // vsubps %ymm7,%ymm1,%ymm10
.byte 197,172,89,192 // vmulps %ymm0,%ymm10,%ymm0
.byte 197,116,92,219 // vsubps %ymm3,%ymm1,%ymm11
@@ -10568,11 +10568,11 @@ _sk_color_skx:
.byte 197,124,89,199 // vmulps %ymm7,%ymm0,%ymm8
.byte 197,116,89,207 // vmulps %ymm7,%ymm1,%ymm9
.byte 197,108,89,223 // vmulps %ymm7,%ymm2,%ymm11
- .byte 196,98,125,24,37,120,134,3,0 // vbroadcastss 0x38678(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,45,115,134,3,0 // vbroadcastss 0x38673(%rip),%ymm13 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,37,244,120,3,0 // vbroadcastss 0x378f4(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,45,239,120,3,0 // vbroadcastss 0x378ef(%rip),%ymm13 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,245 // vmulps %ymm13,%ymm5,%ymm14
.byte 196,66,93,184,244 // vfmadd231ps %ymm12,%ymm4,%ymm14
- .byte 196,98,125,24,61,100,134,3,0 // vbroadcastss 0x38664(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,224,120,3,0 // vbroadcastss 0x378e0(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,66,77,184,247 // vfmadd231ps %ymm15,%ymm6,%ymm14
.byte 196,65,52,89,213 // vmulps %ymm13,%ymm9,%ymm10
.byte 196,66,61,184,212 // vfmadd231ps %ymm12,%ymm8,%ymm10
@@ -10630,7 +10630,7 @@ _sk_color_skx:
.byte 196,65,20,88,219 // vaddps %ymm11,%ymm13,%ymm11
.byte 196,67,53,74,203,192 // vblendvps %ymm12,%ymm11,%ymm9,%ymm9
.byte 98,49,52,40,95,202 // vmaxps %ymm18,%ymm9,%ymm9
- .byte 196,98,125,24,29,245,132,3,0 // vbroadcastss 0x384f5(%rip),%ymm11 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,29,113,119,3,0 // vbroadcastss 0x37771(%rip),%ymm11 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,36,92,231 // vsubps %ymm7,%ymm11,%ymm12
.byte 197,156,89,192 // vmulps %ymm0,%ymm12,%ymm0
.byte 197,36,92,219 // vsubps %ymm3,%ymm11,%ymm11
@@ -10654,11 +10654,11 @@ _sk_luminosity_skx:
.byte 197,100,89,196 // vmulps %ymm4,%ymm3,%ymm8
.byte 197,100,89,205 // vmulps %ymm5,%ymm3,%ymm9
.byte 197,100,89,222 // vmulps %ymm6,%ymm3,%ymm11
- .byte 196,98,125,24,37,192,132,3,0 // vbroadcastss 0x384c0(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,45,187,132,3,0 // vbroadcastss 0x384bb(%rip),%ymm13 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,37,60,119,3,0 // vbroadcastss 0x3773c(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,45,55,119,3,0 // vbroadcastss 0x37737(%rip),%ymm13 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,116,89,245 // vmulps %ymm13,%ymm1,%ymm14
.byte 196,66,125,184,244 // vfmadd231ps %ymm12,%ymm0,%ymm14
- .byte 196,98,125,24,61,172,132,3,0 // vbroadcastss 0x384ac(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,40,119,3,0 // vbroadcastss 0x37728(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,66,109,184,247 // vfmadd231ps %ymm15,%ymm2,%ymm14
.byte 196,65,52,89,213 // vmulps %ymm13,%ymm9,%ymm10
.byte 196,66,61,184,212 // vfmadd231ps %ymm12,%ymm8,%ymm10
@@ -10716,7 +10716,7 @@ _sk_luminosity_skx:
.byte 196,65,20,88,219 // vaddps %ymm11,%ymm13,%ymm11
.byte 196,67,53,74,203,192 // vblendvps %ymm12,%ymm11,%ymm9,%ymm9
.byte 98,49,52,40,95,202 // vmaxps %ymm18,%ymm9,%ymm9
- .byte 196,98,125,24,29,61,131,3,0 // vbroadcastss 0x3833d(%rip),%ymm11 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,29,185,117,3,0 // vbroadcastss 0x375b9(%rip),%ymm11 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,36,92,231 // vsubps %ymm7,%ymm11,%ymm12
.byte 197,156,89,192 // vmulps %ymm0,%ymm12,%ymm0
.byte 197,36,92,219 // vsubps %ymm3,%ymm11,%ymm11
@@ -10759,9 +10759,9 @@ _sk_srcover_rgba_8888_skx:
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,148,130,3,0 // vbroadcastss 0x38294(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,16,117,3,0 // vbroadcastss 0x37510(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,179,130,3,0 // vbroadcastss 0x382b3(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,47,117,3,0 // vbroadcastss 0x3752f(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
@@ -10903,9 +10903,9 @@ _sk_srcover_bgra_8888_skx:
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,87,128,3,0 // vbroadcastss 0x38057(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,211,114,3,0 // vbroadcastss 0x372d3(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,118,128,3,0 // vbroadcastss 0x38076(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,242,114,3,0 // vbroadcastss 0x372f2(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
@@ -11034,7 +11034,7 @@ HIDDEN _sk_clamp_1_skx
.globl _sk_clamp_1_skx
FUNCTION(_sk_clamp_1_skx)
_sk_clamp_1_skx:
- .byte 196,98,125,24,5,90,126,3,0 // vbroadcastss 0x37e5a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,214,112,3,0 // vbroadcastss 0x370d6(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
@@ -11046,7 +11046,7 @@ HIDDEN _sk_clamp_a_skx
.globl _sk_clamp_a_skx
FUNCTION(_sk_clamp_a_skx)
_sk_clamp_a_skx:
- .byte 98,241,100,56,93,29,56,126,3,0 // vminps 0x37e38(%rip){1to8},%ymm3,%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,100,56,93,29,180,112,3,0 // vminps 0x370b4(%rip){1to8},%ymm3,%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,93,195 // vminps %ymm3,%ymm0,%ymm0
.byte 197,244,93,203 // vminps %ymm3,%ymm1,%ymm1
.byte 197,236,93,211 // vminps %ymm3,%ymm2,%ymm2
@@ -11057,7 +11057,7 @@ HIDDEN _sk_clamp_a_dst_skx
.globl _sk_clamp_a_dst_skx
FUNCTION(_sk_clamp_a_dst_skx)
_sk_clamp_a_dst_skx:
- .byte 98,241,68,56,93,61,30,126,3,0 // vminps 0x37e1e(%rip){1to8},%ymm7,%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,68,56,93,61,154,112,3,0 // vminps 0x3709a(%rip){1to8},%ymm7,%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,220,93,231 // vminps %ymm7,%ymm4,%ymm4
.byte 197,212,93,239 // vminps %ymm7,%ymm5,%ymm5
.byte 197,204,93,247 // vminps %ymm7,%ymm6,%ymm6
@@ -11089,7 +11089,7 @@ HIDDEN _sk_invert_skx
.globl _sk_invert_skx
FUNCTION(_sk_invert_skx)
_sk_invert_skx:
- .byte 196,98,125,24,5,222,125,3,0 // vbroadcastss 0x37dde(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,90,112,3,0 // vbroadcastss 0x3705a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,188,92,192 // vsubps %ymm0,%ymm8,%ymm0
.byte 197,188,92,201 // vsubps %ymm1,%ymm8,%ymm1
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
@@ -11143,9 +11143,9 @@ HIDDEN _sk_unpremul_skx
.globl _sk_unpremul_skx
FUNCTION(_sk_unpremul_skx)
_sk_unpremul_skx:
- .byte 196,98,125,24,5,121,125,3,0 // vbroadcastss 0x37d79(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,245,111,3,0 // vbroadcastss 0x36ff5(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,94,195 // vdivps %ymm3,%ymm8,%ymm8
- .byte 98,241,60,56,194,5,154,125,3,0,1 // vcmpltps 0x37d9a(%rip){1to8},%ymm8,%k0 # 39768 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
+ .byte 98,241,60,56,194,5,22,112,3,0,1 // vcmpltps 0x37016(%rip){1to8},%ymm8,%k0 # 389e4 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
.byte 98,114,126,40,56,200 // vpmovm2d %k0,%ymm9
.byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
.byte 196,67,45,74,192,144 // vblendvps %ymm9,%ymm8,%ymm10,%ymm8
@@ -11159,16 +11159,16 @@ HIDDEN _sk_from_srgb_skx
.globl _sk_from_srgb_skx
FUNCTION(_sk_from_srgb_skx)
_sk_from_srgb_skx:
- .byte 196,98,125,24,5,116,125,3,0 // vbroadcastss 0x37d74(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,240,111,3,0 // vbroadcastss 0x36ff0(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 197,124,89,208 // vmulps %ymm0,%ymm0,%ymm10
- .byte 196,98,125,24,29,102,125,3,0 // vbroadcastss 0x37d66(%rip),%ymm11 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
- .byte 196,98,125,24,37,69,125,3,0 // vbroadcastss 0x37d45(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,226,111,3,0 // vbroadcastss 0x36fe2(%rip),%ymm11 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,37,193,111,3,0 // vbroadcastss 0x36fc1(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,125,168,235 // vfmadd213ps %ymm11,%ymm0,%ymm13
- .byte 196,98,125,24,53,78,125,3,0 // vbroadcastss 0x37d4e(%rip),%ymm14 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,53,202,111,3,0 // vbroadcastss 0x36fca(%rip),%ymm14 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,66,45,168,238 // vfmadd213ps %ymm14,%ymm10,%ymm13
- .byte 196,98,125,24,21,68,125,3,0 // vbroadcastss 0x37d44(%rip),%ymm10 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,21,192,111,3,0 // vbroadcastss 0x36fc0(%rip),%ymm10 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 98,209,124,40,194,194,1 // vcmpltps %ymm10,%ymm0,%k0
.byte 98,242,126,40,56,192 // vpmovm2d %k0,%ymm0
.byte 196,195,21,74,193,0 // vblendvps %ymm0,%ymm9,%ymm13,%ymm0
@@ -11194,16 +11194,16 @@ HIDDEN _sk_from_srgb_dst_skx
.globl _sk_from_srgb_dst_skx
FUNCTION(_sk_from_srgb_dst_skx)
_sk_from_srgb_dst_skx:
- .byte 196,98,125,24,5,199,124,3,0 // vbroadcastss 0x37cc7(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,67,111,3,0 // vbroadcastss 0x36f43(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,92,89,200 // vmulps %ymm8,%ymm4,%ymm9
.byte 197,92,89,212 // vmulps %ymm4,%ymm4,%ymm10
- .byte 196,98,125,24,29,185,124,3,0 // vbroadcastss 0x37cb9(%rip),%ymm11 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
- .byte 196,98,125,24,37,152,124,3,0 // vbroadcastss 0x37c98(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,53,111,3,0 // vbroadcastss 0x36f35(%rip),%ymm11 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,37,20,111,3,0 // vbroadcastss 0x36f14(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,93,168,235 // vfmadd213ps %ymm11,%ymm4,%ymm13
- .byte 196,98,125,24,53,161,124,3,0 // vbroadcastss 0x37ca1(%rip),%ymm14 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,53,29,111,3,0 // vbroadcastss 0x36f1d(%rip),%ymm14 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,66,45,168,238 // vfmadd213ps %ymm14,%ymm10,%ymm13
- .byte 196,98,125,24,21,151,124,3,0 // vbroadcastss 0x37c97(%rip),%ymm10 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,21,19,111,3,0 // vbroadcastss 0x36f13(%rip),%ymm10 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 98,209,92,40,194,194,1 // vcmpltps %ymm10,%ymm4,%k0
.byte 98,242,126,40,56,224 // vpmovm2d %k0,%ymm4
.byte 196,195,21,74,225,64 // vblendvps %ymm4,%ymm9,%ymm13,%ymm4
@@ -11230,19 +11230,19 @@ HIDDEN _sk_to_srgb_skx
FUNCTION(_sk_to_srgb_skx)
_sk_to_srgb_skx:
.byte 98,114,125,40,78,192 // vrsqrt14ps %ymm0,%ymm8
- .byte 196,98,125,24,13,36,124,3,0 // vbroadcastss 0x37c24(%rip),%ymm9 # 3977c <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
+ .byte 196,98,125,24,13,160,110,3,0 // vbroadcastss 0x36ea0(%rip),%ymm9 # 389f8 <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
.byte 196,65,124,89,209 // vmulps %ymm9,%ymm0,%ymm10
- .byte 196,98,125,24,29,26,124,3,0 // vbroadcastss 0x37c1a(%rip),%ymm11 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
- .byte 196,98,125,24,37,21,124,3,0 // vbroadcastss 0x37c15(%rip),%ymm12 # 39784 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
+ .byte 196,98,125,24,29,150,110,3,0 // vbroadcastss 0x36e96(%rip),%ymm11 # 389fc <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
+ .byte 196,98,125,24,37,145,110,3,0 // vbroadcastss 0x36e91(%rip),%ymm12 # 38a00 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,61,168,235 // vfmadd213ps %ymm11,%ymm8,%ymm13
- .byte 196,98,125,24,53,6,124,3,0 // vbroadcastss 0x37c06(%rip),%ymm14 # 39788 <_sk_srcover_bgra_8888_sse2_lowp+0x300>
+ .byte 196,98,125,24,53,130,110,3,0 // vbroadcastss 0x36e82(%rip),%ymm14 # 38a04 <_sk_srcover_bgra_8888_sse2_lowp+0x300>
.byte 196,66,61,168,238 // vfmadd213ps %ymm14,%ymm8,%ymm13
- .byte 196,98,125,24,61,252,123,3,0 // vbroadcastss 0x37bfc(%rip),%ymm15 # 3978c <_sk_srcover_bgra_8888_sse2_lowp+0x304>
+ .byte 196,98,125,24,61,120,110,3,0 // vbroadcastss 0x36e78(%rip),%ymm15 # 38a08 <_sk_srcover_bgra_8888_sse2_lowp+0x304>
.byte 196,65,60,88,199 // vaddps %ymm15,%ymm8,%ymm8
.byte 98,82,125,40,76,192 // vrcp14ps %ymm8,%ymm8
.byte 196,65,20,89,192 // vmulps %ymm8,%ymm13,%ymm8
- .byte 196,98,125,24,45,231,123,3,0 // vbroadcastss 0x37be7(%rip),%ymm13 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x308>
+ .byte 196,98,125,24,45,99,110,3,0 // vbroadcastss 0x36e63(%rip),%ymm13 # 38a0c <_sk_srcover_bgra_8888_sse2_lowp+0x308>
.byte 98,209,124,40,194,197,1 // vcmpltps %ymm13,%ymm0,%k0
.byte 98,242,126,40,56,192 // vpmovm2d %k0,%ymm0
.byte 196,195,61,74,194,0 // vblendvps %ymm0,%ymm10,%ymm8,%ymm0
@@ -11280,7 +11280,7 @@ _sk_rgb_to_hsl_skx:
.byte 197,116,93,194 // vminps %ymm2,%ymm1,%ymm8
.byte 196,65,124,93,208 // vminps %ymm8,%ymm0,%ymm10
.byte 98,193,52,40,92,194 // vsubps %ymm10,%ymm9,%ymm16
- .byte 196,98,125,24,5,216,122,3,0 // vbroadcastss 0x37ad8(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,84,109,3,0 // vbroadcastss 0x36d54(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 98,49,60,40,94,224 // vdivps %ymm16,%ymm8,%ymm12
.byte 98,209,52,40,194,194,0 // vcmpeqps %ymm10,%ymm9,%k0
.byte 98,114,126,40,56,192 // vpmovm2d %k0,%ymm8
@@ -11289,23 +11289,23 @@ _sk_rgb_to_hsl_skx:
.byte 197,116,92,242 // vsubps %ymm2,%ymm1,%ymm14
.byte 98,241,116,40,194,194,1 // vcmpltps %ymm2,%ymm1,%k0
.byte 98,114,126,40,56,248 // vpmovm2d %k0,%ymm15
- .byte 196,98,125,24,29,250,122,3,0 // vbroadcastss 0x37afa(%rip),%ymm11 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,98,125,24,29,118,109,3,0 // vbroadcastss 0x36d76(%rip),%ymm11 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
.byte 196,67,69,74,219,240 // vblendvps %ymm15,%ymm11,%ymm7,%ymm11
.byte 196,66,29,168,243 // vfmadd213ps %ymm11,%ymm12,%ymm14
.byte 98,241,52,40,194,193,0 // vcmpeqps %ymm1,%ymm9,%k0
.byte 98,114,126,40,56,216 // vpmovm2d %k0,%ymm11
.byte 197,236,92,208 // vsubps %ymm0,%ymm2,%ymm2
- .byte 196,98,125,24,61,213,122,3,0 // vbroadcastss 0x37ad5(%rip),%ymm15 # 39798 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
+ .byte 196,98,125,24,61,81,109,3,0 // vbroadcastss 0x36d51(%rip),%ymm15 # 38a14 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
.byte 196,194,29,168,215 // vfmadd213ps %ymm15,%ymm12,%ymm2
.byte 197,252,92,193 // vsubps %ymm1,%ymm0,%ymm0
- .byte 98,242,29,56,168,5,198,122,3,0 // vfmadd213ps 0x37ac6(%rip){1to8},%ymm12,%ymm0 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 98,242,29,56,168,5,66,109,3,0 // vfmadd213ps 0x36d42(%rip){1to8},%ymm12,%ymm0 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 196,227,125,74,194,176 // vblendvps %ymm11,%ymm2,%ymm0,%ymm0
.byte 196,195,125,74,198,208 // vblendvps %ymm13,%ymm14,%ymm0,%ymm0
.byte 196,227,125,74,199,128 // vblendvps %ymm8,%ymm7,%ymm0,%ymm0
- .byte 98,241,124,56,89,5,174,122,3,0 // vmulps 0x37aae(%rip){1to8},%ymm0,%ymm0 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 98,241,124,56,89,5,42,109,3,0 // vmulps 0x36d2a(%rip){1to8},%ymm0,%ymm0 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 196,193,52,88,202 // vaddps %ymm10,%ymm9,%ymm1
- .byte 196,98,125,24,29,52,122,3,0 // vbroadcastss 0x37a34(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,176,108,3,0 // vbroadcastss 0x36cb0(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,116,89,211 // vmulps %ymm11,%ymm1,%ymm2
.byte 98,241,36,40,194,194,1 // vcmpltps %ymm2,%ymm11,%k0
.byte 98,114,126,40,56,216 // vpmovm2d %k0,%ymm11
@@ -11324,7 +11324,7 @@ FUNCTION(_sk_hsl_to_rgb_skx)
_sk_hsl_to_rgb_skx:
.byte 98,225,124,40,40,215 // vmovaps %ymm7,%ymm18
.byte 98,225,124,40,40,230 // vmovaps %ymm6,%ymm20
- .byte 98,226,125,40,24,5,230,121,3,0 // vbroadcastss 0x379e6(%rip),%ymm16 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,226,125,40,24,5,98,108,3,0 // vbroadcastss 0x36c62(%rip),%ymm16 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 98,241,124,32,194,194,2 // vcmpleps %ymm2,%ymm16,%k0
.byte 98,114,126,40,56,192 // vpmovm2d %k0,%ymm8
.byte 197,116,89,202 // vmulps %ymm2,%ymm1,%ymm9
@@ -11332,27 +11332,27 @@ _sk_hsl_to_rgb_skx:
.byte 196,67,53,74,194,128 // vblendvps %ymm8,%ymm10,%ymm9,%ymm8
.byte 197,60,88,218 // vaddps %ymm2,%ymm8,%ymm11
.byte 196,65,124,40,203 // vmovaps %ymm11,%ymm9
- .byte 98,114,109,56,186,13,27,122,3,0 // vfmsub231ps 0x37a1b(%rip){1to8},%ymm2,%ymm9 # 39798 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
+ .byte 98,114,109,56,186,13,151,108,3,0 // vfmsub231ps 0x36c97(%rip){1to8},%ymm2,%ymm9 # 38a14 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 98,209,116,40,194,192,0 // vcmpeqps %ymm8,%ymm1,%k0
.byte 98,242,126,40,56,248 // vpmovm2d %k0,%ymm7
- .byte 98,241,124,56,88,13,11,122,3,0 // vaddps 0x37a0b(%rip){1to8},%ymm0,%ymm1 # 397a4 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
+ .byte 98,241,124,56,88,13,135,108,3,0 // vaddps 0x36c87(%rip){1to8},%ymm0,%ymm1 # 38a20 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
.byte 196,99,125,8,193,1 // vroundps $0x1,%ymm1,%ymm8
.byte 196,193,116,92,200 // vsubps %ymm8,%ymm1,%ymm1
- .byte 98,226,125,40,24,29,250,121,3,0 // vbroadcastss 0x379fa(%rip),%ymm19 # 397a8 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
+ .byte 98,226,125,40,24,29,118,108,3,0 // vbroadcastss 0x36c76(%rip),%ymm19 # 38a24 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
.byte 98,241,100,32,194,193,2 // vcmpleps %ymm1,%ymm19,%k0
.byte 98,114,126,40,56,192 // vpmovm2d %k0,%ymm8
.byte 196,65,36,92,241 // vsubps %ymm9,%ymm11,%ymm14
- .byte 196,98,125,24,61,203,121,3,0 // vbroadcastss 0x379cb(%rip),%ymm15 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,98,125,24,61,71,108,3,0 // vbroadcastss 0x36c47(%rip),%ymm15 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 196,65,116,89,231 // vmulps %ymm15,%ymm1,%ymm12
- .byte 98,226,125,40,24,13,196,121,3,0 // vbroadcastss 0x379c4(%rip),%ymm17 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 98,226,125,40,24,13,64,108,3,0 // vbroadcastss 0x36c40(%rip),%ymm17 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 98,81,116,32,92,212 // vsubps %ymm12,%ymm17,%ymm10
.byte 196,66,13,168,209 // vfmadd213ps %ymm9,%ymm14,%ymm10
.byte 196,67,45,74,193,128 // vblendvps %ymm8,%ymm9,%ymm10,%ymm8
.byte 98,241,124,32,194,193,2 // vcmpleps %ymm1,%ymm16,%k0
.byte 98,114,126,40,56,208 // vpmovm2d %k0,%ymm10
.byte 196,67,37,74,192,160 // vblendvps %ymm10,%ymm8,%ymm11,%ymm8
- .byte 196,98,125,24,21,155,121,3,0 // vbroadcastss 0x3799b(%rip),%ymm10 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 196,98,125,24,21,23,108,3,0 // vbroadcastss 0x36c17(%rip),%ymm10 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 98,241,44,40,194,193,2 // vcmpleps %ymm1,%ymm10,%k0
.byte 98,242,126,40,56,200 // vpmovm2d %k0,%ymm1
.byte 196,66,13,168,225 // vfmadd213ps %ymm9,%ymm14,%ymm12
@@ -11374,7 +11374,7 @@ _sk_hsl_to_rgb_skx:
.byte 196,66,13,168,233 // vfmadd213ps %ymm9,%ymm14,%ymm13
.byte 196,195,21,74,200,16 // vblendvps %ymm1,%ymm8,%ymm13,%ymm1
.byte 196,227,117,74,202,112 // vblendvps %ymm7,%ymm2,%ymm1,%ymm1
- .byte 98,241,124,56,88,5,33,121,3,0 // vaddps 0x37921(%rip){1to8},%ymm0,%ymm0 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 98,241,124,56,88,5,157,107,3,0 // vaddps 0x36b9d(%rip){1to8},%ymm0,%ymm0 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,99,125,8,192,1 // vroundps $0x1,%ymm0,%ymm8
.byte 196,193,124,92,192 // vsubps %ymm8,%ymm0,%ymm0
.byte 98,241,100,32,194,192,2 // vcmpleps %ymm0,%ymm19,%k0
@@ -11423,10 +11423,10 @@ _sk_scale_u8_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,54 // jne 1f67 <_sk_scale_u8_skx+0x4e>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,209,128,3,0 // vpand 0x380d1(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,81,115,3,0 // vpand 0x37351(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 98,113,60,56,89,5,93,120,3,0 // vmulps 0x3785d(%rip){1to8},%ymm8,%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,113,60,56,89,5,217,106,3,0 // vmulps 0x36ad9(%rip){1to8},%ymm8,%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
.byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
@@ -11503,15 +11503,15 @@ _sk_scale_565_skx:
.byte 15,133,145,0,0,0 // jne 20e0 <_sk_scale_565_skx+0xb0>
.byte 196,1,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm8
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 98,113,61,56,219,13,80,119,3,0 // vpandd 0x37750(%rip){1to8},%ymm8,%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 98,113,61,56,219,13,204,105,3,0 // vpandd 0x369cc(%rip){1to8},%ymm8,%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 98,113,52,56,89,13,69,119,3,0 // vmulps 0x37745(%rip){1to8},%ymm9,%ymm9 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
- .byte 98,113,61,56,219,21,63,119,3,0 // vpandd 0x3773f(%rip){1to8},%ymm8,%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 98,113,52,56,89,13,193,105,3,0 // vmulps 0x369c1(%rip){1to8},%ymm9,%ymm9 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 98,113,61,56,219,21,187,105,3,0 // vpandd 0x369bb(%rip){1to8},%ymm8,%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 98,113,44,56,89,21,52,119,3,0 // vmulps 0x37734(%rip){1to8},%ymm10,%ymm10 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
- .byte 98,113,61,56,219,5,46,119,3,0 // vpandd 0x3772e(%rip){1to8},%ymm8,%ymm8 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 98,113,44,56,89,21,176,105,3,0 // vmulps 0x369b0(%rip){1to8},%ymm10,%ymm10 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 98,113,61,56,219,5,170,105,3,0 // vpandd 0x369aa(%rip){1to8},%ymm8,%ymm8 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 98,113,60,56,89,5,35,119,3,0 // vmulps 0x37723(%rip){1to8},%ymm8,%ymm8 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 98,113,60,56,89,5,159,105,3,0 // vmulps 0x3699f(%rip){1to8},%ymm8,%ymm8 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 98,241,100,40,194,199,1 // vcmpltps %ymm7,%ymm3,%k0
.byte 98,114,126,40,56,216 // vpmovm2d %k0,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -11607,10 +11607,10 @@ _sk_lerp_u8_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,74 // jne 2235 <_sk_lerp_u8_skx+0x62>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,23,126,3,0 // vpand 0x37e17(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,151,112,3,0 // vpand 0x37097(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 98,113,60,56,89,5,163,117,3,0 // vmulps 0x375a3(%rip){1to8},%ymm8,%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,113,60,56,89,5,31,104,3,0 // vmulps 0x3681f(%rip){1to8},%ymm8,%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
.byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
.byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
@@ -11690,15 +11690,15 @@ _sk_lerp_565_skx:
.byte 15,133,165,0,0,0 // jne 23c0 <_sk_lerp_565_skx+0xc4>
.byte 196,1,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm8
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 98,113,61,56,219,13,132,116,3,0 // vpandd 0x37484(%rip){1to8},%ymm8,%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 98,113,61,56,219,13,0,103,3,0 // vpandd 0x36700(%rip){1to8},%ymm8,%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 98,113,52,56,89,13,121,116,3,0 // vmulps 0x37479(%rip){1to8},%ymm9,%ymm9 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
- .byte 98,113,61,56,219,21,115,116,3,0 // vpandd 0x37473(%rip){1to8},%ymm8,%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 98,113,52,56,89,13,245,102,3,0 // vmulps 0x366f5(%rip){1to8},%ymm9,%ymm9 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 98,113,61,56,219,21,239,102,3,0 // vpandd 0x366ef(%rip){1to8},%ymm8,%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 98,113,44,56,89,21,104,116,3,0 // vmulps 0x37468(%rip){1to8},%ymm10,%ymm10 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
- .byte 98,113,61,56,219,5,98,116,3,0 // vpandd 0x37462(%rip){1to8},%ymm8,%ymm8 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 98,113,44,56,89,21,228,102,3,0 // vmulps 0x366e4(%rip){1to8},%ymm10,%ymm10 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 98,113,61,56,219,5,222,102,3,0 // vpandd 0x366de(%rip){1to8},%ymm8,%ymm8 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 98,113,60,56,89,5,87,116,3,0 // vmulps 0x37457(%rip){1to8},%ymm8,%ymm8 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 98,113,60,56,89,5,211,102,3,0 // vmulps 0x366d3(%rip){1to8},%ymm8,%ymm8 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 98,241,100,40,194,199,1 // vcmpltps %ymm7,%ymm3,%k0
.byte 98,114,126,40,56,216 // vpmovm2d %k0,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -11798,7 +11798,7 @@ _sk_load_tables_skx:
.byte 196,162,61,146,20,136 // vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
.byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 98,241,100,56,89,29,170,114,3,0 // vmulps 0x372aa(%rip){1to8},%ymm3,%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,100,56,89,29,38,101,3,0 // vmulps 0x36526(%rip){1to8},%ymm3,%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -11877,7 +11877,7 @@ _sk_load_tables_u16_be_skx:
.byte 197,113,105,219 // vpunpckhwd %xmm3,%xmm1,%xmm11
.byte 197,185,108,200 // vpunpcklqdq %xmm0,%xmm8,%xmm1
.byte 197,57,109,192 // vpunpckhqdq %xmm0,%xmm8,%xmm8
- .byte 197,121,111,21,213,121,3,0 // vmovdqa 0x379d5(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,121,111,21,85,108,3,0 // vmovdqa 0x36c55(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,193,113,219,194 // vpand %xmm10,%xmm1,%xmm0
.byte 196,226,125,51,200 // vpmovzxwd %xmm0,%ymm1
.byte 76,139,64,8 // mov 0x8(%rax),%r8
@@ -11903,7 +11903,7 @@ _sk_load_tables_u16_be_skx:
.byte 197,185,235,219 // vpor %xmm3,%xmm8,%xmm3
.byte 196,226,125,51,219 // vpmovzxwd %xmm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 98,241,100,56,89,29,16,113,3,0 // vmulps 0x37110(%rip){1to8},%ymm3,%ymm3 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 98,241,100,56,89,29,140,99,3,0 // vmulps 0x3638c(%rip){1to8},%ymm3,%ymm3 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 196,1,123,16,4,72 // vmovsd (%r8,%r9,2),%xmm8
@@ -11960,7 +11960,7 @@ _sk_load_tables_rgb_u16_be_skx:
.byte 197,249,105,193 // vpunpckhwd %xmm1,%xmm0,%xmm0
.byte 197,105,108,200 // vpunpcklqdq %xmm0,%xmm2,%xmm9
.byte 197,185,108,195 // vpunpcklqdq %xmm3,%xmm8,%xmm0
- .byte 197,121,111,21,94,120,3,0 // vmovdqa 0x3785e(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,121,111,21,222,106,3,0 // vmovdqa 0x36ade(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,193,121,219,194 // vpand %xmm10,%xmm0,%xmm0
.byte 196,226,125,51,200 // vpmovzxwd %xmm0,%ymm1
.byte 76,139,64,8 // mov 0x8(%rax),%r8
@@ -11981,7 +11981,7 @@ _sk_load_tables_rgb_u16_be_skx:
.byte 196,226,125,51,219 // vpmovzxwd %xmm3,%ymm3
.byte 196,226,29,146,20,152 // vgatherdps %ymm12,(%rax,%ymm3,4),%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,33,111,3,0 // vbroadcastss 0x36f21(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,157,97,3,0 // vbroadcastss 0x3619d(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 196,129,121,110,4,72 // vmovd (%r8,%r9,2),%xmm0
.byte 196,65,49,87,201 // vxorpd %xmm9,%xmm9,%xmm9
@@ -12050,7 +12050,7 @@ HIDDEN _sk_byte_tables_skx
FUNCTION(_sk_byte_tables_skx)
_sk_byte_tables_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,22,110,3,0 // vbroadcastss 0x36e16(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,146,96,3,0 // vbroadcastss 0x36092(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,125,91,200 // vcvtps2dq %ymm0,%ymm9
.byte 196,65,249,126,200 // vmovq %xmm9,%r8
@@ -12180,7 +12180,7 @@ _sk_byte_tables_skx:
.byte 67,15,182,4,17 // movzbl (%r9,%r10,1),%eax
.byte 196,194,125,49,193 // vpmovzxbd %xmm9,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,232,107,3,0 // vbroadcastss 0x36be8(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,100,94,3,0 // vbroadcastss 0x35e64(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 196,226,125,49,201 // vpmovzxbd %xmm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
@@ -12301,7 +12301,7 @@ _sk_byte_tables_rgb_skx:
.byte 67,15,182,4,17 // movzbl (%r9,%r10,1),%eax
.byte 196,194,125,49,193 // vpmovzxbd %xmm9,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,191,105,3,0 // vbroadcastss 0x369bf(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,59,92,3,0 // vbroadcastss 0x35c3b(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 196,194,125,49,202 // vpmovzxbd %xmm10,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
@@ -12398,33 +12398,33 @@ _sk_parametric_r_skx:
.byte 98,114,125,56,168,80,2 // vfmadd213ps 0x8(%rax){1to8},%ymm0,%ymm10
.byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,179,104,3,0 // vbroadcastss 0x368b3(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 98,114,37,56,168,37,177,104,3,0 // vfmadd213ps 0x368b1(%rip){1to8},%ymm11,%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
- .byte 98,113,44,56,84,21,163,104,3,0 // vandps 0x368a3(%rip){1to8},%ymm10,%ymm10 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
- .byte 98,113,44,56,86,21,249,103,3,0 // vorps 0x367f9(%rip){1to8},%ymm10,%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,45,56,188,37,151,104,3,0 // vfnmadd231ps 0x36897(%rip){1to8},%ymm10,%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
- .byte 98,113,44,56,88,21,145,104,3,0 // vaddps 0x36891(%rip){1to8},%ymm10,%ymm10 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
- .byte 196,98,125,24,29,140,104,3,0 // vbroadcastss 0x3688c(%rip),%ymm11 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,47,91,3,0 // vbroadcastss 0x35b2f(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 98,114,37,56,168,37,45,91,3,0 // vfmadd213ps 0x35b2d(%rip){1to8},%ymm11,%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 98,113,44,56,84,21,31,91,3,0 // vandps 0x35b1f(%rip){1to8},%ymm10,%ymm10 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 98,113,44,56,86,21,117,90,3,0 // vorps 0x35a75(%rip){1to8},%ymm10,%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,45,56,188,37,19,91,3,0 // vfnmadd231ps 0x35b13(%rip){1to8},%ymm10,%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 98,113,44,56,88,21,13,91,3,0 // vaddps 0x35b0d(%rip){1to8},%ymm10,%ymm10 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,29,8,91,3,0 // vbroadcastss 0x35b08(%rip),%ymm11 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,65,28,92,210 // vsubps %ymm10,%ymm12,%ymm10
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 196,99,125,8,208,1 // vroundps $0x1,%ymm0,%ymm10
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 98,241,124,56,88,5,108,104,3,0 // vaddps 0x3686c(%rip){1to8},%ymm0,%ymm0 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
- .byte 98,242,45,56,188,5,102,104,3,0 // vfnmadd231ps 0x36866(%rip){1to8},%ymm10,%ymm0 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
- .byte 196,98,125,24,29,97,104,3,0 // vbroadcastss 0x36861(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 98,241,124,56,88,5,232,90,3,0 // vaddps 0x35ae8(%rip){1to8},%ymm0,%ymm0 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 98,242,45,56,188,5,226,90,3,0 // vfnmadd231ps 0x35ae2(%rip){1to8},%ymm10,%ymm0 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,221,90,3,0 // vbroadcastss 0x35add(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,87,104,3,0 // vbroadcastss 0x36857(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,211,90,3,0 // vbroadcastss 0x35ad3(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,124,88,194 // vaddps %ymm10,%ymm0,%ymm0
- .byte 98,241,124,56,89,5,71,104,3,0 // vmulps 0x36847(%rip){1to8},%ymm0,%ymm0 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 98,241,124,56,89,5,195,90,3,0 // vmulps 0x35ac3(%rip){1to8},%ymm0,%ymm0 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,253,91,192 // vcvtps2dq %ymm0,%ymm0
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
.byte 197,172,88,192 // vaddps %ymm0,%ymm10,%ymm0
.byte 196,195,125,74,193,128 // vblendvps %ymm8,%ymm9,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 98,241,124,56,93,5,95,103,3,0 // vminps 0x3675f(%rip){1to8},%ymm0,%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,93,5,219,89,3,0 // vminps 0x359db(%rip){1to8},%ymm0,%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -12441,33 +12441,33 @@ _sk_parametric_g_skx:
.byte 98,114,117,56,168,80,2 // vfmadd213ps 0x8(%rax){1to8},%ymm1,%ymm10
.byte 196,226,125,24,8 // vbroadcastss (%rax),%ymm1
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,182,103,3,0 // vbroadcastss 0x367b6(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 98,114,37,56,168,37,180,103,3,0 // vfmadd213ps 0x367b4(%rip){1to8},%ymm11,%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
- .byte 98,113,44,56,84,21,166,103,3,0 // vandps 0x367a6(%rip){1to8},%ymm10,%ymm10 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
- .byte 98,113,44,56,86,21,252,102,3,0 // vorps 0x366fc(%rip){1to8},%ymm10,%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,45,56,188,37,154,103,3,0 // vfnmadd231ps 0x3679a(%rip){1to8},%ymm10,%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
- .byte 98,113,44,56,88,21,148,103,3,0 // vaddps 0x36794(%rip){1to8},%ymm10,%ymm10 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
- .byte 196,98,125,24,29,143,103,3,0 // vbroadcastss 0x3678f(%rip),%ymm11 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,50,90,3,0 // vbroadcastss 0x35a32(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 98,114,37,56,168,37,48,90,3,0 // vfmadd213ps 0x35a30(%rip){1to8},%ymm11,%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 98,113,44,56,84,21,34,90,3,0 // vandps 0x35a22(%rip){1to8},%ymm10,%ymm10 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 98,113,44,56,86,21,120,89,3,0 // vorps 0x35978(%rip){1to8},%ymm10,%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,45,56,188,37,22,90,3,0 // vfnmadd231ps 0x35a16(%rip){1to8},%ymm10,%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 98,113,44,56,88,21,16,90,3,0 // vaddps 0x35a10(%rip){1to8},%ymm10,%ymm10 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,29,11,90,3,0 // vbroadcastss 0x35a0b(%rip),%ymm11 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,65,28,92,210 // vsubps %ymm10,%ymm12,%ymm10
.byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
.byte 196,99,125,8,209,1 // vroundps $0x1,%ymm1,%ymm10
.byte 196,65,116,92,210 // vsubps %ymm10,%ymm1,%ymm10
- .byte 98,241,116,56,88,13,111,103,3,0 // vaddps 0x3676f(%rip){1to8},%ymm1,%ymm1 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
- .byte 98,242,45,56,188,13,105,103,3,0 // vfnmadd231ps 0x36769(%rip){1to8},%ymm10,%ymm1 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
- .byte 196,98,125,24,29,100,103,3,0 // vbroadcastss 0x36764(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 98,241,116,56,88,13,235,89,3,0 // vaddps 0x359eb(%rip){1to8},%ymm1,%ymm1 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 98,242,45,56,188,13,229,89,3,0 // vfnmadd231ps 0x359e5(%rip){1to8},%ymm10,%ymm1 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,224,89,3,0 // vbroadcastss 0x359e0(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,90,103,3,0 // vbroadcastss 0x3675a(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,214,89,3,0 // vbroadcastss 0x359d6(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,116,88,202 // vaddps %ymm10,%ymm1,%ymm1
- .byte 98,241,116,56,89,13,74,103,3,0 // vmulps 0x3674a(%rip){1to8},%ymm1,%ymm1 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 98,241,116,56,89,13,198,89,3,0 // vmulps 0x359c6(%rip){1to8},%ymm1,%ymm1 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,253,91,201 // vcvtps2dq %ymm1,%ymm1
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
.byte 197,172,88,201 // vaddps %ymm1,%ymm10,%ymm1
.byte 196,195,117,74,201,128 // vblendvps %ymm8,%ymm9,%ymm1,%ymm1
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
- .byte 98,241,116,56,93,13,98,102,3,0 // vminps 0x36662(%rip){1to8},%ymm1,%ymm1 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,116,56,93,13,222,88,3,0 // vminps 0x358de(%rip){1to8},%ymm1,%ymm1 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -12484,33 +12484,33 @@ _sk_parametric_b_skx:
.byte 98,114,109,56,168,80,2 // vfmadd213ps 0x8(%rax){1to8},%ymm2,%ymm10
.byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,185,102,3,0 // vbroadcastss 0x366b9(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 98,114,37,56,168,37,183,102,3,0 // vfmadd213ps 0x366b7(%rip){1to8},%ymm11,%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
- .byte 98,113,44,56,84,21,169,102,3,0 // vandps 0x366a9(%rip){1to8},%ymm10,%ymm10 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
- .byte 98,113,44,56,86,21,255,101,3,0 // vorps 0x365ff(%rip){1to8},%ymm10,%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,45,56,188,37,157,102,3,0 // vfnmadd231ps 0x3669d(%rip){1to8},%ymm10,%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
- .byte 98,113,44,56,88,21,151,102,3,0 // vaddps 0x36697(%rip){1to8},%ymm10,%ymm10 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
- .byte 196,98,125,24,29,146,102,3,0 // vbroadcastss 0x36692(%rip),%ymm11 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,53,89,3,0 // vbroadcastss 0x35935(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 98,114,37,56,168,37,51,89,3,0 // vfmadd213ps 0x35933(%rip){1to8},%ymm11,%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 98,113,44,56,84,21,37,89,3,0 // vandps 0x35925(%rip){1to8},%ymm10,%ymm10 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 98,113,44,56,86,21,123,88,3,0 // vorps 0x3587b(%rip){1to8},%ymm10,%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,45,56,188,37,25,89,3,0 // vfnmadd231ps 0x35919(%rip){1to8},%ymm10,%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 98,113,44,56,88,21,19,89,3,0 // vaddps 0x35913(%rip){1to8},%ymm10,%ymm10 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,29,14,89,3,0 // vbroadcastss 0x3590e(%rip),%ymm11 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,65,28,92,210 // vsubps %ymm10,%ymm12,%ymm10
.byte 196,193,108,89,210 // vmulps %ymm10,%ymm2,%ymm2
.byte 196,99,125,8,210,1 // vroundps $0x1,%ymm2,%ymm10
.byte 196,65,108,92,210 // vsubps %ymm10,%ymm2,%ymm10
- .byte 98,241,108,56,88,21,114,102,3,0 // vaddps 0x36672(%rip){1to8},%ymm2,%ymm2 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
- .byte 98,242,45,56,188,21,108,102,3,0 // vfnmadd231ps 0x3666c(%rip){1to8},%ymm10,%ymm2 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
- .byte 196,98,125,24,29,103,102,3,0 // vbroadcastss 0x36667(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 98,241,108,56,88,21,238,88,3,0 // vaddps 0x358ee(%rip){1to8},%ymm2,%ymm2 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 98,242,45,56,188,21,232,88,3,0 // vfnmadd231ps 0x358e8(%rip){1to8},%ymm10,%ymm2 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,227,88,3,0 // vbroadcastss 0x358e3(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,93,102,3,0 // vbroadcastss 0x3665d(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,217,88,3,0 // vbroadcastss 0x358d9(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,108,88,210 // vaddps %ymm10,%ymm2,%ymm2
- .byte 98,241,108,56,89,21,77,102,3,0 // vmulps 0x3664d(%rip){1to8},%ymm2,%ymm2 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 98,241,108,56,89,21,201,88,3,0 // vmulps 0x358c9(%rip){1to8},%ymm2,%ymm2 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,253,91,210 // vcvtps2dq %ymm2,%ymm2
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
.byte 197,172,88,210 // vaddps %ymm2,%ymm10,%ymm2
.byte 196,195,109,74,209,128 // vblendvps %ymm8,%ymm9,%ymm2,%ymm2
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 98,241,108,56,93,21,101,101,3,0 // vminps 0x36565(%rip){1to8},%ymm2,%ymm2 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,108,56,93,21,225,87,3,0 // vminps 0x357e1(%rip){1to8},%ymm2,%ymm2 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -12527,33 +12527,33 @@ _sk_parametric_a_skx:
.byte 98,114,101,56,168,80,2 // vfmadd213ps 0x8(%rax){1to8},%ymm3,%ymm10
.byte 196,226,125,24,24 // vbroadcastss (%rax),%ymm3
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,188,101,3,0 // vbroadcastss 0x365bc(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 98,114,37,56,168,37,186,101,3,0 // vfmadd213ps 0x365ba(%rip){1to8},%ymm11,%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
- .byte 98,113,44,56,84,21,172,101,3,0 // vandps 0x365ac(%rip){1to8},%ymm10,%ymm10 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
- .byte 98,113,44,56,86,21,2,101,3,0 // vorps 0x36502(%rip){1to8},%ymm10,%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,45,56,188,37,160,101,3,0 // vfnmadd231ps 0x365a0(%rip){1to8},%ymm10,%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
- .byte 98,113,44,56,88,21,154,101,3,0 // vaddps 0x3659a(%rip){1to8},%ymm10,%ymm10 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
- .byte 196,98,125,24,29,149,101,3,0 // vbroadcastss 0x36595(%rip),%ymm11 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,56,88,3,0 // vbroadcastss 0x35838(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 98,114,37,56,168,37,54,88,3,0 // vfmadd213ps 0x35836(%rip){1to8},%ymm11,%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 98,113,44,56,84,21,40,88,3,0 // vandps 0x35828(%rip){1to8},%ymm10,%ymm10 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 98,113,44,56,86,21,126,87,3,0 // vorps 0x3577e(%rip){1to8},%ymm10,%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,45,56,188,37,28,88,3,0 // vfnmadd231ps 0x3581c(%rip){1to8},%ymm10,%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 98,113,44,56,88,21,22,88,3,0 // vaddps 0x35816(%rip){1to8},%ymm10,%ymm10 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,29,17,88,3,0 // vbroadcastss 0x35811(%rip),%ymm11 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,65,28,92,210 // vsubps %ymm10,%ymm12,%ymm10
.byte 196,193,100,89,218 // vmulps %ymm10,%ymm3,%ymm3
.byte 196,99,125,8,211,1 // vroundps $0x1,%ymm3,%ymm10
.byte 196,65,100,92,210 // vsubps %ymm10,%ymm3,%ymm10
- .byte 98,241,100,56,88,29,117,101,3,0 // vaddps 0x36575(%rip){1to8},%ymm3,%ymm3 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
- .byte 98,242,45,56,188,29,111,101,3,0 // vfnmadd231ps 0x3656f(%rip){1to8},%ymm10,%ymm3 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
- .byte 196,98,125,24,29,106,101,3,0 // vbroadcastss 0x3656a(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 98,241,100,56,88,29,241,87,3,0 // vaddps 0x357f1(%rip){1to8},%ymm3,%ymm3 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 98,242,45,56,188,29,235,87,3,0 // vfnmadd231ps 0x357eb(%rip){1to8},%ymm10,%ymm3 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,230,87,3,0 // vbroadcastss 0x357e6(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,96,101,3,0 // vbroadcastss 0x36560(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,220,87,3,0 // vbroadcastss 0x357dc(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,100,88,218 // vaddps %ymm10,%ymm3,%ymm3
- .byte 98,241,100,56,89,29,80,101,3,0 // vmulps 0x36550(%rip){1to8},%ymm3,%ymm3 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 98,241,100,56,89,29,204,87,3,0 // vmulps 0x357cc(%rip){1to8},%ymm3,%ymm3 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,253,91,219 // vcvtps2dq %ymm3,%ymm3
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
.byte 197,172,88,219 // vaddps %ymm3,%ymm10,%ymm3
.byte 196,195,101,74,217,128 // vblendvps %ymm8,%ymm9,%ymm3,%ymm3
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
- .byte 98,241,100,56,93,29,104,100,3,0 // vminps 0x36468(%rip){1to8},%ymm3,%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,100,56,93,29,228,86,3,0 // vminps 0x356e4(%rip){1to8},%ymm3,%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -12564,33 +12564,33 @@ _sk_gamma_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 98,226,125,40,24,48 // vbroadcastss (%rax),%ymm22
.byte 197,124,91,200 // vcvtdq2ps %ymm0,%ymm9
- .byte 196,98,125,24,21,231,100,3,0 // vbroadcastss 0x364e7(%rip),%ymm10 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,98,125,24,29,226,100,3,0 // vbroadcastss 0x364e2(%rip),%ymm11 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,21,99,87,3,0 // vbroadcastss 0x35763(%rip),%ymm10 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,29,94,87,3,0 // vbroadcastss 0x3575e(%rip),%ymm11 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,124,84,195 // vandps %ymm11,%ymm0,%ymm0
- .byte 196,98,125,24,37,52,100,3,0 // vbroadcastss 0x36434(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,176,86,3,0 // vbroadcastss 0x356b0(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,124,86,196 // vorps %ymm12,%ymm0,%ymm0
- .byte 196,98,125,24,45,202,100,3,0 // vbroadcastss 0x364ca(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,70,87,3,0 // vbroadcastss 0x35746(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,45,168,205 // vfmadd213ps %ymm13,%ymm10,%ymm9
- .byte 196,98,125,24,53,192,100,3,0 // vbroadcastss 0x364c0(%rip),%ymm14 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,53,60,87,3,0 // vbroadcastss 0x3573c(%rip),%ymm14 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,125,188,206 // vfnmadd231ps %ymm14,%ymm0,%ymm9
- .byte 196,98,125,24,61,182,100,3,0 // vbroadcastss 0x364b6(%rip),%ymm15 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,61,50,87,3,0 // vbroadcastss 0x35732(%rip),%ymm15 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,124,88,199 // vaddps %ymm15,%ymm0,%ymm0
- .byte 98,226,125,40,24,5,171,100,3,0 // vbroadcastss 0x364ab(%rip),%ymm16 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 98,226,125,40,24,5,39,87,3,0 // vbroadcastss 0x35727(%rip),%ymm16 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 98,241,124,32,94,192 // vdivps %ymm0,%ymm16,%ymm0
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
.byte 98,177,124,40,89,198 // vmulps %ymm22,%ymm0,%ymm0
.byte 196,99,125,8,200,1 // vroundps $0x1,%ymm0,%ymm9
.byte 196,65,124,92,201 // vsubps %ymm9,%ymm0,%ymm9
- .byte 98,226,125,40,24,13,138,100,3,0 // vbroadcastss 0x3648a(%rip),%ymm17 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 98,226,125,40,24,13,6,87,3,0 // vbroadcastss 0x35706(%rip),%ymm17 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 98,177,124,40,88,193 // vaddps %ymm17,%ymm0,%ymm0
- .byte 98,226,125,40,24,21,126,100,3,0 // vbroadcastss 0x3647e(%rip),%ymm18 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 98,226,125,40,24,21,250,86,3,0 // vbroadcastss 0x356fa(%rip),%ymm18 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 98,178,53,40,188,194 // vfnmadd231ps %ymm18,%ymm9,%ymm0
- .byte 98,226,125,40,24,29,114,100,3,0 // vbroadcastss 0x36472(%rip),%ymm19 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 98,226,125,40,24,29,238,86,3,0 // vbroadcastss 0x356ee(%rip),%ymm19 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 98,81,100,32,92,201 // vsubps %ymm9,%ymm19,%ymm9
- .byte 98,226,125,40,24,37,102,100,3,0 // vbroadcastss 0x36466(%rip),%ymm20 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 98,226,125,40,24,37,226,86,3,0 // vbroadcastss 0x356e2(%rip),%ymm20 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 98,81,92,32,94,201 // vdivps %ymm9,%ymm20,%ymm9
.byte 196,193,124,88,193 // vaddps %ymm9,%ymm0,%ymm0
- .byte 196,98,125,24,13,86,100,3,0 // vbroadcastss 0x36456(%rip),%ymm9 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,13,210,86,3,0 // vbroadcastss 0x356d2(%rip),%ymm9 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,253,91,192 // vcvtps2dq %ymm0,%ymm0
.byte 98,225,124,40,91,233 // vcvtdq2ps %ymm1,%ymm21
@@ -12636,23 +12636,23 @@ HIDDEN _sk_lab_to_xyz_skx
.globl _sk_lab_to_xyz_skx
FUNCTION(_sk_lab_to_xyz_skx)
_sk_lab_to_xyz_skx:
- .byte 196,98,125,24,5,130,99,3,0 // vbroadcastss 0x36382(%rip),%ymm8 # 397fc <_sk_srcover_bgra_8888_sse2_lowp+0x374>
- .byte 196,98,125,24,13,225,98,3,0 // vbroadcastss 0x362e1(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,98,125,24,21,116,99,3,0 // vbroadcastss 0x36374(%rip),%ymm10 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x378>
+ .byte 196,98,125,24,5,254,85,3,0 // vbroadcastss 0x355fe(%rip),%ymm8 # 38a78 <_sk_srcover_bgra_8888_sse2_lowp+0x374>
+ .byte 196,98,125,24,13,93,85,3,0 // vbroadcastss 0x3555d(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,21,240,85,3,0 // vbroadcastss 0x355f0(%rip),%ymm10 # 38a7c <_sk_srcover_bgra_8888_sse2_lowp+0x378>
.byte 196,194,53,168,202 // vfmadd213ps %ymm10,%ymm9,%ymm1
.byte 196,194,53,168,210 // vfmadd213ps %ymm10,%ymm9,%ymm2
- .byte 98,114,125,56,168,5,100,99,3,0 // vfmadd213ps 0x36364(%rip){1to8},%ymm0,%ymm8 # 39804 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
- .byte 98,241,60,56,89,5,94,99,3,0 // vmulps 0x3635e(%rip){1to8},%ymm8,%ymm0 # 39808 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
- .byte 98,242,125,56,152,13,88,99,3,0 // vfmadd132ps 0x36358(%rip){1to8},%ymm0,%ymm1 # 3980c <_sk_srcover_bgra_8888_sse2_lowp+0x384>
- .byte 98,242,125,56,156,21,82,99,3,0 // vfnmadd132ps 0x36352(%rip){1to8},%ymm0,%ymm2 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x388>
+ .byte 98,114,125,56,168,5,224,85,3,0 // vfmadd213ps 0x355e0(%rip){1to8},%ymm0,%ymm8 # 38a80 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
+ .byte 98,241,60,56,89,5,218,85,3,0 // vmulps 0x355da(%rip){1to8},%ymm8,%ymm0 # 38a84 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
+ .byte 98,242,125,56,152,13,212,85,3,0 // vfmadd132ps 0x355d4(%rip){1to8},%ymm0,%ymm1 # 38a88 <_sk_srcover_bgra_8888_sse2_lowp+0x384>
+ .byte 98,242,125,56,156,21,206,85,3,0 // vfnmadd132ps 0x355ce(%rip){1to8},%ymm0,%ymm2 # 38a8c <_sk_srcover_bgra_8888_sse2_lowp+0x388>
.byte 197,116,89,193 // vmulps %ymm1,%ymm1,%ymm8
.byte 196,65,116,89,192 // vmulps %ymm8,%ymm1,%ymm8
- .byte 196,98,125,24,13,68,99,3,0 // vbroadcastss 0x36344(%rip),%ymm9 # 39814 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
+ .byte 196,98,125,24,13,192,85,3,0 // vbroadcastss 0x355c0(%rip),%ymm9 # 38a90 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
.byte 98,209,52,40,194,192,1 // vcmpltps %ymm8,%ymm9,%k0
.byte 98,114,126,40,56,208 // vpmovm2d %k0,%ymm10
- .byte 196,98,125,24,29,50,99,3,0 // vbroadcastss 0x36332(%rip),%ymm11 # 39818 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
+ .byte 196,98,125,24,29,174,85,3,0 // vbroadcastss 0x355ae(%rip),%ymm11 # 38a94 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,37,40,99,3,0 // vbroadcastss 0x36328(%rip),%ymm12 # 3981c <_sk_srcover_bgra_8888_sse2_lowp+0x394>
+ .byte 196,98,125,24,37,164,85,3,0 // vbroadcastss 0x355a4(%rip),%ymm12 # 38a98 <_sk_srcover_bgra_8888_sse2_lowp+0x394>
.byte 196,193,116,89,204 // vmulps %ymm12,%ymm1,%ymm1
.byte 196,67,117,74,192,160 // vblendvps %ymm10,%ymm8,%ymm1,%ymm8
.byte 197,252,89,200 // vmulps %ymm0,%ymm0,%ymm1
@@ -12669,8 +12669,8 @@ _sk_lab_to_xyz_skx:
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
.byte 196,193,108,89,212 // vmulps %ymm12,%ymm2,%ymm2
.byte 196,227,109,74,208,144 // vblendvps %ymm9,%ymm0,%ymm2,%ymm2
- .byte 98,241,60,56,89,5,205,98,3,0 // vmulps 0x362cd(%rip){1to8},%ymm8,%ymm0 # 39820 <_sk_srcover_bgra_8888_sse2_lowp+0x398>
- .byte 98,241,108,56,89,21,199,98,3,0 // vmulps 0x362c7(%rip){1to8},%ymm2,%ymm2 # 39824 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
+ .byte 98,241,60,56,89,5,73,85,3,0 // vmulps 0x35549(%rip){1to8},%ymm8,%ymm0 # 38a9c <_sk_srcover_bgra_8888_sse2_lowp+0x398>
+ .byte 98,241,108,56,89,21,67,85,3,0 // vmulps 0x35543(%rip){1to8},%ymm2,%ymm2 # 38aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -12687,10 +12687,10 @@ _sk_load_a8_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,49 // jne 35aa <_sk_load_a8_skx+0x49>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,137,106,3,0 // vpand 0x36a89(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,9,93,3,0 // vpand 0x35d09(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,29,22,98,3,0 // vmulps 0x36216(%rip){1to8},%ymm0,%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,124,56,89,29,146,84,3,0 // vmulps 0x35492(%rip){1to8},%ymm0,%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
@@ -12764,10 +12764,10 @@ _sk_load_a8_dst_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,49 // jne 36b5 <_sk_load_a8_dst_skx+0x49>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,126,105,3,0 // vpand 0x3697e(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,254,91,3,0 // vpand 0x35bfe(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,228 // vpmovzxwd %xmm4,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 98,241,92,56,89,61,11,97,3,0 // vmulps 0x3610b(%rip){1to8},%ymm4,%ymm7 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,92,56,89,61,135,83,3,0 // vmulps 0x35387(%rip){1to8},%ymm4,%ymm7 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -12883,7 +12883,7 @@ _sk_gather_a8_skx:
.byte 196,227,121,32,192,7 // vpinsrb $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,49,192 // vpmovzxbd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,29,78,95,3,0 // vmulps 0x35f4e(%rip){1to8},%ymm0,%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,124,56,89,29,202,81,3,0 // vmulps 0x351ca(%rip){1to8},%ymm0,%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
@@ -12904,7 +12904,7 @@ _sk_store_a8_skx:
.byte 77,15,175,193 // imul %r9,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 98,113,100,56,89,5,206,94,3,0 // vmulps 0x35ece(%rip){1to8},%ymm3,%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 98,113,100,56,89,5,74,81,3,0 // vmulps 0x3514a(%rip){1to8},%ymm3,%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,66,57,43,193 // vpackusdw %xmm9,%xmm8,%xmm8
@@ -12971,12 +12971,12 @@ _sk_load_g8_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,54 // jne 3992 <_sk_load_g8_skx+0x4e>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,166,102,3,0 // vpand 0x366a6(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,38,89,3,0 // vpand 0x35926(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,51,94,3,0 // vmulps 0x35e33(%rip){1to8},%ymm0,%ymm0 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,124,56,89,5,175,80,3,0 // vmulps 0x350af(%rip){1to8},%ymm0,%ymm0 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,176,93,3,0 // vbroadcastss 0x35db0(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,44,80,3,0 // vbroadcastss 0x3502c(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 255,224 // jmpq *%rax
@@ -13048,12 +13048,12 @@ _sk_load_g8_dst_skx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,54 // jne 3aa2 <_sk_load_g8_dst_skx+0x4e>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,150,101,3,0 // vpand 0x36596(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,22,88,3,0 // vpand 0x35816(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,228 // vpmovzxwd %xmm4,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 98,241,92,56,89,37,35,93,3,0 // vmulps 0x35d23(%rip){1to8},%ymm4,%ymm4 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,92,56,89,37,159,79,3,0 // vmulps 0x34f9f(%rip){1to8},%ymm4,%ymm4 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,160,92,3,0 // vbroadcastss 0x35ca0(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,28,79,3,0 // vbroadcastss 0x34f1c(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,236 // vmovaps %ymm4,%ymm5
.byte 197,252,40,244 // vmovaps %ymm4,%ymm6
.byte 255,224 // jmpq *%rax
@@ -13167,9 +13167,9 @@ _sk_gather_g8_skx:
.byte 196,227,121,32,192,7 // vpinsrb $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,49,192 // vpmovzxbd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,94,91,3,0 // vmulps 0x35b5e(%rip){1to8},%ymm0,%ymm0 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 98,241,124,56,89,5,218,77,3,0 // vmulps 0x34dda(%rip){1to8},%ymm0,%ymm0 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,219,90,3,0 // vbroadcastss 0x35adb(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,87,77,3,0 // vbroadcastss 0x34d57(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 91 // pop %rbx
@@ -13193,17 +13193,17 @@ _sk_load_565_skx:
.byte 117,96 // jne 3ce9 <_sk_load_565_skx+0x7b>
.byte 196,129,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm0
.byte 196,226,125,51,208 // vpmovzxwd %xmm0,%ymm2
- .byte 98,241,109,56,219,5,22,91,3,0 // vpandd 0x35b16(%rip){1to8},%ymm2,%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 98,241,109,56,219,5,146,77,3,0 // vpandd 0x34d92(%rip){1to8},%ymm2,%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,12,91,3,0 // vmulps 0x35b0c(%rip){1to8},%ymm0,%ymm0 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
- .byte 98,241,109,56,219,13,6,91,3,0 // vpandd 0x35b06(%rip){1to8},%ymm2,%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 98,241,124,56,89,5,136,77,3,0 // vmulps 0x34d88(%rip){1to8},%ymm0,%ymm0 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 98,241,109,56,219,13,130,77,3,0 // vpandd 0x34d82(%rip){1to8},%ymm2,%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 98,241,116,56,89,13,252,90,3,0 // vmulps 0x35afc(%rip){1to8},%ymm1,%ymm1 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
- .byte 98,241,109,56,219,21,246,90,3,0 // vpandd 0x35af6(%rip){1to8},%ymm2,%ymm2 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 98,241,116,56,89,13,120,77,3,0 // vmulps 0x34d78(%rip){1to8},%ymm1,%ymm1 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 98,241,109,56,219,21,114,77,3,0 // vpandd 0x34d72(%rip){1to8},%ymm2,%ymm2 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 98,241,108,56,89,21,236,90,3,0 // vmulps 0x35aec(%rip){1to8},%ymm2,%ymm2 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 98,241,108,56,89,21,104,77,3,0 // vmulps 0x34d68(%rip){1to8},%ymm2,%ymm2 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,81,90,3,0 // vbroadcastss 0x35a51(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,205,76,3,0 // vbroadcastss 0x34ccd(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -13271,17 +13271,17 @@ _sk_load_565_dst_skx:
.byte 117,96 // jne 3e1b <_sk_load_565_dst_skx+0x7b>
.byte 196,129,122,111,36,88 // vmovdqu (%r8,%r11,2),%xmm4
.byte 196,226,125,51,244 // vpmovzxwd %xmm4,%ymm6
- .byte 98,241,77,56,219,37,228,89,3,0 // vpandd 0x359e4(%rip){1to8},%ymm6,%ymm4 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 98,241,77,56,219,37,96,76,3,0 // vpandd 0x34c60(%rip){1to8},%ymm6,%ymm4 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 98,241,92,56,89,37,218,89,3,0 // vmulps 0x359da(%rip){1to8},%ymm4,%ymm4 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
- .byte 98,241,77,56,219,45,212,89,3,0 // vpandd 0x359d4(%rip){1to8},%ymm6,%ymm5 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 98,241,92,56,89,37,86,76,3,0 // vmulps 0x34c56(%rip){1to8},%ymm4,%ymm4 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 98,241,77,56,219,45,80,76,3,0 // vpandd 0x34c50(%rip){1to8},%ymm6,%ymm5 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 98,241,84,56,89,45,202,89,3,0 // vmulps 0x359ca(%rip){1to8},%ymm5,%ymm5 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
- .byte 98,241,77,56,219,53,196,89,3,0 // vpandd 0x359c4(%rip){1to8},%ymm6,%ymm6 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 98,241,84,56,89,45,70,76,3,0 // vmulps 0x34c46(%rip){1to8},%ymm5,%ymm5 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 98,241,77,56,219,53,64,76,3,0 // vpandd 0x34c40(%rip){1to8},%ymm6,%ymm6 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 98,241,76,56,89,53,186,89,3,0 // vmulps 0x359ba(%rip){1to8},%ymm6,%ymm6 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 98,241,76,56,89,53,54,76,3,0 // vmulps 0x34c36(%rip){1to8},%ymm6,%ymm6 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,31,89,3,0 // vbroadcastss 0x3591f(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,155,75,3,0 // vbroadcastss 0x34b9b(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -13389,17 +13389,17 @@ _sk_gather_565_skx:
.byte 65,15,183,4,88 // movzwl (%r8,%rbx,2),%eax
.byte 197,249,196,192,7 // vpinsrw $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,51,208 // vpmovzxwd %xmm0,%ymm2
- .byte 98,241,109,56,219,5,255,87,3,0 // vpandd 0x357ff(%rip){1to8},%ymm2,%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 98,241,109,56,219,5,123,74,3,0 // vpandd 0x34a7b(%rip){1to8},%ymm2,%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,245,87,3,0 // vmulps 0x357f5(%rip){1to8},%ymm0,%ymm0 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
- .byte 98,241,109,56,219,13,239,87,3,0 // vpandd 0x357ef(%rip){1to8},%ymm2,%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 98,241,124,56,89,5,113,74,3,0 // vmulps 0x34a71(%rip){1to8},%ymm0,%ymm0 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 98,241,109,56,219,13,107,74,3,0 // vpandd 0x34a6b(%rip){1to8},%ymm2,%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 98,241,116,56,89,13,229,87,3,0 // vmulps 0x357e5(%rip){1to8},%ymm1,%ymm1 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
- .byte 98,241,109,56,219,21,223,87,3,0 // vpandd 0x357df(%rip){1to8},%ymm2,%ymm2 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 98,241,116,56,89,13,97,74,3,0 // vmulps 0x34a61(%rip){1to8},%ymm1,%ymm1 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 98,241,109,56,219,21,91,74,3,0 // vpandd 0x34a5b(%rip){1to8},%ymm2,%ymm2 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 98,241,108,56,89,21,213,87,3,0 // vmulps 0x357d5(%rip){1to8},%ymm2,%ymm2 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 98,241,108,56,89,21,81,74,3,0 // vmulps 0x34a51(%rip){1to8},%ymm2,%ymm2 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,58,87,3,0 // vbroadcastss 0x3573a(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,182,73,3,0 // vbroadcastss 0x349b6(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
.byte 65,94 // pop %r14
@@ -13417,11 +13417,11 @@ _sk_store_565_skx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,2,88,3,0 // vbroadcastss 0x35802(%rip),%ymm8 # 39828 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
+ .byte 196,98,125,24,5,126,74,3,0 // vbroadcastss 0x34a7e(%rip),%ymm8 # 38aa4 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,53,114,241,11 // vpslld $0xb,%ymm9,%ymm9
- .byte 98,113,116,56,89,21,236,87,3,0 // vmulps 0x357ec(%rip){1to8},%ymm1,%ymm10 # 3982c <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
+ .byte 98,113,116,56,89,21,104,74,3,0 // vmulps 0x34a68(%rip){1to8},%ymm1,%ymm10 # 38aa8 <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
.byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
.byte 196,193,45,114,242,5 // vpslld $0x5,%ymm10,%ymm10
.byte 196,65,45,235,201 // vpor %ymm9,%ymm10,%ymm9
@@ -13468,7 +13468,7 @@ _sk_store_565_skx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,225 // jmpq ffffffffe20040f8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe1fcac70>
+ .byte 233,255,255,255,225 // jmpq ffffffffe20040f8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe1fcb9f4>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
@@ -13491,18 +13491,18 @@ _sk_load_4444_skx:
.byte 117,111 // jne 418a <_sk_load_4444_skx+0x8a>
.byte 196,129,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm0
.byte 196,226,125,51,216 // vpmovzxwd %xmm0,%ymm3
- .byte 98,241,101,56,219,5,0,87,3,0 // vpandd 0x35700(%rip){1to8},%ymm3,%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 98,241,101,56,219,5,124,73,3,0 // vpandd 0x3497c(%rip){1to8},%ymm3,%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,246,86,3,0 // vmulps 0x356f6(%rip){1to8},%ymm0,%ymm0 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
- .byte 98,241,101,56,219,13,240,86,3,0 // vpandd 0x356f0(%rip){1to8},%ymm3,%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 98,241,124,56,89,5,114,73,3,0 // vmulps 0x34972(%rip){1to8},%ymm0,%ymm0 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 98,241,101,56,219,13,108,73,3,0 // vpandd 0x3496c(%rip){1to8},%ymm3,%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 98,241,116,56,89,13,230,86,3,0 // vmulps 0x356e6(%rip){1to8},%ymm1,%ymm1 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
- .byte 98,241,101,56,219,21,224,86,3,0 // vpandd 0x356e0(%rip){1to8},%ymm3,%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 98,241,116,56,89,13,98,73,3,0 // vmulps 0x34962(%rip){1to8},%ymm1,%ymm1 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 98,241,101,56,219,21,92,73,3,0 // vpandd 0x3495c(%rip){1to8},%ymm3,%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 98,241,108,56,89,21,214,86,3,0 // vmulps 0x356d6(%rip){1to8},%ymm2,%ymm2 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
- .byte 98,241,101,56,219,29,208,86,3,0 // vpandd 0x356d0(%rip){1to8},%ymm3,%ymm3 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 98,241,108,56,89,21,82,73,3,0 // vmulps 0x34952(%rip){1to8},%ymm2,%ymm2 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 98,241,101,56,219,29,76,73,3,0 // vpandd 0x3494c(%rip){1to8},%ymm3,%ymm3 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 98,241,100,56,89,29,198,86,3,0 // vmulps 0x356c6(%rip){1to8},%ymm3,%ymm3 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 98,241,100,56,89,29,66,73,3,0 // vmulps 0x34942(%rip){1to8},%ymm3,%ymm3 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -13572,18 +13572,18 @@ _sk_load_4444_dst_skx:
.byte 117,111 // jne 42ca <_sk_load_4444_dst_skx+0x8a>
.byte 196,129,122,111,36,88 // vmovdqu (%r8,%r11,2),%xmm4
.byte 196,226,125,51,252 // vpmovzxwd %xmm4,%ymm7
- .byte 98,241,69,56,219,37,192,85,3,0 // vpandd 0x355c0(%rip){1to8},%ymm7,%ymm4 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 98,241,69,56,219,37,60,72,3,0 // vpandd 0x3483c(%rip){1to8},%ymm7,%ymm4 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 98,241,92,56,89,37,182,85,3,0 // vmulps 0x355b6(%rip){1to8},%ymm4,%ymm4 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
- .byte 98,241,69,56,219,45,176,85,3,0 // vpandd 0x355b0(%rip){1to8},%ymm7,%ymm5 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 98,241,92,56,89,37,50,72,3,0 // vmulps 0x34832(%rip){1to8},%ymm4,%ymm4 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 98,241,69,56,219,45,44,72,3,0 // vpandd 0x3482c(%rip){1to8},%ymm7,%ymm5 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 98,241,84,56,89,45,166,85,3,0 // vmulps 0x355a6(%rip){1to8},%ymm5,%ymm5 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
- .byte 98,241,69,56,219,53,160,85,3,0 // vpandd 0x355a0(%rip){1to8},%ymm7,%ymm6 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 98,241,84,56,89,45,34,72,3,0 // vmulps 0x34822(%rip){1to8},%ymm5,%ymm5 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 98,241,69,56,219,53,28,72,3,0 // vpandd 0x3481c(%rip){1to8},%ymm7,%ymm6 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 98,241,76,56,89,53,150,85,3,0 // vmulps 0x35596(%rip){1to8},%ymm6,%ymm6 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
- .byte 98,241,69,56,219,61,144,85,3,0 // vpandd 0x35590(%rip){1to8},%ymm7,%ymm7 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 98,241,76,56,89,53,18,72,3,0 // vmulps 0x34812(%rip){1to8},%ymm6,%ymm6 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 98,241,69,56,219,61,12,72,3,0 // vpandd 0x3480c(%rip){1to8},%ymm7,%ymm7 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 98,241,68,56,89,61,134,85,3,0 // vmulps 0x35586(%rip){1to8},%ymm7,%ymm7 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 98,241,68,56,89,61,2,72,3,0 // vmulps 0x34802(%rip){1to8},%ymm7,%ymm7 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -13692,18 +13692,18 @@ _sk_gather_4444_skx:
.byte 65,15,183,4,88 // movzwl (%r8,%rbx,2),%eax
.byte 197,249,196,192,7 // vpinsrw $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,51,216 // vpmovzxwd %xmm0,%ymm3
- .byte 98,241,101,56,219,5,203,83,3,0 // vpandd 0x353cb(%rip){1to8},%ymm3,%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 98,241,101,56,219,5,71,70,3,0 // vpandd 0x34647(%rip){1to8},%ymm3,%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 98,241,124,56,89,5,193,83,3,0 // vmulps 0x353c1(%rip){1to8},%ymm0,%ymm0 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
- .byte 98,241,101,56,219,13,187,83,3,0 // vpandd 0x353bb(%rip){1to8},%ymm3,%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 98,241,124,56,89,5,61,70,3,0 // vmulps 0x3463d(%rip){1to8},%ymm0,%ymm0 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 98,241,101,56,219,13,55,70,3,0 // vpandd 0x34637(%rip){1to8},%ymm3,%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 98,241,116,56,89,13,177,83,3,0 // vmulps 0x353b1(%rip){1to8},%ymm1,%ymm1 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
- .byte 98,241,101,56,219,21,171,83,3,0 // vpandd 0x353ab(%rip){1to8},%ymm3,%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 98,241,116,56,89,13,45,70,3,0 // vmulps 0x3462d(%rip){1to8},%ymm1,%ymm1 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 98,241,101,56,219,21,39,70,3,0 // vpandd 0x34627(%rip){1to8},%ymm3,%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 98,241,108,56,89,21,161,83,3,0 // vmulps 0x353a1(%rip){1to8},%ymm2,%ymm2 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
- .byte 98,241,101,56,219,29,155,83,3,0 // vpandd 0x3539b(%rip){1to8},%ymm3,%ymm3 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 98,241,108,56,89,21,29,70,3,0 // vmulps 0x3461d(%rip){1to8},%ymm2,%ymm2 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 98,241,101,56,219,29,23,70,3,0 // vpandd 0x34617(%rip){1to8},%ymm3,%ymm3 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 98,241,100,56,89,29,145,83,3,0 // vmulps 0x35391(%rip){1to8},%ymm3,%ymm3 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 98,241,100,56,89,29,13,70,3,0 // vmulps 0x3460d(%rip){1to8},%ymm3,%ymm3 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
@@ -13722,7 +13722,7 @@ _sk_store_4444_skx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,107,83,3,0 // vbroadcastss 0x3536b(%rip),%ymm8 # 39850 <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
+ .byte 196,98,125,24,5,231,69,3,0 // vbroadcastss 0x345e7(%rip),%ymm8 # 38acc <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,53,114,241,12 // vpslld $0xc,%ymm9,%ymm9
@@ -13777,7 +13777,7 @@ _sk_store_4444_skx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,224 // callq ffffffffe10045c8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fcb140>
+ .byte 232,255,255,255,224 // callq ffffffffe10045c8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fcbec4>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
@@ -13803,7 +13803,7 @@ _sk_load_8888_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,241,127,169,111,195 // vmovdqu8 %ymm3,%ymm0{%k1}{z}
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,162,81,3,0 // vbroadcastss 0x351a2(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,30,68,3,0 // vbroadcastss 0x3441e(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,245,114,211,8 // vpsrld $0x8,%ymm3,%ymm1
.byte 98,241,127,169,111,201 // vmovdqu8 %ymm1,%ymm1{%k1}{z}
@@ -13890,7 +13890,7 @@ _sk_load_8888_dst_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,241,127,169,111,231 // vmovdqu8 %ymm7,%ymm4{%k1}{z}
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,98,125,24,5,94,80,3,0 // vbroadcastss 0x3505e(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,218,66,3,0 // vbroadcastss 0x342da(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,92,89,224 // vmulps %ymm8,%ymm4,%ymm4
.byte 197,213,114,215,8 // vpsrld $0x8,%ymm7,%ymm5
.byte 98,241,127,169,111,237 // vmovdqu8 %ymm5,%ymm5{%k1}{z}
@@ -13984,7 +13984,7 @@ _sk_gather_8888_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,209,127,169,111,192 // vmovdqu8 %ymm8,%ymm0{%k1}{z}
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,29,241,78,3,0 // vbroadcastss 0x34ef1(%rip),%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,29,109,65,3,0 // vbroadcastss 0x3416d(%rip),%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
.byte 196,193,117,114,208,8 // vpsrld $0x8,%ymm8,%ymm1
.byte 98,241,127,169,111,201 // vmovdqu8 %ymm1,%ymm1{%k1}{z}
@@ -14011,7 +14011,7 @@ _sk_store_8888_skx:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,70,78,3,0 // vbroadcastss 0x34e46(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,194,64,3,0 // vbroadcastss 0x340c2(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -14067,7 +14067,7 @@ _sk_store_8888_skx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,219 // jmpq ffffffffdc004a00 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfcb578>
+ .byte 233,255,255,255,219 // jmpq ffffffffdc004a00 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfcc2fc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,205 // dec %ebp
@@ -14093,7 +14093,7 @@ _sk_load_bgra_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,241,127,169,111,195 // vmovdqu8 %ymm3,%ymm0{%k1}{z}
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,106,77,3,0 // vbroadcastss 0x34d6a(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,230,63,3,0 // vbroadcastss 0x33fe6(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,208 // vmulps %ymm8,%ymm0,%ymm2
.byte 197,253,114,211,8 // vpsrld $0x8,%ymm3,%ymm0
.byte 98,241,127,169,111,192 // vmovdqu8 %ymm0,%ymm0{%k1}{z}
@@ -14180,7 +14180,7 @@ _sk_load_bgra_dst_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,241,127,169,111,231 // vmovdqu8 %ymm7,%ymm4{%k1}{z}
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,98,125,24,5,38,76,3,0 // vbroadcastss 0x34c26(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,162,62,3,0 // vbroadcastss 0x33ea2(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,92,89,240 // vmulps %ymm8,%ymm4,%ymm6
.byte 197,221,114,215,8 // vpsrld $0x8,%ymm7,%ymm4
.byte 98,241,127,169,111,228 // vmovdqu8 %ymm4,%ymm4{%k1}{z}
@@ -14274,7 +14274,7 @@ _sk_gather_bgra_skx:
.byte 197,251,146,200 // kmovd %eax,%k1
.byte 98,209,127,169,111,192 // vmovdqu8 %ymm8,%ymm0{%k1}{z}
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,29,185,74,3,0 // vbroadcastss 0x34ab9(%rip),%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,29,53,61,3,0 // vbroadcastss 0x33d35(%rip),%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,211 // vmulps %ymm3,%ymm0,%ymm2
.byte 196,193,125,114,208,8 // vpsrld $0x8,%ymm8,%ymm0
.byte 98,241,127,169,111,192 // vmovdqu8 %ymm0,%ymm0{%k1}{z}
@@ -14301,7 +14301,7 @@ _sk_store_bgra_skx:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,14,74,3,0 // vbroadcastss 0x34a0e(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,138,60,3,0 // vbroadcastss 0x33c8a(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,108,89,200 // vmulps %ymm8,%ymm2,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -14357,7 +14357,7 @@ _sk_store_bgra_skx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,219 // jmpq ffffffffdc004e38 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfcb9b0>
+ .byte 233,255,255,255,219 // jmpq ffffffffdc004e38 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfcc734>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,205 // dec %ebp
@@ -14620,7 +14620,7 @@ _sk_load_u16_be_skx:
.byte 197,241,235,192 // vpor %xmm0,%xmm1,%xmm0
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,21,119,69,3,0 // vbroadcastss 0x34577(%rip),%ymm10 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,21,243,55,3,0 // vbroadcastss 0x337f3(%rip),%ymm10 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 197,185,109,202 // vpunpckhqdq %xmm2,%xmm8,%xmm1
.byte 197,233,113,241,8 // vpsllw $0x8,%xmm1,%xmm2
@@ -14711,7 +14711,7 @@ _sk_load_rgb_u16_be_skx:
.byte 197,241,235,192 // vpor %xmm0,%xmm1,%xmm0
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,233,67,3,0 // vbroadcastss 0x343e9(%rip),%ymm9 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,13,101,54,3,0 // vbroadcastss 0x33665(%rip),%ymm9 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,185,109,203 // vpunpckhqdq %xmm3,%xmm8,%xmm1
.byte 197,225,113,241,8 // vpsllw $0x8,%xmm1,%xmm3
@@ -14727,7 +14727,7 @@ _sk_load_rgb_u16_be_skx:
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 196,193,108,89,209 // vmulps %ymm9,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,9,67,3,0 // vbroadcastss 0x34309(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,133,53,3,0 // vbroadcastss 0x33585(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 196,193,121,110,4,64 // vmovd (%r8,%rax,2),%xmm0
.byte 196,65,49,87,201 // vxorpd %xmm9,%xmm9,%xmm9
@@ -14803,7 +14803,7 @@ _sk_store_u16_be_skx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 73,99,193 // movslq %r9d,%rax
- .byte 196,98,125,24,5,210,66,3,0 // vbroadcastss 0x342d2(%rip),%ymm8 # 39854 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
+ .byte 196,98,125,24,5,78,53,3,0 // vbroadcastss 0x3354e(%rip),%ymm8 # 38ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
@@ -15145,7 +15145,7 @@ _sk_mirror_x_skx:
.byte 196,65,124,92,218 // vsubps %ymm10,%ymm0,%ymm11
.byte 196,193,58,88,192 // vaddss %xmm8,%xmm8,%xmm0
.byte 196,98,125,24,192 // vbroadcastss %xmm0,%ymm8
- .byte 197,178,89,5,63,60,3,0 // vmulss 0x33c3f(%rip),%xmm9,%xmm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,5,187,46,3,0 // vmulss 0x32ebb(%rip),%xmm9,%xmm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,89,192 // vmulps %ymm0,%ymm11,%ymm0
.byte 196,227,125,8,192,1 // vroundps $0x1,%ymm0,%ymm0
@@ -15168,7 +15168,7 @@ _sk_mirror_y_skx:
.byte 196,65,116,92,218 // vsubps %ymm10,%ymm1,%ymm11
.byte 196,193,58,88,200 // vaddss %xmm8,%xmm8,%xmm1
.byte 196,98,125,24,193 // vbroadcastss %xmm1,%ymm8
- .byte 197,178,89,13,238,59,3,0 // vmulss 0x33bee(%rip),%xmm9,%xmm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,13,106,46,3,0 // vmulss 0x32e6a(%rip),%xmm9,%xmm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,226,125,24,201 // vbroadcastss %xmm1,%ymm1
.byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
.byte 196,227,125,8,201,1 // vroundps $0x1,%ymm1,%ymm1
@@ -15186,7 +15186,7 @@ FUNCTION(_sk_clamp_x_1_skx)
_sk_clamp_x_1_skx:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 98,241,124,56,93,5,181,59,3,0 // vminps 0x33bb5(%rip){1to8},%ymm0,%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,93,5,49,46,3,0 // vminps 0x32e31(%rip){1to8},%ymm0,%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15198,7 +15198,7 @@ _sk_repeat_x_1_skx:
.byte 196,193,124,92,192 // vsubps %ymm8,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 98,241,124,56,93,5,147,59,3,0 // vminps 0x33b93(%rip){1to8},%ymm0,%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,93,5,15,46,3,0 // vminps 0x32e0f(%rip){1to8},%ymm0,%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15206,9 +15206,9 @@ HIDDEN _sk_mirror_x_1_skx
.globl _sk_mirror_x_1_skx
FUNCTION(_sk_mirror_x_1_skx)
_sk_mirror_x_1_skx:
- .byte 196,98,125,24,5,158,59,3,0 // vbroadcastss 0x33b9e(%rip),%ymm8 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,5,26,46,3,0 // vbroadcastss 0x32e1a(%rip),%ymm8 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
- .byte 98,113,124,56,89,13,115,59,3,0 // vmulps 0x33b73(%rip){1to8},%ymm0,%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,113,124,56,89,13,239,45,3,0 // vmulps 0x32def(%rip){1to8},%ymm0,%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
.byte 196,65,52,88,201 // vaddps %ymm9,%ymm9,%ymm9
.byte 196,193,124,92,193 // vsubps %ymm9,%ymm0,%ymm0
@@ -15217,7 +15217,7 @@ _sk_mirror_x_1_skx:
.byte 197,60,92,200 // vsubps %ymm0,%ymm8,%ymm9
.byte 197,180,84,192 // vandps %ymm0,%ymm9,%ymm0
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 98,241,124,56,93,5,71,59,3,0 // vminps 0x33b47(%rip){1to8},%ymm0,%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,93,5,195,45,3,0 // vminps 0x32dc3(%rip){1to8},%ymm0,%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15225,9 +15225,9 @@ HIDDEN _sk_luminance_to_alpha_skx
.globl _sk_luminance_to_alpha_skx
FUNCTION(_sk_luminance_to_alpha_skx)
_sk_luminance_to_alpha_skx:
- .byte 98,241,116,56,89,29,89,60,3,0 // vmulps 0x33c59(%rip){1to8},%ymm1,%ymm3 # 39858 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
- .byte 98,242,125,56,184,29,83,60,3,0 // vfmadd231ps 0x33c53(%rip){1to8},%ymm0,%ymm3 # 3985c <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
- .byte 98,242,109,56,184,29,77,60,3,0 // vfmadd231ps 0x33c4d(%rip){1to8},%ymm2,%ymm3 # 39860 <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
+ .byte 98,241,116,56,89,29,213,46,3,0 // vmulps 0x32ed5(%rip){1to8},%ymm1,%ymm3 # 38ad4 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
+ .byte 98,242,125,56,184,29,207,46,3,0 // vfmadd231ps 0x32ecf(%rip){1to8},%ymm0,%ymm3 # 38ad8 <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
+ .byte 98,242,109,56,184,29,201,46,3,0 // vfmadd231ps 0x32ec9(%rip){1to8},%ymm2,%ymm3 # 38adc <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
@@ -15476,7 +15476,7 @@ _sk_gradient_skx:
.byte 76,139,72,72 // mov 0x48(%rax),%r9
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 65,186,1,0,0,0 // mov $0x1,%r10d
- .byte 196,226,125,24,21,86,55,3,0 // vbroadcastss 0x33756(%rip),%ymm2 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,226,125,24,21,210,41,3,0 // vbroadcastss 0x329d2(%rip),%ymm2 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 196,65,53,239,201 // vpxor %ymm9,%ymm9,%ymm9
.byte 196,130,125,24,28,145 // vbroadcastss (%r9,%r10,4),%ymm3
.byte 98,241,100,40,194,192,2 // vcmpleps %ymm0,%ymm3,%k0
@@ -15576,24 +15576,24 @@ _sk_xy_to_unit_angle_skx:
.byte 196,65,52,95,226 // vmaxps %ymm10,%ymm9,%ymm12
.byte 196,65,36,94,220 // vdivps %ymm12,%ymm11,%ymm11
.byte 196,65,36,89,227 // vmulps %ymm11,%ymm11,%ymm12
- .byte 196,98,125,24,45,216,54,3,0 // vbroadcastss 0x336d8(%rip),%ymm13 # 39864 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
- .byte 98,114,29,56,168,45,210,54,3,0 // vfmadd213ps 0x336d2(%rip){1to8},%ymm12,%ymm13 # 39868 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
- .byte 98,114,29,56,168,45,204,54,3,0 // vfmadd213ps 0x336cc(%rip){1to8},%ymm12,%ymm13 # 3986c <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
- .byte 98,114,29,56,168,45,198,54,3,0 // vfmadd213ps 0x336c6(%rip){1to8},%ymm12,%ymm13 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
+ .byte 196,98,125,24,45,84,41,3,0 // vbroadcastss 0x32954(%rip),%ymm13 # 38ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
+ .byte 98,114,29,56,168,45,78,41,3,0 // vfmadd213ps 0x3294e(%rip){1to8},%ymm12,%ymm13 # 38ae4 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
+ .byte 98,114,29,56,168,45,72,41,3,0 // vfmadd213ps 0x32948(%rip){1to8},%ymm12,%ymm13 # 38ae8 <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
+ .byte 98,114,29,56,168,45,66,41,3,0 // vfmadd213ps 0x32942(%rip){1to8},%ymm12,%ymm13 # 38aec <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
.byte 196,65,36,89,221 // vmulps %ymm13,%ymm11,%ymm11
.byte 98,209,52,40,194,194,1 // vcmpltps %ymm10,%ymm9,%k0
.byte 98,114,126,40,56,200 // vpmovm2d %k0,%ymm9
- .byte 196,98,125,24,21,175,54,3,0 // vbroadcastss 0x336af(%rip),%ymm10 # 39874 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
+ .byte 196,98,125,24,21,43,41,3,0 // vbroadcastss 0x3292b(%rip),%ymm10 # 38af0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
.byte 196,65,44,92,211 // vsubps %ymm11,%ymm10,%ymm10
.byte 196,67,37,74,202,144 // vblendvps %ymm9,%ymm10,%ymm11,%ymm9
.byte 98,209,124,40,194,192,1 // vcmpltps %ymm8,%ymm0,%k0
.byte 98,242,126,40,56,192 // vpmovm2d %k0,%ymm0
- .byte 196,98,125,24,21,78,53,3,0 // vbroadcastss 0x3354e(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,202,39,3,0 // vbroadcastss 0x327ca(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,92,209 // vsubps %ymm9,%ymm10,%ymm10
.byte 196,195,53,74,194,0 // vblendvps %ymm0,%ymm10,%ymm9,%ymm0
.byte 98,209,116,40,194,192,1 // vcmpltps %ymm8,%ymm1,%k0
.byte 98,114,126,40,56,200 // vpmovm2d %k0,%ymm9
- .byte 196,98,125,24,21,49,53,3,0 // vbroadcastss 0x33531(%rip),%ymm10 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,21,173,39,3,0 // vbroadcastss 0x327ad(%rip),%ymm10 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,44,92,208 // vsubps %ymm0,%ymm10,%ymm10
.byte 196,195,125,74,194,144 // vblendvps %ymm9,%ymm10,%ymm0,%ymm0
.byte 98,209,124,40,194,192,3 // vcmpunordps %ymm8,%ymm0,%k0
@@ -15622,20 +15622,20 @@ _sk_xy_to_2pt_conical_quadratic_max_skx:
.byte 197,50,89,80,76 // vmulss 0x4c(%rax),%xmm9,%xmm10
.byte 196,66,125,24,210 // vbroadcastss %xmm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 98,113,44,56,89,21,25,54,3,0 // vmulps 0x33619(%rip){1to8},%ymm10,%ymm10 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 98,113,44,56,89,21,149,40,3,0 // vmulps 0x32895(%rip){1to8},%ymm10,%ymm10 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
.byte 196,98,125,184,216 // vfmadd231ps %ymm0,%ymm0,%ymm11
.byte 196,193,50,89,193 // vmulss %xmm9,%xmm9,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 98,113,60,56,89,5,252,53,3,0 // vmulps 0x335fc(%rip){1to8},%ymm8,%ymm8 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 98,113,60,56,89,5,120,40,3,0 // vmulps 0x32878(%rip){1to8},%ymm8,%ymm8 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,194,45,184,194 // vfmadd231ps %ymm10,%ymm10,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 98,113,44,56,87,13,227,53,3,0 // vxorps 0x335e3(%rip){1to8},%ymm10,%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 98,113,44,56,87,13,95,40,3,0 // vxorps 0x3285f(%rip){1to8},%ymm10,%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 98,113,60,56,89,5,136,52,3,0 // vmulps 0x33488(%rip){1to8},%ymm8,%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,113,60,56,89,5,4,39,3,0 // vmulps 0x32704(%rip){1to8},%ymm8,%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,89,208 // vmulps %ymm8,%ymm10,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
@@ -15653,20 +15653,20 @@ _sk_xy_to_2pt_conical_quadratic_min_skx:
.byte 197,50,89,80,76 // vmulss 0x4c(%rax),%xmm9,%xmm10
.byte 196,66,125,24,210 // vbroadcastss %xmm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 98,113,44,56,89,21,145,53,3,0 // vmulps 0x33591(%rip){1to8},%ymm10,%ymm10 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 98,113,44,56,89,21,13,40,3,0 // vmulps 0x3280d(%rip){1to8},%ymm10,%ymm10 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
.byte 196,98,125,184,216 // vfmadd231ps %ymm0,%ymm0,%ymm11
.byte 196,193,50,89,193 // vmulss %xmm9,%xmm9,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 98,113,60,56,89,5,116,53,3,0 // vmulps 0x33574(%rip){1to8},%ymm8,%ymm8 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 98,113,60,56,89,5,240,39,3,0 // vmulps 0x327f0(%rip){1to8},%ymm8,%ymm8 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,194,45,184,194 // vfmadd231ps %ymm10,%ymm10,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 98,113,44,56,87,13,91,53,3,0 // vxorps 0x3355b(%rip){1to8},%ymm10,%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 98,113,44,56,87,13,215,39,3,0 // vxorps 0x327d7(%rip){1to8},%ymm10,%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 98,113,60,56,89,5,0,52,3,0 // vmulps 0x33400(%rip){1to8},%ymm8,%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,113,60,56,89,5,124,38,3,0 // vmulps 0x3267c(%rip){1to8},%ymm8,%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,89,208 // vmulps %ymm8,%ymm10,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
@@ -15683,13 +15683,13 @@ _sk_xy_to_2pt_conical_linear_skx:
.byte 197,58,89,72,76 // vmulss 0x4c(%rax),%xmm8,%xmm9
.byte 196,66,125,24,201 // vbroadcastss %xmm9,%ymm9
.byte 197,52,88,200 // vaddps %ymm0,%ymm9,%ymm9
- .byte 98,113,52,56,89,13,15,53,3,0 // vmulps 0x3350f(%rip){1to8},%ymm9,%ymm9 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 98,113,52,56,89,13,139,39,3,0 // vmulps 0x3278b(%rip){1to8},%ymm9,%ymm9 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 197,116,89,209 // vmulps %ymm1,%ymm1,%ymm10
.byte 196,98,125,184,208 // vfmadd231ps %ymm0,%ymm0,%ymm10
.byte 196,193,58,89,192 // vmulss %xmm8,%xmm8,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,172,92,192 // vsubps %ymm0,%ymm10,%ymm0
- .byte 98,241,124,56,87,5,246,52,3,0 // vxorps 0x334f6(%rip){1to8},%ymm0,%ymm0 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 98,241,124,56,87,5,114,39,3,0 // vxorps 0x32772(%rip){1to8},%ymm0,%ymm0 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,193,124,94,193 // vdivps %ymm9,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15731,7 +15731,7 @@ HIDDEN _sk_save_xy_skx
FUNCTION(_sk_save_xy_skx)
_sk_save_xy_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,55,51,3,0 // vbroadcastss 0x33337(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,179,37,3,0 // vbroadcastss 0x325b3(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,124,88,200 // vaddps %ymm8,%ymm0,%ymm9
.byte 196,67,125,8,209,1 // vroundps $0x1,%ymm9,%ymm10
.byte 196,65,52,92,202 // vsubps %ymm10,%ymm9,%ymm9
@@ -15766,8 +15766,8 @@ FUNCTION(_sk_bilinear_nx_skx)
_sk_bilinear_nx_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,0 // vmovups (%rax),%ymm0
- .byte 98,241,124,56,88,5,12,52,3,0 // vaddps 0x3340c(%rip){1to8},%ymm0,%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
- .byte 196,98,125,24,5,183,50,3,0 // vbroadcastss 0x332b7(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,88,5,136,38,3,0 // vaddps 0x32688(%rip){1to8},%ymm0,%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,98,125,24,5,51,37,3,0 // vbroadcastss 0x32533(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -15780,7 +15780,7 @@ _sk_bilinear_px_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,0 // vmovups (%rax),%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
- .byte 98,241,124,56,88,5,135,50,3,0 // vaddps 0x33287(%rip){1to8},%ymm0,%ymm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,241,124,56,88,5,3,37,3,0 // vaddps 0x32503(%rip){1to8},%ymm0,%ymm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15791,8 +15791,8 @@ FUNCTION(_sk_bilinear_ny_skx)
_sk_bilinear_ny_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,72,64 // vmovups 0x40(%rax),%ymm1
- .byte 98,241,116,56,88,13,186,51,3,0 // vaddps 0x333ba(%rip){1to8},%ymm1,%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
- .byte 196,98,125,24,5,101,50,3,0 // vbroadcastss 0x33265(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,116,56,88,13,54,38,3,0 // vaddps 0x32636(%rip){1to8},%ymm1,%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,98,125,24,5,225,36,3,0 // vbroadcastss 0x324e1(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -15805,7 +15805,7 @@ _sk_bilinear_py_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,72,64 // vmovups 0x40(%rax),%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
- .byte 98,241,116,56,88,13,52,50,3,0 // vaddps 0x33234(%rip){1to8},%ymm1,%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,241,116,56,88,13,176,36,3,0 // vaddps 0x324b0(%rip){1to8},%ymm1,%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15816,12 +15816,12 @@ FUNCTION(_sk_bicubic_n3x_skx)
_sk_bicubic_n3x_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,0 // vmovups (%rax),%ymm0
- .byte 98,241,124,56,88,5,108,51,3,0 // vaddps 0x3336c(%rip){1to8},%ymm0,%ymm0 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
- .byte 196,98,125,24,5,19,50,3,0 // vbroadcastss 0x33213(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,88,5,232,37,3,0 // vaddps 0x325e8(%rip){1to8},%ymm0,%ymm0 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,98,125,24,5,143,36,3,0 // vbroadcastss 0x3248f(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,81,51,3,0 // vbroadcastss 0x33351(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
- .byte 98,114,61,56,168,21,103,50,3,0 // vfmadd213ps 0x33267(%rip){1to8},%ymm8,%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,205,37,3,0 // vbroadcastss 0x325cd(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 98,114,61,56,168,21,227,36,3,0 // vfmadd213ps 0x324e3(%rip){1to8},%ymm8,%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,44,89,193 // vmulps %ymm9,%ymm10,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -15833,13 +15833,13 @@ FUNCTION(_sk_bicubic_n1x_skx)
_sk_bicubic_n1x_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,0 // vmovups (%rax),%ymm0
- .byte 98,241,124,56,88,5,30,51,3,0 // vaddps 0x3331e(%rip){1to8},%ymm0,%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
- .byte 196,98,125,24,5,201,49,3,0 // vbroadcastss 0x331c9(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,124,56,88,5,154,37,3,0 // vaddps 0x3259a(%rip){1to8},%ymm0,%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,98,125,24,5,69,36,3,0 // vbroadcastss 0x32445(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,16,51,3,0 // vbroadcastss 0x33310(%rip),%ymm9 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
- .byte 98,114,61,56,168,13,10,51,3,0 // vfmadd213ps 0x3330a(%rip){1to8},%ymm8,%ymm9 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 98,114,61,56,168,13,160,49,3,0 // vfmadd213ps 0x331a0(%rip){1to8},%ymm8,%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,61,56,168,13,250,50,3,0 // vfmadd213ps 0x332fa(%rip){1to8},%ymm8,%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,140,37,3,0 // vbroadcastss 0x3258c(%rip),%ymm9 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 98,114,61,56,168,13,134,37,3,0 // vfmadd213ps 0x32586(%rip){1to8},%ymm8,%ymm9 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 98,114,61,56,168,13,28,36,3,0 // vfmadd213ps 0x3241c(%rip){1to8},%ymm8,%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,61,56,168,13,118,37,3,0 // vfmadd213ps 0x32576(%rip){1to8},%ymm8,%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 197,124,17,136,0,1,0,0 // vmovups %ymm9,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15849,13 +15849,13 @@ HIDDEN _sk_bicubic_p1x_skx
FUNCTION(_sk_bicubic_p1x_skx)
_sk_bicubic_p1x_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,127,49,3,0 // vbroadcastss 0x3317f(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,251,35,3,0 // vbroadcastss 0x323fb(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,0 // vaddps (%rax),%ymm8,%ymm0
.byte 197,124,16,136,128,0,0,0 // vmovups 0x80(%rax),%ymm9
- .byte 196,98,125,24,21,198,50,3,0 // vbroadcastss 0x332c6(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
- .byte 98,114,53,56,168,21,192,50,3,0 // vfmadd213ps 0x332c0(%rip){1to8},%ymm9,%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,66,37,3,0 // vbroadcastss 0x32542(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 98,114,53,56,168,21,60,37,3,0 // vfmadd213ps 0x3253c(%rip){1to8},%ymm9,%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,66,53,168,208 // vfmadd213ps %ymm8,%ymm9,%ymm10
- .byte 98,114,53,56,168,21,181,50,3,0 // vfmadd213ps 0x332b5(%rip){1to8},%ymm9,%ymm10 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 98,114,53,56,168,21,49,37,3,0 // vfmadd213ps 0x32531(%rip){1to8},%ymm9,%ymm10 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 197,124,17,144,0,1,0,0 // vmovups %ymm10,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15867,10 +15867,10 @@ _sk_bicubic_p3x_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,0 // vmovups (%rax),%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
- .byte 98,241,124,56,88,5,141,50,3,0 // vaddps 0x3328d(%rip){1to8},%ymm0,%ymm0 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 98,241,124,56,88,5,9,37,3,0 // vaddps 0x32509(%rip){1to8},%ymm0,%ymm0 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,119,50,3,0 // vbroadcastss 0x33277(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
- .byte 98,114,61,56,168,21,141,49,3,0 // vfmadd213ps 0x3318d(%rip){1to8},%ymm8,%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,243,36,3,0 // vbroadcastss 0x324f3(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 98,114,61,56,168,21,9,36,3,0 // vfmadd213ps 0x32409(%rip){1to8},%ymm8,%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,52,89,194 // vmulps %ymm10,%ymm9,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -15882,12 +15882,12 @@ FUNCTION(_sk_bicubic_n3y_skx)
_sk_bicubic_n3y_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,72,64 // vmovups 0x40(%rax),%ymm1
- .byte 98,241,116,56,88,13,71,50,3,0 // vaddps 0x33247(%rip){1to8},%ymm1,%ymm1 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
- .byte 196,98,125,24,5,238,48,3,0 // vbroadcastss 0x330ee(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,116,56,88,13,195,36,3,0 // vaddps 0x324c3(%rip){1to8},%ymm1,%ymm1 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,98,125,24,5,106,35,3,0 // vbroadcastss 0x3236a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,44,50,3,0 // vbroadcastss 0x3322c(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
- .byte 98,114,61,56,168,21,66,49,3,0 // vfmadd213ps 0x33142(%rip){1to8},%ymm8,%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,168,36,3,0 // vbroadcastss 0x324a8(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 98,114,61,56,168,21,190,35,3,0 // vfmadd213ps 0x323be(%rip){1to8},%ymm8,%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,44,89,193 // vmulps %ymm9,%ymm10,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -15899,13 +15899,13 @@ FUNCTION(_sk_bicubic_n1y_skx)
_sk_bicubic_n1y_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,72,64 // vmovups 0x40(%rax),%ymm1
- .byte 98,241,116,56,88,13,248,49,3,0 // vaddps 0x331f8(%rip){1to8},%ymm1,%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
- .byte 196,98,125,24,5,163,48,3,0 // vbroadcastss 0x330a3(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 98,241,116,56,88,13,116,36,3,0 // vaddps 0x32474(%rip){1to8},%ymm1,%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,98,125,24,5,31,35,3,0 // vbroadcastss 0x3231f(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,234,49,3,0 // vbroadcastss 0x331ea(%rip),%ymm9 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
- .byte 98,114,61,56,168,13,228,49,3,0 // vfmadd213ps 0x331e4(%rip){1to8},%ymm8,%ymm9 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 98,114,61,56,168,13,122,48,3,0 // vfmadd213ps 0x3307a(%rip){1to8},%ymm8,%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 98,114,61,56,168,13,212,49,3,0 // vfmadd213ps 0x331d4(%rip){1to8},%ymm8,%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,102,36,3,0 // vbroadcastss 0x32466(%rip),%ymm9 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 98,114,61,56,168,13,96,36,3,0 // vfmadd213ps 0x32460(%rip){1to8},%ymm8,%ymm9 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 98,114,61,56,168,13,246,34,3,0 // vfmadd213ps 0x322f6(%rip){1to8},%ymm8,%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 98,114,61,56,168,13,80,36,3,0 // vfmadd213ps 0x32450(%rip){1to8},%ymm8,%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 197,124,17,136,64,1,0,0 // vmovups %ymm9,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15915,13 +15915,13 @@ HIDDEN _sk_bicubic_p1y_skx
FUNCTION(_sk_bicubic_p1y_skx)
_sk_bicubic_p1y_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,89,48,3,0 // vbroadcastss 0x33059(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,213,34,3,0 // vbroadcastss 0x322d5(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,72,64 // vaddps 0x40(%rax),%ymm8,%ymm1
.byte 197,124,16,136,192,0,0,0 // vmovups 0xc0(%rax),%ymm9
- .byte 196,98,125,24,21,159,49,3,0 // vbroadcastss 0x3319f(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
- .byte 98,114,53,56,168,21,153,49,3,0 // vfmadd213ps 0x33199(%rip){1to8},%ymm9,%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,27,36,3,0 // vbroadcastss 0x3241b(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 98,114,53,56,168,21,21,36,3,0 // vfmadd213ps 0x32415(%rip){1to8},%ymm9,%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,66,53,168,208 // vfmadd213ps %ymm8,%ymm9,%ymm10
- .byte 98,114,53,56,168,21,142,49,3,0 // vfmadd213ps 0x3318e(%rip){1to8},%ymm9,%ymm10 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 98,114,53,56,168,21,10,36,3,0 // vfmadd213ps 0x3240a(%rip){1to8},%ymm9,%ymm10 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 197,124,17,144,64,1,0,0 // vmovups %ymm10,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -15933,10 +15933,10 @@ _sk_bicubic_p3y_skx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,72,64 // vmovups 0x40(%rax),%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
- .byte 98,241,116,56,88,13,101,49,3,0 // vaddps 0x33165(%rip){1to8},%ymm1,%ymm1 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 98,241,116,56,88,13,225,35,3,0 // vaddps 0x323e1(%rip){1to8},%ymm1,%ymm1 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,79,49,3,0 // vbroadcastss 0x3314f(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
- .byte 98,114,61,56,168,21,101,48,3,0 // vfmadd213ps 0x33065(%rip){1to8},%ymm8,%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,203,35,3,0 // vbroadcastss 0x323cb(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 98,114,61,56,168,21,225,34,3,0 // vfmadd213ps 0x322e1(%rip){1to8},%ymm8,%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,52,89,194 // vmulps %ymm10,%ymm9,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -16062,7 +16062,7 @@ _sk_clut_3D_skx:
.byte 98,226,61,40,64,224 // vpmulld %ymm0,%ymm8,%ymm20
.byte 98,209,93,32,254,193 // vpaddd %ymm9,%ymm20,%ymm0
.byte 72,139,0 // mov (%rax),%rax
- .byte 98,98,125,40,88,29,89,47,3,0 // vpbroadcastd 0x32f59(%rip),%ymm27 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 98,98,125,40,88,29,213,33,3,0 // vpbroadcastd 0x321d5(%rip),%ymm27 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 98,146,125,40,64,195 // vpmulld %ymm27,%ymm0,%ymm0
.byte 196,65,45,239,210 // vpxor %ymm10,%ymm10,%ymm10
.byte 197,245,118,201 // vpcmpeqd %ymm1,%ymm1,%ymm1
@@ -16075,9 +16075,9 @@ _sk_clut_3D_skx:
.byte 196,65,37,118,219 // vpcmpeqd %ymm11,%ymm11,%ymm11
.byte 196,162,37,146,12,144 // vgatherdps %ymm11,(%rax,%ymm10,4),%ymm1
.byte 98,97,124,40,40,233 // vmovaps %ymm1,%ymm29
- .byte 196,98,125,24,21,15,47,3,0 // vbroadcastss 0x32f0f(%rip),%ymm10 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,98,125,24,21,139,33,3,0 // vbroadcastss 0x3218b(%rip),%ymm10 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 98,81,60,32,88,218 // vaddps %ymm10,%ymm24,%ymm11
- .byte 98,226,125,40,88,5,167,45,3,0 // vpbroadcastd 0x32da7(%rip),%ymm16 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 98,226,125,40,88,5,35,32,3,0 // vpbroadcastd 0x32023(%rip),%ymm16 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 98,177,125,40,254,192 // vpaddd %ymm16,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 196,65,29,118,228 // vpcmpeqd %ymm12,%ymm12,%ymm12
@@ -16317,7 +16317,7 @@ _sk_clut_4D_skx:
.byte 98,162,45,40,64,229 // vpmulld %ymm21,%ymm10,%ymm20
.byte 98,241,93,32,254,193 // vpaddd %ymm1,%ymm20,%ymm0
.byte 72,139,0 // mov (%rax),%rax
- .byte 98,226,125,40,88,5,235,41,3,0 // vpbroadcastd 0x329eb(%rip),%ymm16 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 98,226,125,40,88,5,103,28,3,0 // vpbroadcastd 0x31c67(%rip),%ymm16 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 98,50,125,40,64,216 // vpmulld %ymm16,%ymm0,%ymm11
.byte 196,65,28,87,228 // vxorps %ymm12,%ymm12,%ymm12
.byte 197,253,118,192 // vpcmpeqd %ymm0,%ymm0,%ymm0
@@ -16329,9 +16329,9 @@ _sk_clut_4D_skx:
.byte 196,65,29,118,228 // vpcmpeqd %ymm12,%ymm12,%ymm12
.byte 196,98,29,146,44,128 // vgatherdps %ymm12,(%rax,%ymm0,4),%ymm13
.byte 197,124,17,172,36,192,4,0,0 // vmovups %ymm13,0x4c0(%rsp)
- .byte 196,226,125,24,5,157,41,3,0 // vbroadcastss 0x3299d(%rip),%ymm0 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,226,125,24,5,25,28,3,0 // vbroadcastss 0x31c19(%rip),%ymm0 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 98,113,28,32,88,224 // vaddps %ymm0,%ymm28,%ymm12
- .byte 98,226,125,40,88,13,53,40,3,0 // vpbroadcastd 0x32835(%rip),%ymm17 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 98,226,125,40,88,13,177,26,3,0 // vpbroadcastd 0x31ab1(%rip),%ymm17 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 98,49,37,40,254,217 // vpaddd %ymm17,%ymm11,%ymm11
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
.byte 196,65,21,118,237 // vpcmpeqd %ymm13,%ymm13,%ymm13
@@ -16736,7 +16736,7 @@ _sk_clut_4D_skx:
.byte 197,228,92,214 // vsubps %ymm6,%ymm3,%ymm2
.byte 196,226,93,168,214 // vfmadd213ps %ymm6,%ymm4,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,211,30,3,0 // vbroadcastss 0x31ed3(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,79,17,3,0 // vbroadcastss 0x3114f(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 98,145,124,40,40,227 // vmovaps %ymm27,%ymm4
.byte 197,252,16,108,36,128 // vmovups -0x80(%rsp),%ymm5
.byte 197,252,16,116,36,160 // vmovups -0x60(%rsp),%ymm6
@@ -16748,11 +16748,11 @@ HIDDEN _sk_gauss_a_to_rgba_skx
.globl _sk_gauss_a_to_rgba_skx
FUNCTION(_sk_gauss_a_to_rgba_skx)
_sk_gauss_a_to_rgba_skx:
- .byte 196,226,125,24,5,21,32,3,0 // vbroadcastss 0x32015(%rip),%ymm0 # 398a4 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
- .byte 98,242,101,56,168,5,15,32,3,0 // vfmadd213ps 0x3200f(%rip){1to8},%ymm3,%ymm0 # 398a8 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
- .byte 98,242,101,56,168,5,9,32,3,0 // vfmadd213ps 0x32009(%rip){1to8},%ymm3,%ymm0 # 398ac <_sk_srcover_bgra_8888_sse2_lowp+0x424>
- .byte 98,242,101,56,168,5,3,32,3,0 // vfmadd213ps 0x32003(%rip){1to8},%ymm3,%ymm0 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x428>
- .byte 98,242,101,56,168,5,253,31,3,0 // vfmadd213ps 0x31ffd(%rip){1to8},%ymm3,%ymm0 # 398b4 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
+ .byte 196,226,125,24,5,145,18,3,0 // vbroadcastss 0x31291(%rip),%ymm0 # 38b20 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
+ .byte 98,242,101,56,168,5,139,18,3,0 // vfmadd213ps 0x3128b(%rip){1to8},%ymm3,%ymm0 # 38b24 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
+ .byte 98,242,101,56,168,5,133,18,3,0 // vfmadd213ps 0x31285(%rip){1to8},%ymm3,%ymm0 # 38b28 <_sk_srcover_bgra_8888_sse2_lowp+0x424>
+ .byte 98,242,101,56,168,5,127,18,3,0 // vfmadd213ps 0x3127f(%rip){1to8},%ymm3,%ymm0 # 38b2c <_sk_srcover_bgra_8888_sse2_lowp+0x428>
+ .byte 98,242,101,56,168,5,121,18,3,0 // vfmadd213ps 0x31279(%rip){1to8},%ymm3,%ymm0 # 38b30 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
@@ -16852,10 +16852,10 @@ _sk_seed_shader_hsw:
.byte 197,249,110,201 // vmovd %ecx,%xmm1
.byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,102,29,3,0 // vbroadcastss 0x31d66(%rip),%ymm2 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,21,226,15,3,0 // vbroadcastss 0x30fe2(%rip),%ymm2 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,244,88,202 // vaddps %ymm2,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,21,91,29,3,0 // vbroadcastss 0x31d5b(%rip),%ymm2 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,21,215,15,3,0 // vbroadcastss 0x30fd7(%rip),%ymm2 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
.byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
@@ -16870,17 +16870,17 @@ _sk_dither_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,121,110,194 // vmovd %edx,%xmm8
.byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
- .byte 197,61,254,5,250,30,3,0 // vpaddd 0x31efa(%rip),%ymm8,%ymm8 # 39900 <_sk_srcover_bgra_8888_sse2_lowp+0x478>
+ .byte 197,61,254,5,122,17,3,0 // vpaddd 0x3117a(%rip),%ymm8,%ymm8 # 38b80 <_sk_srcover_bgra_8888_sse2_lowp+0x47c>
.byte 197,121,110,201 // vmovd %ecx,%xmm9
.byte 196,66,125,88,201 // vpbroadcastd %xmm9,%ymm9
.byte 196,65,53,239,200 // vpxor %ymm8,%ymm9,%ymm9
- .byte 196,98,125,88,21,31,29,3,0 // vpbroadcastd 0x31d1f(%rip),%ymm10 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,98,125,88,21,155,15,3,0 // vpbroadcastd 0x30f9b(%rip),%ymm10 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 196,65,53,219,218 // vpand %ymm10,%ymm9,%ymm11
.byte 196,193,37,114,243,5 // vpslld $0x5,%ymm11,%ymm11
.byte 196,65,61,219,210 // vpand %ymm10,%ymm8,%ymm10
.byte 196,193,45,114,242,4 // vpslld $0x4,%ymm10,%ymm10
- .byte 196,98,125,88,37,4,29,3,0 // vpbroadcastd 0x31d04(%rip),%ymm12 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
- .byte 196,98,125,88,45,255,28,3,0 // vpbroadcastd 0x31cff(%rip),%ymm13 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,125,88,37,128,15,3,0 // vpbroadcastd 0x30f80(%rip),%ymm12 # 389bc <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
+ .byte 196,98,125,88,45,123,15,3,0 // vpbroadcastd 0x30f7b(%rip),%ymm13 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,65,53,219,245 // vpand %ymm13,%ymm9,%ymm14
.byte 196,193,13,114,246,2 // vpslld $0x2,%ymm14,%ymm14
.byte 196,65,37,235,222 // vpor %ymm14,%ymm11,%ymm11
@@ -16895,8 +16895,8 @@ _sk_dither_hsw:
.byte 196,65,61,235,195 // vpor %ymm11,%ymm8,%ymm8
.byte 196,65,61,235,193 // vpor %ymm9,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,177,28,3,0 // vbroadcastss 0x31cb1(%rip),%ymm9 # 39748 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
- .byte 196,98,125,24,21,172,28,3,0 // vbroadcastss 0x31cac(%rip),%ymm10 # 3974c <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
+ .byte 196,98,125,24,13,45,15,3,0 // vbroadcastss 0x30f2d(%rip),%ymm9 # 389c4 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
+ .byte 196,98,125,24,21,40,15,3,0 // vbroadcastss 0x30f28(%rip),%ymm10 # 389c8 <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
.byte 196,66,61,184,209 // vfmadd231ps %ymm9,%ymm8,%ymm10
.byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
.byte 196,65,44,89,192 // vmulps %ymm8,%ymm10,%ymm8
@@ -16930,7 +16930,7 @@ HIDDEN _sk_black_color_hsw
FUNCTION(_sk_black_color_hsw)
_sk_black_color_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,52,28,3,0 // vbroadcastss 0x31c34(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,176,14,3,0 // vbroadcastss 0x30eb0(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
@@ -16941,7 +16941,7 @@ HIDDEN _sk_white_color_hsw
FUNCTION(_sk_white_color_hsw)
_sk_white_color_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,27,28,3,0 // vbroadcastss 0x31c1b(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,151,14,3,0 // vbroadcastss 0x30e97(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 197,252,40,216 // vmovaps %ymm0,%ymm3
@@ -16987,7 +16987,7 @@ HIDDEN _sk_srcatop_hsw
FUNCTION(_sk_srcatop_hsw)
_sk_srcatop_hsw:
.byte 197,252,89,199 // vmulps %ymm7,%ymm0,%ymm0
- .byte 196,98,125,24,5,186,27,3,0 // vbroadcastss 0x31bba(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,54,14,3,0 // vbroadcastss 0x30e36(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,226,61,184,196 // vfmadd231ps %ymm4,%ymm8,%ymm0
.byte 197,244,89,207 // vmulps %ymm7,%ymm1,%ymm1
@@ -17003,7 +17003,7 @@ HIDDEN _sk_dstatop_hsw
.globl _sk_dstatop_hsw
FUNCTION(_sk_dstatop_hsw)
_sk_dstatop_hsw:
- .byte 196,98,125,24,5,137,27,3,0 // vbroadcastss 0x31b89(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,5,14,3,0 // vbroadcastss 0x30e05(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,226,101,184,196 // vfmadd231ps %ymm4,%ymm3,%ymm0
@@ -17042,7 +17042,7 @@ HIDDEN _sk_srcout_hsw
.globl _sk_srcout_hsw
FUNCTION(_sk_srcout_hsw)
_sk_srcout_hsw:
- .byte 196,98,125,24,5,44,27,3,0 // vbroadcastss 0x31b2c(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,168,13,3,0 // vbroadcastss 0x30da8(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
@@ -17055,7 +17055,7 @@ HIDDEN _sk_dstout_hsw
.globl _sk_dstout_hsw
FUNCTION(_sk_dstout_hsw)
_sk_dstout_hsw:
- .byte 196,226,125,24,5,11,27,3,0 // vbroadcastss 0x31b0b(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,135,13,3,0 // vbroadcastss 0x30d87(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,92,219 // vsubps %ymm3,%ymm0,%ymm3
.byte 197,228,89,196 // vmulps %ymm4,%ymm3,%ymm0
.byte 197,228,89,205 // vmulps %ymm5,%ymm3,%ymm1
@@ -17068,7 +17068,7 @@ HIDDEN _sk_srcover_hsw
.globl _sk_srcover_hsw
FUNCTION(_sk_srcover_hsw)
_sk_srcover_hsw:
- .byte 196,98,125,24,5,234,26,3,0 // vbroadcastss 0x31aea(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,102,13,3,0 // vbroadcastss 0x30d66(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,194,85,184,200 // vfmadd231ps %ymm8,%ymm5,%ymm1
@@ -17081,7 +17081,7 @@ HIDDEN _sk_dstover_hsw
.globl _sk_dstover_hsw
FUNCTION(_sk_dstover_hsw)
_sk_dstover_hsw:
- .byte 196,98,125,24,5,197,26,3,0 // vbroadcastss 0x31ac5(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,65,13,3,0 // vbroadcastss 0x30d41(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
.byte 196,226,61,168,205 // vfmadd213ps %ymm5,%ymm8,%ymm1
@@ -17105,7 +17105,7 @@ HIDDEN _sk_multiply_hsw
.globl _sk_multiply_hsw
FUNCTION(_sk_multiply_hsw)
_sk_multiply_hsw:
- .byte 196,98,125,24,5,140,26,3,0 // vbroadcastss 0x31a8c(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,8,13,3,0 // vbroadcastss 0x30d08(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,208 // vmulps %ymm0,%ymm9,%ymm10
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -17128,7 +17128,7 @@ HIDDEN _sk_plus__hsw
FUNCTION(_sk_plus__hsw)
_sk_plus__hsw:
.byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
- .byte 196,98,125,24,5,59,26,3,0 // vbroadcastss 0x31a3b(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,183,12,3,0 // vbroadcastss 0x30cb7(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
@@ -17158,7 +17158,7 @@ HIDDEN _sk_xor__hsw
.globl _sk_xor__hsw
FUNCTION(_sk_xor__hsw)
_sk_xor__hsw:
- .byte 196,98,125,24,5,230,25,3,0 // vbroadcastss 0x319e6(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,98,12,3,0 // vbroadcastss 0x30c62(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -17192,7 +17192,7 @@ _sk_darken_hsw:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,95,209 // vmaxps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,106,25,3,0 // vbroadcastss 0x3196a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,230,11,3,0 // vbroadcastss 0x30be6(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -17217,7 +17217,7 @@ _sk_lighten_hsw:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,21,25,3,0 // vbroadcastss 0x31915(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,145,11,3,0 // vbroadcastss 0x30b91(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -17245,7 +17245,7 @@ _sk_difference_hsw:
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,180,24,3,0 // vbroadcastss 0x318b4(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,48,11,3,0 // vbroadcastss 0x30b30(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -17267,7 +17267,7 @@ _sk_exclusion_hsw:
.byte 197,236,89,214 // vmulps %ymm6,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,110,24,3,0 // vbroadcastss 0x3186e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,234,10,3,0 // vbroadcastss 0x30aea(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -17277,7 +17277,7 @@ HIDDEN _sk_colorburn_hsw
.globl _sk_colorburn_hsw
FUNCTION(_sk_colorburn_hsw)
_sk_colorburn_hsw:
- .byte 196,98,125,24,5,88,24,3,0 // vbroadcastss 0x31858(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,212,10,3,0 // vbroadcastss 0x30ad4(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,216 // vmulps %ymm0,%ymm9,%ymm11
.byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
@@ -17335,7 +17335,7 @@ HIDDEN _sk_colordodge_hsw
FUNCTION(_sk_colordodge_hsw)
_sk_colordodge_hsw:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
- .byte 196,98,125,24,13,104,23,3,0 // vbroadcastss 0x31768(%rip),%ymm9 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,13,228,9,3,0 // vbroadcastss 0x309e4(%rip),%ymm9 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,52,92,215 // vsubps %ymm7,%ymm9,%ymm10
.byte 197,44,89,216 // vmulps %ymm0,%ymm10,%ymm11
.byte 197,52,92,203 // vsubps %ymm3,%ymm9,%ymm9
@@ -17388,7 +17388,7 @@ HIDDEN _sk_hardlight_hsw
.globl _sk_hardlight_hsw
FUNCTION(_sk_hardlight_hsw)
_sk_hardlight_hsw:
- .byte 196,98,125,24,5,130,22,3,0 // vbroadcastss 0x31682(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,254,8,3,0 // vbroadcastss 0x308fe(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,215 // vsubps %ymm7,%ymm8,%ymm10
.byte 197,44,89,216 // vmulps %ymm0,%ymm10,%ymm11
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -17439,7 +17439,7 @@ HIDDEN _sk_overlay_hsw
.globl _sk_overlay_hsw
FUNCTION(_sk_overlay_hsw)
_sk_overlay_hsw:
- .byte 196,98,125,24,5,182,21,3,0 // vbroadcastss 0x315b6(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,50,8,3,0 // vbroadcastss 0x30832(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,215 // vsubps %ymm7,%ymm8,%ymm10
.byte 197,44,89,216 // vmulps %ymm0,%ymm10,%ymm11
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -17500,10 +17500,10 @@ _sk_softlight_hsw:
.byte 196,65,20,88,197 // vaddps %ymm13,%ymm13,%ymm8
.byte 196,65,60,88,192 // vaddps %ymm8,%ymm8,%ymm8
.byte 196,66,61,168,192 // vfmadd213ps %ymm8,%ymm8,%ymm8
- .byte 196,98,125,24,29,209,20,3,0 // vbroadcastss 0x314d1(%rip),%ymm11 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,29,77,7,3,0 // vbroadcastss 0x3074d(%rip),%ymm11 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,65,20,88,227 // vaddps %ymm11,%ymm13,%ymm12
.byte 196,65,28,89,192 // vmulps %ymm8,%ymm12,%ymm8
- .byte 196,98,125,24,37,194,20,3,0 // vbroadcastss 0x314c2(%rip),%ymm12 # 39754 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
+ .byte 196,98,125,24,37,62,7,3,0 // vbroadcastss 0x3073e(%rip),%ymm12 # 389d0 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
.byte 196,66,21,184,196 // vfmadd231ps %ymm12,%ymm13,%ymm8
.byte 196,65,124,82,245 // vrsqrtps %ymm13,%ymm14
.byte 196,65,124,83,246 // vrcpps %ymm14,%ymm14
@@ -17513,7 +17513,7 @@ _sk_softlight_hsw:
.byte 197,4,194,255,2 // vcmpleps %ymm7,%ymm15,%ymm15
.byte 196,67,13,74,240,240 // vblendvps %ymm15,%ymm8,%ymm14,%ymm14
.byte 197,116,88,249 // vaddps %ymm1,%ymm1,%ymm15
- .byte 196,98,125,24,5,113,20,3,0 // vbroadcastss 0x31471(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,237,6,3,0 // vbroadcastss 0x306ed(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,65,60,92,237 // vsubps %ymm13,%ymm8,%ymm13
.byte 197,132,92,195 // vsubps %ymm3,%ymm15,%ymm0
.byte 196,98,125,168,235 // vfmadd213ps %ymm3,%ymm0,%ymm13
@@ -17626,11 +17626,11 @@ _sk_hue_hsw:
.byte 196,65,28,89,210 // vmulps %ymm10,%ymm12,%ymm10
.byte 196,65,44,94,214 // vdivps %ymm14,%ymm10,%ymm10
.byte 196,67,45,74,224,240 // vblendvps %ymm15,%ymm8,%ymm10,%ymm12
- .byte 196,98,125,24,53,137,18,3,0 // vbroadcastss 0x31289(%rip),%ymm14 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,61,132,18,3,0 // vbroadcastss 0x31284(%rip),%ymm15 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,53,5,5,3,0 // vbroadcastss 0x30505(%rip),%ymm14 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,61,0,5,3,0 // vbroadcastss 0x30500(%rip),%ymm15 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,239 // vmulps %ymm15,%ymm5,%ymm13
.byte 196,66,93,184,238 // vfmadd231ps %ymm14,%ymm4,%ymm13
- .byte 196,226,125,24,5,117,18,3,0 // vbroadcastss 0x31275(%rip),%ymm0 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,226,125,24,5,241,4,3,0 // vbroadcastss 0x304f1(%rip),%ymm0 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,98,77,184,232 // vfmadd231ps %ymm0,%ymm6,%ymm13
.byte 196,65,116,89,215 // vmulps %ymm15,%ymm1,%ymm10
.byte 196,66,53,184,214 // vfmadd231ps %ymm14,%ymm9,%ymm10
@@ -17685,7 +17685,7 @@ _sk_hue_hsw:
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
.byte 196,65,36,95,200 // vmaxps %ymm8,%ymm11,%ymm9
.byte 196,65,116,95,192 // vmaxps %ymm8,%ymm1,%ymm8
- .byte 196,226,125,24,13,54,17,3,0 // vbroadcastss 0x31136(%rip),%ymm1 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,13,178,3,3,0 // vbroadcastss 0x303b2(%rip),%ymm1 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,116,92,215 // vsubps %ymm7,%ymm1,%ymm10
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 197,116,92,219 // vsubps %ymm3,%ymm1,%ymm11
@@ -17739,11 +17739,11 @@ _sk_saturation_hsw:
.byte 196,65,28,89,210 // vmulps %ymm10,%ymm12,%ymm10
.byte 196,65,44,94,214 // vdivps %ymm14,%ymm10,%ymm10
.byte 196,67,45,74,224,240 // vblendvps %ymm15,%ymm8,%ymm10,%ymm12
- .byte 196,98,125,24,53,105,16,3,0 // vbroadcastss 0x31069(%rip),%ymm14 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,61,100,16,3,0 // vbroadcastss 0x31064(%rip),%ymm15 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,53,229,2,3,0 // vbroadcastss 0x302e5(%rip),%ymm14 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,61,224,2,3,0 // vbroadcastss 0x302e0(%rip),%ymm15 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,239 // vmulps %ymm15,%ymm5,%ymm13
.byte 196,66,93,184,238 // vfmadd231ps %ymm14,%ymm4,%ymm13
- .byte 196,226,125,24,5,85,16,3,0 // vbroadcastss 0x31055(%rip),%ymm0 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,226,125,24,5,209,2,3,0 // vbroadcastss 0x302d1(%rip),%ymm0 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,98,77,184,232 // vfmadd231ps %ymm0,%ymm6,%ymm13
.byte 196,65,116,89,215 // vmulps %ymm15,%ymm1,%ymm10
.byte 196,66,53,184,214 // vfmadd231ps %ymm14,%ymm9,%ymm10
@@ -17798,7 +17798,7 @@ _sk_saturation_hsw:
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
.byte 196,65,36,95,200 // vmaxps %ymm8,%ymm11,%ymm9
.byte 196,65,116,95,192 // vmaxps %ymm8,%ymm1,%ymm8
- .byte 196,226,125,24,13,22,15,3,0 // vbroadcastss 0x30f16(%rip),%ymm1 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,13,146,1,3,0 // vbroadcastss 0x30192(%rip),%ymm1 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,116,92,215 // vsubps %ymm7,%ymm1,%ymm10
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 197,116,92,219 // vsubps %ymm3,%ymm1,%ymm11
@@ -17826,11 +17826,11 @@ _sk_color_hsw:
.byte 197,108,89,199 // vmulps %ymm7,%ymm2,%ymm8
.byte 197,116,89,215 // vmulps %ymm7,%ymm1,%ymm10
.byte 197,52,89,223 // vmulps %ymm7,%ymm9,%ymm11
- .byte 196,98,125,24,45,203,14,3,0 // vbroadcastss 0x30ecb(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,53,198,14,3,0 // vbroadcastss 0x30ec6(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,45,71,1,3,0 // vbroadcastss 0x30147(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,66,1,3,0 // vbroadcastss 0x30142(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,230 // vmulps %ymm14,%ymm5,%ymm12
.byte 196,66,93,184,229 // vfmadd231ps %ymm13,%ymm4,%ymm12
- .byte 196,98,125,24,61,183,14,3,0 // vbroadcastss 0x30eb7(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,51,1,3,0 // vbroadcastss 0x30133(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,66,77,184,231 // vfmadd231ps %ymm15,%ymm6,%ymm12
.byte 196,65,44,89,206 // vmulps %ymm14,%ymm10,%ymm9
.byte 196,66,61,184,205 // vfmadd231ps %ymm13,%ymm8,%ymm9
@@ -17886,7 +17886,7 @@ _sk_color_hsw:
.byte 196,193,116,95,206 // vmaxps %ymm14,%ymm1,%ymm1
.byte 196,65,44,95,198 // vmaxps %ymm14,%ymm10,%ymm8
.byte 196,65,124,95,206 // vmaxps %ymm14,%ymm0,%ymm9
- .byte 196,226,125,24,5,109,13,3,0 // vbroadcastss 0x30d6d(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,233,255,2,0 // vbroadcastss 0x2ffe9(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,124,92,215 // vsubps %ymm7,%ymm0,%ymm10
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 197,124,92,219 // vsubps %ymm3,%ymm0,%ymm11
@@ -17914,11 +17914,11 @@ _sk_luminosity_hsw:
.byte 197,100,89,196 // vmulps %ymm4,%ymm3,%ymm8
.byte 197,100,89,213 // vmulps %ymm5,%ymm3,%ymm10
.byte 197,100,89,222 // vmulps %ymm6,%ymm3,%ymm11
- .byte 196,98,125,24,45,34,13,3,0 // vbroadcastss 0x30d22(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
- .byte 196,98,125,24,53,29,13,3,0 // vbroadcastss 0x30d1d(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,45,158,255,2,0 // vbroadcastss 0x2ff9e(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,153,255,2,0 // vbroadcastss 0x2ff99(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,116,89,230 // vmulps %ymm14,%ymm1,%ymm12
.byte 196,66,109,184,229 // vfmadd231ps %ymm13,%ymm2,%ymm12
- .byte 196,98,125,24,61,14,13,3,0 // vbroadcastss 0x30d0e(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,138,255,2,0 // vbroadcastss 0x2ff8a(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,66,53,184,231 // vfmadd231ps %ymm15,%ymm9,%ymm12
.byte 196,65,44,89,206 // vmulps %ymm14,%ymm10,%ymm9
.byte 196,66,61,184,205 // vfmadd231ps %ymm13,%ymm8,%ymm9
@@ -17974,7 +17974,7 @@ _sk_luminosity_hsw:
.byte 196,193,116,95,206 // vmaxps %ymm14,%ymm1,%ymm1
.byte 196,65,44,95,198 // vmaxps %ymm14,%ymm10,%ymm8
.byte 196,65,124,95,206 // vmaxps %ymm14,%ymm0,%ymm9
- .byte 196,226,125,24,5,196,11,3,0 // vbroadcastss 0x30bc4(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,64,254,2,0 // vbroadcastss 0x2fe40(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,124,92,215 // vsubps %ymm7,%ymm0,%ymm10
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 197,124,92,219 // vsubps %ymm3,%ymm0,%ymm11
@@ -18005,17 +18005,17 @@ _sk_srcover_rgba_8888_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,179,0,0,0 // jne 8c89 <_sk_srcover_rgba_8888_hsw+0xd3>
.byte 196,129,126,111,60,152 // vmovdqu (%r8,%r11,4),%ymm7
- .byte 197,197,219,37,60,13,3,0 // vpand 0x30d3c(%rip),%ymm7,%ymm4 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x498>
+ .byte 197,197,219,37,188,255,2,0 // vpand 0x2ffbc(%rip),%ymm7,%ymm4 # 38ba0 <_sk_srcover_bgra_8888_sse2_lowp+0x49c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,69,0,45,79,13,3,0 // vpshufb 0x30d4f(%rip),%ymm7,%ymm5 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x4b8>
+ .byte 196,226,69,0,45,207,255,2,0 // vpshufb 0x2ffcf(%rip),%ymm7,%ymm5 # 38bc0 <_sk_srcover_bgra_8888_sse2_lowp+0x4bc>
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,226,69,0,53,98,13,3,0 // vpshufb 0x30d62(%rip),%ymm7,%ymm6 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x4d8>
+ .byte 196,226,69,0,53,226,255,2,0 // vpshufb 0x2ffe2(%rip),%ymm7,%ymm6 # 38be0 <_sk_srcover_bgra_8888_sse2_lowp+0x4dc>
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,36,11,3,0 // vbroadcastss 0x30b24(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,160,253,2,0 // vbroadcastss 0x2fda0(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,67,11,3,0 // vbroadcastss 0x30b43(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,191,253,2,0 // vbroadcastss 0x2fdbf(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
@@ -18141,17 +18141,17 @@ _sk_srcover_bgra_8888_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,179,0,0,0 // jne 8eb7 <_sk_srcover_bgra_8888_hsw+0xd3>
.byte 196,129,126,111,60,152 // vmovdqu (%r8,%r11,4),%ymm7
- .byte 197,197,219,37,110,11,3,0 // vpand 0x30b6e(%rip),%ymm7,%ymm4 # 39980 <_sk_srcover_bgra_8888_sse2_lowp+0x4f8>
+ .byte 197,197,219,37,238,253,2,0 // vpand 0x2fdee(%rip),%ymm7,%ymm4 # 38c00 <_sk_srcover_bgra_8888_sse2_lowp+0x4fc>
.byte 197,252,91,244 // vcvtdq2ps %ymm4,%ymm6
- .byte 196,226,69,0,37,129,11,3,0 // vpshufb 0x30b81(%rip),%ymm7,%ymm4 # 399a0 <_sk_srcover_bgra_8888_sse2_lowp+0x518>
+ .byte 196,226,69,0,37,1,254,2,0 // vpshufb 0x2fe01(%rip),%ymm7,%ymm4 # 38c20 <_sk_srcover_bgra_8888_sse2_lowp+0x51c>
.byte 197,252,91,236 // vcvtdq2ps %ymm4,%ymm5
- .byte 196,226,69,0,37,148,11,3,0 // vpshufb 0x30b94(%rip),%ymm7,%ymm4 # 399c0 <_sk_srcover_bgra_8888_sse2_lowp+0x538>
+ .byte 196,226,69,0,37,20,254,2,0 // vpshufb 0x2fe14(%rip),%ymm7,%ymm4 # 38c40 <_sk_srcover_bgra_8888_sse2_lowp+0x53c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,246,8,3,0 // vbroadcastss 0x308f6(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,114,251,2,0 // vbroadcastss 0x2fb72(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,21,9,3,0 // vbroadcastss 0x30915(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,145,251,2,0 // vbroadcastss 0x2fb91(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
.byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
@@ -18279,7 +18279,7 @@ HIDDEN _sk_clamp_1_hsw
.globl _sk_clamp_1_hsw
FUNCTION(_sk_clamp_1_hsw)
_sk_clamp_1_hsw:
- .byte 196,98,125,24,5,2,7,3,0 // vbroadcastss 0x30702(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,126,249,2,0 // vbroadcastss 0x2f97e(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
@@ -18291,7 +18291,7 @@ HIDDEN _sk_clamp_a_hsw
.globl _sk_clamp_a_hsw
FUNCTION(_sk_clamp_a_hsw)
_sk_clamp_a_hsw:
- .byte 196,98,125,24,5,225,6,3,0 // vbroadcastss 0x306e1(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,93,249,2,0 // vbroadcastss 0x2f95d(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 197,252,93,195 // vminps %ymm3,%ymm0,%ymm0
.byte 197,244,93,203 // vminps %ymm3,%ymm1,%ymm1
@@ -18303,7 +18303,7 @@ HIDDEN _sk_clamp_a_dst_hsw
.globl _sk_clamp_a_dst_hsw
FUNCTION(_sk_clamp_a_dst_hsw)
_sk_clamp_a_dst_hsw:
- .byte 196,98,125,24,5,195,6,3,0 // vbroadcastss 0x306c3(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,63,249,2,0 // vbroadcastss 0x2f93f(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,68,93,248 // vminps %ymm8,%ymm7,%ymm7
.byte 197,220,93,231 // vminps %ymm7,%ymm4,%ymm4
.byte 197,212,93,239 // vminps %ymm7,%ymm5,%ymm5
@@ -18336,7 +18336,7 @@ HIDDEN _sk_invert_hsw
.globl _sk_invert_hsw
FUNCTION(_sk_invert_hsw)
_sk_invert_hsw:
- .byte 196,98,125,24,5,126,6,3,0 // vbroadcastss 0x3067e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,250,248,2,0 // vbroadcastss 0x2f8fa(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,188,92,192 // vsubps %ymm0,%ymm8,%ymm0
.byte 197,188,92,201 // vsubps %ymm1,%ymm8,%ymm1
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
@@ -18390,9 +18390,9 @@ HIDDEN _sk_unpremul_hsw
.globl _sk_unpremul_hsw
FUNCTION(_sk_unpremul_hsw)
_sk_unpremul_hsw:
- .byte 196,98,125,24,5,25,6,3,0 // vbroadcastss 0x30619(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,149,248,2,0 // vbroadcastss 0x2f895(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,94,195 // vdivps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,60,6,3,0 // vbroadcastss 0x3063c(%rip),%ymm9 # 39768 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
+ .byte 196,98,125,24,13,184,248,2,0 // vbroadcastss 0x2f8b8(%rip),%ymm9 # 389e4 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
.byte 196,65,60,194,201,1 // vcmpltps %ymm9,%ymm8,%ymm9
.byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
.byte 196,67,45,74,192,144 // vblendvps %ymm9,%ymm8,%ymm10,%ymm8
@@ -18406,16 +18406,16 @@ HIDDEN _sk_from_srgb_hsw
.globl _sk_from_srgb_hsw
FUNCTION(_sk_from_srgb_hsw)
_sk_from_srgb_hsw:
- .byte 196,98,125,24,5,22,6,3,0 // vbroadcastss 0x30616(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,146,248,2,0 // vbroadcastss 0x2f892(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 197,124,89,208 // vmulps %ymm0,%ymm0,%ymm10
- .byte 196,98,125,24,29,8,6,3,0 // vbroadcastss 0x30608(%rip),%ymm11 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
- .byte 196,98,125,24,37,231,5,3,0 // vbroadcastss 0x305e7(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,132,248,2,0 // vbroadcastss 0x2f884(%rip),%ymm11 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,37,99,248,2,0 // vbroadcastss 0x2f863(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,125,168,235 // vfmadd213ps %ymm11,%ymm0,%ymm13
- .byte 196,98,125,24,53,240,5,3,0 // vbroadcastss 0x305f0(%rip),%ymm14 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,53,108,248,2,0 // vbroadcastss 0x2f86c(%rip),%ymm14 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,66,45,168,238 // vfmadd213ps %ymm14,%ymm10,%ymm13
- .byte 196,98,125,24,21,230,5,3,0 // vbroadcastss 0x305e6(%rip),%ymm10 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,21,98,248,2,0 // vbroadcastss 0x2f862(%rip),%ymm10 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 196,193,124,194,194,1 // vcmpltps %ymm10,%ymm0,%ymm0
.byte 196,195,21,74,193,0 // vblendvps %ymm0,%ymm9,%ymm13,%ymm0
.byte 196,65,116,89,200 // vmulps %ymm8,%ymm1,%ymm9
@@ -18438,16 +18438,16 @@ HIDDEN _sk_from_srgb_dst_hsw
.globl _sk_from_srgb_dst_hsw
FUNCTION(_sk_from_srgb_dst_hsw)
_sk_from_srgb_dst_hsw:
- .byte 196,98,125,24,5,126,5,3,0 // vbroadcastss 0x3057e(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,250,247,2,0 // vbroadcastss 0x2f7fa(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,92,89,200 // vmulps %ymm8,%ymm4,%ymm9
.byte 197,92,89,212 // vmulps %ymm4,%ymm4,%ymm10
- .byte 196,98,125,24,29,112,5,3,0 // vbroadcastss 0x30570(%rip),%ymm11 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
- .byte 196,98,125,24,37,79,5,3,0 // vbroadcastss 0x3054f(%rip),%ymm12 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,236,247,2,0 // vbroadcastss 0x2f7ec(%rip),%ymm11 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,37,203,247,2,0 // vbroadcastss 0x2f7cb(%rip),%ymm12 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,93,168,235 // vfmadd213ps %ymm11,%ymm4,%ymm13
- .byte 196,98,125,24,53,88,5,3,0 // vbroadcastss 0x30558(%rip),%ymm14 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,53,212,247,2,0 // vbroadcastss 0x2f7d4(%rip),%ymm14 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,66,45,168,238 // vfmadd213ps %ymm14,%ymm10,%ymm13
- .byte 196,98,125,24,21,78,5,3,0 // vbroadcastss 0x3054e(%rip),%ymm10 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,21,202,247,2,0 // vbroadcastss 0x2f7ca(%rip),%ymm10 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 196,193,92,194,226,1 // vcmpltps %ymm10,%ymm4,%ymm4
.byte 196,195,21,74,225,64 // vblendvps %ymm4,%ymm9,%ymm13,%ymm4
.byte 196,65,84,89,200 // vmulps %ymm8,%ymm5,%ymm9
@@ -18471,19 +18471,19 @@ HIDDEN _sk_to_srgb_hsw
FUNCTION(_sk_to_srgb_hsw)
_sk_to_srgb_hsw:
.byte 197,124,82,200 // vrsqrtps %ymm0,%ymm9
- .byte 196,98,125,24,5,242,4,3,0 // vbroadcastss 0x304f2(%rip),%ymm8 # 3977c <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
+ .byte 196,98,125,24,5,110,247,2,0 // vbroadcastss 0x2f76e(%rip),%ymm8 # 389f8 <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
.byte 196,65,124,89,208 // vmulps %ymm8,%ymm0,%ymm10
- .byte 196,98,125,24,29,232,4,3,0 // vbroadcastss 0x304e8(%rip),%ymm11 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
- .byte 196,98,125,24,37,227,4,3,0 // vbroadcastss 0x304e3(%rip),%ymm12 # 39784 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
+ .byte 196,98,125,24,29,100,247,2,0 // vbroadcastss 0x2f764(%rip),%ymm11 # 389fc <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
+ .byte 196,98,125,24,37,95,247,2,0 // vbroadcastss 0x2f75f(%rip),%ymm12 # 38a00 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
.byte 196,65,124,40,236 // vmovaps %ymm12,%ymm13
.byte 196,66,53,168,235 // vfmadd213ps %ymm11,%ymm9,%ymm13
- .byte 196,98,125,24,53,4,6,3,0 // vbroadcastss 0x30604(%rip),%ymm14 # 398b8 <_sk_srcover_bgra_8888_sse2_lowp+0x430>
+ .byte 196,98,125,24,53,128,248,2,0 // vbroadcastss 0x2f880(%rip),%ymm14 # 38b34 <_sk_srcover_bgra_8888_sse2_lowp+0x430>
.byte 196,66,53,168,238 // vfmadd213ps %ymm14,%ymm9,%ymm13
- .byte 196,98,125,24,61,250,5,3,0 // vbroadcastss 0x305fa(%rip),%ymm15 # 398bc <_sk_srcover_bgra_8888_sse2_lowp+0x434>
+ .byte 196,98,125,24,61,118,248,2,0 // vbroadcastss 0x2f876(%rip),%ymm15 # 38b38 <_sk_srcover_bgra_8888_sse2_lowp+0x434>
.byte 196,65,52,88,207 // vaddps %ymm15,%ymm9,%ymm9
.byte 196,65,124,83,201 // vrcpps %ymm9,%ymm9
.byte 196,65,20,89,201 // vmulps %ymm9,%ymm13,%ymm9
- .byte 196,98,125,24,45,182,4,3,0 // vbroadcastss 0x304b6(%rip),%ymm13 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x308>
+ .byte 196,98,125,24,45,50,247,2,0 // vbroadcastss 0x2f732(%rip),%ymm13 # 38a0c <_sk_srcover_bgra_8888_sse2_lowp+0x308>
.byte 196,193,124,194,197,1 // vcmpltps %ymm13,%ymm0,%ymm0
.byte 196,195,53,74,194,0 // vblendvps %ymm0,%ymm10,%ymm9,%ymm0
.byte 197,124,82,201 // vrsqrtps %ymm1,%ymm9
@@ -18517,26 +18517,26 @@ _sk_rgb_to_hsl_hsw:
.byte 197,116,93,202 // vminps %ymm2,%ymm1,%ymm9
.byte 196,65,124,93,201 // vminps %ymm9,%ymm0,%ymm9
.byte 196,65,60,92,209 // vsubps %ymm9,%ymm8,%ymm10
- .byte 196,98,125,24,29,205,3,3,0 // vbroadcastss 0x303cd(%rip),%ymm11 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,29,73,246,2,0 // vbroadcastss 0x2f649(%rip),%ymm11 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,65,36,94,218 // vdivps %ymm10,%ymm11,%ymm11
.byte 197,116,92,226 // vsubps %ymm2,%ymm1,%ymm12
.byte 197,116,194,234,1 // vcmpltps %ymm2,%ymm1,%ymm13
- .byte 196,98,125,24,53,18,4,3,0 // vbroadcastss 0x30412(%rip),%ymm14 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,98,125,24,53,142,246,2,0 // vbroadcastss 0x2f68e(%rip),%ymm14 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 196,65,4,87,255 // vxorps %ymm15,%ymm15,%ymm15
.byte 196,67,5,74,238,208 // vblendvps %ymm13,%ymm14,%ymm15,%ymm13
.byte 196,66,37,168,229 // vfmadd213ps %ymm13,%ymm11,%ymm12
.byte 197,236,92,208 // vsubps %ymm0,%ymm2,%ymm2
.byte 197,124,92,233 // vsubps %ymm1,%ymm0,%ymm13
- .byte 196,98,125,24,53,249,3,3,0 // vbroadcastss 0x303f9(%rip),%ymm14 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 196,98,125,24,53,117,246,2,0 // vbroadcastss 0x2f675(%rip),%ymm14 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 196,66,37,168,238 // vfmadd213ps %ymm14,%ymm11,%ymm13
- .byte 196,98,125,24,53,231,3,3,0 // vbroadcastss 0x303e7(%rip),%ymm14 # 39798 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
+ .byte 196,98,125,24,53,99,246,2,0 // vbroadcastss 0x2f663(%rip),%ymm14 # 38a14 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
.byte 196,194,37,168,214 // vfmadd213ps %ymm14,%ymm11,%ymm2
.byte 197,188,194,201,0 // vcmpeqps %ymm1,%ymm8,%ymm1
.byte 196,227,21,74,202,16 // vblendvps %ymm1,%ymm2,%ymm13,%ymm1
.byte 197,188,194,192,0 // vcmpeqps %ymm0,%ymm8,%ymm0
.byte 196,195,117,74,196,0 // vblendvps %ymm0,%ymm12,%ymm1,%ymm0
.byte 196,193,60,88,201 // vaddps %ymm9,%ymm8,%ymm1
- .byte 196,98,125,24,29,90,3,3,0 // vbroadcastss 0x3035a(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,214,245,2,0 // vbroadcastss 0x2f5d6(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,116,89,211 // vmulps %ymm11,%ymm1,%ymm2
.byte 197,36,194,218,1 // vcmpltps %ymm2,%ymm11,%ymm11
.byte 196,65,12,92,224 // vsubps %ymm8,%ymm14,%ymm12
@@ -18546,7 +18546,7 @@ _sk_rgb_to_hsl_hsw:
.byte 197,172,94,201 // vdivps %ymm1,%ymm10,%ymm1
.byte 196,195,125,74,199,128 // vblendvps %ymm8,%ymm15,%ymm0,%ymm0
.byte 196,195,117,74,207,128 // vblendvps %ymm8,%ymm15,%ymm1,%ymm1
- .byte 196,98,125,24,5,141,3,3,0 // vbroadcastss 0x3038d(%rip),%ymm8 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 196,98,125,24,5,9,246,2,0 // vbroadcastss 0x2f609(%rip),%ymm8 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -18563,30 +18563,30 @@ _sk_hsl_to_rgb_hsw:
.byte 197,252,17,92,36,128 // vmovups %ymm3,-0x80(%rsp)
.byte 197,252,40,233 // vmovaps %ymm1,%ymm5
.byte 197,252,40,224 // vmovaps %ymm0,%ymm4
- .byte 196,98,125,24,5,230,2,3,0 // vbroadcastss 0x302e6(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,98,245,2,0 // vbroadcastss 0x2f562(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,60,194,202,2 // vcmpleps %ymm2,%ymm8,%ymm9
.byte 197,84,89,210 // vmulps %ymm2,%ymm5,%ymm10
.byte 196,65,84,92,218 // vsubps %ymm10,%ymm5,%ymm11
.byte 196,67,45,74,203,144 // vblendvps %ymm9,%ymm11,%ymm10,%ymm9
.byte 197,52,88,210 // vaddps %ymm2,%ymm9,%ymm10
- .byte 196,98,125,24,13,41,3,3,0 // vbroadcastss 0x30329(%rip),%ymm9 # 39798 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
+ .byte 196,98,125,24,13,165,245,2,0 // vbroadcastss 0x2f5a5(%rip),%ymm9 # 38a14 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
.byte 196,66,109,170,202 // vfmsub213ps %ymm10,%ymm2,%ymm9
- .byte 196,98,125,24,29,39,3,3,0 // vbroadcastss 0x30327(%rip),%ymm11 # 397a4 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
+ .byte 196,98,125,24,29,163,245,2,0 // vbroadcastss 0x2f5a3(%rip),%ymm11 # 38a20 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
.byte 196,65,92,88,219 // vaddps %ymm11,%ymm4,%ymm11
.byte 196,67,125,8,227,1 // vroundps $0x1,%ymm11,%ymm12
.byte 196,65,36,92,252 // vsubps %ymm12,%ymm11,%ymm15
.byte 196,65,44,92,217 // vsubps %ymm9,%ymm10,%ymm11
- .byte 196,98,125,24,45,249,2,3,0 // vbroadcastss 0x302f9(%rip),%ymm13 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,98,125,24,45,117,245,2,0 // vbroadcastss 0x2f575(%rip),%ymm13 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 196,193,4,89,197 // vmulps %ymm13,%ymm15,%ymm0
- .byte 196,98,125,24,53,243,2,3,0 // vbroadcastss 0x302f3(%rip),%ymm14 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 196,98,125,24,53,111,245,2,0 // vbroadcastss 0x2f56f(%rip),%ymm14 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 197,12,92,224 // vsubps %ymm0,%ymm14,%ymm12
.byte 196,66,37,168,225 // vfmadd213ps %ymm9,%ymm11,%ymm12
- .byte 196,226,125,24,29,237,2,3,0 // vbroadcastss 0x302ed(%rip),%ymm3 # 397a8 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
+ .byte 196,226,125,24,29,105,245,2,0 // vbroadcastss 0x2f569(%rip),%ymm3 # 38a24 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
.byte 196,193,100,194,255,2 // vcmpleps %ymm15,%ymm3,%ymm7
.byte 196,195,29,74,249,112 // vblendvps %ymm7,%ymm9,%ymm12,%ymm7
.byte 196,65,60,194,231,2 // vcmpleps %ymm15,%ymm8,%ymm12
.byte 196,227,45,74,255,192 // vblendvps %ymm12,%ymm7,%ymm10,%ymm7
- .byte 196,98,125,24,37,196,2,3,0 // vbroadcastss 0x302c4(%rip),%ymm12 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 196,98,125,24,37,64,245,2,0 // vbroadcastss 0x2f540(%rip),%ymm12 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 196,65,28,194,255,2 // vcmpleps %ymm15,%ymm12,%ymm15
.byte 196,194,37,168,193 // vfmadd213ps %ymm9,%ymm11,%ymm0
.byte 196,99,125,74,255,240 // vblendvps %ymm15,%ymm7,%ymm0,%ymm15
@@ -18602,7 +18602,7 @@ _sk_hsl_to_rgb_hsw:
.byte 197,156,194,192,2 // vcmpleps %ymm0,%ymm12,%ymm0
.byte 196,194,37,168,249 // vfmadd213ps %ymm9,%ymm11,%ymm7
.byte 196,227,69,74,201,0 // vblendvps %ymm0,%ymm1,%ymm7,%ymm1
- .byte 196,226,125,24,5,120,2,3,0 // vbroadcastss 0x30278(%rip),%ymm0 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,226,125,24,5,244,244,2,0 // vbroadcastss 0x2f4f4(%rip),%ymm0 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 197,220,88,192 // vaddps %ymm0,%ymm4,%ymm0
.byte 196,227,125,8,224,1 // vroundps $0x1,%ymm0,%ymm4
.byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
@@ -18656,10 +18656,10 @@ _sk_scale_u8_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,58 // jne 9623 <_sk_scale_u8_hsw+0x52>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,25,10,3,0 // vpand 0x30a19(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,153,252,2,0 // vpand 0x2fc99(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,166,1,3,0 // vbroadcastss 0x301a6(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,34,244,2,0 // vbroadcastss 0x2f422(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
@@ -18736,20 +18736,20 @@ _sk_scale_565_hsw:
.byte 15,133,161,0,0,0 // jne 9798 <_sk_scale_565_hsw+0xc0>
.byte 196,1,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm8
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 196,98,125,88,13,169,0,3,0 // vpbroadcastd 0x300a9(%rip),%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,98,125,88,13,37,243,2,0 // vpbroadcastd 0x2f325(%rip),%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,61,219,201 // vpand %ymm9,%ymm8,%ymm9
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 196,98,125,24,21,154,0,3,0 // vbroadcastss 0x3009a(%rip),%ymm10 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,98,125,24,21,22,243,2,0 // vbroadcastss 0x2f316(%rip),%ymm10 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,88,21,144,0,3,0 // vpbroadcastd 0x30090(%rip),%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,98,125,88,21,12,243,2,0 // vpbroadcastd 0x2f30c(%rip),%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,61,219,210 // vpand %ymm10,%ymm8,%ymm10
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 196,98,125,24,29,129,0,3,0 // vbroadcastss 0x30081(%rip),%ymm11 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,98,125,24,29,253,242,2,0 // vbroadcastss 0x2f2fd(%rip),%ymm11 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,88,29,119,0,3,0 // vpbroadcastd 0x30077(%rip),%ymm11 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,98,125,88,29,243,242,2,0 // vpbroadcastd 0x2f2f3(%rip),%ymm11 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,61,219,195 // vpand %ymm11,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,29,104,0,3,0 // vbroadcastss 0x30068(%rip),%ymm11 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,98,125,24,29,228,242,2,0 // vbroadcastss 0x2f2e4(%rip),%ymm11 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 197,100,194,223,1 // vcmpltps %ymm7,%ymm3,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -18840,10 +18840,10 @@ _sk_lerp_u8_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,78 // jne 98d1 <_sk_lerp_u8_hsw+0x66>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,127,7,3,0 // vpand 0x3077f(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,255,249,2,0 // vpand 0x2f9ff(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,12,255,2,0 // vbroadcastss 0x2ff0c(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,136,241,2,0 // vbroadcastss 0x2f188(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
.byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
@@ -18895,7 +18895,7 @@ _sk_lerp_u8_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,221 // callq ffffffffde00997c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd04f4>
+ .byte 232,255,255,255,221 // callq ffffffffde00997c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd1278>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,210 // callq *%rdx
@@ -18921,20 +18921,20 @@ _sk_lerp_565_hsw:
.byte 15,133,181,0,0,0 // jne 9a5c <_sk_lerp_565_hsw+0xd4>
.byte 196,1,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm8
.byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 196,98,125,88,13,249,253,2,0 // vpbroadcastd 0x2fdf9(%rip),%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,98,125,88,13,117,240,2,0 // vpbroadcastd 0x2f075(%rip),%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,61,219,201 // vpand %ymm9,%ymm8,%ymm9
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 196,98,125,24,21,234,253,2,0 // vbroadcastss 0x2fdea(%rip),%ymm10 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,98,125,24,21,102,240,2,0 // vbroadcastss 0x2f066(%rip),%ymm10 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,88,21,224,253,2,0 // vpbroadcastd 0x2fde0(%rip),%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,98,125,88,21,92,240,2,0 // vpbroadcastd 0x2f05c(%rip),%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,61,219,210 // vpand %ymm10,%ymm8,%ymm10
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 196,98,125,24,29,209,253,2,0 // vbroadcastss 0x2fdd1(%rip),%ymm11 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,98,125,24,29,77,240,2,0 // vbroadcastss 0x2f04d(%rip),%ymm11 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,88,29,199,253,2,0 // vpbroadcastd 0x2fdc7(%rip),%ymm11 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,98,125,88,29,67,240,2,0 // vpbroadcastd 0x2f043(%rip),%ymm11 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,61,219,195 // vpand %ymm11,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,29,184,253,2,0 // vbroadcastss 0x2fdb8(%rip),%ymm11 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,98,125,24,29,52,240,2,0 // vbroadcastss 0x2f034(%rip),%ymm11 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 197,100,194,223,1 // vcmpltps %ymm7,%ymm3,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -19008,7 +19008,7 @@ _sk_load_tables_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,116 // jne 9b7e <_sk_load_tables_hsw+0x7e>
.byte 196,193,126,111,28,144 // vmovdqu (%r8,%rdx,4),%ymm3
- .byte 197,229,219,13,200,254,2,0 // vpand 0x2fec8(%rip),%ymm3,%ymm1 # 399e0 <_sk_srcover_bgra_8888_sse2_lowp+0x558>
+ .byte 197,229,219,13,72,241,2,0 // vpand 0x2f148(%rip),%ymm3,%ymm1 # 38c60 <_sk_srcover_bgra_8888_sse2_lowp+0x55c>
.byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
.byte 76,139,64,8 // mov 0x8(%rax),%r8
.byte 76,139,72,16 // mov 0x10(%rax),%r9
@@ -19016,16 +19016,16 @@ _sk_load_tables_hsw:
.byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
.byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
.byte 196,194,53,146,4,136 // vgatherdps %ymm9,(%r8,%ymm1,4),%ymm0
- .byte 196,98,101,0,13,191,254,2,0 // vpshufb 0x2febf(%rip),%ymm3,%ymm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x578>
+ .byte 196,98,101,0,13,63,241,2,0 // vpshufb 0x2f13f(%rip),%ymm3,%ymm9 # 38c80 <_sk_srcover_bgra_8888_sse2_lowp+0x57c>
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 196,65,45,118,210 // vpcmpeqd %ymm10,%ymm10,%ymm10
.byte 196,130,45,146,12,137 // vgatherdps %ymm10,(%r9,%ymm9,4),%ymm1
.byte 72,139,64,24 // mov 0x18(%rax),%rax
- .byte 196,98,101,0,13,195,254,2,0 // vpshufb 0x2fec3(%rip),%ymm3,%ymm9 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x598>
+ .byte 196,98,101,0,13,67,241,2,0 // vpshufb 0x2f143(%rip),%ymm3,%ymm9 # 38ca0 <_sk_srcover_bgra_8888_sse2_lowp+0x59c>
.byte 196,162,61,146,20,136 // vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
.byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,59,252,2,0 // vbroadcastss 0x2fc3b(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,183,238,2,0 // vbroadcastss 0x2eeb7(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19106,7 +19106,7 @@ _sk_load_tables_u16_be_hsw:
.byte 197,185,108,200 // vpunpcklqdq %xmm0,%xmm8,%xmm1
.byte 197,57,109,192 // vpunpckhqdq %xmm0,%xmm8,%xmm8
.byte 196,65,49,108,214 // vpunpcklqdq %xmm14,%xmm9,%xmm10
- .byte 197,121,111,29,96,3,3,0 // vmovdqa 0x30360(%rip),%xmm11 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,121,111,29,224,245,2,0 // vmovdqa 0x2f5e0(%rip),%xmm11 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,193,113,219,195 // vpand %xmm11,%xmm1,%xmm0
.byte 196,226,125,51,200 // vpmovzxwd %xmm0,%ymm1
.byte 196,65,29,118,228 // vpcmpeqd %ymm12,%ymm12,%ymm12
@@ -19131,7 +19131,7 @@ _sk_load_tables_u16_be_hsw:
.byte 197,185,235,219 // vpor %xmm3,%xmm8,%xmm3
.byte 196,226,125,51,219 // vpmovzxwd %xmm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,160,250,2,0 // vbroadcastss 0x2faa0(%rip),%ymm8 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,5,28,237,2,0 // vbroadcastss 0x2ed1c(%rip),%ymm8 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19191,7 +19191,7 @@ _sk_load_tables_rgb_u16_be_hsw:
.byte 197,105,108,192 // vpunpcklqdq %xmm0,%xmm2,%xmm8
.byte 197,241,108,195 // vpunpcklqdq %xmm3,%xmm1,%xmm0
.byte 197,241,109,203 // vpunpckhqdq %xmm3,%xmm1,%xmm1
- .byte 197,249,111,29,224,1,3,0 // vmovdqa 0x301e0(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,111,29,96,244,2,0 // vmovdqa 0x2f460(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 197,249,219,195 // vpand %xmm3,%xmm0,%xmm0
.byte 196,98,125,51,200 // vpmovzxwd %xmm0,%ymm9
.byte 196,65,45,118,210 // vpcmpeqd %ymm10,%ymm10,%ymm10
@@ -19211,7 +19211,7 @@ _sk_load_tables_rgb_u16_be_hsw:
.byte 196,226,125,51,219 // vpmovzxwd %xmm3,%ymm3
.byte 196,226,45,146,20,152 // vgatherdps %ymm10,(%rax,%ymm3,4),%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,169,248,2,0 // vbroadcastss 0x2f8a9(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,37,235,2,0 // vbroadcastss 0x2eb25(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 196,129,121,110,4,72 // vmovd (%r8,%r9,2),%xmm0
.byte 196,65,49,239,201 // vpxor %xmm9,%xmm9,%xmm9
@@ -19280,7 +19280,7 @@ HIDDEN _sk_byte_tables_hsw
FUNCTION(_sk_byte_tables_hsw)
_sk_byte_tables_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,158,247,2,0 // vbroadcastss 0x2f79e(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,26,234,2,0 // vbroadcastss 0x2ea1a(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,125,91,200 // vcvtps2dq %ymm0,%ymm9
.byte 196,65,249,126,200 // vmovq %xmm9,%r8
@@ -19410,7 +19410,7 @@ _sk_byte_tables_hsw:
.byte 67,15,182,4,17 // movzbl (%r9,%r10,1),%eax
.byte 196,194,125,49,193 // vpmovzxbd %xmm9,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,112,245,2,0 // vbroadcastss 0x2f570(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,236,231,2,0 // vbroadcastss 0x2e7ec(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 196,226,125,49,201 // vpmovzxbd %xmm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
@@ -19532,7 +19532,7 @@ _sk_byte_tables_rgb_hsw:
.byte 67,15,182,4,17 // movzbl (%r9,%r10,1),%eax
.byte 196,194,125,49,193 // vpmovzxbd %xmm9,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,68,243,2,0 // vbroadcastss 0x2f344(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,192,229,2,0 // vbroadcastss 0x2e5c0(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 196,226,125,49,201 // vpmovzxbd %xmm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
@@ -19635,33 +19635,33 @@ _sk_parametric_r_hsw:
.byte 196,66,125,168,211 // vfmadd213ps %ymm11,%ymm0,%ymm10
.byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,42,242,2,0 // vbroadcastss 0x2f22a(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,98,125,24,45,37,242,2,0 // vbroadcastss 0x2f225(%rip),%ymm13 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,166,228,2,0 // vbroadcastss 0x2e4a6(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,45,161,228,2,0 // vbroadcastss 0x2e4a1(%rip),%ymm13 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,65,44,84,213 // vandps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,119,241,2,0 // vbroadcastss 0x2f177(%rip),%ymm13 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,45,243,227,2,0 // vbroadcastss 0x2e3f3(%rip),%ymm13 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,86,213 // vorps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,13,242,2,0 // vbroadcastss 0x2f20d(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,137,228,2,0 // vbroadcastss 0x2e489(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,37,184,236 // vfmadd231ps %ymm12,%ymm11,%ymm13
- .byte 196,98,125,24,29,3,242,2,0 // vbroadcastss 0x2f203(%rip),%ymm11 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,29,127,228,2,0 // vbroadcastss 0x2e47f(%rip),%ymm11 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,45,172,221 // vfnmadd213ps %ymm13,%ymm10,%ymm11
- .byte 196,98,125,24,37,249,241,2,0 // vbroadcastss 0x2f1f9(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,117,228,2,0 // vbroadcastss 0x2e475(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,239,241,2,0 // vbroadcastss 0x2f1ef(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,107,228,2,0 // vbroadcastss 0x2e46b(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,28,94,210 // vdivps %ymm10,%ymm12,%ymm10
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 196,99,125,8,208,1 // vroundps $0x1,%ymm0,%ymm10
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,208,241,2,0 // vbroadcastss 0x2f1d0(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,76,228,2,0 // vbroadcastss 0x2e44c(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,124,88,195 // vaddps %ymm11,%ymm0,%ymm0
- .byte 196,98,125,24,29,198,241,2,0 // vbroadcastss 0x2f1c6(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,66,228,2,0 // vbroadcastss 0x2e442(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,98,45,172,216 // vfnmadd213ps %ymm0,%ymm10,%ymm11
- .byte 196,226,125,24,5,188,241,2,0 // vbroadcastss 0x2f1bc(%rip),%ymm0 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,5,56,228,2,0 // vbroadcastss 0x2e438(%rip),%ymm0 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,193,124,92,194 // vsubps %ymm10,%ymm0,%ymm0
- .byte 196,98,125,24,21,178,241,2,0 // vbroadcastss 0x2f1b2(%rip),%ymm10 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,21,46,228,2,0 // vbroadcastss 0x2e42e(%rip),%ymm10 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 197,172,94,192 // vdivps %ymm0,%ymm10,%ymm0
.byte 197,164,88,192 // vaddps %ymm0,%ymm11,%ymm0
- .byte 196,98,125,24,21,165,241,2,0 // vbroadcastss 0x2f1a5(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,33,228,2,0 // vbroadcastss 0x2e421(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 197,253,91,192 // vcvtps2dq %ymm0,%ymm0
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -19669,7 +19669,7 @@ _sk_parametric_r_hsw:
.byte 196,195,125,74,193,128 // vblendvps %ymm8,%ymm9,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,185,240,2,0 // vbroadcastss 0x2f0b9(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,53,227,2,0 // vbroadcastss 0x2e335(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19689,33 +19689,33 @@ _sk_parametric_g_hsw:
.byte 196,66,117,168,211 // vfmadd213ps %ymm11,%ymm1,%ymm10
.byte 196,226,125,24,8 // vbroadcastss (%rax),%ymm1
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,5,241,2,0 // vbroadcastss 0x2f105(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,98,125,24,45,0,241,2,0 // vbroadcastss 0x2f100(%rip),%ymm13 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,129,227,2,0 // vbroadcastss 0x2e381(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,45,124,227,2,0 // vbroadcastss 0x2e37c(%rip),%ymm13 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,65,44,84,213 // vandps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,82,240,2,0 // vbroadcastss 0x2f052(%rip),%ymm13 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,45,206,226,2,0 // vbroadcastss 0x2e2ce(%rip),%ymm13 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,86,213 // vorps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,232,240,2,0 // vbroadcastss 0x2f0e8(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,100,227,2,0 // vbroadcastss 0x2e364(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,37,184,236 // vfmadd231ps %ymm12,%ymm11,%ymm13
- .byte 196,98,125,24,29,222,240,2,0 // vbroadcastss 0x2f0de(%rip),%ymm11 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,29,90,227,2,0 // vbroadcastss 0x2e35a(%rip),%ymm11 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,45,172,221 // vfnmadd213ps %ymm13,%ymm10,%ymm11
- .byte 196,98,125,24,37,212,240,2,0 // vbroadcastss 0x2f0d4(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,80,227,2,0 // vbroadcastss 0x2e350(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,202,240,2,0 // vbroadcastss 0x2f0ca(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,70,227,2,0 // vbroadcastss 0x2e346(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,28,94,210 // vdivps %ymm10,%ymm12,%ymm10
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
.byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
.byte 196,99,125,8,209,1 // vroundps $0x1,%ymm1,%ymm10
.byte 196,65,116,92,210 // vsubps %ymm10,%ymm1,%ymm10
- .byte 196,98,125,24,29,171,240,2,0 // vbroadcastss 0x2f0ab(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,39,227,2,0 // vbroadcastss 0x2e327(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,29,161,240,2,0 // vbroadcastss 0x2f0a1(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,29,227,2,0 // vbroadcastss 0x2e31d(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,98,45,172,217 // vfnmadd213ps %ymm1,%ymm10,%ymm11
- .byte 196,226,125,24,13,151,240,2,0 // vbroadcastss 0x2f097(%rip),%ymm1 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,13,19,227,2,0 // vbroadcastss 0x2e313(%rip),%ymm1 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,193,116,92,202 // vsubps %ymm10,%ymm1,%ymm1
- .byte 196,98,125,24,21,141,240,2,0 // vbroadcastss 0x2f08d(%rip),%ymm10 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,21,9,227,2,0 // vbroadcastss 0x2e309(%rip),%ymm10 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 197,172,94,201 // vdivps %ymm1,%ymm10,%ymm1
.byte 197,164,88,201 // vaddps %ymm1,%ymm11,%ymm1
- .byte 196,98,125,24,21,128,240,2,0 // vbroadcastss 0x2f080(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,252,226,2,0 // vbroadcastss 0x2e2fc(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
.byte 197,253,91,201 // vcvtps2dq %ymm1,%ymm1
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -19723,7 +19723,7 @@ _sk_parametric_g_hsw:
.byte 196,195,117,74,201,128 // vblendvps %ymm8,%ymm9,%ymm1,%ymm1
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
- .byte 196,98,125,24,5,148,239,2,0 // vbroadcastss 0x2ef94(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,16,226,2,0 // vbroadcastss 0x2e210(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19743,33 +19743,33 @@ _sk_parametric_b_hsw:
.byte 196,66,109,168,211 // vfmadd213ps %ymm11,%ymm2,%ymm10
.byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,224,239,2,0 // vbroadcastss 0x2efe0(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,98,125,24,45,219,239,2,0 // vbroadcastss 0x2efdb(%rip),%ymm13 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,92,226,2,0 // vbroadcastss 0x2e25c(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,45,87,226,2,0 // vbroadcastss 0x2e257(%rip),%ymm13 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,65,44,84,213 // vandps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,45,239,2,0 // vbroadcastss 0x2ef2d(%rip),%ymm13 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,45,169,225,2,0 // vbroadcastss 0x2e1a9(%rip),%ymm13 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,86,213 // vorps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,195,239,2,0 // vbroadcastss 0x2efc3(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,63,226,2,0 // vbroadcastss 0x2e23f(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,37,184,236 // vfmadd231ps %ymm12,%ymm11,%ymm13
- .byte 196,98,125,24,29,185,239,2,0 // vbroadcastss 0x2efb9(%rip),%ymm11 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,29,53,226,2,0 // vbroadcastss 0x2e235(%rip),%ymm11 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,45,172,221 // vfnmadd213ps %ymm13,%ymm10,%ymm11
- .byte 196,98,125,24,37,175,239,2,0 // vbroadcastss 0x2efaf(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,43,226,2,0 // vbroadcastss 0x2e22b(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,165,239,2,0 // vbroadcastss 0x2efa5(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,33,226,2,0 // vbroadcastss 0x2e221(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,28,94,210 // vdivps %ymm10,%ymm12,%ymm10
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
.byte 196,193,108,89,210 // vmulps %ymm10,%ymm2,%ymm2
.byte 196,99,125,8,210,1 // vroundps $0x1,%ymm2,%ymm10
.byte 196,65,108,92,210 // vsubps %ymm10,%ymm2,%ymm10
- .byte 196,98,125,24,29,134,239,2,0 // vbroadcastss 0x2ef86(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,2,226,2,0 // vbroadcastss 0x2e202(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
- .byte 196,98,125,24,29,124,239,2,0 // vbroadcastss 0x2ef7c(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,248,225,2,0 // vbroadcastss 0x2e1f8(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,98,45,172,218 // vfnmadd213ps %ymm2,%ymm10,%ymm11
- .byte 196,226,125,24,21,114,239,2,0 // vbroadcastss 0x2ef72(%rip),%ymm2 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,21,238,225,2,0 // vbroadcastss 0x2e1ee(%rip),%ymm2 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,193,108,92,210 // vsubps %ymm10,%ymm2,%ymm2
- .byte 196,98,125,24,21,104,239,2,0 // vbroadcastss 0x2ef68(%rip),%ymm10 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,21,228,225,2,0 // vbroadcastss 0x2e1e4(%rip),%ymm10 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 197,172,94,210 // vdivps %ymm2,%ymm10,%ymm2
.byte 197,164,88,210 // vaddps %ymm2,%ymm11,%ymm2
- .byte 196,98,125,24,21,91,239,2,0 // vbroadcastss 0x2ef5b(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,215,225,2,0 // vbroadcastss 0x2e1d7(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,108,89,210 // vmulps %ymm10,%ymm2,%ymm2
.byte 197,253,91,210 // vcvtps2dq %ymm2,%ymm2
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -19777,7 +19777,7 @@ _sk_parametric_b_hsw:
.byte 196,195,109,74,209,128 // vblendvps %ymm8,%ymm9,%ymm2,%ymm2
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,111,238,2,0 // vbroadcastss 0x2ee6f(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,235,224,2,0 // vbroadcastss 0x2e0eb(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19797,33 +19797,33 @@ _sk_parametric_a_hsw:
.byte 196,66,101,168,211 // vfmadd213ps %ymm11,%ymm3,%ymm10
.byte 196,226,125,24,24 // vbroadcastss (%rax),%ymm3
.byte 196,65,124,91,218 // vcvtdq2ps %ymm10,%ymm11
- .byte 196,98,125,24,37,187,238,2,0 // vbroadcastss 0x2eebb(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,98,125,24,45,182,238,2,0 // vbroadcastss 0x2eeb6(%rip),%ymm13 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,55,225,2,0 // vbroadcastss 0x2e137(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,45,50,225,2,0 // vbroadcastss 0x2e132(%rip),%ymm13 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,65,44,84,213 // vandps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,8,238,2,0 // vbroadcastss 0x2ee08(%rip),%ymm13 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,45,132,224,2,0 // vbroadcastss 0x2e084(%rip),%ymm13 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,86,213 // vorps %ymm13,%ymm10,%ymm10
- .byte 196,98,125,24,45,158,238,2,0 // vbroadcastss 0x2ee9e(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,26,225,2,0 // vbroadcastss 0x2e11a(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,37,184,236 // vfmadd231ps %ymm12,%ymm11,%ymm13
- .byte 196,98,125,24,29,148,238,2,0 // vbroadcastss 0x2ee94(%rip),%ymm11 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,29,16,225,2,0 // vbroadcastss 0x2e110(%rip),%ymm11 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,45,172,221 // vfnmadd213ps %ymm13,%ymm10,%ymm11
- .byte 196,98,125,24,37,138,238,2,0 // vbroadcastss 0x2ee8a(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,6,225,2,0 // vbroadcastss 0x2e106(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,128,238,2,0 // vbroadcastss 0x2ee80(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,252,224,2,0 // vbroadcastss 0x2e0fc(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,28,94,210 // vdivps %ymm10,%ymm12,%ymm10
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
.byte 196,193,100,89,218 // vmulps %ymm10,%ymm3,%ymm3
.byte 196,99,125,8,211,1 // vroundps $0x1,%ymm3,%ymm10
.byte 196,65,100,92,210 // vsubps %ymm10,%ymm3,%ymm10
- .byte 196,98,125,24,29,97,238,2,0 // vbroadcastss 0x2ee61(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,221,224,2,0 // vbroadcastss 0x2e0dd(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,100,88,219 // vaddps %ymm11,%ymm3,%ymm3
- .byte 196,98,125,24,29,87,238,2,0 // vbroadcastss 0x2ee57(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,211,224,2,0 // vbroadcastss 0x2e0d3(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,98,45,172,219 // vfnmadd213ps %ymm3,%ymm10,%ymm11
- .byte 196,226,125,24,29,77,238,2,0 // vbroadcastss 0x2ee4d(%rip),%ymm3 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,29,201,224,2,0 // vbroadcastss 0x2e0c9(%rip),%ymm3 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,193,100,92,218 // vsubps %ymm10,%ymm3,%ymm3
- .byte 196,98,125,24,21,67,238,2,0 // vbroadcastss 0x2ee43(%rip),%ymm10 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,21,191,224,2,0 // vbroadcastss 0x2e0bf(%rip),%ymm10 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 197,172,94,219 // vdivps %ymm3,%ymm10,%ymm3
.byte 197,164,88,219 // vaddps %ymm3,%ymm11,%ymm3
- .byte 196,98,125,24,21,54,238,2,0 // vbroadcastss 0x2ee36(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,178,224,2,0 // vbroadcastss 0x2e0b2(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,100,89,218 // vmulps %ymm10,%ymm3,%ymm3
.byte 197,253,91,219 // vcvtps2dq %ymm3,%ymm3
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -19831,7 +19831,7 @@ _sk_parametric_a_hsw:
.byte 196,195,101,74,217,128 // vblendvps %ymm8,%ymm9,%ymm3,%ymm3
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
- .byte 196,98,125,24,5,74,237,2,0 // vbroadcastss 0x2ed4a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,198,223,2,0 // vbroadcastss 0x2dfc6(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19850,35 +19850,35 @@ _sk_gamma_hsw:
.byte 197,252,40,241 // vmovaps %ymm1,%ymm6
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,124,91,208 // vcvtdq2ps %ymm0,%ymm10
- .byte 196,98,125,24,29,161,237,2,0 // vbroadcastss 0x2eda1(%rip),%ymm11 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
- .byte 196,226,125,24,45,156,237,2,0 // vbroadcastss 0x2ed9c(%rip),%ymm5 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,29,29,224,2,0 // vbroadcastss 0x2e01d(%rip),%ymm11 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,226,125,24,45,24,224,2,0 // vbroadcastss 0x2e018(%rip),%ymm5 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 197,124,84,205 // vandps %ymm5,%ymm0,%ymm9
- .byte 196,226,125,24,37,239,236,2,0 // vbroadcastss 0x2ecef(%rip),%ymm4 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,37,107,223,2,0 // vbroadcastss 0x2df6b(%rip),%ymm4 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,52,86,228 // vorps %ymm4,%ymm9,%ymm12
- .byte 196,98,125,24,45,134,237,2,0 // vbroadcastss 0x2ed86(%rip),%ymm13 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,45,2,224,2,0 // vbroadcastss 0x2e002(%rip),%ymm13 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,66,37,168,213 // vfmadd213ps %ymm13,%ymm11,%ymm10
- .byte 196,98,125,24,13,124,237,2,0 // vbroadcastss 0x2ed7c(%rip),%ymm9 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,13,248,223,2,0 // vbroadcastss 0x2dff8(%rip),%ymm9 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,66,29,188,209 // vfnmadd231ps %ymm9,%ymm12,%ymm10
.byte 197,124,91,246 // vcvtdq2ps %ymm6,%ymm14
.byte 196,66,37,168,245 // vfmadd213ps %ymm13,%ymm11,%ymm14
.byte 197,124,91,255 // vcvtdq2ps %ymm7,%ymm15
.byte 196,66,37,168,253 // vfmadd213ps %ymm13,%ymm11,%ymm15
- .byte 196,98,125,24,29,96,237,2,0 // vbroadcastss 0x2ed60(%rip),%ymm11 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,29,220,223,2,0 // vbroadcastss 0x2dfdc(%rip),%ymm11 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,65,28,88,227 // vaddps %ymm11,%ymm12,%ymm12
- .byte 196,98,125,24,45,86,237,2,0 // vbroadcastss 0x2ed56(%rip),%ymm13 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,45,210,223,2,0 // vbroadcastss 0x2dfd2(%rip),%ymm13 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 196,65,20,94,228 // vdivps %ymm12,%ymm13,%ymm12
.byte 196,65,44,92,212 // vsubps %ymm12,%ymm10,%ymm10
.byte 196,98,125,24,32 // vbroadcastss (%rax),%ymm12
.byte 196,65,44,89,212 // vmulps %ymm12,%ymm10,%ymm10
.byte 196,67,125,8,194,1 // vroundps $0x1,%ymm10,%ymm8
.byte 196,65,44,92,192 // vsubps %ymm8,%ymm10,%ymm8
- .byte 196,226,125,24,21,50,237,2,0 // vbroadcastss 0x2ed32(%rip),%ymm2 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,226,125,24,21,174,223,2,0 // vbroadcastss 0x2dfae(%rip),%ymm2 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 197,44,88,210 // vaddps %ymm2,%ymm10,%ymm10
- .byte 196,226,125,24,29,41,237,2,0 // vbroadcastss 0x2ed29(%rip),%ymm3 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,226,125,24,29,165,223,2,0 // vbroadcastss 0x2dfa5(%rip),%ymm3 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,98,61,188,211 // vfnmadd231ps %ymm3,%ymm8,%ymm10
- .byte 196,226,125,24,13,31,237,2,0 // vbroadcastss 0x2ed1f(%rip),%ymm1 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,13,155,223,2,0 // vbroadcastss 0x2df9b(%rip),%ymm1 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,116,92,192 // vsubps %ymm8,%ymm1,%ymm8
- .byte 196,226,125,24,5,21,237,2,0 // vbroadcastss 0x2ed15(%rip),%ymm0 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,226,125,24,5,145,223,2,0 // vbroadcastss 0x2df91(%rip),%ymm0 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,124,94,192 // vdivps %ymm8,%ymm0,%ymm8
.byte 196,65,44,88,192 // vaddps %ymm8,%ymm10,%ymm8
.byte 197,204,84,245 // vandps %ymm5,%ymm6,%ymm6
@@ -19909,7 +19909,7 @@ _sk_gamma_hsw:
.byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
.byte 197,252,94,193 // vdivps %ymm1,%ymm0,%ymm0
.byte 197,236,88,192 // vaddps %ymm0,%ymm2,%ymm0
- .byte 196,226,125,24,13,134,236,2,0 // vbroadcastss 0x2ec86(%rip),%ymm1 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,226,125,24,13,2,223,2,0 // vbroadcastss 0x2df02(%rip),%ymm1 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,188,89,209 // vmulps %ymm1,%ymm8,%ymm2
.byte 197,204,89,217 // vmulps %ymm1,%ymm6,%ymm3
.byte 197,252,89,225 // vmulps %ymm1,%ymm0,%ymm4
@@ -19929,26 +19929,26 @@ HIDDEN _sk_lab_to_xyz_hsw
.globl _sk_lab_to_xyz_hsw
FUNCTION(_sk_lab_to_xyz_hsw)
_sk_lab_to_xyz_hsw:
- .byte 196,98,125,24,5,68,236,2,0 // vbroadcastss 0x2ec44(%rip),%ymm8 # 397fc <_sk_srcover_bgra_8888_sse2_lowp+0x374>
- .byte 196,98,125,24,13,163,235,2,0 // vbroadcastss 0x2eba3(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,98,125,24,21,54,236,2,0 // vbroadcastss 0x2ec36(%rip),%ymm10 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x378>
+ .byte 196,98,125,24,5,192,222,2,0 // vbroadcastss 0x2dec0(%rip),%ymm8 # 38a78 <_sk_srcover_bgra_8888_sse2_lowp+0x374>
+ .byte 196,98,125,24,13,31,222,2,0 // vbroadcastss 0x2de1f(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,21,178,222,2,0 // vbroadcastss 0x2deb2(%rip),%ymm10 # 38a7c <_sk_srcover_bgra_8888_sse2_lowp+0x378>
.byte 196,194,53,168,202 // vfmadd213ps %ymm10,%ymm9,%ymm1
.byte 196,194,53,168,210 // vfmadd213ps %ymm10,%ymm9,%ymm2
- .byte 196,98,125,24,13,39,236,2,0 // vbroadcastss 0x2ec27(%rip),%ymm9 # 39804 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
+ .byte 196,98,125,24,13,163,222,2,0 // vbroadcastss 0x2dea3(%rip),%ymm9 # 38a80 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
.byte 196,66,125,184,200 // vfmadd231ps %ymm8,%ymm0,%ymm9
- .byte 196,226,125,24,5,29,236,2,0 // vbroadcastss 0x2ec1d(%rip),%ymm0 # 39808 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
+ .byte 196,226,125,24,5,153,222,2,0 // vbroadcastss 0x2de99(%rip),%ymm0 # 38a84 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
.byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
- .byte 196,98,125,24,5,20,236,2,0 // vbroadcastss 0x2ec14(%rip),%ymm8 # 3980c <_sk_srcover_bgra_8888_sse2_lowp+0x384>
+ .byte 196,98,125,24,5,144,222,2,0 // vbroadcastss 0x2de90(%rip),%ymm8 # 38a88 <_sk_srcover_bgra_8888_sse2_lowp+0x384>
.byte 196,98,117,168,192 // vfmadd213ps %ymm0,%ymm1,%ymm8
- .byte 196,98,125,24,13,10,236,2,0 // vbroadcastss 0x2ec0a(%rip),%ymm9 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x388>
+ .byte 196,98,125,24,13,134,222,2,0 // vbroadcastss 0x2de86(%rip),%ymm9 # 38a8c <_sk_srcover_bgra_8888_sse2_lowp+0x388>
.byte 196,98,109,172,200 // vfnmadd213ps %ymm0,%ymm2,%ymm9
.byte 196,193,60,89,200 // vmulps %ymm8,%ymm8,%ymm1
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
- .byte 196,226,125,24,21,247,235,2,0 // vbroadcastss 0x2ebf7(%rip),%ymm2 # 39814 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
+ .byte 196,226,125,24,21,115,222,2,0 // vbroadcastss 0x2de73(%rip),%ymm2 # 38a90 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
.byte 197,108,194,209,1 // vcmpltps %ymm1,%ymm2,%ymm10
- .byte 196,98,125,24,29,237,235,2,0 // vbroadcastss 0x2ebed(%rip),%ymm11 # 39818 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
+ .byte 196,98,125,24,29,105,222,2,0 // vbroadcastss 0x2de69(%rip),%ymm11 # 38a94 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
.byte 196,65,60,88,195 // vaddps %ymm11,%ymm8,%ymm8
- .byte 196,98,125,24,37,227,235,2,0 // vbroadcastss 0x2ebe3(%rip),%ymm12 # 3981c <_sk_srcover_bgra_8888_sse2_lowp+0x394>
+ .byte 196,98,125,24,37,95,222,2,0 // vbroadcastss 0x2de5f(%rip),%ymm12 # 38a98 <_sk_srcover_bgra_8888_sse2_lowp+0x394>
.byte 196,65,60,89,196 // vmulps %ymm12,%ymm8,%ymm8
.byte 196,99,61,74,193,160 // vblendvps %ymm10,%ymm1,%ymm8,%ymm8
.byte 197,252,89,200 // vmulps %ymm0,%ymm0,%ymm1
@@ -19963,9 +19963,9 @@ _sk_lab_to_xyz_hsw:
.byte 196,65,52,88,203 // vaddps %ymm11,%ymm9,%ymm9
.byte 196,65,52,89,204 // vmulps %ymm12,%ymm9,%ymm9
.byte 196,227,53,74,208,32 // vblendvps %ymm2,%ymm0,%ymm9,%ymm2
- .byte 196,226,125,24,5,152,235,2,0 // vbroadcastss 0x2eb98(%rip),%ymm0 # 39820 <_sk_srcover_bgra_8888_sse2_lowp+0x398>
+ .byte 196,226,125,24,5,20,222,2,0 // vbroadcastss 0x2de14(%rip),%ymm0 # 38a9c <_sk_srcover_bgra_8888_sse2_lowp+0x398>
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,143,235,2,0 // vbroadcastss 0x2eb8f(%rip),%ymm8 # 39824 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
+ .byte 196,98,125,24,5,11,222,2,0 // vbroadcastss 0x2de0b(%rip),%ymm8 # 38aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -19983,10 +19983,10 @@ _sk_load_a8_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,52 // jne acea <_sk_load_a8_hsw+0x4c>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,76,243,2,0 // vpand 0x2f34c(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,204,229,2,0 // vpand 0x2e5cc(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,218,234,2,0 // vbroadcastss 0x2eada(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,86,221,2,0 // vbroadcastss 0x2dd56(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
@@ -20059,10 +20059,10 @@ _sk_load_a8_dst_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,52 // jne ade8 <_sk_load_a8_dst_hsw+0x4c>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,78,242,2,0 // vpand 0x2f24e(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,206,228,2,0 // vpand 0x2e4ce(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,228 // vpmovzxwd %xmm4,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,220,233,2,0 // vbroadcastss 0x2e9dc(%rip),%ymm5 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,45,88,220,2,0 // vbroadcastss 0x2dc58(%rip),%ymm5 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,220,89,253 // vmulps %ymm5,%ymm4,%ymm7
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
@@ -20110,7 +20110,7 @@ _sk_load_a8_dst_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,221 // callq ffffffffde00ae90 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd1a08>
+ .byte 232,255,255,255,221 // callq ffffffffde00ae90 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd278c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,210 // callq *%rdx
@@ -20177,7 +20177,7 @@ _sk_gather_a8_hsw:
.byte 196,227,121,32,192,7 // vpinsrb $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,49,192 // vpmovzxbd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,35,232,2,0 // vbroadcastss 0x2e823(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,159,218,2,0 // vbroadcastss 0x2da9f(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
@@ -20199,7 +20199,7 @@ _sk_store_a8_hsw:
.byte 77,15,175,193 // imul %r9,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,160,231,2,0 // vbroadcastss 0x2e7a0(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,28,218,2,0 // vbroadcastss 0x2da1c(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,100,89,192 // vmulps %ymm8,%ymm3,%ymm8
.byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
.byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
@@ -20224,13 +20224,13 @@ _sk_store_a8_hsw:
.byte 196,3,121,20,4,24,0 // vpextrb $0x0,%xmm8,(%r8,%r11,1)
.byte 235,202 // jmp afe8 <_sk_store_a8_hsw+0x40>
.byte 196,3,121,20,68,24,2,4 // vpextrb $0x4,%xmm8,0x2(%r8,%r11,1)
- .byte 196,98,57,0,5,241,239,2,0 // vpshufb 0x2eff1(%rip),%xmm8,%xmm8 # 3a020 <_sk_srcover_bgra_8888_sse2_lowp+0xb98>
+ .byte 196,98,57,0,5,113,226,2,0 // vpshufb 0x2e271(%rip),%xmm8,%xmm8 # 392a0 <_sk_srcover_bgra_8888_sse2_lowp+0xb9c>
.byte 196,3,121,21,4,24,0 // vpextrw $0x0,%xmm8,(%r8,%r11,1)
.byte 235,176 // jmp afe8 <_sk_store_a8_hsw+0x40>
.byte 196,3,121,20,68,24,6,12 // vpextrb $0xc,%xmm8,0x6(%r8,%r11,1)
.byte 196,3,121,20,68,24,5,10 // vpextrb $0xa,%xmm8,0x5(%r8,%r11,1)
.byte 196,3,121,20,68,24,4,8 // vpextrb $0x8,%xmm8,0x4(%r8,%r11,1)
- .byte 196,98,57,0,5,215,239,2,0 // vpshufb 0x2efd7(%rip),%xmm8,%xmm8 # 3a030 <_sk_srcover_bgra_8888_sse2_lowp+0xba8>
+ .byte 196,98,57,0,5,87,226,2,0 // vpshufb 0x2e257(%rip),%xmm8,%xmm8 # 392b0 <_sk_srcover_bgra_8888_sse2_lowp+0xbac>
.byte 196,1,121,126,4,24 // vmovd %xmm8,(%r8,%r11,1)
.byte 235,135 // jmp afe8 <_sk_store_a8_hsw+0x40>
.byte 15,31,0 // nopl (%rax)
@@ -20267,13 +20267,13 @@ _sk_load_g8_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,57 // jne b0d1 <_sk_load_g8_hsw+0x51>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,106,239,2,0 // vpand 0x2ef6a(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,234,225,2,0 // vpand 0x2e1ea(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,248,230,2,0 // vbroadcastss 0x2e6f8(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,116,217,2,0 // vbroadcastss 0x2d974(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,113,230,2,0 // vbroadcastss 0x2e671(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,237,216,2,0 // vbroadcastss 0x2d8ed(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 255,224 // jmpq *%rax
@@ -20318,7 +20318,7 @@ _sk_load_g8_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,222 // jmpq ffffffffdf00b178 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefd1cf0>
+ .byte 233,255,255,255,222 // jmpq ffffffffdf00b178 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefd2a74>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,211 // callq *%rbx
@@ -20342,13 +20342,13 @@ _sk_load_g8_dst_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,57 // jne b1d5 <_sk_load_g8_dst_hsw+0x51>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,102,238,2,0 // vpand 0x2ee66(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,230,224,2,0 // vpand 0x2e0e6(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,125,51,228 // vpmovzxwd %xmm4,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,244,229,2,0 // vbroadcastss 0x2e5f4(%rip),%ymm5 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,45,112,216,2,0 // vbroadcastss 0x2d870(%rip),%ymm5 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,109,229,2,0 // vbroadcastss 0x2e56d(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,233,215,2,0 // vbroadcastss 0x2d7e9(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,236 // vmovaps %ymm4,%ymm5
.byte 197,252,40,244 // vmovaps %ymm4,%ymm6
.byte 255,224 // jmpq *%rax
@@ -20393,7 +20393,7 @@ _sk_load_g8_dst_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,222 // jmpq ffffffffdf00b27c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefd1df4>
+ .byte 233,255,255,255,222 // jmpq ffffffffdf00b27c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefd2b78>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,211 // callq *%rbx
@@ -20460,10 +20460,10 @@ _sk_gather_g8_hsw:
.byte 196,227,121,32,192,7 // vpinsrb $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,49,192 // vpmovzxbd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,55,228,2,0 // vbroadcastss 0x2e437(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,179,214,2,0 // vbroadcastss 0x2d6b3(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,176,227,2,0 // vbroadcastss 0x2e3b0(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,44,214,2,0 // vbroadcastss 0x2d62c(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 91 // pop %rbx
@@ -20487,23 +20487,23 @@ _sk_load_565_hsw:
.byte 117,114 // jne b426 <_sk_load_565_hsw+0x8d>
.byte 196,129,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm0
.byte 196,226,125,51,208 // vpmovzxwd %xmm0,%ymm2
- .byte 196,226,125,88,5,236,227,2,0 // vpbroadcastd 0x2e3ec(%rip),%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,88,5,104,214,2,0 // vpbroadcastd 0x2d668(%rip),%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,237,219,192 // vpand %ymm0,%ymm2,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,223,227,2,0 // vbroadcastss 0x2e3df(%rip),%ymm1 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,13,91,214,2,0 // vbroadcastss 0x2d65b(%rip),%ymm1 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,88,13,214,227,2,0 // vpbroadcastd 0x2e3d6(%rip),%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,88,13,82,214,2,0 // vpbroadcastd 0x2d652(%rip),%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,237,219,201 // vpand %ymm1,%ymm2,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,29,201,227,2,0 // vbroadcastss 0x2e3c9(%rip),%ymm3 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,29,69,214,2,0 // vbroadcastss 0x2d645(%rip),%ymm3 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
- .byte 196,226,125,88,29,192,227,2,0 // vpbroadcastd 0x2e3c0(%rip),%ymm3 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,88,29,60,214,2,0 // vpbroadcastd 0x2d63c(%rip),%ymm3 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,226,125,24,29,179,227,2,0 // vbroadcastss 0x2e3b3(%rip),%ymm3 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,29,47,214,2,0 // vbroadcastss 0x2d62f(%rip),%ymm3 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,20,227,2,0 // vbroadcastss 0x2e314(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,144,213,2,0 // vbroadcastss 0x2d590(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -20566,23 +20566,23 @@ _sk_load_565_dst_hsw:
.byte 117,114 // jne b551 <_sk_load_565_dst_hsw+0x8d>
.byte 196,129,122,111,36,88 // vmovdqu (%r8,%r11,2),%xmm4
.byte 196,226,125,51,244 // vpmovzxwd %xmm4,%ymm6
- .byte 196,226,125,88,37,193,226,2,0 // vpbroadcastd 0x2e2c1(%rip),%ymm4 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,88,37,61,213,2,0 // vpbroadcastd 0x2d53d(%rip),%ymm4 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,205,219,228 // vpand %ymm4,%ymm6,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,180,226,2,0 // vbroadcastss 0x2e2b4(%rip),%ymm5 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,45,48,213,2,0 // vbroadcastss 0x2d530(%rip),%ymm5 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
- .byte 196,226,125,88,45,171,226,2,0 // vpbroadcastd 0x2e2ab(%rip),%ymm5 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,88,45,39,213,2,0 // vpbroadcastd 0x2d527(%rip),%ymm5 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,205,219,237 // vpand %ymm5,%ymm6,%ymm5
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,226,125,24,61,158,226,2,0 // vbroadcastss 0x2e29e(%rip),%ymm7 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,61,26,213,2,0 // vbroadcastss 0x2d51a(%rip),%ymm7 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,212,89,239 // vmulps %ymm7,%ymm5,%ymm5
- .byte 196,226,125,88,61,149,226,2,0 // vpbroadcastd 0x2e295(%rip),%ymm7 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,88,61,17,213,2,0 // vpbroadcastd 0x2d511(%rip),%ymm7 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,205,219,247 // vpand %ymm7,%ymm6,%ymm6
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 196,226,125,24,61,136,226,2,0 // vbroadcastss 0x2e288(%rip),%ymm7 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,61,4,213,2,0 // vbroadcastss 0x2d504(%rip),%ymm7 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,204,89,247 // vmulps %ymm7,%ymm6,%ymm6
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,233,225,2,0 // vbroadcastss 0x2e1e9(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,101,212,2,0 // vbroadcastss 0x2d465(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -20688,23 +20688,23 @@ _sk_gather_565_hsw:
.byte 65,15,183,4,88 // movzwl (%r8,%rbx,2),%eax
.byte 197,249,196,192,7 // vpinsrw $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,51,208 // vpmovzxwd %xmm0,%ymm2
- .byte 196,226,125,88,5,220,224,2,0 // vpbroadcastd 0x2e0dc(%rip),%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,88,5,88,211,2,0 // vpbroadcastd 0x2d358(%rip),%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,237,219,192 // vpand %ymm0,%ymm2,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,207,224,2,0 // vbroadcastss 0x2e0cf(%rip),%ymm1 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,13,75,211,2,0 // vbroadcastss 0x2d34b(%rip),%ymm1 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,88,13,198,224,2,0 // vpbroadcastd 0x2e0c6(%rip),%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,88,13,66,211,2,0 // vpbroadcastd 0x2d342(%rip),%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,237,219,201 // vpand %ymm1,%ymm2,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,29,185,224,2,0 // vbroadcastss 0x2e0b9(%rip),%ymm3 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,29,53,211,2,0 // vbroadcastss 0x2d335(%rip),%ymm3 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
- .byte 196,226,125,88,29,176,224,2,0 // vpbroadcastd 0x2e0b0(%rip),%ymm3 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,88,29,44,211,2,0 // vpbroadcastd 0x2d32c(%rip),%ymm3 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,226,125,24,29,163,224,2,0 // vbroadcastss 0x2e0a3(%rip),%ymm3 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,29,31,211,2,0 // vbroadcastss 0x2d31f(%rip),%ymm3 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,4,224,2,0 // vbroadcastss 0x2e004(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,128,210,2,0 // vbroadcastss 0x2d280(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
.byte 65,94 // pop %r14
@@ -20722,11 +20722,11 @@ _sk_store_565_hsw:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,204,224,2,0 // vbroadcastss 0x2e0cc(%rip),%ymm8 # 39828 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
+ .byte 196,98,125,24,5,72,211,2,0 // vbroadcastss 0x2d348(%rip),%ymm8 # 38aa4 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,53,114,241,11 // vpslld $0xb,%ymm9,%ymm9
- .byte 196,98,125,24,21,183,224,2,0 // vbroadcastss 0x2e0b7(%rip),%ymm10 # 3982c <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
+ .byte 196,98,125,24,21,51,211,2,0 // vbroadcastss 0x2d333(%rip),%ymm10 # 38aa8 <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
.byte 196,65,116,89,210 // vmulps %ymm10,%ymm1,%ymm10
.byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
.byte 196,193,45,114,242,5 // vpslld $0x5,%ymm10,%ymm10
@@ -20778,7 +20778,7 @@ _sk_store_565_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,224 // callq ffffffffe100b828 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fd23a0>
+ .byte 232,255,255,255,224 // callq ffffffffe100b828 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fd3124>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // .byte 0xff
@@ -20798,25 +20798,25 @@ _sk_load_4444_hsw:
.byte 15,133,138,0,0,0 // jne b8d5 <_sk_load_4444_hsw+0xa9>
.byte 196,129,122,111,4,88 // vmovdqu (%r8,%r11,2),%xmm0
.byte 196,226,125,51,216 // vpmovzxwd %xmm0,%ymm3
- .byte 196,226,125,88,5,209,223,2,0 // vpbroadcastd 0x2dfd1(%rip),%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,88,5,77,210,2,0 // vpbroadcastd 0x2d24d(%rip),%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,229,219,192 // vpand %ymm0,%ymm3,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,196,223,2,0 // vbroadcastss 0x2dfc4(%rip),%ymm1 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,13,64,210,2,0 // vbroadcastss 0x2d240(%rip),%ymm1 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,88,13,187,223,2,0 // vpbroadcastd 0x2dfbb(%rip),%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,88,13,55,210,2,0 // vpbroadcastd 0x2d237(%rip),%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,229,219,201 // vpand %ymm1,%ymm3,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,174,223,2,0 // vbroadcastss 0x2dfae(%rip),%ymm2 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,21,42,210,2,0 // vbroadcastss 0x2d22a(%rip),%ymm2 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,244,89,202 // vmulps %ymm2,%ymm1,%ymm1
- .byte 196,226,125,88,21,165,223,2,0 // vpbroadcastd 0x2dfa5(%rip),%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,88,21,33,210,2,0 // vpbroadcastd 0x2d221(%rip),%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,229,219,210 // vpand %ymm2,%ymm3,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,98,125,24,5,152,223,2,0 // vbroadcastss 0x2df98(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,20,210,2,0 // vbroadcastss 0x2d214(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,88,5,142,223,2,0 // vpbroadcastd 0x2df8e(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,88,5,10,210,2,0 // vpbroadcastd 0x2d20a(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,101,219,216 // vpand %ymm8,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,128,223,2,0 // vbroadcastss 0x2df80(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,252,209,2,0 // vbroadcastss 0x2d1fc(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -20884,25 +20884,25 @@ _sk_load_4444_dst_hsw:
.byte 15,133,138,0,0,0 // jne ba21 <_sk_load_4444_dst_hsw+0xa9>
.byte 196,129,122,111,36,88 // vmovdqu (%r8,%r11,2),%xmm4
.byte 196,226,125,51,252 // vpmovzxwd %xmm4,%ymm7
- .byte 196,226,125,88,37,133,222,2,0 // vpbroadcastd 0x2de85(%rip),%ymm4 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,88,37,1,209,2,0 // vpbroadcastd 0x2d101(%rip),%ymm4 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,197,219,228 // vpand %ymm4,%ymm7,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,120,222,2,0 // vbroadcastss 0x2de78(%rip),%ymm5 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,45,244,208,2,0 // vbroadcastss 0x2d0f4(%rip),%ymm5 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
- .byte 196,226,125,88,45,111,222,2,0 // vpbroadcastd 0x2de6f(%rip),%ymm5 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,88,45,235,208,2,0 // vpbroadcastd 0x2d0eb(%rip),%ymm5 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,197,219,237 // vpand %ymm5,%ymm7,%ymm5
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,226,125,24,53,98,222,2,0 // vbroadcastss 0x2de62(%rip),%ymm6 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,53,222,208,2,0 // vbroadcastss 0x2d0de(%rip),%ymm6 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,212,89,238 // vmulps %ymm6,%ymm5,%ymm5
- .byte 196,226,125,88,53,89,222,2,0 // vpbroadcastd 0x2de59(%rip),%ymm6 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,88,53,213,208,2,0 // vpbroadcastd 0x2d0d5(%rip),%ymm6 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,197,219,246 // vpand %ymm6,%ymm7,%ymm6
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 196,98,125,24,5,76,222,2,0 // vbroadcastss 0x2de4c(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,200,208,2,0 // vbroadcastss 0x2d0c8(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,76,89,240 // vmulps %ymm8,%ymm6,%ymm6
- .byte 196,98,125,88,5,66,222,2,0 // vpbroadcastd 0x2de42(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,88,5,190,208,2,0 // vpbroadcastd 0x2d0be(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,69,219,248 // vpand %ymm8,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,52,222,2,0 // vbroadcastss 0x2de34(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,176,208,2,0 // vbroadcastss 0x2d0b0(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,68,89,248 // vmulps %ymm8,%ymm7,%ymm7
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -21010,25 +21010,25 @@ _sk_gather_4444_hsw:
.byte 65,15,183,4,88 // movzwl (%r8,%rbx,2),%eax
.byte 197,249,196,192,7 // vpinsrw $0x7,%eax,%xmm0,%xmm0
.byte 196,226,125,51,216 // vpmovzxwd %xmm0,%ymm3
- .byte 196,226,125,88,5,132,220,2,0 // vpbroadcastd 0x2dc84(%rip),%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,88,5,0,207,2,0 // vpbroadcastd 0x2cf00(%rip),%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,229,219,192 // vpand %ymm0,%ymm3,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,119,220,2,0 // vbroadcastss 0x2dc77(%rip),%ymm1 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,13,243,206,2,0 // vbroadcastss 0x2cef3(%rip),%ymm1 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,88,13,110,220,2,0 // vpbroadcastd 0x2dc6e(%rip),%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,88,13,234,206,2,0 // vpbroadcastd 0x2ceea(%rip),%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,229,219,201 // vpand %ymm1,%ymm3,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,97,220,2,0 // vbroadcastss 0x2dc61(%rip),%ymm2 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,21,221,206,2,0 // vbroadcastss 0x2cedd(%rip),%ymm2 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,244,89,202 // vmulps %ymm2,%ymm1,%ymm1
- .byte 196,226,125,88,21,88,220,2,0 // vpbroadcastd 0x2dc58(%rip),%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,88,21,212,206,2,0 // vpbroadcastd 0x2ced4(%rip),%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,229,219,210 // vpand %ymm2,%ymm3,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,98,125,24,5,75,220,2,0 // vbroadcastss 0x2dc4b(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,199,206,2,0 // vbroadcastss 0x2cec7(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,88,5,65,220,2,0 // vpbroadcastd 0x2dc41(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,88,5,189,206,2,0 // vpbroadcastd 0x2cebd(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,101,219,216 // vpand %ymm8,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,51,220,2,0 // vbroadcastss 0x2dc33(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,175,206,2,0 // vbroadcastss 0x2ceaf(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
@@ -21048,7 +21048,7 @@ _sk_store_4444_hsw:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,8,220,2,0 // vbroadcastss 0x2dc08(%rip),%ymm8 # 39850 <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
+ .byte 196,98,125,24,5,132,206,2,0 // vbroadcastss 0x2ce84(%rip),%ymm8 # 38acc <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,53,114,241,12 // vpslld $0xc,%ymm9,%ymm9
@@ -21107,7 +21107,7 @@ _sk_store_4444_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,224 // callq ffffffffe100bd20 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fd2898>
+ .byte 232,255,255,255,224 // callq ffffffffe100bd20 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fd361c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // .byte 0xff
@@ -21126,14 +21126,14 @@ _sk_load_8888_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,86 // jne bd96 <_sk_load_8888_hsw+0x72>
.byte 196,129,126,111,28,152 // vmovdqu (%r8,%r11,4),%ymm3
- .byte 197,229,219,5,242,220,2,0 // vpand 0x2dcf2(%rip),%ymm3,%ymm0 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x5b8>
+ .byte 197,229,219,5,114,207,2,0 // vpand 0x2cf72(%rip),%ymm3,%ymm0 # 38cc0 <_sk_srcover_bgra_8888_sse2_lowp+0x5bc>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,85,218,2,0 // vbroadcastss 0x2da55(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,209,204,2,0 // vbroadcastss 0x2ccd1(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
- .byte 196,226,101,0,13,247,220,2,0 // vpshufb 0x2dcf7(%rip),%ymm3,%ymm1 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x5d8>
+ .byte 196,226,101,0,13,119,207,2,0 // vpshufb 0x2cf77(%rip),%ymm3,%ymm1 # 38ce0 <_sk_srcover_bgra_8888_sse2_lowp+0x5dc>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
.byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
- .byte 196,226,101,0,21,5,221,2,0 // vpshufb 0x2dd05(%rip),%ymm3,%ymm2 # 39a80 <_sk_srcover_bgra_8888_sse2_lowp+0x5f8>
+ .byte 196,226,101,0,21,133,207,2,0 // vpshufb 0x2cf85(%rip),%ymm3,%ymm2 # 38d00 <_sk_srcover_bgra_8888_sse2_lowp+0x5fc>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
.byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
@@ -21209,14 +21209,14 @@ _sk_load_8888_dst_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,86 // jne beca <_sk_load_8888_dst_hsw+0x72>
.byte 196,129,126,111,60,152 // vmovdqu (%r8,%r11,4),%ymm7
- .byte 197,197,219,37,30,220,2,0 // vpand 0x2dc1e(%rip),%ymm7,%ymm4 # 39aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x618>
+ .byte 197,197,219,37,158,206,2,0 // vpand 0x2ce9e(%rip),%ymm7,%ymm4 # 38d20 <_sk_srcover_bgra_8888_sse2_lowp+0x61c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,98,125,24,5,33,217,2,0 // vbroadcastss 0x2d921(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,157,203,2,0 // vbroadcastss 0x2cb9d(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,92,89,224 // vmulps %ymm8,%ymm4,%ymm4
- .byte 196,226,69,0,45,35,220,2,0 // vpshufb 0x2dc23(%rip),%ymm7,%ymm5 # 39ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x638>
+ .byte 196,226,69,0,45,163,206,2,0 // vpshufb 0x2cea3(%rip),%ymm7,%ymm5 # 38d40 <_sk_srcover_bgra_8888_sse2_lowp+0x63c>
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
.byte 196,193,84,89,232 // vmulps %ymm8,%ymm5,%ymm5
- .byte 196,226,69,0,53,49,220,2,0 // vpshufb 0x2dc31(%rip),%ymm7,%ymm6 # 39ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x658>
+ .byte 196,226,69,0,53,177,206,2,0 // vpshufb 0x2ceb1(%rip),%ymm7,%ymm6 # 38d60 <_sk_srcover_bgra_8888_sse2_lowp+0x65c>
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
.byte 196,193,76,89,240 // vmulps %ymm8,%ymm6,%ymm6
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
@@ -21300,14 +21300,14 @@ _sk_gather_8888_hsw:
.byte 197,254,91,192 // vcvttps2dq %ymm0,%ymm0
.byte 197,245,254,192 // vpaddd %ymm0,%ymm1,%ymm0
.byte 196,66,101,144,4,128 // vpgatherdd %ymm3,(%r8,%ymm0,4),%ymm8
- .byte 197,189,219,5,29,219,2,0 // vpand 0x2db1d(%rip),%ymm8,%ymm0 # 39b00 <_sk_srcover_bgra_8888_sse2_lowp+0x678>
+ .byte 197,189,219,5,157,205,2,0 // vpand 0x2cd9d(%rip),%ymm8,%ymm0 # 38d80 <_sk_srcover_bgra_8888_sse2_lowp+0x67c>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,29,192,215,2,0 // vbroadcastss 0x2d7c0(%rip),%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,29,60,202,2,0 // vbroadcastss 0x2ca3c(%rip),%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
- .byte 196,226,61,0,13,35,219,2,0 // vpshufb 0x2db23(%rip),%ymm8,%ymm1 # 39b20 <_sk_srcover_bgra_8888_sse2_lowp+0x698>
+ .byte 196,226,61,0,13,163,205,2,0 // vpshufb 0x2cda3(%rip),%ymm8,%ymm1 # 38da0 <_sk_srcover_bgra_8888_sse2_lowp+0x69c>
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
- .byte 196,226,61,0,21,50,219,2,0 // vpshufb 0x2db32(%rip),%ymm8,%ymm2 # 39b40 <_sk_srcover_bgra_8888_sse2_lowp+0x6b8>
+ .byte 196,226,61,0,21,178,205,2,0 // vpshufb 0x2cdb2(%rip),%ymm8,%ymm2 # 38dc0 <_sk_srcover_bgra_8888_sse2_lowp+0x6bc>
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
.byte 196,193,61,114,208,24 // vpsrld $0x18,%ymm8,%ymm8
@@ -21327,7 +21327,7 @@ _sk_store_8888_hsw:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,27,215,2,0 // vbroadcastss 0x2d71b(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,151,201,2,0 // vbroadcastss 0x2c997(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -21381,7 +21381,7 @@ _sk_store_8888_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,219 // jmpq ffffffffdc00c124 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfd2c9c>
+ .byte 233,255,255,255,219 // jmpq ffffffffdc00c124 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfd3a20>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,205 // dec %ebp
@@ -21403,14 +21403,14 @@ _sk_load_bgra_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,86 // jne c19e <_sk_load_bgra_hsw+0x72>
.byte 196,129,126,111,28,152 // vmovdqu (%r8,%r11,4),%ymm3
- .byte 197,229,219,5,10,218,2,0 // vpand 0x2da0a(%rip),%ymm3,%ymm0 # 39b60 <_sk_srcover_bgra_8888_sse2_lowp+0x6d8>
+ .byte 197,229,219,5,138,204,2,0 // vpand 0x2cc8a(%rip),%ymm3,%ymm0 # 38de0 <_sk_srcover_bgra_8888_sse2_lowp+0x6dc>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,77,214,2,0 // vbroadcastss 0x2d64d(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,201,200,2,0 // vbroadcastss 0x2c8c9(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,208 // vmulps %ymm8,%ymm0,%ymm2
- .byte 196,226,101,0,5,15,218,2,0 // vpshufb 0x2da0f(%rip),%ymm3,%ymm0 # 39b80 <_sk_srcover_bgra_8888_sse2_lowp+0x6f8>
+ .byte 196,226,101,0,5,143,204,2,0 // vpshufb 0x2cc8f(%rip),%ymm3,%ymm0 # 38e00 <_sk_srcover_bgra_8888_sse2_lowp+0x6fc>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 196,193,124,89,200 // vmulps %ymm8,%ymm0,%ymm1
- .byte 196,226,101,0,5,29,218,2,0 // vpshufb 0x2da1d(%rip),%ymm3,%ymm0 # 39ba0 <_sk_srcover_bgra_8888_sse2_lowp+0x718>
+ .byte 196,226,101,0,5,157,204,2,0 // vpshufb 0x2cc9d(%rip),%ymm3,%ymm0 # 38e20 <_sk_srcover_bgra_8888_sse2_lowp+0x71c>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
@@ -21486,14 +21486,14 @@ _sk_load_bgra_dst_hsw:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,86 // jne c2d2 <_sk_load_bgra_dst_hsw+0x72>
.byte 196,129,126,111,60,152 // vmovdqu (%r8,%r11,4),%ymm7
- .byte 197,197,219,37,54,217,2,0 // vpand 0x2d936(%rip),%ymm7,%ymm4 # 39bc0 <_sk_srcover_bgra_8888_sse2_lowp+0x738>
+ .byte 197,197,219,37,182,203,2,0 // vpand 0x2cbb6(%rip),%ymm7,%ymm4 # 38e40 <_sk_srcover_bgra_8888_sse2_lowp+0x73c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,98,125,24,5,25,213,2,0 // vbroadcastss 0x2d519(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,149,199,2,0 // vbroadcastss 0x2c795(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,92,89,240 // vmulps %ymm8,%ymm4,%ymm6
- .byte 196,226,69,0,37,59,217,2,0 // vpshufb 0x2d93b(%rip),%ymm7,%ymm4 # 39be0 <_sk_srcover_bgra_8888_sse2_lowp+0x758>
+ .byte 196,226,69,0,37,187,203,2,0 // vpshufb 0x2cbbb(%rip),%ymm7,%ymm4 # 38e60 <_sk_srcover_bgra_8888_sse2_lowp+0x75c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
.byte 196,193,92,89,232 // vmulps %ymm8,%ymm4,%ymm5
- .byte 196,226,69,0,37,73,217,2,0 // vpshufb 0x2d949(%rip),%ymm7,%ymm4 # 39c00 <_sk_srcover_bgra_8888_sse2_lowp+0x778>
+ .byte 196,226,69,0,37,201,203,2,0 // vpshufb 0x2cbc9(%rip),%ymm7,%ymm4 # 38e80 <_sk_srcover_bgra_8888_sse2_lowp+0x77c>
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
.byte 196,193,92,89,224 // vmulps %ymm8,%ymm4,%ymm4
.byte 197,197,114,215,24 // vpsrld $0x18,%ymm7,%ymm7
@@ -21577,14 +21577,14 @@ _sk_gather_bgra_hsw:
.byte 197,254,91,192 // vcvttps2dq %ymm0,%ymm0
.byte 197,245,254,192 // vpaddd %ymm0,%ymm1,%ymm0
.byte 196,66,101,144,4,128 // vpgatherdd %ymm3,(%r8,%ymm0,4),%ymm8
- .byte 197,189,219,5,53,216,2,0 // vpand 0x2d835(%rip),%ymm8,%ymm0 # 39c20 <_sk_srcover_bgra_8888_sse2_lowp+0x798>
+ .byte 197,189,219,5,181,202,2,0 // vpand 0x2cab5(%rip),%ymm8,%ymm0 # 38ea0 <_sk_srcover_bgra_8888_sse2_lowp+0x79c>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,29,184,211,2,0 // vbroadcastss 0x2d3b8(%rip),%ymm3 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,29,52,198,2,0 // vbroadcastss 0x2c634(%rip),%ymm3 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,211 // vmulps %ymm3,%ymm0,%ymm2
- .byte 196,226,61,0,5,59,216,2,0 // vpshufb 0x2d83b(%rip),%ymm8,%ymm0 # 39c40 <_sk_srcover_bgra_8888_sse2_lowp+0x7b8>
+ .byte 196,226,61,0,5,187,202,2,0 // vpshufb 0x2cabb(%rip),%ymm8,%ymm0 # 38ec0 <_sk_srcover_bgra_8888_sse2_lowp+0x7bc>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 197,252,89,203 // vmulps %ymm3,%ymm0,%ymm1
- .byte 196,226,61,0,5,74,216,2,0 // vpshufb 0x2d84a(%rip),%ymm8,%ymm0 # 39c60 <_sk_srcover_bgra_8888_sse2_lowp+0x7d8>
+ .byte 196,226,61,0,5,202,202,2,0 // vpshufb 0x2caca(%rip),%ymm8,%ymm0 # 38ee0 <_sk_srcover_bgra_8888_sse2_lowp+0x7dc>
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
.byte 196,193,61,114,208,24 // vpsrld $0x18,%ymm8,%ymm8
@@ -21604,7 +21604,7 @@ _sk_store_bgra_hsw:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,19,211,2,0 // vbroadcastss 0x2d313(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,143,197,2,0 // vbroadcastss 0x2c58f(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,108,89,200 // vmulps %ymm8,%ymm2,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -21658,7 +21658,7 @@ _sk_store_bgra_hsw:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,219 // jmpq ffffffffdc00c52c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfd30a4>
+ .byte 233,255,255,255,219 // jmpq ffffffffdc00c52c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdbfd3e28>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,205 // dec %ebp
@@ -21919,7 +21919,7 @@ _sk_load_u16_be_hsw:
.byte 197,241,235,192 // vpor %xmm0,%xmm1,%xmm0
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,21,148,206,2,0 // vbroadcastss 0x2ce94(%rip),%ymm10 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,21,16,193,2,0 // vbroadcastss 0x2c110(%rip),%ymm10 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 197,185,109,202 // vpunpckhqdq %xmm2,%xmm8,%xmm1
.byte 197,233,113,241,8 // vpsllw $0x8,%xmm1,%xmm2
@@ -22011,7 +22011,7 @@ _sk_load_rgb_u16_be_hsw:
.byte 197,241,235,192 // vpor %xmm0,%xmm1,%xmm0
.byte 196,226,125,51,192 // vpmovzxwd %xmm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,3,205,2,0 // vbroadcastss 0x2cd03(%rip),%ymm9 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,13,127,191,2,0 // vbroadcastss 0x2bf7f(%rip),%ymm9 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,185,109,203 // vpunpckhqdq %xmm3,%xmm8,%xmm1
.byte 197,225,113,241,8 // vpsllw $0x8,%xmm1,%xmm3
@@ -22027,7 +22027,7 @@ _sk_load_rgb_u16_be_hsw:
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 196,193,108,89,209 // vmulps %ymm9,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,35,204,2,0 // vbroadcastss 0x2cc23(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,159,190,2,0 // vbroadcastss 0x2be9f(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 196,193,121,110,4,64 // vmovd (%r8,%rax,2),%xmm0
.byte 196,65,49,239,201 // vpxor %xmm9,%xmm9,%xmm9
@@ -22103,7 +22103,7 @@ _sk_store_u16_be_hsw:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 73,99,193 // movslq %r9d,%rax
- .byte 196,98,125,24,5,236,203,2,0 // vbroadcastss 0x2cbec(%rip),%ymm8 # 39854 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
+ .byte 196,98,125,24,5,104,190,2,0 // vbroadcastss 0x2be68(%rip),%ymm8 # 38ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,67,125,25,202,1 // vextractf128 $0x1,%ymm9,%xmm10
@@ -22262,7 +22262,7 @@ _sk_load_f32_hsw:
.byte 255,92,255,255 // lcall *-0x1(%rdi,%rdi,8)
.byte 255,70,255 // incl -0x1(%rsi)
.byte 255 // (bad)
- .byte 255,53,255,255,255,40 // pushq 0x28ffffff(%rip) # 2900cf38 <_sk_srcover_bgra_8888_sse2_lowp+0x28fd3ab0>
+ .byte 255,53,255,255,255,40 // pushq 0x28ffffff(%rip) # 2900cf38 <_sk_srcover_bgra_8888_sse2_lowp+0x28fd4834>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // .byte 0xff
@@ -22457,7 +22457,7 @@ _sk_mirror_x_hsw:
.byte 196,65,124,92,218 // vsubps %ymm10,%ymm0,%ymm11
.byte 196,193,58,88,192 // vaddss %xmm8,%xmm8,%xmm0
.byte 196,98,125,24,192 // vbroadcastss %xmm0,%ymm8
- .byte 197,178,89,5,33,197,2,0 // vmulss 0x2c521(%rip),%xmm9,%xmm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,5,157,183,2,0 // vmulss 0x2b79d(%rip),%xmm9,%xmm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,89,192 // vmulps %ymm0,%ymm11,%ymm0
.byte 196,227,125,8,192,1 // vroundps $0x1,%ymm0,%ymm0
@@ -22480,7 +22480,7 @@ _sk_mirror_y_hsw:
.byte 196,65,116,92,218 // vsubps %ymm10,%ymm1,%ymm11
.byte 196,193,58,88,200 // vaddss %xmm8,%xmm8,%xmm1
.byte 196,98,125,24,193 // vbroadcastss %xmm1,%ymm8
- .byte 197,178,89,13,208,196,2,0 // vmulss 0x2c4d0(%rip),%xmm9,%xmm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,13,76,183,2,0 // vmulss 0x2b74c(%rip),%xmm9,%xmm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,226,125,24,201 // vbroadcastss %xmm1,%ymm1
.byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
.byte 196,227,125,8,201,1 // vroundps $0x1,%ymm1,%ymm1
@@ -22498,7 +22498,7 @@ FUNCTION(_sk_clamp_x_1_hsw)
_sk_clamp_x_1_hsw:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,152,196,2,0 // vbroadcastss 0x2c498(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,20,183,2,0 // vbroadcastss 0x2b714(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -22511,7 +22511,7 @@ _sk_repeat_x_1_hsw:
.byte 196,193,124,92,192 // vsubps %ymm8,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,114,196,2,0 // vbroadcastss 0x2c472(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,238,182,2,0 // vbroadcastss 0x2b6ee(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -22520,9 +22520,9 @@ HIDDEN _sk_mirror_x_1_hsw
.globl _sk_mirror_x_1_hsw
FUNCTION(_sk_mirror_x_1_hsw)
_sk_mirror_x_1_hsw:
- .byte 196,98,125,24,5,120,196,2,0 // vbroadcastss 0x2c478(%rip),%ymm8 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,5,244,182,2,0 // vbroadcastss 0x2b6f4(%rip),%ymm8 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,13,78,196,2,0 // vbroadcastss 0x2c44e(%rip),%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,13,202,182,2,0 // vbroadcastss 0x2b6ca(%rip),%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,124,89,201 // vmulps %ymm9,%ymm0,%ymm9
.byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
.byte 196,65,52,88,201 // vaddps %ymm9,%ymm9,%ymm9
@@ -22532,7 +22532,7 @@ _sk_mirror_x_1_hsw:
.byte 197,60,92,200 // vsubps %ymm0,%ymm8,%ymm9
.byte 197,180,84,192 // vandps %ymm0,%ymm9,%ymm0
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,30,196,2,0 // vbroadcastss 0x2c41e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,154,182,2,0 // vbroadcastss 0x2b69a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -22541,11 +22541,11 @@ HIDDEN _sk_luminance_to_alpha_hsw
.globl _sk_luminance_to_alpha_hsw
FUNCTION(_sk_luminance_to_alpha_hsw)
_sk_luminance_to_alpha_hsw:
- .byte 196,226,125,24,29,48,197,2,0 // vbroadcastss 0x2c530(%rip),%ymm3 # 3985c <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
- .byte 196,98,125,24,5,35,197,2,0 // vbroadcastss 0x2c523(%rip),%ymm8 # 39858 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
+ .byte 196,226,125,24,29,172,183,2,0 // vbroadcastss 0x2b7ac(%rip),%ymm3 # 38ad8 <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
+ .byte 196,98,125,24,5,159,183,2,0 // vbroadcastss 0x2b79f(%rip),%ymm8 # 38ad4 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
.byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
.byte 196,226,125,184,203 // vfmadd231ps %ymm3,%ymm0,%ymm1
- .byte 196,226,125,24,29,24,197,2,0 // vbroadcastss 0x2c518(%rip),%ymm3 # 39860 <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
+ .byte 196,226,125,24,29,148,183,2,0 // vbroadcastss 0x2b794(%rip),%ymm3 # 38adc <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
.byte 196,226,109,168,217 // vfmadd213ps %ymm1,%ymm2,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
@@ -22824,7 +22824,7 @@ _sk_gradient_hsw:
.byte 76,139,72,72 // mov 0x48(%rax),%r9
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 65,186,1,0,0,0 // mov $0x1,%r10d
- .byte 196,226,125,24,21,181,191,2,0 // vbroadcastss 0x2bfb5(%rip),%ymm2 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,226,125,24,21,49,178,2,0 // vbroadcastss 0x2b231(%rip),%ymm2 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 196,65,53,239,201 // vpxor %ymm9,%ymm9,%ymm9
.byte 196,130,125,24,28,145 // vbroadcastss (%r9,%r10,4),%ymm3
.byte 197,228,194,216,2 // vcmpleps %ymm0,%ymm3,%ymm3
@@ -22927,24 +22927,24 @@ _sk_xy_to_unit_angle_hsw:
.byte 196,65,52,95,226 // vmaxps %ymm10,%ymm9,%ymm12
.byte 196,65,36,94,220 // vdivps %ymm12,%ymm11,%ymm11
.byte 196,65,36,89,227 // vmulps %ymm11,%ymm11,%ymm12
- .byte 196,98,125,24,45,48,191,2,0 // vbroadcastss 0x2bf30(%rip),%ymm13 # 39864 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
- .byte 196,98,125,24,53,43,191,2,0 // vbroadcastss 0x2bf2b(%rip),%ymm14 # 39868 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
+ .byte 196,98,125,24,45,172,177,2,0 // vbroadcastss 0x2b1ac(%rip),%ymm13 # 38ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
+ .byte 196,98,125,24,53,167,177,2,0 // vbroadcastss 0x2b1a7(%rip),%ymm14 # 38ae4 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
.byte 196,66,29,184,245 // vfmadd231ps %ymm13,%ymm12,%ymm14
- .byte 196,98,125,24,45,33,191,2,0 // vbroadcastss 0x2bf21(%rip),%ymm13 # 3986c <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
+ .byte 196,98,125,24,45,157,177,2,0 // vbroadcastss 0x2b19d(%rip),%ymm13 # 38ae8 <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
.byte 196,66,29,184,238 // vfmadd231ps %ymm14,%ymm12,%ymm13
- .byte 196,98,125,24,53,23,191,2,0 // vbroadcastss 0x2bf17(%rip),%ymm14 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
+ .byte 196,98,125,24,53,147,177,2,0 // vbroadcastss 0x2b193(%rip),%ymm14 # 38aec <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
.byte 196,66,29,184,245 // vfmadd231ps %ymm13,%ymm12,%ymm14
.byte 196,65,36,89,222 // vmulps %ymm14,%ymm11,%ymm11
.byte 196,65,52,194,202,1 // vcmpltps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,24,21,2,191,2,0 // vbroadcastss 0x2bf02(%rip),%ymm10 # 39874 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
+ .byte 196,98,125,24,21,126,177,2,0 // vbroadcastss 0x2b17e(%rip),%ymm10 # 38af0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
.byte 196,65,44,92,211 // vsubps %ymm11,%ymm10,%ymm10
.byte 196,67,37,74,202,144 // vblendvps %ymm9,%ymm10,%ymm11,%ymm9
.byte 196,193,124,194,192,1 // vcmpltps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,21,168,189,2,0 // vbroadcastss 0x2bda8(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,36,176,2,0 // vbroadcastss 0x2b024(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,92,209 // vsubps %ymm9,%ymm10,%ymm10
.byte 196,195,53,74,194,0 // vblendvps %ymm0,%ymm10,%ymm9,%ymm0
.byte 196,65,116,194,200,1 // vcmpltps %ymm8,%ymm1,%ymm9
- .byte 196,98,125,24,21,146,189,2,0 // vbroadcastss 0x2bd92(%rip),%ymm10 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,21,14,176,2,0 // vbroadcastss 0x2b00e(%rip),%ymm10 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,44,92,208 // vsubps %ymm0,%ymm10,%ymm10
.byte 196,195,125,74,194,144 // vblendvps %ymm9,%ymm10,%ymm0,%ymm0
.byte 196,65,124,194,200,3 // vcmpunordps %ymm8,%ymm0,%ymm9
@@ -22972,23 +22972,23 @@ _sk_xy_to_2pt_conical_quadratic_max_hsw:
.byte 197,50,89,80,76 // vmulss 0x4c(%rax),%xmm9,%xmm10
.byte 196,66,125,24,210 // vbroadcastss %xmm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 196,98,125,24,29,130,190,2,0 // vbroadcastss 0x2be82(%rip),%ymm11 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,29,254,176,2,0 // vbroadcastss 0x2b0fe(%rip),%ymm11 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
.byte 196,98,125,184,216 // vfmadd231ps %ymm0,%ymm0,%ymm11
.byte 196,193,50,89,193 // vmulss %xmm9,%xmm9,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 196,98,125,24,13,97,190,2,0 // vbroadcastss 0x2be61(%rip),%ymm9 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 196,98,125,24,13,221,176,2,0 // vbroadcastss 0x2b0dd(%rip),%ymm9 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,194,45,184,194 // vfmadd231ps %ymm10,%ymm10,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 196,98,125,24,13,68,190,2,0 // vbroadcastss 0x2be44(%rip),%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,13,192,176,2,0 // vbroadcastss 0x2b0c0(%rip),%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,44,87,201 // vxorps %ymm9,%ymm10,%ymm9
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,229,188,2,0 // vbroadcastss 0x2bce5(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,97,175,2,0 // vbroadcastss 0x2af61(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 196,65,44,89,208 // vmulps %ymm8,%ymm10,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
@@ -23007,23 +23007,23 @@ _sk_xy_to_2pt_conical_quadratic_min_hsw:
.byte 197,50,89,80,76 // vmulss 0x4c(%rax),%xmm9,%xmm10
.byte 196,66,125,24,210 // vbroadcastss %xmm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 196,98,125,24,29,234,189,2,0 // vbroadcastss 0x2bdea(%rip),%ymm11 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,29,102,176,2,0 // vbroadcastss 0x2b066(%rip),%ymm11 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
.byte 196,98,125,184,216 // vfmadd231ps %ymm0,%ymm0,%ymm11
.byte 196,193,50,89,193 // vmulss %xmm9,%xmm9,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 196,98,125,24,13,201,189,2,0 // vbroadcastss 0x2bdc9(%rip),%ymm9 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 196,98,125,24,13,69,176,2,0 // vbroadcastss 0x2b045(%rip),%ymm9 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,194,45,184,194 // vfmadd231ps %ymm10,%ymm10,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 196,98,125,24,13,172,189,2,0 // vbroadcastss 0x2bdac(%rip),%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,13,40,176,2,0 // vbroadcastss 0x2b028(%rip),%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,44,87,201 // vxorps %ymm9,%ymm10,%ymm9
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,77,188,2,0 // vbroadcastss 0x2bc4d(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,201,174,2,0 // vbroadcastss 0x2aec9(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 196,65,44,89,208 // vmulps %ymm8,%ymm10,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
@@ -23041,14 +23041,14 @@ _sk_xy_to_2pt_conical_linear_hsw:
.byte 197,58,89,72,76 // vmulss 0x4c(%rax),%xmm8,%xmm9
.byte 196,66,125,24,201 // vbroadcastss %xmm9,%ymm9
.byte 197,52,88,200 // vaddps %ymm0,%ymm9,%ymm9
- .byte 196,98,125,24,21,88,189,2,0 // vbroadcastss 0x2bd58(%rip),%ymm10 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,21,212,175,2,0 // vbroadcastss 0x2afd4(%rip),%ymm10 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
.byte 197,116,89,209 // vmulps %ymm1,%ymm1,%ymm10
.byte 196,98,125,184,208 // vfmadd231ps %ymm0,%ymm0,%ymm10
.byte 196,193,58,89,192 // vmulss %xmm8,%xmm8,%xmm0
.byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
.byte 197,172,92,192 // vsubps %ymm0,%ymm10,%ymm0
- .byte 196,98,125,24,5,59,189,2,0 // vbroadcastss 0x2bd3b(%rip),%ymm8 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,5,183,175,2,0 // vbroadcastss 0x2afb7(%rip),%ymm8 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,193,124,87,192 // vxorps %ymm8,%ymm0,%ymm0
.byte 196,193,124,94,193 // vdivps %ymm9,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23091,7 +23091,7 @@ HIDDEN _sk_save_xy_hsw
FUNCTION(_sk_save_xy_hsw)
_sk_save_xy_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,122,187,2,0 // vbroadcastss 0x2bb7a(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,246,173,2,0 // vbroadcastss 0x2adf6(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,124,88,200 // vaddps %ymm8,%ymm0,%ymm9
.byte 196,67,125,8,209,1 // vroundps $0x1,%ymm9,%ymm10
.byte 196,65,52,92,202 // vsubps %ymm10,%ymm9,%ymm9
@@ -23125,9 +23125,9 @@ HIDDEN _sk_bilinear_nx_hsw
FUNCTION(_sk_bilinear_nx_hsw)
_sk_bilinear_nx_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,84,188,2,0 // vbroadcastss 0x2bc54(%rip),%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,5,208,174,2,0 // vbroadcastss 0x2aed0(%rip),%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,251,186,2,0 // vbroadcastss 0x2bafb(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,119,173,2,0 // vbroadcastss 0x2ad77(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23138,7 +23138,7 @@ HIDDEN _sk_bilinear_px_hsw
FUNCTION(_sk_bilinear_px_hsw)
_sk_bilinear_px_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,216,186,2,0 // vbroadcastss 0x2bad8(%rip),%ymm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,5,84,173,2,0 // vbroadcastss 0x2ad54(%rip),%ymm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -23150,9 +23150,9 @@ HIDDEN _sk_bilinear_ny_hsw
FUNCTION(_sk_bilinear_ny_hsw)
_sk_bilinear_ny_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,5,188,2,0 // vbroadcastss 0x2bc05(%rip),%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,13,129,174,2,0 // vbroadcastss 0x2ae81(%rip),%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,171,186,2,0 // vbroadcastss 0x2baab(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,39,173,2,0 // vbroadcastss 0x2ad27(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23163,7 +23163,7 @@ HIDDEN _sk_bilinear_py_hsw
FUNCTION(_sk_bilinear_py_hsw)
_sk_bilinear_py_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,136,186,2,0 // vbroadcastss 0x2ba88(%rip),%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,13,4,173,2,0 // vbroadcastss 0x2ad04(%rip),%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -23175,13 +23175,13 @@ HIDDEN _sk_bicubic_n3x_hsw
FUNCTION(_sk_bicubic_n3x_hsw)
_sk_bicubic_n3x_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,184,187,2,0 // vbroadcastss 0x2bbb8(%rip),%ymm0 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,226,125,24,5,52,174,2,0 // vbroadcastss 0x2ae34(%rip),%ymm0 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,91,186,2,0 // vbroadcastss 0x2ba5b(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,215,172,2,0 // vbroadcastss 0x2acd7(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,185,186,2,0 // vbroadcastss 0x2bab9(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
- .byte 196,98,125,24,29,144,187,2,0 // vbroadcastss 0x2bb90(%rip),%ymm11 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,53,173,2,0 // vbroadcastss 0x2ad35(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,29,12,174,2,0 // vbroadcastss 0x2ae0c(%rip),%ymm11 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,66,61,168,218 // vfmadd213ps %ymm10,%ymm8,%ymm11
.byte 196,65,36,89,193 // vmulps %ymm9,%ymm11,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -23193,16 +23193,16 @@ HIDDEN _sk_bicubic_n1x_hsw
FUNCTION(_sk_bicubic_n1x_hsw)
_sk_bicubic_n1x_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,103,187,2,0 // vbroadcastss 0x2bb67(%rip),%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,5,227,173,2,0 // vbroadcastss 0x2ade3(%rip),%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,14,186,2,0 // vbroadcastss 0x2ba0e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,138,172,2,0 // vbroadcastss 0x2ac8a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,89,187,2,0 // vbroadcastss 0x2bb59(%rip),%ymm9 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 196,98,125,24,21,76,187,2,0 // vbroadcastss 0x2bb4c(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,13,213,173,2,0 // vbroadcastss 0x2add5(%rip),%ymm9 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,200,173,2,0 // vbroadcastss 0x2adc8(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,66,61,168,209 // vfmadd213ps %ymm9,%ymm8,%ymm10
- .byte 196,98,125,24,13,226,185,2,0 // vbroadcastss 0x2b9e2(%rip),%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,13,94,172,2,0 // vbroadcastss 0x2ac5e(%rip),%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,66,61,184,202 // vfmadd231ps %ymm10,%ymm8,%ymm9
- .byte 196,98,125,24,21,56,187,2,0 // vbroadcastss 0x2bb38(%rip),%ymm10 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,21,180,173,2,0 // vbroadcastss 0x2adb4(%rip),%ymm10 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,66,61,184,209 // vfmadd231ps %ymm9,%ymm8,%ymm10
.byte 197,124,17,144,0,1,0,0 // vmovups %ymm10,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23213,14 +23213,14 @@ HIDDEN _sk_bicubic_p1x_hsw
FUNCTION(_sk_bicubic_p1x_hsw)
_sk_bicubic_p1x_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,184,185,2,0 // vbroadcastss 0x2b9b8(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,52,172,2,0 // vbroadcastss 0x2ac34(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,0 // vaddps (%rax),%ymm8,%ymm0
.byte 197,124,16,136,128,0,0,0 // vmovups 0x80(%rax),%ymm9
- .byte 196,98,125,24,21,3,187,2,0 // vbroadcastss 0x2bb03(%rip),%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 196,98,125,24,29,246,186,2,0 // vbroadcastss 0x2baf6(%rip),%ymm11 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,21,127,173,2,0 // vbroadcastss 0x2ad7f(%rip),%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,29,114,173,2,0 // vbroadcastss 0x2ad72(%rip),%ymm11 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,66,53,168,218 // vfmadd213ps %ymm10,%ymm9,%ymm11
.byte 196,66,53,168,216 // vfmadd213ps %ymm8,%ymm9,%ymm11
- .byte 196,98,125,24,5,235,186,2,0 // vbroadcastss 0x2baeb(%rip),%ymm8 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,5,103,173,2,0 // vbroadcastss 0x2ad67(%rip),%ymm8 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,66,53,184,195 // vfmadd231ps %ymm11,%ymm9,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23231,12 +23231,12 @@ HIDDEN _sk_bicubic_p3x_hsw
FUNCTION(_sk_bicubic_p3x_hsw)
_sk_bicubic_p3x_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,203,186,2,0 // vbroadcastss 0x2bacb(%rip),%ymm0 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,226,125,24,5,71,173,2,0 // vbroadcastss 0x2ad47(%rip),%ymm0 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,201,185,2,0 // vbroadcastss 0x2b9c9(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
- .byte 196,98,125,24,29,160,186,2,0 // vbroadcastss 0x2baa0(%rip),%ymm11 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,69,172,2,0 // vbroadcastss 0x2ac45(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,29,28,173,2,0 // vbroadcastss 0x2ad1c(%rip),%ymm11 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,66,61,168,218 // vfmadd213ps %ymm10,%ymm8,%ymm11
.byte 196,65,52,89,195 // vmulps %ymm11,%ymm9,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -23248,13 +23248,13 @@ HIDDEN _sk_bicubic_n3y_hsw
FUNCTION(_sk_bicubic_n3y_hsw)
_sk_bicubic_n3y_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,123,186,2,0 // vbroadcastss 0x2ba7b(%rip),%ymm1 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,226,125,24,13,247,172,2,0 // vbroadcastss 0x2acf7(%rip),%ymm1 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,29,185,2,0 // vbroadcastss 0x2b91d(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,153,171,2,0 // vbroadcastss 0x2ab99(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,123,185,2,0 // vbroadcastss 0x2b97b(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
- .byte 196,98,125,24,29,82,186,2,0 // vbroadcastss 0x2ba52(%rip),%ymm11 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,247,171,2,0 // vbroadcastss 0x2abf7(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,29,206,172,2,0 // vbroadcastss 0x2acce(%rip),%ymm11 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,66,61,168,218 // vfmadd213ps %ymm10,%ymm8,%ymm11
.byte 196,65,36,89,193 // vmulps %ymm9,%ymm11,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -23266,16 +23266,16 @@ HIDDEN _sk_bicubic_n1y_hsw
FUNCTION(_sk_bicubic_n1y_hsw)
_sk_bicubic_n1y_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,41,186,2,0 // vbroadcastss 0x2ba29(%rip),%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,13,165,172,2,0 // vbroadcastss 0x2aca5(%rip),%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,207,184,2,0 // vbroadcastss 0x2b8cf(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,75,171,2,0 // vbroadcastss 0x2ab4b(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,26,186,2,0 // vbroadcastss 0x2ba1a(%rip),%ymm9 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 196,98,125,24,21,13,186,2,0 // vbroadcastss 0x2ba0d(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,13,150,172,2,0 // vbroadcastss 0x2ac96(%rip),%ymm9 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,137,172,2,0 // vbroadcastss 0x2ac89(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,66,61,168,209 // vfmadd213ps %ymm9,%ymm8,%ymm10
- .byte 196,98,125,24,13,163,184,2,0 // vbroadcastss 0x2b8a3(%rip),%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,13,31,171,2,0 // vbroadcastss 0x2ab1f(%rip),%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,66,61,184,202 // vfmadd231ps %ymm10,%ymm8,%ymm9
- .byte 196,98,125,24,21,249,185,2,0 // vbroadcastss 0x2b9f9(%rip),%ymm10 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,21,117,172,2,0 // vbroadcastss 0x2ac75(%rip),%ymm10 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,66,61,184,209 // vfmadd231ps %ymm9,%ymm8,%ymm10
.byte 197,124,17,144,64,1,0,0 // vmovups %ymm10,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23286,14 +23286,14 @@ HIDDEN _sk_bicubic_p1y_hsw
FUNCTION(_sk_bicubic_p1y_hsw)
_sk_bicubic_p1y_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,121,184,2,0 // vbroadcastss 0x2b879(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,245,170,2,0 // vbroadcastss 0x2aaf5(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,72,64 // vaddps 0x40(%rax),%ymm8,%ymm1
.byte 197,124,16,136,192,0,0,0 // vmovups 0xc0(%rax),%ymm9
- .byte 196,98,125,24,21,195,185,2,0 // vbroadcastss 0x2b9c3(%rip),%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
- .byte 196,98,125,24,29,182,185,2,0 // vbroadcastss 0x2b9b6(%rip),%ymm11 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,21,63,172,2,0 // vbroadcastss 0x2ac3f(%rip),%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,29,50,172,2,0 // vbroadcastss 0x2ac32(%rip),%ymm11 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,66,53,168,218 // vfmadd213ps %ymm10,%ymm9,%ymm11
.byte 196,66,53,168,216 // vfmadd213ps %ymm8,%ymm9,%ymm11
- .byte 196,98,125,24,5,171,185,2,0 // vbroadcastss 0x2b9ab(%rip),%ymm8 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,5,39,172,2,0 // vbroadcastss 0x2ac27(%rip),%ymm8 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,66,53,184,195 // vfmadd231ps %ymm11,%ymm9,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -23304,12 +23304,12 @@ HIDDEN _sk_bicubic_p3y_hsw
FUNCTION(_sk_bicubic_p3y_hsw)
_sk_bicubic_p3y_hsw:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,139,185,2,0 // vbroadcastss 0x2b98b(%rip),%ymm1 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,226,125,24,13,7,172,2,0 // vbroadcastss 0x2ac07(%rip),%ymm1 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,136,184,2,0 // vbroadcastss 0x2b888(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
- .byte 196,98,125,24,29,95,185,2,0 // vbroadcastss 0x2b95f(%rip),%ymm11 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,4,171,2,0 // vbroadcastss 0x2ab04(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,29,219,171,2,0 // vbroadcastss 0x2abdb(%rip),%ymm11 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,66,61,168,218 // vfmadd213ps %ymm10,%ymm8,%ymm11
.byte 196,65,52,89,195 // vmulps %ymm11,%ymm9,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -23449,7 +23449,7 @@ _sk_clut_3D_hsw:
.byte 196,98,93,64,218 // vpmulld %ymm2,%ymm4,%ymm11
.byte 72,139,0 // mov (%rax),%rax
.byte 197,165,254,193 // vpaddd %ymm1,%ymm11,%ymm0
- .byte 196,98,125,88,61,33,183,2,0 // vpbroadcastd 0x2b721(%rip),%ymm15 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 196,98,125,88,61,157,169,2,0 // vpbroadcastd 0x2a99d(%rip),%ymm15 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 196,194,125,64,199 // vpmulld %ymm15,%ymm0,%ymm0
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
.byte 197,237,118,210 // vpcmpeqd %ymm2,%ymm2,%ymm2
@@ -23461,13 +23461,13 @@ _sk_clut_3D_hsw:
.byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
.byte 196,226,85,146,60,144 // vgatherdps %ymm5,(%rax,%ymm2,4),%ymm7
.byte 197,252,17,188,36,0,2,0,0 // vmovups %ymm7,0x200(%rsp)
- .byte 196,226,125,88,61,127,181,2,0 // vpbroadcastd 0x2b57f(%rip),%ymm7 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,226,125,88,61,251,167,2,0 // vpbroadcastd 0x2a7fb(%rip),%ymm7 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 197,253,254,199 // vpaddd %ymm7,%ymm0,%ymm0
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
.byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
.byte 196,226,85,146,20,128 // vgatherdps %ymm5,(%rax,%ymm0,4),%ymm2
.byte 197,252,17,148,36,32,1,0,0 // vmovups %ymm2,0x120(%rsp)
- .byte 196,226,125,24,5,179,182,2,0 // vbroadcastss 0x2b6b3(%rip),%ymm0 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,226,125,24,5,47,169,2,0 // vbroadcastss 0x2a92f(%rip),%ymm0 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 197,172,88,232 // vaddps %ymm0,%ymm10,%ymm5
.byte 197,254,91,237 // vcvttps2dq %ymm5,%ymm5
.byte 196,226,93,64,213 // vpmulld %ymm5,%ymm4,%ymm2
@@ -23729,7 +23729,7 @@ _sk_clut_4D_hsw:
.byte 197,254,127,132,36,0,1,0,0 // vmovdqu %ymm0,0x100(%rsp)
.byte 196,98,109,64,200 // vpmulld %ymm0,%ymm2,%ymm9
.byte 197,181,254,199 // vpaddd %ymm7,%ymm9,%ymm0
- .byte 196,98,125,88,21,76,177,2,0 // vpbroadcastd 0x2b14c(%rip),%ymm10 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 196,98,125,88,21,200,163,2,0 // vpbroadcastd 0x2a3c8(%rip),%ymm10 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 196,194,125,64,194 // vpmulld %ymm10,%ymm0,%ymm0
.byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
@@ -23741,13 +23741,13 @@ _sk_clut_4D_hsw:
.byte 196,65,36,87,219 // vxorps %ymm11,%ymm11,%ymm11
.byte 196,98,61,146,28,168 // vgatherdps %ymm8,(%rax,%ymm5,4),%ymm11
.byte 197,124,17,156,36,192,0,0,0 // vmovups %ymm11,0xc0(%rsp)
- .byte 196,98,125,88,29,167,175,2,0 // vpbroadcastd 0x2afa7(%rip),%ymm11 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,125,88,29,35,162,2,0 // vpbroadcastd 0x2a223(%rip),%ymm11 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,193,125,254,195 // vpaddd %ymm11,%ymm0,%ymm0
.byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
.byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
.byte 196,226,61,146,44,128 // vgatherdps %ymm8,(%rax,%ymm0,4),%ymm5
.byte 197,252,17,108,36,32 // vmovups %ymm5,0x20(%rsp)
- .byte 196,226,125,24,5,220,176,2,0 // vbroadcastss 0x2b0dc(%rip),%ymm0 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,226,125,24,5,88,163,2,0 // vbroadcastss 0x2a358(%rip),%ymm0 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 197,116,88,192 // vaddps %ymm0,%ymm1,%ymm8
.byte 196,65,126,91,192 // vcvttps2dq %ymm8,%ymm8
.byte 196,194,109,64,232 // vpmulld %ymm8,%ymm2,%ymm5
@@ -24180,7 +24180,7 @@ _sk_clut_4D_hsw:
.byte 196,193,100,92,210 // vsubps %ymm10,%ymm3,%ymm2
.byte 196,194,77,168,210 // vfmadd213ps %ymm10,%ymm6,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,202,165,2,0 // vbroadcastss 0x2a5ca(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,70,152,2,0 // vbroadcastss 0x29846(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,16,164,36,96,6,0,0 // vmovups 0x660(%rsp),%ymm4
.byte 197,252,16,172,36,128,6,0,0 // vmovups 0x680(%rsp),%ymm5
.byte 197,252,16,180,36,160,6,0,0 // vmovups 0x6a0(%rsp),%ymm6
@@ -24192,14 +24192,14 @@ HIDDEN _sk_gauss_a_to_rgba_hsw
.globl _sk_gauss_a_to_rgba_hsw
FUNCTION(_sk_gauss_a_to_rgba_hsw)
_sk_gauss_a_to_rgba_hsw:
- .byte 196,226,125,24,5,4,167,2,0 // vbroadcastss 0x2a704(%rip),%ymm0 # 398a8 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
- .byte 196,226,125,24,13,247,166,2,0 // vbroadcastss 0x2a6f7(%rip),%ymm1 # 398a4 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
+ .byte 196,226,125,24,5,128,153,2,0 // vbroadcastss 0x29980(%rip),%ymm0 # 38b24 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
+ .byte 196,226,125,24,13,115,153,2,0 // vbroadcastss 0x29973(%rip),%ymm1 # 38b20 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
.byte 196,226,101,168,200 // vfmadd213ps %ymm0,%ymm3,%ymm1
- .byte 196,226,125,24,5,241,166,2,0 // vbroadcastss 0x2a6f1(%rip),%ymm0 # 398ac <_sk_srcover_bgra_8888_sse2_lowp+0x424>
+ .byte 196,226,125,24,5,109,153,2,0 // vbroadcastss 0x2996d(%rip),%ymm0 # 38b28 <_sk_srcover_bgra_8888_sse2_lowp+0x424>
.byte 196,226,101,184,193 // vfmadd231ps %ymm1,%ymm3,%ymm0
- .byte 196,226,125,24,13,231,166,2,0 // vbroadcastss 0x2a6e7(%rip),%ymm1 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x428>
+ .byte 196,226,125,24,13,99,153,2,0 // vbroadcastss 0x29963(%rip),%ymm1 # 38b2c <_sk_srcover_bgra_8888_sse2_lowp+0x428>
.byte 196,226,101,184,200 // vfmadd231ps %ymm0,%ymm3,%ymm1
- .byte 196,226,125,24,5,221,166,2,0 // vbroadcastss 0x2a6dd(%rip),%ymm0 # 398b4 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
+ .byte 196,226,125,24,5,89,153,2,0 // vbroadcastss 0x29959(%rip),%ymm0 # 38b30 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
.byte 196,226,101,184,193 // vfmadd231ps %ymm1,%ymm3,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
@@ -24301,10 +24301,10 @@ _sk_seed_shader_avx:
.byte 197,249,112,201,0 // vpshufd $0x0,%xmm1,%xmm1
.byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,54,164,2,0 // vbroadcastss 0x2a436(%rip),%ymm2 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,21,178,150,2,0 // vbroadcastss 0x296b2(%rip),%ymm2 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,244,88,202 // vaddps %ymm2,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,21,43,164,2,0 // vbroadcastss 0x2a42b(%rip),%ymm2 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,21,167,150,2,0 // vbroadcastss 0x296a7(%rip),%ymm2 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
.byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
@@ -24318,46 +24318,46 @@ FUNCTION(_sk_dither_avx)
_sk_dither_avx:
.byte 197,121,110,194 // vmovd %edx,%xmm8
.byte 196,65,121,112,192,0 // vpshufd $0x0,%xmm8,%xmm8
- .byte 197,57,254,13,11,173,2,0 // vpaddd 0x2ad0b(%rip),%xmm8,%xmm9 # 3a040 <_sk_srcover_bgra_8888_sse2_lowp+0xbb8>
- .byte 197,57,254,5,19,173,2,0 // vpaddd 0x2ad13(%rip),%xmm8,%xmm8 # 3a050 <_sk_srcover_bgra_8888_sse2_lowp+0xbc8>
+ .byte 197,57,254,13,139,159,2,0 // vpaddd 0x29f8b(%rip),%xmm8,%xmm9 # 392c0 <_sk_srcover_bgra_8888_sse2_lowp+0xbbc>
+ .byte 197,57,254,5,147,159,2,0 // vpaddd 0x29f93(%rip),%xmm8,%xmm8 # 392d0 <_sk_srcover_bgra_8888_sse2_lowp+0xbcc>
.byte 196,67,53,24,208,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm10
.byte 197,121,110,217 // vmovd %ecx,%xmm11
.byte 196,65,121,112,219,0 // vpshufd $0x0,%xmm11,%xmm11
.byte 196,67,37,24,219,1 // vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
.byte 196,65,36,87,218 // vxorps %ymm10,%ymm11,%ymm11
- .byte 196,98,125,24,21,219,163,2,0 // vbroadcastss 0x2a3db(%rip),%ymm10 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,98,125,24,21,87,150,2,0 // vbroadcastss 0x29657(%rip),%ymm10 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 196,65,36,84,210 // vandps %ymm10,%ymm11,%ymm10
.byte 196,193,25,114,242,5 // vpslld $0x5,%xmm10,%xmm12
.byte 196,67,125,25,210,1 // vextractf128 $0x1,%ymm10,%xmm10
.byte 196,193,41,114,242,5 // vpslld $0x5,%xmm10,%xmm10
.byte 196,67,29,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm12,%ymm10
- .byte 197,123,18,37,10,181,2,0 // vmovddup 0x2b50a(%rip),%xmm12 # 3a890 <_sk_srcover_bgra_8888_sse2_lowp+0x1408>
+ .byte 197,123,18,37,138,167,2,0 // vmovddup 0x2a78a(%rip),%xmm12 # 39b10 <_sk_srcover_bgra_8888_sse2_lowp+0x140c>
.byte 196,65,49,219,236 // vpand %xmm12,%xmm9,%xmm13
.byte 196,193,17,114,245,4 // vpslld $0x4,%xmm13,%xmm13
.byte 196,65,57,219,228 // vpand %xmm12,%xmm8,%xmm12
.byte 196,193,25,114,244,4 // vpslld $0x4,%xmm12,%xmm12
.byte 196,67,21,24,228,1 // vinsertf128 $0x1,%xmm12,%ymm13,%ymm12
- .byte 196,98,125,24,45,153,163,2,0 // vbroadcastss 0x2a399(%rip),%ymm13 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,125,24,45,21,150,2,0 // vbroadcastss 0x29615(%rip),%ymm13 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,65,36,84,237 // vandps %ymm13,%ymm11,%ymm13
.byte 196,193,9,114,245,2 // vpslld $0x2,%xmm13,%xmm14
.byte 196,67,125,25,237,1 // vextractf128 $0x1,%ymm13,%xmm13
.byte 196,193,17,114,245,2 // vpslld $0x2,%xmm13,%xmm13
.byte 196,67,13,24,237,1 // vinsertf128 $0x1,%xmm13,%ymm14,%ymm13
- .byte 197,123,18,53,200,180,2,0 // vmovddup 0x2b4c8(%rip),%xmm14 # 3a898 <_sk_srcover_bgra_8888_sse2_lowp+0x1410>
+ .byte 197,123,18,53,72,167,2,0 // vmovddup 0x2a748(%rip),%xmm14 # 39b18 <_sk_srcover_bgra_8888_sse2_lowp+0x1414>
.byte 196,65,49,219,254 // vpand %xmm14,%xmm9,%xmm15
.byte 196,65,1,254,255 // vpaddd %xmm15,%xmm15,%xmm15
.byte 196,65,57,219,246 // vpand %xmm14,%xmm8,%xmm14
.byte 196,65,9,254,246 // vpaddd %xmm14,%xmm14,%xmm14
.byte 196,67,5,24,246,1 // vinsertf128 $0x1,%xmm14,%ymm15,%ymm14
.byte 196,65,12,86,228 // vorps %ymm12,%ymm14,%ymm12
- .byte 196,98,125,24,53,72,163,2,0 // vbroadcastss 0x2a348(%rip),%ymm14 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
+ .byte 196,98,125,24,53,196,149,2,0 // vbroadcastss 0x295c4(%rip),%ymm14 # 389bc <_sk_srcover_bgra_8888_sse2_lowp+0x2b8>
.byte 196,65,36,84,222 // vandps %ymm14,%ymm11,%ymm11
.byte 196,193,9,114,211,1 // vpsrld $0x1,%xmm11,%xmm14
.byte 196,67,125,25,219,1 // vextractf128 $0x1,%ymm11,%xmm11
.byte 196,193,33,114,211,1 // vpsrld $0x1,%xmm11,%xmm11
.byte 196,67,13,24,219,1 // vinsertf128 $0x1,%xmm11,%ymm14,%ymm11
.byte 196,65,20,86,219 // vorps %ymm11,%ymm13,%ymm11
- .byte 197,123,18,45,126,180,2,0 // vmovddup 0x2b47e(%rip),%xmm13 # 3a8a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1418>
+ .byte 197,123,18,45,254,166,2,0 // vmovddup 0x2a6fe(%rip),%xmm13 # 39b20 <_sk_srcover_bgra_8888_sse2_lowp+0x141c>
.byte 196,65,49,219,205 // vpand %xmm13,%xmm9,%xmm9
.byte 196,65,57,219,197 // vpand %xmm13,%xmm8,%xmm8
.byte 196,193,49,114,209,2 // vpsrld $0x2,%xmm9,%xmm9
@@ -24368,9 +24368,9 @@ _sk_dither_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 196,65,60,86,195 // vorps %ymm11,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,235,162,2,0 // vbroadcastss 0x2a2eb(%rip),%ymm9 # 39748 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
+ .byte 196,98,125,24,13,103,149,2,0 // vbroadcastss 0x29567(%rip),%ymm9 # 389c4 <_sk_srcover_bgra_8888_sse2_lowp+0x2c0>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
- .byte 196,98,125,24,13,225,162,2,0 // vbroadcastss 0x2a2e1(%rip),%ymm9 # 3974c <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
+ .byte 196,98,125,24,13,93,149,2,0 // vbroadcastss 0x2955d(%rip),%ymm9 # 389c8 <_sk_srcover_bgra_8888_sse2_lowp+0x2c4>
.byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
.byte 196,98,125,24,8 // vbroadcastss (%rax),%ymm9
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
@@ -24404,7 +24404,7 @@ HIDDEN _sk_black_color_avx
FUNCTION(_sk_black_color_avx)
_sk_black_color_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,105,162,2,0 // vbroadcastss 0x2a269(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,229,148,2,0 // vbroadcastss 0x294e5(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
@@ -24415,7 +24415,7 @@ HIDDEN _sk_white_color_avx
FUNCTION(_sk_white_color_avx)
_sk_white_color_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,80,162,2,0 // vbroadcastss 0x2a250(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,204,148,2,0 // vbroadcastss 0x294cc(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 197,252,40,216 // vmovaps %ymm0,%ymm3
@@ -24461,7 +24461,7 @@ HIDDEN _sk_srcatop_avx
FUNCTION(_sk_srcatop_avx)
_sk_srcatop_avx:
.byte 197,252,89,199 // vmulps %ymm7,%ymm0,%ymm0
- .byte 196,98,125,24,5,239,161,2,0 // vbroadcastss 0x2a1ef(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,107,148,2,0 // vbroadcastss 0x2946b(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,204 // vmulps %ymm4,%ymm8,%ymm9
.byte 197,180,88,192 // vaddps %ymm0,%ymm9,%ymm0
@@ -24482,7 +24482,7 @@ HIDDEN _sk_dstatop_avx
FUNCTION(_sk_dstatop_avx)
_sk_dstatop_avx:
.byte 197,100,89,196 // vmulps %ymm4,%ymm3,%ymm8
- .byte 196,98,125,24,13,173,161,2,0 // vbroadcastss 0x2a1ad(%rip),%ymm9 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,13,41,148,2,0 // vbroadcastss 0x29429(%rip),%ymm9 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,52,92,207 // vsubps %ymm7,%ymm9,%ymm9
.byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
.byte 197,188,88,192 // vaddps %ymm0,%ymm8,%ymm0
@@ -24524,7 +24524,7 @@ HIDDEN _sk_srcout_avx
.globl _sk_srcout_avx
FUNCTION(_sk_srcout_avx)
_sk_srcout_avx:
- .byte 196,98,125,24,5,72,161,2,0 // vbroadcastss 0x2a148(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,196,147,2,0 // vbroadcastss 0x293c4(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
@@ -24537,7 +24537,7 @@ HIDDEN _sk_dstout_avx
.globl _sk_dstout_avx
FUNCTION(_sk_dstout_avx)
_sk_dstout_avx:
- .byte 196,226,125,24,5,39,161,2,0 // vbroadcastss 0x2a127(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,163,147,2,0 // vbroadcastss 0x293a3(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,92,219 // vsubps %ymm3,%ymm0,%ymm3
.byte 197,228,89,196 // vmulps %ymm4,%ymm3,%ymm0
.byte 197,228,89,205 // vmulps %ymm5,%ymm3,%ymm1
@@ -24550,7 +24550,7 @@ HIDDEN _sk_srcover_avx
.globl _sk_srcover_avx
FUNCTION(_sk_srcover_avx)
_sk_srcover_avx:
- .byte 196,98,125,24,5,6,161,2,0 // vbroadcastss 0x2a106(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,130,147,2,0 // vbroadcastss 0x29382(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,204 // vmulps %ymm4,%ymm8,%ymm9
.byte 197,180,88,192 // vaddps %ymm0,%ymm9,%ymm0
@@ -24567,7 +24567,7 @@ HIDDEN _sk_dstover_avx
.globl _sk_dstover_avx
FUNCTION(_sk_dstover_avx)
_sk_dstover_avx:
- .byte 196,98,125,24,5,213,160,2,0 // vbroadcastss 0x2a0d5(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,81,147,2,0 // vbroadcastss 0x29351(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
@@ -24595,7 +24595,7 @@ HIDDEN _sk_multiply_avx
.globl _sk_multiply_avx
FUNCTION(_sk_multiply_avx)
_sk_multiply_avx:
- .byte 196,98,125,24,5,144,160,2,0 // vbroadcastss 0x2a090(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,12,147,2,0 // vbroadcastss 0x2930c(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,208 // vmulps %ymm0,%ymm9,%ymm10
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -24626,7 +24626,7 @@ HIDDEN _sk_plus__avx
FUNCTION(_sk_plus__avx)
_sk_plus__avx:
.byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
- .byte 196,98,125,24,5,31,160,2,0 // vbroadcastss 0x2a01f(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,155,146,2,0 // vbroadcastss 0x2929b(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
@@ -24660,7 +24660,7 @@ HIDDEN _sk_xor__avx
.globl _sk_xor__avx
FUNCTION(_sk_xor__avx)
_sk_xor__avx:
- .byte 196,98,125,24,5,190,159,2,0 // vbroadcastss 0x29fbe(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,58,146,2,0 // vbroadcastss 0x2923a(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -24697,7 +24697,7 @@ _sk_darken_avx:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,95,209 // vmaxps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,58,159,2,0 // vbroadcastss 0x29f3a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,182,145,2,0 // vbroadcastss 0x291b6(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,199 // vmulps %ymm7,%ymm8,%ymm8
.byte 197,188,88,219 // vaddps %ymm3,%ymm8,%ymm3
@@ -24723,7 +24723,7 @@ _sk_lighten_avx:
.byte 197,100,89,206 // vmulps %ymm6,%ymm3,%ymm9
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,226,158,2,0 // vbroadcastss 0x29ee2(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,94,145,2,0 // vbroadcastss 0x2915e(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,199 // vmulps %ymm7,%ymm8,%ymm8
.byte 197,188,88,219 // vaddps %ymm3,%ymm8,%ymm3
@@ -24752,7 +24752,7 @@ _sk_difference_avx:
.byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,126,158,2,0 // vbroadcastss 0x29e7e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,250,144,2,0 // vbroadcastss 0x290fa(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,199 // vmulps %ymm7,%ymm8,%ymm8
.byte 197,188,88,219 // vaddps %ymm3,%ymm8,%ymm3
@@ -24775,7 +24775,7 @@ _sk_exclusion_avx:
.byte 197,236,89,214 // vmulps %ymm6,%ymm2,%ymm2
.byte 197,236,88,210 // vaddps %ymm2,%ymm2,%ymm2
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
- .byte 196,98,125,24,5,53,158,2,0 // vbroadcastss 0x29e35(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,177,144,2,0 // vbroadcastss 0x290b1(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
.byte 197,60,89,199 // vmulps %ymm7,%ymm8,%ymm8
.byte 197,188,88,219 // vaddps %ymm3,%ymm8,%ymm3
@@ -24786,7 +24786,7 @@ HIDDEN _sk_colorburn_avx
.globl _sk_colorburn_avx
FUNCTION(_sk_colorburn_avx)
_sk_colorburn_avx:
- .byte 196,98,125,24,5,28,158,2,0 // vbroadcastss 0x29e1c(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,152,144,2,0 // vbroadcastss 0x29098(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,52,89,216 // vmulps %ymm0,%ymm9,%ymm11
.byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
@@ -24848,7 +24848,7 @@ HIDDEN _sk_colordodge_avx
FUNCTION(_sk_colordodge_avx)
_sk_colordodge_avx:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
- .byte 196,98,125,24,13,32,157,2,0 // vbroadcastss 0x29d20(%rip),%ymm9 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,13,156,143,2,0 // vbroadcastss 0x28f9c(%rip),%ymm9 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,52,92,215 // vsubps %ymm7,%ymm9,%ymm10
.byte 197,44,89,216 // vmulps %ymm0,%ymm10,%ymm11
.byte 197,52,92,203 // vsubps %ymm3,%ymm9,%ymm9
@@ -24905,7 +24905,7 @@ HIDDEN _sk_hardlight_avx
.globl _sk_hardlight_avx
FUNCTION(_sk_hardlight_avx)
_sk_hardlight_avx:
- .byte 196,98,125,24,5,43,156,2,0 // vbroadcastss 0x29c2b(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,167,142,2,0 // vbroadcastss 0x28ea7(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,215 // vsubps %ymm7,%ymm8,%ymm10
.byte 197,44,89,200 // vmulps %ymm0,%ymm10,%ymm9
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -24960,7 +24960,7 @@ HIDDEN _sk_overlay_avx
.globl _sk_overlay_avx
FUNCTION(_sk_overlay_avx)
_sk_overlay_avx:
- .byte 196,98,125,24,5,80,155,2,0 // vbroadcastss 0x29b50(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,204,141,2,0 // vbroadcastss 0x28dcc(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,215 // vsubps %ymm7,%ymm8,%ymm10
.byte 197,44,89,200 // vmulps %ymm0,%ymm10,%ymm9
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -25026,10 +25026,10 @@ _sk_softlight_avx:
.byte 196,65,60,88,192 // vaddps %ymm8,%ymm8,%ymm8
.byte 196,65,60,89,216 // vmulps %ymm8,%ymm8,%ymm11
.byte 196,65,60,88,195 // vaddps %ymm11,%ymm8,%ymm8
- .byte 196,98,125,24,29,87,154,2,0 // vbroadcastss 0x29a57(%rip),%ymm11 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,29,211,140,2,0 // vbroadcastss 0x28cd3(%rip),%ymm11 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,65,28,88,235 // vaddps %ymm11,%ymm12,%ymm13
.byte 196,65,20,89,192 // vmulps %ymm8,%ymm13,%ymm8
- .byte 196,98,125,24,45,72,154,2,0 // vbroadcastss 0x29a48(%rip),%ymm13 # 39754 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
+ .byte 196,98,125,24,45,196,140,2,0 // vbroadcastss 0x28cc4(%rip),%ymm13 # 389d0 <_sk_srcover_bgra_8888_sse2_lowp+0x2cc>
.byte 196,65,28,89,245 // vmulps %ymm13,%ymm12,%ymm14
.byte 196,65,12,88,192 // vaddps %ymm8,%ymm14,%ymm8
.byte 196,65,124,82,244 // vrsqrtps %ymm12,%ymm14
@@ -25040,7 +25040,7 @@ _sk_softlight_avx:
.byte 197,4,194,255,2 // vcmpleps %ymm7,%ymm15,%ymm15
.byte 196,67,13,74,240,240 // vblendvps %ymm15,%ymm8,%ymm14,%ymm14
.byte 197,116,88,249 // vaddps %ymm1,%ymm1,%ymm15
- .byte 196,98,125,24,5,242,153,2,0 // vbroadcastss 0x299f2(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,110,140,2,0 // vbroadcastss 0x28c6e(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,65,60,92,228 // vsubps %ymm12,%ymm8,%ymm12
.byte 197,132,92,195 // vsubps %ymm3,%ymm15,%ymm0
.byte 196,65,124,89,228 // vmulps %ymm12,%ymm0,%ymm12
@@ -25167,12 +25167,12 @@ _sk_hue_avx:
.byte 196,65,28,89,219 // vmulps %ymm11,%ymm12,%ymm11
.byte 196,65,36,94,222 // vdivps %ymm14,%ymm11,%ymm11
.byte 196,67,37,74,224,240 // vblendvps %ymm15,%ymm8,%ymm11,%ymm12
- .byte 196,98,125,24,53,213,151,2,0 // vbroadcastss 0x297d5(%rip),%ymm14 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,81,138,2,0 // vbroadcastss 0x28a51(%rip),%ymm14 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,92,89,222 // vmulps %ymm14,%ymm4,%ymm11
- .byte 196,98,125,24,61,203,151,2,0 // vbroadcastss 0x297cb(%rip),%ymm15 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,61,71,138,2,0 // vbroadcastss 0x28a47(%rip),%ymm15 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,239 // vmulps %ymm15,%ymm5,%ymm13
.byte 196,65,36,88,221 // vaddps %ymm13,%ymm11,%ymm11
- .byte 196,226,125,24,5,188,151,2,0 // vbroadcastss 0x297bc(%rip),%ymm0 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,226,125,24,5,56,138,2,0 // vbroadcastss 0x28a38(%rip),%ymm0 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 197,76,89,232 // vmulps %ymm0,%ymm6,%ymm13
.byte 196,65,36,88,221 // vaddps %ymm13,%ymm11,%ymm11
.byte 196,65,52,89,238 // vmulps %ymm14,%ymm9,%ymm13
@@ -25233,7 +25233,7 @@ _sk_hue_avx:
.byte 196,65,36,95,208 // vmaxps %ymm8,%ymm11,%ymm10
.byte 196,195,109,74,209,240 // vblendvps %ymm15,%ymm9,%ymm2,%ymm2
.byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,105,150,2,0 // vbroadcastss 0x29669(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,229,136,2,0 // vbroadcastss 0x288e5(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,180,89,201 // vmulps %ymm1,%ymm9,%ymm1
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -25290,12 +25290,12 @@ _sk_saturation_avx:
.byte 196,65,28,89,219 // vmulps %ymm11,%ymm12,%ymm11
.byte 196,65,36,94,222 // vdivps %ymm14,%ymm11,%ymm11
.byte 196,67,37,74,224,240 // vblendvps %ymm15,%ymm8,%ymm11,%ymm12
- .byte 196,98,125,24,53,147,149,2,0 // vbroadcastss 0x29593(%rip),%ymm14 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,53,15,136,2,0 // vbroadcastss 0x2880f(%rip),%ymm14 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,92,89,222 // vmulps %ymm14,%ymm4,%ymm11
- .byte 196,98,125,24,61,137,149,2,0 // vbroadcastss 0x29589(%rip),%ymm15 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,61,5,136,2,0 // vbroadcastss 0x28805(%rip),%ymm15 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,239 // vmulps %ymm15,%ymm5,%ymm13
.byte 196,65,36,88,221 // vaddps %ymm13,%ymm11,%ymm11
- .byte 196,226,125,24,5,122,149,2,0 // vbroadcastss 0x2957a(%rip),%ymm0 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,226,125,24,5,246,135,2,0 // vbroadcastss 0x287f6(%rip),%ymm0 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 197,76,89,232 // vmulps %ymm0,%ymm6,%ymm13
.byte 196,65,36,88,221 // vaddps %ymm13,%ymm11,%ymm11
.byte 196,65,52,89,238 // vmulps %ymm14,%ymm9,%ymm13
@@ -25356,7 +25356,7 @@ _sk_saturation_avx:
.byte 196,65,36,95,208 // vmaxps %ymm8,%ymm11,%ymm10
.byte 196,195,109,74,209,240 // vblendvps %ymm15,%ymm9,%ymm2,%ymm2
.byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,39,148,2,0 // vbroadcastss 0x29427(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,163,134,2,0 // vbroadcastss 0x286a3(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,207 // vsubps %ymm7,%ymm8,%ymm9
.byte 197,180,89,201 // vmulps %ymm1,%ymm9,%ymm1
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
@@ -25385,12 +25385,12 @@ _sk_color_avx:
.byte 197,252,17,68,36,168 // vmovups %ymm0,-0x58(%rsp)
.byte 197,124,89,199 // vmulps %ymm7,%ymm0,%ymm8
.byte 197,116,89,207 // vmulps %ymm7,%ymm1,%ymm9
- .byte 196,98,125,24,45,217,147,2,0 // vbroadcastss 0x293d9(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,45,85,134,2,0 // vbroadcastss 0x28655(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,92,89,213 // vmulps %ymm13,%ymm4,%ymm10
- .byte 196,98,125,24,53,207,147,2,0 // vbroadcastss 0x293cf(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,53,75,134,2,0 // vbroadcastss 0x2864b(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,84,89,222 // vmulps %ymm14,%ymm5,%ymm11
.byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,24,61,192,147,2,0 // vbroadcastss 0x293c0(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,60,134,2,0 // vbroadcastss 0x2863c(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,65,76,89,223 // vmulps %ymm15,%ymm6,%ymm11
.byte 196,193,44,88,195 // vaddps %ymm11,%ymm10,%ymm0
.byte 196,65,60,89,221 // vmulps %ymm13,%ymm8,%ymm11
@@ -25453,7 +25453,7 @@ _sk_color_avx:
.byte 196,65,44,95,207 // vmaxps %ymm15,%ymm10,%ymm9
.byte 196,195,37,74,192,0 // vblendvps %ymm0,%ymm8,%ymm11,%ymm0
.byte 196,65,124,95,199 // vmaxps %ymm15,%ymm0,%ymm8
- .byte 196,226,125,24,5,91,146,2,0 // vbroadcastss 0x2925b(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,215,132,2,0 // vbroadcastss 0x284d7(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,124,92,215 // vsubps %ymm7,%ymm0,%ymm10
.byte 197,172,89,84,36,168 // vmulps -0x58(%rsp),%ymm10,%ymm2
.byte 197,124,92,219 // vsubps %ymm3,%ymm0,%ymm11
@@ -25483,12 +25483,12 @@ _sk_luminosity_avx:
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 197,100,89,196 // vmulps %ymm4,%ymm3,%ymm8
.byte 197,100,89,205 // vmulps %ymm5,%ymm3,%ymm9
- .byte 196,98,125,24,45,9,146,2,0 // vbroadcastss 0x29209(%rip),%ymm13 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,45,133,132,2,0 // vbroadcastss 0x28485(%rip),%ymm13 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,108,89,213 // vmulps %ymm13,%ymm2,%ymm10
- .byte 196,98,125,24,53,255,145,2,0 // vbroadcastss 0x291ff(%rip),%ymm14 # 3975c <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
+ .byte 196,98,125,24,53,123,132,2,0 // vbroadcastss 0x2847b(%rip),%ymm14 # 389d8 <_sk_srcover_bgra_8888_sse2_lowp+0x2d4>
.byte 196,65,116,89,222 // vmulps %ymm14,%ymm1,%ymm11
.byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,24,61,240,145,2,0 // vbroadcastss 0x291f0(%rip),%ymm15 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
+ .byte 196,98,125,24,61,108,132,2,0 // vbroadcastss 0x2846c(%rip),%ymm15 # 389dc <_sk_srcover_bgra_8888_sse2_lowp+0x2d8>
.byte 196,65,28,89,223 // vmulps %ymm15,%ymm12,%ymm11
.byte 196,193,44,88,195 // vaddps %ymm11,%ymm10,%ymm0
.byte 196,65,60,89,221 // vmulps %ymm13,%ymm8,%ymm11
@@ -25551,7 +25551,7 @@ _sk_luminosity_avx:
.byte 196,65,44,95,207 // vmaxps %ymm15,%ymm10,%ymm9
.byte 196,195,37,74,192,0 // vblendvps %ymm0,%ymm8,%ymm11,%ymm0
.byte 196,65,124,95,199 // vmaxps %ymm15,%ymm0,%ymm8
- .byte 196,226,125,24,5,139,144,2,0 // vbroadcastss 0x2908b(%rip),%ymm0 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,5,7,131,2,0 // vbroadcastss 0x28307(%rip),%ymm0 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,124,92,215 // vsubps %ymm7,%ymm0,%ymm10
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 197,124,92,219 // vsubps %ymm3,%ymm0,%ymm11
@@ -25585,7 +25585,7 @@ _sk_srcover_rgba_8888_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,42,1,0,0 // jne 10842 <_sk_srcover_rgba_8888_avx+0x14a>
.byte 196,1,125,16,4,152 // vmovupd (%r8,%r11,4),%ymm8
- .byte 197,125,40,13,90,149,2,0 // vmovapd 0x2955a(%rip),%ymm9 # 39c80 <_sk_srcover_bgra_8888_sse2_lowp+0x7f8>
+ .byte 197,125,40,13,218,135,2,0 // vmovapd 0x287da(%rip),%ymm9 # 38f00 <_sk_srcover_bgra_8888_sse2_lowp+0x7fc>
.byte 196,193,61,84,225 // vandpd %ymm9,%ymm8,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
.byte 196,193,81,114,208,8 // vpsrld $0x8,%xmm8,%xmm5
@@ -25603,9 +25603,9 @@ _sk_srcover_rgba_8888_avx:
.byte 197,193,114,215,24 // vpsrld $0x18,%xmm7,%xmm7
.byte 196,227,61,24,255,1 // vinsertf128 $0x1,%xmm7,%ymm8,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,177,143,2,0 // vbroadcastss 0x28fb1(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,45,130,2,0 // vbroadcastss 0x2822d(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,208,143,2,0 // vbroadcastss 0x28fd0(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,76,130,2,0 // vbroadcastss 0x2824c(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,60,89,212 // vmulps %ymm4,%ymm8,%ymm10
.byte 196,193,124,88,194 // vaddps %ymm10,%ymm0,%ymm0
@@ -25747,7 +25747,7 @@ _sk_srcover_bgra_8888_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,42,1,0,0 // jne 10aee <_sk_srcover_bgra_8888_avx+0x14a>
.byte 196,1,125,16,4,152 // vmovupd (%r8,%r11,4),%ymm8
- .byte 197,125,40,13,206,146,2,0 // vmovapd 0x292ce(%rip),%ymm9 # 39ca0 <_sk_srcover_bgra_8888_sse2_lowp+0x818>
+ .byte 197,125,40,13,78,133,2,0 // vmovapd 0x2854e(%rip),%ymm9 # 38f20 <_sk_srcover_bgra_8888_sse2_lowp+0x81c>
.byte 196,193,61,84,233 // vandpd %ymm9,%ymm8,%ymm5
.byte 197,252,91,245 // vcvtdq2ps %ymm5,%ymm6
.byte 196,193,81,114,208,8 // vpsrld $0x8,%xmm8,%xmm5
@@ -25765,9 +25765,9 @@ _sk_srcover_bgra_8888_avx:
.byte 197,193,114,215,24 // vpsrld $0x18,%xmm7,%xmm7
.byte 196,227,61,24,255,1 // vinsertf128 $0x1,%xmm7,%ymm8,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,5,141,2,0 // vbroadcastss 0x28d05(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,129,127,2,0 // vbroadcastss 0x27f81(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,36,141,2,0 // vbroadcastss 0x28d24(%rip),%ymm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,13,160,127,2,0 // vbroadcastss 0x27fa0(%rip),%ymm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,60,89,212 // vmulps %ymm4,%ymm8,%ymm10
.byte 196,193,124,88,194 // vaddps %ymm10,%ymm0,%ymm0
@@ -25911,7 +25911,7 @@ HIDDEN _sk_clamp_1_avx
.globl _sk_clamp_1_avx
FUNCTION(_sk_clamp_1_avx)
_sk_clamp_1_avx:
- .byte 196,98,125,24,5,194,138,2,0 // vbroadcastss 0x28ac2(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,62,125,2,0 // vbroadcastss 0x27d3e(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
@@ -25923,7 +25923,7 @@ HIDDEN _sk_clamp_a_avx
.globl _sk_clamp_a_avx
FUNCTION(_sk_clamp_a_avx)
_sk_clamp_a_avx:
- .byte 196,98,125,24,5,161,138,2,0 // vbroadcastss 0x28aa1(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,29,125,2,0 // vbroadcastss 0x27d1d(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 197,252,93,195 // vminps %ymm3,%ymm0,%ymm0
.byte 197,244,93,203 // vminps %ymm3,%ymm1,%ymm1
@@ -25935,7 +25935,7 @@ HIDDEN _sk_clamp_a_dst_avx
.globl _sk_clamp_a_dst_avx
FUNCTION(_sk_clamp_a_dst_avx)
_sk_clamp_a_dst_avx:
- .byte 196,98,125,24,5,131,138,2,0 // vbroadcastss 0x28a83(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,255,124,2,0 // vbroadcastss 0x27cff(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,68,93,248 // vminps %ymm8,%ymm7,%ymm7
.byte 197,220,93,231 // vminps %ymm7,%ymm4,%ymm4
.byte 197,212,93,239 // vminps %ymm7,%ymm5,%ymm5
@@ -25968,7 +25968,7 @@ HIDDEN _sk_invert_avx
.globl _sk_invert_avx
FUNCTION(_sk_invert_avx)
_sk_invert_avx:
- .byte 196,98,125,24,5,62,138,2,0 // vbroadcastss 0x28a3e(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,186,124,2,0 // vbroadcastss 0x27cba(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,188,92,192 // vsubps %ymm0,%ymm8,%ymm0
.byte 197,188,92,201 // vsubps %ymm1,%ymm8,%ymm1
.byte 197,188,92,210 // vsubps %ymm2,%ymm8,%ymm2
@@ -26022,9 +26022,9 @@ HIDDEN _sk_unpremul_avx
.globl _sk_unpremul_avx
FUNCTION(_sk_unpremul_avx)
_sk_unpremul_avx:
- .byte 196,98,125,24,5,217,137,2,0 // vbroadcastss 0x289d9(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,85,124,2,0 // vbroadcastss 0x27c55(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,94,195 // vdivps %ymm3,%ymm8,%ymm8
- .byte 196,98,125,24,13,252,137,2,0 // vbroadcastss 0x289fc(%rip),%ymm9 # 39768 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
+ .byte 196,98,125,24,13,120,124,2,0 // vbroadcastss 0x27c78(%rip),%ymm9 # 389e4 <_sk_srcover_bgra_8888_sse2_lowp+0x2e0>
.byte 196,65,60,194,201,1 // vcmpltps %ymm9,%ymm8,%ymm9
.byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
.byte 196,67,45,74,192,144 // vblendvps %ymm9,%ymm8,%ymm10,%ymm8
@@ -26038,17 +26038,17 @@ HIDDEN _sk_from_srgb_avx
.globl _sk_from_srgb_avx
FUNCTION(_sk_from_srgb_avx)
_sk_from_srgb_avx:
- .byte 196,98,125,24,5,214,137,2,0 // vbroadcastss 0x289d6(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,82,124,2,0 // vbroadcastss 0x27c52(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 197,124,89,208 // vmulps %ymm0,%ymm0,%ymm10
- .byte 196,98,125,24,29,176,137,2,0 // vbroadcastss 0x289b0(%rip),%ymm11 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,44,124,2,0 // vbroadcastss 0x27c2c(%rip),%ymm11 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,124,89,227 // vmulps %ymm11,%ymm0,%ymm12
- .byte 196,98,125,24,45,186,137,2,0 // vbroadcastss 0x289ba(%rip),%ymm13 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,45,54,124,2,0 // vbroadcastss 0x27c36(%rip),%ymm13 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
.byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
.byte 196,65,44,89,212 // vmulps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,171,137,2,0 // vbroadcastss 0x289ab(%rip),%ymm12 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,37,39,124,2,0 // vbroadcastss 0x27c27(%rip),%ymm12 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,53,161,137,2,0 // vbroadcastss 0x289a1(%rip),%ymm14 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,53,29,124,2,0 // vbroadcastss 0x27c1d(%rip),%ymm14 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 196,193,124,194,198,1 // vcmpltps %ymm14,%ymm0,%ymm0
.byte 196,195,45,74,193,0 // vblendvps %ymm0,%ymm9,%ymm10,%ymm0
.byte 196,65,116,89,200 // vmulps %ymm8,%ymm1,%ymm9
@@ -26074,17 +26074,17 @@ HIDDEN _sk_from_srgb_dst_avx
.globl _sk_from_srgb_dst_avx
FUNCTION(_sk_from_srgb_dst_avx)
_sk_from_srgb_dst_avx:
- .byte 196,98,125,24,5,42,137,2,0 // vbroadcastss 0x2892a(%rip),%ymm8 # 3976c <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
+ .byte 196,98,125,24,5,166,123,2,0 // vbroadcastss 0x27ba6(%rip),%ymm8 # 389e8 <_sk_srcover_bgra_8888_sse2_lowp+0x2e4>
.byte 196,65,92,89,200 // vmulps %ymm8,%ymm4,%ymm9
.byte 197,92,89,212 // vmulps %ymm4,%ymm4,%ymm10
- .byte 196,98,125,24,29,4,137,2,0 // vbroadcastss 0x28904(%rip),%ymm11 # 39758 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
+ .byte 196,98,125,24,29,128,123,2,0 // vbroadcastss 0x27b80(%rip),%ymm11 # 389d4 <_sk_srcover_bgra_8888_sse2_lowp+0x2d0>
.byte 196,65,92,89,227 // vmulps %ymm11,%ymm4,%ymm12
- .byte 196,98,125,24,45,14,137,2,0 // vbroadcastss 0x2890e(%rip),%ymm13 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
+ .byte 196,98,125,24,45,138,123,2,0 // vbroadcastss 0x27b8a(%rip),%ymm13 # 389ec <_sk_srcover_bgra_8888_sse2_lowp+0x2e8>
.byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
.byte 196,65,44,89,212 // vmulps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,37,255,136,2,0 // vbroadcastss 0x288ff(%rip),%ymm12 # 39774 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
+ .byte 196,98,125,24,37,123,123,2,0 // vbroadcastss 0x27b7b(%rip),%ymm12 # 389f0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ec>
.byte 196,65,44,88,212 // vaddps %ymm12,%ymm10,%ymm10
- .byte 196,98,125,24,53,245,136,2,0 // vbroadcastss 0x288f5(%rip),%ymm14 # 39778 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
+ .byte 196,98,125,24,53,113,123,2,0 // vbroadcastss 0x27b71(%rip),%ymm14 # 389f4 <_sk_srcover_bgra_8888_sse2_lowp+0x2f0>
.byte 196,193,92,194,230,1 // vcmpltps %ymm14,%ymm4,%ymm4
.byte 196,195,45,74,225,64 // vblendvps %ymm4,%ymm9,%ymm10,%ymm4
.byte 196,65,84,89,200 // vmulps %ymm8,%ymm5,%ymm9
@@ -26111,20 +26111,20 @@ HIDDEN _sk_to_srgb_avx
FUNCTION(_sk_to_srgb_avx)
_sk_to_srgb_avx:
.byte 197,124,82,200 // vrsqrtps %ymm0,%ymm9
- .byte 196,98,125,24,5,138,136,2,0 // vbroadcastss 0x2888a(%rip),%ymm8 # 3977c <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
+ .byte 196,98,125,24,5,6,123,2,0 // vbroadcastss 0x27b06(%rip),%ymm8 # 389f8 <_sk_srcover_bgra_8888_sse2_lowp+0x2f4>
.byte 196,65,124,89,208 // vmulps %ymm8,%ymm0,%ymm10
- .byte 196,98,125,24,29,132,136,2,0 // vbroadcastss 0x28884(%rip),%ymm11 # 39784 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
+ .byte 196,98,125,24,29,0,123,2,0 // vbroadcastss 0x27b00(%rip),%ymm11 # 38a00 <_sk_srcover_bgra_8888_sse2_lowp+0x2fc>
.byte 196,65,52,89,227 // vmulps %ymm11,%ymm9,%ymm12
- .byte 196,98,125,24,45,114,136,2,0 // vbroadcastss 0x28872(%rip),%ymm13 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
+ .byte 196,98,125,24,45,238,122,2,0 // vbroadcastss 0x27aee(%rip),%ymm13 # 389fc <_sk_srcover_bgra_8888_sse2_lowp+0x2f8>
.byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
.byte 196,65,52,89,228 // vmulps %ymm12,%ymm9,%ymm12
- .byte 196,98,125,24,53,151,137,2,0 // vbroadcastss 0x28997(%rip),%ymm14 # 398b8 <_sk_srcover_bgra_8888_sse2_lowp+0x430>
+ .byte 196,98,125,24,53,19,124,2,0 // vbroadcastss 0x27c13(%rip),%ymm14 # 38b34 <_sk_srcover_bgra_8888_sse2_lowp+0x430>
.byte 196,65,28,88,230 // vaddps %ymm14,%ymm12,%ymm12
- .byte 196,98,125,24,61,141,137,2,0 // vbroadcastss 0x2898d(%rip),%ymm15 # 398bc <_sk_srcover_bgra_8888_sse2_lowp+0x434>
+ .byte 196,98,125,24,61,9,124,2,0 // vbroadcastss 0x27c09(%rip),%ymm15 # 38b38 <_sk_srcover_bgra_8888_sse2_lowp+0x434>
.byte 196,65,52,88,207 // vaddps %ymm15,%ymm9,%ymm9
.byte 196,65,124,83,201 // vrcpps %ymm9,%ymm9
.byte 196,65,52,89,204 // vmulps %ymm12,%ymm9,%ymm9
- .byte 196,98,125,24,37,73,136,2,0 // vbroadcastss 0x28849(%rip),%ymm12 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x308>
+ .byte 196,98,125,24,37,197,122,2,0 // vbroadcastss 0x27ac5(%rip),%ymm12 # 38a0c <_sk_srcover_bgra_8888_sse2_lowp+0x308>
.byte 196,193,124,194,196,1 // vcmpltps %ymm12,%ymm0,%ymm0
.byte 196,195,53,74,194,0 // vblendvps %ymm0,%ymm10,%ymm9,%ymm0
.byte 197,124,82,201 // vrsqrtps %ymm1,%ymm9
@@ -26161,7 +26161,7 @@ _sk_rgb_to_hsl_avx:
.byte 197,116,93,202 // vminps %ymm2,%ymm1,%ymm9
.byte 196,65,124,93,201 // vminps %ymm9,%ymm0,%ymm9
.byte 196,65,60,92,209 // vsubps %ymm9,%ymm8,%ymm10
- .byte 196,98,125,24,29,81,135,2,0 // vbroadcastss 0x28751(%rip),%ymm11 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,29,205,121,2,0 // vbroadcastss 0x279cd(%rip),%ymm11 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,65,36,94,218 // vdivps %ymm10,%ymm11,%ymm11
.byte 197,116,92,226 // vsubps %ymm2,%ymm1,%ymm12
.byte 196,65,28,89,227 // vmulps %ymm11,%ymm12,%ymm12
@@ -26171,19 +26171,19 @@ _sk_rgb_to_hsl_avx:
.byte 196,193,108,89,211 // vmulps %ymm11,%ymm2,%ymm2
.byte 197,252,92,201 // vsubps %ymm1,%ymm0,%ymm1
.byte 196,193,116,89,203 // vmulps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,29,130,135,2,0 // vbroadcastss 0x28782(%rip),%ymm11 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 196,98,125,24,29,254,121,2,0 // vbroadcastss 0x279fe(%rip),%ymm11 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,29,112,135,2,0 // vbroadcastss 0x28770(%rip),%ymm11 # 39798 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
+ .byte 196,98,125,24,29,236,121,2,0 // vbroadcastss 0x279ec(%rip),%ymm11 # 38a14 <_sk_srcover_bgra_8888_sse2_lowp+0x310>
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
.byte 196,227,117,74,202,224 // vblendvps %ymm14,%ymm2,%ymm1,%ymm1
- .byte 196,226,125,24,21,88,135,2,0 // vbroadcastss 0x28758(%rip),%ymm2 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,226,125,24,21,212,121,2,0 // vbroadcastss 0x279d4(%rip),%ymm2 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 196,65,12,87,246 // vxorps %ymm14,%ymm14,%ymm14
.byte 196,227,13,74,210,208 // vblendvps %ymm13,%ymm2,%ymm14,%ymm2
.byte 197,188,194,192,0 // vcmpeqps %ymm0,%ymm8,%ymm0
.byte 196,193,108,88,212 // vaddps %ymm12,%ymm2,%ymm2
.byte 196,227,117,74,194,0 // vblendvps %ymm0,%ymm2,%ymm1,%ymm0
.byte 196,193,60,88,201 // vaddps %ymm9,%ymm8,%ymm1
- .byte 196,98,125,24,37,207,134,2,0 // vbroadcastss 0x286cf(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,75,121,2,0 // vbroadcastss 0x2794b(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,116,89,212 // vmulps %ymm12,%ymm1,%ymm2
.byte 197,28,194,226,1 // vcmpltps %ymm2,%ymm12,%ymm12
.byte 196,65,36,92,216 // vsubps %ymm8,%ymm11,%ymm11
@@ -26193,7 +26193,7 @@ _sk_rgb_to_hsl_avx:
.byte 197,172,94,201 // vdivps %ymm1,%ymm10,%ymm1
.byte 196,195,125,74,198,128 // vblendvps %ymm8,%ymm14,%ymm0,%ymm0
.byte 196,195,117,74,206,128 // vblendvps %ymm8,%ymm14,%ymm1,%ymm1
- .byte 196,98,125,24,5,2,135,2,0 // vbroadcastss 0x28702(%rip),%ymm8 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 196,98,125,24,5,126,121,2,0 // vbroadcastss 0x2797e(%rip),%ymm8 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -26210,7 +26210,7 @@ _sk_hsl_to_rgb_avx:
.byte 197,252,17,92,36,128 // vmovups %ymm3,-0x80(%rsp)
.byte 197,252,40,225 // vmovaps %ymm1,%ymm4
.byte 197,252,40,216 // vmovaps %ymm0,%ymm3
- .byte 196,98,125,24,5,91,134,2,0 // vbroadcastss 0x2865b(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,215,120,2,0 // vbroadcastss 0x278d7(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,60,194,202,2 // vcmpleps %ymm2,%ymm8,%ymm9
.byte 197,92,89,210 // vmulps %ymm2,%ymm4,%ymm10
.byte 196,65,92,92,218 // vsubps %ymm10,%ymm4,%ymm11
@@ -26218,23 +26218,23 @@ _sk_hsl_to_rgb_avx:
.byte 197,52,88,210 // vaddps %ymm2,%ymm9,%ymm10
.byte 197,108,88,202 // vaddps %ymm2,%ymm2,%ymm9
.byte 196,65,52,92,202 // vsubps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,24,29,161,134,2,0 // vbroadcastss 0x286a1(%rip),%ymm11 # 397a4 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
+ .byte 196,98,125,24,29,29,121,2,0 // vbroadcastss 0x2791d(%rip),%ymm11 # 38a20 <_sk_srcover_bgra_8888_sse2_lowp+0x31c>
.byte 196,65,100,88,219 // vaddps %ymm11,%ymm3,%ymm11
.byte 196,67,125,8,227,1 // vroundps $0x1,%ymm11,%ymm12
.byte 196,65,36,92,252 // vsubps %ymm12,%ymm11,%ymm15
.byte 196,65,44,92,217 // vsubps %ymm9,%ymm10,%ymm11
- .byte 196,98,125,24,37,115,134,2,0 // vbroadcastss 0x28673(%rip),%ymm12 # 39794 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
+ .byte 196,98,125,24,37,239,120,2,0 // vbroadcastss 0x278ef(%rip),%ymm12 # 38a10 <_sk_srcover_bgra_8888_sse2_lowp+0x30c>
.byte 196,193,4,89,196 // vmulps %ymm12,%ymm15,%ymm0
- .byte 196,98,125,24,45,109,134,2,0 // vbroadcastss 0x2866d(%rip),%ymm13 # 3979c <_sk_srcover_bgra_8888_sse2_lowp+0x314>
+ .byte 196,98,125,24,45,233,120,2,0 // vbroadcastss 0x278e9(%rip),%ymm13 # 38a18 <_sk_srcover_bgra_8888_sse2_lowp+0x314>
.byte 197,20,92,240 // vsubps %ymm0,%ymm13,%ymm14
.byte 196,65,36,89,246 // vmulps %ymm14,%ymm11,%ymm14
.byte 196,65,52,88,246 // vaddps %ymm14,%ymm9,%ymm14
- .byte 196,226,125,24,13,98,134,2,0 // vbroadcastss 0x28662(%rip),%ymm1 # 397a8 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
+ .byte 196,226,125,24,13,222,120,2,0 // vbroadcastss 0x278de(%rip),%ymm1 # 38a24 <_sk_srcover_bgra_8888_sse2_lowp+0x320>
.byte 196,193,116,194,255,2 // vcmpleps %ymm15,%ymm1,%ymm7
.byte 196,195,13,74,249,112 // vblendvps %ymm7,%ymm9,%ymm14,%ymm7
.byte 196,65,60,194,247,2 // vcmpleps %ymm15,%ymm8,%ymm14
.byte 196,227,45,74,255,224 // vblendvps %ymm14,%ymm7,%ymm10,%ymm7
- .byte 196,98,125,24,53,57,134,2,0 // vbroadcastss 0x28639(%rip),%ymm14 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x318>
+ .byte 196,98,125,24,53,181,120,2,0 // vbroadcastss 0x278b5(%rip),%ymm14 # 38a1c <_sk_srcover_bgra_8888_sse2_lowp+0x318>
.byte 196,65,12,194,255,2 // vcmpleps %ymm15,%ymm14,%ymm15
.byte 196,193,124,89,195 // vmulps %ymm11,%ymm0,%ymm0
.byte 197,180,88,192 // vaddps %ymm0,%ymm9,%ymm0
@@ -26253,7 +26253,7 @@ _sk_hsl_to_rgb_avx:
.byte 197,164,89,247 // vmulps %ymm7,%ymm11,%ymm6
.byte 197,180,88,246 // vaddps %ymm6,%ymm9,%ymm6
.byte 196,227,77,74,237,0 // vblendvps %ymm0,%ymm5,%ymm6,%ymm5
- .byte 196,226,125,24,5,227,133,2,0 // vbroadcastss 0x285e3(%rip),%ymm0 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,226,125,24,5,95,120,2,0 // vbroadcastss 0x2785f(%rip),%ymm0 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 197,228,88,192 // vaddps %ymm0,%ymm3,%ymm0
.byte 196,227,125,8,216,1 // vroundps $0x1,%ymm0,%ymm3
.byte 197,252,92,195 // vsubps %ymm3,%ymm0,%ymm0
@@ -26309,13 +26309,13 @@ _sk_scale_u8_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,75 // jne 112cf <_sk_scale_u8_avx+0x63>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,126,141,2,0 // vpand 0x28d7e(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,254,127,2,0 // vpand 0x27ffe(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,121,51,200 // vpmovzxwd %xmm8,%xmm9
.byte 196,65,121,112,192,78 // vpshufd $0x4e,%xmm8,%xmm8
.byte 196,66,121,51,192 // vpmovzxwd %xmm8,%xmm8
.byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,250,132,2,0 // vbroadcastss 0x284fa(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,118,119,2,0 // vbroadcastss 0x27776(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
@@ -26395,20 +26395,20 @@ _sk_scale_565_avx:
.byte 196,65,121,112,192,78 // vpshufd $0x4e,%xmm8,%xmm8
.byte 196,66,121,51,192 // vpmovzxwd %xmm8,%xmm8
.byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- .byte 196,98,125,24,13,236,131,2,0 // vbroadcastss 0x283ec(%rip),%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,98,125,24,13,104,118,2,0 // vbroadcastss 0x27668(%rip),%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,60,84,201 // vandps %ymm9,%ymm8,%ymm9
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 196,98,125,24,21,221,131,2,0 // vbroadcastss 0x283dd(%rip),%ymm10 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,98,125,24,21,89,118,2,0 // vbroadcastss 0x27659(%rip),%ymm10 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,24,21,211,131,2,0 // vbroadcastss 0x283d3(%rip),%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,98,125,24,21,79,118,2,0 // vbroadcastss 0x2764f(%rip),%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,60,84,210 // vandps %ymm10,%ymm8,%ymm10
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 196,98,125,24,29,196,131,2,0 // vbroadcastss 0x283c4(%rip),%ymm11 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,98,125,24,29,64,118,2,0 // vbroadcastss 0x27640(%rip),%ymm11 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,24,29,186,131,2,0 // vbroadcastss 0x283ba(%rip),%ymm11 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,98,125,24,29,54,118,2,0 // vbroadcastss 0x27636(%rip),%ymm11 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,60,84,195 // vandps %ymm11,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,29,171,131,2,0 // vbroadcastss 0x283ab(%rip),%ymm11 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,98,125,24,29,39,118,2,0 // vbroadcastss 0x27627(%rip),%ymm11 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 197,100,194,223,1 // vcmpltps %ymm7,%ymm3,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -26501,13 +26501,13 @@ _sk_lerp_u8_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,111 // jne 115be <_sk_lerp_u8_avx+0x87>
.byte 196,2,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm8
- .byte 197,57,219,5,179,138,2,0 // vpand 0x28ab3(%rip),%xmm8,%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,57,219,5,51,125,2,0 // vpand 0x27d33(%rip),%xmm8,%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,66,121,51,200 // vpmovzxwd %xmm8,%xmm9
.byte 196,65,121,112,192,78 // vpshufd $0x4e,%xmm8,%xmm8
.byte 196,66,121,51,192 // vpmovzxwd %xmm8,%xmm8
.byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,13,47,130,2,0 // vbroadcastss 0x2822f(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,171,116,2,0 // vbroadcastss 0x274ab(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
@@ -26563,7 +26563,7 @@ _sk_lerp_u8_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,221 // callq ffffffffde01166c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd81e4>
+ .byte 232,255,255,255,221 // callq ffffffffde01166c <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd8f68>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,210 // callq *%rdx
@@ -26592,20 +26592,20 @@ _sk_lerp_565_avx:
.byte 196,65,121,112,192,78 // vpshufd $0x4e,%xmm8,%xmm8
.byte 196,66,121,51,192 // vpmovzxwd %xmm8,%xmm8
.byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- .byte 196,98,125,24,13,248,128,2,0 // vbroadcastss 0x280f8(%rip),%ymm9 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,98,125,24,13,116,115,2,0 // vbroadcastss 0x27374(%rip),%ymm9 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 196,65,60,84,201 // vandps %ymm9,%ymm8,%ymm9
.byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
- .byte 196,98,125,24,21,233,128,2,0 // vbroadcastss 0x280e9(%rip),%ymm10 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,98,125,24,21,101,115,2,0 // vbroadcastss 0x27365(%rip),%ymm10 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,24,21,223,128,2,0 // vbroadcastss 0x280df(%rip),%ymm10 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,98,125,24,21,91,115,2,0 // vbroadcastss 0x2735b(%rip),%ymm10 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 196,65,60,84,210 // vandps %ymm10,%ymm8,%ymm10
.byte 196,65,124,91,210 // vcvtdq2ps %ymm10,%ymm10
- .byte 196,98,125,24,29,208,128,2,0 // vbroadcastss 0x280d0(%rip),%ymm11 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,98,125,24,29,76,115,2,0 // vbroadcastss 0x2734c(%rip),%ymm11 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
- .byte 196,98,125,24,29,198,128,2,0 // vbroadcastss 0x280c6(%rip),%ymm11 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,98,125,24,29,66,115,2,0 // vbroadcastss 0x27342(%rip),%ymm11 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 196,65,60,84,195 // vandps %ymm11,%ymm8,%ymm8
.byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
- .byte 196,98,125,24,29,183,128,2,0 // vbroadcastss 0x280b7(%rip),%ymm11 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,98,125,24,29,51,115,2,0 // vbroadcastss 0x27333(%rip),%ymm11 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 197,100,194,223,1 // vcmpltps %ymm7,%ymm3,%ymm11
.byte 196,65,44,93,224 // vminps %ymm8,%ymm10,%ymm12
@@ -26682,7 +26682,7 @@ _sk_load_tables_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,4,2,0,0 // jne 11a28 <_sk_load_tables_avx+0x218>
.byte 196,65,125,16,20,144 // vmovupd (%r8,%rdx,4),%ymm10
- .byte 197,125,40,13,142,132,2,0 // vmovapd 0x2848e(%rip),%ymm9 # 39cc0 <_sk_srcover_bgra_8888_sse2_lowp+0x838>
+ .byte 197,125,40,13,14,119,2,0 // vmovapd 0x2770e(%rip),%ymm9 # 38f40 <_sk_srcover_bgra_8888_sse2_lowp+0x83c>
.byte 196,193,45,84,201 // vandpd %ymm9,%ymm10,%ymm1
.byte 196,227,125,25,200,1 // vextractf128 $0x1,%ymm1,%xmm0
.byte 196,193,249,126,192 // vmovq %xmm0,%r8
@@ -26774,7 +26774,7 @@ _sk_load_tables_avx:
.byte 196,193,65,114,213,24 // vpsrld $0x18,%xmm13,%xmm7
.byte 196,227,101,24,223,1 // vinsertf128 $0x1,%xmm7,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,226,125,24,61,150,125,2,0 // vbroadcastss 0x27d96(%rip),%ymm7 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,61,18,112,2,0 // vbroadcastss 0x27012(%rip),%ymm7 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,228,89,223 // vmulps %ymm7,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,124,36,200 // vmovups -0x38(%rsp),%ymm7
@@ -26860,7 +26860,7 @@ _sk_load_tables_u16_be_avx:
.byte 197,177,108,208 // vpunpcklqdq %xmm0,%xmm9,%xmm2
.byte 197,49,109,232 // vpunpckhqdq %xmm0,%xmm9,%xmm13
.byte 196,65,57,108,212 // vpunpcklqdq %xmm12,%xmm8,%xmm10
- .byte 197,121,111,29,160,132,2,0 // vmovdqa 0x284a0(%rip),%xmm11 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,121,111,29,32,119,2,0 // vmovdqa 0x27720(%rip),%xmm11 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,193,105,219,195 // vpand %xmm11,%xmm2,%xmm0
.byte 197,249,112,208,78 // vpshufd $0x4e,%xmm0,%xmm2
.byte 196,226,121,51,210 // vpmovzxwd %xmm2,%xmm2
@@ -26954,7 +26954,7 @@ _sk_load_tables_u16_be_avx:
.byte 196,226,121,51,219 // vpmovzxwd %xmm3,%xmm3
.byte 196,227,77,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm6,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,226,125,24,53,119,122,2,0 // vbroadcastss 0x27a77(%rip),%ymm6 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,226,125,24,53,243,108,2,0 // vbroadcastss 0x26cf3(%rip),%ymm6 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 197,228,89,222 // vmulps %ymm6,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,16,116,36,168 // vmovups -0x58(%rsp),%ymm6
@@ -27018,7 +27018,7 @@ _sk_load_tables_rgb_u16_be_avx:
.byte 197,105,108,216 // vpunpcklqdq %xmm0,%xmm2,%xmm11
.byte 197,241,108,211 // vpunpcklqdq %xmm3,%xmm1,%xmm2
.byte 197,241,109,203 // vpunpckhqdq %xmm3,%xmm1,%xmm1
- .byte 197,121,111,53,160,129,2,0 // vmovdqa 0x281a0(%rip),%xmm14 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,121,111,53,32,116,2,0 // vmovdqa 0x27420(%rip),%xmm14 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,193,105,219,214 // vpand %xmm14,%xmm2,%xmm2
.byte 197,249,112,218,78 // vpshufd $0x4e,%xmm2,%xmm3
.byte 196,226,121,51,219 // vpmovzxwd %xmm3,%xmm3
@@ -27104,7 +27104,7 @@ _sk_load_tables_rgb_u16_be_avx:
.byte 196,227,105,33,215,48 // vinsertps $0x30,%xmm7,%xmm2,%xmm2
.byte 196,227,77,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm6,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,13,119,2,0 // vbroadcastss 0x2770d(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,137,105,2,0 // vbroadcastss 0x26989(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,16,116,36,168 // vmovups -0x58(%rsp),%ymm6
.byte 197,252,16,124,36,200 // vmovups -0x38(%rsp),%ymm7
.byte 255,224 // jmpq *%rax
@@ -27175,7 +27175,7 @@ HIDDEN _sk_byte_tables_avx
FUNCTION(_sk_byte_tables_avx)
_sk_byte_tables_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,246,117,2,0 // vbroadcastss 0x275f6(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,114,104,2,0 // vbroadcastss 0x26872(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,125,91,200 // vcvtps2dq %ymm0,%ymm9
.byte 196,65,249,126,200 // vmovq %xmm9,%r8
@@ -27310,7 +27310,7 @@ _sk_byte_tables_avx:
.byte 196,194,121,49,204 // vpmovzxbd %xmm12,%xmm1
.byte 196,194,121,49,213 // vpmovzxbd %xmm13,%xmm2
.byte 196,227,117,24,202,1 // vinsertf128 $0x1,%xmm2,%ymm1,%ymm1
- .byte 196,98,125,24,13,178,115,2,0 // vbroadcastss 0x273b2(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,46,102,2,0 // vbroadcastss 0x2662e(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
.byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
@@ -27438,7 +27438,7 @@ _sk_byte_tables_rgb_avx:
.byte 196,194,121,49,203 // vpmovzxbd %xmm11,%xmm1
.byte 196,227,125,24,193,1 // vinsertf128 $0x1,%xmm1,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,103,113,2,0 // vbroadcastss 0x27167(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,227,99,2,0 // vbroadcastss 0x263e3(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,194,121,49,202 // vpmovzxbd %xmm10,%xmm1
.byte 196,194,121,49,212 // vpmovzxbd %xmm12,%xmm2
@@ -27675,36 +27675,36 @@ _sk_parametric_r_avx:
.byte 196,193,124,88,195 // vaddps %ymm11,%ymm0,%ymm0
.byte 196,98,125,24,16 // vbroadcastss (%rax),%ymm10
.byte 197,124,91,216 // vcvtdq2ps %ymm0,%ymm11
- .byte 196,98,125,24,37,248,109,2,0 // vbroadcastss 0x26df8(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,37,116,96,2,0 // vbroadcastss 0x26074(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
.byte 196,65,36,89,220 // vmulps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,238,109,2,0 // vbroadcastss 0x26dee(%rip),%ymm12 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,106,96,2,0 // vbroadcastss 0x2606a(%rip),%ymm12 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,124,84,196 // vandps %ymm12,%ymm0,%ymm0
- .byte 196,98,125,24,37,64,109,2,0 // vbroadcastss 0x26d40(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,188,95,2,0 // vbroadcastss 0x25fbc(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,124,86,196 // vorps %ymm12,%ymm0,%ymm0
- .byte 196,98,125,24,37,214,109,2,0 // vbroadcastss 0x26dd6(%rip),%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,37,82,96,2,0 // vbroadcastss 0x26052(%rip),%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,204,109,2,0 // vbroadcastss 0x26dcc(%rip),%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,37,72,96,2,0 // vbroadcastss 0x26048(%rip),%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,65,124,89,228 // vmulps %ymm12,%ymm0,%ymm12
.byte 196,65,36,92,220 // vsubps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,189,109,2,0 // vbroadcastss 0x26dbd(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,57,96,2,0 // vbroadcastss 0x26039(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,124,88,196 // vaddps %ymm12,%ymm0,%ymm0
- .byte 196,98,125,24,37,179,109,2,0 // vbroadcastss 0x26db3(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,47,96,2,0 // vbroadcastss 0x2602f(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 197,156,94,192 // vdivps %ymm0,%ymm12,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
.byte 197,172,89,192 // vmulps %ymm0,%ymm10,%ymm0
.byte 196,99,125,8,208,1 // vroundps $0x1,%ymm0,%ymm10
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,151,109,2,0 // vbroadcastss 0x26d97(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,19,96,2,0 // vbroadcastss 0x26013(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,124,88,195 // vaddps %ymm11,%ymm0,%ymm0
- .byte 196,98,125,24,29,141,109,2,0 // vbroadcastss 0x26d8d(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,9,96,2,0 // vbroadcastss 0x26009(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,65,44,89,219 // vmulps %ymm11,%ymm10,%ymm11
.byte 196,193,124,92,195 // vsubps %ymm11,%ymm0,%ymm0
- .byte 196,98,125,24,29,126,109,2,0 // vbroadcastss 0x26d7e(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,98,125,24,29,250,95,2,0 // vbroadcastss 0x25ffa(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,116,109,2,0 // vbroadcastss 0x26d74(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,240,95,2,0 // vbroadcastss 0x25ff0(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,124,88,194 // vaddps %ymm10,%ymm0,%ymm0
- .byte 196,98,125,24,21,101,109,2,0 // vbroadcastss 0x26d65(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,225,95,2,0 // vbroadcastss 0x25fe1(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,124,89,194 // vmulps %ymm10,%ymm0,%ymm0
.byte 197,253,91,192 // vcvtps2dq %ymm0,%ymm0
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -27712,7 +27712,7 @@ _sk_parametric_r_avx:
.byte 196,195,125,74,193,128 // vblendvps %ymm8,%ymm9,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,121,108,2,0 // vbroadcastss 0x26c79(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,245,94,2,0 // vbroadcastss 0x25ef5(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -27734,36 +27734,36 @@ _sk_parametric_g_avx:
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
.byte 196,98,125,24,16 // vbroadcastss (%rax),%ymm10
.byte 197,124,91,217 // vcvtdq2ps %ymm1,%ymm11
- .byte 196,98,125,24,37,190,108,2,0 // vbroadcastss 0x26cbe(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,37,58,95,2,0 // vbroadcastss 0x25f3a(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
.byte 196,65,36,89,220 // vmulps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,180,108,2,0 // vbroadcastss 0x26cb4(%rip),%ymm12 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,48,95,2,0 // vbroadcastss 0x25f30(%rip),%ymm12 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,116,84,204 // vandps %ymm12,%ymm1,%ymm1
- .byte 196,98,125,24,37,6,108,2,0 // vbroadcastss 0x26c06(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,130,94,2,0 // vbroadcastss 0x25e82(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,116,86,204 // vorps %ymm12,%ymm1,%ymm1
- .byte 196,98,125,24,37,156,108,2,0 // vbroadcastss 0x26c9c(%rip),%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,37,24,95,2,0 // vbroadcastss 0x25f18(%rip),%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,146,108,2,0 // vbroadcastss 0x26c92(%rip),%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,37,14,95,2,0 // vbroadcastss 0x25f0e(%rip),%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,65,116,89,228 // vmulps %ymm12,%ymm1,%ymm12
.byte 196,65,36,92,220 // vsubps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,131,108,2,0 // vbroadcastss 0x26c83(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,255,94,2,0 // vbroadcastss 0x25eff(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,116,88,204 // vaddps %ymm12,%ymm1,%ymm1
- .byte 196,98,125,24,37,121,108,2,0 // vbroadcastss 0x26c79(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,245,94,2,0 // vbroadcastss 0x25ef5(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 197,156,94,201 // vdivps %ymm1,%ymm12,%ymm1
.byte 197,164,92,201 // vsubps %ymm1,%ymm11,%ymm1
.byte 197,172,89,201 // vmulps %ymm1,%ymm10,%ymm1
.byte 196,99,125,8,209,1 // vroundps $0x1,%ymm1,%ymm10
.byte 196,65,116,92,210 // vsubps %ymm10,%ymm1,%ymm10
- .byte 196,98,125,24,29,93,108,2,0 // vbroadcastss 0x26c5d(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,217,94,2,0 // vbroadcastss 0x25ed9(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,29,83,108,2,0 // vbroadcastss 0x26c53(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,207,94,2,0 // vbroadcastss 0x25ecf(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,65,44,89,219 // vmulps %ymm11,%ymm10,%ymm11
.byte 196,193,116,92,203 // vsubps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,29,68,108,2,0 // vbroadcastss 0x26c44(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,98,125,24,29,192,94,2,0 // vbroadcastss 0x25ec0(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,58,108,2,0 // vbroadcastss 0x26c3a(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,182,94,2,0 // vbroadcastss 0x25eb6(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,116,88,202 // vaddps %ymm10,%ymm1,%ymm1
- .byte 196,98,125,24,21,43,108,2,0 // vbroadcastss 0x26c2b(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,167,94,2,0 // vbroadcastss 0x25ea7(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
.byte 197,253,91,201 // vcvtps2dq %ymm1,%ymm1
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -27771,7 +27771,7 @@ _sk_parametric_g_avx:
.byte 196,195,117,74,201,128 // vblendvps %ymm8,%ymm9,%ymm1,%ymm1
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
- .byte 196,98,125,24,5,63,107,2,0 // vbroadcastss 0x26b3f(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,187,93,2,0 // vbroadcastss 0x25dbb(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -27793,36 +27793,36 @@ _sk_parametric_b_avx:
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
.byte 196,98,125,24,16 // vbroadcastss (%rax),%ymm10
.byte 197,124,91,218 // vcvtdq2ps %ymm2,%ymm11
- .byte 196,98,125,24,37,132,107,2,0 // vbroadcastss 0x26b84(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,37,0,94,2,0 // vbroadcastss 0x25e00(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
.byte 196,65,36,89,220 // vmulps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,122,107,2,0 // vbroadcastss 0x26b7a(%rip),%ymm12 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,246,93,2,0 // vbroadcastss 0x25df6(%rip),%ymm12 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,108,84,212 // vandps %ymm12,%ymm2,%ymm2
- .byte 196,98,125,24,37,204,106,2,0 // vbroadcastss 0x26acc(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,72,93,2,0 // vbroadcastss 0x25d48(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,108,86,212 // vorps %ymm12,%ymm2,%ymm2
- .byte 196,98,125,24,37,98,107,2,0 // vbroadcastss 0x26b62(%rip),%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,37,222,93,2,0 // vbroadcastss 0x25dde(%rip),%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,88,107,2,0 // vbroadcastss 0x26b58(%rip),%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,37,212,93,2,0 // vbroadcastss 0x25dd4(%rip),%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,65,108,89,228 // vmulps %ymm12,%ymm2,%ymm12
.byte 196,65,36,92,220 // vsubps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,73,107,2,0 // vbroadcastss 0x26b49(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,197,93,2,0 // vbroadcastss 0x25dc5(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,108,88,212 // vaddps %ymm12,%ymm2,%ymm2
- .byte 196,98,125,24,37,63,107,2,0 // vbroadcastss 0x26b3f(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,187,93,2,0 // vbroadcastss 0x25dbb(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 197,156,94,210 // vdivps %ymm2,%ymm12,%ymm2
.byte 197,164,92,210 // vsubps %ymm2,%ymm11,%ymm2
.byte 197,172,89,210 // vmulps %ymm2,%ymm10,%ymm2
.byte 196,99,125,8,210,1 // vroundps $0x1,%ymm2,%ymm10
.byte 196,65,108,92,210 // vsubps %ymm10,%ymm2,%ymm10
- .byte 196,98,125,24,29,35,107,2,0 // vbroadcastss 0x26b23(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,159,93,2,0 // vbroadcastss 0x25d9f(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
- .byte 196,98,125,24,29,25,107,2,0 // vbroadcastss 0x26b19(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,149,93,2,0 // vbroadcastss 0x25d95(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,65,44,89,219 // vmulps %ymm11,%ymm10,%ymm11
.byte 196,193,108,92,211 // vsubps %ymm11,%ymm2,%ymm2
- .byte 196,98,125,24,29,10,107,2,0 // vbroadcastss 0x26b0a(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,98,125,24,29,134,93,2,0 // vbroadcastss 0x25d86(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,0,107,2,0 // vbroadcastss 0x26b00(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,124,93,2,0 // vbroadcastss 0x25d7c(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,108,88,210 // vaddps %ymm10,%ymm2,%ymm2
- .byte 196,98,125,24,21,241,106,2,0 // vbroadcastss 0x26af1(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,109,93,2,0 // vbroadcastss 0x25d6d(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,108,89,210 // vmulps %ymm10,%ymm2,%ymm2
.byte 197,253,91,210 // vcvtps2dq %ymm2,%ymm2
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -27830,7 +27830,7 @@ _sk_parametric_b_avx:
.byte 196,195,109,74,209,128 // vblendvps %ymm8,%ymm9,%ymm2,%ymm2
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,5,106,2,0 // vbroadcastss 0x26a05(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,129,92,2,0 // vbroadcastss 0x25c81(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -27852,36 +27852,36 @@ _sk_parametric_a_avx:
.byte 196,193,100,88,219 // vaddps %ymm11,%ymm3,%ymm3
.byte 196,98,125,24,16 // vbroadcastss (%rax),%ymm10
.byte 197,124,91,219 // vcvtdq2ps %ymm3,%ymm11
- .byte 196,98,125,24,37,74,106,2,0 // vbroadcastss 0x26a4a(%rip),%ymm12 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,37,198,92,2,0 // vbroadcastss 0x25cc6(%rip),%ymm12 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
.byte 196,65,36,89,220 // vmulps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,64,106,2,0 // vbroadcastss 0x26a40(%rip),%ymm12 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,37,188,92,2,0 // vbroadcastss 0x25cbc(%rip),%ymm12 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,100,84,220 // vandps %ymm12,%ymm3,%ymm3
- .byte 196,98,125,24,37,146,105,2,0 // vbroadcastss 0x26992(%rip),%ymm12 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,37,14,92,2,0 // vbroadcastss 0x25c0e(%rip),%ymm12 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,100,86,220 // vorps %ymm12,%ymm3,%ymm3
- .byte 196,98,125,24,37,40,106,2,0 // vbroadcastss 0x26a28(%rip),%ymm12 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,37,164,92,2,0 // vbroadcastss 0x25ca4(%rip),%ymm12 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,30,106,2,0 // vbroadcastss 0x26a1e(%rip),%ymm12 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,37,154,92,2,0 // vbroadcastss 0x25c9a(%rip),%ymm12 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,65,100,89,228 // vmulps %ymm12,%ymm3,%ymm12
.byte 196,65,36,92,220 // vsubps %ymm12,%ymm11,%ymm11
- .byte 196,98,125,24,37,15,106,2,0 // vbroadcastss 0x26a0f(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,139,92,2,0 // vbroadcastss 0x25c8b(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,100,88,220 // vaddps %ymm12,%ymm3,%ymm3
- .byte 196,98,125,24,37,5,106,2,0 // vbroadcastss 0x26a05(%rip),%ymm12 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,37,129,92,2,0 // vbroadcastss 0x25c81(%rip),%ymm12 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 197,156,94,219 // vdivps %ymm3,%ymm12,%ymm3
.byte 197,164,92,219 // vsubps %ymm3,%ymm11,%ymm3
.byte 197,172,89,219 // vmulps %ymm3,%ymm10,%ymm3
.byte 196,99,125,8,211,1 // vroundps $0x1,%ymm3,%ymm10
.byte 196,65,100,92,210 // vsubps %ymm10,%ymm3,%ymm10
- .byte 196,98,125,24,29,233,105,2,0 // vbroadcastss 0x269e9(%rip),%ymm11 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,29,101,92,2,0 // vbroadcastss 0x25c65(%rip),%ymm11 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,100,88,219 // vaddps %ymm11,%ymm3,%ymm3
- .byte 196,98,125,24,29,223,105,2,0 // vbroadcastss 0x269df(%rip),%ymm11 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,98,125,24,29,91,92,2,0 // vbroadcastss 0x25c5b(%rip),%ymm11 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 196,65,44,89,219 // vmulps %ymm11,%ymm10,%ymm11
.byte 196,193,100,92,219 // vsubps %ymm11,%ymm3,%ymm3
- .byte 196,98,125,24,29,208,105,2,0 // vbroadcastss 0x269d0(%rip),%ymm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,98,125,24,29,76,92,2,0 // vbroadcastss 0x25c4c(%rip),%ymm11 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 196,65,36,92,210 // vsubps %ymm10,%ymm11,%ymm10
- .byte 196,98,125,24,29,198,105,2,0 // vbroadcastss 0x269c6(%rip),%ymm11 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,98,125,24,29,66,92,2,0 // vbroadcastss 0x25c42(%rip),%ymm11 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 196,65,36,94,210 // vdivps %ymm10,%ymm11,%ymm10
.byte 196,193,100,88,218 // vaddps %ymm10,%ymm3,%ymm3
- .byte 196,98,125,24,21,183,105,2,0 // vbroadcastss 0x269b7(%rip),%ymm10 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,98,125,24,21,51,92,2,0 // vbroadcastss 0x25c33(%rip),%ymm10 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 196,193,100,89,218 // vmulps %ymm10,%ymm3,%ymm3
.byte 197,253,91,219 // vcvtps2dq %ymm3,%ymm3
.byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
@@ -27889,7 +27889,7 @@ _sk_parametric_a_avx:
.byte 196,195,101,74,217,128 // vblendvps %ymm8,%ymm9,%ymm3,%ymm3
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
- .byte 196,98,125,24,5,203,104,2,0 // vbroadcastss 0x268cb(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,71,91,2,0 // vbroadcastss 0x25b47(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -27908,34 +27908,34 @@ _sk_gamma_avx:
.byte 197,252,40,233 // vmovaps %ymm1,%ymm5
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,91,200 // vcvtdq2ps %ymm0,%ymm1
- .byte 196,98,125,24,5,34,105,2,0 // vbroadcastss 0x26922(%rip),%ymm8 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x348>
+ .byte 196,98,125,24,5,158,91,2,0 // vbroadcastss 0x25b9e(%rip),%ymm8 # 38a4c <_sk_srcover_bgra_8888_sse2_lowp+0x348>
.byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
- .byte 196,98,125,24,13,24,105,2,0 // vbroadcastss 0x26918(%rip),%ymm9 # 397d4 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
+ .byte 196,98,125,24,13,148,91,2,0 // vbroadcastss 0x25b94(%rip),%ymm9 # 38a50 <_sk_srcover_bgra_8888_sse2_lowp+0x34c>
.byte 196,193,124,84,193 // vandps %ymm9,%ymm0,%ymm0
- .byte 196,226,125,24,37,106,104,2,0 // vbroadcastss 0x2686a(%rip),%ymm4 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,37,230,90,2,0 // vbroadcastss 0x25ae6(%rip),%ymm4 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,252,86,196 // vorps %ymm4,%ymm0,%ymm0
- .byte 196,98,125,24,21,1,105,2,0 // vbroadcastss 0x26901(%rip),%ymm10 # 397d8 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
+ .byte 196,98,125,24,21,125,91,2,0 // vbroadcastss 0x25b7d(%rip),%ymm10 # 38a54 <_sk_srcover_bgra_8888_sse2_lowp+0x350>
.byte 196,193,116,88,202 // vaddps %ymm10,%ymm1,%ymm1
- .byte 196,98,125,24,29,247,104,2,0 // vbroadcastss 0x268f7(%rip),%ymm11 # 397dc <_sk_srcover_bgra_8888_sse2_lowp+0x354>
+ .byte 196,98,125,24,29,115,91,2,0 // vbroadcastss 0x25b73(%rip),%ymm11 # 38a58 <_sk_srcover_bgra_8888_sse2_lowp+0x354>
.byte 196,193,124,89,211 // vmulps %ymm11,%ymm0,%ymm2
.byte 197,244,92,202 // vsubps %ymm2,%ymm1,%ymm1
- .byte 196,98,125,24,37,233,104,2,0 // vbroadcastss 0x268e9(%rip),%ymm12 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x358>
+ .byte 196,98,125,24,37,101,91,2,0 // vbroadcastss 0x25b65(%rip),%ymm12 # 38a5c <_sk_srcover_bgra_8888_sse2_lowp+0x358>
.byte 196,193,124,88,196 // vaddps %ymm12,%ymm0,%ymm0
- .byte 196,98,125,24,45,223,104,2,0 // vbroadcastss 0x268df(%rip),%ymm13 # 397e4 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
+ .byte 196,98,125,24,45,91,91,2,0 // vbroadcastss 0x25b5b(%rip),%ymm13 # 38a60 <_sk_srcover_bgra_8888_sse2_lowp+0x35c>
.byte 197,148,94,192 // vdivps %ymm0,%ymm13,%ymm0
.byte 197,244,92,192 // vsubps %ymm0,%ymm1,%ymm0
.byte 196,98,125,24,48 // vbroadcastss (%rax),%ymm14
.byte 196,193,124,89,198 // vmulps %ymm14,%ymm0,%ymm0
.byte 196,227,125,8,200,1 // vroundps $0x1,%ymm0,%ymm1
.byte 197,252,92,241 // vsubps %ymm1,%ymm0,%ymm6
- .byte 196,98,125,24,61,190,104,2,0 // vbroadcastss 0x268be(%rip),%ymm15 # 397e8 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
+ .byte 196,98,125,24,61,58,91,2,0 // vbroadcastss 0x25b3a(%rip),%ymm15 # 38a64 <_sk_srcover_bgra_8888_sse2_lowp+0x360>
.byte 196,193,124,88,199 // vaddps %ymm15,%ymm0,%ymm0
- .byte 196,226,125,24,29,180,104,2,0 // vbroadcastss 0x268b4(%rip),%ymm3 # 397ec <_sk_srcover_bgra_8888_sse2_lowp+0x364>
+ .byte 196,226,125,24,29,48,91,2,0 // vbroadcastss 0x25b30(%rip),%ymm3 # 38a68 <_sk_srcover_bgra_8888_sse2_lowp+0x364>
.byte 197,204,89,203 // vmulps %ymm3,%ymm6,%ymm1
.byte 197,252,92,201 // vsubps %ymm1,%ymm0,%ymm1
- .byte 196,226,125,24,21,167,104,2,0 // vbroadcastss 0x268a7(%rip),%ymm2 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x368>
+ .byte 196,226,125,24,21,35,91,2,0 // vbroadcastss 0x25b23(%rip),%ymm2 # 38a6c <_sk_srcover_bgra_8888_sse2_lowp+0x368>
.byte 197,236,92,198 // vsubps %ymm6,%ymm2,%ymm0
- .byte 196,226,125,24,53,158,104,2,0 // vbroadcastss 0x2689e(%rip),%ymm6 # 397f4 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
+ .byte 196,226,125,24,53,26,91,2,0 // vbroadcastss 0x25b1a(%rip),%ymm6 # 38a70 <_sk_srcover_bgra_8888_sse2_lowp+0x36c>
.byte 197,204,94,192 // vdivps %ymm0,%ymm6,%ymm0
.byte 197,244,88,192 // vaddps %ymm0,%ymm1,%ymm0
.byte 197,252,17,68,36,128 // vmovups %ymm0,-0x80(%rsp)
@@ -27977,7 +27977,7 @@ _sk_gamma_avx:
.byte 197,236,92,212 // vsubps %ymm4,%ymm2,%ymm2
.byte 197,204,94,210 // vdivps %ymm2,%ymm6,%ymm2
.byte 197,244,88,202 // vaddps %ymm2,%ymm1,%ymm1
- .byte 196,226,125,24,21,225,103,2,0 // vbroadcastss 0x267e1(%rip),%ymm2 # 397f8 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
+ .byte 196,226,125,24,21,93,90,2,0 // vbroadcastss 0x25a5d(%rip),%ymm2 # 38a74 <_sk_srcover_bgra_8888_sse2_lowp+0x370>
.byte 197,236,89,92,36,128 // vmulps -0x80(%rsp),%ymm2,%ymm3
.byte 197,252,89,226 // vmulps %ymm2,%ymm0,%ymm4
.byte 197,244,89,210 // vmulps %ymm2,%ymm1,%ymm2
@@ -27997,31 +27997,31 @@ HIDDEN _sk_lab_to_xyz_avx
.globl _sk_lab_to_xyz_avx
FUNCTION(_sk_lab_to_xyz_avx)
_sk_lab_to_xyz_avx:
- .byte 196,98,125,24,5,157,103,2,0 // vbroadcastss 0x2679d(%rip),%ymm8 # 397fc <_sk_srcover_bgra_8888_sse2_lowp+0x374>
+ .byte 196,98,125,24,5,25,90,2,0 // vbroadcastss 0x25a19(%rip),%ymm8 # 38a78 <_sk_srcover_bgra_8888_sse2_lowp+0x374>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,247,102,2,0 // vbroadcastss 0x266f7(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,115,89,2,0 // vbroadcastss 0x25973(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
- .byte 196,98,125,24,13,133,103,2,0 // vbroadcastss 0x26785(%rip),%ymm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x378>
+ .byte 196,98,125,24,13,1,90,2,0 // vbroadcastss 0x25a01(%rip),%ymm9 # 38a7c <_sk_srcover_bgra_8888_sse2_lowp+0x378>
.byte 196,193,116,88,201 // vaddps %ymm9,%ymm1,%ymm1
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
.byte 196,193,108,88,209 // vaddps %ymm9,%ymm2,%ymm2
- .byte 196,98,125,24,5,113,103,2,0 // vbroadcastss 0x26771(%rip),%ymm8 # 39804 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
+ .byte 196,98,125,24,5,237,89,2,0 // vbroadcastss 0x259ed(%rip),%ymm8 # 38a80 <_sk_srcover_bgra_8888_sse2_lowp+0x37c>
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,103,103,2,0 // vbroadcastss 0x26767(%rip),%ymm8 # 39808 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
+ .byte 196,98,125,24,5,227,89,2,0 // vbroadcastss 0x259e3(%rip),%ymm8 # 38a84 <_sk_srcover_bgra_8888_sse2_lowp+0x380>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,93,103,2,0 // vbroadcastss 0x2675d(%rip),%ymm8 # 3980c <_sk_srcover_bgra_8888_sse2_lowp+0x384>
+ .byte 196,98,125,24,5,217,89,2,0 // vbroadcastss 0x259d9(%rip),%ymm8 # 38a88 <_sk_srcover_bgra_8888_sse2_lowp+0x384>
.byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
.byte 197,252,88,201 // vaddps %ymm1,%ymm0,%ymm1
- .byte 196,98,125,24,5,79,103,2,0 // vbroadcastss 0x2674f(%rip),%ymm8 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x388>
+ .byte 196,98,125,24,5,203,89,2,0 // vbroadcastss 0x259cb(%rip),%ymm8 # 38a8c <_sk_srcover_bgra_8888_sse2_lowp+0x388>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
.byte 197,252,92,210 // vsubps %ymm2,%ymm0,%ymm2
.byte 197,116,89,193 // vmulps %ymm1,%ymm1,%ymm8
.byte 196,65,116,89,192 // vmulps %ymm8,%ymm1,%ymm8
- .byte 196,98,125,24,13,56,103,2,0 // vbroadcastss 0x26738(%rip),%ymm9 # 39814 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
+ .byte 196,98,125,24,13,180,89,2,0 // vbroadcastss 0x259b4(%rip),%ymm9 # 38a90 <_sk_srcover_bgra_8888_sse2_lowp+0x38c>
.byte 196,65,52,194,208,1 // vcmpltps %ymm8,%ymm9,%ymm10
- .byte 196,98,125,24,29,45,103,2,0 // vbroadcastss 0x2672d(%rip),%ymm11 # 39818 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
+ .byte 196,98,125,24,29,169,89,2,0 // vbroadcastss 0x259a9(%rip),%ymm11 # 38a94 <_sk_srcover_bgra_8888_sse2_lowp+0x390>
.byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
- .byte 196,98,125,24,37,35,103,2,0 // vbroadcastss 0x26723(%rip),%ymm12 # 3981c <_sk_srcover_bgra_8888_sse2_lowp+0x394>
+ .byte 196,98,125,24,37,159,89,2,0 // vbroadcastss 0x2599f(%rip),%ymm12 # 38a98 <_sk_srcover_bgra_8888_sse2_lowp+0x394>
.byte 196,193,116,89,204 // vmulps %ymm12,%ymm1,%ymm1
.byte 196,67,117,74,192,160 // vblendvps %ymm10,%ymm8,%ymm1,%ymm8
.byte 197,252,89,200 // vmulps %ymm0,%ymm0,%ymm1
@@ -28036,9 +28036,9 @@ _sk_lab_to_xyz_avx:
.byte 196,193,108,88,211 // vaddps %ymm11,%ymm2,%ymm2
.byte 196,193,108,89,212 // vmulps %ymm12,%ymm2,%ymm2
.byte 196,227,109,74,208,144 // vblendvps %ymm9,%ymm0,%ymm2,%ymm2
- .byte 196,226,125,24,5,217,102,2,0 // vbroadcastss 0x266d9(%rip),%ymm0 # 39820 <_sk_srcover_bgra_8888_sse2_lowp+0x398>
+ .byte 196,226,125,24,5,85,89,2,0 // vbroadcastss 0x25955(%rip),%ymm0 # 38a9c <_sk_srcover_bgra_8888_sse2_lowp+0x398>
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,208,102,2,0 // vbroadcastss 0x266d0(%rip),%ymm8 # 39824 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
+ .byte 196,98,125,24,5,76,89,2,0 // vbroadcastss 0x2594c(%rip),%ymm8 # 38aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x39c>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -28056,13 +28056,13 @@ _sk_load_a8_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,68 // jne 131b9 <_sk_load_a8_avx+0x5c>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,141,110,2,0 // vpand 0x26e8d(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,13,97,2,0 // vpand 0x2610d(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,121,51,200 // vpmovzxwd %xmm0,%xmm1
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,11,102,2,0 // vbroadcastss 0x2660b(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,135,88,2,0 // vbroadcastss 0x25887(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
@@ -28110,7 +28110,7 @@ _sk_load_a8_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,222 // jmpq ffffffffdf013260 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefd9dd8>
+ .byte 233,255,255,255,222 // jmpq ffffffffdf013260 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefdab5c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,211 // callq *%rbx
@@ -28134,13 +28134,13 @@ _sk_load_a8_dst_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,68 // jne 132c8 <_sk_load_a8_dst_avx+0x5c>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,126,109,2,0 // vpand 0x26d7e(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,254,95,2,0 // vpand 0x25ffe(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,121,51,236 // vpmovzxwd %xmm4,%xmm5
.byte 197,249,112,228,78 // vpshufd $0x4e,%xmm4,%xmm4
.byte 196,226,121,51,228 // vpmovzxwd %xmm4,%xmm4
.byte 196,227,85,24,228,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,252,100,2,0 // vbroadcastss 0x264fc(%rip),%ymm5 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,45,120,87,2,0 // vbroadcastss 0x25778(%rip),%ymm5 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,220,89,253 // vmulps %ymm5,%ymm4,%ymm7
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
@@ -28188,7 +28188,7 @@ _sk_load_a8_dst_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,221 // callq ffffffffde013370 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfd9ee8>
+ .byte 232,255,255,255,221 // callq ffffffffde013370 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddfdac6c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,210 // callq *%rdx
@@ -28268,7 +28268,7 @@ _sk_gather_a8_avx:
.byte 196,226,121,49,201 // vpmovzxbd %xmm1,%xmm1
.byte 196,227,125,24,193,1 // vinsertf128 $0x1,%xmm1,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,2,99,2,0 // vbroadcastss 0x26302(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,126,85,2,0 // vbroadcastss 0x2557e(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
@@ -28291,7 +28291,7 @@ _sk_store_a8_avx:
.byte 77,15,175,193 // imul %r9,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,126,98,2,0 // vbroadcastss 0x2627e(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,250,84,2,0 // vbroadcastss 0x254fa(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,100,89,192 // vmulps %ymm8,%ymm3,%ymm8
.byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
.byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
@@ -28316,13 +28316,13 @@ _sk_store_a8_avx:
.byte 196,3,121,20,4,24,0 // vpextrb $0x0,%xmm8,(%r8,%r11,1)
.byte 235,202 // jmp 1350a <_sk_store_a8_avx+0x40>
.byte 196,3,121,20,68,24,2,4 // vpextrb $0x4,%xmm8,0x2(%r8,%r11,1)
- .byte 196,98,57,0,5,207,106,2,0 // vpshufb 0x26acf(%rip),%xmm8,%xmm8 # 3a020 <_sk_srcover_bgra_8888_sse2_lowp+0xb98>
+ .byte 196,98,57,0,5,79,93,2,0 // vpshufb 0x25d4f(%rip),%xmm8,%xmm8 # 392a0 <_sk_srcover_bgra_8888_sse2_lowp+0xb9c>
.byte 196,3,121,21,4,24,0 // vpextrw $0x0,%xmm8,(%r8,%r11,1)
.byte 235,176 // jmp 1350a <_sk_store_a8_avx+0x40>
.byte 196,3,121,20,68,24,6,12 // vpextrb $0xc,%xmm8,0x6(%r8,%r11,1)
.byte 196,3,121,20,68,24,5,10 // vpextrb $0xa,%xmm8,0x5(%r8,%r11,1)
.byte 196,3,121,20,68,24,4,8 // vpextrb $0x8,%xmm8,0x4(%r8,%r11,1)
- .byte 196,98,57,0,5,181,106,2,0 // vpshufb 0x26ab5(%rip),%xmm8,%xmm8 # 3a030 <_sk_srcover_bgra_8888_sse2_lowp+0xba8>
+ .byte 196,98,57,0,5,53,93,2,0 // vpshufb 0x25d35(%rip),%xmm8,%xmm8 # 392b0 <_sk_srcover_bgra_8888_sse2_lowp+0xbac>
.byte 196,1,121,126,4,24 // vmovd %xmm8,(%r8,%r11,1)
.byte 235,135 // jmp 1350a <_sk_store_a8_avx+0x40>
.byte 144 // nop
@@ -28359,16 +28359,16 @@ _sk_load_g8_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,73 // jne 13601 <_sk_load_g8_avx+0x61>
.byte 196,130,121,48,4,24 // vpmovzxbw (%r8,%r11,1),%xmm0
- .byte 197,249,219,5,74,106,2,0 // vpand 0x26a4a(%rip),%xmm0,%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,249,219,5,202,92,2,0 // vpand 0x25cca(%rip),%xmm0,%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,121,51,200 // vpmovzxwd %xmm0,%xmm1
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,200,97,2,0 // vbroadcastss 0x261c8(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,68,84,2,0 // vbroadcastss 0x25444(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,65,97,2,0 // vbroadcastss 0x26141(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,189,83,2,0 // vbroadcastss 0x253bd(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 255,224 // jmpq *%rax
@@ -28413,7 +28413,7 @@ _sk_load_g8_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,222 // jmpq ffffffffdf0136a8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefda220>
+ .byte 233,255,255,255,222 // jmpq ffffffffdf0136a8 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefdafa4>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,211 // callq *%rbx
@@ -28437,16 +28437,16 @@ _sk_load_g8_dst_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,73 // jne 13715 <_sk_load_g8_dst_avx+0x61>
.byte 196,130,121,48,36,24 // vpmovzxbw (%r8,%r11,1),%xmm4
- .byte 197,217,219,37,54,105,2,0 // vpand 0x26936(%rip),%xmm4,%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 197,217,219,37,182,91,2,0 // vpand 0x25bb6(%rip),%xmm4,%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 196,226,121,51,236 // vpmovzxwd %xmm4,%xmm5
.byte 197,249,112,228,78 // vpshufd $0x4e,%xmm4,%xmm4
.byte 196,226,121,51,228 // vpmovzxwd %xmm4,%xmm4
.byte 196,227,85,24,228,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,180,96,2,0 // vbroadcastss 0x260b4(%rip),%ymm5 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,45,48,83,2,0 // vbroadcastss 0x25330(%rip),%ymm5 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,45,96,2,0 // vbroadcastss 0x2602d(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,169,82,2,0 // vbroadcastss 0x252a9(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,236 // vmovaps %ymm4,%ymm5
.byte 197,252,40,244 // vmovaps %ymm4,%ymm6
.byte 255,224 // jmpq *%rax
@@ -28491,7 +28491,7 @@ _sk_load_g8_dst_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,222 // jmpq ffffffffdf0137bc <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefda334>
+ .byte 233,255,255,255,222 // jmpq ffffffffdf0137bc <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdefdb0b8>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,211 // callq *%rbx
@@ -28571,10 +28571,10 @@ _sk_gather_g8_avx:
.byte 196,226,121,49,201 // vpmovzxbd %xmm1,%xmm1
.byte 196,227,125,24,193,1 // vinsertf128 $0x1,%xmm1,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,182,94,2,0 // vbroadcastss 0x25eb6(%rip),%ymm1 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,226,125,24,13,50,81,2,0 // vbroadcastss 0x25132(%rip),%ymm1 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,47,94,2,0 // vbroadcastss 0x25e2f(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,171,80,2,0 // vbroadcastss 0x250ab(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
.byte 197,252,40,208 // vmovaps %ymm0,%ymm2
.byte 91 // pop %rbx
@@ -28602,23 +28602,23 @@ _sk_load_565_avx:
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,208,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm2
- .byte 196,226,125,24,5,86,94,2,0 // vbroadcastss 0x25e56(%rip),%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,24,5,210,80,2,0 // vbroadcastss 0x250d2(%rip),%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,236,84,192 // vandps %ymm0,%ymm2,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,73,94,2,0 // vbroadcastss 0x25e49(%rip),%ymm1 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,13,197,80,2,0 // vbroadcastss 0x250c5(%rip),%ymm1 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,24,13,64,94,2,0 // vbroadcastss 0x25e40(%rip),%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,24,13,188,80,2,0 // vbroadcastss 0x250bc(%rip),%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,236,84,201 // vandps %ymm1,%ymm2,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,29,51,94,2,0 // vbroadcastss 0x25e33(%rip),%ymm3 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,29,175,80,2,0 // vbroadcastss 0x250af(%rip),%ymm3 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
- .byte 196,226,125,24,29,42,94,2,0 // vbroadcastss 0x25e2a(%rip),%ymm3 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,24,29,166,80,2,0 // vbroadcastss 0x250a6(%rip),%ymm3 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,236,84,211 // vandps %ymm3,%ymm2,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,226,125,24,29,29,94,2,0 // vbroadcastss 0x25e1d(%rip),%ymm3 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,29,153,80,2,0 // vbroadcastss 0x25099(%rip),%ymm3 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,126,93,2,0 // vbroadcastss 0x25d7e(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,250,79,2,0 // vbroadcastss 0x24ffa(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -28683,23 +28683,23 @@ _sk_load_565_dst_avx:
.byte 197,249,112,228,78 // vpshufd $0x4e,%xmm4,%xmm4
.byte 196,226,121,51,228 // vpmovzxwd %xmm4,%xmm4
.byte 196,227,85,24,244,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm6
- .byte 196,226,125,24,37,21,93,2,0 // vbroadcastss 0x25d15(%rip),%ymm4 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,24,37,145,79,2,0 // vbroadcastss 0x24f91(%rip),%ymm4 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,204,84,228 // vandps %ymm4,%ymm6,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,8,93,2,0 // vbroadcastss 0x25d08(%rip),%ymm5 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,45,132,79,2,0 // vbroadcastss 0x24f84(%rip),%ymm5 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
- .byte 196,226,125,24,45,255,92,2,0 // vbroadcastss 0x25cff(%rip),%ymm5 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,24,45,123,79,2,0 // vbroadcastss 0x24f7b(%rip),%ymm5 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,204,84,237 // vandps %ymm5,%ymm6,%ymm5
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,226,125,24,61,242,92,2,0 // vbroadcastss 0x25cf2(%rip),%ymm7 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,61,110,79,2,0 // vbroadcastss 0x24f6e(%rip),%ymm7 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,212,89,239 // vmulps %ymm7,%ymm5,%ymm5
- .byte 196,226,125,24,61,233,92,2,0 // vbroadcastss 0x25ce9(%rip),%ymm7 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,24,61,101,79,2,0 // vbroadcastss 0x24f65(%rip),%ymm7 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,204,84,247 // vandps %ymm7,%ymm6,%ymm6
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 196,226,125,24,61,220,92,2,0 // vbroadcastss 0x25cdc(%rip),%ymm7 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,61,88,79,2,0 // vbroadcastss 0x24f58(%rip),%ymm7 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,204,89,247 // vmulps %ymm7,%ymm6,%ymm6
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,61,61,92,2,0 // vbroadcastss 0x25c3d(%rip),%ymm7 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,61,185,78,2,0 // vbroadcastss 0x24eb9(%rip),%ymm7 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,7 // and $0x7,%r9b
@@ -28819,23 +28819,23 @@ _sk_gather_565_avx:
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,208,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm2
- .byte 196,226,125,24,5,229,90,2,0 // vbroadcastss 0x25ae5(%rip),%ymm0 # 397b4 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
+ .byte 196,226,125,24,5,97,77,2,0 // vbroadcastss 0x24d61(%rip),%ymm0 # 38a30 <_sk_srcover_bgra_8888_sse2_lowp+0x32c>
.byte 197,236,84,192 // vandps %ymm0,%ymm2,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,216,90,2,0 // vbroadcastss 0x25ad8(%rip),%ymm1 # 397b8 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
+ .byte 196,226,125,24,13,84,77,2,0 // vbroadcastss 0x24d54(%rip),%ymm1 # 38a34 <_sk_srcover_bgra_8888_sse2_lowp+0x330>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,24,13,207,90,2,0 // vbroadcastss 0x25acf(%rip),%ymm1 # 397bc <_sk_srcover_bgra_8888_sse2_lowp+0x334>
+ .byte 196,226,125,24,13,75,77,2,0 // vbroadcastss 0x24d4b(%rip),%ymm1 # 38a38 <_sk_srcover_bgra_8888_sse2_lowp+0x334>
.byte 197,236,84,201 // vandps %ymm1,%ymm2,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,29,194,90,2,0 // vbroadcastss 0x25ac2(%rip),%ymm3 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x338>
+ .byte 196,226,125,24,29,62,77,2,0 // vbroadcastss 0x24d3e(%rip),%ymm3 # 38a3c <_sk_srcover_bgra_8888_sse2_lowp+0x338>
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
- .byte 196,226,125,24,29,185,90,2,0 // vbroadcastss 0x25ab9(%rip),%ymm3 # 397c4 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
+ .byte 196,226,125,24,29,53,77,2,0 // vbroadcastss 0x24d35(%rip),%ymm3 # 38a40 <_sk_srcover_bgra_8888_sse2_lowp+0x33c>
.byte 197,236,84,211 // vandps %ymm3,%ymm2,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,226,125,24,29,172,90,2,0 // vbroadcastss 0x25aac(%rip),%ymm3 # 397c8 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
+ .byte 196,226,125,24,29,40,77,2,0 // vbroadcastss 0x24d28(%rip),%ymm3 # 38a44 <_sk_srcover_bgra_8888_sse2_lowp+0x340>
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,13,90,2,0 // vbroadcastss 0x25a0d(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,137,76,2,0 // vbroadcastss 0x24c89(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
.byte 65,94 // pop %r14
@@ -28854,14 +28854,14 @@ _sk_store_565_avx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,212,90,2,0 // vbroadcastss 0x25ad4(%rip),%ymm8 # 39828 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
+ .byte 196,98,125,24,5,80,77,2,0 // vbroadcastss 0x24d50(%rip),%ymm8 # 38aa4 <_sk_srcover_bgra_8888_sse2_lowp+0x3a0>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,41,114,241,11 // vpslld $0xb,%xmm9,%xmm10
.byte 196,67,125,25,201,1 // vextractf128 $0x1,%ymm9,%xmm9
.byte 196,193,49,114,241,11 // vpslld $0xb,%xmm9,%xmm9
.byte 196,67,45,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
- .byte 196,98,125,24,21,173,90,2,0 // vbroadcastss 0x25aad(%rip),%ymm10 # 3982c <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
+ .byte 196,98,125,24,21,41,77,2,0 // vbroadcastss 0x24d29(%rip),%ymm10 # 38aa8 <_sk_srcover_bgra_8888_sse2_lowp+0x3a4>
.byte 196,65,116,89,210 // vmulps %ymm10,%ymm1,%ymm10
.byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
.byte 196,193,33,114,242,5 // vpslld $0x5,%xmm10,%xmm11
@@ -28916,7 +28916,7 @@ _sk_store_565_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,224 // callq ffffffffe1013e44 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fda9bc>
+ .byte 232,255,255,255,224 // callq ffffffffe1013e44 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe0fdb740>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // .byte 0xff
@@ -28939,25 +28939,25 @@ _sk_load_4444_avx:
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm3
- .byte 196,226,125,24,5,165,89,2,0 // vbroadcastss 0x259a5(%rip),%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,24,5,33,76,2,0 // vbroadcastss 0x24c21(%rip),%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,228,84,192 // vandps %ymm0,%ymm3,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,152,89,2,0 // vbroadcastss 0x25998(%rip),%ymm1 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,13,20,76,2,0 // vbroadcastss 0x24c14(%rip),%ymm1 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,24,13,143,89,2,0 // vbroadcastss 0x2598f(%rip),%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,24,13,11,76,2,0 // vbroadcastss 0x24c0b(%rip),%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,228,84,201 // vandps %ymm1,%ymm3,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,130,89,2,0 // vbroadcastss 0x25982(%rip),%ymm2 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,21,254,75,2,0 // vbroadcastss 0x24bfe(%rip),%ymm2 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,244,89,202 // vmulps %ymm2,%ymm1,%ymm1
- .byte 196,226,125,24,21,121,89,2,0 // vbroadcastss 0x25979(%rip),%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,24,21,245,75,2,0 // vbroadcastss 0x24bf5(%rip),%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,228,84,210 // vandps %ymm2,%ymm3,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,98,125,24,5,108,89,2,0 // vbroadcastss 0x2596c(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,232,75,2,0 // vbroadcastss 0x24be8(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,98,89,2,0 // vbroadcastss 0x25962(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,24,5,222,75,2,0 // vbroadcastss 0x24bde(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,84,89,2,0 // vbroadcastss 0x25954(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,208,75,2,0 // vbroadcastss 0x24bd0(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -29028,25 +29028,25 @@ _sk_load_4444_dst_avx:
.byte 197,249,112,228,78 // vpshufd $0x4e,%xmm4,%xmm4
.byte 196,226,121,51,228 // vpmovzxwd %xmm4,%xmm4
.byte 196,227,85,24,252,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm7
- .byte 196,226,125,24,37,73,88,2,0 // vbroadcastss 0x25849(%rip),%ymm4 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,24,37,197,74,2,0 // vbroadcastss 0x24ac5(%rip),%ymm4 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,196,84,228 // vandps %ymm4,%ymm7,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,226,125,24,45,60,88,2,0 // vbroadcastss 0x2583c(%rip),%ymm5 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,45,184,74,2,0 // vbroadcastss 0x24ab8(%rip),%ymm5 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,220,89,229 // vmulps %ymm5,%ymm4,%ymm4
- .byte 196,226,125,24,45,51,88,2,0 // vbroadcastss 0x25833(%rip),%ymm5 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,24,45,175,74,2,0 // vbroadcastss 0x24aaf(%rip),%ymm5 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,196,84,237 // vandps %ymm5,%ymm7,%ymm5
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,226,125,24,53,38,88,2,0 // vbroadcastss 0x25826(%rip),%ymm6 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,53,162,74,2,0 // vbroadcastss 0x24aa2(%rip),%ymm6 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,212,89,238 // vmulps %ymm6,%ymm5,%ymm5
- .byte 196,226,125,24,53,29,88,2,0 // vbroadcastss 0x2581d(%rip),%ymm6 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,24,53,153,74,2,0 // vbroadcastss 0x24a99(%rip),%ymm6 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,196,84,246 // vandps %ymm6,%ymm7,%ymm6
.byte 197,252,91,246 // vcvtdq2ps %ymm6,%ymm6
- .byte 196,98,125,24,5,16,88,2,0 // vbroadcastss 0x25810(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,140,74,2,0 // vbroadcastss 0x24a8c(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,76,89,240 // vmulps %ymm8,%ymm6,%ymm6
- .byte 196,98,125,24,5,6,88,2,0 // vbroadcastss 0x25806(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,24,5,130,74,2,0 // vbroadcastss 0x24a82(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,68,84,248 // vandps %ymm8,%ymm7,%ymm7
.byte 197,252,91,255 // vcvtdq2ps %ymm7,%ymm7
- .byte 196,98,125,24,5,248,87,2,0 // vbroadcastss 0x257f8(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,116,74,2,0 // vbroadcastss 0x24a74(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,68,89,248 // vmulps %ymm8,%ymm7,%ymm7
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -29168,25 +29168,25 @@ _sk_gather_4444_avx:
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm3
- .byte 196,226,125,24,5,1,86,2,0 // vbroadcastss 0x25601(%rip),%ymm0 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
+ .byte 196,226,125,24,5,125,72,2,0 // vbroadcastss 0x2487d(%rip),%ymm0 # 38aac <_sk_srcover_bgra_8888_sse2_lowp+0x3a8>
.byte 197,228,84,192 // vandps %ymm0,%ymm3,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,226,125,24,13,244,85,2,0 // vbroadcastss 0x255f4(%rip),%ymm1 # 39834 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
+ .byte 196,226,125,24,13,112,72,2,0 // vbroadcastss 0x24870(%rip),%ymm1 # 38ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ac>
.byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,24,13,235,85,2,0 // vbroadcastss 0x255eb(%rip),%ymm1 # 39838 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
+ .byte 196,226,125,24,13,103,72,2,0 // vbroadcastss 0x24867(%rip),%ymm1 # 38ab4 <_sk_srcover_bgra_8888_sse2_lowp+0x3b0>
.byte 197,228,84,201 // vandps %ymm1,%ymm3,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,226,125,24,21,222,85,2,0 // vbroadcastss 0x255de(%rip),%ymm2 # 3983c <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
+ .byte 196,226,125,24,21,90,72,2,0 // vbroadcastss 0x2485a(%rip),%ymm2 # 38ab8 <_sk_srcover_bgra_8888_sse2_lowp+0x3b4>
.byte 197,244,89,202 // vmulps %ymm2,%ymm1,%ymm1
- .byte 196,226,125,24,21,213,85,2,0 // vbroadcastss 0x255d5(%rip),%ymm2 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
+ .byte 196,226,125,24,21,81,72,2,0 // vbroadcastss 0x24851(%rip),%ymm2 # 38abc <_sk_srcover_bgra_8888_sse2_lowp+0x3b8>
.byte 197,228,84,210 // vandps %ymm2,%ymm3,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,98,125,24,5,200,85,2,0 // vbroadcastss 0x255c8(%rip),%ymm8 # 39844 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
+ .byte 196,98,125,24,5,68,72,2,0 // vbroadcastss 0x24844(%rip),%ymm8 # 38ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x3bc>
.byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
- .byte 196,98,125,24,5,190,85,2,0 // vbroadcastss 0x255be(%rip),%ymm8 # 39848 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
+ .byte 196,98,125,24,5,58,72,2,0 // vbroadcastss 0x2483a(%rip),%ymm8 # 38ac4 <_sk_srcover_bgra_8888_sse2_lowp+0x3c0>
.byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,5,176,85,2,0 // vbroadcastss 0x255b0(%rip),%ymm8 # 3984c <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
+ .byte 196,98,125,24,5,44,72,2,0 // vbroadcastss 0x2482c(%rip),%ymm8 # 38ac8 <_sk_srcover_bgra_8888_sse2_lowp+0x3c4>
.byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
@@ -29207,7 +29207,7 @@ _sk_store_4444_avx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,132,85,2,0 // vbroadcastss 0x25584(%rip),%ymm8 # 39850 <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
+ .byte 196,98,125,24,5,0,72,2,0 // vbroadcastss 0x24800(%rip),%ymm8 # 38acc <_sk_srcover_bgra_8888_sse2_lowp+0x3c8>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,193,41,114,241,12 // vpslld $0xc,%xmm9,%xmm10
@@ -29295,10 +29295,10 @@ _sk_load_8888_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,135,0,0,0 // jne 14487 <_sk_load_8888_avx+0xa7>
.byte 196,1,125,16,12,152 // vmovupd (%r8,%r11,4),%ymm9
- .byte 197,125,40,21,210,88,2,0 // vmovapd 0x258d2(%rip),%ymm10 # 39ce0 <_sk_srcover_bgra_8888_sse2_lowp+0x858>
+ .byte 197,125,40,21,82,75,2,0 // vmovapd 0x24b52(%rip),%ymm10 # 38f60 <_sk_srcover_bgra_8888_sse2_lowp+0x85c>
.byte 196,193,53,84,194 // vandpd %ymm10,%ymm9,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,5,144,83,2,0 // vbroadcastss 0x25390(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,12,70,2,0 // vbroadcastss 0x2460c(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 196,193,113,114,209,8 // vpsrld $0x8,%xmm9,%xmm1
.byte 196,99,125,25,203,1 // vextractf128 $0x1,%ymm9,%xmm3
@@ -29390,10 +29390,10 @@ _sk_load_8888_dst_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,135,0,0,0 // jne 145ff <_sk_load_8888_dst_avx+0xa7>
.byte 196,1,125,16,12,152 // vmovupd (%r8,%r11,4),%ymm9
- .byte 197,125,40,21,122,87,2,0 // vmovapd 0x2577a(%rip),%ymm10 # 39d00 <_sk_srcover_bgra_8888_sse2_lowp+0x878>
+ .byte 197,125,40,21,250,73,2,0 // vmovapd 0x249fa(%rip),%ymm10 # 38f80 <_sk_srcover_bgra_8888_sse2_lowp+0x87c>
.byte 196,193,53,84,226 // vandpd %ymm10,%ymm9,%ymm4
.byte 197,252,91,228 // vcvtdq2ps %ymm4,%ymm4
- .byte 196,98,125,24,5,24,82,2,0 // vbroadcastss 0x25218(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,148,68,2,0 // vbroadcastss 0x24494(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,92,89,224 // vmulps %ymm8,%ymm4,%ymm4
.byte 196,193,81,114,209,8 // vpsrld $0x8,%xmm9,%xmm5
.byte 196,99,125,25,207,1 // vextractf128 $0x1,%ymm9,%xmm7
@@ -29528,10 +29528,10 @@ _sk_gather_8888_avx:
.byte 196,131,121,34,4,176,2 // vpinsrd $0x2,(%r8,%r14,4),%xmm0,%xmm0
.byte 196,195,121,34,28,152,3 // vpinsrd $0x3,(%r8,%rbx,4),%xmm0,%xmm3
.byte 196,227,61,24,195,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm0
- .byte 197,124,40,21,76,85,2,0 // vmovaps 0x2554c(%rip),%ymm10 # 39d20 <_sk_srcover_bgra_8888_sse2_lowp+0x898>
+ .byte 197,124,40,21,204,71,2,0 // vmovaps 0x247cc(%rip),%ymm10 # 38fa0 <_sk_srcover_bgra_8888_sse2_lowp+0x89c>
.byte 196,193,124,84,194 // vandps %ymm10,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,202,79,2,0 // vbroadcastss 0x24fca(%rip),%ymm9 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,13,70,66,2,0 // vbroadcastss 0x24246(%rip),%ymm9 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 196,193,113,114,208,8 // vpsrld $0x8,%xmm8,%xmm1
.byte 197,233,114,211,8 // vpsrld $0x8,%xmm3,%xmm2
@@ -29568,7 +29568,7 @@ _sk_store_8888_avx:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,246,78,2,0 // vbroadcastss 0x24ef6(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,114,65,2,0 // vbroadcastss 0x24172(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -29631,7 +29631,7 @@ _sk_store_8888_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,218 // callq ffffffffdb014980 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdafdb4f8>
+ .byte 232,255,255,255,218 // callq ffffffffdb014980 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdafdc27c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,204 // dec %esp
@@ -29653,10 +29653,10 @@ _sk_load_bgra_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,135,0,0,0 // jne 14a2f <_sk_load_bgra_avx+0xa7>
.byte 196,1,125,16,12,152 // vmovupd (%r8,%r11,4),%ymm9
- .byte 197,125,40,21,138,83,2,0 // vmovapd 0x2538a(%rip),%ymm10 # 39d40 <_sk_srcover_bgra_8888_sse2_lowp+0x8b8>
+ .byte 197,125,40,21,10,70,2,0 // vmovapd 0x2460a(%rip),%ymm10 # 38fc0 <_sk_srcover_bgra_8888_sse2_lowp+0x8bc>
.byte 196,193,53,84,202 // vandpd %ymm10,%ymm9,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
- .byte 196,98,125,24,5,232,77,2,0 // vbroadcastss 0x24de8(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,100,64,2,0 // vbroadcastss 0x24064(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,116,89,208 // vmulps %ymm8,%ymm1,%ymm2
.byte 196,193,113,114,209,8 // vpsrld $0x8,%xmm9,%xmm1
.byte 196,99,125,25,203,1 // vextractf128 $0x1,%ymm9,%xmm3
@@ -29748,10 +29748,10 @@ _sk_load_bgra_dst_avx:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,135,0,0,0 // jne 14ba7 <_sk_load_bgra_dst_avx+0xa7>
.byte 196,1,125,16,12,152 // vmovupd (%r8,%r11,4),%ymm9
- .byte 197,125,40,21,50,82,2,0 // vmovapd 0x25232(%rip),%ymm10 # 39d60 <_sk_srcover_bgra_8888_sse2_lowp+0x8d8>
+ .byte 197,125,40,21,178,68,2,0 // vmovapd 0x244b2(%rip),%ymm10 # 38fe0 <_sk_srcover_bgra_8888_sse2_lowp+0x8dc>
.byte 196,193,53,84,234 // vandpd %ymm10,%ymm9,%ymm5
.byte 197,252,91,237 // vcvtdq2ps %ymm5,%ymm5
- .byte 196,98,125,24,5,112,76,2,0 // vbroadcastss 0x24c70(%rip),%ymm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,5,236,62,2,0 // vbroadcastss 0x23eec(%rip),%ymm8 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,84,89,240 // vmulps %ymm8,%ymm5,%ymm6
.byte 196,193,81,114,209,8 // vpsrld $0x8,%xmm9,%xmm5
.byte 196,99,125,25,207,1 // vextractf128 $0x1,%ymm9,%xmm7
@@ -29886,10 +29886,10 @@ _sk_gather_bgra_avx:
.byte 196,131,121,34,4,176,2 // vpinsrd $0x2,(%r8,%r14,4),%xmm0,%xmm0
.byte 196,195,121,34,28,152,3 // vpinsrd $0x3,(%r8,%rbx,4),%xmm0,%xmm3
.byte 196,227,61,24,195,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm0
- .byte 197,124,40,13,4,80,2,0 // vmovaps 0x25004(%rip),%ymm9 # 39d80 <_sk_srcover_bgra_8888_sse2_lowp+0x8f8>
+ .byte 197,124,40,13,132,66,2,0 // vmovaps 0x24284(%rip),%ymm9 # 39000 <_sk_srcover_bgra_8888_sse2_lowp+0x8fc>
.byte 196,193,124,84,193 // vandps %ymm9,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,21,34,74,2,0 // vbroadcastss 0x24a22(%rip),%ymm10 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x328>
+ .byte 196,98,125,24,21,158,60,2,0 // vbroadcastss 0x23c9e(%rip),%ymm10 # 38a2c <_sk_srcover_bgra_8888_sse2_lowp+0x328>
.byte 196,193,124,89,210 // vmulps %ymm10,%ymm0,%ymm2
.byte 196,193,121,114,208,8 // vpsrld $0x8,%xmm8,%xmm0
.byte 197,241,114,211,8 // vpsrld $0x8,%xmm3,%xmm1
@@ -29926,7 +29926,7 @@ _sk_store_bgra_avx:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,218 // movslq %edx,%r11
- .byte 196,98,125,24,5,78,73,2,0 // vbroadcastss 0x2494e(%rip),%ymm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,5,202,59,2,0 // vbroadcastss 0x23bca(%rip),%ymm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 196,65,108,89,200 // vmulps %ymm8,%ymm2,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,65,116,89,208 // vmulps %ymm8,%ymm1,%ymm10
@@ -29989,7 +29989,7 @@ _sk_store_bgra_avx:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,255,255,255,218 // callq ffffffffdb014f28 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdafdbaa0>
+ .byte 232,255,255,255,218 // callq ffffffffdb014f28 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffdafdc824>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,204 // dec %esp
@@ -30031,14 +30031,14 @@ _sk_load_f16_avx:
.byte 196,226,121,51,200 // vpmovzxwd %xmm0,%xmm1
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
- .byte 197,123,18,21,227,88,2,0 // vmovddup 0x258e3(%rip),%xmm10 # 3a8a8 <_sk_srcover_bgra_8888_sse2_lowp+0x1420>
+ .byte 197,123,18,21,99,75,2,0 // vmovddup 0x24b63(%rip),%xmm10 # 39b28 <_sk_srcover_bgra_8888_sse2_lowp+0x1424>
.byte 196,193,121,219,218 // vpand %xmm10,%xmm0,%xmm3
- .byte 196,98,121,24,29,237,72,2,0 // vbroadcastss 0x248ed(%rip),%xmm11 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x438>
+ .byte 196,98,121,24,29,105,59,2,0 // vbroadcastss 0x23b69(%rip),%xmm11 # 38b3c <_sk_srcover_bgra_8888_sse2_lowp+0x438>
.byte 197,33,102,227 // vpcmpgtd %xmm3,%xmm11,%xmm12
.byte 196,193,113,219,210 // vpand %xmm10,%xmm1,%xmm2
.byte 197,33,102,234 // vpcmpgtd %xmm2,%xmm11,%xmm13
.byte 196,195,21,24,252,1 // vinsertf128 $0x1,%xmm12,%ymm13,%ymm7
- .byte 197,123,18,37,194,88,2,0 // vmovddup 0x258c2(%rip),%xmm12 # 3a8b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1428>
+ .byte 197,123,18,37,66,75,2,0 // vmovddup 0x24b42(%rip),%xmm12 # 39b30 <_sk_srcover_bgra_8888_sse2_lowp+0x142c>
.byte 196,193,113,219,204 // vpand %xmm12,%xmm1,%xmm1
.byte 197,241,114,241,16 // vpslld $0x10,%xmm1,%xmm1
.byte 196,193,121,219,196 // vpand %xmm12,%xmm0,%xmm0
@@ -30047,7 +30047,7 @@ _sk_load_f16_avx:
.byte 197,241,235,202 // vpor %xmm2,%xmm1,%xmm1
.byte 197,233,114,243,13 // vpslld $0xd,%xmm3,%xmm2
.byte 197,249,235,194 // vpor %xmm2,%xmm0,%xmm0
- .byte 196,226,121,24,29,167,72,2,0 // vbroadcastss 0x248a7(%rip),%xmm3 # 398c4 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
+ .byte 196,226,121,24,29,35,59,2,0 // vbroadcastss 0x23b23(%rip),%xmm3 # 38b40 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
.byte 197,249,254,195 // vpaddd %xmm3,%xmm0,%xmm0
.byte 197,241,254,203 // vpaddd %xmm3,%xmm1,%xmm1
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
@@ -30184,14 +30184,14 @@ _sk_load_f16_dst_avx:
.byte 196,226,121,51,236 // vpmovzxwd %xmm4,%xmm5
.byte 197,249,112,228,78 // vpshufd $0x4e,%xmm4,%xmm4
.byte 196,226,121,51,228 // vpmovzxwd %xmm4,%xmm4
- .byte 197,123,18,21,12,86,2,0 // vmovddup 0x2560c(%rip),%xmm10 # 3a8a8 <_sk_srcover_bgra_8888_sse2_lowp+0x1420>
+ .byte 197,123,18,21,140,72,2,0 // vmovddup 0x2488c(%rip),%xmm10 # 39b28 <_sk_srcover_bgra_8888_sse2_lowp+0x1424>
.byte 196,193,89,219,250 // vpand %xmm10,%xmm4,%xmm7
- .byte 196,98,121,24,29,22,70,2,0 // vbroadcastss 0x24616(%rip),%xmm11 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x438>
+ .byte 196,98,121,24,29,146,56,2,0 // vbroadcastss 0x23892(%rip),%xmm11 # 38b3c <_sk_srcover_bgra_8888_sse2_lowp+0x438>
.byte 197,33,102,231 // vpcmpgtd %xmm7,%xmm11,%xmm12
.byte 196,193,81,219,242 // vpand %xmm10,%xmm5,%xmm6
.byte 197,33,102,238 // vpcmpgtd %xmm6,%xmm11,%xmm13
.byte 196,195,21,24,220,1 // vinsertf128 $0x1,%xmm12,%ymm13,%ymm3
- .byte 197,123,18,37,235,85,2,0 // vmovddup 0x255eb(%rip),%xmm12 # 3a8b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1428>
+ .byte 197,123,18,37,107,72,2,0 // vmovddup 0x2486b(%rip),%xmm12 # 39b30 <_sk_srcover_bgra_8888_sse2_lowp+0x142c>
.byte 196,193,81,219,236 // vpand %xmm12,%xmm5,%xmm5
.byte 197,209,114,245,16 // vpslld $0x10,%xmm5,%xmm5
.byte 196,193,89,219,228 // vpand %xmm12,%xmm4,%xmm4
@@ -30200,7 +30200,7 @@ _sk_load_f16_dst_avx:
.byte 197,209,235,238 // vpor %xmm6,%xmm5,%xmm5
.byte 197,201,114,247,13 // vpslld $0xd,%xmm7,%xmm6
.byte 197,217,235,230 // vpor %xmm6,%xmm4,%xmm4
- .byte 196,226,121,24,61,208,69,2,0 // vbroadcastss 0x245d0(%rip),%xmm7 # 398c4 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
+ .byte 196,226,121,24,61,76,56,2,0 // vbroadcastss 0x2384c(%rip),%xmm7 # 38b40 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
.byte 197,217,254,231 // vpaddd %xmm7,%xmm4,%xmm4
.byte 197,209,254,239 // vpaddd %xmm7,%xmm5,%xmm5
.byte 196,227,85,24,228,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm4
@@ -30380,14 +30380,14 @@ _sk_gather_f16_avx:
.byte 196,226,121,51,208 // vpmovzxwd %xmm0,%xmm2
.byte 197,249,112,192,78 // vpshufd $0x4e,%xmm0,%xmm0
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
- .byte 197,123,18,29,114,82,2,0 // vmovddup 0x25272(%rip),%xmm11 # 3a8a8 <_sk_srcover_bgra_8888_sse2_lowp+0x1420>
+ .byte 197,123,18,29,242,68,2,0 // vmovddup 0x244f2(%rip),%xmm11 # 39b28 <_sk_srcover_bgra_8888_sse2_lowp+0x1424>
.byte 196,193,121,219,219 // vpand %xmm11,%xmm0,%xmm3
- .byte 196,98,121,24,37,124,66,2,0 // vbroadcastss 0x2427c(%rip),%xmm12 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x438>
+ .byte 196,98,121,24,37,248,52,2,0 // vbroadcastss 0x234f8(%rip),%xmm12 # 38b3c <_sk_srcover_bgra_8888_sse2_lowp+0x438>
.byte 197,25,102,235 // vpcmpgtd %xmm3,%xmm12,%xmm13
.byte 196,193,105,219,203 // vpand %xmm11,%xmm2,%xmm1
.byte 197,25,102,193 // vpcmpgtd %xmm1,%xmm12,%xmm8
.byte 196,67,61,24,197,1 // vinsertf128 $0x1,%xmm13,%ymm8,%ymm8
- .byte 197,123,18,45,81,82,2,0 // vmovddup 0x25251(%rip),%xmm13 # 3a8b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1428>
+ .byte 197,123,18,45,209,68,2,0 // vmovddup 0x244d1(%rip),%xmm13 # 39b30 <_sk_srcover_bgra_8888_sse2_lowp+0x142c>
.byte 196,193,105,219,213 // vpand %xmm13,%xmm2,%xmm2
.byte 197,233,114,242,16 // vpslld $0x10,%xmm2,%xmm2
.byte 197,241,114,241,13 // vpslld $0xd,%xmm1,%xmm1
@@ -30396,7 +30396,7 @@ _sk_gather_f16_avx:
.byte 197,249,114,240,16 // vpslld $0x10,%xmm0,%xmm0
.byte 197,233,114,243,13 // vpslld $0xd,%xmm3,%xmm2
.byte 197,249,235,194 // vpor %xmm2,%xmm0,%xmm0
- .byte 196,226,121,24,29,54,66,2,0 // vbroadcastss 0x24236(%rip),%xmm3 # 398c4 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
+ .byte 196,226,121,24,29,178,52,2,0 // vbroadcastss 0x234b2(%rip),%xmm3 # 38b40 <_sk_srcover_bgra_8888_sse2_lowp+0x43c>
.byte 197,249,254,195 // vpaddd %xmm3,%xmm0,%xmm0
.byte 197,241,254,203 // vpaddd %xmm3,%xmm1,%xmm1
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
@@ -30486,12 +30486,12 @@ _sk_store_f16_avx:
.byte 197,252,17,108,36,192 // vmovups %ymm5,-0x40(%rsp)
.byte 197,252,17,100,36,160 // vmovups %ymm4,-0x60(%rsp)
.byte 197,252,40,225 // vmovaps %ymm1,%ymm4
- .byte 196,98,125,24,5,88,64,2,0 // vbroadcastss 0x24058(%rip),%ymm8 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,5,212,50,2,0 // vbroadcastss 0x232d4(%rip),%ymm8 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,124,84,216 // vandps %ymm8,%ymm0,%ymm11
- .byte 196,98,125,24,21,146,64,2,0 // vbroadcastss 0x24092(%rip),%ymm10 # 398c8 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
+ .byte 196,98,125,24,21,14,51,2,0 // vbroadcastss 0x2330e(%rip),%ymm10 # 38b44 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
.byte 196,65,124,84,226 // vandps %ymm10,%ymm0,%ymm12
.byte 196,67,125,25,229,1 // vextractf128 $0x1,%ymm12,%xmm13
- .byte 196,98,121,24,13,130,64,2,0 // vbroadcastss 0x24082(%rip),%xmm9 # 398cc <_sk_srcover_bgra_8888_sse2_lowp+0x444>
+ .byte 196,98,121,24,13,254,50,2,0 // vbroadcastss 0x232fe(%rip),%xmm9 # 38b48 <_sk_srcover_bgra_8888_sse2_lowp+0x444>
.byte 196,65,49,102,245 // vpcmpgtd %xmm13,%xmm9,%xmm14
.byte 196,65,49,102,252 // vpcmpgtd %xmm12,%xmm9,%xmm15
.byte 196,67,5,24,246,1 // vinsertf128 $0x1,%xmm14,%ymm15,%ymm14
@@ -30502,7 +30502,7 @@ _sk_store_f16_avx:
.byte 196,65,17,254,239 // vpaddd %xmm15,%xmm13,%xmm13
.byte 196,193,25,114,212,13 // vpsrld $0xd,%xmm12,%xmm12
.byte 196,65,25,254,227 // vpaddd %xmm11,%xmm12,%xmm12
- .byte 196,98,121,24,29,69,64,2,0 // vbroadcastss 0x24045(%rip),%xmm11 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x448>
+ .byte 196,98,121,24,29,193,50,2,0 // vbroadcastss 0x232c1(%rip),%xmm11 # 38b4c <_sk_srcover_bgra_8888_sse2_lowp+0x448>
.byte 196,65,17,254,235 // vpaddd %xmm11,%xmm13,%xmm13
.byte 196,65,25,254,227 // vpaddd %xmm11,%xmm12,%xmm12
.byte 196,67,29,24,237,1 // vinsertf128 $0x1,%xmm13,%ymm12,%ymm13
@@ -30653,7 +30653,7 @@ _sk_load_u16_be_avx:
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,132,60,2,0 // vbroadcastss 0x23c84(%rip),%ymm9 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,13,0,47,2,0 // vbroadcastss 0x22f00(%rip),%ymm9 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,169,109,202 // vpunpckhqdq %xmm2,%xmm10,%xmm1
.byte 197,233,113,241,8 // vpsllw $0x8,%xmm1,%xmm2
@@ -30757,7 +30757,7 @@ _sk_load_rgb_u16_be_avx:
.byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
.byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
- .byte 196,98,125,24,13,179,58,2,0 // vbroadcastss 0x23ab3(%rip),%ymm9 # 397cc <_sk_srcover_bgra_8888_sse2_lowp+0x344>
+ .byte 196,98,125,24,13,47,45,2,0 // vbroadcastss 0x22d2f(%rip),%ymm9 # 38a48 <_sk_srcover_bgra_8888_sse2_lowp+0x344>
.byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
.byte 197,185,109,203 // vpunpckhqdq %xmm3,%xmm8,%xmm1
.byte 197,225,113,241,8 // vpsllw $0x8,%xmm1,%xmm3
@@ -30779,7 +30779,7 @@ _sk_load_rgb_u16_be_avx:
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 196,193,108,89,209 // vmulps %ymm9,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,179,57,2,0 // vbroadcastss 0x239b3(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,47,44,2,0 // vbroadcastss 0x22c2f(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 255,224 // jmpq *%rax
.byte 196,193,121,110,4,64 // vmovd (%r8,%rax,2),%xmm0
.byte 196,65,49,239,201 // vpxor %xmm9,%xmm9,%xmm9
@@ -30855,7 +30855,7 @@ _sk_store_u16_be_avx:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 73,99,193 // movslq %r9d,%rax
- .byte 196,98,125,24,5,124,57,2,0 // vbroadcastss 0x2397c(%rip),%ymm8 # 39854 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
+ .byte 196,98,125,24,5,248,43,2,0 // vbroadcastss 0x22bf8(%rip),%ymm8 # 38ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x3cc>
.byte 196,65,124,89,200 // vmulps %ymm8,%ymm0,%ymm9
.byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
.byte 196,67,125,25,202,1 // vextractf128 $0x1,%ymm9,%xmm10
@@ -31014,7 +31014,7 @@ _sk_load_f32_avx:
.byte 255,92,255,255 // lcall *-0x1(%rdi,%rdi,8)
.byte 255,70,255 // incl -0x1(%rsi)
.byte 255 // (bad)
- .byte 255,53,255,255,255,40 // pushq 0x28ffffff(%rip) # 290161a8 <_sk_srcover_bgra_8888_sse2_lowp+0x28fdcd20>
+ .byte 255,53,255,255,255,40 // pushq 0x28ffffff(%rip) # 290161a8 <_sk_srcover_bgra_8888_sse2_lowp+0x28fddaa4>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // .byte 0xff
@@ -31213,7 +31213,7 @@ _sk_mirror_x_avx:
.byte 196,193,58,88,192 // vaddss %xmm8,%xmm8,%xmm0
.byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
.byte 196,99,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm8
- .byte 197,178,89,5,153,50,2,0 // vmulss 0x23299(%rip),%xmm9,%xmm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,5,21,37,2,0 // vmulss 0x22515(%rip),%xmm9,%xmm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
.byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
.byte 197,164,89,192 // vmulps %ymm0,%ymm11,%ymm0
@@ -31240,7 +31240,7 @@ _sk_mirror_y_avx:
.byte 196,193,58,88,200 // vaddss %xmm8,%xmm8,%xmm1
.byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
.byte 196,99,117,24,193,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm8
- .byte 197,178,89,13,47,50,2,0 // vmulss 0x2322f(%rip),%xmm9,%xmm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,178,89,13,171,36,2,0 // vmulss 0x224ab(%rip),%xmm9,%xmm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
.byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
.byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
@@ -31260,7 +31260,7 @@ FUNCTION(_sk_clamp_x_1_avx)
_sk_clamp_x_1_avx:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,236,49,2,0 // vbroadcastss 0x231ec(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,104,36,2,0 // vbroadcastss 0x22468(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -31273,7 +31273,7 @@ _sk_repeat_x_1_avx:
.byte 196,193,124,92,192 // vsubps %ymm8,%ymm0,%ymm0
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,198,49,2,0 // vbroadcastss 0x231c6(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,66,36,2,0 // vbroadcastss 0x22442(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -31282,9 +31282,9 @@ HIDDEN _sk_mirror_x_1_avx
.globl _sk_mirror_x_1_avx
FUNCTION(_sk_mirror_x_1_avx)
_sk_mirror_x_1_avx:
- .byte 196,98,125,24,5,204,49,2,0 // vbroadcastss 0x231cc(%rip),%ymm8 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,5,72,36,2,0 // vbroadcastss 0x22448(%rip),%ymm8 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,13,162,49,2,0 // vbroadcastss 0x231a2(%rip),%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,13,30,36,2,0 // vbroadcastss 0x2241e(%rip),%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,124,89,201 // vmulps %ymm9,%ymm0,%ymm9
.byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
.byte 196,65,52,88,201 // vaddps %ymm9,%ymm9,%ymm9
@@ -31294,7 +31294,7 @@ _sk_mirror_x_1_avx:
.byte 197,60,92,200 // vsubps %ymm0,%ymm8,%ymm9
.byte 197,180,84,192 // vandps %ymm0,%ymm9,%ymm0
.byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
- .byte 196,98,125,24,5,114,49,2,0 // vbroadcastss 0x23172(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,238,35,2,0 // vbroadcastss 0x223ee(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -31303,12 +31303,12 @@ HIDDEN _sk_luminance_to_alpha_avx
.globl _sk_luminance_to_alpha_avx
FUNCTION(_sk_luminance_to_alpha_avx)
_sk_luminance_to_alpha_avx:
- .byte 196,226,125,24,29,132,50,2,0 // vbroadcastss 0x23284(%rip),%ymm3 # 3985c <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
+ .byte 196,226,125,24,29,0,37,2,0 // vbroadcastss 0x22500(%rip),%ymm3 # 38ad8 <_sk_srcover_bgra_8888_sse2_lowp+0x3d4>
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
- .byte 196,226,125,24,29,115,50,2,0 // vbroadcastss 0x23273(%rip),%ymm3 # 39858 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
+ .byte 196,226,125,24,29,239,36,2,0 // vbroadcastss 0x224ef(%rip),%ymm3 # 38ad4 <_sk_srcover_bgra_8888_sse2_lowp+0x3d0>
.byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
- .byte 196,226,125,24,13,106,50,2,0 // vbroadcastss 0x2326a(%rip),%ymm1 # 39860 <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
+ .byte 196,226,125,24,13,230,36,2,0 // vbroadcastss 0x224e6(%rip),%ymm1 # 38adc <_sk_srcover_bgra_8888_sse2_lowp+0x3d8>
.byte 197,236,89,201 // vmulps %ymm1,%ymm2,%ymm1
.byte 197,252,88,217 // vaddps %ymm1,%ymm0,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -31731,7 +31731,7 @@ _sk_gradient_avx:
.byte 73,255,200 // dec %r8
.byte 72,131,195,4 // add $0x4,%rbx
.byte 196,65,52,87,201 // vxorps %ymm9,%ymm9,%ymm9
- .byte 196,98,125,24,21,43,42,2,0 // vbroadcastss 0x22a2b(%rip),%ymm10 # 3973c <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
+ .byte 196,98,125,24,21,167,28,2,0 // vbroadcastss 0x21ca7(%rip),%ymm10 # 389b8 <_sk_srcover_bgra_8888_sse2_lowp+0x2b4>
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 196,98,125,24,3 // vbroadcastss (%rbx),%ymm8
.byte 197,60,194,192,2 // vcmpleps %ymm0,%ymm8,%ymm8
@@ -31925,27 +31925,27 @@ _sk_xy_to_unit_angle_avx:
.byte 196,65,52,95,226 // vmaxps %ymm10,%ymm9,%ymm12
.byte 196,65,36,94,220 // vdivps %ymm12,%ymm11,%ymm11
.byte 196,65,36,89,227 // vmulps %ymm11,%ymm11,%ymm12
- .byte 196,98,125,24,45,69,39,2,0 // vbroadcastss 0x22745(%rip),%ymm13 # 39864 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
+ .byte 196,98,125,24,45,193,25,2,0 // vbroadcastss 0x219c1(%rip),%ymm13 # 38ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
.byte 196,65,28,89,237 // vmulps %ymm13,%ymm12,%ymm13
- .byte 196,98,125,24,53,59,39,2,0 // vbroadcastss 0x2273b(%rip),%ymm14 # 39868 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
+ .byte 196,98,125,24,53,183,25,2,0 // vbroadcastss 0x219b7(%rip),%ymm14 # 38ae4 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
.byte 196,65,20,88,238 // vaddps %ymm14,%ymm13,%ymm13
.byte 196,65,28,89,237 // vmulps %ymm13,%ymm12,%ymm13
- .byte 196,98,125,24,53,44,39,2,0 // vbroadcastss 0x2272c(%rip),%ymm14 # 3986c <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
+ .byte 196,98,125,24,53,168,25,2,0 // vbroadcastss 0x219a8(%rip),%ymm14 # 38ae8 <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
.byte 196,65,20,88,238 // vaddps %ymm14,%ymm13,%ymm13
.byte 196,65,28,89,229 // vmulps %ymm13,%ymm12,%ymm12
- .byte 196,98,125,24,45,29,39,2,0 // vbroadcastss 0x2271d(%rip),%ymm13 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
+ .byte 196,98,125,24,45,153,25,2,0 // vbroadcastss 0x21999(%rip),%ymm13 # 38aec <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
.byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
.byte 196,65,36,89,220 // vmulps %ymm12,%ymm11,%ymm11
.byte 196,65,52,194,202,1 // vcmpltps %ymm10,%ymm9,%ymm9
- .byte 196,98,125,24,21,8,39,2,0 // vbroadcastss 0x22708(%rip),%ymm10 # 39874 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
+ .byte 196,98,125,24,21,132,25,2,0 // vbroadcastss 0x21984(%rip),%ymm10 # 38af0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
.byte 196,65,44,92,211 // vsubps %ymm11,%ymm10,%ymm10
.byte 196,67,37,74,202,144 // vblendvps %ymm9,%ymm10,%ymm11,%ymm9
.byte 196,193,124,194,192,1 // vcmpltps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,21,174,37,2,0 // vbroadcastss 0x225ae(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,42,24,2,0 // vbroadcastss 0x2182a(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,44,92,209 // vsubps %ymm9,%ymm10,%ymm10
.byte 196,195,53,74,194,0 // vblendvps %ymm0,%ymm10,%ymm9,%ymm0
.byte 196,65,116,194,200,1 // vcmpltps %ymm8,%ymm1,%ymm9
- .byte 196,98,125,24,21,152,37,2,0 // vbroadcastss 0x22598(%rip),%ymm10 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,21,20,24,2,0 // vbroadcastss 0x21814(%rip),%ymm10 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,44,92,208 // vsubps %ymm0,%ymm10,%ymm10
.byte 196,195,125,74,194,144 // vblendvps %ymm9,%ymm10,%ymm0,%ymm0
.byte 196,65,124,194,200,3 // vcmpunordps %ymm8,%ymm0,%ymm9
@@ -31975,7 +31975,7 @@ _sk_xy_to_2pt_conical_quadratic_max_avx:
.byte 196,67,121,4,210,0 // vpermilps $0x0,%xmm10,%xmm10
.byte 196,67,45,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 196,98,125,24,29,126,38,2,0 // vbroadcastss 0x2267e(%rip),%ymm11 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,29,250,24,2,0 // vbroadcastss 0x218fa(%rip),%ymm11 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
.byte 197,252,89,192 // vmulps %ymm0,%ymm0,%ymm0
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
@@ -31984,17 +31984,17 @@ _sk_xy_to_2pt_conical_quadratic_max_avx:
.byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
.byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 196,98,125,24,13,82,38,2,0 // vbroadcastss 0x22652(%rip),%ymm9 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 196,98,125,24,13,206,24,2,0 // vbroadcastss 0x218ce(%rip),%ymm9 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,65,44,89,194 // vmulps %ymm10,%ymm10,%ymm8
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 196,98,125,24,13,48,38,2,0 // vbroadcastss 0x22630(%rip),%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,13,172,24,2,0 // vbroadcastss 0x218ac(%rip),%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,44,87,201 // vxorps %ymm9,%ymm10,%ymm9
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,209,36,2,0 // vbroadcastss 0x224d1(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,77,23,2,0 // vbroadcastss 0x2174d(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 196,65,60,89,210 // vmulps %ymm10,%ymm8,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
@@ -32014,7 +32014,7 @@ _sk_xy_to_2pt_conical_quadratic_min_avx:
.byte 196,67,121,4,210,0 // vpermilps $0x0,%xmm10,%xmm10
.byte 196,67,45,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
.byte 197,44,88,208 // vaddps %ymm0,%ymm10,%ymm10
- .byte 196,98,125,24,29,208,37,2,0 // vbroadcastss 0x225d0(%rip),%ymm11 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,29,76,24,2,0 // vbroadcastss 0x2184c(%rip),%ymm11 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,44,89,211 // vmulps %ymm11,%ymm10,%ymm10
.byte 197,252,89,192 // vmulps %ymm0,%ymm0,%ymm0
.byte 197,116,89,217 // vmulps %ymm1,%ymm1,%ymm11
@@ -32023,17 +32023,17 @@ _sk_xy_to_2pt_conical_quadratic_min_avx:
.byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
.byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
.byte 197,164,92,192 // vsubps %ymm0,%ymm11,%ymm0
- .byte 196,98,125,24,13,164,37,2,0 // vbroadcastss 0x225a4(%rip),%ymm9 # 3987c <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
+ .byte 196,98,125,24,13,32,24,2,0 // vbroadcastss 0x21820(%rip),%ymm9 # 38af8 <_sk_srcover_bgra_8888_sse2_lowp+0x3f4>
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 196,65,44,89,194 // vmulps %ymm10,%ymm10,%ymm8
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
.byte 197,252,81,192 // vsqrtps %ymm0,%ymm0
.byte 196,98,125,24,64,68 // vbroadcastss 0x44(%rax),%ymm8
- .byte 196,98,125,24,13,130,37,2,0 // vbroadcastss 0x22582(%rip),%ymm9 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,13,254,23,2,0 // vbroadcastss 0x217fe(%rip),%ymm9 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,65,44,87,201 // vxorps %ymm9,%ymm10,%ymm9
.byte 196,65,124,92,210 // vsubps %ymm10,%ymm0,%ymm10
- .byte 196,98,125,24,29,35,36,2,0 // vbroadcastss 0x22423(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,29,159,22,2,0 // vbroadcastss 0x2169f(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,60,89,195 // vmulps %ymm11,%ymm8,%ymm8
.byte 196,65,60,89,210 // vmulps %ymm10,%ymm8,%ymm10
.byte 197,180,92,192 // vsubps %ymm0,%ymm9,%ymm0
@@ -32052,7 +32052,7 @@ _sk_xy_to_2pt_conical_linear_avx:
.byte 196,67,121,4,201,0 // vpermilps $0x0,%xmm9,%xmm9
.byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
.byte 197,52,88,200 // vaddps %ymm0,%ymm9,%ymm9
- .byte 196,98,125,24,21,40,37,2,0 // vbroadcastss 0x22528(%rip),%ymm10 # 39878 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
+ .byte 196,98,125,24,21,164,23,2,0 // vbroadcastss 0x217a4(%rip),%ymm10 # 38af4 <_sk_srcover_bgra_8888_sse2_lowp+0x3f0>
.byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
.byte 197,252,89,192 // vmulps %ymm0,%ymm0,%ymm0
.byte 197,116,89,209 // vmulps %ymm1,%ymm1,%ymm10
@@ -32061,7 +32061,7 @@ _sk_xy_to_2pt_conical_linear_avx:
.byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
.byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
.byte 197,172,92,192 // vsubps %ymm0,%ymm10,%ymm0
- .byte 196,98,125,24,5,0,37,2,0 // vbroadcastss 0x22500(%rip),%ymm8 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
+ .byte 196,98,125,24,5,124,23,2,0 // vbroadcastss 0x2177c(%rip),%ymm8 # 38afc <_sk_srcover_bgra_8888_sse2_lowp+0x3f8>
.byte 196,193,124,87,192 // vxorps %ymm8,%ymm0,%ymm0
.byte 196,193,124,94,193 // vdivps %ymm9,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32105,7 +32105,7 @@ HIDDEN _sk_save_xy_avx
FUNCTION(_sk_save_xy_avx)
_sk_save_xy_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,58,35,2,0 // vbroadcastss 0x2233a(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,182,21,2,0 // vbroadcastss 0x215b6(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,124,88,200 // vaddps %ymm8,%ymm0,%ymm9
.byte 196,67,125,8,209,1 // vroundps $0x1,%ymm9,%ymm10
.byte 196,65,52,92,202 // vsubps %ymm10,%ymm9,%ymm9
@@ -32142,9 +32142,9 @@ HIDDEN _sk_bilinear_nx_avx
FUNCTION(_sk_bilinear_nx_avx)
_sk_bilinear_nx_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,12,36,2,0 // vbroadcastss 0x2240c(%rip),%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,5,136,22,2,0 // vbroadcastss 0x21688(%rip),%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,179,34,2,0 // vbroadcastss 0x222b3(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,47,21,2,0 // vbroadcastss 0x2152f(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32155,7 +32155,7 @@ HIDDEN _sk_bilinear_px_avx
FUNCTION(_sk_bilinear_px_avx)
_sk_bilinear_px_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,144,34,2,0 // vbroadcastss 0x22290(%rip),%ymm0 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,5,12,21,2,0 // vbroadcastss 0x2150c(%rip),%ymm0 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -32167,9 +32167,9 @@ HIDDEN _sk_bilinear_ny_avx
FUNCTION(_sk_bilinear_ny_avx)
_sk_bilinear_ny_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,189,35,2,0 // vbroadcastss 0x223bd(%rip),%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,13,57,22,2,0 // vbroadcastss 0x21639(%rip),%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,99,34,2,0 // vbroadcastss 0x22263(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,223,20,2,0 // vbroadcastss 0x214df(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32180,7 +32180,7 @@ HIDDEN _sk_bilinear_py_avx
FUNCTION(_sk_bilinear_py_avx)
_sk_bilinear_py_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,64,34,2,0 // vbroadcastss 0x22240(%rip),%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,13,188,20,2,0 // vbroadcastss 0x214bc(%rip),%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -32192,14 +32192,14 @@ HIDDEN _sk_bicubic_n3x_avx
FUNCTION(_sk_bicubic_n3x_avx)
_sk_bicubic_n3x_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,112,35,2,0 // vbroadcastss 0x22370(%rip),%ymm0 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,226,125,24,5,236,21,2,0 // vbroadcastss 0x215ec(%rip),%ymm0 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,19,34,2,0 // vbroadcastss 0x22213(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,143,20,2,0 // vbroadcastss 0x2148f(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,81,35,2,0 // vbroadcastss 0x22351(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,205,21,2,0 // vbroadcastss 0x215cd(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,65,60,89,194 // vmulps %ymm10,%ymm8,%ymm8
- .byte 196,98,125,24,21,99,34,2,0 // vbroadcastss 0x22263(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,223,20,2,0 // vbroadcastss 0x214df(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,60,88,194 // vaddps %ymm10,%ymm8,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -32211,19 +32211,19 @@ HIDDEN _sk_bicubic_n1x_avx
FUNCTION(_sk_bicubic_n1x_avx)
_sk_bicubic_n1x_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,26,35,2,0 // vbroadcastss 0x2231a(%rip),%ymm0 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,5,150,21,2,0 // vbroadcastss 0x21596(%rip),%ymm0 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
- .byte 196,98,125,24,5,193,33,2,0 // vbroadcastss 0x221c1(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,61,20,2,0 // vbroadcastss 0x2143d(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,128,0,0,0 // vsubps 0x80(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,8,35,2,0 // vbroadcastss 0x22308(%rip),%ymm9 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,13,132,21,2,0 // vbroadcastss 0x21584(%rip),%ymm9 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,65,60,89,201 // vmulps %ymm9,%ymm8,%ymm9
- .byte 196,98,125,24,21,254,34,2,0 // vbroadcastss 0x222fe(%rip),%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,122,21,2,0 // vbroadcastss 0x2157a(%rip),%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
.byte 196,65,60,89,201 // vmulps %ymm9,%ymm8,%ymm9
- .byte 196,98,125,24,21,139,33,2,0 // vbroadcastss 0x2218b(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,7,20,2,0 // vbroadcastss 0x21407(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
- .byte 196,98,125,24,13,220,34,2,0 // vbroadcastss 0x222dc(%rip),%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,88,21,2,0 // vbroadcastss 0x21558(%rip),%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32234,17 +32234,17 @@ HIDDEN _sk_bicubic_p1x_avx
FUNCTION(_sk_bicubic_p1x_avx)
_sk_bicubic_p1x_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,92,33,2,0 // vbroadcastss 0x2215c(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,216,19,2,0 // vbroadcastss 0x213d8(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,0 // vaddps (%rax),%ymm8,%ymm0
.byte 197,124,16,136,128,0,0,0 // vmovups 0x80(%rax),%ymm9
- .byte 196,98,125,24,21,163,34,2,0 // vbroadcastss 0x222a3(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,21,31,21,2,0 // vbroadcastss 0x2151f(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,65,52,89,210 // vmulps %ymm10,%ymm9,%ymm10
- .byte 196,98,125,24,29,153,34,2,0 // vbroadcastss 0x22299(%rip),%ymm11 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,29,21,21,2,0 // vbroadcastss 0x21515(%rip),%ymm11 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
.byte 196,65,52,89,210 // vmulps %ymm10,%ymm9,%ymm10
.byte 196,65,44,88,192 // vaddps %ymm8,%ymm10,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
- .byte 196,98,125,24,13,128,34,2,0 // vbroadcastss 0x22280(%rip),%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,252,20,2,0 // vbroadcastss 0x214fc(%rip),%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32255,13 +32255,13 @@ HIDDEN _sk_bicubic_p3x_avx
FUNCTION(_sk_bicubic_p3x_avx)
_sk_bicubic_p3x_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,5,96,34,2,0 // vbroadcastss 0x22260(%rip),%ymm0 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,226,125,24,5,220,20,2,0 // vbroadcastss 0x214dc(%rip),%ymm0 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 197,252,88,0 // vaddps (%rax),%ymm0,%ymm0
.byte 197,124,16,128,128,0,0,0 // vmovups 0x80(%rax),%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,62,34,2,0 // vbroadcastss 0x2223e(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,186,20,2,0 // vbroadcastss 0x214ba(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,65,60,89,194 // vmulps %ymm10,%ymm8,%ymm8
- .byte 196,98,125,24,21,80,33,2,0 // vbroadcastss 0x22150(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,204,19,2,0 // vbroadcastss 0x213cc(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,60,88,194 // vaddps %ymm10,%ymm8,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
.byte 197,124,17,128,0,1,0,0 // vmovups %ymm8,0x100(%rax)
@@ -32273,14 +32273,14 @@ HIDDEN _sk_bicubic_n3y_avx
FUNCTION(_sk_bicubic_n3y_avx)
_sk_bicubic_n3y_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,11,34,2,0 // vbroadcastss 0x2220b(%rip),%ymm1 # 39888 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
+ .byte 196,226,125,24,13,135,20,2,0 // vbroadcastss 0x21487(%rip),%ymm1 # 38b04 <_sk_srcover_bgra_8888_sse2_lowp+0x400>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,173,32,2,0 // vbroadcastss 0x220ad(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,41,19,2,0 // vbroadcastss 0x21329(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,235,33,2,0 // vbroadcastss 0x221eb(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,103,20,2,0 // vbroadcastss 0x21467(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,65,60,89,194 // vmulps %ymm10,%ymm8,%ymm8
- .byte 196,98,125,24,21,253,32,2,0 // vbroadcastss 0x220fd(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,121,19,2,0 // vbroadcastss 0x21379(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,60,88,194 // vaddps %ymm10,%ymm8,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -32292,19 +32292,19 @@ HIDDEN _sk_bicubic_n1y_avx
FUNCTION(_sk_bicubic_n1y_avx)
_sk_bicubic_n1y_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,180,33,2,0 // vbroadcastss 0x221b4(%rip),%ymm1 # 39884 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
+ .byte 196,226,125,24,13,48,20,2,0 // vbroadcastss 0x21430(%rip),%ymm1 # 38b00 <_sk_srcover_bgra_8888_sse2_lowp+0x3fc>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
- .byte 196,98,125,24,5,90,32,2,0 // vbroadcastss 0x2205a(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,214,18,2,0 // vbroadcastss 0x212d6(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,128,192,0,0,0 // vsubps 0xc0(%rax),%ymm8,%ymm8
- .byte 196,98,125,24,13,161,33,2,0 // vbroadcastss 0x221a1(%rip),%ymm9 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,13,29,20,2,0 // vbroadcastss 0x2141d(%rip),%ymm9 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,65,60,89,201 // vmulps %ymm9,%ymm8,%ymm9
- .byte 196,98,125,24,21,151,33,2,0 // vbroadcastss 0x22197(%rip),%ymm10 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,21,19,20,2,0 // vbroadcastss 0x21413(%rip),%ymm10 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
.byte 196,65,60,89,201 // vmulps %ymm9,%ymm8,%ymm9
- .byte 196,98,125,24,21,36,32,2,0 // vbroadcastss 0x22024(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,160,18,2,0 // vbroadcastss 0x212a0(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
.byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
- .byte 196,98,125,24,13,117,33,2,0 // vbroadcastss 0x22175(%rip),%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,241,19,2,0 // vbroadcastss 0x213f1(%rip),%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32315,17 +32315,17 @@ HIDDEN _sk_bicubic_p1y_avx
FUNCTION(_sk_bicubic_p1y_avx)
_sk_bicubic_p1y_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,98,125,24,5,245,31,2,0 // vbroadcastss 0x21ff5(%rip),%ymm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,5,113,18,2,0 // vbroadcastss 0x21271(%rip),%ymm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,188,88,72,64 // vaddps 0x40(%rax),%ymm8,%ymm1
.byte 197,124,16,136,192,0,0,0 // vmovups 0xc0(%rax),%ymm9
- .byte 196,98,125,24,21,59,33,2,0 // vbroadcastss 0x2213b(%rip),%ymm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x408>
+ .byte 196,98,125,24,21,183,19,2,0 // vbroadcastss 0x213b7(%rip),%ymm10 # 38b0c <_sk_srcover_bgra_8888_sse2_lowp+0x408>
.byte 196,65,52,89,210 // vmulps %ymm10,%ymm9,%ymm10
- .byte 196,98,125,24,29,49,33,2,0 // vbroadcastss 0x22131(%rip),%ymm11 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,98,125,24,29,173,19,2,0 // vbroadcastss 0x213ad(%rip),%ymm11 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
.byte 196,65,52,89,210 // vmulps %ymm10,%ymm9,%ymm10
.byte 196,65,44,88,192 // vaddps %ymm8,%ymm10,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
- .byte 196,98,125,24,13,24,33,2,0 // vbroadcastss 0x22118(%rip),%ymm9 # 39898 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
+ .byte 196,98,125,24,13,148,19,2,0 // vbroadcastss 0x21394(%rip),%ymm9 # 38b14 <_sk_srcover_bgra_8888_sse2_lowp+0x410>
.byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -32336,13 +32336,13 @@ HIDDEN _sk_bicubic_p3y_avx
FUNCTION(_sk_bicubic_p3y_avx)
_sk_bicubic_p3y_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,13,248,32,2,0 // vbroadcastss 0x220f8(%rip),%ymm1 # 39894 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
+ .byte 196,226,125,24,13,116,19,2,0 // vbroadcastss 0x21374(%rip),%ymm1 # 38b10 <_sk_srcover_bgra_8888_sse2_lowp+0x40c>
.byte 197,244,88,72,64 // vaddps 0x40(%rax),%ymm1,%ymm1
.byte 197,124,16,128,192,0,0,0 // vmovups 0xc0(%rax),%ymm8
.byte 196,65,60,89,200 // vmulps %ymm8,%ymm8,%ymm9
- .byte 196,98,125,24,21,213,32,2,0 // vbroadcastss 0x220d5(%rip),%ymm10 # 3988c <_sk_srcover_bgra_8888_sse2_lowp+0x404>
+ .byte 196,98,125,24,21,81,19,2,0 // vbroadcastss 0x21351(%rip),%ymm10 # 38b08 <_sk_srcover_bgra_8888_sse2_lowp+0x404>
.byte 196,65,60,89,194 // vmulps %ymm10,%ymm8,%ymm8
- .byte 196,98,125,24,21,231,31,2,0 // vbroadcastss 0x21fe7(%rip),%ymm10 # 397ac <_sk_srcover_bgra_8888_sse2_lowp+0x324>
+ .byte 196,98,125,24,21,99,18,2,0 // vbroadcastss 0x21263(%rip),%ymm10 # 38a28 <_sk_srcover_bgra_8888_sse2_lowp+0x324>
.byte 196,65,60,88,194 // vaddps %ymm10,%ymm8,%ymm8
.byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
.byte 197,124,17,128,64,1,0,0 // vmovups %ymm8,0x140(%rax)
@@ -32490,7 +32490,7 @@ _sk_clut_3D_avx:
.byte 72,139,0 // mov (%rax),%rax
.byte 197,217,254,207 // vpaddd %xmm7,%xmm4,%xmm1
.byte 197,249,254,193 // vpaddd %xmm1,%xmm0,%xmm0
- .byte 196,226,121,24,37,109,30,2,0 // vbroadcastss 0x21e6d(%rip),%xmm4 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 196,226,121,24,37,233,16,2,0 // vbroadcastss 0x210e9(%rip),%xmm4 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 196,98,121,64,228 // vpmulld %xmm4,%xmm0,%xmm12
.byte 196,226,49,64,198 // vpmulld %xmm6,%xmm9,%xmm0
.byte 197,249,127,132,36,128,0,0,0 // vmovdqa %xmm0,0x80(%rsp)
@@ -32544,7 +32544,7 @@ _sk_clut_3D_avx:
.byte 196,161,122,16,60,128 // vmovss (%rax,%r8,4),%xmm7
.byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
.byte 196,227,77,24,237,1 // vinsertf128 $0x1,%xmm5,%ymm6,%ymm5
- .byte 196,98,121,24,21,230,27,2,0 // vbroadcastss 0x21be6(%rip),%xmm10 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,121,24,21,98,14,2,0 // vbroadcastss 0x20e62(%rip),%xmm10 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,193,105,254,210 // vpaddd %xmm10,%xmm2,%xmm2
.byte 196,195,249,22,208,1 // vpextrq $0x1,%xmm2,%r8
.byte 196,193,249,126,209 // vmovq %xmm2,%r9
@@ -32572,7 +32572,7 @@ _sk_clut_3D_avx:
.byte 196,161,122,16,60,128 // vmovss (%rax,%r8,4),%xmm7
.byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
.byte 196,227,77,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm6,%ymm2
- .byte 196,226,125,24,61,170,28,2,0 // vbroadcastss 0x21caa(%rip),%ymm7 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,226,125,24,61,38,15,2,0 // vbroadcastss 0x20f26(%rip),%ymm7 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 197,148,88,247 // vaddps %ymm7,%ymm13,%ymm6
.byte 197,124,40,231 // vmovaps %ymm7,%ymm12
.byte 197,124,17,36,36 // vmovups %ymm12,(%rsp)
@@ -33315,7 +33315,7 @@ _sk_clut_4D_avx:
.byte 197,249,127,68,36,32 // vmovdqa %xmm0,0x20(%rsp)
.byte 197,225,254,216 // vpaddd %xmm0,%xmm3,%xmm3
.byte 197,233,254,195 // vpaddd %xmm3,%xmm2,%xmm0
- .byte 196,98,121,24,13,110,13,2,0 // vbroadcastss 0x20d6e(%rip),%xmm9 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x418>
+ .byte 196,98,121,24,13,234,255,1,0 // vbroadcastss 0x1ffea(%rip),%xmm9 # 38b1c <_sk_srcover_bgra_8888_sse2_lowp+0x418>
.byte 196,194,121,64,209 // vpmulld %xmm9,%xmm0,%xmm2
.byte 196,226,81,64,199 // vpmulld %xmm7,%xmm5,%xmm0
.byte 197,249,127,132,36,96,1,0,0 // vmovdqa %xmm0,0x160(%rsp)
@@ -33372,7 +33372,7 @@ _sk_clut_4D_avx:
.byte 196,161,122,16,60,128 // vmovss (%rax,%r8,4),%xmm7
.byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
.byte 196,227,77,24,237,1 // vinsertf128 $0x1,%xmm5,%ymm6,%ymm5
- .byte 196,98,121,24,21,215,10,2,0 // vbroadcastss 0x20ad7(%rip),%xmm10 # 39744 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
+ .byte 196,98,121,24,21,83,253,1,0 // vbroadcastss 0x1fd53(%rip),%xmm10 # 389c0 <_sk_srcover_bgra_8888_sse2_lowp+0x2bc>
.byte 196,193,121,254,194 // vpaddd %xmm10,%xmm0,%xmm0
.byte 196,195,249,22,192,1 // vpextrq $0x1,%xmm0,%r8
.byte 196,193,249,126,193 // vmovq %xmm0,%r9
@@ -33400,7 +33400,7 @@ _sk_clut_4D_avx:
.byte 196,161,122,16,52,128 // vmovss (%rax,%r8,4),%xmm6
.byte 196,227,105,33,214,48 // vinsertps $0x30,%xmm6,%xmm2,%xmm2
.byte 196,227,109,24,208,1 // vinsertf128 $0x1,%xmm0,%ymm2,%ymm2
- .byte 196,98,125,24,37,155,11,2,0 // vbroadcastss 0x20b9b(%rip),%ymm12 # 3989c <_sk_srcover_bgra_8888_sse2_lowp+0x414>
+ .byte 196,98,125,24,37,23,254,1,0 // vbroadcastss 0x1fe17(%rip),%ymm12 # 38b18 <_sk_srcover_bgra_8888_sse2_lowp+0x414>
.byte 196,193,4,88,196 // vaddps %ymm12,%ymm15,%ymm0
.byte 197,124,17,36,36 // vmovups %ymm12,(%rsp)
.byte 197,254,91,192 // vcvttps2dq %ymm0,%ymm0
@@ -34798,7 +34798,7 @@ _sk_clut_4D_avx:
.byte 197,228,89,210 // vmulps %ymm2,%ymm3,%ymm2
.byte 197,220,88,210 // vaddps %ymm2,%ymm4,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,24,29,248,236,1,0 // vbroadcastss 0x1ecf8(%rip),%ymm3 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,226,125,24,29,116,223,1,0 // vbroadcastss 0x1df74(%rip),%ymm3 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,252,16,164,36,64,2,0,0 // vmovups 0x240(%rsp),%ymm4
.byte 197,252,16,172,36,96,2,0,0 // vmovups 0x260(%rsp),%ymm5
.byte 197,252,16,180,36,128,2,0,0 // vmovups 0x280(%rsp),%ymm6
@@ -34814,18 +34814,18 @@ HIDDEN _sk_gauss_a_to_rgba_avx
.globl _sk_gauss_a_to_rgba_avx
FUNCTION(_sk_gauss_a_to_rgba_avx)
_sk_gauss_a_to_rgba_avx:
- .byte 196,226,125,24,5,39,238,1,0 // vbroadcastss 0x1ee27(%rip),%ymm0 # 398a4 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
+ .byte 196,226,125,24,5,163,224,1,0 // vbroadcastss 0x1e0a3(%rip),%ymm0 # 38b20 <_sk_srcover_bgra_8888_sse2_lowp+0x41c>
.byte 197,228,89,192 // vmulps %ymm0,%ymm3,%ymm0
- .byte 196,226,125,24,13,30,238,1,0 // vbroadcastss 0x1ee1e(%rip),%ymm1 # 398a8 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
+ .byte 196,226,125,24,13,154,224,1,0 // vbroadcastss 0x1e09a(%rip),%ymm1 # 38b24 <_sk_srcover_bgra_8888_sse2_lowp+0x420>
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
- .byte 196,226,125,24,13,17,238,1,0 // vbroadcastss 0x1ee11(%rip),%ymm1 # 398ac <_sk_srcover_bgra_8888_sse2_lowp+0x424>
+ .byte 196,226,125,24,13,141,224,1,0 // vbroadcastss 0x1e08d(%rip),%ymm1 # 38b28 <_sk_srcover_bgra_8888_sse2_lowp+0x424>
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
- .byte 196,226,125,24,13,4,238,1,0 // vbroadcastss 0x1ee04(%rip),%ymm1 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x428>
+ .byte 196,226,125,24,13,128,224,1,0 // vbroadcastss 0x1e080(%rip),%ymm1 # 38b2c <_sk_srcover_bgra_8888_sse2_lowp+0x428>
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
.byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
- .byte 196,226,125,24,13,247,237,1,0 // vbroadcastss 0x1edf7(%rip),%ymm1 # 398b4 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
+ .byte 196,226,125,24,13,115,224,1,0 // vbroadcastss 0x1e073(%rip),%ymm1 # 38b30 <_sk_srcover_bgra_8888_sse2_lowp+0x42c>
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
@@ -34925,9 +34925,9 @@ _sk_seed_shader_sse41:
.byte 102,15,110,201 // movd %ecx,%xmm1
.byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,88,13,162,244,1,0 // addps 0x1f4a2(%rip),%xmm1 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,13,34,231,1,0 // addps 0x1e722(%rip),%xmm1 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,21,169,244,1,0 // movaps 0x1f4a9(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,41,231,1,0 // movaps 0x1e729(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,87,219 // xorps %xmm3,%xmm3
.byte 15,87,228 // xorps %xmm4,%xmm4
.byte 15,87,237 // xorps %xmm5,%xmm5
@@ -34942,18 +34942,18 @@ _sk_dither_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 102,68,15,110,194 // movd %edx,%xmm8
.byte 102,69,15,112,192,0 // pshufd $0x0,%xmm8,%xmm8
- .byte 102,68,15,254,5,82,244,1,0 // paddd 0x1f452(%rip),%xmm8 # 3a040 <_sk_srcover_bgra_8888_sse2_lowp+0xbb8>
+ .byte 102,68,15,254,5,210,230,1,0 // paddd 0x1e6d2(%rip),%xmm8 # 392c0 <_sk_srcover_bgra_8888_sse2_lowp+0xbbc>
.byte 102,68,15,110,201 // movd %ecx,%xmm9
.byte 102,69,15,112,201,0 // pshufd $0x0,%xmm9,%xmm9
.byte 102,69,15,239,200 // pxor %xmm8,%xmm9
- .byte 102,68,15,111,21,121,244,1,0 // movdqa 0x1f479(%rip),%xmm10 # 3a080 <_sk_srcover_bgra_8888_sse2_lowp+0xbf8>
+ .byte 102,68,15,111,21,249,230,1,0 // movdqa 0x1e6f9(%rip),%xmm10 # 39300 <_sk_srcover_bgra_8888_sse2_lowp+0xbfc>
.byte 102,69,15,111,217 // movdqa %xmm9,%xmm11
.byte 102,69,15,219,218 // pand %xmm10,%xmm11
.byte 102,65,15,114,243,5 // pslld $0x5,%xmm11
.byte 102,69,15,219,208 // pand %xmm8,%xmm10
.byte 102,65,15,114,242,4 // pslld $0x4,%xmm10
- .byte 102,68,15,111,37,101,244,1,0 // movdqa 0x1f465(%rip),%xmm12 # 3a090 <_sk_srcover_bgra_8888_sse2_lowp+0xc08>
- .byte 102,68,15,111,45,108,244,1,0 // movdqa 0x1f46c(%rip),%xmm13 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,37,229,230,1,0 // movdqa 0x1e6e5(%rip),%xmm12 # 39310 <_sk_srcover_bgra_8888_sse2_lowp+0xc0c>
+ .byte 102,68,15,111,45,236,230,1,0 // movdqa 0x1e6ec(%rip),%xmm13 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,69,15,111,241 // movdqa %xmm9,%xmm14
.byte 102,69,15,219,245 // pand %xmm13,%xmm14
.byte 102,65,15,114,246,2 // pslld $0x2,%xmm14
@@ -34969,8 +34969,8 @@ _sk_dither_sse41:
.byte 102,69,15,235,198 // por %xmm14,%xmm8
.byte 102,69,15,235,193 // por %xmm9,%xmm8
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,39,244,1,0 // mulps 0x1f427(%rip),%xmm8 # 3a0b0 <_sk_srcover_bgra_8888_sse2_lowp+0xc28>
- .byte 68,15,88,5,47,244,1,0 // addps 0x1f42f(%rip),%xmm8 # 3a0c0 <_sk_srcover_bgra_8888_sse2_lowp+0xc38>
+ .byte 68,15,89,5,167,230,1,0 // mulps 0x1e6a7(%rip),%xmm8 # 39330 <_sk_srcover_bgra_8888_sse2_lowp+0xc2c>
+ .byte 68,15,88,5,175,230,1,0 // addps 0x1e6af(%rip),%xmm8 # 39340 <_sk_srcover_bgra_8888_sse2_lowp+0xc3c>
.byte 243,68,15,16,16 // movss (%rax),%xmm10
.byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
.byte 69,15,89,208 // mulps %xmm8,%xmm10
@@ -35012,7 +35012,7 @@ HIDDEN _sk_black_color_sse41
FUNCTION(_sk_black_color_sse41)
_sk_black_color_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,101,243,1,0 // movaps 0x1f365(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,229,229,1,0 // movaps 0x1e5e5(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -35023,7 +35023,7 @@ HIDDEN _sk_white_color_sse41
FUNCTION(_sk_white_color_sse41)
_sk_white_color_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,5,81,243,1,0 // movaps 0x1f351(%rip),%xmm0 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,5,209,229,1,0 // movaps 0x1e5d1(%rip),%xmm0 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,40,216 // movaps %xmm0,%xmm3
@@ -35069,7 +35069,7 @@ HIDDEN _sk_srcatop_sse41
FUNCTION(_sk_srcatop_sse41)
_sk_srcatop_sse41:
.byte 15,89,199 // mulps %xmm7,%xmm0
- .byte 68,15,40,5,1,243,1,0 // movaps 0x1f301(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,129,229,1,0 // movaps 0x1e581(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -35094,7 +35094,7 @@ FUNCTION(_sk_dstatop_sse41)
_sk_dstatop_sse41:
.byte 68,15,40,195 // movaps %xmm3,%xmm8
.byte 68,15,89,196 // mulps %xmm4,%xmm8
- .byte 68,15,40,13,180,242,1,0 // movaps 0x1f2b4(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,52,229,1,0 // movaps 0x1e534(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,88,192 // addps %xmm8,%xmm0
@@ -35141,7 +35141,7 @@ HIDDEN _sk_srcout_sse41
.globl _sk_srcout_sse41
FUNCTION(_sk_srcout_sse41)
_sk_srcout_sse41:
- .byte 68,15,40,5,72,242,1,0 // movaps 0x1f248(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,200,228,1,0 // movaps 0x1e4c8(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,199 // subps %xmm7,%xmm8
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
@@ -35154,7 +35154,7 @@ HIDDEN _sk_dstout_sse41
.globl _sk_dstout_sse41
FUNCTION(_sk_dstout_sse41)
_sk_dstout_sse41:
- .byte 68,15,40,5,40,242,1,0 // movaps 0x1f228(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,168,228,1,0 // movaps 0x1e4a8(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 15,89,196 // mulps %xmm4,%xmm0
@@ -35171,7 +35171,7 @@ HIDDEN _sk_srcover_sse41
.globl _sk_srcover_sse41
FUNCTION(_sk_srcover_sse41)
_sk_srcover_sse41:
- .byte 68,15,40,5,251,241,1,0 // movaps 0x1f1fb(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,123,228,1,0 // movaps 0x1e47b(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -35191,7 +35191,7 @@ HIDDEN _sk_dstover_sse41
.globl _sk_dstover_sse41
FUNCTION(_sk_dstover_sse41)
_sk_dstover_sse41:
- .byte 68,15,40,5,191,241,1,0 // movaps 0x1f1bf(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,63,228,1,0 // movaps 0x1e43f(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,199 // subps %xmm7,%xmm8
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 15,88,196 // addps %xmm4,%xmm0
@@ -35219,7 +35219,7 @@ HIDDEN _sk_multiply_sse41
.globl _sk_multiply_sse41
FUNCTION(_sk_multiply_sse41)
_sk_multiply_sse41:
- .byte 68,15,40,5,131,241,1,0 // movaps 0x1f183(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,3,228,1,0 // movaps 0x1e403(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 69,15,40,209 // movaps %xmm9,%xmm10
@@ -35257,7 +35257,7 @@ HIDDEN _sk_plus__sse41
FUNCTION(_sk_plus__sse41)
_sk_plus__sse41:
.byte 15,88,196 // addps %xmm4,%xmm0
- .byte 68,15,40,5,4,241,1,0 // movaps 0x1f104(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,132,227,1,0 // movaps 0x1e384(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 15,88,205 // addps %xmm5,%xmm1
.byte 65,15,93,200 // minps %xmm8,%xmm1
@@ -35299,7 +35299,7 @@ HIDDEN _sk_xor__sse41
FUNCTION(_sk_xor__sse41)
_sk_xor__sse41:
.byte 68,15,40,195 // movaps %xmm3,%xmm8
- .byte 15,40,29,144,240,1,0 // movaps 0x1f090(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,16,227,1,0 // movaps 0x1e310(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,40,203 // movaps %xmm3,%xmm9
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
@@ -35347,7 +35347,7 @@ _sk_darken_sse41:
.byte 68,15,89,206 // mulps %xmm6,%xmm9
.byte 65,15,95,209 // maxps %xmm9,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,235,239,1,0 // movaps 0x1efeb(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,107,226,1,0 // movaps 0x1e26b(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -35381,7 +35381,7 @@ _sk_lighten_sse41:
.byte 68,15,89,206 // mulps %xmm6,%xmm9
.byte 65,15,93,209 // minps %xmm9,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,128,239,1,0 // movaps 0x1ef80(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,0,226,1,0 // movaps 0x1e200(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -35418,7 +35418,7 @@ _sk_difference_sse41:
.byte 65,15,93,209 // minps %xmm9,%xmm2
.byte 15,88,210 // addps %xmm2,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,10,239,1,0 // movaps 0x1ef0a(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,138,225,1,0 // movaps 0x1e18a(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -35446,7 +35446,7 @@ _sk_exclusion_sse41:
.byte 15,89,214 // mulps %xmm6,%xmm2
.byte 15,88,210 // addps %xmm2,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,186,238,1,0 // movaps 0x1eeba(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,58,225,1,0 // movaps 0x1e13a(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -35458,7 +35458,7 @@ HIDDEN _sk_colorburn_sse41
.globl _sk_colorburn_sse41
FUNCTION(_sk_colorburn_sse41)
_sk_colorburn_sse41:
- .byte 68,15,40,29,161,238,1,0 // movaps 0x1eea1(%rip),%xmm11 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,29,33,225,1,0 // movaps 0x1e121(%rip),%xmm11 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,227 // movaps %xmm11,%xmm12
.byte 68,15,92,231 // subps %xmm7,%xmm12
.byte 69,15,40,204 // movaps %xmm12,%xmm9
@@ -35542,7 +35542,7 @@ HIDDEN _sk_colordodge_sse41
FUNCTION(_sk_colordodge_sse41)
_sk_colordodge_sse41:
.byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 68,15,40,21,107,237,1,0 // movaps 0x1ed6b(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,235,223,1,0 // movaps 0x1dfeb(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,218 // movaps %xmm10,%xmm11
.byte 68,15,92,223 // subps %xmm7,%xmm11
.byte 69,15,40,227 // movaps %xmm11,%xmm12
@@ -35624,7 +35624,7 @@ _sk_hardlight_sse41:
.byte 15,40,244 // movaps %xmm4,%xmm6
.byte 15,40,227 // movaps %xmm3,%xmm4
.byte 68,15,40,200 // movaps %xmm0,%xmm9
- .byte 68,15,40,21,54,236,1,0 // movaps 0x1ec36(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,182,222,1,0 // movaps 0x1deb6(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,40,234 // movaps %xmm10,%xmm5
.byte 15,92,239 // subps %xmm7,%xmm5
.byte 15,40,197 // movaps %xmm5,%xmm0
@@ -35707,7 +35707,7 @@ FUNCTION(_sk_overlay_sse41)
_sk_overlay_sse41:
.byte 68,15,40,201 // movaps %xmm1,%xmm9
.byte 68,15,40,240 // movaps %xmm0,%xmm14
- .byte 68,15,40,21,11,235,1,0 // movaps 0x1eb0b(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,139,221,1,0 // movaps 0x1dd8b(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,218 // movaps %xmm10,%xmm11
.byte 68,15,92,223 // subps %xmm7,%xmm11
.byte 65,15,40,195 // movaps %xmm11,%xmm0
@@ -35792,7 +35792,7 @@ _sk_softlight_sse41:
.byte 15,40,198 // movaps %xmm6,%xmm0
.byte 15,94,199 // divps %xmm7,%xmm0
.byte 65,15,84,193 // andps %xmm9,%xmm0
- .byte 15,40,13,210,233,1,0 // movaps 0x1e9d2(%rip),%xmm1 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,13,82,220,1,0 // movaps 0x1dc52(%rip),%xmm1 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,40,209 // movaps %xmm1,%xmm10
.byte 68,15,92,208 // subps %xmm0,%xmm10
.byte 68,15,40,240 // movaps %xmm0,%xmm14
@@ -35805,10 +35805,10 @@ _sk_softlight_sse41:
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,89,210 // mulps %xmm2,%xmm2
.byte 15,88,208 // addps %xmm0,%xmm2
- .byte 68,15,40,45,0,234,1,0 // movaps 0x1ea00(%rip),%xmm13 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,45,128,220,1,0 // movaps 0x1dc80(%rip),%xmm13 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 69,15,88,245 // addps %xmm13,%xmm14
.byte 68,15,89,242 // mulps %xmm2,%xmm14
- .byte 68,15,40,37,0,234,1,0 // movaps 0x1ea00(%rip),%xmm12 # 3a0e0 <_sk_srcover_bgra_8888_sse2_lowp+0xc58>
+ .byte 68,15,40,37,128,220,1,0 // movaps 0x1dc80(%rip),%xmm12 # 39360 <_sk_srcover_bgra_8888_sse2_lowp+0xc5c>
.byte 69,15,89,252 // mulps %xmm12,%xmm15
.byte 69,15,88,254 // addps %xmm14,%xmm15
.byte 15,40,198 // movaps %xmm6,%xmm0
@@ -35994,12 +35994,12 @@ _sk_hue_sse41:
.byte 68,15,84,208 // andps %xmm0,%xmm10
.byte 15,84,200 // andps %xmm0,%xmm1
.byte 68,15,84,232 // andps %xmm0,%xmm13
- .byte 15,40,5,107,231,1,0 // movaps 0x1e76b(%rip),%xmm0 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 15,40,5,235,217,1,0 // movaps 0x1d9eb(%rip),%xmm0 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,89,224 // mulps %xmm0,%xmm12
- .byte 15,40,21,112,231,1,0 // movaps 0x1e770(%rip),%xmm2 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 15,40,21,240,217,1,0 // movaps 0x1d9f0(%rip),%xmm2 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 15,89,250 // mulps %xmm2,%xmm7
.byte 65,15,88,252 // addps %xmm12,%xmm7
- .byte 68,15,40,53,113,231,1,0 // movaps 0x1e771(%rip),%xmm14 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,53,241,217,1,0 // movaps 0x1d9f1(%rip),%xmm14 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 68,15,40,252 // movaps %xmm4,%xmm15
.byte 69,15,89,254 // mulps %xmm14,%xmm15
.byte 68,15,88,255 // addps %xmm7,%xmm15
@@ -36082,7 +36082,7 @@ _sk_hue_sse41:
.byte 65,15,88,214 // addps %xmm14,%xmm2
.byte 15,40,196 // movaps %xmm4,%xmm0
.byte 102,15,56,20,202 // blendvps %xmm0,%xmm2,%xmm1
- .byte 68,15,40,13,133,229,1,0 // movaps 0x1e585(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,5,216,1,0 // movaps 0x1d805(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,40,225 // movaps %xmm9,%xmm4
.byte 15,92,229 // subps %xmm5,%xmm4
.byte 15,40,68,36,200 // movaps -0x38(%rsp),%xmm0
@@ -36176,14 +36176,14 @@ _sk_saturation_sse41:
.byte 68,15,84,215 // andps %xmm7,%xmm10
.byte 68,15,84,223 // andps %xmm7,%xmm11
.byte 68,15,84,199 // andps %xmm7,%xmm8
- .byte 15,40,21,175,228,1,0 // movaps 0x1e4af(%rip),%xmm2 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 15,40,21,47,215,1,0 // movaps 0x1d72f(%rip),%xmm2 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 15,40,221 // movaps %xmm5,%xmm3
.byte 15,89,218 // mulps %xmm2,%xmm3
- .byte 15,40,13,178,228,1,0 // movaps 0x1e4b2(%rip),%xmm1 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 15,40,13,50,215,1,0 // movaps 0x1d732(%rip),%xmm1 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 15,40,254 // movaps %xmm6,%xmm7
.byte 15,89,249 // mulps %xmm1,%xmm7
.byte 15,88,251 // addps %xmm3,%xmm7
- .byte 68,15,40,45,177,228,1,0 // movaps 0x1e4b1(%rip),%xmm13 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,45,49,215,1,0 // movaps 0x1d731(%rip),%xmm13 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 69,15,89,245 // mulps %xmm13,%xmm14
.byte 68,15,88,247 // addps %xmm7,%xmm14
.byte 65,15,40,218 // movaps %xmm10,%xmm3
@@ -36264,7 +36264,7 @@ _sk_saturation_sse41:
.byte 65,15,88,253 // addps %xmm13,%xmm7
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 102,68,15,56,20,223 // blendvps %xmm0,%xmm7,%xmm11
- .byte 68,15,40,13,199,226,1,0 // movaps 0x1e2c7(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,71,213,1,0 // movaps 0x1d547(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,193 // movaps %xmm9,%xmm8
.byte 68,15,92,204 // subps %xmm4,%xmm9
.byte 15,40,124,36,168 // movaps -0x58(%rsp),%xmm7
@@ -36319,14 +36319,14 @@ _sk_color_sse41:
.byte 15,40,231 // movaps %xmm7,%xmm4
.byte 68,15,89,244 // mulps %xmm4,%xmm14
.byte 15,89,204 // mulps %xmm4,%xmm1
- .byte 68,15,40,13,130,226,1,0 // movaps 0x1e282(%rip),%xmm9 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,13,2,213,1,0 // movaps 0x1d502(%rip),%xmm9 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 65,15,40,250 // movaps %xmm10,%xmm7
.byte 65,15,89,249 // mulps %xmm9,%xmm7
- .byte 68,15,40,21,130,226,1,0 // movaps 0x1e282(%rip),%xmm10 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 68,15,40,21,2,213,1,0 // movaps 0x1d502(%rip),%xmm10 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 65,15,40,219 // movaps %xmm11,%xmm3
.byte 65,15,89,218 // mulps %xmm10,%xmm3
.byte 15,88,223 // addps %xmm7,%xmm3
- .byte 68,15,40,29,127,226,1,0 // movaps 0x1e27f(%rip),%xmm11 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,29,255,212,1,0 // movaps 0x1d4ff(%rip),%xmm11 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 69,15,40,236 // movaps %xmm12,%xmm13
.byte 69,15,89,235 // mulps %xmm11,%xmm13
.byte 68,15,88,235 // addps %xmm3,%xmm13
@@ -36411,7 +36411,7 @@ _sk_color_sse41:
.byte 65,15,88,251 // addps %xmm11,%xmm7
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 102,15,56,20,207 // blendvps %xmm0,%xmm7,%xmm1
- .byte 68,15,40,13,139,224,1,0 // movaps 0x1e08b(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,11,211,1,0 // movaps 0x1d30b(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,92,196 // subps %xmm4,%xmm0
.byte 68,15,89,192 // mulps %xmm0,%xmm8
@@ -36463,13 +36463,13 @@ _sk_luminosity_sse41:
.byte 69,15,89,216 // mulps %xmm8,%xmm11
.byte 68,15,40,203 // movaps %xmm3,%xmm9
.byte 68,15,89,205 // mulps %xmm5,%xmm9
- .byte 68,15,40,5,83,224,1,0 // movaps 0x1e053(%rip),%xmm8 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,5,211,210,1,0 // movaps 0x1d2d3(%rip),%xmm8 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
- .byte 68,15,40,21,87,224,1,0 // movaps 0x1e057(%rip),%xmm10 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 68,15,40,21,215,210,1,0 // movaps 0x1d2d7(%rip),%xmm10 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 15,40,233 // movaps %xmm1,%xmm5
.byte 65,15,89,234 // mulps %xmm10,%xmm5
.byte 15,88,232 // addps %xmm0,%xmm5
- .byte 68,15,40,37,85,224,1,0 // movaps 0x1e055(%rip),%xmm12 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,37,213,210,1,0 // movaps 0x1d2d5(%rip),%xmm12 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 68,15,40,242 // movaps %xmm2,%xmm14
.byte 69,15,89,244 // mulps %xmm12,%xmm14
.byte 68,15,88,245 // addps %xmm5,%xmm14
@@ -36554,7 +36554,7 @@ _sk_luminosity_sse41:
.byte 65,15,88,244 // addps %xmm12,%xmm6
.byte 65,15,40,195 // movaps %xmm11,%xmm0
.byte 102,68,15,56,20,206 // blendvps %xmm0,%xmm6,%xmm9
- .byte 15,40,5,91,222,1,0 // movaps 0x1de5b(%rip),%xmm0 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,5,219,208,1,0 // movaps 0x1d0db(%rip),%xmm0 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,92,215 // subps %xmm7,%xmm2
.byte 15,89,226 // mulps %xmm2,%xmm4
@@ -36602,20 +36602,20 @@ _sk_srcover_rgba_8888_sse41:
.byte 15,133,219,0,0,0 // jne 1c37d <_sk_srcover_rgba_8888_sse41+0xfb>
.byte 243,65,15,111,60,128 // movdqu (%r8,%rax,4),%xmm7
.byte 72,133,255 // test %rdi,%rdi
- .byte 102,15,111,37,109,222,1,0 // movdqa 0x1de6d(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,237,208,1,0 // movdqa 0x1d0ed(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
.byte 102,15,111,239 // movdqa %xmm7,%xmm5
- .byte 102,15,56,0,45,105,222,1,0 // pshufb 0x1de69(%rip),%xmm5 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,45,233,208,1,0 // pshufb 0x1d0e9(%rip),%xmm5 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
.byte 102,15,111,247 // movdqa %xmm7,%xmm6
- .byte 102,15,56,0,53,105,222,1,0 // pshufb 0x1de69(%rip),%xmm6 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,53,233,208,1,0 // pshufb 0x1d0e9(%rip),%xmm6 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
.byte 102,15,114,215,24 // psrld $0x18,%xmm7
.byte 15,91,255 // cvtdq2ps %xmm7,%xmm7
- .byte 68,15,40,5,134,221,1,0 // movaps 0x1dd86(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,6,208,1,0 // movaps 0x1d006(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
- .byte 68,15,40,37,90,222,1,0 // movaps 0x1de5a(%rip),%xmm12 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,37,218,208,1,0 // movaps 0x1d0da(%rip),%xmm12 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,196 // mulps %xmm12,%xmm0
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -36694,20 +36694,20 @@ _sk_srcover_bgra_8888_sse41:
.byte 15,133,219,0,0,0 // jne 1c4f6 <_sk_srcover_bgra_8888_sse41+0xfb>
.byte 243,65,15,111,60,128 // movdqu (%r8,%rax,4),%xmm7
.byte 72,133,255 // test %rdi,%rdi
- .byte 102,15,111,37,244,220,1,0 // movdqa 0x1dcf4(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,116,207,1,0 // movdqa 0x1cf74(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,244 // cvtdq2ps %xmm4,%xmm6
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
- .byte 102,15,56,0,37,240,220,1,0 // pshufb 0x1dcf0(%rip),%xmm4 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,37,112,207,1,0 // pshufb 0x1cf70(%rip),%xmm4 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,236 // cvtdq2ps %xmm4,%xmm5
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
- .byte 102,15,56,0,37,240,220,1,0 // pshufb 0x1dcf0(%rip),%xmm4 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,37,112,207,1,0 // pshufb 0x1cf70(%rip),%xmm4 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
.byte 102,15,114,215,24 // psrld $0x18,%xmm7
.byte 15,91,255 // cvtdq2ps %xmm7,%xmm7
- .byte 68,15,40,5,13,220,1,0 // movaps 0x1dc0d(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,141,206,1,0 // movaps 0x1ce8d(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
- .byte 68,15,40,37,225,220,1,0 // movaps 0x1dce1(%rip),%xmm12 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,37,97,207,1,0 // movaps 0x1cf61(%rip),%xmm12 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,196 // mulps %xmm12,%xmm0
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -36787,7 +36787,7 @@ HIDDEN _sk_clamp_1_sse41
.globl _sk_clamp_1_sse41
FUNCTION(_sk_clamp_1_sse41)
_sk_clamp_1_sse41:
- .byte 68,15,40,5,220,218,1,0 // movaps 0x1dadc(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,92,205,1,0 // movaps 0x1cd5c(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 65,15,93,208 // minps %xmm8,%xmm2
@@ -36799,7 +36799,7 @@ HIDDEN _sk_clamp_a_sse41
.globl _sk_clamp_a_sse41
FUNCTION(_sk_clamp_a_sse41)
_sk_clamp_a_sse41:
- .byte 15,93,29,193,218,1,0 // minps 0x1dac1(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,29,65,205,1,0 // minps 0x1cd41(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,93,195 // minps %xmm3,%xmm0
.byte 15,93,203 // minps %xmm3,%xmm1
.byte 15,93,211 // minps %xmm3,%xmm2
@@ -36810,7 +36810,7 @@ HIDDEN _sk_clamp_a_dst_sse41
.globl _sk_clamp_a_dst_sse41
FUNCTION(_sk_clamp_a_dst_sse41)
_sk_clamp_a_dst_sse41:
- .byte 15,93,61,173,218,1,0 // minps 0x1daad(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,61,45,205,1,0 // minps 0x1cd2d(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,93,231 // minps %xmm7,%xmm4
.byte 15,93,239 // minps %xmm7,%xmm5
.byte 15,93,247 // minps %xmm7,%xmm6
@@ -36845,7 +36845,7 @@ HIDDEN _sk_invert_sse41
.globl _sk_invert_sse41
FUNCTION(_sk_invert_sse41)
_sk_invert_sse41:
- .byte 68,15,40,5,105,218,1,0 // movaps 0x1da69(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,233,204,1,0 // movaps 0x1cce9(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,92,200 // subps %xmm0,%xmm9
.byte 69,15,40,208 // movaps %xmm8,%xmm10
@@ -36906,10 +36906,10 @@ HIDDEN _sk_unpremul_sse41
.globl _sk_unpremul_sse41
FUNCTION(_sk_unpremul_sse41)
_sk_unpremul_sse41:
- .byte 68,15,40,5,247,217,1,0 // movaps 0x1d9f7(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,119,204,1,0 // movaps 0x1cc77(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,94,195 // divps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
- .byte 68,15,194,13,214,218,1,0,1 // cmpltps 0x1dad6(%rip),%xmm9 # 3a160 <_sk_srcover_bgra_8888_sse2_lowp+0xcd8>
+ .byte 68,15,194,13,86,205,1,0,1 // cmpltps 0x1cd56(%rip),%xmm9 # 393e0 <_sk_srcover_bgra_8888_sse2_lowp+0xcdc>
.byte 69,15,84,200 // andps %xmm8,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,89,201 // mulps %xmm9,%xmm1
@@ -36921,20 +36921,20 @@ HIDDEN _sk_from_srgb_sse41
.globl _sk_from_srgb_sse41
FUNCTION(_sk_from_srgb_sse41)
_sk_from_srgb_sse41:
- .byte 68,15,40,29,202,218,1,0 // movaps 0x1daca(%rip),%xmm11 # 3a170 <_sk_srcover_bgra_8888_sse2_lowp+0xce8>
+ .byte 68,15,40,29,74,205,1,0 // movaps 0x1cd4a(%rip),%xmm11 # 393f0 <_sk_srcover_bgra_8888_sse2_lowp+0xcec>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
.byte 68,15,40,208 // movaps %xmm0,%xmm10
.byte 69,15,89,210 // mulps %xmm10,%xmm10
- .byte 68,15,40,37,50,218,1,0 // movaps 0x1da32(%rip),%xmm12 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,37,178,204,1,0 // movaps 0x1ccb2(%rip),%xmm12 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,40,192 // movaps %xmm0,%xmm8
.byte 69,15,89,196 // mulps %xmm12,%xmm8
- .byte 68,15,40,45,178,218,1,0 // movaps 0x1dab2(%rip),%xmm13 # 3a180 <_sk_srcover_bgra_8888_sse2_lowp+0xcf8>
+ .byte 68,15,40,45,50,205,1,0 // movaps 0x1cd32(%rip),%xmm13 # 39400 <_sk_srcover_bgra_8888_sse2_lowp+0xcfc>
.byte 69,15,88,197 // addps %xmm13,%xmm8
.byte 69,15,89,194 // mulps %xmm10,%xmm8
- .byte 68,15,40,53,178,218,1,0 // movaps 0x1dab2(%rip),%xmm14 # 3a190 <_sk_srcover_bgra_8888_sse2_lowp+0xd08>
+ .byte 68,15,40,53,50,205,1,0 // movaps 0x1cd32(%rip),%xmm14 # 39410 <_sk_srcover_bgra_8888_sse2_lowp+0xd0c>
.byte 69,15,88,198 // addps %xmm14,%xmm8
- .byte 68,15,40,61,182,218,1,0 // movaps 0x1dab6(%rip),%xmm15 # 3a1a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd18>
+ .byte 68,15,40,61,54,205,1,0 // movaps 0x1cd36(%rip),%xmm15 # 39420 <_sk_srcover_bgra_8888_sse2_lowp+0xd1c>
.byte 65,15,194,199,1 // cmpltps %xmm15,%xmm0
.byte 102,69,15,56,20,193 // blendvps %xmm0,%xmm9,%xmm8
.byte 68,15,40,209 // movaps %xmm1,%xmm10
@@ -36971,19 +36971,19 @@ FUNCTION(_sk_from_srgb_dst_sse41)
_sk_from_srgb_dst_sse41:
.byte 68,15,40,204 // movaps %xmm4,%xmm9
.byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 68,15,40,29,1,218,1,0 // movaps 0x1da01(%rip),%xmm11 # 3a170 <_sk_srcover_bgra_8888_sse2_lowp+0xce8>
+ .byte 68,15,40,29,129,204,1,0 // movaps 0x1cc81(%rip),%xmm11 # 393f0 <_sk_srcover_bgra_8888_sse2_lowp+0xcec>
.byte 69,15,40,209 // movaps %xmm9,%xmm10
.byte 69,15,89,211 // mulps %xmm11,%xmm10
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
- .byte 68,15,40,37,106,217,1,0 // movaps 0x1d96a(%rip),%xmm12 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,37,234,203,1,0 // movaps 0x1cbea(%rip),%xmm12 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 65,15,89,228 // mulps %xmm12,%xmm4
- .byte 68,15,40,45,238,217,1,0 // movaps 0x1d9ee(%rip),%xmm13 # 3a180 <_sk_srcover_bgra_8888_sse2_lowp+0xcf8>
+ .byte 68,15,40,45,110,204,1,0 // movaps 0x1cc6e(%rip),%xmm13 # 39400 <_sk_srcover_bgra_8888_sse2_lowp+0xcfc>
.byte 65,15,88,229 // addps %xmm13,%xmm4
.byte 15,89,224 // mulps %xmm0,%xmm4
- .byte 68,15,40,53,239,217,1,0 // movaps 0x1d9ef(%rip),%xmm14 # 3a190 <_sk_srcover_bgra_8888_sse2_lowp+0xd08>
+ .byte 68,15,40,53,111,204,1,0 // movaps 0x1cc6f(%rip),%xmm14 # 39410 <_sk_srcover_bgra_8888_sse2_lowp+0xd0c>
.byte 65,15,88,230 // addps %xmm14,%xmm4
- .byte 68,15,40,61,243,217,1,0 // movaps 0x1d9f3(%rip),%xmm15 # 3a1a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd18>
+ .byte 68,15,40,61,115,204,1,0 // movaps 0x1cc73(%rip),%xmm15 # 39420 <_sk_srcover_bgra_8888_sse2_lowp+0xd1c>
.byte 69,15,194,207,1 // cmpltps %xmm15,%xmm9
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 102,65,15,56,20,226 // blendvps %xmm0,%xmm10,%xmm4
@@ -37027,22 +37027,22 @@ _sk_to_srgb_sse41:
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 15,40,209 // movaps %xmm1,%xmm2
.byte 68,15,82,192 // rsqrtps %xmm0,%xmm8
- .byte 68,15,40,29,103,217,1,0 // movaps 0x1d967(%rip),%xmm11 # 3a1b0 <_sk_srcover_bgra_8888_sse2_lowp+0xd28>
+ .byte 68,15,40,29,231,203,1,0 // movaps 0x1cbe7(%rip),%xmm11 # 39430 <_sk_srcover_bgra_8888_sse2_lowp+0xd2c>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
- .byte 68,15,40,37,103,217,1,0 // movaps 0x1d967(%rip),%xmm12 # 3a1c0 <_sk_srcover_bgra_8888_sse2_lowp+0xd38>
+ .byte 68,15,40,37,231,203,1,0 // movaps 0x1cbe7(%rip),%xmm12 # 39440 <_sk_srcover_bgra_8888_sse2_lowp+0xd3c>
.byte 69,15,40,248 // movaps %xmm8,%xmm15
.byte 69,15,89,252 // mulps %xmm12,%xmm15
- .byte 68,15,40,21,103,217,1,0 // movaps 0x1d967(%rip),%xmm10 # 3a1d0 <_sk_srcover_bgra_8888_sse2_lowp+0xd48>
+ .byte 68,15,40,21,231,203,1,0 // movaps 0x1cbe7(%rip),%xmm10 # 39450 <_sk_srcover_bgra_8888_sse2_lowp+0xd4c>
.byte 69,15,88,250 // addps %xmm10,%xmm15
.byte 69,15,89,248 // mulps %xmm8,%xmm15
- .byte 68,15,40,45,103,217,1,0 // movaps 0x1d967(%rip),%xmm13 # 3a1e0 <_sk_srcover_bgra_8888_sse2_lowp+0xd58>
+ .byte 68,15,40,45,231,203,1,0 // movaps 0x1cbe7(%rip),%xmm13 # 39460 <_sk_srcover_bgra_8888_sse2_lowp+0xd5c>
.byte 69,15,88,253 // addps %xmm13,%xmm15
- .byte 68,15,40,53,107,217,1,0 // movaps 0x1d96b(%rip),%xmm14 # 3a1f0 <_sk_srcover_bgra_8888_sse2_lowp+0xd68>
+ .byte 68,15,40,53,235,203,1,0 // movaps 0x1cbeb(%rip),%xmm14 # 39470 <_sk_srcover_bgra_8888_sse2_lowp+0xd6c>
.byte 69,15,88,198 // addps %xmm14,%xmm8
.byte 69,15,83,192 // rcpps %xmm8,%xmm8
.byte 69,15,89,199 // mulps %xmm15,%xmm8
- .byte 68,15,40,61,103,217,1,0 // movaps 0x1d967(%rip),%xmm15 # 3a200 <_sk_srcover_bgra_8888_sse2_lowp+0xd78>
+ .byte 68,15,40,61,231,203,1,0 // movaps 0x1cbe7(%rip),%xmm15 # 39480 <_sk_srcover_bgra_8888_sse2_lowp+0xd7c>
.byte 65,15,194,199,1 // cmpltps %xmm15,%xmm0
.byte 102,69,15,56,20,193 // blendvps %xmm0,%xmm9,%xmm8
.byte 68,15,82,202 // rsqrtps %xmm2,%xmm9
@@ -37097,7 +37097,7 @@ _sk_rgb_to_hsl_sse41:
.byte 68,15,93,224 // minps %xmm0,%xmm12
.byte 65,15,40,203 // movaps %xmm11,%xmm1
.byte 65,15,92,204 // subps %xmm12,%xmm1
- .byte 68,15,40,53,14,215,1,0 // movaps 0x1d70e(%rip),%xmm14 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,53,142,201,1,0 // movaps 0x1c98e(%rip),%xmm14 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,94,241 // divps %xmm1,%xmm14
.byte 69,15,40,211 // movaps %xmm11,%xmm10
.byte 69,15,194,208,0 // cmpeqps %xmm8,%xmm10
@@ -37106,27 +37106,27 @@ _sk_rgb_to_hsl_sse41:
.byte 65,15,89,198 // mulps %xmm14,%xmm0
.byte 69,15,40,249 // movaps %xmm9,%xmm15
.byte 68,15,194,250,1 // cmpltps %xmm2,%xmm15
- .byte 68,15,84,61,133,216,1,0 // andps 0x1d885(%rip),%xmm15 # 3a210 <_sk_srcover_bgra_8888_sse2_lowp+0xd88>
+ .byte 68,15,84,61,5,203,1,0 // andps 0x1cb05(%rip),%xmm15 # 39490 <_sk_srcover_bgra_8888_sse2_lowp+0xd8c>
.byte 68,15,88,248 // addps %xmm0,%xmm15
.byte 65,15,40,195 // movaps %xmm11,%xmm0
.byte 65,15,194,193,0 // cmpeqps %xmm9,%xmm0
.byte 65,15,92,208 // subps %xmm8,%xmm2
.byte 65,15,89,214 // mulps %xmm14,%xmm2
- .byte 68,15,40,45,120,216,1,0 // movaps 0x1d878(%rip),%xmm13 # 3a220 <_sk_srcover_bgra_8888_sse2_lowp+0xd98>
+ .byte 68,15,40,45,248,202,1,0 // movaps 0x1caf8(%rip),%xmm13 # 394a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd9c>
.byte 65,15,88,213 // addps %xmm13,%xmm2
.byte 69,15,92,193 // subps %xmm9,%xmm8
.byte 69,15,89,198 // mulps %xmm14,%xmm8
- .byte 68,15,88,5,116,216,1,0 // addps 0x1d874(%rip),%xmm8 # 3a230 <_sk_srcover_bgra_8888_sse2_lowp+0xda8>
+ .byte 68,15,88,5,244,202,1,0 // addps 0x1caf4(%rip),%xmm8 # 394b0 <_sk_srcover_bgra_8888_sse2_lowp+0xdac>
.byte 102,68,15,56,20,194 // blendvps %xmm0,%xmm2,%xmm8
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 102,69,15,56,20,199 // blendvps %xmm0,%xmm15,%xmm8
- .byte 68,15,89,5,108,216,1,0 // mulps 0x1d86c(%rip),%xmm8 # 3a240 <_sk_srcover_bgra_8888_sse2_lowp+0xdb8>
+ .byte 68,15,89,5,236,202,1,0 // mulps 0x1caec(%rip),%xmm8 # 394c0 <_sk_srcover_bgra_8888_sse2_lowp+0xdbc>
.byte 69,15,40,203 // movaps %xmm11,%xmm9
.byte 69,15,194,204,4 // cmpneqps %xmm12,%xmm9
.byte 69,15,84,193 // andps %xmm9,%xmm8
.byte 69,15,92,235 // subps %xmm11,%xmm13
.byte 69,15,88,220 // addps %xmm12,%xmm11
- .byte 15,40,5,112,214,1,0 // movaps 0x1d670(%rip),%xmm0 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,40,5,240,200,1,0 // movaps 0x1c8f0(%rip),%xmm0 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,40,211 // movaps %xmm11,%xmm2
.byte 15,89,208 // mulps %xmm0,%xmm2
.byte 15,194,194,1 // cmpltps %xmm2,%xmm0
@@ -37148,7 +37148,7 @@ _sk_hsl_to_rgb_sse41:
.byte 15,41,100,36,184 // movaps %xmm4,-0x48(%rsp)
.byte 15,41,92,36,168 // movaps %xmm3,-0x58(%rsp)
.byte 68,15,40,208 // movaps %xmm0,%xmm10
- .byte 68,15,40,13,38,214,1,0 // movaps 0x1d626(%rip),%xmm9 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,13,166,200,1,0 // movaps 0x1c8a6(%rip),%xmm9 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,194,194,2 // cmpleps %xmm2,%xmm0
.byte 15,40,217 // movaps %xmm1,%xmm3
@@ -37161,19 +37161,19 @@ _sk_hsl_to_rgb_sse41:
.byte 15,41,84,36,152 // movaps %xmm2,-0x68(%rsp)
.byte 69,15,88,192 // addps %xmm8,%xmm8
.byte 68,15,92,197 // subps %xmm5,%xmm8
- .byte 68,15,40,53,225,215,1,0 // movaps 0x1d7e1(%rip),%xmm14 # 3a250 <_sk_srcover_bgra_8888_sse2_lowp+0xdc8>
+ .byte 68,15,40,53,97,202,1,0 // movaps 0x1ca61(%rip),%xmm14 # 394d0 <_sk_srcover_bgra_8888_sse2_lowp+0xdcc>
.byte 69,15,88,242 // addps %xmm10,%xmm14
.byte 102,65,15,58,8,198,1 // roundps $0x1,%xmm14,%xmm0
.byte 68,15,92,240 // subps %xmm0,%xmm14
- .byte 68,15,40,29,218,215,1,0 // movaps 0x1d7da(%rip),%xmm11 # 3a260 <_sk_srcover_bgra_8888_sse2_lowp+0xdd8>
+ .byte 68,15,40,29,90,202,1,0 // movaps 0x1ca5a(%rip),%xmm11 # 394e0 <_sk_srcover_bgra_8888_sse2_lowp+0xddc>
.byte 65,15,40,195 // movaps %xmm11,%xmm0
.byte 65,15,194,198,2 // cmpleps %xmm14,%xmm0
.byte 15,40,245 // movaps %xmm5,%xmm6
.byte 65,15,92,240 // subps %xmm8,%xmm6
- .byte 15,40,61,115,215,1,0 // movaps 0x1d773(%rip),%xmm7 # 3a210 <_sk_srcover_bgra_8888_sse2_lowp+0xd88>
+ .byte 15,40,61,243,201,1,0 // movaps 0x1c9f3(%rip),%xmm7 # 39490 <_sk_srcover_bgra_8888_sse2_lowp+0xd8c>
.byte 69,15,40,238 // movaps %xmm14,%xmm13
.byte 68,15,89,239 // mulps %xmm7,%xmm13
- .byte 15,40,29,132,215,1,0 // movaps 0x1d784(%rip),%xmm3 # 3a230 <_sk_srcover_bgra_8888_sse2_lowp+0xda8>
+ .byte 15,40,29,4,202,1,0 // movaps 0x1ca04(%rip),%xmm3 # 394b0 <_sk_srcover_bgra_8888_sse2_lowp+0xdac>
.byte 68,15,40,227 // movaps %xmm3,%xmm12
.byte 69,15,92,229 // subps %xmm13,%xmm12
.byte 68,15,89,230 // mulps %xmm6,%xmm12
@@ -37183,7 +37183,7 @@ _sk_hsl_to_rgb_sse41:
.byte 65,15,194,198,2 // cmpleps %xmm14,%xmm0
.byte 68,15,40,253 // movaps %xmm5,%xmm15
.byte 102,69,15,56,20,252 // blendvps %xmm0,%xmm12,%xmm15
- .byte 68,15,40,37,99,215,1,0 // movaps 0x1d763(%rip),%xmm12 # 3a240 <_sk_srcover_bgra_8888_sse2_lowp+0xdb8>
+ .byte 68,15,40,37,227,201,1,0 // movaps 0x1c9e3(%rip),%xmm12 # 394c0 <_sk_srcover_bgra_8888_sse2_lowp+0xdbc>
.byte 65,15,40,196 // movaps %xmm12,%xmm0
.byte 65,15,194,198,2 // cmpleps %xmm14,%xmm0
.byte 68,15,89,238 // mulps %xmm6,%xmm13
@@ -37217,7 +37217,7 @@ _sk_hsl_to_rgb_sse41:
.byte 65,15,40,198 // movaps %xmm14,%xmm0
.byte 15,40,84,36,152 // movaps -0x68(%rsp),%xmm2
.byte 102,15,56,20,202 // blendvps %xmm0,%xmm2,%xmm1
- .byte 68,15,88,21,251,214,1,0 // addps 0x1d6fb(%rip),%xmm10 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,88,21,123,201,1,0 // addps 0x1c97b(%rip),%xmm10 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 102,65,15,58,8,194,1 // roundps $0x1,%xmm10,%xmm0
.byte 68,15,92,208 // subps %xmm0,%xmm10
.byte 69,15,194,218,2 // cmpleps %xmm10,%xmm11
@@ -37274,9 +37274,9 @@ _sk_scale_u8_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,52 // jne 1cc5c <_sk_scale_u8_sse41+0x4c>
.byte 102,71,15,56,49,4,16 // pmovzxbd (%r8,%r10,1),%xmm8
- .byte 102,68,15,219,5,232,212,1,0 // pand 0x1d4e8(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,219,5,104,199,1,0 // pand 0x1c768(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,60,214,1,0 // mulps 0x1d63c(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,89,5,188,200,1,0 // mulps 0x1c8bc(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 65,15,89,208 // mulps %xmm8,%xmm2
@@ -37320,17 +37320,17 @@ _sk_scale_565_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,159,0,0,0 // jne 1cd76 <_sk_scale_565_sse41+0xc3>
.byte 102,71,15,56,51,28,80 // pmovzxwd (%r8,%r10,2),%xmm11
- .byte 102,15,111,5,170,213,1,0 // movdqa 0x1d5aa(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,42,200,1,0 // movdqa 0x1c82a(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,65,15,219,195 // pand %xmm11,%xmm0
.byte 68,15,91,200 // cvtdq2ps %xmm0,%xmm9
- .byte 68,15,89,13,169,213,1,0 // mulps 0x1d5a9(%rip),%xmm9 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,5,177,213,1,0 // movdqa 0x1d5b1(%rip),%xmm0 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 68,15,89,13,41,200,1,0 // mulps 0x1c829(%rip),%xmm9 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,5,49,200,1,0 // movdqa 0x1c831(%rip),%xmm0 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,65,15,219,195 // pand %xmm11,%xmm0
.byte 68,15,91,208 // cvtdq2ps %xmm0,%xmm10
- .byte 68,15,89,21,176,213,1,0 // mulps 0x1d5b0(%rip),%xmm10 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,68,15,219,29,183,213,1,0 // pand 0x1d5b7(%rip),%xmm11 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 68,15,89,21,48,200,1,0 // mulps 0x1c830(%rip),%xmm10 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,68,15,219,29,55,200,1,0 // pand 0x1c837(%rip),%xmm11 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 69,15,91,219 // cvtdq2ps %xmm11,%xmm11
- .byte 68,15,89,29,187,213,1,0 // mulps 0x1d5bb(%rip),%xmm11 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 68,15,89,29,59,200,1,0 // mulps 0x1c83b(%rip),%xmm11 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 15,40,195 // movaps %xmm3,%xmm0
.byte 15,194,199,1 // cmpltps %xmm7,%xmm0
.byte 69,15,40,226 // movaps %xmm10,%xmm12
@@ -37407,9 +37407,9 @@ _sk_lerp_u8_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,72 // jne 1ce66 <_sk_lerp_u8_sse41+0x60>
.byte 102,71,15,56,49,4,16 // pmovzxbd (%r8,%r10,1),%xmm8
- .byte 102,68,15,219,5,242,210,1,0 // pand 0x1d2f2(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,219,5,114,197,1,0 // pand 0x1c572(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,70,212,1,0 // mulps 0x1d446(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,89,5,198,198,1,0 // mulps 0x1c6c6(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 15,92,196 // subps %xmm4,%xmm0
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 15,88,196 // addps %xmm4,%xmm0
@@ -37460,17 +37460,17 @@ _sk_lerp_565_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 15,133,165,0,0,0 // jne 1cf89 <_sk_lerp_565_sse41+0xc9>
.byte 102,71,15,56,51,20,80 // pmovzxwd (%r8,%r10,2),%xmm10
- .byte 102,15,111,5,157,211,1,0 // movdqa 0x1d39d(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,29,198,1,0 // movdqa 0x1c61d(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,65,15,219,194 // pand %xmm10,%xmm0
.byte 68,15,91,200 // cvtdq2ps %xmm0,%xmm9
- .byte 68,15,89,13,156,211,1,0 // mulps 0x1d39c(%rip),%xmm9 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,5,164,211,1,0 // movdqa 0x1d3a4(%rip),%xmm0 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 68,15,89,13,28,198,1,0 // mulps 0x1c61c(%rip),%xmm9 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,5,36,198,1,0 // movdqa 0x1c624(%rip),%xmm0 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,65,15,219,194 // pand %xmm10,%xmm0
.byte 68,15,91,216 // cvtdq2ps %xmm0,%xmm11
- .byte 68,15,89,29,163,211,1,0 // mulps 0x1d3a3(%rip),%xmm11 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,68,15,219,21,170,211,1,0 // pand 0x1d3aa(%rip),%xmm10 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 68,15,89,29,35,198,1,0 // mulps 0x1c623(%rip),%xmm11 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,68,15,219,21,42,198,1,0 // pand 0x1c62a(%rip),%xmm10 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
- .byte 68,15,89,21,174,211,1,0 // mulps 0x1d3ae(%rip),%xmm10 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 68,15,89,21,46,198,1,0 // mulps 0x1c62e(%rip),%xmm10 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 15,40,195 // movaps %xmm3,%xmm0
.byte 15,194,199,1 // cmpltps %xmm7,%xmm0
.byte 69,15,40,227 // movaps %xmm11,%xmm12
@@ -37526,7 +37526,7 @@ _sk_load_tables_sse41:
.byte 243,69,15,111,4,144 // movdqu (%r8,%rdx,4),%xmm8
.byte 65,86 // push %r14
.byte 83 // push %rbx
- .byte 102,15,111,5,32,209,1,0 // movdqa 0x1d120(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,160,195,1,0 // movdqa 0x1c3a0(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,219,192 // pand %xmm8,%xmm0
.byte 102,73,15,58,22,192,1 // pextrq $0x1,%xmm0,%r8
.byte 102,73,15,126,193 // movq %xmm0,%r9
@@ -37541,7 +37541,7 @@ _sk_load_tables_sse41:
.byte 102,66,15,58,33,4,155,32 // insertps $0x20,(%rbx,%r11,4),%xmm0
.byte 102,66,15,58,33,4,3,48 // insertps $0x30,(%rbx,%r8,1),%xmm0
.byte 102,65,15,111,200 // movdqa %xmm8,%xmm1
- .byte 102,15,56,0,13,219,208,1,0 // pshufb 0x1d0db(%rip),%xmm1 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,13,91,195,1,0 // pshufb 0x1c35b(%rip),%xmm1 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 102,73,15,58,22,200,1 // pextrq $0x1,%xmm1,%r8
.byte 102,72,15,126,203 // movq %xmm1,%rbx
.byte 68,15,182,203 // movzbl %bl,%r9d
@@ -37556,7 +37556,7 @@ _sk_load_tables_sse41:
.byte 102,15,58,33,202,48 // insertps $0x30,%xmm2,%xmm1
.byte 76,139,64,24 // mov 0x18(%rax),%r8
.byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
- .byte 102,15,56,0,21,151,208,1,0 // pshufb 0x1d097(%rip),%xmm2 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,21,23,195,1,0 // pshufb 0x1c317(%rip),%xmm2 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 102,72,15,58,22,211,1 // pextrq $0x1,%xmm2,%rbx
.byte 102,72,15,126,208 // movq %xmm2,%rax
.byte 68,15,182,200 // movzbl %al,%r9d
@@ -37571,7 +37571,7 @@ _sk_load_tables_sse41:
.byte 102,15,58,33,211,48 // insertps $0x30,%xmm3,%xmm2
.byte 102,65,15,114,208,24 // psrld $0x18,%xmm8
.byte 65,15,91,216 // cvtdq2ps %xmm8,%xmm3
- .byte 15,89,29,132,209,1,0 // mulps 0x1d184(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,4,196,1,0 // mulps 0x1c404(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
@@ -37612,7 +37612,7 @@ _sk_load_tables_u16_be_sse41:
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,97,200 // punpcklwd %xmm0,%xmm1
.byte 102,68,15,105,200 // punpckhwd %xmm0,%xmm9
- .byte 102,68,15,111,5,86,209,1,0 // movdqa 0x1d156(%rip),%xmm8 # 3a2f0 <_sk_srcover_bgra_8888_sse2_lowp+0xe68>
+ .byte 102,68,15,111,5,214,195,1,0 // movdqa 0x1c3d6(%rip),%xmm8 # 39570 <_sk_srcover_bgra_8888_sse2_lowp+0xe6c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,192 // pand %xmm8,%xmm0
.byte 102,15,56,51,192 // pmovzxwd %xmm0,%xmm0
@@ -37630,7 +37630,7 @@ _sk_load_tables_u16_be_sse41:
.byte 102,15,58,33,194,32 // insertps $0x20,%xmm2,%xmm0
.byte 243,66,15,16,20,3 // movss (%rbx,%r8,1),%xmm2
.byte 102,15,58,33,194,48 // insertps $0x30,%xmm2,%xmm0
- .byte 102,15,56,0,13,5,209,1,0 // pshufb 0x1d105(%rip),%xmm1 # 3a300 <_sk_srcover_bgra_8888_sse2_lowp+0xe78>
+ .byte 102,15,56,0,13,133,195,1,0 // pshufb 0x1c385(%rip),%xmm1 # 39580 <_sk_srcover_bgra_8888_sse2_lowp+0xe7c>
.byte 102,15,56,51,201 // pmovzxwd %xmm1,%xmm1
.byte 102,73,15,58,22,200,1 // pextrq $0x1,%xmm1,%r8
.byte 102,72,15,126,203 // movq %xmm1,%rbx
@@ -37666,7 +37666,7 @@ _sk_load_tables_u16_be_sse41:
.byte 102,65,15,235,216 // por %xmm8,%xmm3
.byte 102,15,56,51,219 // pmovzxwd %xmm3,%xmm3
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,83,208,1,0 // mulps 0x1d053(%rip),%xmm3 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 15,89,29,211,194,1,0 // mulps 0x1c2d3(%rip),%xmm3 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
@@ -37706,7 +37706,7 @@ _sk_load_tables_rgb_u16_be_sse41:
.byte 102,15,97,209 // punpcklwd %xmm1,%xmm2
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,65,15,97,201 // punpcklwd %xmm9,%xmm1
- .byte 102,68,15,111,5,156,207,1,0 // movdqa 0x1cf9c(%rip),%xmm8 # 3a2f0 <_sk_srcover_bgra_8888_sse2_lowp+0xe68>
+ .byte 102,68,15,111,5,28,194,1,0 // movdqa 0x1c21c(%rip),%xmm8 # 39570 <_sk_srcover_bgra_8888_sse2_lowp+0xe6c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,192 // pand %xmm8,%xmm0
.byte 102,15,56,51,192 // pmovzxwd %xmm0,%xmm0
@@ -37724,7 +37724,7 @@ _sk_load_tables_rgb_u16_be_sse41:
.byte 102,15,58,33,195,32 // insertps $0x20,%xmm3,%xmm0
.byte 243,66,15,16,28,3 // movss (%rbx,%r8,1),%xmm3
.byte 102,15,58,33,195,48 // insertps $0x30,%xmm3,%xmm0
- .byte 102,15,56,0,13,75,207,1,0 // pshufb 0x1cf4b(%rip),%xmm1 # 3a300 <_sk_srcover_bgra_8888_sse2_lowp+0xe78>
+ .byte 102,15,56,0,13,203,193,1,0 // pshufb 0x1c1cb(%rip),%xmm1 # 39580 <_sk_srcover_bgra_8888_sse2_lowp+0xe7c>
.byte 102,15,56,51,201 // pmovzxwd %xmm1,%xmm1
.byte 102,73,15,58,22,200,1 // pextrq $0x1,%xmm1,%r8
.byte 102,72,15,126,203 // movq %xmm1,%rbx
@@ -37755,7 +37755,7 @@ _sk_load_tables_rgb_u16_be_sse41:
.byte 243,65,15,16,28,24 // movss (%r8,%rbx,1),%xmm3
.byte 102,15,58,33,211,48 // insertps $0x30,%xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,22,204,1,0 // movaps 0x1cc16(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,150,190,1,0 // movaps 0x1be96(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
.byte 255,224 // jmpq *%rax
@@ -37787,7 +37787,7 @@ _sk_byte_tables_sse41:
.byte 65,86 // push %r14
.byte 83 // push %rbx
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,131,204,1,0 // movaps 0x1cc83(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,3,191,1,0 // movaps 0x1bf03(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,91,192 // cvtps2dq %xmm0,%xmm0
.byte 102,73,15,58,22,192,1 // pextrq $0x1,%xmm0,%r8
@@ -37810,7 +37810,7 @@ _sk_byte_tables_sse41:
.byte 102,15,58,32,197,3 // pinsrb $0x3,%ebp,%xmm0
.byte 102,15,56,49,192 // pmovzxbd %xmm0,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,13,73,205,1,0 // movaps 0x1cd49(%rip),%xmm9 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,13,201,191,1,0 // movaps 0x1bfc9(%rip),%xmm9 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,91,201 // cvtps2dq %xmm1,%xmm1
@@ -37912,7 +37912,7 @@ _sk_byte_tables_rgb_sse41:
.byte 102,15,58,32,197,3 // pinsrb $0x3,%ebp,%xmm0
.byte 102,15,56,49,192 // pmovzxbd %xmm0,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,13,165,203,1,0 // movaps 0x1cba5(%rip),%xmm9 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,13,37,190,1,0 // movaps 0x1be25(%rip),%xmm9 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,91,201 // cvtps2dq %xmm1,%xmm1
@@ -38094,31 +38094,31 @@ _sk_parametric_r_sse41:
.byte 69,15,88,208 // addps %xmm8,%xmm10
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,91,194 // cvtdq2ps %xmm10,%xmm8
- .byte 68,15,89,5,128,201,1,0 // mulps 0x1c980(%rip),%xmm8 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,21,136,201,1,0 // andps 0x1c988(%rip),%xmm10 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,21,176,198,1,0 // orps 0x1c6b0(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,5,136,201,1,0 // addps 0x1c988(%rip),%xmm8 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 68,15,40,37,144,201,1,0 // movaps 0x1c990(%rip),%xmm12 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,5,0,188,1,0 // mulps 0x1bc00(%rip),%xmm8 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,21,8,188,1,0 // andps 0x1bc08(%rip),%xmm10 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,21,48,185,1,0 // orps 0x1b930(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,5,8,188,1,0 // addps 0x1bc08(%rip),%xmm8 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 68,15,40,37,16,188,1,0 // movaps 0x1bc10(%rip),%xmm12 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 69,15,92,196 // subps %xmm12,%xmm8
- .byte 68,15,88,21,144,201,1,0 // addps 0x1c990(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 68,15,40,37,152,201,1,0 // movaps 0x1c998(%rip),%xmm12 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,21,16,188,1,0 // addps 0x1bc10(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 68,15,40,37,24,188,1,0 // movaps 0x1bc18(%rip),%xmm12 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,92,196 // subps %xmm12,%xmm8
.byte 69,15,89,195 // mulps %xmm11,%xmm8
.byte 102,69,15,58,8,208,1 // roundps $0x1,%xmm8,%xmm10
.byte 69,15,40,216 // movaps %xmm8,%xmm11
.byte 69,15,92,218 // subps %xmm10,%xmm11
- .byte 68,15,88,5,133,201,1,0 // addps 0x1c985(%rip),%xmm8 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 68,15,40,21,141,201,1,0 // movaps 0x1c98d(%rip),%xmm10 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,5,5,188,1,0 // addps 0x1bc05(%rip),%xmm8 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 68,15,40,21,13,188,1,0 // movaps 0x1bc0d(%rip),%xmm10 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 69,15,89,211 // mulps %xmm11,%xmm10
.byte 69,15,92,194 // subps %xmm10,%xmm8
- .byte 68,15,40,21,141,201,1,0 // movaps 0x1c98d(%rip),%xmm10 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,21,13,188,1,0 // movaps 0x1bc0d(%rip),%xmm10 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,211 // subps %xmm11,%xmm10
- .byte 68,15,40,29,145,201,1,0 // movaps 0x1c991(%rip),%xmm11 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,29,17,188,1,0 // movaps 0x1bc11(%rip),%xmm11 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 69,15,94,218 // divps %xmm10,%xmm11
.byte 69,15,88,216 // addps %xmm8,%xmm11
- .byte 68,15,89,29,145,201,1,0 // mulps 0x1c991(%rip),%xmm11 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 68,15,89,29,17,188,1,0 // mulps 0x1bc11(%rip),%xmm11 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,69,15,91,211 // cvtps2dq %xmm11,%xmm10
.byte 243,68,15,16,64,20 // movss 0x14(%rax),%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
@@ -38126,7 +38126,7 @@ _sk_parametric_r_sse41:
.byte 102,69,15,56,20,193 // blendvps %xmm0,%xmm9,%xmm8
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 68,15,95,192 // maxps %xmm0,%xmm8
- .byte 68,15,93,5,24,198,1,0 // minps 0x1c618(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,93,5,152,184,1,0 // minps 0x1b898(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -38156,31 +38156,31 @@ _sk_parametric_g_sse41:
.byte 68,15,88,217 // addps %xmm1,%xmm11
.byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
.byte 69,15,91,227 // cvtdq2ps %xmm11,%xmm12
- .byte 68,15,89,37,89,200,1,0 // mulps 0x1c859(%rip),%xmm12 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,29,97,200,1,0 // andps 0x1c861(%rip),%xmm11 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,29,137,197,1,0 // orps 0x1c589(%rip),%xmm11 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,37,97,200,1,0 // addps 0x1c861(%rip),%xmm12 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 15,40,13,106,200,1,0 // movaps 0x1c86a(%rip),%xmm1 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,37,217,186,1,0 // mulps 0x1bad9(%rip),%xmm12 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,29,225,186,1,0 // andps 0x1bae1(%rip),%xmm11 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,29,9,184,1,0 // orps 0x1b809(%rip),%xmm11 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,37,225,186,1,0 // addps 0x1bae1(%rip),%xmm12 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 15,40,13,234,186,1,0 // movaps 0x1baea(%rip),%xmm1 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 65,15,89,203 // mulps %xmm11,%xmm1
.byte 68,15,92,225 // subps %xmm1,%xmm12
- .byte 68,15,88,29,106,200,1,0 // addps 0x1c86a(%rip),%xmm11 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 15,40,13,115,200,1,0 // movaps 0x1c873(%rip),%xmm1 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,29,234,186,1,0 // addps 0x1baea(%rip),%xmm11 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 15,40,13,243,186,1,0 // movaps 0x1baf3(%rip),%xmm1 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 65,15,94,203 // divps %xmm11,%xmm1
.byte 68,15,92,225 // subps %xmm1,%xmm12
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 102,69,15,58,8,212,1 // roundps $0x1,%xmm12,%xmm10
.byte 69,15,40,220 // movaps %xmm12,%xmm11
.byte 69,15,92,218 // subps %xmm10,%xmm11
- .byte 68,15,88,37,96,200,1,0 // addps 0x1c860(%rip),%xmm12 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 15,40,13,105,200,1,0 // movaps 0x1c869(%rip),%xmm1 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,37,224,186,1,0 // addps 0x1bae0(%rip),%xmm12 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 15,40,13,233,186,1,0 // movaps 0x1bae9(%rip),%xmm1 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 65,15,89,203 // mulps %xmm11,%xmm1
.byte 68,15,92,225 // subps %xmm1,%xmm12
- .byte 68,15,40,21,105,200,1,0 // movaps 0x1c869(%rip),%xmm10 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,21,233,186,1,0 // movaps 0x1bae9(%rip),%xmm10 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,211 // subps %xmm11,%xmm10
- .byte 15,40,13,110,200,1,0 // movaps 0x1c86e(%rip),%xmm1 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 15,40,13,238,186,1,0 // movaps 0x1baee(%rip),%xmm1 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 65,15,94,202 // divps %xmm10,%xmm1
.byte 65,15,88,204 // addps %xmm12,%xmm1
- .byte 15,89,13,111,200,1,0 // mulps 0x1c86f(%rip),%xmm1 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 15,89,13,239,186,1,0 // mulps 0x1baef(%rip),%xmm1 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,68,15,91,209 // cvtps2dq %xmm1,%xmm10
.byte 243,15,16,72,20 // movss 0x14(%rax),%xmm1
.byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
@@ -38188,7 +38188,7 @@ _sk_parametric_g_sse41:
.byte 102,65,15,56,20,201 // blendvps %xmm0,%xmm9,%xmm1
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,95,200 // maxps %xmm0,%xmm1
- .byte 15,93,13,250,196,1,0 // minps 0x1c4fa(%rip),%xmm1 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,13,122,183,1,0 // minps 0x1b77a(%rip),%xmm1 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -38218,31 +38218,31 @@ _sk_parametric_b_sse41:
.byte 68,15,88,218 // addps %xmm2,%xmm11
.byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
.byte 69,15,91,227 // cvtdq2ps %xmm11,%xmm12
- .byte 68,15,89,37,59,199,1,0 // mulps 0x1c73b(%rip),%xmm12 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,29,67,199,1,0 // andps 0x1c743(%rip),%xmm11 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,29,107,196,1,0 // orps 0x1c46b(%rip),%xmm11 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,37,67,199,1,0 // addps 0x1c743(%rip),%xmm12 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 15,40,21,76,199,1,0 // movaps 0x1c74c(%rip),%xmm2 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,37,187,185,1,0 // mulps 0x1b9bb(%rip),%xmm12 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,29,195,185,1,0 // andps 0x1b9c3(%rip),%xmm11 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,29,235,182,1,0 // orps 0x1b6eb(%rip),%xmm11 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,37,195,185,1,0 // addps 0x1b9c3(%rip),%xmm12 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 15,40,21,204,185,1,0 // movaps 0x1b9cc(%rip),%xmm2 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 65,15,89,211 // mulps %xmm11,%xmm2
.byte 68,15,92,226 // subps %xmm2,%xmm12
- .byte 68,15,88,29,76,199,1,0 // addps 0x1c74c(%rip),%xmm11 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 15,40,21,85,199,1,0 // movaps 0x1c755(%rip),%xmm2 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,29,204,185,1,0 // addps 0x1b9cc(%rip),%xmm11 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 15,40,21,213,185,1,0 // movaps 0x1b9d5(%rip),%xmm2 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 65,15,94,211 // divps %xmm11,%xmm2
.byte 68,15,92,226 // subps %xmm2,%xmm12
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 102,69,15,58,8,212,1 // roundps $0x1,%xmm12,%xmm10
.byte 69,15,40,220 // movaps %xmm12,%xmm11
.byte 69,15,92,218 // subps %xmm10,%xmm11
- .byte 68,15,88,37,66,199,1,0 // addps 0x1c742(%rip),%xmm12 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 15,40,21,75,199,1,0 // movaps 0x1c74b(%rip),%xmm2 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,37,194,185,1,0 // addps 0x1b9c2(%rip),%xmm12 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 15,40,21,203,185,1,0 // movaps 0x1b9cb(%rip),%xmm2 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 65,15,89,211 // mulps %xmm11,%xmm2
.byte 68,15,92,226 // subps %xmm2,%xmm12
- .byte 68,15,40,21,75,199,1,0 // movaps 0x1c74b(%rip),%xmm10 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,21,203,185,1,0 // movaps 0x1b9cb(%rip),%xmm10 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,211 // subps %xmm11,%xmm10
- .byte 15,40,21,80,199,1,0 // movaps 0x1c750(%rip),%xmm2 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 15,40,21,208,185,1,0 // movaps 0x1b9d0(%rip),%xmm2 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 65,15,94,210 // divps %xmm10,%xmm2
.byte 65,15,88,212 // addps %xmm12,%xmm2
- .byte 15,89,21,81,199,1,0 // mulps 0x1c751(%rip),%xmm2 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 15,89,21,209,185,1,0 // mulps 0x1b9d1(%rip),%xmm2 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,68,15,91,210 // cvtps2dq %xmm2,%xmm10
.byte 243,15,16,80,20 // movss 0x14(%rax),%xmm2
.byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
@@ -38250,7 +38250,7 @@ _sk_parametric_b_sse41:
.byte 102,65,15,56,20,209 // blendvps %xmm0,%xmm9,%xmm2
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,95,208 // maxps %xmm0,%xmm2
- .byte 15,93,21,220,195,1,0 // minps 0x1c3dc(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,21,92,182,1,0 // minps 0x1b65c(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -38280,31 +38280,31 @@ _sk_parametric_a_sse41:
.byte 68,15,88,219 // addps %xmm3,%xmm11
.byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
.byte 69,15,91,227 // cvtdq2ps %xmm11,%xmm12
- .byte 68,15,89,37,29,198,1,0 // mulps 0x1c61d(%rip),%xmm12 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,29,37,198,1,0 // andps 0x1c625(%rip),%xmm11 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,29,77,195,1,0 // orps 0x1c34d(%rip),%xmm11 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,37,37,198,1,0 // addps 0x1c625(%rip),%xmm12 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 15,40,29,46,198,1,0 // movaps 0x1c62e(%rip),%xmm3 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,37,157,184,1,0 // mulps 0x1b89d(%rip),%xmm12 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,29,165,184,1,0 // andps 0x1b8a5(%rip),%xmm11 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,29,205,181,1,0 // orps 0x1b5cd(%rip),%xmm11 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,37,165,184,1,0 // addps 0x1b8a5(%rip),%xmm12 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 15,40,29,174,184,1,0 // movaps 0x1b8ae(%rip),%xmm3 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 65,15,89,219 // mulps %xmm11,%xmm3
.byte 68,15,92,227 // subps %xmm3,%xmm12
- .byte 68,15,88,29,46,198,1,0 // addps 0x1c62e(%rip),%xmm11 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 15,40,29,55,198,1,0 // movaps 0x1c637(%rip),%xmm3 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,29,174,184,1,0 // addps 0x1b8ae(%rip),%xmm11 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 15,40,29,183,184,1,0 // movaps 0x1b8b7(%rip),%xmm3 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 65,15,94,219 // divps %xmm11,%xmm3
.byte 68,15,92,227 // subps %xmm3,%xmm12
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 102,69,15,58,8,212,1 // roundps $0x1,%xmm12,%xmm10
.byte 69,15,40,220 // movaps %xmm12,%xmm11
.byte 69,15,92,218 // subps %xmm10,%xmm11
- .byte 68,15,88,37,36,198,1,0 // addps 0x1c624(%rip),%xmm12 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 15,40,29,45,198,1,0 // movaps 0x1c62d(%rip),%xmm3 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,37,164,184,1,0 // addps 0x1b8a4(%rip),%xmm12 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 15,40,29,173,184,1,0 // movaps 0x1b8ad(%rip),%xmm3 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 65,15,89,219 // mulps %xmm11,%xmm3
.byte 68,15,92,227 // subps %xmm3,%xmm12
- .byte 68,15,40,21,45,198,1,0 // movaps 0x1c62d(%rip),%xmm10 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,21,173,184,1,0 // movaps 0x1b8ad(%rip),%xmm10 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,211 // subps %xmm11,%xmm10
- .byte 15,40,29,50,198,1,0 // movaps 0x1c632(%rip),%xmm3 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 15,40,29,178,184,1,0 // movaps 0x1b8b2(%rip),%xmm3 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 65,15,94,218 // divps %xmm10,%xmm3
.byte 65,15,88,220 // addps %xmm12,%xmm3
- .byte 15,89,29,51,198,1,0 // mulps 0x1c633(%rip),%xmm3 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 15,89,29,179,184,1,0 // mulps 0x1b8b3(%rip),%xmm3 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,68,15,91,211 // cvtps2dq %xmm3,%xmm10
.byte 243,15,16,88,20 // movss 0x14(%rax),%xmm3
.byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
@@ -38312,7 +38312,7 @@ _sk_parametric_a_sse41:
.byte 102,65,15,56,20,217 // blendvps %xmm0,%xmm9,%xmm3
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,95,216 // maxps %xmm0,%xmm3
- .byte 15,93,29,190,194,1,0 // minps 0x1c2be(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,29,62,181,1,0 // minps 0x1b53e(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -38329,20 +38329,20 @@ _sk_gamma_sse41:
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,91,194 // cvtdq2ps %xmm2,%xmm0
- .byte 15,89,5,61,197,1,0 // mulps 0x1c53d(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,40,53,69,197,1,0 // movaps 0x1c545(%rip),%xmm14 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
+ .byte 15,89,5,189,183,1,0 // mulps 0x1b7bd(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,40,53,197,183,1,0 // movaps 0x1b7c5(%rip),%xmm14 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
.byte 65,15,84,214 // andps %xmm14,%xmm2
- .byte 68,15,40,37,105,194,1,0 // movaps 0x1c269(%rip),%xmm12 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,37,233,180,1,0 // movaps 0x1b4e9(%rip),%xmm12 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,86,212 // orps %xmm12,%xmm2
- .byte 68,15,40,21,61,197,1,0 // movaps 0x1c53d(%rip),%xmm10 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
+ .byte 68,15,40,21,189,183,1,0 // movaps 0x1b7bd(%rip),%xmm10 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
.byte 65,15,88,194 // addps %xmm10,%xmm0
- .byte 68,15,40,29,65,197,1,0 // movaps 0x1c541(%rip),%xmm11 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,40,29,193,183,1,0 // movaps 0x1b7c1(%rip),%xmm11 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 15,40,226 // movaps %xmm2,%xmm4
.byte 65,15,89,227 // mulps %xmm11,%xmm4
.byte 15,92,196 // subps %xmm4,%xmm0
- .byte 68,15,40,13,63,197,1,0 // movaps 0x1c53f(%rip),%xmm9 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
+ .byte 68,15,40,13,191,183,1,0 // movaps 0x1b7bf(%rip),%xmm9 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
.byte 65,15,88,209 // addps %xmm9,%xmm2
- .byte 68,15,40,45,67,197,1,0 // movaps 0x1c543(%rip),%xmm13 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,40,45,195,183,1,0 // movaps 0x1b7c3(%rip),%xmm13 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 65,15,40,229 // movaps %xmm13,%xmm4
.byte 15,94,226 // divps %xmm2,%xmm4
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -38353,21 +38353,21 @@ _sk_gamma_sse41:
.byte 102,15,58,8,208,1 // roundps $0x1,%xmm0,%xmm2
.byte 15,40,224 // movaps %xmm0,%xmm4
.byte 15,92,226 // subps %xmm2,%xmm4
- .byte 15,40,53,38,197,1,0 // movaps 0x1c526(%rip),%xmm6 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
+ .byte 15,40,53,166,183,1,0 // movaps 0x1b7a6(%rip),%xmm6 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
.byte 15,88,198 // addps %xmm6,%xmm0
- .byte 15,40,61,60,197,1,0 // movaps 0x1c53c(%rip),%xmm7 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 15,40,61,188,183,1,0 // movaps 0x1b7bc(%rip),%xmm7 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 15,40,239 // movaps %xmm7,%xmm5
.byte 15,92,236 // subps %xmm4,%xmm5
.byte 15,40,212 // movaps %xmm4,%xmm2
- .byte 15,40,37,28,197,1,0 // movaps 0x1c51c(%rip),%xmm4 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 15,40,37,156,183,1,0 // movaps 0x1b79c(%rip),%xmm4 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 15,89,212 // mulps %xmm4,%xmm2
.byte 15,92,194 // subps %xmm2,%xmm0
- .byte 68,15,40,61,46,197,1,0 // movaps 0x1c52e(%rip),%xmm15 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,61,174,183,1,0 // movaps 0x1b7ae(%rip),%xmm15 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 65,15,40,215 // movaps %xmm15,%xmm2
.byte 15,94,213 // divps %xmm5,%xmm2
.byte 15,88,208 // addps %xmm0,%xmm2
.byte 15,91,193 // cvtdq2ps %xmm1,%xmm0
- .byte 15,89,5,138,196,1,0 // mulps 0x1c48a(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
+ .byte 15,89,5,10,183,1,0 // mulps 0x1b70a(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
.byte 65,15,84,206 // andps %xmm14,%xmm1
.byte 65,15,86,204 // orps %xmm12,%xmm1
.byte 65,15,88,194 // addps %xmm10,%xmm0
@@ -38393,8 +38393,8 @@ _sk_gamma_sse41:
.byte 15,94,206 // divps %xmm6,%xmm1
.byte 15,88,200 // addps %xmm0,%xmm1
.byte 15,91,195 // cvtdq2ps %xmm3,%xmm0
- .byte 15,89,5,43,196,1,0 // mulps 0x1c42b(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 15,84,29,52,196,1,0 // andps 0x1c434(%rip),%xmm3 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
+ .byte 15,89,5,171,182,1,0 // mulps 0x1b6ab(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 15,84,29,180,182,1,0 // andps 0x1b6b4(%rip),%xmm3 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
.byte 65,15,86,220 // orps %xmm12,%xmm3
.byte 65,15,88,194 // addps %xmm10,%xmm0
.byte 68,15,89,219 // mulps %xmm3,%xmm11
@@ -38412,7 +38412,7 @@ _sk_gamma_sse41:
.byte 15,92,253 // subps %xmm5,%xmm7
.byte 68,15,94,255 // divps %xmm7,%xmm15
.byte 68,15,88,248 // addps %xmm0,%xmm15
- .byte 15,40,5,124,196,1,0 // movaps 0x1c47c(%rip),%xmm0 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 15,40,5,252,182,1,0 // movaps 0x1b6fc(%rip),%xmm0 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 15,89,208 // mulps %xmm0,%xmm2
.byte 15,89,200 // mulps %xmm0,%xmm1
.byte 68,15,89,248 // mulps %xmm0,%xmm15
@@ -38432,29 +38432,29 @@ HIDDEN _sk_lab_to_xyz_sse41
FUNCTION(_sk_lab_to_xyz_sse41)
_sk_lab_to_xyz_sse41:
.byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 68,15,89,5,76,196,1,0 // mulps 0x1c44c(%rip),%xmm8 # 3a3d0 <_sk_srcover_bgra_8888_sse2_lowp+0xf48>
- .byte 68,15,40,13,196,193,1,0 // movaps 0x1c1c4(%rip),%xmm9 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,89,5,204,182,1,0 // mulps 0x1b6cc(%rip),%xmm8 # 39650 <_sk_srcover_bgra_8888_sse2_lowp+0xf4c>
+ .byte 68,15,40,13,68,180,1,0 // movaps 0x1b444(%rip),%xmm9 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,201 // mulps %xmm9,%xmm1
- .byte 15,40,5,73,196,1,0 // movaps 0x1c449(%rip),%xmm0 # 3a3e0 <_sk_srcover_bgra_8888_sse2_lowp+0xf58>
+ .byte 15,40,5,201,182,1,0 // movaps 0x1b6c9(%rip),%xmm0 # 39660 <_sk_srcover_bgra_8888_sse2_lowp+0xf5c>
.byte 15,88,200 // addps %xmm0,%xmm1
.byte 65,15,89,209 // mulps %xmm9,%xmm2
.byte 15,88,208 // addps %xmm0,%xmm2
- .byte 68,15,88,5,71,196,1,0 // addps 0x1c447(%rip),%xmm8 # 3a3f0 <_sk_srcover_bgra_8888_sse2_lowp+0xf68>
- .byte 68,15,89,5,79,196,1,0 // mulps 0x1c44f(%rip),%xmm8 # 3a400 <_sk_srcover_bgra_8888_sse2_lowp+0xf78>
- .byte 15,89,13,88,196,1,0 // mulps 0x1c458(%rip),%xmm1 # 3a410 <_sk_srcover_bgra_8888_sse2_lowp+0xf88>
+ .byte 68,15,88,5,199,182,1,0 // addps 0x1b6c7(%rip),%xmm8 # 39670 <_sk_srcover_bgra_8888_sse2_lowp+0xf6c>
+ .byte 68,15,89,5,207,182,1,0 // mulps 0x1b6cf(%rip),%xmm8 # 39680 <_sk_srcover_bgra_8888_sse2_lowp+0xf7c>
+ .byte 15,89,13,216,182,1,0 // mulps 0x1b6d8(%rip),%xmm1 # 39690 <_sk_srcover_bgra_8888_sse2_lowp+0xf8c>
.byte 65,15,88,200 // addps %xmm8,%xmm1
- .byte 15,89,21,93,196,1,0 // mulps 0x1c45d(%rip),%xmm2 # 3a420 <_sk_srcover_bgra_8888_sse2_lowp+0xf98>
+ .byte 15,89,21,221,182,1,0 // mulps 0x1b6dd(%rip),%xmm2 # 396a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf9c>
.byte 69,15,40,208 // movaps %xmm8,%xmm10
.byte 68,15,92,210 // subps %xmm2,%xmm10
.byte 68,15,40,217 // movaps %xmm1,%xmm11
.byte 69,15,89,219 // mulps %xmm11,%xmm11
.byte 68,15,89,217 // mulps %xmm1,%xmm11
- .byte 68,15,40,13,81,196,1,0 // movaps 0x1c451(%rip),%xmm9 # 3a430 <_sk_srcover_bgra_8888_sse2_lowp+0xfa8>
+ .byte 68,15,40,13,209,182,1,0 // movaps 0x1b6d1(%rip),%xmm9 # 396b0 <_sk_srcover_bgra_8888_sse2_lowp+0xfac>
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 65,15,194,195,1 // cmpltps %xmm11,%xmm0
- .byte 15,40,21,81,196,1,0 // movaps 0x1c451(%rip),%xmm2 # 3a440 <_sk_srcover_bgra_8888_sse2_lowp+0xfb8>
+ .byte 15,40,21,209,182,1,0 // movaps 0x1b6d1(%rip),%xmm2 # 396c0 <_sk_srcover_bgra_8888_sse2_lowp+0xfbc>
.byte 15,88,202 // addps %xmm2,%xmm1
- .byte 68,15,40,37,86,196,1,0 // movaps 0x1c456(%rip),%xmm12 # 3a450 <_sk_srcover_bgra_8888_sse2_lowp+0xfc8>
+ .byte 68,15,40,37,214,182,1,0 // movaps 0x1b6d6(%rip),%xmm12 # 396d0 <_sk_srcover_bgra_8888_sse2_lowp+0xfcc>
.byte 65,15,89,204 // mulps %xmm12,%xmm1
.byte 102,65,15,56,20,203 // blendvps %xmm0,%xmm11,%xmm1
.byte 69,15,40,216 // movaps %xmm8,%xmm11
@@ -38473,8 +38473,8 @@ _sk_lab_to_xyz_sse41:
.byte 65,15,89,212 // mulps %xmm12,%xmm2
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 102,65,15,56,20,211 // blendvps %xmm0,%xmm11,%xmm2
- .byte 15,89,13,15,196,1,0 // mulps 0x1c40f(%rip),%xmm1 # 3a460 <_sk_srcover_bgra_8888_sse2_lowp+0xfd8>
- .byte 15,89,21,24,196,1,0 // mulps 0x1c418(%rip),%xmm2 # 3a470 <_sk_srcover_bgra_8888_sse2_lowp+0xfe8>
+ .byte 15,89,13,143,182,1,0 // mulps 0x1b68f(%rip),%xmm1 # 396e0 <_sk_srcover_bgra_8888_sse2_lowp+0xfdc>
+ .byte 15,89,21,152,182,1,0 // mulps 0x1b698(%rip),%xmm2 # 396f0 <_sk_srcover_bgra_8888_sse2_lowp+0xfec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,40,193 // movaps %xmm1,%xmm0
.byte 65,15,40,200 // movaps %xmm8,%xmm1
@@ -38493,9 +38493,9 @@ _sk_load_a8_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,39 // jne 1e0a2 <_sk_load_a8_sse41+0x3f>
.byte 102,67,15,56,49,4,16 // pmovzxbd (%r8,%r10,1),%xmm0
- .byte 102,15,219,5,150,192,1,0 // pand 0x1c096(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,5,22,179,1,0 // pand 0x1b316(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,216 // cvtdq2ps %xmm0,%xmm3
- .byte 15,89,29,236,193,1,0 // mulps 0x1c1ec(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,108,180,1,0 // mulps 0x1b46c(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 102,15,239,201 // pxor %xmm1,%xmm1
@@ -38535,9 +38535,9 @@ _sk_load_a8_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,39 // jne 1e12e <_sk_load_a8_dst_sse41+0x3f>
.byte 102,67,15,56,49,36,16 // pmovzxbd (%r8,%r10,1),%xmm4
- .byte 102,15,219,37,10,192,1,0 // pand 0x1c00a(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,37,138,178,1,0 // pand 0x1b28a(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,252 // cvtdq2ps %xmm4,%xmm7
- .byte 15,89,61,96,193,1,0 // mulps 0x1c160(%rip),%xmm7 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,61,224,179,1,0 // mulps 0x1b3e0(%rip),%xmm7 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,228 // xorps %xmm4,%xmm4
.byte 102,15,239,237 // pxor %xmm5,%xmm5
@@ -38605,7 +38605,7 @@ _sk_gather_a8_sse41:
.byte 102,15,58,32,192,3 // pinsrb $0x3,%eax,%xmm0
.byte 102,15,56,49,192 // pmovzxbd %xmm0,%xmm0
.byte 15,91,216 // cvtdq2ps %xmm0,%xmm3
- .byte 15,89,29,88,192,1,0 // mulps 0x1c058(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,216,178,1,0 // mulps 0x1b2d8(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 102,15,239,201 // pxor %xmm1,%xmm1
@@ -38622,7 +38622,7 @@ _sk_store_a8_sse41:
.byte 77,15,175,193 // imul %r9,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,210 // movslq %edx,%r10
- .byte 68,15,40,5,255,190,1,0 // movaps 0x1beff(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,127,177,1,0 // movaps 0x1b17f(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,89,195 // mulps %xmm3,%xmm8
.byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
.byte 102,68,15,56,43,192 // packusdw %xmm0,%xmm8
@@ -38643,7 +38643,7 @@ _sk_store_a8_sse41:
.byte 65,128,249,3 // cmp $0x3,%r9b
.byte 117,221 // jne 1e273 <_sk_store_a8_sse41+0x3d>
.byte 102,71,15,58,20,68,16,2,8 // pextrb $0x8,%xmm8,0x2(%r8,%r10,1)
- .byte 102,68,15,56,0,5,215,193,1,0 // pshufb 0x1c1d7(%rip),%xmm8 # 3a480 <_sk_srcover_bgra_8888_sse2_lowp+0xff8>
+ .byte 102,68,15,56,0,5,87,180,1,0 // pshufb 0x1b457(%rip),%xmm8 # 39700 <_sk_srcover_bgra_8888_sse2_lowp+0xffc>
.byte 102,71,15,58,21,4,16,0 // pextrw $0x0,%xmm8,(%r8,%r10,1)
.byte 235,192 // jmp 1e273 <_sk_store_a8_sse41+0x3d>
.byte 102,71,15,58,20,4,16,0 // pextrb $0x0,%xmm8,(%r8,%r10,1)
@@ -38662,11 +38662,11 @@ _sk_load_g8_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,42 // jne 1e2ff <_sk_load_g8_sse41+0x42>
.byte 102,67,15,56,49,4,16 // pmovzxbd (%r8,%r10,1),%xmm0
- .byte 102,15,219,5,60,190,1,0 // pand 0x1be3c(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,5,188,176,1,0 // pand 0x1b0bc(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,146,191,1,0 // mulps 0x1bf92(%rip),%xmm0 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,5,18,178,1,0 // mulps 0x1b212(%rip),%xmm0 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,121,189,1,0 // movaps 0x1bd79(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,249,175,1,0 // movaps 0x1aff9(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
@@ -38704,11 +38704,11 @@ _sk_load_g8_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,42 // jne 1e38e <_sk_load_g8_dst_sse41+0x42>
.byte 102,67,15,56,49,36,16 // pmovzxbd (%r8,%r10,1),%xmm4
- .byte 102,15,219,37,173,189,1,0 // pand 0x1bdad(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,37,45,176,1,0 // pand 0x1b02d(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,3,191,1,0 // mulps 0x1bf03(%rip),%xmm4 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,37,131,177,1,0 // mulps 0x1b183(%rip),%xmm4 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,234,188,1,0 // movaps 0x1bcea(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,61,106,175,1,0 // movaps 0x1af6a(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,236 // movaps %xmm4,%xmm5
.byte 15,40,244 // movaps %xmm4,%xmm6
.byte 255,224 // jmpq *%rax
@@ -38774,9 +38774,9 @@ _sk_gather_g8_sse41:
.byte 102,15,58,32,192,3 // pinsrb $0x3,%eax,%xmm0
.byte 102,15,56,49,192 // pmovzxbd %xmm0,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,248,189,1,0 // mulps 0x1bdf8(%rip),%xmm0 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,5,120,176,1,0 // mulps 0x1b078(%rip),%xmm0 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,223,187,1,0 // movaps 0x1bbdf(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,95,174,1,0 // movaps 0x1ae5f(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
@@ -38795,19 +38795,19 @@ _sk_load_565_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,80 // jne 1e504 <_sk_load_565_sse41+0x6b>
.byte 102,67,15,56,51,20,80 // pmovzxwd (%r8,%r10,2),%xmm2
- .byte 102,15,111,5,205,189,1,0 // movdqa 0x1bdcd(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,77,176,1,0 // movdqa 0x1b04d(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,207,189,1,0 // mulps 0x1bdcf(%rip),%xmm0 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,13,215,189,1,0 // movdqa 0x1bdd7(%rip),%xmm1 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,5,79,176,1,0 // mulps 0x1b04f(%rip),%xmm0 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,13,87,176,1,0 // movdqa 0x1b057(%rip),%xmm1 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,202 // pand %xmm2,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,217,189,1,0 // mulps 0x1bdd9(%rip),%xmm1 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,21,225,189,1,0 // pand 0x1bde1(%rip),%xmm2 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,13,89,176,1,0 // mulps 0x1b059(%rip),%xmm1 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,21,97,176,1,0 // pand 0x1b061(%rip),%xmm2 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,231,189,1,0 // mulps 0x1bde7(%rip),%xmm2 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,21,103,176,1,0 // mulps 0x1b067(%rip),%xmm2 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,110,187,1,0 // movaps 0x1bb6e(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,238,173,1,0 // movaps 0x1adee(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,3 // and $0x3,%r9b
@@ -38843,19 +38843,19 @@ _sk_load_565_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,80 // jne 1e5bf <_sk_load_565_dst_sse41+0x6b>
.byte 102,67,15,56,51,52,80 // pmovzxwd (%r8,%r10,2),%xmm6
- .byte 102,15,111,37,18,189,1,0 // movdqa 0x1bd12(%rip),%xmm4 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,37,146,175,1,0 // movdqa 0x1af92(%rip),%xmm4 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,230 // pand %xmm6,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,20,189,1,0 // mulps 0x1bd14(%rip),%xmm4 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,45,28,189,1,0 // movdqa 0x1bd1c(%rip),%xmm5 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,37,148,175,1,0 // mulps 0x1af94(%rip),%xmm4 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,45,156,175,1,0 // movdqa 0x1af9c(%rip),%xmm5 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,238 // pand %xmm6,%xmm5
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
- .byte 15,89,45,30,189,1,0 // mulps 0x1bd1e(%rip),%xmm5 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,53,38,189,1,0 // pand 0x1bd26(%rip),%xmm6 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,45,158,175,1,0 // mulps 0x1af9e(%rip),%xmm5 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,53,166,175,1,0 // pand 0x1afa6(%rip),%xmm6 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
- .byte 15,89,53,44,189,1,0 // mulps 0x1bd2c(%rip),%xmm6 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,53,172,175,1,0 // mulps 0x1afac(%rip),%xmm6 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,179,186,1,0 // movaps 0x1bab3(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,61,51,173,1,0 // movaps 0x1ad33(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,3 // and $0x3,%r9b
@@ -38917,19 +38917,19 @@ _sk_gather_565_sse41:
.byte 65,15,183,4,64 // movzwl (%r8,%rax,2),%eax
.byte 102,15,196,192,3 // pinsrw $0x3,%eax,%xmm0
.byte 102,15,56,51,208 // pmovzxwd %xmm0,%xmm2
- .byte 102,15,111,5,217,187,1,0 // movdqa 0x1bbd9(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,89,174,1,0 // movdqa 0x1ae59(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,219,187,1,0 // mulps 0x1bbdb(%rip),%xmm0 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,13,227,187,1,0 // movdqa 0x1bbe3(%rip),%xmm1 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,5,91,174,1,0 // mulps 0x1ae5b(%rip),%xmm0 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,13,99,174,1,0 // movdqa 0x1ae63(%rip),%xmm1 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,202 // pand %xmm2,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,229,187,1,0 // mulps 0x1bbe5(%rip),%xmm1 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,21,237,187,1,0 // pand 0x1bbed(%rip),%xmm2 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,13,101,174,1,0 // mulps 0x1ae65(%rip),%xmm1 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,21,109,174,1,0 // pand 0x1ae6d(%rip),%xmm2 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,243,187,1,0 // mulps 0x1bbf3(%rip),%xmm2 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,21,115,174,1,0 // mulps 0x1ae73(%rip),%xmm2 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,122,185,1,0 // movaps 0x1b97a(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,250,171,1,0 // movaps 0x1abfa(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
HIDDEN _sk_store_565_sse41
@@ -38943,12 +38943,12 @@ _sk_store_565_sse41:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,122,189,1,0 // movaps 0x1bd7a(%rip),%xmm8 # 3a490 <_sk_srcover_bgra_8888_sse2_lowp+0x1008>
+ .byte 68,15,40,5,250,175,1,0 // movaps 0x1affa(%rip),%xmm8 # 39710 <_sk_srcover_bgra_8888_sse2_lowp+0x100c>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
.byte 102,65,15,114,241,11 // pslld $0xb,%xmm9
- .byte 68,15,40,21,111,189,1,0 // movaps 0x1bd6f(%rip),%xmm10 # 3a4a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1018>
+ .byte 68,15,40,21,239,175,1,0 // movaps 0x1afef(%rip),%xmm10 # 39720 <_sk_srcover_bgra_8888_sse2_lowp+0x101c>
.byte 68,15,89,209 // mulps %xmm1,%xmm10
.byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
.byte 102,65,15,114,242,5 // pslld $0x5,%xmm10
@@ -38992,21 +38992,21 @@ _sk_load_4444_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,95 // jne 1e822 <_sk_load_4444_sse41+0x7a>
.byte 102,67,15,56,51,28,80 // pmovzxwd (%r8,%r10,2),%xmm3
- .byte 102,15,111,5,222,188,1,0 // movdqa 0x1bcde(%rip),%xmm0 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,5,94,175,1,0 // movdqa 0x1af5e(%rip),%xmm0 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,224,188,1,0 // mulps 0x1bce0(%rip),%xmm0 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,13,232,188,1,0 // movdqa 0x1bce8(%rip),%xmm1 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,5,96,175,1,0 // mulps 0x1af60(%rip),%xmm0 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,13,104,175,1,0 // movdqa 0x1af68(%rip),%xmm1 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,203 // pand %xmm3,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,234,188,1,0 // mulps 0x1bcea(%rip),%xmm1 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,21,242,188,1,0 // movdqa 0x1bcf2(%rip),%xmm2 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,13,106,175,1,0 // mulps 0x1af6a(%rip),%xmm1 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,21,114,175,1,0 // movdqa 0x1af72(%rip),%xmm2 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,211 // pand %xmm3,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,244,188,1,0 // mulps 0x1bcf4(%rip),%xmm2 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,29,252,188,1,0 // pand 0x1bcfc(%rip),%xmm3 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,21,116,175,1,0 // mulps 0x1af74(%rip),%xmm2 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,29,124,175,1,0 // pand 0x1af7c(%rip),%xmm3 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,2,189,1,0 // mulps 0x1bd02(%rip),%xmm3 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,29,130,175,1,0 // mulps 0x1af82(%rip),%xmm3 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -39043,21 +39043,21 @@ _sk_load_4444_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,95 // jne 1e8ec <_sk_load_4444_dst_sse41+0x7a>
.byte 102,67,15,56,51,60,80 // pmovzxwd (%r8,%r10,2),%xmm7
- .byte 102,15,111,37,20,188,1,0 // movdqa 0x1bc14(%rip),%xmm4 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,37,148,174,1,0 // movdqa 0x1ae94(%rip),%xmm4 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,22,188,1,0 // mulps 0x1bc16(%rip),%xmm4 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,45,30,188,1,0 // movdqa 0x1bc1e(%rip),%xmm5 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,37,150,174,1,0 // mulps 0x1ae96(%rip),%xmm4 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,45,158,174,1,0 // movdqa 0x1ae9e(%rip),%xmm5 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,239 // pand %xmm7,%xmm5
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
- .byte 15,89,45,32,188,1,0 // mulps 0x1bc20(%rip),%xmm5 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,53,40,188,1,0 // movdqa 0x1bc28(%rip),%xmm6 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,45,160,174,1,0 // mulps 0x1aea0(%rip),%xmm5 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,53,168,174,1,0 // movdqa 0x1aea8(%rip),%xmm6 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,247 // pand %xmm7,%xmm6
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
- .byte 15,89,53,42,188,1,0 // mulps 0x1bc2a(%rip),%xmm6 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,61,50,188,1,0 // pand 0x1bc32(%rip),%xmm7 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,53,170,174,1,0 // mulps 0x1aeaa(%rip),%xmm6 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,61,178,174,1,0 // pand 0x1aeb2(%rip),%xmm7 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,255 // cvtdq2ps %xmm7,%xmm7
- .byte 15,89,61,56,188,1,0 // mulps 0x1bc38(%rip),%xmm7 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,61,184,174,1,0 // mulps 0x1aeb8(%rip),%xmm7 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -39120,21 +39120,21 @@ _sk_gather_4444_sse41:
.byte 65,15,183,4,64 // movzwl (%r8,%rax,2),%eax
.byte 102,15,196,192,3 // pinsrw $0x3,%eax,%xmm0
.byte 102,15,56,51,216 // pmovzxwd %xmm0,%xmm3
- .byte 102,15,111,5,204,186,1,0 // movdqa 0x1bacc(%rip),%xmm0 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,5,76,173,1,0 // movdqa 0x1ad4c(%rip),%xmm0 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,206,186,1,0 // mulps 0x1bace(%rip),%xmm0 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,13,214,186,1,0 // movdqa 0x1bad6(%rip),%xmm1 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,5,78,173,1,0 // mulps 0x1ad4e(%rip),%xmm0 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,13,86,173,1,0 // movdqa 0x1ad56(%rip),%xmm1 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,203 // pand %xmm3,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,216,186,1,0 // mulps 0x1bad8(%rip),%xmm1 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,21,224,186,1,0 // movdqa 0x1bae0(%rip),%xmm2 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,13,88,173,1,0 // mulps 0x1ad58(%rip),%xmm1 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,21,96,173,1,0 // movdqa 0x1ad60(%rip),%xmm2 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,211 // pand %xmm3,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,226,186,1,0 // mulps 0x1bae2(%rip),%xmm2 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,29,234,186,1,0 // pand 0x1baea(%rip),%xmm3 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,21,98,173,1,0 // mulps 0x1ad62(%rip),%xmm2 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,29,106,173,1,0 // pand 0x1ad6a(%rip),%xmm3 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,240,186,1,0 // mulps 0x1baf0(%rip),%xmm3 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,29,112,173,1,0 // mulps 0x1ad70(%rip),%xmm3 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -39149,7 +39149,7 @@ _sk_store_4444_sse41:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,222,186,1,0 // movaps 0x1bade(%rip),%xmm8 # 3a530 <_sk_srcover_bgra_8888_sse2_lowp+0x10a8>
+ .byte 68,15,40,5,94,173,1,0 // movaps 0x1ad5e(%rip),%xmm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ac>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -39203,17 +39203,17 @@ _sk_load_8888_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,89 // jne 1eb6d <_sk_load_8888_sse41+0x75>
.byte 243,65,15,111,28,128 // movdqu (%r8,%rax,4),%xmm3
- .byte 102,15,111,5,254,181,1,0 // movdqa 0x1b5fe(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,126,168,1,0 // movdqa 0x1a87e(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,79,183,1,0 // movaps 0x1b74f(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,207,169,1,0 // movaps 0x1a9cf(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
- .byte 102,15,56,0,13,238,181,1,0 // pshufb 0x1b5ee(%rip),%xmm1 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,13,110,168,1,0 // pshufb 0x1a86e(%rip),%xmm1 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,111,211 // movdqa %xmm3,%xmm2
- .byte 102,15,56,0,21,234,181,1,0 // pshufb 0x1b5ea(%rip),%xmm2 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,21,106,168,1,0 // pshufb 0x1a86a(%rip),%xmm2 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,15,114,211,24 // psrld $0x18,%xmm3
@@ -39252,17 +39252,17 @@ _sk_load_8888_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,89 // jne 1ec27 <_sk_load_8888_dst_sse41+0x75>
.byte 243,65,15,111,60,128 // movdqu (%r8,%rax,4),%xmm7
- .byte 102,15,111,37,68,181,1,0 // movdqa 0x1b544(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,196,167,1,0 // movdqa 0x1a7c4(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 68,15,40,5,149,182,1,0 // movaps 0x1b695(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,21,169,1,0 // movaps 0x1a915(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,224 // mulps %xmm8,%xmm4
.byte 102,15,111,239 // movdqa %xmm7,%xmm5
- .byte 102,15,56,0,45,52,181,1,0 // pshufb 0x1b534(%rip),%xmm5 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,45,180,167,1,0 // pshufb 0x1a7b4(%rip),%xmm5 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
.byte 65,15,89,232 // mulps %xmm8,%xmm5
.byte 102,15,111,247 // movdqa %xmm7,%xmm6
- .byte 102,15,56,0,53,48,181,1,0 // pshufb 0x1b530(%rip),%xmm6 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,53,176,167,1,0 // pshufb 0x1a7b0(%rip),%xmm6 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
.byte 65,15,89,240 // mulps %xmm8,%xmm6
.byte 102,15,114,215,24 // psrld $0x18,%xmm7
@@ -39322,17 +39322,17 @@ _sk_gather_8888_sse41:
.byte 102,65,15,58,34,28,128,1 // pinsrd $0x1,(%r8,%rax,4),%xmm3
.byte 102,67,15,58,34,28,152,2 // pinsrd $0x2,(%r8,%r11,4),%xmm3
.byte 102,67,15,58,34,28,144,3 // pinsrd $0x3,(%r8,%r10,4),%xmm3
- .byte 102,15,111,5,29,180,1,0 // movdqa 0x1b41d(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,157,166,1,0 // movdqa 0x1a69d(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,110,181,1,0 // movaps 0x1b56e(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,238,167,1,0 // movaps 0x1a7ee(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
- .byte 102,15,56,0,13,13,180,1,0 // pshufb 0x1b40d(%rip),%xmm1 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,13,141,166,1,0 // pshufb 0x1a68d(%rip),%xmm1 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,111,211 // movdqa %xmm3,%xmm2
- .byte 102,15,56,0,21,9,180,1,0 // pshufb 0x1b409(%rip),%xmm2 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,21,137,166,1,0 // pshufb 0x1a689(%rip),%xmm2 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,15,114,211,24 // psrld $0x18,%xmm3
@@ -39352,7 +39352,7 @@ _sk_store_8888_sse41:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,227,179,1,0 // movaps 0x1b3e3(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,99,166,1,0 // movaps 0x1a663(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -39403,17 +39403,17 @@ _sk_load_bgra_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,89 // jne 1ee74 <_sk_load_bgra_sse41+0x75>
.byte 243,65,15,111,28,128 // movdqu (%r8,%rax,4),%xmm3
- .byte 102,15,111,5,247,178,1,0 // movdqa 0x1b2f7(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,119,165,1,0 // movdqa 0x1a577(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,208 // cvtdq2ps %xmm0,%xmm2
- .byte 68,15,40,5,72,180,1,0 // movaps 0x1b448(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,200,166,1,0 // movaps 0x1a6c8(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,15,56,0,5,231,178,1,0 // pshufb 0x1b2e7(%rip),%xmm0 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,5,103,165,1,0 // pshufb 0x1a567(%rip),%xmm0 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,15,56,0,5,227,178,1,0 // pshufb 0x1b2e3(%rip),%xmm0 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,5,99,165,1,0 // pshufb 0x1a563(%rip),%xmm0 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,114,211,24 // psrld $0x18,%xmm3
@@ -39452,17 +39452,17 @@ _sk_load_bgra_dst_sse41:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,89 // jne 1ef2e <_sk_load_bgra_dst_sse41+0x75>
.byte 243,65,15,111,60,128 // movdqu (%r8,%rax,4),%xmm7
- .byte 102,15,111,37,61,178,1,0 // movdqa 0x1b23d(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,189,164,1,0 // movdqa 0x1a4bd(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,244 // cvtdq2ps %xmm4,%xmm6
- .byte 68,15,40,5,142,179,1,0 // movaps 0x1b38e(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,14,166,1,0 // movaps 0x1a60e(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,240 // mulps %xmm8,%xmm6
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
- .byte 102,15,56,0,37,45,178,1,0 // pshufb 0x1b22d(%rip),%xmm4 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,37,173,164,1,0 // pshufb 0x1a4ad(%rip),%xmm4 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,236 // cvtdq2ps %xmm4,%xmm5
.byte 65,15,89,232 // mulps %xmm8,%xmm5
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
- .byte 102,15,56,0,37,41,178,1,0 // pshufb 0x1b229(%rip),%xmm4 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,37,169,164,1,0 // pshufb 0x1a4a9(%rip),%xmm4 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
.byte 65,15,89,224 // mulps %xmm8,%xmm4
.byte 102,15,114,215,24 // psrld $0x18,%xmm7
@@ -39522,17 +39522,17 @@ _sk_gather_bgra_sse41:
.byte 102,65,15,58,34,28,128,1 // pinsrd $0x1,(%r8,%rax,4),%xmm3
.byte 102,67,15,58,34,28,152,2 // pinsrd $0x2,(%r8,%r11,4),%xmm3
.byte 102,67,15,58,34,28,144,3 // pinsrd $0x3,(%r8,%r10,4),%xmm3
- .byte 102,15,111,5,22,177,1,0 // movdqa 0x1b116(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,150,163,1,0 // movdqa 0x1a396(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,208 // cvtdq2ps %xmm0,%xmm2
- .byte 68,15,40,5,103,178,1,0 // movaps 0x1b267(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,231,164,1,0 // movaps 0x1a4e7(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,15,56,0,5,6,177,1,0 // pshufb 0x1b106(%rip),%xmm0 # 3a130 <_sk_srcover_bgra_8888_sse2_lowp+0xca8>
+ .byte 102,15,56,0,5,134,163,1,0 // pshufb 0x1a386(%rip),%xmm0 # 393b0 <_sk_srcover_bgra_8888_sse2_lowp+0xcac>
.byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,15,56,0,5,2,177,1,0 // pshufb 0x1b102(%rip),%xmm0 # 3a140 <_sk_srcover_bgra_8888_sse2_lowp+0xcb8>
+ .byte 102,15,56,0,5,130,163,1,0 // pshufb 0x1a382(%rip),%xmm0 # 393c0 <_sk_srcover_bgra_8888_sse2_lowp+0xcbc>
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,114,211,24 // psrld $0x18,%xmm3
@@ -39552,7 +39552,7 @@ _sk_store_bgra_sse41:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,220,176,1,0 // movaps 0x1b0dc(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,92,163,1,0 // movaps 0x1a35c(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,40,202 // movaps %xmm2,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -39611,19 +39611,19 @@ _sk_load_f16_sse41:
.byte 102,68,15,97,224 // punpcklwd %xmm0,%xmm12
.byte 102,68,15,105,200 // punpckhwd %xmm0,%xmm9
.byte 102,65,15,56,51,212 // pmovzxwd %xmm12,%xmm2
- .byte 102,68,15,111,5,225,179,1,0 // movdqa 0x1b3e1(%rip),%xmm8 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,5,97,166,1,0 // movdqa 0x1a661(%rip),%xmm8 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,65,15,219,200 // pand %xmm8,%xmm1
- .byte 102,68,15,111,21,223,179,1,0 // movdqa 0x1b3df(%rip),%xmm10 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,21,95,166,1,0 // movdqa 0x1a65f(%rip),%xmm10 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,210 // pand %xmm10,%xmm2
- .byte 102,15,111,29,226,179,1,0 // movdqa 0x1b3e2(%rip),%xmm3 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,29,98,166,1,0 // movdqa 0x1a662(%rip),%xmm3 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,15,114,241,16 // pslld $0x10,%xmm1
.byte 102,15,111,194 // movdqa %xmm2,%xmm0
.byte 102,15,56,63,195 // pmaxud %xmm3,%xmm0
.byte 102,15,118,194 // pcmpeqd %xmm2,%xmm0
.byte 102,15,114,242,13 // pslld $0xd,%xmm2
.byte 102,15,235,209 // por %xmm1,%xmm2
- .byte 102,68,15,111,29,206,179,1,0 // movdqa 0x1b3ce(%rip),%xmm11 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,29,78,166,1,0 // movdqa 0x1a64e(%rip),%xmm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,211 // paddd %xmm11,%xmm2
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 102,65,15,112,204,238 // pshufd $0xee,%xmm12,%xmm1
@@ -39699,19 +39699,19 @@ _sk_load_f16_dst_sse41:
.byte 102,68,15,97,228 // punpcklwd %xmm4,%xmm12
.byte 102,68,15,105,204 // punpckhwd %xmm4,%xmm9
.byte 102,65,15,56,51,244 // pmovzxwd %xmm12,%xmm6
- .byte 102,68,15,111,5,67,178,1,0 // movdqa 0x1b243(%rip),%xmm8 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,5,195,164,1,0 // movdqa 0x1a4c3(%rip),%xmm8 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,15,111,238 // movdqa %xmm6,%xmm5
.byte 102,65,15,219,232 // pand %xmm8,%xmm5
- .byte 102,68,15,111,21,65,178,1,0 // movdqa 0x1b241(%rip),%xmm10 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,21,193,164,1,0 // movdqa 0x1a4c1(%rip),%xmm10 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,242 // pand %xmm10,%xmm6
- .byte 102,15,111,61,68,178,1,0 // movdqa 0x1b244(%rip),%xmm7 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,61,196,164,1,0 // movdqa 0x1a4c4(%rip),%xmm7 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,15,114,245,16 // pslld $0x10,%xmm5
.byte 102,15,111,230 // movdqa %xmm6,%xmm4
.byte 102,15,56,63,231 // pmaxud %xmm7,%xmm4
.byte 102,15,118,230 // pcmpeqd %xmm6,%xmm4
.byte 102,15,114,246,13 // pslld $0xd,%xmm6
.byte 102,15,235,245 // por %xmm5,%xmm6
- .byte 102,68,15,111,29,48,178,1,0 // movdqa 0x1b230(%rip),%xmm11 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,29,176,164,1,0 // movdqa 0x1a4b0(%rip),%xmm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,243 // paddd %xmm11,%xmm6
.byte 102,15,219,230 // pand %xmm6,%xmm4
.byte 102,65,15,112,236,238 // pshufd $0xee,%xmm12,%xmm5
@@ -39809,19 +39809,19 @@ _sk_gather_f16_sse41:
.byte 102,68,15,97,226 // punpcklwd %xmm2,%xmm12
.byte 102,68,15,105,202 // punpckhwd %xmm2,%xmm9
.byte 102,65,15,56,51,212 // pmovzxwd %xmm12,%xmm2
- .byte 102,68,15,111,5,65,176,1,0 // movdqa 0x1b041(%rip),%xmm8 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,5,193,162,1,0 // movdqa 0x1a2c1(%rip),%xmm8 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,65,15,219,200 // pand %xmm8,%xmm1
- .byte 102,68,15,111,21,63,176,1,0 // movdqa 0x1b03f(%rip),%xmm10 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,21,191,162,1,0 // movdqa 0x1a2bf(%rip),%xmm10 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,210 // pand %xmm10,%xmm2
- .byte 102,15,111,29,66,176,1,0 // movdqa 0x1b042(%rip),%xmm3 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,29,194,162,1,0 // movdqa 0x1a2c2(%rip),%xmm3 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,15,114,241,16 // pslld $0x10,%xmm1
.byte 102,15,111,194 // movdqa %xmm2,%xmm0
.byte 102,15,56,63,195 // pmaxud %xmm3,%xmm0
.byte 102,15,118,194 // pcmpeqd %xmm2,%xmm0
.byte 102,15,114,242,13 // pslld $0xd,%xmm2
.byte 102,15,235,209 // por %xmm1,%xmm2
- .byte 102,68,15,111,29,46,176,1,0 // movdqa 0x1b02e(%rip),%xmm11 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,29,174,162,1,0 // movdqa 0x1a2ae(%rip),%xmm11 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,211 // paddd %xmm11,%xmm2
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 102,65,15,112,204,238 // pshufd $0xee,%xmm12,%xmm1
@@ -39874,20 +39874,20 @@ _sk_store_f16_sse41:
.byte 73,193,224,3 // shl $0x3,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 102,68,15,111,13,86,175,1,0 // movdqa 0x1af56(%rip),%xmm9 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
+ .byte 102,68,15,111,13,214,161,1,0 // movdqa 0x1a1d6(%rip),%xmm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
.byte 102,68,15,111,224 // movdqa %xmm0,%xmm12
.byte 102,69,15,219,225 // pand %xmm9,%xmm12
- .byte 102,68,15,111,29,83,175,1,0 // movdqa 0x1af53(%rip),%xmm11 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 102,68,15,111,29,211,161,1,0 // movdqa 0x1a1d3(%rip),%xmm11 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 102,68,15,111,232 // movdqa %xmm0,%xmm13
.byte 102,69,15,219,235 // pand %xmm11,%xmm13
- .byte 102,68,15,111,21,80,175,1,0 // movdqa 0x1af50(%rip),%xmm10 # 3a5a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1118>
+ .byte 102,68,15,111,21,208,161,1,0 // movdqa 0x1a1d0(%rip),%xmm10 # 39820 <_sk_srcover_bgra_8888_sse2_lowp+0x111c>
.byte 102,65,15,114,212,16 // psrld $0x10,%xmm12
.byte 102,69,15,111,197 // movdqa %xmm13,%xmm8
.byte 102,69,15,56,63,194 // pmaxud %xmm10,%xmm8
.byte 102,69,15,118,197 // pcmpeqd %xmm13,%xmm8
.byte 102,65,15,114,213,13 // psrld $0xd,%xmm13
.byte 102,69,15,254,236 // paddd %xmm12,%xmm13
- .byte 102,68,15,111,37,54,175,1,0 // movdqa 0x1af36(%rip),%xmm12 # 3a5b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1128>
+ .byte 102,68,15,111,37,182,161,1,0 // movdqa 0x1a1b6(%rip),%xmm12 # 39830 <_sk_srcover_bgra_8888_sse2_lowp+0x112c>
.byte 102,69,15,254,236 // paddd %xmm12,%xmm13
.byte 102,69,15,219,197 // pand %xmm13,%xmm8
.byte 102,68,15,56,43,192 // packusdw %xmm0,%xmm8
@@ -39977,7 +39977,7 @@ _sk_load_u16_be_sse41:
.byte 102,15,235,200 // por %xmm0,%xmm1
.byte 102,15,56,51,193 // pmovzxwd %xmm1,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,254,170,1,0 // movaps 0x1aafe(%rip),%xmm8 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 68,15,40,5,126,157,1,0 // movaps 0x19d7e(%rip),%xmm8 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
.byte 102,15,113,241,8 // psllw $0x8,%xmm1
@@ -40051,7 +40051,7 @@ _sk_load_rgb_u16_be_sse41:
.byte 102,15,235,200 // por %xmm0,%xmm1
.byte 102,15,56,51,193 // pmovzxwd %xmm1,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,205,169,1,0 // movaps 0x1a9cd(%rip),%xmm8 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 68,15,40,5,77,156,1,0 // movaps 0x19c4d(%rip),%xmm8 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
.byte 102,15,113,241,8 // psllw $0x8,%xmm1
@@ -40068,7 +40068,7 @@ _sk_load_rgb_u16_be_sse41:
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,228,166,1,0 // movaps 0x1a6e4(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,100,153,1,0 // movaps 0x19964(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,20,64 // movd (%r8,%rax,2),%xmm2
.byte 102,15,239,201 // pxor %xmm1,%xmm1
@@ -40102,7 +40102,7 @@ _sk_store_u16_be_sse41:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 73,99,193 // movslq %r9d,%rax
- .byte 68,15,40,21,173,171,1,0 // movaps 0x1abad(%rip),%xmm10 # 3a5c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1138>
+ .byte 68,15,40,21,45,158,1,0 // movaps 0x19e2d(%rip),%xmm10 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x113c>
.byte 68,15,40,192 // movaps %xmm0,%xmm8
.byte 69,15,89,194 // mulps %xmm10,%xmm8
.byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
@@ -40333,7 +40333,7 @@ _sk_mirror_x_sse41:
.byte 65,15,92,194 // subps %xmm10,%xmm0
.byte 243,69,15,88,192 // addss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
- .byte 243,68,15,89,13,223,153,1,0 // mulss 0x199df(%rip),%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,13,91,140,1,0 // mulss 0x18c5b(%rip),%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,89,200 // mulps %xmm0,%xmm9
.byte 102,69,15,58,8,201,1 // roundps $0x1,%xmm9,%xmm9
@@ -40358,7 +40358,7 @@ _sk_mirror_y_sse41:
.byte 65,15,92,202 // subps %xmm10,%xmm1
.byte 243,69,15,88,192 // addss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
- .byte 243,68,15,89,13,134,153,1,0 // mulss 0x19986(%rip),%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,13,2,140,1,0 // mulss 0x18c02(%rip),%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,89,201 // mulps %xmm1,%xmm9
.byte 102,69,15,58,8,201,1 // roundps $0x1,%xmm9,%xmm9
@@ -40377,7 +40377,7 @@ FUNCTION(_sk_clamp_x_1_sse41)
_sk_clamp_x_1_sse41:
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 68,15,95,192 // maxps %xmm0,%xmm8
- .byte 68,15,93,5,134,162,1,0 // minps 0x1a286(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,93,5,6,149,1,0 // minps 0x19506(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -40390,7 +40390,7 @@ _sk_repeat_x_1_sse41:
.byte 65,15,92,192 // subps %xmm8,%xmm0
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 68,15,95,192 // maxps %xmm0,%xmm8
- .byte 68,15,93,5,99,162,1,0 // minps 0x1a263(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,93,5,227,148,1,0 // minps 0x194e3(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -40399,9 +40399,9 @@ HIDDEN _sk_mirror_x_1_sse41
.globl _sk_mirror_x_1_sse41
FUNCTION(_sk_mirror_x_1_sse41)
_sk_mirror_x_1_sse41:
- .byte 68,15,40,5,179,162,1,0 // movaps 0x1a2b3(%rip),%xmm8 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,5,51,149,1,0 // movaps 0x19533(%rip),%xmm8 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,13,55,162,1,0 // movaps 0x1a237(%rip),%xmm9 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,13,183,148,1,0 // movaps 0x194b7(%rip),%xmm9 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,89,200 // mulps %xmm0,%xmm9
.byte 102,69,15,58,8,201,1 // roundps $0x1,%xmm9,%xmm9
.byte 69,15,88,201 // addps %xmm9,%xmm9
@@ -40412,7 +40412,7 @@ _sk_mirror_x_1_sse41:
.byte 68,15,92,200 // subps %xmm0,%xmm9
.byte 68,15,84,200 // andps %xmm0,%xmm9
.byte 69,15,95,193 // maxps %xmm9,%xmm8
- .byte 68,15,93,5,20,162,1,0 // minps 0x1a214(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,93,5,148,148,1,0 // minps 0x19494(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -40422,10 +40422,10 @@ HIDDEN _sk_luminance_to_alpha_sse41
FUNCTION(_sk_luminance_to_alpha_sse41)
_sk_luminance_to_alpha_sse41:
.byte 15,40,218 // movaps %xmm2,%xmm3
- .byte 15,89,5,98,167,1,0 // mulps 0x1a762(%rip),%xmm0 # 3a5d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1148>
- .byte 15,89,13,107,167,1,0 // mulps 0x1a76b(%rip),%xmm1 # 3a5e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1158>
+ .byte 15,89,5,226,153,1,0 // mulps 0x199e2(%rip),%xmm0 # 39850 <_sk_srcover_bgra_8888_sse2_lowp+0x114c>
+ .byte 15,89,13,235,153,1,0 // mulps 0x199eb(%rip),%xmm1 # 39860 <_sk_srcover_bgra_8888_sse2_lowp+0x115c>
.byte 15,88,200 // addps %xmm0,%xmm1
- .byte 15,89,29,113,167,1,0 // mulps 0x1a771(%rip),%xmm3 # 3a5f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1168>
+ .byte 15,89,29,241,153,1,0 // mulps 0x199f1(%rip),%xmm3 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x116c>
.byte 15,88,217 // addps %xmm1,%xmm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
@@ -40974,26 +40974,26 @@ _sk_xy_to_unit_angle_sse41:
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,40,236 // movaps %xmm12,%xmm13
.byte 69,15,89,237 // mulps %xmm13,%xmm13
- .byte 68,15,40,21,155,158,1,0 // movaps 0x19e9b(%rip),%xmm10 # 3a600 <_sk_srcover_bgra_8888_sse2_lowp+0x1178>
+ .byte 68,15,40,21,27,145,1,0 // movaps 0x1911b(%rip),%xmm10 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x117c>
.byte 69,15,89,213 // mulps %xmm13,%xmm10
- .byte 68,15,88,21,159,158,1,0 // addps 0x19e9f(%rip),%xmm10 # 3a610 <_sk_srcover_bgra_8888_sse2_lowp+0x1188>
+ .byte 68,15,88,21,31,145,1,0 // addps 0x1911f(%rip),%xmm10 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x118c>
.byte 69,15,89,213 // mulps %xmm13,%xmm10
- .byte 68,15,88,21,163,158,1,0 // addps 0x19ea3(%rip),%xmm10 # 3a620 <_sk_srcover_bgra_8888_sse2_lowp+0x1198>
+ .byte 68,15,88,21,35,145,1,0 // addps 0x19123(%rip),%xmm10 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x119c>
.byte 69,15,89,213 // mulps %xmm13,%xmm10
- .byte 68,15,88,21,167,158,1,0 // addps 0x19ea7(%rip),%xmm10 # 3a630 <_sk_srcover_bgra_8888_sse2_lowp+0x11a8>
+ .byte 68,15,88,21,39,145,1,0 // addps 0x19127(%rip),%xmm10 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ac>
.byte 69,15,89,212 // mulps %xmm12,%xmm10
.byte 65,15,194,195,1 // cmpltps %xmm11,%xmm0
- .byte 68,15,40,29,166,158,1,0 // movaps 0x19ea6(%rip),%xmm11 # 3a640 <_sk_srcover_bgra_8888_sse2_lowp+0x11b8>
+ .byte 68,15,40,29,38,145,1,0 // movaps 0x19126(%rip),%xmm11 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x11bc>
.byte 69,15,92,218 // subps %xmm10,%xmm11
.byte 102,69,15,56,20,211 // blendvps %xmm0,%xmm11,%xmm10
.byte 69,15,194,200,1 // cmpltps %xmm8,%xmm9
- .byte 68,15,40,29,175,152,1,0 // movaps 0x198af(%rip),%xmm11 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,29,47,139,1,0 // movaps 0x18b2f(%rip),%xmm11 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,92,218 // subps %xmm10,%xmm11
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 102,69,15,56,20,211 // blendvps %xmm0,%xmm11,%xmm10
.byte 15,40,193 // movaps %xmm1,%xmm0
.byte 65,15,194,192,1 // cmpltps %xmm8,%xmm0
- .byte 68,15,40,13,161,152,1,0 // movaps 0x198a1(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,33,139,1,0 // movaps 0x18b21(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,202 // subps %xmm10,%xmm9
.byte 102,69,15,56,20,209 // blendvps %xmm0,%xmm9,%xmm10
.byte 69,15,194,194,7 // cmpordps %xmm10,%xmm8
@@ -41027,7 +41027,7 @@ _sk_xy_to_2pt_conical_quadratic_max_sse41:
.byte 243,69,15,89,203 // mulss %xmm11,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,26,158,1,0 // mulps 0x19e1a(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,154,144,1,0 // mulps 0x1909a(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
@@ -41035,7 +41035,7 @@ _sk_xy_to_2pt_conical_quadratic_max_sse41:
.byte 243,69,15,89,219 // mulss %xmm11,%xmm11
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,92,227 // subps %xmm11,%xmm12
- .byte 68,15,89,21,5,158,1,0 // mulps 0x19e05(%rip),%xmm10 # 3a660 <_sk_srcover_bgra_8888_sse2_lowp+0x11d8>
+ .byte 68,15,89,21,133,144,1,0 // mulps 0x19085(%rip),%xmm10 # 398e0 <_sk_srcover_bgra_8888_sse2_lowp+0x11dc>
.byte 69,15,89,212 // mulps %xmm12,%xmm10
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
@@ -41044,8 +41044,8 @@ _sk_xy_to_2pt_conical_quadratic_max_sse41:
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 65,15,92,193 // subps %xmm9,%xmm0
- .byte 68,15,87,13,253,156,1,0 // xorps 0x19cfd(%rip),%xmm9 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
- .byte 68,15,89,5,213,151,1,0 // mulps 0x197d5(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,87,13,125,143,1,0 // xorps 0x18f7d(%rip),%xmm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
+ .byte 68,15,89,5,85,138,1,0 // mulps 0x18a55(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 69,15,92,202 // subps %xmm10,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
@@ -41066,7 +41066,7 @@ _sk_xy_to_2pt_conical_quadratic_min_sse41:
.byte 243,69,15,89,203 // mulss %xmm11,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,124,157,1,0 // mulps 0x19d7c(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,252,143,1,0 // mulps 0x18ffc(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
@@ -41074,7 +41074,7 @@ _sk_xy_to_2pt_conical_quadratic_min_sse41:
.byte 243,69,15,89,219 // mulss %xmm11,%xmm11
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,92,227 // subps %xmm11,%xmm12
- .byte 68,15,89,21,103,157,1,0 // mulps 0x19d67(%rip),%xmm10 # 3a660 <_sk_srcover_bgra_8888_sse2_lowp+0x11d8>
+ .byte 68,15,89,21,231,143,1,0 // mulps 0x18fe7(%rip),%xmm10 # 398e0 <_sk_srcover_bgra_8888_sse2_lowp+0x11dc>
.byte 69,15,89,212 // mulps %xmm12,%xmm10
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
@@ -41083,8 +41083,8 @@ _sk_xy_to_2pt_conical_quadratic_min_sse41:
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 65,15,92,193 // subps %xmm9,%xmm0
- .byte 68,15,87,13,95,156,1,0 // xorps 0x19c5f(%rip),%xmm9 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
- .byte 68,15,89,5,55,151,1,0 // mulps 0x19737(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,87,13,223,142,1,0 // xorps 0x18edf(%rip),%xmm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
+ .byte 68,15,89,5,183,137,1,0 // mulps 0x189b7(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 69,15,92,202 // subps %xmm10,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
@@ -41102,7 +41102,7 @@ _sk_xy_to_2pt_conical_linear_sse41:
.byte 243,69,15,89,200 // mulss %xmm8,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,239,156,1,0 // mulps 0x19cef(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,111,143,1,0 // mulps 0x18f6f(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,209 // movaps %xmm1,%xmm10
.byte 69,15,89,210 // mulps %xmm10,%xmm10
@@ -41110,7 +41110,7 @@ _sk_xy_to_2pt_conical_linear_sse41:
.byte 243,69,15,89,192 // mulss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,92,192 // subps %xmm8,%xmm0
- .byte 15,87,5,251,155,1,0 // xorps 0x19bfb(%rip),%xmm0 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
+ .byte 15,87,5,123,142,1,0 // xorps 0x18e7b(%rip),%xmm0 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
.byte 65,15,94,193 // divps %xmm9,%xmm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41158,7 +41158,7 @@ HIDDEN _sk_save_xy_sse41
FUNCTION(_sk_save_xy_sse41)
_sk_save_xy_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,90,150,1,0 // movaps 0x1965a(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,218,136,1,0 // movaps 0x188da(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,17,0 // movups %xmm0,(%rax)
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,88,200 // addps %xmm8,%xmm9
@@ -41202,8 +41202,8 @@ _sk_bilinear_nx_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,211,155,1,0 // addps 0x19bd3(%rip),%xmm0 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,203,149,1,0 // movaps 0x195cb(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,83,142,1,0 // addps 0x18e53(%rip),%xmm0 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,75,136,1,0 // movaps 0x1884b(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 68,15,17,136,0,1,0,0 // movups %xmm9,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41216,7 +41216,7 @@ _sk_bilinear_px_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,151,149,1,0 // addps 0x19597(%rip),%xmm0 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,5,23,136,1,0 // addps 0x18817(%rip),%xmm0 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41228,8 +41228,8 @@ _sk_bilinear_ny_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,134,155,1,0 // addps 0x19b86(%rip),%xmm1 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,126,149,1,0 // movaps 0x1957e(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,6,142,1,0 // addps 0x18e06(%rip),%xmm1 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,254,135,1,0 // movaps 0x187fe(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 68,15,17,136,64,1,0,0 // movups %xmm9,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41242,7 +41242,7 @@ _sk_bilinear_py_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,73,149,1,0 // addps 0x19549(%rip),%xmm1 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,13,201,135,1,0 // addps 0x187c9(%rip),%xmm1 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41254,13 +41254,13 @@ _sk_bicubic_n3x_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,73,155,1,0 // addps 0x19b49(%rip),%xmm0 # 3a680 <_sk_srcover_bgra_8888_sse2_lowp+0x11f8>
- .byte 68,15,40,13,49,149,1,0 // movaps 0x19531(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,201,141,1,0 // addps 0x18dc9(%rip),%xmm0 # 39900 <_sk_srcover_bgra_8888_sse2_lowp+0x11fc>
+ .byte 68,15,40,13,177,135,1,0 // movaps 0x187b1(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 69,15,40,193 // movaps %xmm9,%xmm8
.byte 69,15,89,192 // mulps %xmm8,%xmm8
- .byte 68,15,89,13,61,155,1,0 // mulps 0x19b3d(%rip),%xmm9 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,13,21,151,1,0 // addps 0x19715(%rip),%xmm9 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,13,189,141,1,0 // mulps 0x18dbd(%rip),%xmm9 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,13,149,137,1,0 // addps 0x18995(%rip),%xmm9 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 68,15,17,136,0,1,0,0 // movups %xmm9,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41273,16 +41273,16 @@ _sk_bicubic_n1x_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,241,154,1,0 // addps 0x19af1(%rip),%xmm0 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,233,148,1,0 // movaps 0x194e9(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,113,141,1,0 // addps 0x18d71(%rip),%xmm0 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,105,135,1,0 // movaps 0x18769(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
- .byte 68,15,40,5,13,155,1,0 // movaps 0x19b0d(%rip),%xmm8 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,5,141,141,1,0 // movaps 0x18d8d(%rip),%xmm8 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,17,155,1,0 // addps 0x19b11(%rip),%xmm8 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,5,145,141,1,0 // addps 0x18d91(%rip),%xmm8 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,181,148,1,0 // addps 0x194b5(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,88,5,53,135,1,0 // addps 0x18735(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,9,155,1,0 // addps 0x19b09(%rip),%xmm8 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,5,137,141,1,0 // addps 0x18d89(%rip),%xmm8 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41292,17 +41292,17 @@ HIDDEN _sk_bicubic_p1x_sse41
FUNCTION(_sk_bicubic_p1x_sse41)
_sk_bicubic_p1x_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,147,148,1,0 // movaps 0x19493(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,19,135,1,0 // movaps 0x18713(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,136,128,0,0,0 // movups 0x80(%rax),%xmm9
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,21,188,154,1,0 // movaps 0x19abc(%rip),%xmm10 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,21,60,141,1,0 // movaps 0x18d3c(%rip),%xmm10 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,192,154,1,0 // addps 0x19ac0(%rip),%xmm10 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,21,64,141,1,0 // addps 0x18d40(%rip),%xmm10 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
.byte 69,15,88,208 // addps %xmm8,%xmm10
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,188,154,1,0 // addps 0x19abc(%rip),%xmm10 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,21,60,141,1,0 // addps 0x18d3c(%rip),%xmm10 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,144,0,1,0,0 // movups %xmm10,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41314,11 +41314,11 @@ _sk_bicubic_p3x_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,140,154,1,0 // addps 0x19a8c(%rip),%xmm0 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 15,88,5,12,141,1,0 // addps 0x18d0c(%rip),%xmm0 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 69,15,89,201 // mulps %xmm9,%xmm9
- .byte 68,15,89,5,92,154,1,0 // mulps 0x19a5c(%rip),%xmm8 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,5,52,150,1,0 // addps 0x19634(%rip),%xmm8 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,5,220,140,1,0 // mulps 0x18cdc(%rip),%xmm8 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,5,180,136,1,0 // addps 0x188b4(%rip),%xmm8 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41331,13 +41331,13 @@ _sk_bicubic_n3y_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,31,154,1,0 // addps 0x19a1f(%rip),%xmm1 # 3a680 <_sk_srcover_bgra_8888_sse2_lowp+0x11f8>
- .byte 68,15,40,13,7,148,1,0 // movaps 0x19407(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,159,140,1,0 // addps 0x18c9f(%rip),%xmm1 # 39900 <_sk_srcover_bgra_8888_sse2_lowp+0x11fc>
+ .byte 68,15,40,13,135,134,1,0 // movaps 0x18687(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 69,15,40,193 // movaps %xmm9,%xmm8
.byte 69,15,89,192 // mulps %xmm8,%xmm8
- .byte 68,15,89,13,19,154,1,0 // mulps 0x19a13(%rip),%xmm9 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,13,235,149,1,0 // addps 0x195eb(%rip),%xmm9 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,13,147,140,1,0 // mulps 0x18c93(%rip),%xmm9 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,13,107,136,1,0 // addps 0x1886b(%rip),%xmm9 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 68,15,17,136,64,1,0,0 // movups %xmm9,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41350,16 +41350,16 @@ _sk_bicubic_n1y_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,198,153,1,0 // addps 0x199c6(%rip),%xmm1 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,190,147,1,0 // movaps 0x193be(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,70,140,1,0 // addps 0x18c46(%rip),%xmm1 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,62,134,1,0 // movaps 0x1863e(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
- .byte 68,15,40,5,226,153,1,0 // movaps 0x199e2(%rip),%xmm8 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,5,98,140,1,0 // movaps 0x18c62(%rip),%xmm8 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,230,153,1,0 // addps 0x199e6(%rip),%xmm8 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,5,102,140,1,0 // addps 0x18c66(%rip),%xmm8 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,138,147,1,0 // addps 0x1938a(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,88,5,10,134,1,0 // addps 0x1860a(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,222,153,1,0 // addps 0x199de(%rip),%xmm8 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,5,94,140,1,0 // addps 0x18c5e(%rip),%xmm8 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41369,17 +41369,17 @@ HIDDEN _sk_bicubic_p1y_sse41
FUNCTION(_sk_bicubic_p1y_sse41)
_sk_bicubic_p1y_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,104,147,1,0 // movaps 0x19368(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,232,133,1,0 // movaps 0x185e8(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,136,192,0,0,0 // movups 0xc0(%rax),%xmm9
.byte 65,15,88,200 // addps %xmm8,%xmm1
- .byte 68,15,40,21,144,153,1,0 // movaps 0x19990(%rip),%xmm10 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,21,16,140,1,0 // movaps 0x18c10(%rip),%xmm10 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,148,153,1,0 // addps 0x19994(%rip),%xmm10 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,21,20,140,1,0 // addps 0x18c14(%rip),%xmm10 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
.byte 69,15,88,208 // addps %xmm8,%xmm10
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,144,153,1,0 // addps 0x19990(%rip),%xmm10 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,21,16,140,1,0 // addps 0x18c10(%rip),%xmm10 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,144,64,1,0,0 // movups %xmm10,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -41391,11 +41391,11 @@ _sk_bicubic_p3y_sse41:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,95,153,1,0 // addps 0x1995f(%rip),%xmm1 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 15,88,13,223,139,1,0 // addps 0x18bdf(%rip),%xmm1 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 69,15,89,201 // mulps %xmm9,%xmm9
- .byte 68,15,89,5,47,153,1,0 // mulps 0x1992f(%rip),%xmm8 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,5,7,149,1,0 // addps 0x19507(%rip),%xmm8 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,5,175,139,1,0 // mulps 0x18baf(%rip),%xmm8 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,5,135,135,1,0 // addps 0x18787(%rip),%xmm8 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -41506,7 +41506,7 @@ _sk_clut_3D_sse41:
.byte 15,89,194 // mulps %xmm2,%xmm0
.byte 15,41,68,36,208 // movaps %xmm0,-0x30(%rsp)
.byte 243,15,91,240 // cvttps2dq %xmm0,%xmm6
- .byte 15,40,37,26,152,1,0 // movaps 0x1981a(%rip),%xmm4 # 3a6d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1248>
+ .byte 15,40,37,154,138,1,0 // movaps 0x18a9a(%rip),%xmm4 # 39950 <_sk_srcover_bgra_8888_sse2_lowp+0x124c>
.byte 15,88,196 // addps %xmm4,%xmm0
.byte 15,41,68,36,176 // movaps %xmm0,-0x50(%rsp)
.byte 102,65,15,110,208 // movd %r8d,%xmm2
@@ -41545,7 +41545,7 @@ _sk_clut_3D_sse41:
.byte 102,68,15,56,64,192 // pmulld %xmm0,%xmm8
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,254,218 // paddd %xmm2,%xmm3
- .byte 102,68,15,111,37,113,151,1,0 // movdqa 0x19771(%rip),%xmm12 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,68,15,111,37,241,137,1,0 // movdqa 0x189f1(%rip),%xmm12 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,65,15,56,64,220 // pmulld %xmm12,%xmm3
.byte 102,15,118,228 // pcmpeqd %xmm4,%xmm4
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
@@ -41568,7 +41568,7 @@ _sk_clut_3D_sse41:
.byte 102,65,15,58,22,217,1 // pextrd $0x1,%xmm3,%r9d
.byte 102,65,15,58,22,218,2 // pextrd $0x2,%xmm3,%r10d
.byte 102,65,15,58,22,219,3 // pextrd $0x3,%xmm3,%r11d
- .byte 102,15,111,37,181,144,1,0 // movdqa 0x190b5(%rip),%xmm4 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,37,53,131,1,0 // movdqa 0x18335(%rip),%xmm4 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,220 // paddd %xmm4,%xmm3
.byte 102,15,111,252 // movdqa %xmm4,%xmm7
.byte 102,73,15,58,22,222,1 // pextrq $0x1,%xmm3,%r14
@@ -41741,7 +41741,7 @@ _sk_clut_3D_sse41:
.byte 102,68,15,254,202 // paddd %xmm2,%xmm9
.byte 102,65,15,111,192 // movdqa %xmm8,%xmm0
.byte 102,65,15,254,193 // paddd %xmm9,%xmm0
- .byte 102,15,111,45,128,147,1,0 // movdqa 0x19380(%rip),%xmm5 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,45,0,134,1,0 // movdqa 0x18600(%rip),%xmm5 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,56,64,197 // pmulld %xmm5,%xmm0
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,118,246 // pcmpeqd %xmm6,%xmm6
@@ -41762,7 +41762,7 @@ _sk_clut_3D_sse41:
.byte 102,65,15,58,22,193,1 // pextrd $0x1,%xmm0,%r9d
.byte 102,65,15,58,22,194,2 // pextrd $0x2,%xmm0,%r10d
.byte 102,65,15,58,22,195,3 // pextrd $0x3,%xmm0,%r11d
- .byte 102,15,111,61,209,140,1,0 // movdqa 0x18cd1(%rip),%xmm7 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,61,81,127,1,0 // movdqa 0x17f51(%rip),%xmm7 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,199 // paddd %xmm7,%xmm0
.byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
.byte 102,72,15,126,195 // movq %xmm0,%rbx
@@ -41828,7 +41828,7 @@ _sk_clut_3D_sse41:
.byte 15,88,248 // addps %xmm0,%xmm7
.byte 102,68,15,254,210 // paddd %xmm2,%xmm10
.byte 102,69,15,254,194 // paddd %xmm10,%xmm8
- .byte 102,15,111,13,188,145,1,0 // movdqa 0x191bc(%rip),%xmm1 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,13,60,132,1,0 // movdqa 0x1843c(%rip),%xmm1 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,68,15,56,64,193 // pmulld %xmm1,%xmm8
.byte 102,65,15,111,192 // movdqa %xmm8,%xmm0
.byte 102,15,118,237 // pcmpeqd %xmm5,%xmm5
@@ -41849,7 +41849,7 @@ _sk_clut_3D_sse41:
.byte 102,69,15,58,22,193,1 // pextrd $0x1,%xmm8,%r9d
.byte 102,69,15,58,22,194,2 // pextrd $0x2,%xmm8,%r10d
.byte 102,69,15,58,22,195,3 // pextrd $0x3,%xmm8,%r11d
- .byte 102,15,111,21,7,139,1,0 // movdqa 0x18b07(%rip),%xmm2 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,21,135,125,1,0 // movdqa 0x17d87(%rip),%xmm2 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,68,15,254,194 // paddd %xmm2,%xmm8
.byte 102,76,15,58,22,193,1 // pextrq $0x1,%xmm8,%rcx
.byte 102,76,15,126,195 // movq %xmm8,%rbx
@@ -41976,7 +41976,7 @@ _sk_clut_4D_sse41:
.byte 15,89,203 // mulps %xmm3,%xmm1
.byte 15,41,76,36,48 // movaps %xmm1,0x30(%rsp)
.byte 243,15,91,249 // cvttps2dq %xmm1,%xmm7
- .byte 15,40,29,26,143,1,0 // movaps 0x18f1a(%rip),%xmm3 # 3a6d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1248>
+ .byte 15,40,29,154,129,1,0 // movaps 0x1819a(%rip),%xmm3 # 39950 <_sk_srcover_bgra_8888_sse2_lowp+0x124c>
.byte 15,88,203 // addps %xmm3,%xmm1
.byte 15,41,76,36,160 // movaps %xmm1,-0x60(%rsp)
.byte 102,65,15,110,232 // movd %r8d,%xmm5
@@ -42033,7 +42033,7 @@ _sk_clut_4D_sse41:
.byte 102,68,15,56,64,253 // pmulld %xmm5,%xmm15
.byte 102,65,15,111,223 // movdqa %xmm15,%xmm3
.byte 102,15,254,217 // paddd %xmm1,%xmm3
- .byte 102,68,15,111,29,25,142,1,0 // movdqa 0x18e19(%rip),%xmm11 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,68,15,111,29,153,128,1,0 // movdqa 0x18099(%rip),%xmm11 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,65,15,56,64,219 // pmulld %xmm11,%xmm3
.byte 102,15,118,192 // pcmpeqd %xmm0,%xmm0
.byte 102,15,111,243 // movdqa %xmm3,%xmm6
@@ -42055,7 +42055,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,217,1 // pextrd $0x1,%xmm3,%r9d
.byte 102,65,15,58,22,218,2 // pextrd $0x2,%xmm3,%r10d
.byte 102,65,15,58,22,219,3 // pextrd $0x3,%xmm3,%r11d
- .byte 102,68,15,111,21,97,135,1,0 // movdqa 0x18761(%rip),%xmm10 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,21,225,121,1,0 // movdqa 0x179e1(%rip),%xmm10 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,218 // paddd %xmm10,%xmm3
.byte 102,73,15,58,22,222,1 // pextrq $0x1,%xmm3,%r14
.byte 102,72,15,126,219 // movq %xmm3,%rbx
@@ -42073,7 +42073,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,254,201 // paddd %xmm9,%xmm1
.byte 102,65,15,56,64,203 // pmulld %xmm11,%xmm1
.byte 102,15,111,217 // movdqa %xmm1,%xmm3
- .byte 102,15,250,29,79,141,1,0 // psubd 0x18d4f(%rip),%xmm3 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,29,207,127,1,0 // psubd 0x17fcf(%rip),%xmm3 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,73,15,58,22,222,1 // pextrq $0x1,%xmm3,%r14
.byte 102,72,15,126,219 // movq %xmm3,%rbx
.byte 65,137,223 // mov %ebx,%r15d
@@ -42128,7 +42128,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,254,199 // paddd %xmm15,%xmm0
.byte 102,65,15,56,64,195 // pmulld %xmm11,%xmm0
.byte 102,15,111,232 // movdqa %xmm0,%xmm5
- .byte 102,15,250,45,47,140,1,0 // psubd 0x18c2f(%rip),%xmm5 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,45,175,126,1,0 // psubd 0x17eaf(%rip),%xmm5 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,69,15,118,192 // pcmpeqd %xmm8,%xmm8
.byte 102,72,15,58,22,233,1 // pextrq $0x1,%xmm5,%rcx
.byte 102,72,15,126,237 // movq %xmm5,%rbp
@@ -42253,7 +42253,7 @@ _sk_clut_4D_sse41:
.byte 72,193,233,32 // shr $0x20,%rcx
.byte 243,15,16,36,136 // movss (%rax,%rcx,4),%xmm4
.byte 102,68,15,58,33,220,48 // insertps $0x30,%xmm4,%xmm11
- .byte 102,15,111,61,83,131,1,0 // movdqa 0x18353(%rip),%xmm7 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,61,211,117,1,0 // movdqa 0x175d3(%rip),%xmm7 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,223 // paddd %xmm7,%xmm3
.byte 102,72,15,58,22,217,1 // pextrq $0x1,%xmm3,%rcx
.byte 102,72,15,126,221 // movq %xmm3,%rbp
@@ -42339,7 +42339,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,201,1 // pextrd $0x1,%xmm1,%r9d
.byte 102,65,15,58,22,202,2 // pextrd $0x2,%xmm1,%r10d
.byte 102,65,15,58,22,203,3 // pextrd $0x3,%xmm1,%r11d
- .byte 102,68,15,111,21,151,129,1,0 // movdqa 0x18197(%rip),%xmm10 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,21,23,116,1,0 // movdqa 0x17417(%rip),%xmm10 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,202 // paddd %xmm10,%xmm1
.byte 102,72,15,58,22,201,1 // pextrq $0x1,%xmm1,%rcx
.byte 102,72,15,126,203 // movq %xmm1,%rbx
@@ -42435,7 +42435,7 @@ _sk_clut_4D_sse41:
.byte 102,15,111,124,36,192 // movdqa -0x40(%rsp),%xmm7
.byte 102,15,111,199 // movdqa %xmm7,%xmm0
.byte 102,15,254,195 // paddd %xmm3,%xmm0
- .byte 102,15,111,21,241,133,1,0 // movdqa 0x185f1(%rip),%xmm2 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,21,113,120,1,0 // movdqa 0x17871(%rip),%xmm2 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,56,64,194 // pmulld %xmm2,%xmm0
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,69,15,118,192 // pcmpeqd %xmm8,%xmm8
@@ -42456,7 +42456,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,193,1 // pextrd $0x1,%xmm0,%r9d
.byte 102,65,15,58,22,194,2 // pextrd $0x2,%xmm0,%r10d
.byte 102,65,15,58,22,195,3 // pextrd $0x3,%xmm0,%r11d
- .byte 102,15,111,37,64,127,1,0 // movdqa 0x17f40(%rip),%xmm4 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,37,192,113,1,0 // movdqa 0x171c0(%rip),%xmm4 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,196 // paddd %xmm4,%xmm0
.byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
.byte 102,72,15,126,195 // movq %xmm0,%rbx
@@ -42528,7 +42528,7 @@ _sk_clut_4D_sse41:
.byte 102,68,15,111,226 // movdqa %xmm2,%xmm12
.byte 102,65,15,56,64,204 // pmulld %xmm12,%xmm1
.byte 102,15,111,209 // movdqa %xmm1,%xmm2
- .byte 102,15,250,21,28,132,1,0 // psubd 0x1841c(%rip),%xmm2 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,21,156,118,1,0 // psubd 0x1769c(%rip),%xmm2 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,72,15,58,22,209,1 // pextrq $0x1,%xmm2,%rcx
.byte 102,72,15,126,213 // movq %xmm2,%rbp
.byte 137,235 // mov %ebp,%ebx
@@ -42562,7 +42562,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,111,249 // movdqa %xmm9,%xmm7
.byte 102,69,15,56,64,236 // pmulld %xmm12,%xmm13
.byte 102,65,15,111,205 // movdqa %xmm13,%xmm1
- .byte 102,15,250,13,104,131,1,0 // psubd 0x18368(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,232,117,1,0 // psubd 0x175e8(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,72,15,58,22,201,1 // pextrq $0x1,%xmm1,%rcx
.byte 102,72,15,126,203 // movq %xmm1,%rbx
.byte 137,221 // mov %ebx,%ebp
@@ -42627,7 +42627,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,111,206 // movdqa %xmm14,%xmm1
.byte 102,15,111,108,36,192 // movdqa -0x40(%rsp),%xmm5
.byte 102,15,254,205 // paddd %xmm5,%xmm1
- .byte 102,15,111,37,255,129,1,0 // movdqa 0x181ff(%rip),%xmm4 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,37,127,116,1,0 // movdqa 0x1747f(%rip),%xmm4 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,56,64,204 // pmulld %xmm4,%xmm1
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,15,118,246 // pcmpeqd %xmm6,%xmm6
@@ -42648,7 +42648,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,201,1 // pextrd $0x1,%xmm1,%r9d
.byte 102,65,15,58,22,202,2 // pextrd $0x2,%xmm1,%r10d
.byte 102,65,15,58,22,203,3 // pextrd $0x3,%xmm1,%r11d
- .byte 102,15,111,29,80,123,1,0 // movdqa 0x17b50(%rip),%xmm3 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,29,208,109,1,0 // movdqa 0x16dd0(%rip),%xmm3 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,203 // paddd %xmm3,%xmm1
.byte 102,72,15,58,22,201,1 // pextrq $0x1,%xmm1,%rcx
.byte 102,72,15,126,203 // movq %xmm1,%rbx
@@ -42720,7 +42720,7 @@ _sk_clut_4D_sse41:
.byte 102,15,56,64,239 // pmulld %xmm7,%xmm5
.byte 102,15,111,205 // movdqa %xmm5,%xmm1
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
- .byte 102,15,250,5,48,128,1,0 // psubd 0x18030(%rip),%xmm0 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,5,176,114,1,0 // psubd 0x172b0(%rip),%xmm0 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
.byte 102,72,15,126,197 // movq %xmm0,%rbp
.byte 137,235 // mov %ebp,%ebx
@@ -42738,7 +42738,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,193,1 // pextrd $0x1,%xmm0,%r9d
.byte 102,65,15,58,22,194,2 // pextrd $0x2,%xmm0,%r10d
.byte 102,65,15,58,22,195,3 // pextrd $0x3,%xmm0,%r11d
- .byte 102,15,254,5,126,121,1,0 // paddd 0x1797e(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,5,254,107,1,0 // paddd 0x16bfe(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
.byte 102,72,15,126,195 // movq %xmm0,%rbx
.byte 137,221 // mov %ebx,%ebp
@@ -42754,7 +42754,7 @@ _sk_clut_4D_sse41:
.byte 102,15,254,84,36,176 // paddd -0x50(%rsp),%xmm2
.byte 102,15,56,64,215 // pmulld %xmm7,%xmm2
.byte 102,15,111,194 // movdqa %xmm2,%xmm0
- .byte 102,15,250,5,119,127,1,0 // psubd 0x17f77(%rip),%xmm0 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,5,247,113,1,0 // psubd 0x171f7(%rip),%xmm0 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
.byte 102,72,15,126,195 // movq %xmm0,%rbx
.byte 137,221 // mov %ebx,%ebp
@@ -42772,7 +42772,7 @@ _sk_clut_4D_sse41:
.byte 102,65,15,58,22,199,1 // pextrd $0x1,%xmm0,%r15d
.byte 102,65,15,58,22,196,2 // pextrd $0x2,%xmm0,%r12d
.byte 102,65,15,58,22,198,3 // pextrd $0x3,%xmm0,%r14d
- .byte 102,15,254,5,197,120,1,0 // paddd 0x178c5(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,5,69,107,1,0 // paddd 0x16b45(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,72,15,58,22,195,1 // pextrq $0x1,%xmm0,%rbx
.byte 102,72,15,126,197 // movq %xmm0,%rbp
.byte 137,233 // mov %ebp,%ecx
@@ -42835,7 +42835,7 @@ _sk_clut_4D_sse41:
.byte 15,89,212 // mulps %xmm4,%xmm2
.byte 65,15,88,211 // addps %xmm11,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,130,119,1,0 // movaps 0x17782(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,2,106,1,0 // movaps 0x16a02(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,139,76,36,248 // mov -0x8(%rsp),%rcx
.byte 15,40,100,36,64 // movaps 0x40(%rsp),%xmm4
.byte 15,40,108,36,80 // movaps 0x50(%rsp),%xmm5
@@ -42854,15 +42854,15 @@ HIDDEN _sk_gauss_a_to_rgba_sse41
.globl _sk_gauss_a_to_rgba_sse41
FUNCTION(_sk_gauss_a_to_rgba_sse41)
_sk_gauss_a_to_rgba_sse41:
- .byte 15,40,5,223,125,1,0 // movaps 0x17ddf(%rip),%xmm0 # 3a700 <_sk_srcover_bgra_8888_sse2_lowp+0x1278>
+ .byte 15,40,5,95,112,1,0 // movaps 0x1705f(%rip),%xmm0 # 39980 <_sk_srcover_bgra_8888_sse2_lowp+0x127c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,229,125,1,0 // addps 0x17de5(%rip),%xmm0 # 3a710 <_sk_srcover_bgra_8888_sse2_lowp+0x1288>
+ .byte 15,88,5,101,112,1,0 // addps 0x17065(%rip),%xmm0 # 39990 <_sk_srcover_bgra_8888_sse2_lowp+0x128c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,235,125,1,0 // addps 0x17deb(%rip),%xmm0 # 3a720 <_sk_srcover_bgra_8888_sse2_lowp+0x1298>
+ .byte 15,88,5,107,112,1,0 // addps 0x1706b(%rip),%xmm0 # 399a0 <_sk_srcover_bgra_8888_sse2_lowp+0x129c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,241,125,1,0 // addps 0x17df1(%rip),%xmm0 # 3a730 <_sk_srcover_bgra_8888_sse2_lowp+0x12a8>
+ .byte 15,88,5,113,112,1,0 // addps 0x17071(%rip),%xmm0 # 399b0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ac>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,247,125,1,0 // addps 0x17df7(%rip),%xmm0 # 3a740 <_sk_srcover_bgra_8888_sse2_lowp+0x12b8>
+ .byte 15,88,5,119,112,1,0 // addps 0x17077(%rip),%xmm0 # 399c0 <_sk_srcover_bgra_8888_sse2_lowp+0x12bc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
@@ -42961,9 +42961,9 @@ _sk_seed_shader_sse2:
.byte 102,15,110,201 // movd %ecx,%xmm1
.byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,88,13,29,118,1,0 // addps 0x1761d(%rip),%xmm1 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,13,157,104,1,0 // addps 0x1689d(%rip),%xmm1 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,21,36,118,1,0 // movaps 0x17624(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,164,104,1,0 // movaps 0x168a4(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,87,219 // xorps %xmm3,%xmm3
.byte 15,87,228 // xorps %xmm4,%xmm4
.byte 15,87,237 // xorps %xmm5,%xmm5
@@ -42978,18 +42978,18 @@ _sk_dither_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 102,68,15,110,194 // movd %edx,%xmm8
.byte 102,69,15,112,192,0 // pshufd $0x0,%xmm8,%xmm8
- .byte 102,68,15,254,5,205,117,1,0 // paddd 0x175cd(%rip),%xmm8 # 3a040 <_sk_srcover_bgra_8888_sse2_lowp+0xbb8>
+ .byte 102,68,15,254,5,77,104,1,0 // paddd 0x1684d(%rip),%xmm8 # 392c0 <_sk_srcover_bgra_8888_sse2_lowp+0xbbc>
.byte 102,68,15,110,201 // movd %ecx,%xmm9
.byte 102,69,15,112,201,0 // pshufd $0x0,%xmm9,%xmm9
.byte 102,69,15,239,200 // pxor %xmm8,%xmm9
- .byte 102,68,15,111,21,244,117,1,0 // movdqa 0x175f4(%rip),%xmm10 # 3a080 <_sk_srcover_bgra_8888_sse2_lowp+0xbf8>
+ .byte 102,68,15,111,21,116,104,1,0 // movdqa 0x16874(%rip),%xmm10 # 39300 <_sk_srcover_bgra_8888_sse2_lowp+0xbfc>
.byte 102,69,15,111,217 // movdqa %xmm9,%xmm11
.byte 102,69,15,219,218 // pand %xmm10,%xmm11
.byte 102,65,15,114,243,5 // pslld $0x5,%xmm11
.byte 102,69,15,219,208 // pand %xmm8,%xmm10
.byte 102,65,15,114,242,4 // pslld $0x4,%xmm10
- .byte 102,68,15,111,37,224,117,1,0 // movdqa 0x175e0(%rip),%xmm12 # 3a090 <_sk_srcover_bgra_8888_sse2_lowp+0xc08>
- .byte 102,68,15,111,45,231,117,1,0 // movdqa 0x175e7(%rip),%xmm13 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,37,96,104,1,0 // movdqa 0x16860(%rip),%xmm12 # 39310 <_sk_srcover_bgra_8888_sse2_lowp+0xc0c>
+ .byte 102,68,15,111,45,103,104,1,0 // movdqa 0x16867(%rip),%xmm13 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,69,15,111,241 // movdqa %xmm9,%xmm14
.byte 102,69,15,219,245 // pand %xmm13,%xmm14
.byte 102,65,15,114,246,2 // pslld $0x2,%xmm14
@@ -43005,8 +43005,8 @@ _sk_dither_sse2:
.byte 102,69,15,235,198 // por %xmm14,%xmm8
.byte 102,69,15,235,193 // por %xmm9,%xmm8
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,162,117,1,0 // mulps 0x175a2(%rip),%xmm8 # 3a0b0 <_sk_srcover_bgra_8888_sse2_lowp+0xc28>
- .byte 68,15,88,5,170,117,1,0 // addps 0x175aa(%rip),%xmm8 # 3a0c0 <_sk_srcover_bgra_8888_sse2_lowp+0xc38>
+ .byte 68,15,89,5,34,104,1,0 // mulps 0x16822(%rip),%xmm8 # 39330 <_sk_srcover_bgra_8888_sse2_lowp+0xc2c>
+ .byte 68,15,88,5,42,104,1,0 // addps 0x1682a(%rip),%xmm8 # 39340 <_sk_srcover_bgra_8888_sse2_lowp+0xc3c>
.byte 243,68,15,16,16 // movss (%rax),%xmm10
.byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
.byte 69,15,89,208 // mulps %xmm8,%xmm10
@@ -43048,7 +43048,7 @@ HIDDEN _sk_black_color_sse2
FUNCTION(_sk_black_color_sse2)
_sk_black_color_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,224,116,1,0 // movaps 0x174e0(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,96,103,1,0 // movaps 0x16760(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -43059,7 +43059,7 @@ HIDDEN _sk_white_color_sse2
FUNCTION(_sk_white_color_sse2)
_sk_white_color_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,5,204,116,1,0 // movaps 0x174cc(%rip),%xmm0 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,5,76,103,1,0 // movaps 0x1674c(%rip),%xmm0 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,40,216 // movaps %xmm0,%xmm3
@@ -43105,7 +43105,7 @@ HIDDEN _sk_srcatop_sse2
FUNCTION(_sk_srcatop_sse2)
_sk_srcatop_sse2:
.byte 15,89,199 // mulps %xmm7,%xmm0
- .byte 68,15,40,5,124,116,1,0 // movaps 0x1747c(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,252,102,1,0 // movaps 0x166fc(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -43130,7 +43130,7 @@ FUNCTION(_sk_dstatop_sse2)
_sk_dstatop_sse2:
.byte 68,15,40,195 // movaps %xmm3,%xmm8
.byte 68,15,89,196 // mulps %xmm4,%xmm8
- .byte 68,15,40,13,47,116,1,0 // movaps 0x1742f(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,175,102,1,0 // movaps 0x166af(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,88,192 // addps %xmm8,%xmm0
@@ -43177,7 +43177,7 @@ HIDDEN _sk_srcout_sse2
.globl _sk_srcout_sse2
FUNCTION(_sk_srcout_sse2)
_sk_srcout_sse2:
- .byte 68,15,40,5,195,115,1,0 // movaps 0x173c3(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,67,102,1,0 // movaps 0x16643(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,199 // subps %xmm7,%xmm8
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
@@ -43190,7 +43190,7 @@ HIDDEN _sk_dstout_sse2
.globl _sk_dstout_sse2
FUNCTION(_sk_dstout_sse2)
_sk_dstout_sse2:
- .byte 68,15,40,5,163,115,1,0 // movaps 0x173a3(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,35,102,1,0 // movaps 0x16623(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 15,89,196 // mulps %xmm4,%xmm0
@@ -43207,7 +43207,7 @@ HIDDEN _sk_srcover_sse2
.globl _sk_srcover_sse2
FUNCTION(_sk_srcover_sse2)
_sk_srcover_sse2:
- .byte 68,15,40,5,118,115,1,0 // movaps 0x17376(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,246,101,1,0 // movaps 0x165f6(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -43227,7 +43227,7 @@ HIDDEN _sk_dstover_sse2
.globl _sk_dstover_sse2
FUNCTION(_sk_dstover_sse2)
_sk_dstover_sse2:
- .byte 68,15,40,5,58,115,1,0 // movaps 0x1733a(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,186,101,1,0 // movaps 0x165ba(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,199 // subps %xmm7,%xmm8
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 15,88,196 // addps %xmm4,%xmm0
@@ -43255,7 +43255,7 @@ HIDDEN _sk_multiply_sse2
.globl _sk_multiply_sse2
FUNCTION(_sk_multiply_sse2)
_sk_multiply_sse2:
- .byte 68,15,40,5,254,114,1,0 // movaps 0x172fe(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,126,101,1,0 // movaps 0x1657e(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 69,15,40,209 // movaps %xmm9,%xmm10
@@ -43293,7 +43293,7 @@ HIDDEN _sk_plus__sse2
FUNCTION(_sk_plus__sse2)
_sk_plus__sse2:
.byte 15,88,196 // addps %xmm4,%xmm0
- .byte 68,15,40,5,127,114,1,0 // movaps 0x1727f(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,255,100,1,0 // movaps 0x164ff(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 15,88,205 // addps %xmm5,%xmm1
.byte 65,15,93,200 // minps %xmm8,%xmm1
@@ -43335,7 +43335,7 @@ HIDDEN _sk_xor__sse2
FUNCTION(_sk_xor__sse2)
_sk_xor__sse2:
.byte 68,15,40,195 // movaps %xmm3,%xmm8
- .byte 15,40,29,11,114,1,0 // movaps 0x1720b(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,139,100,1,0 // movaps 0x1648b(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,40,203 // movaps %xmm3,%xmm9
.byte 68,15,92,207 // subps %xmm7,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
@@ -43383,7 +43383,7 @@ _sk_darken_sse2:
.byte 68,15,89,206 // mulps %xmm6,%xmm9
.byte 65,15,95,209 // maxps %xmm9,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,102,113,1,0 // movaps 0x17166(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,230,99,1,0 // movaps 0x163e6(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -43417,7 +43417,7 @@ _sk_lighten_sse2:
.byte 68,15,89,206 // mulps %xmm6,%xmm9
.byte 65,15,93,209 // minps %xmm9,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,251,112,1,0 // movaps 0x170fb(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,123,99,1,0 // movaps 0x1637b(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -43454,7 +43454,7 @@ _sk_difference_sse2:
.byte 65,15,93,209 // minps %xmm9,%xmm2
.byte 15,88,210 // addps %xmm2,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,133,112,1,0 // movaps 0x17085(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,5,99,1,0 // movaps 0x16305(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -43482,7 +43482,7 @@ _sk_exclusion_sse2:
.byte 15,89,214 // mulps %xmm6,%xmm2
.byte 15,88,210 // addps %xmm2,%xmm2
.byte 68,15,92,194 // subps %xmm2,%xmm8
- .byte 15,40,21,53,112,1,0 // movaps 0x17035(%rip),%xmm2 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,21,181,98,1,0 // movaps 0x162b5(%rip),%xmm2 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,211 // subps %xmm3,%xmm2
.byte 15,89,215 // mulps %xmm7,%xmm2
.byte 15,88,218 // addps %xmm2,%xmm3
@@ -43495,7 +43495,7 @@ HIDDEN _sk_colorburn_sse2
FUNCTION(_sk_colorburn_sse2)
_sk_colorburn_sse2:
.byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 68,15,40,13,24,112,1,0 // movaps 0x17018(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,152,98,1,0 // movaps 0x16298(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,209 // movaps %xmm9,%xmm10
.byte 68,15,92,215 // subps %xmm7,%xmm10
.byte 69,15,40,218 // movaps %xmm10,%xmm11
@@ -43590,7 +43590,7 @@ HIDDEN _sk_colordodge_sse2
FUNCTION(_sk_colordodge_sse2)
_sk_colordodge_sse2:
.byte 68,15,40,200 // movaps %xmm0,%xmm9
- .byte 68,15,40,21,192,110,1,0 // movaps 0x16ec0(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,64,97,1,0 // movaps 0x16140(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,218 // movaps %xmm10,%xmm11
.byte 68,15,92,223 // subps %xmm7,%xmm11
.byte 65,15,40,195 // movaps %xmm11,%xmm0
@@ -43682,7 +43682,7 @@ _sk_hardlight_sse2:
.byte 15,41,116,36,232 // movaps %xmm6,-0x18(%rsp)
.byte 15,40,245 // movaps %xmm5,%xmm6
.byte 15,40,236 // movaps %xmm4,%xmm5
- .byte 68,15,40,29,105,109,1,0 // movaps 0x16d69(%rip),%xmm11 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,29,233,95,1,0 // movaps 0x15fe9(%rip),%xmm11 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,211 // movaps %xmm11,%xmm10
.byte 68,15,92,215 // subps %xmm7,%xmm10
.byte 69,15,40,194 // movaps %xmm10,%xmm8
@@ -43770,7 +43770,7 @@ FUNCTION(_sk_overlay_sse2)
_sk_overlay_sse2:
.byte 68,15,40,193 // movaps %xmm1,%xmm8
.byte 68,15,40,232 // movaps %xmm0,%xmm13
- .byte 68,15,40,13,39,108,1,0 // movaps 0x16c27(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,167,94,1,0 // movaps 0x15ea7(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,209 // movaps %xmm9,%xmm10
.byte 68,15,92,215 // subps %xmm7,%xmm10
.byte 69,15,40,218 // movaps %xmm10,%xmm11
@@ -43861,7 +43861,7 @@ _sk_softlight_sse2:
.byte 68,15,40,213 // movaps %xmm5,%xmm10
.byte 68,15,94,215 // divps %xmm7,%xmm10
.byte 69,15,84,212 // andps %xmm12,%xmm10
- .byte 68,15,40,13,212,106,1,0 // movaps 0x16ad4(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,84,93,1,0 // movaps 0x15d54(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,249 // movaps %xmm9,%xmm15
.byte 69,15,92,250 // subps %xmm10,%xmm15
.byte 69,15,40,218 // movaps %xmm10,%xmm11
@@ -43874,10 +43874,10 @@ _sk_softlight_sse2:
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 65,15,88,194 // addps %xmm10,%xmm0
- .byte 68,15,40,53,254,106,1,0 // movaps 0x16afe(%rip),%xmm14 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,53,126,93,1,0 // movaps 0x15d7e(%rip),%xmm14 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 69,15,88,222 // addps %xmm14,%xmm11
.byte 68,15,89,216 // mulps %xmm0,%xmm11
- .byte 68,15,40,21,254,106,1,0 // movaps 0x16afe(%rip),%xmm10 # 3a0e0 <_sk_srcover_bgra_8888_sse2_lowp+0xc58>
+ .byte 68,15,40,21,126,93,1,0 // movaps 0x15d7e(%rip),%xmm10 # 39360 <_sk_srcover_bgra_8888_sse2_lowp+0xc5c>
.byte 69,15,89,234 // mulps %xmm10,%xmm13
.byte 69,15,88,235 // addps %xmm11,%xmm13
.byte 15,88,228 // addps %xmm4,%xmm4
@@ -44022,7 +44022,7 @@ _sk_hue_sse2:
.byte 68,15,40,209 // movaps %xmm1,%xmm10
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 68,15,89,211 // mulps %xmm3,%xmm10
- .byte 68,15,40,5,145,104,1,0 // movaps 0x16891(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,17,91,1,0 // movaps 0x15b11(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,216 // movaps %xmm8,%xmm11
.byte 15,40,207 // movaps %xmm7,%xmm1
.byte 68,15,92,217 // subps %xmm1,%xmm11
@@ -44070,12 +44070,12 @@ _sk_hue_sse2:
.byte 69,15,84,206 // andps %xmm14,%xmm9
.byte 69,15,84,214 // andps %xmm14,%xmm10
.byte 65,15,84,214 // andps %xmm14,%xmm2
- .byte 68,15,40,61,78,104,1,0 // movaps 0x1684e(%rip),%xmm15 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,61,206,90,1,0 // movaps 0x15ace(%rip),%xmm15 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 65,15,89,231 // mulps %xmm15,%xmm4
- .byte 15,40,5,83,104,1,0 // movaps 0x16853(%rip),%xmm0 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 15,40,5,211,90,1,0 // movaps 0x15ad3(%rip),%xmm0 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 15,89,240 // mulps %xmm0,%xmm6
.byte 15,88,244 // addps %xmm4,%xmm6
- .byte 68,15,40,53,85,104,1,0 // movaps 0x16855(%rip),%xmm14 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,53,213,90,1,0 // movaps 0x15ad5(%rip),%xmm14 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 68,15,40,239 // movaps %xmm7,%xmm13
.byte 69,15,89,238 // mulps %xmm14,%xmm13
.byte 68,15,88,238 // addps %xmm6,%xmm13
@@ -44253,14 +44253,14 @@ _sk_saturation_sse2:
.byte 68,15,84,211 // andps %xmm3,%xmm10
.byte 68,15,84,203 // andps %xmm3,%xmm9
.byte 15,84,195 // andps %xmm3,%xmm0
- .byte 68,15,40,5,167,101,1,0 // movaps 0x165a7(%rip),%xmm8 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,5,39,88,1,0 // movaps 0x15827(%rip),%xmm8 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 15,40,214 // movaps %xmm6,%xmm2
.byte 65,15,89,208 // mulps %xmm8,%xmm2
- .byte 15,40,13,169,101,1,0 // movaps 0x165a9(%rip),%xmm1 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 15,40,13,41,88,1,0 // movaps 0x15829(%rip),%xmm1 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 15,40,221 // movaps %xmm5,%xmm3
.byte 15,89,217 // mulps %xmm1,%xmm3
.byte 15,88,218 // addps %xmm2,%xmm3
- .byte 68,15,40,37,168,101,1,0 // movaps 0x165a8(%rip),%xmm12 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,37,40,88,1,0 // movaps 0x15828(%rip),%xmm12 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 69,15,89,236 // mulps %xmm12,%xmm13
.byte 68,15,88,235 // addps %xmm3,%xmm13
.byte 65,15,40,210 // movaps %xmm10,%xmm2
@@ -44305,7 +44305,7 @@ _sk_saturation_sse2:
.byte 15,40,223 // movaps %xmm7,%xmm3
.byte 15,40,236 // movaps %xmm4,%xmm5
.byte 15,89,221 // mulps %xmm5,%xmm3
- .byte 68,15,40,5,93,100,1,0 // movaps 0x1645d(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,221,86,1,0 // movaps 0x156dd(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,40,224 // movaps %xmm8,%xmm4
.byte 68,15,92,199 // subps %xmm7,%xmm8
.byte 15,88,253 // addps %xmm5,%xmm7
@@ -44406,14 +44406,14 @@ _sk_color_sse2:
.byte 68,15,40,213 // movaps %xmm5,%xmm10
.byte 69,15,89,208 // mulps %xmm8,%xmm10
.byte 65,15,40,208 // movaps %xmm8,%xmm2
- .byte 68,15,40,45,107,99,1,0 // movaps 0x1636b(%rip),%xmm13 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,45,235,85,1,0 // movaps 0x155eb(%rip),%xmm13 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,40,198 // movaps %xmm6,%xmm8
.byte 69,15,89,197 // mulps %xmm13,%xmm8
- .byte 68,15,40,53,107,99,1,0 // movaps 0x1636b(%rip),%xmm14 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 68,15,40,53,235,85,1,0 // movaps 0x155eb(%rip),%xmm14 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 65,15,40,195 // movaps %xmm11,%xmm0
.byte 65,15,89,198 // mulps %xmm14,%xmm0
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,29,103,99,1,0 // movaps 0x16367(%rip),%xmm11 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
+ .byte 68,15,40,29,231,85,1,0 // movaps 0x155e7(%rip),%xmm11 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
.byte 69,15,89,227 // mulps %xmm11,%xmm12
.byte 68,15,88,224 // addps %xmm0,%xmm12
.byte 65,15,40,193 // movaps %xmm9,%xmm0
@@ -44421,7 +44421,7 @@ _sk_color_sse2:
.byte 69,15,40,250 // movaps %xmm10,%xmm15
.byte 69,15,89,254 // mulps %xmm14,%xmm15
.byte 68,15,88,248 // addps %xmm0,%xmm15
- .byte 68,15,40,5,163,98,1,0 // movaps 0x162a3(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,35,85,1,0 // movaps 0x15523(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,40,224 // movaps %xmm8,%xmm4
.byte 15,92,226 // subps %xmm2,%xmm4
.byte 15,89,252 // mulps %xmm4,%xmm7
@@ -44557,15 +44557,15 @@ _sk_luminosity_sse2:
.byte 68,15,40,205 // movaps %xmm5,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
.byte 15,89,222 // mulps %xmm6,%xmm3
- .byte 68,15,40,37,42,97,1,0 // movaps 0x1612a(%rip),%xmm12 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,37,170,83,1,0 // movaps 0x153aa(%rip),%xmm12 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,40,199 // movaps %xmm7,%xmm8
.byte 69,15,89,196 // mulps %xmm12,%xmm8
- .byte 68,15,40,45,42,97,1,0 // movaps 0x1612a(%rip),%xmm13 # 3a100 <_sk_srcover_bgra_8888_sse2_lowp+0xc78>
+ .byte 68,15,40,45,170,83,1,0 // movaps 0x153aa(%rip),%xmm13 # 39380 <_sk_srcover_bgra_8888_sse2_lowp+0xc7c>
.byte 68,15,40,241 // movaps %xmm1,%xmm14
.byte 69,15,89,245 // mulps %xmm13,%xmm14
.byte 69,15,88,240 // addps %xmm8,%xmm14
- .byte 68,15,40,29,38,97,1,0 // movaps 0x16126(%rip),%xmm11 # 3a110 <_sk_srcover_bgra_8888_sse2_lowp+0xc88>
- .byte 68,15,40,5,126,96,1,0 // movaps 0x1607e(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,29,166,83,1,0 // movaps 0x153a6(%rip),%xmm11 # 39390 <_sk_srcover_bgra_8888_sse2_lowp+0xc8c>
+ .byte 68,15,40,5,254,82,1,0 // movaps 0x152fe(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,248 // movaps %xmm8,%xmm15
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 68,15,92,248 // subps %xmm0,%xmm15
@@ -44709,7 +44709,7 @@ _sk_srcover_rgba_8888_sse2:
.byte 15,133,228,0,0,0 // jne 242dd <_sk_srcover_rgba_8888_sse2+0x104>
.byte 243,69,15,111,4,128 // movdqu (%r8,%rax,4),%xmm8
.byte 72,133,255 // test %rdi,%rdi
- .byte 102,15,111,53,22,95,1,0 // movdqa 0x15f16(%rip),%xmm6 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,53,150,81,1,0 // movdqa 0x15196(%rip),%xmm6 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,224 // movdqa %xmm8,%xmm4
.byte 102,15,219,230 // pand %xmm6,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
@@ -44723,9 +44723,9 @@ _sk_srcover_rgba_8888_sse2:
.byte 15,91,247 // cvtdq2ps %xmm7,%xmm6
.byte 102,65,15,114,208,24 // psrld $0x18,%xmm8
.byte 65,15,91,248 // cvtdq2ps %xmm8,%xmm7
- .byte 68,15,40,5,38,94,1,0 // movaps 0x15e26(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,166,80,1,0 // movaps 0x150a6(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
- .byte 68,15,40,37,250,94,1,0 // movaps 0x15efa(%rip),%xmm12 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,37,122,81,1,0 // movaps 0x1517a(%rip),%xmm12 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,196 // mulps %xmm12,%xmm0
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -44804,7 +44804,7 @@ _sk_srcover_bgra_8888_sse2:
.byte 15,133,228,0,0,0 // jne 2445e <_sk_srcover_bgra_8888_sse2+0x104>
.byte 243,69,15,111,4,128 // movdqu (%r8,%rax,4),%xmm8
.byte 72,133,255 // test %rdi,%rdi
- .byte 102,15,111,37,149,93,1,0 // movdqa 0x15d95(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,21,80,1,0 // movdqa 0x15015(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,232 // movdqa %xmm8,%xmm5
.byte 102,15,219,236 // pand %xmm4,%xmm5
.byte 15,91,245 // cvtdq2ps %xmm5,%xmm6
@@ -44818,9 +44818,9 @@ _sk_srcover_bgra_8888_sse2:
.byte 15,91,231 // cvtdq2ps %xmm7,%xmm4
.byte 102,65,15,114,208,24 // psrld $0x18,%xmm8
.byte 65,15,91,248 // cvtdq2ps %xmm8,%xmm7
- .byte 68,15,40,5,165,92,1,0 // movaps 0x15ca5(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,37,79,1,0 // movaps 0x14f25(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,92,195 // subps %xmm3,%xmm8
- .byte 68,15,40,37,121,93,1,0 // movaps 0x15d79(%rip),%xmm12 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,37,249,79,1,0 // movaps 0x14ff9(%rip),%xmm12 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,196 // mulps %xmm12,%xmm0
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,89,204 // mulps %xmm4,%xmm9
@@ -44900,7 +44900,7 @@ HIDDEN _sk_clamp_1_sse2
.globl _sk_clamp_1_sse2
FUNCTION(_sk_clamp_1_sse2)
_sk_clamp_1_sse2:
- .byte 68,15,40,5,117,91,1,0 // movaps 0x15b75(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,245,77,1,0 // movaps 0x14df5(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 65,15,93,208 // minps %xmm8,%xmm2
@@ -44912,7 +44912,7 @@ HIDDEN _sk_clamp_a_sse2
.globl _sk_clamp_a_sse2
FUNCTION(_sk_clamp_a_sse2)
_sk_clamp_a_sse2:
- .byte 15,93,29,90,91,1,0 // minps 0x15b5a(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,29,218,77,1,0 // minps 0x14dda(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,93,195 // minps %xmm3,%xmm0
.byte 15,93,203 // minps %xmm3,%xmm1
.byte 15,93,211 // minps %xmm3,%xmm2
@@ -44923,7 +44923,7 @@ HIDDEN _sk_clamp_a_dst_sse2
.globl _sk_clamp_a_dst_sse2
FUNCTION(_sk_clamp_a_dst_sse2)
_sk_clamp_a_dst_sse2:
- .byte 15,93,61,70,91,1,0 // minps 0x15b46(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,93,61,198,77,1,0 // minps 0x14dc6(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,93,231 // minps %xmm7,%xmm4
.byte 15,93,239 // minps %xmm7,%xmm5
.byte 15,93,247 // minps %xmm7,%xmm6
@@ -44958,7 +44958,7 @@ HIDDEN _sk_invert_sse2
.globl _sk_invert_sse2
FUNCTION(_sk_invert_sse2)
_sk_invert_sse2:
- .byte 68,15,40,5,2,91,1,0 // movaps 0x15b02(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,130,77,1,0 // movaps 0x14d82(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 68,15,92,200 // subps %xmm0,%xmm9
.byte 69,15,40,208 // movaps %xmm8,%xmm10
@@ -45019,10 +45019,10 @@ HIDDEN _sk_unpremul_sse2
.globl _sk_unpremul_sse2
FUNCTION(_sk_unpremul_sse2)
_sk_unpremul_sse2:
- .byte 68,15,40,5,144,90,1,0 // movaps 0x15a90(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,16,77,1,0 // movaps 0x14d10(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,94,195 // divps %xmm3,%xmm8
.byte 69,15,40,200 // movaps %xmm8,%xmm9
- .byte 68,15,194,13,111,91,1,0,1 // cmpltps 0x15b6f(%rip),%xmm9 # 3a160 <_sk_srcover_bgra_8888_sse2_lowp+0xcd8>
+ .byte 68,15,194,13,239,77,1,0,1 // cmpltps 0x14def(%rip),%xmm9 # 393e0 <_sk_srcover_bgra_8888_sse2_lowp+0xcdc>
.byte 69,15,84,200 // andps %xmm8,%xmm9
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 65,15,89,201 // mulps %xmm9,%xmm1
@@ -45034,20 +45034,20 @@ HIDDEN _sk_from_srgb_sse2
.globl _sk_from_srgb_sse2
FUNCTION(_sk_from_srgb_sse2)
_sk_from_srgb_sse2:
- .byte 68,15,40,5,99,91,1,0 // movaps 0x15b63(%rip),%xmm8 # 3a170 <_sk_srcover_bgra_8888_sse2_lowp+0xce8>
+ .byte 68,15,40,5,227,77,1,0 // movaps 0x14de3(%rip),%xmm8 # 393f0 <_sk_srcover_bgra_8888_sse2_lowp+0xcec>
.byte 68,15,40,232 // movaps %xmm0,%xmm13
.byte 69,15,89,232 // mulps %xmm8,%xmm13
.byte 68,15,40,216 // movaps %xmm0,%xmm11
.byte 69,15,89,219 // mulps %xmm11,%xmm11
- .byte 68,15,40,13,203,90,1,0 // movaps 0x15acb(%rip),%xmm9 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,13,75,77,1,0 // movaps 0x14d4b(%rip),%xmm9 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,40,240 // movaps %xmm0,%xmm14
.byte 69,15,89,241 // mulps %xmm9,%xmm14
- .byte 68,15,40,21,75,91,1,0 // movaps 0x15b4b(%rip),%xmm10 # 3a180 <_sk_srcover_bgra_8888_sse2_lowp+0xcf8>
+ .byte 68,15,40,21,203,77,1,0 // movaps 0x14dcb(%rip),%xmm10 # 39400 <_sk_srcover_bgra_8888_sse2_lowp+0xcfc>
.byte 69,15,88,242 // addps %xmm10,%xmm14
.byte 69,15,89,243 // mulps %xmm11,%xmm14
- .byte 68,15,40,29,75,91,1,0 // movaps 0x15b4b(%rip),%xmm11 # 3a190 <_sk_srcover_bgra_8888_sse2_lowp+0xd08>
+ .byte 68,15,40,29,203,77,1,0 // movaps 0x14dcb(%rip),%xmm11 # 39410 <_sk_srcover_bgra_8888_sse2_lowp+0xd0c>
.byte 69,15,88,243 // addps %xmm11,%xmm14
- .byte 68,15,40,37,79,91,1,0 // movaps 0x15b4f(%rip),%xmm12 # 3a1a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd18>
+ .byte 68,15,40,37,207,77,1,0 // movaps 0x14dcf(%rip),%xmm12 # 39420 <_sk_srcover_bgra_8888_sse2_lowp+0xd1c>
.byte 65,15,194,196,1 // cmpltps %xmm12,%xmm0
.byte 68,15,84,232 // andps %xmm0,%xmm13
.byte 65,15,85,198 // andnps %xmm14,%xmm0
@@ -45083,20 +45083,20 @@ HIDDEN _sk_from_srgb_dst_sse2
.globl _sk_from_srgb_dst_sse2
FUNCTION(_sk_from_srgb_dst_sse2)
_sk_from_srgb_dst_sse2:
- .byte 68,15,40,5,160,90,1,0 // movaps 0x15aa0(%rip),%xmm8 # 3a170 <_sk_srcover_bgra_8888_sse2_lowp+0xce8>
+ .byte 68,15,40,5,32,77,1,0 // movaps 0x14d20(%rip),%xmm8 # 393f0 <_sk_srcover_bgra_8888_sse2_lowp+0xcec>
.byte 68,15,40,236 // movaps %xmm4,%xmm13
.byte 69,15,89,232 // mulps %xmm8,%xmm13
.byte 68,15,40,220 // movaps %xmm4,%xmm11
.byte 69,15,89,219 // mulps %xmm11,%xmm11
- .byte 68,15,40,13,8,90,1,0 // movaps 0x15a08(%rip),%xmm9 # 3a0f0 <_sk_srcover_bgra_8888_sse2_lowp+0xc68>
+ .byte 68,15,40,13,136,76,1,0 // movaps 0x14c88(%rip),%xmm9 # 39370 <_sk_srcover_bgra_8888_sse2_lowp+0xc6c>
.byte 68,15,40,244 // movaps %xmm4,%xmm14
.byte 69,15,89,241 // mulps %xmm9,%xmm14
- .byte 68,15,40,21,136,90,1,0 // movaps 0x15a88(%rip),%xmm10 # 3a180 <_sk_srcover_bgra_8888_sse2_lowp+0xcf8>
+ .byte 68,15,40,21,8,77,1,0 // movaps 0x14d08(%rip),%xmm10 # 39400 <_sk_srcover_bgra_8888_sse2_lowp+0xcfc>
.byte 69,15,88,242 // addps %xmm10,%xmm14
.byte 69,15,89,243 // mulps %xmm11,%xmm14
- .byte 68,15,40,29,136,90,1,0 // movaps 0x15a88(%rip),%xmm11 # 3a190 <_sk_srcover_bgra_8888_sse2_lowp+0xd08>
+ .byte 68,15,40,29,8,77,1,0 // movaps 0x14d08(%rip),%xmm11 # 39410 <_sk_srcover_bgra_8888_sse2_lowp+0xd0c>
.byte 69,15,88,243 // addps %xmm11,%xmm14
- .byte 68,15,40,37,140,90,1,0 // movaps 0x15a8c(%rip),%xmm12 # 3a1a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd18>
+ .byte 68,15,40,37,12,77,1,0 // movaps 0x14d0c(%rip),%xmm12 # 39420 <_sk_srcover_bgra_8888_sse2_lowp+0xd1c>
.byte 65,15,194,228,1 // cmpltps %xmm12,%xmm4
.byte 68,15,84,236 // andps %xmm4,%xmm13
.byte 65,15,85,230 // andnps %xmm14,%xmm4
@@ -45133,22 +45133,22 @@ HIDDEN _sk_to_srgb_sse2
FUNCTION(_sk_to_srgb_sse2)
_sk_to_srgb_sse2:
.byte 68,15,82,232 // rsqrtps %xmm0,%xmm13
- .byte 68,15,40,5,25,90,1,0 // movaps 0x15a19(%rip),%xmm8 # 3a1b0 <_sk_srcover_bgra_8888_sse2_lowp+0xd28>
+ .byte 68,15,40,5,153,76,1,0 // movaps 0x14c99(%rip),%xmm8 # 39430 <_sk_srcover_bgra_8888_sse2_lowp+0xd2c>
.byte 68,15,40,240 // movaps %xmm0,%xmm14
.byte 69,15,89,240 // mulps %xmm8,%xmm14
- .byte 68,15,40,13,25,90,1,0 // movaps 0x15a19(%rip),%xmm9 # 3a1c0 <_sk_srcover_bgra_8888_sse2_lowp+0xd38>
+ .byte 68,15,40,13,153,76,1,0 // movaps 0x14c99(%rip),%xmm9 # 39440 <_sk_srcover_bgra_8888_sse2_lowp+0xd3c>
.byte 69,15,40,253 // movaps %xmm13,%xmm15
.byte 69,15,89,249 // mulps %xmm9,%xmm15
- .byte 68,15,40,21,25,90,1,0 // movaps 0x15a19(%rip),%xmm10 # 3a1d0 <_sk_srcover_bgra_8888_sse2_lowp+0xd48>
+ .byte 68,15,40,21,153,76,1,0 // movaps 0x14c99(%rip),%xmm10 # 39450 <_sk_srcover_bgra_8888_sse2_lowp+0xd4c>
.byte 69,15,88,250 // addps %xmm10,%xmm15
.byte 69,15,89,253 // mulps %xmm13,%xmm15
- .byte 68,15,40,29,25,90,1,0 // movaps 0x15a19(%rip),%xmm11 # 3a1e0 <_sk_srcover_bgra_8888_sse2_lowp+0xd58>
+ .byte 68,15,40,29,153,76,1,0 // movaps 0x14c99(%rip),%xmm11 # 39460 <_sk_srcover_bgra_8888_sse2_lowp+0xd5c>
.byte 69,15,88,251 // addps %xmm11,%xmm15
- .byte 68,15,40,37,29,90,1,0 // movaps 0x15a1d(%rip),%xmm12 # 3a1f0 <_sk_srcover_bgra_8888_sse2_lowp+0xd68>
+ .byte 68,15,40,37,157,76,1,0 // movaps 0x14c9d(%rip),%xmm12 # 39470 <_sk_srcover_bgra_8888_sse2_lowp+0xd6c>
.byte 69,15,88,236 // addps %xmm12,%xmm13
.byte 69,15,83,237 // rcpps %xmm13,%xmm13
.byte 69,15,89,239 // mulps %xmm15,%xmm13
- .byte 68,15,40,61,25,90,1,0 // movaps 0x15a19(%rip),%xmm15 # 3a200 <_sk_srcover_bgra_8888_sse2_lowp+0xd78>
+ .byte 68,15,40,61,153,76,1,0 // movaps 0x14c99(%rip),%xmm15 # 39480 <_sk_srcover_bgra_8888_sse2_lowp+0xd7c>
.byte 65,15,194,199,1 // cmpltps %xmm15,%xmm0
.byte 68,15,84,240 // andps %xmm0,%xmm14
.byte 65,15,85,197 // andnps %xmm13,%xmm0
@@ -45200,7 +45200,7 @@ _sk_rgb_to_hsl_sse2:
.byte 68,15,93,216 // minps %xmm0,%xmm11
.byte 65,15,40,202 // movaps %xmm10,%xmm1
.byte 65,15,92,203 // subps %xmm11,%xmm1
- .byte 68,15,40,45,204,87,1,0 // movaps 0x157cc(%rip),%xmm13 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,45,76,74,1,0 // movaps 0x14a4c(%rip),%xmm13 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 68,15,94,233 // divps %xmm1,%xmm13
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 65,15,194,192,0 // cmpeqps %xmm8,%xmm0
@@ -45209,30 +45209,30 @@ _sk_rgb_to_hsl_sse2:
.byte 69,15,89,229 // mulps %xmm13,%xmm12
.byte 69,15,40,241 // movaps %xmm9,%xmm14
.byte 68,15,194,242,1 // cmpltps %xmm2,%xmm14
- .byte 68,15,84,53,66,89,1,0 // andps 0x15942(%rip),%xmm14 # 3a210 <_sk_srcover_bgra_8888_sse2_lowp+0xd88>
+ .byte 68,15,84,53,194,75,1,0 // andps 0x14bc2(%rip),%xmm14 # 39490 <_sk_srcover_bgra_8888_sse2_lowp+0xd8c>
.byte 69,15,88,244 // addps %xmm12,%xmm14
.byte 69,15,40,250 // movaps %xmm10,%xmm15
.byte 69,15,194,249,0 // cmpeqps %xmm9,%xmm15
.byte 65,15,92,208 // subps %xmm8,%xmm2
.byte 65,15,89,213 // mulps %xmm13,%xmm2
- .byte 68,15,40,37,53,89,1,0 // movaps 0x15935(%rip),%xmm12 # 3a220 <_sk_srcover_bgra_8888_sse2_lowp+0xd98>
+ .byte 68,15,40,37,181,75,1,0 // movaps 0x14bb5(%rip),%xmm12 # 394a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd9c>
.byte 65,15,88,212 // addps %xmm12,%xmm2
.byte 69,15,92,193 // subps %xmm9,%xmm8
.byte 69,15,89,197 // mulps %xmm13,%xmm8
- .byte 68,15,88,5,49,89,1,0 // addps 0x15931(%rip),%xmm8 # 3a230 <_sk_srcover_bgra_8888_sse2_lowp+0xda8>
+ .byte 68,15,88,5,177,75,1,0 // addps 0x14bb1(%rip),%xmm8 # 394b0 <_sk_srcover_bgra_8888_sse2_lowp+0xdac>
.byte 65,15,84,215 // andps %xmm15,%xmm2
.byte 69,15,85,248 // andnps %xmm8,%xmm15
.byte 68,15,86,250 // orps %xmm2,%xmm15
.byte 68,15,84,240 // andps %xmm0,%xmm14
.byte 65,15,85,199 // andnps %xmm15,%xmm0
.byte 65,15,86,198 // orps %xmm14,%xmm0
- .byte 15,89,5,34,89,1,0 // mulps 0x15922(%rip),%xmm0 # 3a240 <_sk_srcover_bgra_8888_sse2_lowp+0xdb8>
+ .byte 15,89,5,162,75,1,0 // mulps 0x14ba2(%rip),%xmm0 # 394c0 <_sk_srcover_bgra_8888_sse2_lowp+0xdbc>
.byte 69,15,40,194 // movaps %xmm10,%xmm8
.byte 69,15,194,195,4 // cmpneqps %xmm11,%xmm8
.byte 65,15,84,192 // andps %xmm8,%xmm0
.byte 69,15,92,226 // subps %xmm10,%xmm12
.byte 69,15,88,211 // addps %xmm11,%xmm10
- .byte 68,15,40,13,37,87,1,0 // movaps 0x15725(%rip),%xmm9 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,13,165,73,1,0 // movaps 0x149a5(%rip),%xmm9 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,40,210 // movaps %xmm10,%xmm2
.byte 65,15,89,209 // mulps %xmm9,%xmm2
.byte 68,15,194,202,1 // cmpltps %xmm2,%xmm9
@@ -45256,7 +45256,7 @@ _sk_hsl_to_rgb_sse2:
.byte 15,41,92,36,168 // movaps %xmm3,-0x58(%rsp)
.byte 68,15,40,218 // movaps %xmm2,%xmm11
.byte 15,40,240 // movaps %xmm0,%xmm6
- .byte 68,15,40,13,212,86,1,0 // movaps 0x156d4(%rip),%xmm9 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,13,84,73,1,0 // movaps 0x14954(%rip),%xmm9 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,40,209 // movaps %xmm9,%xmm10
.byte 69,15,194,211,2 // cmpleps %xmm11,%xmm10
.byte 15,40,193 // movaps %xmm1,%xmm0
@@ -45273,28 +45273,28 @@ _sk_hsl_to_rgb_sse2:
.byte 69,15,88,211 // addps %xmm11,%xmm10
.byte 69,15,88,219 // addps %xmm11,%xmm11
.byte 69,15,92,218 // subps %xmm10,%xmm11
- .byte 15,40,5,125,88,1,0 // movaps 0x1587d(%rip),%xmm0 # 3a250 <_sk_srcover_bgra_8888_sse2_lowp+0xdc8>
+ .byte 15,40,5,253,74,1,0 // movaps 0x14afd(%rip),%xmm0 # 394d0 <_sk_srcover_bgra_8888_sse2_lowp+0xdcc>
.byte 15,88,198 // addps %xmm6,%xmm0
.byte 243,15,91,200 // cvttps2dq %xmm0,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
.byte 15,40,216 // movaps %xmm0,%xmm3
.byte 15,194,217,1 // cmpltps %xmm1,%xmm3
- .byte 15,84,29,133,86,1,0 // andps 0x15685(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,84,29,5,73,1,0 // andps 0x14905(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,203 // subps %xmm3,%xmm1
.byte 15,92,193 // subps %xmm1,%xmm0
- .byte 68,15,40,45,103,88,1,0 // movaps 0x15867(%rip),%xmm13 # 3a260 <_sk_srcover_bgra_8888_sse2_lowp+0xdd8>
+ .byte 68,15,40,45,231,74,1,0 // movaps 0x14ae7(%rip),%xmm13 # 394e0 <_sk_srcover_bgra_8888_sse2_lowp+0xddc>
.byte 69,15,40,197 // movaps %xmm13,%xmm8
.byte 68,15,194,192,2 // cmpleps %xmm0,%xmm8
.byte 69,15,40,242 // movaps %xmm10,%xmm14
.byte 69,15,92,243 // subps %xmm11,%xmm14
.byte 65,15,40,217 // movaps %xmm9,%xmm3
.byte 15,194,216,2 // cmpleps %xmm0,%xmm3
- .byte 15,40,21,39,88,1,0 // movaps 0x15827(%rip),%xmm2 # 3a240 <_sk_srcover_bgra_8888_sse2_lowp+0xdb8>
+ .byte 15,40,21,167,74,1,0 // movaps 0x14aa7(%rip),%xmm2 # 394c0 <_sk_srcover_bgra_8888_sse2_lowp+0xdbc>
.byte 68,15,40,250 // movaps %xmm2,%xmm15
.byte 68,15,194,248,2 // cmpleps %xmm0,%xmm15
- .byte 15,40,13,231,87,1,0 // movaps 0x157e7(%rip),%xmm1 # 3a210 <_sk_srcover_bgra_8888_sse2_lowp+0xd88>
+ .byte 15,40,13,103,74,1,0 // movaps 0x14a67(%rip),%xmm1 # 39490 <_sk_srcover_bgra_8888_sse2_lowp+0xd8c>
.byte 15,89,193 // mulps %xmm1,%xmm0
- .byte 15,40,45,253,87,1,0 // movaps 0x157fd(%rip),%xmm5 # 3a230 <_sk_srcover_bgra_8888_sse2_lowp+0xda8>
+ .byte 15,40,45,125,74,1,0 // movaps 0x14a7d(%rip),%xmm5 # 394b0 <_sk_srcover_bgra_8888_sse2_lowp+0xdac>
.byte 15,40,229 // movaps %xmm5,%xmm4
.byte 15,92,224 // subps %xmm0,%xmm4
.byte 65,15,89,230 // mulps %xmm14,%xmm4
@@ -45317,7 +45317,7 @@ _sk_hsl_to_rgb_sse2:
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
.byte 15,40,222 // movaps %xmm6,%xmm3
.byte 15,194,216,1 // cmpltps %xmm0,%xmm3
- .byte 15,84,29,226,85,1,0 // andps 0x155e2(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,84,29,98,72,1,0 // andps 0x14862(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,195 // subps %xmm3,%xmm0
.byte 68,15,40,230 // movaps %xmm6,%xmm12
.byte 68,15,92,224 // subps %xmm0,%xmm12
@@ -45347,12 +45347,12 @@ _sk_hsl_to_rgb_sse2:
.byte 15,40,124,36,136 // movaps -0x78(%rsp),%xmm7
.byte 15,40,231 // movaps %xmm7,%xmm4
.byte 15,85,227 // andnps %xmm3,%xmm4
- .byte 15,88,53,106,87,1,0 // addps 0x1576a(%rip),%xmm6 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 15,88,53,234,73,1,0 // addps 0x149ea(%rip),%xmm6 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 243,15,91,198 // cvttps2dq %xmm6,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
.byte 15,40,222 // movaps %xmm6,%xmm3
.byte 15,194,216,1 // cmpltps %xmm0,%xmm3
- .byte 15,84,29,85,85,1,0 // andps 0x15555(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,84,29,213,71,1,0 // andps 0x147d5(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,92,195 // subps %xmm3,%xmm0
.byte 15,92,240 // subps %xmm0,%xmm6
.byte 15,89,206 // mulps %xmm6,%xmm1
@@ -45420,9 +45420,9 @@ _sk_scale_u8_sse2:
.byte 102,71,15,110,4,16 // movd (%r8,%r10,1),%xmm8
.byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
.byte 102,68,15,97,192 // punpcklwd %xmm0,%xmm8
- .byte 102,68,15,219,5,40,85,1,0 // pand 0x15528(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,219,5,168,71,1,0 // pand 0x147a8(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,124,86,1,0 // mulps 0x1567c(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,89,5,252,72,1,0 // mulps 0x148fc(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 65,15,89,208 // mulps %xmm8,%xmm2
@@ -45467,17 +45467,17 @@ _sk_scale_565_sse2:
.byte 15,133,173,0,0,0 // jne 24d41 <_sk_scale_565_sse2+0xcc>
.byte 243,71,15,126,20,80 // movq (%r8,%r10,2),%xmm10
.byte 102,68,15,97,208 // punpcklwd %xmm0,%xmm10
- .byte 102,68,15,111,5,232,85,1,0 // movdqa 0x155e8(%rip),%xmm8 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,68,15,111,5,104,72,1,0 // movdqa 0x14868(%rip),%xmm8 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,69,15,219,194 // pand %xmm10,%xmm8
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,231,85,1,0 // mulps 0x155e7(%rip),%xmm8 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,68,15,111,13,238,85,1,0 // movdqa 0x155ee(%rip),%xmm9 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 68,15,89,5,103,72,1,0 // mulps 0x14867(%rip),%xmm8 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,68,15,111,13,110,72,1,0 // movdqa 0x1486e(%rip),%xmm9 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,69,15,219,202 // pand %xmm10,%xmm9
.byte 69,15,91,201 // cvtdq2ps %xmm9,%xmm9
- .byte 68,15,89,13,237,85,1,0 // mulps 0x155ed(%rip),%xmm9 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,68,15,219,21,244,85,1,0 // pand 0x155f4(%rip),%xmm10 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 68,15,89,13,109,72,1,0 // mulps 0x1486d(%rip),%xmm9 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,68,15,219,21,116,72,1,0 // pand 0x14874(%rip),%xmm10 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
- .byte 68,15,89,21,248,85,1,0 // mulps 0x155f8(%rip),%xmm10 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 68,15,89,21,120,72,1,0 // mulps 0x14878(%rip),%xmm10 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 68,15,40,219 // movaps %xmm3,%xmm11
.byte 68,15,194,223,1 // cmpltps %xmm7,%xmm11
.byte 69,15,40,225 // movaps %xmm9,%xmm12
@@ -45558,9 +45558,9 @@ _sk_lerp_u8_sse2:
.byte 102,71,15,110,4,16 // movd (%r8,%r10,1),%xmm8
.byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
.byte 102,68,15,97,192 // punpcklwd %xmm0,%xmm8
- .byte 102,68,15,219,5,30,83,1,0 // pand 0x1531e(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,219,5,158,69,1,0 // pand 0x1459e(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,114,84,1,0 // mulps 0x15472(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,89,5,242,70,1,0 // mulps 0x146f2(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 15,92,196 // subps %xmm4,%xmm0
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 15,88,196 // addps %xmm4,%xmm0
@@ -45612,17 +45612,17 @@ _sk_lerp_565_sse2:
.byte 15,133,173,0,0,0 // jne 24f62 <_sk_lerp_565_sse2+0xcc>
.byte 243,71,15,126,4,80 // movq (%r8,%r10,2),%xmm8
.byte 102,68,15,97,192 // punpcklwd %xmm0,%xmm8
- .byte 102,68,15,111,13,199,83,1,0 // movdqa 0x153c7(%rip),%xmm9 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,68,15,111,13,71,70,1,0 // movdqa 0x14647(%rip),%xmm9 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,69,15,219,200 // pand %xmm8,%xmm9
.byte 69,15,91,201 // cvtdq2ps %xmm9,%xmm9
- .byte 68,15,89,13,198,83,1,0 // mulps 0x153c6(%rip),%xmm9 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,68,15,111,21,205,83,1,0 // movdqa 0x153cd(%rip),%xmm10 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 68,15,89,13,70,70,1,0 // mulps 0x14646(%rip),%xmm9 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,68,15,111,21,77,70,1,0 // movdqa 0x1464d(%rip),%xmm10 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,69,15,219,208 // pand %xmm8,%xmm10
.byte 69,15,91,218 // cvtdq2ps %xmm10,%xmm11
- .byte 68,15,89,29,204,83,1,0 // mulps 0x153cc(%rip),%xmm11 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,68,15,219,5,211,83,1,0 // pand 0x153d3(%rip),%xmm8 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 68,15,89,29,76,70,1,0 // mulps 0x1464c(%rip),%xmm11 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,68,15,219,5,83,70,1,0 // pand 0x14653(%rip),%xmm8 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
- .byte 68,15,89,5,215,83,1,0 // mulps 0x153d7(%rip),%xmm8 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 68,15,89,5,87,70,1,0 // mulps 0x14657(%rip),%xmm8 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 68,15,40,211 // movaps %xmm3,%xmm10
.byte 68,15,194,215,1 // cmpltps %xmm7,%xmm10
.byte 69,15,40,227 // movaps %xmm11,%xmm12
@@ -45679,7 +45679,7 @@ _sk_load_tables_sse2:
.byte 243,69,15,111,12,144 // movdqu (%r8,%rdx,4),%xmm9
.byte 65,86 // push %r14
.byte 83 // push %rbx
- .byte 102,68,15,111,5,70,81,1,0 // movdqa 0x15146(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,111,5,198,67,1,0 // movdqa 0x143c6(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,65,15,219,192 // pand %xmm8,%xmm0
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
@@ -45735,7 +45735,7 @@ _sk_load_tables_sse2:
.byte 102,15,20,211 // unpcklpd %xmm3,%xmm2
.byte 102,65,15,114,209,24 // psrld $0x18,%xmm9
.byte 65,15,91,217 // cvtdq2ps %xmm9,%xmm3
- .byte 15,89,29,157,81,1,0 // mulps 0x1519d(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,29,68,1,0 // mulps 0x1441d(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
@@ -45775,7 +45775,7 @@ _sk_load_tables_u16_be_sse2:
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,97,200 // punpcklwd %xmm0,%xmm1
.byte 102,68,15,105,200 // punpckhwd %xmm0,%xmm9
- .byte 102,15,111,21,119,81,1,0 // movdqa 0x15177(%rip),%xmm2 # 3a2f0 <_sk_srcover_bgra_8888_sse2_lowp+0xe68>
+ .byte 102,15,111,21,247,67,1,0 // movdqa 0x143f7(%rip),%xmm2 # 39570 <_sk_srcover_bgra_8888_sse2_lowp+0xe6c>
.byte 102,15,112,217,238 // pshufd $0xee,%xmm1,%xmm3
.byte 102,15,219,202 // pand %xmm2,%xmm1
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
@@ -45836,7 +45836,7 @@ _sk_load_tables_u16_be_sse2:
.byte 102,65,15,235,217 // por %xmm9,%xmm3
.byte 102,65,15,97,216 // punpcklwd %xmm8,%xmm3
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,117,80,1,0 // mulps 0x15075(%rip),%xmm3 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 15,89,29,245,66,1,0 // mulps 0x142f5(%rip),%xmm3 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
@@ -45877,7 +45877,7 @@ _sk_load_tables_rgb_u16_be_sse2:
.byte 102,15,111,194 // movdqa %xmm2,%xmm0
.byte 102,65,15,97,194 // punpcklwd %xmm10,%xmm0
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
- .byte 102,68,15,111,5,185,79,1,0 // movdqa 0x14fb9(%rip),%xmm8 # 3a2f0 <_sk_srcover_bgra_8888_sse2_lowp+0xe68>
+ .byte 102,68,15,111,5,57,66,1,0 // movdqa 0x14239(%rip),%xmm8 # 39570 <_sk_srcover_bgra_8888_sse2_lowp+0xe6c>
.byte 102,65,15,219,192 // pand %xmm8,%xmm0
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,65,15,97,193 // punpcklwd %xmm9,%xmm0
@@ -45932,7 +45932,7 @@ _sk_load_tables_rgb_u16_be_sse2:
.byte 65,15,20,216 // unpcklps %xmm8,%xmm3
.byte 102,15,20,211 // unpcklpd %xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,54,76,1,0 // movaps 0x14c36(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,182,62,1,0 // movaps 0x13eb6(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 91 // pop %rbx
.byte 65,94 // pop %r14
.byte 255,224 // jmpq *%rax
@@ -45964,7 +45964,7 @@ _sk_byte_tables_sse2:
.byte 65,86 // push %r14
.byte 83 // push %rbx
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,163,76,1,0 // movaps 0x14ca3(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,35,63,1,0 // movaps 0x13f23(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,91,192 // cvtps2dq %xmm0,%xmm0
.byte 102,73,15,126,192 // movq %xmm0,%r8
@@ -45992,7 +45992,7 @@ _sk_byte_tables_sse2:
.byte 102,65,15,96,193 // punpcklbw %xmm9,%xmm0
.byte 102,65,15,97,193 // punpcklwd %xmm9,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,21,94,77,1,0 // movaps 0x14d5e(%rip),%xmm10 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,21,222,63,1,0 // movaps 0x13fde(%rip),%xmm10 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,194 // mulps %xmm10,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,91,201 // cvtps2dq %xmm1,%xmm1
@@ -46111,7 +46111,7 @@ _sk_byte_tables_rgb_sse2:
.byte 102,65,15,96,193 // punpcklbw %xmm9,%xmm0
.byte 102,65,15,97,193 // punpcklwd %xmm9,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,21,158,75,1,0 // movaps 0x14b9e(%rip),%xmm10 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,21,30,62,1,0 // movaps 0x13e1e(%rip),%xmm10 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,194 // mulps %xmm10,%xmm0
.byte 65,15,89,200 // mulps %xmm8,%xmm1
.byte 102,15,91,201 // cvtps2dq %xmm1,%xmm1
@@ -46309,15 +46309,15 @@ _sk_parametric_r_sse2:
.byte 69,15,88,209 // addps %xmm9,%xmm10
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,91,202 // cvtdq2ps %xmm10,%xmm9
- .byte 68,15,89,13,106,73,1,0 // mulps 0x1496a(%rip),%xmm9 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,21,114,73,1,0 // andps 0x14972(%rip),%xmm10 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,21,154,70,1,0 // orps 0x1469a(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,13,114,73,1,0 // addps 0x14972(%rip),%xmm9 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 68,15,40,37,122,73,1,0 // movaps 0x1497a(%rip),%xmm12 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,13,234,59,1,0 // mulps 0x13bea(%rip),%xmm9 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,21,242,59,1,0 // andps 0x13bf2(%rip),%xmm10 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,21,26,57,1,0 // orps 0x1391a(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,13,242,59,1,0 // addps 0x13bf2(%rip),%xmm9 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 68,15,40,37,250,59,1,0 // movaps 0x13bfa(%rip),%xmm12 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,88,21,122,73,1,0 // addps 0x1497a(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 68,15,40,37,130,73,1,0 // movaps 0x14982(%rip),%xmm12 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,21,250,59,1,0 // addps 0x13bfa(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 68,15,40,37,2,60,1,0 // movaps 0x13c02(%rip),%xmm12 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
@@ -46325,22 +46325,22 @@ _sk_parametric_r_sse2:
.byte 69,15,91,226 // cvtdq2ps %xmm10,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,194,236,1 // cmpltps %xmm12,%xmm13
- .byte 68,15,40,21,92,70,1,0 // movaps 0x1465c(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,220,56,1,0 // movaps 0x138dc(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,234 // andps %xmm10,%xmm13
.byte 69,15,87,219 // xorps %xmm11,%xmm11
.byte 69,15,92,229 // subps %xmm13,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,92,236 // subps %xmm12,%xmm13
- .byte 68,15,88,13,80,73,1,0 // addps 0x14950(%rip),%xmm9 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 68,15,40,37,88,73,1,0 // movaps 0x14958(%rip),%xmm12 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,13,208,59,1,0 // addps 0x13bd0(%rip),%xmm9 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 68,15,40,37,216,59,1,0 // movaps 0x13bd8(%rip),%xmm12 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 69,15,89,229 // mulps %xmm13,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,40,37,88,73,1,0 // movaps 0x14958(%rip),%xmm12 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,37,216,59,1,0 // movaps 0x13bd8(%rip),%xmm12 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,229 // subps %xmm13,%xmm12
- .byte 68,15,40,45,92,73,1,0 // movaps 0x1495c(%rip),%xmm13 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,45,220,59,1,0 // movaps 0x13bdc(%rip),%xmm13 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 69,15,94,236 // divps %xmm12,%xmm13
.byte 69,15,88,233 // addps %xmm9,%xmm13
- .byte 68,15,89,45,92,73,1,0 // mulps 0x1495c(%rip),%xmm13 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 68,15,89,45,220,59,1,0 // mulps 0x13bdc(%rip),%xmm13 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,69,15,91,205 // cvtps2dq %xmm13,%xmm9
.byte 243,68,15,16,96,20 // movss 0x14(%rax),%xmm12
.byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
@@ -46376,15 +46376,15 @@ _sk_parametric_g_sse2:
.byte 69,15,88,209 // addps %xmm9,%xmm10
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,91,202 // cvtdq2ps %xmm10,%xmm9
- .byte 68,15,89,13,44,72,1,0 // mulps 0x1482c(%rip),%xmm9 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,21,52,72,1,0 // andps 0x14834(%rip),%xmm10 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,21,92,69,1,0 // orps 0x1455c(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,13,52,72,1,0 // addps 0x14834(%rip),%xmm9 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 68,15,40,37,60,72,1,0 // movaps 0x1483c(%rip),%xmm12 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,13,172,58,1,0 // mulps 0x13aac(%rip),%xmm9 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,21,180,58,1,0 // andps 0x13ab4(%rip),%xmm10 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,21,220,55,1,0 // orps 0x137dc(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,13,180,58,1,0 // addps 0x13ab4(%rip),%xmm9 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 68,15,40,37,188,58,1,0 // movaps 0x13abc(%rip),%xmm12 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,88,21,60,72,1,0 // addps 0x1483c(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 68,15,40,37,68,72,1,0 // movaps 0x14844(%rip),%xmm12 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,21,188,58,1,0 // addps 0x13abc(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 68,15,40,37,196,58,1,0 // movaps 0x13ac4(%rip),%xmm12 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
@@ -46392,22 +46392,22 @@ _sk_parametric_g_sse2:
.byte 69,15,91,226 // cvtdq2ps %xmm10,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,194,236,1 // cmpltps %xmm12,%xmm13
- .byte 68,15,40,21,30,69,1,0 // movaps 0x1451e(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,158,55,1,0 // movaps 0x1379e(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,234 // andps %xmm10,%xmm13
.byte 69,15,87,219 // xorps %xmm11,%xmm11
.byte 69,15,92,229 // subps %xmm13,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,92,236 // subps %xmm12,%xmm13
- .byte 68,15,88,13,18,72,1,0 // addps 0x14812(%rip),%xmm9 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 68,15,40,37,26,72,1,0 // movaps 0x1481a(%rip),%xmm12 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,13,146,58,1,0 // addps 0x13a92(%rip),%xmm9 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 68,15,40,37,154,58,1,0 // movaps 0x13a9a(%rip),%xmm12 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 69,15,89,229 // mulps %xmm13,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,40,37,26,72,1,0 // movaps 0x1481a(%rip),%xmm12 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,37,154,58,1,0 // movaps 0x13a9a(%rip),%xmm12 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,229 // subps %xmm13,%xmm12
- .byte 68,15,40,45,30,72,1,0 // movaps 0x1481e(%rip),%xmm13 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,45,158,58,1,0 // movaps 0x13a9e(%rip),%xmm13 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 69,15,94,236 // divps %xmm12,%xmm13
.byte 69,15,88,233 // addps %xmm9,%xmm13
- .byte 68,15,89,45,30,72,1,0 // mulps 0x1481e(%rip),%xmm13 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 68,15,89,45,158,58,1,0 // mulps 0x13a9e(%rip),%xmm13 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,69,15,91,205 // cvtps2dq %xmm13,%xmm9
.byte 243,68,15,16,96,20 // movss 0x14(%rax),%xmm12
.byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
@@ -46443,15 +46443,15 @@ _sk_parametric_b_sse2:
.byte 69,15,88,209 // addps %xmm9,%xmm10
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,91,202 // cvtdq2ps %xmm10,%xmm9
- .byte 68,15,89,13,238,70,1,0 // mulps 0x146ee(%rip),%xmm9 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,21,246,70,1,0 // andps 0x146f6(%rip),%xmm10 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,21,30,68,1,0 // orps 0x1441e(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,13,246,70,1,0 // addps 0x146f6(%rip),%xmm9 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 68,15,40,37,254,70,1,0 // movaps 0x146fe(%rip),%xmm12 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,13,110,57,1,0 // mulps 0x1396e(%rip),%xmm9 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,21,118,57,1,0 // andps 0x13976(%rip),%xmm10 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,21,158,54,1,0 // orps 0x1369e(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,13,118,57,1,0 // addps 0x13976(%rip),%xmm9 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 68,15,40,37,126,57,1,0 // movaps 0x1397e(%rip),%xmm12 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,88,21,254,70,1,0 // addps 0x146fe(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 68,15,40,37,6,71,1,0 // movaps 0x14706(%rip),%xmm12 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,21,126,57,1,0 // addps 0x1397e(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 68,15,40,37,134,57,1,0 // movaps 0x13986(%rip),%xmm12 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
@@ -46459,22 +46459,22 @@ _sk_parametric_b_sse2:
.byte 69,15,91,226 // cvtdq2ps %xmm10,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,194,236,1 // cmpltps %xmm12,%xmm13
- .byte 68,15,40,21,224,67,1,0 // movaps 0x143e0(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,96,54,1,0 // movaps 0x13660(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,234 // andps %xmm10,%xmm13
.byte 69,15,87,219 // xorps %xmm11,%xmm11
.byte 69,15,92,229 // subps %xmm13,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,92,236 // subps %xmm12,%xmm13
- .byte 68,15,88,13,212,70,1,0 // addps 0x146d4(%rip),%xmm9 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 68,15,40,37,220,70,1,0 // movaps 0x146dc(%rip),%xmm12 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,13,84,57,1,0 // addps 0x13954(%rip),%xmm9 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 68,15,40,37,92,57,1,0 // movaps 0x1395c(%rip),%xmm12 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 69,15,89,229 // mulps %xmm13,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,40,37,220,70,1,0 // movaps 0x146dc(%rip),%xmm12 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,37,92,57,1,0 // movaps 0x1395c(%rip),%xmm12 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,229 // subps %xmm13,%xmm12
- .byte 68,15,40,45,224,70,1,0 // movaps 0x146e0(%rip),%xmm13 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,45,96,57,1,0 // movaps 0x13960(%rip),%xmm13 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 69,15,94,236 // divps %xmm12,%xmm13
.byte 69,15,88,233 // addps %xmm9,%xmm13
- .byte 68,15,89,45,224,70,1,0 // mulps 0x146e0(%rip),%xmm13 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 68,15,89,45,96,57,1,0 // mulps 0x13960(%rip),%xmm13 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,69,15,91,205 // cvtps2dq %xmm13,%xmm9
.byte 243,68,15,16,96,20 // movss 0x14(%rax),%xmm12
.byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
@@ -46510,15 +46510,15 @@ _sk_parametric_a_sse2:
.byte 69,15,88,209 // addps %xmm9,%xmm10
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,91,202 // cvtdq2ps %xmm10,%xmm9
- .byte 68,15,89,13,176,69,1,0 // mulps 0x145b0(%rip),%xmm9 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 68,15,84,21,184,69,1,0 // andps 0x145b8(%rip),%xmm10 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,86,21,224,66,1,0 // orps 0x142e0(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
- .byte 68,15,88,13,184,69,1,0 // addps 0x145b8(%rip),%xmm9 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
- .byte 68,15,40,37,192,69,1,0 // movaps 0x145c0(%rip),%xmm12 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,89,13,48,56,1,0 // mulps 0x13830(%rip),%xmm9 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 68,15,84,21,56,56,1,0 // andps 0x13838(%rip),%xmm10 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,86,21,96,53,1,0 // orps 0x13560(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
+ .byte 68,15,88,13,56,56,1,0 // addps 0x13838(%rip),%xmm9 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
+ .byte 68,15,40,37,64,56,1,0 // movaps 0x13840(%rip),%xmm12 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 69,15,89,226 // mulps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,88,21,192,69,1,0 // addps 0x145c0(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
- .byte 68,15,40,37,200,69,1,0 // movaps 0x145c8(%rip),%xmm12 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,88,21,64,56,1,0 // addps 0x13840(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
+ .byte 68,15,40,37,72,56,1,0 // movaps 0x13848(%rip),%xmm12 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 69,15,94,226 // divps %xmm10,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
.byte 69,15,89,203 // mulps %xmm11,%xmm9
@@ -46526,22 +46526,22 @@ _sk_parametric_a_sse2:
.byte 69,15,91,226 // cvtdq2ps %xmm10,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,194,236,1 // cmpltps %xmm12,%xmm13
- .byte 68,15,40,21,162,66,1,0 // movaps 0x142a2(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,34,53,1,0 // movaps 0x13522(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,234 // andps %xmm10,%xmm13
.byte 69,15,87,219 // xorps %xmm11,%xmm11
.byte 69,15,92,229 // subps %xmm13,%xmm12
.byte 69,15,40,233 // movaps %xmm9,%xmm13
.byte 69,15,92,236 // subps %xmm12,%xmm13
- .byte 68,15,88,13,150,69,1,0 // addps 0x14596(%rip),%xmm9 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
- .byte 68,15,40,37,158,69,1,0 // movaps 0x1459e(%rip),%xmm12 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 68,15,88,13,22,56,1,0 // addps 0x13816(%rip),%xmm9 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
+ .byte 68,15,40,37,30,56,1,0 // movaps 0x1381e(%rip),%xmm12 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 69,15,89,229 // mulps %xmm13,%xmm12
.byte 69,15,92,204 // subps %xmm12,%xmm9
- .byte 68,15,40,37,158,69,1,0 // movaps 0x1459e(%rip),%xmm12 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 68,15,40,37,30,56,1,0 // movaps 0x1381e(%rip),%xmm12 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 69,15,92,229 // subps %xmm13,%xmm12
- .byte 68,15,40,45,162,69,1,0 // movaps 0x145a2(%rip),%xmm13 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,45,34,56,1,0 // movaps 0x13822(%rip),%xmm13 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 69,15,94,236 // divps %xmm12,%xmm13
.byte 69,15,88,233 // addps %xmm9,%xmm13
- .byte 68,15,89,45,162,69,1,0 // mulps 0x145a2(%rip),%xmm13 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 68,15,89,45,34,56,1,0 // mulps 0x13822(%rip),%xmm13 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 102,69,15,91,205 // cvtps2dq %xmm13,%xmm9
.byte 243,68,15,16,96,20 // movss 0x14(%rax),%xmm12
.byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
@@ -46566,19 +46566,19 @@ _sk_gamma_sse2:
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,91,194 // cvtdq2ps %xmm2,%xmm0
- .byte 15,89,5,173,68,1,0 // mulps 0x144ad(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 15,84,21,182,68,1,0 // andps 0x144b6(%rip),%xmm2 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 68,15,40,53,222,65,1,0 // movaps 0x141de(%rip),%xmm14 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,89,5,45,55,1,0 // mulps 0x1372d(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 15,84,21,54,55,1,0 // andps 0x13736(%rip),%xmm2 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 68,15,40,53,94,52,1,0 // movaps 0x1345e(%rip),%xmm14 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,86,214 // orps %xmm14,%xmm2
- .byte 68,15,40,37,178,68,1,0 // movaps 0x144b2(%rip),%xmm12 # 3a340 <_sk_srcover_bgra_8888_sse2_lowp+0xeb8>
+ .byte 68,15,40,37,50,55,1,0 // movaps 0x13732(%rip),%xmm12 # 395c0 <_sk_srcover_bgra_8888_sse2_lowp+0xebc>
.byte 65,15,88,196 // addps %xmm12,%xmm0
- .byte 68,15,40,29,182,68,1,0 // movaps 0x144b6(%rip),%xmm11 # 3a350 <_sk_srcover_bgra_8888_sse2_lowp+0xec8>
+ .byte 68,15,40,29,54,55,1,0 // movaps 0x13736(%rip),%xmm11 # 395d0 <_sk_srcover_bgra_8888_sse2_lowp+0xecc>
.byte 15,40,226 // movaps %xmm2,%xmm4
.byte 65,15,89,227 // mulps %xmm11,%xmm4
.byte 15,92,196 // subps %xmm4,%xmm0
- .byte 68,15,40,21,180,68,1,0 // movaps 0x144b4(%rip),%xmm10 # 3a360 <_sk_srcover_bgra_8888_sse2_lowp+0xed8>
+ .byte 68,15,40,21,52,55,1,0 // movaps 0x13734(%rip),%xmm10 # 395e0 <_sk_srcover_bgra_8888_sse2_lowp+0xedc>
.byte 65,15,88,210 // addps %xmm10,%xmm2
- .byte 68,15,40,45,184,68,1,0 // movaps 0x144b8(%rip),%xmm13 # 3a370 <_sk_srcover_bgra_8888_sse2_lowp+0xee8>
+ .byte 68,15,40,45,56,55,1,0 // movaps 0x13738(%rip),%xmm13 # 395f0 <_sk_srcover_bgra_8888_sse2_lowp+0xeec>
.byte 65,15,40,229 // movaps %xmm13,%xmm4
.byte 15,94,226 // divps %xmm2,%xmm4
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -46590,27 +46590,27 @@ _sk_gamma_sse2:
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
.byte 15,40,224 // movaps %xmm0,%xmm4
.byte 15,194,226,1 // cmpltps %xmm2,%xmm4
- .byte 68,15,40,13,136,65,1,0 // movaps 0x14188(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,8,52,1,0 // movaps 0x13408(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,84,225 // andps %xmm9,%xmm4
.byte 15,92,212 // subps %xmm4,%xmm2
.byte 15,40,224 // movaps %xmm0,%xmm4
.byte 15,92,226 // subps %xmm2,%xmm4
- .byte 15,40,53,132,68,1,0 // movaps 0x14484(%rip),%xmm6 # 3a380 <_sk_srcover_bgra_8888_sse2_lowp+0xef8>
+ .byte 15,40,53,4,55,1,0 // movaps 0x13704(%rip),%xmm6 # 39600 <_sk_srcover_bgra_8888_sse2_lowp+0xefc>
.byte 15,88,198 // addps %xmm6,%xmm0
- .byte 15,40,61,154,68,1,0 // movaps 0x1449a(%rip),%xmm7 # 3a3a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf18>
+ .byte 15,40,61,26,55,1,0 // movaps 0x1371a(%rip),%xmm7 # 39620 <_sk_srcover_bgra_8888_sse2_lowp+0xf1c>
.byte 15,40,239 // movaps %xmm7,%xmm5
.byte 15,92,236 // subps %xmm4,%xmm5
.byte 15,40,212 // movaps %xmm4,%xmm2
- .byte 15,40,37,122,68,1,0 // movaps 0x1447a(%rip),%xmm4 # 3a390 <_sk_srcover_bgra_8888_sse2_lowp+0xf08>
+ .byte 15,40,37,250,54,1,0 // movaps 0x136fa(%rip),%xmm4 # 39610 <_sk_srcover_bgra_8888_sse2_lowp+0xf0c>
.byte 15,89,212 // mulps %xmm4,%xmm2
.byte 15,92,194 // subps %xmm2,%xmm0
- .byte 68,15,40,61,140,68,1,0 // movaps 0x1448c(%rip),%xmm15 # 3a3b0 <_sk_srcover_bgra_8888_sse2_lowp+0xf28>
+ .byte 68,15,40,61,12,55,1,0 // movaps 0x1370c(%rip),%xmm15 # 39630 <_sk_srcover_bgra_8888_sse2_lowp+0xf2c>
.byte 65,15,40,215 // movaps %xmm15,%xmm2
.byte 15,94,213 // divps %xmm5,%xmm2
.byte 15,88,208 // addps %xmm0,%xmm2
.byte 15,91,193 // cvtdq2ps %xmm1,%xmm0
- .byte 15,89,5,232,67,1,0 // mulps 0x143e8(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 15,84,13,241,67,1,0 // andps 0x143f1(%rip),%xmm1 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
+ .byte 15,89,5,104,54,1,0 // mulps 0x13668(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 15,84,13,113,54,1,0 // andps 0x13671(%rip),%xmm1 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
.byte 65,15,86,206 // orps %xmm14,%xmm1
.byte 65,15,88,196 // addps %xmm12,%xmm0
.byte 15,40,233 // movaps %xmm1,%xmm5
@@ -46640,9 +46640,9 @@ _sk_gamma_sse2:
.byte 15,94,206 // divps %xmm6,%xmm1
.byte 15,88,200 // addps %xmm0,%xmm1
.byte 15,91,195 // cvtdq2ps %xmm3,%xmm0
- .byte 15,89,5,119,67,1,0 // mulps 0x14377(%rip),%xmm0 # 3a320 <_sk_srcover_bgra_8888_sse2_lowp+0xe98>
- .byte 15,84,29,128,67,1,0 // andps 0x14380(%rip),%xmm3 # 3a330 <_sk_srcover_bgra_8888_sse2_lowp+0xea8>
- .byte 15,86,29,169,64,1,0 // orps 0x140a9(%rip),%xmm3 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,89,5,247,53,1,0 // mulps 0x135f7(%rip),%xmm0 # 395a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe9c>
+ .byte 15,84,29,0,54,1,0 // andps 0x13600(%rip),%xmm3 # 395b0 <_sk_srcover_bgra_8888_sse2_lowp+0xeac>
+ .byte 15,86,29,41,51,1,0 // orps 0x13329(%rip),%xmm3 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,88,196 // addps %xmm12,%xmm0
.byte 68,15,89,219 // mulps %xmm3,%xmm11
.byte 65,15,92,195 // subps %xmm11,%xmm0
@@ -46664,7 +46664,7 @@ _sk_gamma_sse2:
.byte 15,92,253 // subps %xmm5,%xmm7
.byte 68,15,94,255 // divps %xmm7,%xmm15
.byte 68,15,88,248 // addps %xmm0,%xmm15
- .byte 15,40,5,182,67,1,0 // movaps 0x143b6(%rip),%xmm0 # 3a3c0 <_sk_srcover_bgra_8888_sse2_lowp+0xf38>
+ .byte 15,40,5,54,54,1,0 // movaps 0x13636(%rip),%xmm0 # 39640 <_sk_srcover_bgra_8888_sse2_lowp+0xf3c>
.byte 15,89,208 // mulps %xmm0,%xmm2
.byte 15,89,200 // mulps %xmm0,%xmm1
.byte 68,15,89,248 // mulps %xmm0,%xmm15
@@ -46683,29 +46683,29 @@ HIDDEN _sk_lab_to_xyz_sse2
.globl _sk_lab_to_xyz_sse2
FUNCTION(_sk_lab_to_xyz_sse2)
_sk_lab_to_xyz_sse2:
- .byte 15,89,5,139,67,1,0 // mulps 0x1438b(%rip),%xmm0 # 3a3d0 <_sk_srcover_bgra_8888_sse2_lowp+0xf48>
- .byte 68,15,40,5,3,65,1,0 // movaps 0x14103(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,89,5,11,54,1,0 // mulps 0x1360b(%rip),%xmm0 # 39650 <_sk_srcover_bgra_8888_sse2_lowp+0xf4c>
+ .byte 68,15,40,5,131,51,1,0 // movaps 0x13383(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,200 // mulps %xmm8,%xmm1
- .byte 68,15,40,13,135,67,1,0 // movaps 0x14387(%rip),%xmm9 # 3a3e0 <_sk_srcover_bgra_8888_sse2_lowp+0xf58>
+ .byte 68,15,40,13,7,54,1,0 // movaps 0x13607(%rip),%xmm9 # 39660 <_sk_srcover_bgra_8888_sse2_lowp+0xf5c>
.byte 65,15,88,201 // addps %xmm9,%xmm1
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 65,15,88,209 // addps %xmm9,%xmm2
- .byte 15,88,5,132,67,1,0 // addps 0x14384(%rip),%xmm0 # 3a3f0 <_sk_srcover_bgra_8888_sse2_lowp+0xf68>
- .byte 15,89,5,141,67,1,0 // mulps 0x1438d(%rip),%xmm0 # 3a400 <_sk_srcover_bgra_8888_sse2_lowp+0xf78>
- .byte 15,89,13,150,67,1,0 // mulps 0x14396(%rip),%xmm1 # 3a410 <_sk_srcover_bgra_8888_sse2_lowp+0xf88>
+ .byte 15,88,5,4,54,1,0 // addps 0x13604(%rip),%xmm0 # 39670 <_sk_srcover_bgra_8888_sse2_lowp+0xf6c>
+ .byte 15,89,5,13,54,1,0 // mulps 0x1360d(%rip),%xmm0 # 39680 <_sk_srcover_bgra_8888_sse2_lowp+0xf7c>
+ .byte 15,89,13,22,54,1,0 // mulps 0x13616(%rip),%xmm1 # 39690 <_sk_srcover_bgra_8888_sse2_lowp+0xf8c>
.byte 15,88,200 // addps %xmm0,%xmm1
- .byte 15,89,21,156,67,1,0 // mulps 0x1439c(%rip),%xmm2 # 3a420 <_sk_srcover_bgra_8888_sse2_lowp+0xf98>
+ .byte 15,89,21,28,54,1,0 // mulps 0x1361c(%rip),%xmm2 # 396a0 <_sk_srcover_bgra_8888_sse2_lowp+0xf9c>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 68,15,92,202 // subps %xmm2,%xmm9
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
.byte 68,15,89,225 // mulps %xmm1,%xmm12
- .byte 15,40,21,145,67,1,0 // movaps 0x14391(%rip),%xmm2 # 3a430 <_sk_srcover_bgra_8888_sse2_lowp+0xfa8>
+ .byte 15,40,21,17,54,1,0 // movaps 0x13611(%rip),%xmm2 # 396b0 <_sk_srcover_bgra_8888_sse2_lowp+0xfac>
.byte 68,15,40,194 // movaps %xmm2,%xmm8
.byte 69,15,194,196,1 // cmpltps %xmm12,%xmm8
- .byte 68,15,40,21,144,67,1,0 // movaps 0x14390(%rip),%xmm10 # 3a440 <_sk_srcover_bgra_8888_sse2_lowp+0xfb8>
+ .byte 68,15,40,21,16,54,1,0 // movaps 0x13610(%rip),%xmm10 # 396c0 <_sk_srcover_bgra_8888_sse2_lowp+0xfbc>
.byte 65,15,88,202 // addps %xmm10,%xmm1
- .byte 68,15,40,29,148,67,1,0 // movaps 0x14394(%rip),%xmm11 # 3a450 <_sk_srcover_bgra_8888_sse2_lowp+0xfc8>
+ .byte 68,15,40,29,20,54,1,0 // movaps 0x13614(%rip),%xmm11 # 396d0 <_sk_srcover_bgra_8888_sse2_lowp+0xfcc>
.byte 65,15,89,203 // mulps %xmm11,%xmm1
.byte 69,15,84,224 // andps %xmm8,%xmm12
.byte 68,15,85,193 // andnps %xmm1,%xmm8
@@ -46729,8 +46729,8 @@ _sk_lab_to_xyz_sse2:
.byte 15,84,194 // andps %xmm2,%xmm0
.byte 65,15,85,209 // andnps %xmm9,%xmm2
.byte 15,86,208 // orps %xmm0,%xmm2
- .byte 68,15,89,5,68,67,1,0 // mulps 0x14344(%rip),%xmm8 # 3a460 <_sk_srcover_bgra_8888_sse2_lowp+0xfd8>
- .byte 15,89,21,77,67,1,0 // mulps 0x1434d(%rip),%xmm2 # 3a470 <_sk_srcover_bgra_8888_sse2_lowp+0xfe8>
+ .byte 68,15,89,5,196,53,1,0 // mulps 0x135c4(%rip),%xmm8 # 396e0 <_sk_srcover_bgra_8888_sse2_lowp+0xfdc>
+ .byte 15,89,21,205,53,1,0 // mulps 0x135cd(%rip),%xmm2 # 396f0 <_sk_srcover_bgra_8888_sse2_lowp+0xfec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -46750,9 +46750,9 @@ _sk_load_a8_sse2:
.byte 102,67,15,110,4,16 // movd (%r8,%r10,1),%xmm0
.byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
.byte 102,15,97,192 // punpcklwd %xmm0,%xmm0
- .byte 102,15,219,5,199,63,1,0 // pand 0x13fc7(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,5,71,50,1,0 // pand 0x13247(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,216 // cvtdq2ps %xmm0,%xmm3
- .byte 15,89,29,29,65,1,0 // mulps 0x1411d(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,157,51,1,0 // mulps 0x1339d(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 102,15,87,201 // xorpd %xmm1,%xmm1
@@ -46795,9 +46795,9 @@ _sk_load_a8_dst_sse2:
.byte 102,67,15,110,36,16 // movd (%r8,%r10,1),%xmm4
.byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
.byte 102,15,97,224 // punpcklwd %xmm0,%xmm4
- .byte 102,15,219,37,51,63,1,0 // pand 0x13f33(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,37,179,49,1,0 // pand 0x131b3(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,252 // cvtdq2ps %xmm4,%xmm7
- .byte 15,89,61,137,64,1,0 // mulps 0x14089(%rip),%xmm7 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,61,9,51,1,0 // mulps 0x13309(%rip),%xmm7 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,228 // xorps %xmm4,%xmm4
.byte 102,15,87,237 // xorpd %xmm5,%xmm5
@@ -46876,7 +46876,7 @@ _sk_gather_a8_sse2:
.byte 102,65,15,96,192 // punpcklbw %xmm8,%xmm0
.byte 102,65,15,97,192 // punpcklwd %xmm8,%xmm0
.byte 15,91,216 // cvtdq2ps %xmm0,%xmm3
- .byte 15,89,29,99,63,1,0 // mulps 0x13f63(%rip),%xmm3 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,29,227,49,1,0 // mulps 0x131e3(%rip),%xmm3 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 102,15,239,201 // pxor %xmm1,%xmm1
@@ -46893,7 +46893,7 @@ _sk_store_a8_sse2:
.byte 77,15,175,193 // imul %r9,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,210 // movslq %edx,%r10
- .byte 68,15,40,5,10,62,1,0 // movaps 0x13e0a(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,138,48,1,0 // movaps 0x1308a(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,89,195 // mulps %xmm3,%xmm8
.byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
.byte 102,65,15,114,240,16 // pslld $0x10,%xmm8
@@ -46918,7 +46918,7 @@ _sk_store_a8_sse2:
.byte 117,217 // jne 26373 <_sk_store_a8_sse2+0x48>
.byte 102,65,15,197,192,4 // pextrw $0x4,%xmm8,%eax
.byte 67,136,68,16,2 // mov %al,0x2(%r8,%r10,1)
- .byte 102,68,15,219,5,114,61,1,0 // pand 0x13d72(%rip),%xmm8 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,68,15,219,5,242,47,1,0 // pand 0x12ff2(%rip),%xmm8 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
.byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
.byte 102,68,15,126,192 // movd %xmm8,%eax
@@ -46943,11 +46943,11 @@ _sk_load_g8_sse2:
.byte 102,67,15,110,4,16 // movd (%r8,%r10,1),%xmm0
.byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
.byte 102,15,97,192 // punpcklwd %xmm0,%xmm0
- .byte 102,15,219,5,35,61,1,0 // pand 0x13d23(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,5,163,47,1,0 // pand 0x12fa3(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,121,62,1,0 // mulps 0x13e79(%rip),%xmm0 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,5,249,48,1,0 // mulps 0x130f9(%rip),%xmm0 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,96,60,1,0 // movaps 0x13c60(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,224,46,1,0 // movaps 0x12ee0(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
@@ -46988,11 +46988,11 @@ _sk_load_g8_dst_sse2:
.byte 102,67,15,110,36,16 // movd (%r8,%r10,1),%xmm4
.byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
.byte 102,15,97,224 // punpcklwd %xmm0,%xmm4
- .byte 102,15,219,37,140,60,1,0 // pand 0x13c8c(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,219,37,12,47,1,0 // pand 0x12f0c(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,226,61,1,0 // mulps 0x13de2(%rip),%xmm4 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,37,98,48,1,0 // mulps 0x13062(%rip),%xmm4 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,201,59,1,0 // movaps 0x13bc9(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,61,73,46,1,0 // movaps 0x12e49(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,236 // movaps %xmm4,%xmm5
.byte 15,40,244 // movaps %xmm4,%xmm6
.byte 255,224 // jmpq *%rax
@@ -47069,9 +47069,9 @@ _sk_gather_g8_sse2:
.byte 102,65,15,96,192 // punpcklbw %xmm8,%xmm0
.byte 102,65,15,97,192 // punpcklwd %xmm8,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,185,60,1,0 // mulps 0x13cb9(%rip),%xmm0 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 15,89,5,57,47,1,0 // mulps 0x12f39(%rip),%xmm0 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,160,58,1,0 // movaps 0x13aa0(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,32,45,1,0 // movaps 0x12d20(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
@@ -47091,19 +47091,19 @@ _sk_load_565_sse2:
.byte 117,83 // jne 26646 <_sk_load_565_sse2+0x6e>
.byte 243,67,15,126,20,80 // movq (%r8,%r10,2),%xmm2
.byte 102,15,97,208 // punpcklwd %xmm0,%xmm2
- .byte 102,15,111,5,139,60,1,0 // movdqa 0x13c8b(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,11,47,1,0 // movdqa 0x12f0b(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,141,60,1,0 // mulps 0x13c8d(%rip),%xmm0 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,13,149,60,1,0 // movdqa 0x13c95(%rip),%xmm1 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,5,13,47,1,0 // mulps 0x12f0d(%rip),%xmm0 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,13,21,47,1,0 // movdqa 0x12f15(%rip),%xmm1 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,202 // pand %xmm2,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,151,60,1,0 // mulps 0x13c97(%rip),%xmm1 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,21,159,60,1,0 // pand 0x13c9f(%rip),%xmm2 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,13,23,47,1,0 // mulps 0x12f17(%rip),%xmm1 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,21,31,47,1,0 // pand 0x12f1f(%rip),%xmm2 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,165,60,1,0 // mulps 0x13ca5(%rip),%xmm2 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,21,37,47,1,0 // mulps 0x12f25(%rip),%xmm2 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,44,58,1,0 // movaps 0x13a2c(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,172,44,1,0 // movaps 0x12cac(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,3 // and $0x3,%r9b
@@ -47140,19 +47140,19 @@ _sk_load_565_dst_sse2:
.byte 117,83 // jne 26702 <_sk_load_565_dst_sse2+0x6e>
.byte 243,67,15,126,52,80 // movq (%r8,%r10,2),%xmm6
.byte 102,15,97,240 // punpcklwd %xmm0,%xmm6
- .byte 102,15,111,37,207,59,1,0 // movdqa 0x13bcf(%rip),%xmm4 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,37,79,46,1,0 // movdqa 0x12e4f(%rip),%xmm4 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,230 // pand %xmm6,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,209,59,1,0 // mulps 0x13bd1(%rip),%xmm4 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,45,217,59,1,0 // movdqa 0x13bd9(%rip),%xmm5 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,37,81,46,1,0 // mulps 0x12e51(%rip),%xmm4 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,45,89,46,1,0 // movdqa 0x12e59(%rip),%xmm5 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,238 // pand %xmm6,%xmm5
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
- .byte 15,89,45,219,59,1,0 // mulps 0x13bdb(%rip),%xmm5 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,53,227,59,1,0 // pand 0x13be3(%rip),%xmm6 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,45,91,46,1,0 // mulps 0x12e5b(%rip),%xmm5 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,53,99,46,1,0 // pand 0x12e63(%rip),%xmm6 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
- .byte 15,89,53,233,59,1,0 // mulps 0x13be9(%rip),%xmm6 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,53,105,46,1,0 // mulps 0x12e69(%rip),%xmm6 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,112,57,1,0 // movaps 0x13970(%rip),%xmm7 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,61,240,43,1,0 // movaps 0x12bf0(%rip),%xmm7 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
.byte 65,128,225,3 // and $0x3,%r9b
@@ -47221,19 +47221,19 @@ _sk_gather_565_sse2:
.byte 65,15,183,4,64 // movzwl (%r8,%rax,2),%eax
.byte 102,15,196,208,3 // pinsrw $0x3,%eax,%xmm2
.byte 102,65,15,97,208 // punpcklwd %xmm8,%xmm2
- .byte 102,15,111,5,126,58,1,0 // movdqa 0x13a7e(%rip),%xmm0 # 3a290 <_sk_srcover_bgra_8888_sse2_lowp+0xe08>
+ .byte 102,15,111,5,254,44,1,0 // movdqa 0x12cfe(%rip),%xmm0 # 39510 <_sk_srcover_bgra_8888_sse2_lowp+0xe0c>
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,128,58,1,0 // mulps 0x13a80(%rip),%xmm0 # 3a2a0 <_sk_srcover_bgra_8888_sse2_lowp+0xe18>
- .byte 102,15,111,13,136,58,1,0 // movdqa 0x13a88(%rip),%xmm1 # 3a2b0 <_sk_srcover_bgra_8888_sse2_lowp+0xe28>
+ .byte 15,89,5,0,45,1,0 // mulps 0x12d00(%rip),%xmm0 # 39520 <_sk_srcover_bgra_8888_sse2_lowp+0xe1c>
+ .byte 102,15,111,13,8,45,1,0 // movdqa 0x12d08(%rip),%xmm1 # 39530 <_sk_srcover_bgra_8888_sse2_lowp+0xe2c>
.byte 102,15,219,202 // pand %xmm2,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,138,58,1,0 // mulps 0x13a8a(%rip),%xmm1 # 3a2c0 <_sk_srcover_bgra_8888_sse2_lowp+0xe38>
- .byte 102,15,219,21,146,58,1,0 // pand 0x13a92(%rip),%xmm2 # 3a2d0 <_sk_srcover_bgra_8888_sse2_lowp+0xe48>
+ .byte 15,89,13,10,45,1,0 // mulps 0x12d0a(%rip),%xmm1 # 39540 <_sk_srcover_bgra_8888_sse2_lowp+0xe3c>
+ .byte 102,15,219,21,18,45,1,0 // pand 0x12d12(%rip),%xmm2 # 39550 <_sk_srcover_bgra_8888_sse2_lowp+0xe4c>
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,152,58,1,0 // mulps 0x13a98(%rip),%xmm2 # 3a2e0 <_sk_srcover_bgra_8888_sse2_lowp+0xe58>
+ .byte 15,89,21,24,45,1,0 // mulps 0x12d18(%rip),%xmm2 # 39560 <_sk_srcover_bgra_8888_sse2_lowp+0xe5c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,31,56,1,0 // movaps 0x1381f(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,159,42,1,0 // movaps 0x12a9f(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
HIDDEN _sk_store_565_sse2
@@ -47247,12 +47247,12 @@ _sk_store_565_sse2:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,210 // movslq %edx,%r10
- .byte 68,15,40,5,31,60,1,0 // movaps 0x13c1f(%rip),%xmm8 # 3a490 <_sk_srcover_bgra_8888_sse2_lowp+0x1008>
+ .byte 68,15,40,5,159,46,1,0 // movaps 0x12e9f(%rip),%xmm8 # 39710 <_sk_srcover_bgra_8888_sse2_lowp+0x100c>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
.byte 102,65,15,114,241,11 // pslld $0xb,%xmm9
- .byte 68,15,40,21,20,60,1,0 // movaps 0x13c14(%rip),%xmm10 # 3a4a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1018>
+ .byte 68,15,40,21,148,46,1,0 // movaps 0x12e94(%rip),%xmm10 # 39720 <_sk_srcover_bgra_8888_sse2_lowp+0x101c>
.byte 68,15,89,209 // mulps %xmm1,%xmm10
.byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
.byte 102,65,15,114,242,5 // pslld $0x5,%xmm10
@@ -47301,21 +47301,21 @@ _sk_load_4444_sse2:
.byte 117,98 // jne 2698f <_sk_load_4444_sse2+0x7d>
.byte 243,67,15,126,28,80 // movq (%r8,%r10,2),%xmm3
.byte 102,15,97,216 // punpcklwd %xmm0,%xmm3
- .byte 102,15,111,5,113,59,1,0 // movdqa 0x13b71(%rip),%xmm0 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,5,241,45,1,0 // movdqa 0x12df1(%rip),%xmm0 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,115,59,1,0 // mulps 0x13b73(%rip),%xmm0 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,13,123,59,1,0 // movdqa 0x13b7b(%rip),%xmm1 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,5,243,45,1,0 // mulps 0x12df3(%rip),%xmm0 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,13,251,45,1,0 // movdqa 0x12dfb(%rip),%xmm1 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,203 // pand %xmm3,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,125,59,1,0 // mulps 0x13b7d(%rip),%xmm1 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,21,133,59,1,0 // movdqa 0x13b85(%rip),%xmm2 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,13,253,45,1,0 // mulps 0x12dfd(%rip),%xmm1 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,21,5,46,1,0 // movdqa 0x12e05(%rip),%xmm2 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,211 // pand %xmm3,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,135,59,1,0 // mulps 0x13b87(%rip),%xmm2 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,29,143,59,1,0 // pand 0x13b8f(%rip),%xmm3 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,21,7,46,1,0 // mulps 0x12e07(%rip),%xmm2 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,29,15,46,1,0 // pand 0x12e0f(%rip),%xmm3 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,149,59,1,0 // mulps 0x13b95(%rip),%xmm3 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,29,21,46,1,0 // mulps 0x12e15(%rip),%xmm3 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -47353,21 +47353,21 @@ _sk_load_4444_dst_sse2:
.byte 117,98 // jne 26a5a <_sk_load_4444_dst_sse2+0x7d>
.byte 243,67,15,126,60,80 // movq (%r8,%r10,2),%xmm7
.byte 102,15,97,248 // punpcklwd %xmm0,%xmm7
- .byte 102,15,111,37,166,58,1,0 // movdqa 0x13aa6(%rip),%xmm4 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,37,38,45,1,0 // movdqa 0x12d26(%rip),%xmm4 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,231 // pand %xmm7,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 15,89,37,168,58,1,0 // mulps 0x13aa8(%rip),%xmm4 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,45,176,58,1,0 // movdqa 0x13ab0(%rip),%xmm5 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,37,40,45,1,0 // mulps 0x12d28(%rip),%xmm4 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,45,48,45,1,0 // movdqa 0x12d30(%rip),%xmm5 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,239 // pand %xmm7,%xmm5
.byte 15,91,237 // cvtdq2ps %xmm5,%xmm5
- .byte 15,89,45,178,58,1,0 // mulps 0x13ab2(%rip),%xmm5 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,53,186,58,1,0 // movdqa 0x13aba(%rip),%xmm6 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,45,50,45,1,0 // mulps 0x12d32(%rip),%xmm5 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,53,58,45,1,0 // movdqa 0x12d3a(%rip),%xmm6 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,247 // pand %xmm7,%xmm6
.byte 15,91,246 // cvtdq2ps %xmm6,%xmm6
- .byte 15,89,53,188,58,1,0 // mulps 0x13abc(%rip),%xmm6 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,61,196,58,1,0 // pand 0x13ac4(%rip),%xmm7 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,53,60,45,1,0 // mulps 0x12d3c(%rip),%xmm6 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,61,68,45,1,0 // pand 0x12d44(%rip),%xmm7 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,255 // cvtdq2ps %xmm7,%xmm7
- .byte 15,89,61,202,58,1,0 // mulps 0x13aca(%rip),%xmm7 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,61,74,45,1,0 // mulps 0x12d4a(%rip),%xmm7 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 65,137,249 // mov %edi,%r9d
@@ -47437,21 +47437,21 @@ _sk_gather_4444_sse2:
.byte 65,15,183,4,64 // movzwl (%r8,%rax,2),%eax
.byte 102,15,196,216,3 // pinsrw $0x3,%eax,%xmm3
.byte 102,65,15,97,216 // punpcklwd %xmm8,%xmm3
- .byte 102,15,111,5,70,57,1,0 // movdqa 0x13946(%rip),%xmm0 # 3a4b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1028>
+ .byte 102,15,111,5,198,43,1,0 // movdqa 0x12bc6(%rip),%xmm0 # 39730 <_sk_srcover_bgra_8888_sse2_lowp+0x102c>
.byte 102,15,219,195 // pand %xmm3,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 15,89,5,72,57,1,0 // mulps 0x13948(%rip),%xmm0 # 3a4c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1038>
- .byte 102,15,111,13,80,57,1,0 // movdqa 0x13950(%rip),%xmm1 # 3a4d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1048>
+ .byte 15,89,5,200,43,1,0 // mulps 0x12bc8(%rip),%xmm0 # 39740 <_sk_srcover_bgra_8888_sse2_lowp+0x103c>
+ .byte 102,15,111,13,208,43,1,0 // movdqa 0x12bd0(%rip),%xmm1 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x104c>
.byte 102,15,219,203 // pand %xmm3,%xmm1
.byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
- .byte 15,89,13,82,57,1,0 // mulps 0x13952(%rip),%xmm1 # 3a4e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1058>
- .byte 102,15,111,21,90,57,1,0 // movdqa 0x1395a(%rip),%xmm2 # 3a4f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1068>
+ .byte 15,89,13,210,43,1,0 // mulps 0x12bd2(%rip),%xmm1 # 39760 <_sk_srcover_bgra_8888_sse2_lowp+0x105c>
+ .byte 102,15,111,21,218,43,1,0 // movdqa 0x12bda(%rip),%xmm2 # 39770 <_sk_srcover_bgra_8888_sse2_lowp+0x106c>
.byte 102,15,219,211 // pand %xmm3,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,89,21,92,57,1,0 // mulps 0x1395c(%rip),%xmm2 # 3a500 <_sk_srcover_bgra_8888_sse2_lowp+0x1078>
- .byte 102,15,219,29,100,57,1,0 // pand 0x13964(%rip),%xmm3 # 3a510 <_sk_srcover_bgra_8888_sse2_lowp+0x1088>
+ .byte 15,89,21,220,43,1,0 // mulps 0x12bdc(%rip),%xmm2 # 39780 <_sk_srcover_bgra_8888_sse2_lowp+0x107c>
+ .byte 102,15,219,29,228,43,1,0 // pand 0x12be4(%rip),%xmm3 # 39790 <_sk_srcover_bgra_8888_sse2_lowp+0x108c>
.byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
- .byte 15,89,29,106,57,1,0 // mulps 0x1396a(%rip),%xmm3 # 3a520 <_sk_srcover_bgra_8888_sse2_lowp+0x1098>
+ .byte 15,89,29,234,43,1,0 // mulps 0x12bea(%rip),%xmm3 # 397a0 <_sk_srcover_bgra_8888_sse2_lowp+0x109c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -47466,7 +47466,7 @@ _sk_store_4444_sse2:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 76,99,210 // movslq %edx,%r10
- .byte 68,15,40,5,88,57,1,0 // movaps 0x13958(%rip),%xmm8 # 3a530 <_sk_srcover_bgra_8888_sse2_lowp+0x10a8>
+ .byte 68,15,40,5,216,43,1,0 // movaps 0x12bd8(%rip),%xmm8 # 397b0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ac>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -47524,11 +47524,11 @@ _sk_load_8888_sse2:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,98 // jne 26d0b <_sk_load_8888_sse2+0x7e>
.byte 243,69,15,111,12,128 // movdqu (%r8,%rax,4),%xmm9
- .byte 102,15,111,21,105,52,1,0 // movdqa 0x13469(%rip),%xmm2 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,21,233,38,1,0 // movdqa 0x126e9(%rip),%xmm2 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,181,53,1,0 // movaps 0x135b5(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,53,40,1,0 // movaps 0x12835(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,114,209,8 // psrld $0x8,%xmm1
@@ -47575,11 +47575,11 @@ _sk_load_8888_dst_sse2:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,98 // jne 26dca <_sk_load_8888_dst_sse2+0x7e>
.byte 243,69,15,111,12,128 // movdqu (%r8,%rax,4),%xmm9
- .byte 102,15,111,53,170,51,1,0 // movdqa 0x133aa(%rip),%xmm6 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,53,42,38,1,0 // movdqa 0x1262a(%rip),%xmm6 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,225 // movdqa %xmm9,%xmm4
.byte 102,15,219,230 // pand %xmm6,%xmm4
.byte 15,91,228 // cvtdq2ps %xmm4,%xmm4
- .byte 68,15,40,5,246,52,1,0 // movaps 0x134f6(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,118,39,1,0 // movaps 0x12776(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,224 // mulps %xmm8,%xmm4
.byte 102,65,15,111,233 // movdqa %xmm9,%xmm5
.byte 102,15,114,213,8 // psrld $0x8,%xmm5
@@ -47656,11 +47656,11 @@ _sk_gather_8888_sse2:
.byte 102,67,15,110,12,136 // movd (%r8,%r9,4),%xmm1
.byte 102,15,98,200 // punpckldq %xmm0,%xmm1
.byte 102,68,15,108,201 // punpcklqdq %xmm1,%xmm9
- .byte 102,15,111,21,93,50,1,0 // movdqa 0x1325d(%rip),%xmm2 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,21,221,36,1,0 // movdqa 0x124dd(%rip),%xmm2 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,219,194 // pand %xmm2,%xmm0
.byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
- .byte 68,15,40,5,169,51,1,0 // movaps 0x133a9(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,41,38,1,0 // movaps 0x12629(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,114,209,8 // psrld $0x8,%xmm1
@@ -47689,7 +47689,7 @@ _sk_store_8888_sse2:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,26,50,1,0 // movaps 0x1321a(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,154,36,1,0 // movaps 0x1249a(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -47741,11 +47741,11 @@ _sk_load_bgra_sse2:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,98 // jne 2704a <_sk_load_bgra_sse2+0x7e>
.byte 243,69,15,111,12,128 // movdqu (%r8,%rax,4),%xmm9
- .byte 102,15,111,5,42,49,1,0 // movdqa 0x1312a(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,170,35,1,0 // movdqa 0x123aa(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,219,200 // pand %xmm0,%xmm1
.byte 15,91,209 // cvtdq2ps %xmm1,%xmm2
- .byte 68,15,40,5,118,50,1,0 // movaps 0x13276(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,246,36,1,0 // movaps 0x124f6(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,114,209,8 // psrld $0x8,%xmm1
@@ -47792,11 +47792,11 @@ _sk_load_bgra_dst_sse2:
.byte 72,133,255 // test %rdi,%rdi
.byte 117,98 // jne 27109 <_sk_load_bgra_dst_sse2+0x7e>
.byte 243,69,15,111,12,128 // movdqu (%r8,%rax,4),%xmm9
- .byte 102,15,111,37,107,48,1,0 // movdqa 0x1306b(%rip),%xmm4 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,37,235,34,1,0 // movdqa 0x122eb(%rip),%xmm4 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,233 // movdqa %xmm9,%xmm5
.byte 102,15,219,236 // pand %xmm4,%xmm5
.byte 15,91,245 // cvtdq2ps %xmm5,%xmm6
- .byte 68,15,40,5,183,49,1,0 // movaps 0x131b7(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,55,36,1,0 // movaps 0x12437(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,240 // mulps %xmm8,%xmm6
.byte 102,65,15,111,233 // movdqa %xmm9,%xmm5
.byte 102,15,114,213,8 // psrld $0x8,%xmm5
@@ -47873,11 +47873,11 @@ _sk_gather_bgra_sse2:
.byte 102,67,15,110,12,136 // movd (%r8,%r9,4),%xmm1
.byte 102,15,98,200 // punpckldq %xmm0,%xmm1
.byte 102,68,15,108,201 // punpcklqdq %xmm1,%xmm9
- .byte 102,15,111,5,30,47,1,0 // movdqa 0x12f1e(%rip),%xmm0 # 3a120 <_sk_srcover_bgra_8888_sse2_lowp+0xc98>
+ .byte 102,15,111,5,158,33,1,0 // movdqa 0x1219e(%rip),%xmm0 # 393a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc9c>
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,219,200 // pand %xmm0,%xmm1
.byte 15,91,209 // cvtdq2ps %xmm1,%xmm2
- .byte 68,15,40,5,106,48,1,0 // movaps 0x1306a(%rip),%xmm8 # 3a280 <_sk_srcover_bgra_8888_sse2_lowp+0xdf8>
+ .byte 68,15,40,5,234,34,1,0 // movaps 0x122ea(%rip),%xmm8 # 39500 <_sk_srcover_bgra_8888_sse2_lowp+0xdfc>
.byte 65,15,89,208 // mulps %xmm8,%xmm2
.byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
.byte 102,15,114,209,8 // psrld $0x8,%xmm1
@@ -47906,7 +47906,7 @@ _sk_store_bgra_sse2:
.byte 73,193,224,2 // shl $0x2,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 72,99,194 // movslq %edx,%rax
- .byte 68,15,40,5,219,46,1,0 // movaps 0x12edb(%rip),%xmm8 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,5,91,33,1,0 // movaps 0x1215b(%rip),%xmm8 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,40,202 // movaps %xmm2,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
@@ -47968,18 +47968,18 @@ _sk_load_f16_sse2:
.byte 102,69,15,239,210 // pxor %xmm10,%xmm10
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,65,15,97,202 // punpcklwd %xmm10,%xmm1
- .byte 102,68,15,111,13,213,49,1,0 // movdqa 0x131d5(%rip),%xmm9 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,13,85,36,1,0 // movdqa 0x12455(%rip),%xmm9 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,68,15,111,225 // movdqa %xmm1,%xmm12
.byte 102,69,15,219,225 // pand %xmm9,%xmm12
- .byte 102,68,15,111,29,210,49,1,0 // movdqa 0x131d2(%rip),%xmm11 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,29,82,36,1,0 // movdqa 0x12452(%rip),%xmm11 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,203 // pand %xmm11,%xmm1
- .byte 102,15,111,29,213,49,1,0 // movdqa 0x131d5(%rip),%xmm3 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,29,85,36,1,0 // movdqa 0x12455(%rip),%xmm3 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,65,15,114,244,16 // pslld $0x10,%xmm12
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
.byte 102,15,102,193 // pcmpgtd %xmm1,%xmm0
.byte 102,15,114,241,13 // pslld $0xd,%xmm1
.byte 102,65,15,235,204 // por %xmm12,%xmm1
- .byte 102,68,15,111,37,196,49,1,0 // movdqa 0x131c4(%rip),%xmm12 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,37,68,36,1,0 // movdqa 0x12444(%rip),%xmm12 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,204 // paddd %xmm12,%xmm1
.byte 102,15,223,193 // pandn %xmm1,%xmm0
.byte 102,65,15,105,210 // punpckhwd %xmm10,%xmm2
@@ -48053,18 +48053,18 @@ _sk_load_f16_dst_sse2:
.byte 102,69,15,239,210 // pxor %xmm10,%xmm10
.byte 102,15,111,238 // movdqa %xmm6,%xmm5
.byte 102,65,15,97,234 // punpcklwd %xmm10,%xmm5
- .byte 102,68,15,111,13,76,48,1,0 // movdqa 0x1304c(%rip),%xmm9 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,13,204,34,1,0 // movdqa 0x122cc(%rip),%xmm9 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,68,15,111,229 // movdqa %xmm5,%xmm12
.byte 102,69,15,219,225 // pand %xmm9,%xmm12
- .byte 102,68,15,111,29,73,48,1,0 // movdqa 0x13049(%rip),%xmm11 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,29,201,34,1,0 // movdqa 0x122c9(%rip),%xmm11 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,235 // pand %xmm11,%xmm5
- .byte 102,15,111,61,76,48,1,0 // movdqa 0x1304c(%rip),%xmm7 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,61,204,34,1,0 // movdqa 0x122cc(%rip),%xmm7 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,65,15,114,244,16 // pslld $0x10,%xmm12
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
.byte 102,15,102,229 // pcmpgtd %xmm5,%xmm4
.byte 102,15,114,245,13 // pslld $0xd,%xmm5
.byte 102,65,15,235,236 // por %xmm12,%xmm5
- .byte 102,68,15,111,37,59,48,1,0 // movdqa 0x1303b(%rip),%xmm12 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,37,187,34,1,0 // movdqa 0x122bb(%rip),%xmm12 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,236 // paddd %xmm12,%xmm5
.byte 102,15,223,229 // pandn %xmm5,%xmm4
.byte 102,65,15,105,242 // punpckhwd %xmm10,%xmm6
@@ -48166,18 +48166,18 @@ _sk_gather_f16_sse2:
.byte 102,68,15,105,201 // punpckhwd %xmm1,%xmm9
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,65,15,97,200 // punpcklwd %xmm8,%xmm1
- .byte 102,68,15,111,21,74,46,1,0 // movdqa 0x12e4a(%rip),%xmm10 # 3a540 <_sk_srcover_bgra_8888_sse2_lowp+0x10b8>
+ .byte 102,68,15,111,21,202,32,1,0 // movdqa 0x120ca(%rip),%xmm10 # 397c0 <_sk_srcover_bgra_8888_sse2_lowp+0x10bc>
.byte 102,68,15,111,225 // movdqa %xmm1,%xmm12
.byte 102,69,15,219,226 // pand %xmm10,%xmm12
- .byte 102,68,15,111,29,71,46,1,0 // movdqa 0x12e47(%rip),%xmm11 # 3a550 <_sk_srcover_bgra_8888_sse2_lowp+0x10c8>
+ .byte 102,68,15,111,29,199,32,1,0 // movdqa 0x120c7(%rip),%xmm11 # 397d0 <_sk_srcover_bgra_8888_sse2_lowp+0x10cc>
.byte 102,65,15,219,203 // pand %xmm11,%xmm1
- .byte 102,15,111,29,74,46,1,0 // movdqa 0x12e4a(%rip),%xmm3 # 3a560 <_sk_srcover_bgra_8888_sse2_lowp+0x10d8>
+ .byte 102,15,111,29,202,32,1,0 // movdqa 0x120ca(%rip),%xmm3 # 397e0 <_sk_srcover_bgra_8888_sse2_lowp+0x10dc>
.byte 102,65,15,114,244,16 // pslld $0x10,%xmm12
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
.byte 102,15,102,193 // pcmpgtd %xmm1,%xmm0
.byte 102,15,114,241,13 // pslld $0xd,%xmm1
.byte 102,65,15,235,204 // por %xmm12,%xmm1
- .byte 102,68,15,111,37,57,46,1,0 // movdqa 0x12e39(%rip),%xmm12 # 3a570 <_sk_srcover_bgra_8888_sse2_lowp+0x10e8>
+ .byte 102,68,15,111,37,185,32,1,0 // movdqa 0x120b9(%rip),%xmm12 # 397f0 <_sk_srcover_bgra_8888_sse2_lowp+0x10ec>
.byte 102,65,15,254,204 // paddd %xmm12,%xmm1
.byte 102,15,223,193 // pandn %xmm1,%xmm0
.byte 102,65,15,105,208 // punpckhwd %xmm8,%xmm2
@@ -48220,19 +48220,19 @@ HIDDEN _sk_store_f16_sse2
FUNCTION(_sk_store_f16_sse2)
_sk_store_f16_sse2:
.byte 15,41,124,36,232 // movaps %xmm7,-0x18(%rsp)
- .byte 102,68,15,111,29,141,45,1,0 // movdqa 0x12d8d(%rip),%xmm11 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
+ .byte 102,68,15,111,29,13,32,1,0 // movdqa 0x1200d(%rip),%xmm11 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
.byte 102,69,15,219,195 // pand %xmm11,%xmm8
- .byte 102,68,15,111,21,138,45,1,0 // movdqa 0x12d8a(%rip),%xmm10 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 102,68,15,111,21,10,32,1,0 // movdqa 0x1200a(%rip),%xmm10 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 102,68,15,111,240 // movdqa %xmm0,%xmm14
.byte 102,69,15,219,242 // pand %xmm10,%xmm14
- .byte 102,15,111,61,136,45,1,0 // movdqa 0x12d88(%rip),%xmm7 # 3a5a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1118>
- .byte 102,68,15,111,37,47,47,1,0 // movdqa 0x12f2f(%rip),%xmm12 # 3a750 <_sk_srcover_bgra_8888_sse2_lowp+0x12c8>
+ .byte 102,15,111,61,8,32,1,0 // movdqa 0x12008(%rip),%xmm7 # 39820 <_sk_srcover_bgra_8888_sse2_lowp+0x111c>
+ .byte 102,68,15,111,37,175,33,1,0 // movdqa 0x121af(%rip),%xmm12 # 399d0 <_sk_srcover_bgra_8888_sse2_lowp+0x12cc>
.byte 102,68,15,111,248 // movdqa %xmm0,%xmm15
.byte 102,65,15,114,247,3 // pslld $0x3,%xmm15
.byte 102,69,15,219,252 // pand %xmm12,%xmm15
.byte 102,69,15,254,248 // paddd %xmm8,%xmm15
- .byte 102,68,15,111,45,225,41,1,0 // movdqa 0x129e1(%rip),%xmm13 # 3a220 <_sk_srcover_bgra_8888_sse2_lowp+0xd98>
+ .byte 102,68,15,111,45,97,28,1,0 // movdqa 0x11c61(%rip),%xmm13 # 394a0 <_sk_srcover_bgra_8888_sse2_lowp+0xd9c>
.byte 102,69,15,254,253 // paddd %xmm13,%xmm15
.byte 102,65,15,114,231,16 // psrad $0x10,%xmm15
.byte 102,68,15,111,199 // movdqa %xmm7,%xmm8
@@ -48339,7 +48339,7 @@ _sk_load_u16_be_sse2:
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,65,15,97,201 // punpcklwd %xmm9,%xmm1
.byte 15,91,193 // cvtdq2ps %xmm1,%xmm0
- .byte 68,15,40,5,246,40,1,0 // movaps 0x128f6(%rip),%xmm8 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 68,15,40,5,118,27,1,0 // movaps 0x11b76(%rip),%xmm8 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
.byte 102,15,113,241,8 // psllw $0x8,%xmm1
@@ -48414,7 +48414,7 @@ _sk_load_rgb_u16_be_sse2:
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,65,15,97,200 // punpcklwd %xmm8,%xmm1
.byte 15,91,193 // cvtdq2ps %xmm1,%xmm0
- .byte 68,15,40,13,192,39,1,0 // movaps 0x127c0(%rip),%xmm9 # 3a310 <_sk_srcover_bgra_8888_sse2_lowp+0xe88>
+ .byte 68,15,40,13,64,26,1,0 // movaps 0x11a40(%rip),%xmm9 # 39590 <_sk_srcover_bgra_8888_sse2_lowp+0xe8c>
.byte 65,15,89,193 // mulps %xmm9,%xmm0
.byte 102,15,111,203 // movdqa %xmm3,%xmm1
.byte 102,15,113,241,8 // psllw $0x8,%xmm1
@@ -48431,7 +48431,7 @@ _sk_load_rgb_u16_be_sse2:
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
.byte 65,15,89,209 // mulps %xmm9,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,215,36,1,0 // movaps 0x124d7(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,87,23,1,0 // movaps 0x11757(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,20,64 // movd (%r8,%rax,2),%xmm2
.byte 102,15,239,201 // pxor %xmm1,%xmm1
@@ -48465,7 +48465,7 @@ _sk_store_u16_be_sse2:
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 73,99,193 // movslq %r9d,%rax
- .byte 68,15,40,21,160,41,1,0 // movaps 0x129a0(%rip),%xmm10 # 3a5c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1138>
+ .byte 68,15,40,21,32,28,1,0 // movaps 0x11c20(%rip),%xmm10 # 39840 <_sk_srcover_bgra_8888_sse2_lowp+0x113c>
.byte 68,15,40,192 // movaps %xmm0,%xmm8
.byte 69,15,89,194 // mulps %xmm10,%xmm8
.byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
@@ -48672,7 +48672,7 @@ _sk_repeat_x_sse2:
.byte 243,69,15,91,209 // cvttps2dq %xmm9,%xmm10
.byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
.byte 69,15,194,202,1 // cmpltps %xmm10,%xmm9
- .byte 68,15,84,13,63,33,1,0 // andps 0x1213f(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,84,13,191,19,1,0 // andps 0x113bf(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,209 // subps %xmm9,%xmm10
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 69,15,89,194 // mulps %xmm10,%xmm8
@@ -48692,7 +48692,7 @@ _sk_repeat_y_sse2:
.byte 243,69,15,91,209 // cvttps2dq %xmm9,%xmm10
.byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
.byte 69,15,194,202,1 // cmpltps %xmm10,%xmm9
- .byte 68,15,84,13,254,32,1,0 // andps 0x120fe(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,84,13,126,19,1,0 // andps 0x1137e(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,209 // subps %xmm9,%xmm10
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 69,15,89,194 // mulps %xmm10,%xmm8
@@ -48712,13 +48712,13 @@ _sk_mirror_x_sse2:
.byte 65,15,92,194 // subps %xmm10,%xmm0
.byte 243,69,15,88,192 // addss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
- .byte 243,68,15,89,13,128,23,1,0 // mulss 0x11780(%rip),%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,13,252,9,1,0 // mulss 0x109fc(%rip),%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,89,200 // mulps %xmm0,%xmm9
.byte 243,69,15,91,217 // cvttps2dq %xmm9,%xmm11
.byte 69,15,91,219 // cvtdq2ps %xmm11,%xmm11
.byte 69,15,194,203,1 // cmpltps %xmm11,%xmm9
- .byte 68,15,84,13,157,32,1,0 // andps 0x1209d(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,84,13,29,19,1,0 // andps 0x1131d(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,87,228 // xorps %xmm12,%xmm12
.byte 69,15,92,217 // subps %xmm9,%xmm11
.byte 69,15,89,216 // mulps %xmm8,%xmm11
@@ -48741,13 +48741,13 @@ _sk_mirror_y_sse2:
.byte 65,15,92,202 // subps %xmm10,%xmm1
.byte 243,69,15,88,192 // addss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
- .byte 243,68,15,89,13,20,23,1,0 // mulss 0x11714(%rip),%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,13,144,9,1,0 // mulss 0x10990(%rip),%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,89,201 // mulps %xmm1,%xmm9
.byte 243,69,15,91,217 // cvttps2dq %xmm9,%xmm11
.byte 69,15,91,219 // cvtdq2ps %xmm11,%xmm11
.byte 69,15,194,203,1 // cmpltps %xmm11,%xmm9
- .byte 68,15,84,13,49,32,1,0 // andps 0x12031(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,84,13,177,18,1,0 // andps 0x112b1(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,87,228 // xorps %xmm12,%xmm12
.byte 69,15,92,217 // subps %xmm9,%xmm11
.byte 69,15,89,216 // mulps %xmm8,%xmm11
@@ -48764,7 +48764,7 @@ FUNCTION(_sk_clamp_x_1_sse2)
_sk_clamp_x_1_sse2:
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 68,15,95,192 // maxps %xmm0,%xmm8
- .byte 68,15,93,5,1,32,1,0 // minps 0x12001(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,93,5,129,18,1,0 // minps 0x11281(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 65,15,40,192 // movaps %xmm8,%xmm0
.byte 255,224 // jmpq *%rax
@@ -48777,7 +48777,7 @@ _sk_repeat_x_1_sse2:
.byte 69,15,91,200 // cvtdq2ps %xmm8,%xmm9
.byte 68,15,40,208 // movaps %xmm0,%xmm10
.byte 69,15,194,209,1 // cmpltps %xmm9,%xmm10
- .byte 68,15,40,29,223,31,1,0 // movaps 0x11fdf(%rip),%xmm11 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,29,95,18,1,0 // movaps 0x1125f(%rip),%xmm11 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,211 // andps %xmm11,%xmm10
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 69,15,92,202 // subps %xmm10,%xmm9
@@ -48792,14 +48792,14 @@ HIDDEN _sk_mirror_x_1_sse2
.globl _sk_mirror_x_1_sse2
FUNCTION(_sk_mirror_x_1_sse2)
_sk_mirror_x_1_sse2:
- .byte 68,15,40,13,23,32,1,0 // movaps 0x12017(%rip),%xmm9 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,13,151,18,1,0 // movaps 0x11297(%rip),%xmm9 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 65,15,88,193 // addps %xmm9,%xmm0
- .byte 68,15,40,21,155,31,1,0 // movaps 0x11f9b(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,21,27,18,1,0 // movaps 0x1121b(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,89,208 // mulps %xmm0,%xmm10
.byte 243,69,15,91,194 // cvttps2dq %xmm10,%xmm8
.byte 69,15,91,216 // cvtdq2ps %xmm8,%xmm11
.byte 69,15,194,211,1 // cmpltps %xmm11,%xmm10
- .byte 68,15,40,37,145,31,1,0 // movaps 0x11f91(%rip),%xmm12 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,37,17,18,1,0 // movaps 0x11211(%rip),%xmm12 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,212 // andps %xmm12,%xmm10
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 69,15,92,218 // subps %xmm10,%xmm11
@@ -48820,10 +48820,10 @@ HIDDEN _sk_luminance_to_alpha_sse2
FUNCTION(_sk_luminance_to_alpha_sse2)
_sk_luminance_to_alpha_sse2:
.byte 15,40,218 // movaps %xmm2,%xmm3
- .byte 15,89,5,179,36,1,0 // mulps 0x124b3(%rip),%xmm0 # 3a5d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1148>
- .byte 15,89,13,188,36,1,0 // mulps 0x124bc(%rip),%xmm1 # 3a5e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1158>
+ .byte 15,89,5,51,23,1,0 // mulps 0x11733(%rip),%xmm0 # 39850 <_sk_srcover_bgra_8888_sse2_lowp+0x114c>
+ .byte 15,89,13,60,23,1,0 // mulps 0x1173c(%rip),%xmm1 # 39860 <_sk_srcover_bgra_8888_sse2_lowp+0x115c>
.byte 15,88,200 // addps %xmm0,%xmm1
- .byte 15,89,29,194,36,1,0 // mulps 0x124c2(%rip),%xmm3 # 3a5f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1168>
+ .byte 15,89,29,66,23,1,0 // mulps 0x11742(%rip),%xmm3 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x116c>
.byte 15,88,217 // addps %xmm1,%xmm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
@@ -49390,29 +49390,29 @@ _sk_xy_to_unit_angle_sse2:
.byte 69,15,94,220 // divps %xmm12,%xmm11
.byte 69,15,40,227 // movaps %xmm11,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
- .byte 68,15,40,45,30,28,1,0 // movaps 0x11c1e(%rip),%xmm13 # 3a600 <_sk_srcover_bgra_8888_sse2_lowp+0x1178>
+ .byte 68,15,40,45,158,14,1,0 // movaps 0x10e9e(%rip),%xmm13 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x117c>
.byte 69,15,89,236 // mulps %xmm12,%xmm13
- .byte 68,15,88,45,34,28,1,0 // addps 0x11c22(%rip),%xmm13 # 3a610 <_sk_srcover_bgra_8888_sse2_lowp+0x1188>
+ .byte 68,15,88,45,162,14,1,0 // addps 0x10ea2(%rip),%xmm13 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x118c>
.byte 69,15,89,236 // mulps %xmm12,%xmm13
- .byte 68,15,88,45,38,28,1,0 // addps 0x11c26(%rip),%xmm13 # 3a620 <_sk_srcover_bgra_8888_sse2_lowp+0x1198>
+ .byte 68,15,88,45,166,14,1,0 // addps 0x10ea6(%rip),%xmm13 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x119c>
.byte 69,15,89,236 // mulps %xmm12,%xmm13
- .byte 68,15,88,45,42,28,1,0 // addps 0x11c2a(%rip),%xmm13 # 3a630 <_sk_srcover_bgra_8888_sse2_lowp+0x11a8>
+ .byte 68,15,88,45,170,14,1,0 // addps 0x10eaa(%rip),%xmm13 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ac>
.byte 69,15,89,235 // mulps %xmm11,%xmm13
.byte 69,15,194,202,1 // cmpltps %xmm10,%xmm9
- .byte 68,15,40,21,41,28,1,0 // movaps 0x11c29(%rip),%xmm10 # 3a640 <_sk_srcover_bgra_8888_sse2_lowp+0x11b8>
+ .byte 68,15,40,21,169,14,1,0 // movaps 0x10ea9(%rip),%xmm10 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x11bc>
.byte 69,15,92,213 // subps %xmm13,%xmm10
.byte 69,15,84,209 // andps %xmm9,%xmm10
.byte 69,15,85,205 // andnps %xmm13,%xmm9
.byte 69,15,86,202 // orps %xmm10,%xmm9
.byte 68,15,194,192,1 // cmpltps %xmm0,%xmm8
- .byte 68,15,40,21,44,22,1,0 // movaps 0x1162c(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,21,172,8,1,0 // movaps 0x108ac(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,92,209 // subps %xmm9,%xmm10
.byte 69,15,84,208 // andps %xmm8,%xmm10
.byte 69,15,85,193 // andnps %xmm9,%xmm8
.byte 69,15,86,194 // orps %xmm10,%xmm8
.byte 68,15,40,201 // movaps %xmm1,%xmm9
.byte 68,15,194,200,1 // cmpltps %xmm0,%xmm9
- .byte 68,15,40,21,27,22,1,0 // movaps 0x1161b(%rip),%xmm10 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,21,155,8,1,0 // movaps 0x1089b(%rip),%xmm10 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,208 // subps %xmm8,%xmm10
.byte 69,15,84,209 // andps %xmm9,%xmm10
.byte 69,15,85,200 // andnps %xmm8,%xmm9
@@ -49447,7 +49447,7 @@ _sk_xy_to_2pt_conical_quadratic_max_sse2:
.byte 243,69,15,89,203 // mulss %xmm11,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,146,27,1,0 // mulps 0x11b92(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,18,14,1,0 // mulps 0x10e12(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
@@ -49455,7 +49455,7 @@ _sk_xy_to_2pt_conical_quadratic_max_sse2:
.byte 243,69,15,89,219 // mulss %xmm11,%xmm11
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,92,227 // subps %xmm11,%xmm12
- .byte 68,15,89,21,125,27,1,0 // mulps 0x11b7d(%rip),%xmm10 # 3a660 <_sk_srcover_bgra_8888_sse2_lowp+0x11d8>
+ .byte 68,15,89,21,253,13,1,0 // mulps 0x10dfd(%rip),%xmm10 # 398e0 <_sk_srcover_bgra_8888_sse2_lowp+0x11dc>
.byte 69,15,89,212 // mulps %xmm12,%xmm10
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
@@ -49464,8 +49464,8 @@ _sk_xy_to_2pt_conical_quadratic_max_sse2:
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 65,15,92,193 // subps %xmm9,%xmm0
- .byte 68,15,87,13,117,26,1,0 // xorps 0x11a75(%rip),%xmm9 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
- .byte 68,15,89,5,77,21,1,0 // mulps 0x1154d(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,87,13,245,12,1,0 // xorps 0x10cf5(%rip),%xmm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
+ .byte 68,15,89,5,205,7,1,0 // mulps 0x107cd(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 69,15,92,202 // subps %xmm10,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
@@ -49486,7 +49486,7 @@ _sk_xy_to_2pt_conical_quadratic_min_sse2:
.byte 243,69,15,89,203 // mulss %xmm11,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,244,26,1,0 // mulps 0x11af4(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,116,13,1,0 // mulps 0x10d74(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,225 // movaps %xmm1,%xmm12
.byte 69,15,89,228 // mulps %xmm12,%xmm12
@@ -49494,7 +49494,7 @@ _sk_xy_to_2pt_conical_quadratic_min_sse2:
.byte 243,69,15,89,219 // mulss %xmm11,%xmm11
.byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
.byte 69,15,92,227 // subps %xmm11,%xmm12
- .byte 68,15,89,21,223,26,1,0 // mulps 0x11adf(%rip),%xmm10 # 3a660 <_sk_srcover_bgra_8888_sse2_lowp+0x11d8>
+ .byte 68,15,89,21,95,13,1,0 // mulps 0x10d5f(%rip),%xmm10 # 398e0 <_sk_srcover_bgra_8888_sse2_lowp+0x11dc>
.byte 69,15,89,212 // mulps %xmm12,%xmm10
.byte 65,15,40,193 // movaps %xmm9,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
@@ -49503,8 +49503,8 @@ _sk_xy_to_2pt_conical_quadratic_min_sse2:
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,40,194 // movaps %xmm10,%xmm0
.byte 65,15,92,193 // subps %xmm9,%xmm0
- .byte 68,15,87,13,215,25,1,0 // xorps 0x119d7(%rip),%xmm9 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
- .byte 68,15,89,5,175,20,1,0 // mulps 0x114af(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,87,13,87,12,1,0 // xorps 0x10c57(%rip),%xmm9 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
+ .byte 68,15,89,5,47,7,1,0 // mulps 0x1072f(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,89,192 // mulps %xmm8,%xmm0
.byte 69,15,92,202 // subps %xmm10,%xmm9
.byte 69,15,89,200 // mulps %xmm8,%xmm9
@@ -49522,7 +49522,7 @@ _sk_xy_to_2pt_conical_linear_sse2:
.byte 243,69,15,89,200 // mulss %xmm8,%xmm9
.byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
.byte 68,15,88,200 // addps %xmm0,%xmm9
- .byte 68,15,89,13,103,26,1,0 // mulps 0x11a67(%rip),%xmm9 # 3a650 <_sk_srcover_bgra_8888_sse2_lowp+0x11c8>
+ .byte 68,15,89,13,231,12,1,0 // mulps 0x10ce7(%rip),%xmm9 # 398d0 <_sk_srcover_bgra_8888_sse2_lowp+0x11cc>
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,40,209 // movaps %xmm1,%xmm10
.byte 69,15,89,210 // mulps %xmm10,%xmm10
@@ -49530,7 +49530,7 @@ _sk_xy_to_2pt_conical_linear_sse2:
.byte 243,69,15,89,192 // mulss %xmm8,%xmm8
.byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
.byte 65,15,92,192 // subps %xmm8,%xmm0
- .byte 15,87,5,115,25,1,0 // xorps 0x11973(%rip),%xmm0 # 3a580 <_sk_srcover_bgra_8888_sse2_lowp+0x10f8>
+ .byte 15,87,5,243,11,1,0 // xorps 0x10bf3(%rip),%xmm0 # 39800 <_sk_srcover_bgra_8888_sse2_lowp+0x10fc>
.byte 65,15,94,193 // divps %xmm9,%xmm0
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49576,7 +49576,7 @@ HIDDEN _sk_save_xy_sse2
FUNCTION(_sk_save_xy_sse2)
_sk_save_xy_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,218,19,1,0 // movaps 0x113da(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,90,6,1,0 // movaps 0x1065a(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,17,0 // movups %xmm0,(%rax)
.byte 68,15,40,200 // movaps %xmm0,%xmm9
.byte 69,15,88,200 // addps %xmm8,%xmm9
@@ -49584,7 +49584,7 @@ _sk_save_xy_sse2:
.byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
.byte 69,15,40,217 // movaps %xmm9,%xmm11
.byte 69,15,194,218,1 // cmpltps %xmm10,%xmm11
- .byte 68,15,40,37,197,19,1,0 // movaps 0x113c5(%rip),%xmm12 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,37,69,6,1,0 // movaps 0x10645(%rip),%xmm12 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,220 // andps %xmm12,%xmm11
.byte 69,15,92,211 // subps %xmm11,%xmm10
.byte 69,15,92,202 // subps %xmm10,%xmm9
@@ -49631,8 +49631,8 @@ _sk_bilinear_nx_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,37,25,1,0 // addps 0x11925(%rip),%xmm0 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,29,19,1,0 // movaps 0x1131d(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,165,11,1,0 // addps 0x10ba5(%rip),%xmm0 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,157,5,1,0 // movaps 0x1059d(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 68,15,17,136,0,1,0,0 // movups %xmm9,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49645,7 +49645,7 @@ _sk_bilinear_px_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,233,18,1,0 // addps 0x112e9(%rip),%xmm0 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,5,105,5,1,0 // addps 0x10569(%rip),%xmm0 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49657,8 +49657,8 @@ _sk_bilinear_ny_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,216,24,1,0 // addps 0x118d8(%rip),%xmm1 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,208,18,1,0 // movaps 0x112d0(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,88,11,1,0 // addps 0x10b58(%rip),%xmm1 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,80,5,1,0 // movaps 0x10550(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 68,15,17,136,64,1,0,0 // movups %xmm9,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49671,7 +49671,7 @@ _sk_bilinear_py_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,155,18,1,0 // addps 0x1129b(%rip),%xmm1 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,13,27,5,1,0 // addps 0x1051b(%rip),%xmm1 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49683,13 +49683,13 @@ _sk_bicubic_n3x_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,155,24,1,0 // addps 0x1189b(%rip),%xmm0 # 3a680 <_sk_srcover_bgra_8888_sse2_lowp+0x11f8>
- .byte 68,15,40,13,131,18,1,0 // movaps 0x11283(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,27,11,1,0 // addps 0x10b1b(%rip),%xmm0 # 39900 <_sk_srcover_bgra_8888_sse2_lowp+0x11fc>
+ .byte 68,15,40,13,3,5,1,0 // movaps 0x10503(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 69,15,40,193 // movaps %xmm9,%xmm8
.byte 69,15,89,192 // mulps %xmm8,%xmm8
- .byte 68,15,89,13,143,24,1,0 // mulps 0x1188f(%rip),%xmm9 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,13,103,20,1,0 // addps 0x11467(%rip),%xmm9 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,13,15,11,1,0 // mulps 0x10b0f(%rip),%xmm9 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,13,231,6,1,0 // addps 0x106e7(%rip),%xmm9 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 68,15,17,136,0,1,0,0 // movups %xmm9,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49702,16 +49702,16 @@ _sk_bicubic_n1x_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,67,24,1,0 // addps 0x11843(%rip),%xmm0 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,59,18,1,0 // movaps 0x1123b(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,5,195,10,1,0 // addps 0x10ac3(%rip),%xmm0 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,187,4,1,0 // movaps 0x104bb(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
- .byte 68,15,40,5,95,24,1,0 // movaps 0x1185f(%rip),%xmm8 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,5,223,10,1,0 // movaps 0x10adf(%rip),%xmm8 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,99,24,1,0 // addps 0x11863(%rip),%xmm8 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,5,227,10,1,0 // addps 0x10ae3(%rip),%xmm8 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,7,18,1,0 // addps 0x11207(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,88,5,135,4,1,0 // addps 0x10487(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,91,24,1,0 // addps 0x1185b(%rip),%xmm8 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,5,219,10,1,0 // addps 0x10adb(%rip),%xmm8 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49721,17 +49721,17 @@ HIDDEN _sk_bicubic_p1x_sse2
FUNCTION(_sk_bicubic_p1x_sse2)
_sk_bicubic_p1x_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,229,17,1,0 // movaps 0x111e5(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,101,4,1,0 // movaps 0x10465(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,136,128,0,0,0 // movups 0x80(%rax),%xmm9
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,21,14,24,1,0 // movaps 0x1180e(%rip),%xmm10 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,21,142,10,1,0 // movaps 0x10a8e(%rip),%xmm10 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,18,24,1,0 // addps 0x11812(%rip),%xmm10 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,21,146,10,1,0 // addps 0x10a92(%rip),%xmm10 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
.byte 69,15,88,208 // addps %xmm8,%xmm10
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,14,24,1,0 // addps 0x1180e(%rip),%xmm10 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,21,142,10,1,0 // addps 0x10a8e(%rip),%xmm10 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,144,0,1,0,0 // movups %xmm10,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49743,11 +49743,11 @@ _sk_bicubic_p3x_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,0 // movups (%rax),%xmm0
.byte 68,15,16,128,128,0,0,0 // movups 0x80(%rax),%xmm8
- .byte 15,88,5,222,23,1,0 // addps 0x117de(%rip),%xmm0 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 15,88,5,94,10,1,0 // addps 0x10a5e(%rip),%xmm0 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 69,15,89,201 // mulps %xmm9,%xmm9
- .byte 68,15,89,5,174,23,1,0 // mulps 0x117ae(%rip),%xmm8 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,5,134,19,1,0 // addps 0x11386(%rip),%xmm8 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,5,46,10,1,0 // mulps 0x10a2e(%rip),%xmm8 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,5,6,6,1,0 // addps 0x10606(%rip),%xmm8 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
.byte 68,15,17,128,0,1,0,0 // movups %xmm8,0x100(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49760,13 +49760,13 @@ _sk_bicubic_n3y_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,113,23,1,0 // addps 0x11771(%rip),%xmm1 # 3a680 <_sk_srcover_bgra_8888_sse2_lowp+0x11f8>
- .byte 68,15,40,13,89,17,1,0 // movaps 0x11159(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,241,9,1,0 // addps 0x109f1(%rip),%xmm1 # 39900 <_sk_srcover_bgra_8888_sse2_lowp+0x11fc>
+ .byte 68,15,40,13,217,3,1,0 // movaps 0x103d9(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
.byte 69,15,40,193 // movaps %xmm9,%xmm8
.byte 69,15,89,192 // mulps %xmm8,%xmm8
- .byte 68,15,89,13,101,23,1,0 // mulps 0x11765(%rip),%xmm9 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,13,61,19,1,0 // addps 0x1133d(%rip),%xmm9 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,13,229,9,1,0 // mulps 0x109e5(%rip),%xmm9 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,13,189,5,1,0 // addps 0x105bd(%rip),%xmm9 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,200 // mulps %xmm8,%xmm9
.byte 68,15,17,136,64,1,0,0 // movups %xmm9,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49779,16 +49779,16 @@ _sk_bicubic_n1y_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,24,23,1,0 // addps 0x11718(%rip),%xmm1 # 3a670 <_sk_srcover_bgra_8888_sse2_lowp+0x11e8>
- .byte 68,15,40,13,16,17,1,0 // movaps 0x11110(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,88,13,152,9,1,0 // addps 0x10998(%rip),%xmm1 # 398f0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ec>
+ .byte 68,15,40,13,144,3,1,0 // movaps 0x10390(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,92,200 // subps %xmm8,%xmm9
- .byte 68,15,40,5,52,23,1,0 // movaps 0x11734(%rip),%xmm8 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,5,180,9,1,0 // movaps 0x109b4(%rip),%xmm8 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,56,23,1,0 // addps 0x11738(%rip),%xmm8 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,5,184,9,1,0 // addps 0x109b8(%rip),%xmm8 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,220,16,1,0 // addps 0x110dc(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,88,5,92,3,1,0 // addps 0x1035c(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
- .byte 68,15,88,5,48,23,1,0 // addps 0x11730(%rip),%xmm8 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,5,176,9,1,0 // addps 0x109b0(%rip),%xmm8 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49798,17 +49798,17 @@ HIDDEN _sk_bicubic_p1y_sse2
FUNCTION(_sk_bicubic_p1y_sse2)
_sk_bicubic_p1y_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 68,15,40,5,186,16,1,0 // movaps 0x110ba(%rip),%xmm8 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,5,58,3,1,0 // movaps 0x1033a(%rip),%xmm8 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,136,192,0,0,0 // movups 0xc0(%rax),%xmm9
.byte 65,15,88,200 // addps %xmm8,%xmm1
- .byte 68,15,40,21,226,22,1,0 // movaps 0x116e2(%rip),%xmm10 # 3a6a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1218>
+ .byte 68,15,40,21,98,9,1,0 // movaps 0x10962(%rip),%xmm10 # 39920 <_sk_srcover_bgra_8888_sse2_lowp+0x121c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,230,22,1,0 // addps 0x116e6(%rip),%xmm10 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 68,15,88,21,102,9,1,0 // addps 0x10966(%rip),%xmm10 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,89,209 // mulps %xmm9,%xmm10
.byte 69,15,88,208 // addps %xmm8,%xmm10
.byte 69,15,89,209 // mulps %xmm9,%xmm10
- .byte 68,15,88,21,226,22,1,0 // addps 0x116e2(%rip),%xmm10 # 3a6c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1238>
+ .byte 68,15,88,21,98,9,1,0 // addps 0x10962(%rip),%xmm10 # 39940 <_sk_srcover_bgra_8888_sse2_lowp+0x123c>
.byte 68,15,17,144,64,1,0,0 // movups %xmm10,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -49820,11 +49820,11 @@ _sk_bicubic_p3y_sse2:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,16,72,64 // movups 0x40(%rax),%xmm1
.byte 68,15,16,128,192,0,0,0 // movups 0xc0(%rax),%xmm8
- .byte 15,88,13,177,22,1,0 // addps 0x116b1(%rip),%xmm1 # 3a6b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1228>
+ .byte 15,88,13,49,9,1,0 // addps 0x10931(%rip),%xmm1 # 39930 <_sk_srcover_bgra_8888_sse2_lowp+0x122c>
.byte 69,15,40,200 // movaps %xmm8,%xmm9
.byte 69,15,89,201 // mulps %xmm9,%xmm9
- .byte 68,15,89,5,129,22,1,0 // mulps 0x11681(%rip),%xmm8 # 3a690 <_sk_srcover_bgra_8888_sse2_lowp+0x1208>
- .byte 68,15,88,5,89,18,1,0 // addps 0x11259(%rip),%xmm8 # 3a270 <_sk_srcover_bgra_8888_sse2_lowp+0xde8>
+ .byte 68,15,89,5,1,9,1,0 // mulps 0x10901(%rip),%xmm8 # 39910 <_sk_srcover_bgra_8888_sse2_lowp+0x120c>
+ .byte 68,15,88,5,217,4,1,0 // addps 0x104d9(%rip),%xmm8 # 394f0 <_sk_srcover_bgra_8888_sse2_lowp+0xdec>
.byte 69,15,89,193 // mulps %xmm9,%xmm8
.byte 68,15,17,128,64,1,0,0 // movups %xmm8,0x140(%rax)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -49969,7 +49969,7 @@ _sk_clut_3D_sse2:
.byte 102,65,15,254,205 // paddd %xmm13,%xmm1
.byte 102,68,15,127,108,36,208 // movdqa %xmm13,-0x30(%rsp)
.byte 102,15,254,217 // paddd %xmm1,%xmm3
- .byte 102,68,15,111,5,200,20,1,0 // movdqa 0x114c8(%rip),%xmm8 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,68,15,111,5,72,7,1,0 // movdqa 0x10748(%rip),%xmm8 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,112,195,245 // pshufd $0xf5,%xmm3,%xmm0
.byte 102,65,15,244,216 // pmuludq %xmm8,%xmm3
.byte 102,65,15,244,192 // pmuludq %xmm8,%xmm0
@@ -50008,7 +50008,7 @@ _sk_clut_3D_sse2:
.byte 243,66,15,16,44,139 // movss (%rbx,%r9,4),%xmm5
.byte 15,20,236 // unpcklps %xmm4,%xmm5
.byte 102,15,20,221 // unpcklpd %xmm5,%xmm3
- .byte 102,68,15,111,37,207,13,1,0 // movdqa 0x10dcf(%rip),%xmm12 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,37,79,0,1,0 // movdqa 0x1004f(%rip),%xmm12 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,196 // paddd %xmm12,%xmm0
.byte 102,15,112,224,78 // pshufd $0x4e,%xmm0,%xmm4
.byte 102,73,15,126,224 // movq %xmm4,%r8
@@ -50024,7 +50024,7 @@ _sk_clut_3D_sse2:
.byte 243,66,15,16,36,139 // movss (%rbx,%r9,4),%xmm4
.byte 15,20,224 // unpcklps %xmm0,%xmm4
.byte 102,15,20,236 // unpcklpd %xmm4,%xmm5
- .byte 15,40,37,180,19,1,0 // movaps 0x113b4(%rip),%xmm4 # 3a6d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1248>
+ .byte 15,40,37,52,6,1,0 // movaps 0x10634(%rip),%xmm4 # 39950 <_sk_srcover_bgra_8888_sse2_lowp+0x124c>
.byte 68,15,88,220 // addps %xmm4,%xmm11
.byte 68,15,41,92,36,192 // movaps %xmm11,-0x40(%rsp)
.byte 68,15,40,223 // movaps %xmm7,%xmm11
@@ -50060,7 +50060,7 @@ _sk_clut_3D_sse2:
.byte 15,20,249 // unpcklps %xmm1,%xmm7
.byte 102,15,20,252 // unpcklpd %xmm4,%xmm7
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
- .byte 102,15,250,13,41,19,1,0 // psubd 0x11329(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,169,5,1,0 // psubd 0x105a9(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,225,78 // pshufd $0x4e,%xmm1,%xmm4
.byte 102,73,15,126,224 // movq %xmm4,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -50150,7 +50150,7 @@ _sk_clut_3D_sse2:
.byte 243,15,16,44,131 // movss (%rbx,%rax,4),%xmm5
.byte 15,20,233 // unpcklps %xmm1,%xmm5
.byte 102,15,20,221 // unpcklpd %xmm5,%xmm3
- .byte 102,68,15,111,45,65,11,1,0 // movdqa 0x10b41(%rip),%xmm13 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,45,193,253,0,0 // movdqa 0xfdc1(%rip),%xmm13 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,197 // paddd %xmm13,%xmm0
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
.byte 102,73,15,126,200 // movq %xmm1,%r8
@@ -50247,7 +50247,7 @@ _sk_clut_3D_sse2:
.byte 102,15,111,206 // movdqa %xmm6,%xmm1
.byte 102,65,15,254,202 // paddd %xmm10,%xmm1
.byte 102,15,112,209,245 // pshufd $0xf5,%xmm1,%xmm2
- .byte 102,15,111,29,196,15,1,0 // movdqa 0x10fc4(%rip),%xmm3 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,29,68,2,1,0 // movdqa 0x10244(%rip),%xmm3 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,203 // pmuludq %xmm3,%xmm1
.byte 102,15,244,211 // pmuludq %xmm3,%xmm2
.byte 102,15,111,251 // movdqa %xmm3,%xmm7
@@ -50285,7 +50285,7 @@ _sk_clut_3D_sse2:
.byte 243,15,16,44,131 // movss (%rbx,%rax,4),%xmm5
.byte 15,20,236 // unpcklps %xmm4,%xmm5
.byte 102,15,20,213 // unpcklpd %xmm5,%xmm2
- .byte 102,15,111,5,212,8,1,0 // movdqa 0x108d4(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,111,5,84,251,0,0 // movdqa 0xfb54(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,254,216 // paddd %xmm0,%xmm3
.byte 102,15,112,227,78 // pshufd $0x4e,%xmm3,%xmm4
.byte 102,73,15,126,224 // movq %xmm4,%r8
@@ -50366,7 +50366,7 @@ _sk_clut_3D_sse2:
.byte 102,15,111,214 // movdqa %xmm6,%xmm2
.byte 102,65,15,254,212 // paddd %xmm12,%xmm2
.byte 102,15,112,194,245 // pshufd $0xf5,%xmm2,%xmm0
- .byte 102,15,111,13,161,13,1,0 // movdqa 0x10da1(%rip),%xmm1 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,13,33,0,1,0 // movdqa 0x10021(%rip),%xmm1 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,209 // pmuludq %xmm1,%xmm2
.byte 102,15,244,193 // pmuludq %xmm1,%xmm0
.byte 102,15,111,241 // movdqa %xmm1,%xmm6
@@ -50388,7 +50388,7 @@ _sk_clut_3D_sse2:
.byte 68,15,20,209 // unpcklps %xmm1,%xmm10
.byte 102,68,15,20,210 // unpcklpd %xmm2,%xmm10
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
- .byte 102,15,250,13,70,13,1,0 // psubd 0x10d46(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,198,255,0,0 // psubd 0xffc6(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,209,78 // pshufd $0x4e,%xmm1,%xmm2
.byte 102,73,15,126,208 // movq %xmm2,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -50403,7 +50403,7 @@ _sk_clut_3D_sse2:
.byte 243,15,16,20,131 // movss (%rbx,%rax,4),%xmm2
.byte 15,20,209 // unpcklps %xmm1,%xmm2
.byte 102,15,20,226 // unpcklpd %xmm2,%xmm4
- .byte 102,15,254,5,177,6,1,0 // paddd 0x106b1(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,5,49,249,0,0 // paddd 0xf931(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
.byte 102,73,15,126,200 // movq %xmm1,%r8
.byte 102,72,15,126,192 // movq %xmm0,%rax
@@ -50440,7 +50440,7 @@ _sk_clut_3D_sse2:
.byte 15,20,199 // unpcklps %xmm7,%xmm0
.byte 102,15,20,193 // unpcklpd %xmm1,%xmm0
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
- .byte 102,15,250,13,80,12,1,0 // psubd 0x10c50(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,208,254,0,0 // psubd 0xfed0(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,249,78 // pshufd $0x4e,%xmm1,%xmm7
.byte 102,73,15,126,248 // movq %xmm7,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -50455,7 +50455,7 @@ _sk_clut_3D_sse2:
.byte 243,15,16,52,131 // movss (%rbx,%rax,4),%xmm6
.byte 15,20,247 // unpcklps %xmm7,%xmm6
.byte 102,15,20,206 // unpcklpd %xmm6,%xmm1
- .byte 102,15,254,21,187,5,1,0 // paddd 0x105bb(%rip),%xmm2 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,21,59,248,0,0 // paddd 0xf83b(%rip),%xmm2 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,242,78 // pshufd $0x4e,%xmm2,%xmm6
.byte 102,73,15,126,240 // movq %xmm6,%r8
.byte 102,72,15,126,208 // movq %xmm2,%rax
@@ -50601,7 +50601,7 @@ _sk_clut_4D_sse2:
.byte 102,65,15,254,210 // paddd %xmm10,%xmm2
.byte 102,15,111,220 // movdqa %xmm4,%xmm3
.byte 102,15,254,218 // paddd %xmm2,%xmm3
- .byte 102,15,111,37,147,9,1,0 // movdqa 0x10993(%rip),%xmm4 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,37,19,252,0,0 // movdqa 0xfc13(%rip),%xmm4 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,112,195,245 // pshufd $0xf5,%xmm3,%xmm0
.byte 102,15,244,220 // pmuludq %xmm4,%xmm3
.byte 102,15,244,196 // pmuludq %xmm4,%xmm0
@@ -50640,7 +50640,7 @@ _sk_clut_4D_sse2:
.byte 243,66,15,16,52,139 // movss (%rbx,%r9,4),%xmm6
.byte 15,20,244 // unpcklps %xmm4,%xmm6
.byte 102,15,20,238 // unpcklpd %xmm6,%xmm5
- .byte 102,15,254,5,157,2,1,0 // paddd 0x1029d(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,5,29,245,0,0 // paddd 0xf51d(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,224,78 // pshufd $0x4e,%xmm0,%xmm4
.byte 102,73,15,126,224 // movq %xmm4,%r8
.byte 102,73,15,126,193 // movq %xmm0,%r9
@@ -50655,7 +50655,7 @@ _sk_clut_4D_sse2:
.byte 243,66,15,16,52,139 // movss (%rbx,%r9,4),%xmm6
.byte 15,20,244 // unpcklps %xmm4,%xmm6
.byte 102,15,20,222 // unpcklpd %xmm6,%xmm3
- .byte 15,40,53,135,8,1,0 // movaps 0x10887(%rip),%xmm6 # 3a6d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1248>
+ .byte 15,40,53,7,251,0,0 // movaps 0xfb07(%rip),%xmm6 # 39950 <_sk_srcover_bgra_8888_sse2_lowp+0x124c>
.byte 68,15,88,198 // addps %xmm6,%xmm8
.byte 68,15,41,68,36,176 // movaps %xmm8,-0x50(%rsp)
.byte 68,15,88,254 // addps %xmm6,%xmm15
@@ -50673,7 +50673,7 @@ _sk_clut_4D_sse2:
.byte 102,65,15,254,215 // paddd %xmm15,%xmm2
.byte 102,68,15,127,124,36,224 // movdqa %xmm15,-0x20(%rsp)
.byte 102,15,112,194,245 // pshufd $0xf5,%xmm2,%xmm0
- .byte 102,15,111,13,61,8,1,0 // movdqa 0x1083d(%rip),%xmm1 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,13,189,250,0,0 // movdqa 0xfabd(%rip),%xmm1 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,209 // pmuludq %xmm1,%xmm2
.byte 102,15,244,193 // pmuludq %xmm1,%xmm0
.byte 102,15,112,192,232 // pshufd $0xe8,%xmm0,%xmm0
@@ -50694,7 +50694,7 @@ _sk_clut_4D_sse2:
.byte 15,20,240 // unpcklps %xmm0,%xmm6
.byte 102,15,20,242 // unpcklpd %xmm2,%xmm6
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
- .byte 102,15,250,5,232,7,1,0 // psubd 0x107e8(%rip),%xmm0 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,5,104,250,0,0 // psubd 0xfa68(%rip),%xmm0 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,208,78 // pshufd $0x4e,%xmm0,%xmm2
.byte 102,73,15,126,208 // movq %xmm2,%r8
.byte 102,72,15,126,192 // movq %xmm0,%rax
@@ -50709,7 +50709,7 @@ _sk_clut_4D_sse2:
.byte 243,68,15,16,28,131 // movss (%rbx,%rax,4),%xmm11
.byte 68,15,20,216 // unpcklps %xmm0,%xmm11
.byte 102,65,15,20,211 // unpcklpd %xmm11,%xmm2
- .byte 102,15,254,13,80,1,1,0 // paddd 0x10150(%rip),%xmm1 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,13,208,243,0,0 // paddd 0xf3d0(%rip),%xmm1 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,193,78 // pshufd $0x4e,%xmm1,%xmm0
.byte 102,73,15,126,192 // movq %xmm0,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -50747,7 +50747,7 @@ _sk_clut_4D_sse2:
.byte 102,65,15,111,194 // movdqa %xmm10,%xmm0
.byte 102,15,254,68,36,160 // paddd -0x60(%rsp),%xmm0
.byte 102,15,112,216,245 // pshufd $0xf5,%xmm0,%xmm3
- .byte 102,15,111,37,231,6,1,0 // movdqa 0x106e7(%rip),%xmm4 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,37,103,249,0,0 // movdqa 0xf967(%rip),%xmm4 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,196 // pmuludq %xmm4,%xmm0
.byte 102,15,244,220 // pmuludq %xmm4,%xmm3
.byte 102,68,15,111,220 // movdqa %xmm4,%xmm11
@@ -50785,7 +50785,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,60,131 // movss (%rbx,%rax,4),%xmm7
.byte 15,20,253 // unpcklps %xmm5,%xmm7
.byte 102,15,20,199 // unpcklpd %xmm7,%xmm0
- .byte 102,68,15,111,5,243,255,0,0 // movdqa 0xfff3(%rip),%xmm8 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,5,115,242,0,0 // movdqa 0xf273(%rip),%xmm8 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,224 // paddd %xmm8,%xmm4
.byte 102,15,112,236,78 // pshufd $0x4e,%xmm4,%xmm5
.byte 102,73,15,126,232 // movq %xmm5,%r8
@@ -50931,7 +50931,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,36,131 // movss (%rbx,%rax,4),%xmm4
.byte 15,20,226 // unpcklps %xmm2,%xmm4
.byte 102,15,20,252 // unpcklpd %xmm4,%xmm7
- .byte 102,68,15,111,61,77,253,0,0 // movdqa 0xfd4d(%rip),%xmm15 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,61,205,239,0,0 // movdqa 0xefcd(%rip),%xmm15 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,207 // paddd %xmm15,%xmm1
.byte 102,15,112,209,78 // pshufd $0x4e,%xmm1,%xmm2
.byte 102,73,15,126,208 // movq %xmm2,%r8
@@ -51014,7 +51014,7 @@ _sk_clut_4D_sse2:
.byte 102,65,15,111,217 // movdqa %xmm9,%xmm3
.byte 102,65,15,254,216 // paddd %xmm8,%xmm3
.byte 102,15,112,211,245 // pshufd $0xf5,%xmm3,%xmm2
- .byte 102,15,111,45,28,2,1,0 // movdqa 0x1021c(%rip),%xmm5 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,45,156,244,0,0 // movdqa 0xf49c(%rip),%xmm5 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,221 // pmuludq %xmm5,%xmm3
.byte 102,15,244,213 // pmuludq %xmm5,%xmm2
.byte 102,15,112,234,232 // pshufd $0xe8,%xmm2,%xmm5
@@ -51035,7 +51035,7 @@ _sk_clut_4D_sse2:
.byte 68,15,20,198 // unpcklps %xmm6,%xmm8
.byte 102,68,15,20,197 // unpcklpd %xmm5,%xmm8
.byte 102,15,111,234 // movdqa %xmm2,%xmm5
- .byte 102,15,250,45,197,1,1,0 // psubd 0x101c5(%rip),%xmm5 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,45,69,244,0,0 // psubd 0xf445(%rip),%xmm5 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,245,78 // pshufd $0x4e,%xmm5,%xmm6
.byte 102,73,15,126,240 // movq %xmm6,%r8
.byte 102,72,15,126,232 // movq %xmm5,%rax
@@ -51067,7 +51067,7 @@ _sk_clut_4D_sse2:
.byte 102,15,20,214 // unpcklpd %xmm6,%xmm2
.byte 102,69,15,254,205 // paddd %xmm13,%xmm9
.byte 102,65,15,112,233,245 // pshufd $0xf5,%xmm9,%xmm5
- .byte 102,15,111,29,35,1,1,0 // movdqa 0x10123(%rip),%xmm3 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,29,163,243,0,0 // movdqa 0xf3a3(%rip),%xmm3 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,68,15,244,203 // pmuludq %xmm3,%xmm9
.byte 102,15,244,235 // pmuludq %xmm3,%xmm5
.byte 102,15,112,237,232 // pshufd $0xe8,%xmm5,%xmm5
@@ -51088,7 +51088,7 @@ _sk_clut_4D_sse2:
.byte 69,15,20,233 // unpcklps %xmm9,%xmm13
.byte 102,68,15,20,237 // unpcklpd %xmm5,%xmm13
.byte 102,65,15,111,239 // movdqa %xmm15,%xmm5
- .byte 102,15,250,45,195,0,1,0 // psubd 0x100c3(%rip),%xmm5 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,45,67,243,0,0 // psubd 0xf343(%rip),%xmm5 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,221,78 // pshufd $0x4e,%xmm5,%xmm3
.byte 102,73,15,126,216 // movq %xmm3,%r8
.byte 102,72,15,126,232 // movq %xmm5,%rax
@@ -51103,7 +51103,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,44,131 // movss (%rbx,%rax,4),%xmm5
.byte 15,20,235 // unpcklps %xmm3,%xmm5
.byte 102,15,20,245 // unpcklpd %xmm5,%xmm6
- .byte 102,68,15,111,13,45,250,0,0 // movdqa 0xfa2d(%rip),%xmm9 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,13,173,236,0,0 // movdqa 0xecad(%rip),%xmm9 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,69,15,254,249 // paddd %xmm9,%xmm15
.byte 102,65,15,112,223,78 // pshufd $0x4e,%xmm15,%xmm3
.byte 102,73,15,126,216 // movq %xmm3,%r8
@@ -51165,7 +51165,7 @@ _sk_clut_4D_sse2:
.byte 102,65,15,111,195 // movdqa %xmm11,%xmm0
.byte 102,15,254,195 // paddd %xmm3,%xmm0
.byte 102,15,112,200,245 // pshufd $0xf5,%xmm0,%xmm1
- .byte 102,15,111,21,94,255,0,0 // movdqa 0xff5e(%rip),%xmm2 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,21,222,241,0,0 // movdqa 0xf1de(%rip),%xmm2 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,194 // pmuludq %xmm2,%xmm0
.byte 102,15,244,202 // pmuludq %xmm2,%xmm1
.byte 102,68,15,111,234 // movdqa %xmm2,%xmm13
@@ -51358,7 +51358,7 @@ _sk_clut_4D_sse2:
.byte 68,15,20,238 // unpcklps %xmm6,%xmm13
.byte 102,68,15,20,234 // unpcklpd %xmm2,%xmm13
.byte 102,15,111,212 // movdqa %xmm4,%xmm2
- .byte 102,15,250,21,250,251,0,0 // psubd 0xfbfa(%rip),%xmm2 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,21,122,238,0,0 // psubd 0xee7a(%rip),%xmm2 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,242,78 // pshufd $0x4e,%xmm2,%xmm6
.byte 102,73,15,126,240 // movq %xmm6,%r8
.byte 102,72,15,126,208 // movq %xmm2,%rax
@@ -51418,7 +51418,7 @@ _sk_clut_4D_sse2:
.byte 102,68,15,111,116,36,160 // movdqa -0x60(%rsp),%xmm14
.byte 102,65,15,254,198 // paddd %xmm14,%xmm0
.byte 102,15,112,200,245 // pshufd $0xf5,%xmm0,%xmm1
- .byte 102,68,15,111,5,217,250,0,0 // movdqa 0xfad9(%rip),%xmm8 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,68,15,111,5,89,237,0,0 // movdqa 0xed59(%rip),%xmm8 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,65,15,244,192 // pmuludq %xmm8,%xmm0
.byte 102,65,15,244,200 // pmuludq %xmm8,%xmm1
.byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
@@ -51456,7 +51456,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,44,131 // movss (%rbx,%rax,4),%xmm5
.byte 15,20,235 // unpcklps %xmm3,%xmm5
.byte 102,15,20,205 // unpcklpd %xmm5,%xmm1
- .byte 102,68,15,111,5,231,243,0,0 // movdqa 0xf3e7(%rip),%xmm8 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,68,15,111,5,103,230,0,0 // movdqa 0xe667(%rip),%xmm8 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,65,15,254,208 // paddd %xmm8,%xmm2
.byte 102,15,112,218,78 // pshufd $0x4e,%xmm2,%xmm3
.byte 102,73,15,126,216 // movq %xmm3,%r8
@@ -51539,7 +51539,7 @@ _sk_clut_4D_sse2:
.byte 102,15,254,208 // paddd %xmm0,%xmm2
.byte 102,15,111,240 // movdqa %xmm0,%xmm6
.byte 102,15,112,194,245 // pshufd $0xf5,%xmm2,%xmm0
- .byte 102,15,111,37,179,248,0,0 // movdqa 0xf8b3(%rip),%xmm4 # 3a6e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1258>
+ .byte 102,15,111,37,51,235,0,0 // movdqa 0xeb33(%rip),%xmm4 # 39960 <_sk_srcover_bgra_8888_sse2_lowp+0x125c>
.byte 102,15,244,212 // pmuludq %xmm4,%xmm2
.byte 102,15,244,196 // pmuludq %xmm4,%xmm0
.byte 102,15,112,200,232 // pshufd $0xe8,%xmm0,%xmm1
@@ -51560,7 +51560,7 @@ _sk_clut_4D_sse2:
.byte 68,15,20,193 // unpcklps %xmm1,%xmm8
.byte 102,68,15,20,194 // unpcklpd %xmm2,%xmm8
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
- .byte 102,15,250,13,92,248,0,0 // psubd 0xf85c(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,220,234,0,0 // psubd 0xeadc(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,209,78 // pshufd $0x4e,%xmm1,%xmm2
.byte 102,73,15,126,208 // movq %xmm2,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -51575,7 +51575,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,20,131 // movss (%rbx,%rax,4),%xmm2
.byte 15,20,209 // unpcklps %xmm1,%xmm2
.byte 102,68,15,20,242 // unpcklpd %xmm2,%xmm14
- .byte 102,15,254,5,197,241,0,0 // paddd 0xf1c5(%rip),%xmm0 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,5,69,228,0,0 // paddd 0xe445(%rip),%xmm0 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
.byte 102,73,15,126,200 // movq %xmm1,%r8
.byte 102,72,15,126,192 // movq %xmm0,%rax
@@ -51613,7 +51613,7 @@ _sk_clut_4D_sse2:
.byte 15,20,198 // unpcklps %xmm6,%xmm0
.byte 102,15,20,193 // unpcklpd %xmm1,%xmm0
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
- .byte 102,15,250,13,100,247,0,0 // psubd 0xf764(%rip),%xmm1 # 3a6f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1268>
+ .byte 102,15,250,13,228,233,0,0 // psubd 0xe9e4(%rip),%xmm1 # 39970 <_sk_srcover_bgra_8888_sse2_lowp+0x126c>
.byte 102,15,112,241,78 // pshufd $0x4e,%xmm1,%xmm6
.byte 102,73,15,126,240 // movq %xmm6,%r8
.byte 102,72,15,126,200 // movq %xmm1,%rax
@@ -51628,7 +51628,7 @@ _sk_clut_4D_sse2:
.byte 243,15,16,36,131 // movss (%rbx,%rax,4),%xmm4
.byte 15,20,230 // unpcklps %xmm6,%xmm4
.byte 102,15,20,204 // unpcklpd %xmm4,%xmm1
- .byte 102,15,254,21,207,240,0,0 // paddd 0xf0cf(%rip),%xmm2 # 3a0a0 <_sk_srcover_bgra_8888_sse2_lowp+0xc18>
+ .byte 102,15,254,21,79,227,0,0 // paddd 0xe34f(%rip),%xmm2 # 39320 <_sk_srcover_bgra_8888_sse2_lowp+0xc1c>
.byte 102,15,112,226,78 // pshufd $0x4e,%xmm2,%xmm4
.byte 102,73,15,126,224 // movq %xmm4,%r8
.byte 102,72,15,126,208 // movq %xmm2,%rax
@@ -51687,7 +51687,7 @@ _sk_clut_4D_sse2:
.byte 15,89,212 // mulps %xmm4,%xmm2
.byte 65,15,88,215 // addps %xmm15,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,187,239,0,0 // movaps 0xefbb(%rip),%xmm3 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,29,59,226,0,0 // movaps 0xe23b(%rip),%xmm3 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,100,36,96 // movaps 0x60(%rsp),%xmm4
.byte 15,40,108,36,112 // movaps 0x70(%rsp),%xmm5
.byte 15,40,180,36,128,0,0,0 // movaps 0x80(%rsp),%xmm6
@@ -51700,15 +51700,15 @@ HIDDEN _sk_gauss_a_to_rgba_sse2
.globl _sk_gauss_a_to_rgba_sse2
FUNCTION(_sk_gauss_a_to_rgba_sse2)
_sk_gauss_a_to_rgba_sse2:
- .byte 15,40,5,32,246,0,0 // movaps 0xf620(%rip),%xmm0 # 3a700 <_sk_srcover_bgra_8888_sse2_lowp+0x1278>
+ .byte 15,40,5,160,232,0,0 // movaps 0xe8a0(%rip),%xmm0 # 39980 <_sk_srcover_bgra_8888_sse2_lowp+0x127c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,38,246,0,0 // addps 0xf626(%rip),%xmm0 # 3a710 <_sk_srcover_bgra_8888_sse2_lowp+0x1288>
+ .byte 15,88,5,166,232,0,0 // addps 0xe8a6(%rip),%xmm0 # 39990 <_sk_srcover_bgra_8888_sse2_lowp+0x128c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,44,246,0,0 // addps 0xf62c(%rip),%xmm0 # 3a720 <_sk_srcover_bgra_8888_sse2_lowp+0x1298>
+ .byte 15,88,5,172,232,0,0 // addps 0xe8ac(%rip),%xmm0 # 399a0 <_sk_srcover_bgra_8888_sse2_lowp+0x129c>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,50,246,0,0 // addps 0xf632(%rip),%xmm0 # 3a730 <_sk_srcover_bgra_8888_sse2_lowp+0x12a8>
+ .byte 15,88,5,178,232,0,0 // addps 0xe8b2(%rip),%xmm0 # 399b0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ac>
.byte 15,89,195 // mulps %xmm3,%xmm0
- .byte 15,88,5,56,246,0,0 // addps 0xf638(%rip),%xmm0 # 3a740 <_sk_srcover_bgra_8888_sse2_lowp+0x12b8>
+ .byte 15,88,5,184,232,0,0 // addps 0xe8b8(%rip),%xmm0 # 399c0 <_sk_srcover_bgra_8888_sse2_lowp+0x12bc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
@@ -51811,7 +51811,7 @@ _sk_seed_shader_hsw_lowp:
.byte 197,249,110,209 // vmovd %ecx,%xmm2
.byte 196,226,125,88,210 // vpbroadcastd %xmm2,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
- .byte 196,226,125,24,29,17,229,0,0 // vbroadcastss 0xe511(%rip),%ymm3 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,29,141,215,0,0 // vbroadcastss 0xd78d(%rip),%ymm3 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,236,88,211 // vaddps %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,252,40,218 // vmovaps %ymm2,%ymm3
@@ -51927,7 +51927,7 @@ HIDDEN _sk_black_color_hsw_lowp
FUNCTION(_sk_black_color_hsw_lowp)
_sk_black_color_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,248,233,0,0 // vpbroadcastw 0xe9f8(%rip),%ymm3 # 39da0 <_sk_srcover_bgra_8888_sse2_lowp+0x918>
+ .byte 196,226,125,121,29,120,220,0,0 // vpbroadcastw 0xdc78(%rip),%ymm3 # 39020 <_sk_srcover_bgra_8888_sse2_lowp+0x91c>
.byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
.byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
.byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
@@ -51938,7 +51938,7 @@ HIDDEN _sk_white_color_hsw_lowp
FUNCTION(_sk_white_color_hsw_lowp)
_sk_white_color_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,5,225,233,0,0 // vpbroadcastw 0xe9e1(%rip),%ymm0 # 39da2 <_sk_srcover_bgra_8888_sse2_lowp+0x91a>
+ .byte 196,226,125,121,5,97,220,0,0 // vpbroadcastw 0xdc61(%rip),%ymm0 # 39022 <_sk_srcover_bgra_8888_sse2_lowp+0x91e>
.byte 197,253,111,200 // vmovdqa %ymm0,%ymm1
.byte 197,253,111,208 // vmovdqa %ymm0,%ymm2
.byte 197,253,111,216 // vmovdqa %ymm0,%ymm3
@@ -51949,8 +51949,8 @@ HIDDEN _sk_set_rgb_hsw_lowp
FUNCTION(_sk_set_rgb_hsw_lowp)
_sk_set_rgb_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,122,16,5,91,227,0,0 // vmovss 0xe35b(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 197,250,16,21,131,227,0,0 // vmovss 0xe383(%rip),%xmm2 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 197,122,16,5,215,213,0,0 // vmovss 0xd5d7(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,250,16,21,255,213,0,0 // vmovss 0xd5ff(%rip),%xmm2 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 197,250,16,0 // vmovss (%rax),%xmm0
.byte 196,226,57,153,194 // vfmadd132ss %xmm2,%xmm8,%xmm0
.byte 197,122,44,192 // vcvttss2si %xmm0,%r8d
@@ -51993,7 +51993,7 @@ HIDDEN _sk_premul_hsw_lowp
FUNCTION(_sk_premul_hsw_lowp)
_sk_premul_hsw_lowp:
.byte 197,229,213,192 // vpmullw %ymm0,%ymm3,%ymm0
- .byte 196,98,125,121,5,74,233,0,0 // vpbroadcastw 0xe94a(%rip),%ymm8 # 39da4 <_sk_srcover_bgra_8888_sse2_lowp+0x91c>
+ .byte 196,98,125,121,5,202,219,0,0 // vpbroadcastw 0xdbca(%rip),%ymm8 # 39024 <_sk_srcover_bgra_8888_sse2_lowp+0x920>
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,229,213,201 // vpmullw %ymm1,%ymm3,%ymm1
@@ -52010,7 +52010,7 @@ HIDDEN _sk_premul_dst_hsw_lowp
FUNCTION(_sk_premul_dst_hsw_lowp)
_sk_premul_dst_hsw_lowp:
.byte 197,197,213,228 // vpmullw %ymm4,%ymm7,%ymm4
- .byte 196,98,125,121,5,21,233,0,0 // vpbroadcastw 0xe915(%rip),%ymm8 # 39da6 <_sk_srcover_bgra_8888_sse2_lowp+0x91e>
+ .byte 196,98,125,121,5,149,219,0,0 // vpbroadcastw 0xdb95(%rip),%ymm8 # 39026 <_sk_srcover_bgra_8888_sse2_lowp+0x922>
.byte 196,193,93,253,224 // vpaddw %ymm8,%ymm4,%ymm4
.byte 197,221,113,212,8 // vpsrlw $0x8,%ymm4,%ymm4
.byte 197,197,213,237 // vpmullw %ymm5,%ymm7,%ymm5
@@ -52058,7 +52058,7 @@ HIDDEN _sk_invert_hsw_lowp
.globl _sk_invert_hsw_lowp
FUNCTION(_sk_invert_hsw_lowp)
_sk_invert_hsw_lowp:
- .byte 196,98,125,121,5,172,232,0,0 // vpbroadcastw 0xe8ac(%rip),%ymm8 # 39da8 <_sk_srcover_bgra_8888_sse2_lowp+0x920>
+ .byte 196,98,125,121,5,44,219,0,0 // vpbroadcastw 0xdb2c(%rip),%ymm8 # 39028 <_sk_srcover_bgra_8888_sse2_lowp+0x924>
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
.byte 197,189,249,201 // vpsubw %ymm1,%ymm8,%ymm1
.byte 197,189,249,210 // vpsubw %ymm2,%ymm8,%ymm2
@@ -52082,7 +52082,7 @@ HIDDEN _sk_srcatop_hsw_lowp
FUNCTION(_sk_srcatop_hsw_lowp)
_sk_srcatop_hsw_lowp:
.byte 197,197,213,192 // vpmullw %ymm0,%ymm7,%ymm0
- .byte 196,98,125,121,5,121,232,0,0 // vpbroadcastw 0xe879(%rip),%ymm8 # 39daa <_sk_srcover_bgra_8888_sse2_lowp+0x922>
+ .byte 196,98,125,121,5,249,218,0,0 // vpbroadcastw 0xdaf9(%rip),%ymm8 # 3902a <_sk_srcover_bgra_8888_sse2_lowp+0x926>
.byte 197,189,249,219 // vpsubw %ymm3,%ymm8,%ymm3
.byte 197,101,213,204 // vpmullw %ymm4,%ymm3,%ymm9
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
@@ -52109,7 +52109,7 @@ HIDDEN _sk_dstatop_hsw_lowp
FUNCTION(_sk_dstatop_hsw_lowp)
_sk_dstatop_hsw_lowp:
.byte 197,93,213,195 // vpmullw %ymm3,%ymm4,%ymm8
- .byte 196,98,125,121,13,25,232,0,0 // vpbroadcastw 0xe819(%rip),%ymm9 # 39dac <_sk_srcover_bgra_8888_sse2_lowp+0x924>
+ .byte 196,98,125,121,13,153,218,0,0 // vpbroadcastw 0xda99(%rip),%ymm9 # 3902c <_sk_srcover_bgra_8888_sse2_lowp+0x928>
.byte 197,53,249,215 // vpsubw %ymm7,%ymm9,%ymm10
.byte 197,173,213,192 // vpmullw %ymm0,%ymm10,%ymm0
.byte 196,65,61,253,193 // vpaddw %ymm9,%ymm8,%ymm8
@@ -52136,7 +52136,7 @@ HIDDEN _sk_srcin_hsw_lowp
FUNCTION(_sk_srcin_hsw_lowp)
_sk_srcin_hsw_lowp:
.byte 197,197,213,192 // vpmullw %ymm0,%ymm7,%ymm0
- .byte 196,98,125,121,5,182,231,0,0 // vpbroadcastw 0xe7b6(%rip),%ymm8 # 39dae <_sk_srcover_bgra_8888_sse2_lowp+0x926>
+ .byte 196,98,125,121,5,54,218,0,0 // vpbroadcastw 0xda36(%rip),%ymm8 # 3902e <_sk_srcover_bgra_8888_sse2_lowp+0x92a>
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,197,213,201 // vpmullw %ymm1,%ymm7,%ymm1
@@ -52156,7 +52156,7 @@ HIDDEN _sk_dstin_hsw_lowp
FUNCTION(_sk_dstin_hsw_lowp)
_sk_dstin_hsw_lowp:
.byte 197,221,213,195 // vpmullw %ymm3,%ymm4,%ymm0
- .byte 196,98,125,121,5,115,231,0,0 // vpbroadcastw 0xe773(%rip),%ymm8 # 39db0 <_sk_srcover_bgra_8888_sse2_lowp+0x928>
+ .byte 196,98,125,121,5,243,217,0,0 // vpbroadcastw 0xd9f3(%rip),%ymm8 # 39030 <_sk_srcover_bgra_8888_sse2_lowp+0x92c>
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,213,213,203 // vpmullw %ymm3,%ymm5,%ymm1
@@ -52175,7 +52175,7 @@ HIDDEN _sk_srcout_hsw_lowp
.globl _sk_srcout_hsw_lowp
FUNCTION(_sk_srcout_hsw_lowp)
_sk_srcout_hsw_lowp:
- .byte 196,98,125,121,5,52,231,0,0 // vpbroadcastw 0xe734(%rip),%ymm8 # 39db2 <_sk_srcover_bgra_8888_sse2_lowp+0x92a>
+ .byte 196,98,125,121,5,180,217,0,0 // vpbroadcastw 0xd9b4(%rip),%ymm8 # 39032 <_sk_srcover_bgra_8888_sse2_lowp+0x92e>
.byte 197,61,249,207 // vpsubw %ymm7,%ymm8,%ymm9
.byte 197,181,213,192 // vpmullw %ymm0,%ymm9,%ymm0
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
@@ -52196,7 +52196,7 @@ HIDDEN _sk_dstout_hsw_lowp
.globl _sk_dstout_hsw_lowp
FUNCTION(_sk_dstout_hsw_lowp)
_sk_dstout_hsw_lowp:
- .byte 196,98,125,121,5,237,230,0,0 // vpbroadcastw 0xe6ed(%rip),%ymm8 # 39db4 <_sk_srcover_bgra_8888_sse2_lowp+0x92c>
+ .byte 196,98,125,121,5,109,217,0,0 // vpbroadcastw 0xd96d(%rip),%ymm8 # 39034 <_sk_srcover_bgra_8888_sse2_lowp+0x930>
.byte 197,189,249,219 // vpsubw %ymm3,%ymm8,%ymm3
.byte 197,229,213,196 // vpmullw %ymm4,%ymm3,%ymm0
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
@@ -52217,7 +52217,7 @@ HIDDEN _sk_srcover_hsw_lowp
.globl _sk_srcover_hsw_lowp
FUNCTION(_sk_srcover_hsw_lowp)
_sk_srcover_hsw_lowp:
- .byte 196,98,125,121,5,166,230,0,0 // vpbroadcastw 0xe6a6(%rip),%ymm8 # 39db6 <_sk_srcover_bgra_8888_sse2_lowp+0x92e>
+ .byte 196,98,125,121,5,38,217,0,0 // vpbroadcastw 0xd926(%rip),%ymm8 # 39036 <_sk_srcover_bgra_8888_sse2_lowp+0x932>
.byte 197,61,249,203 // vpsubw %ymm3,%ymm8,%ymm9
.byte 197,53,213,212 // vpmullw %ymm4,%ymm9,%ymm10
.byte 196,65,45,253,208 // vpaddw %ymm8,%ymm10,%ymm10
@@ -52242,7 +52242,7 @@ HIDDEN _sk_dstover_hsw_lowp
.globl _sk_dstover_hsw_lowp
FUNCTION(_sk_dstover_hsw_lowp)
_sk_dstover_hsw_lowp:
- .byte 196,98,125,121,5,75,230,0,0 // vpbroadcastw 0xe64b(%rip),%ymm8 # 39db8 <_sk_srcover_bgra_8888_sse2_lowp+0x930>
+ .byte 196,98,125,121,5,203,216,0,0 // vpbroadcastw 0xd8cb(%rip),%ymm8 # 39038 <_sk_srcover_bgra_8888_sse2_lowp+0x934>
.byte 197,61,249,207 // vpsubw %ymm7,%ymm8,%ymm9
.byte 197,181,213,192 // vpmullw %ymm0,%ymm9,%ymm0
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
@@ -52268,7 +52268,7 @@ HIDDEN _sk_modulate_hsw_lowp
FUNCTION(_sk_modulate_hsw_lowp)
_sk_modulate_hsw_lowp:
.byte 197,221,213,192 // vpmullw %ymm0,%ymm4,%ymm0
- .byte 196,98,125,121,5,240,229,0,0 // vpbroadcastw 0xe5f0(%rip),%ymm8 # 39dba <_sk_srcover_bgra_8888_sse2_lowp+0x932>
+ .byte 196,98,125,121,5,112,216,0,0 // vpbroadcastw 0xd870(%rip),%ymm8 # 3903a <_sk_srcover_bgra_8888_sse2_lowp+0x936>
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,213,213,201 // vpmullw %ymm1,%ymm5,%ymm1
@@ -52287,7 +52287,7 @@ HIDDEN _sk_multiply_hsw_lowp
.globl _sk_multiply_hsw_lowp
FUNCTION(_sk_multiply_hsw_lowp)
_sk_multiply_hsw_lowp:
- .byte 196,98,125,121,5,177,229,0,0 // vpbroadcastw 0xe5b1(%rip),%ymm8 # 39dbc <_sk_srcover_bgra_8888_sse2_lowp+0x934>
+ .byte 196,98,125,121,5,49,216,0,0 // vpbroadcastw 0xd831(%rip),%ymm8 # 3903c <_sk_srcover_bgra_8888_sse2_lowp+0x938>
.byte 197,61,249,203 // vpsubw %ymm3,%ymm8,%ymm9
.byte 197,53,213,212 // vpmullw %ymm4,%ymm9,%ymm10
.byte 196,65,93,253,216 // vpaddw %ymm8,%ymm4,%ymm11
@@ -52323,7 +52323,7 @@ HIDDEN _sk_plus__hsw_lowp
FUNCTION(_sk_plus__hsw_lowp)
_sk_plus__hsw_lowp:
.byte 197,221,253,192 // vpaddw %ymm0,%ymm4,%ymm0
- .byte 196,98,125,121,5,39,229,0,0 // vpbroadcastw 0xe527(%rip),%ymm8 # 39dbe <_sk_srcover_bgra_8888_sse2_lowp+0x936>
+ .byte 196,98,125,121,5,167,215,0,0 // vpbroadcastw 0xd7a7(%rip),%ymm8 # 3903e <_sk_srcover_bgra_8888_sse2_lowp+0x93a>
.byte 196,194,125,58,192 // vpminuw %ymm8,%ymm0,%ymm0
.byte 197,213,253,201 // vpaddw %ymm1,%ymm5,%ymm1
.byte 196,194,117,58,200 // vpminuw %ymm8,%ymm1,%ymm1
@@ -52340,7 +52340,7 @@ FUNCTION(_sk_screen_hsw_lowp)
_sk_screen_hsw_lowp:
.byte 197,93,253,192 // vpaddw %ymm0,%ymm4,%ymm8
.byte 197,221,213,192 // vpmullw %ymm0,%ymm4,%ymm0
- .byte 196,98,125,121,13,244,228,0,0 // vpbroadcastw 0xe4f4(%rip),%ymm9 # 39dc0 <_sk_srcover_bgra_8888_sse2_lowp+0x938>
+ .byte 196,98,125,121,13,116,215,0,0 // vpbroadcastw 0xd774(%rip),%ymm9 # 39040 <_sk_srcover_bgra_8888_sse2_lowp+0x93c>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
@@ -52366,7 +52366,7 @@ HIDDEN _sk_xor__hsw_lowp
.globl _sk_xor__hsw_lowp
FUNCTION(_sk_xor__hsw_lowp)
_sk_xor__hsw_lowp:
- .byte 196,98,125,121,5,153,228,0,0 // vpbroadcastw 0xe499(%rip),%ymm8 # 39dc2 <_sk_srcover_bgra_8888_sse2_lowp+0x93a>
+ .byte 196,98,125,121,5,25,215,0,0 // vpbroadcastw 0xd719(%rip),%ymm8 # 39042 <_sk_srcover_bgra_8888_sse2_lowp+0x93e>
.byte 197,61,249,207 // vpsubw %ymm7,%ymm8,%ymm9
.byte 197,181,213,192 // vpmullw %ymm0,%ymm9,%ymm0
.byte 197,61,249,211 // vpsubw %ymm3,%ymm8,%ymm10
@@ -52400,7 +52400,7 @@ _sk_darken_hsw_lowp:
.byte 197,197,213,192 // vpmullw %ymm0,%ymm7,%ymm0
.byte 197,93,213,203 // vpmullw %ymm3,%ymm4,%ymm9
.byte 196,194,125,62,193 // vpmaxuw %ymm9,%ymm0,%ymm0
- .byte 196,98,125,121,13,29,228,0,0 // vpbroadcastw 0xe41d(%rip),%ymm9 # 39dc4 <_sk_srcover_bgra_8888_sse2_lowp+0x93c>
+ .byte 196,98,125,121,13,157,214,0,0 // vpbroadcastw 0xd69d(%rip),%ymm9 # 39044 <_sk_srcover_bgra_8888_sse2_lowp+0x940>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
@@ -52434,7 +52434,7 @@ _sk_lighten_hsw_lowp:
.byte 197,197,213,192 // vpmullw %ymm0,%ymm7,%ymm0
.byte 197,93,213,203 // vpmullw %ymm3,%ymm4,%ymm9
.byte 196,194,125,58,193 // vpminuw %ymm9,%ymm0,%ymm0
- .byte 196,98,125,121,13,158,227,0,0 // vpbroadcastw 0xe39e(%rip),%ymm9 # 39dc6 <_sk_srcover_bgra_8888_sse2_lowp+0x93e>
+ .byte 196,98,125,121,13,30,214,0,0 // vpbroadcastw 0xd61e(%rip),%ymm9 # 39046 <_sk_srcover_bgra_8888_sse2_lowp+0x942>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
@@ -52468,10 +52468,10 @@ _sk_difference_hsw_lowp:
.byte 197,197,213,192 // vpmullw %ymm0,%ymm7,%ymm0
.byte 197,93,213,203 // vpmullw %ymm3,%ymm4,%ymm9
.byte 196,194,125,58,193 // vpminuw %ymm9,%ymm0,%ymm0
- .byte 196,98,125,121,13,31,227,0,0 // vpbroadcastw 0xe31f(%rip),%ymm9 # 39dc8 <_sk_srcover_bgra_8888_sse2_lowp+0x940>
+ .byte 196,98,125,121,13,159,213,0,0 // vpbroadcastw 0xd59f(%rip),%ymm9 # 39048 <_sk_srcover_bgra_8888_sse2_lowp+0x944>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,7 // vpsrlw $0x7,%ymm0,%ymm0
- .byte 196,98,125,121,21,14,227,0,0 // vpbroadcastw 0xe30e(%rip),%ymm10 # 39dca <_sk_srcover_bgra_8888_sse2_lowp+0x942>
+ .byte 196,98,125,121,21,142,213,0,0 // vpbroadcastw 0xd58e(%rip),%ymm10 # 3904a <_sk_srcover_bgra_8888_sse2_lowp+0x946>
.byte 196,193,125,219,194 // vpand %ymm10,%ymm0,%ymm0
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
.byte 197,85,253,193 // vpaddw %ymm1,%ymm5,%ymm8
@@ -52504,10 +52504,10 @@ FUNCTION(_sk_exclusion_hsw_lowp)
_sk_exclusion_hsw_lowp:
.byte 197,93,253,192 // vpaddw %ymm0,%ymm4,%ymm8
.byte 197,221,213,192 // vpmullw %ymm0,%ymm4,%ymm0
- .byte 196,98,125,121,13,147,226,0,0 // vpbroadcastw 0xe293(%rip),%ymm9 # 39dcc <_sk_srcover_bgra_8888_sse2_lowp+0x944>
+ .byte 196,98,125,121,13,19,213,0,0 // vpbroadcastw 0xd513(%rip),%ymm9 # 3904c <_sk_srcover_bgra_8888_sse2_lowp+0x948>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,7 // vpsrlw $0x7,%ymm0,%ymm0
- .byte 196,98,125,121,21,130,226,0,0 // vpbroadcastw 0xe282(%rip),%ymm10 # 39dce <_sk_srcover_bgra_8888_sse2_lowp+0x946>
+ .byte 196,98,125,121,21,2,213,0,0 // vpbroadcastw 0xd502(%rip),%ymm10 # 3904e <_sk_srcover_bgra_8888_sse2_lowp+0x94a>
.byte 196,193,125,219,194 // vpand %ymm10,%ymm0,%ymm0
.byte 197,189,249,192 // vpsubw %ymm0,%ymm8,%ymm0
.byte 197,85,253,193 // vpaddw %ymm1,%ymm5,%ymm8
@@ -52535,7 +52535,7 @@ HIDDEN _sk_hardlight_hsw_lowp
FUNCTION(_sk_hardlight_hsw_lowp)
_sk_hardlight_hsw_lowp:
.byte 197,125,253,192 // vpaddw %ymm0,%ymm0,%ymm8
- .byte 196,98,125,121,13,31,226,0,0 // vpbroadcastw 0xe21f(%rip),%ymm9 # 39dd2 <_sk_srcover_bgra_8888_sse2_lowp+0x94a>
+ .byte 196,98,125,121,13,159,212,0,0 // vpbroadcastw 0xd49f(%rip),%ymm9 # 39052 <_sk_srcover_bgra_8888_sse2_lowp+0x94e>
.byte 196,65,101,239,209 // vpxor %ymm9,%ymm3,%ymm10
.byte 196,65,61,239,217 // vpxor %ymm9,%ymm8,%ymm11
.byte 196,65,37,101,218 // vpcmpgtw %ymm10,%ymm11,%ymm11
@@ -52563,7 +52563,7 @@ _sk_hardlight_hsw_lowp:
.byte 197,101,249,210 // vpsubw %ymm2,%ymm3,%ymm10
.byte 197,69,249,246 // vpsubw %ymm6,%ymm7,%ymm14
.byte 196,65,45,213,214 // vpmullw %ymm14,%ymm10,%ymm10
- .byte 196,98,125,121,53,150,225,0,0 // vpbroadcastw 0xe196(%rip),%ymm14 # 39dd0 <_sk_srcover_bgra_8888_sse2_lowp+0x948>
+ .byte 196,98,125,121,53,22,212,0,0 // vpbroadcastw 0xd416(%rip),%ymm14 # 39050 <_sk_srcover_bgra_8888_sse2_lowp+0x94c>
.byte 196,65,45,253,210 // vpaddw %ymm10,%ymm10,%ymm10
.byte 196,65,61,249,194 // vpsubw %ymm10,%ymm8,%ymm8
.byte 197,13,249,215 // vpsubw %ymm7,%ymm14,%ymm10
@@ -52600,7 +52600,7 @@ HIDDEN _sk_overlay_hsw_lowp
FUNCTION(_sk_overlay_hsw_lowp)
_sk_overlay_hsw_lowp:
.byte 197,93,253,196 // vpaddw %ymm4,%ymm4,%ymm8
- .byte 196,98,125,121,13,11,225,0,0 // vpbroadcastw 0xe10b(%rip),%ymm9 # 39dd6 <_sk_srcover_bgra_8888_sse2_lowp+0x94e>
+ .byte 196,98,125,121,13,139,211,0,0 // vpbroadcastw 0xd38b(%rip),%ymm9 # 39056 <_sk_srcover_bgra_8888_sse2_lowp+0x952>
.byte 196,65,69,239,209 // vpxor %ymm9,%ymm7,%ymm10
.byte 196,65,61,239,193 // vpxor %ymm9,%ymm8,%ymm8
.byte 196,65,61,101,218 // vpcmpgtw %ymm10,%ymm8,%ymm11
@@ -52624,7 +52624,7 @@ _sk_overlay_hsw_lowp:
.byte 196,65,21,253,237 // vpaddw %ymm13,%ymm13,%ymm13
.byte 196,65,61,249,237 // vpsubw %ymm13,%ymm8,%ymm13
.byte 196,67,13,76,229,192 // vpblendvb %ymm12,%ymm13,%ymm14,%ymm12
- .byte 196,98,125,121,45,147,224,0,0 // vpbroadcastw 0xe093(%rip),%ymm13 # 39dd4 <_sk_srcover_bgra_8888_sse2_lowp+0x94c>
+ .byte 196,98,125,121,45,19,211,0,0 // vpbroadcastw 0xd313(%rip),%ymm13 # 39054 <_sk_srcover_bgra_8888_sse2_lowp+0x950>
.byte 197,77,253,246 // vpaddw %ymm6,%ymm6,%ymm14
.byte 196,65,13,239,201 // vpxor %ymm9,%ymm14,%ymm9
.byte 196,65,53,101,202 // vpcmpgtw %ymm10,%ymm9,%ymm9
@@ -52747,7 +52747,7 @@ _sk_load_8888_hsw_lowp:
.byte 196,227,101,14,200,170 // vpblendw $0xaa,%ymm0,%ymm3,%ymm1
.byte 196,227,109,14,192,170 // vpblendw $0xaa,%ymm0,%ymm2,%ymm0
.byte 196,226,125,43,201 // vpackusdw %ymm1,%ymm0,%ymm1
- .byte 197,125,111,5,89,222,0,0 // vmovdqa 0xde59(%rip),%ymm8 # 39de0 <_sk_srcover_bgra_8888_sse2_lowp+0x958>
+ .byte 197,125,111,5,217,208,0,0 // vmovdqa 0xd0d9(%rip),%ymm8 # 39060 <_sk_srcover_bgra_8888_sse2_lowp+0x95c>
.byte 196,193,117,219,192 // vpand %ymm8,%ymm1,%ymm0
.byte 197,245,113,209,8 // vpsrlw $0x8,%ymm1,%ymm1
.byte 197,229,114,211,16 // vpsrld $0x10,%ymm3,%ymm3
@@ -52877,7 +52877,7 @@ _sk_load_8888_dst_hsw_lowp:
.byte 196,227,69,14,236,170 // vpblendw $0xaa,%ymm4,%ymm7,%ymm5
.byte 196,227,77,14,228,170 // vpblendw $0xaa,%ymm4,%ymm6,%ymm4
.byte 196,226,93,43,237 // vpackusdw %ymm5,%ymm4,%ymm5
- .byte 197,125,111,5,114,220,0,0 // vmovdqa 0xdc72(%rip),%ymm8 # 39e00 <_sk_srcover_bgra_8888_sse2_lowp+0x978>
+ .byte 197,125,111,5,242,206,0,0 // vmovdqa 0xcef2(%rip),%ymm8 # 39080 <_sk_srcover_bgra_8888_sse2_lowp+0x97c>
.byte 196,193,85,219,224 // vpand %ymm8,%ymm5,%ymm4
.byte 197,213,113,213,8 // vpsrlw $0x8,%ymm5,%ymm5
.byte 197,197,114,215,16 // vpsrld $0x10,%ymm7,%ymm7
@@ -53103,7 +53103,7 @@ _sk_load_bgra_hsw_lowp:
.byte 196,227,125,14,209,170 // vpblendw $0xaa,%ymm1,%ymm0,%ymm2
.byte 196,227,101,14,201,170 // vpblendw $0xaa,%ymm1,%ymm3,%ymm1
.byte 196,226,117,43,202 // vpackusdw %ymm2,%ymm1,%ymm1
- .byte 197,125,111,5,30,217,0,0 // vmovdqa 0xd91e(%rip),%ymm8 # 39e20 <_sk_srcover_bgra_8888_sse2_lowp+0x998>
+ .byte 197,125,111,5,158,203,0,0 // vmovdqa 0xcb9e(%rip),%ymm8 # 390a0 <_sk_srcover_bgra_8888_sse2_lowp+0x99c>
.byte 196,193,117,219,208 // vpand %ymm8,%ymm1,%ymm2
.byte 197,245,113,209,8 // vpsrlw $0x8,%ymm1,%ymm1
.byte 197,253,114,208,16 // vpsrld $0x10,%ymm0,%ymm0
@@ -53232,7 +53232,7 @@ _sk_load_bgra_dst_hsw_lowp:
.byte 196,227,93,14,245,170 // vpblendw $0xaa,%ymm5,%ymm4,%ymm6
.byte 196,227,69,14,237,170 // vpblendw $0xaa,%ymm5,%ymm7,%ymm5
.byte 196,226,85,43,238 // vpackusdw %ymm6,%ymm5,%ymm5
- .byte 197,125,111,5,54,215,0,0 // vmovdqa 0xd736(%rip),%ymm8 # 39e40 <_sk_srcover_bgra_8888_sse2_lowp+0x9b8>
+ .byte 197,125,111,5,182,201,0,0 // vmovdqa 0xc9b6(%rip),%ymm8 # 390c0 <_sk_srcover_bgra_8888_sse2_lowp+0x9bc>
.byte 196,193,85,219,240 // vpand %ymm8,%ymm5,%ymm6
.byte 197,213,113,213,8 // vpsrlw $0x8,%ymm5,%ymm5
.byte 197,221,114,212,16 // vpsrld $0x10,%ymm4,%ymm4
@@ -53381,83 +53381,44 @@ _sk_gather_8888_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 196,98,125,88,64,12 // vpbroadcastd 0xc(%rax),%ymm8
.byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
- .byte 196,65,61,254,209 // vpaddd %ymm9,%ymm8,%ymm10
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
- .byte 196,193,116,93,202 // vminps %ymm10,%ymm1,%ymm1
- .byte 196,193,124,93,194 // vminps %ymm10,%ymm0,%ymm0
- .byte 196,98,125,88,80,16 // vpbroadcastd 0x10(%rax),%ymm10
- .byte 196,65,45,254,201 // vpaddd %ymm9,%ymm10,%ymm9
- .byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
- .byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
- .byte 196,193,100,93,217 // vminps %ymm9,%ymm3,%ymm3
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
+ .byte 196,193,124,95,194 // vmaxps %ymm10,%ymm0,%ymm0
+ .byte 196,193,116,95,202 // vmaxps %ymm10,%ymm1,%ymm1
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 196,98,125,88,64,16 // vpbroadcastd 0x10(%rax),%ymm8
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,193,100,95,218 // vmaxps %ymm10,%ymm3,%ymm3
+ .byte 196,193,108,95,210 // vmaxps %ymm10,%ymm2,%ymm2
+ .byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 76,139,0 // mov (%rax),%r8
.byte 197,254,91,219 // vcvttps2dq %ymm3,%ymm3
.byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
- .byte 196,98,125,88,72,8 // vpbroadcastd 0x8(%rax),%ymm9
- .byte 196,226,53,64,210 // vpmulld %ymm2,%ymm9,%ymm2
- .byte 196,226,53,64,219 // vpmulld %ymm3,%ymm9,%ymm3
+ .byte 196,98,125,88,64,8 // vpbroadcastd 0x8(%rax),%ymm8
+ .byte 196,226,61,64,210 // vpmulld %ymm2,%ymm8,%ymm2
+ .byte 196,226,61,64,219 // vpmulld %ymm3,%ymm8,%ymm3
.byte 197,254,91,192 // vcvttps2dq %ymm0,%ymm0
- .byte 197,109,254,208 // vpaddd %ymm0,%ymm2,%ymm10
+ .byte 197,237,254,192 // vpaddd %ymm0,%ymm2,%ymm0
.byte 197,254,91,201 // vcvttps2dq %ymm1,%ymm1
.byte 197,229,254,201 // vpaddd %ymm1,%ymm3,%ymm1
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,227,125,57,203,1 // vextracti128 $0x1,%ymm1,%xmm3
- .byte 196,225,249,126,216 // vmovq %xmm3,%rax
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,20,136 // vmovd (%r8,%r9,4),%xmm2
- .byte 196,195,249,22,217,1 // vpextrq $0x1,%xmm3,%r9
- .byte 196,195,105,34,20,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm2,%xmm2
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,105,34,20,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm2,%xmm2
- .byte 196,225,249,126,200 // vmovq %xmm1,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,105,34,20,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm2,%xmm2
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,28,136 // vmovd (%r8,%r9,4),%xmm3
- .byte 196,195,249,22,201,1 // vpextrq $0x1,%xmm1,%r9
- .byte 196,195,97,34,12,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm3,%xmm1
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,113,34,12,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm1,%xmm1
- .byte 196,97,249,126,200 // vmovq %xmm9,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,113,34,12,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm1,%xmm1
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,28,136 // vmovd (%r8,%r9,4),%xmm3
- .byte 196,67,249,22,201,1 // vpextrq $0x1,%xmm9,%r9
- .byte 196,195,97,34,28,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm3,%xmm3
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,97,34,28,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm3,%xmm3
- .byte 196,97,249,126,208 // vmovq %xmm10,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,97,34,28,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm3,%xmm3
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,4,136 // vmovd (%r8,%r9,4),%xmm0
- .byte 196,67,249,22,209,1 // vpextrq $0x1,%xmm10,%r9
- .byte 196,195,121,34,4,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm0,%xmm0
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,121,34,4,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm0,%xmm0
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,121,34,4,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm0,%xmm0
- .byte 196,227,117,56,210,1 // vinserti128 $0x1,%xmm2,%ymm1,%ymm2
- .byte 196,227,125,56,195,1 // vinserti128 $0x1,%xmm3,%ymm0,%ymm0
- .byte 196,227,125,56,217,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm3
- .byte 196,227,125,70,210,49 // vperm2i128 $0x31,%ymm2,%ymm0,%ymm2
- .byte 196,195,109,14,192,170 // vpblendw $0xaa,%ymm8,%ymm2,%ymm0
- .byte 196,195,101,14,200,170 // vpblendw $0xaa,%ymm8,%ymm3,%ymm1
+ .byte 197,237,118,210 // vpcmpeqd %ymm2,%ymm2,%ymm2
+ .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
+ .byte 196,194,109,144,28,128 // vpgatherdd %ymm2,(%r8,%ymm0,4),%ymm3
+ .byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
+ .byte 196,194,53,144,4,136 // vpgatherdd %ymm9,(%r8,%ymm1,4),%ymm0
+ .byte 196,227,101,56,208,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm2
+ .byte 196,227,101,70,216,49 // vperm2i128 $0x31,%ymm0,%ymm3,%ymm3
+ .byte 196,195,101,14,194,170 // vpblendw $0xaa,%ymm10,%ymm3,%ymm0
+ .byte 196,195,109,14,202,170 // vpblendw $0xaa,%ymm10,%ymm2,%ymm1
.byte 196,226,117,43,200 // vpackusdw %ymm0,%ymm1,%ymm1
- .byte 197,125,111,5,2,212,0,0 // vmovdqa 0xd402(%rip),%ymm8 # 39e60 <_sk_srcover_bgra_8888_sse2_lowp+0x9d8>
+ .byte 197,125,111,5,82,199,0,0 // vmovdqa 0xc752(%rip),%ymm8 # 390e0 <_sk_srcover_bgra_8888_sse2_lowp+0x9dc>
.byte 196,193,117,219,192 // vpand %ymm8,%ymm1,%ymm0
.byte 197,245,113,209,8 // vpsrlw $0x8,%ymm1,%ymm1
- .byte 197,237,114,210,16 // vpsrld $0x10,%ymm2,%ymm2
.byte 197,229,114,211,16 // vpsrld $0x10,%ymm3,%ymm3
- .byte 196,226,101,43,218 // vpackusdw %ymm2,%ymm3,%ymm3
+ .byte 197,237,114,210,16 // vpsrld $0x10,%ymm2,%ymm2
+ .byte 196,226,109,43,219 // vpackusdw %ymm3,%ymm2,%ymm3
.byte 196,193,101,219,208 // vpand %ymm8,%ymm3,%ymm2
.byte 197,229,113,211,8 // vpsrlw $0x8,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -53470,85 +53431,46 @@ _sk_gather_bgra_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 196,98,125,88,64,12 // vpbroadcastd 0xc(%rax),%ymm8
.byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
- .byte 196,65,61,254,209 // vpaddd %ymm9,%ymm8,%ymm10
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
- .byte 196,193,116,93,202 // vminps %ymm10,%ymm1,%ymm1
- .byte 196,193,124,93,194 // vminps %ymm10,%ymm0,%ymm0
- .byte 196,98,125,88,80,16 // vpbroadcastd 0x10(%rax),%ymm10
- .byte 196,65,45,254,201 // vpaddd %ymm9,%ymm10,%ymm9
- .byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
- .byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
- .byte 196,193,108,93,209 // vminps %ymm9,%ymm2,%ymm2
- .byte 196,193,100,93,217 // vminps %ymm9,%ymm3,%ymm3
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
+ .byte 196,193,124,95,194 // vmaxps %ymm10,%ymm0,%ymm0
+ .byte 196,193,116,95,202 // vmaxps %ymm10,%ymm1,%ymm1
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 196,98,125,88,64,16 // vpbroadcastd 0x10(%rax),%ymm8
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,193,100,95,218 // vmaxps %ymm10,%ymm3,%ymm3
+ .byte 196,193,108,95,210 // vmaxps %ymm10,%ymm2,%ymm2
+ .byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
.byte 76,139,0 // mov (%rax),%r8
.byte 197,254,91,219 // vcvttps2dq %ymm3,%ymm3
.byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
- .byte 196,98,125,88,72,8 // vpbroadcastd 0x8(%rax),%ymm9
- .byte 196,226,53,64,210 // vpmulld %ymm2,%ymm9,%ymm2
- .byte 196,226,53,64,219 // vpmulld %ymm3,%ymm9,%ymm3
+ .byte 196,98,125,88,64,8 // vpbroadcastd 0x8(%rax),%ymm8
+ .byte 196,226,61,64,210 // vpmulld %ymm2,%ymm8,%ymm2
+ .byte 196,226,61,64,219 // vpmulld %ymm3,%ymm8,%ymm3
.byte 197,254,91,192 // vcvttps2dq %ymm0,%ymm0
- .byte 197,109,254,208 // vpaddd %ymm0,%ymm2,%ymm10
+ .byte 197,237,254,192 // vpaddd %ymm0,%ymm2,%ymm0
.byte 197,254,91,201 // vcvttps2dq %ymm1,%ymm1
.byte 197,229,254,201 // vpaddd %ymm1,%ymm3,%ymm1
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,227,125,57,203,1 // vextracti128 $0x1,%ymm1,%xmm3
- .byte 196,225,249,126,216 // vmovq %xmm3,%rax
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,20,136 // vmovd (%r8,%r9,4),%xmm2
- .byte 196,195,249,22,217,1 // vpextrq $0x1,%xmm3,%r9
- .byte 196,195,105,34,20,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm2,%xmm2
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,105,34,20,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm2,%xmm2
- .byte 196,225,249,126,200 // vmovq %xmm1,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,105,34,20,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm2,%xmm2
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,28,136 // vmovd (%r8,%r9,4),%xmm3
- .byte 196,195,249,22,201,1 // vpextrq $0x1,%xmm1,%r9
- .byte 196,195,97,34,12,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm3,%xmm1
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,113,34,12,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm1,%xmm1
- .byte 196,97,249,126,200 // vmovq %xmm9,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,113,34,12,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm1,%xmm1
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,28,136 // vmovd (%r8,%r9,4),%xmm3
- .byte 196,67,249,22,201,1 // vpextrq $0x1,%xmm9,%r9
- .byte 196,195,97,34,28,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm3,%xmm3
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,97,34,28,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm3,%xmm3
- .byte 196,97,249,126,208 // vmovq %xmm10,%rax
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,97,34,28,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm3,%xmm3
- .byte 65,137,193 // mov %eax,%r9d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 196,129,121,110,4,136 // vmovd (%r8,%r9,4),%xmm0
- .byte 196,67,249,22,209,1 // vpextrq $0x1,%xmm10,%r9
- .byte 196,195,121,34,4,128,1 // vpinsrd $0x1,(%r8,%rax,4),%xmm0,%xmm0
- .byte 68,137,200 // mov %r9d,%eax
- .byte 196,195,121,34,4,128,2 // vpinsrd $0x2,(%r8,%rax,4),%xmm0,%xmm0
- .byte 73,193,233,32 // shr $0x20,%r9
- .byte 196,131,121,34,4,136,3 // vpinsrd $0x3,(%r8,%r9,4),%xmm0,%xmm0
- .byte 196,227,117,56,210,1 // vinserti128 $0x1,%xmm2,%ymm1,%ymm2
- .byte 196,227,125,56,195,1 // vinserti128 $0x1,%xmm3,%ymm0,%ymm0
- .byte 196,227,125,56,217,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm3
- .byte 196,227,125,70,194,49 // vperm2i128 $0x31,%ymm2,%ymm0,%ymm0
- .byte 196,195,125,14,200,170 // vpblendw $0xaa,%ymm8,%ymm0,%ymm1
- .byte 196,195,101,14,208,170 // vpblendw $0xaa,%ymm8,%ymm3,%ymm2
+ .byte 197,237,118,210 // vpcmpeqd %ymm2,%ymm2,%ymm2
+ .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
+ .byte 196,194,109,144,28,128 // vpgatherdd %ymm2,(%r8,%ymm0,4),%ymm3
+ .byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
+ .byte 196,194,53,144,4,136 // vpgatherdd %ymm9,(%r8,%ymm1,4),%ymm0
+ .byte 196,99,101,56,192,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm8
+ .byte 196,227,101,70,192,49 // vperm2i128 $0x31,%ymm0,%ymm3,%ymm0
+ .byte 196,195,125,14,202,170 // vpblendw $0xaa,%ymm10,%ymm0,%ymm1
+ .byte 196,195,61,14,210,170 // vpblendw $0xaa,%ymm10,%ymm8,%ymm2
.byte 196,226,109,43,201 // vpackusdw %ymm1,%ymm2,%ymm1
- .byte 197,125,111,5,121,210,0,0 // vmovdqa 0xd279(%rip),%ymm8 # 39e80 <_sk_srcover_bgra_8888_sse2_lowp+0x9f8>
- .byte 196,193,117,219,208 // vpand %ymm8,%ymm1,%ymm2
+ .byte 197,253,111,29,153,198,0,0 // vmovdqa 0xc699(%rip),%ymm3 # 39100 <_sk_srcover_bgra_8888_sse2_lowp+0x9fc>
+ .byte 197,245,219,211 // vpand %ymm3,%ymm1,%ymm2
.byte 197,245,113,209,8 // vpsrlw $0x8,%ymm1,%ymm1
.byte 197,253,114,208,16 // vpsrld $0x10,%ymm0,%ymm0
- .byte 197,229,114,211,16 // vpsrld $0x10,%ymm3,%ymm3
- .byte 196,226,101,43,216 // vpackusdw %ymm0,%ymm3,%ymm3
- .byte 196,193,101,219,192 // vpand %ymm8,%ymm3,%ymm0
- .byte 197,229,113,211,8 // vpsrlw $0x8,%ymm3,%ymm3
+ .byte 196,193,61,114,208,16 // vpsrld $0x10,%ymm8,%ymm8
+ .byte 196,98,61,43,192 // vpackusdw %ymm0,%ymm8,%ymm8
+ .byte 197,189,219,195 // vpand %ymm3,%ymm8,%ymm0
+ .byte 196,193,101,113,208,8 // vpsrlw $0x8,%ymm8,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -53565,24 +53487,24 @@ _sk_load_565_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2cc6d <_sk_load_565_hsw_lowp+0x3f>
+ .byte 119,37 // ja 2cacd <_sk_load_565_hsw_lowp+0x3f>
.byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,130,1,0,0 // lea 0x182(%rip),%r9 # 2cdd8 <_sk_load_565_hsw_lowp+0x1aa>
+ .byte 76,141,13,130,1,0,0 // lea 0x182(%rip),%r9 # 2cc38 <_sk_load_565_hsw_lowp+0x1aa>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 233,8,1,0,0 // jmpq 2cd75 <_sk_load_565_hsw_lowp+0x147>
+ .byte 233,8,1,0,0 // jmpq 2cbd5 <_sk_load_565_hsw_lowp+0x147>
.byte 196,193,126,111,4,80 // vmovdqu (%r8,%rdx,2),%ymm0
- .byte 233,253,0,0,0 // jmpq 2cd75 <_sk_load_565_hsw_lowp+0x147>
+ .byte 233,253,0,0,0 // jmpq 2cbd5 <_sk_load_565_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
.byte 196,193,121,110,12,80 // vmovd (%r8,%rdx,2),%xmm1
.byte 196,227,125,2,193,1 // vpblendd $0x1,%ymm1,%ymm0,%ymm0
- .byte 233,218,0,0,0 // jmpq 2cd75 <_sk_load_565_hsw_lowp+0x147>
+ .byte 233,218,0,0,0 // jmpq 2cbd5 <_sk_load_565_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
@@ -53592,7 +53514,7 @@ _sk_load_565_hsw_lowp:
.byte 196,227,125,2,193,15 // vpblendd $0xf,%ymm1,%ymm0,%ymm0
.byte 196,193,122,126,12,80 // vmovq (%r8,%rdx,2),%xmm1
.byte 196,227,125,2,193,3 // vpblendd $0x3,%ymm1,%ymm0,%ymm0
- .byte 233,155,0,0,0 // jmpq 2cd75 <_sk_load_565_hsw_lowp+0x147>
+ .byte 233,155,0,0,0 // jmpq 2cbd5 <_sk_load_565_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
@@ -53604,7 +53526,7 @@ _sk_load_565_hsw_lowp:
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
.byte 196,193,122,111,12,80 // vmovdqu (%r8,%rdx,2),%xmm1
.byte 196,227,117,2,192,240 // vpblendd $0xf0,%ymm0,%ymm1,%ymm0
- .byte 235,83 // jmp 2cd75 <_sk_load_565_hsw_lowp+0x147>
+ .byte 235,83 // jmp 2cbd5 <_sk_load_565_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
@@ -53618,13 +53540,13 @@ _sk_load_565_hsw_lowp:
.byte 196,193,122,111,20,80 // vmovdqu (%r8,%rdx,2),%xmm2
.byte 196,227,109,56,201,1 // vinserti128 $0x1,%xmm1,%ymm2,%ymm1
.byte 196,227,117,2,192,192 // vpblendd $0xc0,%ymm0,%ymm1,%ymm0
- .byte 196,226,125,121,13,34,209,0,0 // vpbroadcastw 0xd122(%rip),%ymm1 # 39ea0 <_sk_srcover_bgra_8888_sse2_lowp+0xa18>
+ .byte 196,226,125,121,13,66,197,0,0 // vpbroadcastw 0xc542(%rip),%ymm1 # 39120 <_sk_srcover_bgra_8888_sse2_lowp+0xa1c>
.byte 197,237,113,208,8 // vpsrlw $0x8,%ymm0,%ymm2
.byte 197,237,219,201 // vpand %ymm1,%ymm2,%ymm1
.byte 197,237,113,208,5 // vpsrlw $0x5,%ymm0,%ymm2
- .byte 196,226,125,121,29,13,209,0,0 // vpbroadcastw 0xd10d(%rip),%ymm3 # 39ea2 <_sk_srcover_bgra_8888_sse2_lowp+0xa1a>
+ .byte 196,226,125,121,29,45,197,0,0 // vpbroadcastw 0xc52d(%rip),%ymm3 # 39122 <_sk_srcover_bgra_8888_sse2_lowp+0xa1e>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
- .byte 196,226,125,121,29,2,209,0,0 // vpbroadcastw 0xd102(%rip),%ymm3 # 39ea4 <_sk_srcover_bgra_8888_sse2_lowp+0xa1c>
+ .byte 196,226,125,121,29,34,197,0,0 // vpbroadcastw 0xc522(%rip),%ymm3 # 39124 <_sk_srcover_bgra_8888_sse2_lowp+0xa20>
.byte 197,253,219,219 // vpand %ymm3,%ymm0,%ymm3
.byte 197,253,113,208,13 // vpsrlw $0xd,%ymm0,%ymm0
.byte 197,245,235,192 // vpor %ymm0,%ymm1,%ymm0
@@ -53635,7 +53557,7 @@ _sk_load_565_hsw_lowp:
.byte 197,229,113,211,2 // vpsrlw $0x2,%ymm3,%ymm3
.byte 197,237,235,211 // vpor %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,208,208,0,0 // vpbroadcastw 0xd0d0(%rip),%ymm3 # 39ea6 <_sk_srcover_bgra_8888_sse2_lowp+0xa1e>
+ .byte 196,226,125,121,29,240,196,0,0 // vpbroadcastw 0xc4f0(%rip),%ymm3 # 39126 <_sk_srcover_bgra_8888_sse2_lowp+0xa22>
.byte 255,224 // jmpq *%rax
.byte 135,254 // xchg %edi,%esi
.byte 255 // (bad)
@@ -53685,24 +53607,24 @@ _sk_load_565_dst_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2ce53 <_sk_load_565_dst_hsw_lowp+0x3f>
+ .byte 119,37 // ja 2ccb3 <_sk_load_565_dst_hsw_lowp+0x3f>
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,132,1,0,0 // lea 0x184(%rip),%r9 # 2cfc0 <_sk_load_565_dst_hsw_lowp+0x1ac>
+ .byte 76,141,13,132,1,0,0 // lea 0x184(%rip),%r9 # 2ce20 <_sk_load_565_dst_hsw_lowp+0x1ac>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 233,8,1,0,0 // jmpq 2cf5b <_sk_load_565_dst_hsw_lowp+0x147>
+ .byte 233,8,1,0,0 // jmpq 2cdbb <_sk_load_565_dst_hsw_lowp+0x147>
.byte 196,193,126,111,36,80 // vmovdqu (%r8,%rdx,2),%ymm4
- .byte 233,253,0,0,0 // jmpq 2cf5b <_sk_load_565_dst_hsw_lowp+0x147>
+ .byte 233,253,0,0,0 // jmpq 2cdbb <_sk_load_565_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
.byte 196,193,121,110,44,80 // vmovd (%r8,%rdx,2),%xmm5
.byte 196,227,93,2,229,1 // vpblendd $0x1,%ymm5,%ymm4,%ymm4
- .byte 233,218,0,0,0 // jmpq 2cf5b <_sk_load_565_dst_hsw_lowp+0x147>
+ .byte 233,218,0,0,0 // jmpq 2cdbb <_sk_load_565_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
@@ -53712,7 +53634,7 @@ _sk_load_565_dst_hsw_lowp:
.byte 196,227,93,2,229,15 // vpblendd $0xf,%ymm5,%ymm4,%ymm4
.byte 196,193,122,126,44,80 // vmovq (%r8,%rdx,2),%xmm5
.byte 196,227,93,2,229,3 // vpblendd $0x3,%ymm5,%ymm4,%ymm4
- .byte 233,155,0,0,0 // jmpq 2cf5b <_sk_load_565_dst_hsw_lowp+0x147>
+ .byte 233,155,0,0,0 // jmpq 2cdbb <_sk_load_565_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
@@ -53724,7 +53646,7 @@ _sk_load_565_dst_hsw_lowp:
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
.byte 196,193,122,111,44,80 // vmovdqu (%r8,%rdx,2),%xmm5
.byte 196,227,85,2,228,240 // vpblendd $0xf0,%ymm4,%ymm5,%ymm4
- .byte 235,83 // jmp 2cf5b <_sk_load_565_dst_hsw_lowp+0x147>
+ .byte 235,83 // jmp 2cdbb <_sk_load_565_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
@@ -53738,13 +53660,13 @@ _sk_load_565_dst_hsw_lowp:
.byte 196,193,122,111,52,80 // vmovdqu (%r8,%rdx,2),%xmm6
.byte 196,227,77,56,237,1 // vinserti128 $0x1,%xmm5,%ymm6,%ymm5
.byte 196,227,85,2,228,192 // vpblendd $0xc0,%ymm4,%ymm5,%ymm4
- .byte 196,226,125,121,45,68,207,0,0 // vpbroadcastw 0xcf44(%rip),%ymm5 # 39ea8 <_sk_srcover_bgra_8888_sse2_lowp+0xa20>
+ .byte 196,226,125,121,45,100,195,0,0 // vpbroadcastw 0xc364(%rip),%ymm5 # 39128 <_sk_srcover_bgra_8888_sse2_lowp+0xa24>
.byte 197,205,113,212,8 // vpsrlw $0x8,%ymm4,%ymm6
.byte 197,205,219,237 // vpand %ymm5,%ymm6,%ymm5
.byte 197,205,113,212,5 // vpsrlw $0x5,%ymm4,%ymm6
- .byte 196,226,125,121,61,47,207,0,0 // vpbroadcastw 0xcf2f(%rip),%ymm7 # 39eaa <_sk_srcover_bgra_8888_sse2_lowp+0xa22>
+ .byte 196,226,125,121,61,79,195,0,0 // vpbroadcastw 0xc34f(%rip),%ymm7 # 3912a <_sk_srcover_bgra_8888_sse2_lowp+0xa26>
.byte 197,205,219,247 // vpand %ymm7,%ymm6,%ymm6
- .byte 196,226,125,121,61,36,207,0,0 // vpbroadcastw 0xcf24(%rip),%ymm7 # 39eac <_sk_srcover_bgra_8888_sse2_lowp+0xa24>
+ .byte 196,226,125,121,61,68,195,0,0 // vpbroadcastw 0xc344(%rip),%ymm7 # 3912c <_sk_srcover_bgra_8888_sse2_lowp+0xa28>
.byte 197,221,219,255 // vpand %ymm7,%ymm4,%ymm7
.byte 197,221,113,212,13 // vpsrlw $0xd,%ymm4,%ymm4
.byte 197,213,235,228 // vpor %ymm4,%ymm5,%ymm4
@@ -53755,7 +53677,7 @@ _sk_load_565_dst_hsw_lowp:
.byte 197,197,113,215,2 // vpsrlw $0x2,%ymm7,%ymm7
.byte 197,205,235,247 // vpor %ymm7,%ymm6,%ymm6
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,61,242,206,0,0 // vpbroadcastw 0xcef2(%rip),%ymm7 # 39eae <_sk_srcover_bgra_8888_sse2_lowp+0xa26>
+ .byte 196,226,125,121,61,18,195,0,0 // vpbroadcastw 0xc312(%rip),%ymm7 # 3912e <_sk_srcover_bgra_8888_sse2_lowp+0xa2a>
.byte 255,224 // jmpq *%rax
.byte 102,144 // xchg %ax,%ax
.byte 133,254 // test %edi,%esi
@@ -53806,10 +53728,10 @@ _sk_store_565_hsw_lowp:
.byte 76,15,175,193 // imul %rcx,%r8
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
- .byte 196,98,125,121,5,155,206,0,0 // vpbroadcastw 0xce9b(%rip),%ymm8 # 39eb0 <_sk_srcover_bgra_8888_sse2_lowp+0xa28>
+ .byte 196,98,125,121,5,187,194,0,0 // vpbroadcastw 0xc2bb(%rip),%ymm8 # 39130 <_sk_srcover_bgra_8888_sse2_lowp+0xa2c>
.byte 197,181,113,240,8 // vpsllw $0x8,%ymm0,%ymm9
.byte 196,65,53,219,192 // vpand %ymm8,%ymm9,%ymm8
- .byte 196,98,125,121,13,138,206,0,0 // vpbroadcastw 0xce8a(%rip),%ymm9 # 39eb2 <_sk_srcover_bgra_8888_sse2_lowp+0xa2a>
+ .byte 196,98,125,121,13,170,194,0,0 // vpbroadcastw 0xc2aa(%rip),%ymm9 # 39132 <_sk_srcover_bgra_8888_sse2_lowp+0xa2e>
.byte 197,173,113,241,3 // vpsllw $0x3,%ymm1,%ymm10
.byte 196,65,45,219,201 // vpand %ymm9,%ymm10,%ymm9
.byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
@@ -53819,31 +53741,31 @@ _sk_store_565_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,31 // ja 2d06a <_sk_store_565_hsw_lowp+0x6e>
+ .byte 119,31 // ja 2ceca <_sk_store_565_hsw_lowp+0x6e>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,195,0,0,0 // lea 0xc3(%rip),%r9 # 2d118 <_sk_store_565_hsw_lowp+0x11c>
+ .byte 76,141,13,195,0,0,0 // lea 0xc3(%rip),%r9 # 2cf78 <_sk_store_565_hsw_lowp+0x11c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 196,67,121,21,4,80,0 // vpextrw $0x0,%xmm8,(%r8,%rdx,2)
- .byte 233,167,0,0,0 // jmpq 2d111 <_sk_store_565_hsw_lowp+0x115>
+ .byte 233,167,0,0,0 // jmpq 2cf71 <_sk_store_565_hsw_lowp+0x115>
.byte 196,65,126,127,4,80 // vmovdqu %ymm8,(%r8,%rdx,2)
- .byte 233,156,0,0,0 // jmpq 2d111 <_sk_store_565_hsw_lowp+0x115>
+ .byte 233,156,0,0,0 // jmpq 2cf71 <_sk_store_565_hsw_lowp+0x115>
.byte 196,67,121,21,68,80,4,2 // vpextrw $0x2,%xmm8,0x4(%r8,%rdx,2)
.byte 196,65,121,126,4,80 // vmovd %xmm8,(%r8,%rdx,2)
- .byte 233,137,0,0,0 // jmpq 2d111 <_sk_store_565_hsw_lowp+0x115>
+ .byte 233,137,0,0,0 // jmpq 2cf71 <_sk_store_565_hsw_lowp+0x115>
.byte 196,67,121,21,68,80,12,6 // vpextrw $0x6,%xmm8,0xc(%r8,%rdx,2)
.byte 196,67,121,21,68,80,10,5 // vpextrw $0x5,%xmm8,0xa(%r8,%rdx,2)
.byte 196,67,121,21,68,80,8,4 // vpextrw $0x4,%xmm8,0x8(%r8,%rdx,2)
.byte 196,65,121,214,4,80 // vmovq %xmm8,(%r8,%rdx,2)
- .byte 235,105 // jmp 2d111 <_sk_store_565_hsw_lowp+0x115>
+ .byte 235,105 // jmp 2cf71 <_sk_store_565_hsw_lowp+0x115>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,20,2 // vpextrw $0x2,%xmm9,0x14(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,18,1 // vpextrw $0x1,%xmm9,0x12(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,16,0 // vpextrw $0x0,%xmm9,0x10(%r8,%rdx,2)
- .byte 235,55 // jmp 2d10b <_sk_store_565_hsw_lowp+0x10f>
+ .byte 235,55 // jmp 2cf6b <_sk_store_565_hsw_lowp+0x10f>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,28,6 // vpextrw $0x6,%xmm9,0x1c(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
@@ -53866,7 +53788,7 @@ _sk_store_565_hsw_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 120,255 // js 2d12d <_sk_store_565_hsw_lowp+0x131>
+ .byte 120,255 // js 2cf8d <_sk_store_565_hsw_lowp+0x131>
.byte 255 // (bad)
.byte 255,112,255 // pushq -0x1(%rax)
.byte 255 // (bad)
@@ -53978,13 +53900,13 @@ _sk_gather_565_hsw_lowp:
.byte 67,15,183,4,80 // movzwl (%r8,%r10,2),%eax
.byte 197,249,196,192,7 // vpinsrw $0x7,%eax,%xmm0,%xmm0
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
- .byte 196,226,125,121,13,206,203,0,0 // vpbroadcastw 0xcbce(%rip),%ymm1 # 39eb4 <_sk_srcover_bgra_8888_sse2_lowp+0xa2c>
+ .byte 196,226,125,121,13,238,191,0,0 // vpbroadcastw 0xbfee(%rip),%ymm1 # 39134 <_sk_srcover_bgra_8888_sse2_lowp+0xa30>
.byte 197,237,113,208,8 // vpsrlw $0x8,%ymm0,%ymm2
.byte 197,237,219,201 // vpand %ymm1,%ymm2,%ymm1
.byte 197,237,113,208,5 // vpsrlw $0x5,%ymm0,%ymm2
- .byte 196,226,125,121,29,185,203,0,0 // vpbroadcastw 0xcbb9(%rip),%ymm3 # 39eb6 <_sk_srcover_bgra_8888_sse2_lowp+0xa2e>
+ .byte 196,226,125,121,29,217,191,0,0 // vpbroadcastw 0xbfd9(%rip),%ymm3 # 39136 <_sk_srcover_bgra_8888_sse2_lowp+0xa32>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
- .byte 196,226,125,121,29,174,203,0,0 // vpbroadcastw 0xcbae(%rip),%ymm3 # 39eb8 <_sk_srcover_bgra_8888_sse2_lowp+0xa30>
+ .byte 196,226,125,121,29,206,191,0,0 // vpbroadcastw 0xbfce(%rip),%ymm3 # 39138 <_sk_srcover_bgra_8888_sse2_lowp+0xa34>
.byte 197,253,219,219 // vpand %ymm3,%ymm0,%ymm3
.byte 197,253,113,208,13 // vpsrlw $0xd,%ymm0,%ymm0
.byte 197,245,235,192 // vpor %ymm0,%ymm1,%ymm0
@@ -53995,7 +53917,7 @@ _sk_gather_565_hsw_lowp:
.byte 197,229,113,211,2 // vpsrlw $0x2,%ymm3,%ymm3
.byte 197,237,235,211 // vpor %ymm3,%ymm2,%ymm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,124,203,0,0 // vpbroadcastw 0xcb7c(%rip),%ymm3 # 39eba <_sk_srcover_bgra_8888_sse2_lowp+0xa32>
+ .byte 196,226,125,121,29,156,191,0,0 // vpbroadcastw 0xbf9c(%rip),%ymm3 # 3913a <_sk_srcover_bgra_8888_sse2_lowp+0xa36>
.byte 255,224 // jmpq *%rax
HIDDEN _sk_load_4444_hsw_lowp
@@ -54011,24 +53933,24 @@ _sk_load_4444_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2d37f <_sk_load_4444_hsw_lowp+0x3f>
+ .byte 119,37 // ja 2d1df <_sk_load_4444_hsw_lowp+0x3f>
.byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,112,1,0,0 // lea 0x170(%rip),%r9 # 2d4d8 <_sk_load_4444_hsw_lowp+0x198>
+ .byte 76,141,13,112,1,0,0 // lea 0x170(%rip),%r9 # 2d338 <_sk_load_4444_hsw_lowp+0x198>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 233,8,1,0,0 // jmpq 2d487 <_sk_load_4444_hsw_lowp+0x147>
+ .byte 233,8,1,0,0 // jmpq 2d2e7 <_sk_load_4444_hsw_lowp+0x147>
.byte 196,193,126,111,4,80 // vmovdqu (%r8,%rdx,2),%ymm0
- .byte 233,253,0,0,0 // jmpq 2d487 <_sk_load_4444_hsw_lowp+0x147>
+ .byte 233,253,0,0,0 // jmpq 2d2e7 <_sk_load_4444_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
.byte 196,193,121,110,12,80 // vmovd (%r8,%rdx,2),%xmm1
.byte 196,227,125,2,193,1 // vpblendd $0x1,%ymm1,%ymm0,%ymm0
- .byte 233,218,0,0,0 // jmpq 2d487 <_sk_load_4444_hsw_lowp+0x147>
+ .byte 233,218,0,0,0 // jmpq 2d2e7 <_sk_load_4444_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
@@ -54038,7 +53960,7 @@ _sk_load_4444_hsw_lowp:
.byte 196,227,125,2,193,15 // vpblendd $0xf,%ymm1,%ymm0,%ymm0
.byte 196,193,122,126,12,80 // vmovq (%r8,%rdx,2),%xmm1
.byte 196,227,125,2,193,3 // vpblendd $0x3,%ymm1,%ymm0,%ymm0
- .byte 233,155,0,0,0 // jmpq 2d487 <_sk_load_4444_hsw_lowp+0x147>
+ .byte 233,155,0,0,0 // jmpq 2d2e7 <_sk_load_4444_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
@@ -54050,7 +53972,7 @@ _sk_load_4444_hsw_lowp:
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
.byte 196,193,122,111,12,80 // vmovdqu (%r8,%rdx,2),%xmm1
.byte 196,227,117,2,192,240 // vpblendd $0xf0,%ymm0,%ymm1,%ymm0
- .byte 235,83 // jmp 2d487 <_sk_load_4444_hsw_lowp+0x147>
+ .byte 235,83 // jmp 2d2e7 <_sk_load_4444_hsw_lowp+0x147>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,193,121,196,76,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm0,%xmm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
@@ -54066,7 +53988,7 @@ _sk_load_4444_hsw_lowp:
.byte 196,227,117,2,192,192 // vpblendd $0xc0,%ymm0,%ymm1,%ymm0
.byte 197,245,113,208,12 // vpsrlw $0xc,%ymm0,%ymm1
.byte 197,237,113,208,8 // vpsrlw $0x8,%ymm0,%ymm2
- .byte 196,226,125,121,29,34,202,0,0 // vpbroadcastw 0xca22(%rip),%ymm3 # 39ebc <_sk_srcover_bgra_8888_sse2_lowp+0xa34>
+ .byte 196,226,125,121,29,66,190,0,0 // vpbroadcastw 0xbe42(%rip),%ymm3 # 3913c <_sk_srcover_bgra_8888_sse2_lowp+0xa38>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
.byte 197,189,113,208,4 // vpsrlw $0x4,%ymm0,%ymm8
.byte 197,61,219,195 // vpand %ymm3,%ymm8,%ymm8
@@ -54131,24 +54053,24 @@ _sk_load_4444_dst_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2d553 <_sk_load_4444_dst_hsw_lowp+0x3f>
+ .byte 119,37 // ja 2d3b3 <_sk_load_4444_dst_hsw_lowp+0x3f>
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,112,1,0,0 // lea 0x170(%rip),%r9 # 2d6ac <_sk_load_4444_dst_hsw_lowp+0x198>
+ .byte 76,141,13,112,1,0,0 // lea 0x170(%rip),%r9 # 2d50c <_sk_load_4444_dst_hsw_lowp+0x198>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 233,8,1,0,0 // jmpq 2d65b <_sk_load_4444_dst_hsw_lowp+0x147>
+ .byte 233,8,1,0,0 // jmpq 2d4bb <_sk_load_4444_dst_hsw_lowp+0x147>
.byte 196,193,126,111,36,80 // vmovdqu (%r8,%rdx,2),%ymm4
- .byte 233,253,0,0,0 // jmpq 2d65b <_sk_load_4444_dst_hsw_lowp+0x147>
+ .byte 233,253,0,0,0 // jmpq 2d4bb <_sk_load_4444_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
.byte 196,193,121,110,44,80 // vmovd (%r8,%rdx,2),%xmm5
.byte 196,227,93,2,229,1 // vpblendd $0x1,%ymm5,%ymm4,%ymm4
- .byte 233,218,0,0,0 // jmpq 2d65b <_sk_load_4444_dst_hsw_lowp+0x147>
+ .byte 233,218,0,0,0 // jmpq 2d4bb <_sk_load_4444_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
@@ -54158,7 +54080,7 @@ _sk_load_4444_dst_hsw_lowp:
.byte 196,227,93,2,229,15 // vpblendd $0xf,%ymm5,%ymm4,%ymm4
.byte 196,193,122,126,44,80 // vmovq (%r8,%rdx,2),%xmm5
.byte 196,227,93,2,229,3 // vpblendd $0x3,%ymm5,%ymm4,%ymm4
- .byte 233,155,0,0,0 // jmpq 2d65b <_sk_load_4444_dst_hsw_lowp+0x147>
+ .byte 233,155,0,0,0 // jmpq 2d4bb <_sk_load_4444_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
@@ -54170,7 +54092,7 @@ _sk_load_4444_dst_hsw_lowp:
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
.byte 196,193,122,111,44,80 // vmovdqu (%r8,%rdx,2),%xmm5
.byte 196,227,85,2,228,240 // vpblendd $0xf0,%ymm4,%ymm5,%ymm4
- .byte 235,83 // jmp 2d65b <_sk_load_4444_dst_hsw_lowp+0x147>
+ .byte 235,83 // jmp 2d4bb <_sk_load_4444_dst_hsw_lowp+0x147>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,193,89,196,108,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm4,%xmm5
.byte 196,227,93,56,229,1 // vinserti128 $0x1,%xmm5,%ymm4,%ymm4
@@ -54186,7 +54108,7 @@ _sk_load_4444_dst_hsw_lowp:
.byte 196,227,85,2,228,192 // vpblendd $0xc0,%ymm4,%ymm5,%ymm4
.byte 197,213,113,212,12 // vpsrlw $0xc,%ymm4,%ymm5
.byte 197,205,113,212,8 // vpsrlw $0x8,%ymm4,%ymm6
- .byte 196,226,125,121,61,80,200,0,0 // vpbroadcastw 0xc850(%rip),%ymm7 # 39ebe <_sk_srcover_bgra_8888_sse2_lowp+0xa36>
+ .byte 196,226,125,121,61,112,188,0,0 // vpbroadcastw 0xbc70(%rip),%ymm7 # 3913e <_sk_srcover_bgra_8888_sse2_lowp+0xa3a>
.byte 197,205,219,247 // vpand %ymm7,%ymm6,%ymm6
.byte 197,189,113,212,4 // vpsrlw $0x4,%ymm4,%ymm8
.byte 197,61,219,199 // vpand %ymm7,%ymm8,%ymm8
@@ -54247,13 +54169,13 @@ _sk_store_4444_hsw_lowp:
.byte 76,15,175,193 // imul %rcx,%r8
.byte 77,1,192 // add %r8,%r8
.byte 76,3,0 // add (%rax),%r8
- .byte 196,98,125,121,5,191,199,0,0 // vpbroadcastw 0xc7bf(%rip),%ymm8 # 39ec0 <_sk_srcover_bgra_8888_sse2_lowp+0xa38>
+ .byte 196,98,125,121,5,223,187,0,0 // vpbroadcastw 0xbbdf(%rip),%ymm8 # 39140 <_sk_srcover_bgra_8888_sse2_lowp+0xa3c>
.byte 197,181,113,240,8 // vpsllw $0x8,%ymm0,%ymm9
.byte 196,65,53,219,192 // vpand %ymm8,%ymm9,%ymm8
.byte 197,181,113,241,4 // vpsllw $0x4,%ymm1,%ymm9
- .byte 197,53,219,13,200,199,0,0 // vpand 0xc7c8(%rip),%ymm9,%ymm9 # 39ee0 <_sk_srcover_bgra_8888_sse2_lowp+0xa58>
+ .byte 197,53,219,13,232,187,0,0 // vpand 0xbbe8(%rip),%ymm9,%ymm9 # 39160 <_sk_srcover_bgra_8888_sse2_lowp+0xa5c>
.byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
- .byte 196,98,125,121,13,218,199,0,0 // vpbroadcastw 0xc7da(%rip),%ymm9 # 39f00 <_sk_srcover_bgra_8888_sse2_lowp+0xa78>
+ .byte 196,98,125,121,13,250,187,0,0 // vpbroadcastw 0xbbfa(%rip),%ymm9 # 39180 <_sk_srcover_bgra_8888_sse2_lowp+0xa7c>
.byte 196,65,109,219,201 // vpand %ymm9,%ymm2,%ymm9
.byte 197,173,113,211,4 // vpsrlw $0x4,%ymm3,%ymm10
.byte 196,65,53,235,202 // vpor %ymm10,%ymm9,%ymm9
@@ -54262,31 +54184,31 @@ _sk_store_4444_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,31 // ja 2d763 <_sk_store_4444_hsw_lowp+0x7b>
+ .byte 119,31 // ja 2d5c3 <_sk_store_4444_hsw_lowp+0x7b>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,194,0,0,0 // lea 0xc2(%rip),%r9 # 2d810 <_sk_store_4444_hsw_lowp+0x128>
+ .byte 76,141,13,194,0,0,0 // lea 0xc2(%rip),%r9 # 2d670 <_sk_store_4444_hsw_lowp+0x128>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 196,67,121,21,4,80,0 // vpextrw $0x0,%xmm8,(%r8,%rdx,2)
- .byte 233,167,0,0,0 // jmpq 2d80a <_sk_store_4444_hsw_lowp+0x122>
+ .byte 233,167,0,0,0 // jmpq 2d66a <_sk_store_4444_hsw_lowp+0x122>
.byte 196,65,126,127,4,80 // vmovdqu %ymm8,(%r8,%rdx,2)
- .byte 233,156,0,0,0 // jmpq 2d80a <_sk_store_4444_hsw_lowp+0x122>
+ .byte 233,156,0,0,0 // jmpq 2d66a <_sk_store_4444_hsw_lowp+0x122>
.byte 196,67,121,21,68,80,4,2 // vpextrw $0x2,%xmm8,0x4(%r8,%rdx,2)
.byte 196,65,121,126,4,80 // vmovd %xmm8,(%r8,%rdx,2)
- .byte 233,137,0,0,0 // jmpq 2d80a <_sk_store_4444_hsw_lowp+0x122>
+ .byte 233,137,0,0,0 // jmpq 2d66a <_sk_store_4444_hsw_lowp+0x122>
.byte 196,67,121,21,68,80,12,6 // vpextrw $0x6,%xmm8,0xc(%r8,%rdx,2)
.byte 196,67,121,21,68,80,10,5 // vpextrw $0x5,%xmm8,0xa(%r8,%rdx,2)
.byte 196,67,121,21,68,80,8,4 // vpextrw $0x4,%xmm8,0x8(%r8,%rdx,2)
.byte 196,65,121,214,4,80 // vmovq %xmm8,(%r8,%rdx,2)
- .byte 235,105 // jmp 2d80a <_sk_store_4444_hsw_lowp+0x122>
+ .byte 235,105 // jmp 2d66a <_sk_store_4444_hsw_lowp+0x122>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,20,2 // vpextrw $0x2,%xmm9,0x14(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,18,1 // vpextrw $0x1,%xmm9,0x12(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,16,0 // vpextrw $0x0,%xmm9,0x10(%r8,%rdx,2)
- .byte 235,55 // jmp 2d804 <_sk_store_4444_hsw_lowp+0x11c>
+ .byte 235,55 // jmp 2d664 <_sk_store_4444_hsw_lowp+0x11c>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,21,76,80,28,6 // vpextrw $0x6,%xmm9,0x1c(%r8,%rdx,2)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
@@ -54309,7 +54231,7 @@ _sk_store_4444_hsw_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 121,255 // jns 2d825 <_sk_store_4444_hsw_lowp+0x13d>
+ .byte 121,255 // jns 2d685 <_sk_store_4444_hsw_lowp+0x13d>
.byte 255 // (bad)
.byte 255,113,255 // pushq -0x1(%rcx)
.byte 255 // (bad)
@@ -54424,7 +54346,7 @@ _sk_gather_4444_hsw_lowp:
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
.byte 197,245,113,208,12 // vpsrlw $0xc,%ymm0,%ymm1
.byte 197,237,113,208,8 // vpsrlw $0x8,%ymm0,%ymm2
- .byte 196,226,125,121,29,26,197,0,0 // vpbroadcastw 0xc51a(%rip),%ymm3 # 39f02 <_sk_srcover_bgra_8888_sse2_lowp+0xa7a>
+ .byte 196,226,125,121,29,58,185,0,0 // vpbroadcastw 0xb93a(%rip),%ymm3 # 39182 <_sk_srcover_bgra_8888_sse2_lowp+0xa7e>
.byte 197,237,219,211 // vpand %ymm3,%ymm2,%ymm2
.byte 197,189,113,208,4 // vpsrlw $0x4,%ymm0,%ymm8
.byte 197,61,219,195 // vpand %ymm3,%ymm8,%ymm8
@@ -54452,38 +54374,38 @@ _sk_load_a8_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2da5f <_sk_load_a8_hsw_lowp+0x3c>
+ .byte 119,37 // ja 2d8bf <_sk_load_a8_hsw_lowp+0x3c>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,220,0,0,0 // lea 0xdc(%rip),%r9 # 2db24 <_sk_load_a8_hsw_lowp+0x101>
+ .byte 76,141,13,220,0,0,0 // lea 0xdc(%rip),%r9 # 2d984 <_sk_load_a8_hsw_lowp+0x101>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 233,175,0,0,0 // jmpq 2db0e <_sk_load_a8_hsw_lowp+0xeb>
+ .byte 233,175,0,0,0 // jmpq 2d96e <_sk_load_a8_hsw_lowp+0xeb>
.byte 196,193,122,111,4,16 // vmovdqu (%r8,%rdx,1),%xmm0
- .byte 233,164,0,0,0 // jmpq 2db0e <_sk_load_a8_hsw_lowp+0xeb>
+ .byte 233,164,0,0,0 // jmpq 2d96e <_sk_load_a8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm0,%xmm0
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,249,110,200 // vmovd %eax,%xmm1
.byte 196,227,121,14,193,1 // vpblendw $0x1,%xmm1,%xmm0,%xmm0
- .byte 233,132,0,0,0 // jmpq 2db0e <_sk_load_a8_hsw_lowp+0xeb>
+ .byte 233,132,0,0,0 // jmpq 2d96e <_sk_load_a8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,193,121,110,12,16 // vmovd (%r8,%rdx,1),%xmm1
.byte 196,227,121,2,193,1 // vpblendd $0x1,%xmm1,%xmm0,%xmm0
- .byte 235,90 // jmp 2db0e <_sk_load_a8_hsw_lowp+0xeb>
+ .byte 235,90 // jmp 2d96e <_sk_load_a8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,193,122,126,12,16 // vmovq (%r8,%rdx,1),%xmm1
.byte 196,227,113,2,192,12 // vpblendd $0xc,%xmm0,%xmm1,%xmm0
- .byte 235,48 // jmp 2db0e <_sk_load_a8_hsw_lowp+0xeb>
+ .byte 235,48 // jmp 2d96e <_sk_load_a8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm0,%xmm0
@@ -54541,38 +54463,38 @@ _sk_load_a8_dst_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2db9c <_sk_load_a8_dst_hsw_lowp+0x3c>
+ .byte 119,37 // ja 2d9fc <_sk_load_a8_dst_hsw_lowp+0x3c>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,219,0,0,0 // lea 0xdb(%rip),%r9 # 2dc60 <_sk_load_a8_dst_hsw_lowp+0x100>
+ .byte 76,141,13,219,0,0,0 // lea 0xdb(%rip),%r9 # 2dac0 <_sk_load_a8_dst_hsw_lowp+0x100>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 233,175,0,0,0 // jmpq 2dc4b <_sk_load_a8_dst_hsw_lowp+0xeb>
+ .byte 233,175,0,0,0 // jmpq 2daab <_sk_load_a8_dst_hsw_lowp+0xeb>
.byte 196,193,122,111,36,16 // vmovdqu (%r8,%rdx,1),%xmm4
- .byte 233,164,0,0,0 // jmpq 2dc4b <_sk_load_a8_dst_hsw_lowp+0xeb>
+ .byte 233,164,0,0,0 // jmpq 2daab <_sk_load_a8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm4,%xmm4
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,249,110,232 // vmovd %eax,%xmm5
.byte 196,227,89,14,229,1 // vpblendw $0x1,%xmm5,%xmm4,%xmm4
- .byte 233,132,0,0,0 // jmpq 2dc4b <_sk_load_a8_dst_hsw_lowp+0xeb>
+ .byte 233,132,0,0,0 // jmpq 2daab <_sk_load_a8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,193,121,110,44,16 // vmovd (%r8,%rdx,1),%xmm5
.byte 196,227,89,2,229,1 // vpblendd $0x1,%xmm5,%xmm4,%xmm4
- .byte 235,90 // jmp 2dc4b <_sk_load_a8_dst_hsw_lowp+0xeb>
+ .byte 235,90 // jmp 2daab <_sk_load_a8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,193,122,126,44,16 // vmovq (%r8,%rdx,1),%xmm5
.byte 196,227,81,2,228,12 // vpblendd $0xc,%xmm4,%xmm5,%xmm4
- .byte 235,48 // jmp 2dc4b <_sk_load_a8_dst_hsw_lowp+0xeb>
+ .byte 235,48 // jmp 2daab <_sk_load_a8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm4,%xmm4
@@ -54629,7 +54551,7 @@ _sk_store_a8_hsw_lowp:
.byte 76,15,175,193 // imul %rcx,%r8
.byte 76,3,0 // add (%rax),%r8
.byte 196,195,125,57,216,1 // vextracti128 $0x1,%ymm3,%xmm8
- .byte 197,121,111,13,169,202,0,0 // vmovdqa 0xcaa9(%rip),%xmm9 # 3a760 <_sk_srcover_bgra_8888_sse2_lowp+0x12d8>
+ .byte 197,121,111,13,201,190,0,0 // vmovdqa 0xbec9(%rip),%xmm9 # 399e0 <_sk_srcover_bgra_8888_sse2_lowp+0x12dc>
.byte 196,66,57,0,193 // vpshufb %xmm9,%xmm8,%xmm8
.byte 196,66,97,0,201 // vpshufb %xmm9,%xmm3,%xmm9
.byte 196,65,49,108,192 // vpunpcklqdq %xmm8,%xmm9,%xmm8
@@ -54637,28 +54559,28 @@ _sk_store_a8_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,28 // ja 2dcec <_sk_store_a8_hsw_lowp+0x50>
+ .byte 119,28 // ja 2db4c <_sk_store_a8_hsw_lowp+0x50>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,146,0,0,0 // lea 0x92(%rip),%r9 # 2dd6c <_sk_store_a8_hsw_lowp+0xd0>
+ .byte 76,141,13,146,0,0,0 // lea 0x92(%rip),%r9 # 2dbcc <_sk_store_a8_hsw_lowp+0xd0>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 196,67,121,20,4,16,0 // vpextrb $0x0,%xmm8,(%r8,%rdx,1)
- .byte 235,121 // jmp 2dd65 <_sk_store_a8_hsw_lowp+0xc9>
+ .byte 235,121 // jmp 2dbc5 <_sk_store_a8_hsw_lowp+0xc9>
.byte 196,65,122,127,4,16 // vmovdqu %xmm8,(%r8,%rdx,1)
- .byte 235,113 // jmp 2dd65 <_sk_store_a8_hsw_lowp+0xc9>
+ .byte 235,113 // jmp 2dbc5 <_sk_store_a8_hsw_lowp+0xc9>
.byte 196,67,121,20,68,16,2,2 // vpextrb $0x2,%xmm8,0x2(%r8,%rdx,1)
.byte 196,67,121,21,4,16,0 // vpextrw $0x0,%xmm8,(%r8,%rdx,1)
- .byte 235,96 // jmp 2dd65 <_sk_store_a8_hsw_lowp+0xc9>
+ .byte 235,96 // jmp 2dbc5 <_sk_store_a8_hsw_lowp+0xc9>
.byte 196,67,121,20,68,16,6,6 // vpextrb $0x6,%xmm8,0x6(%r8,%rdx,1)
.byte 196,67,121,20,68,16,5,5 // vpextrb $0x5,%xmm8,0x5(%r8,%rdx,1)
.byte 196,67,121,20,68,16,4,4 // vpextrb $0x4,%xmm8,0x4(%r8,%rdx,1)
.byte 196,65,121,126,4,16 // vmovd %xmm8,(%r8,%rdx,1)
- .byte 235,64 // jmp 2dd65 <_sk_store_a8_hsw_lowp+0xc9>
+ .byte 235,64 // jmp 2dbc5 <_sk_store_a8_hsw_lowp+0xc9>
.byte 196,67,121,20,68,16,10,10 // vpextrb $0xa,%xmm8,0xa(%r8,%rdx,1)
.byte 196,67,121,20,68,16,9,9 // vpextrb $0x9,%xmm8,0x9(%r8,%rdx,1)
.byte 196,67,121,20,68,16,8,8 // vpextrb $0x8,%xmm8,0x8(%r8,%rdx,1)
- .byte 235,32 // jmp 2dd5f <_sk_store_a8_hsw_lowp+0xc3>
+ .byte 235,32 // jmp 2dbbf <_sk_store_a8_hsw_lowp+0xc3>
.byte 196,67,121,20,68,16,14,14 // vpextrb $0xe,%xmm8,0xe(%r8,%rdx,1)
.byte 196,67,121,20,68,16,13,13 // vpextrb $0xd,%xmm8,0xd(%r8,%rdx,1)
.byte 196,67,121,20,68,16,12,12 // vpextrb $0xc,%xmm8,0xc(%r8,%rdx,1)
@@ -54667,7 +54589,7 @@ _sk_store_a8_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 119,255 // ja 2dd6d <_sk_store_a8_hsw_lowp+0xd1>
+ .byte 119,255 // ja 2dbcd <_sk_store_a8_hsw_lowp+0xd1>
.byte 255 // (bad)
.byte 255,144,255,255,255,136 // callq *-0x77000001(%rax)
.byte 255 // (bad)
@@ -54808,38 +54730,38 @@ _sk_load_g8_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2df8b <_sk_load_g8_hsw_lowp+0x3c>
+ .byte 119,37 // ja 2ddeb <_sk_load_g8_hsw_lowp+0x3c>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,224,0,0,0 // lea 0xe0(%rip),%r9 # 2e054 <_sk_load_g8_hsw_lowp+0x105>
+ .byte 76,141,13,224,0,0,0 // lea 0xe0(%rip),%r9 # 2deb4 <_sk_load_g8_hsw_lowp+0x105>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 233,175,0,0,0 // jmpq 2e03a <_sk_load_g8_hsw_lowp+0xeb>
+ .byte 233,175,0,0,0 // jmpq 2de9a <_sk_load_g8_hsw_lowp+0xeb>
.byte 196,193,122,111,4,16 // vmovdqu (%r8,%rdx,1),%xmm0
- .byte 233,164,0,0,0 // jmpq 2e03a <_sk_load_g8_hsw_lowp+0xeb>
+ .byte 233,164,0,0,0 // jmpq 2de9a <_sk_load_g8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm0,%xmm0
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,249,110,200 // vmovd %eax,%xmm1
.byte 196,227,121,14,193,1 // vpblendw $0x1,%xmm1,%xmm0,%xmm0
- .byte 233,132,0,0,0 // jmpq 2e03a <_sk_load_g8_hsw_lowp+0xeb>
+ .byte 233,132,0,0,0 // jmpq 2de9a <_sk_load_g8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,193,121,110,12,16 // vmovd (%r8,%rdx,1),%xmm1
.byte 196,227,121,2,193,1 // vpblendd $0x1,%xmm1,%xmm0,%xmm0
- .byte 235,90 // jmp 2e03a <_sk_load_g8_hsw_lowp+0xeb>
+ .byte 235,90 // jmp 2de9a <_sk_load_g8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,193,122,126,12,16 // vmovq (%r8,%rdx,1),%xmm1
.byte 196,227,113,2,192,12 // vpblendd $0xc,%xmm0,%xmm1,%xmm0
- .byte 235,48 // jmp 2e03a <_sk_load_g8_hsw_lowp+0xeb>
+ .byte 235,48 // jmp 2de9a <_sk_load_g8_hsw_lowp+0xeb>
.byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
.byte 196,195,121,32,68,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm0,%xmm0
.byte 196,195,121,32,68,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm0,%xmm0
@@ -54849,7 +54771,7 @@ _sk_load_g8_hsw_lowp:
.byte 196,227,113,2,192,8 // vpblendd $0x8,%xmm0,%xmm1,%xmm0
.byte 196,226,125,48,192 // vpmovzxbw %xmm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,186,190,0,0 // vpbroadcastw 0xbeba(%rip),%ymm3 # 39f04 <_sk_srcover_bgra_8888_sse2_lowp+0xa7c>
+ .byte 196,226,125,121,29,218,178,0,0 // vpbroadcastw 0xb2da(%rip),%ymm3 # 39184 <_sk_srcover_bgra_8888_sse2_lowp+0xa80>
.byte 197,253,111,200 // vmovdqa %ymm0,%ymm1
.byte 197,253,111,208 // vmovdqa %ymm0,%ymm2
.byte 255,224 // jmpq *%rax
@@ -54860,7 +54782,7 @@ _sk_load_g8_hsw_lowp:
.byte 255,66,255 // incl -0x1(%rdx)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 126,255 // jle 2e061 <_sk_load_g8_hsw_lowp+0x112>
+ .byte 126,255 // jle 2dec1 <_sk_load_g8_hsw_lowp+0x112>
.byte 255 // (bad)
.byte 255,118,255 // pushq -0x1(%rsi)
.byte 255 // (bad)
@@ -54900,38 +54822,38 @@ _sk_load_g8_dst_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,37 // ja 2e0cc <_sk_load_g8_dst_hsw_lowp+0x3c>
+ .byte 119,37 // ja 2df2c <_sk_load_g8_dst_hsw_lowp+0x3c>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,227,0,0,0 // lea 0xe3(%rip),%r9 # 2e198 <_sk_load_g8_dst_hsw_lowp+0x108>
+ .byte 76,141,13,227,0,0,0 // lea 0xe3(%rip),%r9 # 2dff8 <_sk_load_g8_dst_hsw_lowp+0x108>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 233,175,0,0,0 // jmpq 2e17b <_sk_load_g8_dst_hsw_lowp+0xeb>
+ .byte 233,175,0,0,0 // jmpq 2dfdb <_sk_load_g8_dst_hsw_lowp+0xeb>
.byte 196,193,122,111,36,16 // vmovdqu (%r8,%rdx,1),%xmm4
- .byte 233,164,0,0,0 // jmpq 2e17b <_sk_load_g8_dst_hsw_lowp+0xeb>
+ .byte 233,164,0,0,0 // jmpq 2dfdb <_sk_load_g8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm4,%xmm4
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,249,110,232 // vmovd %eax,%xmm5
.byte 196,227,89,14,229,1 // vpblendw $0x1,%xmm5,%xmm4,%xmm4
- .byte 233,132,0,0,0 // jmpq 2e17b <_sk_load_g8_dst_hsw_lowp+0xeb>
+ .byte 233,132,0,0,0 // jmpq 2dfdb <_sk_load_g8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,193,121,110,44,16 // vmovd (%r8,%rdx,1),%xmm5
.byte 196,227,89,2,229,1 // vpblendd $0x1,%xmm5,%xmm4,%xmm4
- .byte 235,90 // jmp 2e17b <_sk_load_g8_dst_hsw_lowp+0xeb>
+ .byte 235,90 // jmp 2dfdb <_sk_load_g8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,193,122,126,44,16 // vmovq (%r8,%rdx,1),%xmm5
.byte 196,227,81,2,228,12 // vpblendd $0xc,%xmm4,%xmm5,%xmm4
- .byte 235,48 // jmp 2e17b <_sk_load_g8_dst_hsw_lowp+0xeb>
+ .byte 235,48 // jmp 2dfdb <_sk_load_g8_dst_hsw_lowp+0xeb>
.byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
.byte 196,195,89,32,100,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm4,%xmm4
.byte 196,195,89,32,100,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm4,%xmm4
@@ -54941,7 +54863,7 @@ _sk_load_g8_dst_hsw_lowp:
.byte 196,227,81,2,228,8 // vpblendd $0x8,%xmm4,%xmm5,%xmm4
.byte 196,226,125,48,228 // vpmovzxbw %xmm4,%ymm4
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,61,123,189,0,0 // vpbroadcastw 0xbd7b(%rip),%ymm7 # 39f06 <_sk_srcover_bgra_8888_sse2_lowp+0xa7e>
+ .byte 196,226,125,121,61,155,177,0,0 // vpbroadcastw 0xb19b(%rip),%ymm7 # 39186 <_sk_srcover_bgra_8888_sse2_lowp+0xa82>
.byte 197,253,111,236 // vmovdqa %ymm4,%ymm5
.byte 197,253,111,244 // vmovdqa %ymm4,%ymm6
.byte 255,224 // jmpq *%rax
@@ -54955,7 +54877,7 @@ _sk_load_g8_dst_hsw_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 2e1a5 <_sk_load_g8_dst_hsw_lowp+0x115>
+ .byte 123,255 // jnp 2e005 <_sk_load_g8_dst_hsw_lowp+0x115>
.byte 255 // (bad)
.byte 255,115,255 // pushq -0x1(%rbx)
.byte 255 // (bad)
@@ -54985,12 +54907,12 @@ HIDDEN _sk_luminance_to_alpha_hsw_lowp
.globl _sk_luminance_to_alpha_hsw_lowp
FUNCTION(_sk_luminance_to_alpha_hsw_lowp)
_sk_luminance_to_alpha_hsw_lowp:
- .byte 196,226,125,121,29,43,189,0,0 // vpbroadcastw 0xbd2b(%rip),%ymm3 # 39f08 <_sk_srcover_bgra_8888_sse2_lowp+0xa80>
+ .byte 196,226,125,121,29,75,177,0,0 // vpbroadcastw 0xb14b(%rip),%ymm3 # 39188 <_sk_srcover_bgra_8888_sse2_lowp+0xa84>
.byte 197,253,213,195 // vpmullw %ymm3,%ymm0,%ymm0
- .byte 196,226,125,121,29,32,189,0,0 // vpbroadcastw 0xbd20(%rip),%ymm3 # 39f0a <_sk_srcover_bgra_8888_sse2_lowp+0xa82>
+ .byte 196,226,125,121,29,64,177,0,0 // vpbroadcastw 0xb140(%rip),%ymm3 # 3918a <_sk_srcover_bgra_8888_sse2_lowp+0xa86>
.byte 197,245,213,203 // vpmullw %ymm3,%ymm1,%ymm1
.byte 197,245,253,192 // vpaddw %ymm0,%ymm1,%ymm0
- .byte 196,226,125,121,13,17,189,0,0 // vpbroadcastw 0xbd11(%rip),%ymm1 # 39f0c <_sk_srcover_bgra_8888_sse2_lowp+0xa84>
+ .byte 196,226,125,121,13,49,177,0,0 // vpbroadcastw 0xb131(%rip),%ymm1 # 3918c <_sk_srcover_bgra_8888_sse2_lowp+0xa88>
.byte 197,237,213,201 // vpmullw %ymm1,%ymm2,%ymm1
.byte 197,253,253,193 // vpaddw %ymm1,%ymm0,%ymm0
.byte 197,229,113,208,8 // vpsrlw $0x8,%ymm0,%ymm3
@@ -55089,7 +55011,7 @@ _sk_gather_g8_hsw_lowp:
.byte 196,227,121,32,192,15 // vpinsrb $0xf,%eax,%xmm0,%xmm0
.byte 196,226,125,48,192 // vpmovzxbw %xmm0,%ymm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,84,187,0,0 // vpbroadcastw 0xbb54(%rip),%ymm3 # 39f0e <_sk_srcover_bgra_8888_sse2_lowp+0xa86>
+ .byte 196,226,125,121,29,116,175,0,0 // vpbroadcastw 0xaf74(%rip),%ymm3 # 3918e <_sk_srcover_bgra_8888_sse2_lowp+0xa8a>
.byte 197,253,111,200 // vmovdqa %ymm0,%ymm1
.byte 197,253,111,208 // vmovdqa %ymm0,%ymm2
.byte 255,224 // jmpq *%rax
@@ -55100,13 +55022,13 @@ FUNCTION(_sk_scale_1_float_hsw_lowp)
_sk_scale_1_float_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,122,16,0 // vmovss (%rax),%xmm8
- .byte 197,122,16,13,146,179,0,0 // vmovss 0xb392(%rip),%xmm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,98,57,169,13,89,179,0,0 // vfmadd213ss 0xb359(%rip),%xmm8,%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,122,16,13,174,167,0,0 // vmovss 0xa7ae(%rip),%xmm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,57,169,13,117,167,0,0 // vfmadd213ss 0xa775(%rip),%xmm8,%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,122,44,193 // vcvttss2si %xmm9,%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
.byte 196,66,125,121,192 // vpbroadcastw %xmm8,%ymm8
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
- .byte 196,98,125,121,13,26,187,0,0 // vpbroadcastw 0xbb1a(%rip),%ymm9 # 39f10 <_sk_srcover_bgra_8888_sse2_lowp+0xa88>
+ .byte 196,98,125,121,13,58,175,0,0 // vpbroadcastw 0xaf3a(%rip),%ymm9 # 39190 <_sk_srcover_bgra_8888_sse2_lowp+0xa8c>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,189,213,201 // vpmullw %ymm1,%ymm8,%ymm1
@@ -55127,12 +55049,12 @@ FUNCTION(_sk_lerp_1_float_hsw_lowp)
_sk_lerp_1_float_hsw_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,122,16,0 // vmovss (%rax),%xmm8
- .byte 197,122,16,13,40,179,0,0 // vmovss 0xb328(%rip),%xmm9 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,98,57,169,13,239,178,0,0 // vfmadd213ss 0xb2ef(%rip),%xmm8,%xmm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,122,16,13,68,167,0,0 // vmovss 0xa744(%rip),%xmm9 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,57,169,13,11,167,0,0 // vfmadd213ss 0xa70b(%rip),%xmm8,%xmm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,193,122,44,193 // vcvttss2si %xmm9,%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
.byte 196,66,125,121,192 // vpbroadcastw %xmm8,%ymm8
- .byte 196,98,125,121,13,182,186,0,0 // vpbroadcastw 0xbab6(%rip),%ymm9 # 39f12 <_sk_srcover_bgra_8888_sse2_lowp+0xa8a>
+ .byte 196,98,125,121,13,214,174,0,0 // vpbroadcastw 0xaed6(%rip),%ymm9 # 39192 <_sk_srcover_bgra_8888_sse2_lowp+0xa8e>
.byte 196,65,53,249,208 // vpsubw %ymm8,%ymm9,%ymm10
.byte 197,45,213,220 // vpmullw %ymm4,%ymm10,%ymm11
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
@@ -55169,38 +55091,38 @@ _sk_scale_u8_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,38 // ja 2e4fa <_sk_scale_u8_hsw_lowp+0x3d>
+ .byte 119,38 // ja 2e35a <_sk_scale_u8_hsw_lowp+0x3d>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,21,1,0,0 // lea 0x115(%rip),%r9 # 2e5f8 <_sk_scale_u8_hsw_lowp+0x13b>
+ .byte 76,141,13,21,1,0,0 // lea 0x115(%rip),%r9 # 2e458 <_sk_scale_u8_hsw_lowp+0x13b>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 233,179,0,0,0 // jmpq 2e5ad <_sk_scale_u8_hsw_lowp+0xf0>
+ .byte 233,179,0,0,0 // jmpq 2e40d <_sk_scale_u8_hsw_lowp+0xf0>
.byte 196,65,122,111,4,16 // vmovdqu (%r8,%rdx,1),%xmm8
- .byte 233,168,0,0,0 // jmpq 2e5ad <_sk_scale_u8_hsw_lowp+0xf0>
+ .byte 233,168,0,0,0 // jmpq 2e40d <_sk_scale_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm8,%xmm8
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,121,110,200 // vmovd %eax,%xmm9
.byte 196,67,57,14,193,1 // vpblendw $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,135,0,0,0 // jmpq 2e5ad <_sk_scale_u8_hsw_lowp+0xf0>
+ .byte 233,135,0,0,0 // jmpq 2e40d <_sk_scale_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,65,121,110,12,16 // vmovd (%r8,%rdx,1),%xmm9
.byte 196,67,57,2,193,1 // vpblendd $0x1,%xmm9,%xmm8,%xmm8
- .byte 235,92 // jmp 2e5ad <_sk_scale_u8_hsw_lowp+0xf0>
+ .byte 235,92 // jmp 2e40d <_sk_scale_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,65,122,126,12,16 // vmovq (%r8,%rdx,1),%xmm9
.byte 196,67,49,2,192,12 // vpblendd $0xc,%xmm8,%xmm9,%xmm8
- .byte 235,49 // jmp 2e5ad <_sk_scale_u8_hsw_lowp+0xf0>
+ .byte 235,49 // jmp 2e40d <_sk_scale_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm8,%xmm8
@@ -55210,7 +55132,7 @@ _sk_scale_u8_hsw_lowp:
.byte 196,67,49,2,192,8 // vpblendd $0x8,%xmm8,%xmm9,%xmm8
.byte 196,66,125,48,192 // vpmovzxbw %xmm8,%ymm8
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
- .byte 196,98,125,121,13,85,185,0,0 // vpbroadcastw 0xb955(%rip),%ymm9 # 39f14 <_sk_srcover_bgra_8888_sse2_lowp+0xa8c>
+ .byte 196,98,125,121,13,117,173,0,0 // vpbroadcastw 0xad75(%rip),%ymm9 # 39194 <_sk_srcover_bgra_8888_sse2_lowp+0xa90>
.byte 196,193,125,253,193 // vpaddw %ymm9,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,189,213,201 // vpmullw %ymm1,%ymm8,%ymm1
@@ -55231,7 +55153,7 @@ _sk_scale_u8_hsw_lowp:
.byte 255,26 // lcall *(%rdx)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,75 // decl 0x4bffffff(%rip) # 4c02e604 <_sk_srcover_bgra_8888_sse2_lowp+0x4bff517c>
+ .byte 255,13,255,255,255,75 // decl 0x4bffffff(%rip) # 4c02e464 <_sk_srcover_bgra_8888_sse2_lowp+0x4bff5d60>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,67,255 // incl -0x1(%rbx)
@@ -55270,38 +55192,38 @@ _sk_lerp_u8_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,38 // ja 2e671 <_sk_lerp_u8_hsw_lowp+0x3d>
+ .byte 119,38 // ja 2e4d1 <_sk_lerp_u8_hsw_lowp+0x3d>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,58,1,0,0 // lea 0x13a(%rip),%r9 # 2e794 <_sk_lerp_u8_hsw_lowp+0x160>
+ .byte 76,141,13,58,1,0,0 // lea 0x13a(%rip),%r9 # 2e5f4 <_sk_lerp_u8_hsw_lowp+0x160>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 233,179,0,0,0 // jmpq 2e724 <_sk_lerp_u8_hsw_lowp+0xf0>
+ .byte 233,179,0,0,0 // jmpq 2e584 <_sk_lerp_u8_hsw_lowp+0xf0>
.byte 196,65,122,111,4,16 // vmovdqu (%r8,%rdx,1),%xmm8
- .byte 233,168,0,0,0 // jmpq 2e724 <_sk_lerp_u8_hsw_lowp+0xf0>
+ .byte 233,168,0,0,0 // jmpq 2e584 <_sk_lerp_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,2,2 // vpinsrb $0x2,0x2(%r8,%rdx,1),%xmm8,%xmm8
.byte 65,15,183,4,16 // movzwl (%r8,%rdx,1),%eax
.byte 197,121,110,200 // vmovd %eax,%xmm9
.byte 196,67,57,14,193,1 // vpblendw $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,135,0,0,0 // jmpq 2e724 <_sk_lerp_u8_hsw_lowp+0xf0>
+ .byte 233,135,0,0,0 // jmpq 2e584 <_sk_lerp_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,6,6 // vpinsrb $0x6,0x6(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,5,5 // vpinsrb $0x5,0x5(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,4,4 // vpinsrb $0x4,0x4(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,65,121,110,12,16 // vmovd (%r8,%rdx,1),%xmm9
.byte 196,67,57,2,193,1 // vpblendd $0x1,%xmm9,%xmm8,%xmm8
- .byte 235,92 // jmp 2e724 <_sk_lerp_u8_hsw_lowp+0xf0>
+ .byte 235,92 // jmp 2e584 <_sk_lerp_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,10,10 // vpinsrb $0xa,0xa(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,9,9 // vpinsrb $0x9,0x9(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,8,8 // vpinsrb $0x8,0x8(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,65,122,126,12,16 // vmovq (%r8,%rdx,1),%xmm9
.byte 196,67,49,2,192,12 // vpblendd $0xc,%xmm8,%xmm9,%xmm8
- .byte 235,49 // jmp 2e724 <_sk_lerp_u8_hsw_lowp+0xf0>
+ .byte 235,49 // jmp 2e584 <_sk_lerp_u8_hsw_lowp+0xf0>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,67,57,32,68,16,14,14 // vpinsrb $0xe,0xe(%r8,%rdx,1),%xmm8,%xmm8
.byte 196,67,57,32,68,16,13,13 // vpinsrb $0xd,0xd(%r8,%rdx,1),%xmm8,%xmm8
@@ -55310,7 +55232,7 @@ _sk_lerp_u8_hsw_lowp:
.byte 196,67,49,34,76,16,8,2 // vpinsrd $0x2,0x8(%r8,%rdx,1),%xmm9,%xmm9
.byte 196,67,49,2,192,8 // vpblendd $0x8,%xmm8,%xmm9,%xmm8
.byte 196,66,125,48,192 // vpmovzxbw %xmm8,%ymm8
- .byte 196,98,125,121,13,228,183,0,0 // vpbroadcastw 0xb7e4(%rip),%ymm9 # 39f16 <_sk_srcover_bgra_8888_sse2_lowp+0xa8e>
+ .byte 196,98,125,121,13,4,172,0,0 // vpbroadcastw 0xac04(%rip),%ymm9 # 39196 <_sk_srcover_bgra_8888_sse2_lowp+0xa92>
.byte 196,65,61,239,209 // vpxor %ymm9,%ymm8,%ymm10
.byte 197,45,213,220 // vpmullw %ymm4,%ymm10,%ymm11
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
@@ -55342,7 +55264,7 @@ _sk_lerp_u8_hsw_lowp:
.byte 254 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 232,254,255,255,38 // callq 2702e79f <_sk_srcover_bgra_8888_sse2_lowp+0x26ff5317>
+ .byte 232,254,255,255,38 // callq 2702e5ff <_sk_srcover_bgra_8888_sse2_lowp+0x26ff5efb>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,30 // lcall *(%rsi)
@@ -55363,7 +55285,7 @@ _sk_lerp_u8_hsw_lowp:
.byte 255,52,255 // pushq (%rdi,%rdi,8)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 124,255 // jl 2e7c1 <_sk_lerp_u8_hsw_lowp+0x18d>
+ .byte 124,255 // jl 2e621 <_sk_lerp_u8_hsw_lowp+0x18d>
.byte 255 // (bad)
.byte 255,116,255,255 // pushq -0x1(%rdi,%rdi,8)
.byte 255,108,255,255 // ljmp *-0x1(%rdi,%rdi,8)
@@ -55384,24 +55306,24 @@ _sk_scale_565_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,38 // ja 2e810 <_sk_scale_565_hsw_lowp+0x40>
+ .byte 119,38 // ja 2e670 <_sk_scale_565_hsw_lowp+0x40>
.byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,255,1,0,0 // lea 0x1ff(%rip),%r9 # 2e9f8 <_sk_scale_565_hsw_lowp+0x228>
+ .byte 76,141,13,255,1,0,0 // lea 0x1ff(%rip),%r9 # 2e858 <_sk_scale_565_hsw_lowp+0x228>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 233,12,1,0,0 // jmpq 2e91c <_sk_scale_565_hsw_lowp+0x14c>
+ .byte 233,12,1,0,0 // jmpq 2e77c <_sk_scale_565_hsw_lowp+0x14c>
.byte 196,65,126,111,4,80 // vmovdqu (%r8,%rdx,2),%ymm8
- .byte 233,1,1,0,0 // jmpq 2e91c <_sk_scale_565_hsw_lowp+0x14c>
+ .byte 233,1,1,0,0 // jmpq 2e77c <_sk_scale_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,53,56,192,1 // vinserti128 $0x1,%xmm8,%ymm9,%ymm8
.byte 196,65,121,110,12,80 // vmovd (%r8,%rdx,2),%xmm9
.byte 196,67,61,2,193,1 // vpblendd $0x1,%ymm9,%ymm8,%ymm8
- .byte 233,221,0,0,0 // jmpq 2e91c <_sk_scale_565_hsw_lowp+0x14c>
+ .byte 233,221,0,0,0 // jmpq 2e77c <_sk_scale_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,53,56,192,1 // vinserti128 $0x1,%xmm8,%ymm9,%ymm8
@@ -55411,7 +55333,7 @@ _sk_scale_565_hsw_lowp:
.byte 196,67,61,2,193,15 // vpblendd $0xf,%ymm9,%ymm8,%ymm8
.byte 196,65,122,126,12,80 // vmovq (%r8,%rdx,2),%xmm9
.byte 196,67,61,2,193,3 // vpblendd $0x3,%ymm9,%ymm8,%ymm8
- .byte 233,157,0,0,0 // jmpq 2e91c <_sk_scale_565_hsw_lowp+0x14c>
+ .byte 233,157,0,0,0 // jmpq 2e77c <_sk_scale_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
@@ -55423,7 +55345,7 @@ _sk_scale_565_hsw_lowp:
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
.byte 196,65,122,111,12,80 // vmovdqu (%r8,%rdx,2),%xmm9
.byte 196,67,53,2,192,240 // vpblendd $0xf0,%ymm8,%ymm9,%ymm8
- .byte 235,84 // jmp 2e91c <_sk_scale_565_hsw_lowp+0x14c>
+ .byte 235,84 // jmp 2e77c <_sk_scale_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
@@ -55437,13 +55359,13 @@ _sk_scale_565_hsw_lowp:
.byte 196,65,122,111,20,80 // vmovdqu (%r8,%rdx,2),%xmm10
.byte 196,67,45,56,201,1 // vinserti128 $0x1,%xmm9,%ymm10,%ymm9
.byte 196,67,53,2,192,192 // vpblendd $0xc0,%ymm8,%ymm9,%ymm8
- .byte 196,98,125,121,13,243,181,0,0 // vpbroadcastw 0xb5f3(%rip),%ymm9 # 39f18 <_sk_srcover_bgra_8888_sse2_lowp+0xa90>
+ .byte 196,98,125,121,13,19,170,0,0 // vpbroadcastw 0xaa13(%rip),%ymm9 # 39198 <_sk_srcover_bgra_8888_sse2_lowp+0xa94>
.byte 196,193,45,113,208,8 // vpsrlw $0x8,%ymm8,%ymm10
.byte 196,65,45,219,201 // vpand %ymm9,%ymm10,%ymm9
.byte 196,193,45,113,208,5 // vpsrlw $0x5,%ymm8,%ymm10
- .byte 196,98,125,121,29,219,181,0,0 // vpbroadcastw 0xb5db(%rip),%ymm11 # 39f1a <_sk_srcover_bgra_8888_sse2_lowp+0xa92>
+ .byte 196,98,125,121,29,251,169,0,0 // vpbroadcastw 0xa9fb(%rip),%ymm11 # 3919a <_sk_srcover_bgra_8888_sse2_lowp+0xa96>
.byte 196,65,45,219,211 // vpand %ymm11,%ymm10,%ymm10
- .byte 196,98,125,121,29,207,181,0,0 // vpbroadcastw 0xb5cf(%rip),%ymm11 # 39f1c <_sk_srcover_bgra_8888_sse2_lowp+0xa94>
+ .byte 196,98,125,121,29,239,169,0,0 // vpbroadcastw 0xa9ef(%rip),%ymm11 # 3919c <_sk_srcover_bgra_8888_sse2_lowp+0xa98>
.byte 196,65,61,219,219 // vpand %ymm11,%ymm8,%ymm11
.byte 196,193,61,113,208,13 // vpsrlw $0xd,%ymm8,%ymm8
.byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
@@ -55453,7 +55375,7 @@ _sk_scale_565_hsw_lowp:
.byte 196,193,45,113,243,3 // vpsllw $0x3,%ymm11,%ymm10
.byte 196,193,37,113,211,2 // vpsrlw $0x2,%ymm11,%ymm11
.byte 196,65,45,235,211 // vpor %ymm11,%ymm10,%ymm10
- .byte 196,98,125,121,29,150,181,0,0 // vpbroadcastw 0xb596(%rip),%ymm11 # 39f1e <_sk_srcover_bgra_8888_sse2_lowp+0xa96>
+ .byte 196,98,125,121,29,182,169,0,0 // vpbroadcastw 0xa9b6(%rip),%ymm11 # 3919e <_sk_srcover_bgra_8888_sse2_lowp+0xa9a>
.byte 196,65,101,239,227 // vpxor %ymm11,%ymm3,%ymm12
.byte 196,65,69,239,219 // vpxor %ymm11,%ymm7,%ymm11
.byte 196,65,37,101,220 // vpcmpgtw %ymm12,%ymm11,%ymm11
@@ -55463,7 +55385,7 @@ _sk_scale_565_hsw_lowp:
.byte 196,66,21,62,232 // vpmaxuw %ymm8,%ymm13,%ymm13
.byte 196,67,21,76,220,176 // vpblendvb %ymm11,%ymm12,%ymm13,%ymm11
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
- .byte 196,98,125,121,5,98,181,0,0 // vpbroadcastw 0xb562(%rip),%ymm8 # 39f20 <_sk_srcover_bgra_8888_sse2_lowp+0xa98>
+ .byte 196,98,125,121,5,130,169,0,0 // vpbroadcastw 0xa982(%rip),%ymm8 # 391a0 <_sk_srcover_bgra_8888_sse2_lowp+0xa9c>
.byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
.byte 197,253,113,208,8 // vpsrlw $0x8,%ymm0,%ymm0
.byte 197,181,213,201 // vpmullw %ymm1,%ymm9,%ymm1
@@ -55527,24 +55449,24 @@ _sk_lerp_565_hsw_lowp:
.byte 36,15 // and $0xf,%al
.byte 254,200 // dec %al
.byte 60,14 // cmp $0xe,%al
- .byte 119,38 // ja 2ea74 <_sk_lerp_565_hsw_lowp+0x40>
+ .byte 119,38 // ja 2e8d4 <_sk_lerp_565_hsw_lowp+0x40>
.byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,51,2,0,0 // lea 0x233(%rip),%r9 # 2ec90 <_sk_lerp_565_hsw_lowp+0x25c>
+ .byte 76,141,13,51,2,0,0 // lea 0x233(%rip),%r9 # 2eaf0 <_sk_lerp_565_hsw_lowp+0x25c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 233,12,1,0,0 // jmpq 2eb80 <_sk_lerp_565_hsw_lowp+0x14c>
+ .byte 233,12,1,0,0 // jmpq 2e9e0 <_sk_lerp_565_hsw_lowp+0x14c>
.byte 196,65,126,111,4,80 // vmovdqu (%r8,%rdx,2),%ymm8
- .byte 233,1,1,0,0 // jmpq 2eb80 <_sk_lerp_565_hsw_lowp+0x14c>
+ .byte 233,1,1,0,0 // jmpq 2e9e0 <_sk_lerp_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,4,2 // vpinsrw $0x2,0x4(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,53,56,192,1 // vinserti128 $0x1,%xmm8,%ymm9,%ymm8
.byte 196,65,121,110,12,80 // vmovd (%r8,%rdx,2),%xmm9
.byte 196,67,61,2,193,1 // vpblendd $0x1,%ymm9,%ymm8,%ymm8
- .byte 233,221,0,0,0 // jmpq 2eb80 <_sk_lerp_565_hsw_lowp+0x14c>
+ .byte 233,221,0,0,0 // jmpq 2e9e0 <_sk_lerp_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,12,6 // vpinsrw $0x6,0xc(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,53,56,192,1 // vinserti128 $0x1,%xmm8,%ymm9,%ymm8
@@ -55554,7 +55476,7 @@ _sk_lerp_565_hsw_lowp:
.byte 196,67,61,2,193,15 // vpblendd $0xf,%ymm9,%ymm8,%ymm8
.byte 196,65,122,126,12,80 // vmovq (%r8,%rdx,2),%xmm9
.byte 196,67,61,2,193,3 // vpblendd $0x3,%ymm9,%ymm8,%ymm8
- .byte 233,157,0,0,0 // jmpq 2eb80 <_sk_lerp_565_hsw_lowp+0x14c>
+ .byte 233,157,0,0,0 // jmpq 2e9e0 <_sk_lerp_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,20,2 // vpinsrw $0x2,0x14(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
@@ -55566,7 +55488,7 @@ _sk_lerp_565_hsw_lowp:
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
.byte 196,65,122,111,12,80 // vmovdqu (%r8,%rdx,2),%xmm9
.byte 196,67,53,2,192,240 // vpblendd $0xf0,%ymm8,%ymm9,%ymm8
- .byte 235,84 // jmp 2eb80 <_sk_lerp_565_hsw_lowp+0x14c>
+ .byte 235,84 // jmp 2e9e0 <_sk_lerp_565_hsw_lowp+0x14c>
.byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
.byte 196,65,57,196,76,80,28,6 // vpinsrw $0x6,0x1c(%r8,%rdx,2),%xmm8,%xmm9
.byte 196,67,61,56,193,1 // vinserti128 $0x1,%xmm9,%ymm8,%ymm8
@@ -55580,13 +55502,13 @@ _sk_lerp_565_hsw_lowp:
.byte 196,65,122,111,20,80 // vmovdqu (%r8,%rdx,2),%xmm10
.byte 196,67,45,56,201,1 // vinserti128 $0x1,%xmm9,%ymm10,%ymm9
.byte 196,67,53,2,192,192 // vpblendd $0xc0,%ymm8,%ymm9,%ymm8
- .byte 196,98,125,121,13,153,179,0,0 // vpbroadcastw 0xb399(%rip),%ymm9 # 39f22 <_sk_srcover_bgra_8888_sse2_lowp+0xa9a>
+ .byte 196,98,125,121,13,185,167,0,0 // vpbroadcastw 0xa7b9(%rip),%ymm9 # 391a2 <_sk_srcover_bgra_8888_sse2_lowp+0xa9e>
.byte 196,193,45,113,208,8 // vpsrlw $0x8,%ymm8,%ymm10
.byte 196,65,45,219,201 // vpand %ymm9,%ymm10,%ymm9
.byte 196,193,45,113,208,5 // vpsrlw $0x5,%ymm8,%ymm10
- .byte 196,98,125,121,29,129,179,0,0 // vpbroadcastw 0xb381(%rip),%ymm11 # 39f24 <_sk_srcover_bgra_8888_sse2_lowp+0xa9c>
+ .byte 196,98,125,121,29,161,167,0,0 // vpbroadcastw 0xa7a1(%rip),%ymm11 # 391a4 <_sk_srcover_bgra_8888_sse2_lowp+0xaa0>
.byte 196,65,45,219,211 // vpand %ymm11,%ymm10,%ymm10
- .byte 196,98,125,121,29,117,179,0,0 // vpbroadcastw 0xb375(%rip),%ymm11 # 39f26 <_sk_srcover_bgra_8888_sse2_lowp+0xa9e>
+ .byte 196,98,125,121,29,149,167,0,0 // vpbroadcastw 0xa795(%rip),%ymm11 # 391a6 <_sk_srcover_bgra_8888_sse2_lowp+0xaa2>
.byte 196,65,61,219,219 // vpand %ymm11,%ymm8,%ymm11
.byte 196,193,61,113,208,13 // vpsrlw $0xd,%ymm8,%ymm8
.byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
@@ -55596,7 +55518,7 @@ _sk_lerp_565_hsw_lowp:
.byte 196,193,45,113,243,3 // vpsllw $0x3,%ymm11,%ymm10
.byte 196,193,37,113,211,2 // vpsrlw $0x2,%ymm11,%ymm11
.byte 196,65,45,235,211 // vpor %ymm11,%ymm10,%ymm10
- .byte 196,98,125,121,29,60,179,0,0 // vpbroadcastw 0xb33c(%rip),%ymm11 # 39f28 <_sk_srcover_bgra_8888_sse2_lowp+0xaa0>
+ .byte 196,98,125,121,29,92,167,0,0 // vpbroadcastw 0xa75c(%rip),%ymm11 # 391a8 <_sk_srcover_bgra_8888_sse2_lowp+0xaa4>
.byte 196,65,101,239,227 // vpxor %ymm11,%ymm3,%ymm12
.byte 196,65,69,239,219 // vpxor %ymm11,%ymm7,%ymm11
.byte 196,65,37,101,220 // vpcmpgtw %ymm12,%ymm11,%ymm11
@@ -55605,7 +55527,7 @@ _sk_lerp_565_hsw_lowp:
.byte 196,66,53,62,234 // vpmaxuw %ymm10,%ymm9,%ymm13
.byte 196,66,21,62,232 // vpmaxuw %ymm8,%ymm13,%ymm13
.byte 196,67,21,76,220,176 // vpblendvb %ymm11,%ymm12,%ymm13,%ymm11
- .byte 196,98,125,121,37,12,179,0,0 // vpbroadcastw 0xb30c(%rip),%ymm12 # 39f2a <_sk_srcover_bgra_8888_sse2_lowp+0xaa2>
+ .byte 196,98,125,121,37,44,167,0,0 // vpbroadcastw 0xa72c(%rip),%ymm12 # 391aa <_sk_srcover_bgra_8888_sse2_lowp+0xaa6>
.byte 196,65,61,239,236 // vpxor %ymm12,%ymm8,%ymm13
.byte 197,21,213,236 // vpmullw %ymm4,%ymm13,%ymm13
.byte 197,189,213,192 // vpmullw %ymm0,%ymm8,%ymm0
@@ -55678,7 +55600,7 @@ _sk_clamp_x_1_hsw_lowp:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,84,170,0,0 // vbroadcastss 0xaa54(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,112,158,0,0 // vbroadcastss 0x9e70(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -55695,7 +55617,7 @@ _sk_repeat_x_1_hsw_lowp:
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,24,170,0,0 // vbroadcastss 0xaa18(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,52,158,0,0 // vbroadcastss 0x9e34(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -55705,10 +55627,10 @@ HIDDEN _sk_mirror_x_1_hsw_lowp
.globl _sk_mirror_x_1_hsw_lowp
FUNCTION(_sk_mirror_x_1_hsw_lowp)
_sk_mirror_x_1_hsw_lowp:
- .byte 196,98,125,24,5,25,170,0,0 // vbroadcastss 0xaa19(%rip),%ymm8 # 39750 <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
+ .byte 196,98,125,24,5,53,158,0,0 // vbroadcastss 0x9e35(%rip),%ymm8 # 389cc <_sk_srcover_bgra_8888_sse2_lowp+0x2c8>
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
.byte 196,193,116,88,200 // vaddps %ymm8,%ymm1,%ymm1
- .byte 196,98,125,24,13,234,169,0,0 // vbroadcastss 0xa9ea(%rip),%ymm9 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,13,6,158,0,0 // vbroadcastss 0x9e06(%rip),%ymm9 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,65,116,89,209 // vmulps %ymm9,%ymm1,%ymm10
.byte 196,65,124,89,201 // vmulps %ymm9,%ymm0,%ymm9
.byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
@@ -55719,13 +55641,13 @@ _sk_mirror_x_1_hsw_lowp:
.byte 196,193,116,92,202 // vsubps %ymm10,%ymm1,%ymm1
.byte 196,193,116,88,200 // vaddps %ymm8,%ymm1,%ymm1
.byte 196,193,124,88,192 // vaddps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,65,171,0,0 // vbroadcastss 0xab41(%rip),%ymm8 # 398c8 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
+ .byte 196,98,125,24,5,93,159,0,0 // vbroadcastss 0x9f5d(%rip),%ymm8 # 38b44 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
.byte 196,193,124,84,192 // vandps %ymm8,%ymm0,%ymm0
.byte 196,193,116,84,200 // vandps %ymm8,%ymm1,%ymm1
.byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
.byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
.byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
- .byte 196,98,125,24,5,143,169,0,0 // vbroadcastss 0xa98f(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,171,157,0,0 // vbroadcastss 0x9dab(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
.byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -55735,824 +55657,286 @@ HIDDEN _sk_gradient_hsw_lowp
.globl _sk_gradient_hsw_lowp
FUNCTION(_sk_gradient_hsw_lowp)
_sk_gradient_hsw_lowp:
- .byte 85 // push %rbp
- .byte 65,87 // push %r15
- .byte 65,86 // push %r14
- .byte 65,85 // push %r13
- .byte 65,84 // push %r12
- .byte 83 // push %rbx
- .byte 72,129,236,40,1,0,0 // sub $0x128,%rsp
- .byte 197,252,17,188,36,0,1,0,0 // vmovups %ymm7,0x100(%rsp)
- .byte 197,252,17,180,36,224,0,0,0 // vmovups %ymm6,0xe0(%rsp)
- .byte 197,252,17,172,36,192,0,0,0 // vmovups %ymm5,0xc0(%rsp)
- .byte 197,252,17,164,36,160,0,0,0 // vmovups %ymm4,0xa0(%rsp)
- .byte 72,137,76,36,56 // mov %rcx,0x38(%rsp)
- .byte 72,137,84,36,48 // mov %rdx,0x30(%rsp)
- .byte 72,137,124,36,40 // mov %rdi,0x28(%rsp)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,137,116,36,32 // mov %rsi,0x20(%rsp)
- .byte 73,137,193 // mov %rax,%r9
- .byte 73,139,1 // mov (%r9),%rax
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
+ .byte 72,129,236,184,0,0,0 // sub $0xb8,%rsp
+ .byte 197,252,17,188,36,128,0,0,0 // vmovups %ymm7,0x80(%rsp)
+ .byte 197,252,17,116,36,96 // vmovups %ymm6,0x60(%rsp)
+ .byte 197,254,127,108,36,64 // vmovdqu %ymm5,0x40(%rsp)
+ .byte 197,252,17,100,36,32 // vmovups %ymm4,0x20(%rsp)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,0 // mov (%rax),%r8
+ .byte 196,65,28,87,228 // vxorps %ymm12,%ymm12,%ymm12
.byte 197,237,239,210 // vpxor %ymm2,%ymm2,%ymm2
- .byte 72,131,248,2 // cmp $0x2,%rax
- .byte 114,51 // jb 2ee49 <_sk_gradient_hsw_lowp+0x92>
- .byte 73,139,73,72 // mov 0x48(%r9),%rcx
- .byte 72,255,200 // dec %rax
- .byte 72,131,193,4 // add $0x4,%rcx
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
+ .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
+ .byte 73,131,248,2 // cmp $0x2,%r8
+ .byte 114,51 // jb 2ec84 <_sk_gradient_hsw_lowp+0x6d>
+ .byte 76,139,72,72 // mov 0x48(%rax),%r9
+ .byte 73,255,200 // dec %r8
+ .byte 73,131,193,4 // add $0x4,%r9
.byte 197,237,239,210 // vpxor %ymm2,%ymm2,%ymm2
- .byte 196,226,125,24,41 // vbroadcastss (%rcx),%ymm5
- .byte 197,212,194,240,2 // vcmpleps %ymm0,%ymm5,%ymm6
- .byte 197,229,250,222 // vpsubd %ymm6,%ymm3,%ymm3
- .byte 197,212,194,233,2 // vcmpleps %ymm1,%ymm5,%ymm5
- .byte 197,237,250,213 // vpsubd %ymm5,%ymm2,%ymm2
- .byte 72,131,193,4 // add $0x4,%rcx
- .byte 72,255,200 // dec %rax
- .byte 117,224 // jne 2ee29 <_sk_gradient_hsw_lowp+0x72>
- .byte 196,227,249,22,219,1 // vpextrq $0x1,%xmm3,%rbx
- .byte 137,216 // mov %ebx,%eax
- .byte 72,137,68,36,160 // mov %rax,-0x60(%rsp)
- .byte 72,193,235,32 // shr $0x20,%rbx
- .byte 72,137,92,36,240 // mov %rbx,-0x10(%rsp)
- .byte 196,225,249,126,223 // vmovq %xmm3,%rdi
- .byte 137,248 // mov %edi,%eax
- .byte 72,137,68,36,224 // mov %rax,-0x20(%rsp)
- .byte 72,193,239,32 // shr $0x20,%rdi
- .byte 72,137,124,36,232 // mov %rdi,-0x18(%rsp)
- .byte 196,227,125,57,219,1 // vextracti128 $0x1,%ymm3,%xmm3
- .byte 196,227,249,22,218,1 // vpextrq $0x1,%xmm3,%rdx
- .byte 137,215 // mov %edx,%edi
- .byte 72,193,234,32 // shr $0x20,%rdx
- .byte 196,225,249,126,216 // vmovq %xmm3,%rax
- .byte 65,137,198 // mov %eax,%r14d
- .byte 72,193,232,32 // shr $0x20,%rax
- .byte 72,137,195 // mov %rax,%rbx
- .byte 196,195,249,22,208,1 // vpextrq $0x1,%xmm2,%r8
- .byte 68,137,193 // mov %r8d,%ecx
- .byte 72,137,76,36,152 // mov %rcx,-0x68(%rsp)
- .byte 73,193,232,32 // shr $0x20,%r8
- .byte 76,137,68,36,200 // mov %r8,-0x38(%rsp)
- .byte 196,193,249,126,210 // vmovq %xmm2,%r10
- .byte 69,137,212 // mov %r10d,%r12d
- .byte 73,193,234,32 // shr $0x20,%r10
- .byte 76,137,84,36,168 // mov %r10,-0x58(%rsp)
- .byte 196,227,125,57,210,1 // vextracti128 $0x1,%ymm2,%xmm2
- .byte 196,195,249,22,215,1 // vpextrq $0x1,%xmm2,%r15
- .byte 68,137,254 // mov %r15d,%esi
- .byte 73,193,239,32 // shr $0x20,%r15
- .byte 196,193,249,126,213 // vmovq %xmm2,%r13
- .byte 69,137,235 // mov %r13d,%r11d
- .byte 73,193,237,32 // shr $0x20,%r13
- .byte 76,137,76,36,192 // mov %r9,-0x40(%rsp)
- .byte 77,139,73,8 // mov 0x8(%r9),%r9
- .byte 196,129,122,16,20,153 // vmovss (%r9,%r11,4),%xmm2
- .byte 196,3,105,33,4,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm2,%xmm8
- .byte 196,65,122,16,12,177 // vmovss (%r9,%rsi,4),%xmm9
- .byte 72,137,245 // mov %rsi,%rbp
- .byte 72,137,108,36,176 // mov %rbp,-0x50(%rsp)
- .byte 196,1,122,16,20,185 // vmovss (%r9,%r15,4),%xmm10
- .byte 196,129,122,16,52,161 // vmovss (%r9,%r12,4),%xmm6
- .byte 196,3,73,33,28,145,16 // vinsertps $0x10,(%r9,%r10,4),%xmm6,%xmm11
- .byte 196,65,122,16,36,137 // vmovss (%r9,%rcx,4),%xmm12
- .byte 196,1,122,16,44,129 // vmovss (%r9,%r8,4),%xmm13
- .byte 76,137,246 // mov %r14,%rsi
- .byte 196,193,122,16,28,177 // vmovss (%r9,%rsi,4),%xmm3
- .byte 196,195,97,33,28,153,16 // vinsertps $0x10,(%r9,%rbx,4),%xmm3,%xmm3
- .byte 72,137,92,36,208 // mov %rbx,-0x30(%rsp)
- .byte 72,137,124,36,184 // mov %rdi,-0x48(%rsp)
- .byte 196,67,97,33,52,185,32 // vinsertps $0x20,(%r9,%rdi,4),%xmm3,%xmm14
- .byte 196,193,122,16,44,145 // vmovss (%r9,%rdx,4),%xmm5
- .byte 76,139,84,36,224 // mov -0x20(%rsp),%r10
- .byte 196,129,122,16,52,145 // vmovss (%r9,%r10,4),%xmm6
- .byte 76,139,68,36,232 // mov -0x18(%rsp),%r8
- .byte 196,131,73,33,52,129,16 // vinsertps $0x10,(%r9,%r8,4),%xmm6,%xmm6
- .byte 72,139,76,36,160 // mov -0x60(%rsp),%rcx
- .byte 196,195,73,33,52,137,32 // vinsertps $0x20,(%r9,%rcx,4),%xmm6,%xmm6
- .byte 72,139,68,36,240 // mov -0x10(%rsp),%rax
- .byte 196,193,122,16,20,129 // vmovss (%r9,%rax,4),%xmm2
- .byte 196,195,57,33,249,32 // vinsertps $0x20,%xmm9,%xmm8,%xmm7
- .byte 196,195,65,33,250,48 // vinsertps $0x30,%xmm10,%xmm7,%xmm7
- .byte 196,195,33,33,220,32 // vinsertps $0x20,%xmm12,%xmm11,%xmm3
- .byte 196,195,97,33,221,48 // vinsertps $0x30,%xmm13,%xmm3,%xmm3
- .byte 196,227,101,24,223,1 // vinsertf128 $0x1,%xmm7,%ymm3,%ymm3
- .byte 197,252,17,156,36,128,0,0,0 // vmovups %ymm3,0x80(%rsp)
- .byte 196,227,9,33,221,48 // vinsertps $0x30,%xmm5,%xmm14,%xmm3
- .byte 196,227,73,33,210,48 // vinsertps $0x30,%xmm2,%xmm6,%xmm2
- .byte 196,99,109,24,211,1 // vinsertf128 $0x1,%xmm3,%ymm2,%ymm10
- .byte 76,139,116,36,192 // mov -0x40(%rsp),%r14
- .byte 77,139,78,16 // mov 0x10(%r14),%r9
- .byte 196,129,122,16,20,153 // vmovss (%r9,%r11,4),%xmm2
- .byte 196,131,105,33,20,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm2,%xmm2
- .byte 196,193,122,16,28,169 // vmovss (%r9,%rbp,4),%xmm3
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,185 // vmovss (%r9,%r15,4),%xmm3
- .byte 196,99,105,33,195,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm8
- .byte 196,129,122,16,28,161 // vmovss (%r9,%r12,4),%xmm3
- .byte 72,139,108,36,168 // mov -0x58(%rsp),%rbp
- .byte 196,67,97,33,12,169,16 // vinsertps $0x10,(%r9,%rbp,4),%xmm3,%xmm9
- .byte 72,139,108,36,152 // mov -0x68(%rsp),%rbp
- .byte 196,193,122,16,44,169 // vmovss (%r9,%rbp,4),%xmm5
- .byte 72,139,108,36,200 // mov -0x38(%rsp),%rbp
- .byte 196,193,122,16,52,169 // vmovss (%r9,%rbp,4),%xmm6
- .byte 196,193,122,16,60,177 // vmovss (%r9,%rsi,4),%xmm7
- .byte 72,137,245 // mov %rsi,%rbp
- .byte 72,137,108,36,248 // mov %rbp,-0x8(%rsp)
- .byte 196,195,65,33,60,153,16 // vinsertps $0x10,(%r9,%rbx,4),%xmm7,%xmm7
- .byte 196,195,65,33,60,185,32 // vinsertps $0x20,(%r9,%rdi,4),%xmm7,%xmm7
- .byte 196,193,122,16,20,145 // vmovss (%r9,%rdx,4),%xmm2
- .byte 72,137,214 // mov %rdx,%rsi
- .byte 72,137,116,36,216 // mov %rsi,-0x28(%rsp)
- .byte 76,137,210 // mov %r10,%rdx
- .byte 196,193,122,16,36,145 // vmovss (%r9,%rdx,4),%xmm4
- .byte 76,137,199 // mov %r8,%rdi
- .byte 196,195,89,33,36,185,16 // vinsertps $0x10,(%r9,%rdi,4),%xmm4,%xmm4
- .byte 196,195,89,33,36,137,32 // vinsertps $0x20,(%r9,%rcx,4),%xmm4,%xmm4
- .byte 196,193,122,16,28,129 // vmovss (%r9,%rax,4),%xmm3
- .byte 72,137,195 // mov %rax,%rbx
- .byte 196,227,49,33,237,32 // vinsertps $0x20,%xmm5,%xmm9,%xmm5
- .byte 196,227,81,33,238,48 // vinsertps $0x30,%xmm6,%xmm5,%xmm5
- .byte 196,195,85,24,232,1 // vinsertf128 $0x1,%xmm8,%ymm5,%ymm5
- .byte 197,252,17,108,36,96 // vmovups %ymm5,0x60(%rsp)
- .byte 196,227,65,33,210,48 // vinsertps $0x30,%xmm2,%xmm7,%xmm2
- .byte 196,227,89,33,219,48 // vinsertps $0x30,%xmm3,%xmm4,%xmm3
- .byte 196,99,101,24,218,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm11
- .byte 76,137,241 // mov %r14,%rcx
- .byte 76,139,73,24 // mov 0x18(%rcx),%r9
- .byte 196,129,122,16,20,153 // vmovss (%r9,%r11,4),%xmm2
- .byte 196,131,105,33,20,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm2,%xmm2
- .byte 72,139,68,36,176 // mov -0x50(%rsp),%rax
- .byte 196,193,122,16,28,129 // vmovss (%r9,%rax,4),%xmm3
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,185 // vmovss (%r9,%r15,4),%xmm3
- .byte 196,99,105,33,195,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm8
- .byte 196,129,122,16,28,161 // vmovss (%r9,%r12,4),%xmm3
- .byte 77,137,224 // mov %r12,%r8
- .byte 76,139,84,36,168 // mov -0x58(%rsp),%r10
- .byte 196,131,97,33,28,145,16 // vinsertps $0x10,(%r9,%r10,4),%xmm3,%xmm3
- .byte 76,139,116,36,152 // mov -0x68(%rsp),%r14
- .byte 196,129,122,16,36,177 // vmovss (%r9,%r14,4),%xmm4
- .byte 196,227,97,33,220,32 // vinsertps $0x20,%xmm4,%xmm3,%xmm3
- .byte 76,139,100,36,200 // mov -0x38(%rsp),%r12
- .byte 196,129,122,16,36,161 // vmovss (%r9,%r12,4),%xmm4
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,193,122,16,36,169 // vmovss (%r9,%rbp,4),%xmm4
- .byte 72,139,108,36,208 // mov -0x30(%rsp),%rbp
- .byte 196,195,89,33,36,169,16 // vinsertps $0x10,(%r9,%rbp,4),%xmm4,%xmm4
- .byte 72,139,68,36,184 // mov -0x48(%rsp),%rax
- .byte 196,195,89,33,36,129,32 // vinsertps $0x20,(%r9,%rax,4),%xmm4,%xmm4
- .byte 196,193,122,16,44,177 // vmovss (%r9,%rsi,4),%xmm5
- .byte 72,137,214 // mov %rdx,%rsi
- .byte 196,193,122,16,60,177 // vmovss (%r9,%rsi,4),%xmm7
- .byte 196,195,65,33,60,185,16 // vinsertps $0x10,(%r9,%rdi,4),%xmm7,%xmm7
- .byte 72,139,84,36,160 // mov -0x60(%rsp),%rdx
- .byte 196,195,65,33,60,145,32 // vinsertps $0x20,(%r9,%rdx,4),%xmm7,%xmm7
- .byte 196,193,122,16,20,153 // vmovss (%r9,%rbx,4),%xmm2
- .byte 196,195,101,24,216,1 // vinsertf128 $0x1,%xmm8,%ymm3,%ymm3
- .byte 197,252,17,92,36,64 // vmovups %ymm3,0x40(%rsp)
- .byte 196,227,89,33,221,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm3
- .byte 196,227,65,33,210,48 // vinsertps $0x30,%xmm2,%xmm7,%xmm2
- .byte 196,99,109,24,227,1 // vinsertf128 $0x1,%xmm3,%ymm2,%ymm12
- .byte 76,139,73,32 // mov 0x20(%rcx),%r9
- .byte 196,129,122,16,20,153 // vmovss (%r9,%r11,4),%xmm2
- .byte 76,137,92,36,24 // mov %r11,0x18(%rsp)
- .byte 196,131,105,33,20,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm2,%xmm2
- .byte 76,137,108,36,16 // mov %r13,0x10(%rsp)
- .byte 72,139,108,36,176 // mov -0x50(%rsp),%rbp
- .byte 196,193,122,16,28,169 // vmovss (%r9,%rbp,4),%xmm3
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,185 // vmovss (%r9,%r15,4),%xmm3
- .byte 76,137,60,36 // mov %r15,(%rsp)
- .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,129 // vmovss (%r9,%r8,4),%xmm3
- .byte 76,137,193 // mov %r8,%rcx
- .byte 72,137,76,36,8 // mov %rcx,0x8(%rsp)
- .byte 196,131,97,33,28,145,16 // vinsertps $0x10,(%r9,%r10,4),%xmm3,%xmm3
- .byte 196,129,122,16,36,177 // vmovss (%r9,%r14,4),%xmm4
- .byte 76,137,242 // mov %r14,%rdx
- .byte 196,227,97,33,220,32 // vinsertps $0x20,%xmm4,%xmm3,%xmm3
- .byte 196,129,122,16,36,161 // vmovss (%r9,%r12,4),%xmm4
- .byte 76,137,231 // mov %r12,%rdi
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,99,101,24,202,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm9
- .byte 76,139,116,36,248 // mov -0x8(%rsp),%r14
- .byte 196,129,122,16,20,177 // vmovss (%r9,%r14,4),%xmm2
- .byte 72,139,92,36,208 // mov -0x30(%rsp),%rbx
- .byte 196,195,105,33,20,153,16 // vinsertps $0x10,(%r9,%rbx,4),%xmm2,%xmm2
- .byte 196,195,105,33,20,129,32 // vinsertps $0x20,(%r9,%rax,4),%xmm2,%xmm2
- .byte 72,139,68,36,216 // mov -0x28(%rsp),%rax
- .byte 196,193,122,16,28,129 // vmovss (%r9,%rax,4),%xmm3
- .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
- .byte 196,193,122,16,28,177 // vmovss (%r9,%rsi,4),%xmm3
- .byte 72,139,68,36,232 // mov -0x18(%rsp),%rax
- .byte 196,195,97,33,28,129,16 // vinsertps $0x10,(%r9,%rax,4),%xmm3,%xmm3
- .byte 72,139,116,36,160 // mov -0x60(%rsp),%rsi
- .byte 196,195,97,33,28,177,32 // vinsertps $0x20,(%r9,%rsi,4),%xmm3,%xmm3
- .byte 76,139,100,36,240 // mov -0x10(%rsp),%r12
- .byte 196,129,122,16,36,161 // vmovss (%r9,%r12,4),%xmm4
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,99,101,24,242,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm14
- .byte 76,139,68,36,192 // mov -0x40(%rsp),%r8
- .byte 77,139,72,40 // mov 0x28(%r8),%r9
- .byte 196,129,122,16,20,153 // vmovss (%r9,%r11,4),%xmm2
- .byte 196,131,105,33,20,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm2,%xmm2
- .byte 196,193,122,16,28,169 // vmovss (%r9,%rbp,4),%xmm3
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,185 // vmovss (%r9,%r15,4),%xmm3
- .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
- .byte 196,193,122,16,28,137 // vmovss (%r9,%rcx,4),%xmm3
- .byte 196,131,97,33,28,145,16 // vinsertps $0x10,(%r9,%r10,4),%xmm3,%xmm3
- .byte 196,193,122,16,36,145 // vmovss (%r9,%rdx,4),%xmm4
- .byte 196,227,97,33,220,32 // vinsertps $0x20,%xmm4,%xmm3,%xmm3
- .byte 196,193,122,16,36,185 // vmovss (%r9,%rdi,4),%xmm4
- .byte 73,137,253 // mov %rdi,%r13
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,99,101,24,234,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm13
- .byte 196,129,122,16,20,177 // vmovss (%r9,%r14,4),%xmm2
- .byte 77,137,242 // mov %r14,%r10
- .byte 196,195,105,33,20,153,16 // vinsertps $0x10,(%r9,%rbx,4),%xmm2,%xmm2
- .byte 72,139,76,36,184 // mov -0x48(%rsp),%rcx
- .byte 196,195,105,33,20,137,32 // vinsertps $0x20,(%r9,%rcx,4),%xmm2,%xmm2
- .byte 72,139,124,36,216 // mov -0x28(%rsp),%rdi
- .byte 196,193,122,16,28,185 // vmovss (%r9,%rdi,4),%xmm3
- .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
- .byte 76,139,116,36,224 // mov -0x20(%rsp),%r14
- .byte 196,129,122,16,28,177 // vmovss (%r9,%r14,4),%xmm3
- .byte 196,195,97,33,28,129,16 // vinsertps $0x10,(%r9,%rax,4),%xmm3,%xmm3
- .byte 73,137,199 // mov %rax,%r15
- .byte 72,137,245 // mov %rsi,%rbp
- .byte 196,195,97,33,28,169,32 // vinsertps $0x20,(%r9,%rbp,4),%xmm3,%xmm3
- .byte 196,129,122,16,36,161 // vmovss (%r9,%r12,4),%xmm4
- .byte 77,137,227 // mov %r12,%r11
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,99,101,24,250,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm15
- .byte 77,139,72,48 // mov 0x30(%r8),%r9
- .byte 196,129,122,16,20,145 // vmovss (%r9,%r10,4),%xmm2
- .byte 196,195,105,33,20,153,16 // vinsertps $0x10,(%r9,%rbx,4),%xmm2,%xmm2
- .byte 196,195,105,33,20,137,32 // vinsertps $0x20,(%r9,%rcx,4),%xmm2,%xmm2
- .byte 72,137,206 // mov %rcx,%rsi
- .byte 73,137,252 // mov %rdi,%r12
- .byte 196,129,122,16,28,161 // vmovss (%r9,%r12,4),%xmm3
- .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
- .byte 196,129,122,16,28,177 // vmovss (%r9,%r14,4),%xmm3
- .byte 76,137,241 // mov %r14,%rcx
- .byte 196,131,97,33,28,185,16 // vinsertps $0x10,(%r9,%r15,4),%xmm3,%xmm3
- .byte 196,195,97,33,28,169,32 // vinsertps $0x20,(%r9,%rbp,4),%xmm3,%xmm3
- .byte 196,129,122,16,36,153 // vmovss (%r9,%r11,4),%xmm4
- .byte 196,227,97,33,220,48 // vinsertps $0x30,%xmm4,%xmm3,%xmm3
- .byte 196,99,101,24,194,1 // vinsertf128 $0x1,%xmm2,%ymm3,%ymm8
- .byte 77,139,112,56 // mov 0x38(%r8),%r14
- .byte 196,129,122,16,20,150 // vmovss (%r14,%r10,4),%xmm2
- .byte 196,195,105,33,20,158,16 // vinsertps $0x10,(%r14,%rbx,4),%xmm2,%xmm2
- .byte 196,195,105,33,20,182,32 // vinsertps $0x20,(%r14,%rsi,4),%xmm2,%xmm2
- .byte 196,129,122,16,36,166 // vmovss (%r14,%r12,4),%xmm4
- .byte 76,137,224 // mov %r12,%rax
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 196,193,122,16,36,142 // vmovss (%r14,%rcx,4),%xmm4
- .byte 196,131,89,33,36,190,16 // vinsertps $0x10,(%r14,%r15,4),%xmm4,%xmm4
- .byte 196,195,89,33,36,174,32 // vinsertps $0x20,(%r14,%rbp,4),%xmm4,%xmm4
- .byte 196,129,122,16,44,158 // vmovss (%r14,%r11,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,227,93,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm2
- .byte 77,139,96,64 // mov 0x40(%r8),%r12
- .byte 196,129,122,16,36,148 // vmovss (%r12,%r10,4),%xmm4
- .byte 196,195,89,33,36,156,16 // vinsertps $0x10,(%r12,%rbx,4),%xmm4,%xmm4
- .byte 196,195,89,33,36,180,32 // vinsertps $0x20,(%r12,%rsi,4),%xmm4,%xmm4
- .byte 196,193,122,16,44,132 // vmovss (%r12,%rax,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,193,122,16,44,140 // vmovss (%r12,%rcx,4),%xmm5
- .byte 196,131,81,33,44,188,16 // vinsertps $0x10,(%r12,%r15,4),%xmm5,%xmm5
- .byte 196,195,81,33,44,172,32 // vinsertps $0x20,(%r12,%rbp,4),%xmm5,%xmm5
- .byte 196,129,122,16,60,156 // vmovss (%r12,%r11,4),%xmm7
- .byte 196,227,81,33,239,48 // vinsertps $0x30,%xmm7,%xmm5,%xmm5
- .byte 196,227,85,24,236,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm5
- .byte 196,66,125,184,250 // vfmadd231ps %ymm10,%ymm0,%ymm15
- .byte 196,66,125,184,195 // vfmadd231ps %ymm11,%ymm0,%ymm8
- .byte 196,194,125,184,212 // vfmadd231ps %ymm12,%ymm0,%ymm2
- .byte 196,194,125,184,238 // vfmadd231ps %ymm14,%ymm0,%ymm5
- .byte 72,139,116,36,24 // mov 0x18(%rsp),%rsi
- .byte 196,193,122,16,4,177 // vmovss (%r9,%rsi,4),%xmm0
- .byte 76,139,68,36,16 // mov 0x10(%rsp),%r8
- .byte 196,131,121,33,4,129,16 // vinsertps $0x10,(%r9,%r8,4),%xmm0,%xmm0
- .byte 72,139,84,36,176 // mov -0x50(%rsp),%rdx
- .byte 196,193,122,16,36,145 // vmovss (%r9,%rdx,4),%xmm4
- .byte 196,227,121,33,196,32 // vinsertps $0x20,%xmm4,%xmm0,%xmm0
- .byte 72,139,28,36 // mov (%rsp),%rbx
- .byte 196,193,122,16,36,153 // vmovss (%r9,%rbx,4),%xmm4
- .byte 196,227,121,33,196,48 // vinsertps $0x30,%xmm4,%xmm0,%xmm0
- .byte 72,139,76,36,8 // mov 0x8(%rsp),%rcx
- .byte 196,193,122,16,36,137 // vmovss (%r9,%rcx,4),%xmm4
- .byte 72,139,108,36,168 // mov -0x58(%rsp),%rbp
- .byte 196,195,89,33,36,169,16 // vinsertps $0x10,(%r9,%rbp,4),%xmm4,%xmm4
- .byte 72,139,68,36,152 // mov -0x68(%rsp),%rax
- .byte 196,193,122,16,60,129 // vmovss (%r9,%rax,4),%xmm7
- .byte 196,227,89,33,231,32 // vinsertps $0x20,%xmm7,%xmm4,%xmm4
- .byte 196,129,122,16,60,169 // vmovss (%r9,%r13,4),%xmm7
- .byte 196,227,89,33,231,48 // vinsertps $0x30,%xmm7,%xmm4,%xmm4
- .byte 196,227,93,24,224,1 // vinsertf128 $0x1,%xmm0,%ymm4,%ymm4
- .byte 196,193,122,16,4,182 // vmovss (%r14,%rsi,4),%xmm0
- .byte 196,131,121,33,4,134,16 // vinsertps $0x10,(%r14,%r8,4),%xmm0,%xmm0
- .byte 196,193,122,16,60,150 // vmovss (%r14,%rdx,4),%xmm7
- .byte 196,227,121,33,199,32 // vinsertps $0x20,%xmm7,%xmm0,%xmm0
- .byte 196,193,122,16,60,158 // vmovss (%r14,%rbx,4),%xmm7
- .byte 196,227,121,33,199,48 // vinsertps $0x30,%xmm7,%xmm0,%xmm0
- .byte 196,193,122,16,60,142 // vmovss (%r14,%rcx,4),%xmm7
- .byte 196,195,65,33,60,174,16 // vinsertps $0x10,(%r14,%rbp,4),%xmm7,%xmm7
- .byte 196,193,122,16,52,134 // vmovss (%r14,%rax,4),%xmm6
- .byte 196,227,65,33,246,32 // vinsertps $0x20,%xmm6,%xmm7,%xmm6
- .byte 196,129,122,16,60,174 // vmovss (%r14,%r13,4),%xmm7
- .byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
- .byte 196,227,77,24,240,1 // vinsertf128 $0x1,%xmm0,%ymm6,%ymm6
- .byte 196,193,122,16,4,180 // vmovss (%r12,%rsi,4),%xmm0
- .byte 196,131,121,33,4,132,16 // vinsertps $0x10,(%r12,%r8,4),%xmm0,%xmm0
- .byte 196,193,122,16,60,148 // vmovss (%r12,%rdx,4),%xmm7
- .byte 196,227,121,33,199,32 // vinsertps $0x20,%xmm7,%xmm0,%xmm0
- .byte 196,193,122,16,60,156 // vmovss (%r12,%rbx,4),%xmm7
- .byte 196,227,121,33,199,48 // vinsertps $0x30,%xmm7,%xmm0,%xmm0
- .byte 196,193,122,16,60,140 // vmovss (%r12,%rcx,4),%xmm7
- .byte 196,195,65,33,60,172,16 // vinsertps $0x10,(%r12,%rbp,4),%xmm7,%xmm7
- .byte 196,193,122,16,28,132 // vmovss (%r12,%rax,4),%xmm3
- .byte 196,227,65,33,219,32 // vinsertps $0x20,%xmm3,%xmm7,%xmm3
- .byte 196,129,122,16,60,172 // vmovss (%r12,%r13,4),%xmm7
- .byte 196,227,97,33,223,48 // vinsertps $0x30,%xmm7,%xmm3,%xmm3
- .byte 196,227,101,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,98,117,184,172,36,128,0,0,0 // vfmadd231ps 0x80(%rsp),%ymm1,%ymm13
- .byte 196,226,117,184,100,36,96 // vfmadd231ps 0x60(%rsp),%ymm1,%ymm4
- .byte 196,226,117,184,116,36,64 // vfmadd231ps 0x40(%rsp),%ymm1,%ymm6
- .byte 196,194,117,184,217 // vfmadd231ps %ymm9,%ymm1,%ymm3
- .byte 196,226,125,24,5,82,162,0,0 // vbroadcastss 0xa252(%rip),%ymm0 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,226,125,24,13,25,162,0,0 // vbroadcastss 0xa219(%rip),%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
- .byte 196,98,125,168,233 // vfmadd213ps %ymm1,%ymm0,%ymm13
+ .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
+ .byte 196,194,125,24,25 // vbroadcastss (%r9),%ymm3
+ .byte 197,228,194,224,2 // vcmpleps %ymm0,%ymm3,%ymm4
+ .byte 197,237,250,212 // vpsubd %ymm4,%ymm2,%ymm2
+ .byte 197,228,194,217,2 // vcmpleps %ymm1,%ymm3,%ymm3
+ .byte 197,213,250,235 // vpsubd %ymm3,%ymm5,%ymm5
+ .byte 73,131,193,4 // add $0x4,%r9
+ .byte 73,255,200 // dec %r8
+ .byte 117,224 // jne 2ec64 <_sk_gradient_hsw_lowp+0x4d>
+ .byte 76,139,64,8 // mov 0x8(%rax),%r8
+ .byte 76,139,72,16 // mov 0x10(%rax),%r9
+ .byte 196,65,52,87,201 // vxorps %ymm9,%ymm9,%ymm9
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,12,144 // vgatherdps %ymm3,(%r8,%ymm2,4),%ymm9
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,36,168 // vgatherdps %ymm3,(%r8,%ymm5,4),%ymm4
+ .byte 197,252,17,36,36 // vmovups %ymm4,(%rsp)
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,36,145 // vgatherdps %ymm3,(%r9,%ymm2,4),%ymm4
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,52,169 // vgatherdps %ymm3,(%r9,%ymm5,4),%ymm6
+ .byte 197,252,17,116,36,224 // vmovups %ymm6,-0x20(%rsp)
+ .byte 76,139,64,24 // mov 0x18(%rax),%r8
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,4,144 // vgatherdps %ymm3,(%r8,%ymm2,4),%ymm8
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,52,168 // vgatherdps %ymm3,(%r8,%ymm5,4),%ymm6
+ .byte 197,252,17,116,36,192 // vmovups %ymm6,-0x40(%rsp)
+ .byte 76,139,64,32 // mov 0x20(%rax),%r8
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,52,144 // vgatherdps %ymm3,(%r8,%ymm2,4),%ymm6
+ .byte 197,252,17,116,36,160 // vmovups %ymm6,-0x60(%rsp)
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,194,101,146,52,168 // vgatherdps %ymm3,(%r8,%ymm5,4),%ymm6
+ .byte 197,252,17,116,36,128 // vmovups %ymm6,-0x80(%rsp)
+ .byte 76,139,64,40 // mov 0x28(%rax),%r8
+ .byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,20,144 // vgatherdps %ymm3,(%r8,%ymm2,4),%ymm10
+ .byte 196,65,36,87,219 // vxorps %ymm11,%ymm11,%ymm11
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,28,168 // vgatherdps %ymm3,(%r8,%ymm5,4),%ymm11
+ .byte 76,139,64,48 // mov 0x30(%rax),%r8
+ .byte 196,65,12,87,246 // vxorps %ymm14,%ymm14,%ymm14
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,52,144 // vgatherdps %ymm3,(%r8,%ymm2,4),%ymm14
+ .byte 196,65,4,87,255 // vxorps %ymm15,%ymm15,%ymm15
+ .byte 197,229,118,219 // vpcmpeqd %ymm3,%ymm3,%ymm3
+ .byte 196,66,101,146,60,168 // vgatherdps %ymm3,(%r8,%ymm5,4),%ymm15
+ .byte 76,139,64,56 // mov 0x38(%rax),%r8
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,205,118,246 // vpcmpeqd %ymm6,%ymm6,%ymm6
+ .byte 196,194,77,146,28,144 // vgatherdps %ymm6,(%r8,%ymm2,4),%ymm3
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 197,205,118,246 // vpcmpeqd %ymm6,%ymm6,%ymm6
+ .byte 196,194,77,146,60,168 // vgatherdps %ymm6,(%r8,%ymm5,4),%ymm7
+ .byte 72,139,64,64 // mov 0x40(%rax),%rax
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 196,65,21,118,237 // vpcmpeqd %ymm13,%ymm13,%ymm13
+ .byte 196,226,21,146,52,144 // vgatherdps %ymm13,(%rax,%ymm2,4),%ymm6
+ .byte 197,237,118,210 // vpcmpeqd %ymm2,%ymm2,%ymm2
+ .byte 196,98,109,146,36,168 // vgatherdps %ymm2,(%rax,%ymm5,4),%ymm12
+ .byte 196,66,125,184,209 // vfmadd231ps %ymm9,%ymm0,%ymm10
+ .byte 196,98,125,184,244 // vfmadd231ps %ymm4,%ymm0,%ymm14
+ .byte 196,194,125,184,216 // vfmadd231ps %ymm8,%ymm0,%ymm3
+ .byte 196,226,125,184,116,36,160 // vfmadd231ps -0x60(%rsp),%ymm0,%ymm6
+ .byte 196,98,117,184,28,36 // vfmadd231ps (%rsp),%ymm1,%ymm11
+ .byte 196,98,117,184,124,36,224 // vfmadd231ps -0x20(%rsp),%ymm1,%ymm15
+ .byte 196,226,117,184,124,36,192 // vfmadd231ps -0x40(%rsp),%ymm1,%ymm7
+ .byte 196,98,117,184,100,36,128 // vfmadd231ps -0x80(%rsp),%ymm1,%ymm12
+ .byte 196,226,125,24,5,2,156,0,0 // vbroadcastss 0x9c02(%rip),%ymm0 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,226,125,24,13,201,155,0,0 // vbroadcastss 0x9bc9(%rip),%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,168,217 // vfmadd213ps %ymm1,%ymm0,%ymm11
+ .byte 196,98,125,168,209 // vfmadd213ps %ymm1,%ymm0,%ymm10
.byte 196,98,125,168,249 // vfmadd213ps %ymm1,%ymm0,%ymm15
- .byte 196,226,125,168,225 // vfmadd213ps %ymm1,%ymm0,%ymm4
- .byte 196,98,125,168,193 // vfmadd213ps %ymm1,%ymm0,%ymm8
- .byte 196,226,125,168,241 // vfmadd213ps %ymm1,%ymm0,%ymm6
- .byte 196,226,125,168,209 // vfmadd213ps %ymm1,%ymm0,%ymm2
+ .byte 196,98,125,168,241 // vfmadd213ps %ymm1,%ymm0,%ymm14
+ .byte 196,226,125,168,249 // vfmadd213ps %ymm1,%ymm0,%ymm7
.byte 196,226,125,168,217 // vfmadd213ps %ymm1,%ymm0,%ymm3
- .byte 196,226,125,168,233 // vfmadd213ps %ymm1,%ymm0,%ymm5
- .byte 196,193,126,91,199 // vcvttps2dq %ymm15,%ymm0
- .byte 197,253,111,61,240,169,0,0 // vmovdqa 0xa9f0(%rip),%ymm7 # 39f40 <_sk_srcover_bgra_8888_sse2_lowp+0xab8>
- .byte 196,226,125,0,199 // vpshufb %ymm7,%ymm0,%ymm0
+ .byte 196,98,125,168,225 // vfmadd213ps %ymm1,%ymm0,%ymm12
+ .byte 196,226,125,168,241 // vfmadd213ps %ymm1,%ymm0,%ymm6
+ .byte 196,193,126,91,194 // vcvttps2dq %ymm10,%ymm0
+ .byte 197,253,111,37,164,163,0,0 // vmovdqa 0xa3a4(%rip),%ymm4 # 391c0 <_sk_srcover_bgra_8888_sse2_lowp+0xabc>
+ .byte 196,226,125,0,196 // vpshufb %ymm4,%ymm0,%ymm0
.byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,193,126,91,205 // vcvttps2dq %ymm13,%ymm1
- .byte 196,226,117,0,207 // vpshufb %ymm7,%ymm1,%ymm1
+ .byte 196,193,126,91,203 // vcvttps2dq %ymm11,%ymm1
+ .byte 196,226,117,0,204 // vpshufb %ymm4,%ymm1,%ymm1
.byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
- .byte 196,193,126,91,200 // vcvttps2dq %ymm8,%ymm1
- .byte 196,226,117,0,207 // vpshufb %ymm7,%ymm1,%ymm1
+ .byte 196,193,126,91,206 // vcvttps2dq %ymm14,%ymm1
+ .byte 196,226,117,0,204 // vpshufb %ymm4,%ymm1,%ymm1
.byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 197,254,91,228 // vcvttps2dq %ymm4,%ymm4
- .byte 196,226,93,0,231 // vpshufb %ymm7,%ymm4,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,117,56,204,1 // vinserti128 $0x1,%xmm4,%ymm1,%ymm1
- .byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
- .byte 196,226,109,0,215 // vpshufb %ymm7,%ymm2,%ymm2
+ .byte 196,193,126,91,215 // vcvttps2dq %ymm15,%ymm2
+ .byte 196,226,109,0,212 // vpshufb %ymm4,%ymm2,%ymm2
.byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
- .byte 197,254,91,230 // vcvttps2dq %ymm6,%ymm4
- .byte 196,226,93,0,231 // vpshufb %ymm7,%ymm4,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,109,56,212,1 // vinserti128 $0x1,%xmm4,%ymm2,%ymm2
- .byte 197,254,91,229 // vcvttps2dq %ymm5,%ymm4
- .byte 196,226,93,0,231 // vpshufb %ymm7,%ymm4,%ymm4
- .byte 197,254,91,219 // vcvttps2dq %ymm3,%ymm3
- .byte 196,226,101,0,223 // vpshufb %ymm7,%ymm3,%ymm3
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
+ .byte 196,227,117,56,202,1 // vinserti128 $0x1,%xmm2,%ymm1,%ymm1
+ .byte 197,254,91,211 // vcvttps2dq %ymm3,%ymm2
+ .byte 196,226,109,0,212 // vpshufb %ymm4,%ymm2,%ymm2
+ .byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
+ .byte 197,254,91,223 // vcvttps2dq %ymm7,%ymm3
+ .byte 196,226,101,0,220 // vpshufb %ymm4,%ymm3,%ymm3
.byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
- .byte 196,227,93,56,219,1 // vinserti128 $0x1,%xmm3,%ymm4,%ymm3
- .byte 72,139,116,36,32 // mov 0x20(%rsp),%rsi
+ .byte 196,227,109,56,211,1 // vinserti128 $0x1,%xmm3,%ymm2,%ymm2
+ .byte 197,254,91,222 // vcvttps2dq %ymm6,%ymm3
+ .byte 196,226,101,0,220 // vpshufb %ymm4,%ymm3,%ymm3
+ .byte 196,193,126,91,236 // vcvttps2dq %ymm12,%ymm5
+ .byte 196,226,85,0,228 // vpshufb %ymm4,%ymm5,%ymm4
+ .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
+ .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
+ .byte 196,227,101,56,220,1 // vinserti128 $0x1,%xmm4,%ymm3,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,139,124,36,40 // mov 0x28(%rsp),%rdi
- .byte 72,139,84,36,48 // mov 0x30(%rsp),%rdx
- .byte 72,139,76,36,56 // mov 0x38(%rsp),%rcx
- .byte 197,252,16,164,36,160,0,0,0 // vmovups 0xa0(%rsp),%ymm4
- .byte 197,252,16,172,36,192,0,0,0 // vmovups 0xc0(%rsp),%ymm5
- .byte 197,252,16,180,36,224,0,0,0 // vmovups 0xe0(%rsp),%ymm6
- .byte 197,252,16,188,36,0,1,0,0 // vmovups 0x100(%rsp),%ymm7
- .byte 72,129,196,40,1,0,0 // add $0x128,%rsp
- .byte 91 // pop %rbx
- .byte 65,92 // pop %r12
- .byte 65,93 // pop %r13
- .byte 65,94 // pop %r14
- .byte 65,95 // pop %r15
- .byte 93 // pop %rbp
+ .byte 197,252,16,100,36,32 // vmovups 0x20(%rsp),%ymm4
+ .byte 197,252,16,108,36,64 // vmovups 0x40(%rsp),%ymm5
+ .byte 197,252,16,116,36,96 // vmovups 0x60(%rsp),%ymm6
+ .byte 197,252,16,188,36,128,0,0,0 // vmovups 0x80(%rsp),%ymm7
+ .byte 72,129,196,184,0,0,0 // add $0xb8,%rsp
.byte 255,224 // jmpq *%rax
HIDDEN _sk_evenly_spaced_gradient_hsw_lowp
.globl _sk_evenly_spaced_gradient_hsw_lowp
FUNCTION(_sk_evenly_spaced_gradient_hsw_lowp)
_sk_evenly_spaced_gradient_hsw_lowp:
- .byte 85 // push %rbp
- .byte 65,87 // push %r15
- .byte 65,86 // push %r14
- .byte 65,85 // push %r13
- .byte 65,84 // push %r12
- .byte 83 // push %rbx
- .byte 72,129,236,232,0,0,0 // sub $0xe8,%rsp
- .byte 197,252,17,188,36,192,0,0,0 // vmovups %ymm7,0xc0(%rsp)
- .byte 197,252,17,180,36,160,0,0,0 // vmovups %ymm6,0xa0(%rsp)
- .byte 197,252,17,172,36,128,0,0,0 // vmovups %ymm5,0x80(%rsp)
- .byte 197,252,17,100,36,96 // vmovups %ymm4,0x60(%rsp)
- .byte 72,137,76,36,24 // mov %rcx,0x18(%rsp)
- .byte 72,137,84,36,16 // mov %rdx,0x10(%rsp)
- .byte 72,137,124,36,8 // mov %rdi,0x8(%rsp)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,137,52,36 // mov %rsi,(%rsp)
- .byte 72,139,16 // mov (%rax),%rdx
- .byte 72,137,68,36,144 // mov %rax,-0x70(%rsp)
- .byte 72,139,104,8 // mov 0x8(%rax),%rbp
- .byte 72,255,202 // dec %rdx
- .byte 120,7 // js 2f68a <_sk_evenly_spaced_gradient_hsw_lowp+0x5f>
- .byte 196,225,234,42,210 // vcvtsi2ss %rdx,%xmm2,%xmm2
- .byte 235,21 // jmp 2f69f <_sk_evenly_spaced_gradient_hsw_lowp+0x74>
- .byte 72,137,209 // mov %rdx,%rcx
- .byte 72,209,233 // shr %rcx
- .byte 131,226,1 // and $0x1,%edx
- .byte 72,9,202 // or %rcx,%rdx
- .byte 196,225,234,42,210 // vcvtsi2ss %rdx,%xmm2,%xmm2
+ .byte 72,129,236,152,0,0,0 // sub $0x98,%rsp
+ .byte 197,252,17,124,36,96 // vmovups %ymm7,0x60(%rsp)
+ .byte 197,252,17,116,36,64 // vmovups %ymm6,0x40(%rsp)
+ .byte 197,252,17,108,36,32 // vmovups %ymm5,0x20(%rsp)
+ .byte 197,252,17,36,36 // vmovups %ymm4,(%rsp)
+ .byte 197,124,40,217 // vmovaps %ymm1,%ymm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 76,139,64,8 // mov 0x8(%rax),%r8
+ .byte 73,255,201 // dec %r9
+ .byte 120,7 // js 2ef09 <_sk_evenly_spaced_gradient_hsw_lowp+0x37>
+ .byte 196,193,234,42,209 // vcvtsi2ss %r9,%xmm2,%xmm2
+ .byte 235,22 // jmp 2ef1f <_sk_evenly_spaced_gradient_hsw_lowp+0x4d>
+ .byte 77,137,202 // mov %r9,%r10
+ .byte 73,209,234 // shr %r10
+ .byte 65,131,225,1 // and $0x1,%r9d
+ .byte 77,9,209 // or %r10,%r9
+ .byte 196,193,234,42,209 // vcvtsi2ss %r9,%xmm2,%xmm2
.byte 197,234,88,210 // vaddss %xmm2,%xmm2,%xmm2
.byte 196,226,125,24,210 // vbroadcastss %xmm2,%ymm2
.byte 197,252,89,218 // vmulps %ymm2,%ymm0,%ymm3
- .byte 197,244,89,210 // vmulps %ymm2,%ymm1,%ymm2
+ .byte 197,164,89,210 // vmulps %ymm2,%ymm11,%ymm2
.byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
- .byte 197,254,91,219 // vcvttps2dq %ymm3,%ymm3
- .byte 196,227,249,22,219,1 // vpextrq $0x1,%xmm3,%rbx
- .byte 137,216 // mov %ebx,%eax
- .byte 72,137,68,36,200 // mov %rax,-0x38(%rsp)
- .byte 72,193,235,32 // shr $0x20,%rbx
- .byte 72,137,92,36,224 // mov %rbx,-0x20(%rsp)
- .byte 196,225,249,126,223 // vmovq %xmm3,%rdi
- .byte 137,248 // mov %edi,%eax
- .byte 72,137,68,36,128 // mov %rax,-0x80(%rsp)
- .byte 72,193,239,32 // shr $0x20,%rdi
- .byte 196,227,125,57,219,1 // vextracti128 $0x1,%ymm3,%xmm3
- .byte 196,227,249,22,219,1 // vpextrq $0x1,%xmm3,%rbx
- .byte 137,222 // mov %ebx,%esi
- .byte 72,193,235,32 // shr $0x20,%rbx
- .byte 196,225,249,126,218 // vmovq %xmm3,%rdx
- .byte 137,209 // mov %edx,%ecx
- .byte 72,137,76,36,136 // mov %rcx,-0x78(%rsp)
- .byte 72,193,234,32 // shr $0x20,%rdx
- .byte 196,195,249,22,208,1 // vpextrq $0x1,%xmm2,%r8
- .byte 69,137,193 // mov %r8d,%r9d
- .byte 73,193,232,32 // shr $0x20,%r8
- .byte 76,137,192 // mov %r8,%rax
- .byte 72,137,68,36,168 // mov %rax,-0x58(%rsp)
- .byte 196,193,249,126,210 // vmovq %xmm2,%r10
- .byte 69,137,214 // mov %r10d,%r14d
- .byte 73,193,234,32 // shr $0x20,%r10
- .byte 76,137,84,36,152 // mov %r10,-0x68(%rsp)
- .byte 196,227,125,57,210,1 // vextracti128 $0x1,%ymm2,%xmm2
- .byte 196,195,249,22,215,1 // vpextrq $0x1,%xmm2,%r15
- .byte 69,137,252 // mov %r15d,%r12d
- .byte 73,193,239,32 // shr $0x20,%r15
- .byte 196,193,249,126,213 // vmovq %xmm2,%r13
- .byte 69,137,232 // mov %r13d,%r8d
- .byte 73,193,237,32 // shr $0x20,%r13
- .byte 196,161,122,16,84,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm2
- .byte 196,163,105,33,84,173,0,16 // vinsertps $0x10,0x0(%rbp,%r13,4),%xmm2,%xmm2
- .byte 196,161,122,16,92,165,0 // vmovss 0x0(%rbp,%r12,4),%xmm3
- .byte 76,137,100,36,160 // mov %r12,-0x60(%rsp)
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,161,122,16,92,189,0 // vmovss 0x0(%rbp,%r15,4),%xmm3
- .byte 196,99,105,33,195,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm8
- .byte 196,161,122,16,92,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm3
- .byte 196,35,97,33,76,149,0,16 // vinsertps $0x10,0x0(%rbp,%r10,4),%xmm3,%xmm9
- .byte 196,33,122,16,84,141,0 // vmovss 0x0(%rbp,%r9,4),%xmm10
- .byte 77,137,202 // mov %r9,%r10
- .byte 76,137,84,36,208 // mov %r10,-0x30(%rsp)
- .byte 197,122,16,108,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm13
- .byte 197,250,16,124,141,0 // vmovss 0x0(%rbp,%rcx,4),%xmm7
- .byte 196,99,65,33,92,149,0,16 // vinsertps $0x10,0x0(%rbp,%rdx,4),%xmm7,%xmm11
- .byte 72,137,241 // mov %rsi,%rcx
- .byte 197,122,16,100,141,0 // vmovss 0x0(%rbp,%rcx,4),%xmm12
- .byte 72,137,216 // mov %rbx,%rax
- .byte 72,137,68,36,176 // mov %rax,-0x50(%rsp)
- .byte 197,250,16,92,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm3
- .byte 72,139,92,36,128 // mov -0x80(%rsp),%rbx
- .byte 197,250,16,108,157,0 // vmovss 0x0(%rbp,%rbx,4),%xmm5
- .byte 73,137,251 // mov %rdi,%r11
- .byte 76,137,92,36,184 // mov %r11,-0x48(%rsp)
- .byte 196,163,81,33,108,157,0,16 // vinsertps $0x10,0x0(%rbp,%r11,4),%xmm5,%xmm5
- .byte 72,139,116,36,200 // mov -0x38(%rsp),%rsi
- .byte 197,250,16,124,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm7
- .byte 76,139,76,36,224 // mov -0x20(%rsp),%r9
- .byte 196,161,122,16,84,141,0 // vmovss 0x0(%rbp,%r9,4),%xmm2
- .byte 196,195,49,33,242,32 // vinsertps $0x20,%xmm10,%xmm9,%xmm6
- .byte 196,195,73,33,245,48 // vinsertps $0x30,%xmm13,%xmm6,%xmm6
- .byte 196,195,77,24,224,1 // vinsertf128 $0x1,%xmm8,%ymm6,%ymm4
- .byte 197,252,17,100,36,64 // vmovups %ymm4,0x40(%rsp)
- .byte 196,195,33,33,244,32 // vinsertps $0x20,%xmm12,%xmm11,%xmm6
- .byte 196,227,73,33,219,48 // vinsertps $0x30,%xmm3,%xmm6,%xmm3
- .byte 196,227,81,33,239,32 // vinsertps $0x20,%xmm7,%xmm5,%xmm5
- .byte 196,227,81,33,210,48 // vinsertps $0x30,%xmm2,%xmm5,%xmm2
- .byte 196,99,109,24,203,1 // vinsertf128 $0x1,%xmm3,%ymm2,%ymm9
- .byte 72,139,124,36,144 // mov -0x70(%rsp),%rdi
- .byte 72,139,111,16 // mov 0x10(%rdi),%rbp
- .byte 196,161,122,16,84,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm2
- .byte 196,163,105,33,84,173,0,16 // vinsertps $0x10,0x0(%rbp,%r13,4),%xmm2,%xmm2
- .byte 196,161,122,16,92,165,0 // vmovss 0x0(%rbp,%r12,4),%xmm3
- .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
- .byte 196,161,122,16,92,189,0 // vmovss 0x0(%rbp,%r15,4),%xmm3
- .byte 196,99,105,33,195,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm8
- .byte 196,161,122,16,92,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm3
- .byte 72,139,124,36,152 // mov -0x68(%rsp),%rdi
- .byte 196,227,97,33,92,189,0,16 // vinsertps $0x10,0x0(%rbp,%rdi,4),%xmm3,%xmm3
- .byte 196,161,122,16,108,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm5
- .byte 196,227,97,33,221,32 // vinsertps $0x20,%xmm5,%xmm3,%xmm3
- .byte 72,139,124,36,168 // mov -0x58(%rsp),%rdi
- .byte 197,250,16,108,189,0 // vmovss 0x0(%rbp,%rdi,4),%xmm5
- .byte 196,227,97,33,221,48 // vinsertps $0x30,%xmm5,%xmm3,%xmm3
- .byte 72,139,124,36,136 // mov -0x78(%rsp),%rdi
- .byte 197,250,16,108,189,0 // vmovss 0x0(%rbp,%rdi,4),%xmm5
- .byte 196,99,81,33,84,149,0,16 // vinsertps $0x10,0x0(%rbp,%rdx,4),%xmm5,%xmm10
- .byte 72,137,84,36,192 // mov %rdx,-0x40(%rsp)
- .byte 197,250,16,116,141,0 // vmovss 0x0(%rbp,%rcx,4),%xmm6
- .byte 197,250,16,124,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm7
- .byte 197,250,16,100,157,0 // vmovss 0x0(%rbp,%rbx,4),%xmm4
- .byte 196,163,89,33,100,157,0,16 // vinsertps $0x10,0x0(%rbp,%r11,4),%xmm4,%xmm4
- .byte 197,250,16,84,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm2
- .byte 73,137,242 // mov %rsi,%r10
- .byte 77,137,203 // mov %r9,%r11
- .byte 196,161,122,16,108,157,0 // vmovss 0x0(%rbp,%r11,4),%xmm5
- .byte 196,195,101,24,216,1 // vinsertf128 $0x1,%xmm8,%ymm3,%ymm3
- .byte 197,252,17,92,36,32 // vmovups %ymm3,0x20(%rsp)
- .byte 196,227,41,33,246,32 // vinsertps $0x20,%xmm6,%xmm10,%xmm6
- .byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
- .byte 196,227,89,33,210,32 // vinsertps $0x20,%xmm2,%xmm4,%xmm2
- .byte 196,227,105,33,213,48 // vinsertps $0x30,%xmm5,%xmm2,%xmm2
- .byte 196,227,109,24,222,1 // vinsertf128 $0x1,%xmm6,%ymm2,%ymm3
- .byte 72,139,124,36,144 // mov -0x70(%rsp),%rdi
- .byte 72,139,111,24 // mov 0x18(%rdi),%rbp
- .byte 196,161,122,16,84,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm2
- .byte 77,137,196 // mov %r8,%r12
- .byte 196,163,105,33,84,173,0,16 // vinsertps $0x10,0x0(%rbp,%r13,4),%xmm2,%xmm2
- .byte 76,137,235 // mov %r13,%rbx
- .byte 72,139,68,36,160 // mov -0x60(%rsp),%rax
- .byte 197,250,16,100,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 196,161,122,16,100,189,0 // vmovss 0x0(%rbp,%r15,4),%xmm4
- .byte 77,137,253 // mov %r15,%r13
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 196,161,122,16,100,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm4
- .byte 76,137,240 // mov %r14,%rax
- .byte 76,139,116,36,152 // mov -0x68(%rsp),%r14
- .byte 196,163,89,33,100,181,0,16 // vinsertps $0x10,0x0(%rbp,%r14,4),%xmm4,%xmm4
- .byte 76,139,76,36,208 // mov -0x30(%rsp),%r9
- .byte 196,161,122,16,108,141,0 // vmovss 0x0(%rbp,%r9,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 76,139,68,36,168 // mov -0x58(%rsp),%r8
- .byte 196,161,122,16,108,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,194,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm8
- .byte 72,139,116,36,136 // mov -0x78(%rsp),%rsi
- .byte 197,250,16,84,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm2
- .byte 196,227,105,33,84,149,0,16 // vinsertps $0x10,0x0(%rbp,%rdx,4),%xmm2,%xmm2
- .byte 72,137,76,36,216 // mov %rcx,-0x28(%rsp)
- .byte 197,250,16,100,141,0 // vmovss 0x0(%rbp,%rcx,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 72,139,84,36,176 // mov -0x50(%rsp),%rdx
- .byte 197,250,16,100,149,0 // vmovss 0x0(%rbp,%rdx,4),%xmm4
- .byte 72,139,84,36,128 // mov -0x80(%rsp),%rdx
- .byte 197,250,16,108,149,0 // vmovss 0x0(%rbp,%rdx,4),%xmm5
- .byte 72,139,84,36,184 // mov -0x48(%rsp),%rdx
- .byte 196,227,81,33,108,149,0,16 // vinsertps $0x10,0x0(%rbp,%rdx,4),%xmm5,%xmm5
- .byte 196,161,122,16,116,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm6
- .byte 77,137,218 // mov %r11,%r10
- .byte 196,161,122,16,124,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm7
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 196,227,81,33,230,32 // vinsertps $0x20,%xmm6,%xmm5,%xmm4
- .byte 196,227,89,33,231,48 // vinsertps $0x30,%xmm7,%xmm4,%xmm4
- .byte 196,99,93,24,226,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm12
- .byte 72,139,111,32 // mov 0x20(%rdi),%rbp
- .byte 196,161,122,16,84,165,0 // vmovss 0x0(%rbp,%r12,4),%xmm2
- .byte 77,137,227 // mov %r12,%r11
- .byte 76,137,92,36,248 // mov %r11,-0x8(%rsp)
- .byte 196,227,105,33,84,157,0,16 // vinsertps $0x10,0x0(%rbp,%rbx,4),%xmm2,%xmm2
- .byte 73,137,223 // mov %rbx,%r15
- .byte 76,137,124,36,232 // mov %r15,-0x18(%rsp)
- .byte 76,139,100,36,160 // mov -0x60(%rsp),%r12
- .byte 196,161,122,16,100,165,0 // vmovss 0x0(%rbp,%r12,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 196,161,122,16,100,173,0 // vmovss 0x0(%rbp,%r13,4),%xmm4
- .byte 76,137,235 // mov %r13,%rbx
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 197,250,16,100,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm4
- .byte 72,137,194 // mov %rax,%rdx
- .byte 72,137,84,36,240 // mov %rdx,-0x10(%rsp)
- .byte 196,163,89,33,100,181,0,16 // vinsertps $0x10,0x0(%rbp,%r14,4),%xmm4,%xmm4
- .byte 76,137,240 // mov %r14,%rax
- .byte 196,161,122,16,108,141,0 // vmovss 0x0(%rbp,%r9,4),%xmm5
- .byte 76,137,207 // mov %r9,%rdi
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,161,122,16,108,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm10
- .byte 197,250,16,84,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm2
- .byte 72,139,116,36,192 // mov -0x40(%rsp),%rsi
- .byte 196,227,105,33,84,181,0,16 // vinsertps $0x10,0x0(%rbp,%rsi,4),%xmm2,%xmm2
- .byte 197,250,16,100,141,0 // vmovss 0x0(%rbp,%rcx,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 72,139,116,36,176 // mov -0x50(%rsp),%rsi
- .byte 197,250,16,100,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm4
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 76,139,108,36,128 // mov -0x80(%rsp),%r13
- .byte 196,161,122,16,100,173,0 // vmovss 0x0(%rbp,%r13,4),%xmm4
- .byte 76,139,76,36,184 // mov -0x48(%rsp),%r9
- .byte 196,163,89,33,100,141,0,16 // vinsertps $0x10,0x0(%rbp,%r9,4),%xmm4,%xmm4
- .byte 76,139,116,36,200 // mov -0x38(%rsp),%r14
- .byte 196,161,122,16,108,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,161,122,16,108,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,242,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm14
- .byte 72,139,76,36,144 // mov -0x70(%rsp),%rcx
- .byte 72,139,105,40 // mov 0x28(%rcx),%rbp
- .byte 196,161,122,16,84,157,0 // vmovss 0x0(%rbp,%r11,4),%xmm2
- .byte 196,163,105,33,84,189,0,16 // vinsertps $0x10,0x0(%rbp,%r15,4),%xmm2,%xmm2
- .byte 196,161,122,16,100,165,0 // vmovss 0x0(%rbp,%r12,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 197,250,16,100,157,0 // vmovss 0x0(%rbp,%rbx,4),%xmm4
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 197,250,16,100,149,0 // vmovss 0x0(%rbp,%rdx,4),%xmm4
- .byte 196,227,89,33,100,133,0,16 // vinsertps $0x10,0x0(%rbp,%rax,4),%xmm4,%xmm4
- .byte 197,250,16,108,189,0 // vmovss 0x0(%rbp,%rdi,4),%xmm5
- .byte 73,137,255 // mov %rdi,%r15
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,161,122,16,108,133,0 // vmovss 0x0(%rbp,%r8,4),%xmm5
- .byte 77,137,195 // mov %r8,%r11
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,234,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm13
- .byte 72,139,84,36,136 // mov -0x78(%rsp),%rdx
- .byte 197,250,16,84,149,0 // vmovss 0x0(%rbp,%rdx,4),%xmm2
- .byte 72,139,124,36,192 // mov -0x40(%rsp),%rdi
- .byte 196,227,105,33,84,189,0,16 // vinsertps $0x10,0x0(%rbp,%rdi,4),%xmm2,%xmm2
- .byte 72,139,68,36,216 // mov -0x28(%rsp),%rax
- .byte 197,250,16,100,133,0 // vmovss 0x0(%rbp,%rax,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 197,250,16,100,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm4
- .byte 73,137,240 // mov %rsi,%r8
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 73,137,204 // mov %rcx,%r12
- .byte 76,137,238 // mov %r13,%rsi
- .byte 197,250,16,100,181,0 // vmovss 0x0(%rbp,%rsi,4),%xmm4
- .byte 196,163,89,33,100,141,0,16 // vinsertps $0x10,0x0(%rbp,%r9,4),%xmm4,%xmm4
- .byte 77,137,205 // mov %r9,%r13
- .byte 196,161,122,16,108,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,161,122,16,108,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,250,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm15
- .byte 77,139,76,36,48 // mov 0x30(%r12),%r9
- .byte 196,193,122,16,20,145 // vmovss (%r9,%rdx,4),%xmm2
- .byte 196,195,105,33,20,185,16 // vinsertps $0x10,(%r9,%rdi,4),%xmm2,%xmm2
- .byte 196,193,122,16,36,129 // vmovss (%r9,%rax,4),%xmm4
- .byte 72,137,193 // mov %rax,%rcx
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 196,129,122,16,36,129 // vmovss (%r9,%r8,4),%xmm4
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 196,193,122,16,36,177 // vmovss (%r9,%rsi,4),%xmm4
- .byte 196,131,89,33,36,169,16 // vinsertps $0x10,(%r9,%r13,4),%xmm4,%xmm4
- .byte 76,137,245 // mov %r14,%rbp
- .byte 196,193,122,16,44,169 // vmovss (%r9,%rbp,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,129,122,16,44,145 // vmovss (%r9,%r10,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,99,93,24,218,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm11
- .byte 77,139,116,36,56 // mov 0x38(%r12),%r14
- .byte 196,193,122,16,20,150 // vmovss (%r14,%rdx,4),%xmm2
- .byte 196,195,105,33,20,190,16 // vinsertps $0x10,(%r14,%rdi,4),%xmm2,%xmm2
- .byte 196,193,122,16,36,142 // vmovss (%r14,%rcx,4),%xmm4
- .byte 196,227,105,33,212,32 // vinsertps $0x20,%xmm4,%xmm2,%xmm2
- .byte 196,129,122,16,36,134 // vmovss (%r14,%r8,4),%xmm4
- .byte 196,227,105,33,212,48 // vinsertps $0x30,%xmm4,%xmm2,%xmm2
- .byte 196,193,122,16,36,182 // vmovss (%r14,%rsi,4),%xmm4
- .byte 196,131,89,33,36,174,16 // vinsertps $0x10,(%r14,%r13,4),%xmm4,%xmm4
- .byte 196,193,122,16,44,174 // vmovss (%r14,%rbp,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,129,122,16,44,150 // vmovss (%r14,%r10,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,227,93,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm4,%ymm2
- .byte 77,139,100,36,64 // mov 0x40(%r12),%r12
- .byte 196,193,122,16,36,148 // vmovss (%r12,%rdx,4),%xmm4
- .byte 196,195,89,33,36,188,16 // vinsertps $0x10,(%r12,%rdi,4),%xmm4,%xmm4
- .byte 196,193,122,16,44,140 // vmovss (%r12,%rcx,4),%xmm5
- .byte 196,227,89,33,229,32 // vinsertps $0x20,%xmm5,%xmm4,%xmm4
- .byte 196,129,122,16,44,132 // vmovss (%r12,%r8,4),%xmm5
- .byte 196,227,89,33,229,48 // vinsertps $0x30,%xmm5,%xmm4,%xmm4
- .byte 196,193,122,16,44,180 // vmovss (%r12,%rsi,4),%xmm5
- .byte 196,131,81,33,44,172,16 // vinsertps $0x10,(%r12,%r13,4),%xmm5,%xmm5
- .byte 196,193,122,16,52,172 // vmovss (%r12,%rbp,4),%xmm6
- .byte 196,227,81,33,238,32 // vinsertps $0x20,%xmm6,%xmm5,%xmm5
- .byte 196,129,122,16,52,148 // vmovss (%r12,%r10,4),%xmm6
- .byte 196,227,81,33,238,48 // vinsertps $0x30,%xmm6,%xmm5,%xmm5
- .byte 196,227,85,24,236,1 // vinsertf128 $0x1,%xmm4,%ymm5,%ymm5
- .byte 196,66,125,184,249 // vfmadd231ps %ymm9,%ymm0,%ymm15
- .byte 196,98,125,184,219 // vfmadd231ps %ymm3,%ymm0,%ymm11
- .byte 196,194,125,184,212 // vfmadd231ps %ymm12,%ymm0,%ymm2
- .byte 196,194,125,184,238 // vfmadd231ps %ymm14,%ymm0,%ymm5
- .byte 72,139,116,36,248 // mov -0x8(%rsp),%rsi
- .byte 196,193,122,16,4,177 // vmovss (%r9,%rsi,4),%xmm0
- .byte 76,139,68,36,232 // mov -0x18(%rsp),%r8
- .byte 196,131,121,33,4,129,16 // vinsertps $0x10,(%r9,%r8,4),%xmm0,%xmm0
- .byte 72,139,84,36,160 // mov -0x60(%rsp),%rdx
- .byte 196,193,122,16,36,145 // vmovss (%r9,%rdx,4),%xmm4
- .byte 196,227,121,33,196,32 // vinsertps $0x20,%xmm4,%xmm0,%xmm0
- .byte 196,193,122,16,36,153 // vmovss (%r9,%rbx,4),%xmm4
- .byte 196,227,121,33,196,48 // vinsertps $0x30,%xmm4,%xmm0,%xmm0
- .byte 72,139,76,36,240 // mov -0x10(%rsp),%rcx
- .byte 196,193,122,16,36,137 // vmovss (%r9,%rcx,4),%xmm4
- .byte 72,139,108,36,152 // mov -0x68(%rsp),%rbp
- .byte 196,195,89,33,36,169,16 // vinsertps $0x10,(%r9,%rbp,4),%xmm4,%xmm4
- .byte 196,129,122,16,52,185 // vmovss (%r9,%r15,4),%xmm6
- .byte 196,227,89,33,230,32 // vinsertps $0x20,%xmm6,%xmm4,%xmm4
- .byte 196,129,122,16,52,153 // vmovss (%r9,%r11,4),%xmm6
- .byte 196,227,89,33,230,48 // vinsertps $0x30,%xmm6,%xmm4,%xmm4
- .byte 196,227,93,24,224,1 // vinsertf128 $0x1,%xmm0,%ymm4,%ymm4
- .byte 196,193,122,16,4,182 // vmovss (%r14,%rsi,4),%xmm0
- .byte 196,131,121,33,4,134,16 // vinsertps $0x10,(%r14,%r8,4),%xmm0,%xmm0
- .byte 196,193,122,16,52,150 // vmovss (%r14,%rdx,4),%xmm6
- .byte 196,227,121,33,198,32 // vinsertps $0x20,%xmm6,%xmm0,%xmm0
- .byte 196,193,122,16,52,158 // vmovss (%r14,%rbx,4),%xmm6
- .byte 196,227,121,33,198,48 // vinsertps $0x30,%xmm6,%xmm0,%xmm0
- .byte 196,193,122,16,52,142 // vmovss (%r14,%rcx,4),%xmm6
- .byte 196,195,73,33,52,174,16 // vinsertps $0x10,(%r14,%rbp,4),%xmm6,%xmm6
- .byte 196,129,122,16,28,190 // vmovss (%r14,%r15,4),%xmm3
- .byte 196,227,73,33,219,32 // vinsertps $0x20,%xmm3,%xmm6,%xmm3
- .byte 196,129,122,16,52,158 // vmovss (%r14,%r11,4),%xmm6
- .byte 196,227,97,33,222,48 // vinsertps $0x30,%xmm6,%xmm3,%xmm3
- .byte 196,227,101,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,193,122,16,4,180 // vmovss (%r12,%rsi,4),%xmm0
- .byte 196,131,121,33,4,132,16 // vinsertps $0x10,(%r12,%r8,4),%xmm0,%xmm0
- .byte 196,193,122,16,52,148 // vmovss (%r12,%rdx,4),%xmm6
- .byte 196,227,121,33,198,32 // vinsertps $0x20,%xmm6,%xmm0,%xmm0
- .byte 196,193,122,16,52,156 // vmovss (%r12,%rbx,4),%xmm6
- .byte 196,227,121,33,198,48 // vinsertps $0x30,%xmm6,%xmm0,%xmm0
- .byte 196,193,122,16,52,140 // vmovss (%r12,%rcx,4),%xmm6
- .byte 196,195,73,33,52,172,16 // vinsertps $0x10,(%r12,%rbp,4),%xmm6,%xmm6
- .byte 196,129,122,16,60,188 // vmovss (%r12,%r15,4),%xmm7
- .byte 196,227,73,33,247,32 // vinsertps $0x20,%xmm7,%xmm6,%xmm6
- .byte 196,129,122,16,60,156 // vmovss (%r12,%r11,4),%xmm7
- .byte 196,227,73,33,247,48 // vinsertps $0x30,%xmm7,%xmm6,%xmm6
- .byte 196,227,77,24,240,1 // vinsertf128 $0x1,%xmm0,%ymm6,%ymm6
- .byte 196,98,117,184,108,36,64 // vfmadd231ps 0x40(%rsp),%ymm1,%ymm13
- .byte 196,226,117,184,100,36,32 // vfmadd231ps 0x20(%rsp),%ymm1,%ymm4
- .byte 196,194,117,184,216 // vfmadd231ps %ymm8,%ymm1,%ymm3
- .byte 196,194,117,184,242 // vfmadd231ps %ymm10,%ymm1,%ymm6
- .byte 196,226,125,24,5,124,153,0,0 // vbroadcastss 0x997c(%rip),%ymm0 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,226,125,24,13,67,153,0,0 // vbroadcastss 0x9943(%rip),%ymm1 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 197,254,91,243 // vcvttps2dq %ymm3,%ymm6
+ .byte 197,245,118,201 // vpcmpeqd %ymm1,%ymm1,%ymm1
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 196,194,117,146,28,176 // vgatherdps %ymm1,(%r8,%ymm6,4),%ymm3
+ .byte 197,245,118,201 // vpcmpeqd %ymm1,%ymm1,%ymm1
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 196,194,117,146,36,144 // vgatherdps %ymm1,(%r8,%ymm2,4),%ymm4
+ .byte 197,252,17,100,36,224 // vmovups %ymm4,-0x20(%rsp)
+ .byte 76,139,64,16 // mov 0x10(%rax),%r8
+ .byte 197,221,118,228 // vpcmpeqd %ymm4,%ymm4,%ymm4
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 196,194,93,146,12,176 // vgatherdps %ymm4,(%r8,%ymm6,4),%ymm1
+ .byte 197,221,118,228 // vpcmpeqd %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 196,194,93,146,44,144 // vgatherdps %ymm4,(%r8,%ymm2,4),%ymm5
+ .byte 197,252,17,108,36,192 // vmovups %ymm5,-0x40(%rsp)
+ .byte 76,139,64,24 // mov 0x18(%rax),%r8
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 196,194,85,146,36,176 // vgatherdps %ymm5,(%r8,%ymm6,4),%ymm4
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 196,194,85,146,60,144 // vgatherdps %ymm5,(%r8,%ymm2,4),%ymm7
+ .byte 197,252,17,124,36,160 // vmovups %ymm7,-0x60(%rsp)
+ .byte 76,139,64,32 // mov 0x20(%rax),%r8
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 196,194,85,146,60,176 // vgatherdps %ymm5,(%r8,%ymm6,4),%ymm7
+ .byte 197,252,17,124,36,128 // vmovups %ymm7,-0x80(%rsp)
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,28,87,228 // vxorps %ymm12,%ymm12,%ymm12
+ .byte 196,66,85,146,36,144 // vgatherdps %ymm5,(%r8,%ymm2,4),%ymm12
+ .byte 76,139,64,40 // mov 0x28(%rax),%r8
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,52,87,201 // vxorps %ymm9,%ymm9,%ymm9
+ .byte 196,66,85,146,12,176 // vgatherdps %ymm5,(%r8,%ymm6,4),%ymm9
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,44,87,210 // vxorps %ymm10,%ymm10,%ymm10
+ .byte 196,66,85,146,20,144 // vgatherdps %ymm5,(%r8,%ymm2,4),%ymm10
+ .byte 76,139,64,48 // mov 0x30(%rax),%r8
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,20,87,237 // vxorps %ymm13,%ymm13,%ymm13
+ .byte 196,66,85,146,44,176 // vgatherdps %ymm5,(%r8,%ymm6,4),%ymm13
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,12,87,246 // vxorps %ymm14,%ymm14,%ymm14
+ .byte 196,66,85,146,52,144 // vgatherdps %ymm5,(%r8,%ymm2,4),%ymm14
+ .byte 76,139,64,56 // mov 0x38(%rax),%r8
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 196,65,4,87,255 // vxorps %ymm15,%ymm15,%ymm15
+ .byte 196,66,85,146,60,176 // vgatherdps %ymm5,(%r8,%ymm6,4),%ymm15
+ .byte 197,213,118,237 // vpcmpeqd %ymm5,%ymm5,%ymm5
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 196,194,85,146,60,144 // vgatherdps %ymm5,(%r8,%ymm2,4),%ymm7
+ .byte 72,139,64,64 // mov 0x40(%rax),%rax
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 196,226,61,146,44,176 // vgatherdps %ymm8,(%rax,%ymm6,4),%ymm5
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 196,226,61,146,52,144 // vgatherdps %ymm8,(%rax,%ymm2,4),%ymm6
+ .byte 196,98,125,184,203 // vfmadd231ps %ymm3,%ymm0,%ymm9
+ .byte 196,98,125,184,233 // vfmadd231ps %ymm1,%ymm0,%ymm13
+ .byte 196,98,125,184,252 // vfmadd231ps %ymm4,%ymm0,%ymm15
+ .byte 196,226,125,184,108,36,128 // vfmadd231ps -0x80(%rsp),%ymm0,%ymm5
+ .byte 196,98,37,184,84,36,224 // vfmadd231ps -0x20(%rsp),%ymm11,%ymm10
+ .byte 196,98,37,184,116,36,192 // vfmadd231ps -0x40(%rsp),%ymm11,%ymm14
+ .byte 196,226,37,184,124,36,160 // vfmadd231ps -0x60(%rsp),%ymm11,%ymm7
+ .byte 196,194,37,184,244 // vfmadd231ps %ymm12,%ymm11,%ymm6
+ .byte 196,226,125,24,5,87,153,0,0 // vbroadcastss 0x9957(%rip),%ymm0 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,226,125,24,13,30,153,0,0 // vbroadcastss 0x991e(%rip),%ymm1 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,168,209 // vfmadd213ps %ymm1,%ymm0,%ymm10
+ .byte 196,98,125,168,201 // vfmadd213ps %ymm1,%ymm0,%ymm9
+ .byte 196,98,125,168,241 // vfmadd213ps %ymm1,%ymm0,%ymm14
.byte 196,98,125,168,233 // vfmadd213ps %ymm1,%ymm0,%ymm13
+ .byte 196,226,125,168,249 // vfmadd213ps %ymm1,%ymm0,%ymm7
.byte 196,98,125,168,249 // vfmadd213ps %ymm1,%ymm0,%ymm15
- .byte 196,226,125,168,225 // vfmadd213ps %ymm1,%ymm0,%ymm4
- .byte 196,98,125,168,217 // vfmadd213ps %ymm1,%ymm0,%ymm11
- .byte 196,226,125,168,217 // vfmadd213ps %ymm1,%ymm0,%ymm3
- .byte 196,226,125,168,209 // vfmadd213ps %ymm1,%ymm0,%ymm2
.byte 196,226,125,168,241 // vfmadd213ps %ymm1,%ymm0,%ymm6
.byte 196,226,125,168,233 // vfmadd213ps %ymm1,%ymm0,%ymm5
- .byte 196,193,126,91,199 // vcvttps2dq %ymm15,%ymm0
- .byte 197,253,111,61,58,161,0,0 // vmovdqa 0xa13a(%rip),%ymm7 # 39f60 <_sk_srcover_bgra_8888_sse2_lowp+0xad8>
- .byte 196,226,125,0,199 // vpshufb %ymm7,%ymm0,%ymm0
+ .byte 196,193,126,91,193 // vcvttps2dq %ymm9,%ymm0
+ .byte 197,253,111,29,25,161,0,0 // vmovdqa 0xa119(%rip),%ymm3 # 391e0 <_sk_srcover_bgra_8888_sse2_lowp+0xadc>
+ .byte 196,226,125,0,195 // vpshufb %ymm3,%ymm0,%ymm0
.byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,193,126,91,205 // vcvttps2dq %ymm13,%ymm1
- .byte 196,226,117,0,207 // vpshufb %ymm7,%ymm1,%ymm1
+ .byte 196,193,126,91,202 // vcvttps2dq %ymm10,%ymm1
+ .byte 196,226,117,0,203 // vpshufb %ymm3,%ymm1,%ymm1
.byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
.byte 196,227,125,56,193,1 // vinserti128 $0x1,%xmm1,%ymm0,%ymm0
- .byte 196,193,126,91,203 // vcvttps2dq %ymm11,%ymm1
- .byte 196,226,117,0,207 // vpshufb %ymm7,%ymm1,%ymm1
+ .byte 196,193,126,91,205 // vcvttps2dq %ymm13,%ymm1
+ .byte 196,226,117,0,203 // vpshufb %ymm3,%ymm1,%ymm1
.byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 197,254,91,228 // vcvttps2dq %ymm4,%ymm4
- .byte 196,226,93,0,231 // vpshufb %ymm7,%ymm4,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,117,56,204,1 // vinserti128 $0x1,%xmm4,%ymm1,%ymm1
- .byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
- .byte 196,226,109,0,215 // vpshufb %ymm7,%ymm2,%ymm2
+ .byte 196,193,126,91,214 // vcvttps2dq %ymm14,%ymm2
+ .byte 196,226,109,0,211 // vpshufb %ymm3,%ymm2,%ymm2
.byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
- .byte 197,254,91,219 // vcvttps2dq %ymm3,%ymm3
- .byte 196,226,101,0,223 // vpshufb %ymm7,%ymm3,%ymm3
- .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
- .byte 196,227,109,56,211,1 // vinserti128 $0x1,%xmm3,%ymm2,%ymm2
- .byte 197,254,91,221 // vcvttps2dq %ymm5,%ymm3
- .byte 196,226,101,0,223 // vpshufb %ymm7,%ymm3,%ymm3
- .byte 197,254,91,230 // vcvttps2dq %ymm6,%ymm4
- .byte 196,226,93,0,231 // vpshufb %ymm7,%ymm4,%ymm4
- .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
+ .byte 196,227,117,56,202,1 // vinserti128 $0x1,%xmm2,%ymm1,%ymm1
+ .byte 196,193,126,91,215 // vcvttps2dq %ymm15,%ymm2
+ .byte 196,226,109,0,211 // vpshufb %ymm3,%ymm2,%ymm2
+ .byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
+ .byte 197,254,91,231 // vcvttps2dq %ymm7,%ymm4
+ .byte 196,226,93,0,227 // vpshufb %ymm3,%ymm4,%ymm4
.byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,101,56,220,1 // vinserti128 $0x1,%xmm4,%ymm3,%ymm3
- .byte 72,139,52,36 // mov (%rsp),%rsi
+ .byte 196,227,109,56,212,1 // vinserti128 $0x1,%xmm4,%ymm2,%ymm2
+ .byte 197,254,91,229 // vcvttps2dq %ymm5,%ymm4
+ .byte 196,226,93,0,227 // vpshufb %ymm3,%ymm4,%ymm4
+ .byte 197,254,91,238 // vcvttps2dq %ymm6,%ymm5
+ .byte 196,226,85,0,219 // vpshufb %ymm3,%ymm5,%ymm3
+ .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
+ .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
+ .byte 196,227,93,56,219,1 // vinserti128 $0x1,%xmm3,%ymm4,%ymm3
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,139,124,36,8 // mov 0x8(%rsp),%rdi
- .byte 72,139,84,36,16 // mov 0x10(%rsp),%rdx
- .byte 72,139,76,36,24 // mov 0x18(%rsp),%rcx
- .byte 197,252,16,100,36,96 // vmovups 0x60(%rsp),%ymm4
- .byte 197,252,16,172,36,128,0,0,0 // vmovups 0x80(%rsp),%ymm5
- .byte 197,252,16,180,36,160,0,0,0 // vmovups 0xa0(%rsp),%ymm6
- .byte 197,252,16,188,36,192,0,0,0 // vmovups 0xc0(%rsp),%ymm7
- .byte 72,129,196,232,0,0,0 // add $0xe8,%rsp
- .byte 91 // pop %rbx
- .byte 65,92 // pop %r12
- .byte 65,93 // pop %r13
- .byte 65,94 // pop %r14
- .byte 65,95 // pop %r15
- .byte 93 // pop %rbp
+ .byte 197,252,16,36,36 // vmovups (%rsp),%ymm4
+ .byte 197,252,16,108,36,32 // vmovups 0x20(%rsp),%ymm5
+ .byte 197,252,16,116,36,64 // vmovups 0x40(%rsp),%ymm6
+ .byte 197,252,16,124,36,96 // vmovups 0x60(%rsp),%ymm7
+ .byte 72,129,196,152,0,0,0 // add $0x98,%rsp
.byte 255,224 // jmpq *%rax
HIDDEN _sk_evenly_spaced_2_stop_gradient_hsw_lowp
@@ -56565,12 +55949,12 @@ _sk_evenly_spaced_2_stop_gradient_hsw_lowp:
.byte 197,124,40,194 // vmovaps %ymm2,%ymm8
.byte 196,98,125,168,195 // vfmadd213ps %ymm3,%ymm0,%ymm8
.byte 196,226,117,168,211 // vfmadd213ps %ymm3,%ymm1,%ymm2
- .byte 196,226,125,24,29,67,152,0,0 // vbroadcastss 0x9843(%rip),%ymm3 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 196,98,125,24,29,10,152,0,0 // vbroadcastss 0x980a(%rip),%ymm11 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,226,125,24,29,67,152,0,0 // vbroadcastss 0x9843(%rip),%ymm3 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 196,98,125,24,29,10,152,0,0 // vbroadcastss 0x980a(%rip),%ymm11 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 196,194,101,168,211 // vfmadd213ps %ymm11,%ymm3,%ymm2
.byte 196,66,101,168,195 // vfmadd213ps %ymm11,%ymm3,%ymm8
.byte 196,65,126,91,192 // vcvttps2dq %ymm8,%ymm8
- .byte 197,125,111,21,63,160,0,0 // vmovdqa 0xa03f(%rip),%ymm10 # 39f80 <_sk_srcover_bgra_8888_sse2_lowp+0xaf8>
+ .byte 197,125,111,21,67,160,0,0 // vmovdqa 0xa043(%rip),%ymm10 # 39200 <_sk_srcover_bgra_8888_sse2_lowp+0xafc>
.byte 196,66,61,0,194 // vpshufb %ymm10,%ymm8,%ymm8
.byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
.byte 197,254,91,210 // vcvttps2dq %ymm2,%ymm2
@@ -56634,7 +56018,7 @@ _sk_xy_to_unit_angle_hsw_lowp:
.byte 197,252,40,227 // vmovaps %ymm3,%ymm4
.byte 197,252,40,217 // vmovaps %ymm1,%ymm3
.byte 197,252,40,200 // vmovaps %ymm0,%ymm1
- .byte 196,98,125,24,5,93,152,0,0 // vbroadcastss 0x985d(%rip),%ymm8 # 398c8 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
+ .byte 196,98,125,24,5,93,152,0,0 // vbroadcastss 0x985d(%rip),%ymm8 # 38b44 <_sk_srcover_bgra_8888_sse2_lowp+0x440>
.byte 196,65,116,84,200 // vandps %ymm8,%ymm1,%ymm9
.byte 196,65,100,84,208 // vandps %ymm8,%ymm3,%ymm10
.byte 196,65,108,84,216 // vandps %ymm8,%ymm2,%ymm11
@@ -56648,20 +56032,20 @@ _sk_xy_to_unit_angle_hsw_lowp:
.byte 196,67,45,74,192,192 // vblendvps %ymm12,%ymm8,%ymm10,%ymm8
.byte 196,65,4,94,192 // vdivps %ymm8,%ymm15,%ymm8
.byte 196,65,60,89,208 // vmulps %ymm8,%ymm8,%ymm10
- .byte 196,98,125,24,29,169,151,0,0 // vbroadcastss 0x97a9(%rip),%ymm11 # 39864 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
- .byte 196,98,125,24,53,164,151,0,0 // vbroadcastss 0x97a4(%rip),%ymm14 # 39868 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
+ .byte 196,98,125,24,29,169,151,0,0 // vbroadcastss 0x97a9(%rip),%ymm11 # 38ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x3dc>
+ .byte 196,98,125,24,53,164,151,0,0 // vbroadcastss 0x97a4(%rip),%ymm14 # 38ae4 <_sk_srcover_bgra_8888_sse2_lowp+0x3e0>
.byte 196,65,124,40,251 // vmovaps %ymm11,%ymm15
.byte 196,66,45,168,254 // vfmadd213ps %ymm14,%ymm10,%ymm15
.byte 196,193,52,89,193 // vmulps %ymm9,%ymm9,%ymm0
.byte 196,66,125,168,222 // vfmadd213ps %ymm14,%ymm0,%ymm11
- .byte 196,98,125,24,53,139,151,0,0 // vbroadcastss 0x978b(%rip),%ymm14 # 3986c <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
+ .byte 196,98,125,24,53,139,151,0,0 // vbroadcastss 0x978b(%rip),%ymm14 # 38ae8 <_sk_srcover_bgra_8888_sse2_lowp+0x3e4>
.byte 196,66,125,168,222 // vfmadd213ps %ymm14,%ymm0,%ymm11
.byte 196,66,45,168,254 // vfmadd213ps %ymm14,%ymm10,%ymm15
- .byte 196,98,125,24,53,124,151,0,0 // vbroadcastss 0x977c(%rip),%ymm14 # 39870 <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
+ .byte 196,98,125,24,53,124,151,0,0 // vbroadcastss 0x977c(%rip),%ymm14 # 38aec <_sk_srcover_bgra_8888_sse2_lowp+0x3e8>
.byte 196,66,45,168,254 // vfmadd213ps %ymm14,%ymm10,%ymm15
.byte 196,66,125,168,222 // vfmadd213ps %ymm14,%ymm0,%ymm11
.byte 196,193,52,89,195 // vmulps %ymm11,%ymm9,%ymm0
- .byte 196,98,125,24,13,104,151,0,0 // vbroadcastss 0x9768(%rip),%ymm9 # 39874 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
+ .byte 196,98,125,24,13,104,151,0,0 // vbroadcastss 0x9768(%rip),%ymm9 # 38af0 <_sk_srcover_bgra_8888_sse2_lowp+0x3ec>
.byte 197,52,92,208 // vsubps %ymm0,%ymm9,%ymm10
.byte 196,195,125,74,194,208 // vblendvps %ymm13,%ymm10,%ymm0,%ymm0
.byte 196,65,60,89,199 // vmulps %ymm15,%ymm8,%ymm8
@@ -56669,14 +56053,14 @@ _sk_xy_to_unit_angle_hsw_lowp:
.byte 196,67,61,74,193,192 // vblendvps %ymm12,%ymm9,%ymm8,%ymm8
.byte 196,65,52,87,201 // vxorps %ymm9,%ymm9,%ymm9
.byte 196,193,116,194,201,1 // vcmpltps %ymm9,%ymm1,%ymm1
- .byte 196,98,125,24,21,250,149,0,0 // vbroadcastss 0x95fa(%rip),%ymm10 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 196,98,125,24,21,250,149,0,0 // vbroadcastss 0x95fa(%rip),%ymm10 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 197,44,92,216 // vsubps %ymm0,%ymm10,%ymm11
.byte 196,195,125,74,195,16 // vblendvps %ymm1,%ymm11,%ymm0,%ymm0
.byte 196,193,100,194,201,1 // vcmpltps %ymm9,%ymm3,%ymm1
.byte 196,193,44,92,216 // vsubps %ymm8,%ymm10,%ymm3
.byte 196,227,61,74,203,16 // vblendvps %ymm1,%ymm3,%ymm8,%ymm1
.byte 196,193,108,194,217,1 // vcmpltps %ymm9,%ymm2,%ymm3
- .byte 196,98,125,24,5,212,149,0,0 // vbroadcastss 0x95d4(%rip),%ymm8 # 39738 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
+ .byte 196,98,125,24,5,212,149,0,0 // vbroadcastss 0x95d4(%rip),%ymm8 # 389b4 <_sk_srcover_bgra_8888_sse2_lowp+0x2b0>
.byte 197,60,92,208 // vsubps %ymm0,%ymm8,%ymm10
.byte 196,195,125,74,194,48 // vblendvps %ymm3,%ymm10,%ymm0,%ymm0
.byte 196,193,92,194,217,1 // vcmpltps %ymm9,%ymm4,%ymm3
@@ -56722,9 +56106,9 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,194 // movzbl %r10b,%r8d
.byte 65,128,248,14 // cmp $0xe,%r8b
- .byte 119,51 // ja 30228 <_sk_srcover_rgba_8888_hsw_lowp+0x5c>
+ .byte 119,51 // ja 2f4a4 <_sk_srcover_rgba_8888_hsw_lowp+0x5c>
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
- .byte 76,141,29,236,2,0,0 // lea 0x2ec(%rip),%r11 # 304ec <_sk_srcover_rgba_8888_hsw_lowp+0x320>
+ .byte 76,141,29,236,2,0,0 // lea 0x2ec(%rip),%r11 # 2f768 <_sk_srcover_rgba_8888_hsw_lowp+0x320>
.byte 75,99,4,131 // movslq (%r11,%r8,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56734,10 +56118,10 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 255,224 // jmpq *%rax
.byte 196,193,121,110,41 // vmovd (%r9),%xmm5
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
- .byte 233,22,1,0,0 // jmpq 3033e <_sk_srcover_rgba_8888_hsw_lowp+0x172>
+ .byte 233,22,1,0,0 // jmpq 2f5ba <_sk_srcover_rgba_8888_hsw_lowp+0x172>
.byte 196,193,126,111,41 // vmovdqu (%r9),%ymm5
.byte 196,193,126,111,97,32 // vmovdqu 0x20(%r9),%ymm4
- .byte 233,6,1,0,0 // jmpq 3033e <_sk_srcover_rgba_8888_hsw_lowp+0x172>
+ .byte 233,6,1,0,0 // jmpq 2f5ba <_sk_srcover_rgba_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,8 // vmovd 0x8(%r9),%xmm4
.byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
.byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
@@ -56746,7 +56130,7 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 197,249,112,237,232 // vpshufd $0xe8,%xmm5,%xmm5
.byte 196,227,93,2,237,3 // vpblendd $0x3,%ymm5,%ymm4,%ymm5
.byte 197,253,111,231 // vmovdqa %ymm7,%ymm4
- .byte 233,216,0,0,0 // jmpq 3033e <_sk_srcover_rgba_8888_hsw_lowp+0x172>
+ .byte 233,216,0,0,0 // jmpq 2f5ba <_sk_srcover_rgba_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,24 // vmovd 0x18(%r9),%xmm4
.byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56762,7 +56146,7 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 196,193,122,111,41 // vmovdqu (%r9),%xmm5
.byte 196,227,85,2,236,240 // vpblendd $0xf0,%ymm4,%ymm5,%ymm5
.byte 197,125,127,196 // vmovdqa %ymm8,%ymm4
- .byte 233,129,0,0,0 // jmpq 3033e <_sk_srcover_rgba_8888_hsw_lowp+0x172>
+ .byte 233,129,0,0,0 // jmpq 2f5ba <_sk_srcover_rgba_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,40 // vmovd 0x28(%r9),%xmm4
.byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56772,7 +56156,7 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 196,193,121,110,105,32 // vmovd 0x20(%r9),%xmm5
.byte 196,227,93,2,229,1 // vpblendd $0x1,%ymm5,%ymm4,%ymm4
.byte 196,193,126,111,41 // vmovdqu (%r9),%ymm5
- .byte 235,76 // jmp 3033e <_sk_srcover_rgba_8888_hsw_lowp+0x172>
+ .byte 235,76 // jmp 2f5ba <_sk_srcover_rgba_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,56 // vmovd 0x38(%r9),%xmm4
.byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56792,9 +56176,9 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 196,227,69,14,236,170 // vpblendw $0xaa,%ymm4,%ymm7,%ymm5
.byte 196,227,77,14,228,170 // vpblendw $0xaa,%ymm4,%ymm6,%ymm4
.byte 196,226,93,43,237 // vpackusdw %ymm5,%ymm4,%ymm5
- .byte 197,125,111,5,57,156,0,0 // vmovdqa 0x9c39(%rip),%ymm8 # 39fa0 <_sk_srcover_bgra_8888_sse2_lowp+0xb18>
+ .byte 197,125,111,5,61,156,0,0 // vmovdqa 0x9c3d(%rip),%ymm8 # 39220 <_sk_srcover_bgra_8888_sse2_lowp+0xb1c>
.byte 196,193,85,219,224 // vpand %ymm8,%ymm5,%ymm4
- .byte 196,98,125,121,13,75,156,0,0 // vpbroadcastw 0x9c4b(%rip),%ymm9 # 39fc0 <_sk_srcover_bgra_8888_sse2_lowp+0xb38>
+ .byte 196,98,125,121,13,79,156,0,0 // vpbroadcastw 0x9c4f(%rip),%ymm9 # 39240 <_sk_srcover_bgra_8888_sse2_lowp+0xb3c>
.byte 197,213,113,213,8 // vpsrlw $0x8,%ymm5,%ymm5
.byte 197,197,114,215,16 // vpsrld $0x10,%ymm7,%ymm7
.byte 197,205,114,214,16 // vpsrld $0x10,%ymm6,%ymm6
@@ -56833,19 +56217,19 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 196,193,53,114,243,16 // vpslld $0x10,%ymm11,%ymm9
.byte 196,65,53,235,202 // vpor %ymm10,%ymm9,%ymm9
.byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,26 // ja 3044f <_sk_srcover_rgba_8888_hsw_lowp+0x283>
- .byte 76,141,21,236,0,0,0 // lea 0xec(%rip),%r10 # 30528 <_sk_srcover_rgba_8888_hsw_lowp+0x35c>
+ .byte 119,26 // ja 2f6cb <_sk_srcover_rgba_8888_hsw_lowp+0x283>
+ .byte 76,141,21,236,0,0,0 // lea 0xec(%rip),%r10 # 2f7a4 <_sk_srcover_rgba_8888_hsw_lowp+0x35c>
.byte 75,99,4,130 // movslq (%r10,%r8,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 196,65,121,126,1 // vmovd %xmm8,(%r9)
- .byte 233,151,0,0,0 // jmpq 304e6 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
+ .byte 233,151,0,0,0 // jmpq 2f762 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
.byte 196,65,126,127,1 // vmovdqu %ymm8,(%r9)
.byte 196,65,126,127,73,32 // vmovdqu %ymm9,0x20(%r9)
- .byte 233,135,0,0,0 // jmpq 304e6 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
+ .byte 233,135,0,0,0 // jmpq 2f762 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
.byte 196,67,121,22,65,8,2 // vpextrd $0x2,%xmm8,0x8(%r9)
.byte 196,65,121,214,1 // vmovq %xmm8,(%r9)
- .byte 235,121 // jmp 304e6 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
+ .byte 235,121 // jmp 2f762 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,22,73,24,2 // vpextrd $0x2,%xmm9,0x18(%r9)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
@@ -56853,12 +56237,12 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,65,121,126,73,16 // vmovd %xmm9,0x10(%r9)
.byte 196,65,122,127,1 // vmovdqu %xmm8,(%r9)
- .byte 235,76 // jmp 304e6 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
+ .byte 235,76 // jmp 2f762 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
.byte 196,67,121,22,73,40,2 // vpextrd $0x2,%xmm9,0x28(%r9)
.byte 196,67,121,22,73,36,1 // vpextrd $0x1,%xmm9,0x24(%r9)
.byte 196,65,121,126,73,32 // vmovd %xmm9,0x20(%r9)
.byte 196,65,126,127,1 // vmovdqu %ymm8,(%r9)
- .byte 235,49 // jmp 304e6 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
+ .byte 235,49 // jmp 2f762 <_sk_srcover_rgba_8888_hsw_lowp+0x31a>
.byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
.byte 196,67,121,22,81,56,2 // vpextrd $0x2,%xmm10,0x38(%r9)
.byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
@@ -56906,7 +56290,7 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 255,6 // incl (%rsi)
.byte 254 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,62 // lcall *0x3effffff(%rip) # 3f03052c <_sk_srcover_bgra_8888_sse2_lowp+0x3eff70a4>
+ .byte 255,29,255,255,255,62 // lcall *0x3effffff(%rip) # 3f02f7a8 <_sk_srcover_bgra_8888_sse2_lowp+0x3eff70a4>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,55 // pushq (%rdi)
@@ -56924,7 +56308,7 @@ _sk_srcover_rgba_8888_hsw_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 121,255 // jns 3054d <_sk_srcover_rgba_8888_hsw_lowp+0x381>
+ .byte 121,255 // jns 2f7c9 <_sk_srcover_rgba_8888_hsw_lowp+0x381>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -56951,9 +56335,9 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,194 // movzbl %r10b,%r8d
.byte 65,128,248,14 // cmp $0xe,%r8b
- .byte 119,51 // ja 305c0 <_sk_srcover_bgra_8888_hsw_lowp+0x5c>
+ .byte 119,51 // ja 2f83c <_sk_srcover_bgra_8888_hsw_lowp+0x5c>
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
- .byte 76,141,29,236,2,0,0 // lea 0x2ec(%rip),%r11 # 30884 <_sk_srcover_bgra_8888_hsw_lowp+0x320>
+ .byte 76,141,29,236,2,0,0 // lea 0x2ec(%rip),%r11 # 2fb00 <_sk_srcover_bgra_8888_hsw_lowp+0x320>
.byte 75,99,4,131 // movslq (%r11,%r8,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56963,10 +56347,10 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 255,224 // jmpq *%rax
.byte 196,193,121,110,41 // vmovd (%r9),%xmm5
.byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
- .byte 233,22,1,0,0 // jmpq 306d6 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
+ .byte 233,22,1,0,0 // jmpq 2f952 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
.byte 196,193,126,111,41 // vmovdqu (%r9),%ymm5
.byte 196,193,126,111,97,32 // vmovdqu 0x20(%r9),%ymm4
- .byte 233,6,1,0,0 // jmpq 306d6 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
+ .byte 233,6,1,0,0 // jmpq 2f952 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,8 // vmovd 0x8(%r9),%xmm4
.byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
.byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
@@ -56975,7 +56359,7 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 197,249,112,237,232 // vpshufd $0xe8,%xmm5,%xmm5
.byte 196,227,93,2,237,3 // vpblendd $0x3,%ymm5,%ymm4,%ymm5
.byte 197,253,111,231 // vmovdqa %ymm7,%ymm4
- .byte 233,216,0,0,0 // jmpq 306d6 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
+ .byte 233,216,0,0,0 // jmpq 2f952 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,24 // vmovd 0x18(%r9),%xmm4
.byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -56991,7 +56375,7 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 196,193,122,111,41 // vmovdqu (%r9),%xmm5
.byte 196,227,85,2,236,240 // vpblendd $0xf0,%ymm4,%ymm5,%ymm5
.byte 197,125,127,196 // vmovdqa %ymm8,%ymm4
- .byte 233,129,0,0,0 // jmpq 306d6 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
+ .byte 233,129,0,0,0 // jmpq 2f952 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,40 // vmovd 0x28(%r9),%xmm4
.byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -57001,7 +56385,7 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 196,193,121,110,105,32 // vmovd 0x20(%r9),%xmm5
.byte 196,227,93,2,229,1 // vpblendd $0x1,%ymm5,%ymm4,%ymm4
.byte 196,193,126,111,41 // vmovdqu (%r9),%ymm5
- .byte 235,76 // jmp 306d6 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
+ .byte 235,76 // jmp 2f952 <_sk_srcover_bgra_8888_hsw_lowp+0x172>
.byte 196,193,121,110,97,56 // vmovd 0x38(%r9),%xmm4
.byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
.byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
@@ -57021,9 +56405,9 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 196,227,93,14,245,170 // vpblendw $0xaa,%ymm5,%ymm4,%ymm6
.byte 196,227,69,14,237,170 // vpblendw $0xaa,%ymm5,%ymm7,%ymm5
.byte 196,226,85,43,238 // vpackusdw %ymm6,%ymm5,%ymm5
- .byte 197,125,111,5,225,152,0,0 // vmovdqa 0x98e1(%rip),%ymm8 # 39fe0 <_sk_srcover_bgra_8888_sse2_lowp+0xb58>
+ .byte 197,125,111,5,229,152,0,0 // vmovdqa 0x98e5(%rip),%ymm8 # 39260 <_sk_srcover_bgra_8888_sse2_lowp+0xb5c>
.byte 196,193,85,219,240 // vpand %ymm8,%ymm5,%ymm6
- .byte 196,98,125,121,13,243,152,0,0 // vpbroadcastw 0x98f3(%rip),%ymm9 # 3a000 <_sk_srcover_bgra_8888_sse2_lowp+0xb78>
+ .byte 196,98,125,121,13,247,152,0,0 // vpbroadcastw 0x98f7(%rip),%ymm9 # 39280 <_sk_srcover_bgra_8888_sse2_lowp+0xb7c>
.byte 197,213,113,213,8 // vpsrlw $0x8,%ymm5,%ymm5
.byte 197,221,114,212,16 // vpsrld $0x10,%ymm4,%ymm4
.byte 197,197,114,215,16 // vpsrld $0x10,%ymm7,%ymm7
@@ -57062,19 +56446,19 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 196,193,53,114,243,16 // vpslld $0x10,%ymm11,%ymm9
.byte 196,65,53,235,202 // vpor %ymm10,%ymm9,%ymm9
.byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,26 // ja 307e7 <_sk_srcover_bgra_8888_hsw_lowp+0x283>
- .byte 76,141,21,236,0,0,0 // lea 0xec(%rip),%r10 # 308c0 <_sk_srcover_bgra_8888_hsw_lowp+0x35c>
+ .byte 119,26 // ja 2fa63 <_sk_srcover_bgra_8888_hsw_lowp+0x283>
+ .byte 76,141,21,236,0,0,0 // lea 0xec(%rip),%r10 # 2fb3c <_sk_srcover_bgra_8888_hsw_lowp+0x35c>
.byte 75,99,4,130 // movslq (%r10,%r8,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 196,65,121,126,1 // vmovd %xmm8,(%r9)
- .byte 233,151,0,0,0 // jmpq 3087e <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
+ .byte 233,151,0,0,0 // jmpq 2fafa <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
.byte 196,65,126,127,1 // vmovdqu %ymm8,(%r9)
.byte 196,65,126,127,73,32 // vmovdqu %ymm9,0x20(%r9)
- .byte 233,135,0,0,0 // jmpq 3087e <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
+ .byte 233,135,0,0,0 // jmpq 2fafa <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
.byte 196,67,121,22,65,8,2 // vpextrd $0x2,%xmm8,0x8(%r9)
.byte 196,65,121,214,1 // vmovq %xmm8,(%r9)
- .byte 235,121 // jmp 3087e <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
+ .byte 235,121 // jmp 2fafa <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,67,121,22,73,24,2 // vpextrd $0x2,%xmm9,0x18(%r9)
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
@@ -57082,12 +56466,12 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
.byte 196,65,121,126,73,16 // vmovd %xmm9,0x10(%r9)
.byte 196,65,122,127,1 // vmovdqu %xmm8,(%r9)
- .byte 235,76 // jmp 3087e <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
+ .byte 235,76 // jmp 2fafa <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
.byte 196,67,121,22,73,40,2 // vpextrd $0x2,%xmm9,0x28(%r9)
.byte 196,67,121,22,73,36,1 // vpextrd $0x1,%xmm9,0x24(%r9)
.byte 196,65,121,126,73,32 // vmovd %xmm9,0x20(%r9)
.byte 196,65,126,127,1 // vmovdqu %ymm8,(%r9)
- .byte 235,49 // jmp 3087e <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
+ .byte 235,49 // jmp 2fafa <_sk_srcover_bgra_8888_hsw_lowp+0x31a>
.byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
.byte 196,67,121,22,81,56,2 // vpextrd $0x2,%xmm10,0x38(%r9)
.byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
@@ -57135,7 +56519,7 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 255,6 // incl (%rsi)
.byte 254 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,62 // lcall *0x3effffff(%rip) # 3f0308c4 <_sk_srcover_bgra_8888_sse2_lowp+0x3eff743c>
+ .byte 255,29,255,255,255,62 // lcall *0x3effffff(%rip) # 3f02fb40 <_sk_srcover_bgra_8888_sse2_lowp+0x3eff743c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,55 // pushq (%rdi)
@@ -57153,7 +56537,7 @@ _sk_srcover_bgra_8888_hsw_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 121,255 // jns 308e5 <_sk_srcover_bgra_8888_hsw_lowp+0x381>
+ .byte 121,255 // jns 2fb61 <_sk_srcover_bgra_8888_hsw_lowp+0x381>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -57186,13 +56570,13 @@ _sk_start_pipeline_sse41_lowp:
.byte 73,137,246 // mov %rsi,%r14
.byte 72,137,77,192 // mov %rcx,-0x40(%rbp)
.byte 72,57,203 // cmp %rcx,%rbx
- .byte 15,131,131,0,0,0 // jae 309b2 <_sk_start_pipeline_sse41_lowp+0xb6>
+ .byte 15,131,131,0,0,0 // jae 2fc2e <_sk_start_pipeline_sse41_lowp+0xb6>
.byte 72,139,69,208 // mov -0x30(%rbp),%rax
.byte 72,141,64,8 // lea 0x8(%rax),%rax
.byte 72,137,69,200 // mov %rax,-0x38(%rbp)
.byte 76,57,125,200 // cmp %r15,-0x38(%rbp)
.byte 72,139,85,208 // mov -0x30(%rbp),%rdx
- .byte 119,59 // ja 30980 <_sk_start_pipeline_sse41_lowp+0x84>
+ .byte 119,59 // ja 2fbfc <_sk_start_pipeline_sse41_lowp+0x84>
.byte 76,139,101,208 // mov -0x30(%rbp),%r12
.byte 49,255 // xor %edi,%edi
.byte 15,87,192 // xorps %xmm0,%xmm0
@@ -57211,10 +56595,10 @@ _sk_start_pipeline_sse41_lowp:
.byte 73,131,196,16 // add $0x10,%r12
.byte 77,57,252 // cmp %r15,%r12
.byte 73,137,212 // mov %rdx,%r12
- .byte 118,201 // jbe 30949 <_sk_start_pipeline_sse41_lowp+0x4d>
+ .byte 118,201 // jbe 2fbc5 <_sk_start_pipeline_sse41_lowp+0x4d>
.byte 76,137,255 // mov %r15,%rdi
.byte 72,41,215 // sub %rdx,%rdi
- .byte 116,33 // je 309a9 <_sk_start_pipeline_sse41_lowp+0xad>
+ .byte 116,33 // je 2fc25 <_sk_start_pipeline_sse41_lowp+0xad>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -57228,7 +56612,7 @@ _sk_start_pipeline_sse41_lowp:
.byte 65,255,213 // callq *%r13
.byte 72,255,195 // inc %rbx
.byte 72,59,93,192 // cmp -0x40(%rbp),%rbx
- .byte 117,137 // jne 3093b <_sk_start_pipeline_sse41_lowp+0x3f>
+ .byte 117,137 // jne 2fbb7 <_sk_start_pipeline_sse41_lowp+0x3f>
.byte 72,131,196,24 // add $0x18,%rsp
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
@@ -57259,7 +56643,7 @@ _sk_seed_shader_sse41_lowp:
.byte 102,15,110,209 // movd %ecx,%xmm2
.byte 102,15,112,210,0 // pshufd $0x0,%xmm2,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,88,21,112,150,0,0 // addps 0x9670(%rip),%xmm2 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,21,116,150,0,0 // addps 0x9674(%rip),%xmm2 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 255,224 // jmpq *%rax
@@ -57436,7 +56820,7 @@ HIDDEN _sk_black_color_sse41_lowp
FUNCTION(_sk_black_color_sse41_lowp)
_sk_black_color_sse41_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,181,147,0,0 // movaps 0x93b5(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,185,147,0,0 // movaps 0x93b9(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -57447,7 +56831,7 @@ HIDDEN _sk_white_color_sse41_lowp
FUNCTION(_sk_white_color_sse41_lowp)
_sk_white_color_sse41_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,5,161,147,0,0 // movaps 0x93a1(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,5,165,147,0,0 // movaps 0x93a5(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,40,216 // movaps %xmm0,%xmm3
@@ -57458,10 +56842,10 @@ HIDDEN _sk_set_rgb_sse41_lowp
FUNCTION(_sk_set_rgb_sse41_lowp)
_sk_set_rgb_sse41_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 243,15,16,21,224,138,0,0 // movss 0x8ae0(%rip),%xmm2 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,15,16,21,224,138,0,0 // movss 0x8ae0(%rip),%xmm2 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 243,15,16,0 // movss (%rax),%xmm0
.byte 243,15,89,194 // mulss %xmm2,%xmm0
- .byte 243,68,15,16,5,159,138,0,0 // movss 0x8a9f(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,16,5,159,138,0,0 // movss 0x8a9f(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,88,192 // addss %xmm8,%xmm0
.byte 243,68,15,44,192 // cvttss2si %xmm0,%r8d
.byte 102,65,15,110,192 // movd %r8d,%xmm0
@@ -57508,7 +56892,7 @@ HIDDEN _sk_premul_sse41_lowp
FUNCTION(_sk_premul_sse41_lowp)
_sk_premul_sse41_lowp:
.byte 102,15,213,195 // pmullw %xmm3,%xmm0
- .byte 102,68,15,111,5,237,146,0,0 // movdqa 0x92ed(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,241,146,0,0 // movdqa 0x92f1(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,203 // pmullw %xmm3,%xmm1
@@ -57525,7 +56909,7 @@ HIDDEN _sk_premul_dst_sse41_lowp
FUNCTION(_sk_premul_dst_sse41_lowp)
_sk_premul_dst_sse41_lowp:
.byte 102,15,213,231 // pmullw %xmm7,%xmm4
- .byte 102,68,15,111,5,182,146,0,0 // movdqa 0x92b6(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,186,146,0,0 // movdqa 0x92ba(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,224 // paddw %xmm8,%xmm4
.byte 102,15,113,212,8 // psrlw $0x8,%xmm4
.byte 102,15,213,239 // pmullw %xmm7,%xmm5
@@ -57573,7 +56957,7 @@ HIDDEN _sk_invert_sse41_lowp
.globl _sk_invert_sse41_lowp
FUNCTION(_sk_invert_sse41_lowp)
_sk_invert_sse41_lowp:
- .byte 102,68,15,111,5,84,146,0,0 // movdqa 0x9254(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,88,146,0,0 // movdqa 0x9258(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,200 // psubw %xmm0,%xmm9
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -57605,7 +56989,7 @@ FUNCTION(_sk_srcatop_sse41_lowp)
_sk_srcatop_sse41_lowp:
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
.byte 102,15,213,199 // pmullw %xmm7,%xmm0
- .byte 102,15,111,29,248,145,0,0 // movdqa 0x91f8(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,29,252,145,0,0 // movdqa 0x91fc(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,111,203 // movdqa %xmm3,%xmm9
.byte 102,69,15,249,200 // psubw %xmm8,%xmm9
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
@@ -57637,7 +57021,7 @@ FUNCTION(_sk_dstatop_sse41_lowp)
_sk_dstatop_sse41_lowp:
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
.byte 102,68,15,213,195 // pmullw %xmm3,%xmm8
- .byte 102,68,15,111,13,120,145,0,0 // movdqa 0x9178(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,124,145,0,0 // movdqa 0x917c(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,253,193 // paddw %xmm9,%xmm8
.byte 102,68,15,111,213 // movdqa %xmm5,%xmm10
.byte 102,68,15,213,211 // pmullw %xmm3,%xmm10
@@ -57666,7 +57050,7 @@ HIDDEN _sk_srcin_sse41_lowp
FUNCTION(_sk_srcin_sse41_lowp)
_sk_srcin_sse41_lowp:
.byte 102,15,213,199 // pmullw %xmm7,%xmm0
- .byte 102,68,15,111,5,3,145,0,0 // movdqa 0x9103(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,7,145,0,0 // movdqa 0x9107(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,207 // pmullw %xmm7,%xmm1
@@ -57687,7 +57071,7 @@ FUNCTION(_sk_dstin_sse41_lowp)
_sk_dstin_sse41_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,213,195 // pmullw %xmm3,%xmm0
- .byte 102,68,15,111,5,186,144,0,0 // movdqa 0x90ba(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,190,144,0,0 // movdqa 0x90be(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,111,205 // movdqa %xmm5,%xmm1
@@ -57708,7 +57092,7 @@ HIDDEN _sk_srcout_sse41_lowp
.globl _sk_srcout_sse41_lowp
FUNCTION(_sk_srcout_sse41_lowp)
_sk_srcout_sse41_lowp:
- .byte 102,68,15,111,5,113,144,0,0 // movdqa 0x9071(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,117,144,0,0 // movdqa 0x9075(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,207 // psubw %xmm7,%xmm9
.byte 102,65,15,213,193 // pmullw %xmm9,%xmm0
@@ -57731,7 +57115,7 @@ HIDDEN _sk_dstout_sse41_lowp
FUNCTION(_sk_dstout_sse41_lowp)
_sk_dstout_sse41_lowp:
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,68,15,111,5,26,144,0,0 // movdqa 0x901a(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,30,144,0,0 // movdqa 0x901e(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,249,216 // psubw %xmm0,%xmm3
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
@@ -57756,7 +57140,7 @@ HIDDEN _sk_srcover_sse41_lowp
.globl _sk_srcover_sse41_lowp
FUNCTION(_sk_srcover_sse41_lowp)
_sk_srcover_sse41_lowp:
- .byte 102,68,15,111,13,192,143,0,0 // movdqa 0x8fc0(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,196,143,0,0 // movdqa 0x8fc4(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
.byte 102,68,15,249,195 // psubw %xmm3,%xmm8
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -57785,7 +57169,7 @@ HIDDEN _sk_dstover_sse41_lowp
.globl _sk_dstover_sse41_lowp
FUNCTION(_sk_dstover_sse41_lowp)
_sk_dstover_sse41_lowp:
- .byte 102,68,15,111,5,70,143,0,0 // movdqa 0x8f46(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,74,143,0,0 // movdqa 0x8f4a(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,207 // psubw %xmm7,%xmm9
.byte 102,65,15,213,193 // pmullw %xmm9,%xmm0
@@ -57812,7 +57196,7 @@ HIDDEN _sk_modulate_sse41_lowp
FUNCTION(_sk_modulate_sse41_lowp)
_sk_modulate_sse41_lowp:
.byte 102,15,213,196 // pmullw %xmm4,%xmm0
- .byte 102,68,15,111,5,223,142,0,0 // movdqa 0x8edf(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,227,142,0,0 // movdqa 0x8ee3(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,205 // pmullw %xmm5,%xmm1
@@ -57831,7 +57215,7 @@ HIDDEN _sk_multiply_sse41_lowp
.globl _sk_multiply_sse41_lowp
FUNCTION(_sk_multiply_sse41_lowp)
_sk_multiply_sse41_lowp:
- .byte 102,68,15,111,13,158,142,0,0 // movdqa 0x8e9e(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,162,142,0,0 // movdqa 0x8ea2(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
.byte 102,68,15,249,195 // psubw %xmm3,%xmm8
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -57874,7 +57258,7 @@ HIDDEN _sk_plus__sse41_lowp
FUNCTION(_sk_plus__sse41_lowp)
_sk_plus__sse41_lowp:
.byte 102,15,253,196 // paddw %xmm4,%xmm0
- .byte 102,68,15,111,5,227,141,0,0 // movdqa 0x8de3(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,231,141,0,0 // movdqa 0x8de7(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,56,58,192 // pminuw %xmm8,%xmm0
.byte 102,15,253,205 // paddw %xmm5,%xmm1
.byte 102,65,15,56,58,200 // pminuw %xmm8,%xmm1
@@ -57894,7 +57278,7 @@ _sk_screen_sse41_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,253,193 // paddw %xmm1,%xmm0
.byte 102,15,213,204 // pmullw %xmm4,%xmm1
- .byte 102,68,15,111,21,157,141,0,0 // movdqa 0x8d9d(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,21,161,141,0,0 // movdqa 0x8da1(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,202 // paddw %xmm10,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -57926,7 +57310,7 @@ HIDDEN _sk_xor__sse41_lowp
FUNCTION(_sk_xor__sse41_lowp)
_sk_xor__sse41_lowp:
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,68,15,111,13,27,141,0,0 // movdqa 0x8d1b(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,31,141,0,0 // movdqa 0x8d1f(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,209 // movdqa %xmm9,%xmm10
.byte 102,68,15,249,215 // psubw %xmm7,%xmm10
.byte 102,65,15,213,194 // pmullw %xmm10,%xmm0
@@ -57969,7 +57353,7 @@ _sk_darken_sse41_lowp:
.byte 102,68,15,213,207 // pmullw %xmm7,%xmm9
.byte 102,15,213,203 // pmullw %xmm3,%xmm1
.byte 102,65,15,56,62,201 // pmaxuw %xmm9,%xmm1
- .byte 102,68,15,111,13,98,140,0,0 // movdqa 0x8c62(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,102,140,0,0 // movdqa 0x8c66(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,201 // paddw %xmm9,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -58013,7 +57397,7 @@ _sk_lighten_sse41_lowp:
.byte 102,68,15,213,207 // pmullw %xmm7,%xmm9
.byte 102,15,213,203 // pmullw %xmm3,%xmm1
.byte 102,65,15,56,58,201 // pminuw %xmm9,%xmm1
- .byte 102,68,15,111,13,165,139,0,0 // movdqa 0x8ba5(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,169,139,0,0 // movdqa 0x8ba9(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,201 // paddw %xmm9,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -58057,10 +57441,10 @@ _sk_difference_sse41_lowp:
.byte 102,68,15,213,207 // pmullw %xmm7,%xmm9
.byte 102,15,213,203 // pmullw %xmm3,%xmm1
.byte 102,65,15,56,58,201 // pminuw %xmm9,%xmm1
- .byte 102,68,15,111,13,232,138,0,0 // movdqa 0x8ae8(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,236,138,0,0 // movdqa 0x8aec(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,201 // paddw %xmm9,%xmm1
.byte 102,15,113,209,7 // psrlw $0x7,%xmm1
- .byte 102,68,15,111,21,53,146,0,0 // movdqa 0x9235(%rip),%xmm10 # 3a770 <_sk_srcover_bgra_8888_sse2_lowp+0x12e8>
+ .byte 102,68,15,111,21,57,146,0,0 // movdqa 0x9239(%rip),%xmm10 # 399f0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ec>
.byte 102,65,15,219,202 // pand %xmm10,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
.byte 102,68,15,111,221 // movdqa %xmm5,%xmm11
@@ -58102,10 +57486,10 @@ _sk_exclusion_sse41_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,253,193 // paddw %xmm1,%xmm0
.byte 102,15,213,204 // pmullw %xmm4,%xmm1
- .byte 102,68,15,111,13,36,138,0,0 // movdqa 0x8a24(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,40,138,0,0 // movdqa 0x8a28(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,201 // paddw %xmm9,%xmm1
.byte 102,15,113,209,7 // psrlw $0x7,%xmm1
- .byte 102,68,15,111,21,113,145,0,0 // movdqa 0x9171(%rip),%xmm10 # 3a770 <_sk_srcover_bgra_8888_sse2_lowp+0x12e8>
+ .byte 102,68,15,111,21,117,145,0,0 // movdqa 0x9175(%rip),%xmm10 # 399f0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ec>
.byte 102,65,15,219,202 // pand %xmm10,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
.byte 102,15,111,205 // movdqa %xmm5,%xmm1
@@ -58141,14 +57525,14 @@ _sk_hardlight_sse41_lowp:
.byte 102,15,111,245 // movdqa %xmm5,%xmm6
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 102,68,15,111,29,127,137,0,0 // movdqa 0x897f(%rip),%xmm11 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,29,131,137,0,0 // movdqa 0x8983(%rip),%xmm11 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,249,220 // psubw %xmm12,%xmm11
.byte 102,65,15,111,227 // movdqa %xmm11,%xmm4
.byte 102,65,15,213,224 // pmullw %xmm8,%xmm4
.byte 102,68,15,111,251 // movdqa %xmm3,%xmm15
.byte 102,69,15,249,248 // psubw %xmm8,%xmm15
.byte 102,69,15,253,192 // paddw %xmm8,%xmm8
- .byte 102,68,15,111,13,200,144,0,0 // movdqa 0x90c8(%rip),%xmm9 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,13,204,144,0,0 // movdqa 0x90cc(%rip),%xmm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,243 // movdqa %xmm3,%xmm14
.byte 102,69,15,239,241 // pxor %xmm9,%xmm14
.byte 102,65,15,111,196 // movdqa %xmm12,%xmm0
@@ -58195,7 +57579,7 @@ _sk_hardlight_sse41_lowp:
.byte 102,68,15,111,232 // movdqa %xmm0,%xmm13
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,65,15,56,16,210 // pblendvb %xmm0,%xmm10,%xmm2
- .byte 102,68,15,111,13,112,136,0,0 // movdqa 0x8870(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,116,136,0,0 // movdqa 0x8874(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,249,195 // psubw %xmm3,%xmm0
.byte 102,15,111,248 // movdqa %xmm0,%xmm7
@@ -58234,7 +57618,7 @@ FUNCTION(_sk_overlay_sse41_lowp)
_sk_overlay_sse41_lowp:
.byte 102,68,15,111,231 // movdqa %xmm7,%xmm12
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 102,68,15,111,29,211,135,0,0 // movdqa 0x87d3(%rip),%xmm11 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,29,215,135,0,0 // movdqa 0x87d7(%rip),%xmm11 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,249,220 // psubw %xmm12,%xmm11
.byte 102,65,15,111,195 // movdqa %xmm11,%xmm0
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
@@ -58246,7 +57630,7 @@ _sk_overlay_sse41_lowp:
.byte 102,68,15,249,252 // psubw %xmm4,%xmm15
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,253,192 // paddw %xmm0,%xmm0
- .byte 102,68,15,111,13,4,143,0,0 // movdqa 0x8f04(%rip),%xmm9 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,13,8,143,0,0 // movdqa 0x8f08(%rip),%xmm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,69,15,111,212 // movdqa %xmm12,%xmm10
.byte 102,68,15,213,211 // pmullw %xmm3,%xmm10
.byte 102,69,15,213,254 // pmullw %xmm14,%xmm15
@@ -58292,7 +57676,7 @@ _sk_overlay_sse41_lowp:
.byte 102,15,253,210 // paddw %xmm2,%xmm2
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,65,15,56,16,210 // pblendvb %xmm0,%xmm10,%xmm2
- .byte 102,68,15,111,13,182,134,0,0 // movdqa 0x86b6(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,186,134,0,0 // movdqa 0x86ba(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,249,195 // psubw %xmm3,%xmm0
.byte 102,15,111,248 // movdqa %xmm0,%xmm7
@@ -58335,49 +57719,49 @@ _sk_load_8888_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 31a1c <_sk_load_8888_sse41_lowp+0x44>
+ .byte 119,41 // ja 30c98 <_sk_load_8888_sse41_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 31af4 <_sk_load_8888_sse41_lowp+0x11c>
+ .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 30d70 <_sk_load_8888_sse41_lowp+0x11c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,20,144 // movd (%r8,%rdx,4),%xmm2
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 31a75 <_sk_load_8888_sse41_lowp+0x9d>
+ .byte 235,89 // jmp 30cf1 <_sk_load_8888_sse41_lowp+0x9d>
.byte 243,65,15,111,20,144 // movdqu (%r8,%rdx,4),%xmm2
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 31a75 <_sk_load_8888_sse41_lowp+0x9d>
+ .byte 235,74 // jmp 30cf1 <_sk_load_8888_sse41_lowp+0x9d>
.byte 102,65,15,110,68,144,8 // movd 0x8(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 243,65,15,126,20,144 // movq (%r8,%rdx,4),%xmm2
.byte 102,65,15,58,14,208,240 // pblendw $0xf0,%xmm8,%xmm2
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 235,37 // jmp 31a75 <_sk_load_8888_sse41_lowp+0x9d>
+ .byte 235,37 // jmp 30cf1 <_sk_load_8888_sse41_lowp+0x9d>
.byte 102,65,15,110,68,144,24 // movd 0x18(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,20,144 // movdqu (%r8,%rdx,4),%xmm2
- .byte 102,15,111,5,19,141,0,0 // movdqa 0x8d13(%rip),%xmm0 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,5,23,141,0,0 // movdqa 0x8d17(%rip),%xmm0 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,202 // movdqa %xmm2,%xmm1
.byte 102,15,56,0,200 // pshufb %xmm0,%xmm1
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,56,0,216 // pshufb %xmm0,%xmm3
.byte 102,15,108,203 // punpcklqdq %xmm3,%xmm1
- .byte 102,68,15,111,13,115,133,0,0 // movdqa 0x8573(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,119,133,0,0 // movdqa 0x8577(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,193 // pand %xmm9,%xmm0
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
- .byte 102,68,15,111,21,236,140,0,0 // movdqa 0x8cec(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,240,140,0,0 // movdqa 0x8cf0(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,218 // movdqa %xmm2,%xmm3
.byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,219 // punpcklqdq %xmm11,%xmm3
- .byte 102,68,15,111,21,217,140,0,0 // movdqa 0x8cd9(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,221,140,0,0 // movdqa 0x8cdd(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,210 // pshufb %xmm10,%xmm2
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,208 // punpcklqdq %xmm8,%xmm2
@@ -58393,7 +57777,7 @@ _sk_load_8888_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 31b01 <_sk_load_8888_sse41_lowp+0x129>
+ .byte 123,255 // jnp 30d7d <_sk_load_8888_sse41_lowp+0x129>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -58415,49 +57799,49 @@ _sk_load_8888_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 31b54 <_sk_load_8888_dst_sse41_lowp+0x44>
+ .byte 119,41 // ja 30dd0 <_sk_load_8888_dst_sse41_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 31c2c <_sk_load_8888_dst_sse41_lowp+0x11c>
+ .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 30ea8 <_sk_load_8888_dst_sse41_lowp+0x11c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,52,144 // movd (%r8,%rdx,4),%xmm6
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 31bad <_sk_load_8888_dst_sse41_lowp+0x9d>
+ .byte 235,89 // jmp 30e29 <_sk_load_8888_dst_sse41_lowp+0x9d>
.byte 243,65,15,111,52,144 // movdqu (%r8,%rdx,4),%xmm6
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 31bad <_sk_load_8888_dst_sse41_lowp+0x9d>
+ .byte 235,74 // jmp 30e29 <_sk_load_8888_dst_sse41_lowp+0x9d>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 243,65,15,126,52,144 // movq (%r8,%rdx,4),%xmm6
.byte 102,65,15,58,14,240,240 // pblendw $0xf0,%xmm8,%xmm6
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,37 // jmp 31bad <_sk_load_8888_dst_sse41_lowp+0x9d>
+ .byte 235,37 // jmp 30e29 <_sk_load_8888_dst_sse41_lowp+0x9d>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,52,144 // movdqu (%r8,%rdx,4),%xmm6
- .byte 102,15,111,37,219,139,0,0 // movdqa 0x8bdb(%rip),%xmm4 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,37,223,139,0,0 // movdqa 0x8bdf(%rip),%xmm4 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,238 // movdqa %xmm6,%xmm5
.byte 102,15,56,0,236 // pshufb %xmm4,%xmm5
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,56,0,252 // pshufb %xmm4,%xmm7
.byte 102,15,108,239 // punpcklqdq %xmm7,%xmm5
- .byte 102,68,15,111,13,59,132,0,0 // movdqa 0x843b(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,63,132,0,0 // movdqa 0x843f(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,229 // movdqa %xmm5,%xmm4
.byte 102,65,15,219,225 // pand %xmm9,%xmm4
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
- .byte 102,68,15,111,21,180,139,0,0 // movdqa 0x8bb4(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,184,139,0,0 // movdqa 0x8bb8(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,254 // movdqa %xmm6,%xmm7
.byte 102,65,15,56,0,250 // pshufb %xmm10,%xmm7
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,251 // punpcklqdq %xmm11,%xmm7
- .byte 102,68,15,111,21,161,139,0,0 // movdqa 0x8ba1(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,165,139,0,0 // movdqa 0x8ba5(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,242 // pshufb %xmm10,%xmm6
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,240 // punpcklqdq %xmm8,%xmm6
@@ -58473,7 +57857,7 @@ _sk_load_8888_dst_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 31c39 <_sk_load_8888_dst_sse41_lowp+0x129>
+ .byte 123,255 // jnp 30eb5 <_sk_load_8888_dst_sse41_lowp+0x129>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -58510,20 +57894,20 @@ _sk_store_8888_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,27 // ja 31cce <_sk_store_8888_sse41_lowp+0x86>
+ .byte 119,27 // ja 30f4a <_sk_store_8888_sse41_lowp+0x86>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,87,0,0,0 // lea 0x57(%rip),%r9 # 31d14 <_sk_store_8888_sse41_lowp+0xcc>
+ .byte 76,141,13,87,0,0,0 // lea 0x57(%rip),%r9 # 30f90 <_sk_store_8888_sse41_lowp+0xcc>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,4,144 // movd %xmm8,(%r8,%rdx,4)
- .byte 235,63 // jmp 31d0d <_sk_store_8888_sse41_lowp+0xc5>
+ .byte 235,63 // jmp 30f89 <_sk_store_8888_sse41_lowp+0xc5>
.byte 243,69,15,127,4,144 // movdqu %xmm8,(%r8,%rdx,4)
.byte 243,69,15,127,76,144,16 // movdqu %xmm9,0x10(%r8,%rdx,4)
- .byte 235,48 // jmp 31d0d <_sk_store_8888_sse41_lowp+0xc5>
+ .byte 235,48 // jmp 30f89 <_sk_store_8888_sse41_lowp+0xc5>
.byte 102,69,15,58,22,68,144,8,2 // pextrd $0x2,%xmm8,0x8(%r8,%rdx,4)
.byte 102,69,15,214,4,144 // movq %xmm8,(%r8,%rdx,4)
- .byte 235,31 // jmp 31d0d <_sk_store_8888_sse41_lowp+0xc5>
+ .byte 235,31 // jmp 30f89 <_sk_store_8888_sse41_lowp+0xc5>
.byte 102,69,15,58,22,76,144,24,2 // pextrd $0x2,%xmm9,0x18(%r8,%rdx,4)
.byte 102,69,15,58,22,76,144,20,1 // pextrd $0x1,%xmm9,0x14(%r8,%rdx,4)
.byte 102,69,15,126,76,144,16 // movd %xmm9,0x10(%r8,%rdx,4)
@@ -58567,49 +57951,49 @@ _sk_load_bgra_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 31d74 <_sk_load_bgra_sse41_lowp+0x44>
+ .byte 119,41 // ja 30ff0 <_sk_load_bgra_sse41_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 31e4c <_sk_load_bgra_sse41_lowp+0x11c>
+ .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 310c8 <_sk_load_bgra_sse41_lowp+0x11c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,4,144 // movd (%r8,%rdx,4),%xmm0
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 31dcd <_sk_load_bgra_sse41_lowp+0x9d>
+ .byte 235,89 // jmp 31049 <_sk_load_bgra_sse41_lowp+0x9d>
.byte 243,65,15,111,4,144 // movdqu (%r8,%rdx,4),%xmm0
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 31dcd <_sk_load_bgra_sse41_lowp+0x9d>
+ .byte 235,74 // jmp 31049 <_sk_load_bgra_sse41_lowp+0x9d>
.byte 102,65,15,110,68,144,8 // movd 0x8(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 243,65,15,126,4,144 // movq (%r8,%rdx,4),%xmm0
.byte 102,65,15,58,14,192,240 // pblendw $0xf0,%xmm8,%xmm0
.byte 102,68,15,111,193 // movdqa %xmm1,%xmm8
- .byte 235,37 // jmp 31dcd <_sk_load_bgra_sse41_lowp+0x9d>
+ .byte 235,37 // jmp 31049 <_sk_load_bgra_sse41_lowp+0x9d>
.byte 102,65,15,110,68,144,24 // movd 0x18(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,4,144 // movdqu (%r8,%rdx,4),%xmm0
- .byte 102,15,111,21,187,137,0,0 // movdqa 0x89bb(%rip),%xmm2 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,21,191,137,0,0 // movdqa 0x89bf(%rip),%xmm2 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,56,0,202 // pshufb %xmm2,%xmm1
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,56,0,218 // pshufb %xmm2,%xmm3
.byte 102,15,108,203 // punpcklqdq %xmm3,%xmm1
- .byte 102,68,15,111,13,27,130,0,0 // movdqa 0x821b(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,31,130,0,0 // movdqa 0x821f(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,209 // movdqa %xmm1,%xmm2
.byte 102,65,15,219,209 // pand %xmm9,%xmm2
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
- .byte 102,68,15,111,21,148,137,0,0 // movdqa 0x8994(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,152,137,0,0 // movdqa 0x8998(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,219 // punpcklqdq %xmm11,%xmm3
- .byte 102,68,15,111,21,129,137,0,0 // movdqa 0x8981(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,133,137,0,0 // movdqa 0x8985(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,194 // pshufb %xmm10,%xmm0
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,192 // punpcklqdq %xmm8,%xmm0
@@ -58625,7 +58009,7 @@ _sk_load_bgra_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 31e59 <_sk_load_bgra_sse41_lowp+0x129>
+ .byte 123,255 // jnp 310d5 <_sk_load_bgra_sse41_lowp+0x129>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -58647,49 +58031,49 @@ _sk_load_bgra_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 31eac <_sk_load_bgra_dst_sse41_lowp+0x44>
+ .byte 119,41 // ja 31128 <_sk_load_bgra_dst_sse41_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 31f84 <_sk_load_bgra_dst_sse41_lowp+0x11c>
+ .byte 76,141,13,242,0,0,0 // lea 0xf2(%rip),%r9 # 31200 <_sk_load_bgra_dst_sse41_lowp+0x11c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,36,144 // movd (%r8,%rdx,4),%xmm4
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 31f05 <_sk_load_bgra_dst_sse41_lowp+0x9d>
+ .byte 235,89 // jmp 31181 <_sk_load_bgra_dst_sse41_lowp+0x9d>
.byte 243,65,15,111,36,144 // movdqu (%r8,%rdx,4),%xmm4
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 31f05 <_sk_load_bgra_dst_sse41_lowp+0x9d>
+ .byte 235,74 // jmp 31181 <_sk_load_bgra_dst_sse41_lowp+0x9d>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 243,65,15,126,36,144 // movq (%r8,%rdx,4),%xmm4
.byte 102,65,15,58,14,224,240 // pblendw $0xf0,%xmm8,%xmm4
.byte 102,68,15,111,197 // movdqa %xmm5,%xmm8
- .byte 235,37 // jmp 31f05 <_sk_load_bgra_dst_sse41_lowp+0x9d>
+ .byte 235,37 // jmp 31181 <_sk_load_bgra_dst_sse41_lowp+0x9d>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,36,144 // movdqu (%r8,%rdx,4),%xmm4
- .byte 102,15,111,53,131,136,0,0 // movdqa 0x8883(%rip),%xmm6 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,53,135,136,0,0 // movdqa 0x8887(%rip),%xmm6 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,56,0,238 // pshufb %xmm6,%xmm5
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,56,0,254 // pshufb %xmm6,%xmm7
.byte 102,15,108,239 // punpcklqdq %xmm7,%xmm5
- .byte 102,68,15,111,13,227,128,0,0 // movdqa 0x80e3(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,231,128,0,0 // movdqa 0x80e7(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,245 // movdqa %xmm5,%xmm6
.byte 102,65,15,219,241 // pand %xmm9,%xmm6
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
- .byte 102,68,15,111,21,92,136,0,0 // movdqa 0x885c(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,96,136,0,0 // movdqa 0x8860(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,252 // movdqa %xmm4,%xmm7
.byte 102,65,15,56,0,250 // pshufb %xmm10,%xmm7
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,251 // punpcklqdq %xmm11,%xmm7
- .byte 102,68,15,111,21,73,136,0,0 // movdqa 0x8849(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,77,136,0,0 // movdqa 0x884d(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,226 // pshufb %xmm10,%xmm4
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,224 // punpcklqdq %xmm8,%xmm4
@@ -58705,7 +58089,7 @@ _sk_load_bgra_dst_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 31f91 <_sk_load_bgra_dst_sse41_lowp+0x129>
+ .byte 123,255 // jnp 3120d <_sk_load_bgra_dst_sse41_lowp+0x129>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -58742,20 +58126,20 @@ _sk_store_bgra_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,27 // ja 32026 <_sk_store_bgra_sse41_lowp+0x86>
+ .byte 119,27 // ja 312a2 <_sk_store_bgra_sse41_lowp+0x86>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,87,0,0,0 // lea 0x57(%rip),%r9 # 3206c <_sk_store_bgra_sse41_lowp+0xcc>
+ .byte 76,141,13,87,0,0,0 // lea 0x57(%rip),%r9 # 312e8 <_sk_store_bgra_sse41_lowp+0xcc>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,4,144 // movd %xmm8,(%r8,%rdx,4)
- .byte 235,63 // jmp 32065 <_sk_store_bgra_sse41_lowp+0xc5>
+ .byte 235,63 // jmp 312e1 <_sk_store_bgra_sse41_lowp+0xc5>
.byte 243,69,15,127,4,144 // movdqu %xmm8,(%r8,%rdx,4)
.byte 243,69,15,127,76,144,16 // movdqu %xmm9,0x10(%r8,%rdx,4)
- .byte 235,48 // jmp 32065 <_sk_store_bgra_sse41_lowp+0xc5>
+ .byte 235,48 // jmp 312e1 <_sk_store_bgra_sse41_lowp+0xc5>
.byte 102,69,15,58,22,68,144,8,2 // pextrd $0x2,%xmm8,0x8(%r8,%rdx,4)
.byte 102,69,15,214,4,144 // movq %xmm8,(%r8,%rdx,4)
- .byte 235,31 // jmp 32065 <_sk_store_bgra_sse41_lowp+0xc5>
+ .byte 235,31 // jmp 312e1 <_sk_store_bgra_sse41_lowp+0xc5>
.byte 102,69,15,58,22,76,144,24,2 // pextrd $0x2,%xmm9,0x18(%r8,%rdx,4)
.byte 102,69,15,58,22,76,144,20,1 // pextrd $0x1,%xmm9,0x14(%r8,%rdx,4)
.byte 102,69,15,126,76,144,16 // movd %xmm9,0x10(%r8,%rdx,4)
@@ -58858,17 +58242,17 @@ _sk_gather_8888_sse41_lowp:
.byte 102,65,15,196,202,5 // pinsrw $0x5,%r10d,%xmm1
.byte 102,65,15,196,201,6 // pinsrw $0x6,%r9d,%xmm1
.byte 102,15,196,200,7 // pinsrw $0x7,%eax,%xmm1
- .byte 102,68,15,111,13,76,126,0,0 // movdqa 0x7e4c(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,80,126,0,0 // movdqa 0x7e50(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,193 // pand %xmm9,%xmm0
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
- .byte 102,68,15,111,21,197,133,0,0 // movdqa 0x85c5(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,201,133,0,0 // movdqa 0x85c9(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,218 // movdqa %xmm2,%xmm3
.byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,219 // punpcklqdq %xmm11,%xmm3
- .byte 102,68,15,111,21,178,133,0,0 // movdqa 0x85b2(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,182,133,0,0 // movdqa 0x85b6(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,210 // pshufb %xmm10,%xmm2
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,208 // punpcklqdq %xmm8,%xmm2
@@ -58953,17 +58337,17 @@ _sk_gather_bgra_sse41_lowp:
.byte 102,65,15,196,202,5 // pinsrw $0x5,%r10d,%xmm1
.byte 102,65,15,196,201,6 // pinsrw $0x6,%r9d,%xmm1
.byte 102,15,196,200,7 // pinsrw $0x7,%eax,%xmm1
- .byte 102,68,15,111,13,180,124,0,0 // movdqa 0x7cb4(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,184,124,0,0 // movdqa 0x7cb8(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,209 // movdqa %xmm1,%xmm2
.byte 102,65,15,219,209 // pand %xmm9,%xmm2
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
- .byte 102,68,15,111,21,45,132,0,0 // movdqa 0x842d(%rip),%xmm10 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,21,49,132,0,0 // movdqa 0x8431(%rip),%xmm10 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
.byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
.byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
.byte 102,65,15,108,219 // punpcklqdq %xmm11,%xmm3
- .byte 102,68,15,111,21,26,132,0,0 // movdqa 0x841a(%rip),%xmm10 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,21,30,132,0,0 // movdqa 0x841e(%rip),%xmm10 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,194 // pshufb %xmm10,%xmm0
.byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
.byte 102,65,15,108,192 // punpcklqdq %xmm8,%xmm0
@@ -58989,23 +58373,23 @@ _sk_load_565_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 323f4 <_sk_load_565_sse41_lowp+0x3c>
+ .byte 119,34 // ja 31670 <_sk_load_565_sse41_lowp+0x3c>
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,196,0,0,0 // lea 0xc4(%rip),%r9 # 324a4 <_sk_load_565_sse41_lowp+0xec>
+ .byte 76,141,13,196,0,0,0 // lea 0xc4(%rip),%r9 # 31720 <_sk_load_565_sse41_lowp+0xec>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,192 // movd %eax,%xmm0
- .byte 235,74 // jmp 3243e <_sk_load_565_sse41_lowp+0x86>
+ .byte 235,74 // jmp 316ba <_sk_load_565_sse41_lowp+0x86>
.byte 243,65,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm0
- .byte 235,66 // jmp 3243e <_sk_load_565_sse41_lowp+0x86>
+ .byte 235,66 // jmp 316ba <_sk_load_565_sse41_lowp+0x86>
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 102,65,15,196,76,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm1
.byte 102,65,15,110,4,80 // movd (%r8,%rdx,2),%xmm0
.byte 102,15,58,14,193,252 // pblendw $0xfc,%xmm1,%xmm0
- .byte 235,40 // jmp 3243e <_sk_load_565_sse41_lowp+0x86>
+ .byte 235,40 // jmp 316ba <_sk_load_565_sse41_lowp+0x86>
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 102,65,15,196,76,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm1
.byte 102,65,15,196,76,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm1
@@ -59014,11 +58398,11 @@ _sk_load_565_sse41_lowp:
.byte 102,15,58,14,193,240 // pblendw $0xf0,%xmm1,%xmm0
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
- .byte 102,15,219,29,113,131,0,0 // pand 0x8371(%rip),%xmm3 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,29,117,131,0,0 // pand 0x8375(%rip),%xmm3 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,113,209,5 // psrlw $0x5,%xmm1
- .byte 102,15,219,13,112,131,0,0 // pand 0x8370(%rip),%xmm1 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,21,120,131,0,0 // movdqa 0x8378(%rip),%xmm2 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,13,116,131,0,0 // pand 0x8374(%rip),%xmm1 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,21,124,131,0,0 // movdqa 0x837c(%rip),%xmm2 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,15,113,208,13 // psrlw $0xd,%xmm0
.byte 102,15,235,195 // por %xmm3,%xmm0
@@ -59031,7 +58415,7 @@ _sk_load_565_sse41_lowp:
.byte 102,15,113,210,2 // psrlw $0x2,%xmm2
.byte 102,15,235,211 // por %xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,110,123,0,0 // movaps 0x7b6e(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,114,123,0,0 // movaps 0x7b72(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 255,224 // jmpq *%rax
.byte 69,255 // rex.RB (bad)
.byte 255 // (bad)
@@ -59042,7 +58426,7 @@ _sk_load_565_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 126,255 // jle 324b9 <_sk_load_565_sse41_lowp+0x101>
+ .byte 126,255 // jle 31735 <_sk_load_565_sse41_lowp+0x101>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -59061,23 +58445,23 @@ _sk_load_565_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 324fc <_sk_load_565_dst_sse41_lowp+0x3c>
+ .byte 119,34 // ja 31778 <_sk_load_565_dst_sse41_lowp+0x3c>
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,196,0,0,0 // lea 0xc4(%rip),%r9 # 325ac <_sk_load_565_dst_sse41_lowp+0xec>
+ .byte 76,141,13,196,0,0,0 // lea 0xc4(%rip),%r9 # 31828 <_sk_load_565_dst_sse41_lowp+0xec>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,224 // movd %eax,%xmm4
- .byte 235,74 // jmp 32546 <_sk_load_565_dst_sse41_lowp+0x86>
+ .byte 235,74 // jmp 317c2 <_sk_load_565_dst_sse41_lowp+0x86>
.byte 243,65,15,111,36,80 // movdqu (%r8,%rdx,2),%xmm4
- .byte 235,66 // jmp 32546 <_sk_load_565_dst_sse41_lowp+0x86>
+ .byte 235,66 // jmp 317c2 <_sk_load_565_dst_sse41_lowp+0x86>
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 102,65,15,196,108,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm5
.byte 102,65,15,110,36,80 // movd (%r8,%rdx,2),%xmm4
.byte 102,15,58,14,229,252 // pblendw $0xfc,%xmm5,%xmm4
- .byte 235,40 // jmp 32546 <_sk_load_565_dst_sse41_lowp+0x86>
+ .byte 235,40 // jmp 317c2 <_sk_load_565_dst_sse41_lowp+0x86>
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 102,65,15,196,108,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm5
.byte 102,65,15,196,108,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm5
@@ -59086,11 +58470,11 @@ _sk_load_565_dst_sse41_lowp:
.byte 102,15,58,14,229,240 // pblendw $0xf0,%xmm5,%xmm4
.byte 102,15,111,252 // movdqa %xmm4,%xmm7
.byte 102,15,113,215,8 // psrlw $0x8,%xmm7
- .byte 102,15,219,61,105,130,0,0 // pand 0x8269(%rip),%xmm7 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,61,109,130,0,0 // pand 0x826d(%rip),%xmm7 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,113,213,5 // psrlw $0x5,%xmm5
- .byte 102,15,219,45,104,130,0,0 // pand 0x8268(%rip),%xmm5 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,53,112,130,0,0 // movdqa 0x8270(%rip),%xmm6 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,45,108,130,0,0 // pand 0x826c(%rip),%xmm5 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,53,116,130,0,0 // movdqa 0x8274(%rip),%xmm6 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,244 // pand %xmm4,%xmm6
.byte 102,15,113,212,13 // psrlw $0xd,%xmm4
.byte 102,15,235,231 // por %xmm7,%xmm4
@@ -59103,7 +58487,7 @@ _sk_load_565_dst_sse41_lowp:
.byte 102,15,113,214,2 // psrlw $0x2,%xmm6
.byte 102,15,235,247 // por %xmm7,%xmm6
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,102,122,0,0 // movaps 0x7a66(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,61,106,122,0,0 // movaps 0x7a6a(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 255,224 // jmpq *%rax
.byte 69,255 // rex.RB (bad)
.byte 255 // (bad)
@@ -59114,7 +58498,7 @@ _sk_load_565_dst_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 126,255 // jle 325c1 <_sk_load_565_dst_sse41_lowp+0x101>
+ .byte 126,255 // jle 3183d <_sk_load_565_dst_sse41_lowp+0x101>
.byte 255 // (bad)
.byte 255,114,255 // pushq -0x1(%rdx)
.byte 255 // (bad)
@@ -59131,10 +58515,10 @@ _sk_store_565_sse41_lowp:
.byte 76,3,0 // add (%rax),%r8
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
.byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,219,5,4,130,0,0 // pand 0x8204(%rip),%xmm8 # 3a7f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1368>
+ .byte 102,68,15,219,5,8,130,0,0 // pand 0x8208(%rip),%xmm8 # 39a70 <_sk_srcover_bgra_8888_sse2_lowp+0x136c>
.byte 102,68,15,111,201 // movdqa %xmm1,%xmm9
.byte 102,65,15,113,241,3 // psllw $0x3,%xmm9
- .byte 102,68,15,219,13,0,130,0,0 // pand 0x8200(%rip),%xmm9 # 3a800 <_sk_srcover_bgra_8888_sse2_lowp+0x1378>
+ .byte 102,68,15,219,13,4,130,0,0 // pand 0x8204(%rip),%xmm9 # 39a80 <_sk_srcover_bgra_8888_sse2_lowp+0x137c>
.byte 102,69,15,235,200 // por %xmm8,%xmm9
.byte 102,68,15,111,194 // movdqa %xmm2,%xmm8
.byte 102,65,15,113,208,3 // psrlw $0x3,%xmm8
@@ -59143,19 +58527,19 @@ _sk_store_565_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,29 // ja 3263c <_sk_store_565_sse41_lowp+0x74>
+ .byte 119,29 // ja 318b8 <_sk_store_565_sse41_lowp+0x74>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,83,0,0,0 // lea 0x53(%rip),%r9 # 3267c <_sk_store_565_sse41_lowp+0xb4>
+ .byte 76,141,13,83,0,0,0 // lea 0x53(%rip),%r9 # 318f8 <_sk_store_565_sse41_lowp+0xb4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,58,21,4,80,0 // pextrw $0x0,%xmm8,(%r8,%rdx,2)
- .byte 235,58 // jmp 32676 <_sk_store_565_sse41_lowp+0xae>
+ .byte 235,58 // jmp 318f2 <_sk_store_565_sse41_lowp+0xae>
.byte 243,69,15,127,4,80 // movdqu %xmm8,(%r8,%rdx,2)
- .byte 235,50 // jmp 32676 <_sk_store_565_sse41_lowp+0xae>
+ .byte 235,50 // jmp 318f2 <_sk_store_565_sse41_lowp+0xae>
.byte 102,69,15,58,21,68,80,4,2 // pextrw $0x2,%xmm8,0x4(%r8,%rdx,2)
.byte 102,69,15,126,4,80 // movd %xmm8,(%r8,%rdx,2)
- .byte 235,33 // jmp 32676 <_sk_store_565_sse41_lowp+0xae>
+ .byte 235,33 // jmp 318f2 <_sk_store_565_sse41_lowp+0xae>
.byte 102,69,15,58,21,68,80,12,6 // pextrw $0x6,%xmm8,0xc(%r8,%rdx,2)
.byte 102,69,15,58,21,68,80,10,5 // pextrw $0x5,%xmm8,0xa(%r8,%rdx,2)
.byte 102,69,15,58,21,68,80,8,4 // pextrw $0x4,%xmm8,0x8(%r8,%rdx,2)
@@ -59175,7 +58559,7 @@ _sk_store_565_sse41_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 235,255 // jmp 3268d <_sk_store_565_sse41_lowp+0xc5>
+ .byte 235,255 // jmp 31909 <_sk_store_565_sse41_lowp+0xc5>
.byte 255 // (bad)
.byte 255,226 // jmpq *%rdx
.byte 255 // (bad)
@@ -59252,11 +58636,11 @@ _sk_gather_565_sse41_lowp:
.byte 102,15,196,192,7 // pinsrw $0x7,%eax,%xmm0
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
- .byte 102,15,219,29,14,128,0,0 // pand 0x800e(%rip),%xmm3 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,29,18,128,0,0 // pand 0x8012(%rip),%xmm3 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,113,209,5 // psrlw $0x5,%xmm1
- .byte 102,15,219,13,13,128,0,0 // pand 0x800d(%rip),%xmm1 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,21,21,128,0,0 // movdqa 0x8015(%rip),%xmm2 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,13,17,128,0,0 // pand 0x8011(%rip),%xmm1 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,21,25,128,0,0 // movdqa 0x8019(%rip),%xmm2 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,15,113,208,13 // psrlw $0xd,%xmm0
.byte 102,15,235,195 // por %xmm3,%xmm0
@@ -59269,7 +58653,7 @@ _sk_gather_565_sse41_lowp:
.byte 102,15,113,210,2 // psrlw $0x2,%xmm2
.byte 102,15,235,211 // por %xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,11,120,0,0 // movaps 0x780b(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,15,120,0,0 // movaps 0x780f(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
.byte 65,94 // pop %r14
@@ -59290,23 +58674,23 @@ _sk_load_4444_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,35 // ja 3284c <_sk_load_4444_sse41_lowp+0x3d>
+ .byte 119,35 // ja 31ac8 <_sk_load_4444_sse41_lowp+0x3d>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,209,0,0,0 // lea 0xd1(%rip),%r9 # 32908 <_sk_load_4444_sse41_lowp+0xf9>
+ .byte 76,141,13,209,0,0,0 // lea 0xd1(%rip),%r9 # 31b84 <_sk_load_4444_sse41_lowp+0xf9>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,76 // jmp 32898 <_sk_load_4444_sse41_lowp+0x89>
+ .byte 235,76 // jmp 31b14 <_sk_load_4444_sse41_lowp+0x89>
.byte 243,69,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm8
- .byte 235,68 // jmp 32898 <_sk_load_4444_sse41_lowp+0x89>
+ .byte 235,68 // jmp 31b14 <_sk_load_4444_sse41_lowp+0x89>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm0
.byte 102,69,15,110,4,80 // movd (%r8,%rdx,2),%xmm8
.byte 102,68,15,58,14,192,252 // pblendw $0xfc,%xmm0,%xmm8
- .byte 235,41 // jmp 32898 <_sk_load_4444_sse41_lowp+0x89>
+ .byte 235,41 // jmp 31b14 <_sk_load_4444_sse41_lowp+0x89>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm0
.byte 102,65,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm0
@@ -59317,7 +58701,7 @@ _sk_load_4444_sse41_lowp:
.byte 102,15,113,209,12 // psrlw $0xc,%xmm1
.byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
.byte 102,15,113,210,8 // psrlw $0x8,%xmm2
- .byte 102,15,111,5,92,127,0,0 // movdqa 0x7f5c(%rip),%xmm0 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,5,96,127,0,0 // movdqa 0x7f60(%rip),%xmm0 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,113,211,4 // psrlw $0x4,%xmm3
@@ -59365,23 +58749,23 @@ _sk_load_4444_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,35 // ja 32961 <_sk_load_4444_dst_sse41_lowp+0x3d>
+ .byte 119,35 // ja 31bdd <_sk_load_4444_dst_sse41_lowp+0x3d>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,208,0,0,0 // lea 0xd0(%rip),%r9 # 32a1c <_sk_load_4444_dst_sse41_lowp+0xf8>
+ .byte 76,141,13,208,0,0,0 // lea 0xd0(%rip),%r9 # 31c98 <_sk_load_4444_dst_sse41_lowp+0xf8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,76 // jmp 329ad <_sk_load_4444_dst_sse41_lowp+0x89>
+ .byte 235,76 // jmp 31c29 <_sk_load_4444_dst_sse41_lowp+0x89>
.byte 243,69,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm8
- .byte 235,68 // jmp 329ad <_sk_load_4444_dst_sse41_lowp+0x89>
+ .byte 235,68 // jmp 31c29 <_sk_load_4444_dst_sse41_lowp+0x89>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,65,15,196,100,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm4
.byte 102,69,15,110,4,80 // movd (%r8,%rdx,2),%xmm8
.byte 102,68,15,58,14,196,252 // pblendw $0xfc,%xmm4,%xmm8
- .byte 235,41 // jmp 329ad <_sk_load_4444_dst_sse41_lowp+0x89>
+ .byte 235,41 // jmp 31c29 <_sk_load_4444_dst_sse41_lowp+0x89>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,65,15,196,100,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm4
.byte 102,65,15,196,100,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm4
@@ -59392,7 +58776,7 @@ _sk_load_4444_dst_sse41_lowp:
.byte 102,15,113,213,12 // psrlw $0xc,%xmm5
.byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
.byte 102,15,113,214,8 // psrlw $0x8,%xmm6
- .byte 102,15,111,37,71,126,0,0 // movdqa 0x7e47(%rip),%xmm4 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,37,75,126,0,0 // movdqa 0x7e4b(%rip),%xmm4 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,244 // pand %xmm4,%xmm6
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,113,215,4 // psrlw $0x4,%xmm7
@@ -59437,12 +58821,12 @@ _sk_store_4444_sse41_lowp:
.byte 76,3,0 // add (%rax),%r8
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
.byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,219,5,196,125,0,0 // pand 0x7dc4(%rip),%xmm8 # 3a820 <_sk_srcover_bgra_8888_sse2_lowp+0x1398>
+ .byte 102,68,15,219,5,200,125,0,0 // pand 0x7dc8(%rip),%xmm8 # 39aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x139c>
.byte 102,68,15,111,201 // movdqa %xmm1,%xmm9
.byte 102,65,15,113,241,4 // psllw $0x4,%xmm9
- .byte 102,68,15,219,13,192,125,0,0 // pand 0x7dc0(%rip),%xmm9 # 3a830 <_sk_srcover_bgra_8888_sse2_lowp+0x13a8>
+ .byte 102,68,15,219,13,196,125,0,0 // pand 0x7dc4(%rip),%xmm9 # 39ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x13ac>
.byte 102,69,15,235,200 // por %xmm8,%xmm9
- .byte 102,68,15,111,21,194,125,0,0 // movdqa 0x7dc2(%rip),%xmm10 # 3a840 <_sk_srcover_bgra_8888_sse2_lowp+0x13b8>
+ .byte 102,68,15,111,21,198,125,0,0 // movdqa 0x7dc6(%rip),%xmm10 # 39ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x13bc>
.byte 102,68,15,219,210 // pand %xmm2,%xmm10
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
.byte 102,65,15,113,208,4 // psrlw $0x4,%xmm8
@@ -59452,19 +58836,19 @@ _sk_store_4444_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,29 // ja 32abf <_sk_store_4444_sse41_lowp+0x87>
+ .byte 119,29 // ja 31d3b <_sk_store_4444_sse41_lowp+0x87>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,84,0,0,0 // lea 0x54(%rip),%r9 # 32b00 <_sk_store_4444_sse41_lowp+0xc8>
+ .byte 76,141,13,84,0,0,0 // lea 0x54(%rip),%r9 # 31d7c <_sk_store_4444_sse41_lowp+0xc8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,58,21,4,80,0 // pextrw $0x0,%xmm8,(%r8,%rdx,2)
- .byte 235,58 // jmp 32af9 <_sk_store_4444_sse41_lowp+0xc1>
+ .byte 235,58 // jmp 31d75 <_sk_store_4444_sse41_lowp+0xc1>
.byte 243,69,15,127,4,80 // movdqu %xmm8,(%r8,%rdx,2)
- .byte 235,50 // jmp 32af9 <_sk_store_4444_sse41_lowp+0xc1>
+ .byte 235,50 // jmp 31d75 <_sk_store_4444_sse41_lowp+0xc1>
.byte 102,69,15,58,21,68,80,4,2 // pextrw $0x2,%xmm8,0x4(%r8,%rdx,2)
.byte 102,69,15,126,4,80 // movd %xmm8,(%r8,%rdx,2)
- .byte 235,33 // jmp 32af9 <_sk_store_4444_sse41_lowp+0xc1>
+ .byte 235,33 // jmp 31d75 <_sk_store_4444_sse41_lowp+0xc1>
.byte 102,69,15,58,21,68,80,12,6 // pextrw $0x6,%xmm8,0xc(%r8,%rdx,2)
.byte 102,69,15,58,21,68,80,10,5 // pextrw $0x5,%xmm8,0xa(%r8,%rdx,2)
.byte 102,69,15,58,21,68,80,8,4 // pextrw $0x4,%xmm8,0x8(%r8,%rdx,2)
@@ -59564,7 +58948,7 @@ _sk_gather_4444_sse41_lowp:
.byte 102,15,113,209,12 // psrlw $0xc,%xmm1
.byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
.byte 102,15,113,210,8 // psrlw $0x8,%xmm2
- .byte 102,15,111,5,200,123,0,0 // movdqa 0x7bc8(%rip),%xmm0 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,5,204,123,0,0 // movdqa 0x7bcc(%rip),%xmm0 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,113,211,4 // psrlw $0x4,%xmm3
@@ -59602,18 +58986,18 @@ _sk_load_a8_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 32cda <_sk_load_a8_sse41_lowp+0x39>
+ .byte 119,34 // ja 31f56 <_sk_load_a8_sse41_lowp+0x39>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,146,0,0,0 // lea 0x92(%rip),%r9 # 32d58 <_sk_load_a8_sse41_lowp+0xb7>
+ .byte 76,141,13,146,0,0,0 // lea 0x92(%rip),%r9 # 31fd4 <_sk_load_a8_sse41_lowp+0xb7>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,216 // movd %eax,%xmm3
- .byte 235,100 // jmp 32d3e <_sk_load_a8_sse41_lowp+0x9d>
+ .byte 235,100 // jmp 31fba <_sk_load_a8_sse41_lowp+0x9d>
.byte 102,65,15,56,48,28,16 // pmovzxbw (%r8,%rdx,1),%xmm3
- .byte 235,91 // jmp 32d3e <_sk_load_a8_sse41_lowp+0x9d>
+ .byte 235,91 // jmp 31fba <_sk_load_a8_sse41_lowp+0x9d>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,15,196,192,2 // pinsrw $0x2,%eax,%xmm0
@@ -59621,7 +59005,7 @@ _sk_load_a8_sse41_lowp:
.byte 102,15,110,200 // movd %eax,%xmm1
.byte 102,15,56,48,217 // pmovzxbw %xmm1,%xmm3
.byte 102,15,58,14,216,252 // pblendw $0xfc,%xmm0,%xmm3
- .byte 235,54 // jmp 32d3e <_sk_load_a8_sse41_lowp+0x9d>
+ .byte 235,54 // jmp 31fba <_sk_load_a8_sse41_lowp+0x9d>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,15,196,192,6 // pinsrw $0x6,%eax,%xmm0
@@ -59632,14 +59016,14 @@ _sk_load_a8_sse41_lowp:
.byte 102,65,15,110,12,16 // movd (%r8,%rdx,1),%xmm1
.byte 102,15,56,48,217 // pmovzxbw %xmm1,%xmm3
.byte 102,15,58,14,216,240 // pblendw $0xf0,%xmm0,%xmm3
- .byte 102,15,219,29,202,114,0,0 // pand 0x72ca(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,29,206,114,0,0 // pand 0x72ce(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 119,255 // ja 32d59 <_sk_load_a8_sse41_lowp+0xb8>
+ .byte 119,255 // ja 31fd5 <_sk_load_a8_sse41_lowp+0xb8>
.byte 255 // (bad)
.byte 255,154,255,255,255,139 // lcall *-0x74000001(%rdx)
.byte 255 // (bad)
@@ -59668,18 +59052,18 @@ _sk_load_a8_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 32dad <_sk_load_a8_dst_sse41_lowp+0x39>
+ .byte 119,34 // ja 32029 <_sk_load_a8_dst_sse41_lowp+0x39>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,143,0,0,0 // lea 0x8f(%rip),%r9 # 32e28 <_sk_load_a8_dst_sse41_lowp+0xb4>
+ .byte 76,141,13,143,0,0,0 // lea 0x8f(%rip),%r9 # 320a4 <_sk_load_a8_dst_sse41_lowp+0xb4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,248 // movd %eax,%xmm7
- .byte 235,100 // jmp 32e11 <_sk_load_a8_dst_sse41_lowp+0x9d>
+ .byte 235,100 // jmp 3208d <_sk_load_a8_dst_sse41_lowp+0x9d>
.byte 102,65,15,56,48,60,16 // pmovzxbw (%r8,%rdx,1),%xmm7
- .byte 235,91 // jmp 32e11 <_sk_load_a8_dst_sse41_lowp+0x9d>
+ .byte 235,91 // jmp 3208d <_sk_load_a8_dst_sse41_lowp+0x9d>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,15,196,224,2 // pinsrw $0x2,%eax,%xmm4
@@ -59687,7 +59071,7 @@ _sk_load_a8_dst_sse41_lowp:
.byte 102,15,110,232 // movd %eax,%xmm5
.byte 102,15,56,48,253 // pmovzxbw %xmm5,%xmm7
.byte 102,15,58,14,252,252 // pblendw $0xfc,%xmm4,%xmm7
- .byte 235,54 // jmp 32e11 <_sk_load_a8_dst_sse41_lowp+0x9d>
+ .byte 235,54 // jmp 3208d <_sk_load_a8_dst_sse41_lowp+0x9d>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,15,196,224,6 // pinsrw $0x6,%eax,%xmm4
@@ -59698,13 +59082,13 @@ _sk_load_a8_dst_sse41_lowp:
.byte 102,65,15,110,44,16 // movd (%r8,%rdx,1),%xmm5
.byte 102,15,56,48,253 // pmovzxbw %xmm5,%xmm7
.byte 102,15,58,14,252,240 // pblendw $0xf0,%xmm4,%xmm7
- .byte 102,15,219,61,247,113,0,0 // pand 0x71f7(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,61,251,113,0,0 // pand 0x71fb(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 15,87,246 // xorps %xmm6,%xmm6
.byte 255,224 // jmpq *%rax
- .byte 122,255 // jp 32e29 <_sk_load_a8_dst_sse41_lowp+0xb5>
+ .byte 122,255 // jp 320a5 <_sk_load_a8_dst_sse41_lowp+0xb5>
.byte 255 // (bad)
.byte 255,157,255,255,255,142 // lcall *-0x71000001(%rbp)
.byte 255 // (bad)
@@ -59735,28 +59119,28 @@ _sk_store_a8_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,29 // ja 32e78 <_sk_store_a8_sse41_lowp+0x34>
+ .byte 119,29 // ja 320f4 <_sk_store_a8_sse41_lowp+0x34>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,131,0,0,0 // lea 0x83(%rip),%r9 # 32ee8 <_sk_store_a8_sse41_lowp+0xa4>
+ .byte 76,141,13,131,0,0,0 // lea 0x83(%rip),%r9 # 32164 <_sk_store_a8_sse41_lowp+0xa4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,65,15,58,20,28,16,0 // pextrb $0x0,%xmm3,(%r8,%rdx,1)
- .byte 235,105 // jmp 32ee1 <_sk_store_a8_sse41_lowp+0x9d>
+ .byte 235,105 // jmp 3215d <_sk_store_a8_sse41_lowp+0x9d>
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,68,15,56,0,5,217,120,0,0 // pshufb 0x78d9(%rip),%xmm8 # 3a760 <_sk_srcover_bgra_8888_sse2_lowp+0x12d8>
+ .byte 102,68,15,56,0,5,221,120,0,0 // pshufb 0x78dd(%rip),%xmm8 # 399e0 <_sk_srcover_bgra_8888_sse2_lowp+0x12dc>
.byte 102,69,15,214,4,16 // movq %xmm8,(%r8,%rdx,1)
- .byte 235,82 // jmp 32ee1 <_sk_store_a8_sse41_lowp+0x9d>
+ .byte 235,82 // jmp 3215d <_sk_store_a8_sse41_lowp+0x9d>
.byte 102,65,15,58,20,92,16,2,4 // pextrb $0x4,%xmm3,0x2(%r8,%rdx,1)
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,68,15,56,0,5,121,113,0,0 // pshufb 0x7179(%rip),%xmm8 # 3a020 <_sk_srcover_bgra_8888_sse2_lowp+0xb98>
+ .byte 102,68,15,56,0,5,125,113,0,0 // pshufb 0x717d(%rip),%xmm8 # 392a0 <_sk_srcover_bgra_8888_sse2_lowp+0xb9c>
.byte 102,69,15,58,21,4,16,0 // pextrw $0x0,%xmm8,(%r8,%rdx,1)
- .byte 235,48 // jmp 32ee1 <_sk_store_a8_sse41_lowp+0x9d>
+ .byte 235,48 // jmp 3215d <_sk_store_a8_sse41_lowp+0x9d>
.byte 102,65,15,58,20,92,16,6,12 // pextrb $0xc,%xmm3,0x6(%r8,%rdx,1)
.byte 102,65,15,58,20,92,16,5,10 // pextrb $0xa,%xmm3,0x5(%r8,%rdx,1)
.byte 102,65,15,58,20,92,16,4,8 // pextrb $0x8,%xmm3,0x4(%r8,%rdx,1)
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,68,15,56,0,5,85,113,0,0 // pshufb 0x7155(%rip),%xmm8 # 3a030 <_sk_srcover_bgra_8888_sse2_lowp+0xba8>
+ .byte 102,68,15,56,0,5,89,113,0,0 // pshufb 0x7159(%rip),%xmm8 # 392b0 <_sk_srcover_bgra_8888_sse2_lowp+0xbac>
.byte 102,69,15,126,4,16 // movd %xmm8,(%r8,%rdx,1)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
@@ -59869,18 +59253,18 @@ _sk_load_g8_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 3306a <_sk_load_g8_sse41_lowp+0x39>
+ .byte 119,34 // ja 322e6 <_sk_load_g8_sse41_lowp+0x39>
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,150,0,0,0 // lea 0x96(%rip),%r9 # 330ec <_sk_load_g8_sse41_lowp+0xbb>
+ .byte 76,141,13,150,0,0,0 // lea 0x96(%rip),%r9 # 32368 <_sk_load_g8_sse41_lowp+0xbb>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,192 // movd %eax,%xmm0
- .byte 235,100 // jmp 330ce <_sk_load_g8_sse41_lowp+0x9d>
+ .byte 235,100 // jmp 3234a <_sk_load_g8_sse41_lowp+0x9d>
.byte 102,65,15,56,48,4,16 // pmovzxbw (%r8,%rdx,1),%xmm0
- .byte 235,91 // jmp 330ce <_sk_load_g8_sse41_lowp+0x9d>
+ .byte 235,91 // jmp 3234a <_sk_load_g8_sse41_lowp+0x9d>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 102,15,196,200,2 // pinsrw $0x2,%eax,%xmm1
@@ -59888,7 +59272,7 @@ _sk_load_g8_sse41_lowp:
.byte 102,15,110,192 // movd %eax,%xmm0
.byte 102,15,56,48,192 // pmovzxbw %xmm0,%xmm0
.byte 102,15,58,14,193,252 // pblendw $0xfc,%xmm1,%xmm0
- .byte 235,54 // jmp 330ce <_sk_load_g8_sse41_lowp+0x9d>
+ .byte 235,54 // jmp 3234a <_sk_load_g8_sse41_lowp+0x9d>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,201 // pxor %xmm1,%xmm1
.byte 102,15,196,200,6 // pinsrw $0x6,%eax,%xmm1
@@ -59899,14 +59283,14 @@ _sk_load_g8_sse41_lowp:
.byte 102,65,15,110,4,16 // movd (%r8,%rdx,1),%xmm0
.byte 102,15,56,48,192 // pmovzxbw %xmm0,%xmm0
.byte 102,15,58,14,193,240 // pblendw $0xf0,%xmm1,%xmm0
- .byte 102,15,219,5,58,111,0,0 // pand 0x6f3a(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,5,62,111,0,0 // pand 0x6f3e(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,49,111,0,0 // movaps 0x6f31(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,53,111,0,0 // movaps 0x6f35(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,111,208 // movdqa %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 115,255 // jae 330ed <_sk_load_g8_sse41_lowp+0xbc>
+ .byte 115,255 // jae 32369 <_sk_load_g8_sse41_lowp+0xbc>
.byte 255 // (bad)
.byte 255,150,255,255,255,135 // callq *-0x78000001(%rsi)
.byte 255 // (bad)
@@ -59935,18 +59319,18 @@ _sk_load_g8_dst_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 33141 <_sk_load_g8_dst_sse41_lowp+0x39>
+ .byte 119,34 // ja 323bd <_sk_load_g8_dst_sse41_lowp+0x39>
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 331c0 <_sk_load_g8_dst_sse41_lowp+0xb8>
+ .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 3243c <_sk_load_g8_dst_sse41_lowp+0xb8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,224 // movd %eax,%xmm4
- .byte 235,100 // jmp 331a5 <_sk_load_g8_dst_sse41_lowp+0x9d>
+ .byte 235,100 // jmp 32421 <_sk_load_g8_dst_sse41_lowp+0x9d>
.byte 102,65,15,56,48,36,16 // pmovzxbw (%r8,%rdx,1),%xmm4
- .byte 235,91 // jmp 331a5 <_sk_load_g8_dst_sse41_lowp+0x9d>
+ .byte 235,91 // jmp 32421 <_sk_load_g8_dst_sse41_lowp+0x9d>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 102,15,196,232,2 // pinsrw $0x2,%eax,%xmm5
@@ -59954,7 +59338,7 @@ _sk_load_g8_dst_sse41_lowp:
.byte 102,15,110,224 // movd %eax,%xmm4
.byte 102,15,56,48,228 // pmovzxbw %xmm4,%xmm4
.byte 102,15,58,14,229,252 // pblendw $0xfc,%xmm5,%xmm4
- .byte 235,54 // jmp 331a5 <_sk_load_g8_dst_sse41_lowp+0x9d>
+ .byte 235,54 // jmp 32421 <_sk_load_g8_dst_sse41_lowp+0x9d>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 102,15,196,232,6 // pinsrw $0x6,%eax,%xmm5
@@ -59965,13 +59349,13 @@ _sk_load_g8_dst_sse41_lowp:
.byte 102,65,15,110,36,16 // movd (%r8,%rdx,1),%xmm4
.byte 102,15,56,48,228 // pmovzxbw %xmm4,%xmm4
.byte 102,15,58,14,229,240 // pblendw $0xf0,%xmm5,%xmm4
- .byte 102,15,219,37,99,110,0,0 // pand 0x6e63(%rip),%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,37,103,110,0,0 // pand 0x6e67(%rip),%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,90,110,0,0 // movaps 0x6e5a(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,61,94,110,0,0 // movaps 0x6e5e(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,111,244 // movdqa %xmm4,%xmm6
.byte 255,224 // jmpq *%rax
- .byte 118,255 // jbe 331c1 <_sk_load_g8_dst_sse41_lowp+0xb9>
+ .byte 118,255 // jbe 3243d <_sk_load_g8_dst_sse41_lowp+0xb9>
.byte 255 // (bad)
.byte 255,153,255,255,255,138 // lcall *-0x75000001(%rcx)
.byte 255 // (bad)
@@ -59993,10 +59377,10 @@ HIDDEN _sk_luminance_to_alpha_sse41_lowp
FUNCTION(_sk_luminance_to_alpha_sse41_lowp)
_sk_luminance_to_alpha_sse41_lowp:
.byte 102,15,111,218 // movdqa %xmm2,%xmm3
- .byte 102,15,213,5,104,118,0,0 // pmullw 0x7668(%rip),%xmm0 # 3a850 <_sk_srcover_bgra_8888_sse2_lowp+0x13c8>
- .byte 102,15,213,13,112,118,0,0 // pmullw 0x7670(%rip),%xmm1 # 3a860 <_sk_srcover_bgra_8888_sse2_lowp+0x13d8>
+ .byte 102,15,213,5,108,118,0,0 // pmullw 0x766c(%rip),%xmm0 # 39ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x13cc>
+ .byte 102,15,213,13,116,118,0,0 // pmullw 0x7674(%rip),%xmm1 # 39ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x13dc>
.byte 102,15,253,200 // paddw %xmm0,%xmm1
- .byte 102,15,213,29,116,118,0,0 // pmullw 0x7674(%rip),%xmm3 # 3a870 <_sk_srcover_bgra_8888_sse2_lowp+0x13e8>
+ .byte 102,15,213,29,120,118,0,0 // pmullw 0x7678(%rip),%xmm3 # 39af0 <_sk_srcover_bgra_8888_sse2_lowp+0x13ec>
.byte 102,15,253,217 // paddw %xmm1,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -60072,7 +59456,7 @@ _sk_gather_g8_sse41_lowp:
.byte 102,15,58,32,192,7 // pinsrb $0x7,%eax,%xmm0
.byte 102,15,56,48,192 // pmovzxbw %xmm0,%xmm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,221,108,0,0 // movaps 0x6cdd(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,225,108,0,0 // movaps 0x6ce1(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,111,208 // movdqa %xmm0,%xmm2
.byte 91 // pop %rbx
@@ -60088,14 +59472,14 @@ FUNCTION(_sk_scale_1_float_sse41_lowp)
_sk_scale_1_float_sse41_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,89,5,15,100,0,0 // mulss 0x640f(%rip),%xmm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 243,68,15,88,5,214,99,0,0 // addss 0x63d6(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,5,15,100,0,0 // mulss 0x640f(%rip),%xmm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,68,15,88,5,214,99,0,0 // addss 0x63d6(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,44,192 // cvttss2si %xmm8,%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
.byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
- .byte 102,68,15,111,13,142,108,0,0 // movdqa 0x6c8e(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,146,108,0,0 // movdqa 0x6c92(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -60116,13 +59500,13 @@ FUNCTION(_sk_lerp_1_float_sse41_lowp)
_sk_lerp_1_float_sse41_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,89,5,151,99,0,0 // mulss 0x6397(%rip),%xmm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 243,68,15,88,5,94,99,0,0 // addss 0x635e(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,5,151,99,0,0 // mulss 0x6397(%rip),%xmm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,68,15,88,5,94,99,0,0 // addss 0x635e(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,44,192 // cvttss2si %xmm8,%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
.byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
- .byte 102,68,15,111,13,27,108,0,0 // movdqa 0x6c1b(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,31,108,0,0 // movdqa 0x6c1f(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -60162,18 +59546,18 @@ _sk_scale_u8_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 334ac <_sk_scale_u8_sse41_lowp+0x3b>
+ .byte 119,36 // ja 32728 <_sk_scale_u8_sse41_lowp+0x3b>
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,221,0,0,0 // lea 0xdd(%rip),%r9 # 33574 <_sk_scale_u8_sse41_lowp+0x103>
+ .byte 76,141,13,221,0,0,0 // lea 0xdd(%rip),%r9 # 327f0 <_sk_scale_u8_sse41_lowp+0x103>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,111 // jmp 3351b <_sk_scale_u8_sse41_lowp+0xaa>
+ .byte 235,111 // jmp 32797 <_sk_scale_u8_sse41_lowp+0xaa>
.byte 102,69,15,56,48,4,16 // pmovzxbw (%r8,%rdx,1),%xmm8
- .byte 235,102 // jmp 3351b <_sk_scale_u8_sse41_lowp+0xaa>
+ .byte 235,102 // jmp 32797 <_sk_scale_u8_sse41_lowp+0xaa>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,68,15,196,200,2 // pinsrw $0x2,%eax,%xmm9
@@ -60181,7 +59565,7 @@ _sk_scale_u8_sse41_lowp:
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 102,69,15,56,48,192 // pmovzxbw %xmm8,%xmm8
.byte 102,69,15,58,14,193,252 // pblendw $0xfc,%xmm9,%xmm8
- .byte 235,60 // jmp 3351b <_sk_scale_u8_sse41_lowp+0xaa>
+ .byte 235,60 // jmp 32797 <_sk_scale_u8_sse41_lowp+0xaa>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,68,15,196,200,6 // pinsrw $0x6,%eax,%xmm9
@@ -60192,9 +59576,9 @@ _sk_scale_u8_sse41_lowp:
.byte 102,69,15,110,4,16 // movd (%r8,%rdx,1),%xmm8
.byte 102,69,15,56,48,192 // pmovzxbw %xmm8,%xmm8
.byte 102,69,15,58,14,193,240 // pblendw $0xf0,%xmm9,%xmm8
- .byte 102,68,15,219,5,236,106,0,0 // pand 0x6aec(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,219,5,240,106,0,0 // pand 0x6af0(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
- .byte 102,68,15,111,13,222,106,0,0 // movdqa 0x6ade(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,226,106,0,0 // movdqa 0x6ae2(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -60219,7 +59603,7 @@ _sk_scale_u8_sse41_lowp:
.byte 255,148,255,255,255,136,255 // callq *-0x770001(%rdi,%rdi,8)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 124,255 // jl 33589 <_sk_scale_u8_sse41_lowp+0x118>
+ .byte 124,255 // jl 32805 <_sk_scale_u8_sse41_lowp+0x118>
.byte 255 // (bad)
.byte 255,107,255 // ljmp *-0x1(%rbx)
.byte 255 // (bad)
@@ -60237,18 +59621,18 @@ _sk_lerp_u8_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 335cb <_sk_lerp_u8_sse41_lowp+0x3b>
+ .byte 119,36 // ja 32847 <_sk_lerp_u8_sse41_lowp+0x3b>
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,30,1,0,0 // lea 0x11e(%rip),%r9 # 336d4 <_sk_lerp_u8_sse41_lowp+0x144>
+ .byte 76,141,13,30,1,0,0 // lea 0x11e(%rip),%r9 # 32950 <_sk_lerp_u8_sse41_lowp+0x144>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,111 // jmp 3363a <_sk_lerp_u8_sse41_lowp+0xaa>
+ .byte 235,111 // jmp 328b6 <_sk_lerp_u8_sse41_lowp+0xaa>
.byte 102,69,15,56,48,4,16 // pmovzxbw (%r8,%rdx,1),%xmm8
- .byte 235,102 // jmp 3363a <_sk_lerp_u8_sse41_lowp+0xaa>
+ .byte 235,102 // jmp 328b6 <_sk_lerp_u8_sse41_lowp+0xaa>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,68,15,196,200,2 // pinsrw $0x2,%eax,%xmm9
@@ -60256,7 +59640,7 @@ _sk_lerp_u8_sse41_lowp:
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 102,69,15,56,48,192 // pmovzxbw %xmm8,%xmm8
.byte 102,69,15,58,14,193,252 // pblendw $0xfc,%xmm9,%xmm8
- .byte 235,60 // jmp 3363a <_sk_lerp_u8_sse41_lowp+0xaa>
+ .byte 235,60 // jmp 328b6 <_sk_lerp_u8_sse41_lowp+0xaa>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 102,68,15,196,200,6 // pinsrw $0x6,%eax,%xmm9
@@ -60267,8 +59651,8 @@ _sk_lerp_u8_sse41_lowp:
.byte 102,69,15,110,4,16 // movd (%r8,%rdx,1),%xmm8
.byte 102,69,15,56,48,192 // pmovzxbw %xmm8,%xmm8
.byte 102,69,15,58,14,193,240 // pblendw $0xf0,%xmm9,%xmm8
- .byte 102,68,15,219,5,205,105,0,0 // pand 0x69cd(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
- .byte 102,68,15,111,21,196,105,0,0 // movdqa 0x69c4(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,219,5,209,105,0,0 // pand 0x69d1(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
+ .byte 102,68,15,111,21,200,105,0,0 // movdqa 0x69c8(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,69,15,239,202 // pxor %xmm10,%xmm9
.byte 102,69,15,111,217 // movdqa %xmm9,%xmm11
@@ -60298,7 +59682,7 @@ _sk_lerp_u8_sse41_lowp:
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 255,224 // jmpq *%rax
.byte 144 // nop
- .byte 235,254 // jmp 336d4 <_sk_lerp_u8_sse41_lowp+0x144>
+ .byte 235,254 // jmp 32950 <_sk_lerp_u8_sse41_lowp+0x144>
.byte 255 // (bad)
.byte 255,17 // callq *(%rcx)
.byte 255 // (bad)
@@ -60332,23 +59716,23 @@ _sk_scale_565_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,35 // ja 33732 <_sk_scale_565_sse41_lowp+0x42>
+ .byte 119,35 // ja 329ae <_sk_scale_565_sse41_lowp+0x42>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,111,1,0,0 // lea 0x16f(%rip),%r9 # 3388c <_sk_scale_565_sse41_lowp+0x19c>
+ .byte 76,141,13,111,1,0,0 // lea 0x16f(%rip),%r9 # 32b08 <_sk_scale_565_sse41_lowp+0x19c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,200 // movd %eax,%xmm9
- .byte 235,76 // jmp 3377e <_sk_scale_565_sse41_lowp+0x8e>
+ .byte 235,76 // jmp 329fa <_sk_scale_565_sse41_lowp+0x8e>
.byte 243,69,15,111,12,80 // movdqu (%r8,%rdx,2),%xmm9
- .byte 235,68 // jmp 3377e <_sk_scale_565_sse41_lowp+0x8e>
+ .byte 235,68 // jmp 329fa <_sk_scale_565_sse41_lowp+0x8e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm0
.byte 102,69,15,110,12,80 // movd (%r8,%rdx,2),%xmm9
.byte 102,68,15,58,14,200,252 // pblendw $0xfc,%xmm0,%xmm9
- .byte 235,41 // jmp 3377e <_sk_scale_565_sse41_lowp+0x8e>
+ .byte 235,41 // jmp 329fa <_sk_scale_565_sse41_lowp+0x8e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm0
.byte 102,65,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm0
@@ -60357,11 +59741,11 @@ _sk_scale_565_sse41_lowp:
.byte 102,68,15,58,14,200,240 // pblendw $0xf0,%xmm0,%xmm9
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
- .byte 102,15,219,5,48,112,0,0 // pand 0x7030(%rip),%xmm0 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,5,52,112,0,0 // pand 0x7034(%rip),%xmm0 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,69,15,111,209 // movdqa %xmm9,%xmm10
.byte 102,65,15,113,210,5 // psrlw $0x5,%xmm10
- .byte 102,68,15,219,21,44,112,0,0 // pand 0x702c(%rip),%xmm10 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,68,15,111,29,51,112,0,0 // movdqa 0x7033(%rip),%xmm11 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,68,15,219,21,48,112,0,0 // pand 0x7030(%rip),%xmm10 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,68,15,111,29,55,112,0,0 // movdqa 0x7037(%rip),%xmm11 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,69,15,219,217 // pand %xmm9,%xmm11
.byte 102,65,15,113,209,13 // psrlw $0xd,%xmm9
.byte 102,68,15,235,200 // por %xmm0,%xmm9
@@ -60373,7 +59757,7 @@ _sk_scale_565_sse41_lowp:
.byte 102,15,113,240,3 // psllw $0x3,%xmm0
.byte 102,65,15,113,211,2 // psrlw $0x2,%xmm11
.byte 102,68,15,235,216 // por %xmm0,%xmm11
- .byte 102,15,111,5,145,111,0,0 // movdqa 0x6f91(%rip),%xmm0 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,15,111,5,149,111,0,0 // movdqa 0x6f95(%rip),%xmm0 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,227 // movdqa %xmm3,%xmm12
.byte 102,68,15,239,224 // pxor %xmm0,%xmm12
.byte 102,15,239,199 // pxor %xmm7,%xmm0
@@ -60386,7 +59770,7 @@ _sk_scale_565_sse41_lowp:
.byte 102,69,15,56,62,225 // pmaxuw %xmm9,%xmm12
.byte 102,69,15,56,16,229 // pblendvb %xmm0,%xmm13,%xmm12
.byte 102,69,15,213,200 // pmullw %xmm8,%xmm9
- .byte 102,15,111,5,217,103,0,0 // movdqa 0x67d9(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,5,221,103,0,0 // movdqa 0x67dd(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,253,200 // paddw %xmm0,%xmm9
.byte 102,65,15,113,209,8 // psrlw $0x8,%xmm9
.byte 102,68,15,213,209 // pmullw %xmm1,%xmm10
@@ -60440,23 +59824,23 @@ _sk_lerp_565_sse41_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,35 // ja 338ea <_sk_lerp_565_sse41_lowp+0x42>
+ .byte 119,35 // ja 32b66 <_sk_lerp_565_sse41_lowp+0x42>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,187,1,0,0 // lea 0x1bb(%rip),%r9 # 33a90 <_sk_lerp_565_sse41_lowp+0x1e8>
+ .byte 76,141,13,187,1,0,0 // lea 0x1bb(%rip),%r9 # 32d0c <_sk_lerp_565_sse41_lowp+0x1e8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,200 // movd %eax,%xmm9
- .byte 235,76 // jmp 33936 <_sk_lerp_565_sse41_lowp+0x8e>
+ .byte 235,76 // jmp 32bb2 <_sk_lerp_565_sse41_lowp+0x8e>
.byte 243,69,15,111,12,80 // movdqu (%r8,%rdx,2),%xmm9
- .byte 235,68 // jmp 33936 <_sk_lerp_565_sse41_lowp+0x8e>
+ .byte 235,68 // jmp 32bb2 <_sk_lerp_565_sse41_lowp+0x8e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm0
.byte 102,69,15,110,12,80 // movd (%r8,%rdx,2),%xmm9
.byte 102,68,15,58,14,200,252 // pblendw $0xfc,%xmm0,%xmm9
- .byte 235,41 // jmp 33936 <_sk_lerp_565_sse41_lowp+0x8e>
+ .byte 235,41 // jmp 32bb2 <_sk_lerp_565_sse41_lowp+0x8e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm0
.byte 102,65,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm0
@@ -60465,11 +59849,11 @@ _sk_lerp_565_sse41_lowp:
.byte 102,68,15,58,14,200,240 // pblendw $0xf0,%xmm0,%xmm9
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
- .byte 102,15,219,5,120,110,0,0 // pand 0x6e78(%rip),%xmm0 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,5,124,110,0,0 // pand 0x6e7c(%rip),%xmm0 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,69,15,111,209 // movdqa %xmm9,%xmm10
.byte 102,65,15,113,210,5 // psrlw $0x5,%xmm10
- .byte 102,68,15,219,21,116,110,0,0 // pand 0x6e74(%rip),%xmm10 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,68,15,111,29,123,110,0,0 // movdqa 0x6e7b(%rip),%xmm11 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,68,15,219,21,120,110,0,0 // pand 0x6e78(%rip),%xmm10 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,68,15,111,29,127,110,0,0 // movdqa 0x6e7f(%rip),%xmm11 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,69,15,219,217 // pand %xmm9,%xmm11
.byte 102,65,15,113,209,13 // psrlw $0xd,%xmm9
.byte 102,68,15,235,200 // por %xmm0,%xmm9
@@ -60481,7 +59865,7 @@ _sk_lerp_565_sse41_lowp:
.byte 102,15,113,240,3 // psllw $0x3,%xmm0
.byte 102,65,15,113,211,2 // psrlw $0x2,%xmm11
.byte 102,68,15,235,216 // por %xmm0,%xmm11
- .byte 102,15,111,5,217,109,0,0 // movdqa 0x6dd9(%rip),%xmm0 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,15,111,5,221,109,0,0 // movdqa 0x6ddd(%rip),%xmm0 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,227 // movdqa %xmm3,%xmm12
.byte 102,68,15,239,224 // pxor %xmm0,%xmm12
.byte 102,15,239,199 // pxor %xmm7,%xmm0
@@ -60493,7 +59877,7 @@ _sk_lerp_565_sse41_lowp:
.byte 102,69,15,56,62,227 // pmaxuw %xmm11,%xmm12
.byte 102,69,15,56,62,225 // pmaxuw %xmm9,%xmm12
.byte 102,69,15,56,16,229 // pblendvb %xmm0,%xmm13,%xmm12
- .byte 102,68,15,111,45,37,102,0,0 // movdqa 0x6625(%rip),%xmm13 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,45,41,102,0,0 // movdqa 0x6629(%rip),%xmm13 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,111,193 // movdqa %xmm9,%xmm0
.byte 102,65,15,239,197 // pxor %xmm13,%xmm0
.byte 102,15,213,196 // pmullw %xmm4,%xmm0
@@ -60550,7 +59934,7 @@ _sk_clamp_x_1_sse41_lowp:
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 65,15,95,200 // maxps %xmm8,%xmm1
.byte 65,15,95,192 // maxps %xmm8,%xmm0
- .byte 68,15,40,5,176,101,0,0 // movaps 0x65b0(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,180,101,0,0 // movaps 0x65b4(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -60567,7 +59951,7 @@ _sk_repeat_x_1_sse41_lowp:
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 65,15,95,200 // maxps %xmm8,%xmm1
.byte 65,15,95,192 // maxps %xmm8,%xmm0
- .byte 68,15,40,5,122,101,0,0 // movaps 0x657a(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,126,101,0,0 // movaps 0x657e(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -60577,10 +59961,10 @@ HIDDEN _sk_mirror_x_1_sse41_lowp
.globl _sk_mirror_x_1_sse41_lowp
FUNCTION(_sk_mirror_x_1_sse41_lowp)
_sk_mirror_x_1_sse41_lowp:
- .byte 68,15,40,5,198,101,0,0 // movaps 0x65c6(%rip),%xmm8 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,5,202,101,0,0 // movaps 0x65ca(%rip),%xmm8 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 65,15,88,192 // addps %xmm8,%xmm0
.byte 65,15,88,200 // addps %xmm8,%xmm1
- .byte 68,15,40,13,70,101,0,0 // movaps 0x6546(%rip),%xmm9 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,13,74,101,0,0 // movaps 0x654a(%rip),%xmm9 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,40,209 // movaps %xmm1,%xmm10
.byte 69,15,89,209 // mulps %xmm9,%xmm10
.byte 68,15,89,200 // mulps %xmm0,%xmm9
@@ -60592,13 +59976,13 @@ _sk_mirror_x_1_sse41_lowp:
.byte 65,15,92,202 // subps %xmm10,%xmm1
.byte 65,15,88,200 // addps %xmm8,%xmm1
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,5,60,106,0,0 // movaps 0x6a3c(%rip),%xmm8 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 68,15,40,5,64,106,0,0 // movaps 0x6a40(%rip),%xmm8 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 65,15,84,192 // andps %xmm8,%xmm0
.byte 65,15,84,200 // andps %xmm8,%xmm1
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 65,15,95,200 // maxps %xmm8,%xmm1
.byte 65,15,95,192 // maxps %xmm8,%xmm0
- .byte 68,15,40,5,0,101,0,0 // movaps 0x6500(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,4,101,0,0 // movaps 0x6504(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -60625,7 +60009,7 @@ _sk_gradient_sse41_lowp:
.byte 102,69,15,239,201 // pxor %xmm9,%xmm9
.byte 72,131,249,2 // cmp $0x2,%rcx
.byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 114,57 // jb 33bf1 <_sk_gradient_sse41_lowp+0x75>
+ .byte 114,57 // jb 32e6d <_sk_gradient_sse41_lowp+0x75>
.byte 72,139,80,72 // mov 0x48(%rax),%rdx
.byte 72,255,201 // dec %rcx
.byte 72,131,194,4 // add $0x4,%rdx
@@ -60640,7 +60024,7 @@ _sk_gradient_sse41_lowp:
.byte 102,15,250,218 // psubd %xmm2,%xmm3
.byte 72,131,194,4 // add $0x4,%rdx
.byte 72,255,201 // dec %rcx
- .byte 117,219 // jne 33bcc <_sk_gradient_sse41_lowp+0x50>
+ .byte 117,219 // jne 32e48 <_sk_gradient_sse41_lowp+0x50>
.byte 102,72,15,58,22,219,1 // pextrq $0x1,%xmm3,%rbx
.byte 65,137,221 // mov %ebx,%r13d
.byte 72,193,235,32 // shr $0x20,%rbx
@@ -60719,14 +60103,14 @@ _sk_gradient_sse41_lowp:
.byte 69,15,88,239 // addps %xmm15,%xmm13
.byte 15,89,216 // mulps %xmm0,%xmm3
.byte 15,88,223 // addps %xmm7,%xmm3
- .byte 15,40,61,131,99,0,0 // movaps 0x6383(%rip),%xmm7 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,61,135,99,0,0 // movaps 0x6387(%rip),%xmm7 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 68,15,89,239 // mulps %xmm7,%xmm13
.byte 15,89,223 // mulps %xmm7,%xmm3
- .byte 68,15,40,61,132,98,0,0 // movaps 0x6284(%rip),%xmm15 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,61,136,98,0,0 // movaps 0x6288(%rip),%xmm15 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,88,239 // addps %xmm15,%xmm13
.byte 65,15,88,223 // addps %xmm15,%xmm3
.byte 243,65,15,91,245 // cvttps2dq %xmm13,%xmm6
- .byte 102,68,15,111,45,158,105,0,0 // movdqa 0x699e(%rip),%xmm13 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,68,15,111,45,162,105,0,0 // movdqa 0x69a2(%rip),%xmm13 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,65,15,56,0,245 // pshufb %xmm13,%xmm6
.byte 243,68,15,91,195 // cvttps2dq %xmm3,%xmm8
.byte 102,69,15,56,0,197 // pshufb %xmm13,%xmm8
@@ -60830,9 +60214,9 @@ _sk_evenly_spaced_gradient_sse41_lowp:
.byte 72,139,24 // mov (%rax),%rbx
.byte 72,139,104,8 // mov 0x8(%rax),%rbp
.byte 72,255,203 // dec %rbx
- .byte 120,7 // js 33fca <_sk_evenly_spaced_gradient_sse41_lowp+0x34>
+ .byte 120,7 // js 33246 <_sk_evenly_spaced_gradient_sse41_lowp+0x34>
.byte 243,72,15,42,211 // cvtsi2ss %rbx,%xmm2
- .byte 235,21 // jmp 33fdf <_sk_evenly_spaced_gradient_sse41_lowp+0x49>
+ .byte 235,21 // jmp 3325b <_sk_evenly_spaced_gradient_sse41_lowp+0x49>
.byte 73,137,216 // mov %rbx,%r8
.byte 73,209,232 // shr %r8
.byte 131,227,1 // and $0x1,%ebx
@@ -60970,14 +60354,14 @@ _sk_evenly_spaced_gradient_sse41_lowp:
.byte 15,40,100,36,176 // movaps -0x50(%rsp),%xmm4
.byte 68,15,89,196 // mulps %xmm4,%xmm8
.byte 68,15,88,192 // addps %xmm0,%xmm8
- .byte 15,40,5,96,94,0,0 // movaps 0x5e60(%rip),%xmm0 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,5,100,94,0,0 // movaps 0x5e64(%rip),%xmm0 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 15,89,216 // mulps %xmm0,%xmm3
.byte 68,15,89,192 // mulps %xmm0,%xmm8
- .byte 68,15,40,53,97,93,0,0 // movaps 0x5d61(%rip),%xmm14 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,53,101,93,0,0 // movaps 0x5d65(%rip),%xmm14 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 69,15,88,198 // addps %xmm14,%xmm8
.byte 65,15,88,222 // addps %xmm14,%xmm3
.byte 243,15,91,219 // cvttps2dq %xmm3,%xmm3
- .byte 102,15,111,5,125,100,0,0 // movdqa 0x647d(%rip),%xmm0 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,5,129,100,0,0 // movdqa 0x6481(%rip),%xmm0 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,56,0,216 // pshufb %xmm0,%xmm3
.byte 243,69,15,91,192 // cvttps2dq %xmm8,%xmm8
.byte 102,68,15,56,0,192 // pshufb %xmm0,%xmm8
@@ -60987,7 +60371,7 @@ _sk_evenly_spaced_gradient_sse41_lowp:
.byte 15,40,220 // movaps %xmm4,%xmm3
.byte 68,15,89,203 // mulps %xmm3,%xmm9
.byte 68,15,88,201 // addps %xmm1,%xmm9
- .byte 15,40,13,16,94,0,0 // movaps 0x5e10(%rip),%xmm1 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,13,20,94,0,0 // movaps 0x5e14(%rip),%xmm1 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 15,89,209 // mulps %xmm1,%xmm2
.byte 68,15,89,201 // mulps %xmm1,%xmm9
.byte 15,40,225 // movaps %xmm1,%xmm4
@@ -61056,14 +60440,14 @@ _sk_evenly_spaced_2_stop_gradient_sse41_lowp:
.byte 15,89,216 // mulps %xmm0,%xmm3
.byte 65,15,88,216 // addps %xmm8,%xmm3
.byte 65,15,88,208 // addps %xmm8,%xmm2
- .byte 68,15,40,21,10,93,0,0 // movaps 0x5d0a(%rip),%xmm10 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,21,14,93,0,0 // movaps 0x5d0e(%rip),%xmm10 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,210 // mulps %xmm10,%xmm2
.byte 65,15,89,218 // mulps %xmm10,%xmm3
- .byte 68,15,40,37,10,92,0,0 // movaps 0x5c0a(%rip),%xmm12 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,37,14,92,0,0 // movaps 0x5c0e(%rip),%xmm12 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,88,220 // addps %xmm12,%xmm3
.byte 65,15,88,212 // addps %xmm12,%xmm2
.byte 243,15,91,210 // cvttps2dq %xmm2,%xmm2
- .byte 102,68,15,111,29,37,99,0,0 // movdqa 0x6325(%rip),%xmm11 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,68,15,111,29,41,99,0,0 // movdqa 0x6329(%rip),%xmm11 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,65,15,56,0,211 // pshufb %xmm11,%xmm2
.byte 243,68,15,91,195 // cvttps2dq %xmm3,%xmm8
.byte 102,69,15,56,0,195 // pshufb %xmm11,%xmm8
@@ -61136,7 +60520,7 @@ _sk_xy_to_unit_angle_sse41_lowp:
.byte 15,40,251 // movaps %xmm3,%xmm7
.byte 15,40,242 // movaps %xmm2,%xmm6
.byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 15,40,37,251,95,0,0 // movaps 0x5ffb(%rip),%xmm4 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 15,40,37,255,95,0,0 // movaps 0x5fff(%rip),%xmm4 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 68,15,40,233 // movaps %xmm1,%xmm13
.byte 68,15,84,236 // andps %xmm4,%xmm13
.byte 69,15,40,216 // movaps %xmm8,%xmm11
@@ -61161,18 +60545,18 @@ _sk_xy_to_unit_angle_sse41_lowp:
.byte 69,15,94,251 // divps %xmm11,%xmm15
.byte 65,15,40,199 // movaps %xmm15,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
- .byte 68,15,40,29,251,95,0,0 // movaps 0x5ffb(%rip),%xmm11 # 3a600 <_sk_srcover_bgra_8888_sse2_lowp+0x1178>
+ .byte 68,15,40,29,255,95,0,0 // movaps 0x5fff(%rip),%xmm11 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x117c>
.byte 68,15,40,224 // movaps %xmm0,%xmm12
.byte 69,15,89,227 // mulps %xmm11,%xmm12
- .byte 68,15,88,37,251,95,0,0 // addps 0x5ffb(%rip),%xmm12 # 3a610 <_sk_srcover_bgra_8888_sse2_lowp+0x1188>
+ .byte 68,15,88,37,255,95,0,0 // addps 0x5fff(%rip),%xmm12 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x118c>
.byte 68,15,89,224 // mulps %xmm0,%xmm12
- .byte 15,40,45,0,96,0,0 // movaps 0x6000(%rip),%xmm5 # 3a620 <_sk_srcover_bgra_8888_sse2_lowp+0x1198>
+ .byte 15,40,45,4,96,0,0 // movaps 0x6004(%rip),%xmm5 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x119c>
.byte 68,15,88,229 // addps %xmm5,%xmm12
.byte 68,15,89,224 // mulps %xmm0,%xmm12
- .byte 15,40,37,1,96,0,0 // movaps 0x6001(%rip),%xmm4 # 3a630 <_sk_srcover_bgra_8888_sse2_lowp+0x11a8>
+ .byte 15,40,37,5,96,0,0 // movaps 0x6005(%rip),%xmm4 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ac>
.byte 68,15,88,228 // addps %xmm4,%xmm12
.byte 69,15,89,231 // mulps %xmm15,%xmm12
- .byte 15,40,21,2,96,0,0 // movaps 0x6002(%rip),%xmm2 # 3a640 <_sk_srcover_bgra_8888_sse2_lowp+0x11b8>
+ .byte 15,40,21,6,96,0,0 // movaps 0x6006(%rip),%xmm2 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x11bc>
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 65,15,92,212 // subps %xmm12,%xmm2
.byte 65,15,40,194 // movaps %xmm10,%xmm0
@@ -61181,7 +60565,7 @@ _sk_xy_to_unit_angle_sse41_lowp:
.byte 65,15,40,198 // movaps %xmm14,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
.byte 68,15,89,216 // mulps %xmm0,%xmm11
- .byte 68,15,88,29,170,95,0,0 // addps 0x5faa(%rip),%xmm11 # 3a610 <_sk_srcover_bgra_8888_sse2_lowp+0x1188>
+ .byte 68,15,88,29,174,95,0,0 // addps 0x5fae(%rip),%xmm11 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x118c>
.byte 68,15,89,216 // mulps %xmm0,%xmm11
.byte 68,15,88,221 // addps %xmm5,%xmm11
.byte 68,15,89,216 // mulps %xmm0,%xmm11
@@ -61192,7 +60576,7 @@ _sk_xy_to_unit_angle_sse41_lowp:
.byte 102,68,15,56,20,219 // blendvps %xmm0,%xmm3,%xmm11
.byte 69,15,87,201 // xorps %xmm9,%xmm9
.byte 69,15,194,193,1 // cmpltps %xmm9,%xmm8
- .byte 15,40,21,200,89,0,0 // movaps 0x59c8(%rip),%xmm2 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,40,21,204,89,0,0 // movaps 0x59cc(%rip),%xmm2 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 65,15,92,212 // subps %xmm12,%xmm2
.byte 65,15,40,192 // movaps %xmm8,%xmm0
@@ -61203,7 +60587,7 @@ _sk_xy_to_unit_angle_sse41_lowp:
.byte 102,68,15,56,20,219 // blendvps %xmm0,%xmm3,%xmm11
.byte 15,40,198 // movaps %xmm6,%xmm0
.byte 65,15,194,193,1 // cmpltps %xmm9,%xmm0
- .byte 15,40,13,166,89,0,0 // movaps 0x59a6(%rip),%xmm1 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 15,40,13,170,89,0,0 // movaps 0x59aa(%rip),%xmm1 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 15,40,209 // movaps %xmm1,%xmm2
.byte 65,15,92,204 // subps %xmm12,%xmm1
.byte 102,68,15,56,20,225 // blendvps %xmm0,%xmm1,%xmm12
@@ -61257,48 +60641,48 @@ _sk_srcover_rgba_8888_sse41_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,202 // movzbl %r10b,%r9d
.byte 65,128,249,6 // cmp $0x6,%r9b
- .byte 119,38 // ja 34796 <_sk_srcover_rgba_8888_sse41_lowp+0x4b>
+ .byte 119,38 // ja 33a12 <_sk_srcover_rgba_8888_sse41_lowp+0x4b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 76,141,29,20,2,0,0 // lea 0x214(%rip),%r11 # 34990 <_sk_srcover_rgba_8888_sse41_lowp+0x245>
+ .byte 76,141,29,20,2,0,0 // lea 0x214(%rip),%r11 # 33c0c <_sk_srcover_rgba_8888_sse41_lowp+0x245>
.byte 75,99,4,139 // movslq (%r11,%r9,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,52,144 // movd (%r8,%rdx,4),%xmm6
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 347ef <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
+ .byte 235,89 // jmp 33a6b <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
.byte 243,65,15,111,52,144 // movdqu (%r8,%rdx,4),%xmm6
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 347ef <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
+ .byte 235,74 // jmp 33a6b <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 243,65,15,126,52,144 // movq (%r8,%rdx,4),%xmm6
.byte 102,65,15,58,14,240,240 // pblendw $0xf0,%xmm8,%xmm6
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,37 // jmp 347ef <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
+ .byte 235,37 // jmp 33a6b <_sk_srcover_rgba_8888_sse41_lowp+0xa4>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,52,144 // movdqu (%r8,%rdx,4),%xmm6
- .byte 102,15,111,37,153,95,0,0 // movdqa 0x5f99(%rip),%xmm4 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,37,157,95,0,0 // movdqa 0x5f9d(%rip),%xmm4 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,238 // movdqa %xmm6,%xmm5
.byte 102,15,56,0,236 // pshufb %xmm4,%xmm5
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,56,0,252 // pshufb %xmm4,%xmm7
.byte 102,15,108,239 // punpcklqdq %xmm7,%xmm5
- .byte 102,68,15,111,37,249,87,0,0 // movdqa 0x57f9(%rip),%xmm12 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,37,253,87,0,0 // movdqa 0x57fd(%rip),%xmm12 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,229 // movdqa %xmm5,%xmm4
.byte 102,65,15,219,228 // pand %xmm12,%xmm4
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
- .byte 102,68,15,111,13,114,95,0,0 // movdqa 0x5f72(%rip),%xmm9 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,13,118,95,0,0 // movdqa 0x5f76(%rip),%xmm9 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,254 // movdqa %xmm6,%xmm7
.byte 102,65,15,56,0,249 // pshufb %xmm9,%xmm7
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
.byte 102,69,15,56,0,209 // pshufb %xmm9,%xmm10
.byte 102,65,15,108,250 // punpcklqdq %xmm10,%xmm7
- .byte 102,68,15,111,13,95,95,0,0 // movdqa 0x5f5f(%rip),%xmm9 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,13,99,95,0,0 // movdqa 0x5f63(%rip),%xmm9 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,241 // pshufb %xmm9,%xmm6
.byte 102,69,15,56,0,193 // pshufb %xmm9,%xmm8
.byte 102,65,15,108,240 // punpcklqdq %xmm8,%xmm6
@@ -61340,19 +60724,19 @@ _sk_srcover_rgba_8888_sse41_lowp:
.byte 102,15,97,195 // punpcklwd %xmm3,%xmm0
.byte 102,65,15,235,196 // por %xmm12,%xmm0
.byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,24 // ja 34939 <_sk_srcover_rgba_8888_sse41_lowp+0x1ee>
- .byte 76,141,21,132,0,0,0 // lea 0x84(%rip),%r10 # 349ac <_sk_srcover_rgba_8888_sse41_lowp+0x261>
+ .byte 119,24 // ja 33bb5 <_sk_srcover_rgba_8888_sse41_lowp+0x1ee>
+ .byte 76,141,21,132,0,0,0 // lea 0x84(%rip),%r10 # 33c28 <_sk_srcover_rgba_8888_sse41_lowp+0x261>
.byte 75,99,4,138 // movslq (%r10,%r9,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 102,65,15,126,4,144 // movd %xmm0,(%r8,%rdx,4)
- .byte 235,63 // jmp 34978 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
+ .byte 235,63 // jmp 33bf4 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
.byte 243,65,15,127,4,144 // movdqu %xmm0,(%r8,%rdx,4)
.byte 243,65,15,127,76,144,16 // movdqu %xmm1,0x10(%r8,%rdx,4)
- .byte 235,48 // jmp 34978 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
+ .byte 235,48 // jmp 33bf4 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
.byte 102,65,15,58,22,68,144,8,2 // pextrd $0x2,%xmm0,0x8(%r8,%rdx,4)
.byte 102,65,15,214,4,144 // movq %xmm0,(%r8,%rdx,4)
- .byte 235,31 // jmp 34978 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
+ .byte 235,31 // jmp 33bf4 <_sk_srcover_rgba_8888_sse41_lowp+0x22d>
.byte 102,65,15,58,22,76,144,24,2 // pextrd $0x2,%xmm1,0x18(%r8,%rdx,4)
.byte 102,65,15,58,22,76,144,20,1 // pextrd $0x1,%xmm1,0x14(%r8,%rdx,4)
.byte 102,65,15,126,76,144,16 // movd %xmm1,0x10(%r8,%rdx,4)
@@ -61369,7 +60753,7 @@ _sk_srcover_rgba_8888_sse41_lowp:
.byte 255,38 // jmpq *(%rsi)
.byte 254 // (bad)
.byte 255 // (bad)
- .byte 255,21,254,255,255,89 // callq *0x59fffffe(%rip) # 5a03499b <_sk_srcover_bgra_8888_sse2_lowp+0x59ffb513>
+ .byte 255,21,254,255,255,89 // callq *0x59fffffe(%rip) # 5a033c17 <_sk_srcover_bgra_8888_sse2_lowp+0x59ffb513>
.byte 254 // (bad)
.byte 255 // (bad)
.byte 255,80,254 // callq *-0x2(%rax)
@@ -61408,48 +60792,48 @@ _sk_srcover_bgra_8888_sse41_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,202 // movzbl %r10b,%r9d
.byte 65,128,249,6 // cmp $0x6,%r9b
- .byte 119,38 // ja 34a13 <_sk_srcover_bgra_8888_sse41_lowp+0x4b>
+ .byte 119,38 // ja 33c8f <_sk_srcover_bgra_8888_sse41_lowp+0x4b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 76,141,29,23,2,0,0 // lea 0x217(%rip),%r11 # 34c10 <_sk_srcover_bgra_8888_sse41_lowp+0x248>
+ .byte 76,141,29,23,2,0,0 // lea 0x217(%rip),%r11 # 33e8c <_sk_srcover_bgra_8888_sse41_lowp+0x248>
.byte 75,99,4,139 // movslq (%r11,%r9,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 255,224 // jmpq *%rax
.byte 102,65,15,110,36,144 // movd (%r8,%rdx,4),%xmm4
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,89 // jmp 34a6c <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
+ .byte 235,89 // jmp 33ce8 <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
.byte 243,65,15,111,36,144 // movdqu (%r8,%rdx,4),%xmm4
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,74 // jmp 34a6c <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
+ .byte 235,74 // jmp 33ce8 <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,237 // pxor %xmm5,%xmm5
.byte 243,65,15,126,36,144 // movq (%r8,%rdx,4),%xmm4
.byte 102,65,15,58,14,224,240 // pblendw $0xf0,%xmm8,%xmm4
.byte 102,68,15,111,197 // movdqa %xmm5,%xmm8
- .byte 235,37 // jmp 34a6c <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
+ .byte 235,37 // jmp 33ce8 <_sk_srcover_bgra_8888_sse41_lowp+0xa4>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,69,15,58,34,68,144,20,1 // pinsrd $0x1,0x14(%r8,%rdx,4),%xmm8
.byte 102,69,15,58,34,68,144,16,0 // pinsrd $0x0,0x10(%r8,%rdx,4),%xmm8
.byte 243,65,15,111,36,144 // movdqu (%r8,%rdx,4),%xmm4
- .byte 102,15,111,53,28,93,0,0 // movdqa 0x5d1c(%rip),%xmm6 # 3a790 <_sk_srcover_bgra_8888_sse2_lowp+0x1308>
+ .byte 102,15,111,53,32,93,0,0 // movdqa 0x5d20(%rip),%xmm6 # 39a10 <_sk_srcover_bgra_8888_sse2_lowp+0x130c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,56,0,238 // pshufb %xmm6,%xmm5
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,56,0,254 // pshufb %xmm6,%xmm7
.byte 102,15,108,239 // punpcklqdq %xmm7,%xmm5
- .byte 102,68,15,111,37,124,85,0,0 // movdqa 0x557c(%rip),%xmm12 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,37,128,85,0,0 // movdqa 0x5580(%rip),%xmm12 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,245 // movdqa %xmm5,%xmm6
.byte 102,65,15,219,244 // pand %xmm12,%xmm6
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
- .byte 102,68,15,111,13,245,92,0,0 // movdqa 0x5cf5(%rip),%xmm9 # 3a7a0 <_sk_srcover_bgra_8888_sse2_lowp+0x1318>
+ .byte 102,68,15,111,13,249,92,0,0 // movdqa 0x5cf9(%rip),%xmm9 # 39a20 <_sk_srcover_bgra_8888_sse2_lowp+0x131c>
.byte 102,15,111,252 // movdqa %xmm4,%xmm7
.byte 102,65,15,56,0,249 // pshufb %xmm9,%xmm7
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
.byte 102,69,15,56,0,209 // pshufb %xmm9,%xmm10
.byte 102,65,15,108,250 // punpcklqdq %xmm10,%xmm7
- .byte 102,68,15,111,13,226,92,0,0 // movdqa 0x5ce2(%rip),%xmm9 # 3a7b0 <_sk_srcover_bgra_8888_sse2_lowp+0x1328>
+ .byte 102,68,15,111,13,230,92,0,0 // movdqa 0x5ce6(%rip),%xmm9 # 39a30 <_sk_srcover_bgra_8888_sse2_lowp+0x132c>
.byte 102,65,15,56,0,225 // pshufb %xmm9,%xmm4
.byte 102,69,15,56,0,193 // pshufb %xmm9,%xmm8
.byte 102,65,15,108,224 // punpcklqdq %xmm8,%xmm4
@@ -61491,19 +60875,19 @@ _sk_srcover_bgra_8888_sse41_lowp:
.byte 102,15,97,195 // punpcklwd %xmm3,%xmm0
.byte 102,65,15,235,196 // por %xmm12,%xmm0
.byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,24 // ja 34bb6 <_sk_srcover_bgra_8888_sse41_lowp+0x1ee>
- .byte 76,141,21,135,0,0,0 // lea 0x87(%rip),%r10 # 34c2c <_sk_srcover_bgra_8888_sse41_lowp+0x264>
+ .byte 119,24 // ja 33e32 <_sk_srcover_bgra_8888_sse41_lowp+0x1ee>
+ .byte 76,141,21,135,0,0,0 // lea 0x87(%rip),%r10 # 33ea8 <_sk_srcover_bgra_8888_sse41_lowp+0x264>
.byte 75,99,4,138 // movslq (%r10,%r9,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 102,65,15,126,4,144 // movd %xmm0,(%r8,%rdx,4)
- .byte 235,63 // jmp 34bf5 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
+ .byte 235,63 // jmp 33e71 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
.byte 243,65,15,127,4,144 // movdqu %xmm0,(%r8,%rdx,4)
.byte 243,65,15,127,76,144,16 // movdqu %xmm1,0x10(%r8,%rdx,4)
- .byte 235,48 // jmp 34bf5 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
+ .byte 235,48 // jmp 33e71 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
.byte 102,65,15,58,22,68,144,8,2 // pextrd $0x2,%xmm0,0x8(%r8,%rdx,4)
.byte 102,65,15,214,4,144 // movq %xmm0,(%r8,%rdx,4)
- .byte 235,31 // jmp 34bf5 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
+ .byte 235,31 // jmp 33e71 <_sk_srcover_bgra_8888_sse41_lowp+0x22d>
.byte 102,65,15,58,22,76,144,24,2 // pextrd $0x2,%xmm1,0x18(%r8,%rdx,4)
.byte 102,65,15,58,22,76,144,20,1 // pextrd $0x1,%xmm1,0x14(%r8,%rdx,4)
.byte 102,65,15,126,76,144,16 // movd %xmm1,0x10(%r8,%rdx,4)
@@ -61568,13 +60952,13 @@ _sk_start_pipeline_sse2_lowp:
.byte 73,137,246 // mov %rsi,%r14
.byte 72,137,77,192 // mov %rcx,-0x40(%rbp)
.byte 72,57,203 // cmp %rcx,%rbx
- .byte 15,131,131,0,0,0 // jae 34cfe <_sk_start_pipeline_sse2_lowp+0xb6>
+ .byte 15,131,131,0,0,0 // jae 33f7a <_sk_start_pipeline_sse2_lowp+0xb6>
.byte 72,139,69,208 // mov -0x30(%rbp),%rax
.byte 72,141,64,8 // lea 0x8(%rax),%rax
.byte 72,137,69,200 // mov %rax,-0x38(%rbp)
.byte 76,57,125,200 // cmp %r15,-0x38(%rbp)
.byte 72,139,85,208 // mov -0x30(%rbp),%rdx
- .byte 119,59 // ja 34ccc <_sk_start_pipeline_sse2_lowp+0x84>
+ .byte 119,59 // ja 33f48 <_sk_start_pipeline_sse2_lowp+0x84>
.byte 76,139,101,208 // mov -0x30(%rbp),%r12
.byte 49,255 // xor %edi,%edi
.byte 15,87,192 // xorps %xmm0,%xmm0
@@ -61593,10 +60977,10 @@ _sk_start_pipeline_sse2_lowp:
.byte 73,131,196,16 // add $0x10,%r12
.byte 77,57,252 // cmp %r15,%r12
.byte 73,137,212 // mov %rdx,%r12
- .byte 118,201 // jbe 34c95 <_sk_start_pipeline_sse2_lowp+0x4d>
+ .byte 118,201 // jbe 33f11 <_sk_start_pipeline_sse2_lowp+0x4d>
.byte 76,137,255 // mov %r15,%rdi
.byte 72,41,215 // sub %rdx,%rdi
- .byte 116,33 // je 34cf5 <_sk_start_pipeline_sse2_lowp+0xad>
+ .byte 116,33 // je 33f71 <_sk_start_pipeline_sse2_lowp+0xad>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -61610,7 +60994,7 @@ _sk_start_pipeline_sse2_lowp:
.byte 65,255,213 // callq *%r13
.byte 72,255,195 // inc %rbx
.byte 72,59,93,192 // cmp -0x40(%rbp),%rbx
- .byte 117,137 // jne 34c87 <_sk_start_pipeline_sse2_lowp+0x3f>
+ .byte 117,137 // jne 33f03 <_sk_start_pipeline_sse2_lowp+0x3f>
.byte 72,131,196,24 // add $0x18,%rsp
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
@@ -61641,7 +61025,7 @@ _sk_seed_shader_sse2_lowp:
.byte 102,15,110,209 // movd %ecx,%xmm2
.byte 102,15,112,210,0 // pshufd $0x0,%xmm2,%xmm2
.byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
- .byte 15,88,21,36,83,0,0 // addps 0x5324(%rip),%xmm2 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,88,21,40,83,0,0 // addps 0x5328(%rip),%xmm2 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 255,224 // jmpq *%rax
@@ -61818,7 +61202,7 @@ HIDDEN _sk_black_color_sse2_lowp
FUNCTION(_sk_black_color_sse2_lowp)
_sk_black_color_sse2_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,105,80,0,0 // movaps 0x5069(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,109,80,0,0 // movaps 0x506d(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
@@ -61829,7 +61213,7 @@ HIDDEN _sk_white_color_sse2_lowp
FUNCTION(_sk_white_color_sse2_lowp)
_sk_white_color_sse2_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,5,85,80,0,0 // movaps 0x5055(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,5,89,80,0,0 // movaps 0x5059(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 15,40,200 // movaps %xmm0,%xmm1
.byte 15,40,208 // movaps %xmm0,%xmm2
.byte 15,40,216 // movaps %xmm0,%xmm3
@@ -61840,10 +61224,10 @@ HIDDEN _sk_set_rgb_sse2_lowp
FUNCTION(_sk_set_rgb_sse2_lowp)
_sk_set_rgb_sse2_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 243,15,16,21,148,71,0,0 // movss 0x4794(%rip),%xmm2 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,15,16,21,148,71,0,0 // movss 0x4794(%rip),%xmm2 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
.byte 243,15,16,0 // movss (%rax),%xmm0
.byte 243,15,89,194 // mulss %xmm2,%xmm0
- .byte 243,68,15,16,5,83,71,0,0 // movss 0x4753(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,16,5,83,71,0,0 // movss 0x4753(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,88,192 // addss %xmm8,%xmm0
.byte 243,68,15,44,192 // cvttss2si %xmm0,%r8d
.byte 102,65,15,110,192 // movd %r8d,%xmm0
@@ -61869,7 +61253,7 @@ HIDDEN _sk_clamp_a_sse2_lowp
.globl _sk_clamp_a_sse2_lowp
FUNCTION(_sk_clamp_a_sse2_lowp)
_sk_clamp_a_sse2_lowp:
- .byte 102,68,15,111,5,59,87,0,0 // movdqa 0x573b(%rip),%xmm8 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,5,63,87,0,0 // movdqa 0x573f(%rip),%xmm8 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,200 // movdqa %xmm0,%xmm9
.byte 102,69,15,239,200 // pxor %xmm8,%xmm9
.byte 102,68,15,111,211 // movdqa %xmm3,%xmm10
@@ -61898,7 +61282,7 @@ HIDDEN _sk_clamp_a_dst_sse2_lowp
.globl _sk_clamp_a_dst_sse2_lowp
FUNCTION(_sk_clamp_a_dst_sse2_lowp)
_sk_clamp_a_dst_sse2_lowp:
- .byte 102,68,15,111,5,197,86,0,0 // movdqa 0x56c5(%rip),%xmm8 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,5,201,86,0,0 // movdqa 0x56c9(%rip),%xmm8 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,204 // movdqa %xmm4,%xmm9
.byte 102,69,15,239,200 // pxor %xmm8,%xmm9
.byte 102,68,15,111,215 // movdqa %xmm7,%xmm10
@@ -61928,7 +61312,7 @@ HIDDEN _sk_premul_sse2_lowp
FUNCTION(_sk_premul_sse2_lowp)
_sk_premul_sse2_lowp:
.byte 102,15,213,195 // pmullw %xmm3,%xmm0
- .byte 102,68,15,111,5,219,78,0,0 // movdqa 0x4edb(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,223,78,0,0 // movdqa 0x4edf(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,203 // pmullw %xmm3,%xmm1
@@ -61945,7 +61329,7 @@ HIDDEN _sk_premul_dst_sse2_lowp
FUNCTION(_sk_premul_dst_sse2_lowp)
_sk_premul_dst_sse2_lowp:
.byte 102,15,213,231 // pmullw %xmm7,%xmm4
- .byte 102,68,15,111,5,164,78,0,0 // movdqa 0x4ea4(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,168,78,0,0 // movdqa 0x4ea8(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,224 // paddw %xmm8,%xmm4
.byte 102,15,113,212,8 // psrlw $0x8,%xmm4
.byte 102,15,213,239 // pmullw %xmm7,%xmm5
@@ -61993,7 +61377,7 @@ HIDDEN _sk_invert_sse2_lowp
.globl _sk_invert_sse2_lowp
FUNCTION(_sk_invert_sse2_lowp)
_sk_invert_sse2_lowp:
- .byte 102,68,15,111,5,66,78,0,0 // movdqa 0x4e42(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,70,78,0,0 // movdqa 0x4e46(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,200 // psubw %xmm0,%xmm9
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -62025,7 +61409,7 @@ FUNCTION(_sk_srcatop_sse2_lowp)
_sk_srcatop_sse2_lowp:
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
.byte 102,15,213,199 // pmullw %xmm7,%xmm0
- .byte 102,15,111,29,230,77,0,0 // movdqa 0x4de6(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,29,234,77,0,0 // movdqa 0x4dea(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,111,203 // movdqa %xmm3,%xmm9
.byte 102,69,15,249,200 // psubw %xmm8,%xmm9
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
@@ -62057,7 +61441,7 @@ FUNCTION(_sk_dstatop_sse2_lowp)
_sk_dstatop_sse2_lowp:
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
.byte 102,68,15,213,195 // pmullw %xmm3,%xmm8
- .byte 102,68,15,111,13,102,77,0,0 // movdqa 0x4d66(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,106,77,0,0 // movdqa 0x4d6a(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,253,193 // paddw %xmm9,%xmm8
.byte 102,68,15,111,213 // movdqa %xmm5,%xmm10
.byte 102,68,15,213,211 // pmullw %xmm3,%xmm10
@@ -62086,7 +61470,7 @@ HIDDEN _sk_srcin_sse2_lowp
FUNCTION(_sk_srcin_sse2_lowp)
_sk_srcin_sse2_lowp:
.byte 102,15,213,199 // pmullw %xmm7,%xmm0
- .byte 102,68,15,111,5,241,76,0,0 // movdqa 0x4cf1(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,245,76,0,0 // movdqa 0x4cf5(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,207 // pmullw %xmm7,%xmm1
@@ -62107,7 +61491,7 @@ FUNCTION(_sk_dstin_sse2_lowp)
_sk_dstin_sse2_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,213,195 // pmullw %xmm3,%xmm0
- .byte 102,68,15,111,5,168,76,0,0 // movdqa 0x4ca8(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,172,76,0,0 // movdqa 0x4cac(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,111,205 // movdqa %xmm5,%xmm1
@@ -62128,7 +61512,7 @@ HIDDEN _sk_srcout_sse2_lowp
.globl _sk_srcout_sse2_lowp
FUNCTION(_sk_srcout_sse2_lowp)
_sk_srcout_sse2_lowp:
- .byte 102,68,15,111,5,95,76,0,0 // movdqa 0x4c5f(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,99,76,0,0 // movdqa 0x4c63(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,207 // psubw %xmm7,%xmm9
.byte 102,65,15,213,193 // pmullw %xmm9,%xmm0
@@ -62151,7 +61535,7 @@ HIDDEN _sk_dstout_sse2_lowp
FUNCTION(_sk_dstout_sse2_lowp)
_sk_dstout_sse2_lowp:
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
- .byte 102,68,15,111,5,8,76,0,0 // movdqa 0x4c08(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,12,76,0,0 // movdqa 0x4c0c(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,249,216 // psubw %xmm0,%xmm3
.byte 102,15,111,195 // movdqa %xmm3,%xmm0
@@ -62176,7 +61560,7 @@ HIDDEN _sk_srcover_sse2_lowp
.globl _sk_srcover_sse2_lowp
FUNCTION(_sk_srcover_sse2_lowp)
_sk_srcover_sse2_lowp:
- .byte 102,68,15,111,13,174,75,0,0 // movdqa 0x4bae(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,178,75,0,0 // movdqa 0x4bb2(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
.byte 102,68,15,249,195 // psubw %xmm3,%xmm8
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -62205,7 +61589,7 @@ HIDDEN _sk_dstover_sse2_lowp
.globl _sk_dstover_sse2_lowp
FUNCTION(_sk_dstover_sse2_lowp)
_sk_dstover_sse2_lowp:
- .byte 102,68,15,111,5,52,75,0,0 // movdqa 0x4b34(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,56,75,0,0 // movdqa 0x4b38(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,68,15,249,207 // psubw %xmm7,%xmm9
.byte 102,65,15,213,193 // pmullw %xmm9,%xmm0
@@ -62232,7 +61616,7 @@ HIDDEN _sk_modulate_sse2_lowp
FUNCTION(_sk_modulate_sse2_lowp)
_sk_modulate_sse2_lowp:
.byte 102,15,213,196 // pmullw %xmm4,%xmm0
- .byte 102,68,15,111,5,205,74,0,0 // movdqa 0x4acd(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,209,74,0,0 // movdqa 0x4ad1(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,192 // paddw %xmm8,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,15,213,205 // pmullw %xmm5,%xmm1
@@ -62251,7 +61635,7 @@ HIDDEN _sk_multiply_sse2_lowp
.globl _sk_multiply_sse2_lowp
FUNCTION(_sk_multiply_sse2_lowp)
_sk_multiply_sse2_lowp:
- .byte 102,68,15,111,13,140,74,0,0 // movdqa 0x4a8c(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,144,74,0,0 // movdqa 0x4a90(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,193 // movdqa %xmm9,%xmm8
.byte 102,68,15,249,195 // psubw %xmm3,%xmm8
.byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
@@ -62296,12 +61680,12 @@ _sk_plus__sse2_lowp:
.byte 102,68,15,111,193 // movdqa %xmm1,%xmm8
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,253,204 // paddw %xmm4,%xmm1
- .byte 102,68,15,111,21,56,82,0,0 // movdqa 0x5238(%rip),%xmm10 # 3a880 <_sk_srcover_bgra_8888_sse2_lowp+0x13f8>
+ .byte 102,68,15,111,21,60,82,0,0 // movdqa 0x523c(%rip),%xmm10 # 39b00 <_sk_srcover_bgra_8888_sse2_lowp+0x13fc>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,217,194 // psubusw %xmm10,%xmm0
.byte 102,69,15,239,228 // pxor %xmm12,%xmm12
.byte 102,65,15,117,196 // pcmpeqw %xmm12,%xmm0
- .byte 102,68,15,111,29,172,73,0,0 // movdqa 0x49ac(%rip),%xmm11 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,29,176,73,0,0 // movdqa 0x49b0(%rip),%xmm11 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,219,200 // pand %xmm0,%xmm1
.byte 102,65,15,223,195 // pandn %xmm11,%xmm0
.byte 102,15,235,193 // por %xmm1,%xmm0
@@ -62340,7 +61724,7 @@ _sk_screen_sse2_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,253,193 // paddw %xmm1,%xmm0
.byte 102,15,213,204 // pmullw %xmm4,%xmm1
- .byte 102,68,15,111,21,12,73,0,0 // movdqa 0x490c(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,21,16,73,0,0 // movdqa 0x4910(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,202 // paddw %xmm10,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -62372,7 +61756,7 @@ HIDDEN _sk_xor__sse2_lowp
FUNCTION(_sk_xor__sse2_lowp)
_sk_xor__sse2_lowp:
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,68,15,111,13,138,72,0,0 // movdqa 0x488a(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,142,72,0,0 // movdqa 0x488e(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,209 // movdqa %xmm9,%xmm10
.byte 102,68,15,249,215 // psubw %xmm7,%xmm10
.byte 102,65,15,213,194 // pmullw %xmm10,%xmm0
@@ -62414,7 +61798,7 @@ _sk_darken_sse2_lowp:
.byte 102,65,15,253,194 // paddw %xmm10,%xmm0
.byte 102,68,15,213,215 // pmullw %xmm7,%xmm10
.byte 102,68,15,213,227 // pmullw %xmm3,%xmm12
- .byte 102,68,15,111,13,69,79,0,0 // movdqa 0x4f45(%rip),%xmm9 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,13,73,79,0,0 // movdqa 0x4f49(%rip),%xmm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,69,15,111,218 // movdqa %xmm10,%xmm11
.byte 102,69,15,239,217 // pxor %xmm9,%xmm11
.byte 102,65,15,111,204 // movdqa %xmm12,%xmm1
@@ -62423,7 +61807,7 @@ _sk_darken_sse2_lowp:
.byte 102,68,15,219,225 // pand %xmm1,%xmm12
.byte 102,65,15,223,202 // pandn %xmm10,%xmm1
.byte 102,65,15,235,204 // por %xmm12,%xmm1
- .byte 102,68,15,111,21,164,71,0,0 // movdqa 0x47a4(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,21,168,71,0,0 // movdqa 0x47a8(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,202 // paddw %xmm10,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -62479,7 +61863,7 @@ _sk_lighten_sse2_lowp:
.byte 102,65,15,253,196 // paddw %xmm12,%xmm0
.byte 102,68,15,213,231 // pmullw %xmm7,%xmm12
.byte 102,68,15,213,211 // pmullw %xmm3,%xmm10
- .byte 102,68,15,111,13,28,78,0,0 // movdqa 0x4e1c(%rip),%xmm9 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,13,32,78,0,0 // movdqa 0x4e20(%rip),%xmm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,69,15,111,220 // movdqa %xmm12,%xmm11
.byte 102,69,15,239,217 // pxor %xmm9,%xmm11
.byte 102,65,15,111,202 // movdqa %xmm10,%xmm1
@@ -62488,7 +61872,7 @@ _sk_lighten_sse2_lowp:
.byte 102,68,15,219,225 // pand %xmm1,%xmm12
.byte 102,65,15,223,202 // pandn %xmm10,%xmm1
.byte 102,65,15,235,204 // por %xmm12,%xmm1
- .byte 102,68,15,111,21,123,70,0,0 // movdqa 0x467b(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,21,127,70,0,0 // movdqa 0x467f(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,202 // paddw %xmm10,%xmm1
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
@@ -62544,7 +61928,7 @@ _sk_difference_sse2_lowp:
.byte 102,65,15,253,196 // paddw %xmm12,%xmm0
.byte 102,68,15,213,231 // pmullw %xmm7,%xmm12
.byte 102,68,15,213,211 // pmullw %xmm3,%xmm10
- .byte 102,68,15,111,13,243,76,0,0 // movdqa 0x4cf3(%rip),%xmm9 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,13,247,76,0,0 // movdqa 0x4cf7(%rip),%xmm9 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,69,15,111,220 // movdqa %xmm12,%xmm11
.byte 102,69,15,239,217 // pxor %xmm9,%xmm11
.byte 102,65,15,111,202 // movdqa %xmm10,%xmm1
@@ -62553,10 +61937,10 @@ _sk_difference_sse2_lowp:
.byte 102,68,15,219,225 // pand %xmm1,%xmm12
.byte 102,65,15,223,202 // pandn %xmm10,%xmm1
.byte 102,65,15,235,204 // por %xmm12,%xmm1
- .byte 102,68,15,111,21,82,69,0,0 // movdqa 0x4552(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,21,86,69,0,0 // movdqa 0x4556(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,202 // paddw %xmm10,%xmm1
.byte 102,15,113,209,7 // psrlw $0x7,%xmm1
- .byte 102,68,15,111,29,159,76,0,0 // movdqa 0x4c9f(%rip),%xmm11 # 3a770 <_sk_srcover_bgra_8888_sse2_lowp+0x12e8>
+ .byte 102,68,15,111,29,163,76,0,0 // movdqa 0x4ca3(%rip),%xmm11 # 399f0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ec>
.byte 102,65,15,219,203 // pand %xmm11,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
.byte 102,68,15,111,229 // movdqa %xmm5,%xmm12
@@ -62611,10 +61995,10 @@ _sk_exclusion_sse2_lowp:
.byte 102,15,111,196 // movdqa %xmm4,%xmm0
.byte 102,15,253,193 // paddw %xmm1,%xmm0
.byte 102,15,213,204 // pmullw %xmm4,%xmm1
- .byte 102,68,15,111,13,79,68,0,0 // movdqa 0x444f(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,83,68,0,0 // movdqa 0x4453(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,201 // paddw %xmm9,%xmm1
.byte 102,15,113,209,7 // psrlw $0x7,%xmm1
- .byte 102,68,15,111,21,156,75,0,0 // movdqa 0x4b9c(%rip),%xmm10 # 3a770 <_sk_srcover_bgra_8888_sse2_lowp+0x12e8>
+ .byte 102,68,15,111,21,160,75,0,0 // movdqa 0x4ba0(%rip),%xmm10 # 399f0 <_sk_srcover_bgra_8888_sse2_lowp+0x12ec>
.byte 102,65,15,219,202 // pand %xmm10,%xmm1
.byte 102,15,249,193 // psubw %xmm1,%xmm0
.byte 102,15,111,205 // movdqa %xmm5,%xmm1
@@ -62650,7 +62034,7 @@ _sk_hardlight_sse2_lowp:
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,68,15,111,194 // movdqa %xmm2,%xmm8
.byte 102,68,15,111,200 // movdqa %xmm0,%xmm9
- .byte 102,15,111,21,171,67,0,0 // movdqa 0x43ab(%rip),%xmm2 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,21,175,67,0,0 // movdqa 0x43af(%rip),%xmm2 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,111,226 // movdqa %xmm2,%xmm12
.byte 102,68,15,249,231 // psubw %xmm7,%xmm12
.byte 102,65,15,111,196 // movdqa %xmm12,%xmm0
@@ -62663,7 +62047,7 @@ _sk_hardlight_sse2_lowp:
.byte 102,68,15,111,243 // movdqa %xmm3,%xmm14
.byte 102,69,15,249,241 // psubw %xmm9,%xmm14
.byte 102,69,15,253,201 // paddw %xmm9,%xmm9
- .byte 102,15,111,21,215,74,0,0 // movdqa 0x4ad7(%rip),%xmm2 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,15,111,21,219,74,0,0 // movdqa 0x4adb(%rip),%xmm2 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,15,111,199 // movdqa %xmm7,%xmm0
.byte 102,68,15,111,215 // movdqa %xmm7,%xmm10
.byte 102,15,111,231 // movdqa %xmm7,%xmm4
@@ -62682,7 +62066,7 @@ _sk_hardlight_sse2_lowp:
.byte 102,68,15,219,232 // pand %xmm0,%xmm13
.byte 102,65,15,223,193 // pandn %xmm9,%xmm0
.byte 102,65,15,235,197 // por %xmm13,%xmm0
- .byte 102,15,253,5,10,67,0,0 // paddw 0x430a(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,253,5,14,67,0,0 // paddw 0x430e(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,199 // paddw %xmm15,%xmm0
.byte 102,69,15,111,204 // movdqa %xmm12,%xmm9
.byte 102,68,15,213,201 // pmullw %xmm1,%xmm9
@@ -62705,7 +62089,7 @@ _sk_hardlight_sse2_lowp:
.byte 102,65,15,219,249 // pand %xmm9,%xmm7
.byte 102,68,15,223,201 // pandn %xmm1,%xmm9
.byte 102,68,15,235,207 // por %xmm7,%xmm9
- .byte 102,68,15,253,13,149,66,0,0 // paddw 0x4295(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,253,13,153,66,0,0 // paddw 0x4299(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,253,205 // paddw %xmm13,%xmm9
.byte 102,69,15,213,224 // pmullw %xmm8,%xmm12
.byte 102,69,15,111,251 // movdqa %xmm11,%xmm15
@@ -62727,7 +62111,7 @@ _sk_hardlight_sse2_lowp:
.byte 102,68,15,219,210 // pand %xmm2,%xmm10
.byte 102,65,15,223,208 // pandn %xmm8,%xmm2
.byte 102,65,15,235,210 // por %xmm10,%xmm2
- .byte 102,68,15,111,5,37,66,0,0 // movdqa 0x4225(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,41,66,0,0 // movdqa 0x4229(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,208 // paddw %xmm8,%xmm2
.byte 102,65,15,253,215 // paddw %xmm15,%xmm2
.byte 102,68,15,213,220 // pmullw %xmm4,%xmm11
@@ -62755,7 +62139,7 @@ _sk_overlay_sse2_lowp:
.byte 102,15,127,124,36,232 // movdqa %xmm7,-0x18(%rsp)
.byte 102,68,15,111,193 // movdqa %xmm1,%xmm8
.byte 102,68,15,111,248 // movdqa %xmm0,%xmm15
- .byte 102,15,111,53,182,65,0,0 // movdqa 0x41b6(%rip),%xmm6 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,53,186,65,0,0 // movdqa 0x41ba(%rip),%xmm6 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,111,222 // movdqa %xmm6,%xmm11
.byte 102,69,15,249,222 // psubw %xmm14,%xmm11
.byte 102,65,15,111,203 // movdqa %xmm11,%xmm1
@@ -62777,7 +62161,7 @@ _sk_overlay_sse2_lowp:
.byte 102,15,253,201 // paddw %xmm1,%xmm1
.byte 102,69,15,111,225 // movdqa %xmm9,%xmm12
.byte 102,68,15,249,225 // psubw %xmm1,%xmm12
- .byte 102,15,111,13,186,72,0,0 // movdqa 0x48ba(%rip),%xmm1 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,15,111,13,190,72,0,0 // movdqa 0x48be(%rip),%xmm1 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,69,15,111,238 // movdqa %xmm14,%xmm13
.byte 102,68,15,239,233 // pxor %xmm1,%xmm13
.byte 102,15,239,249 // pxor %xmm1,%xmm7
@@ -62804,7 +62188,7 @@ _sk_overlay_sse2_lowp:
.byte 102,68,15,213,197 // pmullw %xmm5,%xmm8
.byte 102,15,111,229 // movdqa %xmm5,%xmm4
.byte 102,15,253,228 // paddw %xmm4,%xmm4
- .byte 102,15,111,61,57,72,0,0 // movdqa 0x4839(%rip),%xmm7 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,15,111,61,61,72,0,0 // movdqa 0x483d(%rip),%xmm7 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,15,239,231 // pxor %xmm7,%xmm4
.byte 102,65,15,101,229 // pcmpgtw %xmm13,%xmm4
.byte 102,69,15,253,192 // paddw %xmm8,%xmm8
@@ -62864,27 +62248,27 @@ _sk_load_8888_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 3605d <_sk_load_8888_sse2_lowp+0x44>
+ .byte 119,41 // ja 352d9 <_sk_load_8888_sse2_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,25,1,0,0 // lea 0x119(%rip),%r9 # 3615c <_sk_load_8888_sse2_lowp+0x143>
+ .byte 76,141,13,25,1,0,0 // lea 0x119(%rip),%r9 # 353d8 <_sk_load_8888_sse2_lowp+0x143>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,28,144 // movss (%r8,%rdx,4),%xmm3
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 360c3 <_sk_load_8888_sse2_lowp+0xaa>
+ .byte 235,102 // jmp 3533f <_sk_load_8888_sse2_lowp+0xaa>
.byte 102,65,15,16,28,144 // movupd (%r8,%rdx,4),%xmm3
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 360c3 <_sk_load_8888_sse2_lowp+0xaa>
+ .byte 235,87 // jmp 3533f <_sk_load_8888_sse2_lowp+0xaa>
.byte 102,65,15,110,68,144,8 // movd 0x8(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,216 // movapd %xmm8,%xmm3
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 235,52 // jmp 360c3 <_sk_load_8888_sse2_lowp+0xaa>
+ .byte 235,52 // jmp 3533f <_sk_load_8888_sse2_lowp+0xaa>
.byte 102,65,15,110,68,144,24 // movd 0x18(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 243,65,15,16,68,144,20 // movss 0x14(%r8,%rdx,4),%xmm0
@@ -62901,7 +62285,7 @@ _sk_load_8888_sse2_lowp:
.byte 102,15,114,240,16 // pslld $0x10,%xmm0
.byte 102,15,114,224,16 // psrad $0x10,%xmm0
.byte 102,15,107,200 // packssdw %xmm0,%xmm1
- .byte 102,68,15,111,13,35,63,0,0 // movdqa 0x3f23(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,39,63,0,0 // movdqa 0x3f27(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,193 // pand %xmm9,%xmm0
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
@@ -62958,27 +62342,27 @@ _sk_load_8888_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 361bc <_sk_load_8888_dst_sse2_lowp+0x44>
+ .byte 119,41 // ja 35438 <_sk_load_8888_dst_sse2_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 362b8 <_sk_load_8888_dst_sse2_lowp+0x140>
+ .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 35534 <_sk_load_8888_dst_sse2_lowp+0x140>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,60,144 // movss (%r8,%rdx,4),%xmm7
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 36222 <_sk_load_8888_dst_sse2_lowp+0xaa>
+ .byte 235,102 // jmp 3549e <_sk_load_8888_dst_sse2_lowp+0xaa>
.byte 102,65,15,16,60,144 // movupd (%r8,%rdx,4),%xmm7
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 36222 <_sk_load_8888_dst_sse2_lowp+0xaa>
+ .byte 235,87 // jmp 3549e <_sk_load_8888_dst_sse2_lowp+0xaa>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,248 // movapd %xmm8,%xmm7
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,52 // jmp 36222 <_sk_load_8888_dst_sse2_lowp+0xaa>
+ .byte 235,52 // jmp 3549e <_sk_load_8888_dst_sse2_lowp+0xaa>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 243,65,15,16,100,144,20 // movss 0x14(%r8,%rdx,4),%xmm4
@@ -62995,7 +62379,7 @@ _sk_load_8888_dst_sse2_lowp:
.byte 102,15,114,244,16 // pslld $0x10,%xmm4
.byte 102,15,114,228,16 // psrad $0x10,%xmm4
.byte 102,15,107,236 // packssdw %xmm4,%xmm5
- .byte 102,68,15,111,13,196,61,0,0 // movdqa 0x3dc4(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,200,61,0,0 // movdqa 0x3dc8(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,229 // movdqa %xmm5,%xmm4
.byte 102,65,15,219,225 // pand %xmm9,%xmm4
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
@@ -63063,21 +62447,21 @@ _sk_store_8888_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,27 // ja 36357 <_sk_store_8888_sse2_lowp+0x83>
+ .byte 119,27 // ja 355d3 <_sk_store_8888_sse2_lowp+0x83>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,98,0,0,0 // lea 0x62(%rip),%r9 # 363a8 <_sk_store_8888_sse2_lowp+0xd4>
+ .byte 76,141,13,98,0,0,0 // lea 0x62(%rip),%r9 # 35624 <_sk_store_8888_sse2_lowp+0xd4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,12,144 // movd %xmm9,(%r8,%rdx,4)
- .byte 235,75 // jmp 363a2 <_sk_store_8888_sse2_lowp+0xce>
+ .byte 235,75 // jmp 3561e <_sk_store_8888_sse2_lowp+0xce>
.byte 243,69,15,127,12,144 // movdqu %xmm9,(%r8,%rdx,4)
.byte 243,69,15,127,68,144,16 // movdqu %xmm8,0x10(%r8,%rdx,4)
- .byte 235,60 // jmp 363a2 <_sk_store_8888_sse2_lowp+0xce>
+ .byte 235,60 // jmp 3561e <_sk_store_8888_sse2_lowp+0xce>
.byte 102,69,15,112,193,78 // pshufd $0x4e,%xmm9,%xmm8
.byte 102,69,15,126,68,144,8 // movd %xmm8,0x8(%r8,%rdx,4)
.byte 102,69,15,214,12,144 // movq %xmm9,(%r8,%rdx,4)
- .byte 235,39 // jmp 363a2 <_sk_store_8888_sse2_lowp+0xce>
+ .byte 235,39 // jmp 3561e <_sk_store_8888_sse2_lowp+0xce>
.byte 102,69,15,112,208,78 // pshufd $0x4e,%xmm8,%xmm10
.byte 102,69,15,126,84,144,24 // movd %xmm10,0x18(%r8,%rdx,4)
.byte 102,69,15,112,208,229 // pshufd $0xe5,%xmm8,%xmm10
@@ -63122,27 +62506,27 @@ _sk_load_bgra_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 36408 <_sk_load_bgra_sse2_lowp+0x44>
+ .byte 119,41 // ja 35684 <_sk_load_bgra_sse2_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 36504 <_sk_load_bgra_sse2_lowp+0x140>
+ .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 35780 <_sk_load_bgra_sse2_lowp+0x140>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,28,144 // movss (%r8,%rdx,4),%xmm3
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 3646e <_sk_load_bgra_sse2_lowp+0xaa>
+ .byte 235,102 // jmp 356ea <_sk_load_bgra_sse2_lowp+0xaa>
.byte 102,65,15,16,28,144 // movupd (%r8,%rdx,4),%xmm3
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 3646e <_sk_load_bgra_sse2_lowp+0xaa>
+ .byte 235,87 // jmp 356ea <_sk_load_bgra_sse2_lowp+0xaa>
.byte 102,65,15,110,68,144,8 // movd 0x8(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,216 // movapd %xmm8,%xmm3
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 235,52 // jmp 3646e <_sk_load_bgra_sse2_lowp+0xaa>
+ .byte 235,52 // jmp 356ea <_sk_load_bgra_sse2_lowp+0xaa>
.byte 102,65,15,110,68,144,24 // movd 0x18(%r8,%rdx,4),%xmm0
.byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
.byte 243,65,15,16,68,144,20 // movss 0x14(%r8,%rdx,4),%xmm0
@@ -63159,7 +62543,7 @@ _sk_load_bgra_sse2_lowp:
.byte 102,15,114,240,16 // pslld $0x10,%xmm0
.byte 102,15,114,224,16 // psrad $0x10,%xmm0
.byte 102,15,107,200 // packssdw %xmm0,%xmm1
- .byte 102,68,15,111,13,120,59,0,0 // movdqa 0x3b78(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,124,59,0,0 // movdqa 0x3b7c(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,209 // movdqa %xmm1,%xmm2
.byte 102,65,15,219,209 // pand %xmm9,%xmm2
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
@@ -63212,27 +62596,27 @@ _sk_load_bgra_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,41 // ja 36564 <_sk_load_bgra_dst_sse2_lowp+0x44>
+ .byte 119,41 // ja 357e0 <_sk_load_bgra_dst_sse2_lowp+0x44>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 36660 <_sk_load_bgra_dst_sse2_lowp+0x140>
+ .byte 76,141,13,22,1,0,0 // lea 0x116(%rip),%r9 # 358dc <_sk_load_bgra_dst_sse2_lowp+0x140>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,60,144 // movss (%r8,%rdx,4),%xmm7
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 365ca <_sk_load_bgra_dst_sse2_lowp+0xaa>
+ .byte 235,102 // jmp 35846 <_sk_load_bgra_dst_sse2_lowp+0xaa>
.byte 102,65,15,16,60,144 // movupd (%r8,%rdx,4),%xmm7
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 365ca <_sk_load_bgra_dst_sse2_lowp+0xaa>
+ .byte 235,87 // jmp 35846 <_sk_load_bgra_dst_sse2_lowp+0xaa>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,248 // movapd %xmm8,%xmm7
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,52 // jmp 365ca <_sk_load_bgra_dst_sse2_lowp+0xaa>
+ .byte 235,52 // jmp 35846 <_sk_load_bgra_dst_sse2_lowp+0xaa>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 243,65,15,16,100,144,20 // movss 0x14(%r8,%rdx,4),%xmm4
@@ -63249,7 +62633,7 @@ _sk_load_bgra_dst_sse2_lowp:
.byte 102,15,114,244,16 // pslld $0x10,%xmm4
.byte 102,15,114,228,16 // psrad $0x10,%xmm4
.byte 102,15,107,236 // packssdw %xmm4,%xmm5
- .byte 102,68,15,111,13,28,58,0,0 // movdqa 0x3a1c(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,32,58,0,0 // movdqa 0x3a20(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,245 // movdqa %xmm5,%xmm6
.byte 102,65,15,219,241 // pand %xmm9,%xmm6
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
@@ -63317,21 +62701,21 @@ _sk_store_bgra_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,27 // ja 366ff <_sk_store_bgra_sse2_lowp+0x83>
+ .byte 119,27 // ja 3597b <_sk_store_bgra_sse2_lowp+0x83>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,98,0,0,0 // lea 0x62(%rip),%r9 # 36750 <_sk_store_bgra_sse2_lowp+0xd4>
+ .byte 76,141,13,98,0,0,0 // lea 0x62(%rip),%r9 # 359cc <_sk_store_bgra_sse2_lowp+0xd4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,12,144 // movd %xmm9,(%r8,%rdx,4)
- .byte 235,75 // jmp 3674a <_sk_store_bgra_sse2_lowp+0xce>
+ .byte 235,75 // jmp 359c6 <_sk_store_bgra_sse2_lowp+0xce>
.byte 243,69,15,127,12,144 // movdqu %xmm9,(%r8,%rdx,4)
.byte 243,69,15,127,68,144,16 // movdqu %xmm8,0x10(%r8,%rdx,4)
- .byte 235,60 // jmp 3674a <_sk_store_bgra_sse2_lowp+0xce>
+ .byte 235,60 // jmp 359c6 <_sk_store_bgra_sse2_lowp+0xce>
.byte 102,69,15,112,193,78 // pshufd $0x4e,%xmm9,%xmm8
.byte 102,69,15,126,68,144,8 // movd %xmm8,0x8(%r8,%rdx,4)
.byte 102,69,15,214,12,144 // movq %xmm9,(%r8,%rdx,4)
- .byte 235,39 // jmp 3674a <_sk_store_bgra_sse2_lowp+0xce>
+ .byte 235,39 // jmp 359c6 <_sk_store_bgra_sse2_lowp+0xce>
.byte 102,69,15,112,208,78 // pshufd $0x4e,%xmm8,%xmm10
.byte 102,69,15,126,84,144,24 // movd %xmm10,0x18(%r8,%rdx,4)
.byte 102,69,15,112,208,229 // pshufd $0xe5,%xmm8,%xmm10
@@ -63445,7 +62829,7 @@ _sk_gather_8888_sse2_lowp:
.byte 102,15,114,241,16 // pslld $0x10,%xmm1
.byte 102,15,114,225,16 // psrad $0x10,%xmm1
.byte 102,15,107,200 // packssdw %xmm0,%xmm1
- .byte 102,68,15,111,13,56,55,0,0 // movdqa 0x3738(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,60,55,0,0 // movdqa 0x373c(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,193 // movdqa %xmm1,%xmm0
.byte 102,65,15,219,193 // pand %xmm9,%xmm0
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
@@ -63555,7 +62939,7 @@ _sk_gather_bgra_sse2_lowp:
.byte 102,15,114,241,16 // pslld $0x10,%xmm1
.byte 102,15,114,225,16 // psrad $0x10,%xmm1
.byte 102,15,107,200 // packssdw %xmm0,%xmm1
- .byte 102,68,15,111,13,89,53,0,0 // movdqa 0x3559(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,93,53,0,0 // movdqa 0x355d(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,209 // movdqa %xmm1,%xmm2
.byte 102,65,15,219,209 // pand %xmm9,%xmm2
.byte 102,15,113,209,8 // psrlw $0x8,%xmm1
@@ -63596,23 +62980,23 @@ _sk_load_565_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 36b66 <_sk_load_565_sse2_lowp+0x3c>
+ .byte 119,34 // ja 35de2 <_sk_load_565_sse2_lowp+0x3c>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,190,0,0,0 // lea 0xbe(%rip),%r9 # 36c10 <_sk_load_565_sse2_lowp+0xe6>
+ .byte 76,141,13,190,0,0,0 // lea 0xbe(%rip),%r9 # 35e8c <_sk_load_565_sse2_lowp+0xe6>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,192 // movd %eax,%xmm0
- .byte 235,66 // jmp 36ba8 <_sk_load_565_sse2_lowp+0x7e>
+ .byte 235,66 // jmp 35e24 <_sk_load_565_sse2_lowp+0x7e>
.byte 243,65,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm0
- .byte 235,58 // jmp 36ba8 <_sk_load_565_sse2_lowp+0x7e>
+ .byte 235,58 // jmp 35e24 <_sk_load_565_sse2_lowp+0x7e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm0
.byte 243,65,15,16,12,80 // movss (%r8,%rdx,2),%xmm1
.byte 243,15,16,193 // movss %xmm1,%xmm0
- .byte 235,34 // jmp 36ba8 <_sk_load_565_sse2_lowp+0x7e>
+ .byte 235,34 // jmp 35e24 <_sk_load_565_sse2_lowp+0x7e>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,65,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm0
.byte 102,65,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm0
@@ -63620,11 +63004,11 @@ _sk_load_565_sse2_lowp:
.byte 102,65,15,18,4,80 // movlpd (%r8,%rdx,2),%xmm0
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
- .byte 102,15,219,29,7,60,0,0 // pand 0x3c07(%rip),%xmm3 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,29,11,60,0,0 // pand 0x3c0b(%rip),%xmm3 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,113,209,5 // psrlw $0x5,%xmm1
- .byte 102,15,219,13,6,60,0,0 // pand 0x3c06(%rip),%xmm1 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,21,14,60,0,0 // movdqa 0x3c0e(%rip),%xmm2 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,13,10,60,0,0 // pand 0x3c0a(%rip),%xmm1 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,21,18,60,0,0 // movdqa 0x3c12(%rip),%xmm2 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,15,113,208,13 // psrlw $0xd,%xmm0
.byte 102,15,235,195 // por %xmm3,%xmm0
@@ -63637,7 +63021,7 @@ _sk_load_565_sse2_lowp:
.byte 102,15,113,210,2 // psrlw $0x2,%xmm2
.byte 102,15,235,211 // por %xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,4,52,0,0 // movaps 0x3404(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,8,52,0,0 // movaps 0x3408(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 255,224 // jmpq *%rax
.byte 102,144 // xchg %ax,%ax
.byte 75,255 // rex.WXB (bad)
@@ -63667,23 +63051,23 @@ _sk_load_565_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 36c68 <_sk_load_565_dst_sse2_lowp+0x3c>
+ .byte 119,34 // ja 35ee4 <_sk_load_565_dst_sse2_lowp+0x3c>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,188,0,0,0 // lea 0xbc(%rip),%r9 # 36d10 <_sk_load_565_dst_sse2_lowp+0xe4>
+ .byte 76,141,13,188,0,0,0 // lea 0xbc(%rip),%r9 # 35f8c <_sk_load_565_dst_sse2_lowp+0xe4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,224 // movd %eax,%xmm4
- .byte 235,66 // jmp 36caa <_sk_load_565_dst_sse2_lowp+0x7e>
+ .byte 235,66 // jmp 35f26 <_sk_load_565_dst_sse2_lowp+0x7e>
.byte 243,65,15,111,36,80 // movdqu (%r8,%rdx,2),%xmm4
- .byte 235,58 // jmp 36caa <_sk_load_565_dst_sse2_lowp+0x7e>
+ .byte 235,58 // jmp 35f26 <_sk_load_565_dst_sse2_lowp+0x7e>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,65,15,196,100,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm4
.byte 243,65,15,16,44,80 // movss (%r8,%rdx,2),%xmm5
.byte 243,15,16,229 // movss %xmm5,%xmm4
- .byte 235,34 // jmp 36caa <_sk_load_565_dst_sse2_lowp+0x7e>
+ .byte 235,34 // jmp 35f26 <_sk_load_565_dst_sse2_lowp+0x7e>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,65,15,196,100,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm4
.byte 102,65,15,196,100,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm4
@@ -63691,11 +63075,11 @@ _sk_load_565_dst_sse2_lowp:
.byte 102,65,15,18,36,80 // movlpd (%r8,%rdx,2),%xmm4
.byte 102,15,111,252 // movdqa %xmm4,%xmm7
.byte 102,15,113,215,8 // psrlw $0x8,%xmm7
- .byte 102,15,219,61,5,59,0,0 // pand 0x3b05(%rip),%xmm7 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,61,9,59,0,0 // pand 0x3b09(%rip),%xmm7 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,113,213,5 // psrlw $0x5,%xmm5
- .byte 102,15,219,45,4,59,0,0 // pand 0x3b04(%rip),%xmm5 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,53,12,59,0,0 // movdqa 0x3b0c(%rip),%xmm6 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,45,8,59,0,0 // pand 0x3b08(%rip),%xmm5 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,53,16,59,0,0 // movdqa 0x3b10(%rip),%xmm6 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,244 // pand %xmm4,%xmm6
.byte 102,15,113,212,13 // psrlw $0xd,%xmm4
.byte 102,15,235,231 // por %xmm7,%xmm4
@@ -63708,7 +63092,7 @@ _sk_load_565_dst_sse2_lowp:
.byte 102,15,113,214,2 // psrlw $0x2,%xmm6
.byte 102,15,235,247 // por %xmm7,%xmm6
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,2,51,0,0 // movaps 0x3302(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,61,6,51,0,0 // movaps 0x3306(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 255,224 // jmpq *%rax
.byte 77,255 // rex.WRB (bad)
.byte 255 // (bad)
@@ -63732,10 +63116,10 @@ _sk_store_565_sse2_lowp:
.byte 76,3,0 // add (%rax),%r8
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
.byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,219,5,160,58,0,0 // pand 0x3aa0(%rip),%xmm8 # 3a7f0 <_sk_srcover_bgra_8888_sse2_lowp+0x1368>
+ .byte 102,68,15,219,5,164,58,0,0 // pand 0x3aa4(%rip),%xmm8 # 39a70 <_sk_srcover_bgra_8888_sse2_lowp+0x136c>
.byte 102,68,15,111,201 // movdqa %xmm1,%xmm9
.byte 102,65,15,113,241,3 // psllw $0x3,%xmm9
- .byte 102,68,15,219,13,156,58,0,0 // pand 0x3a9c(%rip),%xmm9 # 3a800 <_sk_srcover_bgra_8888_sse2_lowp+0x1378>
+ .byte 102,68,15,219,13,160,58,0,0 // pand 0x3aa0(%rip),%xmm9 # 39a80 <_sk_srcover_bgra_8888_sse2_lowp+0x137c>
.byte 102,69,15,235,200 // por %xmm8,%xmm9
.byte 102,68,15,111,194 // movdqa %xmm2,%xmm8
.byte 102,65,15,113,208,3 // psrlw $0x3,%xmm8
@@ -63744,21 +63128,21 @@ _sk_store_565_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,31 // ja 36da2 <_sk_store_565_sse2_lowp+0x76>
+ .byte 119,31 // ja 3601e <_sk_store_565_sse2_lowp+0x76>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,95,0,0,0 // lea 0x5f(%rip),%r9 # 36dec <_sk_store_565_sse2_lowp+0xc0>
+ .byte 76,141,13,95,0,0,0 // lea 0x5f(%rip),%r9 # 36068 <_sk_store_565_sse2_lowp+0xc0>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,68,15,126,192 // movd %xmm8,%eax
.byte 102,65,137,4,80 // mov %ax,(%r8,%rdx,2)
- .byte 235,70 // jmp 36de8 <_sk_store_565_sse2_lowp+0xbc>
+ .byte 235,70 // jmp 36064 <_sk_store_565_sse2_lowp+0xbc>
.byte 243,69,15,127,4,80 // movdqu %xmm8,(%r8,%rdx,2)
- .byte 235,62 // jmp 36de8 <_sk_store_565_sse2_lowp+0xbc>
+ .byte 235,62 // jmp 36064 <_sk_store_565_sse2_lowp+0xbc>
.byte 102,65,15,197,192,2 // pextrw $0x2,%xmm8,%eax
.byte 102,65,137,68,80,4 // mov %ax,0x4(%r8,%rdx,2)
.byte 102,69,15,126,4,80 // movd %xmm8,(%r8,%rdx,2)
- .byte 235,42 // jmp 36de8 <_sk_store_565_sse2_lowp+0xbc>
+ .byte 235,42 // jmp 36064 <_sk_store_565_sse2_lowp+0xbc>
.byte 102,65,15,197,192,6 // pextrw $0x6,%xmm8,%eax
.byte 102,65,137,68,80,12 // mov %ax,0xc(%r8,%rdx,2)
.byte 102,65,15,197,192,5 // pextrw $0x5,%xmm8,%eax
@@ -63877,11 +63261,11 @@ _sk_gather_565_sse2_lowp:
.byte 102,15,108,193 // punpcklqdq %xmm1,%xmm0
.byte 102,15,111,216 // movdqa %xmm0,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
- .byte 102,15,219,29,73,56,0,0 // pand 0x3849(%rip),%xmm3 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,15,219,29,77,56,0,0 // pand 0x384d(%rip),%xmm3 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,113,209,5 // psrlw $0x5,%xmm1
- .byte 102,15,219,13,72,56,0,0 // pand 0x3848(%rip),%xmm1 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,15,111,21,80,56,0,0 // movdqa 0x3850(%rip),%xmm2 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,15,219,13,76,56,0,0 // pand 0x384c(%rip),%xmm1 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,15,111,21,84,56,0,0 // movdqa 0x3854(%rip),%xmm2 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,15,113,208,13 // psrlw $0xd,%xmm0
.byte 102,15,235,195 // por %xmm3,%xmm0
@@ -63894,7 +63278,7 @@ _sk_gather_565_sse2_lowp:
.byte 102,15,113,210,2 // psrlw $0x2,%xmm2
.byte 102,15,235,211 // por %xmm3,%xmm2
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,70,48,0,0 // movaps 0x3046(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,74,48,0,0 // movaps 0x304a(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 91 // pop %rbx
.byte 65,92 // pop %r12
.byte 65,94 // pop %r14
@@ -63915,23 +63299,23 @@ _sk_load_4444_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 37012 <_sk_load_4444_sse2_lowp+0x3e>
+ .byte 119,36 // ja 3628e <_sk_load_4444_sse2_lowp+0x3e>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,199,0,0,0 // lea 0xc7(%rip),%r9 # 370c4 <_sk_load_4444_sse2_lowp+0xf0>
+ .byte 76,141,13,199,0,0,0 // lea 0xc7(%rip),%r9 # 36340 <_sk_load_4444_sse2_lowp+0xf0>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,69 // jmp 37057 <_sk_load_4444_sse2_lowp+0x83>
+ .byte 235,69 // jmp 362d3 <_sk_load_4444_sse2_lowp+0x83>
.byte 243,69,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm8
- .byte 235,61 // jmp 37057 <_sk_load_4444_sse2_lowp+0x83>
+ .byte 235,61 // jmp 362d3 <_sk_load_4444_sse2_lowp+0x83>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,69,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm8
.byte 243,65,15,16,4,80 // movss (%r8,%rdx,2),%xmm0
.byte 243,68,15,16,192 // movss %xmm0,%xmm8
- .byte 235,35 // jmp 37057 <_sk_load_4444_sse2_lowp+0x83>
+ .byte 235,35 // jmp 362d3 <_sk_load_4444_sse2_lowp+0x83>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,69,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm8
.byte 102,69,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm8
@@ -63941,7 +63325,7 @@ _sk_load_4444_sse2_lowp:
.byte 102,15,113,209,12 // psrlw $0xc,%xmm1
.byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
.byte 102,15,113,210,8 // psrlw $0x8,%xmm2
- .byte 102,15,111,5,157,55,0,0 // movdqa 0x379d(%rip),%xmm0 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,5,161,55,0,0 // movdqa 0x37a1(%rip),%xmm0 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,113,211,4 // psrlw $0x4,%xmm3
@@ -63971,7 +63355,7 @@ _sk_load_4444_sse2_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 125,255 // jge 370d9 <_sk_load_4444_sse2_lowp+0x105>
+ .byte 125,255 // jge 36355 <_sk_load_4444_sse2_lowp+0x105>
.byte 255 // (bad)
.byte 255,112,255 // pushq -0x1(%rax)
.byte 255 // (bad)
@@ -63990,23 +63374,23 @@ _sk_load_4444_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 3711e <_sk_load_4444_dst_sse2_lowp+0x3e>
+ .byte 119,36 // ja 3639a <_sk_load_4444_dst_sse2_lowp+0x3e>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,199,0,0,0 // lea 0xc7(%rip),%r9 # 371d0 <_sk_load_4444_dst_sse2_lowp+0xf0>
+ .byte 76,141,13,199,0,0,0 // lea 0xc7(%rip),%r9 # 3644c <_sk_load_4444_dst_sse2_lowp+0xf0>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,69 // jmp 37163 <_sk_load_4444_dst_sse2_lowp+0x83>
+ .byte 235,69 // jmp 363df <_sk_load_4444_dst_sse2_lowp+0x83>
.byte 243,69,15,111,4,80 // movdqu (%r8,%rdx,2),%xmm8
- .byte 235,61 // jmp 37163 <_sk_load_4444_dst_sse2_lowp+0x83>
+ .byte 235,61 // jmp 363df <_sk_load_4444_dst_sse2_lowp+0x83>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,69,15,196,68,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm8
.byte 243,65,15,16,36,80 // movss (%r8,%rdx,2),%xmm4
.byte 243,68,15,16,196 // movss %xmm4,%xmm8
- .byte 235,35 // jmp 37163 <_sk_load_4444_dst_sse2_lowp+0x83>
+ .byte 235,35 // jmp 363df <_sk_load_4444_dst_sse2_lowp+0x83>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,69,15,196,68,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm8
.byte 102,69,15,196,68,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm8
@@ -64016,7 +63400,7 @@ _sk_load_4444_dst_sse2_lowp:
.byte 102,15,113,213,12 // psrlw $0xc,%xmm5
.byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
.byte 102,15,113,214,8 // psrlw $0x8,%xmm6
- .byte 102,15,111,37,145,54,0,0 // movdqa 0x3691(%rip),%xmm4 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,37,149,54,0,0 // movdqa 0x3695(%rip),%xmm4 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,244 // pand %xmm4,%xmm6
.byte 102,65,15,111,248 // movdqa %xmm8,%xmm7
.byte 102,15,113,215,4 // psrlw $0x4,%xmm7
@@ -64046,7 +63430,7 @@ _sk_load_4444_dst_sse2_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 125,255 // jge 371e5 <_sk_load_4444_dst_sse2_lowp+0x105>
+ .byte 125,255 // jge 36461 <_sk_load_4444_dst_sse2_lowp+0x105>
.byte 255 // (bad)
.byte 255,112,255 // pushq -0x1(%rax)
.byte 255 // (bad)
@@ -64063,12 +63447,12 @@ _sk_store_4444_sse2_lowp:
.byte 76,3,0 // add (%rax),%r8
.byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
.byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,219,5,16,54,0,0 // pand 0x3610(%rip),%xmm8 # 3a820 <_sk_srcover_bgra_8888_sse2_lowp+0x1398>
+ .byte 102,68,15,219,5,20,54,0,0 // pand 0x3614(%rip),%xmm8 # 39aa0 <_sk_srcover_bgra_8888_sse2_lowp+0x139c>
.byte 102,68,15,111,201 // movdqa %xmm1,%xmm9
.byte 102,65,15,113,241,4 // psllw $0x4,%xmm9
- .byte 102,68,15,219,13,12,54,0,0 // pand 0x360c(%rip),%xmm9 # 3a830 <_sk_srcover_bgra_8888_sse2_lowp+0x13a8>
+ .byte 102,68,15,219,13,16,54,0,0 // pand 0x3610(%rip),%xmm9 # 39ab0 <_sk_srcover_bgra_8888_sse2_lowp+0x13ac>
.byte 102,69,15,235,200 // por %xmm8,%xmm9
- .byte 102,68,15,111,21,14,54,0,0 // movdqa 0x360e(%rip),%xmm10 # 3a840 <_sk_srcover_bgra_8888_sse2_lowp+0x13b8>
+ .byte 102,68,15,111,21,18,54,0,0 // movdqa 0x3612(%rip),%xmm10 # 39ac0 <_sk_srcover_bgra_8888_sse2_lowp+0x13bc>
.byte 102,68,15,219,210 // pand %xmm2,%xmm10
.byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
.byte 102,65,15,113,208,4 // psrlw $0x4,%xmm8
@@ -64078,21 +63462,21 @@ _sk_store_4444_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,31 // ja 37275 <_sk_store_4444_sse2_lowp+0x89>
+ .byte 119,31 // ja 364f1 <_sk_store_4444_sse2_lowp+0x89>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,96,0,0,0 // lea 0x60(%rip),%r9 # 372c0 <_sk_store_4444_sse2_lowp+0xd4>
+ .byte 76,141,13,96,0,0,0 // lea 0x60(%rip),%r9 # 3653c <_sk_store_4444_sse2_lowp+0xd4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,68,15,126,192 // movd %xmm8,%eax
.byte 102,65,137,4,80 // mov %ax,(%r8,%rdx,2)
- .byte 235,70 // jmp 372bb <_sk_store_4444_sse2_lowp+0xcf>
+ .byte 235,70 // jmp 36537 <_sk_store_4444_sse2_lowp+0xcf>
.byte 243,69,15,127,4,80 // movdqu %xmm8,(%r8,%rdx,2)
- .byte 235,62 // jmp 372bb <_sk_store_4444_sse2_lowp+0xcf>
+ .byte 235,62 // jmp 36537 <_sk_store_4444_sse2_lowp+0xcf>
.byte 102,65,15,197,192,2 // pextrw $0x2,%xmm8,%eax
.byte 102,65,137,68,80,4 // mov %ax,0x4(%r8,%rdx,2)
.byte 102,69,15,126,4,80 // movd %xmm8,(%r8,%rdx,2)
- .byte 235,42 // jmp 372bb <_sk_store_4444_sse2_lowp+0xcf>
+ .byte 235,42 // jmp 36537 <_sk_store_4444_sse2_lowp+0xcf>
.byte 102,65,15,197,192,6 // pextrw $0x6,%xmm8,%eax
.byte 102,65,137,68,80,12 // mov %ax,0xc(%r8,%rdx,2)
.byte 102,65,15,197,192,5 // pextrw $0x5,%xmm8,%eax
@@ -64111,7 +63495,7 @@ _sk_store_4444_sse2_lowp:
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 233,255,255,255,221 // jmpq ffffffffde0372d4 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddffde4c>
+ .byte 233,255,255,255,221 // jmpq ffffffffde036550 <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffddffde4c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,209 // callq *%rcx
@@ -64208,7 +63592,7 @@ _sk_gather_4444_sse2_lowp:
.byte 102,15,113,209,12 // psrlw $0xc,%xmm1
.byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
.byte 102,15,113,210,8 // psrlw $0x8,%xmm2
- .byte 102,15,111,5,182,51,0,0 // movdqa 0x33b6(%rip),%xmm0 # 3a810 <_sk_srcover_bgra_8888_sse2_lowp+0x1388>
+ .byte 102,15,111,5,186,51,0,0 // movdqa 0x33ba(%rip),%xmm0 # 39a90 <_sk_srcover_bgra_8888_sse2_lowp+0x138c>
.byte 102,15,219,208 // pand %xmm0,%xmm2
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 102,15,113,211,4 // psrlw $0x4,%xmm3
@@ -64246,19 +63630,19 @@ _sk_load_a8_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 374ec <_sk_load_a8_sse2_lowp+0x39>
+ .byte 119,34 // ja 36768 <_sk_load_a8_sse2_lowp+0x39>
.byte 102,15,239,219 // pxor %xmm3,%xmm3
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,140,0,0,0 // lea 0x8c(%rip),%r9 # 37564 <_sk_load_a8_sse2_lowp+0xb1>
+ .byte 76,141,13,140,0,0,0 // lea 0x8c(%rip),%r9 # 367e0 <_sk_load_a8_sse2_lowp+0xb1>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,216 // movd %eax,%xmm3
- .byte 235,97 // jmp 3754d <_sk_load_a8_sse2_lowp+0x9a>
+ .byte 235,97 // jmp 367c9 <_sk_load_a8_sse2_lowp+0x9a>
.byte 243,65,15,126,28,16 // movq (%r8,%rdx,1),%xmm3
.byte 102,15,96,216 // punpcklbw %xmm0,%xmm3
- .byte 235,85 // jmp 3754d <_sk_load_a8_sse2_lowp+0x9a>
+ .byte 235,85 // jmp 367c9 <_sk_load_a8_sse2_lowp+0x9a>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,219 // pxor %xmm3,%xmm3
.byte 102,15,196,216,2 // pinsrw $0x2,%eax,%xmm3
@@ -64266,7 +63650,7 @@ _sk_load_a8_sse2_lowp:
.byte 102,15,110,192 // movd %eax,%xmm0
.byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
.byte 243,15,16,216 // movss %xmm0,%xmm3
- .byte 235,51 // jmp 3754d <_sk_load_a8_sse2_lowp+0x9a>
+ .byte 235,51 // jmp 367c9 <_sk_load_a8_sse2_lowp+0x9a>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,219 // pxor %xmm3,%xmm3
.byte 102,15,196,216,6 // pinsrw $0x6,%eax,%xmm3
@@ -64277,14 +63661,14 @@ _sk_load_a8_sse2_lowp:
.byte 102,65,15,110,4,16 // movd (%r8,%rdx,1),%xmm0
.byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
.byte 242,15,16,216 // movsd %xmm0,%xmm3
- .byte 102,15,219,29,187,42,0,0 // pand 0x2abb(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,29,191,42,0,0 // pand 0x2abf(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,192 // xorps %xmm0,%xmm0
.byte 15,87,201 // xorps %xmm1,%xmm1
.byte 15,87,210 // xorps %xmm2,%xmm2
.byte 255,224 // jmpq *%rax
.byte 102,144 // xchg %ax,%ax
- .byte 125,255 // jge 37565 <_sk_load_a8_sse2_lowp+0xb2>
+ .byte 125,255 // jge 367e1 <_sk_load_a8_sse2_lowp+0xb2>
.byte 255 // (bad)
.byte 255,163,255,255,255,148 // jmpq *-0x6b000001(%rbx)
.byte 255 // (bad)
@@ -64315,19 +63699,19 @@ _sk_load_a8_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 375b9 <_sk_load_a8_dst_sse2_lowp+0x39>
+ .byte 119,34 // ja 36835 <_sk_load_a8_dst_sse2_lowp+0x39>
.byte 102,15,239,255 // pxor %xmm7,%xmm7
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,139,0,0,0 // lea 0x8b(%rip),%r9 # 37630 <_sk_load_a8_dst_sse2_lowp+0xb0>
+ .byte 76,141,13,139,0,0,0 // lea 0x8b(%rip),%r9 # 368ac <_sk_load_a8_dst_sse2_lowp+0xb0>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,248 // movd %eax,%xmm7
- .byte 235,97 // jmp 3761a <_sk_load_a8_dst_sse2_lowp+0x9a>
+ .byte 235,97 // jmp 36896 <_sk_load_a8_dst_sse2_lowp+0x9a>
.byte 243,65,15,126,60,16 // movq (%r8,%rdx,1),%xmm7
.byte 102,15,96,248 // punpcklbw %xmm0,%xmm7
- .byte 235,85 // jmp 3761a <_sk_load_a8_dst_sse2_lowp+0x9a>
+ .byte 235,85 // jmp 36896 <_sk_load_a8_dst_sse2_lowp+0x9a>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,255 // pxor %xmm7,%xmm7
.byte 102,15,196,248,2 // pinsrw $0x2,%eax,%xmm7
@@ -64335,7 +63719,7 @@ _sk_load_a8_dst_sse2_lowp:
.byte 102,15,110,224 // movd %eax,%xmm4
.byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
.byte 243,15,16,252 // movss %xmm4,%xmm7
- .byte 235,51 // jmp 3761a <_sk_load_a8_dst_sse2_lowp+0x9a>
+ .byte 235,51 // jmp 36896 <_sk_load_a8_dst_sse2_lowp+0x9a>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,255 // pxor %xmm7,%xmm7
.byte 102,15,196,248,6 // pinsrw $0x6,%eax,%xmm7
@@ -64346,14 +63730,14 @@ _sk_load_a8_dst_sse2_lowp:
.byte 102,65,15,110,36,16 // movd (%r8,%rdx,1),%xmm4
.byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
.byte 242,15,16,252 // movsd %xmm4,%xmm7
- .byte 102,15,219,61,238,41,0,0 // pand 0x29ee(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,61,242,41,0,0 // pand 0x29f2(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 15,87,228 // xorps %xmm4,%xmm4
.byte 15,87,237 // xorps %xmm5,%xmm5
.byte 15,87,246 // xorps %xmm6,%xmm6
.byte 255,224 // jmpq *%rax
.byte 144 // nop
- .byte 126,255 // jle 37631 <_sk_load_a8_dst_sse2_lowp+0xb1>
+ .byte 126,255 // jle 368ad <_sk_load_a8_dst_sse2_lowp+0xb1>
.byte 255 // (bad)
.byte 255,164,255,255,255,149,255 // jmpq *-0x6a0001(%rdi,%rdi,8)
.byte 255 // (bad)
@@ -64383,42 +63767,42 @@ _sk_store_a8_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,29 // ja 37680 <_sk_store_a8_sse2_lowp+0x34>
+ .byte 119,29 // ja 368fc <_sk_store_a8_sse2_lowp+0x34>
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 37700 <_sk_store_a8_sse2_lowp+0xb4>
+ .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 3697c <_sk_store_a8_sse2_lowp+0xb4>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 102,15,126,216 // movd %xmm3,%eax
.byte 65,136,4,16 // mov %al,(%r8,%rdx,1)
- .byte 235,123 // jmp 376fb <_sk_store_a8_sse2_lowp+0xaf>
- .byte 102,68,15,111,5,135,41,0,0 // movdqa 0x2987(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 235,123 // jmp 36977 <_sk_store_a8_sse2_lowp+0xaf>
+ .byte 102,68,15,111,5,139,41,0,0 // movdqa 0x298b(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,219,195 // pand %xmm3,%xmm8
.byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
.byte 102,69,15,214,4,16 // movq %xmm8,(%r8,%rdx,1)
- .byte 235,96 // jmp 376fb <_sk_store_a8_sse2_lowp+0xaf>
+ .byte 235,96 // jmp 36977 <_sk_store_a8_sse2_lowp+0xaf>
.byte 102,15,197,195,2 // pextrw $0x2,%xmm3,%eax
.byte 65,136,68,16,2 // mov %al,0x2(%r8,%rdx,1)
- .byte 102,68,15,111,5,98,41,0,0 // movdqa 0x2962(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,102,41,0,0 // movdqa 0x2966(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,219,195 // pand %xmm3,%xmm8
.byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
.byte 102,68,15,126,192 // movd %xmm8,%eax
.byte 102,65,137,4,16 // mov %ax,(%r8,%rdx,1)
- .byte 235,55 // jmp 376fb <_sk_store_a8_sse2_lowp+0xaf>
+ .byte 235,55 // jmp 36977 <_sk_store_a8_sse2_lowp+0xaf>
.byte 102,15,197,195,6 // pextrw $0x6,%xmm3,%eax
.byte 65,136,68,16,6 // mov %al,0x6(%r8,%rdx,1)
.byte 102,15,197,195,5 // pextrw $0x5,%xmm3,%eax
.byte 65,136,68,16,5 // mov %al,0x5(%r8,%rdx,1)
.byte 102,15,197,195,4 // pextrw $0x4,%xmm3,%eax
.byte 65,136,68,16,4 // mov %al,0x4(%r8,%rdx,1)
- .byte 102,68,15,111,5,37,41,0,0 // movdqa 0x2925(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,41,41,0,0 // movdqa 0x2929(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,68,15,219,195 // pand %xmm3,%xmm8
.byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
.byte 102,69,15,126,4,16 // movd %xmm8,(%r8,%rdx,1)
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 255,224 // jmpq *%rax
.byte 144 // nop
- .byte 118,255 // jbe 37701 <_sk_store_a8_sse2_lowp+0xb5>
+ .byte 118,255 // jbe 3697d <_sk_store_a8_sse2_lowp+0xb5>
.byte 255 // (bad)
.byte 255,165,255,255,255,155 // jmpq *-0x64000001(%rbp)
.byte 255 // (bad)
@@ -64543,19 +63927,19 @@ _sk_load_g8_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 378bd <_sk_load_g8_sse2_lowp+0x39>
+ .byte 119,34 // ja 36b39 <_sk_load_g8_sse2_lowp+0x39>
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 3793c <_sk_load_g8_sse2_lowp+0xb8>
+ .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 36bb8 <_sk_load_g8_sse2_lowp+0xb8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,192 // movd %eax,%xmm0
- .byte 235,97 // jmp 3791e <_sk_load_g8_sse2_lowp+0x9a>
+ .byte 235,97 // jmp 36b9a <_sk_load_g8_sse2_lowp+0x9a>
.byte 243,65,15,126,4,16 // movq (%r8,%rdx,1),%xmm0
.byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
- .byte 235,85 // jmp 3791e <_sk_load_g8_sse2_lowp+0x9a>
+ .byte 235,85 // jmp 36b9a <_sk_load_g8_sse2_lowp+0x9a>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,15,196,192,2 // pinsrw $0x2,%eax,%xmm0
@@ -64563,7 +63947,7 @@ _sk_load_g8_sse2_lowp:
.byte 102,15,110,200 // movd %eax,%xmm1
.byte 102,15,96,200 // punpcklbw %xmm0,%xmm1
.byte 243,15,16,193 // movss %xmm1,%xmm0
- .byte 235,51 // jmp 3791e <_sk_load_g8_sse2_lowp+0x9a>
+ .byte 235,51 // jmp 36b9a <_sk_load_g8_sse2_lowp+0x9a>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,192 // pxor %xmm0,%xmm0
.byte 102,15,196,192,6 // pinsrw $0x6,%eax,%xmm0
@@ -64574,14 +63958,14 @@ _sk_load_g8_sse2_lowp:
.byte 102,65,15,110,12,16 // movd (%r8,%rdx,1),%xmm1
.byte 102,15,96,200 // punpcklbw %xmm0,%xmm1
.byte 242,15,16,193 // movsd %xmm1,%xmm0
- .byte 102,15,219,5,234,38,0,0 // pand 0x26ea(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,5,238,38,0,0 // pand 0x26ee(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,225,38,0,0 // movaps 0x26e1(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,229,38,0,0 // movaps 0x26e5(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,111,208 // movdqa %xmm0,%xmm2
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 118,255 // jbe 3793d <_sk_load_g8_sse2_lowp+0xb9>
+ .byte 118,255 // jbe 36bb9 <_sk_load_g8_sse2_lowp+0xb9>
.byte 255 // (bad)
.byte 255,156,255,255,255,141,255 // lcall *-0x720001(%rdi,%rdi,8)
.byte 255 // (bad)
@@ -64609,19 +63993,19 @@ _sk_load_g8_dst_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 37991 <_sk_load_g8_dst_sse2_lowp+0x39>
+ .byte 119,34 // ja 36c0d <_sk_load_g8_dst_sse2_lowp+0x39>
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 37a10 <_sk_load_g8_dst_sse2_lowp+0xb8>
+ .byte 76,141,13,147,0,0,0 // lea 0x93(%rip),%r9 # 36c8c <_sk_load_g8_dst_sse2_lowp+0xb8>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,15,110,224 // movd %eax,%xmm4
- .byte 235,97 // jmp 379f2 <_sk_load_g8_dst_sse2_lowp+0x9a>
+ .byte 235,97 // jmp 36c6e <_sk_load_g8_dst_sse2_lowp+0x9a>
.byte 243,65,15,126,36,16 // movq (%r8,%rdx,1),%xmm4
.byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
- .byte 235,85 // jmp 379f2 <_sk_load_g8_dst_sse2_lowp+0x9a>
+ .byte 235,85 // jmp 36c6e <_sk_load_g8_dst_sse2_lowp+0x9a>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,15,196,224,2 // pinsrw $0x2,%eax,%xmm4
@@ -64629,7 +64013,7 @@ _sk_load_g8_dst_sse2_lowp:
.byte 102,15,110,232 // movd %eax,%xmm5
.byte 102,15,96,232 // punpcklbw %xmm0,%xmm5
.byte 243,15,16,229 // movss %xmm5,%xmm4
- .byte 235,51 // jmp 379f2 <_sk_load_g8_dst_sse2_lowp+0x9a>
+ .byte 235,51 // jmp 36c6e <_sk_load_g8_dst_sse2_lowp+0x9a>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,15,196,224,6 // pinsrw $0x6,%eax,%xmm4
@@ -64640,14 +64024,14 @@ _sk_load_g8_dst_sse2_lowp:
.byte 102,65,15,110,44,16 // movd (%r8,%rdx,1),%xmm5
.byte 102,15,96,232 // punpcklbw %xmm0,%xmm5
.byte 242,15,16,229 // movsd %xmm5,%xmm4
- .byte 102,15,219,37,22,38,0,0 // pand 0x2616(%rip),%xmm4 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,219,37,26,38,0,0 // pand 0x261a(%rip),%xmm4 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,13,38,0,0 // movaps 0x260d(%rip),%xmm7 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,61,17,38,0,0 // movaps 0x2611(%rip),%xmm7 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,236 // movdqa %xmm4,%xmm5
.byte 102,15,111,244 // movdqa %xmm4,%xmm6
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 118,255 // jbe 37a11 <_sk_load_g8_dst_sse2_lowp+0xb9>
+ .byte 118,255 // jbe 36c8d <_sk_load_g8_dst_sse2_lowp+0xb9>
.byte 255 // (bad)
.byte 255,156,255,255,255,141,255 // lcall *-0x720001(%rdi,%rdi,8)
.byte 255 // (bad)
@@ -64668,10 +64052,10 @@ HIDDEN _sk_luminance_to_alpha_sse2_lowp
FUNCTION(_sk_luminance_to_alpha_sse2_lowp)
_sk_luminance_to_alpha_sse2_lowp:
.byte 102,15,111,218 // movdqa %xmm2,%xmm3
- .byte 102,15,213,5,24,46,0,0 // pmullw 0x2e18(%rip),%xmm0 # 3a850 <_sk_srcover_bgra_8888_sse2_lowp+0x13c8>
- .byte 102,15,213,13,32,46,0,0 // pmullw 0x2e20(%rip),%xmm1 # 3a860 <_sk_srcover_bgra_8888_sse2_lowp+0x13d8>
+ .byte 102,15,213,5,28,46,0,0 // pmullw 0x2e1c(%rip),%xmm0 # 39ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x13cc>
+ .byte 102,15,213,13,36,46,0,0 // pmullw 0x2e24(%rip),%xmm1 # 39ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x13dc>
.byte 102,15,253,200 // paddw %xmm0,%xmm1
- .byte 102,15,213,29,36,46,0,0 // pmullw 0x2e24(%rip),%xmm3 # 3a870 <_sk_srcover_bgra_8888_sse2_lowp+0x13e8>
+ .byte 102,15,213,29,40,46,0,0 // pmullw 0x2e28(%rip),%xmm3 # 39af0 <_sk_srcover_bgra_8888_sse2_lowp+0x13ec>
.byte 102,15,253,217 // paddw %xmm1,%xmm3
.byte 102,15,113,211,8 // psrlw $0x8,%xmm3
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -64764,7 +64148,7 @@ _sk_gather_g8_sse2_lowp:
.byte 102,15,196,197,3 // pinsrw $0x3,%ebp,%xmm0
.byte 102,65,15,96,192 // punpcklbw %xmm8,%xmm0
.byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,82,36,0,0 // movaps 0x2452(%rip),%xmm3 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 15,40,29,86,36,0,0 // movaps 0x2456(%rip),%xmm3 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,200 // movdqa %xmm0,%xmm1
.byte 102,15,111,208 // movdqa %xmm0,%xmm2
.byte 91 // pop %rbx
@@ -64780,14 +64164,14 @@ FUNCTION(_sk_scale_1_float_sse2_lowp)
_sk_scale_1_float_sse2_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,89,5,132,27,0,0 // mulss 0x1b84(%rip),%xmm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 243,68,15,88,5,75,27,0,0 // addss 0x1b4b(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,5,132,27,0,0 // mulss 0x1b84(%rip),%xmm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,68,15,88,5,75,27,0,0 // addss 0x1b4b(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,44,192 // cvttss2si %xmm8,%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
.byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
- .byte 102,68,15,111,13,3,36,0,0 // movdqa 0x2403(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,7,36,0,0 // movdqa 0x2407(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -64808,13 +64192,13 @@ FUNCTION(_sk_lerp_1_float_sse2_lowp)
_sk_lerp_1_float_sse2_lowp:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,89,5,12,27,0,0 // mulss 0x1b0c(%rip),%xmm8 # 39764 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
- .byte 243,68,15,88,5,211,26,0,0 // addss 0x1ad3(%rip),%xmm8 # 39734 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
+ .byte 243,68,15,89,5,12,27,0,0 // mulss 0x1b0c(%rip),%xmm8 # 389e0 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc>
+ .byte 243,68,15,88,5,211,26,0,0 // addss 0x1ad3(%rip),%xmm8 # 389b0 <_sk_srcover_bgra_8888_sse2_lowp+0x2ac>
.byte 243,65,15,44,192 // cvttss2si %xmm8,%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
.byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
.byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
- .byte 102,68,15,111,13,144,35,0,0 // movdqa 0x2390(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,148,35,0,0 // movdqa 0x2394(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -64854,19 +64238,19 @@ _sk_scale_u8_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 37d37 <_sk_scale_u8_sse2_lowp+0x3b>
+ .byte 119,36 // ja 36fb3 <_sk_scale_u8_sse2_lowp+0x3b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,218,0,0,0 // lea 0xda(%rip),%r9 # 37dfc <_sk_scale_u8_sse2_lowp+0x100>
+ .byte 76,141,13,218,0,0,0 // lea 0xda(%rip),%r9 # 37078 <_sk_scale_u8_sse2_lowp+0x100>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,109 // jmp 37da4 <_sk_scale_u8_sse2_lowp+0xa8>
+ .byte 235,109 // jmp 37020 <_sk_scale_u8_sse2_lowp+0xa8>
.byte 243,69,15,126,4,16 // movq (%r8,%rdx,1),%xmm8
.byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
- .byte 235,96 // jmp 37da4 <_sk_scale_u8_sse2_lowp+0xa8>
+ .byte 235,96 // jmp 37020 <_sk_scale_u8_sse2_lowp+0xa8>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,68,15,196,192,2 // pinsrw $0x2,%eax,%xmm8
@@ -64874,7 +64258,7 @@ _sk_scale_u8_sse2_lowp:
.byte 102,68,15,110,200 // movd %eax,%xmm9
.byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
.byte 243,69,15,16,193 // movss %xmm9,%xmm8
- .byte 235,57 // jmp 37da4 <_sk_scale_u8_sse2_lowp+0xa8>
+ .byte 235,57 // jmp 37020 <_sk_scale_u8_sse2_lowp+0xa8>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,68,15,196,192,6 // pinsrw $0x6,%eax,%xmm8
@@ -64885,9 +64269,9 @@ _sk_scale_u8_sse2_lowp:
.byte 102,69,15,110,12,16 // movd (%r8,%rdx,1),%xmm9
.byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
.byte 242,69,15,16,193 // movsd %xmm9,%xmm8
- .byte 102,68,15,219,5,99,34,0,0 // pand 0x2263(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,219,5,103,34,0,0 // pand 0x2267(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,213,192 // pmullw %xmm8,%xmm0
- .byte 102,68,15,111,13,85,34,0,0 // movdqa 0x2255(%rip),%xmm9 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,13,89,34,0,0 // movdqa 0x2259(%rip),%xmm9 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,253,193 // paddw %xmm9,%xmm0
.byte 102,15,113,208,8 // psrlw $0x8,%xmm0
.byte 102,65,15,213,200 // pmullw %xmm8,%xmm1
@@ -64929,19 +64313,19 @@ _sk_lerp_u8_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,36 // ja 37e53 <_sk_lerp_u8_sse2_lowp+0x3b>
+ .byte 119,36 // ja 370cf <_sk_lerp_u8_sse2_lowp+0x3b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,30,1,0,0 // lea 0x11e(%rip),%r9 # 37f5c <_sk_lerp_u8_sse2_lowp+0x144>
+ .byte 76,141,13,30,1,0,0 // lea 0x11e(%rip),%r9 # 371d8 <_sk_lerp_u8_sse2_lowp+0x144>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,182,4,16 // movzbl (%r8,%rdx,1),%eax
.byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,109 // jmp 37ec0 <_sk_lerp_u8_sse2_lowp+0xa8>
+ .byte 235,109 // jmp 3713c <_sk_lerp_u8_sse2_lowp+0xa8>
.byte 243,69,15,126,4,16 // movq (%r8,%rdx,1),%xmm8
.byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
- .byte 235,96 // jmp 37ec0 <_sk_lerp_u8_sse2_lowp+0xa8>
+ .byte 235,96 // jmp 3713c <_sk_lerp_u8_sse2_lowp+0xa8>
.byte 65,15,182,68,16,2 // movzbl 0x2(%r8,%rdx,1),%eax
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,68,15,196,192,2 // pinsrw $0x2,%eax,%xmm8
@@ -64949,7 +64333,7 @@ _sk_lerp_u8_sse2_lowp:
.byte 102,68,15,110,200 // movd %eax,%xmm9
.byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
.byte 243,69,15,16,193 // movss %xmm9,%xmm8
- .byte 235,57 // jmp 37ec0 <_sk_lerp_u8_sse2_lowp+0xa8>
+ .byte 235,57 // jmp 3713c <_sk_lerp_u8_sse2_lowp+0xa8>
.byte 65,15,182,68,16,6 // movzbl 0x6(%r8,%rdx,1),%eax
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
.byte 102,68,15,196,192,6 // pinsrw $0x6,%eax,%xmm8
@@ -64960,8 +64344,8 @@ _sk_lerp_u8_sse2_lowp:
.byte 102,69,15,110,12,16 // movd (%r8,%rdx,1),%xmm9
.byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
.byte 242,69,15,16,193 // movsd %xmm9,%xmm8
- .byte 102,68,15,219,5,71,33,0,0 // pand 0x2147(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
- .byte 102,68,15,111,21,62,33,0,0 // movdqa 0x213e(%rip),%xmm10 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,219,5,75,33,0,0 // pand 0x214b(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
+ .byte 102,68,15,111,21,66,33,0,0 // movdqa 0x2142(%rip),%xmm10 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
.byte 102,69,15,239,202 // pxor %xmm10,%xmm9
.byte 102,69,15,111,217 // movdqa %xmm9,%xmm11
@@ -64991,9 +64375,9 @@ _sk_lerp_u8_sse2_lowp:
.byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
.byte 255,224 // jmpq *%rax
.byte 15,31,0 // nopl (%rax)
- .byte 235,254 // jmp 37f5c <_sk_lerp_u8_sse2_lowp+0x144>
+ .byte 235,254 // jmp 371d8 <_sk_lerp_u8_sse2_lowp+0x144>
.byte 255 // (bad)
- .byte 255,21,255,255,255,4 // callq *0x4ffffff(%rip) # 5037f64 <_sk_srcover_bgra_8888_sse2_lowp+0x4ffeadc>
+ .byte 255,21,255,255,255,4 // callq *0x4ffffff(%rip) # 50371e0 <_sk_srcover_bgra_8888_sse2_lowp+0x4ffeadc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,84,255,255 // callq *-0x1(%rdi,%rdi,8)
@@ -65023,23 +64407,23 @@ _sk_scale_565_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 37fc3 <_sk_scale_565_sse2_lowp+0x4b>
+ .byte 119,34 // ja 3723f <_sk_scale_565_sse2_lowp+0x4b>
.byte 102,15,239,246 // pxor %xmm6,%xmm6
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,229,1,0,0 // lea 0x1e5(%rip),%r9 # 38194 <_sk_scale_565_sse2_lowp+0x21c>
+ .byte 76,141,13,229,1,0,0 // lea 0x1e5(%rip),%r9 # 37410 <_sk_scale_565_sse2_lowp+0x21c>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,240 // movd %eax,%xmm6
- .byte 235,67 // jmp 38006 <_sk_scale_565_sse2_lowp+0x8e>
+ .byte 235,67 // jmp 37282 <_sk_scale_565_sse2_lowp+0x8e>
.byte 243,65,15,111,52,80 // movdqu (%r8,%rdx,2),%xmm6
- .byte 235,59 // jmp 38006 <_sk_scale_565_sse2_lowp+0x8e>
+ .byte 235,59 // jmp 37282 <_sk_scale_565_sse2_lowp+0x8e>
.byte 102,15,239,246 // pxor %xmm6,%xmm6
.byte 102,65,15,196,116,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm6
.byte 243,69,15,16,12,80 // movss (%r8,%rdx,2),%xmm9
.byte 243,65,15,16,241 // movss %xmm9,%xmm6
- .byte 235,34 // jmp 38006 <_sk_scale_565_sse2_lowp+0x8e>
+ .byte 235,34 // jmp 37282 <_sk_scale_565_sse2_lowp+0x8e>
.byte 102,15,239,246 // pxor %xmm6,%xmm6
.byte 102,65,15,196,116,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm6
.byte 102,65,15,196,116,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm6
@@ -65047,11 +64431,11 @@ _sk_scale_565_sse2_lowp:
.byte 102,65,15,18,52,80 // movlpd (%r8,%rdx,2),%xmm6
.byte 102,68,15,111,222 // movdqa %xmm6,%xmm11
.byte 102,65,15,113,211,8 // psrlw $0x8,%xmm11
- .byte 102,68,15,219,29,166,39,0,0 // pand 0x27a6(%rip),%xmm11 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,68,15,219,29,170,39,0,0 // pand 0x27aa(%rip),%xmm11 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,68,15,111,206 // movdqa %xmm6,%xmm9
.byte 102,65,15,113,209,5 // psrlw $0x5,%xmm9
- .byte 102,68,15,219,13,162,39,0,0 // pand 0x27a2(%rip),%xmm9 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,68,15,111,21,169,39,0,0 // movdqa 0x27a9(%rip),%xmm10 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,68,15,219,13,166,39,0,0 // pand 0x27a6(%rip),%xmm9 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,68,15,111,21,173,39,0,0 // movdqa 0x27ad(%rip),%xmm10 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,68,15,219,214 // pand %xmm6,%xmm10
.byte 102,15,113,214,13 // psrlw $0xd,%xmm6
.byte 102,65,15,235,243 // por %xmm11,%xmm6
@@ -65063,7 +64447,7 @@ _sk_scale_565_sse2_lowp:
.byte 102,65,15,113,243,3 // psllw $0x3,%xmm11
.byte 102,65,15,113,210,2 // psrlw $0x2,%xmm10
.byte 102,69,15,235,211 // por %xmm11,%xmm10
- .byte 102,68,15,111,37,5,39,0,0 // movdqa 0x2705(%rip),%xmm12 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,37,9,39,0,0 // movdqa 0x2709(%rip),%xmm12 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,235 // movdqa %xmm3,%xmm13
.byte 102,69,15,239,236 // pxor %xmm12,%xmm13
.byte 102,68,15,111,223 // movdqa %xmm7,%xmm11
@@ -65104,7 +64488,7 @@ _sk_scale_565_sse2_lowp:
.byte 102,68,15,213,201 // pmullw %xmm1,%xmm9
.byte 102,68,15,213,210 // pmullw %xmm2,%xmm10
.byte 102,68,15,213,219 // pmullw %xmm3,%xmm11
- .byte 102,15,111,5,199,30,0,0 // movdqa 0x1ec7(%rip),%xmm0 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,15,111,5,203,30,0,0 // movdqa 0x1ecb(%rip),%xmm0 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,253,240 // paddw %xmm0,%xmm6
.byte 102,68,15,253,200 // paddw %xmm0,%xmm9
.byte 102,68,15,253,208 // paddw %xmm0,%xmm10
@@ -65151,23 +64535,23 @@ _sk_lerp_565_sse2_lowp:
.byte 36,7 // and $0x7,%al
.byte 254,200 // dec %al
.byte 60,6 // cmp $0x6,%al
- .byte 119,34 // ja 381fb <_sk_lerp_565_sse2_lowp+0x4b>
+ .byte 119,34 // ja 37477 <_sk_lerp_565_sse2_lowp+0x4b>
.byte 102,15,239,210 // pxor %xmm2,%xmm2
.byte 15,182,192 // movzbl %al,%eax
- .byte 76,141,13,45,2,0,0 // lea 0x22d(%rip),%r9 # 38414 <_sk_lerp_565_sse2_lowp+0x264>
+ .byte 76,141,13,45,2,0,0 // lea 0x22d(%rip),%r9 # 37690 <_sk_lerp_565_sse2_lowp+0x264>
.byte 73,99,4,129 // movslq (%r9,%rax,4),%rax
.byte 76,1,200 // add %r9,%rax
.byte 255,224 // jmpq *%rax
.byte 65,15,183,4,80 // movzwl (%r8,%rdx,2),%eax
.byte 102,15,110,208 // movd %eax,%xmm2
- .byte 235,67 // jmp 3823e <_sk_lerp_565_sse2_lowp+0x8e>
+ .byte 235,67 // jmp 374ba <_sk_lerp_565_sse2_lowp+0x8e>
.byte 243,65,15,111,20,80 // movdqu (%r8,%rdx,2),%xmm2
- .byte 235,59 // jmp 3823e <_sk_lerp_565_sse2_lowp+0x8e>
+ .byte 235,59 // jmp 374ba <_sk_lerp_565_sse2_lowp+0x8e>
.byte 102,15,239,210 // pxor %xmm2,%xmm2
.byte 102,65,15,196,84,80,4,2 // pinsrw $0x2,0x4(%r8,%rdx,2),%xmm2
.byte 243,69,15,16,12,80 // movss (%r8,%rdx,2),%xmm9
.byte 243,65,15,16,209 // movss %xmm9,%xmm2
- .byte 235,34 // jmp 3823e <_sk_lerp_565_sse2_lowp+0x8e>
+ .byte 235,34 // jmp 374ba <_sk_lerp_565_sse2_lowp+0x8e>
.byte 102,15,239,210 // pxor %xmm2,%xmm2
.byte 102,65,15,196,84,80,12,6 // pinsrw $0x6,0xc(%r8,%rdx,2),%xmm2
.byte 102,65,15,196,84,80,10,5 // pinsrw $0x5,0xa(%r8,%rdx,2),%xmm2
@@ -65175,11 +64559,11 @@ _sk_lerp_565_sse2_lowp:
.byte 102,65,15,18,20,80 // movlpd (%r8,%rdx,2),%xmm2
.byte 102,68,15,111,218 // movdqa %xmm2,%xmm11
.byte 102,65,15,113,211,8 // psrlw $0x8,%xmm11
- .byte 102,68,15,219,29,110,37,0,0 // pand 0x256e(%rip),%xmm11 # 3a7c0 <_sk_srcover_bgra_8888_sse2_lowp+0x1338>
+ .byte 102,68,15,219,29,114,37,0,0 // pand 0x2572(%rip),%xmm11 # 39a40 <_sk_srcover_bgra_8888_sse2_lowp+0x133c>
.byte 102,68,15,111,202 // movdqa %xmm2,%xmm9
.byte 102,65,15,113,209,5 // psrlw $0x5,%xmm9
- .byte 102,68,15,219,13,106,37,0,0 // pand 0x256a(%rip),%xmm9 # 3a7d0 <_sk_srcover_bgra_8888_sse2_lowp+0x1348>
- .byte 102,68,15,111,21,113,37,0,0 // movdqa 0x2571(%rip),%xmm10 # 3a7e0 <_sk_srcover_bgra_8888_sse2_lowp+0x1358>
+ .byte 102,68,15,219,13,110,37,0,0 // pand 0x256e(%rip),%xmm9 # 39a50 <_sk_srcover_bgra_8888_sse2_lowp+0x134c>
+ .byte 102,68,15,111,21,117,37,0,0 // movdqa 0x2575(%rip),%xmm10 # 39a60 <_sk_srcover_bgra_8888_sse2_lowp+0x135c>
.byte 102,68,15,219,210 // pand %xmm2,%xmm10
.byte 102,15,113,210,13 // psrlw $0xd,%xmm2
.byte 102,65,15,235,211 // por %xmm11,%xmm2
@@ -65191,7 +64575,7 @@ _sk_lerp_565_sse2_lowp:
.byte 102,65,15,113,243,3 // psllw $0x3,%xmm11
.byte 102,65,15,113,210,2 // psrlw $0x2,%xmm10
.byte 102,69,15,235,211 // por %xmm11,%xmm10
- .byte 102,68,15,111,37,205,36,0,0 // movdqa 0x24cd(%rip),%xmm12 # 3a780 <_sk_srcover_bgra_8888_sse2_lowp+0x12f8>
+ .byte 102,68,15,111,37,209,36,0,0 // movdqa 0x24d1(%rip),%xmm12 # 39a00 <_sk_srcover_bgra_8888_sse2_lowp+0x12fc>
.byte 102,68,15,111,235 // movdqa %xmm3,%xmm13
.byte 102,69,15,239,236 // pxor %xmm12,%xmm13
.byte 102,68,15,111,223 // movdqa %xmm7,%xmm11
@@ -65230,7 +64614,7 @@ _sk_lerp_565_sse2_lowp:
.byte 102,69,15,235,222 // por %xmm14,%xmm11
.byte 102,15,111,194 // movdqa %xmm2,%xmm0
.byte 102,15,213,209 // pmullw %xmm1,%xmm2
- .byte 102,68,15,111,5,153,28,0,0 // movdqa 0x1c99(%rip),%xmm8 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,5,157,28,0,0 // movdqa 0x1c9d(%rip),%xmm8 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,65,15,239,192 // pxor %xmm8,%xmm0
.byte 102,15,213,196 // pmullw %xmm4,%xmm0
.byte 102,65,15,253,208 // paddw %xmm8,%xmm2
@@ -65292,7 +64676,7 @@ _sk_clamp_x_1_sse2_lowp:
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 65,15,95,200 // maxps %xmm8,%xmm1
.byte 65,15,95,192 // maxps %xmm8,%xmm0
- .byte 68,15,40,5,44,28,0,0 // movaps 0x1c2c(%rip),%xmm8 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,5,48,28,0,0 // movaps 0x1c30(%rip),%xmm8 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 65,15,93,192 // minps %xmm8,%xmm0
.byte 65,15,93,200 // minps %xmm8,%xmm1
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -65310,7 +64694,7 @@ _sk_repeat_x_1_sse2_lowp:
.byte 69,15,194,208,1 // cmpltps %xmm8,%xmm10
.byte 68,15,40,216 // movaps %xmm0,%xmm11
.byte 69,15,194,217,1 // cmpltps %xmm9,%xmm11
- .byte 68,15,40,37,244,27,0,0 // movaps 0x1bf4(%rip),%xmm12 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,37,248,27,0,0 // movaps 0x1bf8(%rip),%xmm12 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,220 // andps %xmm12,%xmm11
.byte 69,15,84,212 // andps %xmm12,%xmm10
.byte 69,15,87,237 // xorps %xmm13,%xmm13
@@ -65329,10 +64713,10 @@ HIDDEN _sk_mirror_x_1_sse2_lowp
.globl _sk_mirror_x_1_sse2_lowp
FUNCTION(_sk_mirror_x_1_sse2_lowp)
_sk_mirror_x_1_sse2_lowp:
- .byte 68,15,40,5,28,28,0,0 // movaps 0x1c1c(%rip),%xmm8 # 3a0d0 <_sk_srcover_bgra_8888_sse2_lowp+0xc48>
+ .byte 68,15,40,5,32,28,0,0 // movaps 0x1c20(%rip),%xmm8 # 39350 <_sk_srcover_bgra_8888_sse2_lowp+0xc4c>
.byte 65,15,88,192 // addps %xmm8,%xmm0
.byte 65,15,88,200 // addps %xmm8,%xmm1
- .byte 68,15,40,21,156,27,0,0 // movaps 0x1b9c(%rip),%xmm10 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,21,160,27,0,0 // movaps 0x1ba0(%rip),%xmm10 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,40,217 // movaps %xmm1,%xmm11
.byte 69,15,89,218 // mulps %xmm10,%xmm11
.byte 68,15,89,208 // mulps %xmm0,%xmm10
@@ -65342,7 +64726,7 @@ _sk_mirror_x_1_sse2_lowp:
.byte 69,15,91,233 // cvtdq2ps %xmm9,%xmm13
.byte 69,15,194,213,1 // cmpltps %xmm13,%xmm10
.byte 69,15,194,220,1 // cmpltps %xmm12,%xmm11
- .byte 68,15,40,13,124,27,0,0 // movaps 0x1b7c(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,128,27,0,0 // movaps 0x1b80(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,84,217 // andps %xmm9,%xmm11
.byte 69,15,84,209 // andps %xmm9,%xmm10
.byte 69,15,87,246 // xorps %xmm14,%xmm14
@@ -65354,7 +64738,7 @@ _sk_mirror_x_1_sse2_lowp:
.byte 65,15,92,204 // subps %xmm12,%xmm1
.byte 65,15,88,200 // addps %xmm8,%xmm1
.byte 65,15,88,192 // addps %xmm8,%xmm0
- .byte 68,15,40,5,104,32,0,0 // movaps 0x2068(%rip),%xmm8 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 68,15,40,5,108,32,0,0 // movaps 0x206c(%rip),%xmm8 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 65,15,84,192 // andps %xmm8,%xmm0
.byte 65,15,84,200 // andps %xmm8,%xmm1
.byte 65,15,95,206 // maxps %xmm14,%xmm1
@@ -65383,7 +64767,7 @@ _sk_gradient_sse2_lowp:
.byte 102,15,239,210 // pxor %xmm2,%xmm2
.byte 73,131,248,2 // cmp $0x2,%r8
.byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 114,56 // jb 385b0 <_sk_gradient_sse2_lowp+0x6c>
+ .byte 114,56 // jb 3782c <_sk_gradient_sse2_lowp+0x6c>
.byte 72,139,88,72 // mov 0x48(%rax),%rbx
.byte 73,255,200 // dec %r8
.byte 72,131,195,4 // add $0x4,%rbx
@@ -65398,7 +64782,7 @@ _sk_gradient_sse2_lowp:
.byte 102,15,250,220 // psubd %xmm4,%xmm3
.byte 72,131,195,4 // add $0x4,%rbx
.byte 73,255,200 // dec %r8
- .byte 117,219 // jne 3858b <_sk_gradient_sse2_lowp+0x47>
+ .byte 117,219 // jne 37807 <_sk_gradient_sse2_lowp+0x47>
.byte 102,15,112,227,78 // pshufd $0x4e,%xmm3,%xmm4
.byte 102,73,15,126,225 // movq %xmm4,%r9
.byte 69,137,200 // mov %r9d,%r8d
@@ -65541,10 +64925,10 @@ _sk_gradient_sse2_lowp:
.byte 15,88,221 // addps %xmm5,%xmm3
.byte 69,15,89,207 // mulps %xmm15,%xmm9
.byte 68,15,88,204 // addps %xmm4,%xmm9
- .byte 15,40,53,227,24,0,0 // movaps 0x18e3(%rip),%xmm6 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,53,231,24,0,0 // movaps 0x18e7(%rip),%xmm6 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 15,89,222 // mulps %xmm6,%xmm3
.byte 68,15,89,206 // mulps %xmm6,%xmm9
- .byte 15,40,37,229,23,0,0 // movaps 0x17e5(%rip),%xmm4 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,40,37,233,23,0,0 // movaps 0x17e9(%rip),%xmm4 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,88,204 // addps %xmm4,%xmm9
.byte 15,88,220 // addps %xmm4,%xmm3
.byte 243,15,91,219 // cvttps2dq %xmm3,%xmm3
@@ -65640,9 +65024,9 @@ _sk_evenly_spaced_gradient_sse2_lowp:
.byte 72,139,24 // mov (%rax),%rbx
.byte 72,139,104,8 // mov 0x8(%rax),%rbp
.byte 72,255,203 // dec %rbx
- .byte 120,7 // js 389f4 <_sk_evenly_spaced_gradient_sse2_lowp+0x34>
+ .byte 120,7 // js 37c70 <_sk_evenly_spaced_gradient_sse2_lowp+0x34>
.byte 243,72,15,42,211 // cvtsi2ss %rbx,%xmm2
- .byte 235,21 // jmp 38a09 <_sk_evenly_spaced_gradient_sse2_lowp+0x49>
+ .byte 235,21 // jmp 37c85 <_sk_evenly_spaced_gradient_sse2_lowp+0x49>
.byte 73,137,216 // mov %rbx,%r8
.byte 73,209,232 // shr %r8
.byte 131,227,1 // and $0x1,%ebx
@@ -65797,10 +65181,10 @@ _sk_evenly_spaced_gradient_sse2_lowp:
.byte 15,40,68,36,176 // movaps -0x50(%rsp),%xmm0
.byte 68,15,89,192 // mulps %xmm0,%xmm8
.byte 68,15,88,193 // addps %xmm1,%xmm8
- .byte 15,40,13,111,20,0,0 // movaps 0x146f(%rip),%xmm1 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,13,115,20,0,0 // movaps 0x1473(%rip),%xmm1 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 15,89,217 // mulps %xmm1,%xmm3
.byte 68,15,89,193 // mulps %xmm1,%xmm8
- .byte 15,40,13,113,19,0,0 // movaps 0x1371(%rip),%xmm1 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,40,13,117,19,0,0 // movaps 0x1375(%rip),%xmm1 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,88,193 // addps %xmm1,%xmm8
.byte 15,88,217 // addps %xmm1,%xmm3
.byte 243,15,91,219 // cvttps2dq %xmm3,%xmm3
@@ -65817,7 +65201,7 @@ _sk_evenly_spaced_gradient_sse2_lowp:
.byte 15,40,240 // movaps %xmm0,%xmm6
.byte 68,15,89,206 // mulps %xmm6,%xmm9
.byte 68,15,88,204 // addps %xmm4,%xmm9
- .byte 15,40,5,19,20,0,0 // movaps 0x1413(%rip),%xmm0 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 15,40,5,23,20,0,0 // movaps 0x1417(%rip),%xmm0 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 15,89,208 // mulps %xmm0,%xmm2
.byte 68,15,89,200 // mulps %xmm0,%xmm9
.byte 68,15,88,201 // addps %xmm1,%xmm9
@@ -65896,10 +65280,10 @@ _sk_evenly_spaced_2_stop_gradient_sse2_lowp:
.byte 15,89,216 // mulps %xmm0,%xmm3
.byte 65,15,88,216 // addps %xmm8,%xmm3
.byte 65,15,88,208 // addps %xmm8,%xmm2
- .byte 68,15,40,21,216,18,0,0 // movaps 0x12d8(%rip),%xmm10 # 3a150 <_sk_srcover_bgra_8888_sse2_lowp+0xcc8>
+ .byte 68,15,40,21,220,18,0,0 // movaps 0x12dc(%rip),%xmm10 # 393d0 <_sk_srcover_bgra_8888_sse2_lowp+0xccc>
.byte 65,15,89,210 // mulps %xmm10,%xmm2
.byte 65,15,89,218 // mulps %xmm10,%xmm3
- .byte 68,15,40,29,216,17,0,0 // movaps 0x11d8(%rip),%xmm11 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 68,15,40,29,220,17,0,0 // movaps 0x11dc(%rip),%xmm11 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 65,15,88,219 // addps %xmm11,%xmm3
.byte 65,15,88,211 // addps %xmm11,%xmm2
.byte 243,15,91,210 // cvttps2dq %xmm2,%xmm2
@@ -65992,7 +65376,7 @@ _sk_xy_to_unit_angle_sse2_lowp:
.byte 15,40,218 // movaps %xmm2,%xmm3
.byte 15,40,209 // movaps %xmm1,%xmm2
.byte 15,40,200 // movaps %xmm0,%xmm1
- .byte 68,15,40,37,140,21,0,0 // movaps 0x158c(%rip),%xmm12 # 3a590 <_sk_srcover_bgra_8888_sse2_lowp+0x1108>
+ .byte 68,15,40,37,144,21,0,0 // movaps 0x1590(%rip),%xmm12 # 39810 <_sk_srcover_bgra_8888_sse2_lowp+0x110c>
.byte 68,15,40,194 // movaps %xmm2,%xmm8
.byte 69,15,84,196 // andps %xmm12,%xmm8
.byte 68,15,40,217 // movaps %xmm1,%xmm11
@@ -66026,26 +65410,26 @@ _sk_xy_to_unit_angle_sse2_lowp:
.byte 69,15,89,246 // mulps %xmm14,%xmm14
.byte 65,15,40,195 // movaps %xmm11,%xmm0
.byte 15,89,192 // mulps %xmm0,%xmm0
- .byte 68,15,40,37,111,21,0,0 // movaps 0x156f(%rip),%xmm12 # 3a600 <_sk_srcover_bgra_8888_sse2_lowp+0x1178>
+ .byte 68,15,40,37,115,21,0,0 // movaps 0x1573(%rip),%xmm12 # 39880 <_sk_srcover_bgra_8888_sse2_lowp+0x117c>
.byte 68,15,40,232 // movaps %xmm0,%xmm13
.byte 69,15,89,236 // mulps %xmm12,%xmm13
.byte 69,15,89,230 // mulps %xmm14,%xmm12
- .byte 68,15,40,61,107,21,0,0 // movaps 0x156b(%rip),%xmm15 # 3a610 <_sk_srcover_bgra_8888_sse2_lowp+0x1188>
+ .byte 68,15,40,61,111,21,0,0 // movaps 0x156f(%rip),%xmm15 # 39890 <_sk_srcover_bgra_8888_sse2_lowp+0x118c>
.byte 69,15,88,231 // addps %xmm15,%xmm12
.byte 69,15,88,239 // addps %xmm15,%xmm13
.byte 68,15,89,232 // mulps %xmm0,%xmm13
.byte 69,15,89,230 // mulps %xmm14,%xmm12
- .byte 68,15,40,61,99,21,0,0 // movaps 0x1563(%rip),%xmm15 # 3a620 <_sk_srcover_bgra_8888_sse2_lowp+0x1198>
+ .byte 68,15,40,61,103,21,0,0 // movaps 0x1567(%rip),%xmm15 # 398a0 <_sk_srcover_bgra_8888_sse2_lowp+0x119c>
.byte 69,15,88,231 // addps %xmm15,%xmm12
.byte 69,15,88,239 // addps %xmm15,%xmm13
.byte 68,15,89,232 // mulps %xmm0,%xmm13
.byte 69,15,89,230 // mulps %xmm14,%xmm12
- .byte 15,40,5,92,21,0,0 // movaps 0x155c(%rip),%xmm0 # 3a630 <_sk_srcover_bgra_8888_sse2_lowp+0x11a8>
+ .byte 15,40,5,96,21,0,0 // movaps 0x1560(%rip),%xmm0 # 398b0 <_sk_srcover_bgra_8888_sse2_lowp+0x11ac>
.byte 68,15,88,224 // addps %xmm0,%xmm12
.byte 68,15,88,232 // addps %xmm0,%xmm13
.byte 69,15,89,235 // mulps %xmm11,%xmm13
.byte 69,15,89,224 // mulps %xmm8,%xmm12
- .byte 15,40,5,85,21,0,0 // movaps 0x1555(%rip),%xmm0 # 3a640 <_sk_srcover_bgra_8888_sse2_lowp+0x11b8>
+ .byte 15,40,5,89,21,0,0 // movaps 0x1559(%rip),%xmm0 # 398c0 <_sk_srcover_bgra_8888_sse2_lowp+0x11bc>
.byte 68,15,40,192 // movaps %xmm0,%xmm8
.byte 65,15,92,197 // subps %xmm13,%xmm0
.byte 65,15,84,194 // andps %xmm10,%xmm0
@@ -66057,7 +65441,7 @@ _sk_xy_to_unit_angle_sse2_lowp:
.byte 69,15,86,200 // orps %xmm8,%xmm9
.byte 69,15,87,192 // xorps %xmm8,%xmm8
.byte 65,15,194,200,1 // cmpltps %xmm8,%xmm1
- .byte 15,40,5,65,15,0,0 // movaps 0xf41(%rip),%xmm0 # 3a060 <_sk_srcover_bgra_8888_sse2_lowp+0xbd8>
+ .byte 15,40,5,69,15,0,0 // movaps 0xf45(%rip),%xmm0 # 392e0 <_sk_srcover_bgra_8888_sse2_lowp+0xbdc>
.byte 68,15,40,216 // movaps %xmm0,%xmm11
.byte 65,15,92,194 // subps %xmm10,%xmm0
.byte 15,84,193 // andps %xmm1,%xmm0
@@ -66070,7 +65454,7 @@ _sk_xy_to_unit_angle_sse2_lowp:
.byte 65,15,86,211 // orps %xmm11,%xmm2
.byte 68,15,40,219 // movaps %xmm3,%xmm11
.byte 69,15,194,216,1 // cmpltps %xmm8,%xmm11
- .byte 68,15,40,13,25,15,0,0 // movaps 0xf19(%rip),%xmm9 # 3a070 <_sk_srcover_bgra_8888_sse2_lowp+0xbe8>
+ .byte 68,15,40,13,29,15,0,0 // movaps 0xf1d(%rip),%xmm9 # 392f0 <_sk_srcover_bgra_8888_sse2_lowp+0xbec>
.byte 69,15,40,209 // movaps %xmm9,%xmm10
.byte 68,15,92,201 // subps %xmm1,%xmm9
.byte 69,15,84,203 // andps %xmm11,%xmm9
@@ -66128,26 +65512,26 @@ _sk_srcover_rgba_8888_sse2_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,202 // movzbl %r10b,%r9d
.byte 65,128,249,6 // cmp $0x6,%r9b
- .byte 119,38 // ja 39227 <_sk_srcover_rgba_8888_sse2_lowp+0x4b>
+ .byte 119,38 // ja 384a3 <_sk_srcover_rgba_8888_sse2_lowp+0x4b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 76,141,29,67,2,0,0 // lea 0x243(%rip),%r11 # 39450 <_sk_srcover_rgba_8888_sse2_lowp+0x274>
+ .byte 76,141,29,67,2,0,0 // lea 0x243(%rip),%r11 # 386cc <_sk_srcover_rgba_8888_sse2_lowp+0x274>
.byte 75,99,4,139 // movslq (%r11,%r9,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,60,144 // movss (%r8,%rdx,4),%xmm7
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 3928d <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
+ .byte 235,102 // jmp 38509 <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
.byte 102,65,15,16,60,144 // movupd (%r8,%rdx,4),%xmm7
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 3928d <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
+ .byte 235,87 // jmp 38509 <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,248 // movapd %xmm8,%xmm7
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,52 // jmp 3928d <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
+ .byte 235,52 // jmp 38509 <_sk_srcover_rgba_8888_sse2_lowp+0xb1>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 243,65,15,16,100,144,20 // movss 0x14(%r8,%rdx,4),%xmm4
@@ -66164,7 +65548,7 @@ _sk_srcover_rgba_8888_sse2_lowp:
.byte 102,15,114,244,16 // pslld $0x10,%xmm4
.byte 102,15,114,228,16 // psrad $0x10,%xmm4
.byte 102,15,107,236 // packssdw %xmm4,%xmm5
- .byte 102,68,15,111,37,89,13,0,0 // movdqa 0xd59(%rip),%xmm12 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,37,93,13,0,0 // movdqa 0xd5d(%rip),%xmm12 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,229 // movdqa %xmm5,%xmm4
.byte 102,65,15,219,228 // pand %xmm12,%xmm4
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
@@ -66222,20 +65606,20 @@ _sk_srcover_rgba_8888_sse2_lowp:
.byte 102,15,105,193 // punpckhwd %xmm1,%xmm0
.byte 102,15,235,194 // por %xmm2,%xmm0
.byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,24 // ja 393ef <_sk_srcover_rgba_8888_sse2_lowp+0x213>
- .byte 76,141,21,142,0,0,0 // lea 0x8e(%rip),%r10 # 3946c <_sk_srcover_rgba_8888_sse2_lowp+0x290>
+ .byte 119,24 // ja 3866b <_sk_srcover_rgba_8888_sse2_lowp+0x213>
+ .byte 76,141,21,142,0,0,0 // lea 0x8e(%rip),%r10 # 386e8 <_sk_srcover_rgba_8888_sse2_lowp+0x290>
.byte 75,99,4,138 // movslq (%r10,%r9,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,36,144 // movd %xmm12,(%r8,%rdx,4)
- .byte 235,73 // jmp 39438 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
+ .byte 235,73 // jmp 386b4 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
.byte 243,69,15,127,36,144 // movdqu %xmm12,(%r8,%rdx,4)
.byte 243,65,15,127,68,144,16 // movdqu %xmm0,0x10(%r8,%rdx,4)
- .byte 235,58 // jmp 39438 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
+ .byte 235,58 // jmp 386b4 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
.byte 102,65,15,112,196,78 // pshufd $0x4e,%xmm12,%xmm0
.byte 102,65,15,126,68,144,8 // movd %xmm0,0x8(%r8,%rdx,4)
.byte 102,69,15,214,36,144 // movq %xmm12,(%r8,%rdx,4)
- .byte 235,37 // jmp 39438 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
+ .byte 235,37 // jmp 386b4 <_sk_srcover_rgba_8888_sse2_lowp+0x25c>
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
.byte 102,65,15,126,76,144,24 // movd %xmm1,0x18(%r8,%rdx,4)
.byte 102,15,112,200,229 // pshufd $0xe5,%xmm0,%xmm1
@@ -66268,7 +65652,7 @@ _sk_srcover_rgba_8888_sse2_lowp:
.byte 254 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 3946d <_sk_srcover_rgba_8888_sse2_lowp+0x291>
+ .byte 123,255 // jnp 386e9 <_sk_srcover_rgba_8888_sse2_lowp+0x291>
.byte 255 // (bad)
.byte 255,159,255,255,255,146 // lcall *-0x6d000001(%rdi)
.byte 255 // (bad)
@@ -66300,26 +65684,26 @@ _sk_srcover_bgra_8888_sse2_lowp:
.byte 65,254,202 // dec %r10b
.byte 69,15,182,202 // movzbl %r10b,%r9d
.byte 65,128,249,6 // cmp $0x6,%r9b
- .byte 119,38 // ja 394d3 <_sk_srcover_bgra_8888_sse2_lowp+0x4b>
+ .byte 119,38 // ja 3874f <_sk_srcover_bgra_8888_sse2_lowp+0x4b>
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 76,141,29,67,2,0,0 // lea 0x243(%rip),%r11 # 396fc <_sk_srcover_bgra_8888_sse2_lowp+0x274>
+ .byte 76,141,29,67,2,0,0 // lea 0x243(%rip),%r11 # 38978 <_sk_srcover_bgra_8888_sse2_lowp+0x274>
.byte 75,99,4,139 // movslq (%r11,%r9,4),%rax
.byte 76,1,216 // add %r11,%rax
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 255,224 // jmpq *%rax
.byte 243,65,15,16,60,144 // movss (%r8,%rdx,4),%xmm7
.byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 235,102 // jmp 39539 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
+ .byte 235,102 // jmp 387b5 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
.byte 102,65,15,16,60,144 // movupd (%r8,%rdx,4),%xmm7
.byte 243,69,15,111,68,144,16 // movdqu 0x10(%r8,%rdx,4),%xmm8
- .byte 235,87 // jmp 39539 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
+ .byte 235,87 // jmp 387b5 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
.byte 102,65,15,110,100,144,8 // movd 0x8(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 102,15,239,228 // pxor %xmm4,%xmm4
.byte 102,69,15,18,4,144 // movlpd (%r8,%rdx,4),%xmm8
.byte 102,65,15,40,248 // movapd %xmm8,%xmm7
.byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 235,52 // jmp 39539 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
+ .byte 235,52 // jmp 387b5 <_sk_srcover_bgra_8888_sse2_lowp+0xb1>
.byte 102,65,15,110,100,144,24 // movd 0x18(%r8,%rdx,4),%xmm4
.byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
.byte 243,65,15,16,100,144,20 // movss 0x14(%r8,%rdx,4),%xmm4
@@ -66336,7 +65720,7 @@ _sk_srcover_bgra_8888_sse2_lowp:
.byte 102,15,114,244,16 // pslld $0x10,%xmm4
.byte 102,15,114,228,16 // psrad $0x10,%xmm4
.byte 102,15,107,236 // packssdw %xmm4,%xmm5
- .byte 102,68,15,111,37,173,10,0,0 // movdqa 0xaad(%rip),%xmm12 # 3a010 <_sk_srcover_bgra_8888_sse2_lowp+0xb88>
+ .byte 102,68,15,111,37,177,10,0,0 // movdqa 0xab1(%rip),%xmm12 # 39290 <_sk_srcover_bgra_8888_sse2_lowp+0xb8c>
.byte 102,15,111,245 // movdqa %xmm5,%xmm6
.byte 102,65,15,219,244 // pand %xmm12,%xmm6
.byte 102,15,113,213,8 // psrlw $0x8,%xmm5
@@ -66394,20 +65778,20 @@ _sk_srcover_bgra_8888_sse2_lowp:
.byte 102,15,105,193 // punpckhwd %xmm1,%xmm0
.byte 102,15,235,194 // por %xmm2,%xmm0
.byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,24 // ja 3969b <_sk_srcover_bgra_8888_sse2_lowp+0x213>
- .byte 76,141,21,142,0,0,0 // lea 0x8e(%rip),%r10 # 39718 <_sk_srcover_bgra_8888_sse2_lowp+0x290>
+ .byte 119,24 // ja 38917 <_sk_srcover_bgra_8888_sse2_lowp+0x213>
+ .byte 76,141,21,142,0,0,0 // lea 0x8e(%rip),%r10 # 38994 <_sk_srcover_bgra_8888_sse2_lowp+0x290>
.byte 75,99,4,138 // movslq (%r10,%r9,4),%rax
.byte 76,1,208 // add %r10,%rax
.byte 255,224 // jmpq *%rax
.byte 102,69,15,126,36,144 // movd %xmm12,(%r8,%rdx,4)
- .byte 235,73 // jmp 396e4 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
+ .byte 235,73 // jmp 38960 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
.byte 243,69,15,127,36,144 // movdqu %xmm12,(%r8,%rdx,4)
.byte 243,65,15,127,68,144,16 // movdqu %xmm0,0x10(%r8,%rdx,4)
- .byte 235,58 // jmp 396e4 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
+ .byte 235,58 // jmp 38960 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
.byte 102,65,15,112,196,78 // pshufd $0x4e,%xmm12,%xmm0
.byte 102,65,15,126,68,144,8 // movd %xmm0,0x8(%r8,%rdx,4)
.byte 102,69,15,214,36,144 // movq %xmm12,(%r8,%rdx,4)
- .byte 235,37 // jmp 396e4 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
+ .byte 235,37 // jmp 38960 <_sk_srcover_bgra_8888_sse2_lowp+0x25c>
.byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
.byte 102,65,15,126,76,144,24 // movd %xmm1,0x18(%r8,%rdx,4)
.byte 102,15,112,200,229 // pshufd $0xe5,%xmm0,%xmm1
@@ -66440,7 +65824,7 @@ _sk_srcover_bgra_8888_sse2_lowp:
.byte 254 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 123,255 // jnp 39719 <_sk_srcover_bgra_8888_sse2_lowp+0x291>
+ .byte 123,255 // jnp 38995 <_sk_srcover_bgra_8888_sse2_lowp+0x291>
.byte 255 // (bad)
.byte 255,159,255,255,255,146 // lcall *-0x6d000001(%rdi)
.byte 255 // (bad)
@@ -66473,14 +65857,14 @@ BALIGN4
.byte 252 // cld
.byte 190,0,0,128,191 // mov $0xbf800000,%esi
.byte 0,0 // add %al,(%rax)
- .byte 224,64 // loopne 39798 <.literal4+0x64>
+ .byte 224,64 // loopne 38a14 <.literal4+0x64>
.byte 154 // (bad)
.byte 153 // cltd
.byte 153 // cltd
.byte 62,61,10,23,63,174 // ds cmp $0xae3f170a,%eax
- .byte 71,225,61 // rex.RXB loope 397a1 <.literal4+0x6d>
+ .byte 71,225,61 // rex.RXB loope 38a1d <.literal4+0x6d>
.byte 0,0 // add %al,(%rax)
- .byte 127,67 // jg 397ab <.literal4+0x77>
+ .byte 127,67 // jg 38a27 <.literal4+0x77>
.byte 0,0 // add %al,(%rax)
.byte 128,127,145,131 // cmpb $0x83,-0x6f(%rdi)
.byte 158 // sahf
@@ -66517,7 +65901,7 @@ BALIGN4
.byte 0,0 // add %al,(%rax)
.byte 8,33 // or %ah,(%rcx)
.byte 132,55 // test %dh,(%rdi)
- .byte 224,7 // loopne 397c5 <.literal4+0x91>
+ .byte 224,7 // loopne 38a41 <.literal4+0x91>
.byte 0,0 // add %al,(%rax)
.byte 33,8 // and %ecx,(%rax)
.byte 2,58 // add (%rdx),%bh
@@ -66528,8 +65912,8 @@ BALIGN4
.byte 0,0 // add %al,(%rax)
.byte 0,52,255 // add %dh,(%rdi,%rdi,8)
.byte 255 // (bad)
- .byte 127,0 // jg 397d8 <.literal4+0xa4>
- .byte 119,115 // ja 3984d <.literal4+0x119>
+ .byte 127,0 // jg 38a54 <.literal4+0xa4>
+ .byte 119,115 // ja 38ac9 <.literal4+0x119>
.byte 248 // clc
.byte 194,117,191 // retq $0xbf75
.byte 191,63,249,68,180 // mov $0xb444f93f,%edi
@@ -66552,12 +65936,12 @@ BALIGN4
.byte 190,80,128,3,62 // mov $0x3e038050,%esi
.byte 31 // (bad)
.byte 215 // xlat %ds:(%rbx)
- .byte 118,63 // jbe 39863 <.literal4+0x12f>
+ .byte 118,63 // jbe 38adf <.literal4+0x12f>
.byte 246,64,83,63 // testb $0x3f,0x53(%rax)
.byte 0,0 // add %al,(%rax)
.byte 248 // clc
.byte 65,0,0 // add %al,(%r8)
- .byte 124,66 // jl 39872 <.literal4+0x13e>
+ .byte 124,66 // jl 38aee <.literal4+0x13e>
.byte 0,240 // add %dh,%al
.byte 0,0 // add %al,(%rax)
.byte 137,136,136,55,0,15 // mov %ecx,0xf003788(%rax)
@@ -66567,9 +65951,9 @@ BALIGN4
.byte 137,136,136,59,15,0 // mov %ecx,0xf3b88(%rax)
.byte 0,0 // add %al,(%rax)
.byte 137,136,136,61,0,0 // mov %ecx,0x3d88(%rax)
- .byte 112,65 // jo 39895 <.literal4+0x161>
+ .byte 112,65 // jo 38b11 <.literal4+0x161>
.byte 0,255 // add %bh,%bh
- .byte 127,71 // jg 3989f <.literal4+0x16b>
+ .byte 127,71 // jg 38b1b <.literal4+0x16b>
.byte 89 // pop %rcx
.byte 23 // (bad)
.byte 55 // (bad)
@@ -66590,18 +65974,18 @@ BALIGN4
.byte 0,0 // add %al,(%rax)
.byte 128,0,0 // addb $0x0,(%rax)
.byte 0,191,0,0,192,191 // add %bh,-0x40400000(%rdi)
- .byte 114,28 // jb 398aa <.literal4+0x176>
+ .byte 114,28 // jb 38b26 <.literal4+0x176>
.byte 199 // (bad)
.byte 62,85 // ds push %rbp
.byte 85 // push %rbp
.byte 149 // xchg %eax,%ebp
.byte 191,0,0,192,63 // mov $0x3fc00000,%edi
.byte 57,142,99,61,114,249 // cmp %ecx,-0x68dc29d(%rsi)
- .byte 127,63 // jg 398df <_sk_srcover_bgra_8888_sse2_lowp+0x457>
+ .byte 127,63 // jg 38b5b <_sk_srcover_bgra_8888_sse2_lowp+0x457>
.byte 3,0 // add (%rax),%eax
.byte 0,0 // add %al,(%rax)
.byte 45,16,17,192,18 // sub $0x12c01110,%eax
- .byte 120,57 // js 398e4 <_sk_srcover_bgra_8888_sse2_lowp+0x45c>
+ .byte 120,57 // js 38b60 <_sk_srcover_bgra_8888_sse2_lowp+0x45c>
.byte 64,32,148,90,62,4,157,30 // and %dl,0x1e9d043e(%rdx,%rbx,2)
.byte 62,0,24 // add %bl,%ds:(%rax)
.byte 161,57,109,165,144,63,252,191,16 // movabs 0x10bffc3f90a56d39,%eax
@@ -66611,7 +65995,7 @@ BALIGN4
.byte 56,255 // cmp %bh,%bh
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,0 // jg 398cd <.literal4+0x199>
+ .byte 127,0 // jg 38b49 <.literal4+0x199>
.byte 0,128,56,0,64,254 // add %al,-0x1bfffc8(%rax)
.byte 255 // .byte 0xff
@@ -66663,16 +66047,16 @@ BALIGN32
.byte 0,1 // add %al,(%rcx)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039948 <_sk_srcover_bgra_8888_sse2_lowp+0xa0004c0>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038bc8 <_sk_srcover_bgra_8888_sse2_lowp+0xa0004c4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039950 <_sk_srcover_bgra_8888_sse2_lowp+0x120004c8>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038bd0 <_sk_srcover_bgra_8888_sse2_lowp+0x120004cc>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039958 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0004d0>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038bd8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0004d4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039960 <_sk_srcover_bgra_8888_sse2_lowp+0x30004d8>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038be0 <_sk_srcover_bgra_8888_sse2_lowp+0x30004dc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66715,16 +66099,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a0399a8 <_sk_srcover_bgra_8888_sse2_lowp+0xa000520>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038c28 <_sk_srcover_bgra_8888_sse2_lowp+0xa000524>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 120399b0 <_sk_srcover_bgra_8888_sse2_lowp+0x12000528>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038c30 <_sk_srcover_bgra_8888_sse2_lowp+0x1200052c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a0399b8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000530>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038c38 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000534>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 30399c0 <_sk_srcover_bgra_8888_sse2_lowp+0x3000538>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038c40 <_sk_srcover_bgra_8888_sse2_lowp+0x300053c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66767,16 +66151,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039a08 <_sk_srcover_bgra_8888_sse2_lowp+0xa000580>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038c88 <_sk_srcover_bgra_8888_sse2_lowp+0xa000584>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039a10 <_sk_srcover_bgra_8888_sse2_lowp+0x12000588>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038c90 <_sk_srcover_bgra_8888_sse2_lowp+0x1200058c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039a18 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000590>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038c98 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000594>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039a20 <_sk_srcover_bgra_8888_sse2_lowp+0x3000598>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038ca0 <_sk_srcover_bgra_8888_sse2_lowp+0x300059c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66819,16 +66203,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039a68 <_sk_srcover_bgra_8888_sse2_lowp+0xa0005e0>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038ce8 <_sk_srcover_bgra_8888_sse2_lowp+0xa0005e4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039a70 <_sk_srcover_bgra_8888_sse2_lowp+0x120005e8>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038cf0 <_sk_srcover_bgra_8888_sse2_lowp+0x120005ec>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039a78 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0005f0>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038cf8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0005f4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039a80 <_sk_srcover_bgra_8888_sse2_lowp+0x30005f8>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038d00 <_sk_srcover_bgra_8888_sse2_lowp+0x30005fc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66871,16 +66255,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039ac8 <_sk_srcover_bgra_8888_sse2_lowp+0xa000640>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038d48 <_sk_srcover_bgra_8888_sse2_lowp+0xa000644>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039ad0 <_sk_srcover_bgra_8888_sse2_lowp+0x12000648>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038d50 <_sk_srcover_bgra_8888_sse2_lowp+0x1200064c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039ad8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000650>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038d58 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000654>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039ae0 <_sk_srcover_bgra_8888_sse2_lowp+0x3000658>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038d60 <_sk_srcover_bgra_8888_sse2_lowp+0x300065c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66923,16 +66307,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039b28 <_sk_srcover_bgra_8888_sse2_lowp+0xa0006a0>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038da8 <_sk_srcover_bgra_8888_sse2_lowp+0xa0006a4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039b30 <_sk_srcover_bgra_8888_sse2_lowp+0x120006a8>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038db0 <_sk_srcover_bgra_8888_sse2_lowp+0x120006ac>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039b38 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0006b0>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038db8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0006b4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039b40 <_sk_srcover_bgra_8888_sse2_lowp+0x30006b8>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038dc0 <_sk_srcover_bgra_8888_sse2_lowp+0x30006bc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -66975,16 +66359,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039b88 <_sk_srcover_bgra_8888_sse2_lowp+0xa000700>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038e08 <_sk_srcover_bgra_8888_sse2_lowp+0xa000704>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039b90 <_sk_srcover_bgra_8888_sse2_lowp+0x12000708>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038e10 <_sk_srcover_bgra_8888_sse2_lowp+0x1200070c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039b98 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000710>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038e18 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000714>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039ba0 <_sk_srcover_bgra_8888_sse2_lowp+0x3000718>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038e20 <_sk_srcover_bgra_8888_sse2_lowp+0x300071c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -67027,16 +66411,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039be8 <_sk_srcover_bgra_8888_sse2_lowp+0xa000760>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038e68 <_sk_srcover_bgra_8888_sse2_lowp+0xa000764>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039bf0 <_sk_srcover_bgra_8888_sse2_lowp+0x12000768>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038e70 <_sk_srcover_bgra_8888_sse2_lowp+0x1200076c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039bf8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000770>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038e78 <_sk_srcover_bgra_8888_sse2_lowp+0x1a000774>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039c00 <_sk_srcover_bgra_8888_sse2_lowp+0x3000778>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038e80 <_sk_srcover_bgra_8888_sse2_lowp+0x300077c>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -67079,16 +66463,16 @@ BALIGN32
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a039c48 <_sk_srcover_bgra_8888_sse2_lowp+0xa0007c0>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a038ec8 <_sk_srcover_bgra_8888_sse2_lowp+0xa0007c4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12039c50 <_sk_srcover_bgra_8888_sse2_lowp+0x120007c8>
+ .byte 255,13,255,255,255,17 // decl 0x11ffffff(%rip) # 12038ed0 <_sk_srcover_bgra_8888_sse2_lowp+0x120007cc>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a039c58 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0007d0>
+ .byte 255,21,255,255,255,25 // callq *0x19ffffff(%rip) # 1a038ed8 <_sk_srcover_bgra_8888_sse2_lowp+0x1a0007d4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3039c60 <_sk_srcover_bgra_8888_sse2_lowp+0x30007d8>
+ .byte 255,29,255,255,255,2 // lcall *0x2ffffff(%rip) # 3038ee0 <_sk_srcover_bgra_8888_sse2_lowp+0x30007dc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -67393,7 +66777,7 @@ BALIGN32
.byte 0,255 // add %bh,%bh
.byte 0,0 // add %al,(%rax)
.byte 248 // clc
- .byte 224,255 // loopne 39eb3 <.const+0x5d3>
+ .byte 224,255 // loopne 39133 <.const+0x5d3>
.byte 248 // clc
.byte 0,63 // add %bh,(%rdi)
.byte 0,31 // add %bl,(%rdi)
@@ -67654,11 +67038,11 @@ BALIGN16
.byte 0,128,191,0,0,128 // add %al,-0x7fffff41(%rax)
.byte 191,0,0,224,64 // mov $0x40e00000,%edi
.byte 0,0 // add %al,(%rax)
- .byte 224,64 // loopne 3a128 <.literal16+0x118>
+ .byte 224,64 // loopne 393a8 <.literal16+0x118>
.byte 0,0 // add %al,(%rax)
- .byte 224,64 // loopne 3a12c <.literal16+0x11c>
+ .byte 224,64 // loopne 393ac <.literal16+0x11c>
.byte 0,0 // add %al,(%rax)
- .byte 224,64 // loopne 3a130 <.literal16+0x120>
+ .byte 224,64 // loopne 393b0 <.literal16+0x120>
.byte 154 // (bad)
.byte 153 // cltd
.byte 153 // cltd
@@ -67678,13 +67062,13 @@ BALIGN16
.byte 10,23 // or (%rdi),%dl
.byte 63 // (bad)
.byte 174 // scas %es:(%rdi),%al
- .byte 71,225,61 // rex.RXB loope 3a151 <.literal16+0x141>
+ .byte 71,225,61 // rex.RXB loope 393d1 <.literal16+0x141>
.byte 174 // scas %es:(%rdi),%al
- .byte 71,225,61 // rex.RXB loope 3a155 <.literal16+0x145>
+ .byte 71,225,61 // rex.RXB loope 393d5 <.literal16+0x145>
.byte 174 // scas %es:(%rdi),%al
- .byte 71,225,61 // rex.RXB loope 3a159 <.literal16+0x149>
+ .byte 71,225,61 // rex.RXB loope 393d9 <.literal16+0x149>
.byte 174 // scas %es:(%rdi),%al
- .byte 71,225,61 // rex.RXB loope 3a15d <.literal16+0x14d>
+ .byte 71,225,61 // rex.RXB loope 393dd <.literal16+0x14d>
.byte 255,0 // incl (%rax)
.byte 0,0 // add %al,(%rax)
.byte 255,0 // incl (%rax)
@@ -67695,10 +67079,10 @@ BALIGN16
.byte 0,0 // add %al,(%rax)
.byte 1,255 // add %edi,%edi
.byte 255 // (bad)
- .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a03a138 <_sk_srcover_bgra_8888_sse2_lowp+0xa000cb0>
+ .byte 255,5,255,255,255,9 // incl 0x9ffffff(%rip) # a0393b8 <_sk_srcover_bgra_8888_sse2_lowp+0xa000cb4>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,13,255,255,255,2 // decl 0x2ffffff(%rip) # 303a140 <_sk_srcover_bgra_8888_sse2_lowp+0x3000cb8>
+ .byte 255,13,255,255,255,2 // decl 0x2ffffff(%rip) # 30393c0 <_sk_srcover_bgra_8888_sse2_lowp+0x3000cbc>
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255,6 // incl (%rsi)
@@ -67713,11 +67097,11 @@ BALIGN16
.byte 255,0 // incl (%rax)
.byte 0,127,67 // add %bh,0x43(%rdi)
.byte 0,0 // add %al,(%rax)
- .byte 127,67 // jg 3a19b <.literal16+0x18b>
+ .byte 127,67 // jg 3941b <.literal16+0x18b>
.byte 0,0 // add %al,(%rax)
- .byte 127,67 // jg 3a19f <.literal16+0x18f>
+ .byte 127,67 // jg 3941f <.literal16+0x18f>
.byte 0,0 // add %al,(%rax)
- .byte 127,67 // jg 3a1a3 <.literal16+0x193>
+ .byte 127,67 // jg 39423 <.literal16+0x193>
.byte 0,0 // add %al,(%rax)
.byte 128,127,0,0 // cmpb $0x0,0x0(%rdi)
.byte 128,127,0,0 // cmpb $0x0,0x0(%rdi)
@@ -67873,13 +67257,13 @@ BALIGN16
.byte 132,55 // test %dh,(%rdi)
.byte 8,33 // or %ah,(%rcx)
.byte 132,55 // test %dh,(%rdi)
- .byte 224,7 // loopne 3a2b9 <.literal16+0x2a9>
+ .byte 224,7 // loopne 39539 <.literal16+0x2a9>
.byte 0,0 // add %al,(%rax)
- .byte 224,7 // loopne 3a2bd <.literal16+0x2ad>
+ .byte 224,7 // loopne 3953d <.literal16+0x2ad>
.byte 0,0 // add %al,(%rax)
- .byte 224,7 // loopne 3a2c1 <.literal16+0x2b1>
+ .byte 224,7 // loopne 39541 <.literal16+0x2b1>
.byte 0,0 // add %al,(%rax)
- .byte 224,7 // loopne 3a2c5 <.literal16+0x2b5>
+ .byte 224,7 // loopne 39545 <.literal16+0x2b5>
.byte 0,0 // add %al,(%rax)
.byte 33,8 // and %ecx,(%rax)
.byte 2,58 // add (%rdx),%bh
@@ -67933,17 +67317,17 @@ BALIGN16
.byte 0,0 // add %al,(%rax)
.byte 52,255 // xor $0xff,%al
.byte 255 // (bad)
- .byte 127,0 // jg 3a334 <.literal16+0x324>
+ .byte 127,0 // jg 395b4 <.literal16+0x324>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,0 // jg 3a338 <.literal16+0x328>
+ .byte 127,0 // jg 395b8 <.literal16+0x328>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,0 // jg 3a33c <.literal16+0x32c>
+ .byte 127,0 // jg 395bc <.literal16+0x32c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,0 // jg 3a340 <.literal16+0x330>
- .byte 119,115 // ja 3a3b5 <.literal16+0x3a5>
+ .byte 127,0 // jg 395c0 <.literal16+0x330>
+ .byte 119,115 // ja 39635 <.literal16+0x3a5>
.byte 248 // clc
.byte 194,119,115 // retq $0x7377
.byte 248 // clc
@@ -67954,7 +67338,7 @@ BALIGN16
.byte 194,117,191 // retq $0xbf75
.byte 191,63,117,191,191 // mov $0xbfbf753f,%edi
.byte 63 // (bad)
- .byte 117,191 // jne 3a319 <.literal16+0x309>
+ .byte 117,191 // jne 39599 <.literal16+0x309>
.byte 191,63,117,191,191 // mov $0xbfbf753f,%edi
.byte 63 // (bad)
.byte 249 // stc
@@ -67966,7 +67350,7 @@ BALIGN16
.byte 249 // stc
.byte 68,180,62 // rex.R mov $0x3e,%spl
.byte 163,233,220,63,163,233,220,63,163 // movabs %eax,0xa33fdce9a33fdce9
- .byte 233,220,63,163,233 // jmpq ffffffffe9a6e35a <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe9a34ed2>
+ .byte 233,220,63,163,233 // jmpq ffffffffe9a6d5da <_sk_srcover_bgra_8888_sse2_lowp+0xffffffffe9a34ed6>
.byte 220,63 // fdivrl (%rdi)
.byte 81 // push %rcx
.byte 140,242 // mov %?,%edx
@@ -68056,16 +67440,16 @@ BALIGN16
.byte 128,3,62 // addb $0x3e,(%rbx)
.byte 31 // (bad)
.byte 215 // xlat %ds:(%rbx)
- .byte 118,63 // jbe 3a4a3 <.literal16+0x493>
+ .byte 118,63 // jbe 39723 <.literal16+0x493>
.byte 31 // (bad)
.byte 215 // xlat %ds:(%rbx)
- .byte 118,63 // jbe 3a4a7 <.literal16+0x497>
+ .byte 118,63 // jbe 39727 <.literal16+0x497>
.byte 31 // (bad)
.byte 215 // xlat %ds:(%rbx)
- .byte 118,63 // jbe 3a4ab <.literal16+0x49b>
+ .byte 118,63 // jbe 3972b <.literal16+0x49b>
.byte 31 // (bad)
.byte 215 // xlat %ds:(%rbx)
- .byte 118,63 // jbe 3a4af <.literal16+0x49f>
+ .byte 118,63 // jbe 3972f <.literal16+0x49f>
.byte 246,64,83,63 // testb $0x3f,0x53(%rax)
.byte 246,64,83,63 // testb $0x3f,0x53(%rax)
.byte 246,64,83,63 // testb $0x3f,0x53(%rax)
@@ -68086,13 +67470,13 @@ BALIGN16
.byte 65,0,0 // add %al,(%r8)
.byte 248 // clc
.byte 65,0,0 // add %al,(%r8)
- .byte 124,66 // jl 3a4e6 <.literal16+0x4d6>
+ .byte 124,66 // jl 39766 <.literal16+0x4d6>
.byte 0,0 // add %al,(%rax)
- .byte 124,66 // jl 3a4ea <.literal16+0x4da>
+ .byte 124,66 // jl 3976a <.literal16+0x4da>
.byte 0,0 // add %al,(%rax)
- .byte 124,66 // jl 3a4ee <.literal16+0x4de>
+ .byte 124,66 // jl 3976e <.literal16+0x4de>
.byte 0,0 // add %al,(%rax)
- .byte 124,66 // jl 3a4f2 <.literal16+0x4e2>
+ .byte 124,66 // jl 39772 <.literal16+0x4e2>
.byte 0,240 // add %dh,%al
.byte 0,0 // add %al,(%rax)
.byte 0,240 // add %dh,%al
@@ -68138,25 +67522,25 @@ BALIGN16
.byte 136,136,61,137,136,136 // mov %cl,-0x777776c3(%rax)
.byte 61,137,136,136,61 // cmp $0x3d888889,%eax
.byte 0,0 // add %al,(%rax)
- .byte 112,65 // jo 3a575 <.literal16+0x565>
+ .byte 112,65 // jo 397f5 <.literal16+0x565>
.byte 0,0 // add %al,(%rax)
- .byte 112,65 // jo 3a579 <.literal16+0x569>
+ .byte 112,65 // jo 397f9 <.literal16+0x569>
.byte 0,0 // add %al,(%rax)
- .byte 112,65 // jo 3a57d <.literal16+0x56d>
+ .byte 112,65 // jo 397fd <.literal16+0x56d>
.byte 0,0 // add %al,(%rax)
- .byte 112,65 // jo 3a581 <.literal16+0x571>
+ .byte 112,65 // jo 39801 <.literal16+0x571>
.byte 0,128,0,0,0,128 // add %al,-0x80000000(%rax)
.byte 0,0 // add %al,(%rax)
.byte 0,128,0,0,0,128 // add %al,-0x80000000(%rax)
.byte 0,0 // add %al,(%rax)
.byte 255 // (bad)
- .byte 127,0 // jg 3a553 <.literal16+0x543>
+ .byte 127,0 // jg 397d3 <.literal16+0x543>
.byte 0,255 // add %bh,%bh
- .byte 127,0 // jg 3a557 <.literal16+0x547>
+ .byte 127,0 // jg 397d7 <.literal16+0x547>
.byte 0,255 // add %bh,%bh
- .byte 127,0 // jg 3a55b <.literal16+0x54b>
+ .byte 127,0 // jg 397db <.literal16+0x54b>
.byte 0,255 // add %bh,%bh
- .byte 127,0 // jg 3a55f <.literal16+0x54f>
+ .byte 127,0 // jg 397df <.literal16+0x54f>
.byte 0,0 // add %al,(%rax)
.byte 4,0 // add $0x0,%al
.byte 0,0 // add %al,(%rax)
@@ -68182,13 +67566,13 @@ BALIGN16
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,255 // jg 3a598 <.literal16+0x588>
+ .byte 127,255 // jg 39818 <.literal16+0x588>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,255 // jg 3a59c <.literal16+0x58c>
+ .byte 127,255 // jg 3981c <.literal16+0x58c>
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 127,0 // jg 3a5a1 <.literal16+0x591>
+ .byte 127,0 // jg 39821 <.literal16+0x591>
.byte 0,128,56,0,0,128 // add %al,-0x7fffffc8(%rax)
.byte 56,0 // cmp %al,(%rax)
.byte 0,128,56,0,0,128 // add %al,-0x7fffffc8(%rax)
@@ -68202,13 +67586,13 @@ BALIGN16
.byte 64,254 // rex (bad)
.byte 255,0 // incl (%rax)
.byte 255 // (bad)
- .byte 127,71 // jg 3a60b <.literal16+0x5fb>
+ .byte 127,71 // jg 3988b <.literal16+0x5fb>
.byte 0,255 // add %bh,%bh
- .byte 127,71 // jg 3a60f <.literal16+0x5ff>
+ .byte 127,71 // jg 3988f <.literal16+0x5ff>
.byte 0,255 // add %bh,%bh
- .byte 127,71 // jg 3a613 <.literal16+0x603>
+ .byte 127,71 // jg 39893 <.literal16+0x603>
.byte 0,255 // add %bh,%bh
- .byte 127,71 // jg 3a617 <.literal16+0x607>
+ .byte 127,71 // jg 39897 <.literal16+0x607>
.byte 208 // (bad)
.byte 179,89 // mov $0x59,%bl
.byte 62,208 // ds (bad)
@@ -68285,13 +67669,13 @@ BALIGN16
.byte 192,191,0,0,192,191,0 // sarb $0x0,-0x40400000(%rdi)
.byte 0,192 // add %al,%al
.byte 191,0,0,192,191 // mov $0xbfc00000,%edi
- .byte 114,28 // jb 3a6ae <.literal16+0x69e>
+ .byte 114,28 // jb 3992e <.literal16+0x69e>
.byte 199 // (bad)
- .byte 62,114,28 // jb,pt 3a6b2 <.literal16+0x6a2>
+ .byte 62,114,28 // jb,pt 39932 <.literal16+0x6a2>
.byte 199 // (bad)
- .byte 62,114,28 // jb,pt 3a6b6 <.literal16+0x6a6>
+ .byte 62,114,28 // jb,pt 39936 <.literal16+0x6a6>
.byte 199 // (bad)
- .byte 62,114,28 // jb,pt 3a6ba <.literal16+0x6aa>
+ .byte 62,114,28 // jb,pt 3993a <.literal16+0x6aa>
.byte 199 // (bad)
.byte 62,85 // ds push %rbp
.byte 85 // push %rbp
@@ -68310,15 +67694,15 @@ BALIGN16
.byte 0,192 // add %al,%al
.byte 63 // (bad)
.byte 57,142,99,61,57,142 // cmp %ecx,-0x71c6c29d(%rsi)
- .byte 99,61,57,142,99,61 // movslq 0x3d638e39(%rip),%edi # 3d673505 <_sk_srcover_bgra_8888_sse2_lowp+0x3d63a07d>
+ .byte 99,61,57,142,99,61 // movslq 0x3d638e39(%rip),%edi # 3d672785 <_sk_srcover_bgra_8888_sse2_lowp+0x3d63a081>
.byte 57,142,99,61,114,249 // cmp %ecx,-0x68dc29d(%rsi)
- .byte 127,63 // jg 3a713 <.literal16+0x703>
- .byte 114,249 // jb 3a6cf <.literal16+0x6bf>
- .byte 127,63 // jg 3a717 <.literal16+0x707>
- .byte 114,249 // jb 3a6d3 <.literal16+0x6c3>
- .byte 127,63 // jg 3a71b <.literal16+0x70b>
- .byte 114,249 // jb 3a6d7 <.literal16+0x6c7>
- .byte 127,63 // jg 3a71f <.literal16+0x70f>
+ .byte 127,63 // jg 39993 <.literal16+0x703>
+ .byte 114,249 // jb 3994f <.literal16+0x6bf>
+ .byte 127,63 // jg 39997 <.literal16+0x707>
+ .byte 114,249 // jb 39953 <.literal16+0x6c3>
+ .byte 127,63 // jg 3999b <.literal16+0x70b>
+ .byte 114,249 // jb 39957 <.literal16+0x6c7>
+ .byte 127,63 // jg 3999f <.literal16+0x70f>
.byte 3,0 // add (%rax),%eax
.byte 0,0 // add %al,(%rax)
.byte 3,0 // add (%rax),%eax
@@ -68342,9 +67726,9 @@ BALIGN16
.byte 255 // (bad)
.byte 255 // (bad)
.byte 255 // (bad)
- .byte 255,45,16,17,192,45 // ljmp *0x2dc01110(%rip) # 2dc3b815 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc0238d>
+ .byte 255,45,16,17,192,45 // ljmp *0x2dc01110(%rip) # 2dc3aa95 <_sk_srcover_bgra_8888_sse2_lowp+0x2dc02391>
.byte 16,17 // adc %dl,(%rcx)
- .byte 192,45,16,17,192,45,16 // shrb $0x10,0x2dc01110(%rip) # 2dc3b81e <_sk_srcover_bgra_8888_sse2_lowp+0x2dc02396>
+ .byte 192,45,16,17,192,45,16 // shrb $0x10,0x2dc01110(%rip) # 2dc3aa9e <_sk_srcover_bgra_8888_sse2_lowp+0x2dc0239a>
.byte 17,192 // adc %eax,%eax
.byte 18,120,57 // adc 0x39(%rax),%bh
.byte 64,18,120,57 // adc 0x39(%rax),%dil
@@ -68448,14 +67832,14 @@ BALIGN16
.byte 0,248 // add %bh,%al
.byte 0,248 // add %bh,%al
.byte 0,248 // add %bh,%al
- .byte 224,255 // loopne 3a801 <.literal16+0x7f1>
- .byte 224,255 // loopne 3a803 <.literal16+0x7f3>
- .byte 224,255 // loopne 3a805 <.literal16+0x7f5>
- .byte 224,255 // loopne 3a807 <.literal16+0x7f7>
- .byte 224,255 // loopne 3a809 <.literal16+0x7f9>
- .byte 224,255 // loopne 3a80b <.literal16+0x7fb>
- .byte 224,255 // loopne 3a80d <.literal16+0x7fd>
- .byte 224,255 // loopne 3a80f <.literal16+0x7ff>
+ .byte 224,255 // loopne 39a81 <.literal16+0x7f1>
+ .byte 224,255 // loopne 39a83 <.literal16+0x7f3>
+ .byte 224,255 // loopne 39a85 <.literal16+0x7f5>
+ .byte 224,255 // loopne 39a87 <.literal16+0x7f7>
+ .byte 224,255 // loopne 39a89 <.literal16+0x7f9>
+ .byte 224,255 // loopne 39a8b <.literal16+0x7fb>
+ .byte 224,255 // loopne 39a8d <.literal16+0x7fd>
+ .byte 224,255 // loopne 39a8f <.literal16+0x7ff>
.byte 15,0,15 // str (%rdi)
.byte 0,15 // add %cl,(%rdi)
.byte 0,15 // add %cl,(%rdi)
@@ -68528,9 +67912,9 @@ BALIGN8
.byte 4,0 // add $0x0,%al
.byte 0,0 // add %al,(%rax)
.byte 255 // (bad)
- .byte 127,0 // jg 3a8ab <.literal8+0x1b>
+ .byte 127,0 // jg 39b2b <.literal8+0x1b>
.byte 0,255 // add %bh,%bh
- .byte 127,0 // jg 3a8af <.literal8+0x1f>
+ .byte 127,0 // jg 39b2f <.literal8+0x1f>
.byte 0,0 // add %al,(%rax)
.byte 128,0,0 // addb $0x0,(%rax)
.byte 0 // .byte 0x0