aboutsummaryrefslogtreecommitdiffhomepage
path: root/src
diff options
context:
space:
mode:
authorGravatar Mike Klein <mtklein@chromium.org>2017-08-03 22:12:32 -0400
committerGravatar Skia Commit-Bot <skia-commit-bot@chromium.org>2017-08-04 05:19:38 +0000
commit5910ed347a638ded8cd4c06dbfda086695df1112 (patch)
treec17a722b06451d269cd4a867e1959998bcfa718d /src
parent0e914eb7259557a74b15f8a433df287ef4e5a077 (diff)
15-bit lowp is dead, long live 8-bit lowp
Change-Id: Icc4b06094aeba3af99b534746f66286d776ef78a Reviewed-on: https://skia-review.googlesource.com/30920 Reviewed-by: Florin Malita <fmalita@chromium.org> Commit-Queue: Mike Klein <mtklein@chromium.org>
Diffstat (limited to 'src')
-rw-r--r--src/jumper/SkJumper.cpp107
-rw-r--r--src/jumper/SkJumper_generated.S5032
-rw-r--r--src/jumper/SkJumper_generated_win.S4923
-rw-r--r--src/jumper/SkJumper_stages_8bit.cpp7
-rw-r--r--src/jumper/SkJumper_stages_lowp.cpp405
-rwxr-xr-xsrc/jumper/build_stages.py30
6 files changed, 35 insertions, 10469 deletions
diff --git a/src/jumper/SkJumper.cpp b/src/jumper/SkJumper.cpp
index 84cdc20f68..20fbce5625 100644
--- a/src/jumper/SkJumper.cpp
+++ b/src/jumper/SkJumper.cpp
@@ -69,7 +69,7 @@ using StartPipelineFn = void(size_t,size_t,size_t,size_t, void**,K*);
#define ASM(name, suffix) _sk_##name##_##suffix
#endif
-// Some stages have low-precision (~15 bit) versions from SkJumper_stages_lowp.cpp.
+// Some stages have 8-bit versions from SkJumper_stages_8bit.cpp.
#define LOWP_STAGES(M) \
M(black_color) M(white_color) M(uniform_color) \
M(set_rgb) \
@@ -124,8 +124,6 @@ extern "C" {
ASM(start_pipeline,avx ),
ASM(start_pipeline,sse41 ),
ASM(start_pipeline,sse2 ),
- ASM(start_pipeline,hsw_lowp ),
- ASM(start_pipeline,ssse3_lowp),
ASM(start_pipeline,hsw_8bit ),
ASM(start_pipeline,sse41_8bit),
ASM(start_pipeline,sse2_8bit );
@@ -134,8 +132,6 @@ extern "C" {
ASM(just_return,avx),
ASM(just_return,sse41),
ASM(just_return,sse2),
- ASM(just_return,hsw_lowp ),
- ASM(just_return,ssse3_lowp),
ASM(just_return,hsw_8bit ),
ASM(just_return,sse41_8bit),
ASM(just_return,sse2_8bit );
@@ -153,12 +149,6 @@ extern "C" {
SK_RASTER_PIPELINE_STAGES(M)
#undef M
- #define M(st) StageFn ASM(st,hsw_lowp);
- LOWP_STAGES(M)
- #undef M
- #define M(st) StageFn ASM(st,ssse3_lowp);
- LOWP_STAGES(M)
- #undef M
#define M(st) StageFn ASM(st,hsw_8bit);
LOWP_STAGES(M)
#undef M
@@ -188,44 +178,26 @@ extern "C" {
#if !__has_feature(memory_sanitizer) && (defined(__x86_64__) || defined(_M_X64))
template <SkRasterPipeline::StockStage st>
- static constexpr StageFn* hsw_lowp() { return nullptr; }
+ static constexpr StageFn* hsw_8bit() { return nullptr; }
+
+ template <SkRasterPipeline::StockStage st>
+ static constexpr StageFn* sse41_8bit() { return nullptr; }
template <SkRasterPipeline::StockStage st>
- static constexpr StageFn* ssse3_lowp() { return nullptr; }
+ static constexpr StageFn* sse2_8bit() { return nullptr; }
#define M(st) \
- template <> constexpr StageFn* hsw_lowp<SkRasterPipeline::st>() { \
- return ASM(st,hsw_lowp); \
+ template <> constexpr StageFn* hsw_8bit<SkRasterPipeline::st>() { \
+ return ASM(st,hsw_8bit); \
} \
- template <> constexpr StageFn* ssse3_lowp<SkRasterPipeline::st>() { \
- return ASM(st,ssse3_lowp); \
+ template <> constexpr StageFn* sse41_8bit<SkRasterPipeline::st>() { \
+ return ASM(st,sse41_8bit); \
+ } \
+ template <> constexpr StageFn* sse2_8bit<SkRasterPipeline::st>() { \
+ return ASM(st,sse2_8bit); \
}
LOWP_STAGES(M)
#undef M
-
- #if !defined(SK_JUMPER_LEGACY_LOWP)
- template <SkRasterPipeline::StockStage st>
- static constexpr StageFn* hsw_8bit() { return nullptr; }
-
- template <SkRasterPipeline::StockStage st>
- static constexpr StageFn* sse41_8bit() { return nullptr; }
-
- template <SkRasterPipeline::StockStage st>
- static constexpr StageFn* sse2_8bit() { return nullptr; }
-
- #define M(st) \
- template <> constexpr StageFn* hsw_8bit<SkRasterPipeline::st>() { \
- return ASM(st,hsw_8bit); \
- } \
- template <> constexpr StageFn* sse41_8bit<SkRasterPipeline::st>() { \
- return ASM(st,sse41_8bit); \
- } \
- template <> constexpr StageFn* sse2_8bit<SkRasterPipeline::st>() { \
- return ASM(st,sse2_8bit); \
- }
- LOWP_STAGES(M)
- #undef M
- #endif
#endif
// Engines comprise everything we need to run SkRasterPipelines.
@@ -336,51 +308,30 @@ static SkJumper_Engine choose_engine() {
static SkJumper_Engine choose_lowp() {
#if !__has_feature(memory_sanitizer) && (defined(__x86_64__) || defined(_M_X64))
- #if !defined(SK_JUMPER_LEGACY_LOWP)
- if (1 && SkCpu::Supports(SkCpu::HSW)) {
- return {
- #define M(st) hsw_8bit<SkRasterPipeline::st>(),
- { SK_RASTER_PIPELINE_STAGES(M) },
- ASM(start_pipeline,hsw_8bit),
- ASM(just_return ,hsw_8bit)
- #undef M
- };
- }
- if (1 && SkCpu::Supports(SkCpu::SSE41)) {
- return {
- #define M(st) sse41_8bit<SkRasterPipeline::st>(),
- { SK_RASTER_PIPELINE_STAGES(M) },
- ASM(start_pipeline,sse41_8bit),
- ASM(just_return ,sse41_8bit)
- #undef M
- };
- }
- if (1 && SkCpu::Supports(SkCpu::SSE2)) {
- return {
- #define M(st) sse2_8bit<SkRasterPipeline::st>(),
- { SK_RASTER_PIPELINE_STAGES(M) },
- ASM(start_pipeline,sse2_8bit),
- ASM(just_return ,sse2_8bit)
- #undef M
- };
- }
- #endif
-
if (1 && SkCpu::Supports(SkCpu::HSW)) {
return {
- #define M(st) hsw_lowp<SkRasterPipeline::st>(),
+ #define M(st) hsw_8bit<SkRasterPipeline::st>(),
+ { SK_RASTER_PIPELINE_STAGES(M) },
+ ASM(start_pipeline,hsw_8bit),
+ ASM(just_return ,hsw_8bit)
+ #undef M
+ };
+ }
+ if (1 && SkCpu::Supports(SkCpu::SSE41)) {
+ return {
+ #define M(st) sse41_8bit<SkRasterPipeline::st>(),
{ SK_RASTER_PIPELINE_STAGES(M) },
- ASM(start_pipeline,hsw_lowp),
- ASM(just_return ,hsw_lowp)
+ ASM(start_pipeline,sse41_8bit),
+ ASM(just_return ,sse41_8bit)
#undef M
};
}
- if (1 && SkCpu::Supports(SkCpu::SSSE3)) {
+ if (1 && SkCpu::Supports(SkCpu::SSE2)) {
return {
- #define M(st) ssse3_lowp<SkRasterPipeline::st>(),
+ #define M(st) sse2_8bit<SkRasterPipeline::st>(),
{ SK_RASTER_PIPELINE_STAGES(M) },
- ASM(start_pipeline,ssse3_lowp),
- ASM(just_return ,ssse3_lowp)
+ ASM(start_pipeline,sse2_8bit),
+ ASM(just_return ,sse2_8bit)
#undef M
};
}
diff --git a/src/jumper/SkJumper_generated.S b/src/jumper/SkJumper_generated.S
index 0e39cf728f..a30e5df8ee 100644
--- a/src/jumper/SkJumper_generated.S
+++ b/src/jumper/SkJumper_generated.S
@@ -45339,5038 +45339,6 @@ BALIGN4
.byte 0,63 // add %bh,(%rdi)
BALIGN32
-HIDDEN _sk_start_pipeline_hsw_lowp
-.globl _sk_start_pipeline_hsw_lowp
-FUNCTION(_sk_start_pipeline_hsw_lowp)
-_sk_start_pipeline_hsw_lowp:
- .byte 85 // push %rbp
- .byte 72,137,229 // mov %rsp,%rbp
- .byte 65,87 // push %r15
- .byte 65,86 // push %r14
- .byte 65,85 // push %r13
- .byte 65,84 // push %r12
- .byte 83 // push %rbx
- .byte 72,131,236,40 // sub $0x28,%rsp
- .byte 72,137,211 // mov %rdx,%rbx
- .byte 73,137,246 // mov %rsi,%r14
- .byte 72,137,125,208 // mov %rdi,-0x30(%rbp)
- .byte 76,137,198 // mov %r8,%rsi
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,137,194 // mov %rax,%rdx
- .byte 72,137,77,176 // mov %rcx,-0x50(%rbp)
- .byte 73,57,206 // cmp %rcx,%r14
- .byte 115,117 // jae a1 <_sk_start_pipeline_hsw_lowp+0xa1>
- .byte 72,139,69,208 // mov -0x30(%rbp),%rax
- .byte 72,141,64,16 // lea 0x10(%rax),%rax
- .byte 72,137,69,184 // mov %rax,-0x48(%rbp)
- .byte 76,137,77,192 // mov %r9,-0x40(%rbp)
- .byte 72,137,85,200 // mov %rdx,-0x38(%rbp)
- .byte 73,137,221 // mov %rbx,%r13
- .byte 73,137,244 // mov %rsi,%r12
- .byte 72,139,85,208 // mov -0x30(%rbp),%rdx
- .byte 72,57,93,184 // cmp %rbx,-0x48(%rbp)
- .byte 119,45 // ja 7d <_sk_start_pipeline_hsw_lowp+0x7d>
- .byte 76,139,125,208 // mov -0x30(%rbp),%r15
- .byte 69,49,192 // xor %r8d,%r8d
- .byte 76,137,207 // mov %r9,%rdi
- .byte 76,137,230 // mov %r12,%rsi
- .byte 76,137,250 // mov %r15,%rdx
- .byte 76,137,241 // mov %r14,%rcx
- .byte 255,85,200 // callq *-0x38(%rbp)
- .byte 76,137,235 // mov %r13,%rbx
- .byte 76,139,77,192 // mov -0x40(%rbp),%r9
- .byte 73,141,87,16 // lea 0x10(%r15),%rdx
- .byte 73,131,199,32 // add $0x20,%r15
- .byte 73,57,223 // cmp %rbx,%r15
- .byte 73,137,215 // mov %rdx,%r15
- .byte 118,215 // jbe 54 <_sk_start_pipeline_hsw_lowp+0x54>
- .byte 73,137,216 // mov %rbx,%r8
- .byte 73,41,208 // sub %rdx,%r8
- .byte 116,19 // je 98 <_sk_start_pipeline_hsw_lowp+0x98>
- .byte 76,137,207 // mov %r9,%rdi
- .byte 76,137,230 // mov %r12,%rsi
- .byte 76,137,241 // mov %r14,%rcx
- .byte 255,85,200 // callq *-0x38(%rbp)
- .byte 76,137,235 // mov %r13,%rbx
- .byte 76,139,77,192 // mov -0x40(%rbp),%r9
- .byte 73,255,198 // inc %r14
- .byte 76,59,117,176 // cmp -0x50(%rbp),%r14
- .byte 117,165 // jne 46 <_sk_start_pipeline_hsw_lowp+0x46>
- .byte 72,131,196,40 // add $0x28,%rsp
- .byte 91 // pop %rbx
- .byte 65,92 // pop %r12
- .byte 65,93 // pop %r13
- .byte 65,94 // pop %r14
- .byte 65,95 // pop %r15
- .byte 93 // pop %rbp
- .byte 197,248,119 // vzeroupper
- .byte 195 // retq
-
-HIDDEN _sk_just_return_hsw_lowp
-.globl _sk_just_return_hsw_lowp
-FUNCTION(_sk_just_return_hsw_lowp)
-_sk_just_return_hsw_lowp:
- .byte 195 // retq
-
-HIDDEN _sk_uniform_color_hsw_lowp
-.globl _sk_uniform_color_hsw_lowp
-FUNCTION(_sk_uniform_color_hsw_lowp)
-_sk_uniform_color_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,121,24,5,25,34,0,0 // vbroadcastss 0x2219(%rip),%xmm0 # 22d8 <_sk_xor__hsw_lowp+0x78>
- .byte 197,248,88,24 // vaddps (%rax),%xmm0,%xmm3
- .byte 196,226,125,121,195 // vpbroadcastw %xmm3,%ymm0
- .byte 197,251,112,203,234 // vpshuflw $0xea,%xmm3,%xmm1
- .byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
- .byte 196,227,121,4,211,230 // vpermilps $0xe6,%xmm3,%xmm2
- .byte 197,251,112,210,224 // vpshuflw $0xe0,%xmm2,%xmm2
- .byte 196,226,125,88,210 // vpbroadcastd %xmm2,%ymm2
- .byte 196,227,121,4,219,236 // vpermilps $0xec,%xmm3,%xmm3
- .byte 197,251,112,219,234 // vpshuflw $0xea,%xmm3,%xmm3
- .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_black_color_hsw_lowp
-.globl _sk_black_color_hsw_lowp
-FUNCTION(_sk_black_color_hsw_lowp)
-_sk_black_color_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,255,33,0,0 // vpbroadcastw 0x21ff(%rip),%ymm3 # 2300 <_sk_xor__hsw_lowp+0xa0>
- .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
- .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
- .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_white_color_hsw_lowp
-.globl _sk_white_color_hsw_lowp
-FUNCTION(_sk_white_color_hsw_lowp)
-_sk_white_color_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,5,232,33,0,0 // vpbroadcastw 0x21e8(%rip),%ymm0 # 2302 <_sk_xor__hsw_lowp+0xa2>
- .byte 197,253,111,200 // vmovdqa %ymm0,%ymm1
- .byte 197,253,111,208 // vmovdqa %ymm0,%ymm2
- .byte 197,253,111,216 // vmovdqa %ymm0,%ymm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_set_rgb_hsw_lowp
-.globl _sk_set_rgb_hsw_lowp
-FUNCTION(_sk_set_rgb_hsw_lowp)
-_sk_set_rgb_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,250,16,21,170,33,0,0 // vmovss 0x21aa(%rip),%xmm2 # 22dc <_sk_xor__hsw_lowp+0x7c>
- .byte 197,234,88,0 // vaddss (%rax),%xmm2,%xmm0
- .byte 196,193,121,126,193 // vmovd %xmm0,%r9d
- .byte 196,193,121,110,193 // vmovd %r9d,%xmm0
- .byte 196,226,125,121,192 // vpbroadcastw %xmm0,%ymm0
- .byte 197,234,88,72,4 // vaddss 0x4(%rax),%xmm2,%xmm1
- .byte 196,193,121,126,201 // vmovd %xmm1,%r9d
- .byte 196,193,121,110,201 // vmovd %r9d,%xmm1
- .byte 196,226,125,121,201 // vpbroadcastw %xmm1,%ymm1
- .byte 197,234,88,80,8 // vaddss 0x8(%rax),%xmm2,%xmm2
- .byte 197,249,126,208 // vmovd %xmm2,%eax
- .byte 197,249,110,208 // vmovd %eax,%xmm2
- .byte 196,226,125,121,210 // vpbroadcastw %xmm2,%ymm2
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_premul_hsw_lowp
-.globl _sk_premul_hsw_lowp
-FUNCTION(_sk_premul_hsw_lowp)
-_sk_premul_hsw_lowp:
- .byte 196,226,125,11,195 // vpmulhrsw %ymm3,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,226,117,11,203 // vpmulhrsw %ymm3,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,226,109,11,211 // vpmulhrsw %ymm3,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_load_8888_hsw_lowp
-.globl _sk_load_8888_hsw_lowp
-FUNCTION(_sk_load_8888_hsw_lowp)
-_sk_load_8888_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,211,0,0,0 // jne 285 <_sk_load_8888_hsw_lowp+0xf4>
- .byte 196,193,126,111,92,153,32 // vmovdqu 0x20(%r9,%rbx,4),%ymm3
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 197,253,111,5,89,33,0,0 // vmovdqa 0x2159(%rip),%ymm0 # 2320 <_sk_xor__hsw_lowp+0xc0>
- .byte 196,226,61,0,200 // vpshufb %ymm0,%ymm8,%ymm1
- .byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 196,226,101,0,192 // vpshufb %ymm0,%ymm3,%ymm0
- .byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,98,125,121,13,79,33,0,0 // vpbroadcastw 0x214f(%rip),%ymm9 # 2340 <_sk_xor__hsw_lowp+0xe0>
- .byte 196,193,125,228,193 // vpmulhuw %ymm9,%ymm0,%ymm0
- .byte 197,253,111,13,98,33,0,0 // vmovdqa 0x2162(%rip),%ymm1 # 2360 <_sk_xor__hsw_lowp+0x100>
- .byte 196,226,61,0,209 // vpshufb %ymm1,%ymm8,%ymm2
- .byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
- .byte 196,226,101,0,201 // vpshufb %ymm1,%ymm3,%ymm1
- .byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 196,227,109,56,201,1 // vinserti128 $0x1,%xmm1,%ymm2,%ymm1
- .byte 197,245,113,241,8 // vpsllw $0x8,%ymm1,%ymm1
- .byte 196,193,117,228,201 // vpmulhuw %ymm9,%ymm1,%ymm1
- .byte 197,253,111,21,84,33,0,0 // vmovdqa 0x2154(%rip),%ymm2 # 2380 <_sk_xor__hsw_lowp+0x120>
- .byte 196,98,61,0,210 // vpshufb %ymm2,%ymm8,%ymm10
- .byte 196,67,253,0,210,232 // vpermq $0xe8,%ymm10,%ymm10
- .byte 196,226,101,0,210 // vpshufb %ymm2,%ymm3,%ymm2
- .byte 196,227,253,0,210,232 // vpermq $0xe8,%ymm2,%ymm2
- .byte 196,227,45,56,210,1 // vinserti128 $0x1,%xmm2,%ymm10,%ymm2
- .byte 197,237,113,242,8 // vpsllw $0x8,%ymm2,%ymm2
- .byte 196,193,109,228,209 // vpmulhuw %ymm9,%ymm2,%ymm2
- .byte 197,125,111,21,70,33,0,0 // vmovdqa 0x2146(%rip),%ymm10 # 23a0 <_sk_xor__hsw_lowp+0x140>
- .byte 196,66,61,0,194 // vpshufb %ymm10,%ymm8,%ymm8
- .byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
- .byte 196,194,101,0,218 // vpshufb %ymm10,%ymm3,%ymm3
- .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
- .byte 196,227,61,56,219,1 // vinserti128 $0x1,%xmm3,%ymm8,%ymm3
- .byte 197,229,113,243,8 // vpsllw $0x8,%ymm3,%ymm3
- .byte 196,193,101,228,217 // vpmulhuw %ymm9,%ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,29,255,255,255 // ja 1bf <_sk_load_8888_hsw_lowp+0x2e>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,27,1,0,0 // lea 0x11b(%rip),%r11 # 3c8 <_sk_load_8888_hsw_lowp+0x237>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,110,4,153 // vmovd (%r9,%rbx,4),%xmm8
- .byte 233,254,254,255,255 // jmpq 1bf <_sk_load_8888_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,8 // vmovd 0x8(%r9,%rbx,4),%xmm0
- .byte 196,226,121,89,192 // vpbroadcastq %xmm0,%xmm0
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,99,101,2,192,4 // vpblendd $0x4,%ymm0,%ymm3,%ymm8
- .byte 196,194,121,53,4,153 // vpmovzxdq (%r9,%rbx,4),%xmm0
- .byte 197,249,112,192,232 // vpshufd $0xe8,%xmm0,%xmm0
- .byte 196,99,61,2,192,3 // vpblendd $0x3,%ymm0,%ymm8,%ymm8
- .byte 233,210,254,255,255 // jmpq 1bf <_sk_load_8888_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,24 // vmovd 0x18(%r9,%rbx,4),%xmm0
- .byte 196,226,125,89,192 // vpbroadcastq %xmm0,%ymm0
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,99,101,2,192,64 // vpblendd $0x40,%ymm0,%ymm3,%ymm8
- .byte 196,99,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm0
- .byte 196,195,121,34,68,153,20,1 // vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,99,61,56,192,1 // vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- .byte 196,99,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm0
- .byte 196,195,121,34,68,153,16,0 // vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,99,61,56,192,1 // vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- .byte 196,193,122,111,4,153 // vmovdqu (%r9,%rbx,4),%xmm0
- .byte 196,67,125,2,192,240 // vpblendd $0xf0,%ymm8,%ymm0,%ymm8
- .byte 233,131,254,255,255 // jmpq 1bf <_sk_load_8888_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,40 // vmovd 0x28(%r9,%rbx,4),%xmm0
- .byte 196,226,121,89,192 // vpbroadcastq %xmm0,%xmm0
- .byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
- .byte 196,227,117,2,216,4 // vpblendd $0x4,%ymm0,%ymm1,%ymm3
- .byte 196,195,97,34,68,153,36,1 // vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm3,%xmm0
- .byte 196,227,101,2,216,15 // vpblendd $0xf,%ymm0,%ymm3,%ymm3
- .byte 196,193,121,110,68,153,32 // vmovd 0x20(%r9,%rbx,4),%xmm0
- .byte 196,227,101,2,216,1 // vpblendd $0x1,%ymm0,%ymm3,%ymm3
- .byte 233,71,254,255,255 // jmpq 1b9 <_sk_load_8888_hsw_lowp+0x28>
- .byte 196,193,121,110,68,153,56 // vmovd 0x38(%r9,%rbx,4),%xmm0
- .byte 196,226,125,89,192 // vpbroadcastq %xmm0,%ymm0
- .byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
- .byte 196,227,117,2,216,64 // vpblendd $0x40,%ymm0,%ymm1,%ymm3
- .byte 196,227,125,57,216,1 // vextracti128 $0x1,%ymm3,%xmm0
- .byte 196,195,121,34,68,153,52,1 // vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,227,101,56,216,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,227,125,57,216,1 // vextracti128 $0x1,%ymm3,%xmm0
- .byte 196,195,121,34,68,153,48,0 // vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,227,101,56,216,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 196,193,122,111,68,153,32 // vmovdqu 0x20(%r9,%rbx,4),%xmm0
- .byte 196,227,125,2,219,240 // vpblendd $0xf0,%ymm3,%ymm0,%ymm3
- .byte 233,247,253,255,255 // jmpq 1bf <_sk_load_8888_hsw_lowp+0x2e>
- .byte 238 // out %al,(%dx)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 249 // stc
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,99,255 // jmpq *-0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,79,255 // decl -0x1(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 59,255 // cmp %edi,%edi
- .byte 255 // (bad)
- .byte 255,37,255,255,255,241 // jmpq *-0xe000001(%rip) # fffffffff20003e4 <_sk_xor__hsw_lowp+0xfffffffff1ffe184>
- .byte 253 // std
- .byte 255 // (bad)
- .byte 255,152,255,255,255,138 // lcall *-0x75000001(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,116,255,255 // pushq -0x1(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 232,255,255,255,212 // callq ffffffffd50003f8 <_sk_xor__hsw_lowp+0xffffffffd4ffe198>
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,192 // inc %eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
- .byte 170 // stos %al,%es:(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_8888_dst_hsw_lowp
-.globl _sk_load_8888_dst_hsw_lowp
-FUNCTION(_sk_load_8888_dst_hsw_lowp)
-_sk_load_8888_dst_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,211,0,0,0 // jne 4f8 <_sk_load_8888_dst_hsw_lowp+0xf4>
- .byte 196,193,126,111,124,153,32 // vmovdqu 0x20(%r9,%rbx,4),%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 197,253,111,37,134,31,0,0 // vmovdqa 0x1f86(%rip),%ymm4 # 23c0 <_sk_xor__hsw_lowp+0x160>
- .byte 196,226,61,0,236 // vpshufb %ymm4,%ymm8,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,226,69,0,228 // vpshufb %ymm4,%ymm7,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,98,125,121,13,124,31,0,0 // vpbroadcastw 0x1f7c(%rip),%ymm9 # 23e0 <_sk_xor__hsw_lowp+0x180>
- .byte 196,193,93,228,225 // vpmulhuw %ymm9,%ymm4,%ymm4
- .byte 197,253,111,45,143,31,0,0 // vmovdqa 0x1f8f(%rip),%ymm5 # 2400 <_sk_xor__hsw_lowp+0x1a0>
- .byte 196,226,61,0,245 // vpshufb %ymm5,%ymm8,%ymm6
- .byte 196,227,253,0,246,232 // vpermq $0xe8,%ymm6,%ymm6
- .byte 196,226,69,0,237 // vpshufb %ymm5,%ymm7,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,227,77,56,237,1 // vinserti128 $0x1,%xmm5,%ymm6,%ymm5
- .byte 197,213,113,245,8 // vpsllw $0x8,%ymm5,%ymm5
- .byte 196,193,85,228,233 // vpmulhuw %ymm9,%ymm5,%ymm5
- .byte 197,253,111,53,129,31,0,0 // vmovdqa 0x1f81(%rip),%ymm6 # 2420 <_sk_xor__hsw_lowp+0x1c0>
- .byte 196,98,61,0,214 // vpshufb %ymm6,%ymm8,%ymm10
- .byte 196,67,253,0,210,232 // vpermq $0xe8,%ymm10,%ymm10
- .byte 196,226,69,0,246 // vpshufb %ymm6,%ymm7,%ymm6
- .byte 196,227,253,0,246,232 // vpermq $0xe8,%ymm6,%ymm6
- .byte 196,227,45,56,246,1 // vinserti128 $0x1,%xmm6,%ymm10,%ymm6
- .byte 197,205,113,246,8 // vpsllw $0x8,%ymm6,%ymm6
- .byte 196,193,77,228,241 // vpmulhuw %ymm9,%ymm6,%ymm6
- .byte 197,125,111,21,115,31,0,0 // vmovdqa 0x1f73(%rip),%ymm10 # 2440 <_sk_xor__hsw_lowp+0x1e0>
- .byte 196,66,61,0,194 // vpshufb %ymm10,%ymm8,%ymm8
- .byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
- .byte 196,194,69,0,250 // vpshufb %ymm10,%ymm7,%ymm7
- .byte 196,227,253,0,255,232 // vpermq $0xe8,%ymm7,%ymm7
- .byte 196,227,61,56,255,1 // vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- .byte 197,197,113,247,8 // vpsllw $0x8,%ymm7,%ymm7
- .byte 196,193,69,228,249 // vpmulhuw %ymm9,%ymm7,%ymm7
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,29,255,255,255 // ja 432 <_sk_load_8888_dst_hsw_lowp+0x2e>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,28,1,0,0 // lea 0x11c(%rip),%r11 # 63c <_sk_load_8888_dst_hsw_lowp+0x238>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,110,4,153 // vmovd (%r9,%rbx,4),%xmm8
- .byte 233,254,254,255,255 // jmpq 432 <_sk_load_8888_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,8 // vmovd 0x8(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,4 // vpblendd $0x4,%ymm4,%ymm7,%ymm8
- .byte 196,194,121,53,36,153 // vpmovzxdq (%r9,%rbx,4),%xmm4
- .byte 197,249,112,228,232 // vpshufd $0xe8,%xmm4,%xmm4
- .byte 196,99,61,2,196,3 // vpblendd $0x3,%ymm4,%ymm8,%ymm8
- .byte 233,210,254,255,255 // jmpq 432 <_sk_load_8888_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,24 // vmovd 0x18(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,64 // vpblendd $0x40,%ymm4,%ymm7,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,20,1 // vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,16,0 // vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,193,122,111,36,153 // vmovdqu (%r9,%rbx,4),%xmm4
- .byte 196,67,93,2,192,240 // vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- .byte 233,131,254,255,255 // jmpq 432 <_sk_load_8888_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,40 // vmovd 0x28(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,4 // vpblendd $0x4,%ymm4,%ymm5,%ymm7
- .byte 196,195,65,34,100,153,36,1 // vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- .byte 196,227,69,2,252,15 // vpblendd $0xf,%ymm4,%ymm7,%ymm7
- .byte 196,193,121,110,100,153,32 // vmovd 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,69,2,252,1 // vpblendd $0x1,%ymm4,%ymm7,%ymm7
- .byte 233,71,254,255,255 // jmpq 42c <_sk_load_8888_dst_hsw_lowp+0x28>
- .byte 196,193,121,110,100,153,56 // vmovd 0x38(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,64 // vpblendd $0x40,%ymm4,%ymm5,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,52,1 // vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,48,0 // vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 196,193,122,111,100,153,32 // vmovdqu 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,93,2,255,240 // vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- .byte 233,247,253,255,255 // jmpq 432 <_sk_load_8888_dst_hsw_lowp+0x2e>
- .byte 144 // nop
- .byte 237 // in (%dx),%eax
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,14 // decl (%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 248 // clc
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,98,255 // jmpq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,78,255 // decl -0x1(%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 58,255 // cmp %bh,%bh
- .byte 255 // (bad)
- .byte 255,36,255 // jmpq *(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255,240 // push %rax
- .byte 253 // std
- .byte 255 // (bad)
- .byte 255,151,255,255,255,137 // callq *-0x76000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,115,255 // pushq -0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,231 // jmpq *%rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 191,255,255,255,169 // mov $0xa9ffffff,%edi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_8888_hsw_lowp
-.globl _sk_store_8888_hsw_lowp
-FUNCTION(_sk_store_8888_hsw_lowp)
-_sk_store_8888_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 197,189,113,208,7 // vpsrlw $0x7,%ymm0,%ymm8
- .byte 196,98,125,121,13,194,29,0,0 // vpbroadcastw 0x1dc2(%rip),%ymm9 # 2460 <_sk_xor__hsw_lowp+0x200>
- .byte 196,65,61,234,193 // vpminsw %ymm9,%ymm8,%ymm8
- .byte 196,66,125,51,208 // vpmovzxwd %xmm8,%ymm10
- .byte 196,67,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm8
- .byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 197,165,113,209,7 // vpsrlw $0x7,%ymm1,%ymm11
- .byte 196,65,37,234,217 // vpminsw %ymm9,%ymm11,%ymm11
- .byte 196,67,125,57,220,1 // vextracti128 $0x1,%ymm11,%xmm12
- .byte 196,66,125,51,228 // vpmovzxwd %xmm12,%ymm12
- .byte 196,66,125,51,219 // vpmovzxwd %xmm11,%ymm11
- .byte 196,193,37,114,243,8 // vpslld $0x8,%ymm11,%ymm11
- .byte 196,193,29,114,244,8 // vpslld $0x8,%ymm12,%ymm12
- .byte 196,65,29,235,192 // vpor %ymm8,%ymm12,%ymm8
- .byte 196,65,37,235,210 // vpor %ymm10,%ymm11,%ymm10
- .byte 197,165,113,210,7 // vpsrlw $0x7,%ymm2,%ymm11
- .byte 196,65,37,234,217 // vpminsw %ymm9,%ymm11,%ymm11
- .byte 196,66,125,51,227 // vpmovzxwd %xmm11,%ymm12
- .byte 196,67,125,57,219,1 // vextracti128 $0x1,%ymm11,%xmm11
- .byte 196,66,125,51,219 // vpmovzxwd %xmm11,%ymm11
- .byte 196,193,37,114,243,16 // vpslld $0x10,%ymm11,%ymm11
- .byte 196,193,29,114,244,16 // vpslld $0x10,%ymm12,%ymm12
- .byte 197,149,113,211,7 // vpsrlw $0x7,%ymm3,%ymm13
- .byte 196,65,21,234,201 // vpminsw %ymm9,%ymm13,%ymm9
- .byte 196,67,125,57,205,1 // vextracti128 $0x1,%ymm9,%xmm13
- .byte 196,66,125,51,237 // vpmovzxwd %xmm13,%ymm13
- .byte 196,66,125,51,201 // vpmovzxwd %xmm9,%ymm9
- .byte 196,193,13,114,241,24 // vpslld $0x18,%ymm9,%ymm14
- .byte 196,193,53,114,245,24 // vpslld $0x18,%ymm13,%ymm9
- .byte 196,65,37,235,201 // vpor %ymm9,%ymm11,%ymm9
- .byte 196,65,61,235,201 // vpor %ymm9,%ymm8,%ymm9
- .byte 196,65,29,235,198 // vpor %ymm14,%ymm12,%ymm8
- .byte 196,65,45,235,192 // vpor %ymm8,%ymm10,%ymm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,18 // jne 75a <_sk_store_8888_hsw_lowp+0xe2>
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,126,127,76,153,32 // vmovdqu %ymm9,0x20(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,235 // ja 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,175,0,0,0 // lea 0xaf(%rip),%r11 # 824 <_sk_store_8888_hsw_lowp+0x1ac>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,126,4,153 // vmovd %xmm8,(%r9,%rbx,4)
- .byte 235,207 // jmp 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 196,67,121,22,68,153,8,2 // vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- .byte 196,65,121,214,4,153 // vmovq %xmm8,(%r9,%rbx,4)
- .byte 235,191 // jmp 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,24,2 // vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,20,1 // vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,65,121,126,76,153,16 // vmovd %xmm9,0x10(%r9,%rbx,4)
- .byte 196,65,122,127,4,153 // vmovdqu %xmm8,(%r9,%rbx,4)
- .byte 235,142 // jmp 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 196,67,121,22,76,153,40,2 // vpextrd $0x2,%xmm9,0x28(%r9,%rbx,4)
- .byte 196,67,121,22,76,153,36,1 // vpextrd $0x1,%xmm9,0x24(%r9,%rbx,4)
- .byte 196,65,121,126,76,153,32 // vmovd %xmm9,0x20(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 233,108,255,255,255 // jmpq 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
- .byte 196,67,121,22,84,153,56,2 // vpextrd $0x2,%xmm10,0x38(%r9,%rbx,4)
- .byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
- .byte 196,67,121,22,84,153,52,1 // vpextrd $0x1,%xmm10,0x34(%r9,%rbx,4)
- .byte 196,67,125,57,202,1 // vextracti128 $0x1,%ymm9,%xmm10
- .byte 196,65,121,126,84,153,48 // vmovd %xmm10,0x30(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,122,127,76,153,32 // vmovdqu %xmm9,0x20(%r9,%rbx,4)
- .byte 233,49,255,255,255 // jmpq 755 <_sk_store_8888_hsw_lowp+0xdd>
- .byte 90 // pop %rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,106,255 // ljmp *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,98,255 // jmpq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,155,255,255,255,142 // lcall *-0x71000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,128,255,255,255,114 // incl 0x72ffffff(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 186,255,255,255,179 // mov $0xb3ffffff,%edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,171,255,255,255,163 // ljmp *-0x5c000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,225 // jmpq *%rcx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,197 // inc %ebp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_bgra_hsw_lowp
-.globl _sk_load_bgra_hsw_lowp
-FUNCTION(_sk_load_bgra_hsw_lowp)
-_sk_load_bgra_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,211,0,0,0 // jne 954 <_sk_load_bgra_hsw_lowp+0xf4>
- .byte 196,193,126,111,92,153,32 // vmovdqu 0x20(%r9,%rbx,4),%ymm3
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 197,253,111,5,234,27,0,0 // vmovdqa 0x1bea(%rip),%ymm0 # 2480 <_sk_xor__hsw_lowp+0x220>
- .byte 196,226,61,0,200 // vpshufb %ymm0,%ymm8,%ymm1
- .byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 196,226,101,0,192 // vpshufb %ymm0,%ymm3,%ymm0
- .byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,98,125,121,13,224,27,0,0 // vpbroadcastw 0x1be0(%rip),%ymm9 # 24a0 <_sk_xor__hsw_lowp+0x240>
- .byte 196,193,125,228,209 // vpmulhuw %ymm9,%ymm0,%ymm2
- .byte 197,253,111,5,243,27,0,0 // vmovdqa 0x1bf3(%rip),%ymm0 # 24c0 <_sk_xor__hsw_lowp+0x260>
- .byte 196,226,61,0,200 // vpshufb %ymm0,%ymm8,%ymm1
- .byte 196,227,253,0,201,232 // vpermq $0xe8,%ymm1,%ymm1
- .byte 196,226,101,0,192 // vpshufb %ymm0,%ymm3,%ymm0
- .byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,227,117,56,192,1 // vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,193,125,228,201 // vpmulhuw %ymm9,%ymm0,%ymm1
- .byte 197,253,111,5,229,27,0,0 // vmovdqa 0x1be5(%rip),%ymm0 # 24e0 <_sk_xor__hsw_lowp+0x280>
- .byte 196,98,61,0,208 // vpshufb %ymm0,%ymm8,%ymm10
- .byte 196,67,253,0,210,232 // vpermq $0xe8,%ymm10,%ymm10
- .byte 196,226,101,0,192 // vpshufb %ymm0,%ymm3,%ymm0
- .byte 196,227,253,0,192,232 // vpermq $0xe8,%ymm0,%ymm0
- .byte 196,227,45,56,192,1 // vinserti128 $0x1,%xmm0,%ymm10,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,193,125,228,193 // vpmulhuw %ymm9,%ymm0,%ymm0
- .byte 197,125,111,21,215,27,0,0 // vmovdqa 0x1bd7(%rip),%ymm10 # 2500 <_sk_xor__hsw_lowp+0x2a0>
- .byte 196,66,61,0,194 // vpshufb %ymm10,%ymm8,%ymm8
- .byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
- .byte 196,194,101,0,218 // vpshufb %ymm10,%ymm3,%ymm3
- .byte 196,227,253,0,219,232 // vpermq $0xe8,%ymm3,%ymm3
- .byte 196,227,61,56,219,1 // vinserti128 $0x1,%xmm3,%ymm8,%ymm3
- .byte 197,229,113,243,8 // vpsllw $0x8,%ymm3,%ymm3
- .byte 196,193,101,228,217 // vpmulhuw %ymm9,%ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,29,255,255,255 // ja 88e <_sk_load_bgra_hsw_lowp+0x2e>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,28,1,0,0 // lea 0x11c(%rip),%r11 # a98 <_sk_load_bgra_hsw_lowp+0x238>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,110,4,153 // vmovd (%r9,%rbx,4),%xmm8
- .byte 233,254,254,255,255 // jmpq 88e <_sk_load_bgra_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,8 // vmovd 0x8(%r9,%rbx,4),%xmm0
- .byte 196,226,121,89,192 // vpbroadcastq %xmm0,%xmm0
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,99,101,2,192,4 // vpblendd $0x4,%ymm0,%ymm3,%ymm8
- .byte 196,194,121,53,4,153 // vpmovzxdq (%r9,%rbx,4),%xmm0
- .byte 197,249,112,192,232 // vpshufd $0xe8,%xmm0,%xmm0
- .byte 196,99,61,2,192,3 // vpblendd $0x3,%ymm0,%ymm8,%ymm8
- .byte 233,210,254,255,255 // jmpq 88e <_sk_load_bgra_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,24 // vmovd 0x18(%r9,%rbx,4),%xmm0
- .byte 196,226,125,89,192 // vpbroadcastq %xmm0,%ymm0
- .byte 197,229,239,219 // vpxor %ymm3,%ymm3,%ymm3
- .byte 196,99,101,2,192,64 // vpblendd $0x40,%ymm0,%ymm3,%ymm8
- .byte 196,99,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm0
- .byte 196,195,121,34,68,153,20,1 // vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,99,61,56,192,1 // vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- .byte 196,99,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm0
- .byte 196,195,121,34,68,153,16,0 // vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,99,61,56,192,1 // vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- .byte 196,193,122,111,4,153 // vmovdqu (%r9,%rbx,4),%xmm0
- .byte 196,67,125,2,192,240 // vpblendd $0xf0,%ymm8,%ymm0,%ymm8
- .byte 233,131,254,255,255 // jmpq 88e <_sk_load_bgra_hsw_lowp+0x2e>
- .byte 196,193,121,110,68,153,40 // vmovd 0x28(%r9,%rbx,4),%xmm0
- .byte 196,226,121,89,192 // vpbroadcastq %xmm0,%xmm0
- .byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
- .byte 196,227,117,2,216,4 // vpblendd $0x4,%ymm0,%ymm1,%ymm3
- .byte 196,195,97,34,68,153,36,1 // vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm3,%xmm0
- .byte 196,227,101,2,216,15 // vpblendd $0xf,%ymm0,%ymm3,%ymm3
- .byte 196,193,121,110,68,153,32 // vmovd 0x20(%r9,%rbx,4),%xmm0
- .byte 196,227,101,2,216,1 // vpblendd $0x1,%ymm0,%ymm3,%ymm3
- .byte 233,71,254,255,255 // jmpq 888 <_sk_load_bgra_hsw_lowp+0x28>
- .byte 196,193,121,110,68,153,56 // vmovd 0x38(%r9,%rbx,4),%xmm0
- .byte 196,226,125,89,192 // vpbroadcastq %xmm0,%ymm0
- .byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
- .byte 196,227,117,2,216,64 // vpblendd $0x40,%ymm0,%ymm1,%ymm3
- .byte 196,227,125,57,216,1 // vextracti128 $0x1,%ymm3,%xmm0
- .byte 196,195,121,34,68,153,52,1 // vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,227,101,56,216,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,227,125,57,216,1 // vextracti128 $0x1,%ymm3,%xmm0
- .byte 196,195,121,34,68,153,48,0 // vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm0,%xmm0
- .byte 196,227,101,56,216,1 // vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 196,193,122,111,68,153,32 // vmovdqu 0x20(%r9,%rbx,4),%xmm0
- .byte 196,227,125,2,219,240 // vpblendd $0xf0,%ymm3,%ymm0,%ymm3
- .byte 233,247,253,255,255 // jmpq 88e <_sk_load_bgra_hsw_lowp+0x2e>
- .byte 144 // nop
- .byte 237 // in (%dx),%eax
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,14 // decl (%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 248 // clc
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,98,255 // jmpq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,78,255 // decl -0x1(%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 58,255 // cmp %bh,%bh
- .byte 255 // (bad)
- .byte 255,36,255 // jmpq *(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255,240 // push %rax
- .byte 253 // std
- .byte 255 // (bad)
- .byte 255,151,255,255,255,137 // callq *-0x76000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,115,255 // pushq -0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,231 // jmpq *%rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 191,255,255,255,169 // mov $0xa9ffffff,%edi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_bgra_dst_hsw_lowp
-.globl _sk_load_bgra_dst_hsw_lowp
-FUNCTION(_sk_load_bgra_dst_hsw_lowp)
-_sk_load_bgra_dst_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,211,0,0,0 // jne bc8 <_sk_load_bgra_dst_hsw_lowp+0xf4>
- .byte 196,193,126,111,124,153,32 // vmovdqu 0x20(%r9,%rbx,4),%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 197,253,111,37,22,26,0,0 // vmovdqa 0x1a16(%rip),%ymm4 # 2520 <_sk_xor__hsw_lowp+0x2c0>
- .byte 196,226,61,0,236 // vpshufb %ymm4,%ymm8,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,226,69,0,228 // vpshufb %ymm4,%ymm7,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,98,125,121,13,12,26,0,0 // vpbroadcastw 0x1a0c(%rip),%ymm9 # 2540 <_sk_xor__hsw_lowp+0x2e0>
- .byte 196,193,93,228,241 // vpmulhuw %ymm9,%ymm4,%ymm6
- .byte 197,253,111,37,31,26,0,0 // vmovdqa 0x1a1f(%rip),%ymm4 # 2560 <_sk_xor__hsw_lowp+0x300>
- .byte 196,226,61,0,236 // vpshufb %ymm4,%ymm8,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,226,69,0,228 // vpshufb %ymm4,%ymm7,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,193,93,228,233 // vpmulhuw %ymm9,%ymm4,%ymm5
- .byte 197,253,111,37,17,26,0,0 // vmovdqa 0x1a11(%rip),%ymm4 # 2580 <_sk_xor__hsw_lowp+0x320>
- .byte 196,98,61,0,212 // vpshufb %ymm4,%ymm8,%ymm10
- .byte 196,67,253,0,210,232 // vpermq $0xe8,%ymm10,%ymm10
- .byte 196,226,69,0,228 // vpshufb %ymm4,%ymm7,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,45,56,228,1 // vinserti128 $0x1,%xmm4,%ymm10,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,193,93,228,225 // vpmulhuw %ymm9,%ymm4,%ymm4
- .byte 197,125,111,21,3,26,0,0 // vmovdqa 0x1a03(%rip),%ymm10 # 25a0 <_sk_xor__hsw_lowp+0x340>
- .byte 196,66,61,0,194 // vpshufb %ymm10,%ymm8,%ymm8
- .byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
- .byte 196,194,69,0,250 // vpshufb %ymm10,%ymm7,%ymm7
- .byte 196,227,253,0,255,232 // vpermq $0xe8,%ymm7,%ymm7
- .byte 196,227,61,56,255,1 // vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- .byte 197,197,113,247,8 // vpsllw $0x8,%ymm7,%ymm7
- .byte 196,193,69,228,249 // vpmulhuw %ymm9,%ymm7,%ymm7
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,29,255,255,255 // ja b02 <_sk_load_bgra_dst_hsw_lowp+0x2e>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,28,1,0,0 // lea 0x11c(%rip),%r11 # d0c <_sk_load_bgra_dst_hsw_lowp+0x238>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,110,4,153 // vmovd (%r9,%rbx,4),%xmm8
- .byte 233,254,254,255,255 // jmpq b02 <_sk_load_bgra_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,8 // vmovd 0x8(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,4 // vpblendd $0x4,%ymm4,%ymm7,%ymm8
- .byte 196,194,121,53,36,153 // vpmovzxdq (%r9,%rbx,4),%xmm4
- .byte 197,249,112,228,232 // vpshufd $0xe8,%xmm4,%xmm4
- .byte 196,99,61,2,196,3 // vpblendd $0x3,%ymm4,%ymm8,%ymm8
- .byte 233,210,254,255,255 // jmpq b02 <_sk_load_bgra_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,24 // vmovd 0x18(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,64 // vpblendd $0x40,%ymm4,%ymm7,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,20,1 // vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,16,0 // vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,193,122,111,36,153 // vmovdqu (%r9,%rbx,4),%xmm4
- .byte 196,67,93,2,192,240 // vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- .byte 233,131,254,255,255 // jmpq b02 <_sk_load_bgra_dst_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,40 // vmovd 0x28(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,4 // vpblendd $0x4,%ymm4,%ymm5,%ymm7
- .byte 196,195,65,34,100,153,36,1 // vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- .byte 196,227,69,2,252,15 // vpblendd $0xf,%ymm4,%ymm7,%ymm7
- .byte 196,193,121,110,100,153,32 // vmovd 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,69,2,252,1 // vpblendd $0x1,%ymm4,%ymm7,%ymm7
- .byte 233,71,254,255,255 // jmpq afc <_sk_load_bgra_dst_hsw_lowp+0x28>
- .byte 196,193,121,110,100,153,56 // vmovd 0x38(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,64 // vpblendd $0x40,%ymm4,%ymm5,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,52,1 // vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,48,0 // vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 196,193,122,111,100,153,32 // vmovdqu 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,93,2,255,240 // vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- .byte 233,247,253,255,255 // jmpq b02 <_sk_load_bgra_dst_hsw_lowp+0x2e>
- .byte 144 // nop
- .byte 237 // in (%dx),%eax
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,14 // decl (%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 248 // clc
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,98,255 // jmpq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,78,255 // decl -0x1(%rsi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 58,255 // cmp %bh,%bh
- .byte 255 // (bad)
- .byte 255,36,255 // jmpq *(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255,240 // push %rax
- .byte 253 // std
- .byte 255 // (bad)
- .byte 255,151,255,255,255,137 // callq *-0x76000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,115,255 // pushq -0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,231 // jmpq *%rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 191,255,255,255,169 // mov $0xa9ffffff,%edi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_bgra_hsw_lowp
-.globl _sk_store_bgra_hsw_lowp
-FUNCTION(_sk_store_bgra_hsw_lowp)
-_sk_store_bgra_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 197,189,113,210,7 // vpsrlw $0x7,%ymm2,%ymm8
- .byte 196,98,125,121,13,82,24,0,0 // vpbroadcastw 0x1852(%rip),%ymm9 # 25c0 <_sk_xor__hsw_lowp+0x360>
- .byte 196,65,61,234,193 // vpminsw %ymm9,%ymm8,%ymm8
- .byte 196,66,125,51,208 // vpmovzxwd %xmm8,%ymm10
- .byte 196,67,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm8
- .byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 197,165,113,209,7 // vpsrlw $0x7,%ymm1,%ymm11
- .byte 196,65,37,234,217 // vpminsw %ymm9,%ymm11,%ymm11
- .byte 196,66,125,51,227 // vpmovzxwd %xmm11,%ymm12
- .byte 196,67,125,57,219,1 // vextracti128 $0x1,%ymm11,%xmm11
- .byte 196,66,125,51,219 // vpmovzxwd %xmm11,%ymm11
- .byte 196,193,37,114,243,8 // vpslld $0x8,%ymm11,%ymm11
- .byte 196,193,29,114,244,8 // vpslld $0x8,%ymm12,%ymm12
- .byte 197,149,113,208,7 // vpsrlw $0x7,%ymm0,%ymm13
- .byte 196,65,21,234,233 // vpminsw %ymm9,%ymm13,%ymm13
- .byte 196,67,125,57,238,1 // vextracti128 $0x1,%ymm13,%xmm14
- .byte 196,66,125,51,246 // vpmovzxwd %xmm14,%ymm14
- .byte 196,66,125,51,237 // vpmovzxwd %xmm13,%ymm13
- .byte 196,193,21,114,245,16 // vpslld $0x10,%ymm13,%ymm13
- .byte 196,193,13,114,246,16 // vpslld $0x10,%ymm14,%ymm14
- .byte 197,133,113,211,7 // vpsrlw $0x7,%ymm3,%ymm15
- .byte 196,65,5,234,201 // vpminsw %ymm9,%ymm15,%ymm9
- .byte 196,67,125,57,207,1 // vextracti128 $0x1,%ymm9,%xmm15
- .byte 196,66,125,51,255 // vpmovzxwd %xmm15,%ymm15
- .byte 196,66,125,51,201 // vpmovzxwd %xmm9,%ymm9
- .byte 196,193,53,114,241,24 // vpslld $0x18,%ymm9,%ymm9
- .byte 196,193,5,114,247,24 // vpslld $0x18,%ymm15,%ymm15
- .byte 196,65,61,235,198 // vpor %ymm14,%ymm8,%ymm8
- .byte 196,65,45,235,237 // vpor %ymm13,%ymm10,%ymm13
- .byte 196,65,37,235,215 // vpor %ymm15,%ymm11,%ymm10
- .byte 196,65,61,235,210 // vpor %ymm10,%ymm8,%ymm10
- .byte 196,65,29,235,193 // vpor %ymm9,%ymm12,%ymm8
- .byte 196,65,21,235,192 // vpor %ymm8,%ymm13,%ymm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,18 // jne e2a <_sk_store_bgra_hsw_lowp+0xe2>
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,126,127,84,153,32 // vmovdqu %ymm10,0x20(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,235 // ja e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,175,0,0,0 // lea 0xaf(%rip),%r11 # ef4 <_sk_store_bgra_hsw_lowp+0x1ac>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,126,4,153 // vmovd %xmm8,(%r9,%rbx,4)
- .byte 235,207 // jmp e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 196,67,121,22,68,153,8,2 // vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- .byte 196,65,121,214,4,153 // vmovq %xmm8,(%r9,%rbx,4)
- .byte 235,191 // jmp e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,24,2 // vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,20,1 // vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,65,121,126,76,153,16 // vmovd %xmm9,0x10(%r9,%rbx,4)
- .byte 196,65,122,127,4,153 // vmovdqu %xmm8,(%r9,%rbx,4)
- .byte 235,142 // jmp e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 196,67,121,22,84,153,40,2 // vpextrd $0x2,%xmm10,0x28(%r9,%rbx,4)
- .byte 196,67,121,22,84,153,36,1 // vpextrd $0x1,%xmm10,0x24(%r9,%rbx,4)
- .byte 196,65,121,126,84,153,32 // vmovd %xmm10,0x20(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 233,108,255,255,255 // jmpq e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,67,121,22,76,153,56,2 // vpextrd $0x2,%xmm9,0x38(%r9,%rbx,4)
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,67,121,22,76,153,52,1 // vpextrd $0x1,%xmm9,0x34(%r9,%rbx,4)
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,65,121,126,76,153,48 // vmovd %xmm9,0x30(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,122,127,84,153,32 // vmovdqu %xmm10,0x20(%r9,%rbx,4)
- .byte 233,49,255,255,255 // jmpq e25 <_sk_store_bgra_hsw_lowp+0xdd>
- .byte 90 // pop %rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,106,255 // ljmp *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,98,255 // jmpq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,155,255,255,255,142 // lcall *-0x71000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,128,255,255,255,114 // incl 0x72ffffff(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 186,255,255,255,179 // mov $0xb3ffffff,%edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,171,255,255,255,163 // ljmp *-0x5c000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,225 // jmpq *%rcx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,197 // inc %ebp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_a8_hsw_lowp
-.globl _sk_load_a8_hsw_lowp
-FUNCTION(_sk_load_a8_hsw_lowp)
-_sk_load_a8_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,46 // jne f77 <_sk_load_a8_hsw_lowp+0x47>
- .byte 196,193,122,111,4,25 // vmovdqu (%r9,%rbx,1),%xmm0
- .byte 196,226,125,48,192 // vpmovzxbw %xmm0,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,226,125,121,13,96,22,0,0 // vpbroadcastw 0x1660(%rip),%ymm1 # 25c2 <_sk_xor__hsw_lowp+0x362>
- .byte 197,253,228,217 // vpmulhuw %ymm1,%ymm0,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
- .byte 197,245,239,201 // vpxor %ymm1,%ymm1,%ymm1
- .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,196 // ja f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,194,0,0,0 // lea 0xc2(%rip),%r11 # 1058 <_sk_load_a8_hsw_lowp+0x128>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 235,165 // jmp f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm0,%xmm0
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,249,110,200 // vmovd %eax,%xmm1
- .byte 196,227,121,14,193,1 // vpblendw $0x1,%xmm1,%xmm0,%xmm0
- .byte 235,136 // jmp f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,121,110,12,25 // vmovd (%r9,%rbx,1),%xmm1
- .byte 196,227,121,2,193,1 // vpblendd $0x1,%xmm1,%xmm0,%xmm0
- .byte 233,91,255,255,255 // jmpq f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm1
- .byte 196,227,113,2,192,12 // vpblendd $0xc,%xmm0,%xmm1,%xmm0
- .byte 233,46,255,255,255 // jmpq f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm1
- .byte 196,195,113,34,76,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm1,%xmm1
- .byte 196,227,113,2,192,8 // vpblendd $0x8,%xmm0,%xmm1,%xmm0
- .byte 233,249,254,255,255 // jmpq f4f <_sk_load_a8_hsw_lowp+0x1f>
- .byte 102,144 // xchg %ax,%ax
- .byte 71,255 // rex.RXB (bad)
- .byte 255 // (bad)
- .byte 255,94,255 // lcall *-0x1(%rsi)
- .byte 255 // (bad)
- .byte 255,82,255 // callq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,139,255,255,255,131 // decl -0x7c000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 123,255 // jnp 106d <_sk_load_a8_hsw_lowp+0x13d>
- .byte 255 // (bad)
- .byte 255,111,255 // ljmp *-0x1(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 184,255,255,255,176 // mov $0xb0ffffff,%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,168,255,255,255,156 // ljmp *-0x63000001(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,229 // jmpq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 221,255 // (bad)
- .byte 255 // (bad)
- .byte 255,213 // callq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,201 // dec %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_a8_dst_hsw_lowp
-.globl _sk_load_a8_dst_hsw_lowp
-FUNCTION(_sk_load_a8_dst_hsw_lowp)
-_sk_load_a8_dst_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,46 // jne 10db <_sk_load_a8_dst_hsw_lowp+0x47>
- .byte 196,193,122,111,36,25 // vmovdqu (%r9,%rbx,1),%xmm4
- .byte 196,226,125,48,228 // vpmovzxbw %xmm4,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,226,125,121,45,254,20,0,0 // vpbroadcastw 0x14fe(%rip),%ymm5 # 25c4 <_sk_xor__hsw_lowp+0x364>
- .byte 197,221,228,253 // vpmulhuw %ymm5,%ymm4,%ymm7
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,221,239,228 // vpxor %ymm4,%ymm4,%ymm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,196 // ja 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,194,0,0,0 // lea 0xc2(%rip),%r11 # 11bc <_sk_load_a8_dst_hsw_lowp+0x128>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 235,165 // jmp 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm4,%xmm4
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,249,110,232 // vmovd %eax,%xmm5
- .byte 196,227,89,14,229,1 // vpblendw $0x1,%xmm5,%xmm4,%xmm4
- .byte 235,136 // jmp 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,121,110,44,25 // vmovd (%r9,%rbx,1),%xmm5
- .byte 196,227,89,2,229,1 // vpblendd $0x1,%xmm5,%xmm4,%xmm4
- .byte 233,91,255,255,255 // jmpq 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,122,126,44,25 // vmovq (%r9,%rbx,1),%xmm5
- .byte 196,227,81,2,228,12 // vpblendd $0xc,%xmm4,%xmm5,%xmm4
- .byte 233,46,255,255,255 // jmpq 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,122,126,44,25 // vmovq (%r9,%rbx,1),%xmm5
- .byte 196,195,81,34,108,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm5,%xmm5
- .byte 196,227,81,2,228,8 // vpblendd $0x8,%xmm4,%xmm5,%xmm4
- .byte 233,249,254,255,255 // jmpq 10b3 <_sk_load_a8_dst_hsw_lowp+0x1f>
- .byte 102,144 // xchg %ax,%ax
- .byte 71,255 // rex.RXB (bad)
- .byte 255 // (bad)
- .byte 255,94,255 // lcall *-0x1(%rsi)
- .byte 255 // (bad)
- .byte 255,82,255 // callq *-0x1(%rdx)
- .byte 255 // (bad)
- .byte 255,139,255,255,255,131 // decl -0x7c000001(%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 123,255 // jnp 11d1 <_sk_load_a8_dst_hsw_lowp+0x13d>
- .byte 255 // (bad)
- .byte 255,111,255 // ljmp *-0x1(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 184,255,255,255,176 // mov $0xb0ffffff,%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,168,255,255,255,156 // ljmp *-0x63000001(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,229 // jmpq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 221,255 // (bad)
- .byte 255 // (bad)
- .byte 255,213 // callq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,201 // dec %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_a8_hsw_lowp
-.globl _sk_store_a8_hsw_lowp
-FUNCTION(_sk_store_a8_hsw_lowp)
-_sk_store_a8_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 197,189,113,211,7 // vpsrlw $0x7,%ymm3,%ymm8
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,65,57,103,193 // vpackuswb %xmm9,%xmm8,%xmm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,11 // jne 122c <_sk_store_a8_hsw_lowp+0x34>
- .byte 196,65,122,127,4,25 // vmovdqu %xmm8,(%r9,%rbx,1)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,235 // ja 1227 <_sk_store_a8_hsw_lowp+0x2f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,137,0,0,0 // lea 0x89(%rip),%r11 # 12d0 <_sk_store_a8_hsw_lowp+0xd8>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,67,121,20,4,25,0 // vpextrb $0x0,%xmm8,(%r9,%rbx,1)
- .byte 235,206 // jmp 1227 <_sk_store_a8_hsw_lowp+0x2f>
- .byte 196,67,121,20,68,25,2,2 // vpextrb $0x2,%xmm8,0x2(%r9,%rbx,1)
- .byte 196,67,121,21,4,25,0 // vpextrw $0x0,%xmm8,(%r9,%rbx,1)
- .byte 235,189 // jmp 1227 <_sk_store_a8_hsw_lowp+0x2f>
- .byte 196,67,121,20,68,25,6,6 // vpextrb $0x6,%xmm8,0x6(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,5,5 // vpextrb $0x5,%xmm8,0x5(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,4,4 // vpextrb $0x4,%xmm8,0x4(%r9,%rbx,1)
- .byte 196,65,121,126,4,25 // vmovd %xmm8,(%r9,%rbx,1)
- .byte 235,157 // jmp 1227 <_sk_store_a8_hsw_lowp+0x2f>
- .byte 196,67,121,20,68,25,10,10 // vpextrb $0xa,%xmm8,0xa(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,9,9 // vpextrb $0x9,%xmm8,0x9(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,8,8 // vpextrb $0x8,%xmm8,0x8(%r9,%rbx,1)
- .byte 235,32 // jmp 12c4 <_sk_store_a8_hsw_lowp+0xcc>
- .byte 196,67,121,20,68,25,14,14 // vpextrb $0xe,%xmm8,0xe(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,13,13 // vpextrb $0xd,%xmm8,0xd(%r9,%rbx,1)
- .byte 196,67,121,20,68,25,12,12 // vpextrb $0xc,%xmm8,0xc(%r9,%rbx,1)
- .byte 196,67,121,22,68,25,8,2 // vpextrd $0x2,%xmm8,0x8(%r9,%rbx,1)
- .byte 196,65,121,214,4,25 // vmovq %xmm8,(%r9,%rbx,1)
- .byte 233,88,255,255,255 // jmpq 1227 <_sk_store_a8_hsw_lowp+0x2f>
- .byte 144 // nop
- .byte 128,255,255 // cmp $0xff,%bh
- .byte 255,145,255,255,255,137 // callq *-0x76000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,178,255,255,255,170 // pushq -0x55000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,162,255,255,255,154 // jmpq *-0x65000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,244 // push %rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,202 // dec %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,194 // inc %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 186,255,255,255,236 // mov $0xecffffff,%edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,228 // jmpq *%rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 220,255 // fdivr %st,%st(7)
- .byte 255 // (bad)
- .byte 255,212 // callq *%rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_g8_hsw_lowp
-.globl _sk_load_g8_hsw_lowp
-FUNCTION(_sk_load_g8_hsw_lowp)
-_sk_load_g8_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,51 // jne 1358 <_sk_load_g8_hsw_lowp+0x4c>
- .byte 196,193,122,111,4,25 // vmovdqu (%r9,%rbx,1),%xmm0
- .byte 196,226,125,48,192 // vpmovzxbw %xmm0,%ymm0
- .byte 197,253,113,240,8 // vpsllw $0x8,%ymm0,%ymm0
- .byte 196,226,125,121,13,136,18,0,0 // vpbroadcastw 0x1288(%rip),%ymm1 # 25c6 <_sk_xor__hsw_lowp+0x366>
- .byte 197,253,228,193 // vpmulhuw %ymm1,%ymm0,%ymm0
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,29,123,18,0,0 // vpbroadcastw 0x127b(%rip),%ymm3 # 25c8 <_sk_xor__hsw_lowp+0x368>
- .byte 197,253,111,200 // vmovdqa %ymm0,%ymm1
- .byte 197,253,111,208 // vmovdqa %ymm0,%ymm2
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,191 // ja 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,193,0,0,0 // lea 0xc1(%rip),%r11 # 1438 <_sk_load_g8_hsw_lowp+0x12c>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,249,110,192 // vmovd %eax,%xmm0
- .byte 235,160 // jmp 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm0,%xmm0
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,249,110,200 // vmovd %eax,%xmm1
- .byte 196,227,121,14,193,1 // vpblendw $0x1,%xmm1,%xmm0,%xmm0
- .byte 235,131 // jmp 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,121,110,12,25 // vmovd (%r9,%rbx,1),%xmm1
- .byte 196,227,121,2,193,1 // vpblendd $0x1,%xmm1,%xmm0,%xmm0
- .byte 233,86,255,255,255 // jmpq 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm1
- .byte 196,227,113,2,192,12 // vpblendd $0xc,%xmm0,%xmm1,%xmm0
- .byte 233,41,255,255,255 // jmpq 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
- .byte 196,195,121,32,68,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,195,121,32,68,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm0,%xmm0
- .byte 196,193,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm1
- .byte 196,195,113,34,76,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm1,%xmm1
- .byte 196,227,113,2,192,8 // vpblendd $0x8,%xmm0,%xmm1,%xmm0
- .byte 233,244,254,255,255 // jmpq 132b <_sk_load_g8_hsw_lowp+0x1f>
- .byte 144 // nop
- .byte 72,255 // rex.W (bad)
- .byte 255 // (bad)
- .byte 255,95,255 // lcall *-0x1(%rdi)
- .byte 255 // (bad)
- .byte 255,83,255 // callq *-0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,140,255,255,255,132,255 // decl -0x7b0001(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 124,255 // jl 144d <_sk_load_g8_hsw_lowp+0x141>
- .byte 255 // (bad)
- .byte 255,112,255 // pushq -0x1(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 185,255,255,255,177 // mov $0xb1ffffff,%ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,169,255,255,255,157 // ljmp *-0x62000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,230 // jmpq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 222,255 // fdivrp %st,%st(7)
- .byte 255 // (bad)
- .byte 255,214 // callq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,202 // dec %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_g8_dst_hsw_lowp
-.globl _sk_load_g8_dst_hsw_lowp
-FUNCTION(_sk_load_g8_dst_hsw_lowp)
-_sk_load_g8_dst_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,51 // jne 14c0 <_sk_load_g8_dst_hsw_lowp+0x4c>
- .byte 196,193,122,111,36,25 // vmovdqu (%r9,%rbx,1),%xmm4
- .byte 196,226,125,48,228 // vpmovzxbw %xmm4,%ymm4
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,226,125,121,45,36,17,0,0 // vpbroadcastw 0x1124(%rip),%ymm5 # 25ca <_sk_xor__hsw_lowp+0x36a>
- .byte 197,221,228,229 // vpmulhuw %ymm5,%ymm4,%ymm4
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 196,226,125,121,61,23,17,0,0 // vpbroadcastw 0x1117(%rip),%ymm7 # 25cc <_sk_xor__hsw_lowp+0x36c>
- .byte 197,253,111,236 // vmovdqa %ymm4,%ymm5
- .byte 197,253,111,244 // vmovdqa %ymm4,%ymm6
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,191 // ja 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,193,0,0,0 // lea 0xc1(%rip),%r11 # 15a0 <_sk_load_g8_dst_hsw_lowp+0x12c>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,249,110,224 // vmovd %eax,%xmm4
- .byte 235,160 // jmp 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm4,%xmm4
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,249,110,232 // vmovd %eax,%xmm5
- .byte 196,227,89,14,229,1 // vpblendw $0x1,%xmm5,%xmm4,%xmm4
- .byte 235,131 // jmp 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,121,110,44,25 // vmovd (%r9,%rbx,1),%xmm5
- .byte 196,227,89,2,229,1 // vpblendd $0x1,%xmm5,%xmm4,%xmm4
- .byte 233,86,255,255,255 // jmpq 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,122,126,44,25 // vmovq (%r9,%rbx,1),%xmm5
- .byte 196,227,81,2,228,12 // vpblendd $0xc,%xmm4,%xmm5,%xmm4
- .byte 233,41,255,255,255 // jmpq 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 197,217,239,228 // vpxor %xmm4,%xmm4,%xmm4
- .byte 196,195,89,32,100,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,195,89,32,100,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm4,%xmm4
- .byte 196,193,122,126,44,25 // vmovq (%r9,%rbx,1),%xmm5
- .byte 196,195,81,34,108,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm5,%xmm5
- .byte 196,227,81,2,228,8 // vpblendd $0x8,%xmm4,%xmm5,%xmm4
- .byte 233,244,254,255,255 // jmpq 1493 <_sk_load_g8_dst_hsw_lowp+0x1f>
- .byte 144 // nop
- .byte 72,255 // rex.W (bad)
- .byte 255 // (bad)
- .byte 255,95,255 // lcall *-0x1(%rdi)
- .byte 255 // (bad)
- .byte 255,83,255 // callq *-0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,140,255,255,255,132,255 // decl -0x7b0001(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 124,255 // jl 15b5 <_sk_load_g8_dst_hsw_lowp+0x141>
- .byte 255 // (bad)
- .byte 255,112,255 // pushq -0x1(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 185,255,255,255,177 // mov $0xb1ffffff,%ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,169,255,255,255,157 // ljmp *-0x62000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,230 // jmpq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 222,255 // fdivrp %st,%st(7)
- .byte 255 // (bad)
- .byte 255,214 // callq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,202 // dec %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_srcover_rgba_8888_hsw_lowp
-.globl _sk_srcover_rgba_8888_hsw_lowp
-FUNCTION(_sk_srcover_rgba_8888_hsw_lowp)
-_sk_srcover_rgba_8888_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,221,1,0,0 // jne 17da <_sk_srcover_rgba_8888_hsw_lowp+0x1fe>
- .byte 196,193,126,111,124,153,32 // vmovdqu 0x20(%r9,%rbx,4),%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 197,253,111,37,206,15,0,0 // vmovdqa 0xfce(%rip),%ymm4 # 25e0 <_sk_xor__hsw_lowp+0x380>
- .byte 196,226,61,0,236 // vpshufb %ymm4,%ymm8,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,226,69,0,228 // vpshufb %ymm4,%ymm7,%ymm4
- .byte 196,227,253,0,228,232 // vpermq $0xe8,%ymm4,%ymm4
- .byte 196,227,85,56,228,1 // vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- .byte 196,98,125,121,13,201,15,0,0 // vpbroadcastw 0xfc9(%rip),%ymm9 # 2600 <_sk_xor__hsw_lowp+0x3a0>
- .byte 197,221,113,244,8 // vpsllw $0x8,%ymm4,%ymm4
- .byte 196,98,125,121,21,189,15,0,0 // vpbroadcastw 0xfbd(%rip),%ymm10 # 2602 <_sk_xor__hsw_lowp+0x3a2>
- .byte 196,193,93,228,226 // vpmulhuw %ymm10,%ymm4,%ymm4
- .byte 197,253,111,45,206,15,0,0 // vmovdqa 0xfce(%rip),%ymm5 # 2620 <_sk_xor__hsw_lowp+0x3c0>
- .byte 196,226,61,0,245 // vpshufb %ymm5,%ymm8,%ymm6
- .byte 196,227,253,0,246,232 // vpermq $0xe8,%ymm6,%ymm6
- .byte 196,226,69,0,237 // vpshufb %ymm5,%ymm7,%ymm5
- .byte 196,227,253,0,237,232 // vpermq $0xe8,%ymm5,%ymm5
- .byte 196,227,77,56,237,1 // vinserti128 $0x1,%xmm5,%ymm6,%ymm5
- .byte 197,213,113,245,8 // vpsllw $0x8,%ymm5,%ymm5
- .byte 196,193,85,228,234 // vpmulhuw %ymm10,%ymm5,%ymm5
- .byte 197,253,111,53,192,15,0,0 // vmovdqa 0xfc0(%rip),%ymm6 # 2640 <_sk_xor__hsw_lowp+0x3e0>
- .byte 196,98,61,0,222 // vpshufb %ymm6,%ymm8,%ymm11
- .byte 196,67,253,0,219,232 // vpermq $0xe8,%ymm11,%ymm11
- .byte 196,226,69,0,246 // vpshufb %ymm6,%ymm7,%ymm6
- .byte 196,227,253,0,246,232 // vpermq $0xe8,%ymm6,%ymm6
- .byte 196,227,37,56,246,1 // vinserti128 $0x1,%xmm6,%ymm11,%ymm6
- .byte 197,205,113,246,8 // vpsllw $0x8,%ymm6,%ymm6
- .byte 196,193,77,228,242 // vpmulhuw %ymm10,%ymm6,%ymm6
- .byte 197,125,111,29,178,15,0,0 // vmovdqa 0xfb2(%rip),%ymm11 # 2660 <_sk_xor__hsw_lowp+0x400>
- .byte 196,66,61,0,195 // vpshufb %ymm11,%ymm8,%ymm8
- .byte 196,67,253,0,192,232 // vpermq $0xe8,%ymm8,%ymm8
- .byte 196,194,69,0,251 // vpshufb %ymm11,%ymm7,%ymm7
- .byte 196,227,253,0,255,232 // vpermq $0xe8,%ymm7,%ymm7
- .byte 196,227,61,56,255,1 // vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- .byte 197,197,113,247,8 // vpsllw $0x8,%ymm7,%ymm7
- .byte 196,193,69,228,250 // vpmulhuw %ymm10,%ymm7,%ymm7
- .byte 196,98,125,121,5,163,15,0,0 // vpbroadcastw 0xfa3(%rip),%ymm8 # 2680 <_sk_xor__hsw_lowp+0x420>
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,66,93,11,208 // vpmulhrsw %ymm8,%ymm4,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,192 // vpaddw %ymm0,%ymm10,%ymm0
- .byte 196,66,85,11,208 // vpmulhrsw %ymm8,%ymm5,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,201 // vpaddw %ymm1,%ymm10,%ymm1
- .byte 196,66,77,11,208 // vpmulhrsw %ymm8,%ymm6,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,210 // vpaddw %ymm2,%ymm10,%ymm2
- .byte 196,66,69,11,192 // vpmulhrsw %ymm8,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 197,189,113,208,7 // vpsrlw $0x7,%ymm0,%ymm8
- .byte 196,65,61,234,193 // vpminsw %ymm9,%ymm8,%ymm8
- .byte 196,66,125,51,208 // vpmovzxwd %xmm8,%ymm10
- .byte 196,67,125,57,192,1 // vextracti128 $0x1,%ymm8,%xmm8
- .byte 196,66,125,51,192 // vpmovzxwd %xmm8,%ymm8
- .byte 197,165,113,209,7 // vpsrlw $0x7,%ymm1,%ymm11
- .byte 196,65,37,234,217 // vpminsw %ymm9,%ymm11,%ymm11
- .byte 196,67,125,57,220,1 // vextracti128 $0x1,%ymm11,%xmm12
- .byte 196,66,125,51,228 // vpmovzxwd %xmm12,%ymm12
- .byte 196,66,125,51,219 // vpmovzxwd %xmm11,%ymm11
- .byte 196,193,37,114,243,8 // vpslld $0x8,%ymm11,%ymm11
- .byte 196,193,29,114,244,8 // vpslld $0x8,%ymm12,%ymm12
- .byte 197,149,113,210,7 // vpsrlw $0x7,%ymm2,%ymm13
- .byte 196,65,21,234,233 // vpminsw %ymm9,%ymm13,%ymm13
- .byte 196,66,125,51,245 // vpmovzxwd %xmm13,%ymm14
- .byte 196,67,125,57,237,1 // vextracti128 $0x1,%ymm13,%xmm13
- .byte 196,66,125,51,237 // vpmovzxwd %xmm13,%ymm13
- .byte 196,193,21,114,245,16 // vpslld $0x10,%ymm13,%ymm13
- .byte 196,193,13,114,246,16 // vpslld $0x10,%ymm14,%ymm14
- .byte 197,133,113,211,7 // vpsrlw $0x7,%ymm3,%ymm15
- .byte 196,65,5,234,201 // vpminsw %ymm9,%ymm15,%ymm9
- .byte 196,67,125,57,207,1 // vextracti128 $0x1,%ymm9,%xmm15
- .byte 196,66,125,51,255 // vpmovzxwd %xmm15,%ymm15
- .byte 196,66,125,51,201 // vpmovzxwd %xmm9,%ymm9
- .byte 196,193,53,114,241,24 // vpslld $0x18,%ymm9,%ymm9
- .byte 196,193,5,114,247,24 // vpslld $0x18,%ymm15,%ymm15
- .byte 196,65,29,235,192 // vpor %ymm8,%ymm12,%ymm8
- .byte 196,65,37,235,218 // vpor %ymm10,%ymm11,%ymm11
- .byte 196,65,21,235,215 // vpor %ymm15,%ymm13,%ymm10
- .byte 196,65,61,235,210 // vpor %ymm10,%ymm8,%ymm10
- .byte 196,65,13,235,193 // vpor %ymm9,%ymm14,%ymm8
- .byte 196,65,37,235,192 // vpor %ymm8,%ymm11,%ymm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,78 // jne 1816 <_sk_srcover_rgba_8888_hsw_lowp+0x23a>
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,126,127,84,153,32 // vmovdqu %ymm10,0x20(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,65,61,239,192 // vpxor %ymm8,%ymm8,%ymm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,19,254,255,255 // ja 160a <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,238,1,0,0 // lea 0x1ee(%rip),%r11 # 19f0 <_sk_srcover_rgba_8888_hsw_lowp+0x414>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,110,4,153 // vmovd (%r9,%rbx,4),%xmm8
- .byte 233,244,253,255,255 // jmpq 160a <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,175 // ja 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 65,15,182,194 // movzbl %r10b,%eax
- .byte 76,141,21,251,1,0,0 // lea 0x1fb(%rip),%r10 # 1a2c <_sk_srcover_rgba_8888_hsw_lowp+0x450>
- .byte 73,99,4,130 // movslq (%r10,%rax,4),%rax
- .byte 76,1,208 // add %r10,%rax
- .byte 255,224 // jmpq *%rax
- .byte 196,65,121,126,4,153 // vmovd %xmm8,(%r9,%rbx,4)
- .byte 235,147 // jmp 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 196,193,121,110,100,153,8 // vmovd 0x8(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,4 // vpblendd $0x4,%ymm4,%ymm7,%ymm8
- .byte 196,194,121,53,36,153 // vpmovzxdq (%r9,%rbx,4),%xmm4
- .byte 197,249,112,228,232 // vpshufd $0xe8,%xmm4,%xmm4
- .byte 196,99,61,2,196,3 // vpblendd $0x3,%ymm4,%ymm8,%ymm8
- .byte 233,156,253,255,255 // jmpq 160a <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,24 // vmovd 0x18(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,197,239,255 // vpxor %ymm7,%ymm7,%ymm7
- .byte 196,99,69,2,196,64 // vpblendd $0x40,%ymm4,%ymm7,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,20,1 // vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,99,125,57,196,1 // vextracti128 $0x1,%ymm8,%xmm4
- .byte 196,195,89,34,100,153,16,0 // vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,99,61,56,196,1 // vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- .byte 196,193,122,111,36,153 // vmovdqu (%r9,%rbx,4),%xmm4
- .byte 196,67,93,2,192,240 // vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- .byte 233,77,253,255,255 // jmpq 160a <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- .byte 196,193,121,110,100,153,40 // vmovd 0x28(%r9,%rbx,4),%xmm4
- .byte 196,226,121,89,228 // vpbroadcastq %xmm4,%xmm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,4 // vpblendd $0x4,%ymm4,%ymm5,%ymm7
- .byte 196,195,65,34,100,153,36,1 // vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- .byte 196,227,69,2,252,15 // vpblendd $0xf,%ymm4,%ymm7,%ymm7
- .byte 196,193,121,110,100,153,32 // vmovd 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,69,2,252,1 // vpblendd $0x1,%ymm4,%ymm7,%ymm7
- .byte 233,17,253,255,255 // jmpq 1604 <_sk_srcover_rgba_8888_hsw_lowp+0x28>
- .byte 196,193,121,110,100,153,56 // vmovd 0x38(%r9,%rbx,4),%xmm4
- .byte 196,226,125,89,228 // vpbroadcastq %xmm4,%ymm4
- .byte 197,213,239,237 // vpxor %ymm5,%ymm5,%ymm5
- .byte 196,227,85,2,252,64 // vpblendd $0x40,%ymm4,%ymm5,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,52,1 // vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,227,125,57,252,1 // vextracti128 $0x1,%ymm7,%xmm4
- .byte 196,195,89,34,100,153,48,0 // vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- .byte 196,227,69,56,252,1 // vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- .byte 196,65,126,111,4,153 // vmovdqu (%r9,%rbx,4),%ymm8
- .byte 196,193,122,111,100,153,32 // vmovdqu 0x20(%r9,%rbx,4),%xmm4
- .byte 196,227,93,2,255,240 // vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- .byte 233,193,252,255,255 // jmpq 160a <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- .byte 196,67,121,22,68,153,8,2 // vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- .byte 196,65,121,214,4,153 // vmovq %xmm8,(%r9,%rbx,4)
- .byte 233,121,254,255,255 // jmpq 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,24,2 // vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,67,121,22,76,153,20,1 // vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
- .byte 196,65,121,126,76,153,16 // vmovd %xmm9,0x10(%r9,%rbx,4)
- .byte 196,65,122,127,4,153 // vmovdqu %xmm8,(%r9,%rbx,4)
- .byte 233,69,254,255,255 // jmpq 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 196,67,121,22,84,153,40,2 // vpextrd $0x2,%xmm10,0x28(%r9,%rbx,4)
- .byte 196,67,121,22,84,153,36,1 // vpextrd $0x1,%xmm10,0x24(%r9,%rbx,4)
- .byte 196,65,121,126,84,153,32 // vmovd %xmm10,0x20(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 233,35,254,255,255 // jmpq 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,67,121,22,76,153,56,2 // vpextrd $0x2,%xmm9,0x38(%r9,%rbx,4)
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,67,121,22,76,153,52,1 // vpextrd $0x1,%xmm9,0x34(%r9,%rbx,4)
- .byte 196,67,125,57,209,1 // vextracti128 $0x1,%ymm10,%xmm9
- .byte 196,65,121,126,76,153,48 // vmovd %xmm9,0x30(%r9,%rbx,4)
- .byte 196,65,126,127,4,153 // vmovdqu %ymm8,(%r9,%rbx,4)
- .byte 196,65,122,127,84,153,32 // vmovdqu %xmm10,0x20(%r9,%rbx,4)
- .byte 233,232,253,255,255 // jmpq 17d5 <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- .byte 15,31,0 // nopl (%rax)
- .byte 27,254 // sbb %esi,%edi
- .byte 255 // (bad)
- .byte 255,104,254 // ljmp *-0x2(%rax)
- .byte 255 // (bad)
- .byte 255,82,254 // callq *-0x2(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 188,254,255,255,168 // mov $0xa8fffffe,%esp
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,148,254,255,255,126,254 // callq *-0x1810001(%rsi,%rdi,8)
- .byte 255 // (bad)
- .byte 255,20,252 // callq *(%rsp,%rdi,8)
- .byte 255 // (bad)
- .byte 255,241 // push %rcx
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,227 // jmpq *%rbx
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,205 // dec %ebp
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,65,255 // incl -0x1(%rcx)
- .byte 255 // (bad)
- .byte 255,45,255,255,255,25 // ljmp *0x19ffffff(%rip) # 1a001a24 <_sk_xor__hsw_lowp+0x19fff7c4>
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,14 // decl (%rsi)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255,37,255,255,255,29 // jmpq *0x1dffffff(%rip) # 1e001a34 <_sk_xor__hsw_lowp+0x1dfff7d4>
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,89,255 // lcall *-0x1(%rcx)
- .byte 255 // (bad)
- .byte 255,76,255,255 // decl -0x1(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 62,255 // ds (bad)
- .byte 255 // (bad)
- .byte 255,48 // pushq (%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 123,255 // jnp 1a49 <_sk_srcover_rgba_8888_hsw_lowp+0x46d>
- .byte 255 // (bad)
- .byte 255,116,255,255 // pushq -0x1(%rdi,%rdi,8)
- .byte 255,108,255,255 // ljmp *-0x1(%rdi,%rdi,8)
- .byte 255,100,255,255 // jmpq *-0x1(%rdi,%rdi,8)
- .byte 255,175,255,255,255,162 // ljmp *-0x5d000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,148,255,255,255,134,255 // callq *-0x790001(%rdi,%rdi,8)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_scale_1_float_hsw_lowp
-.globl _sk_scale_1_float_hsw_lowp
-FUNCTION(_sk_scale_1_float_hsw_lowp)
-_sk_scale_1_float_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,122,16,0 // vmovss (%rax),%xmm8
- .byte 197,58,88,5,106,8,0,0 // vaddss 0x86a(%rip),%xmm8,%xmm8 # 22e0 <_sk_xor__hsw_lowp+0x80>
- .byte 197,121,126,192 // vmovd %xmm8,%eax
- .byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 196,66,125,121,192 // vpbroadcastw %xmm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_scale_u8_hsw_lowp
-.globl _sk_scale_u8_hsw_lowp
-FUNCTION(_sk_scale_u8_hsw_lowp)
-_sk_scale_u8_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,76 // jne 1b14 <_sk_scale_u8_hsw_lowp+0x65>
- .byte 196,65,122,111,4,25 // vmovdqu (%r9,%rbx,1),%xmm8
- .byte 196,66,125,48,192 // vpmovzxbw %xmm8,%ymm8
- .byte 196,193,61,113,240,8 // vpsllw $0x8,%ymm8,%ymm8
- .byte 196,98,125,121,13,160,11,0,0 // vpbroadcastw 0xba0(%rip),%ymm9 # 2682 <_sk_xor__hsw_lowp+0x422>
- .byte 196,65,61,228,193 // vpmulhuw %ymm9,%ymm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 119,165 // ja 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,200,0,0,0 // lea 0xc8(%rip),%r11 # 1bfc <_sk_scale_u8_hsw_lowp+0x14d>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 235,134 // jmp 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm8,%xmm8
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,121,110,200 // vmovd %eax,%xmm9
- .byte 196,67,57,14,193,1 // vpblendw $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,101,255,255,255 // jmpq 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,121,110,12,25 // vmovd (%r9,%rbx,1),%xmm9
- .byte 196,67,57,2,193,1 // vpblendd $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,55,255,255,255 // jmpq 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm9
- .byte 196,67,49,2,192,12 // vpblendd $0xc,%xmm8,%xmm9,%xmm8
- .byte 233,9,255,255,255 // jmpq 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm9
- .byte 196,67,49,34,76,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm9,%xmm9
- .byte 196,67,49,2,192,8 // vpblendd $0x8,%xmm8,%xmm9,%xmm8
- .byte 233,211,254,255,255 // jmpq 1ace <_sk_scale_u8_hsw_lowp+0x1f>
- .byte 144 // nop
- .byte 65,255 // rex.B (bad)
- .byte 255 // (bad)
- .byte 255,89,255 // lcall *-0x1(%rcx)
- .byte 255 // (bad)
- .byte 255,76,255,255 // decl -0x1(%rdi,%rdi,8)
- .byte 255,138,255,255,255,130 // decl -0x7d000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 122,255 // jp 1c11 <_sk_scale_u8_hsw_lowp+0x162>
- .byte 255 // (bad)
- .byte 255,109,255 // ljmp *-0x1(%rbp)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 184,255,255,255,176 // mov $0xb0ffffff,%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,168,255,255,255,155 // ljmp *-0x64000001(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,230 // jmpq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 222,255 // fdivrp %st,%st(7)
- .byte 255 // (bad)
- .byte 255,214 // callq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,201 // dec %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_lerp_1_float_hsw_lowp
-.globl _sk_lerp_1_float_hsw_lowp
-FUNCTION(_sk_lerp_1_float_hsw_lowp)
-_sk_lerp_1_float_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,122,16,0 // vmovss (%rax),%xmm8
- .byte 197,58,88,5,158,6,0,0 // vaddss 0x69e(%rip),%xmm8,%xmm8 # 22e4 <_sk_xor__hsw_lowp+0x84>
- .byte 197,121,126,192 // vmovd %xmm8,%eax
- .byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 196,66,125,121,192 // vpbroadcastw %xmm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,98,125,121,13,30,10,0,0 // vpbroadcastw 0xa1e(%rip),%ymm9 # 2684 <_sk_xor__hsw_lowp+0x424>
- .byte 196,65,53,249,200 // vpsubw %ymm8,%ymm9,%ymm9
- .byte 196,66,93,11,209 // vpmulhrsw %ymm9,%ymm4,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,192 // vpaddw %ymm0,%ymm10,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,66,85,11,209 // vpmulhrsw %ymm9,%ymm5,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,201 // vpaddw %ymm1,%ymm10,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,66,77,11,209 // vpmulhrsw %ymm9,%ymm6,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,210 // vpaddw %ymm2,%ymm10,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 196,66,69,11,193 // vpmulhrsw %ymm9,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_lerp_u8_hsw_lowp
-.globl _sk_lerp_u8_hsw_lowp
-FUNCTION(_sk_lerp_u8_hsw_lowp)
-_sk_lerp_u8_hsw_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,146,0,0,0 // jne 1d74 <_sk_lerp_u8_hsw_lowp+0xaf>
- .byte 196,65,122,111,4,25 // vmovdqu (%r9,%rbx,1),%xmm8
- .byte 196,66,125,48,192 // vpmovzxbw %xmm8,%ymm8
- .byte 196,193,61,113,240,8 // vpsllw $0x8,%ymm8,%ymm8
- .byte 196,98,125,121,13,138,9,0,0 // vpbroadcastw 0x98a(%rip),%ymm9 # 2686 <_sk_xor__hsw_lowp+0x426>
- .byte 196,65,61,228,193 // vpmulhuw %ymm9,%ymm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,98,125,121,13,116,9,0,0 // vpbroadcastw 0x974(%rip),%ymm9 # 2688 <_sk_xor__hsw_lowp+0x428>
- .byte 196,65,53,249,200 // vpsubw %ymm8,%ymm9,%ymm9
- .byte 196,66,93,11,209 // vpmulhrsw %ymm9,%ymm4,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,192 // vpaddw %ymm0,%ymm10,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,66,85,11,209 // vpmulhrsw %ymm9,%ymm5,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,201 // vpaddw %ymm1,%ymm10,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,66,77,11,209 // vpmulhrsw %ymm9,%ymm6,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,210 // vpaddw %ymm2,%ymm10,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 196,66,69,11,193 // vpmulhrsw %ymm9,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,15 // and $0xf,%r10b
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,14 // cmp $0xe,%r10b
- .byte 15,135,91,255,255,255 // ja 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,204,0,0,0 // lea 0xcc(%rip),%r11 # 1e64 <_sk_lerp_u8_hsw_lowp+0x19f>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 197,121,110,192 // vmovd %eax,%xmm8
- .byte 233,57,255,255,255 // jmpq 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,2,2 // vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm8,%xmm8
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 197,121,110,200 // vmovd %eax,%xmm9
- .byte 196,67,57,14,193,1 // vpblendw $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,24,255,255,255 // jmpq 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,6,6 // vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,5,5 // vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,4,4 // vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,121,110,12,25 // vmovd (%r9,%rbx,1),%xmm9
- .byte 196,67,57,2,193,1 // vpblendd $0x1,%xmm9,%xmm8,%xmm8
- .byte 233,234,254,255,255 // jmpq 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,10,10 // vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,9,9 // vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,8,8 // vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm9
- .byte 196,67,49,2,192,12 // vpblendd $0xc,%xmm8,%xmm9,%xmm8
- .byte 233,188,254,255,255 // jmpq 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
- .byte 196,67,57,32,68,25,14,14 // vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,13,13 // vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,67,57,32,68,25,12,12 // vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm8,%xmm8
- .byte 196,65,122,126,12,25 // vmovq (%r9,%rbx,1),%xmm9
- .byte 196,67,49,34,76,25,8,2 // vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm9,%xmm9
- .byte 196,67,49,2,192,8 // vpblendd $0x8,%xmm8,%xmm9,%xmm8
- .byte 233,134,254,255,255 // jmpq 1ce8 <_sk_lerp_u8_hsw_lowp+0x23>
- .byte 102,144 // xchg %ax,%ax
- .byte 61,255,255,255,88 // cmp $0x58ffffff,%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,75,255 // decl -0x1(%rbx)
- .byte 255 // (bad)
- .byte 255,137,255,255,255,129 // decl -0x7e000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 121,255 // jns 1e79 <_sk_lerp_u8_hsw_lowp+0x1b4>
- .byte 255 // (bad)
- .byte 255,108,255,255 // ljmp *-0x1(%rdi,%rdi,8)
- .byte 255,183,255,255,255,175 // pushq -0x50000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,167,255,255,255,154 // jmpq *-0x65000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,229 // jmpq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 221,255 // (bad)
- .byte 255 // (bad)
- .byte 255,213 // callq *%rbp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,200 // dec %eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_swap_rb_hsw_lowp
-.globl _sk_swap_rb_hsw_lowp
-FUNCTION(_sk_swap_rb_hsw_lowp)
-_sk_swap_rb_hsw_lowp:
- .byte 197,124,40,192 // vmovaps %ymm0,%ymm8
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,252,40,194 // vmovaps %ymm2,%ymm0
- .byte 197,124,41,194 // vmovaps %ymm8,%ymm2
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_move_src_dst_hsw_lowp
-.globl _sk_move_src_dst_hsw_lowp
-FUNCTION(_sk_move_src_dst_hsw_lowp)
-_sk_move_src_dst_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,252,40,224 // vmovaps %ymm0,%ymm4
- .byte 197,252,40,233 // vmovaps %ymm1,%ymm5
- .byte 197,252,40,242 // vmovaps %ymm2,%ymm6
- .byte 197,252,40,251 // vmovaps %ymm3,%ymm7
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_move_dst_src_hsw_lowp
-.globl _sk_move_dst_src_hsw_lowp
-FUNCTION(_sk_move_dst_src_hsw_lowp)
-_sk_move_dst_src_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,252,40,196 // vmovaps %ymm4,%ymm0
- .byte 197,252,40,205 // vmovaps %ymm5,%ymm1
- .byte 197,252,40,214 // vmovaps %ymm6,%ymm2
- .byte 197,252,40,223 // vmovaps %ymm7,%ymm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_clear_hsw_lowp
-.globl _sk_clear_hsw_lowp
-FUNCTION(_sk_clear_hsw_lowp)
-_sk_clear_hsw_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
- .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
- .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
- .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcatop_hsw_lowp
-.globl _sk_srcatop_hsw_lowp
-FUNCTION(_sk_srcatop_hsw_lowp)
-_sk_srcatop_hsw_lowp:
- .byte 196,226,125,11,199 // vpmulhrsw %ymm7,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,98,125,121,5,139,7,0,0 // vpbroadcastw 0x78b(%rip),%ymm8 # 268a <_sk_xor__hsw_lowp+0x42a>
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,66,93,11,200 // vpmulhrsw %ymm8,%ymm4,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,192 // vpaddw %ymm0,%ymm9,%ymm0
- .byte 196,226,117,11,207 // vpmulhrsw %ymm7,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,66,85,11,200 // vpmulhrsw %ymm8,%ymm5,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,201 // vpaddw %ymm1,%ymm9,%ymm1
- .byte 196,226,109,11,215 // vpmulhrsw %ymm7,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,66,77,11,200 // vpmulhrsw %ymm8,%ymm6,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,210 // vpaddw %ymm2,%ymm9,%ymm2
- .byte 196,226,101,11,223 // vpmulhrsw %ymm7,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 196,66,69,11,192 // vpmulhrsw %ymm8,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstatop_hsw_lowp
-.globl _sk_dstatop_hsw_lowp
-FUNCTION(_sk_dstatop_hsw_lowp)
-_sk_dstatop_hsw_lowp:
- .byte 196,98,93,11,195 // vpmulhrsw %ymm3,%ymm4,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 196,98,125,121,13,28,7,0,0 // vpbroadcastw 0x71c(%rip),%ymm9 # 268c <_sk_xor__hsw_lowp+0x42c>
- .byte 197,53,249,207 // vpsubw %ymm7,%ymm9,%ymm9
- .byte 196,194,125,11,193 // vpmulhrsw %ymm9,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,193,125,253,192 // vpaddw %ymm8,%ymm0,%ymm0
- .byte 196,98,85,11,195 // vpmulhrsw %ymm3,%ymm5,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 196,194,117,11,201 // vpmulhrsw %ymm9,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,193,117,253,200 // vpaddw %ymm8,%ymm1,%ymm1
- .byte 196,98,77,11,195 // vpmulhrsw %ymm3,%ymm6,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 196,194,109,11,209 // vpmulhrsw %ymm9,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,193,109,253,208 // vpaddw %ymm8,%ymm2,%ymm2
- .byte 196,98,69,11,195 // vpmulhrsw %ymm3,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 196,194,101,11,217 // vpmulhrsw %ymm9,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 196,193,101,253,216 // vpaddw %ymm8,%ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcin_hsw_lowp
-.globl _sk_srcin_hsw_lowp
-FUNCTION(_sk_srcin_hsw_lowp)
-_sk_srcin_hsw_lowp:
- .byte 196,226,125,11,199 // vpmulhrsw %ymm7,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,226,117,11,207 // vpmulhrsw %ymm7,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,226,109,11,215 // vpmulhrsw %ymm7,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,226,101,11,223 // vpmulhrsw %ymm7,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstin_hsw_lowp
-.globl _sk_dstin_hsw_lowp
-FUNCTION(_sk_dstin_hsw_lowp)
-_sk_dstin_hsw_lowp:
- .byte 196,226,93,11,195 // vpmulhrsw %ymm3,%ymm4,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,226,85,11,203 // vpmulhrsw %ymm3,%ymm5,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,226,77,11,211 // vpmulhrsw %ymm3,%ymm6,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,226,69,11,219 // vpmulhrsw %ymm3,%ymm7,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcout_hsw_lowp
-.globl _sk_srcout_hsw_lowp
-FUNCTION(_sk_srcout_hsw_lowp)
-_sk_srcout_hsw_lowp:
- .byte 196,98,125,121,5,91,6,0,0 // vpbroadcastw 0x65b(%rip),%ymm8 # 268e <_sk_xor__hsw_lowp+0x42e>
- .byte 197,61,249,199 // vpsubw %ymm7,%ymm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstout_hsw_lowp
-.globl _sk_dstout_hsw_lowp
-FUNCTION(_sk_dstout_hsw_lowp)
-_sk_dstout_hsw_lowp:
- .byte 196,226,125,121,5,36,6,0,0 // vpbroadcastw 0x624(%rip),%ymm0 # 2690 <_sk_xor__hsw_lowp+0x430>
- .byte 197,253,249,219 // vpsubw %ymm3,%ymm0,%ymm3
- .byte 196,226,93,11,195 // vpmulhrsw %ymm3,%ymm4,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,226,85,11,203 // vpmulhrsw %ymm3,%ymm5,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,226,77,11,211 // vpmulhrsw %ymm3,%ymm6,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,226,69,11,219 // vpmulhrsw %ymm3,%ymm7,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcover_hsw_lowp
-.globl _sk_srcover_hsw_lowp
-FUNCTION(_sk_srcover_hsw_lowp)
-_sk_srcover_hsw_lowp:
- .byte 196,98,125,121,5,237,5,0,0 // vpbroadcastw 0x5ed(%rip),%ymm8 # 2692 <_sk_xor__hsw_lowp+0x432>
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,66,93,11,200 // vpmulhrsw %ymm8,%ymm4,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,192 // vpaddw %ymm0,%ymm9,%ymm0
- .byte 196,66,85,11,200 // vpmulhrsw %ymm8,%ymm5,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,201 // vpaddw %ymm1,%ymm9,%ymm1
- .byte 196,66,77,11,200 // vpmulhrsw %ymm8,%ymm6,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,210 // vpaddw %ymm2,%ymm9,%ymm2
- .byte 196,66,69,11,192 // vpmulhrsw %ymm8,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstover_hsw_lowp
-.globl _sk_dstover_hsw_lowp
-FUNCTION(_sk_dstover_hsw_lowp)
-_sk_dstover_hsw_lowp:
- .byte 196,98,125,121,5,166,5,0,0 // vpbroadcastw 0x5a6(%rip),%ymm8 # 2694 <_sk_xor__hsw_lowp+0x434>
- .byte 197,61,249,199 // vpsubw %ymm7,%ymm8,%ymm8
- .byte 196,194,125,11,192 // vpmulhrsw %ymm8,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 197,253,253,196 // vpaddw %ymm4,%ymm0,%ymm0
- .byte 196,194,117,11,200 // vpmulhrsw %ymm8,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 197,245,253,205 // vpaddw %ymm5,%ymm1,%ymm1
- .byte 196,194,109,11,208 // vpmulhrsw %ymm8,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 197,237,253,214 // vpaddw %ymm6,%ymm2,%ymm2
- .byte 196,194,101,11,216 // vpmulhrsw %ymm8,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 197,229,253,223 // vpaddw %ymm7,%ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_modulate_hsw_lowp
-.globl _sk_modulate_hsw_lowp
-FUNCTION(_sk_modulate_hsw_lowp)
-_sk_modulate_hsw_lowp:
- .byte 196,226,125,11,196 // vpmulhrsw %ymm4,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 196,226,117,11,205 // vpmulhrsw %ymm5,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,226,109,11,214 // vpmulhrsw %ymm6,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,226,101,11,223 // vpmulhrsw %ymm7,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_multiply_hsw_lowp
-.globl _sk_multiply_hsw_lowp
-FUNCTION(_sk_multiply_hsw_lowp)
-_sk_multiply_hsw_lowp:
- .byte 196,98,125,121,5,51,5,0,0 // vpbroadcastw 0x533(%rip),%ymm8 # 2696 <_sk_xor__hsw_lowp+0x436>
- .byte 197,61,249,207 // vpsubw %ymm7,%ymm8,%ymm9
- .byte 196,66,125,11,209 // vpmulhrsw %ymm9,%ymm0,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,66,93,11,216 // vpmulhrsw %ymm8,%ymm4,%ymm11
- .byte 196,66,125,29,219 // vpabsw %ymm11,%ymm11
- .byte 196,65,37,253,210 // vpaddw %ymm10,%ymm11,%ymm10
- .byte 196,226,125,11,196 // vpmulhrsw %ymm4,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 197,173,253,192 // vpaddw %ymm0,%ymm10,%ymm0
- .byte 196,66,117,11,209 // vpmulhrsw %ymm9,%ymm1,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 196,66,85,11,216 // vpmulhrsw %ymm8,%ymm5,%ymm11
- .byte 196,66,125,29,219 // vpabsw %ymm11,%ymm11
- .byte 196,65,37,253,210 // vpaddw %ymm10,%ymm11,%ymm10
- .byte 196,226,117,11,205 // vpmulhrsw %ymm5,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 197,173,253,201 // vpaddw %ymm1,%ymm10,%ymm1
- .byte 196,66,109,11,209 // vpmulhrsw %ymm9,%ymm2,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 196,66,77,11,216 // vpmulhrsw %ymm8,%ymm6,%ymm11
- .byte 196,66,125,29,219 // vpabsw %ymm11,%ymm11
- .byte 196,65,37,253,210 // vpaddw %ymm10,%ymm11,%ymm10
- .byte 196,226,109,11,214 // vpmulhrsw %ymm6,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 197,173,253,210 // vpaddw %ymm2,%ymm10,%ymm2
- .byte 196,66,101,11,201 // vpmulhrsw %ymm9,%ymm3,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 196,66,69,11,192 // vpmulhrsw %ymm8,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 196,65,61,253,193 // vpaddw %ymm9,%ymm8,%ymm8
- .byte 196,226,101,11,223 // vpmulhrsw %ymm7,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_screen_hsw_lowp
-.globl _sk_screen_hsw_lowp
-FUNCTION(_sk_screen_hsw_lowp)
-_sk_screen_hsw_lowp:
- .byte 196,98,125,121,5,132,4,0,0 // vpbroadcastw 0x484(%rip),%ymm8 # 2698 <_sk_xor__hsw_lowp+0x438>
- .byte 197,61,249,200 // vpsubw %ymm0,%ymm8,%ymm9
- .byte 196,98,53,11,204 // vpmulhrsw %ymm4,%ymm9,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,192 // vpaddw %ymm0,%ymm9,%ymm0
- .byte 197,61,249,201 // vpsubw %ymm1,%ymm8,%ymm9
- .byte 196,98,53,11,205 // vpmulhrsw %ymm5,%ymm9,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,201 // vpaddw %ymm1,%ymm9,%ymm1
- .byte 197,61,249,202 // vpsubw %ymm2,%ymm8,%ymm9
- .byte 196,98,53,11,206 // vpmulhrsw %ymm6,%ymm9,%ymm9
- .byte 196,66,125,29,201 // vpabsw %ymm9,%ymm9
- .byte 197,181,253,210 // vpaddw %ymm2,%ymm9,%ymm2
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,98,61,11,199 // vpmulhrsw %ymm7,%ymm8,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_xor__hsw_lowp
-.globl _sk_xor__hsw_lowp
-FUNCTION(_sk_xor__hsw_lowp)
-_sk_xor__hsw_lowp:
- .byte 196,98,125,121,5,49,4,0,0 // vpbroadcastw 0x431(%rip),%ymm8 # 269a <_sk_xor__hsw_lowp+0x43a>
- .byte 197,61,249,207 // vpsubw %ymm7,%ymm8,%ymm9
- .byte 196,194,125,11,193 // vpmulhrsw %ymm9,%ymm0,%ymm0
- .byte 196,226,125,29,192 // vpabsw %ymm0,%ymm0
- .byte 197,61,249,195 // vpsubw %ymm3,%ymm8,%ymm8
- .byte 196,66,93,11,208 // vpmulhrsw %ymm8,%ymm4,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,192 // vpaddw %ymm0,%ymm10,%ymm0
- .byte 196,194,117,11,201 // vpmulhrsw %ymm9,%ymm1,%ymm1
- .byte 196,226,125,29,201 // vpabsw %ymm1,%ymm1
- .byte 196,66,85,11,208 // vpmulhrsw %ymm8,%ymm5,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,201 // vpaddw %ymm1,%ymm10,%ymm1
- .byte 196,194,109,11,209 // vpmulhrsw %ymm9,%ymm2,%ymm2
- .byte 196,226,125,29,210 // vpabsw %ymm2,%ymm2
- .byte 196,66,77,11,208 // vpmulhrsw %ymm8,%ymm6,%ymm10
- .byte 196,66,125,29,210 // vpabsw %ymm10,%ymm10
- .byte 197,173,253,210 // vpaddw %ymm2,%ymm10,%ymm2
- .byte 196,194,101,11,217 // vpmulhrsw %ymm9,%ymm3,%ymm3
- .byte 196,226,125,29,219 // vpabsw %ymm3,%ymm3
- .byte 196,66,69,11,192 // vpmulhrsw %ymm8,%ymm7,%ymm8
- .byte 196,66,125,29,192 // vpabsw %ymm8,%ymm8
- .byte 197,189,253,219 // vpaddw %ymm3,%ymm8,%ymm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-BALIGN4
- .byte 0,0 // add %al,(%rax)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128 // .byte 0x80
- .byte 67 // rex.XB
-
-BALIGN32
- .byte 0,128,0,128,0,0 // add %al,0x8000(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,0,1 // addb $0x1,(%rax)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,129,128,0,0,0,0 // addb $0x0,0x80(%rcx)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 1,2 // add %eax,(%rdx)
- .byte 5,6,9,10,13 // add $0xd0a0906,%eax
- .byte 14 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,17 // callq *(%rcx)
- .byte 18,21,22,25,26,29 // adc 0x1d1a1916(%rip),%dl # 1d1a3c8d <_sk_xor__hsw_lowp+0x1d1a1a2d>
- .byte 30 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,2 // incl (%rdx)
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,18 // callq *(%rdx)
- .byte 19,22 // adc (%rsi),%edx
- .byte 23 // (bad)
- .byte 26,27 // sbb (%rbx),%bl
- .byte 30 // (bad)
- .byte 31 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,19 // callq *(%rbx)
- .byte 255,23 // callq *(%rdi)
- .byte 255,27 // lcall *(%rbx)
- .byte 255,31 // lcall *(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 1,4,5,8,9,12,13 // add %eax,0xd0c0908(,%rax,1)
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,0,1 // addb $0x1,(%rax)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,129,128,0,0,0,0 // addb $0x0,0x80(%rcx)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 1,2 // add %eax,(%rdx)
- .byte 5,6,9,10,13 // add $0xd0a0906,%eax
- .byte 14 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,17 // callq *(%rcx)
- .byte 18,21,22,25,26,29 // adc 0x1d1a1916(%rip),%dl # 1d1a3d2d <_sk_xor__hsw_lowp+0x1d1a1acd>
- .byte 30 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,2 // incl (%rdx)
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,18 // callq *(%rdx)
- .byte 19,22 // adc (%rsi),%edx
- .byte 23 // (bad)
- .byte 26,27 // sbb (%rbx),%bl
- .byte 30 // (bad)
- .byte 31 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,19 // callq *(%rbx)
- .byte 255,23 // callq *(%rdi)
- .byte 255,27 // lcall *(%rbx)
- .byte 255,31 // lcall *(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,0,1 // addb $0x1,(%rax)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,129,128,0,0,0,0 // addb $0x0,0x80(%rcx)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 1,2 // add %eax,(%rdx)
- .byte 5,6,9,10,13 // add $0xd0a0906,%eax
- .byte 14 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,17 // callq *(%rcx)
- .byte 18,21,22,25,26,29 // adc 0x1d1a1916(%rip),%dl # 1d1a3ded <_sk_xor__hsw_lowp+0x1d1a1b8d>
- .byte 30 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,2 // incl (%rdx)
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,18 // callq *(%rdx)
- .byte 19,22 // adc (%rsi),%edx
- .byte 23 // (bad)
- .byte 26,27 // sbb (%rbx),%bl
- .byte 30 // (bad)
- .byte 31 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,19 // callq *(%rbx)
- .byte 255,23 // callq *(%rdi)
- .byte 255,27 // lcall *(%rbx)
- .byte 255,31 // lcall *(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 1,4,5,8,9,12,13 // add %eax,0xd0c0908(,%rax,1)
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,0,1 // addb $0x1,(%rax)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,129,128,0,0,0,0 // addb $0x0,0x80(%rcx)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 1,2 // add %eax,(%rdx)
- .byte 5,6,9,10,13 // add $0xd0a0906,%eax
- .byte 14 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,17 // callq *(%rcx)
- .byte 18,21,22,25,26,29 // adc 0x1d1a1916(%rip),%dl # 1d1a3e8d <_sk_xor__hsw_lowp+0x1d1a1c2d>
- .byte 30 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,2 // incl (%rdx)
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,18 // callq *(%rdx)
- .byte 19,22 // adc (%rsi),%edx
- .byte 23 // (bad)
- .byte 26,27 // sbb (%rbx),%bl
- .byte 30 // (bad)
- .byte 31 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,19 // callq *(%rbx)
- .byte 255,23 // callq *(%rdi)
- .byte 255,27 // lcall *(%rbx)
- .byte 255,31 // lcall *(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 129,128,129,128,129,128,0,128,129,128// addl $0x80818000,-0x7f7e7f7f(%rax)
- .byte 0,128,0,0,0,0 // add %al,0x0(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,0,1 // addb $0x1,(%rax)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 128,128,128,128,128,128,128 // addb $0x80,-0x7f7f7f80(%rax)
- .byte 128,255,0 // cmp $0x0,%bh
- .byte 129,128,0,0,0,0,0,0,0,0 // addl $0x0,0x0(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 1,2 // add %eax,(%rdx)
- .byte 5,6,9,10,13 // add $0xd0a0906,%eax
- .byte 14 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,17 // callq *(%rcx)
- .byte 18,21,22,25,26,29 // adc 0x1d1a1916(%rip),%dl # 1d1a3f4d <_sk_xor__hsw_lowp+0x1d1a1ced>
- .byte 30 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,2 // incl (%rdx)
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,18 // callq *(%rdx)
- .byte 19,22 // adc (%rsi),%edx
- .byte 23 // (bad)
- .byte 26,27 // sbb (%rbx),%bl
- .byte 30 // (bad)
- .byte 31 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,19 // callq *(%rbx)
- .byte 255,23 // callq *(%rdi)
- .byte 255,27 // lcall *(%rbx)
- .byte 255,31 // lcall *(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 128,129,128,0,128,129,128 // addb $0x80,-0x7e7fff80(%rcx)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0 // .byte 0x0
- .byte 128 // .byte 0x80
-BALIGN32
-
-HIDDEN _sk_start_pipeline_ssse3_lowp
-.globl _sk_start_pipeline_ssse3_lowp
-FUNCTION(_sk_start_pipeline_ssse3_lowp)
-_sk_start_pipeline_ssse3_lowp:
- .byte 85 // push %rbp
- .byte 72,137,229 // mov %rsp,%rbp
- .byte 65,87 // push %r15
- .byte 65,86 // push %r14
- .byte 65,85 // push %r13
- .byte 65,84 // push %r12
- .byte 83 // push %rbx
- .byte 72,131,236,40 // sub $0x28,%rsp
- .byte 72,137,211 // mov %rdx,%rbx
- .byte 73,137,246 // mov %rsi,%r14
- .byte 72,137,125,208 // mov %rdi,-0x30(%rbp)
- .byte 76,137,198 // mov %r8,%rsi
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 72,137,194 // mov %rax,%rdx
- .byte 72,137,77,176 // mov %rcx,-0x50(%rbp)
- .byte 73,57,206 // cmp %rcx,%r14
- .byte 115,117 // jae a1 <_sk_start_pipeline_ssse3_lowp+0xa1>
- .byte 72,139,69,208 // mov -0x30(%rbp),%rax
- .byte 72,141,64,8 // lea 0x8(%rax),%rax
- .byte 72,137,69,184 // mov %rax,-0x48(%rbp)
- .byte 76,137,77,192 // mov %r9,-0x40(%rbp)
- .byte 72,137,85,200 // mov %rdx,-0x38(%rbp)
- .byte 73,137,221 // mov %rbx,%r13
- .byte 73,137,244 // mov %rsi,%r12
- .byte 72,57,93,184 // cmp %rbx,-0x48(%rbp)
- .byte 72,139,85,208 // mov -0x30(%rbp),%rdx
- .byte 119,45 // ja 7d <_sk_start_pipeline_ssse3_lowp+0x7d>
- .byte 76,139,125,208 // mov -0x30(%rbp),%r15
- .byte 69,49,192 // xor %r8d,%r8d
- .byte 76,137,207 // mov %r9,%rdi
- .byte 76,137,230 // mov %r12,%rsi
- .byte 76,137,250 // mov %r15,%rdx
- .byte 76,137,241 // mov %r14,%rcx
- .byte 255,85,200 // callq *-0x38(%rbp)
- .byte 76,137,235 // mov %r13,%rbx
- .byte 76,139,77,192 // mov -0x40(%rbp),%r9
- .byte 73,141,87,8 // lea 0x8(%r15),%rdx
- .byte 73,131,199,16 // add $0x10,%r15
- .byte 73,57,223 // cmp %rbx,%r15
- .byte 73,137,215 // mov %rdx,%r15
- .byte 118,215 // jbe 54 <_sk_start_pipeline_ssse3_lowp+0x54>
- .byte 73,137,216 // mov %rbx,%r8
- .byte 73,41,208 // sub %rdx,%r8
- .byte 116,19 // je 98 <_sk_start_pipeline_ssse3_lowp+0x98>
- .byte 76,137,207 // mov %r9,%rdi
- .byte 76,137,230 // mov %r12,%rsi
- .byte 76,137,241 // mov %r14,%rcx
- .byte 255,85,200 // callq *-0x38(%rbp)
- .byte 76,137,235 // mov %r13,%rbx
- .byte 76,139,77,192 // mov -0x40(%rbp),%r9
- .byte 73,255,198 // inc %r14
- .byte 76,59,117,176 // cmp -0x50(%rbp),%r14
- .byte 117,165 // jne 46 <_sk_start_pipeline_ssse3_lowp+0x46>
- .byte 72,131,196,40 // add $0x28,%rsp
- .byte 91 // pop %rbx
- .byte 65,92 // pop %r12
- .byte 65,93 // pop %r13
- .byte 65,94 // pop %r14
- .byte 65,95 // pop %r15
- .byte 93 // pop %rbp
- .byte 195 // retq
-
-HIDDEN _sk_just_return_ssse3_lowp
-.globl _sk_just_return_ssse3_lowp
-FUNCTION(_sk_just_return_ssse3_lowp)
-_sk_just_return_ssse3_lowp:
- .byte 195 // retq
-
-HIDDEN _sk_uniform_color_ssse3_lowp
-.globl _sk_uniform_color_ssse3_lowp
-FUNCTION(_sk_uniform_color_ssse3_lowp)
-_sk_uniform_color_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,16,24 // movups (%rax),%xmm3
- .byte 15,88,29,211,26,0,0 // addps 0x1ad3(%rip),%xmm3 # 1b90 <_sk_xor__ssse3_lowp+0x9b>
- .byte 242,15,112,195,0 // pshuflw $0x0,%xmm3,%xmm0
- .byte 102,15,112,192,80 // pshufd $0x50,%xmm0,%xmm0
- .byte 242,15,112,203,170 // pshuflw $0xaa,%xmm3,%xmm1
- .byte 102,15,112,201,80 // pshufd $0x50,%xmm1,%xmm1
- .byte 243,15,112,211,0 // pshufhw $0x0,%xmm3,%xmm2
- .byte 102,15,112,210,250 // pshufd $0xfa,%xmm2,%xmm2
- .byte 243,15,112,219,170 // pshufhw $0xaa,%xmm3,%xmm3
- .byte 102,15,112,219,250 // pshufd $0xfa,%xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_black_color_ssse3_lowp
-.globl _sk_black_color_ssse3_lowp
-FUNCTION(_sk_black_color_ssse3_lowp)
-_sk_black_color_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,174,26,0,0 // movaps 0x1aae(%rip),%xmm3 # 1ba0 <_sk_xor__ssse3_lowp+0xab>
- .byte 15,87,192 // xorps %xmm0,%xmm0
- .byte 15,87,201 // xorps %xmm1,%xmm1
- .byte 15,87,210 // xorps %xmm2,%xmm2
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_white_color_ssse3_lowp
-.globl _sk_white_color_ssse3_lowp
-FUNCTION(_sk_white_color_ssse3_lowp)
-_sk_white_color_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,5,170,26,0,0 // movaps 0x1aaa(%rip),%xmm0 # 1bb0 <_sk_xor__ssse3_lowp+0xbb>
- .byte 15,40,200 // movaps %xmm0,%xmm1
- .byte 15,40,208 // movaps %xmm0,%xmm2
- .byte 15,40,216 // movaps %xmm0,%xmm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_set_rgb_ssse3_lowp
-.globl _sk_set_rgb_ssse3_lowp
-FUNCTION(_sk_set_rgb_ssse3_lowp)
-_sk_set_rgb_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 243,15,16,21,197,29,0,0 // movss 0x1dc5(%rip),%xmm2 # 1ee0 <_sk_xor__ssse3_lowp+0x3eb>
- .byte 243,15,16,0 // movss (%rax),%xmm0
- .byte 243,15,88,194 // addss %xmm2,%xmm0
- .byte 102,65,15,126,193 // movd %xmm0,%r9d
- .byte 102,65,15,110,193 // movd %r9d,%xmm0
- .byte 242,15,112,192,0 // pshuflw $0x0,%xmm0,%xmm0
- .byte 102,15,112,192,80 // pshufd $0x50,%xmm0,%xmm0
- .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
- .byte 243,15,88,202 // addss %xmm2,%xmm1
- .byte 102,65,15,126,201 // movd %xmm1,%r9d
- .byte 102,65,15,110,201 // movd %r9d,%xmm1
- .byte 242,15,112,201,0 // pshuflw $0x0,%xmm1,%xmm1
- .byte 102,15,112,201,80 // pshufd $0x50,%xmm1,%xmm1
- .byte 243,15,88,80,8 // addss 0x8(%rax),%xmm2
- .byte 102,15,126,208 // movd %xmm2,%eax
- .byte 102,15,110,208 // movd %eax,%xmm2
- .byte 242,15,112,210,0 // pshuflw $0x0,%xmm2,%xmm2
- .byte 102,15,112,210,80 // pshufd $0x50,%xmm2,%xmm2
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_premul_ssse3_lowp
-.globl _sk_premul_ssse3_lowp
-FUNCTION(_sk_premul_ssse3_lowp)
-_sk_premul_ssse3_lowp:
- .byte 102,15,56,11,195 // pmulhrsw %xmm3,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,56,11,203 // pmulhrsw %xmm3,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,56,11,211 // pmulhrsw %xmm3,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_load_8888_ssse3_lowp
-.globl _sk_load_8888_ssse3_lowp
-FUNCTION(_sk_load_8888_ssse3_lowp)
-_sk_load_8888_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,192,0,0,0 // jne 272 <_sk_load_8888_ssse3_lowp+0xe1>
- .byte 69,15,16,68,153,16 // movups 0x10(%r9,%rbx,4),%xmm8
- .byte 102,65,15,16,28,153 // movupd (%r9,%rbx,4),%xmm3
- .byte 102,15,111,13,250,25,0,0 // movdqa 0x19fa(%rip),%xmm1 # 1bc0 <_sk_xor__ssse3_lowp+0xcb>
- .byte 102,15,40,195 // movapd %xmm3,%xmm0
- .byte 102,15,56,0,193 // pshufb %xmm1,%xmm0
- .byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
- .byte 102,15,56,0,209 // pshufb %xmm1,%xmm2
- .byte 102,15,108,194 // punpcklqdq %xmm2,%xmm0
- .byte 102,15,113,240,8 // psllw $0x8,%xmm0
- .byte 102,68,15,111,13,229,25,0,0 // movdqa 0x19e5(%rip),%xmm9 # 1bd0 <_sk_xor__ssse3_lowp+0xdb>
- .byte 102,65,15,228,193 // pmulhuw %xmm9,%xmm0
- .byte 102,68,15,111,21,231,25,0,0 // movdqa 0x19e7(%rip),%xmm10 # 1be0 <_sk_xor__ssse3_lowp+0xeb>
- .byte 102,15,40,203 // movapd %xmm3,%xmm1
- .byte 102,65,15,56,0,202 // pshufb %xmm10,%xmm1
- .byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
- .byte 102,65,15,56,0,210 // pshufb %xmm10,%xmm2
- .byte 102,15,108,202 // punpcklqdq %xmm2,%xmm1
- .byte 102,15,113,241,8 // psllw $0x8,%xmm1
- .byte 102,65,15,228,201 // pmulhuw %xmm9,%xmm1
- .byte 102,68,15,111,21,203,25,0,0 // movdqa 0x19cb(%rip),%xmm10 # 1bf0 <_sk_xor__ssse3_lowp+0xfb>
- .byte 102,15,40,211 // movapd %xmm3,%xmm2
- .byte 102,65,15,56,0,210 // pshufb %xmm10,%xmm2
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
- .byte 102,65,15,108,211 // punpcklqdq %xmm11,%xmm2
- .byte 102,15,113,242,8 // psllw $0x8,%xmm2
- .byte 102,65,15,228,209 // pmulhuw %xmm9,%xmm2
- .byte 102,68,15,111,21,174,25,0,0 // movdqa 0x19ae(%rip),%xmm10 # 1c00 <_sk_xor__ssse3_lowp+0x10b>
- .byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
- .byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
- .byte 102,65,15,108,216 // punpcklqdq %xmm8,%xmm3
- .byte 102,15,113,243,8 // psllw $0x8,%xmm3
- .byte 102,65,15,228,217 // pmulhuw %xmm9,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,47,255,255,255 // ja 1be <_sk_load_8888_ssse3_lowp+0x2d>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,94,0,0,0 // lea 0x5e(%rip),%r11 # 2f8 <_sk_load_8888_ssse3_lowp+0x167>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 243,65,15,16,28,153 // movss (%r9,%rbx,4),%xmm3
- .byte 233,16,255,255,255 // jmpq 1be <_sk_load_8888_ssse3_lowp+0x2d>
- .byte 102,65,15,110,68,153,8 // movd 0x8(%r9,%rbx,4),%xmm0
- .byte 102,15,112,216,69 // pshufd $0x45,%xmm0,%xmm3
- .byte 102,65,15,18,28,153 // movlpd (%r9,%rbx,4),%xmm3
- .byte 233,249,254,255,255 // jmpq 1be <_sk_load_8888_ssse3_lowp+0x2d>
- .byte 102,65,15,110,68,153,24 // movd 0x18(%r9,%rbx,4),%xmm0
- .byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
- .byte 243,65,15,16,68,153,20 // movss 0x14(%r9,%rbx,4),%xmm0
- .byte 65,15,198,192,0 // shufps $0x0,%xmm8,%xmm0
- .byte 65,15,198,192,226 // shufps $0xe2,%xmm8,%xmm0
- .byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 243,65,15,16,68,153,16 // movss 0x10(%r9,%rbx,4),%xmm0
- .byte 243,68,15,16,192 // movss %xmm0,%xmm8
- .byte 233,192,254,255,255 // jmpq 1b8 <_sk_load_8888_ssse3_lowp+0x27>
- .byte 171 // stos %eax,%es:(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,194 // inc %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,182,255,255,255,192 // pushq -0x3f000001(%rsi)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 239 // out %eax,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 218,255 // (bad)
- .byte 255 // (bad)
- .byte 255,205 // dec %ebp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_8888_dst_ssse3_lowp
-.globl _sk_load_8888_dst_ssse3_lowp
-FUNCTION(_sk_load_8888_dst_ssse3_lowp)
-_sk_load_8888_dst_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,192,0,0,0 // jne 3f5 <_sk_load_8888_dst_ssse3_lowp+0xe1>
- .byte 69,15,16,68,153,16 // movups 0x10(%r9,%rbx,4),%xmm8
- .byte 102,65,15,16,60,153 // movupd (%r9,%rbx,4),%xmm7
- .byte 102,15,111,45,199,24,0,0 // movdqa 0x18c7(%rip),%xmm5 # 1c10 <_sk_xor__ssse3_lowp+0x11b>
- .byte 102,15,40,231 // movapd %xmm7,%xmm4
- .byte 102,15,56,0,229 // pshufb %xmm5,%xmm4
- .byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
- .byte 102,15,56,0,245 // pshufb %xmm5,%xmm6
- .byte 102,15,108,230 // punpcklqdq %xmm6,%xmm4
- .byte 102,15,113,244,8 // psllw $0x8,%xmm4
- .byte 102,68,15,111,13,178,24,0,0 // movdqa 0x18b2(%rip),%xmm9 # 1c20 <_sk_xor__ssse3_lowp+0x12b>
- .byte 102,65,15,228,225 // pmulhuw %xmm9,%xmm4
- .byte 102,68,15,111,21,180,24,0,0 // movdqa 0x18b4(%rip),%xmm10 # 1c30 <_sk_xor__ssse3_lowp+0x13b>
- .byte 102,15,40,239 // movapd %xmm7,%xmm5
- .byte 102,65,15,56,0,234 // pshufb %xmm10,%xmm5
- .byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
- .byte 102,65,15,56,0,242 // pshufb %xmm10,%xmm6
- .byte 102,15,108,238 // punpcklqdq %xmm6,%xmm5
- .byte 102,15,113,245,8 // psllw $0x8,%xmm5
- .byte 102,65,15,228,233 // pmulhuw %xmm9,%xmm5
- .byte 102,68,15,111,21,152,24,0,0 // movdqa 0x1898(%rip),%xmm10 # 1c40 <_sk_xor__ssse3_lowp+0x14b>
- .byte 102,15,40,247 // movapd %xmm7,%xmm6
- .byte 102,65,15,56,0,242 // pshufb %xmm10,%xmm6
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
- .byte 102,65,15,108,243 // punpcklqdq %xmm11,%xmm6
- .byte 102,15,113,246,8 // psllw $0x8,%xmm6
- .byte 102,65,15,228,241 // pmulhuw %xmm9,%xmm6
- .byte 102,68,15,111,21,123,24,0,0 // movdqa 0x187b(%rip),%xmm10 # 1c50 <_sk_xor__ssse3_lowp+0x15b>
- .byte 102,65,15,56,0,250 // pshufb %xmm10,%xmm7
- .byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
- .byte 102,65,15,108,248 // punpcklqdq %xmm8,%xmm7
- .byte 102,15,113,247,8 // psllw $0x8,%xmm7
- .byte 102,65,15,228,249 // pmulhuw %xmm9,%xmm7
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,47,255,255,255 // ja 341 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,95,0,0,0 // lea 0x5f(%rip),%r11 # 47c <_sk_load_8888_dst_ssse3_lowp+0x168>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 243,65,15,16,60,153 // movss (%r9,%rbx,4),%xmm7
- .byte 233,16,255,255,255 // jmpq 341 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- .byte 102,65,15,110,100,153,8 // movd 0x8(%r9,%rbx,4),%xmm4
- .byte 102,15,112,252,69 // pshufd $0x45,%xmm4,%xmm7
- .byte 102,65,15,18,60,153 // movlpd (%r9,%rbx,4),%xmm7
- .byte 233,249,254,255,255 // jmpq 341 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- .byte 102,65,15,110,100,153,24 // movd 0x18(%r9,%rbx,4),%xmm4
- .byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
- .byte 243,65,15,16,100,153,20 // movss 0x14(%r9,%rbx,4),%xmm4
- .byte 65,15,198,224,0 // shufps $0x0,%xmm8,%xmm4
- .byte 65,15,198,224,226 // shufps $0xe2,%xmm8,%xmm4
- .byte 68,15,40,196 // movaps %xmm4,%xmm8
- .byte 243,65,15,16,100,153,16 // movss 0x10(%r9,%rbx,4),%xmm4
- .byte 243,68,15,16,196 // movss %xmm4,%xmm8
- .byte 233,192,254,255,255 // jmpq 33b <_sk_load_8888_dst_ssse3_lowp+0x27>
- .byte 144 // nop
- .byte 170 // stos %al,%es:(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,193 // inc %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,181,255,255,255,191 // pushq -0x40000001(%rbp)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 217,255 // fcos
- .byte 255 // (bad)
- .byte 255,204 // dec %esp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_8888_ssse3_lowp
-.globl _sk_store_8888_ssse3_lowp
-FUNCTION(_sk_store_8888_ssse3_lowp)
-_sk_store_8888_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 102,68,15,111,192 // movdqa %xmm0,%xmm8
- .byte 102,65,15,113,208,7 // psrlw $0x7,%xmm8
- .byte 102,68,15,111,21,156,23,0,0 // movdqa 0x179c(%rip),%xmm10 # 1c60 <_sk_xor__ssse3_lowp+0x16b>
- .byte 102,69,15,234,194 // pminsw %xmm10,%xmm8
- .byte 102,69,15,239,219 // pxor %xmm11,%xmm11
- .byte 102,69,15,111,232 // movdqa %xmm8,%xmm13
- .byte 102,69,15,97,235 // punpcklwd %xmm11,%xmm13
- .byte 102,69,15,105,195 // punpckhwd %xmm11,%xmm8
- .byte 102,68,15,111,201 // movdqa %xmm1,%xmm9
- .byte 102,65,15,113,209,7 // psrlw $0x7,%xmm9
- .byte 102,69,15,234,202 // pminsw %xmm10,%xmm9
- .byte 102,69,15,111,225 // movdqa %xmm9,%xmm12
- .byte 102,69,15,105,227 // punpckhwd %xmm11,%xmm12
- .byte 102,69,15,97,203 // punpcklwd %xmm11,%xmm9
- .byte 102,65,15,114,241,8 // pslld $0x8,%xmm9
- .byte 102,65,15,114,244,8 // pslld $0x8,%xmm12
- .byte 102,69,15,235,224 // por %xmm8,%xmm12
- .byte 102,69,15,235,205 // por %xmm13,%xmm9
- .byte 102,68,15,111,234 // movdqa %xmm2,%xmm13
- .byte 102,65,15,113,213,7 // psrlw $0x7,%xmm13
- .byte 102,69,15,234,234 // pminsw %xmm10,%xmm13
- .byte 102,69,15,111,245 // movdqa %xmm13,%xmm14
- .byte 102,69,15,97,243 // punpcklwd %xmm11,%xmm14
- .byte 102,69,15,105,235 // punpckhwd %xmm11,%xmm13
- .byte 102,65,15,114,245,16 // pslld $0x10,%xmm13
- .byte 102,65,15,114,246,16 // pslld $0x10,%xmm14
- .byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,65,15,113,208,7 // psrlw $0x7,%xmm8
- .byte 102,69,15,234,194 // pminsw %xmm10,%xmm8
- .byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
- .byte 102,69,15,105,211 // punpckhwd %xmm11,%xmm10
- .byte 102,69,15,97,195 // punpcklwd %xmm11,%xmm8
- .byte 102,65,15,114,240,24 // pslld $0x18,%xmm8
- .byte 102,65,15,114,242,24 // pslld $0x18,%xmm10
- .byte 102,69,15,235,213 // por %xmm13,%xmm10
- .byte 102,69,15,235,212 // por %xmm12,%xmm10
- .byte 102,69,15,235,198 // por %xmm14,%xmm8
- .byte 102,69,15,235,193 // por %xmm9,%xmm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,18 // jne 593 <_sk_store_8888_ssse3_lowp+0xfb>
- .byte 243,69,15,127,4,153 // movdqu %xmm8,(%r9,%rbx,4)
- .byte 243,69,15,127,84,153,16 // movdqu %xmm10,0x10(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,235 // ja 58e <_sk_store_8888_ssse3_lowp+0xf6>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,82,0,0,0 // lea 0x52(%rip),%r11 # 600 <_sk_store_8888_ssse3_lowp+0x168>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 102,69,15,126,4,153 // movd %xmm8,(%r9,%rbx,4)
- .byte 235,207 // jmp 58e <_sk_store_8888_ssse3_lowp+0xf6>
- .byte 102,69,15,112,200,78 // pshufd $0x4e,%xmm8,%xmm9
- .byte 102,69,15,126,76,153,8 // movd %xmm9,0x8(%r9,%rbx,4)
- .byte 102,69,15,214,4,153 // movq %xmm8,(%r9,%rbx,4)
- .byte 235,186 // jmp 58e <_sk_store_8888_ssse3_lowp+0xf6>
- .byte 102,69,15,112,202,78 // pshufd $0x4e,%xmm10,%xmm9
- .byte 102,69,15,126,76,153,24 // movd %xmm9,0x18(%r9,%rbx,4)
- .byte 102,69,15,112,202,229 // pshufd $0xe5,%xmm10,%xmm9
- .byte 102,69,15,126,76,153,20 // movd %xmm9,0x14(%r9,%rbx,4)
- .byte 102,69,15,126,84,153,16 // movd %xmm10,0x10(%r9,%rbx,4)
- .byte 243,69,15,127,4,153 // movdqu %xmm8,(%r9,%rbx,4)
- .byte 235,145 // jmp 58e <_sk_store_8888_ssse3_lowp+0xf6>
- .byte 15,31,0 // nopl (%rax)
- .byte 183,255 // mov $0xff,%bh
- .byte 255 // (bad)
- .byte 255,204 // dec %esp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 191,255,255,255,245 // mov $0xf5ffffff,%edi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,225 // jmpq *%rcx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,212 // callq *%rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_bgra_ssse3_lowp
-.globl _sk_load_bgra_ssse3_lowp
-FUNCTION(_sk_load_bgra_ssse3_lowp)
-_sk_load_bgra_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,192,0,0,0 // jne 6fd <_sk_load_bgra_ssse3_lowp+0xe1>
- .byte 69,15,16,68,153,16 // movups 0x10(%r9,%rbx,4),%xmm8
- .byte 102,65,15,16,28,153 // movupd (%r9,%rbx,4),%xmm3
- .byte 102,15,111,5,31,22,0,0 // movdqa 0x161f(%rip),%xmm0 # 1c70 <_sk_xor__ssse3_lowp+0x17b>
- .byte 102,15,40,211 // movapd %xmm3,%xmm2
- .byte 102,15,56,0,208 // pshufb %xmm0,%xmm2
- .byte 102,65,15,111,200 // movdqa %xmm8,%xmm1
- .byte 102,15,56,0,200 // pshufb %xmm0,%xmm1
- .byte 102,15,108,209 // punpcklqdq %xmm1,%xmm2
- .byte 102,15,113,242,8 // psllw $0x8,%xmm2
- .byte 102,68,15,111,13,10,22,0,0 // movdqa 0x160a(%rip),%xmm9 # 1c80 <_sk_xor__ssse3_lowp+0x18b>
- .byte 102,65,15,228,209 // pmulhuw %xmm9,%xmm2
- .byte 102,68,15,111,21,12,22,0,0 // movdqa 0x160c(%rip),%xmm10 # 1c90 <_sk_xor__ssse3_lowp+0x19b>
- .byte 102,15,40,203 // movapd %xmm3,%xmm1
- .byte 102,65,15,56,0,202 // pshufb %xmm10,%xmm1
- .byte 102,65,15,111,192 // movdqa %xmm8,%xmm0
- .byte 102,65,15,56,0,194 // pshufb %xmm10,%xmm0
- .byte 102,15,108,200 // punpcklqdq %xmm0,%xmm1
- .byte 102,15,113,241,8 // psllw $0x8,%xmm1
- .byte 102,65,15,228,201 // pmulhuw %xmm9,%xmm1
- .byte 102,68,15,111,21,240,21,0,0 // movdqa 0x15f0(%rip),%xmm10 # 1ca0 <_sk_xor__ssse3_lowp+0x1ab>
- .byte 102,15,40,195 // movapd %xmm3,%xmm0
- .byte 102,65,15,56,0,194 // pshufb %xmm10,%xmm0
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
- .byte 102,65,15,108,195 // punpcklqdq %xmm11,%xmm0
- .byte 102,15,113,240,8 // psllw $0x8,%xmm0
- .byte 102,65,15,228,193 // pmulhuw %xmm9,%xmm0
- .byte 102,68,15,111,21,211,21,0,0 // movdqa 0x15d3(%rip),%xmm10 # 1cb0 <_sk_xor__ssse3_lowp+0x1bb>
- .byte 102,65,15,56,0,218 // pshufb %xmm10,%xmm3
- .byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
- .byte 102,65,15,108,216 // punpcklqdq %xmm8,%xmm3
- .byte 102,15,113,243,8 // psllw $0x8,%xmm3
- .byte 102,65,15,228,217 // pmulhuw %xmm9,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,47,255,255,255 // ja 649 <_sk_load_bgra_ssse3_lowp+0x2d>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,95,0,0,0 // lea 0x5f(%rip),%r11 # 784 <_sk_load_bgra_ssse3_lowp+0x168>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 243,65,15,16,28,153 // movss (%r9,%rbx,4),%xmm3
- .byte 233,16,255,255,255 // jmpq 649 <_sk_load_bgra_ssse3_lowp+0x2d>
- .byte 102,65,15,110,68,153,8 // movd 0x8(%r9,%rbx,4),%xmm0
- .byte 102,15,112,216,69 // pshufd $0x45,%xmm0,%xmm3
- .byte 102,65,15,18,28,153 // movlpd (%r9,%rbx,4),%xmm3
- .byte 233,249,254,255,255 // jmpq 649 <_sk_load_bgra_ssse3_lowp+0x2d>
- .byte 102,65,15,110,68,153,24 // movd 0x18(%r9,%rbx,4),%xmm0
- .byte 102,68,15,112,192,69 // pshufd $0x45,%xmm0,%xmm8
- .byte 243,65,15,16,68,153,20 // movss 0x14(%r9,%rbx,4),%xmm0
- .byte 65,15,198,192,0 // shufps $0x0,%xmm8,%xmm0
- .byte 65,15,198,192,226 // shufps $0xe2,%xmm8,%xmm0
- .byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 243,65,15,16,68,153,16 // movss 0x10(%r9,%rbx,4),%xmm0
- .byte 243,68,15,16,192 // movss %xmm0,%xmm8
- .byte 233,192,254,255,255 // jmpq 643 <_sk_load_bgra_ssse3_lowp+0x27>
- .byte 144 // nop
- .byte 170 // stos %al,%es:(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,193 // inc %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,181,255,255,255,191 // pushq -0x40000001(%rbp)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 217,255 // fcos
- .byte 255 // (bad)
- .byte 255,204 // dec %esp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_bgra_dst_ssse3_lowp
-.globl _sk_load_bgra_dst_ssse3_lowp
-FUNCTION(_sk_load_bgra_dst_ssse3_lowp)
-_sk_load_bgra_dst_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,192,0,0,0 // jne 881 <_sk_load_bgra_dst_ssse3_lowp+0xe1>
- .byte 69,15,16,68,153,16 // movups 0x10(%r9,%rbx,4),%xmm8
- .byte 102,65,15,16,60,153 // movupd (%r9,%rbx,4),%xmm7
- .byte 102,15,111,37,235,20,0,0 // movdqa 0x14eb(%rip),%xmm4 # 1cc0 <_sk_xor__ssse3_lowp+0x1cb>
- .byte 102,15,40,247 // movapd %xmm7,%xmm6
- .byte 102,15,56,0,244 // pshufb %xmm4,%xmm6
- .byte 102,65,15,111,232 // movdqa %xmm8,%xmm5
- .byte 102,15,56,0,236 // pshufb %xmm4,%xmm5
- .byte 102,15,108,245 // punpcklqdq %xmm5,%xmm6
- .byte 102,15,113,246,8 // psllw $0x8,%xmm6
- .byte 102,68,15,111,13,214,20,0,0 // movdqa 0x14d6(%rip),%xmm9 # 1cd0 <_sk_xor__ssse3_lowp+0x1db>
- .byte 102,65,15,228,241 // pmulhuw %xmm9,%xmm6
- .byte 102,68,15,111,21,216,20,0,0 // movdqa 0x14d8(%rip),%xmm10 # 1ce0 <_sk_xor__ssse3_lowp+0x1eb>
- .byte 102,15,40,239 // movapd %xmm7,%xmm5
- .byte 102,65,15,56,0,234 // pshufb %xmm10,%xmm5
- .byte 102,65,15,111,224 // movdqa %xmm8,%xmm4
- .byte 102,65,15,56,0,226 // pshufb %xmm10,%xmm4
- .byte 102,15,108,236 // punpcklqdq %xmm4,%xmm5
- .byte 102,15,113,245,8 // psllw $0x8,%xmm5
- .byte 102,65,15,228,233 // pmulhuw %xmm9,%xmm5
- .byte 102,68,15,111,21,188,20,0,0 // movdqa 0x14bc(%rip),%xmm10 # 1cf0 <_sk_xor__ssse3_lowp+0x1fb>
- .byte 102,15,40,231 // movapd %xmm7,%xmm4
- .byte 102,65,15,56,0,226 // pshufb %xmm10,%xmm4
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
- .byte 102,65,15,108,227 // punpcklqdq %xmm11,%xmm4
- .byte 102,15,113,244,8 // psllw $0x8,%xmm4
- .byte 102,65,15,228,225 // pmulhuw %xmm9,%xmm4
- .byte 102,68,15,111,21,159,20,0,0 // movdqa 0x149f(%rip),%xmm10 # 1d00 <_sk_xor__ssse3_lowp+0x20b>
- .byte 102,65,15,56,0,250 // pshufb %xmm10,%xmm7
- .byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
- .byte 102,65,15,108,248 // punpcklqdq %xmm8,%xmm7
- .byte 102,15,113,247,8 // psllw $0x8,%xmm7
- .byte 102,65,15,228,249 // pmulhuw %xmm9,%xmm7
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,47,255,255,255 // ja 7cd <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,95,0,0,0 // lea 0x5f(%rip),%r11 # 908 <_sk_load_bgra_dst_ssse3_lowp+0x168>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 243,65,15,16,60,153 // movss (%r9,%rbx,4),%xmm7
- .byte 233,16,255,255,255 // jmpq 7cd <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- .byte 102,65,15,110,100,153,8 // movd 0x8(%r9,%rbx,4),%xmm4
- .byte 102,15,112,252,69 // pshufd $0x45,%xmm4,%xmm7
- .byte 102,65,15,18,60,153 // movlpd (%r9,%rbx,4),%xmm7
- .byte 233,249,254,255,255 // jmpq 7cd <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- .byte 102,65,15,110,100,153,24 // movd 0x18(%r9,%rbx,4),%xmm4
- .byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
- .byte 243,65,15,16,100,153,20 // movss 0x14(%r9,%rbx,4),%xmm4
- .byte 65,15,198,224,0 // shufps $0x0,%xmm8,%xmm4
- .byte 65,15,198,224,226 // shufps $0xe2,%xmm8,%xmm4
- .byte 68,15,40,196 // movaps %xmm4,%xmm8
- .byte 243,65,15,16,100,153,16 // movss 0x10(%r9,%rbx,4),%xmm4
- .byte 243,68,15,16,196 // movss %xmm4,%xmm8
- .byte 233,192,254,255,255 // jmpq 7c7 <_sk_load_bgra_dst_ssse3_lowp+0x27>
- .byte 144 // nop
- .byte 170 // stos %al,%es:(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,193 // inc %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,181,255,255,255,191 // pushq -0x40000001(%rbp)
- .byte 254 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 238 // out %al,(%dx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 217,255 // fcos
- .byte 255 // (bad)
- .byte 255,204 // dec %esp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_bgra_ssse3_lowp
-.globl _sk_store_bgra_ssse3_lowp
-FUNCTION(_sk_store_bgra_ssse3_lowp)
-_sk_store_bgra_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 15,41,124,36,240 // movaps %xmm7,-0x10(%rsp)
- .byte 102,68,15,111,210 // movdqa %xmm2,%xmm10
- .byte 102,65,15,113,210,7 // psrlw $0x7,%xmm10
- .byte 102,68,15,111,13,210,19,0,0 // movdqa 0x13d2(%rip),%xmm9 # 1d10 <_sk_xor__ssse3_lowp+0x21b>
- .byte 102,69,15,234,209 // pminsw %xmm9,%xmm10
- .byte 102,69,15,239,246 // pxor %xmm14,%xmm14
- .byte 102,65,15,111,250 // movdqa %xmm10,%xmm7
- .byte 102,65,15,97,254 // punpcklwd %xmm14,%xmm7
- .byte 102,69,15,105,214 // punpckhwd %xmm14,%xmm10
- .byte 102,68,15,111,225 // movdqa %xmm1,%xmm12
- .byte 102,65,15,113,212,7 // psrlw $0x7,%xmm12
- .byte 102,69,15,234,225 // pminsw %xmm9,%xmm12
- .byte 102,69,15,111,236 // movdqa %xmm12,%xmm13
- .byte 102,69,15,97,238 // punpcklwd %xmm14,%xmm13
- .byte 102,69,15,105,230 // punpckhwd %xmm14,%xmm12
- .byte 102,68,15,111,248 // movdqa %xmm0,%xmm15
- .byte 102,65,15,113,215,7 // psrlw $0x7,%xmm15
- .byte 102,69,15,234,249 // pminsw %xmm9,%xmm15
- .byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,65,15,113,208,7 // psrlw $0x7,%xmm8
- .byte 102,69,15,234,193 // pminsw %xmm9,%xmm8
- .byte 102,69,15,111,207 // movdqa %xmm15,%xmm9
- .byte 102,69,15,105,206 // punpckhwd %xmm14,%xmm9
- .byte 102,69,15,97,254 // punpcklwd %xmm14,%xmm15
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,105,222 // punpckhwd %xmm14,%xmm11
- .byte 102,69,15,97,198 // punpcklwd %xmm14,%xmm8
- .byte 102,65,15,114,241,16 // pslld $0x10,%xmm9
- .byte 102,69,15,235,202 // por %xmm10,%xmm9
- .byte 102,65,15,114,247,16 // pslld $0x10,%xmm15
- .byte 102,68,15,235,255 // por %xmm7,%xmm15
- .byte 102,65,15,114,244,8 // pslld $0x8,%xmm12
- .byte 102,65,15,114,243,24 // pslld $0x18,%xmm11
- .byte 102,69,15,235,220 // por %xmm12,%xmm11
- .byte 102,69,15,235,217 // por %xmm9,%xmm11
- .byte 102,65,15,114,245,8 // pslld $0x8,%xmm13
- .byte 102,65,15,114,240,24 // pslld $0x18,%xmm8
- .byte 102,69,15,235,197 // por %xmm13,%xmm8
- .byte 102,69,15,235,199 // por %xmm15,%xmm8
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,23 // jne a29 <_sk_store_bgra_ssse3_lowp+0x105>
- .byte 243,69,15,127,4,153 // movdqu %xmm8,(%r9,%rbx,4)
- .byte 243,69,15,127,92,153,16 // movdqu %xmm11,0x10(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,124,36,240 // movaps -0x10(%rsp),%xmm7
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,230 // ja a1f <_sk_store_bgra_ssse3_lowp+0xfb>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,80,0,0,0 // lea 0x50(%rip),%r11 # a94 <_sk_store_bgra_ssse3_lowp+0x170>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 102,69,15,126,4,153 // movd %xmm8,(%r9,%rbx,4)
- .byte 235,202 // jmp a1f <_sk_store_bgra_ssse3_lowp+0xfb>
- .byte 102,65,15,112,248,78 // pshufd $0x4e,%xmm8,%xmm7
- .byte 102,65,15,126,124,153,8 // movd %xmm7,0x8(%r9,%rbx,4)
- .byte 102,69,15,214,4,153 // movq %xmm8,(%r9,%rbx,4)
- .byte 235,181 // jmp a1f <_sk_store_bgra_ssse3_lowp+0xfb>
- .byte 102,65,15,112,251,78 // pshufd $0x4e,%xmm11,%xmm7
- .byte 102,65,15,126,124,153,24 // movd %xmm7,0x18(%r9,%rbx,4)
- .byte 102,65,15,112,251,229 // pshufd $0xe5,%xmm11,%xmm7
- .byte 102,65,15,126,124,153,20 // movd %xmm7,0x14(%r9,%rbx,4)
- .byte 102,69,15,126,92,153,16 // movd %xmm11,0x10(%r9,%rbx,4)
- .byte 243,69,15,127,4,153 // movdqu %xmm8,(%r9,%rbx,4)
- .byte 235,140 // jmp a1f <_sk_store_bgra_ssse3_lowp+0xfb>
- .byte 144 // nop
- .byte 185,255,255,255,206 // mov $0xceffffff,%ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,193 // inc %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,247 // push %rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,240 // push %rax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,227 // jmpq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,214 // callq *%rsi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_a8_ssse3_lowp
-.globl _sk_load_a8_ssse3_lowp
-FUNCTION(_sk_load_a8_ssse3_lowp)
-_sk_load_a8_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,38 // jne aef <_sk_load_a8_ssse3_lowp+0x3f>
- .byte 243,65,15,126,28,25 // movq (%r9,%rbx,1),%xmm3
- .byte 102,15,96,216 // punpcklbw %xmm0,%xmm3
- .byte 102,15,113,243,8 // psllw $0x8,%xmm3
- .byte 102,15,228,29,64,18,0,0 // pmulhuw 0x1240(%rip),%xmm3 # 1d20 <_sk_xor__ssse3_lowp+0x22b>
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 102,15,87,192 // xorpd %xmm0,%xmm0
- .byte 15,87,201 // xorps %xmm1,%xmm1
- .byte 15,87,210 // xorps %xmm2,%xmm2
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,208 // ja ad3 <_sk_load_a8_ssse3_lowp+0x23>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,110,0,0,0 // lea 0x6e(%rip),%r11 # b7c <_sk_load_a8_ssse3_lowp+0xcc>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,15,110,216 // movd %eax,%xmm3
- .byte 235,177 // jmp ad3 <_sk_load_a8_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 102,15,196,216,2 // pinsrw $0x2,%eax,%xmm3
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,15,110,192 // movd %eax,%xmm0
- .byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
- .byte 243,15,16,216 // movss %xmm0,%xmm3
- .byte 235,143 // jmp ad3 <_sk_load_a8_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,15,239,219 // pxor %xmm3,%xmm3
- .byte 102,15,196,216,6 // pinsrw $0x6,%eax,%xmm3
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,15,196,216,5 // pinsrw $0x5,%eax,%xmm3
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,15,196,216,4 // pinsrw $0x4,%eax,%xmm3
- .byte 102,65,15,110,4,25 // movd (%r9,%rbx,1),%xmm0
- .byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
- .byte 242,15,16,216 // movsd %xmm0,%xmm3
- .byte 233,87,255,255,255 // jmpq ad3 <_sk_load_a8_ssse3_lowp+0x23>
- .byte 155 // fwait
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,181,255,255,255,166 // pushq -0x59000001(%rbp)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 237 // in (%dx),%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,226 // jmpq *%rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,215 // callq *%rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,200 // dec %eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_a8_dst_ssse3_lowp
-.globl _sk_load_a8_dst_ssse3_lowp
-FUNCTION(_sk_load_a8_dst_ssse3_lowp)
-_sk_load_a8_dst_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,38 // jne bd7 <_sk_load_a8_dst_ssse3_lowp+0x3f>
- .byte 243,65,15,126,60,25 // movq (%r9,%rbx,1),%xmm7
- .byte 102,15,96,248 // punpcklbw %xmm0,%xmm7
- .byte 102,15,113,247,8 // psllw $0x8,%xmm7
- .byte 102,15,228,61,104,17,0,0 // pmulhuw 0x1168(%rip),%xmm7 # 1d30 <_sk_xor__ssse3_lowp+0x23b>
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 102,15,87,228 // xorpd %xmm4,%xmm4
- .byte 15,87,237 // xorps %xmm5,%xmm5
- .byte 15,87,246 // xorps %xmm6,%xmm6
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,208 // ja bbb <_sk_load_a8_dst_ssse3_lowp+0x23>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,110,0,0,0 // lea 0x6e(%rip),%r11 # c64 <_sk_load_a8_dst_ssse3_lowp+0xcc>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,15,110,248 // movd %eax,%xmm7
- .byte 235,177 // jmp bbb <_sk_load_a8_dst_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 102,15,196,248,2 // pinsrw $0x2,%eax,%xmm7
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,15,110,224 // movd %eax,%xmm4
- .byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
- .byte 243,15,16,252 // movss %xmm4,%xmm7
- .byte 235,143 // jmp bbb <_sk_load_a8_dst_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 102,15,196,248,6 // pinsrw $0x6,%eax,%xmm7
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,15,196,248,5 // pinsrw $0x5,%eax,%xmm7
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,15,196,248,4 // pinsrw $0x4,%eax,%xmm7
- .byte 102,65,15,110,36,25 // movd (%r9,%rbx,1),%xmm4
- .byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
- .byte 242,15,16,252 // movsd %xmm4,%xmm7
- .byte 233,87,255,255,255 // jmpq bbb <_sk_load_a8_dst_ssse3_lowp+0x23>
- .byte 155 // fwait
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,181,255,255,255,166 // pushq -0x59000001(%rbp)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 237 // in (%dx),%eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,226 // jmpq *%rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,215 // callq *%rdi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,200 // dec %eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_store_a8_ssse3_lowp
-.globl _sk_store_a8_ssse3_lowp
-FUNCTION(_sk_store_a8_ssse3_lowp)
-_sk_store_a8_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 102,68,15,111,195 // movdqa %xmm3,%xmm8
- .byte 102,65,15,113,208,7 // psrlw $0x7,%xmm8
- .byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
- .byte 77,133,192 // test %r8,%r8
- .byte 117,11 // jne cb4 <_sk_store_a8_ssse3_lowp+0x34>
- .byte 242,69,15,17,4,25 // movsd %xmm8,(%r9,%rbx,1)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,235 // ja caf <_sk_store_a8_ssse3_lowp+0x2f>
- .byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,136,0,0,0 // lea 0x88(%rip),%r11 # d5c <_sk_store_a8_ssse3_lowp+0xdc>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 102,68,15,127,68,36,176 // movdqa %xmm8,-0x50(%rsp)
- .byte 138,68,36,176 // mov -0x50(%rsp),%al
- .byte 65,136,4,25 // mov %al,(%r9,%rbx,1)
- .byte 235,193 // jmp caf <_sk_store_a8_ssse3_lowp+0x2f>
- .byte 102,68,15,127,68,36,192 // movdqa %xmm8,-0x40(%rsp)
- .byte 138,68,36,196 // mov -0x3c(%rsp),%al
- .byte 65,136,68,25,2 // mov %al,0x2(%r9,%rbx,1)
- .byte 102,68,15,56,0,5,72,16,0,0 // pshufb 0x1048(%rip),%xmm8 # 1d50 <_sk_xor__ssse3_lowp+0x25b>
- .byte 102,68,15,126,192 // movd %xmm8,%eax
- .byte 102,65,137,4,25 // mov %ax,(%r9,%rbx,1)
- .byte 235,155 // jmp caf <_sk_store_a8_ssse3_lowp+0x2f>
- .byte 102,68,15,127,68,36,240 // movdqa %xmm8,-0x10(%rsp)
- .byte 138,68,36,252 // mov -0x4(%rsp),%al
- .byte 65,136,68,25,6 // mov %al,0x6(%r9,%rbx,1)
- .byte 102,68,15,127,68,36,224 // movdqa %xmm8,-0x20(%rsp)
- .byte 138,68,36,234 // mov -0x16(%rsp),%al
- .byte 65,136,68,25,5 // mov %al,0x5(%r9,%rbx,1)
- .byte 102,68,15,127,68,36,208 // movdqa %xmm8,-0x30(%rsp)
- .byte 138,68,36,216 // mov -0x28(%rsp),%al
- .byte 65,136,68,25,4 // mov %al,0x4(%r9,%rbx,1)
- .byte 102,68,15,56,0,5,242,15,0,0 // pshufb 0xff2(%rip),%xmm8 # 1d40 <_sk_xor__ssse3_lowp+0x24b>
- .byte 102,69,15,126,4,25 // movd %xmm8,(%r9,%rbx,1)
- .byte 233,86,255,255,255 // jmpq caf <_sk_store_a8_ssse3_lowp+0x2f>
- .byte 15,31,0 // nopl (%rax)
- .byte 129,255,255,255,162,255 // cmp $0xffa2ffff,%edi
- .byte 255 // (bad)
- .byte 255,146,255,255,255,232 // callq *-0x17000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 216,255 // fdivr %st(7),%st
- .byte 255 // (bad)
- .byte 255,200 // dec %eax
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 184 // .byte 0xb8
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_g8_ssse3_lowp
-.globl _sk_load_g8_ssse3_lowp
-FUNCTION(_sk_load_g8_ssse3_lowp)
-_sk_load_g8_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,43 // jne dbc <_sk_load_g8_ssse3_lowp+0x44>
- .byte 243,65,15,126,4,25 // movq (%r9,%rbx,1),%xmm0
- .byte 102,15,96,192 // punpcklbw %xmm0,%xmm0
- .byte 102,15,113,240,8 // psllw $0x8,%xmm0
- .byte 102,15,228,5,184,15,0,0 // pmulhuw 0xfb8(%rip),%xmm0 # 1d60 <_sk_xor__ssse3_lowp+0x26b>
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,29,191,15,0,0 // movaps 0xfbf(%rip),%xmm3 # 1d70 <_sk_xor__ssse3_lowp+0x27b>
- .byte 102,15,111,200 // movdqa %xmm0,%xmm1
- .byte 102,15,111,208 // movdqa %xmm0,%xmm2
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,15,239,192 // pxor %xmm0,%xmm0
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,203 // ja d9b <_sk_load_g8_ssse3_lowp+0x23>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,113,0,0,0 // lea 0x71(%rip),%r11 # e4c <_sk_load_g8_ssse3_lowp+0xd4>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,15,110,192 // movd %eax,%xmm0
- .byte 235,172 // jmp d9b <_sk_load_g8_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,15,239,192 // pxor %xmm0,%xmm0
- .byte 102,15,196,192,2 // pinsrw $0x2,%eax,%xmm0
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,15,110,200 // movd %eax,%xmm1
- .byte 102,15,96,200 // punpcklbw %xmm0,%xmm1
- .byte 243,15,16,193 // movss %xmm1,%xmm0
- .byte 235,138 // jmp d9b <_sk_load_g8_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,15,239,192 // pxor %xmm0,%xmm0
- .byte 102,15,196,192,6 // pinsrw $0x6,%eax,%xmm0
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,15,196,192,5 // pinsrw $0x5,%eax,%xmm0
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,15,196,192,4 // pinsrw $0x4,%eax,%xmm0
- .byte 102,65,15,110,12,25 // movd (%r9,%rbx,1),%xmm1
- .byte 102,15,96,200 // punpcklbw %xmm0,%xmm1
- .byte 242,15,16,193 // movsd %xmm1,%xmm0
- .byte 233,82,255,255,255 // jmpq d9b <_sk_load_g8_ssse3_lowp+0x23>
- .byte 15,31,0 // nopl (%rax)
- .byte 152 // cwtl
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,178,255,255,255,163 // pushq -0x5c000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 234 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 223,255 // (bad)
- .byte 255 // (bad)
- .byte 255,212 // callq *%rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,197 // inc %ebp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_load_g8_dst_ssse3_lowp
-.globl _sk_load_g8_dst_ssse3_lowp
-FUNCTION(_sk_load_g8_dst_ssse3_lowp)
-_sk_load_g8_dst_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,43 // jne eac <_sk_load_g8_dst_ssse3_lowp+0x44>
- .byte 243,65,15,126,36,25 // movq (%r9,%rbx,1),%xmm4
- .byte 102,15,96,224 // punpcklbw %xmm0,%xmm4
- .byte 102,15,113,244,8 // psllw $0x8,%xmm4
- .byte 102,15,228,37,232,14,0,0 // pmulhuw 0xee8(%rip),%xmm4 # 1d80 <_sk_xor__ssse3_lowp+0x28b>
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,61,239,14,0,0 // movaps 0xeef(%rip),%xmm7 # 1d90 <_sk_xor__ssse3_lowp+0x29b>
- .byte 102,15,111,236 // movdqa %xmm4,%xmm5
- .byte 102,15,111,244 // movdqa %xmm4,%xmm6
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,15,239,228 // pxor %xmm4,%xmm4
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,203 // ja e8b <_sk_load_g8_dst_ssse3_lowp+0x23>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,113,0,0,0 // lea 0x71(%rip),%r11 # f3c <_sk_load_g8_dst_ssse3_lowp+0xd4>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,15,110,224 // movd %eax,%xmm4
- .byte 235,172 // jmp e8b <_sk_load_g8_dst_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,15,239,228 // pxor %xmm4,%xmm4
- .byte 102,15,196,224,2 // pinsrw $0x2,%eax,%xmm4
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,15,110,232 // movd %eax,%xmm5
- .byte 102,15,96,232 // punpcklbw %xmm0,%xmm5
- .byte 243,15,16,229 // movss %xmm5,%xmm4
- .byte 235,138 // jmp e8b <_sk_load_g8_dst_ssse3_lowp+0x23>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,15,239,228 // pxor %xmm4,%xmm4
- .byte 102,15,196,224,6 // pinsrw $0x6,%eax,%xmm4
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,15,196,224,5 // pinsrw $0x5,%eax,%xmm4
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,15,196,224,4 // pinsrw $0x4,%eax,%xmm4
- .byte 102,65,15,110,44,25 // movd (%r9,%rbx,1),%xmm5
- .byte 102,15,96,232 // punpcklbw %xmm0,%xmm5
- .byte 242,15,16,229 // movsd %xmm5,%xmm4
- .byte 233,82,255,255,255 // jmpq e8b <_sk_load_g8_dst_ssse3_lowp+0x23>
- .byte 15,31,0 // nopl (%rax)
- .byte 152 // cwtl
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,178,255,255,255,163 // pushq -0x5c000001(%rdx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 234 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 223,255 // (bad)
- .byte 255 // (bad)
- .byte 255,212 // callq *%rsp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,197 // inc %ebp
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_srcover_rgba_8888_ssse3_lowp
-.globl _sk_srcover_rgba_8888_ssse3_lowp
-FUNCTION(_sk_srcover_rgba_8888_ssse3_lowp)
-_sk_srcover_rgba_8888_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 102,68,15,111,225 // movdqa %xmm1,%xmm12
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 73,193,225,2 // shl $0x2,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,12,2,0,0 // jne 118a <_sk_srcover_rgba_8888_ssse3_lowp+0x232>
- .byte 69,15,16,68,153,16 // movups 0x10(%r9,%rbx,4),%xmm8
- .byte 102,65,15,16,60,153 // movupd (%r9,%rbx,4),%xmm7
- .byte 102,15,111,45,14,14,0,0 // movdqa 0xe0e(%rip),%xmm5 # 1da0 <_sk_xor__ssse3_lowp+0x2ab>
- .byte 102,15,40,231 // movapd %xmm7,%xmm4
- .byte 102,15,56,0,229 // pshufb %xmm5,%xmm4
- .byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
- .byte 102,15,56,0,245 // pshufb %xmm5,%xmm6
- .byte 102,15,108,230 // punpcklqdq %xmm6,%xmm4
- .byte 102,15,111,13,255,13,0,0 // movdqa 0xdff(%rip),%xmm1 # 1db0 <_sk_xor__ssse3_lowp+0x2bb>
- .byte 102,15,113,244,8 // psllw $0x8,%xmm4
- .byte 102,68,15,111,13,1,14,0,0 // movdqa 0xe01(%rip),%xmm9 # 1dc0 <_sk_xor__ssse3_lowp+0x2cb>
- .byte 102,65,15,228,225 // pmulhuw %xmm9,%xmm4
- .byte 102,68,15,111,21,3,14,0,0 // movdqa 0xe03(%rip),%xmm10 # 1dd0 <_sk_xor__ssse3_lowp+0x2db>
- .byte 102,15,40,239 // movapd %xmm7,%xmm5
- .byte 102,65,15,56,0,234 // pshufb %xmm10,%xmm5
- .byte 102,65,15,111,240 // movdqa %xmm8,%xmm6
- .byte 102,65,15,56,0,242 // pshufb %xmm10,%xmm6
- .byte 102,15,108,238 // punpcklqdq %xmm6,%xmm5
- .byte 102,15,113,245,8 // psllw $0x8,%xmm5
- .byte 102,65,15,228,233 // pmulhuw %xmm9,%xmm5
- .byte 102,68,15,111,21,231,13,0,0 // movdqa 0xde7(%rip),%xmm10 # 1de0 <_sk_xor__ssse3_lowp+0x2eb>
- .byte 102,15,40,247 // movapd %xmm7,%xmm6
- .byte 102,65,15,56,0,242 // pshufb %xmm10,%xmm6
- .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
- .byte 102,69,15,56,0,218 // pshufb %xmm10,%xmm11
- .byte 102,65,15,108,243 // punpcklqdq %xmm11,%xmm6
- .byte 102,15,113,246,8 // psllw $0x8,%xmm6
- .byte 102,65,15,228,241 // pmulhuw %xmm9,%xmm6
- .byte 102,68,15,111,21,202,13,0,0 // movdqa 0xdca(%rip),%xmm10 # 1df0 <_sk_xor__ssse3_lowp+0x2fb>
- .byte 102,65,15,56,0,250 // pshufb %xmm10,%xmm7
- .byte 102,69,15,56,0,194 // pshufb %xmm10,%xmm8
- .byte 102,65,15,108,248 // punpcklqdq %xmm8,%xmm7
- .byte 102,15,113,247,8 // psllw $0x8,%xmm7
- .byte 102,65,15,228,249 // pmulhuw %xmm9,%xmm7
- .byte 102,68,15,111,29,182,13,0,0 // movdqa 0xdb6(%rip),%xmm11 # 1e00 <_sk_xor__ssse3_lowp+0x30b>
- .byte 102,68,15,249,219 // psubw %xmm3,%xmm11
- .byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 102,69,15,56,11,195 // pmulhrsw %xmm11,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,68,15,253,192 // paddw %xmm0,%xmm8
- .byte 102,15,111,197 // movdqa %xmm5,%xmm0
- .byte 102,65,15,56,11,195 // pmulhrsw %xmm11,%xmm0
- .byte 102,68,15,56,29,200 // pabsw %xmm0,%xmm9
- .byte 102,69,15,253,204 // paddw %xmm12,%xmm9
- .byte 102,15,111,198 // movdqa %xmm6,%xmm0
- .byte 102,65,15,56,11,195 // pmulhrsw %xmm11,%xmm0
- .byte 102,68,15,56,29,224 // pabsw %xmm0,%xmm12
- .byte 102,68,15,253,226 // paddw %xmm2,%xmm12
- .byte 102,68,15,56,11,223 // pmulhrsw %xmm7,%xmm11
- .byte 102,69,15,56,29,219 // pabsw %xmm11,%xmm11
- .byte 102,68,15,253,219 // paddw %xmm3,%xmm11
- .byte 102,68,15,127,68,36,240 // movdqa %xmm8,-0x10(%rsp)
- .byte 102,65,15,111,216 // movdqa %xmm8,%xmm3
- .byte 102,15,113,211,7 // psrlw $0x7,%xmm3
- .byte 102,15,234,217 // pminsw %xmm1,%xmm3
- .byte 102,69,15,239,255 // pxor %xmm15,%xmm15
- .byte 102,68,15,111,235 // movdqa %xmm3,%xmm13
- .byte 102,69,15,97,239 // punpcklwd %xmm15,%xmm13
- .byte 102,65,15,105,223 // punpckhwd %xmm15,%xmm3
- .byte 102,65,15,111,209 // movdqa %xmm9,%xmm2
- .byte 102,15,113,210,7 // psrlw $0x7,%xmm2
- .byte 102,15,234,209 // pminsw %xmm1,%xmm2
- .byte 102,68,15,111,242 // movdqa %xmm2,%xmm14
- .byte 102,69,15,105,247 // punpckhwd %xmm15,%xmm14
- .byte 102,65,15,97,215 // punpcklwd %xmm15,%xmm2
- .byte 102,69,15,111,212 // movdqa %xmm12,%xmm10
- .byte 102,65,15,113,210,7 // psrlw $0x7,%xmm10
- .byte 102,68,15,234,209 // pminsw %xmm1,%xmm10
- .byte 102,69,15,111,194 // movdqa %xmm10,%xmm8
- .byte 102,69,15,97,199 // punpcklwd %xmm15,%xmm8
- .byte 102,69,15,105,215 // punpckhwd %xmm15,%xmm10
- .byte 102,65,15,111,195 // movdqa %xmm11,%xmm0
- .byte 102,15,113,208,7 // psrlw $0x7,%xmm0
- .byte 102,15,234,193 // pminsw %xmm1,%xmm0
- .byte 102,15,111,200 // movdqa %xmm0,%xmm1
- .byte 102,65,15,105,207 // punpckhwd %xmm15,%xmm1
- .byte 102,65,15,97,199 // punpcklwd %xmm15,%xmm0
- .byte 102,65,15,114,246,8 // pslld $0x8,%xmm14
- .byte 102,68,15,235,243 // por %xmm3,%xmm14
- .byte 102,15,114,242,8 // pslld $0x8,%xmm2
- .byte 102,65,15,235,213 // por %xmm13,%xmm2
- .byte 102,65,15,114,242,16 // pslld $0x10,%xmm10
- .byte 102,15,114,241,24 // pslld $0x18,%xmm1
- .byte 102,65,15,235,202 // por %xmm10,%xmm1
- .byte 102,65,15,235,206 // por %xmm14,%xmm1
- .byte 102,65,15,114,240,16 // pslld $0x10,%xmm8
- .byte 102,15,114,240,24 // pslld $0x18,%xmm0
- .byte 102,65,15,235,192 // por %xmm8,%xmm0
- .byte 102,15,235,194 // por %xmm2,%xmm0
- .byte 77,133,192 // test %r8,%r8
- .byte 117,98 // jne 11c6 <_sk_srcover_rgba_8888_ssse3_lowp+0x26e>
- .byte 243,65,15,127,4,153 // movdqu %xmm0,(%r9,%rbx,4)
- .byte 243,65,15,127,76,153,16 // movdqu %xmm1,0x10(%r9,%rbx,4)
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,68,36,240 // movaps -0x10(%rsp),%xmm0
- .byte 102,65,15,111,201 // movdqa %xmm9,%xmm1
- .byte 102,65,15,111,212 // movdqa %xmm12,%xmm2
- .byte 102,65,15,111,219 // movdqa %xmm11,%xmm3
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,15,239,255 // pxor %xmm7,%xmm7
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,227,253,255,255 // ja f8a <_sk_srcover_rgba_8888_ssse3_lowp+0x32>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,206,0,0,0 // lea 0xce(%rip),%r11 # 1280 <_sk_srcover_rgba_8888_ssse3_lowp+0x328>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 243,65,15,16,60,153 // movss (%r9,%rbx,4),%xmm7
- .byte 233,196,253,255,255 // jmpq f8a <_sk_srcover_rgba_8888_ssse3_lowp+0x32>
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,155 // ja 1171 <_sk_srcover_rgba_8888_ssse3_lowp+0x219>
- .byte 65,15,182,194 // movzbl %r10b,%eax
- .byte 76,141,21,187,0,0,0 // lea 0xbb(%rip),%r10 # 129c <_sk_srcover_rgba_8888_ssse3_lowp+0x344>
- .byte 73,99,4,130 // movslq (%r10,%rax,4),%rax
- .byte 76,1,208 // add %r10,%rax
- .byte 255,224 // jmpq *%rax
- .byte 102,65,15,126,4,153 // movd %xmm0,(%r9,%rbx,4)
- .byte 233,124,255,255,255 // jmpq 1171 <_sk_srcover_rgba_8888_ssse3_lowp+0x219>
- .byte 102,65,15,110,100,153,8 // movd 0x8(%r9,%rbx,4),%xmm4
- .byte 102,15,112,252,69 // pshufd $0x45,%xmm4,%xmm7
- .byte 102,65,15,18,60,153 // movlpd (%r9,%rbx,4),%xmm7
- .byte 233,126,253,255,255 // jmpq f8a <_sk_srcover_rgba_8888_ssse3_lowp+0x32>
- .byte 102,65,15,110,100,153,24 // movd 0x18(%r9,%rbx,4),%xmm4
- .byte 102,68,15,112,196,69 // pshufd $0x45,%xmm4,%xmm8
- .byte 243,65,15,16,100,153,20 // movss 0x14(%r9,%rbx,4),%xmm4
- .byte 65,15,198,224,0 // shufps $0x0,%xmm8,%xmm4
- .byte 65,15,198,224,226 // shufps $0xe2,%xmm8,%xmm4
- .byte 68,15,40,196 // movaps %xmm4,%xmm8
- .byte 243,65,15,16,100,153,16 // movss 0x10(%r9,%rbx,4),%xmm4
- .byte 243,68,15,16,196 // movss %xmm4,%xmm8
- .byte 233,69,253,255,255 // jmpq f84 <_sk_srcover_rgba_8888_ssse3_lowp+0x2c>
- .byte 102,15,112,200,78 // pshufd $0x4e,%xmm0,%xmm1
- .byte 102,65,15,126,76,153,8 // movd %xmm1,0x8(%r9,%rbx,4)
- .byte 102,65,15,214,4,153 // movq %xmm0,(%r9,%rbx,4)
- .byte 233,27,255,255,255 // jmpq 1171 <_sk_srcover_rgba_8888_ssse3_lowp+0x219>
- .byte 102,15,112,209,78 // pshufd $0x4e,%xmm1,%xmm2
- .byte 102,65,15,126,84,153,24 // movd %xmm2,0x18(%r9,%rbx,4)
- .byte 102,15,112,209,229 // pshufd $0xe5,%xmm1,%xmm2
- .byte 102,65,15,126,84,153,20 // movd %xmm2,0x14(%r9,%rbx,4)
- .byte 102,65,15,126,76,153,16 // movd %xmm1,0x10(%r9,%rbx,4)
- .byte 243,65,15,127,4,153 // movdqu %xmm0,(%r9,%rbx,4)
- .byte 233,241,254,255,255 // jmpq 1171 <_sk_srcover_rgba_8888_ssse3_lowp+0x219>
- .byte 59,255 // cmp %edi,%edi
- .byte 255 // (bad)
- .byte 255,129,255,255,255,117 // incl 0x75ffffff(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,4,253,255,255,174,255 // incl -0x510001(,%rdi,8)
- .byte 255 // (bad)
- .byte 255,153,255,255,255,140 // lcall *-0x73000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,78,255 // decl -0x1(%rsi)
- .byte 255 // (bad)
- .byte 255,175,255,255,255,163 // ljmp *-0x5c000001(%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 217,255 // fcos
- .byte 255 // (bad)
- .byte 255,210 // callq *%rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,198 // inc %esi
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 186 // .byte 0xba
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_scale_1_float_ssse3_lowp
-.globl _sk_scale_1_float_ssse3_lowp
-FUNCTION(_sk_scale_1_float_ssse3_lowp)
-_sk_scale_1_float_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,88,5,28,12,0,0 // addss 0xc1c(%rip),%xmm8 # 1ee4 <_sk_xor__ssse3_lowp+0x3ef>
- .byte 102,68,15,126,192 // movd %xmm8,%eax
- .byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
- .byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,56,11,216 // pmulhrsw %xmm8,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_scale_u8_ssse3_lowp
-.globl _sk_scale_u8_ssse3_lowp
-FUNCTION(_sk_scale_u8_ssse3_lowp)
-_sk_scale_u8_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 117,75 // jne 1372 <_sk_scale_u8_ssse3_lowp+0x64>
- .byte 243,69,15,126,4,25 // movq (%r9,%rbx,1),%xmm8
- .byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
- .byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,228,5,207,10,0,0 // pmulhuw 0xacf(%rip),%xmm8 # 1e10 <_sk_xor__ssse3_lowp+0x31b>
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,56,11,216 // pmulhrsw %xmm8,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 119,171 // ja 1332 <_sk_scale_u8_ssse3_lowp+0x24>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,126,0,0,0 // lea 0x7e(%rip),%r11 # 1410 <_sk_scale_u8_ssse3_lowp+0x102>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 235,139 // jmp 1332 <_sk_scale_u8_ssse3_lowp+0x24>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,68,15,196,192,2 // pinsrw $0x2,%eax,%xmm8
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,68,15,110,200 // movd %eax,%xmm9
- .byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
- .byte 243,69,15,16,193 // movss %xmm9,%xmm8
- .byte 233,97,255,255,255 // jmpq 1332 <_sk_scale_u8_ssse3_lowp+0x24>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,68,15,196,192,6 // pinsrw $0x6,%eax,%xmm8
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,68,15,196,192,5 // pinsrw $0x5,%eax,%xmm8
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,68,15,196,192,4 // pinsrw $0x4,%eax,%xmm8
- .byte 102,69,15,110,12,25 // movd (%r9,%rbx,1),%xmm9
- .byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
- .byte 242,69,15,16,193 // movsd %xmm9,%xmm8
- .byte 233,35,255,255,255 // jmpq 1332 <_sk_scale_u8_ssse3_lowp+0x24>
- .byte 144 // nop
- .byte 139,255 // mov %edi,%edi
- .byte 255 // (bad)
- .byte 255,168,255,255,255,151 // ljmp *-0x68000001(%rax)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 234 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 222,255 // fdivrp %st,%st(7)
- .byte 255 // (bad)
- .byte 255,210 // callq *%rdx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,193 // inc %ecx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_lerp_1_float_ssse3_lowp
-.globl _sk_lerp_1_float_ssse3_lowp
-FUNCTION(_sk_lerp_1_float_ssse3_lowp)
-_sk_lerp_1_float_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 243,68,15,16,0 // movss (%rax),%xmm8
- .byte 243,68,15,88,5,172,10,0,0 // addss 0xaac(%rip),%xmm8 # 1ee8 <_sk_xor__ssse3_lowp+0x3f3>
- .byte 102,68,15,126,192 // movd %xmm8,%eax
- .byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 242,69,15,112,192,0 // pshuflw $0x0,%xmm8,%xmm8
- .byte 102,69,15,112,192,80 // pshufd $0x50,%xmm8,%xmm8
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,68,15,56,29,200 // pabsw %xmm0,%xmm9
- .byte 102,68,15,111,21,185,9,0,0 // movdqa 0x9b9(%rip),%xmm10 # 1e20 <_sk_xor__ssse3_lowp+0x32b>
- .byte 102,69,15,249,208 // psubw %xmm8,%xmm10
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,65,15,56,11,194 // pmulhrsw %xmm10,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,193 // paddw %xmm9,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,68,15,56,29,201 // pabsw %xmm1,%xmm9
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,65,15,56,11,202 // pmulhrsw %xmm10,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,201 // paddw %xmm9,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,68,15,56,29,202 // pabsw %xmm2,%xmm9
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,65,15,56,11,210 // pmulhrsw %xmm10,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,209 // paddw %xmm9,%xmm2
- .byte 102,65,15,56,11,216 // pmulhrsw %xmm8,%xmm3
- .byte 102,68,15,56,29,195 // pabsw %xmm3,%xmm8
- .byte 102,68,15,56,11,215 // pmulhrsw %xmm7,%xmm10
- .byte 102,65,15,56,29,218 // pabsw %xmm10,%xmm3
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_lerp_u8_ssse3_lowp
-.globl _sk_lerp_u8_ssse3_lowp
-FUNCTION(_sk_lerp_u8_ssse3_lowp)
-_sk_lerp_u8_ssse3_lowp:
- .byte 83 // push %rbx
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 76,99,80,8 // movslq 0x8(%rax),%r10
- .byte 76,99,201 // movslq %ecx,%r9
- .byte 77,15,175,202 // imul %r10,%r9
- .byte 76,3,8 // add (%rax),%r9
- .byte 72,99,218 // movslq %edx,%rbx
- .byte 77,133,192 // test %r8,%r8
- .byte 15,133,170,0,0,0 // jne 15a8 <_sk_lerp_u8_ssse3_lowp+0xc7>
- .byte 243,69,15,126,4,25 // movq (%r9,%rbx,1),%xmm8
- .byte 102,68,15,96,192 // punpcklbw %xmm0,%xmm8
- .byte 102,65,15,113,240,8 // psllw $0x8,%xmm8
- .byte 102,68,15,228,5,24,9,0,0 // pmulhuw 0x918(%rip),%xmm8 # 1e30 <_sk_xor__ssse3_lowp+0x33b>
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,68,15,56,29,200 // pabsw %xmm0,%xmm9
- .byte 102,68,15,111,21,19,9,0,0 // movdqa 0x913(%rip),%xmm10 # 1e40 <_sk_xor__ssse3_lowp+0x34b>
- .byte 102,69,15,249,208 // psubw %xmm8,%xmm10
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,65,15,56,11,194 // pmulhrsw %xmm10,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,193 // paddw %xmm9,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,68,15,56,29,201 // pabsw %xmm1,%xmm9
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,65,15,56,11,202 // pmulhrsw %xmm10,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,201 // paddw %xmm9,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,68,15,56,29,202 // pabsw %xmm2,%xmm9
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,65,15,56,11,210 // pmulhrsw %xmm10,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,209 // paddw %xmm9,%xmm2
- .byte 102,65,15,56,11,216 // pmulhrsw %xmm8,%xmm3
- .byte 102,68,15,56,29,195 // pabsw %xmm3,%xmm8
- .byte 102,68,15,56,11,215 // pmulhrsw %xmm7,%xmm10
- .byte 102,65,15,56,29,218 // pabsw %xmm10,%xmm3
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 91 // pop %rbx
- .byte 255,224 // jmpq *%rax
- .byte 69,137,194 // mov %r8d,%r10d
- .byte 65,128,226,7 // and $0x7,%r10b
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 65,254,202 // dec %r10b
- .byte 65,128,250,6 // cmp $0x6,%r10b
- .byte 15,135,72,255,255,255 // ja 1509 <_sk_lerp_u8_ssse3_lowp+0x28>
- .byte 69,15,182,210 // movzbl %r10b,%r10d
- .byte 76,141,29,128,0,0,0 // lea 0x80(%rip),%r11 # 164c <_sk_lerp_u8_ssse3_lowp+0x16b>
- .byte 75,99,4,147 // movslq (%r11,%r10,4),%rax
- .byte 76,1,216 // add %r11,%rax
- .byte 255,224 // jmpq *%rax
- .byte 65,15,182,4,25 // movzbl (%r9,%rbx,1),%eax
- .byte 102,68,15,110,192 // movd %eax,%xmm8
- .byte 233,37,255,255,255 // jmpq 1509 <_sk_lerp_u8_ssse3_lowp+0x28>
- .byte 65,15,182,68,25,2 // movzbl 0x2(%r9,%rbx,1),%eax
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,68,15,196,192,2 // pinsrw $0x2,%eax,%xmm8
- .byte 65,15,183,4,25 // movzwl (%r9,%rbx,1),%eax
- .byte 102,68,15,110,200 // movd %eax,%xmm9
- .byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
- .byte 243,69,15,16,193 // movss %xmm9,%xmm8
- .byte 233,251,254,255,255 // jmpq 1509 <_sk_lerp_u8_ssse3_lowp+0x28>
- .byte 65,15,182,68,25,6 // movzbl 0x6(%r9,%rbx,1),%eax
- .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
- .byte 102,68,15,196,192,6 // pinsrw $0x6,%eax,%xmm8
- .byte 65,15,182,68,25,5 // movzbl 0x5(%r9,%rbx,1),%eax
- .byte 102,68,15,196,192,5 // pinsrw $0x5,%eax,%xmm8
- .byte 65,15,182,68,25,4 // movzbl 0x4(%r9,%rbx,1),%eax
- .byte 102,68,15,196,192,4 // pinsrw $0x4,%eax,%xmm8
- .byte 102,69,15,110,12,25 // movd (%r9,%rbx,1),%xmm9
- .byte 102,68,15,96,200 // punpcklbw %xmm0,%xmm9
- .byte 242,69,15,16,193 // movsd %xmm9,%xmm8
- .byte 233,189,254,255,255 // jmpq 1509 <_sk_lerp_u8_ssse3_lowp+0x28>
- .byte 137,255 // mov %edi,%edi
- .byte 255 // (bad)
- .byte 255,169,255,255,255,152 // ljmp *-0x67000001(%rcx)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 235,255 // jmp 1659 <_sk_lerp_u8_ssse3_lowp+0x178>
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 223,255 // (bad)
- .byte 255 // (bad)
- .byte 255,211 // callq *%rbx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,194 // inc %edx
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // .byte 0xff
-
-HIDDEN _sk_swap_rb_ssse3_lowp
-.globl _sk_swap_rb_ssse3_lowp
-FUNCTION(_sk_swap_rb_ssse3_lowp)
-_sk_swap_rb_ssse3_lowp:
- .byte 68,15,40,192 // movaps %xmm0,%xmm8
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,194 // movaps %xmm2,%xmm0
- .byte 65,15,40,208 // movaps %xmm8,%xmm2
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_move_src_dst_ssse3_lowp
-.globl _sk_move_src_dst_ssse3_lowp
-FUNCTION(_sk_move_src_dst_ssse3_lowp)
-_sk_move_src_dst_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,224 // movaps %xmm0,%xmm4
- .byte 15,40,233 // movaps %xmm1,%xmm5
- .byte 15,40,242 // movaps %xmm2,%xmm6
- .byte 15,40,251 // movaps %xmm3,%xmm7
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_move_dst_src_ssse3_lowp
-.globl _sk_move_dst_src_ssse3_lowp
-FUNCTION(_sk_move_dst_src_ssse3_lowp)
-_sk_move_dst_src_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,40,196 // movaps %xmm4,%xmm0
- .byte 15,40,205 // movaps %xmm5,%xmm1
- .byte 15,40,214 // movaps %xmm6,%xmm2
- .byte 15,40,223 // movaps %xmm7,%xmm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_clear_ssse3_lowp
-.globl _sk_clear_ssse3_lowp
-FUNCTION(_sk_clear_ssse3_lowp)
-_sk_clear_ssse3_lowp:
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 15,87,192 // xorps %xmm0,%xmm0
- .byte 15,87,201 // xorps %xmm1,%xmm1
- .byte 15,87,210 // xorps %xmm2,%xmm2
- .byte 15,87,219 // xorps %xmm3,%xmm3
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcatop_ssse3_lowp
-.globl _sk_srcatop_ssse3_lowp
-FUNCTION(_sk_srcatop_ssse3_lowp)
-_sk_srcatop_ssse3_lowp:
- .byte 102,15,56,11,199 // pmulhrsw %xmm7,%xmm0
- .byte 102,68,15,56,29,192 // pabsw %xmm0,%xmm8
- .byte 102,68,15,111,13,149,7,0,0 // movdqa 0x795(%rip),%xmm9 # 1e50 <_sk_xor__ssse3_lowp+0x35b>
- .byte 102,68,15,249,203 // psubw %xmm3,%xmm9
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,65,15,56,11,193 // pmulhrsw %xmm9,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,192 // paddw %xmm8,%xmm0
- .byte 102,15,56,11,207 // pmulhrsw %xmm7,%xmm1
- .byte 102,68,15,56,29,193 // pabsw %xmm1,%xmm8
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,65,15,56,11,201 // pmulhrsw %xmm9,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,200 // paddw %xmm8,%xmm1
- .byte 102,15,56,11,215 // pmulhrsw %xmm7,%xmm2
- .byte 102,68,15,56,29,194 // pabsw %xmm2,%xmm8
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,65,15,56,11,209 // pmulhrsw %xmm9,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,208 // paddw %xmm8,%xmm2
- .byte 102,15,56,11,223 // pmulhrsw %xmm7,%xmm3
- .byte 102,68,15,56,29,195 // pabsw %xmm3,%xmm8
- .byte 102,68,15,56,11,207 // pmulhrsw %xmm7,%xmm9
- .byte 102,65,15,56,29,217 // pabsw %xmm9,%xmm3
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstatop_ssse3_lowp
-.globl _sk_dstatop_ssse3_lowp
-FUNCTION(_sk_dstatop_ssse3_lowp)
-_sk_dstatop_ssse3_lowp:
- .byte 102,68,15,111,196 // movdqa %xmm4,%xmm8
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,68,15,111,13,20,7,0,0 // movdqa 0x714(%rip),%xmm9 # 1e60 <_sk_xor__ssse3_lowp+0x36b>
- .byte 102,68,15,249,207 // psubw %xmm7,%xmm9
- .byte 102,65,15,56,11,193 // pmulhrsw %xmm9,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,192 // paddw %xmm8,%xmm0
- .byte 102,68,15,111,197 // movdqa %xmm5,%xmm8
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,65,15,56,11,201 // pmulhrsw %xmm9,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,200 // paddw %xmm8,%xmm1
- .byte 102,68,15,111,198 // movdqa %xmm6,%xmm8
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,65,15,56,11,209 // pmulhrsw %xmm9,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,208 // paddw %xmm8,%xmm2
- .byte 102,68,15,111,199 // movdqa %xmm7,%xmm8
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,68,15,56,11,203 // pmulhrsw %xmm3,%xmm9
- .byte 102,65,15,56,29,217 // pabsw %xmm9,%xmm3
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcin_ssse3_lowp
-.globl _sk_srcin_ssse3_lowp
-FUNCTION(_sk_srcin_ssse3_lowp)
-_sk_srcin_ssse3_lowp:
- .byte 102,15,56,11,199 // pmulhrsw %xmm7,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,56,11,207 // pmulhrsw %xmm7,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,56,11,215 // pmulhrsw %xmm7,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,15,56,11,223 // pmulhrsw %xmm7,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstin_ssse3_lowp
-.globl _sk_dstin_ssse3_lowp
-FUNCTION(_sk_dstin_ssse3_lowp)
-_sk_dstin_ssse3_lowp:
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,15,56,11,195 // pmulhrsw %xmm3,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,15,56,11,203 // pmulhrsw %xmm3,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,15,56,11,211 // pmulhrsw %xmm3,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,15,56,11,223 // pmulhrsw %xmm7,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcout_ssse3_lowp
-.globl _sk_srcout_ssse3_lowp
-FUNCTION(_sk_srcout_ssse3_lowp)
-_sk_srcout_ssse3_lowp:
- .byte 102,68,15,111,5,58,6,0,0 // movdqa 0x63a(%rip),%xmm8 # 1e70 <_sk_xor__ssse3_lowp+0x37b>
- .byte 102,68,15,249,199 // psubw %xmm7,%xmm8
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,65,15,56,29,216 // pabsw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstout_ssse3_lowp
-.globl _sk_dstout_ssse3_lowp
-FUNCTION(_sk_dstout_ssse3_lowp)
-_sk_dstout_ssse3_lowp:
- .byte 102,68,15,111,5,11,6,0,0 // movdqa 0x60b(%rip),%xmm8 # 1e80 <_sk_xor__ssse3_lowp+0x38b>
- .byte 102,68,15,249,195 // psubw %xmm3,%xmm8
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,68,15,56,11,199 // pmulhrsw %xmm7,%xmm8
- .byte 102,65,15,56,29,216 // pabsw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_srcover_ssse3_lowp
-.globl _sk_srcover_ssse3_lowp
-FUNCTION(_sk_srcover_ssse3_lowp)
-_sk_srcover_ssse3_lowp:
- .byte 102,68,15,111,5,208,5,0,0 // movdqa 0x5d0(%rip),%xmm8 # 1e90 <_sk_xor__ssse3_lowp+0x39b>
- .byte 102,68,15,249,195 // psubw %xmm3,%xmm8
- .byte 102,68,15,111,204 // movdqa %xmm4,%xmm9
- .byte 102,69,15,56,11,200 // pmulhrsw %xmm8,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,193 // paddw %xmm9,%xmm0
- .byte 102,68,15,111,205 // movdqa %xmm5,%xmm9
- .byte 102,69,15,56,11,200 // pmulhrsw %xmm8,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,201 // paddw %xmm9,%xmm1
- .byte 102,68,15,111,206 // movdqa %xmm6,%xmm9
- .byte 102,69,15,56,11,200 // pmulhrsw %xmm8,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,209 // paddw %xmm9,%xmm2
- .byte 102,68,15,56,11,199 // pmulhrsw %xmm7,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_dstover_ssse3_lowp
-.globl _sk_dstover_ssse3_lowp
-FUNCTION(_sk_dstover_ssse3_lowp)
-_sk_dstover_ssse3_lowp:
- .byte 102,68,15,111,5,123,5,0,0 // movdqa 0x57b(%rip),%xmm8 # 1ea0 <_sk_xor__ssse3_lowp+0x3ab>
- .byte 102,68,15,249,199 // psubw %xmm7,%xmm8
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,253,196 // paddw %xmm4,%xmm0
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,253,205 // paddw %xmm5,%xmm1
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,15,253,214 // paddw %xmm6,%xmm2
- .byte 102,68,15,56,11,195 // pmulhrsw %xmm3,%xmm8
- .byte 102,65,15,56,29,216 // pabsw %xmm8,%xmm3
- .byte 102,15,253,223 // paddw %xmm7,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_modulate_ssse3_lowp
-.globl _sk_modulate_ssse3_lowp
-FUNCTION(_sk_modulate_ssse3_lowp)
-_sk_modulate_ssse3_lowp:
- .byte 102,15,56,11,196 // pmulhrsw %xmm4,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,15,56,11,205 // pmulhrsw %xmm5,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,15,56,11,214 // pmulhrsw %xmm6,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,15,56,11,223 // pmulhrsw %xmm7,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_multiply_ssse3_lowp
-.globl _sk_multiply_ssse3_lowp
-FUNCTION(_sk_multiply_ssse3_lowp)
-_sk_multiply_ssse3_lowp:
- .byte 102,68,15,111,5,16,5,0,0 // movdqa 0x510(%rip),%xmm8 # 1eb0 <_sk_xor__ssse3_lowp+0x3bb>
- .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
- .byte 102,68,15,249,207 // psubw %xmm7,%xmm9
- .byte 102,68,15,111,208 // movdqa %xmm0,%xmm10
- .byte 102,69,15,56,11,209 // pmulhrsw %xmm9,%xmm10
- .byte 102,69,15,56,29,210 // pabsw %xmm10,%xmm10
- .byte 102,68,15,249,195 // psubw %xmm3,%xmm8
- .byte 102,15,56,11,196 // pmulhrsw %xmm4,%xmm0
- .byte 102,68,15,111,220 // movdqa %xmm4,%xmm11
- .byte 102,69,15,56,11,216 // pmulhrsw %xmm8,%xmm11
- .byte 102,69,15,56,29,219 // pabsw %xmm11,%xmm11
- .byte 102,69,15,253,218 // paddw %xmm10,%xmm11
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,195 // paddw %xmm11,%xmm0
- .byte 102,68,15,111,209 // movdqa %xmm1,%xmm10
- .byte 102,69,15,56,11,209 // pmulhrsw %xmm9,%xmm10
- .byte 102,69,15,56,29,210 // pabsw %xmm10,%xmm10
- .byte 102,15,56,11,205 // pmulhrsw %xmm5,%xmm1
- .byte 102,68,15,111,221 // movdqa %xmm5,%xmm11
- .byte 102,69,15,56,11,216 // pmulhrsw %xmm8,%xmm11
- .byte 102,69,15,56,29,219 // pabsw %xmm11,%xmm11
- .byte 102,69,15,253,218 // paddw %xmm10,%xmm11
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,203 // paddw %xmm11,%xmm1
- .byte 102,68,15,111,210 // movdqa %xmm2,%xmm10
- .byte 102,69,15,56,11,209 // pmulhrsw %xmm9,%xmm10
- .byte 102,69,15,56,29,210 // pabsw %xmm10,%xmm10
- .byte 102,15,56,11,214 // pmulhrsw %xmm6,%xmm2
- .byte 102,68,15,111,222 // movdqa %xmm6,%xmm11
- .byte 102,69,15,56,11,216 // pmulhrsw %xmm8,%xmm11
- .byte 102,69,15,56,29,219 // pabsw %xmm11,%xmm11
- .byte 102,69,15,253,218 // paddw %xmm10,%xmm11
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,211 // paddw %xmm11,%xmm2
- .byte 102,68,15,56,11,203 // pmulhrsw %xmm3,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,68,15,56,11,199 // pmulhrsw %xmm7,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,69,15,253,193 // paddw %xmm9,%xmm8
- .byte 102,15,56,11,223 // pmulhrsw %xmm7,%xmm3
- .byte 102,15,56,29,219 // pabsw %xmm3,%xmm3
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_screen_ssse3_lowp
-.globl _sk_screen_ssse3_lowp
-FUNCTION(_sk_screen_ssse3_lowp)
-_sk_screen_ssse3_lowp:
- .byte 102,68,15,111,5,54,4,0,0 // movdqa 0x436(%rip),%xmm8 # 1ec0 <_sk_xor__ssse3_lowp+0x3cb>
- .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
- .byte 102,68,15,249,200 // psubw %xmm0,%xmm9
- .byte 102,68,15,56,11,204 // pmulhrsw %xmm4,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,193 // paddw %xmm9,%xmm0
- .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
- .byte 102,68,15,249,201 // psubw %xmm1,%xmm9
- .byte 102,68,15,56,11,205 // pmulhrsw %xmm5,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,201 // paddw %xmm9,%xmm1
- .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
- .byte 102,68,15,249,202 // psubw %xmm2,%xmm9
- .byte 102,68,15,56,11,206 // pmulhrsw %xmm6,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,65,15,253,209 // paddw %xmm9,%xmm2
- .byte 102,68,15,249,195 // psubw %xmm3,%xmm8
- .byte 102,68,15,56,11,199 // pmulhrsw %xmm7,%xmm8
- .byte 102,69,15,56,29,192 // pabsw %xmm8,%xmm8
- .byte 102,65,15,253,216 // paddw %xmm8,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-HIDDEN _sk_xor__ssse3_lowp
-.globl _sk_xor__ssse3_lowp
-FUNCTION(_sk_xor__ssse3_lowp)
-_sk_xor__ssse3_lowp:
- .byte 102,68,15,111,5,210,3,0,0 // movdqa 0x3d2(%rip),%xmm8 # 1ed0 <_sk_xor__ssse3_lowp+0x3db>
- .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
- .byte 102,68,15,249,207 // psubw %xmm7,%xmm9
- .byte 102,65,15,56,11,193 // pmulhrsw %xmm9,%xmm0
- .byte 102,68,15,56,29,208 // pabsw %xmm0,%xmm10
- .byte 102,68,15,249,195 // psubw %xmm3,%xmm8
- .byte 102,15,111,196 // movdqa %xmm4,%xmm0
- .byte 102,65,15,56,11,192 // pmulhrsw %xmm8,%xmm0
- .byte 102,15,56,29,192 // pabsw %xmm0,%xmm0
- .byte 102,65,15,253,194 // paddw %xmm10,%xmm0
- .byte 102,65,15,56,11,201 // pmulhrsw %xmm9,%xmm1
- .byte 102,68,15,56,29,209 // pabsw %xmm1,%xmm10
- .byte 102,15,111,205 // movdqa %xmm5,%xmm1
- .byte 102,65,15,56,11,200 // pmulhrsw %xmm8,%xmm1
- .byte 102,15,56,29,201 // pabsw %xmm1,%xmm1
- .byte 102,65,15,253,202 // paddw %xmm10,%xmm1
- .byte 102,65,15,56,11,209 // pmulhrsw %xmm9,%xmm2
- .byte 102,68,15,56,29,210 // pabsw %xmm2,%xmm10
- .byte 102,15,111,214 // movdqa %xmm6,%xmm2
- .byte 102,65,15,56,11,208 // pmulhrsw %xmm8,%xmm2
- .byte 102,15,56,29,210 // pabsw %xmm2,%xmm2
- .byte 102,65,15,253,210 // paddw %xmm10,%xmm2
- .byte 102,68,15,56,11,203 // pmulhrsw %xmm3,%xmm9
- .byte 102,69,15,56,29,201 // pabsw %xmm9,%xmm9
- .byte 102,68,15,56,11,199 // pmulhrsw %xmm7,%xmm8
- .byte 102,65,15,56,29,216 // pabsw %xmm8,%xmm3
- .byte 102,65,15,253,217 // paddw %xmm9,%xmm3
- .byte 72,173 // lods %ds:(%rsi),%rax
- .byte 255,224 // jmpq *%rax
-
-BALIGN16
- .byte 0,0 // add %al,(%rax)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,128 // addb $0x80,0x0(%rbx)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 12,13 // or $0xd,%al
- .byte 14 // (bad)
- .byte 15,129,128,129,128,129 // jno ffffffff81809d55 <_sk_xor__ssse3_lowp+0xffffffff81808260>
- .byte 128,129,128,129,128,129,128 // addb $0x80,-0x7e7f7e80(%rcx)
- .byte 129,128,129,128,1,2,5,6,9,10 // addl $0xa090605,0x2018081(%rax)
- .byte 13,14,9,10,13 // or $0xd0a090e,%eax
- .byte 14 // (bad)
- .byte 13,14,15,255,2 // or $0x2ff0f0e,%eax
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,10 // (bad)
- .byte 11,14 // or (%rsi),%ecx
- .byte 15,14 // femms
- .byte 15,255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 1,4,5,8,9,12,13 // add %eax,0xd0c0908(,%rax,1)
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 12,13 // or $0xd,%al
- .byte 14 // (bad)
- .byte 15,129,128,129,128,129 // jno ffffffff81809da5 <_sk_xor__ssse3_lowp+0xffffffff818082b0>
- .byte 128,129,128,129,128,129,128 // addb $0x80,-0x7e7f7e80(%rcx)
- .byte 129,128,129,128,1,2,5,6,9,10 // addl $0xa090605,0x2018081(%rax)
- .byte 13,14,9,10,13 // or $0xd0a090e,%eax
- .byte 14 // (bad)
- .byte 13,14,15,255,2 // or $0x2ff0f0e,%eax
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,10 // (bad)
- .byte 11,14 // or (%rsi),%ecx
- .byte 15,14 // femms
- .byte 15,255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 12,13 // or $0xd,%al
- .byte 14 // (bad)
- .byte 15,129,128,129,128,129 // jno ffffffff81809e05 <_sk_xor__ssse3_lowp+0xffffffff81808310>
- .byte 128,129,128,129,128,129,128 // addb $0x80,-0x7e7f7e80(%rcx)
- .byte 129,128,129,128,1,2,5,6,9,10 // addl $0xa090605,0x2018081(%rax)
- .byte 13,14,9,10,13 // or $0xd0a090e,%eax
- .byte 14 // (bad)
- .byte 13,14,15,255,2 // or $0x2ff0f0e,%eax
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,10 // (bad)
- .byte 11,14 // or (%rsi),%ecx
- .byte 15,14 // femms
- .byte 15,255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 1,4,5,8,9,12,13 // add %eax,0xd0c0908(,%rax,1)
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 12,13 // or $0xd,%al
- .byte 14 // (bad)
- .byte 15,129,128,129,128,129 // jno ffffffff81809e55 <_sk_xor__ssse3_lowp+0xffffffff81808360>
- .byte 128,129,128,129,128,129,128 // addb $0x80,-0x7e7f7e80(%rcx)
- .byte 129,128,129,128,1,2,5,6,9,10 // addl $0xa090605,0x2018081(%rax)
- .byte 13,14,9,10,13 // or $0xd0a090e,%eax
- .byte 14 // (bad)
- .byte 13,14,15,255,2 // or $0x2ff0f0e,%eax
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,10 // (bad)
- .byte 11,14 // or (%rsi),%ecx
- .byte 15,14 // femms
- .byte 15,255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 255,0 // incl (%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,0,2,4,6,0,0,0,0 // addl $0x0,0x6040200(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,2 // add %al,(%rdx)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 0,0 // add %al,(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,0,128,0,128 // addl $0x80008000,-0x7f7e7f7f(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,0,128,0,128 // addl $0x80008000,-0x7f7e7f7f(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,1 // add %al,(%rcx)
- .byte 4,5 // add $0x5,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 8,9 // or %cl,(%rcx)
- .byte 12,13 // or $0xd,%al
- .byte 12,13 // or $0xd,%al
- .byte 14 // (bad)
- .byte 15,255 // (bad)
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,255 // add %bh,%bh
- .byte 0,129,128,129,128,129 // add %al,-0x7e7f7e80(%rcx)
- .byte 128,129,128,129,128,129,128 // addb $0x80,-0x7e7f7e80(%rcx)
- .byte 129,128,129,128,1,2,5,6,9,10 // addl $0xa090605,0x2018081(%rax)
- .byte 13,14,9,10,13 // or $0xd0a090e,%eax
- .byte 14 // (bad)
- .byte 13,14,15,255,2 // or $0x2ff0f0e,%eax
- .byte 3,6 // add (%rsi),%eax
- .byte 7 // (bad)
- .byte 10,11 // or (%rbx),%cl
- .byte 14 // (bad)
- .byte 15,10 // (bad)
- .byte 11,14 // or (%rsi),%ecx
- .byte 15,14 // femms
- .byte 15,255 // (bad)
- .byte 255,3 // incl (%rbx)
- .byte 255,7 // incl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,11 // decl (%rbx)
- .byte 255,15 // decl (%rdi)
- .byte 255,15 // decl (%rdi)
- .byte 255 // (bad)
- .byte 255 // (bad)
- .byte 255,0 // incl (%rax)
- .byte 128,0,128 // addb $0x80,(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,0,128,0,128 // addl $0x80008000,-0x7f7e7f7f(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 129,128,129,128,129,128,129,128,129,128// addl $0x80818081,-0x7f7e7f7f(%rax)
- .byte 129,128,129,128,129,128,0,128,0,128 // addl $0x80008000,-0x7f7e7f7f(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
- .byte 0,128,0,128,0,128 // add %al,-0x7fff8000(%rax)
-
-BALIGN4
- .byte 0,0 // add %al,(%rax)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128,67,0,0 // addb $0x0,0x0(%rbx)
- .byte 128 // .byte 0x80
- .byte 67 // rex.XB
-BALIGN32
-
HIDDEN _sk_start_pipeline_hsw_8bit
.globl _sk_start_pipeline_hsw_8bit
FUNCTION(_sk_start_pipeline_hsw_8bit)
diff --git a/src/jumper/SkJumper_generated_win.S b/src/jumper/SkJumper_generated_win.S
index 414fef3eac..2765623570 100644
--- a/src/jumper/SkJumper_generated_win.S
+++ b/src/jumper/SkJumper_generated_win.S
@@ -31231,4929 +31231,6 @@ ALIGN 4
DB 0,63 ; add %bh,(%rdi)
ALIGN 32
-PUBLIC _sk_start_pipeline_hsw_lowp
-_sk_start_pipeline_hsw_lowp LABEL PROC
- DB 85 ; push %rbp
- DB 72,137,229 ; mov %rsp,%rbp
- DB 65,87 ; push %r15
- DB 65,86 ; push %r14
- DB 65,85 ; push %r13
- DB 65,84 ; push %r12
- DB 86 ; push %rsi
- DB 87 ; push %rdi
- DB 83 ; push %rbx
- DB 72,129,236,200,0,0,0 ; sub $0xc8,%rsp
- DB 197,120,41,125,176 ; vmovaps %xmm15,-0x50(%rbp)
- DB 197,120,41,117,160 ; vmovaps %xmm14,-0x60(%rbp)
- DB 197,120,41,109,144 ; vmovaps %xmm13,-0x70(%rbp)
- DB 197,120,41,101,128 ; vmovaps %xmm12,-0x80(%rbp)
- DB 197,120,41,157,112,255,255,255 ; vmovaps %xmm11,-0x90(%rbp)
- DB 197,120,41,149,96,255,255,255 ; vmovaps %xmm10,-0xa0(%rbp)
- DB 197,120,41,141,80,255,255,255 ; vmovaps %xmm9,-0xb0(%rbp)
- DB 197,120,41,133,64,255,255,255 ; vmovaps %xmm8,-0xc0(%rbp)
- DB 197,248,41,189,48,255,255,255 ; vmovaps %xmm7,-0xd0(%rbp)
- DB 197,248,41,181,32,255,255,255 ; vmovaps %xmm6,-0xe0(%rbp)
- DB 77,137,199 ; mov %r8,%r15
- DB 72,137,211 ; mov %rdx,%rbx
- DB 72,137,141,24,255,255,255 ; mov %rcx,-0xe8(%rbp)
- DB 72,139,117,48 ; mov 0x30(%rbp),%rsi
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 73,137,197 ; mov %rax,%r13
- DB 73,137,246 ; mov %rsi,%r14
- DB 76,137,141,8,255,255,255 ; mov %r9,-0xf8(%rbp)
- DB 76,57,203 ; cmp %r9,%rbx
- DB 115,110 ; jae ed <_sk_start_pipeline_hsw_lowp+0xed>
- DB 72,139,133,24,255,255,255 ; mov -0xe8(%rbp),%rax
- DB 72,141,64,16 ; lea 0x10(%rax),%rax
- DB 72,137,133,16,255,255,255 ; mov %rax,-0xf0(%rbp)
- DB 72,139,149,24,255,255,255 ; mov -0xe8(%rbp),%rdx
- DB 76,57,189,16,255,255,255 ; cmp %r15,-0xf0(%rbp)
- DB 119,43 ; ja cc <_sk_start_pipeline_hsw_lowp+0xcc>
- DB 76,139,165,24,255,255,255 ; mov -0xe8(%rbp),%r12
- DB 69,49,192 ; xor %r8d,%r8d
- DB 72,139,125,56 ; mov 0x38(%rbp),%rdi
- DB 76,137,246 ; mov %r14,%rsi
- DB 76,137,226 ; mov %r12,%rdx
- DB 72,137,217 ; mov %rbx,%rcx
- DB 65,255,213 ; callq *%r13
- DB 73,141,84,36,16 ; lea 0x10(%r12),%rdx
- DB 73,131,196,32 ; add $0x20,%r12
- DB 77,57,252 ; cmp %r15,%r12
- DB 73,137,212 ; mov %rdx,%r12
- DB 118,220 ; jbe a8 <_sk_start_pipeline_hsw_lowp+0xa8>
- DB 77,137,248 ; mov %r15,%r8
- DB 73,41,208 ; sub %rdx,%r8
- DB 116,13 ; je e1 <_sk_start_pipeline_hsw_lowp+0xe1>
- DB 72,139,125,56 ; mov 0x38(%rbp),%rdi
- DB 76,137,246 ; mov %r14,%rsi
- DB 72,137,217 ; mov %rbx,%rcx
- DB 65,255,213 ; callq *%r13
- DB 72,255,195 ; inc %rbx
- DB 72,59,157,8,255,255,255 ; cmp -0xf8(%rbp),%rbx
- DB 117,164 ; jne 91 <_sk_start_pipeline_hsw_lowp+0x91>
- DB 197,248,40,181,32,255,255,255 ; vmovaps -0xe0(%rbp),%xmm6
- DB 197,248,40,189,48,255,255,255 ; vmovaps -0xd0(%rbp),%xmm7
- DB 197,120,40,133,64,255,255,255 ; vmovaps -0xc0(%rbp),%xmm8
- DB 197,120,40,141,80,255,255,255 ; vmovaps -0xb0(%rbp),%xmm9
- DB 197,120,40,149,96,255,255,255 ; vmovaps -0xa0(%rbp),%xmm10
- DB 197,120,40,157,112,255,255,255 ; vmovaps -0x90(%rbp),%xmm11
- DB 197,120,40,101,128 ; vmovaps -0x80(%rbp),%xmm12
- DB 197,120,40,109,144 ; vmovaps -0x70(%rbp),%xmm13
- DB 197,120,40,117,160 ; vmovaps -0x60(%rbp),%xmm14
- DB 197,120,40,125,176 ; vmovaps -0x50(%rbp),%xmm15
- DB 72,129,196,200,0,0,0 ; add $0xc8,%rsp
- DB 91 ; pop %rbx
- DB 95 ; pop %rdi
- DB 94 ; pop %rsi
- DB 65,92 ; pop %r12
- DB 65,93 ; pop %r13
- DB 65,94 ; pop %r14
- DB 65,95 ; pop %r15
- DB 93 ; pop %rbp
- DB 197,248,119 ; vzeroupper
- DB 195 ; retq
-
-PUBLIC _sk_just_return_hsw_lowp
-_sk_just_return_hsw_lowp LABEL PROC
- DB 195 ; retq
-
-PUBLIC _sk_uniform_color_hsw_lowp
-_sk_uniform_color_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 196,226,121,24,5,28,34,0,0 ; vbroadcastss 0x221c(%rip),%xmm0 # 2370 <_sk_xor__hsw_lowp+0x78>
- DB 197,248,88,24 ; vaddps (%rax),%xmm0,%xmm3
- DB 196,226,125,121,195 ; vpbroadcastw %xmm3,%ymm0
- DB 197,251,112,203,234 ; vpshuflw $0xea,%xmm3,%xmm1
- DB 196,226,125,88,201 ; vpbroadcastd %xmm1,%ymm1
- DB 196,227,121,4,211,230 ; vpermilps $0xe6,%xmm3,%xmm2
- DB 197,251,112,210,224 ; vpshuflw $0xe0,%xmm2,%xmm2
- DB 196,226,125,88,210 ; vpbroadcastd %xmm2,%ymm2
- DB 196,227,121,4,219,236 ; vpermilps $0xec,%xmm3,%xmm3
- DB 197,251,112,219,234 ; vpshuflw $0xea,%xmm3,%xmm3
- DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_black_color_hsw_lowp
-_sk_black_color_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 196,226,125,121,29,234,33,0,0 ; vpbroadcastw 0x21ea(%rip),%ymm3 # 2380 <_sk_xor__hsw_lowp+0x88>
- DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
- DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
- DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_white_color_hsw_lowp
-_sk_white_color_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 196,226,125,121,5,211,33,0,0 ; vpbroadcastw 0x21d3(%rip),%ymm0 # 2382 <_sk_xor__hsw_lowp+0x8a>
- DB 197,253,111,200 ; vmovdqa %ymm0,%ymm1
- DB 197,253,111,208 ; vmovdqa %ymm0,%ymm2
- DB 197,253,111,216 ; vmovdqa %ymm0,%ymm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_set_rgb_hsw_lowp
-_sk_set_rgb_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,250,16,21,173,33,0,0 ; vmovss 0x21ad(%rip),%xmm2 # 2374 <_sk_xor__hsw_lowp+0x7c>
- DB 197,234,88,0 ; vaddss (%rax),%xmm2,%xmm0
- DB 196,193,121,126,193 ; vmovd %xmm0,%r9d
- DB 196,193,121,110,193 ; vmovd %r9d,%xmm0
- DB 196,226,125,121,192 ; vpbroadcastw %xmm0,%ymm0
- DB 197,234,88,72,4 ; vaddss 0x4(%rax),%xmm2,%xmm1
- DB 196,193,121,126,201 ; vmovd %xmm1,%r9d
- DB 196,193,121,110,201 ; vmovd %r9d,%xmm1
- DB 196,226,125,121,201 ; vpbroadcastw %xmm1,%ymm1
- DB 197,234,88,80,8 ; vaddss 0x8(%rax),%xmm2,%xmm2
- DB 197,249,126,208 ; vmovd %xmm2,%eax
- DB 197,249,110,208 ; vmovd %eax,%xmm2
- DB 196,226,125,121,210 ; vpbroadcastw %xmm2,%ymm2
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_premul_hsw_lowp
-_sk_premul_hsw_lowp LABEL PROC
- DB 196,226,125,11,195 ; vpmulhrsw %ymm3,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,226,117,11,203 ; vpmulhrsw %ymm3,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,226,109,11,211 ; vpmulhrsw %ymm3,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_load_8888_hsw_lowp
-_sk_load_8888_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,211,0,0,0 ; jne 31a <_sk_load_8888_hsw_lowp+0xf4>
- DB 196,193,126,111,92,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%ymm3
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 197,253,111,5,68,33,0,0 ; vmovdqa 0x2144(%rip),%ymm0 # 23a0 <_sk_xor__hsw_lowp+0xa8>
- DB 196,226,61,0,200 ; vpshufb %ymm0,%ymm8,%ymm1
- DB 196,227,253,0,201,232 ; vpermq $0xe8,%ymm1,%ymm1
- DB 196,226,101,0,192 ; vpshufb %ymm0,%ymm3,%ymm0
- DB 196,227,253,0,192,232 ; vpermq $0xe8,%ymm0,%ymm0
- DB 196,227,117,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,98,125,121,13,58,33,0,0 ; vpbroadcastw 0x213a(%rip),%ymm9 # 23c0 <_sk_xor__hsw_lowp+0xc8>
- DB 196,193,125,228,193 ; vpmulhuw %ymm9,%ymm0,%ymm0
- DB 197,253,111,13,77,33,0,0 ; vmovdqa 0x214d(%rip),%ymm1 # 23e0 <_sk_xor__hsw_lowp+0xe8>
- DB 196,226,61,0,209 ; vpshufb %ymm1,%ymm8,%ymm2
- DB 196,227,253,0,210,232 ; vpermq $0xe8,%ymm2,%ymm2
- DB 196,226,101,0,201 ; vpshufb %ymm1,%ymm3,%ymm1
- DB 196,227,253,0,201,232 ; vpermq $0xe8,%ymm1,%ymm1
- DB 196,227,109,56,201,1 ; vinserti128 $0x1,%xmm1,%ymm2,%ymm1
- DB 197,245,113,241,8 ; vpsllw $0x8,%ymm1,%ymm1
- DB 196,193,117,228,201 ; vpmulhuw %ymm9,%ymm1,%ymm1
- DB 197,253,111,21,63,33,0,0 ; vmovdqa 0x213f(%rip),%ymm2 # 2400 <_sk_xor__hsw_lowp+0x108>
- DB 196,98,61,0,210 ; vpshufb %ymm2,%ymm8,%ymm10
- DB 196,67,253,0,210,232 ; vpermq $0xe8,%ymm10,%ymm10
- DB 196,226,101,0,210 ; vpshufb %ymm2,%ymm3,%ymm2
- DB 196,227,253,0,210,232 ; vpermq $0xe8,%ymm2,%ymm2
- DB 196,227,45,56,210,1 ; vinserti128 $0x1,%xmm2,%ymm10,%ymm2
- DB 197,237,113,242,8 ; vpsllw $0x8,%ymm2,%ymm2
- DB 196,193,109,228,209 ; vpmulhuw %ymm9,%ymm2,%ymm2
- DB 197,125,111,21,49,33,0,0 ; vmovdqa 0x2131(%rip),%ymm10 # 2420 <_sk_xor__hsw_lowp+0x128>
- DB 196,66,61,0,194 ; vpshufb %ymm10,%ymm8,%ymm8
- DB 196,67,253,0,192,232 ; vpermq $0xe8,%ymm8,%ymm8
- DB 196,194,101,0,218 ; vpshufb %ymm10,%ymm3,%ymm3
- DB 196,227,253,0,219,232 ; vpermq $0xe8,%ymm3,%ymm3
- DB 196,227,61,56,219,1 ; vinserti128 $0x1,%xmm3,%ymm8,%ymm3
- DB 197,229,113,243,8 ; vpsllw $0x8,%ymm3,%ymm3
- DB 196,193,101,228,217 ; vpmulhuw %ymm9,%ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,65,61,239,192 ; vpxor %ymm8,%ymm8,%ymm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,29,255,255,255 ; ja 254 <_sk_load_8888_hsw_lowp+0x2e>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,30,1,0,0 ; lea 0x11e(%rip),%r11 # 460 <_sk_load_8888_hsw_lowp+0x23a>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,110,4,153 ; vmovd (%r9,%rbx,4),%xmm8
- DB 233,254,254,255,255 ; jmpq 254 <_sk_load_8888_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,8 ; vmovd 0x8(%r9,%rbx,4),%xmm0
- DB 196,226,121,89,192 ; vpbroadcastq %xmm0,%xmm0
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,99,101,2,192,4 ; vpblendd $0x4,%ymm0,%ymm3,%ymm8
- DB 196,194,121,53,4,153 ; vpmovzxdq (%r9,%rbx,4),%xmm0
- DB 197,249,112,192,232 ; vpshufd $0xe8,%xmm0,%xmm0
- DB 196,99,61,2,192,3 ; vpblendd $0x3,%ymm0,%ymm8,%ymm8
- DB 233,210,254,255,255 ; jmpq 254 <_sk_load_8888_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,24 ; vmovd 0x18(%r9,%rbx,4),%xmm0
- DB 196,226,125,89,192 ; vpbroadcastq %xmm0,%ymm0
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,99,101,2,192,64 ; vpblendd $0x40,%ymm0,%ymm3,%ymm8
- DB 196,99,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm0
- DB 196,195,121,34,68,153,20,1 ; vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,99,61,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- DB 196,99,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm0
- DB 196,195,121,34,68,153,16,0 ; vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,99,61,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- DB 196,193,122,111,4,153 ; vmovdqu (%r9,%rbx,4),%xmm0
- DB 196,67,125,2,192,240 ; vpblendd $0xf0,%ymm8,%ymm0,%ymm8
- DB 233,131,254,255,255 ; jmpq 254 <_sk_load_8888_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,40 ; vmovd 0x28(%r9,%rbx,4),%xmm0
- DB 196,226,121,89,192 ; vpbroadcastq %xmm0,%xmm0
- DB 197,245,239,201 ; vpxor %ymm1,%ymm1,%ymm1
- DB 196,227,117,2,216,4 ; vpblendd $0x4,%ymm0,%ymm1,%ymm3
- DB 196,195,97,34,68,153,36,1 ; vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm3,%xmm0
- DB 196,227,101,2,216,15 ; vpblendd $0xf,%ymm0,%ymm3,%ymm3
- DB 196,193,121,110,68,153,32 ; vmovd 0x20(%r9,%rbx,4),%xmm0
- DB 196,227,101,2,216,1 ; vpblendd $0x1,%ymm0,%ymm3,%ymm3
- DB 233,71,254,255,255 ; jmpq 24e <_sk_load_8888_hsw_lowp+0x28>
- DB 196,193,121,110,68,153,56 ; vmovd 0x38(%r9,%rbx,4),%xmm0
- DB 196,226,125,89,192 ; vpbroadcastq %xmm0,%ymm0
- DB 197,245,239,201 ; vpxor %ymm1,%ymm1,%ymm1
- DB 196,227,117,2,216,64 ; vpblendd $0x40,%ymm0,%ymm1,%ymm3
- DB 196,227,125,57,216,1 ; vextracti128 $0x1,%ymm3,%xmm0
- DB 196,195,121,34,68,153,52,1 ; vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,227,101,56,216,1 ; vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- DB 196,227,125,57,216,1 ; vextracti128 $0x1,%ymm3,%xmm0
- DB 196,195,121,34,68,153,48,0 ; vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,227,101,56,216,1 ; vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 196,193,122,111,68,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%xmm0
- DB 196,227,125,2,219,240 ; vpblendd $0xf0,%ymm3,%ymm0,%ymm3
- DB 233,247,253,255,255 ; jmpq 254 <_sk_load_8888_hsw_lowp+0x2e>
- DB 15,31,0 ; nopl (%rax)
- DB 235,254 ; jmp 460 <_sk_load_8888_hsw_lowp+0x23a>
- DB 255 ; (bad)
- DB 255,12,255 ; decl (%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255,246 ; push %rsi
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,96,255 ; jmpq *-0x1(%rax)
- DB 255 ; (bad)
- DB 255,76,255,255 ; decl -0x1(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 56,255 ; cmp %bh,%bh
- DB 255 ; (bad)
- DB 255,34 ; jmpq *(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 253 ; std
- DB 255 ; (bad)
- DB 255,149,255,255,255,135 ; callq *-0x78000001(%rbp)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,113,255 ; pushq -0x1(%rcx)
- DB 255 ; (bad)
- DB 255,229 ; jmpq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,209 ; callq *%rcx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 189,255,255,255,167 ; mov $0xa7ffffff,%ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_8888_dst_hsw_lowp
-_sk_load_8888_dst_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,211,0,0,0 ; jne 590 <_sk_load_8888_dst_hsw_lowp+0xf4>
- DB 196,193,126,111,124,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 197,253,111,37,110,31,0,0 ; vmovdqa 0x1f6e(%rip),%ymm4 # 2440 <_sk_xor__hsw_lowp+0x148>
- DB 196,226,61,0,236 ; vpshufb %ymm4,%ymm8,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,226,69,0,228 ; vpshufb %ymm4,%ymm7,%ymm4
- DB 196,227,253,0,228,232 ; vpermq $0xe8,%ymm4,%ymm4
- DB 196,227,85,56,228,1 ; vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,98,125,121,13,100,31,0,0 ; vpbroadcastw 0x1f64(%rip),%ymm9 # 2460 <_sk_xor__hsw_lowp+0x168>
- DB 196,193,93,228,225 ; vpmulhuw %ymm9,%ymm4,%ymm4
- DB 197,253,111,45,119,31,0,0 ; vmovdqa 0x1f77(%rip),%ymm5 # 2480 <_sk_xor__hsw_lowp+0x188>
- DB 196,226,61,0,245 ; vpshufb %ymm5,%ymm8,%ymm6
- DB 196,227,253,0,246,232 ; vpermq $0xe8,%ymm6,%ymm6
- DB 196,226,69,0,237 ; vpshufb %ymm5,%ymm7,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,227,77,56,237,1 ; vinserti128 $0x1,%xmm5,%ymm6,%ymm5
- DB 197,213,113,245,8 ; vpsllw $0x8,%ymm5,%ymm5
- DB 196,193,85,228,233 ; vpmulhuw %ymm9,%ymm5,%ymm5
- DB 197,253,111,53,105,31,0,0 ; vmovdqa 0x1f69(%rip),%ymm6 # 24a0 <_sk_xor__hsw_lowp+0x1a8>
- DB 196,98,61,0,214 ; vpshufb %ymm6,%ymm8,%ymm10
- DB 196,67,253,0,210,232 ; vpermq $0xe8,%ymm10,%ymm10
- DB 196,226,69,0,246 ; vpshufb %ymm6,%ymm7,%ymm6
- DB 196,227,253,0,246,232 ; vpermq $0xe8,%ymm6,%ymm6
- DB 196,227,45,56,246,1 ; vinserti128 $0x1,%xmm6,%ymm10,%ymm6
- DB 197,205,113,246,8 ; vpsllw $0x8,%ymm6,%ymm6
- DB 196,193,77,228,241 ; vpmulhuw %ymm9,%ymm6,%ymm6
- DB 197,125,111,21,91,31,0,0 ; vmovdqa 0x1f5b(%rip),%ymm10 # 24c0 <_sk_xor__hsw_lowp+0x1c8>
- DB 196,66,61,0,194 ; vpshufb %ymm10,%ymm8,%ymm8
- DB 196,67,253,0,192,232 ; vpermq $0xe8,%ymm8,%ymm8
- DB 196,194,69,0,250 ; vpshufb %ymm10,%ymm7,%ymm7
- DB 196,227,253,0,255,232 ; vpermq $0xe8,%ymm7,%ymm7
- DB 196,227,61,56,255,1 ; vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- DB 197,197,113,247,8 ; vpsllw $0x8,%ymm7,%ymm7
- DB 196,193,69,228,249 ; vpmulhuw %ymm9,%ymm7,%ymm7
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,65,61,239,192 ; vpxor %ymm8,%ymm8,%ymm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,29,255,255,255 ; ja 4ca <_sk_load_8888_dst_hsw_lowp+0x2e>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,28,1,0,0 ; lea 0x11c(%rip),%r11 # 6d4 <_sk_load_8888_dst_hsw_lowp+0x238>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,110,4,153 ; vmovd (%r9,%rbx,4),%xmm8
- DB 233,254,254,255,255 ; jmpq 4ca <_sk_load_8888_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,8 ; vmovd 0x8(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,4 ; vpblendd $0x4,%ymm4,%ymm7,%ymm8
- DB 196,194,121,53,36,153 ; vpmovzxdq (%r9,%rbx,4),%xmm4
- DB 197,249,112,228,232 ; vpshufd $0xe8,%xmm4,%xmm4
- DB 196,99,61,2,196,3 ; vpblendd $0x3,%ymm4,%ymm8,%ymm8
- DB 233,210,254,255,255 ; jmpq 4ca <_sk_load_8888_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,24 ; vmovd 0x18(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,64 ; vpblendd $0x40,%ymm4,%ymm7,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,20,1 ; vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,16,0 ; vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,193,122,111,36,153 ; vmovdqu (%r9,%rbx,4),%xmm4
- DB 196,67,93,2,192,240 ; vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- DB 233,131,254,255,255 ; jmpq 4ca <_sk_load_8888_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,40 ; vmovd 0x28(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,4 ; vpblendd $0x4,%ymm4,%ymm5,%ymm7
- DB 196,195,65,34,100,153,36,1 ; vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- DB 196,227,69,2,252,15 ; vpblendd $0xf,%ymm4,%ymm7,%ymm7
- DB 196,193,121,110,100,153,32 ; vmovd 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,69,2,252,1 ; vpblendd $0x1,%ymm4,%ymm7,%ymm7
- DB 233,71,254,255,255 ; jmpq 4c4 <_sk_load_8888_dst_hsw_lowp+0x28>
- DB 196,193,121,110,100,153,56 ; vmovd 0x38(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,64 ; vpblendd $0x40,%ymm4,%ymm5,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,52,1 ; vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,48,0 ; vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 196,193,122,111,100,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,93,2,255,240 ; vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- DB 233,247,253,255,255 ; jmpq 4ca <_sk_load_8888_dst_hsw_lowp+0x2e>
- DB 144 ; nop
- DB 237 ; in (%dx),%eax
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,14 ; decl (%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 248 ; clc
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,98,255 ; jmpq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,78,255 ; decl -0x1(%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 58,255 ; cmp %bh,%bh
- DB 255 ; (bad)
- DB 255,36,255 ; jmpq *(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255,240 ; push %rax
- DB 253 ; std
- DB 255 ; (bad)
- DB 255,151,255,255,255,137 ; callq *-0x76000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,115,255 ; pushq -0x1(%rbx)
- DB 255 ; (bad)
- DB 255,231 ; jmpq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,169 ; mov $0xa9ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_8888_hsw_lowp
-_sk_store_8888_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 197,189,113,208,7 ; vpsrlw $0x7,%ymm0,%ymm8
- DB 196,98,125,121,13,170,29,0,0 ; vpbroadcastw 0x1daa(%rip),%ymm9 # 24e0 <_sk_xor__hsw_lowp+0x1e8>
- DB 196,65,61,234,193 ; vpminsw %ymm9,%ymm8,%ymm8
- DB 196,66,125,51,208 ; vpmovzxwd %xmm8,%ymm10
- DB 196,67,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm8
- DB 196,66,125,51,192 ; vpmovzxwd %xmm8,%ymm8
- DB 197,165,113,209,7 ; vpsrlw $0x7,%ymm1,%ymm11
- DB 196,65,37,234,217 ; vpminsw %ymm9,%ymm11,%ymm11
- DB 196,67,125,57,220,1 ; vextracti128 $0x1,%ymm11,%xmm12
- DB 196,66,125,51,228 ; vpmovzxwd %xmm12,%ymm12
- DB 196,66,125,51,219 ; vpmovzxwd %xmm11,%ymm11
- DB 196,193,37,114,243,8 ; vpslld $0x8,%ymm11,%ymm11
- DB 196,193,29,114,244,8 ; vpslld $0x8,%ymm12,%ymm12
- DB 196,65,29,235,192 ; vpor %ymm8,%ymm12,%ymm8
- DB 196,65,37,235,210 ; vpor %ymm10,%ymm11,%ymm10
- DB 197,165,113,210,7 ; vpsrlw $0x7,%ymm2,%ymm11
- DB 196,65,37,234,217 ; vpminsw %ymm9,%ymm11,%ymm11
- DB 196,66,125,51,227 ; vpmovzxwd %xmm11,%ymm12
- DB 196,67,125,57,219,1 ; vextracti128 $0x1,%ymm11,%xmm11
- DB 196,66,125,51,219 ; vpmovzxwd %xmm11,%ymm11
- DB 196,193,37,114,243,16 ; vpslld $0x10,%ymm11,%ymm11
- DB 196,193,29,114,244,16 ; vpslld $0x10,%ymm12,%ymm12
- DB 197,149,113,211,7 ; vpsrlw $0x7,%ymm3,%ymm13
- DB 196,65,21,234,201 ; vpminsw %ymm9,%ymm13,%ymm9
- DB 196,67,125,57,205,1 ; vextracti128 $0x1,%ymm9,%xmm13
- DB 196,66,125,51,237 ; vpmovzxwd %xmm13,%ymm13
- DB 196,66,125,51,201 ; vpmovzxwd %xmm9,%ymm9
- DB 196,193,13,114,241,24 ; vpslld $0x18,%ymm9,%ymm14
- DB 196,193,53,114,245,24 ; vpslld $0x18,%ymm13,%ymm9
- DB 196,65,37,235,201 ; vpor %ymm9,%ymm11,%ymm9
- DB 196,65,61,235,201 ; vpor %ymm9,%ymm8,%ymm9
- DB 196,65,29,235,198 ; vpor %ymm14,%ymm12,%ymm8
- DB 196,65,45,235,192 ; vpor %ymm8,%ymm10,%ymm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,18 ; jne 7f2 <_sk_store_8888_hsw_lowp+0xe2>
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,126,127,76,153,32 ; vmovdqu %ymm9,0x20(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,235 ; ja 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,175,0,0,0 ; lea 0xaf(%rip),%r11 # 8bc <_sk_store_8888_hsw_lowp+0x1ac>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,126,4,153 ; vmovd %xmm8,(%r9,%rbx,4)
- DB 235,207 ; jmp 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 196,67,121,22,68,153,8,2 ; vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- DB 196,65,121,214,4,153 ; vmovq %xmm8,(%r9,%rbx,4)
- DB 235,191 ; jmp 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,24,2 ; vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,20,1 ; vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,65,121,126,76,153,16 ; vmovd %xmm9,0x10(%r9,%rbx,4)
- DB 196,65,122,127,4,153 ; vmovdqu %xmm8,(%r9,%rbx,4)
- DB 235,142 ; jmp 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 196,67,121,22,76,153,40,2 ; vpextrd $0x2,%xmm9,0x28(%r9,%rbx,4)
- DB 196,67,121,22,76,153,36,1 ; vpextrd $0x1,%xmm9,0x24(%r9,%rbx,4)
- DB 196,65,121,126,76,153,32 ; vmovd %xmm9,0x20(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 233,108,255,255,255 ; jmpq 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 196,67,125,57,202,1 ; vextracti128 $0x1,%ymm9,%xmm10
- DB 196,67,121,22,84,153,56,2 ; vpextrd $0x2,%xmm10,0x38(%r9,%rbx,4)
- DB 196,67,125,57,202,1 ; vextracti128 $0x1,%ymm9,%xmm10
- DB 196,67,121,22,84,153,52,1 ; vpextrd $0x1,%xmm10,0x34(%r9,%rbx,4)
- DB 196,67,125,57,202,1 ; vextracti128 $0x1,%ymm9,%xmm10
- DB 196,65,121,126,84,153,48 ; vmovd %xmm10,0x30(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,122,127,76,153,32 ; vmovdqu %xmm9,0x20(%r9,%rbx,4)
- DB 233,49,255,255,255 ; jmpq 7ed <_sk_store_8888_hsw_lowp+0xdd>
- DB 90 ; pop %rdx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,106,255 ; ljmp *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,98,255 ; jmpq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,155,255,255,255,142 ; lcall *-0x71000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,128,255,255,255,114 ; incl 0x72ffffff(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 186,255,255,255,179 ; mov $0xb3ffffff,%edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,171,255,255,255,163 ; ljmp *-0x5c000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,225 ; jmpq *%rcx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,197 ; inc %ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_bgra_hsw_lowp
-_sk_load_bgra_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,211,0,0,0 ; jne 9ec <_sk_load_bgra_hsw_lowp+0xf4>
- DB 196,193,126,111,92,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%ymm3
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 197,253,111,5,210,27,0,0 ; vmovdqa 0x1bd2(%rip),%ymm0 # 2500 <_sk_xor__hsw_lowp+0x208>
- DB 196,226,61,0,200 ; vpshufb %ymm0,%ymm8,%ymm1
- DB 196,227,253,0,201,232 ; vpermq $0xe8,%ymm1,%ymm1
- DB 196,226,101,0,192 ; vpshufb %ymm0,%ymm3,%ymm0
- DB 196,227,253,0,192,232 ; vpermq $0xe8,%ymm0,%ymm0
- DB 196,227,117,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,98,125,121,13,200,27,0,0 ; vpbroadcastw 0x1bc8(%rip),%ymm9 # 2520 <_sk_xor__hsw_lowp+0x228>
- DB 196,193,125,228,209 ; vpmulhuw %ymm9,%ymm0,%ymm2
- DB 197,253,111,5,219,27,0,0 ; vmovdqa 0x1bdb(%rip),%ymm0 # 2540 <_sk_xor__hsw_lowp+0x248>
- DB 196,226,61,0,200 ; vpshufb %ymm0,%ymm8,%ymm1
- DB 196,227,253,0,201,232 ; vpermq $0xe8,%ymm1,%ymm1
- DB 196,226,101,0,192 ; vpshufb %ymm0,%ymm3,%ymm0
- DB 196,227,253,0,192,232 ; vpermq $0xe8,%ymm0,%ymm0
- DB 196,227,117,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm1,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,193,125,228,201 ; vpmulhuw %ymm9,%ymm0,%ymm1
- DB 197,253,111,5,205,27,0,0 ; vmovdqa 0x1bcd(%rip),%ymm0 # 2560 <_sk_xor__hsw_lowp+0x268>
- DB 196,98,61,0,208 ; vpshufb %ymm0,%ymm8,%ymm10
- DB 196,67,253,0,210,232 ; vpermq $0xe8,%ymm10,%ymm10
- DB 196,226,101,0,192 ; vpshufb %ymm0,%ymm3,%ymm0
- DB 196,227,253,0,192,232 ; vpermq $0xe8,%ymm0,%ymm0
- DB 196,227,45,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm10,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,193,125,228,193 ; vpmulhuw %ymm9,%ymm0,%ymm0
- DB 197,125,111,21,191,27,0,0 ; vmovdqa 0x1bbf(%rip),%ymm10 # 2580 <_sk_xor__hsw_lowp+0x288>
- DB 196,66,61,0,194 ; vpshufb %ymm10,%ymm8,%ymm8
- DB 196,67,253,0,192,232 ; vpermq $0xe8,%ymm8,%ymm8
- DB 196,194,101,0,218 ; vpshufb %ymm10,%ymm3,%ymm3
- DB 196,227,253,0,219,232 ; vpermq $0xe8,%ymm3,%ymm3
- DB 196,227,61,56,219,1 ; vinserti128 $0x1,%xmm3,%ymm8,%ymm3
- DB 197,229,113,243,8 ; vpsllw $0x8,%ymm3,%ymm3
- DB 196,193,101,228,217 ; vpmulhuw %ymm9,%ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,65,61,239,192 ; vpxor %ymm8,%ymm8,%ymm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,29,255,255,255 ; ja 926 <_sk_load_bgra_hsw_lowp+0x2e>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,28,1,0,0 ; lea 0x11c(%rip),%r11 # b30 <_sk_load_bgra_hsw_lowp+0x238>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,110,4,153 ; vmovd (%r9,%rbx,4),%xmm8
- DB 233,254,254,255,255 ; jmpq 926 <_sk_load_bgra_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,8 ; vmovd 0x8(%r9,%rbx,4),%xmm0
- DB 196,226,121,89,192 ; vpbroadcastq %xmm0,%xmm0
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,99,101,2,192,4 ; vpblendd $0x4,%ymm0,%ymm3,%ymm8
- DB 196,194,121,53,4,153 ; vpmovzxdq (%r9,%rbx,4),%xmm0
- DB 197,249,112,192,232 ; vpshufd $0xe8,%xmm0,%xmm0
- DB 196,99,61,2,192,3 ; vpblendd $0x3,%ymm0,%ymm8,%ymm8
- DB 233,210,254,255,255 ; jmpq 926 <_sk_load_bgra_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,24 ; vmovd 0x18(%r9,%rbx,4),%xmm0
- DB 196,226,125,89,192 ; vpbroadcastq %xmm0,%ymm0
- DB 197,229,239,219 ; vpxor %ymm3,%ymm3,%ymm3
- DB 196,99,101,2,192,64 ; vpblendd $0x40,%ymm0,%ymm3,%ymm8
- DB 196,99,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm0
- DB 196,195,121,34,68,153,20,1 ; vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,99,61,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- DB 196,99,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm0
- DB 196,195,121,34,68,153,16,0 ; vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,99,61,56,192,1 ; vinserti128 $0x1,%xmm0,%ymm8,%ymm8
- DB 196,193,122,111,4,153 ; vmovdqu (%r9,%rbx,4),%xmm0
- DB 196,67,125,2,192,240 ; vpblendd $0xf0,%ymm8,%ymm0,%ymm8
- DB 233,131,254,255,255 ; jmpq 926 <_sk_load_bgra_hsw_lowp+0x2e>
- DB 196,193,121,110,68,153,40 ; vmovd 0x28(%r9,%rbx,4),%xmm0
- DB 196,226,121,89,192 ; vpbroadcastq %xmm0,%xmm0
- DB 197,245,239,201 ; vpxor %ymm1,%ymm1,%ymm1
- DB 196,227,117,2,216,4 ; vpblendd $0x4,%ymm0,%ymm1,%ymm3
- DB 196,195,97,34,68,153,36,1 ; vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm3,%xmm0
- DB 196,227,101,2,216,15 ; vpblendd $0xf,%ymm0,%ymm3,%ymm3
- DB 196,193,121,110,68,153,32 ; vmovd 0x20(%r9,%rbx,4),%xmm0
- DB 196,227,101,2,216,1 ; vpblendd $0x1,%ymm0,%ymm3,%ymm3
- DB 233,71,254,255,255 ; jmpq 920 <_sk_load_bgra_hsw_lowp+0x28>
- DB 196,193,121,110,68,153,56 ; vmovd 0x38(%r9,%rbx,4),%xmm0
- DB 196,226,125,89,192 ; vpbroadcastq %xmm0,%ymm0
- DB 197,245,239,201 ; vpxor %ymm1,%ymm1,%ymm1
- DB 196,227,117,2,216,64 ; vpblendd $0x40,%ymm0,%ymm1,%ymm3
- DB 196,227,125,57,216,1 ; vextracti128 $0x1,%ymm3,%xmm0
- DB 196,195,121,34,68,153,52,1 ; vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,227,101,56,216,1 ; vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- DB 196,227,125,57,216,1 ; vextracti128 $0x1,%ymm3,%xmm0
- DB 196,195,121,34,68,153,48,0 ; vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm0,%xmm0
- DB 196,227,101,56,216,1 ; vinserti128 $0x1,%xmm0,%ymm3,%ymm3
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 196,193,122,111,68,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%xmm0
- DB 196,227,125,2,219,240 ; vpblendd $0xf0,%ymm3,%ymm0,%ymm3
- DB 233,247,253,255,255 ; jmpq 926 <_sk_load_bgra_hsw_lowp+0x2e>
- DB 144 ; nop
- DB 237 ; in (%dx),%eax
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,14 ; decl (%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 248 ; clc
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,98,255 ; jmpq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,78,255 ; decl -0x1(%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 58,255 ; cmp %bh,%bh
- DB 255 ; (bad)
- DB 255,36,255 ; jmpq *(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255,240 ; push %rax
- DB 253 ; std
- DB 255 ; (bad)
- DB 255,151,255,255,255,137 ; callq *-0x76000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,115,255 ; pushq -0x1(%rbx)
- DB 255 ; (bad)
- DB 255,231 ; jmpq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,169 ; mov $0xa9ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_bgra_dst_hsw_lowp
-_sk_load_bgra_dst_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,211,0,0,0 ; jne c60 <_sk_load_bgra_dst_hsw_lowp+0xf4>
- DB 196,193,126,111,124,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 197,253,111,37,254,25,0,0 ; vmovdqa 0x19fe(%rip),%ymm4 # 25a0 <_sk_xor__hsw_lowp+0x2a8>
- DB 196,226,61,0,236 ; vpshufb %ymm4,%ymm8,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,226,69,0,228 ; vpshufb %ymm4,%ymm7,%ymm4
- DB 196,227,253,0,228,232 ; vpermq $0xe8,%ymm4,%ymm4
- DB 196,227,85,56,228,1 ; vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,98,125,121,13,244,25,0,0 ; vpbroadcastw 0x19f4(%rip),%ymm9 # 25c0 <_sk_xor__hsw_lowp+0x2c8>
- DB 196,193,93,228,241 ; vpmulhuw %ymm9,%ymm4,%ymm6
- DB 197,253,111,37,7,26,0,0 ; vmovdqa 0x1a07(%rip),%ymm4 # 25e0 <_sk_xor__hsw_lowp+0x2e8>
- DB 196,226,61,0,236 ; vpshufb %ymm4,%ymm8,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,226,69,0,228 ; vpshufb %ymm4,%ymm7,%ymm4
- DB 196,227,253,0,228,232 ; vpermq $0xe8,%ymm4,%ymm4
- DB 196,227,85,56,228,1 ; vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,193,93,228,233 ; vpmulhuw %ymm9,%ymm4,%ymm5
- DB 197,253,111,37,249,25,0,0 ; vmovdqa 0x19f9(%rip),%ymm4 # 2600 <_sk_xor__hsw_lowp+0x308>
- DB 196,98,61,0,212 ; vpshufb %ymm4,%ymm8,%ymm10
- DB 196,67,253,0,210,232 ; vpermq $0xe8,%ymm10,%ymm10
- DB 196,226,69,0,228 ; vpshufb %ymm4,%ymm7,%ymm4
- DB 196,227,253,0,228,232 ; vpermq $0xe8,%ymm4,%ymm4
- DB 196,227,45,56,228,1 ; vinserti128 $0x1,%xmm4,%ymm10,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,193,93,228,225 ; vpmulhuw %ymm9,%ymm4,%ymm4
- DB 197,125,111,21,235,25,0,0 ; vmovdqa 0x19eb(%rip),%ymm10 # 2620 <_sk_xor__hsw_lowp+0x328>
- DB 196,66,61,0,194 ; vpshufb %ymm10,%ymm8,%ymm8
- DB 196,67,253,0,192,232 ; vpermq $0xe8,%ymm8,%ymm8
- DB 196,194,69,0,250 ; vpshufb %ymm10,%ymm7,%ymm7
- DB 196,227,253,0,255,232 ; vpermq $0xe8,%ymm7,%ymm7
- DB 196,227,61,56,255,1 ; vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- DB 197,197,113,247,8 ; vpsllw $0x8,%ymm7,%ymm7
- DB 196,193,69,228,249 ; vpmulhuw %ymm9,%ymm7,%ymm7
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,65,61,239,192 ; vpxor %ymm8,%ymm8,%ymm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,29,255,255,255 ; ja b9a <_sk_load_bgra_dst_hsw_lowp+0x2e>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,28,1,0,0 ; lea 0x11c(%rip),%r11 # da4 <_sk_load_bgra_dst_hsw_lowp+0x238>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,110,4,153 ; vmovd (%r9,%rbx,4),%xmm8
- DB 233,254,254,255,255 ; jmpq b9a <_sk_load_bgra_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,8 ; vmovd 0x8(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,4 ; vpblendd $0x4,%ymm4,%ymm7,%ymm8
- DB 196,194,121,53,36,153 ; vpmovzxdq (%r9,%rbx,4),%xmm4
- DB 197,249,112,228,232 ; vpshufd $0xe8,%xmm4,%xmm4
- DB 196,99,61,2,196,3 ; vpblendd $0x3,%ymm4,%ymm8,%ymm8
- DB 233,210,254,255,255 ; jmpq b9a <_sk_load_bgra_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,24 ; vmovd 0x18(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,64 ; vpblendd $0x40,%ymm4,%ymm7,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,20,1 ; vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,16,0 ; vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,193,122,111,36,153 ; vmovdqu (%r9,%rbx,4),%xmm4
- DB 196,67,93,2,192,240 ; vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- DB 233,131,254,255,255 ; jmpq b9a <_sk_load_bgra_dst_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,40 ; vmovd 0x28(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,4 ; vpblendd $0x4,%ymm4,%ymm5,%ymm7
- DB 196,195,65,34,100,153,36,1 ; vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- DB 196,227,69,2,252,15 ; vpblendd $0xf,%ymm4,%ymm7,%ymm7
- DB 196,193,121,110,100,153,32 ; vmovd 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,69,2,252,1 ; vpblendd $0x1,%ymm4,%ymm7,%ymm7
- DB 233,71,254,255,255 ; jmpq b94 <_sk_load_bgra_dst_hsw_lowp+0x28>
- DB 196,193,121,110,100,153,56 ; vmovd 0x38(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,64 ; vpblendd $0x40,%ymm4,%ymm5,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,52,1 ; vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,48,0 ; vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 196,193,122,111,100,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,93,2,255,240 ; vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- DB 233,247,253,255,255 ; jmpq b9a <_sk_load_bgra_dst_hsw_lowp+0x2e>
- DB 144 ; nop
- DB 237 ; in (%dx),%eax
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,14 ; decl (%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 248 ; clc
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,98,255 ; jmpq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,78,255 ; decl -0x1(%rsi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 58,255 ; cmp %bh,%bh
- DB 255 ; (bad)
- DB 255,36,255 ; jmpq *(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255,240 ; push %rax
- DB 253 ; std
- DB 255 ; (bad)
- DB 255,151,255,255,255,137 ; callq *-0x76000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,115,255 ; pushq -0x1(%rbx)
- DB 255 ; (bad)
- DB 255,231 ; jmpq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,169 ; mov $0xa9ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_bgra_hsw_lowp
-_sk_store_bgra_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 197,189,113,210,7 ; vpsrlw $0x7,%ymm2,%ymm8
- DB 196,98,125,121,13,58,24,0,0 ; vpbroadcastw 0x183a(%rip),%ymm9 # 2640 <_sk_xor__hsw_lowp+0x348>
- DB 196,65,61,234,193 ; vpminsw %ymm9,%ymm8,%ymm8
- DB 196,66,125,51,208 ; vpmovzxwd %xmm8,%ymm10
- DB 196,67,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm8
- DB 196,66,125,51,192 ; vpmovzxwd %xmm8,%ymm8
- DB 197,165,113,209,7 ; vpsrlw $0x7,%ymm1,%ymm11
- DB 196,65,37,234,217 ; vpminsw %ymm9,%ymm11,%ymm11
- DB 196,66,125,51,227 ; vpmovzxwd %xmm11,%ymm12
- DB 196,67,125,57,219,1 ; vextracti128 $0x1,%ymm11,%xmm11
- DB 196,66,125,51,219 ; vpmovzxwd %xmm11,%ymm11
- DB 196,193,37,114,243,8 ; vpslld $0x8,%ymm11,%ymm11
- DB 196,193,29,114,244,8 ; vpslld $0x8,%ymm12,%ymm12
- DB 197,149,113,208,7 ; vpsrlw $0x7,%ymm0,%ymm13
- DB 196,65,21,234,233 ; vpminsw %ymm9,%ymm13,%ymm13
- DB 196,67,125,57,238,1 ; vextracti128 $0x1,%ymm13,%xmm14
- DB 196,66,125,51,246 ; vpmovzxwd %xmm14,%ymm14
- DB 196,66,125,51,237 ; vpmovzxwd %xmm13,%ymm13
- DB 196,193,21,114,245,16 ; vpslld $0x10,%ymm13,%ymm13
- DB 196,193,13,114,246,16 ; vpslld $0x10,%ymm14,%ymm14
- DB 197,133,113,211,7 ; vpsrlw $0x7,%ymm3,%ymm15
- DB 196,65,5,234,201 ; vpminsw %ymm9,%ymm15,%ymm9
- DB 196,67,125,57,207,1 ; vextracti128 $0x1,%ymm9,%xmm15
- DB 196,66,125,51,255 ; vpmovzxwd %xmm15,%ymm15
- DB 196,66,125,51,201 ; vpmovzxwd %xmm9,%ymm9
- DB 196,193,53,114,241,24 ; vpslld $0x18,%ymm9,%ymm9
- DB 196,193,5,114,247,24 ; vpslld $0x18,%ymm15,%ymm15
- DB 196,65,61,235,198 ; vpor %ymm14,%ymm8,%ymm8
- DB 196,65,45,235,237 ; vpor %ymm13,%ymm10,%ymm13
- DB 196,65,37,235,215 ; vpor %ymm15,%ymm11,%ymm10
- DB 196,65,61,235,210 ; vpor %ymm10,%ymm8,%ymm10
- DB 196,65,29,235,193 ; vpor %ymm9,%ymm12,%ymm8
- DB 196,65,21,235,192 ; vpor %ymm8,%ymm13,%ymm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,18 ; jne ec2 <_sk_store_bgra_hsw_lowp+0xe2>
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,126,127,84,153,32 ; vmovdqu %ymm10,0x20(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,235 ; ja ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,175,0,0,0 ; lea 0xaf(%rip),%r11 # f8c <_sk_store_bgra_hsw_lowp+0x1ac>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,126,4,153 ; vmovd %xmm8,(%r9,%rbx,4)
- DB 235,207 ; jmp ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 196,67,121,22,68,153,8,2 ; vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- DB 196,65,121,214,4,153 ; vmovq %xmm8,(%r9,%rbx,4)
- DB 235,191 ; jmp ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,24,2 ; vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,20,1 ; vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,65,121,126,76,153,16 ; vmovd %xmm9,0x10(%r9,%rbx,4)
- DB 196,65,122,127,4,153 ; vmovdqu %xmm8,(%r9,%rbx,4)
- DB 235,142 ; jmp ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 196,67,121,22,84,153,40,2 ; vpextrd $0x2,%xmm10,0x28(%r9,%rbx,4)
- DB 196,67,121,22,84,153,36,1 ; vpextrd $0x1,%xmm10,0x24(%r9,%rbx,4)
- DB 196,65,121,126,84,153,32 ; vmovd %xmm10,0x20(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 233,108,255,255,255 ; jmpq ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,67,121,22,76,153,56,2 ; vpextrd $0x2,%xmm9,0x38(%r9,%rbx,4)
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,67,121,22,76,153,52,1 ; vpextrd $0x1,%xmm9,0x34(%r9,%rbx,4)
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,65,121,126,76,153,48 ; vmovd %xmm9,0x30(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,122,127,84,153,32 ; vmovdqu %xmm10,0x20(%r9,%rbx,4)
- DB 233,49,255,255,255 ; jmpq ebd <_sk_store_bgra_hsw_lowp+0xdd>
- DB 90 ; pop %rdx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,106,255 ; ljmp *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,98,255 ; jmpq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,155,255,255,255,142 ; lcall *-0x71000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,128,255,255,255,114 ; incl 0x72ffffff(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 186,255,255,255,179 ; mov $0xb3ffffff,%edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,171,255,255,255,163 ; ljmp *-0x5c000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,225 ; jmpq *%rcx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,197 ; inc %ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_a8_hsw_lowp
-_sk_load_a8_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,46 ; jne 100f <_sk_load_a8_hsw_lowp+0x47>
- DB 196,193,122,111,4,25 ; vmovdqu (%r9,%rbx,1),%xmm0
- DB 196,226,125,48,192 ; vpmovzxbw %xmm0,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,226,125,121,13,72,22,0,0 ; vpbroadcastw 0x1648(%rip),%ymm1 # 2642 <_sk_xor__hsw_lowp+0x34a>
- DB 197,253,228,217 ; vpmulhuw %ymm1,%ymm0,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,253,239,192 ; vpxor %ymm0,%ymm0,%ymm0
- DB 197,245,239,201 ; vpxor %ymm1,%ymm1,%ymm1
- DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,196 ; ja fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,194,0,0,0 ; lea 0xc2(%rip),%r11 # 10f0 <_sk_load_a8_hsw_lowp+0x128>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,249,110,192 ; vmovd %eax,%xmm0
- DB 235,165 ; jmp fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm0,%xmm0
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,249,110,200 ; vmovd %eax,%xmm1
- DB 196,227,121,14,193,1 ; vpblendw $0x1,%xmm1,%xmm0,%xmm0
- DB 235,136 ; jmp fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,121,110,12,25 ; vmovd (%r9,%rbx,1),%xmm1
- DB 196,227,121,2,193,1 ; vpblendd $0x1,%xmm1,%xmm0,%xmm0
- DB 233,91,255,255,255 ; jmpq fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm1
- DB 196,227,113,2,192,12 ; vpblendd $0xc,%xmm0,%xmm1,%xmm0
- DB 233,46,255,255,255 ; jmpq fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm1
- DB 196,195,113,34,76,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm1,%xmm1
- DB 196,227,113,2,192,8 ; vpblendd $0x8,%xmm0,%xmm1,%xmm0
- DB 233,249,254,255,255 ; jmpq fe7 <_sk_load_a8_hsw_lowp+0x1f>
- DB 102,144 ; xchg %ax,%ax
- DB 71,255 ; rex.RXB (bad)
- DB 255 ; (bad)
- DB 255,94,255 ; lcall *-0x1(%rsi)
- DB 255 ; (bad)
- DB 255,82,255 ; callq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,139,255,255,255,131 ; decl -0x7c000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 123,255 ; jnp 1105 <_sk_load_a8_hsw_lowp+0x13d>
- DB 255 ; (bad)
- DB 255,111,255 ; ljmp *-0x1(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 184,255,255,255,176 ; mov $0xb0ffffff,%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,168,255,255,255,156 ; ljmp *-0x63000001(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,229 ; jmpq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 221,255 ; (bad)
- DB 255 ; (bad)
- DB 255,213 ; callq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,201 ; dec %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_a8_dst_hsw_lowp
-_sk_load_a8_dst_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,46 ; jne 1173 <_sk_load_a8_dst_hsw_lowp+0x47>
- DB 196,193,122,111,36,25 ; vmovdqu (%r9,%rbx,1),%xmm4
- DB 196,226,125,48,228 ; vpmovzxbw %xmm4,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,226,125,121,45,230,20,0,0 ; vpbroadcastw 0x14e6(%rip),%ymm5 # 2644 <_sk_xor__hsw_lowp+0x34c>
- DB 197,221,228,253 ; vpmulhuw %ymm5,%ymm4,%ymm7
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,221,239,228 ; vpxor %ymm4,%ymm4,%ymm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,196 ; ja 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,194,0,0,0 ; lea 0xc2(%rip),%r11 # 1254 <_sk_load_a8_dst_hsw_lowp+0x128>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,249,110,224 ; vmovd %eax,%xmm4
- DB 235,165 ; jmp 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm4,%xmm4
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,249,110,232 ; vmovd %eax,%xmm5
- DB 196,227,89,14,229,1 ; vpblendw $0x1,%xmm5,%xmm4,%xmm4
- DB 235,136 ; jmp 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,121,110,44,25 ; vmovd (%r9,%rbx,1),%xmm5
- DB 196,227,89,2,229,1 ; vpblendd $0x1,%xmm5,%xmm4,%xmm4
- DB 233,91,255,255,255 ; jmpq 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,122,126,44,25 ; vmovq (%r9,%rbx,1),%xmm5
- DB 196,227,81,2,228,12 ; vpblendd $0xc,%xmm4,%xmm5,%xmm4
- DB 233,46,255,255,255 ; jmpq 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,122,126,44,25 ; vmovq (%r9,%rbx,1),%xmm5
- DB 196,195,81,34,108,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm5,%xmm5
- DB 196,227,81,2,228,8 ; vpblendd $0x8,%xmm4,%xmm5,%xmm4
- DB 233,249,254,255,255 ; jmpq 114b <_sk_load_a8_dst_hsw_lowp+0x1f>
- DB 102,144 ; xchg %ax,%ax
- DB 71,255 ; rex.RXB (bad)
- DB 255 ; (bad)
- DB 255,94,255 ; lcall *-0x1(%rsi)
- DB 255 ; (bad)
- DB 255,82,255 ; callq *-0x1(%rdx)
- DB 255 ; (bad)
- DB 255,139,255,255,255,131 ; decl -0x7c000001(%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 123,255 ; jnp 1269 <_sk_load_a8_dst_hsw_lowp+0x13d>
- DB 255 ; (bad)
- DB 255,111,255 ; ljmp *-0x1(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 184,255,255,255,176 ; mov $0xb0ffffff,%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,168,255,255,255,156 ; ljmp *-0x63000001(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,229 ; jmpq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 221,255 ; (bad)
- DB 255 ; (bad)
- DB 255,213 ; callq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,201 ; dec %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_a8_hsw_lowp
-_sk_store_a8_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 197,189,113,211,7 ; vpsrlw $0x7,%ymm3,%ymm8
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,65,57,103,193 ; vpackuswb %xmm9,%xmm8,%xmm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,11 ; jne 12c4 <_sk_store_a8_hsw_lowp+0x34>
- DB 196,65,122,127,4,25 ; vmovdqu %xmm8,(%r9,%rbx,1)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,235 ; ja 12bf <_sk_store_a8_hsw_lowp+0x2f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,137,0,0,0 ; lea 0x89(%rip),%r11 # 1368 <_sk_store_a8_hsw_lowp+0xd8>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,67,121,20,4,25,0 ; vpextrb $0x0,%xmm8,(%r9,%rbx,1)
- DB 235,206 ; jmp 12bf <_sk_store_a8_hsw_lowp+0x2f>
- DB 196,67,121,20,68,25,2,2 ; vpextrb $0x2,%xmm8,0x2(%r9,%rbx,1)
- DB 196,67,121,21,4,25,0 ; vpextrw $0x0,%xmm8,(%r9,%rbx,1)
- DB 235,189 ; jmp 12bf <_sk_store_a8_hsw_lowp+0x2f>
- DB 196,67,121,20,68,25,6,6 ; vpextrb $0x6,%xmm8,0x6(%r9,%rbx,1)
- DB 196,67,121,20,68,25,5,5 ; vpextrb $0x5,%xmm8,0x5(%r9,%rbx,1)
- DB 196,67,121,20,68,25,4,4 ; vpextrb $0x4,%xmm8,0x4(%r9,%rbx,1)
- DB 196,65,121,126,4,25 ; vmovd %xmm8,(%r9,%rbx,1)
- DB 235,157 ; jmp 12bf <_sk_store_a8_hsw_lowp+0x2f>
- DB 196,67,121,20,68,25,10,10 ; vpextrb $0xa,%xmm8,0xa(%r9,%rbx,1)
- DB 196,67,121,20,68,25,9,9 ; vpextrb $0x9,%xmm8,0x9(%r9,%rbx,1)
- DB 196,67,121,20,68,25,8,8 ; vpextrb $0x8,%xmm8,0x8(%r9,%rbx,1)
- DB 235,32 ; jmp 135c <_sk_store_a8_hsw_lowp+0xcc>
- DB 196,67,121,20,68,25,14,14 ; vpextrb $0xe,%xmm8,0xe(%r9,%rbx,1)
- DB 196,67,121,20,68,25,13,13 ; vpextrb $0xd,%xmm8,0xd(%r9,%rbx,1)
- DB 196,67,121,20,68,25,12,12 ; vpextrb $0xc,%xmm8,0xc(%r9,%rbx,1)
- DB 196,67,121,22,68,25,8,2 ; vpextrd $0x2,%xmm8,0x8(%r9,%rbx,1)
- DB 196,65,121,214,4,25 ; vmovq %xmm8,(%r9,%rbx,1)
- DB 233,88,255,255,255 ; jmpq 12bf <_sk_store_a8_hsw_lowp+0x2f>
- DB 144 ; nop
- DB 128,255,255 ; cmp $0xff,%bh
- DB 255,145,255,255,255,137 ; callq *-0x76000001(%rcx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,178,255,255,255,170 ; pushq -0x55000001(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,162,255,255,255,154 ; jmpq *-0x65000001(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,244 ; push %rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,202 ; dec %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,194 ; inc %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 186,255,255,255,236 ; mov $0xecffffff,%edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,228 ; jmpq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 220,255 ; fdivr %st,%st(7)
- DB 255 ; (bad)
- DB 255,212 ; callq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_g8_hsw_lowp
-_sk_load_g8_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,51 ; jne 13f0 <_sk_load_g8_hsw_lowp+0x4c>
- DB 196,193,122,111,4,25 ; vmovdqu (%r9,%rbx,1),%xmm0
- DB 196,226,125,48,192 ; vpmovzxbw %xmm0,%ymm0
- DB 197,253,113,240,8 ; vpsllw $0x8,%ymm0,%ymm0
- DB 196,226,125,121,13,112,18,0,0 ; vpbroadcastw 0x1270(%rip),%ymm1 # 2646 <_sk_xor__hsw_lowp+0x34e>
- DB 197,253,228,193 ; vpmulhuw %ymm1,%ymm0,%ymm0
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 196,226,125,121,29,99,18,0,0 ; vpbroadcastw 0x1263(%rip),%ymm3 # 2648 <_sk_xor__hsw_lowp+0x350>
- DB 197,253,111,200 ; vmovdqa %ymm0,%ymm1
- DB 197,253,111,208 ; vmovdqa %ymm0,%ymm2
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,191 ; ja 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,193,0,0,0 ; lea 0xc1(%rip),%r11 # 14d0 <_sk_load_g8_hsw_lowp+0x12c>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,249,110,192 ; vmovd %eax,%xmm0
- DB 235,160 ; jmp 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm0,%xmm0
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,249,110,200 ; vmovd %eax,%xmm1
- DB 196,227,121,14,193,1 ; vpblendw $0x1,%xmm1,%xmm0,%xmm0
- DB 235,131 ; jmp 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,121,110,12,25 ; vmovd (%r9,%rbx,1),%xmm1
- DB 196,227,121,2,193,1 ; vpblendd $0x1,%xmm1,%xmm0,%xmm0
- DB 233,86,255,255,255 ; jmpq 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm1
- DB 196,227,113,2,192,12 ; vpblendd $0xc,%xmm0,%xmm1,%xmm0
- DB 233,41,255,255,255 ; jmpq 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
- DB 196,195,121,32,68,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,195,121,32,68,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm0,%xmm0
- DB 196,193,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm1
- DB 196,195,113,34,76,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm1,%xmm1
- DB 196,227,113,2,192,8 ; vpblendd $0x8,%xmm0,%xmm1,%xmm0
- DB 233,244,254,255,255 ; jmpq 13c3 <_sk_load_g8_hsw_lowp+0x1f>
- DB 144 ; nop
- DB 72,255 ; rex.W (bad)
- DB 255 ; (bad)
- DB 255,95,255 ; lcall *-0x1(%rdi)
- DB 255 ; (bad)
- DB 255,83,255 ; callq *-0x1(%rbx)
- DB 255 ; (bad)
- DB 255,140,255,255,255,132,255 ; decl -0x7b0001(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 124,255 ; jl 14e5 <_sk_load_g8_hsw_lowp+0x141>
- DB 255 ; (bad)
- DB 255,112,255 ; pushq -0x1(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 185,255,255,255,177 ; mov $0xb1ffffff,%ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,169,255,255,255,157 ; ljmp *-0x62000001(%rcx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,230 ; jmpq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 222,255 ; fdivrp %st,%st(7)
- DB 255 ; (bad)
- DB 255,214 ; callq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,202 ; dec %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_g8_dst_hsw_lowp
-_sk_load_g8_dst_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,51 ; jne 1558 <_sk_load_g8_dst_hsw_lowp+0x4c>
- DB 196,193,122,111,36,25 ; vmovdqu (%r9,%rbx,1),%xmm4
- DB 196,226,125,48,228 ; vpmovzxbw %xmm4,%ymm4
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,226,125,121,45,12,17,0,0 ; vpbroadcastw 0x110c(%rip),%ymm5 # 264a <_sk_xor__hsw_lowp+0x352>
- DB 197,221,228,229 ; vpmulhuw %ymm5,%ymm4,%ymm4
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 196,226,125,121,61,255,16,0,0 ; vpbroadcastw 0x10ff(%rip),%ymm7 # 264c <_sk_xor__hsw_lowp+0x354>
- DB 197,253,111,236 ; vmovdqa %ymm4,%ymm5
- DB 197,253,111,244 ; vmovdqa %ymm4,%ymm6
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,191 ; ja 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,193,0,0,0 ; lea 0xc1(%rip),%r11 # 1638 <_sk_load_g8_dst_hsw_lowp+0x12c>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,249,110,224 ; vmovd %eax,%xmm4
- DB 235,160 ; jmp 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm4,%xmm4
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,249,110,232 ; vmovd %eax,%xmm5
- DB 196,227,89,14,229,1 ; vpblendw $0x1,%xmm5,%xmm4,%xmm4
- DB 235,131 ; jmp 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,121,110,44,25 ; vmovd (%r9,%rbx,1),%xmm5
- DB 196,227,89,2,229,1 ; vpblendd $0x1,%xmm5,%xmm4,%xmm4
- DB 233,86,255,255,255 ; jmpq 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,122,126,44,25 ; vmovq (%r9,%rbx,1),%xmm5
- DB 196,227,81,2,228,12 ; vpblendd $0xc,%xmm4,%xmm5,%xmm4
- DB 233,41,255,255,255 ; jmpq 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 197,217,239,228 ; vpxor %xmm4,%xmm4,%xmm4
- DB 196,195,89,32,100,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,195,89,32,100,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm4,%xmm4
- DB 196,193,122,126,44,25 ; vmovq (%r9,%rbx,1),%xmm5
- DB 196,195,81,34,108,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm5,%xmm5
- DB 196,227,81,2,228,8 ; vpblendd $0x8,%xmm4,%xmm5,%xmm4
- DB 233,244,254,255,255 ; jmpq 152b <_sk_load_g8_dst_hsw_lowp+0x1f>
- DB 144 ; nop
- DB 72,255 ; rex.W (bad)
- DB 255 ; (bad)
- DB 255,95,255 ; lcall *-0x1(%rdi)
- DB 255 ; (bad)
- DB 255,83,255 ; callq *-0x1(%rbx)
- DB 255 ; (bad)
- DB 255,140,255,255,255,132,255 ; decl -0x7b0001(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 124,255 ; jl 164d <_sk_load_g8_dst_hsw_lowp+0x141>
- DB 255 ; (bad)
- DB 255,112,255 ; pushq -0x1(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 185,255,255,255,177 ; mov $0xb1ffffff,%ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,169,255,255,255,157 ; ljmp *-0x62000001(%rcx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,230 ; jmpq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 222,255 ; fdivrp %st,%st(7)
- DB 255 ; (bad)
- DB 255,214 ; callq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,202 ; dec %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_srcover_rgba_8888_hsw_lowp
-_sk_srcover_rgba_8888_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,221,1,0,0 ; jne 1872 <_sk_srcover_rgba_8888_hsw_lowp+0x1fe>
- DB 196,193,126,111,124,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 197,253,111,37,182,15,0,0 ; vmovdqa 0xfb6(%rip),%ymm4 # 2660 <_sk_xor__hsw_lowp+0x368>
- DB 196,226,61,0,236 ; vpshufb %ymm4,%ymm8,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,226,69,0,228 ; vpshufb %ymm4,%ymm7,%ymm4
- DB 196,227,253,0,228,232 ; vpermq $0xe8,%ymm4,%ymm4
- DB 196,227,85,56,228,1 ; vinserti128 $0x1,%xmm4,%ymm5,%ymm4
- DB 196,98,125,121,13,177,15,0,0 ; vpbroadcastw 0xfb1(%rip),%ymm9 # 2680 <_sk_xor__hsw_lowp+0x388>
- DB 197,221,113,244,8 ; vpsllw $0x8,%ymm4,%ymm4
- DB 196,98,125,121,21,165,15,0,0 ; vpbroadcastw 0xfa5(%rip),%ymm10 # 2682 <_sk_xor__hsw_lowp+0x38a>
- DB 196,193,93,228,226 ; vpmulhuw %ymm10,%ymm4,%ymm4
- DB 197,253,111,45,182,15,0,0 ; vmovdqa 0xfb6(%rip),%ymm5 # 26a0 <_sk_xor__hsw_lowp+0x3a8>
- DB 196,226,61,0,245 ; vpshufb %ymm5,%ymm8,%ymm6
- DB 196,227,253,0,246,232 ; vpermq $0xe8,%ymm6,%ymm6
- DB 196,226,69,0,237 ; vpshufb %ymm5,%ymm7,%ymm5
- DB 196,227,253,0,237,232 ; vpermq $0xe8,%ymm5,%ymm5
- DB 196,227,77,56,237,1 ; vinserti128 $0x1,%xmm5,%ymm6,%ymm5
- DB 197,213,113,245,8 ; vpsllw $0x8,%ymm5,%ymm5
- DB 196,193,85,228,234 ; vpmulhuw %ymm10,%ymm5,%ymm5
- DB 197,253,111,53,168,15,0,0 ; vmovdqa 0xfa8(%rip),%ymm6 # 26c0 <_sk_xor__hsw_lowp+0x3c8>
- DB 196,98,61,0,222 ; vpshufb %ymm6,%ymm8,%ymm11
- DB 196,67,253,0,219,232 ; vpermq $0xe8,%ymm11,%ymm11
- DB 196,226,69,0,246 ; vpshufb %ymm6,%ymm7,%ymm6
- DB 196,227,253,0,246,232 ; vpermq $0xe8,%ymm6,%ymm6
- DB 196,227,37,56,246,1 ; vinserti128 $0x1,%xmm6,%ymm11,%ymm6
- DB 197,205,113,246,8 ; vpsllw $0x8,%ymm6,%ymm6
- DB 196,193,77,228,242 ; vpmulhuw %ymm10,%ymm6,%ymm6
- DB 197,125,111,29,154,15,0,0 ; vmovdqa 0xf9a(%rip),%ymm11 # 26e0 <_sk_xor__hsw_lowp+0x3e8>
- DB 196,66,61,0,195 ; vpshufb %ymm11,%ymm8,%ymm8
- DB 196,67,253,0,192,232 ; vpermq $0xe8,%ymm8,%ymm8
- DB 196,194,69,0,251 ; vpshufb %ymm11,%ymm7,%ymm7
- DB 196,227,253,0,255,232 ; vpermq $0xe8,%ymm7,%ymm7
- DB 196,227,61,56,255,1 ; vinserti128 $0x1,%xmm7,%ymm8,%ymm7
- DB 197,197,113,247,8 ; vpsllw $0x8,%ymm7,%ymm7
- DB 196,193,69,228,250 ; vpmulhuw %ymm10,%ymm7,%ymm7
- DB 196,98,125,121,5,139,15,0,0 ; vpbroadcastw 0xf8b(%rip),%ymm8 # 2700 <_sk_xor__hsw_lowp+0x408>
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,66,93,11,208 ; vpmulhrsw %ymm8,%ymm4,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,192 ; vpaddw %ymm0,%ymm10,%ymm0
- DB 196,66,85,11,208 ; vpmulhrsw %ymm8,%ymm5,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,201 ; vpaddw %ymm1,%ymm10,%ymm1
- DB 196,66,77,11,208 ; vpmulhrsw %ymm8,%ymm6,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,210 ; vpaddw %ymm2,%ymm10,%ymm2
- DB 196,66,69,11,192 ; vpmulhrsw %ymm8,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 197,189,113,208,7 ; vpsrlw $0x7,%ymm0,%ymm8
- DB 196,65,61,234,193 ; vpminsw %ymm9,%ymm8,%ymm8
- DB 196,66,125,51,208 ; vpmovzxwd %xmm8,%ymm10
- DB 196,67,125,57,192,1 ; vextracti128 $0x1,%ymm8,%xmm8
- DB 196,66,125,51,192 ; vpmovzxwd %xmm8,%ymm8
- DB 197,165,113,209,7 ; vpsrlw $0x7,%ymm1,%ymm11
- DB 196,65,37,234,217 ; vpminsw %ymm9,%ymm11,%ymm11
- DB 196,67,125,57,220,1 ; vextracti128 $0x1,%ymm11,%xmm12
- DB 196,66,125,51,228 ; vpmovzxwd %xmm12,%ymm12
- DB 196,66,125,51,219 ; vpmovzxwd %xmm11,%ymm11
- DB 196,193,37,114,243,8 ; vpslld $0x8,%ymm11,%ymm11
- DB 196,193,29,114,244,8 ; vpslld $0x8,%ymm12,%ymm12
- DB 197,149,113,210,7 ; vpsrlw $0x7,%ymm2,%ymm13
- DB 196,65,21,234,233 ; vpminsw %ymm9,%ymm13,%ymm13
- DB 196,66,125,51,245 ; vpmovzxwd %xmm13,%ymm14
- DB 196,67,125,57,237,1 ; vextracti128 $0x1,%ymm13,%xmm13
- DB 196,66,125,51,237 ; vpmovzxwd %xmm13,%ymm13
- DB 196,193,21,114,245,16 ; vpslld $0x10,%ymm13,%ymm13
- DB 196,193,13,114,246,16 ; vpslld $0x10,%ymm14,%ymm14
- DB 197,133,113,211,7 ; vpsrlw $0x7,%ymm3,%ymm15
- DB 196,65,5,234,201 ; vpminsw %ymm9,%ymm15,%ymm9
- DB 196,67,125,57,207,1 ; vextracti128 $0x1,%ymm9,%xmm15
- DB 196,66,125,51,255 ; vpmovzxwd %xmm15,%ymm15
- DB 196,66,125,51,201 ; vpmovzxwd %xmm9,%ymm9
- DB 196,193,53,114,241,24 ; vpslld $0x18,%ymm9,%ymm9
- DB 196,193,5,114,247,24 ; vpslld $0x18,%ymm15,%ymm15
- DB 196,65,29,235,192 ; vpor %ymm8,%ymm12,%ymm8
- DB 196,65,37,235,218 ; vpor %ymm10,%ymm11,%ymm11
- DB 196,65,21,235,215 ; vpor %ymm15,%ymm13,%ymm10
- DB 196,65,61,235,210 ; vpor %ymm10,%ymm8,%ymm10
- DB 196,65,13,235,193 ; vpor %ymm9,%ymm14,%ymm8
- DB 196,65,37,235,192 ; vpor %ymm8,%ymm11,%ymm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,78 ; jne 18ae <_sk_srcover_rgba_8888_hsw_lowp+0x23a>
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,126,127,84,153,32 ; vmovdqu %ymm10,0x20(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,65,61,239,192 ; vpxor %ymm8,%ymm8,%ymm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,19,254,255,255 ; ja 16a2 <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,238,1,0,0 ; lea 0x1ee(%rip),%r11 # 1a88 <_sk_srcover_rgba_8888_hsw_lowp+0x414>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,110,4,153 ; vmovd (%r9,%rbx,4),%xmm8
- DB 233,244,253,255,255 ; jmpq 16a2 <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,175 ; ja 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 65,15,182,194 ; movzbl %r10b,%eax
- DB 76,141,21,251,1,0,0 ; lea 0x1fb(%rip),%r10 # 1ac4 <_sk_srcover_rgba_8888_hsw_lowp+0x450>
- DB 73,99,4,130 ; movslq (%r10,%rax,4),%rax
- DB 76,1,208 ; add %r10,%rax
- DB 255,224 ; jmpq *%rax
- DB 196,65,121,126,4,153 ; vmovd %xmm8,(%r9,%rbx,4)
- DB 235,147 ; jmp 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 196,193,121,110,100,153,8 ; vmovd 0x8(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,4 ; vpblendd $0x4,%ymm4,%ymm7,%ymm8
- DB 196,194,121,53,36,153 ; vpmovzxdq (%r9,%rbx,4),%xmm4
- DB 197,249,112,228,232 ; vpshufd $0xe8,%xmm4,%xmm4
- DB 196,99,61,2,196,3 ; vpblendd $0x3,%ymm4,%ymm8,%ymm8
- DB 233,156,253,255,255 ; jmpq 16a2 <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,24 ; vmovd 0x18(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,197,239,255 ; vpxor %ymm7,%ymm7,%ymm7
- DB 196,99,69,2,196,64 ; vpblendd $0x40,%ymm4,%ymm7,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,20,1 ; vpinsrd $0x1,0x14(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,99,125,57,196,1 ; vextracti128 $0x1,%ymm8,%xmm4
- DB 196,195,89,34,100,153,16,0 ; vpinsrd $0x0,0x10(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,99,61,56,196,1 ; vinserti128 $0x1,%xmm4,%ymm8,%ymm8
- DB 196,193,122,111,36,153 ; vmovdqu (%r9,%rbx,4),%xmm4
- DB 196,67,93,2,192,240 ; vpblendd $0xf0,%ymm8,%ymm4,%ymm8
- DB 233,77,253,255,255 ; jmpq 16a2 <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- DB 196,193,121,110,100,153,40 ; vmovd 0x28(%r9,%rbx,4),%xmm4
- DB 196,226,121,89,228 ; vpbroadcastq %xmm4,%xmm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,4 ; vpblendd $0x4,%ymm4,%ymm5,%ymm7
- DB 196,195,65,34,100,153,36,1 ; vpinsrd $0x1,0x24(%r9,%rbx,4),%xmm7,%xmm4
- DB 196,227,69,2,252,15 ; vpblendd $0xf,%ymm4,%ymm7,%ymm7
- DB 196,193,121,110,100,153,32 ; vmovd 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,69,2,252,1 ; vpblendd $0x1,%ymm4,%ymm7,%ymm7
- DB 233,17,253,255,255 ; jmpq 169c <_sk_srcover_rgba_8888_hsw_lowp+0x28>
- DB 196,193,121,110,100,153,56 ; vmovd 0x38(%r9,%rbx,4),%xmm4
- DB 196,226,125,89,228 ; vpbroadcastq %xmm4,%ymm4
- DB 197,213,239,237 ; vpxor %ymm5,%ymm5,%ymm5
- DB 196,227,85,2,252,64 ; vpblendd $0x40,%ymm4,%ymm5,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,52,1 ; vpinsrd $0x1,0x34(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,227,125,57,252,1 ; vextracti128 $0x1,%ymm7,%xmm4
- DB 196,195,89,34,100,153,48,0 ; vpinsrd $0x0,0x30(%r9,%rbx,4),%xmm4,%xmm4
- DB 196,227,69,56,252,1 ; vinserti128 $0x1,%xmm4,%ymm7,%ymm7
- DB 196,65,126,111,4,153 ; vmovdqu (%r9,%rbx,4),%ymm8
- DB 196,193,122,111,100,153,32 ; vmovdqu 0x20(%r9,%rbx,4),%xmm4
- DB 196,227,93,2,255,240 ; vpblendd $0xf0,%ymm7,%ymm4,%ymm7
- DB 233,193,252,255,255 ; jmpq 16a2 <_sk_srcover_rgba_8888_hsw_lowp+0x2e>
- DB 196,67,121,22,68,153,8,2 ; vpextrd $0x2,%xmm8,0x8(%r9,%rbx,4)
- DB 196,65,121,214,4,153 ; vmovq %xmm8,(%r9,%rbx,4)
- DB 233,121,254,255,255 ; jmpq 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,24,2 ; vpextrd $0x2,%xmm9,0x18(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,67,121,22,76,153,20,1 ; vpextrd $0x1,%xmm9,0x14(%r9,%rbx,4)
- DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
- DB 196,65,121,126,76,153,16 ; vmovd %xmm9,0x10(%r9,%rbx,4)
- DB 196,65,122,127,4,153 ; vmovdqu %xmm8,(%r9,%rbx,4)
- DB 233,69,254,255,255 ; jmpq 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 196,67,121,22,84,153,40,2 ; vpextrd $0x2,%xmm10,0x28(%r9,%rbx,4)
- DB 196,67,121,22,84,153,36,1 ; vpextrd $0x1,%xmm10,0x24(%r9,%rbx,4)
- DB 196,65,121,126,84,153,32 ; vmovd %xmm10,0x20(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 233,35,254,255,255 ; jmpq 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,67,121,22,76,153,56,2 ; vpextrd $0x2,%xmm9,0x38(%r9,%rbx,4)
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,67,121,22,76,153,52,1 ; vpextrd $0x1,%xmm9,0x34(%r9,%rbx,4)
- DB 196,67,125,57,209,1 ; vextracti128 $0x1,%ymm10,%xmm9
- DB 196,65,121,126,76,153,48 ; vmovd %xmm9,0x30(%r9,%rbx,4)
- DB 196,65,126,127,4,153 ; vmovdqu %ymm8,(%r9,%rbx,4)
- DB 196,65,122,127,84,153,32 ; vmovdqu %xmm10,0x20(%r9,%rbx,4)
- DB 233,232,253,255,255 ; jmpq 186d <_sk_srcover_rgba_8888_hsw_lowp+0x1f9>
- DB 15,31,0 ; nopl (%rax)
- DB 27,254 ; sbb %esi,%edi
- DB 255 ; (bad)
- DB 255,104,254 ; ljmp *-0x2(%rax)
- DB 255 ; (bad)
- DB 255,82,254 ; callq *-0x2(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 188,254,255,255,168 ; mov $0xa8fffffe,%esp
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,148,254,255,255,126,254 ; callq *-0x1810001(%rsi,%rdi,8)
- DB 255 ; (bad)
- DB 255,20,252 ; callq *(%rsp,%rdi,8)
- DB 255 ; (bad)
- DB 255,241 ; push %rcx
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,227 ; jmpq *%rbx
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,205 ; dec %ebp
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,65,255 ; incl -0x1(%rcx)
- DB 255 ; (bad)
- DB 255,45,255,255,255,25 ; ljmp *0x19ffffff(%rip) # 1a001abc <_sk_xor__hsw_lowp+0x19fff7c4>
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,14 ; decl (%rsi)
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255,37,255,255,255,29 ; jmpq *0x1dffffff(%rip) # 1e001acc <_sk_xor__hsw_lowp+0x1dfff7d4>
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,89,255 ; lcall *-0x1(%rcx)
- DB 255 ; (bad)
- DB 255,76,255,255 ; decl -0x1(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 62,255 ; ds (bad)
- DB 255 ; (bad)
- DB 255,48 ; pushq (%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 123,255 ; jnp 1ae1 <_sk_srcover_rgba_8888_hsw_lowp+0x46d>
- DB 255 ; (bad)
- DB 255,116,255,255 ; pushq -0x1(%rdi,%rdi,8)
- DB 255,108,255,255 ; ljmp *-0x1(%rdi,%rdi,8)
- DB 255,100,255,255 ; jmpq *-0x1(%rdi,%rdi,8)
- DB 255,175,255,255,255,162 ; ljmp *-0x5d000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,148,255,255,255,134,255 ; callq *-0x790001(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_scale_1_float_hsw_lowp
-_sk_scale_1_float_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,122,16,0 ; vmovss (%rax),%xmm8
- DB 197,58,88,5,106,8,0,0 ; vaddss 0x86a(%rip),%xmm8,%xmm8 # 2378 <_sk_xor__hsw_lowp+0x80>
- DB 197,121,126,192 ; vmovd %xmm8,%eax
- DB 197,121,110,192 ; vmovd %eax,%xmm8
- DB 196,66,125,121,192 ; vpbroadcastw %xmm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_scale_u8_hsw_lowp
-_sk_scale_u8_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,76 ; jne 1bac <_sk_scale_u8_hsw_lowp+0x65>
- DB 196,65,122,111,4,25 ; vmovdqu (%r9,%rbx,1),%xmm8
- DB 196,66,125,48,192 ; vpmovzxbw %xmm8,%ymm8
- DB 196,193,61,113,240,8 ; vpsllw $0x8,%ymm8,%ymm8
- DB 196,98,125,121,13,136,11,0,0 ; vpbroadcastw 0xb88(%rip),%ymm9 # 2702 <_sk_xor__hsw_lowp+0x40a>
- DB 196,65,61,228,193 ; vpmulhuw %ymm9,%ymm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 119,165 ; ja 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,200,0,0,0 ; lea 0xc8(%rip),%r11 # 1c94 <_sk_scale_u8_hsw_lowp+0x14d>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,121,110,192 ; vmovd %eax,%xmm8
- DB 235,134 ; jmp 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm8,%xmm8
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,121,110,200 ; vmovd %eax,%xmm9
- DB 196,67,57,14,193,1 ; vpblendw $0x1,%xmm9,%xmm8,%xmm8
- DB 233,101,255,255,255 ; jmpq 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,121,110,12,25 ; vmovd (%r9,%rbx,1),%xmm9
- DB 196,67,57,2,193,1 ; vpblendd $0x1,%xmm9,%xmm8,%xmm8
- DB 233,55,255,255,255 ; jmpq 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm9
- DB 196,67,49,2,192,12 ; vpblendd $0xc,%xmm8,%xmm9,%xmm8
- DB 233,9,255,255,255 ; jmpq 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm9
- DB 196,67,49,34,76,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm9,%xmm9
- DB 196,67,49,2,192,8 ; vpblendd $0x8,%xmm8,%xmm9,%xmm8
- DB 233,211,254,255,255 ; jmpq 1b66 <_sk_scale_u8_hsw_lowp+0x1f>
- DB 144 ; nop
- DB 65,255 ; rex.B (bad)
- DB 255 ; (bad)
- DB 255,89,255 ; lcall *-0x1(%rcx)
- DB 255 ; (bad)
- DB 255,76,255,255 ; decl -0x1(%rdi,%rdi,8)
- DB 255,138,255,255,255,130 ; decl -0x7d000001(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 122,255 ; jp 1ca9 <_sk_scale_u8_hsw_lowp+0x162>
- DB 255 ; (bad)
- DB 255,109,255 ; ljmp *-0x1(%rbp)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 184,255,255,255,176 ; mov $0xb0ffffff,%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,168,255,255,255,155 ; ljmp *-0x64000001(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,230 ; jmpq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 222,255 ; fdivrp %st,%st(7)
- DB 255 ; (bad)
- DB 255,214 ; callq *%rsi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,201 ; dec %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_lerp_1_float_hsw_lowp
-_sk_lerp_1_float_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,122,16,0 ; vmovss (%rax),%xmm8
- DB 197,58,88,5,158,6,0,0 ; vaddss 0x69e(%rip),%xmm8,%xmm8 # 237c <_sk_xor__hsw_lowp+0x84>
- DB 197,121,126,192 ; vmovd %xmm8,%eax
- DB 197,121,110,192 ; vmovd %eax,%xmm8
- DB 196,66,125,121,192 ; vpbroadcastw %xmm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,98,125,121,13,6,10,0,0 ; vpbroadcastw 0xa06(%rip),%ymm9 # 2704 <_sk_xor__hsw_lowp+0x40c>
- DB 196,65,53,249,200 ; vpsubw %ymm8,%ymm9,%ymm9
- DB 196,66,93,11,209 ; vpmulhrsw %ymm9,%ymm4,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,192 ; vpaddw %ymm0,%ymm10,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,66,85,11,209 ; vpmulhrsw %ymm9,%ymm5,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,201 ; vpaddw %ymm1,%ymm10,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,66,77,11,209 ; vpmulhrsw %ymm9,%ymm6,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,210 ; vpaddw %ymm2,%ymm10,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 196,66,69,11,193 ; vpmulhrsw %ymm9,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_lerp_u8_hsw_lowp
-_sk_lerp_u8_hsw_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,146,0,0,0 ; jne 1e0c <_sk_lerp_u8_hsw_lowp+0xaf>
- DB 196,65,122,111,4,25 ; vmovdqu (%r9,%rbx,1),%xmm8
- DB 196,66,125,48,192 ; vpmovzxbw %xmm8,%ymm8
- DB 196,193,61,113,240,8 ; vpsllw $0x8,%ymm8,%ymm8
- DB 196,98,125,121,13,114,9,0,0 ; vpbroadcastw 0x972(%rip),%ymm9 # 2706 <_sk_xor__hsw_lowp+0x40e>
- DB 196,65,61,228,193 ; vpmulhuw %ymm9,%ymm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,98,125,121,13,92,9,0,0 ; vpbroadcastw 0x95c(%rip),%ymm9 # 2708 <_sk_xor__hsw_lowp+0x410>
- DB 196,65,53,249,200 ; vpsubw %ymm8,%ymm9,%ymm9
- DB 196,66,93,11,209 ; vpmulhrsw %ymm9,%ymm4,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,192 ; vpaddw %ymm0,%ymm10,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,66,85,11,209 ; vpmulhrsw %ymm9,%ymm5,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,201 ; vpaddw %ymm1,%ymm10,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,66,77,11,209 ; vpmulhrsw %ymm9,%ymm6,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,210 ; vpaddw %ymm2,%ymm10,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 196,66,69,11,193 ; vpmulhrsw %ymm9,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,15 ; and $0xf,%r10b
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,14 ; cmp $0xe,%r10b
- DB 15,135,91,255,255,255 ; ja 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,204,0,0,0 ; lea 0xcc(%rip),%r11 # 1efc <_sk_lerp_u8_hsw_lowp+0x19f>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 197,121,110,192 ; vmovd %eax,%xmm8
- DB 233,57,255,255,255 ; jmpq 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,2,2 ; vpinsrb $0x2,0x2(%r9,%rbx,1),%xmm8,%xmm8
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 197,121,110,200 ; vmovd %eax,%xmm9
- DB 196,67,57,14,193,1 ; vpblendw $0x1,%xmm9,%xmm8,%xmm8
- DB 233,24,255,255,255 ; jmpq 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,6,6 ; vpinsrb $0x6,0x6(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,5,5 ; vpinsrb $0x5,0x5(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,4,4 ; vpinsrb $0x4,0x4(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,121,110,12,25 ; vmovd (%r9,%rbx,1),%xmm9
- DB 196,67,57,2,193,1 ; vpblendd $0x1,%xmm9,%xmm8,%xmm8
- DB 233,234,254,255,255 ; jmpq 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,10,10 ; vpinsrb $0xa,0xa(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,9,9 ; vpinsrb $0x9,0x9(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,8,8 ; vpinsrb $0x8,0x8(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm9
- DB 196,67,49,2,192,12 ; vpblendd $0xc,%xmm8,%xmm9,%xmm8
- DB 233,188,254,255,255 ; jmpq 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
- DB 196,67,57,32,68,25,14,14 ; vpinsrb $0xe,0xe(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,13,13 ; vpinsrb $0xd,0xd(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,67,57,32,68,25,12,12 ; vpinsrb $0xc,0xc(%r9,%rbx,1),%xmm8,%xmm8
- DB 196,65,122,126,12,25 ; vmovq (%r9,%rbx,1),%xmm9
- DB 196,67,49,34,76,25,8,2 ; vpinsrd $0x2,0x8(%r9,%rbx,1),%xmm9,%xmm9
- DB 196,67,49,2,192,8 ; vpblendd $0x8,%xmm8,%xmm9,%xmm8
- DB 233,134,254,255,255 ; jmpq 1d80 <_sk_lerp_u8_hsw_lowp+0x23>
- DB 102,144 ; xchg %ax,%ax
- DB 61,255,255,255,88 ; cmp $0x58ffffff,%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,75,255 ; decl -0x1(%rbx)
- DB 255 ; (bad)
- DB 255,137,255,255,255,129 ; decl -0x7e000001(%rcx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 121,255 ; jns 1f11 <_sk_lerp_u8_hsw_lowp+0x1b4>
- DB 255 ; (bad)
- DB 255,108,255,255 ; ljmp *-0x1(%rdi,%rdi,8)
- DB 255,183,255,255,255,175 ; pushq -0x50000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,167,255,255,255,154 ; jmpq *-0x65000001(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,229 ; jmpq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 221,255 ; (bad)
- DB 255 ; (bad)
- DB 255,213 ; callq *%rbp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,200 ; dec %eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_swap_rb_hsw_lowp
-_sk_swap_rb_hsw_lowp LABEL PROC
- DB 197,124,40,192 ; vmovaps %ymm0,%ymm8
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,252,40,194 ; vmovaps %ymm2,%ymm0
- DB 197,124,41,194 ; vmovaps %ymm8,%ymm2
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_move_src_dst_hsw_lowp
-_sk_move_src_dst_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,252,40,224 ; vmovaps %ymm0,%ymm4
- DB 197,252,40,233 ; vmovaps %ymm1,%ymm5
- DB 197,252,40,242 ; vmovaps %ymm2,%ymm6
- DB 197,252,40,251 ; vmovaps %ymm3,%ymm7
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_move_dst_src_hsw_lowp
-_sk_move_dst_src_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,252,40,196 ; vmovaps %ymm4,%ymm0
- DB 197,252,40,205 ; vmovaps %ymm5,%ymm1
- DB 197,252,40,214 ; vmovaps %ymm6,%ymm2
- DB 197,252,40,223 ; vmovaps %ymm7,%ymm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_clear_hsw_lowp
-_sk_clear_hsw_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
- DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
- DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
- DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcatop_hsw_lowp
-_sk_srcatop_hsw_lowp LABEL PROC
- DB 196,226,125,11,199 ; vpmulhrsw %ymm7,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,98,125,121,5,115,7,0,0 ; vpbroadcastw 0x773(%rip),%ymm8 # 270a <_sk_xor__hsw_lowp+0x412>
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,66,93,11,200 ; vpmulhrsw %ymm8,%ymm4,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,192 ; vpaddw %ymm0,%ymm9,%ymm0
- DB 196,226,117,11,207 ; vpmulhrsw %ymm7,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,66,85,11,200 ; vpmulhrsw %ymm8,%ymm5,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,201 ; vpaddw %ymm1,%ymm9,%ymm1
- DB 196,226,109,11,215 ; vpmulhrsw %ymm7,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,66,77,11,200 ; vpmulhrsw %ymm8,%ymm6,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,210 ; vpaddw %ymm2,%ymm9,%ymm2
- DB 196,226,101,11,223 ; vpmulhrsw %ymm7,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 196,66,69,11,192 ; vpmulhrsw %ymm8,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstatop_hsw_lowp
-_sk_dstatop_hsw_lowp LABEL PROC
- DB 196,98,93,11,195 ; vpmulhrsw %ymm3,%ymm4,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 196,98,125,121,13,4,7,0,0 ; vpbroadcastw 0x704(%rip),%ymm9 # 270c <_sk_xor__hsw_lowp+0x414>
- DB 197,53,249,207 ; vpsubw %ymm7,%ymm9,%ymm9
- DB 196,194,125,11,193 ; vpmulhrsw %ymm9,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,193,125,253,192 ; vpaddw %ymm8,%ymm0,%ymm0
- DB 196,98,85,11,195 ; vpmulhrsw %ymm3,%ymm5,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 196,194,117,11,201 ; vpmulhrsw %ymm9,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,193,117,253,200 ; vpaddw %ymm8,%ymm1,%ymm1
- DB 196,98,77,11,195 ; vpmulhrsw %ymm3,%ymm6,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 196,194,109,11,209 ; vpmulhrsw %ymm9,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,193,109,253,208 ; vpaddw %ymm8,%ymm2,%ymm2
- DB 196,98,69,11,195 ; vpmulhrsw %ymm3,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 196,194,101,11,217 ; vpmulhrsw %ymm9,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 196,193,101,253,216 ; vpaddw %ymm8,%ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcin_hsw_lowp
-_sk_srcin_hsw_lowp LABEL PROC
- DB 196,226,125,11,199 ; vpmulhrsw %ymm7,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,226,117,11,207 ; vpmulhrsw %ymm7,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,226,109,11,215 ; vpmulhrsw %ymm7,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,226,101,11,223 ; vpmulhrsw %ymm7,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstin_hsw_lowp
-_sk_dstin_hsw_lowp LABEL PROC
- DB 196,226,93,11,195 ; vpmulhrsw %ymm3,%ymm4,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,226,85,11,203 ; vpmulhrsw %ymm3,%ymm5,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,226,77,11,211 ; vpmulhrsw %ymm3,%ymm6,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,226,69,11,219 ; vpmulhrsw %ymm3,%ymm7,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcout_hsw_lowp
-_sk_srcout_hsw_lowp LABEL PROC
- DB 196,98,125,121,5,67,6,0,0 ; vpbroadcastw 0x643(%rip),%ymm8 # 270e <_sk_xor__hsw_lowp+0x416>
- DB 197,61,249,199 ; vpsubw %ymm7,%ymm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstout_hsw_lowp
-_sk_dstout_hsw_lowp LABEL PROC
- DB 196,226,125,121,5,12,6,0,0 ; vpbroadcastw 0x60c(%rip),%ymm0 # 2710 <_sk_xor__hsw_lowp+0x418>
- DB 197,253,249,219 ; vpsubw %ymm3,%ymm0,%ymm3
- DB 196,226,93,11,195 ; vpmulhrsw %ymm3,%ymm4,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,226,85,11,203 ; vpmulhrsw %ymm3,%ymm5,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,226,77,11,211 ; vpmulhrsw %ymm3,%ymm6,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,226,69,11,219 ; vpmulhrsw %ymm3,%ymm7,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcover_hsw_lowp
-_sk_srcover_hsw_lowp LABEL PROC
- DB 196,98,125,121,5,213,5,0,0 ; vpbroadcastw 0x5d5(%rip),%ymm8 # 2712 <_sk_xor__hsw_lowp+0x41a>
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,66,93,11,200 ; vpmulhrsw %ymm8,%ymm4,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,192 ; vpaddw %ymm0,%ymm9,%ymm0
- DB 196,66,85,11,200 ; vpmulhrsw %ymm8,%ymm5,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,201 ; vpaddw %ymm1,%ymm9,%ymm1
- DB 196,66,77,11,200 ; vpmulhrsw %ymm8,%ymm6,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,210 ; vpaddw %ymm2,%ymm9,%ymm2
- DB 196,66,69,11,192 ; vpmulhrsw %ymm8,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstover_hsw_lowp
-_sk_dstover_hsw_lowp LABEL PROC
- DB 196,98,125,121,5,142,5,0,0 ; vpbroadcastw 0x58e(%rip),%ymm8 # 2714 <_sk_xor__hsw_lowp+0x41c>
- DB 197,61,249,199 ; vpsubw %ymm7,%ymm8,%ymm8
- DB 196,194,125,11,192 ; vpmulhrsw %ymm8,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 197,253,253,196 ; vpaddw %ymm4,%ymm0,%ymm0
- DB 196,194,117,11,200 ; vpmulhrsw %ymm8,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 197,245,253,205 ; vpaddw %ymm5,%ymm1,%ymm1
- DB 196,194,109,11,208 ; vpmulhrsw %ymm8,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 197,237,253,214 ; vpaddw %ymm6,%ymm2,%ymm2
- DB 196,194,101,11,216 ; vpmulhrsw %ymm8,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 197,229,253,223 ; vpaddw %ymm7,%ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_modulate_hsw_lowp
-_sk_modulate_hsw_lowp LABEL PROC
- DB 196,226,125,11,196 ; vpmulhrsw %ymm4,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 196,226,117,11,205 ; vpmulhrsw %ymm5,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,226,109,11,214 ; vpmulhrsw %ymm6,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,226,101,11,223 ; vpmulhrsw %ymm7,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_multiply_hsw_lowp
-_sk_multiply_hsw_lowp LABEL PROC
- DB 196,98,125,121,5,27,5,0,0 ; vpbroadcastw 0x51b(%rip),%ymm8 # 2716 <_sk_xor__hsw_lowp+0x41e>
- DB 197,61,249,207 ; vpsubw %ymm7,%ymm8,%ymm9
- DB 196,66,125,11,209 ; vpmulhrsw %ymm9,%ymm0,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,66,93,11,216 ; vpmulhrsw %ymm8,%ymm4,%ymm11
- DB 196,66,125,29,219 ; vpabsw %ymm11,%ymm11
- DB 196,65,37,253,210 ; vpaddw %ymm10,%ymm11,%ymm10
- DB 196,226,125,11,196 ; vpmulhrsw %ymm4,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 197,173,253,192 ; vpaddw %ymm0,%ymm10,%ymm0
- DB 196,66,117,11,209 ; vpmulhrsw %ymm9,%ymm1,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 196,66,85,11,216 ; vpmulhrsw %ymm8,%ymm5,%ymm11
- DB 196,66,125,29,219 ; vpabsw %ymm11,%ymm11
- DB 196,65,37,253,210 ; vpaddw %ymm10,%ymm11,%ymm10
- DB 196,226,117,11,205 ; vpmulhrsw %ymm5,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 197,173,253,201 ; vpaddw %ymm1,%ymm10,%ymm1
- DB 196,66,109,11,209 ; vpmulhrsw %ymm9,%ymm2,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 196,66,77,11,216 ; vpmulhrsw %ymm8,%ymm6,%ymm11
- DB 196,66,125,29,219 ; vpabsw %ymm11,%ymm11
- DB 196,65,37,253,210 ; vpaddw %ymm10,%ymm11,%ymm10
- DB 196,226,109,11,214 ; vpmulhrsw %ymm6,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 197,173,253,210 ; vpaddw %ymm2,%ymm10,%ymm2
- DB 196,66,101,11,201 ; vpmulhrsw %ymm9,%ymm3,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 196,66,69,11,192 ; vpmulhrsw %ymm8,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 196,65,61,253,193 ; vpaddw %ymm9,%ymm8,%ymm8
- DB 196,226,101,11,223 ; vpmulhrsw %ymm7,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_screen_hsw_lowp
-_sk_screen_hsw_lowp LABEL PROC
- DB 196,98,125,121,5,108,4,0,0 ; vpbroadcastw 0x46c(%rip),%ymm8 # 2718 <_sk_xor__hsw_lowp+0x420>
- DB 197,61,249,200 ; vpsubw %ymm0,%ymm8,%ymm9
- DB 196,98,53,11,204 ; vpmulhrsw %ymm4,%ymm9,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,192 ; vpaddw %ymm0,%ymm9,%ymm0
- DB 197,61,249,201 ; vpsubw %ymm1,%ymm8,%ymm9
- DB 196,98,53,11,205 ; vpmulhrsw %ymm5,%ymm9,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,201 ; vpaddw %ymm1,%ymm9,%ymm1
- DB 197,61,249,202 ; vpsubw %ymm2,%ymm8,%ymm9
- DB 196,98,53,11,206 ; vpmulhrsw %ymm6,%ymm9,%ymm9
- DB 196,66,125,29,201 ; vpabsw %ymm9,%ymm9
- DB 197,181,253,210 ; vpaddw %ymm2,%ymm9,%ymm2
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,98,61,11,199 ; vpmulhrsw %ymm7,%ymm8,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_xor__hsw_lowp
-_sk_xor__hsw_lowp LABEL PROC
- DB 196,98,125,121,5,25,4,0,0 ; vpbroadcastw 0x419(%rip),%ymm8 # 271a <_sk_xor__hsw_lowp+0x422>
- DB 197,61,249,207 ; vpsubw %ymm7,%ymm8,%ymm9
- DB 196,194,125,11,193 ; vpmulhrsw %ymm9,%ymm0,%ymm0
- DB 196,226,125,29,192 ; vpabsw %ymm0,%ymm0
- DB 197,61,249,195 ; vpsubw %ymm3,%ymm8,%ymm8
- DB 196,66,93,11,208 ; vpmulhrsw %ymm8,%ymm4,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,192 ; vpaddw %ymm0,%ymm10,%ymm0
- DB 196,194,117,11,201 ; vpmulhrsw %ymm9,%ymm1,%ymm1
- DB 196,226,125,29,201 ; vpabsw %ymm1,%ymm1
- DB 196,66,85,11,208 ; vpmulhrsw %ymm8,%ymm5,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,201 ; vpaddw %ymm1,%ymm10,%ymm1
- DB 196,194,109,11,209 ; vpmulhrsw %ymm9,%ymm2,%ymm2
- DB 196,226,125,29,210 ; vpabsw %ymm2,%ymm2
- DB 196,66,77,11,208 ; vpmulhrsw %ymm8,%ymm6,%ymm10
- DB 196,66,125,29,210 ; vpabsw %ymm10,%ymm10
- DB 197,173,253,210 ; vpaddw %ymm2,%ymm10,%ymm2
- DB 196,194,101,11,217 ; vpmulhrsw %ymm9,%ymm3,%ymm3
- DB 196,226,125,29,219 ; vpabsw %ymm3,%ymm3
- DB 196,66,69,11,192 ; vpmulhrsw %ymm8,%ymm7,%ymm8
- DB 196,66,125,29,192 ; vpabsw %ymm8,%ymm8
- DB 197,189,253,219 ; vpaddw %ymm3,%ymm8,%ymm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-ALIGN 4
- DB 0,0 ; add %al,(%rax)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128 ; .byte 0x80
- DB 67 ; rex.XB
-
-ALIGN 32
- DB 0,128,0,128,0,0 ; add %al,0x8000(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,0,1 ; addb $0x1,(%rax)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,129,128,0,0,0,0 ; addb $0x0,0x80(%rcx)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 1,2 ; add %eax,(%rdx)
- DB 5,6,9,10,13 ; add $0xd0a0906,%eax
- DB 14 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,17 ; callq *(%rcx)
- DB 18,21,22,25,26,29 ; adc 0x1d1a1916(%rip),%dl # 1d1a3d0d <_sk_xor__hsw_lowp+0x1d1a1a15>
- DB 30 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,2 ; incl (%rdx)
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,18 ; callq *(%rdx)
- DB 19,22 ; adc (%rsi),%edx
- DB 23 ; (bad)
- DB 26,27 ; sbb (%rbx),%bl
- DB 30 ; (bad)
- DB 31 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,19 ; callq *(%rbx)
- DB 255,23 ; callq *(%rdi)
- DB 255,27 ; lcall *(%rbx)
- DB 255,31 ; lcall *(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 1,4,5,8,9,12,13 ; add %eax,0xd0c0908(,%rax,1)
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,0,1 ; addb $0x1,(%rax)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,129,128,0,0,0,0 ; addb $0x0,0x80(%rcx)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 1,2 ; add %eax,(%rdx)
- DB 5,6,9,10,13 ; add $0xd0a0906,%eax
- DB 14 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,17 ; callq *(%rcx)
- DB 18,21,22,25,26,29 ; adc 0x1d1a1916(%rip),%dl # 1d1a3dad <_sk_xor__hsw_lowp+0x1d1a1ab5>
- DB 30 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,2 ; incl (%rdx)
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,18 ; callq *(%rdx)
- DB 19,22 ; adc (%rsi),%edx
- DB 23 ; (bad)
- DB 26,27 ; sbb (%rbx),%bl
- DB 30 ; (bad)
- DB 31 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,19 ; callq *(%rbx)
- DB 255,23 ; callq *(%rdi)
- DB 255,27 ; lcall *(%rbx)
- DB 255,31 ; lcall *(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,0,1 ; addb $0x1,(%rax)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,129,128,0,0,0,0 ; addb $0x0,0x80(%rcx)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 1,2 ; add %eax,(%rdx)
- DB 5,6,9,10,13 ; add $0xd0a0906,%eax
- DB 14 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,17 ; callq *(%rcx)
- DB 18,21,22,25,26,29 ; adc 0x1d1a1916(%rip),%dl # 1d1a3e6d <_sk_xor__hsw_lowp+0x1d1a1b75>
- DB 30 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,2 ; incl (%rdx)
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,18 ; callq *(%rdx)
- DB 19,22 ; adc (%rsi),%edx
- DB 23 ; (bad)
- DB 26,27 ; sbb (%rbx),%bl
- DB 30 ; (bad)
- DB 31 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,19 ; callq *(%rbx)
- DB 255,23 ; callq *(%rdi)
- DB 255,27 ; lcall *(%rbx)
- DB 255,31 ; lcall *(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 1,4,5,8,9,12,13 ; add %eax,0xd0c0908(,%rax,1)
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,0,1 ; addb $0x1,(%rax)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,129,128,0,0,0,0 ; addb $0x0,0x80(%rcx)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 1,2 ; add %eax,(%rdx)
- DB 5,6,9,10,13 ; add $0xd0a0906,%eax
- DB 14 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,17 ; callq *(%rcx)
- DB 18,21,22,25,26,29 ; adc 0x1d1a1916(%rip),%dl # 1d1a3f0d <_sk_xor__hsw_lowp+0x1d1a1c15>
- DB 30 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,2 ; incl (%rdx)
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,18 ; callq *(%rdx)
- DB 19,22 ; adc (%rsi),%edx
- DB 23 ; (bad)
- DB 26,27 ; sbb (%rbx),%bl
- DB 30 ; (bad)
- DB 31 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,19 ; callq *(%rbx)
- DB 255,23 ; callq *(%rdi)
- DB 255,27 ; lcall *(%rbx)
- DB 255,31 ; lcall *(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 129,128,129,128,129,128,0,128,129,128; addl $0x80818000,-0x7f7e7f7f(%rax)
- DB 0,128,0,0,0,0 ; add %al,0x0(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,0,1 ; addb $0x1,(%rax)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 128,128,128,128,128,128,128 ; addb $0x80,-0x7f7f7f80(%rax)
- DB 128,255,0 ; cmp $0x0,%bh
- DB 129,128,0,0,0,0,0,0,0,0 ; addl $0x0,0x0(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 1,2 ; add %eax,(%rdx)
- DB 5,6,9,10,13 ; add $0xd0a0906,%eax
- DB 14 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,17 ; callq *(%rcx)
- DB 18,21,22,25,26,29 ; adc 0x1d1a1916(%rip),%dl # 1d1a3fcd <_sk_xor__hsw_lowp+0x1d1a1cd5>
- DB 30 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,2 ; incl (%rdx)
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,18 ; callq *(%rdx)
- DB 19,22 ; adc (%rsi),%edx
- DB 23 ; (bad)
- DB 26,27 ; sbb (%rbx),%bl
- DB 30 ; (bad)
- DB 31 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,19 ; callq *(%rbx)
- DB 255,23 ; callq *(%rdi)
- DB 255,27 ; lcall *(%rbx)
- DB 255,31 ; lcall *(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 128,129,128,0,128,129,128 ; addb $0x80,-0x7e7fff80(%rcx)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0 ; .byte 0x0
- DB 128 ; .byte 0x80
-ALIGN 32
-
-PUBLIC _sk_start_pipeline_ssse3_lowp
-_sk_start_pipeline_ssse3_lowp LABEL PROC
- DB 85 ; push %rbp
- DB 72,137,229 ; mov %rsp,%rbp
- DB 65,87 ; push %r15
- DB 65,86 ; push %r14
- DB 65,85 ; push %r13
- DB 65,84 ; push %r12
- DB 86 ; push %rsi
- DB 87 ; push %rdi
- DB 83 ; push %rbx
- DB 72,129,236,200,0,0,0 ; sub $0xc8,%rsp
- DB 68,15,41,125,176 ; movaps %xmm15,-0x50(%rbp)
- DB 68,15,41,117,160 ; movaps %xmm14,-0x60(%rbp)
- DB 68,15,41,109,144 ; movaps %xmm13,-0x70(%rbp)
- DB 68,15,41,101,128 ; movaps %xmm12,-0x80(%rbp)
- DB 68,15,41,157,112,255,255,255 ; movaps %xmm11,-0x90(%rbp)
- DB 68,15,41,149,96,255,255,255 ; movaps %xmm10,-0xa0(%rbp)
- DB 68,15,41,141,80,255,255,255 ; movaps %xmm9,-0xb0(%rbp)
- DB 68,15,41,133,64,255,255,255 ; movaps %xmm8,-0xc0(%rbp)
- DB 15,41,189,48,255,255,255 ; movaps %xmm7,-0xd0(%rbp)
- DB 15,41,181,32,255,255,255 ; movaps %xmm6,-0xe0(%rbp)
- DB 77,137,199 ; mov %r8,%r15
- DB 72,137,211 ; mov %rdx,%rbx
- DB 72,137,141,24,255,255,255 ; mov %rcx,-0xe8(%rbp)
- DB 72,139,117,48 ; mov 0x30(%rbp),%rsi
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 73,137,197 ; mov %rax,%r13
- DB 73,137,246 ; mov %rsi,%r14
- DB 76,137,141,8,255,255,255 ; mov %r9,-0xf8(%rbp)
- DB 76,57,203 ; cmp %r9,%rbx
- DB 115,110 ; jae eb <_sk_start_pipeline_ssse3_lowp+0xeb>
- DB 72,139,133,24,255,255,255 ; mov -0xe8(%rbp),%rax
- DB 72,141,64,8 ; lea 0x8(%rax),%rax
- DB 72,137,133,16,255,255,255 ; mov %rax,-0xf0(%rbp)
- DB 76,57,189,16,255,255,255 ; cmp %r15,-0xf0(%rbp)
- DB 72,139,149,24,255,255,255 ; mov -0xe8(%rbp),%rdx
- DB 119,43 ; ja ca <_sk_start_pipeline_ssse3_lowp+0xca>
- DB 76,139,165,24,255,255,255 ; mov -0xe8(%rbp),%r12
- DB 69,49,192 ; xor %r8d,%r8d
- DB 72,139,125,56 ; mov 0x38(%rbp),%rdi
- DB 76,137,246 ; mov %r14,%rsi
- DB 76,137,226 ; mov %r12,%rdx
- DB 72,137,217 ; mov %rbx,%rcx
- DB 65,255,213 ; callq *%r13
- DB 73,141,84,36,8 ; lea 0x8(%r12),%rdx
- DB 73,131,196,16 ; add $0x10,%r12
- DB 77,57,252 ; cmp %r15,%r12
- DB 73,137,212 ; mov %rdx,%r12
- DB 118,220 ; jbe a6 <_sk_start_pipeline_ssse3_lowp+0xa6>
- DB 77,137,248 ; mov %r15,%r8
- DB 73,41,208 ; sub %rdx,%r8
- DB 116,13 ; je df <_sk_start_pipeline_ssse3_lowp+0xdf>
- DB 72,139,125,56 ; mov 0x38(%rbp),%rdi
- DB 76,137,246 ; mov %r14,%rsi
- DB 72,137,217 ; mov %rbx,%rcx
- DB 65,255,213 ; callq *%r13
- DB 72,255,195 ; inc %rbx
- DB 72,59,157,8,255,255,255 ; cmp -0xf8(%rbp),%rbx
- DB 117,164 ; jne 8f <_sk_start_pipeline_ssse3_lowp+0x8f>
- DB 15,40,181,32,255,255,255 ; movaps -0xe0(%rbp),%xmm6
- DB 15,40,189,48,255,255,255 ; movaps -0xd0(%rbp),%xmm7
- DB 68,15,40,133,64,255,255,255 ; movaps -0xc0(%rbp),%xmm8
- DB 68,15,40,141,80,255,255,255 ; movaps -0xb0(%rbp),%xmm9
- DB 68,15,40,149,96,255,255,255 ; movaps -0xa0(%rbp),%xmm10
- DB 68,15,40,157,112,255,255,255 ; movaps -0x90(%rbp),%xmm11
- DB 68,15,40,101,128 ; movaps -0x80(%rbp),%xmm12
- DB 68,15,40,109,144 ; movaps -0x70(%rbp),%xmm13
- DB 68,15,40,117,160 ; movaps -0x60(%rbp),%xmm14
- DB 68,15,40,125,176 ; movaps -0x50(%rbp),%xmm15
- DB 72,129,196,200,0,0,0 ; add $0xc8,%rsp
- DB 91 ; pop %rbx
- DB 95 ; pop %rdi
- DB 94 ; pop %rsi
- DB 65,92 ; pop %r12
- DB 65,93 ; pop %r13
- DB 65,94 ; pop %r14
- DB 65,95 ; pop %r15
- DB 93 ; pop %rbp
- DB 195 ; retq
-
-PUBLIC _sk_just_return_ssse3_lowp
-_sk_just_return_ssse3_lowp LABEL PROC
- DB 195 ; retq
-
-PUBLIC _sk_uniform_color_ssse3_lowp
-_sk_uniform_color_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,16,24 ; movups (%rax),%xmm3
- DB 15,88,29,242,26,0,0 ; addps 0x1af2(%rip),%xmm3 # 1c40 <_sk_xor__ssse3_lowp+0xa3>
- DB 242,15,112,195,0 ; pshuflw $0x0,%xmm3,%xmm0
- DB 102,15,112,192,80 ; pshufd $0x50,%xmm0,%xmm0
- DB 242,15,112,203,170 ; pshuflw $0xaa,%xmm3,%xmm1
- DB 102,15,112,201,80 ; pshufd $0x50,%xmm1,%xmm1
- DB 243,15,112,211,0 ; pshufhw $0x0,%xmm3,%xmm2
- DB 102,15,112,210,250 ; pshufd $0xfa,%xmm2,%xmm2
- DB 243,15,112,219,170 ; pshufhw $0xaa,%xmm3,%xmm3
- DB 102,15,112,219,250 ; pshufd $0xfa,%xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_black_color_ssse3_lowp
-_sk_black_color_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,29,205,26,0,0 ; movaps 0x1acd(%rip),%xmm3 # 1c50 <_sk_xor__ssse3_lowp+0xb3>
- DB 15,87,192 ; xorps %xmm0,%xmm0
- DB 15,87,201 ; xorps %xmm1,%xmm1
- DB 15,87,210 ; xorps %xmm2,%xmm2
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_white_color_ssse3_lowp
-_sk_white_color_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,5,201,26,0,0 ; movaps 0x1ac9(%rip),%xmm0 # 1c60 <_sk_xor__ssse3_lowp+0xc3>
- DB 15,40,200 ; movaps %xmm0,%xmm1
- DB 15,40,208 ; movaps %xmm0,%xmm2
- DB 15,40,216 ; movaps %xmm0,%xmm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_set_rgb_ssse3_lowp
-_sk_set_rgb_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 243,15,16,21,228,29,0,0 ; movss 0x1de4(%rip),%xmm2 # 1f90 <_sk_xor__ssse3_lowp+0x3f3>
- DB 243,15,16,0 ; movss (%rax),%xmm0
- DB 243,15,88,194 ; addss %xmm2,%xmm0
- DB 102,65,15,126,193 ; movd %xmm0,%r9d
- DB 102,65,15,110,193 ; movd %r9d,%xmm0
- DB 242,15,112,192,0 ; pshuflw $0x0,%xmm0,%xmm0
- DB 102,15,112,192,80 ; pshufd $0x50,%xmm0,%xmm0
- DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
- DB 243,15,88,202 ; addss %xmm2,%xmm1
- DB 102,65,15,126,201 ; movd %xmm1,%r9d
- DB 102,65,15,110,201 ; movd %r9d,%xmm1
- DB 242,15,112,201,0 ; pshuflw $0x0,%xmm1,%xmm1
- DB 102,15,112,201,80 ; pshufd $0x50,%xmm1,%xmm1
- DB 243,15,88,80,8 ; addss 0x8(%rax),%xmm2
- DB 102,15,126,208 ; movd %xmm2,%eax
- DB 102,15,110,208 ; movd %eax,%xmm2
- DB 242,15,112,210,0 ; pshuflw $0x0,%xmm2,%xmm2
- DB 102,15,112,210,80 ; pshufd $0x50,%xmm2,%xmm2
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_premul_ssse3_lowp
-_sk_premul_ssse3_lowp LABEL PROC
- DB 102,15,56,11,195 ; pmulhrsw %xmm3,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,56,11,203 ; pmulhrsw %xmm3,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,56,11,211 ; pmulhrsw %xmm3,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_load_8888_ssse3_lowp
-_sk_load_8888_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,192,0,0,0 ; jne 303 <_sk_load_8888_ssse3_lowp+0xe1>
- DB 69,15,16,68,153,16 ; movups 0x10(%r9,%rbx,4),%xmm8
- DB 102,65,15,16,28,153 ; movupd (%r9,%rbx,4),%xmm3
- DB 102,15,111,13,25,26,0,0 ; movdqa 0x1a19(%rip),%xmm1 # 1c70 <_sk_xor__ssse3_lowp+0xd3>
- DB 102,15,40,195 ; movapd %xmm3,%xmm0
- DB 102,15,56,0,193 ; pshufb %xmm1,%xmm0
- DB 102,65,15,111,208 ; movdqa %xmm8,%xmm2
- DB 102,15,56,0,209 ; pshufb %xmm1,%xmm2
- DB 102,15,108,194 ; punpcklqdq %xmm2,%xmm0
- DB 102,15,113,240,8 ; psllw $0x8,%xmm0
- DB 102,68,15,111,13,4,26,0,0 ; movdqa 0x1a04(%rip),%xmm9 # 1c80 <_sk_xor__ssse3_lowp+0xe3>
- DB 102,65,15,228,193 ; pmulhuw %xmm9,%xmm0
- DB 102,68,15,111,21,6,26,0,0 ; movdqa 0x1a06(%rip),%xmm10 # 1c90 <_sk_xor__ssse3_lowp+0xf3>
- DB 102,15,40,203 ; movapd %xmm3,%xmm1
- DB 102,65,15,56,0,202 ; pshufb %xmm10,%xmm1
- DB 102,65,15,111,208 ; movdqa %xmm8,%xmm2
- DB 102,65,15,56,0,210 ; pshufb %xmm10,%xmm2
- DB 102,15,108,202 ; punpcklqdq %xmm2,%xmm1
- DB 102,15,113,241,8 ; psllw $0x8,%xmm1
- DB 102,65,15,228,201 ; pmulhuw %xmm9,%xmm1
- DB 102,68,15,111,21,234,25,0,0 ; movdqa 0x19ea(%rip),%xmm10 # 1ca0 <_sk_xor__ssse3_lowp+0x103>
- DB 102,15,40,211 ; movapd %xmm3,%xmm2
- DB 102,65,15,56,0,210 ; pshufb %xmm10,%xmm2
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,56,0,218 ; pshufb %xmm10,%xmm11
- DB 102,65,15,108,211 ; punpcklqdq %xmm11,%xmm2
- DB 102,15,113,242,8 ; psllw $0x8,%xmm2
- DB 102,65,15,228,209 ; pmulhuw %xmm9,%xmm2
- DB 102,68,15,111,21,205,25,0,0 ; movdqa 0x19cd(%rip),%xmm10 # 1cb0 <_sk_xor__ssse3_lowp+0x113>
- DB 102,65,15,56,0,218 ; pshufb %xmm10,%xmm3
- DB 102,69,15,56,0,194 ; pshufb %xmm10,%xmm8
- DB 102,65,15,108,216 ; punpcklqdq %xmm8,%xmm3
- DB 102,15,113,243,8 ; psllw $0x8,%xmm3
- DB 102,65,15,228,217 ; pmulhuw %xmm9,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,15,239,219 ; pxor %xmm3,%xmm3
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,47,255,255,255 ; ja 24f <_sk_load_8888_ssse3_lowp+0x2d>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,97,0,0,0 ; lea 0x61(%rip),%r11 # 38c <_sk_load_8888_ssse3_lowp+0x16a>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 243,65,15,16,28,153 ; movss (%r9,%rbx,4),%xmm3
- DB 233,16,255,255,255 ; jmpq 24f <_sk_load_8888_ssse3_lowp+0x2d>
- DB 102,65,15,110,68,153,8 ; movd 0x8(%r9,%rbx,4),%xmm0
- DB 102,15,112,216,69 ; pshufd $0x45,%xmm0,%xmm3
- DB 102,65,15,18,28,153 ; movlpd (%r9,%rbx,4),%xmm3
- DB 233,249,254,255,255 ; jmpq 24f <_sk_load_8888_ssse3_lowp+0x2d>
- DB 102,65,15,110,68,153,24 ; movd 0x18(%r9,%rbx,4),%xmm0
- DB 102,68,15,112,192,69 ; pshufd $0x45,%xmm0,%xmm8
- DB 243,65,15,16,68,153,20 ; movss 0x14(%r9,%rbx,4),%xmm0
- DB 65,15,198,192,0 ; shufps $0x0,%xmm8,%xmm0
- DB 65,15,198,192,226 ; shufps $0xe2,%xmm8,%xmm0
- DB 68,15,40,192 ; movaps %xmm0,%xmm8
- DB 243,65,15,16,68,153,16 ; movss 0x10(%r9,%rbx,4),%xmm0
- DB 243,68,15,16,192 ; movss %xmm0,%xmm8
- DB 233,192,254,255,255 ; jmpq 249 <_sk_load_8888_ssse3_lowp+0x27>
- DB 15,31,0 ; nopl (%rax)
- DB 168,255 ; test $0xff,%al
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,179 ; mov $0xb3ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 189,254,255,255,236 ; mov $0xecfffffe,%ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,215 ; callq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,202 ; dec %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_8888_dst_ssse3_lowp
-_sk_load_8888_dst_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,192,0,0,0 ; jne 489 <_sk_load_8888_dst_ssse3_lowp+0xe1>
- DB 69,15,16,68,153,16 ; movups 0x10(%r9,%rbx,4),%xmm8
- DB 102,65,15,16,60,153 ; movupd (%r9,%rbx,4),%xmm7
- DB 102,15,111,45,227,24,0,0 ; movdqa 0x18e3(%rip),%xmm5 # 1cc0 <_sk_xor__ssse3_lowp+0x123>
- DB 102,15,40,231 ; movapd %xmm7,%xmm4
- DB 102,15,56,0,229 ; pshufb %xmm5,%xmm4
- DB 102,65,15,111,240 ; movdqa %xmm8,%xmm6
- DB 102,15,56,0,245 ; pshufb %xmm5,%xmm6
- DB 102,15,108,230 ; punpcklqdq %xmm6,%xmm4
- DB 102,15,113,244,8 ; psllw $0x8,%xmm4
- DB 102,68,15,111,13,206,24,0,0 ; movdqa 0x18ce(%rip),%xmm9 # 1cd0 <_sk_xor__ssse3_lowp+0x133>
- DB 102,65,15,228,225 ; pmulhuw %xmm9,%xmm4
- DB 102,68,15,111,21,208,24,0,0 ; movdqa 0x18d0(%rip),%xmm10 # 1ce0 <_sk_xor__ssse3_lowp+0x143>
- DB 102,15,40,239 ; movapd %xmm7,%xmm5
- DB 102,65,15,56,0,234 ; pshufb %xmm10,%xmm5
- DB 102,65,15,111,240 ; movdqa %xmm8,%xmm6
- DB 102,65,15,56,0,242 ; pshufb %xmm10,%xmm6
- DB 102,15,108,238 ; punpcklqdq %xmm6,%xmm5
- DB 102,15,113,245,8 ; psllw $0x8,%xmm5
- DB 102,65,15,228,233 ; pmulhuw %xmm9,%xmm5
- DB 102,68,15,111,21,180,24,0,0 ; movdqa 0x18b4(%rip),%xmm10 # 1cf0 <_sk_xor__ssse3_lowp+0x153>
- DB 102,15,40,247 ; movapd %xmm7,%xmm6
- DB 102,65,15,56,0,242 ; pshufb %xmm10,%xmm6
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,56,0,218 ; pshufb %xmm10,%xmm11
- DB 102,65,15,108,243 ; punpcklqdq %xmm11,%xmm6
- DB 102,15,113,246,8 ; psllw $0x8,%xmm6
- DB 102,65,15,228,241 ; pmulhuw %xmm9,%xmm6
- DB 102,68,15,111,21,151,24,0,0 ; movdqa 0x1897(%rip),%xmm10 # 1d00 <_sk_xor__ssse3_lowp+0x163>
- DB 102,65,15,56,0,250 ; pshufb %xmm10,%xmm7
- DB 102,69,15,56,0,194 ; pshufb %xmm10,%xmm8
- DB 102,65,15,108,248 ; punpcklqdq %xmm8,%xmm7
- DB 102,15,113,247,8 ; psllw $0x8,%xmm7
- DB 102,65,15,228,249 ; pmulhuw %xmm9,%xmm7
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,47,255,255,255 ; ja 3d5 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,95,0,0,0 ; lea 0x5f(%rip),%r11 # 510 <_sk_load_8888_dst_ssse3_lowp+0x168>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 243,65,15,16,60,153 ; movss (%r9,%rbx,4),%xmm7
- DB 233,16,255,255,255 ; jmpq 3d5 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- DB 102,65,15,110,100,153,8 ; movd 0x8(%r9,%rbx,4),%xmm4
- DB 102,15,112,252,69 ; pshufd $0x45,%xmm4,%xmm7
- DB 102,65,15,18,60,153 ; movlpd (%r9,%rbx,4),%xmm7
- DB 233,249,254,255,255 ; jmpq 3d5 <_sk_load_8888_dst_ssse3_lowp+0x2d>
- DB 102,65,15,110,100,153,24 ; movd 0x18(%r9,%rbx,4),%xmm4
- DB 102,68,15,112,196,69 ; pshufd $0x45,%xmm4,%xmm8
- DB 243,65,15,16,100,153,20 ; movss 0x14(%r9,%rbx,4),%xmm4
- DB 65,15,198,224,0 ; shufps $0x0,%xmm8,%xmm4
- DB 65,15,198,224,226 ; shufps $0xe2,%xmm8,%xmm4
- DB 68,15,40,196 ; movaps %xmm4,%xmm8
- DB 243,65,15,16,100,153,16 ; movss 0x10(%r9,%rbx,4),%xmm4
- DB 243,68,15,16,196 ; movss %xmm4,%xmm8
- DB 233,192,254,255,255 ; jmpq 3cf <_sk_load_8888_dst_ssse3_lowp+0x27>
- DB 144 ; nop
- DB 170 ; stos %al,%es:(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,193 ; inc %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,181,255,255,255,191 ; pushq -0x40000001(%rbp)
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 217,255 ; fcos
- DB 255 ; (bad)
- DB 255,204 ; dec %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_8888_ssse3_lowp
-_sk_store_8888_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 102,68,15,111,192 ; movdqa %xmm0,%xmm8
- DB 102,65,15,113,208,7 ; psrlw $0x7,%xmm8
- DB 102,68,15,111,21,184,23,0,0 ; movdqa 0x17b8(%rip),%xmm10 # 1d10 <_sk_xor__ssse3_lowp+0x173>
- DB 102,69,15,234,194 ; pminsw %xmm10,%xmm8
- DB 102,69,15,239,219 ; pxor %xmm11,%xmm11
- DB 102,69,15,111,232 ; movdqa %xmm8,%xmm13
- DB 102,69,15,97,235 ; punpcklwd %xmm11,%xmm13
- DB 102,69,15,105,195 ; punpckhwd %xmm11,%xmm8
- DB 102,68,15,111,201 ; movdqa %xmm1,%xmm9
- DB 102,65,15,113,209,7 ; psrlw $0x7,%xmm9
- DB 102,69,15,234,202 ; pminsw %xmm10,%xmm9
- DB 102,69,15,111,225 ; movdqa %xmm9,%xmm12
- DB 102,69,15,105,227 ; punpckhwd %xmm11,%xmm12
- DB 102,69,15,97,203 ; punpcklwd %xmm11,%xmm9
- DB 102,65,15,114,241,8 ; pslld $0x8,%xmm9
- DB 102,65,15,114,244,8 ; pslld $0x8,%xmm12
- DB 102,69,15,235,224 ; por %xmm8,%xmm12
- DB 102,69,15,235,205 ; por %xmm13,%xmm9
- DB 102,68,15,111,234 ; movdqa %xmm2,%xmm13
- DB 102,65,15,113,213,7 ; psrlw $0x7,%xmm13
- DB 102,69,15,234,234 ; pminsw %xmm10,%xmm13
- DB 102,69,15,111,245 ; movdqa %xmm13,%xmm14
- DB 102,69,15,97,243 ; punpcklwd %xmm11,%xmm14
- DB 102,69,15,105,235 ; punpckhwd %xmm11,%xmm13
- DB 102,65,15,114,245,16 ; pslld $0x10,%xmm13
- DB 102,65,15,114,246,16 ; pslld $0x10,%xmm14
- DB 102,68,15,111,195 ; movdqa %xmm3,%xmm8
- DB 102,65,15,113,208,7 ; psrlw $0x7,%xmm8
- DB 102,69,15,234,194 ; pminsw %xmm10,%xmm8
- DB 102,69,15,111,208 ; movdqa %xmm8,%xmm10
- DB 102,69,15,105,211 ; punpckhwd %xmm11,%xmm10
- DB 102,69,15,97,195 ; punpcklwd %xmm11,%xmm8
- DB 102,65,15,114,240,24 ; pslld $0x18,%xmm8
- DB 102,65,15,114,242,24 ; pslld $0x18,%xmm10
- DB 102,69,15,235,213 ; por %xmm13,%xmm10
- DB 102,69,15,235,212 ; por %xmm12,%xmm10
- DB 102,69,15,235,198 ; por %xmm14,%xmm8
- DB 102,69,15,235,193 ; por %xmm9,%xmm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,18 ; jne 627 <_sk_store_8888_ssse3_lowp+0xfb>
- DB 243,69,15,127,4,153 ; movdqu %xmm8,(%r9,%rbx,4)
- DB 243,69,15,127,84,153,16 ; movdqu %xmm10,0x10(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,235 ; ja 622 <_sk_store_8888_ssse3_lowp+0xf6>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,82,0,0,0 ; lea 0x52(%rip),%r11 # 694 <_sk_store_8888_ssse3_lowp+0x168>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 102,69,15,126,4,153 ; movd %xmm8,(%r9,%rbx,4)
- DB 235,207 ; jmp 622 <_sk_store_8888_ssse3_lowp+0xf6>
- DB 102,69,15,112,200,78 ; pshufd $0x4e,%xmm8,%xmm9
- DB 102,69,15,126,76,153,8 ; movd %xmm9,0x8(%r9,%rbx,4)
- DB 102,69,15,214,4,153 ; movq %xmm8,(%r9,%rbx,4)
- DB 235,186 ; jmp 622 <_sk_store_8888_ssse3_lowp+0xf6>
- DB 102,69,15,112,202,78 ; pshufd $0x4e,%xmm10,%xmm9
- DB 102,69,15,126,76,153,24 ; movd %xmm9,0x18(%r9,%rbx,4)
- DB 102,69,15,112,202,229 ; pshufd $0xe5,%xmm10,%xmm9
- DB 102,69,15,126,76,153,20 ; movd %xmm9,0x14(%r9,%rbx,4)
- DB 102,69,15,126,84,153,16 ; movd %xmm10,0x10(%r9,%rbx,4)
- DB 243,69,15,127,4,153 ; movdqu %xmm8,(%r9,%rbx,4)
- DB 235,145 ; jmp 622 <_sk_store_8888_ssse3_lowp+0xf6>
- DB 15,31,0 ; nopl (%rax)
- DB 183,255 ; mov $0xff,%bh
- DB 255 ; (bad)
- DB 255,204 ; dec %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,245 ; mov $0xf5ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,225 ; jmpq *%rcx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,212 ; callq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_bgra_ssse3_lowp
-_sk_load_bgra_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,192,0,0,0 ; jne 791 <_sk_load_bgra_ssse3_lowp+0xe1>
- DB 69,15,16,68,153,16 ; movups 0x10(%r9,%rbx,4),%xmm8
- DB 102,65,15,16,28,153 ; movupd (%r9,%rbx,4),%xmm3
- DB 102,15,111,5,59,22,0,0 ; movdqa 0x163b(%rip),%xmm0 # 1d20 <_sk_xor__ssse3_lowp+0x183>
- DB 102,15,40,211 ; movapd %xmm3,%xmm2
- DB 102,15,56,0,208 ; pshufb %xmm0,%xmm2
- DB 102,65,15,111,200 ; movdqa %xmm8,%xmm1
- DB 102,15,56,0,200 ; pshufb %xmm0,%xmm1
- DB 102,15,108,209 ; punpcklqdq %xmm1,%xmm2
- DB 102,15,113,242,8 ; psllw $0x8,%xmm2
- DB 102,68,15,111,13,38,22,0,0 ; movdqa 0x1626(%rip),%xmm9 # 1d30 <_sk_xor__ssse3_lowp+0x193>
- DB 102,65,15,228,209 ; pmulhuw %xmm9,%xmm2
- DB 102,68,15,111,21,40,22,0,0 ; movdqa 0x1628(%rip),%xmm10 # 1d40 <_sk_xor__ssse3_lowp+0x1a3>
- DB 102,15,40,203 ; movapd %xmm3,%xmm1
- DB 102,65,15,56,0,202 ; pshufb %xmm10,%xmm1
- DB 102,65,15,111,192 ; movdqa %xmm8,%xmm0
- DB 102,65,15,56,0,194 ; pshufb %xmm10,%xmm0
- DB 102,15,108,200 ; punpcklqdq %xmm0,%xmm1
- DB 102,15,113,241,8 ; psllw $0x8,%xmm1
- DB 102,65,15,228,201 ; pmulhuw %xmm9,%xmm1
- DB 102,68,15,111,21,12,22,0,0 ; movdqa 0x160c(%rip),%xmm10 # 1d50 <_sk_xor__ssse3_lowp+0x1b3>
- DB 102,15,40,195 ; movapd %xmm3,%xmm0
- DB 102,65,15,56,0,194 ; pshufb %xmm10,%xmm0
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,56,0,218 ; pshufb %xmm10,%xmm11
- DB 102,65,15,108,195 ; punpcklqdq %xmm11,%xmm0
- DB 102,15,113,240,8 ; psllw $0x8,%xmm0
- DB 102,65,15,228,193 ; pmulhuw %xmm9,%xmm0
- DB 102,68,15,111,21,239,21,0,0 ; movdqa 0x15ef(%rip),%xmm10 # 1d60 <_sk_xor__ssse3_lowp+0x1c3>
- DB 102,65,15,56,0,218 ; pshufb %xmm10,%xmm3
- DB 102,69,15,56,0,194 ; pshufb %xmm10,%xmm8
- DB 102,65,15,108,216 ; punpcklqdq %xmm8,%xmm3
- DB 102,15,113,243,8 ; psllw $0x8,%xmm3
- DB 102,65,15,228,217 ; pmulhuw %xmm9,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,15,239,219 ; pxor %xmm3,%xmm3
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,47,255,255,255 ; ja 6dd <_sk_load_bgra_ssse3_lowp+0x2d>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,95,0,0,0 ; lea 0x5f(%rip),%r11 # 818 <_sk_load_bgra_ssse3_lowp+0x168>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 243,65,15,16,28,153 ; movss (%r9,%rbx,4),%xmm3
- DB 233,16,255,255,255 ; jmpq 6dd <_sk_load_bgra_ssse3_lowp+0x2d>
- DB 102,65,15,110,68,153,8 ; movd 0x8(%r9,%rbx,4),%xmm0
- DB 102,15,112,216,69 ; pshufd $0x45,%xmm0,%xmm3
- DB 102,65,15,18,28,153 ; movlpd (%r9,%rbx,4),%xmm3
- DB 233,249,254,255,255 ; jmpq 6dd <_sk_load_bgra_ssse3_lowp+0x2d>
- DB 102,65,15,110,68,153,24 ; movd 0x18(%r9,%rbx,4),%xmm0
- DB 102,68,15,112,192,69 ; pshufd $0x45,%xmm0,%xmm8
- DB 243,65,15,16,68,153,20 ; movss 0x14(%r9,%rbx,4),%xmm0
- DB 65,15,198,192,0 ; shufps $0x0,%xmm8,%xmm0
- DB 65,15,198,192,226 ; shufps $0xe2,%xmm8,%xmm0
- DB 68,15,40,192 ; movaps %xmm0,%xmm8
- DB 243,65,15,16,68,153,16 ; movss 0x10(%r9,%rbx,4),%xmm0
- DB 243,68,15,16,192 ; movss %xmm0,%xmm8
- DB 233,192,254,255,255 ; jmpq 6d7 <_sk_load_bgra_ssse3_lowp+0x27>
- DB 144 ; nop
- DB 170 ; stos %al,%es:(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,193 ; inc %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,181,255,255,255,191 ; pushq -0x40000001(%rbp)
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 217,255 ; fcos
- DB 255 ; (bad)
- DB 255,204 ; dec %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_bgra_dst_ssse3_lowp
-_sk_load_bgra_dst_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,192,0,0,0 ; jne 915 <_sk_load_bgra_dst_ssse3_lowp+0xe1>
- DB 69,15,16,68,153,16 ; movups 0x10(%r9,%rbx,4),%xmm8
- DB 102,65,15,16,60,153 ; movupd (%r9,%rbx,4),%xmm7
- DB 102,15,111,37,7,21,0,0 ; movdqa 0x1507(%rip),%xmm4 # 1d70 <_sk_xor__ssse3_lowp+0x1d3>
- DB 102,15,40,247 ; movapd %xmm7,%xmm6
- DB 102,15,56,0,244 ; pshufb %xmm4,%xmm6
- DB 102,65,15,111,232 ; movdqa %xmm8,%xmm5
- DB 102,15,56,0,236 ; pshufb %xmm4,%xmm5
- DB 102,15,108,245 ; punpcklqdq %xmm5,%xmm6
- DB 102,15,113,246,8 ; psllw $0x8,%xmm6
- DB 102,68,15,111,13,242,20,0,0 ; movdqa 0x14f2(%rip),%xmm9 # 1d80 <_sk_xor__ssse3_lowp+0x1e3>
- DB 102,65,15,228,241 ; pmulhuw %xmm9,%xmm6
- DB 102,68,15,111,21,244,20,0,0 ; movdqa 0x14f4(%rip),%xmm10 # 1d90 <_sk_xor__ssse3_lowp+0x1f3>
- DB 102,15,40,239 ; movapd %xmm7,%xmm5
- DB 102,65,15,56,0,234 ; pshufb %xmm10,%xmm5
- DB 102,65,15,111,224 ; movdqa %xmm8,%xmm4
- DB 102,65,15,56,0,226 ; pshufb %xmm10,%xmm4
- DB 102,15,108,236 ; punpcklqdq %xmm4,%xmm5
- DB 102,15,113,245,8 ; psllw $0x8,%xmm5
- DB 102,65,15,228,233 ; pmulhuw %xmm9,%xmm5
- DB 102,68,15,111,21,216,20,0,0 ; movdqa 0x14d8(%rip),%xmm10 # 1da0 <_sk_xor__ssse3_lowp+0x203>
- DB 102,15,40,231 ; movapd %xmm7,%xmm4
- DB 102,65,15,56,0,226 ; pshufb %xmm10,%xmm4
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,56,0,218 ; pshufb %xmm10,%xmm11
- DB 102,65,15,108,227 ; punpcklqdq %xmm11,%xmm4
- DB 102,15,113,244,8 ; psllw $0x8,%xmm4
- DB 102,65,15,228,225 ; pmulhuw %xmm9,%xmm4
- DB 102,68,15,111,21,187,20,0,0 ; movdqa 0x14bb(%rip),%xmm10 # 1db0 <_sk_xor__ssse3_lowp+0x213>
- DB 102,65,15,56,0,250 ; pshufb %xmm10,%xmm7
- DB 102,69,15,56,0,194 ; pshufb %xmm10,%xmm8
- DB 102,65,15,108,248 ; punpcklqdq %xmm8,%xmm7
- DB 102,15,113,247,8 ; psllw $0x8,%xmm7
- DB 102,65,15,228,249 ; pmulhuw %xmm9,%xmm7
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,47,255,255,255 ; ja 861 <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,95,0,0,0 ; lea 0x5f(%rip),%r11 # 99c <_sk_load_bgra_dst_ssse3_lowp+0x168>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 243,65,15,16,60,153 ; movss (%r9,%rbx,4),%xmm7
- DB 233,16,255,255,255 ; jmpq 861 <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- DB 102,65,15,110,100,153,8 ; movd 0x8(%r9,%rbx,4),%xmm4
- DB 102,15,112,252,69 ; pshufd $0x45,%xmm4,%xmm7
- DB 102,65,15,18,60,153 ; movlpd (%r9,%rbx,4),%xmm7
- DB 233,249,254,255,255 ; jmpq 861 <_sk_load_bgra_dst_ssse3_lowp+0x2d>
- DB 102,65,15,110,100,153,24 ; movd 0x18(%r9,%rbx,4),%xmm4
- DB 102,68,15,112,196,69 ; pshufd $0x45,%xmm4,%xmm8
- DB 243,65,15,16,100,153,20 ; movss 0x14(%r9,%rbx,4),%xmm4
- DB 65,15,198,224,0 ; shufps $0x0,%xmm8,%xmm4
- DB 65,15,198,224,226 ; shufps $0xe2,%xmm8,%xmm4
- DB 68,15,40,196 ; movaps %xmm4,%xmm8
- DB 243,65,15,16,100,153,16 ; movss 0x10(%r9,%rbx,4),%xmm4
- DB 243,68,15,16,196 ; movss %xmm4,%xmm8
- DB 233,192,254,255,255 ; jmpq 85b <_sk_load_bgra_dst_ssse3_lowp+0x27>
- DB 144 ; nop
- DB 170 ; stos %al,%es:(%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,193 ; inc %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,181,255,255,255,191 ; pushq -0x40000001(%rbp)
- DB 254 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 217,255 ; fcos
- DB 255 ; (bad)
- DB 255,204 ; dec %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_bgra_ssse3_lowp
-_sk_store_bgra_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,131,236,16 ; sub $0x10,%rsp
- DB 15,41,60,36 ; movaps %xmm7,(%rsp)
- DB 102,68,15,111,210 ; movdqa %xmm2,%xmm10
- DB 102,65,15,113,210,7 ; psrlw $0x7,%xmm10
- DB 102,68,15,111,13,235,19,0,0 ; movdqa 0x13eb(%rip),%xmm9 # 1dc0 <_sk_xor__ssse3_lowp+0x223>
- DB 102,69,15,234,209 ; pminsw %xmm9,%xmm10
- DB 102,69,15,239,246 ; pxor %xmm14,%xmm14
- DB 102,65,15,111,250 ; movdqa %xmm10,%xmm7
- DB 102,65,15,97,254 ; punpcklwd %xmm14,%xmm7
- DB 102,69,15,105,214 ; punpckhwd %xmm14,%xmm10
- DB 102,68,15,111,225 ; movdqa %xmm1,%xmm12
- DB 102,65,15,113,212,7 ; psrlw $0x7,%xmm12
- DB 102,69,15,234,225 ; pminsw %xmm9,%xmm12
- DB 102,69,15,111,236 ; movdqa %xmm12,%xmm13
- DB 102,69,15,97,238 ; punpcklwd %xmm14,%xmm13
- DB 102,69,15,105,230 ; punpckhwd %xmm14,%xmm12
- DB 102,68,15,111,248 ; movdqa %xmm0,%xmm15
- DB 102,65,15,113,215,7 ; psrlw $0x7,%xmm15
- DB 102,69,15,234,249 ; pminsw %xmm9,%xmm15
- DB 102,68,15,111,195 ; movdqa %xmm3,%xmm8
- DB 102,65,15,113,208,7 ; psrlw $0x7,%xmm8
- DB 102,69,15,234,193 ; pminsw %xmm9,%xmm8
- DB 102,69,15,111,207 ; movdqa %xmm15,%xmm9
- DB 102,69,15,105,206 ; punpckhwd %xmm14,%xmm9
- DB 102,69,15,97,254 ; punpcklwd %xmm14,%xmm15
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,105,222 ; punpckhwd %xmm14,%xmm11
- DB 102,69,15,97,198 ; punpcklwd %xmm14,%xmm8
- DB 102,65,15,114,241,16 ; pslld $0x10,%xmm9
- DB 102,69,15,235,202 ; por %xmm10,%xmm9
- DB 102,65,15,114,247,16 ; pslld $0x10,%xmm15
- DB 102,68,15,235,255 ; por %xmm7,%xmm15
- DB 102,65,15,114,244,8 ; pslld $0x8,%xmm12
- DB 102,65,15,114,243,24 ; pslld $0x18,%xmm11
- DB 102,69,15,235,220 ; por %xmm12,%xmm11
- DB 102,69,15,235,217 ; por %xmm9,%xmm11
- DB 102,65,15,114,245,8 ; pslld $0x8,%xmm13
- DB 102,65,15,114,240,24 ; pslld $0x18,%xmm8
- DB 102,69,15,235,197 ; por %xmm13,%xmm8
- DB 102,69,15,235,199 ; por %xmm15,%xmm8
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,26 ; jne ac3 <_sk_store_bgra_ssse3_lowp+0x10b>
- DB 243,69,15,127,4,153 ; movdqu %xmm8,(%r9,%rbx,4)
- DB 243,69,15,127,92,153,16 ; movdqu %xmm11,0x10(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,60,36 ; movaps (%rsp),%xmm7
- DB 72,131,196,16 ; add $0x10,%rsp
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,227 ; ja ab6 <_sk_store_bgra_ssse3_lowp+0xfe>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,82,0,0,0 ; lea 0x52(%rip),%r11 # b30 <_sk_store_bgra_ssse3_lowp+0x178>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 102,69,15,126,4,153 ; movd %xmm8,(%r9,%rbx,4)
- DB 235,199 ; jmp ab6 <_sk_store_bgra_ssse3_lowp+0xfe>
- DB 102,65,15,112,248,78 ; pshufd $0x4e,%xmm8,%xmm7
- DB 102,65,15,126,124,153,8 ; movd %xmm7,0x8(%r9,%rbx,4)
- DB 102,69,15,214,4,153 ; movq %xmm8,(%r9,%rbx,4)
- DB 235,178 ; jmp ab6 <_sk_store_bgra_ssse3_lowp+0xfe>
- DB 102,65,15,112,251,78 ; pshufd $0x4e,%xmm11,%xmm7
- DB 102,65,15,126,124,153,24 ; movd %xmm7,0x18(%r9,%rbx,4)
- DB 102,65,15,112,251,229 ; pshufd $0xe5,%xmm11,%xmm7
- DB 102,65,15,126,124,153,20 ; movd %xmm7,0x14(%r9,%rbx,4)
- DB 102,69,15,126,92,153,16 ; movd %xmm11,0x10(%r9,%rbx,4)
- DB 243,69,15,127,4,153 ; movdqu %xmm8,(%r9,%rbx,4)
- DB 235,137 ; jmp ab6 <_sk_store_bgra_ssse3_lowp+0xfe>
- DB 15,31,0 ; nopl (%rax)
- DB 183,255 ; mov $0xff,%bh
- DB 255 ; (bad)
- DB 255,204 ; dec %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 191,255,255,255,245 ; mov $0xf5ffffff,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 238 ; out %al,(%dx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,225 ; jmpq *%rcx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,212 ; callq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_a8_ssse3_lowp
-_sk_load_a8_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,38 ; jne b8b <_sk_load_a8_ssse3_lowp+0x3f>
- DB 243,65,15,126,28,25 ; movq (%r9,%rbx,1),%xmm3
- DB 102,15,96,216 ; punpcklbw %xmm0,%xmm3
- DB 102,15,113,243,8 ; psllw $0x8,%xmm3
- DB 102,15,228,29,84,18,0,0 ; pmulhuw 0x1254(%rip),%xmm3 # 1dd0 <_sk_xor__ssse3_lowp+0x233>
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 102,15,87,192 ; xorpd %xmm0,%xmm0
- DB 15,87,201 ; xorps %xmm1,%xmm1
- DB 15,87,210 ; xorps %xmm2,%xmm2
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,15,239,219 ; pxor %xmm3,%xmm3
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,208 ; ja b6f <_sk_load_a8_ssse3_lowp+0x23>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,110,0,0,0 ; lea 0x6e(%rip),%r11 # c18 <_sk_load_a8_ssse3_lowp+0xcc>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,15,110,216 ; movd %eax,%xmm3
- DB 235,177 ; jmp b6f <_sk_load_a8_ssse3_lowp+0x23>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,15,239,219 ; pxor %xmm3,%xmm3
- DB 102,15,196,216,2 ; pinsrw $0x2,%eax,%xmm3
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,15,110,192 ; movd %eax,%xmm0
- DB 102,15,96,192 ; punpcklbw %xmm0,%xmm0
- DB 243,15,16,216 ; movss %xmm0,%xmm3
- DB 235,143 ; jmp b6f <_sk_load_a8_ssse3_lowp+0x23>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,15,239,219 ; pxor %xmm3,%xmm3
- DB 102,15,196,216,6 ; pinsrw $0x6,%eax,%xmm3
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,15,196,216,5 ; pinsrw $0x5,%eax,%xmm3
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,15,196,216,4 ; pinsrw $0x4,%eax,%xmm3
- DB 102,65,15,110,4,25 ; movd (%r9,%rbx,1),%xmm0
- DB 102,15,96,192 ; punpcklbw %xmm0,%xmm0
- DB 242,15,16,216 ; movsd %xmm0,%xmm3
- DB 233,87,255,255,255 ; jmpq b6f <_sk_load_a8_ssse3_lowp+0x23>
- DB 155 ; fwait
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,181,255,255,255,166 ; pushq -0x59000001(%rbp)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 237 ; in (%dx),%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,226 ; jmpq *%rdx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,215 ; callq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,200 ; dec %eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_a8_dst_ssse3_lowp
-_sk_load_a8_dst_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,38 ; jne c73 <_sk_load_a8_dst_ssse3_lowp+0x3f>
- DB 243,65,15,126,60,25 ; movq (%r9,%rbx,1),%xmm7
- DB 102,15,96,248 ; punpcklbw %xmm0,%xmm7
- DB 102,15,113,247,8 ; psllw $0x8,%xmm7
- DB 102,15,228,61,124,17,0,0 ; pmulhuw 0x117c(%rip),%xmm7 # 1de0 <_sk_xor__ssse3_lowp+0x243>
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 102,15,87,228 ; xorpd %xmm4,%xmm4
- DB 15,87,237 ; xorps %xmm5,%xmm5
- DB 15,87,246 ; xorps %xmm6,%xmm6
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,208 ; ja c57 <_sk_load_a8_dst_ssse3_lowp+0x23>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,110,0,0,0 ; lea 0x6e(%rip),%r11 # d00 <_sk_load_a8_dst_ssse3_lowp+0xcc>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,15,110,248 ; movd %eax,%xmm7
- DB 235,177 ; jmp c57 <_sk_load_a8_dst_ssse3_lowp+0x23>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 102,15,196,248,2 ; pinsrw $0x2,%eax,%xmm7
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,15,110,224 ; movd %eax,%xmm4
- DB 102,15,96,224 ; punpcklbw %xmm0,%xmm4
- DB 243,15,16,252 ; movss %xmm4,%xmm7
- DB 235,143 ; jmp c57 <_sk_load_a8_dst_ssse3_lowp+0x23>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 102,15,196,248,6 ; pinsrw $0x6,%eax,%xmm7
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,15,196,248,5 ; pinsrw $0x5,%eax,%xmm7
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,15,196,248,4 ; pinsrw $0x4,%eax,%xmm7
- DB 102,65,15,110,36,25 ; movd (%r9,%rbx,1),%xmm4
- DB 102,15,96,224 ; punpcklbw %xmm0,%xmm4
- DB 242,15,16,252 ; movsd %xmm4,%xmm7
- DB 233,87,255,255,255 ; jmpq c57 <_sk_load_a8_dst_ssse3_lowp+0x23>
- DB 155 ; fwait
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,181,255,255,255,166 ; pushq -0x59000001(%rbp)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 237 ; in (%dx),%eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,226 ; jmpq *%rdx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,215 ; callq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,200 ; dec %eax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_store_a8_ssse3_lowp
-_sk_store_a8_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,131,236,80 ; sub $0x50,%rsp
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 102,68,15,111,195 ; movdqa %xmm3,%xmm8
- DB 102,65,15,113,208,7 ; psrlw $0x7,%xmm8
- DB 102,69,15,103,192 ; packuswb %xmm8,%xmm8
- DB 77,133,192 ; test %r8,%r8
- DB 117,15 ; jne d58 <_sk_store_a8_ssse3_lowp+0x3c>
- DB 242,69,15,17,4,25 ; movsd %xmm8,(%r9,%rbx,1)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 72,131,196,80 ; add $0x50,%rsp
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,231 ; ja d4f <_sk_store_a8_ssse3_lowp+0x33>
- DB 102,68,15,96,192 ; punpcklbw %xmm0,%xmm8
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,132,0,0,0 ; lea 0x84(%rip),%r11 # dfc <_sk_store_a8_ssse3_lowp+0xe0>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 102,68,15,127,4,36 ; movdqa %xmm8,(%rsp)
- DB 138,4,36 ; mov (%rsp),%al
- DB 65,136,4,25 ; mov %al,(%r9,%rbx,1)
- DB 235,191 ; jmp d4f <_sk_store_a8_ssse3_lowp+0x33>
- DB 102,68,15,127,68,36,16 ; movdqa %xmm8,0x10(%rsp)
- DB 138,68,36,20 ; mov 0x14(%rsp),%al
- DB 65,136,68,25,2 ; mov %al,0x2(%r9,%rbx,1)
- DB 102,68,15,56,0,5,86,16,0,0 ; pshufb 0x1056(%rip),%xmm8 # 1e00 <_sk_xor__ssse3_lowp+0x263>
- DB 102,68,15,126,192 ; movd %xmm8,%eax
- DB 102,65,137,4,25 ; mov %ax,(%r9,%rbx,1)
- DB 235,153 ; jmp d4f <_sk_store_a8_ssse3_lowp+0x33>
- DB 102,68,15,127,68,36,64 ; movdqa %xmm8,0x40(%rsp)
- DB 138,68,36,76 ; mov 0x4c(%rsp),%al
- DB 65,136,68,25,6 ; mov %al,0x6(%r9,%rbx,1)
- DB 102,68,15,127,68,36,48 ; movdqa %xmm8,0x30(%rsp)
- DB 138,68,36,58 ; mov 0x3a(%rsp),%al
- DB 65,136,68,25,5 ; mov %al,0x5(%r9,%rbx,1)
- DB 102,68,15,127,68,36,32 ; movdqa %xmm8,0x20(%rsp)
- DB 138,68,36,40 ; mov 0x28(%rsp),%al
- DB 65,136,68,25,4 ; mov %al,0x4(%r9,%rbx,1)
- DB 102,68,15,56,0,5,0,16,0,0 ; pshufb 0x1000(%rip),%xmm8 # 1df0 <_sk_xor__ssse3_lowp+0x253>
- DB 102,69,15,126,4,25 ; movd %xmm8,(%r9,%rbx,1)
- DB 233,84,255,255,255 ; jmpq d4f <_sk_store_a8_ssse3_lowp+0x33>
- DB 144 ; nop
- DB 133,255 ; test %edi,%edi
- DB 255 ; (bad)
- DB 255,164,255,255,255,148,255 ; jmpq *-0x6b0001(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 234 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 218,255 ; (bad)
- DB 255 ; (bad)
- DB 255,202 ; dec %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 186 ; .byte 0xba
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_g8_ssse3_lowp
-_sk_load_g8_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,43 ; jne e5c <_sk_load_g8_ssse3_lowp+0x44>
- DB 243,65,15,126,4,25 ; movq (%r9,%rbx,1),%xmm0
- DB 102,15,96,192 ; punpcklbw %xmm0,%xmm0
- DB 102,15,113,240,8 ; psllw $0x8,%xmm0
- DB 102,15,228,5,200,15,0,0 ; pmulhuw 0xfc8(%rip),%xmm0 # 1e10 <_sk_xor__ssse3_lowp+0x273>
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,29,207,15,0,0 ; movaps 0xfcf(%rip),%xmm3 # 1e20 <_sk_xor__ssse3_lowp+0x283>
- DB 102,15,111,200 ; movdqa %xmm0,%xmm1
- DB 102,15,111,208 ; movdqa %xmm0,%xmm2
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,15,239,192 ; pxor %xmm0,%xmm0
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,203 ; ja e3b <_sk_load_g8_ssse3_lowp+0x23>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,113,0,0,0 ; lea 0x71(%rip),%r11 # eec <_sk_load_g8_ssse3_lowp+0xd4>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,15,110,192 ; movd %eax,%xmm0
- DB 235,172 ; jmp e3b <_sk_load_g8_ssse3_lowp+0x23>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,15,239,192 ; pxor %xmm0,%xmm0
- DB 102,15,196,192,2 ; pinsrw $0x2,%eax,%xmm0
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,15,110,200 ; movd %eax,%xmm1
- DB 102,15,96,200 ; punpcklbw %xmm0,%xmm1
- DB 243,15,16,193 ; movss %xmm1,%xmm0
- DB 235,138 ; jmp e3b <_sk_load_g8_ssse3_lowp+0x23>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,15,239,192 ; pxor %xmm0,%xmm0
- DB 102,15,196,192,6 ; pinsrw $0x6,%eax,%xmm0
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,15,196,192,5 ; pinsrw $0x5,%eax,%xmm0
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,15,196,192,4 ; pinsrw $0x4,%eax,%xmm0
- DB 102,65,15,110,12,25 ; movd (%r9,%rbx,1),%xmm1
- DB 102,15,96,200 ; punpcklbw %xmm0,%xmm1
- DB 242,15,16,193 ; movsd %xmm1,%xmm0
- DB 233,82,255,255,255 ; jmpq e3b <_sk_load_g8_ssse3_lowp+0x23>
- DB 15,31,0 ; nopl (%rax)
- DB 152 ; cwtl
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,178,255,255,255,163 ; pushq -0x5c000001(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 234 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 223,255 ; (bad)
- DB 255 ; (bad)
- DB 255,212 ; callq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,197 ; inc %ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_load_g8_dst_ssse3_lowp
-_sk_load_g8_dst_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,43 ; jne f4c <_sk_load_g8_dst_ssse3_lowp+0x44>
- DB 243,65,15,126,36,25 ; movq (%r9,%rbx,1),%xmm4
- DB 102,15,96,224 ; punpcklbw %xmm0,%xmm4
- DB 102,15,113,244,8 ; psllw $0x8,%xmm4
- DB 102,15,228,37,248,14,0,0 ; pmulhuw 0xef8(%rip),%xmm4 # 1e30 <_sk_xor__ssse3_lowp+0x293>
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,61,255,14,0,0 ; movaps 0xeff(%rip),%xmm7 # 1e40 <_sk_xor__ssse3_lowp+0x2a3>
- DB 102,15,111,236 ; movdqa %xmm4,%xmm5
- DB 102,15,111,244 ; movdqa %xmm4,%xmm6
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,15,239,228 ; pxor %xmm4,%xmm4
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,203 ; ja f2b <_sk_load_g8_dst_ssse3_lowp+0x23>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,113,0,0,0 ; lea 0x71(%rip),%r11 # fdc <_sk_load_g8_dst_ssse3_lowp+0xd4>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,15,110,224 ; movd %eax,%xmm4
- DB 235,172 ; jmp f2b <_sk_load_g8_dst_ssse3_lowp+0x23>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,15,239,228 ; pxor %xmm4,%xmm4
- DB 102,15,196,224,2 ; pinsrw $0x2,%eax,%xmm4
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,15,110,232 ; movd %eax,%xmm5
- DB 102,15,96,232 ; punpcklbw %xmm0,%xmm5
- DB 243,15,16,229 ; movss %xmm5,%xmm4
- DB 235,138 ; jmp f2b <_sk_load_g8_dst_ssse3_lowp+0x23>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,15,239,228 ; pxor %xmm4,%xmm4
- DB 102,15,196,224,6 ; pinsrw $0x6,%eax,%xmm4
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,15,196,224,5 ; pinsrw $0x5,%eax,%xmm4
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,15,196,224,4 ; pinsrw $0x4,%eax,%xmm4
- DB 102,65,15,110,44,25 ; movd (%r9,%rbx,1),%xmm5
- DB 102,15,96,232 ; punpcklbw %xmm0,%xmm5
- DB 242,15,16,229 ; movsd %xmm5,%xmm4
- DB 233,82,255,255,255 ; jmpq f2b <_sk_load_g8_dst_ssse3_lowp+0x23>
- DB 15,31,0 ; nopl (%rax)
- DB 152 ; cwtl
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,178,255,255,255,163 ; pushq -0x5c000001(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 234 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 223,255 ; (bad)
- DB 255 ; (bad)
- DB 255,212 ; callq *%rsp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,197 ; inc %ebp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_srcover_rgba_8888_ssse3_lowp
-_sk_srcover_rgba_8888_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,131,236,16 ; sub $0x10,%rsp
- DB 102,68,15,111,225 ; movdqa %xmm1,%xmm12
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 73,193,225,2 ; shl $0x2,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,14,2,0,0 ; jne 1230 <_sk_srcover_rgba_8888_ssse3_lowp+0x238>
- DB 69,15,16,68,153,16 ; movups 0x10(%r9,%rbx,4),%xmm8
- DB 102,65,15,16,60,153 ; movupd (%r9,%rbx,4),%xmm7
- DB 102,15,111,45,26,14,0,0 ; movdqa 0xe1a(%rip),%xmm5 # 1e50 <_sk_xor__ssse3_lowp+0x2b3>
- DB 102,15,40,231 ; movapd %xmm7,%xmm4
- DB 102,15,56,0,229 ; pshufb %xmm5,%xmm4
- DB 102,65,15,111,240 ; movdqa %xmm8,%xmm6
- DB 102,15,56,0,245 ; pshufb %xmm5,%xmm6
- DB 102,15,108,230 ; punpcklqdq %xmm6,%xmm4
- DB 102,15,111,13,11,14,0,0 ; movdqa 0xe0b(%rip),%xmm1 # 1e60 <_sk_xor__ssse3_lowp+0x2c3>
- DB 102,15,113,244,8 ; psllw $0x8,%xmm4
- DB 102,68,15,111,13,13,14,0,0 ; movdqa 0xe0d(%rip),%xmm9 # 1e70 <_sk_xor__ssse3_lowp+0x2d3>
- DB 102,65,15,228,225 ; pmulhuw %xmm9,%xmm4
- DB 102,68,15,111,21,15,14,0,0 ; movdqa 0xe0f(%rip),%xmm10 # 1e80 <_sk_xor__ssse3_lowp+0x2e3>
- DB 102,15,40,239 ; movapd %xmm7,%xmm5
- DB 102,65,15,56,0,234 ; pshufb %xmm10,%xmm5
- DB 102,65,15,111,240 ; movdqa %xmm8,%xmm6
- DB 102,65,15,56,0,242 ; pshufb %xmm10,%xmm6
- DB 102,15,108,238 ; punpcklqdq %xmm6,%xmm5
- DB 102,15,113,245,8 ; psllw $0x8,%xmm5
- DB 102,65,15,228,233 ; pmulhuw %xmm9,%xmm5
- DB 102,68,15,111,21,243,13,0,0 ; movdqa 0xdf3(%rip),%xmm10 # 1e90 <_sk_xor__ssse3_lowp+0x2f3>
- DB 102,15,40,247 ; movapd %xmm7,%xmm6
- DB 102,65,15,56,0,242 ; pshufb %xmm10,%xmm6
- DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
- DB 102,69,15,56,0,218 ; pshufb %xmm10,%xmm11
- DB 102,65,15,108,243 ; punpcklqdq %xmm11,%xmm6
- DB 102,15,113,246,8 ; psllw $0x8,%xmm6
- DB 102,65,15,228,241 ; pmulhuw %xmm9,%xmm6
- DB 102,68,15,111,21,214,13,0,0 ; movdqa 0xdd6(%rip),%xmm10 # 1ea0 <_sk_xor__ssse3_lowp+0x303>
- DB 102,65,15,56,0,250 ; pshufb %xmm10,%xmm7
- DB 102,69,15,56,0,194 ; pshufb %xmm10,%xmm8
- DB 102,65,15,108,248 ; punpcklqdq %xmm8,%xmm7
- DB 102,15,113,247,8 ; psllw $0x8,%xmm7
- DB 102,65,15,228,249 ; pmulhuw %xmm9,%xmm7
- DB 102,68,15,111,29,194,13,0,0 ; movdqa 0xdc2(%rip),%xmm11 # 1eb0 <_sk_xor__ssse3_lowp+0x313>
- DB 102,68,15,249,219 ; psubw %xmm3,%xmm11
- DB 102,68,15,111,196 ; movdqa %xmm4,%xmm8
- DB 102,69,15,56,11,195 ; pmulhrsw %xmm11,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,68,15,253,192 ; paddw %xmm0,%xmm8
- DB 102,15,111,197 ; movdqa %xmm5,%xmm0
- DB 102,65,15,56,11,195 ; pmulhrsw %xmm11,%xmm0
- DB 102,68,15,56,29,200 ; pabsw %xmm0,%xmm9
- DB 102,69,15,253,204 ; paddw %xmm12,%xmm9
- DB 102,15,111,198 ; movdqa %xmm6,%xmm0
- DB 102,65,15,56,11,195 ; pmulhrsw %xmm11,%xmm0
- DB 102,68,15,56,29,224 ; pabsw %xmm0,%xmm12
- DB 102,68,15,253,226 ; paddw %xmm2,%xmm12
- DB 102,68,15,56,11,223 ; pmulhrsw %xmm7,%xmm11
- DB 102,69,15,56,29,219 ; pabsw %xmm11,%xmm11
- DB 102,68,15,253,219 ; paddw %xmm3,%xmm11
- DB 102,68,15,127,4,36 ; movdqa %xmm8,(%rsp)
- DB 102,65,15,111,216 ; movdqa %xmm8,%xmm3
- DB 102,15,113,211,7 ; psrlw $0x7,%xmm3
- DB 102,15,234,217 ; pminsw %xmm1,%xmm3
- DB 102,69,15,239,255 ; pxor %xmm15,%xmm15
- DB 102,68,15,111,235 ; movdqa %xmm3,%xmm13
- DB 102,69,15,97,239 ; punpcklwd %xmm15,%xmm13
- DB 102,65,15,105,223 ; punpckhwd %xmm15,%xmm3
- DB 102,65,15,111,209 ; movdqa %xmm9,%xmm2
- DB 102,15,113,210,7 ; psrlw $0x7,%xmm2
- DB 102,15,234,209 ; pminsw %xmm1,%xmm2
- DB 102,68,15,111,242 ; movdqa %xmm2,%xmm14
- DB 102,69,15,105,247 ; punpckhwd %xmm15,%xmm14
- DB 102,65,15,97,215 ; punpcklwd %xmm15,%xmm2
- DB 102,69,15,111,212 ; movdqa %xmm12,%xmm10
- DB 102,65,15,113,210,7 ; psrlw $0x7,%xmm10
- DB 102,68,15,234,209 ; pminsw %xmm1,%xmm10
- DB 102,69,15,111,194 ; movdqa %xmm10,%xmm8
- DB 102,69,15,97,199 ; punpcklwd %xmm15,%xmm8
- DB 102,69,15,105,215 ; punpckhwd %xmm15,%xmm10
- DB 102,65,15,111,195 ; movdqa %xmm11,%xmm0
- DB 102,15,113,208,7 ; psrlw $0x7,%xmm0
- DB 102,15,234,193 ; pminsw %xmm1,%xmm0
- DB 102,15,111,200 ; movdqa %xmm0,%xmm1
- DB 102,65,15,105,207 ; punpckhwd %xmm15,%xmm1
- DB 102,65,15,97,199 ; punpcklwd %xmm15,%xmm0
- DB 102,65,15,114,246,8 ; pslld $0x8,%xmm14
- DB 102,68,15,235,243 ; por %xmm3,%xmm14
- DB 102,15,114,242,8 ; pslld $0x8,%xmm2
- DB 102,65,15,235,213 ; por %xmm13,%xmm2
- DB 102,65,15,114,242,16 ; pslld $0x10,%xmm10
- DB 102,15,114,241,24 ; pslld $0x18,%xmm1
- DB 102,65,15,235,202 ; por %xmm10,%xmm1
- DB 102,65,15,235,206 ; por %xmm14,%xmm1
- DB 102,65,15,114,240,16 ; pslld $0x10,%xmm8
- DB 102,15,114,240,24 ; pslld $0x18,%xmm0
- DB 102,65,15,235,192 ; por %xmm8,%xmm0
- DB 102,15,235,194 ; por %xmm2,%xmm0
- DB 77,133,192 ; test %r8,%r8
- DB 117,101 ; jne 126c <_sk_srcover_rgba_8888_ssse3_lowp+0x274>
- DB 243,65,15,127,4,153 ; movdqu %xmm0,(%r9,%rbx,4)
- DB 243,65,15,127,76,153,16 ; movdqu %xmm1,0x10(%r9,%rbx,4)
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,4,36 ; movaps (%rsp),%xmm0
- DB 102,65,15,111,201 ; movdqa %xmm9,%xmm1
- DB 102,65,15,111,212 ; movdqa %xmm12,%xmm2
- DB 102,65,15,111,219 ; movdqa %xmm11,%xmm3
- DB 72,131,196,16 ; add $0x10,%rsp
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,15,239,255 ; pxor %xmm7,%xmm7
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,225,253,255,255 ; ja 102e <_sk_srcover_rgba_8888_ssse3_lowp+0x36>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,208,0,0,0 ; lea 0xd0(%rip),%r11 # 1328 <_sk_srcover_rgba_8888_ssse3_lowp+0x330>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 243,65,15,16,60,153 ; movss (%r9,%rbx,4),%xmm7
- DB 233,194,253,255,255 ; jmpq 102e <_sk_srcover_rgba_8888_ssse3_lowp+0x36>
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,152 ; ja 1214 <_sk_srcover_rgba_8888_ssse3_lowp+0x21c>
- DB 65,15,182,194 ; movzbl %r10b,%eax
- DB 76,141,21,189,0,0,0 ; lea 0xbd(%rip),%r10 # 1344 <_sk_srcover_rgba_8888_ssse3_lowp+0x34c>
- DB 73,99,4,130 ; movslq (%r10,%rax,4),%rax
- DB 76,1,208 ; add %r10,%rax
- DB 255,224 ; jmpq *%rax
- DB 102,65,15,126,4,153 ; movd %xmm0,(%r9,%rbx,4)
- DB 233,121,255,255,255 ; jmpq 1214 <_sk_srcover_rgba_8888_ssse3_lowp+0x21c>
- DB 102,65,15,110,100,153,8 ; movd 0x8(%r9,%rbx,4),%xmm4
- DB 102,15,112,252,69 ; pshufd $0x45,%xmm4,%xmm7
- DB 102,65,15,18,60,153 ; movlpd (%r9,%rbx,4),%xmm7
- DB 233,124,253,255,255 ; jmpq 102e <_sk_srcover_rgba_8888_ssse3_lowp+0x36>
- DB 102,65,15,110,100,153,24 ; movd 0x18(%r9,%rbx,4),%xmm4
- DB 102,68,15,112,196,69 ; pshufd $0x45,%xmm4,%xmm8
- DB 243,65,15,16,100,153,20 ; movss 0x14(%r9,%rbx,4),%xmm4
- DB 65,15,198,224,0 ; shufps $0x0,%xmm8,%xmm4
- DB 65,15,198,224,226 ; shufps $0xe2,%xmm8,%xmm4
- DB 68,15,40,196 ; movaps %xmm4,%xmm8
- DB 243,65,15,16,100,153,16 ; movss 0x10(%r9,%rbx,4),%xmm4
- DB 243,68,15,16,196 ; movss %xmm4,%xmm8
- DB 233,67,253,255,255 ; jmpq 1028 <_sk_srcover_rgba_8888_ssse3_lowp+0x30>
- DB 102,15,112,200,78 ; pshufd $0x4e,%xmm0,%xmm1
- DB 102,65,15,126,76,153,8 ; movd %xmm1,0x8(%r9,%rbx,4)
- DB 102,65,15,214,4,153 ; movq %xmm0,(%r9,%rbx,4)
- DB 233,24,255,255,255 ; jmpq 1214 <_sk_srcover_rgba_8888_ssse3_lowp+0x21c>
- DB 102,15,112,209,78 ; pshufd $0x4e,%xmm1,%xmm2
- DB 102,65,15,126,84,153,24 ; movd %xmm2,0x18(%r9,%rbx,4)
- DB 102,15,112,209,229 ; pshufd $0xe5,%xmm1,%xmm2
- DB 102,65,15,126,84,153,20 ; movd %xmm2,0x14(%r9,%rbx,4)
- DB 102,65,15,126,76,153,16 ; movd %xmm1,0x10(%r9,%rbx,4)
- DB 243,65,15,127,4,153 ; movdqu %xmm0,(%r9,%rbx,4)
- DB 233,238,254,255,255 ; jmpq 1214 <_sk_srcover_rgba_8888_ssse3_lowp+0x21c>
- DB 102,144 ; xchg %ax,%ax
- DB 57,255 ; cmp %edi,%edi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 127,255 ; jg 132d <_sk_srcover_rgba_8888_ssse3_lowp+0x335>
- DB 255 ; (bad)
- DB 255,115,255 ; pushq -0x1(%rbx)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 253 ; std
- DB 255 ; (bad)
- DB 255,172,255,255,255,151,255 ; ljmp *-0x680001(%rdi,%rdi,8)
- DB 255 ; (bad)
- DB 255,138,255,255,255,76 ; decl 0x4cffffff(%rdx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,173,255,255,255,161 ; ljmp *-0x5e000001(%rbp)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,215 ; callq *%rdi
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,208 ; callq *%rax
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,196 ; inc %esp
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 184 ; .byte 0xb8
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_scale_1_float_ssse3_lowp
-_sk_scale_1_float_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 243,68,15,16,0 ; movss (%rax),%xmm8
- DB 243,68,15,88,5,36,12,0,0 ; addss 0xc24(%rip),%xmm8 # 1f94 <_sk_xor__ssse3_lowp+0x3f7>
- DB 102,68,15,126,192 ; movd %xmm8,%eax
- DB 102,68,15,110,192 ; movd %eax,%xmm8
- DB 242,69,15,112,192,0 ; pshuflw $0x0,%xmm8,%xmm8
- DB 102,69,15,112,192,80 ; pshufd $0x50,%xmm8,%xmm8
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,56,11,216 ; pmulhrsw %xmm8,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_scale_u8_ssse3_lowp
-_sk_scale_u8_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 117,75 ; jne 141a <_sk_scale_u8_ssse3_lowp+0x64>
- DB 243,69,15,126,4,25 ; movq (%r9,%rbx,1),%xmm8
- DB 102,68,15,96,192 ; punpcklbw %xmm0,%xmm8
- DB 102,65,15,113,240,8 ; psllw $0x8,%xmm8
- DB 102,68,15,228,5,215,10,0,0 ; pmulhuw 0xad7(%rip),%xmm8 # 1ec0 <_sk_xor__ssse3_lowp+0x323>
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,56,11,216 ; pmulhrsw %xmm8,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 119,171 ; ja 13da <_sk_scale_u8_ssse3_lowp+0x24>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,126,0,0,0 ; lea 0x7e(%rip),%r11 # 14b8 <_sk_scale_u8_ssse3_lowp+0x102>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,68,15,110,192 ; movd %eax,%xmm8
- DB 235,139 ; jmp 13da <_sk_scale_u8_ssse3_lowp+0x24>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,68,15,196,192,2 ; pinsrw $0x2,%eax,%xmm8
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,68,15,110,200 ; movd %eax,%xmm9
- DB 102,68,15,96,200 ; punpcklbw %xmm0,%xmm9
- DB 243,69,15,16,193 ; movss %xmm9,%xmm8
- DB 233,97,255,255,255 ; jmpq 13da <_sk_scale_u8_ssse3_lowp+0x24>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,68,15,196,192,6 ; pinsrw $0x6,%eax,%xmm8
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,68,15,196,192,5 ; pinsrw $0x5,%eax,%xmm8
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,68,15,196,192,4 ; pinsrw $0x4,%eax,%xmm8
- DB 102,69,15,110,12,25 ; movd (%r9,%rbx,1),%xmm9
- DB 102,68,15,96,200 ; punpcklbw %xmm0,%xmm9
- DB 242,69,15,16,193 ; movsd %xmm9,%xmm8
- DB 233,35,255,255,255 ; jmpq 13da <_sk_scale_u8_ssse3_lowp+0x24>
- DB 144 ; nop
- DB 139,255 ; mov %edi,%edi
- DB 255 ; (bad)
- DB 255,168,255,255,255,151 ; ljmp *-0x68000001(%rax)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 234 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 222,255 ; fdivrp %st,%st(7)
- DB 255 ; (bad)
- DB 255,210 ; callq *%rdx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,193 ; inc %ecx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_lerp_1_float_ssse3_lowp
-_sk_lerp_1_float_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 243,68,15,16,0 ; movss (%rax),%xmm8
- DB 243,68,15,88,5,180,10,0,0 ; addss 0xab4(%rip),%xmm8 # 1f98 <_sk_xor__ssse3_lowp+0x3fb>
- DB 102,68,15,126,192 ; movd %xmm8,%eax
- DB 102,68,15,110,192 ; movd %eax,%xmm8
- DB 242,69,15,112,192,0 ; pshuflw $0x0,%xmm8,%xmm8
- DB 102,69,15,112,192,80 ; pshufd $0x50,%xmm8,%xmm8
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,68,15,56,29,200 ; pabsw %xmm0,%xmm9
- DB 102,68,15,111,21,193,9,0,0 ; movdqa 0x9c1(%rip),%xmm10 # 1ed0 <_sk_xor__ssse3_lowp+0x333>
- DB 102,69,15,249,208 ; psubw %xmm8,%xmm10
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,65,15,56,11,194 ; pmulhrsw %xmm10,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,193 ; paddw %xmm9,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,68,15,56,29,201 ; pabsw %xmm1,%xmm9
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,65,15,56,11,202 ; pmulhrsw %xmm10,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,201 ; paddw %xmm9,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,68,15,56,29,202 ; pabsw %xmm2,%xmm9
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,65,15,56,11,210 ; pmulhrsw %xmm10,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,209 ; paddw %xmm9,%xmm2
- DB 102,65,15,56,11,216 ; pmulhrsw %xmm8,%xmm3
- DB 102,68,15,56,29,195 ; pabsw %xmm3,%xmm8
- DB 102,68,15,56,11,215 ; pmulhrsw %xmm7,%xmm10
- DB 102,65,15,56,29,218 ; pabsw %xmm10,%xmm3
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_lerp_u8_ssse3_lowp
-_sk_lerp_u8_ssse3_lowp LABEL PROC
- DB 83 ; push %rbx
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 76,99,80,8 ; movslq 0x8(%rax),%r10
- DB 76,99,201 ; movslq %ecx,%r9
- DB 77,15,175,202 ; imul %r10,%r9
- DB 76,3,8 ; add (%rax),%r9
- DB 72,99,218 ; movslq %edx,%rbx
- DB 77,133,192 ; test %r8,%r8
- DB 15,133,170,0,0,0 ; jne 1650 <_sk_lerp_u8_ssse3_lowp+0xc7>
- DB 243,69,15,126,4,25 ; movq (%r9,%rbx,1),%xmm8
- DB 102,68,15,96,192 ; punpcklbw %xmm0,%xmm8
- DB 102,65,15,113,240,8 ; psllw $0x8,%xmm8
- DB 102,68,15,228,5,32,9,0,0 ; pmulhuw 0x920(%rip),%xmm8 # 1ee0 <_sk_xor__ssse3_lowp+0x343>
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,68,15,56,29,200 ; pabsw %xmm0,%xmm9
- DB 102,68,15,111,21,27,9,0,0 ; movdqa 0x91b(%rip),%xmm10 # 1ef0 <_sk_xor__ssse3_lowp+0x353>
- DB 102,69,15,249,208 ; psubw %xmm8,%xmm10
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,65,15,56,11,194 ; pmulhrsw %xmm10,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,193 ; paddw %xmm9,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,68,15,56,29,201 ; pabsw %xmm1,%xmm9
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,65,15,56,11,202 ; pmulhrsw %xmm10,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,201 ; paddw %xmm9,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,68,15,56,29,202 ; pabsw %xmm2,%xmm9
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,65,15,56,11,210 ; pmulhrsw %xmm10,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,209 ; paddw %xmm9,%xmm2
- DB 102,65,15,56,11,216 ; pmulhrsw %xmm8,%xmm3
- DB 102,68,15,56,29,195 ; pabsw %xmm3,%xmm8
- DB 102,68,15,56,11,215 ; pmulhrsw %xmm7,%xmm10
- DB 102,65,15,56,29,218 ; pabsw %xmm10,%xmm3
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 91 ; pop %rbx
- DB 255,224 ; jmpq *%rax
- DB 69,137,194 ; mov %r8d,%r10d
- DB 65,128,226,7 ; and $0x7,%r10b
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 65,254,202 ; dec %r10b
- DB 65,128,250,6 ; cmp $0x6,%r10b
- DB 15,135,72,255,255,255 ; ja 15b1 <_sk_lerp_u8_ssse3_lowp+0x28>
- DB 69,15,182,210 ; movzbl %r10b,%r10d
- DB 76,141,29,128,0,0,0 ; lea 0x80(%rip),%r11 # 16f4 <_sk_lerp_u8_ssse3_lowp+0x16b>
- DB 75,99,4,147 ; movslq (%r11,%r10,4),%rax
- DB 76,1,216 ; add %r11,%rax
- DB 255,224 ; jmpq *%rax
- DB 65,15,182,4,25 ; movzbl (%r9,%rbx,1),%eax
- DB 102,68,15,110,192 ; movd %eax,%xmm8
- DB 233,37,255,255,255 ; jmpq 15b1 <_sk_lerp_u8_ssse3_lowp+0x28>
- DB 65,15,182,68,25,2 ; movzbl 0x2(%r9,%rbx,1),%eax
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,68,15,196,192,2 ; pinsrw $0x2,%eax,%xmm8
- DB 65,15,183,4,25 ; movzwl (%r9,%rbx,1),%eax
- DB 102,68,15,110,200 ; movd %eax,%xmm9
- DB 102,68,15,96,200 ; punpcklbw %xmm0,%xmm9
- DB 243,69,15,16,193 ; movss %xmm9,%xmm8
- DB 233,251,254,255,255 ; jmpq 15b1 <_sk_lerp_u8_ssse3_lowp+0x28>
- DB 65,15,182,68,25,6 ; movzbl 0x6(%r9,%rbx,1),%eax
- DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
- DB 102,68,15,196,192,6 ; pinsrw $0x6,%eax,%xmm8
- DB 65,15,182,68,25,5 ; movzbl 0x5(%r9,%rbx,1),%eax
- DB 102,68,15,196,192,5 ; pinsrw $0x5,%eax,%xmm8
- DB 65,15,182,68,25,4 ; movzbl 0x4(%r9,%rbx,1),%eax
- DB 102,68,15,196,192,4 ; pinsrw $0x4,%eax,%xmm8
- DB 102,69,15,110,12,25 ; movd (%r9,%rbx,1),%xmm9
- DB 102,68,15,96,200 ; punpcklbw %xmm0,%xmm9
- DB 242,69,15,16,193 ; movsd %xmm9,%xmm8
- DB 233,189,254,255,255 ; jmpq 15b1 <_sk_lerp_u8_ssse3_lowp+0x28>
- DB 137,255 ; mov %edi,%edi
- DB 255 ; (bad)
- DB 255,169,255,255,255,152 ; ljmp *-0x67000001(%rcx)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 235,255 ; jmp 1701 <_sk_lerp_u8_ssse3_lowp+0x178>
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 223,255 ; (bad)
- DB 255 ; (bad)
- DB 255,211 ; callq *%rbx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,194 ; inc %edx
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; .byte 0xff
-
-PUBLIC _sk_swap_rb_ssse3_lowp
-_sk_swap_rb_ssse3_lowp LABEL PROC
- DB 68,15,40,192 ; movaps %xmm0,%xmm8
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,194 ; movaps %xmm2,%xmm0
- DB 65,15,40,208 ; movaps %xmm8,%xmm2
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_move_src_dst_ssse3_lowp
-_sk_move_src_dst_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,224 ; movaps %xmm0,%xmm4
- DB 15,40,233 ; movaps %xmm1,%xmm5
- DB 15,40,242 ; movaps %xmm2,%xmm6
- DB 15,40,251 ; movaps %xmm3,%xmm7
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_move_dst_src_ssse3_lowp
-_sk_move_dst_src_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,40,196 ; movaps %xmm4,%xmm0
- DB 15,40,205 ; movaps %xmm5,%xmm1
- DB 15,40,214 ; movaps %xmm6,%xmm2
- DB 15,40,223 ; movaps %xmm7,%xmm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_clear_ssse3_lowp
-_sk_clear_ssse3_lowp LABEL PROC
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 15,87,192 ; xorps %xmm0,%xmm0
- DB 15,87,201 ; xorps %xmm1,%xmm1
- DB 15,87,210 ; xorps %xmm2,%xmm2
- DB 15,87,219 ; xorps %xmm3,%xmm3
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcatop_ssse3_lowp
-_sk_srcatop_ssse3_lowp LABEL PROC
- DB 102,15,56,11,199 ; pmulhrsw %xmm7,%xmm0
- DB 102,68,15,56,29,192 ; pabsw %xmm0,%xmm8
- DB 102,68,15,111,13,157,7,0,0 ; movdqa 0x79d(%rip),%xmm9 # 1f00 <_sk_xor__ssse3_lowp+0x363>
- DB 102,68,15,249,203 ; psubw %xmm3,%xmm9
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,65,15,56,11,193 ; pmulhrsw %xmm9,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,192 ; paddw %xmm8,%xmm0
- DB 102,15,56,11,207 ; pmulhrsw %xmm7,%xmm1
- DB 102,68,15,56,29,193 ; pabsw %xmm1,%xmm8
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,65,15,56,11,201 ; pmulhrsw %xmm9,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,200 ; paddw %xmm8,%xmm1
- DB 102,15,56,11,215 ; pmulhrsw %xmm7,%xmm2
- DB 102,68,15,56,29,194 ; pabsw %xmm2,%xmm8
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,65,15,56,11,209 ; pmulhrsw %xmm9,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,208 ; paddw %xmm8,%xmm2
- DB 102,15,56,11,223 ; pmulhrsw %xmm7,%xmm3
- DB 102,68,15,56,29,195 ; pabsw %xmm3,%xmm8
- DB 102,68,15,56,11,207 ; pmulhrsw %xmm7,%xmm9
- DB 102,65,15,56,29,217 ; pabsw %xmm9,%xmm3
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstatop_ssse3_lowp
-_sk_dstatop_ssse3_lowp LABEL PROC
- DB 102,68,15,111,196 ; movdqa %xmm4,%xmm8
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,68,15,111,13,28,7,0,0 ; movdqa 0x71c(%rip),%xmm9 # 1f10 <_sk_xor__ssse3_lowp+0x373>
- DB 102,68,15,249,207 ; psubw %xmm7,%xmm9
- DB 102,65,15,56,11,193 ; pmulhrsw %xmm9,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,192 ; paddw %xmm8,%xmm0
- DB 102,68,15,111,197 ; movdqa %xmm5,%xmm8
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,65,15,56,11,201 ; pmulhrsw %xmm9,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,200 ; paddw %xmm8,%xmm1
- DB 102,68,15,111,198 ; movdqa %xmm6,%xmm8
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,65,15,56,11,209 ; pmulhrsw %xmm9,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,208 ; paddw %xmm8,%xmm2
- DB 102,68,15,111,199 ; movdqa %xmm7,%xmm8
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,68,15,56,11,203 ; pmulhrsw %xmm3,%xmm9
- DB 102,65,15,56,29,217 ; pabsw %xmm9,%xmm3
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcin_ssse3_lowp
-_sk_srcin_ssse3_lowp LABEL PROC
- DB 102,15,56,11,199 ; pmulhrsw %xmm7,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,56,11,207 ; pmulhrsw %xmm7,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,56,11,215 ; pmulhrsw %xmm7,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,15,56,11,223 ; pmulhrsw %xmm7,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstin_ssse3_lowp
-_sk_dstin_ssse3_lowp LABEL PROC
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,15,56,11,195 ; pmulhrsw %xmm3,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,15,56,11,203 ; pmulhrsw %xmm3,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,15,56,11,211 ; pmulhrsw %xmm3,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,15,56,11,223 ; pmulhrsw %xmm7,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcout_ssse3_lowp
-_sk_srcout_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,66,6,0,0 ; movdqa 0x642(%rip),%xmm8 # 1f20 <_sk_xor__ssse3_lowp+0x383>
- DB 102,68,15,249,199 ; psubw %xmm7,%xmm8
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,65,15,56,29,216 ; pabsw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstout_ssse3_lowp
-_sk_dstout_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,19,6,0,0 ; movdqa 0x613(%rip),%xmm8 # 1f30 <_sk_xor__ssse3_lowp+0x393>
- DB 102,68,15,249,195 ; psubw %xmm3,%xmm8
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,68,15,56,11,199 ; pmulhrsw %xmm7,%xmm8
- DB 102,65,15,56,29,216 ; pabsw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_srcover_ssse3_lowp
-_sk_srcover_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,216,5,0,0 ; movdqa 0x5d8(%rip),%xmm8 # 1f40 <_sk_xor__ssse3_lowp+0x3a3>
- DB 102,68,15,249,195 ; psubw %xmm3,%xmm8
- DB 102,68,15,111,204 ; movdqa %xmm4,%xmm9
- DB 102,69,15,56,11,200 ; pmulhrsw %xmm8,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,193 ; paddw %xmm9,%xmm0
- DB 102,68,15,111,205 ; movdqa %xmm5,%xmm9
- DB 102,69,15,56,11,200 ; pmulhrsw %xmm8,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,201 ; paddw %xmm9,%xmm1
- DB 102,68,15,111,206 ; movdqa %xmm6,%xmm9
- DB 102,69,15,56,11,200 ; pmulhrsw %xmm8,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,209 ; paddw %xmm9,%xmm2
- DB 102,68,15,56,11,199 ; pmulhrsw %xmm7,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_dstover_ssse3_lowp
-_sk_dstover_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,131,5,0,0 ; movdqa 0x583(%rip),%xmm8 # 1f50 <_sk_xor__ssse3_lowp+0x3b3>
- DB 102,68,15,249,199 ; psubw %xmm7,%xmm8
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,253,196 ; paddw %xmm4,%xmm0
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,253,205 ; paddw %xmm5,%xmm1
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,15,253,214 ; paddw %xmm6,%xmm2
- DB 102,68,15,56,11,195 ; pmulhrsw %xmm3,%xmm8
- DB 102,65,15,56,29,216 ; pabsw %xmm8,%xmm3
- DB 102,15,253,223 ; paddw %xmm7,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_modulate_ssse3_lowp
-_sk_modulate_ssse3_lowp LABEL PROC
- DB 102,15,56,11,196 ; pmulhrsw %xmm4,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,15,56,11,205 ; pmulhrsw %xmm5,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,15,56,11,214 ; pmulhrsw %xmm6,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,15,56,11,223 ; pmulhrsw %xmm7,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_multiply_ssse3_lowp
-_sk_multiply_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,24,5,0,0 ; movdqa 0x518(%rip),%xmm8 # 1f60 <_sk_xor__ssse3_lowp+0x3c3>
- DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
- DB 102,68,15,249,207 ; psubw %xmm7,%xmm9
- DB 102,68,15,111,208 ; movdqa %xmm0,%xmm10
- DB 102,69,15,56,11,209 ; pmulhrsw %xmm9,%xmm10
- DB 102,69,15,56,29,210 ; pabsw %xmm10,%xmm10
- DB 102,68,15,249,195 ; psubw %xmm3,%xmm8
- DB 102,15,56,11,196 ; pmulhrsw %xmm4,%xmm0
- DB 102,68,15,111,220 ; movdqa %xmm4,%xmm11
- DB 102,69,15,56,11,216 ; pmulhrsw %xmm8,%xmm11
- DB 102,69,15,56,29,219 ; pabsw %xmm11,%xmm11
- DB 102,69,15,253,218 ; paddw %xmm10,%xmm11
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,195 ; paddw %xmm11,%xmm0
- DB 102,68,15,111,209 ; movdqa %xmm1,%xmm10
- DB 102,69,15,56,11,209 ; pmulhrsw %xmm9,%xmm10
- DB 102,69,15,56,29,210 ; pabsw %xmm10,%xmm10
- DB 102,15,56,11,205 ; pmulhrsw %xmm5,%xmm1
- DB 102,68,15,111,221 ; movdqa %xmm5,%xmm11
- DB 102,69,15,56,11,216 ; pmulhrsw %xmm8,%xmm11
- DB 102,69,15,56,29,219 ; pabsw %xmm11,%xmm11
- DB 102,69,15,253,218 ; paddw %xmm10,%xmm11
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,203 ; paddw %xmm11,%xmm1
- DB 102,68,15,111,210 ; movdqa %xmm2,%xmm10
- DB 102,69,15,56,11,209 ; pmulhrsw %xmm9,%xmm10
- DB 102,69,15,56,29,210 ; pabsw %xmm10,%xmm10
- DB 102,15,56,11,214 ; pmulhrsw %xmm6,%xmm2
- DB 102,68,15,111,222 ; movdqa %xmm6,%xmm11
- DB 102,69,15,56,11,216 ; pmulhrsw %xmm8,%xmm11
- DB 102,69,15,56,29,219 ; pabsw %xmm11,%xmm11
- DB 102,69,15,253,218 ; paddw %xmm10,%xmm11
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,211 ; paddw %xmm11,%xmm2
- DB 102,68,15,56,11,203 ; pmulhrsw %xmm3,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,68,15,56,11,199 ; pmulhrsw %xmm7,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,69,15,253,193 ; paddw %xmm9,%xmm8
- DB 102,15,56,11,223 ; pmulhrsw %xmm7,%xmm3
- DB 102,15,56,29,219 ; pabsw %xmm3,%xmm3
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_screen_ssse3_lowp
-_sk_screen_ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,62,4,0,0 ; movdqa 0x43e(%rip),%xmm8 # 1f70 <_sk_xor__ssse3_lowp+0x3d3>
- DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
- DB 102,68,15,249,200 ; psubw %xmm0,%xmm9
- DB 102,68,15,56,11,204 ; pmulhrsw %xmm4,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,193 ; paddw %xmm9,%xmm0
- DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
- DB 102,68,15,249,201 ; psubw %xmm1,%xmm9
- DB 102,68,15,56,11,205 ; pmulhrsw %xmm5,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,201 ; paddw %xmm9,%xmm1
- DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
- DB 102,68,15,249,202 ; psubw %xmm2,%xmm9
- DB 102,68,15,56,11,206 ; pmulhrsw %xmm6,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,65,15,253,209 ; paddw %xmm9,%xmm2
- DB 102,68,15,249,195 ; psubw %xmm3,%xmm8
- DB 102,68,15,56,11,199 ; pmulhrsw %xmm7,%xmm8
- DB 102,69,15,56,29,192 ; pabsw %xmm8,%xmm8
- DB 102,65,15,253,216 ; paddw %xmm8,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-PUBLIC _sk_xor__ssse3_lowp
-_sk_xor__ssse3_lowp LABEL PROC
- DB 102,68,15,111,5,218,3,0,0 ; movdqa 0x3da(%rip),%xmm8 # 1f80 <_sk_xor__ssse3_lowp+0x3e3>
- DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
- DB 102,68,15,249,207 ; psubw %xmm7,%xmm9
- DB 102,65,15,56,11,193 ; pmulhrsw %xmm9,%xmm0
- DB 102,68,15,56,29,208 ; pabsw %xmm0,%xmm10
- DB 102,68,15,249,195 ; psubw %xmm3,%xmm8
- DB 102,15,111,196 ; movdqa %xmm4,%xmm0
- DB 102,65,15,56,11,192 ; pmulhrsw %xmm8,%xmm0
- DB 102,15,56,29,192 ; pabsw %xmm0,%xmm0
- DB 102,65,15,253,194 ; paddw %xmm10,%xmm0
- DB 102,65,15,56,11,201 ; pmulhrsw %xmm9,%xmm1
- DB 102,68,15,56,29,209 ; pabsw %xmm1,%xmm10
- DB 102,15,111,205 ; movdqa %xmm5,%xmm1
- DB 102,65,15,56,11,200 ; pmulhrsw %xmm8,%xmm1
- DB 102,15,56,29,201 ; pabsw %xmm1,%xmm1
- DB 102,65,15,253,202 ; paddw %xmm10,%xmm1
- DB 102,65,15,56,11,209 ; pmulhrsw %xmm9,%xmm2
- DB 102,68,15,56,29,210 ; pabsw %xmm2,%xmm10
- DB 102,15,111,214 ; movdqa %xmm6,%xmm2
- DB 102,65,15,56,11,208 ; pmulhrsw %xmm8,%xmm2
- DB 102,15,56,29,210 ; pabsw %xmm2,%xmm2
- DB 102,65,15,253,210 ; paddw %xmm10,%xmm2
- DB 102,68,15,56,11,203 ; pmulhrsw %xmm3,%xmm9
- DB 102,69,15,56,29,201 ; pabsw %xmm9,%xmm9
- DB 102,68,15,56,11,199 ; pmulhrsw %xmm7,%xmm8
- DB 102,65,15,56,29,216 ; pabsw %xmm8,%xmm3
- DB 102,65,15,253,217 ; paddw %xmm9,%xmm3
- DB 72,173 ; lods %ds:(%rsi),%rax
- DB 255,224 ; jmpq *%rax
-
-ALIGN 16
- DB 0,0 ; add %al,(%rax)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,128 ; addb $0x80,0x0(%rbx)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 12,13 ; or $0xd,%al
- DB 14 ; (bad)
- DB 15,129,128,129,128,129 ; jno ffffffff81809e05 <_sk_xor__ssse3_lowp+0xffffffff81808268>
- DB 128,129,128,129,128,129,128 ; addb $0x80,-0x7e7f7e80(%rcx)
- DB 129,128,129,128,1,2,5,6,9,10 ; addl $0xa090605,0x2018081(%rax)
- DB 13,14,9,10,13 ; or $0xd0a090e,%eax
- DB 14 ; (bad)
- DB 13,14,15,255,2 ; or $0x2ff0f0e,%eax
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,10 ; (bad)
- DB 11,14 ; or (%rsi),%ecx
- DB 15,14 ; femms
- DB 15,255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 1,4,5,8,9,12,13 ; add %eax,0xd0c0908(,%rax,1)
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 12,13 ; or $0xd,%al
- DB 14 ; (bad)
- DB 15,129,128,129,128,129 ; jno ffffffff81809e55 <_sk_xor__ssse3_lowp+0xffffffff818082b8>
- DB 128,129,128,129,128,129,128 ; addb $0x80,-0x7e7f7e80(%rcx)
- DB 129,128,129,128,1,2,5,6,9,10 ; addl $0xa090605,0x2018081(%rax)
- DB 13,14,9,10,13 ; or $0xd0a090e,%eax
- DB 14 ; (bad)
- DB 13,14,15,255,2 ; or $0x2ff0f0e,%eax
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,10 ; (bad)
- DB 11,14 ; or (%rsi),%ecx
- DB 15,14 ; femms
- DB 15,255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 12,13 ; or $0xd,%al
- DB 14 ; (bad)
- DB 15,129,128,129,128,129 ; jno ffffffff81809eb5 <_sk_xor__ssse3_lowp+0xffffffff81808318>
- DB 128,129,128,129,128,129,128 ; addb $0x80,-0x7e7f7e80(%rcx)
- DB 129,128,129,128,1,2,5,6,9,10 ; addl $0xa090605,0x2018081(%rax)
- DB 13,14,9,10,13 ; or $0xd0a090e,%eax
- DB 14 ; (bad)
- DB 13,14,15,255,2 ; or $0x2ff0f0e,%eax
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,10 ; (bad)
- DB 11,14 ; or (%rsi),%ecx
- DB 15,14 ; femms
- DB 15,255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 1,4,5,8,9,12,13 ; add %eax,0xd0c0908(,%rax,1)
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 12,13 ; or $0xd,%al
- DB 14 ; (bad)
- DB 15,129,128,129,128,129 ; jno ffffffff81809f05 <_sk_xor__ssse3_lowp+0xffffffff81808368>
- DB 128,129,128,129,128,129,128 ; addb $0x80,-0x7e7f7e80(%rcx)
- DB 129,128,129,128,1,2,5,6,9,10 ; addl $0xa090605,0x2018081(%rax)
- DB 13,14,9,10,13 ; or $0xd0a090e,%eax
- DB 14 ; (bad)
- DB 13,14,15,255,2 ; or $0x2ff0f0e,%eax
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,10 ; (bad)
- DB 11,14 ; or (%rsi),%ecx
- DB 15,14 ; femms
- DB 15,255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 255,0 ; incl (%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,0,2,4,6,0,0,0,0 ; addl $0x0,0x6040200(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,2 ; add %al,(%rdx)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 0,0 ; add %al,(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,0,128,0,128 ; addl $0x80008000,-0x7f7e7f7f(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,0,128,0,128 ; addl $0x80008000,-0x7f7e7f7f(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,1 ; add %al,(%rcx)
- DB 4,5 ; add $0x5,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 8,9 ; or %cl,(%rcx)
- DB 12,13 ; or $0xd,%al
- DB 12,13 ; or $0xd,%al
- DB 14 ; (bad)
- DB 15,255 ; (bad)
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,255 ; add %bh,%bh
- DB 0,129,128,129,128,129 ; add %al,-0x7e7f7e80(%rcx)
- DB 128,129,128,129,128,129,128 ; addb $0x80,-0x7e7f7e80(%rcx)
- DB 129,128,129,128,1,2,5,6,9,10 ; addl $0xa090605,0x2018081(%rax)
- DB 13,14,9,10,13 ; or $0xd0a090e,%eax
- DB 14 ; (bad)
- DB 13,14,15,255,2 ; or $0x2ff0f0e,%eax
- DB 3,6 ; add (%rsi),%eax
- DB 7 ; (bad)
- DB 10,11 ; or (%rbx),%cl
- DB 14 ; (bad)
- DB 15,10 ; (bad)
- DB 11,14 ; or (%rsi),%ecx
- DB 15,14 ; femms
- DB 15,255 ; (bad)
- DB 255,3 ; incl (%rbx)
- DB 255,7 ; incl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,11 ; decl (%rbx)
- DB 255,15 ; decl (%rdi)
- DB 255,15 ; decl (%rdi)
- DB 255 ; (bad)
- DB 255 ; (bad)
- DB 255,0 ; incl (%rax)
- DB 128,0,128 ; addb $0x80,(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,0,128,0,128 ; addl $0x80008000,-0x7f7e7f7f(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 129,128,129,128,129,128,129,128,129,128; addl $0x80818081,-0x7f7e7f7f(%rax)
- DB 129,128,129,128,129,128,0,128,0,128 ; addl $0x80008000,-0x7f7e7f7f(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
- DB 0,128,0,128,0,128 ; add %al,-0x7fff8000(%rax)
-
-ALIGN 4
- DB 0,0 ; add %al,(%rax)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128,67,0,0 ; addb $0x0,0x0(%rbx)
- DB 128 ; .byte 0x80
- DB 67 ; rex.XB
-ALIGN 32
-
PUBLIC _sk_start_pipeline_hsw_8bit
_sk_start_pipeline_hsw_8bit LABEL PROC
DB 85 ; push %rbp
diff --git a/src/jumper/SkJumper_stages_8bit.cpp b/src/jumper/SkJumper_stages_8bit.cpp
index 57701c32e4..6f683f132d 100644
--- a/src/jumper/SkJumper_stages_8bit.cpp
+++ b/src/jumper/SkJumper_stages_8bit.cpp
@@ -12,10 +12,9 @@
#include <immintrin.h>
#endif
-// We're going to try going even lower precision than _lowp.cpp,
-// 8-bit per channel, and while we're at it keep our pixels interlaced.
-// This is the natural format for kN32_SkColorType buffers, and we hope
-// the stages in this file can replace many custom legacy routines.
+// This restricted SkJumper backend works on 8-bit per channel interlaced
+// pixels. This is the natural format for kN32_SkColorType buffers, and we
+// hope the stages in this file can replace many custom legacy routines.
#if !defined(JUMPER)
#error "This file must be pre-compiled."
diff --git a/src/jumper/SkJumper_stages_lowp.cpp b/src/jumper/SkJumper_stages_lowp.cpp
deleted file mode 100644
index f90ebdb5d3..0000000000
--- a/src/jumper/SkJumper_stages_lowp.cpp
+++ /dev/null
@@ -1,405 +0,0 @@
-/*
- * Copyright 2017 Google Inc.
- *
- * Use of this source code is governed by a BSD-style license that can be
- * found in the LICENSE file.
- */
-
-#include "SkJumper.h"
-#include "SkJumper_misc.h"
-#include <immintrin.h>
-
-#if !defined(__clang__) || !defined(__x86_64__)
- #error "We're starting with just x86-64 for now, and will always require Clang."
-#endif
-
-using K = const SkJumper_constants;
-
-#if defined(__AVX2__)
- #define WRAP(name) sk_##name##_hsw_lowp
- template <typename T> using V = T __attribute__((ext_vector_type(16)));
- static const size_t kStride = 16;
-#else
- #define WRAP(name) sk_##name##_ssse3_lowp
- template <typename T> using V = T __attribute__((ext_vector_type(8)));
- static const size_t kStride = 8;
-#endif
-
-using U8 = V<uint8_t>;
-using U16 = V<uint16_t>;
-using U32 = V<uint32_t>;
-
-// See SkFixed15.h for details on this format and its operations.
-struct F {
- U16 vec;
-
- F() = default;
- F(float f) {
- // After adding 256.0f, the SkFixed15 value is the bottom two bytes of the float.
- f += 256.0f;
- vec = unaligned_load<uint16_t>(&f);
- }
-
- F(U16 v) : vec(v) {}
- operator U16() const { return vec; }
-};
-
-SI F operator+(F x, F y) { return x.vec + y.vec; }
-SI F operator-(F x, F y) { return x.vec - y.vec; }
-SI F operator*(F x, F y) {
-#if defined(__AVX2__)
- return _mm256_abs_epi16(_mm256_mulhrs_epi16(x.vec, y.vec));
-#else
- return _mm_abs_epi16(_mm_mulhrs_epi16(x.vec, y.vec));
-#endif
-}
-
-SI F mad(F f, F m, F a) { return f*m+a; }
-SI F inv(F v) { return 1.0f - v; }
-SI F two(F v) { return v + v; }
-SI F lerp(F from, F to, F t) { return to*t + from*inv(t); }
-
-SI F operator<<(F x, int bits) { return x.vec << bits; }
-SI F operator>>(F x, int bits) { return x.vec >> bits; }
-
-using Stage = void(K* k, void** program, size_t x, size_t y, size_t tail, F,F,F,F, F,F,F,F);
-
-#if defined(__AVX__)
- // We really want to make sure all paths go through this function's (implicit) vzeroupper.
- // If they don't, we'll experience severe slowdowns when we first use SSE instructions again.
- __attribute__((disable_tail_calls))
-#endif
-MAYBE_MSABI
-extern "C" void WRAP(start_pipeline)(size_t x, size_t y, size_t xlimit, size_t ylimit,
- void** program, K* k) {
- F v;
- auto start = (Stage*)load_and_inc(program);
- size_t x0 = x;
- for (; y < ylimit; y++) {
- x = x0;
- while (x + kStride <= xlimit) {
- start(k,program,x,y,0, v,v,v,v, v,v,v,v);
- x += kStride;
- }
- if (size_t tail = xlimit - x) {
- start(k,program,x,y,tail, v,v,v,v, v,v,v,v);
- }
- }
-}
-
-extern "C" void WRAP(just_return)(K*, void**, size_t,size_t,size_t, F,F,F,F, F,F,F,F) {}
-
-#define STAGE(name) \
- SI void name##_k(K* k, LazyCtx ctx, size_t x, size_t y, size_t tail, \
- F& r, F& g, F& b, F& a, F& dr, F& dg, F& db, F& da); \
- extern "C" void WRAP(name)(K* k, void** program, size_t x, size_t y, size_t tail, \
- F r, F g, F b, F a, F dr, F dg, F db, F da) { \
- LazyCtx ctx(program); \
- name##_k(k,ctx,x,y,tail, r,g,b,a, dr,dg,db,da); \
- auto next = (Stage*)load_and_inc(program); \
- next(k,program,x,y,tail, r,g,b,a, dr,dg,db,da); \
- } \
- SI void name##_k(K* k, LazyCtx ctx, size_t x, size_t y, size_t tail, \
- F& r, F& g, F& b, F& a, F& dr, F& dg, F& db, F& da)
-
-
-// Helper functions used by multiple stages.
-
-template <typename V, typename T>
-SI V load(const T* src, size_t tail) {
- __builtin_assume(tail < kStride);
- if (__builtin_expect(tail, 0)) {
- V v{}; // Any inactive lanes are zeroed.
- switch (tail) {
- case 15: v[14] = src[14];
- case 14: v[13] = src[13];
- case 13: v[12] = src[12];
- case 12: memcpy(&v, src, 12*sizeof(T)); break;
- case 11: v[10] = src[10];
- case 10: v[ 9] = src[ 9];
- case 9: v[ 8] = src[ 8];
- case 8: memcpy(&v, src, 8*sizeof(T)); break;
- case 7: v[6] = src[6];
- case 6: v[5] = src[5];
- case 5: v[4] = src[4];
- case 4: memcpy(&v, src, 4*sizeof(T)); break;
- case 3: v[2] = src[2];
- case 2: memcpy(&v, src, 2*sizeof(T)); break;
- case 1: memcpy(&v, src, 1*sizeof(T)); break;
- }
- return v;
- }
- return unaligned_load<V>(src);
-}
-
-template <typename V, typename T>
-SI void store(T* dst, V v, size_t tail) {
- __builtin_assume(tail < kStride);
- if (__builtin_expect(tail, 0)) {
- switch (tail) {
- case 15: dst[14] = v[14];
- case 14: dst[13] = v[13];
- case 13: dst[12] = v[12];
- case 12: memcpy(dst, &v, 12*sizeof(T)); break;
- case 11: dst[10] = v[10];
- case 10: dst[ 9] = v[ 9];
- case 9: dst[ 8] = v[ 8];
- case 8: memcpy(dst, &v, 8*sizeof(T)); break;
- case 7: dst[6] = v[6];
- case 6: dst[5] = v[5];
- case 5: dst[4] = v[4];
- case 4: memcpy(dst, &v, 4*sizeof(T)); break;
- case 3: dst[2] = v[2];
- case 2: memcpy(dst, &v, 2*sizeof(T)); break;
- case 1: memcpy(dst, &v, 1*sizeof(T)); break;
- }
- return;
- }
- unaligned_store(dst, v);
-}
-
-// TODO: mask loads and stores with AVX2
-
-// Scale from [0,255] up to [0,32768].
-SI F from_wide_byte(U16 bytes) {
- // Ideally we'd scale by 32768/255 = 128.50196, but instead we'll approximate
- // that a little more cheaply as 256*32897/65536 = 128.50391.
- // 0 and 255 map to 0 and 32768 correctly, and nothing else is off by more than 1 bit.
-#if defined(__AVX2__)
- return _mm256_mulhi_epu16(bytes << 8, U16(32897));
-#else
- return _mm_mulhi_epu16(bytes << 8, U16(32897));
-#endif
-}
-SI F from_byte(U8 bytes) {
- return from_wide_byte(__builtin_convertvector(bytes, U16));
-}
-
-// Pack from [0,32768] down to [0,255].
-SI U16 to_wide_byte(F v) {
- // The simplest thing works great: divide by 128 and saturate.
-#if defined(__AVX2__)
- return _mm256_min_epi16(v >> 7, U16(255));
-#else
- return _mm_min_epi16(v >> 7, U16(255));
-#endif
-}
-SI U8 to_byte(F v) {
- // Like to_wide_byte(), but we'll bake the saturation into the 16->8 bit pack.
-#if defined(__AVX2__)
- return _mm_packus_epi16(_mm256_extracti128_si256(v >> 7, 0),
- _mm256_extracti128_si256(v >> 7, 1));
-#else
- // Only the bottom 8 bytes are of interest... it doesn't matter what we pack on top.
- __m128i packed = _mm_packus_epi16(v >> 7, v >> 7);
- return unaligned_load<U8>(&packed);
-#endif
-}
-
-SI void from_8888(U32 rgba, F* r, F* g, F* b, F* a) {
- *r = from_wide_byte(__builtin_convertvector((rgba >> 0) & 0xff, U16));
- *g = from_wide_byte(__builtin_convertvector((rgba >> 8) & 0xff, U16));
- *b = from_wide_byte(__builtin_convertvector((rgba >> 16) & 0xff, U16));
- *a = from_wide_byte(__builtin_convertvector((rgba >> 24) & 0xff, U16));
-}
-
-SI U32 to_8888(F r, F g, F b, F a) {
- return __builtin_convertvector(to_wide_byte(r), U32) << 0
- | __builtin_convertvector(to_wide_byte(g), U32) << 8
- | __builtin_convertvector(to_wide_byte(b), U32) << 16
- | __builtin_convertvector(to_wide_byte(a), U32) << 24;
-}
-
-template <typename T>
-SI T* ptr_at_xy(const SkJumper_MemoryCtx* ctx, int x, int y) {
- return (T*)ctx->pixels + y*ctx->stride + x;
-}
-
-// Stages!
-
-STAGE(uniform_color) {
- auto c = (const SkJumper_UniformColorCtx*)ctx;
- // We're converting to fixed point, which lets us play some IEEE representation tricks,
- // replacing a naive *32768 and float->int conversion with a simple float add.
- using F32x4 = float __attribute__((ext_vector_type(4)));
- using U16x8 = uint16_t __attribute__((ext_vector_type(8)));
- auto bits = (U16x8)(unaligned_load<F32x4>((const float*)&c->r) + 256.0f);
- r = (U16)bits[0];
- g = (U16)bits[2];
- b = (U16)bits[4];
- a = (U16)bits[6];
-}
-
-STAGE(black_color) {
- r = g = b = 0.0f;
- a = 1.0f;
-}
-
-STAGE(white_color) {
- r = g = b = a = 1.0f;
-}
-
-STAGE(set_rgb) {
- auto rgb = (const float*)ctx;
- r = rgb[0];
- g = rgb[1];
- b = rgb[2];
-}
-
-STAGE(premul) {
- r = r * a;
- g = g * a;
- b = b * a;
-}
-
-STAGE(load_8888) {
- auto ptr = ptr_at_xy<const uint32_t>(ctx, x,y);
- from_8888(load<U32>(ptr, tail), &r,&g,&b,&a);
-}
-STAGE(load_8888_dst) {
- auto ptr = ptr_at_xy<const uint32_t>(ctx, x,y);
- from_8888(load<U32>(ptr, tail), &dr,&dg,&db,&da);
-}
-STAGE(store_8888) {
- auto ptr = ptr_at_xy<uint32_t>(ctx, x,y);
- store(ptr, to_8888(r,g,b,a), tail);
-}
-
-STAGE(load_bgra) {
- auto ptr = ptr_at_xy<const uint32_t>(ctx, x,y);
- from_8888(load<U32>(ptr, tail), &b,&g,&r,&a);
-}
-STAGE(load_bgra_dst) {
- auto ptr = ptr_at_xy<const uint32_t>(ctx, x,y);
- from_8888(load<U32>(ptr, tail), &db,&dg,&dr,&da);
-}
-STAGE(store_bgra) {
- auto ptr = ptr_at_xy<uint32_t>(ctx, x,y);
- store(ptr, to_8888(b,g,r,a), tail);
-}
-
-STAGE(load_a8) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
- r = g = b = 0.0f;
- a = from_byte(load<U8>(ptr, tail));
-}
-STAGE(load_a8_dst) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
- dr = dg = db = 0.0f;
- da = from_byte(load<U8>(ptr, tail));
-}
-STAGE(store_a8) {
- auto ptr = ptr_at_xy<uint8_t>(ctx, x,y);
- store(ptr, to_byte(a), tail);
-}
-
-STAGE(load_g8) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
- r = g = b = from_byte(load<U8>(ptr, tail));
- a = 1.0f;
-}
-
-STAGE(load_g8_dst) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
- dr = dg = db = from_byte(load<U8>(ptr, tail));
- da = 1.0f;
-}
-
-STAGE(srcover_rgba_8888) {
- auto ptr = ptr_at_xy<uint32_t>(ctx, x,y);
-
- from_8888(load<U32>(ptr, tail), &dr,&dg,&db,&da);
-
- r = mad(dr, inv(a), r);
- g = mad(dg, inv(a), g);
- b = mad(db, inv(a), b);
- a = mad(da, inv(a), a);
-
- store(ptr, to_8888(r,g,b,a), tail);
-}
-
-STAGE(scale_1_float) {
- float c = *(const float*)ctx;
-
- r = r * c;
- g = g * c;
- b = b * c;
- a = a * c;
-}
-STAGE(scale_u8) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
-
- U8 scales = load<U8>(ptr, tail);
- F c = from_byte(scales);
-
- r = r * c;
- g = g * c;
- b = b * c;
- a = a * c;
-}
-
-STAGE(lerp_1_float) {
- float c = *(const float*)ctx;
-
- r = lerp(dr, r, c);
- g = lerp(dg, g, c);
- b = lerp(db, b, c);
- a = lerp(da, a, c);
-}
-STAGE(lerp_u8) {
- auto ptr = ptr_at_xy<const uint8_t>(ctx, x,y);
-
- U8 scales = load<U8>(ptr, tail);
- F c = from_byte(scales);
-
- r = lerp(dr, r, c);
- g = lerp(dg, g, c);
- b = lerp(db, b, c);
- a = lerp(da, a, c);
-}
-
-STAGE(swap_rb) {
- auto tmp = r;
- r = b;
- b = tmp;
-}
-STAGE(move_src_dst) {
- dr = r;
- dg = g;
- db = b;
- da = a;
-}
-STAGE(move_dst_src) {
- r = dr;
- g = dg;
- b = db;
- a = da;
-}
-
-// Most blend modes apply the same logic to each channel.
-#define BLEND_MODE(name) \
- SI F name##_channel(F s, F d, F sa, F da); \
- STAGE(name) { \
- r = name##_channel(r,dr,a,da); \
- g = name##_channel(g,dg,a,da); \
- b = name##_channel(b,db,a,da); \
- a = name##_channel(a,da,a,da); \
- } \
- SI F name##_channel(F s, F d, F sa, F da)
-
-BLEND_MODE(clear) { return 0.0f; }
-BLEND_MODE(srcatop) { return s*da + d*inv(sa); }
-BLEND_MODE(dstatop) { return d*sa + s*inv(da); }
-BLEND_MODE(srcin) { return s * da; }
-BLEND_MODE(dstin) { return d * sa; }
-BLEND_MODE(srcout) { return s * inv(da); }
-BLEND_MODE(dstout) { return d * inv(sa); }
-BLEND_MODE(srcover) { return mad(d, inv(sa), s); }
-BLEND_MODE(dstover) { return mad(s, inv(da), d); }
-
-BLEND_MODE(modulate) { return s*d; }
-BLEND_MODE(multiply) { return s*inv(da) + d*inv(sa) + s*d; }
-BLEND_MODE(screen) { return s + inv(s)*d; }
-BLEND_MODE(xor_) { return s*inv(da) + d*inv(sa); }
-
-#undef BLEND_MODE
diff --git a/src/jumper/build_stages.py b/src/jumper/build_stages.py
index bf724ce0ac..3c54e8e894 100755
--- a/src/jumper/build_stages.py
+++ b/src/jumper/build_stages.py
@@ -13,7 +13,6 @@ clang = 'clang-4.0'
objdump = 'gobjdump'
ccache = 'ccache'
stages = 'src/jumper/SkJumper_stages.cpp'
-stages_lowp = 'src/jumper/SkJumper_stages_lowp.cpp'
stages_8bit = 'src/jumper/SkJumper_stages_8bit.cpp'
generated = 'src/jumper/SkJumper_generated.S'
generated_win = 'src/jumper/SkJumper_generated_win.S'
@@ -22,10 +21,9 @@ clang = sys.argv[1] if len(sys.argv) > 1 else clang
objdump = sys.argv[2] if len(sys.argv) > 2 else objdump
ccache = sys.argv[3] if len(sys.argv) > 3 else ccache
stages = sys.argv[4] if len(sys.argv) > 4 else stages
-stages_lowp = sys.argv[5] if len(sys.argv) > 5 else stages_lowp
-stages_8bit = sys.argv[6] if len(sys.argv) > 6 else stages_8bit
-generated = sys.argv[7] if len(sys.argv) > 7 else generated
-generated_win = sys.argv[8] if len(sys.argv) > 8 else generated_win
+stages_8bit = sys.argv[5] if len(sys.argv) > 5 else stages_8bit
+generated = sys.argv[6] if len(sys.argv) > 6 else generated
+generated_win = sys.argv[7] if len(sys.argv) > 7 else generated_win
clang = [ccache, clang, '-x', 'c++']
@@ -57,14 +55,6 @@ subprocess.check_call(clang + cflags + sse2 + win +
['-c', stages_8bit] +
['-o', 'win_8bit_sse2.o'])
-ssse3 = ['-mssse3', '-mno-sse4.1']
-subprocess.check_call(clang + cflags + ssse3 +
- ['-c', stages_lowp] +
- ['-o', 'lowp_ssse3.o'])
-subprocess.check_call(clang + cflags + ssse3 + win +
- ['-c', stages_lowp] +
- ['-o', 'win_lowp_ssse3.o'])
-
sse41 = ['-msse4.1']
subprocess.check_call(clang + cflags + sse41 +
['-c', stages] +
@@ -95,12 +85,6 @@ subprocess.check_call(clang + cflags + hsw +
subprocess.check_call(clang + cflags + hsw + win +
['-c', stages] +
['-o', 'win_hsw.o'])
-subprocess.check_call(clang + cflags + hsw +
- ['-c', stages_lowp] +
- ['-o', 'lowp_hsw.o'])
-subprocess.check_call(clang + cflags + hsw + win +
- ['-c', stages_lowp] +
- ['-o', 'win_lowp_hsw.o'])
subprocess.check_call(clang + cflags + hsw +
['-c', stages_8bit] +
@@ -259,10 +243,6 @@ parse_object_file('sse41.o', '.byte')
print 'BALIGN32'
parse_object_file('sse2.o', '.byte')
print 'BALIGN32'
-parse_object_file('lowp_hsw.o', '.byte')
-print 'BALIGN32'
-parse_object_file('lowp_ssse3.o', '.byte')
-print 'BALIGN32'
parse_object_file('8bit_hsw.o', '.byte')
print 'BALIGN32'
parse_object_file('8bit_sse41.o', '.byte')
@@ -295,10 +275,6 @@ parse_object_file('win_sse41.o', 'DB')
print 'ALIGN 32'
parse_object_file('win_sse2.o', 'DB')
print 'ALIGN 32'
-parse_object_file('win_lowp_hsw.o', 'DB')
-print 'ALIGN 32'
-parse_object_file('win_lowp_ssse3.o', 'DB')
-print 'ALIGN 32'
parse_object_file('win_8bit_hsw.o', 'DB')
print 'ALIGN 32'
parse_object_file('win_8bit_sse41.o', 'DB')