aboutsummaryrefslogtreecommitdiffhomepage
path: root/src/jumper
diff options
context:
space:
mode:
authorGravatar Mike Klein <mtklein@chromium.org>2017-03-29 16:33:06 -0400
committerGravatar Skia Commit-Bot <skia-commit-bot@chromium.org>2017-03-29 21:22:11 +0000
commitd7e06aec7e7837ca1dd54680ad37b431e40e4eef (patch)
treeffe17ff5709c7acb4759c3b131de35a2096ea5f6 /src/jumper
parent291932e8e42e0556bfef9118231b4ac364ecf5df (diff)
jumper, revert to generating .S files
I went with the unified-in-one-.cpp approach mostly to make it easy to roll out SkJumper. I no longer see any difficultly rolling out the assembly files, and it's possible the unified .cpp approach just makes things harder. Let's see if it's any easier to get Chrome's official build to work with normal assembly files. It's not going to be a problem to roll out. This is a partial revert of https://skia-review.googlesource.com/c/9336. CQ_INCLUDE_TRYBOTS=skia.primary:Test-Win2k8-MSVC-GCE-CPU-AVX2-x86_64-Debug,Test-Mac-Clang-MacMini6.2-CPU-AVX-x86_64-Debug,Test-Ubuntu-Clang-GCE-CPU-AVX2-x86_64-Debug,Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Debug Change-Id: Idfdbd2d322452b44bc0adaf6dc299cc7649bc51e Reviewed-on: https://skia-review.googlesource.com/10561 Reviewed-by: Herb Derby <herb@google.com> Commit-Queue: Mike Klein <mtklein@chromium.org>
Diffstat (limited to 'src/jumper')
-rw-r--r--src/jumper/SkJumper.cpp8
-rw-r--r--src/jumper/SkJumper_generated.S7498
-rw-r--r--src/jumper/SkJumper_generated.cpp13052
-rw-r--r--src/jumper/SkJumper_generated_win.S5553
-rwxr-xr-xsrc/jumper/build_stages.py99
5 files changed, 13107 insertions, 13103 deletions
diff --git a/src/jumper/SkJumper.cpp b/src/jumper/SkJumper.cpp
index 85f1231b1d..0f895d4b96 100644
--- a/src/jumper/SkJumper.cpp
+++ b/src/jumper/SkJumper.cpp
@@ -91,8 +91,12 @@ static K kConstants = {
// We'll only ever call start_pipeline(), which then chains into the rest for us.
using StageFn = void(void);
-// TODO: maybe don't need this wrapper anymore.
-#define ASM(name, suffix) sk_##name##_##suffix
+// Some platforms expect C "name" maps to asm "_name", others to "name".
+#if defined(__APPLE__)
+ #define ASM(name, suffix) sk_##name##_##suffix
+#else
+ #define ASM(name, suffix) _sk_##name##_##suffix
+#endif
extern "C" {
diff --git a/src/jumper/SkJumper_generated.S b/src/jumper/SkJumper_generated.S
new file mode 100644
index 0000000000..0cf1e26853
--- /dev/null
+++ b/src/jumper/SkJumper_generated.S
@@ -0,0 +1,7498 @@
+# Copyright 2017 Google Inc.
+#
+# Use of this source code is governed by a BSD-style license that can be
+# found in the LICENSE file.
+
+# This file is generated semi-automatically with this command:
+# $ src/jumper/build_stages.py
+
+.text
+#if defined(__aarch64__)
+.balign 4
+
+.globl _sk_start_pipeline_aarch64
+_sk_start_pipeline_aarch64:
+ .long 0xa9bd5bf7 // stp x23, x22, [sp, #-48]!
+ .long 0xa90153f5 // stp x21, x20, [sp, #16]
+ .long 0xa9027bf3 // stp x19, x30, [sp, #32]
+ .long 0xaa0103f4 // mov x20, x1
+ .long 0xf8408697 // ldr x23, [x20], #8
+ .long 0xaa0003f5 // mov x21, x0
+ .long 0xaa0303f3 // mov x19, x3
+ .long 0x910012a8 // add x8, x21, #0x4
+ .long 0xeb13011f // cmp x8, x19
+ .long 0xaa0203f6 // mov x22, x2
+ .long 0x54000069 // b.ls 34 <sk_start_pipeline_aarch64+0x34> // b.plast
+ .long 0xaa1503e0 // mov x0, x21
+ .long 0x14000012 // b 78 <sk_start_pipeline_aarch64+0x78>
+ .long 0x6f00e400 // movi v0.2d, #0x0
+ .long 0x6f00e401 // movi v1.2d, #0x0
+ .long 0x6f00e402 // movi v2.2d, #0x0
+ .long 0x6f00e403 // movi v3.2d, #0x0
+ .long 0x6f00e404 // movi v4.2d, #0x0
+ .long 0x6f00e405 // movi v5.2d, #0x0
+ .long 0x6f00e406 // movi v6.2d, #0x0
+ .long 0x6f00e407 // movi v7.2d, #0x0
+ .long 0xaa1503e0 // mov x0, x21
+ .long 0xaa1403e1 // mov x1, x20
+ .long 0xaa1603e2 // mov x2, x22
+ .long 0xd63f02e0 // blr x23
+ .long 0x910012a0 // add x0, x21, #0x4
+ .long 0x910022a8 // add x8, x21, #0x8
+ .long 0xeb13011f // cmp x8, x19
+ .long 0xaa0003f5 // mov x21, x0
+ .long 0x54fffe09 // b.ls 34 <sk_start_pipeline_aarch64+0x34> // b.plast
+ .long 0xa9427bf3 // ldp x19, x30, [sp, #32]
+ .long 0xa94153f5 // ldp x21, x20, [sp, #16]
+ .long 0xa8c35bf7 // ldp x23, x22, [sp], #48
+ .long 0xd65f03c0 // ret
+
+.globl _sk_just_return_aarch64
+_sk_just_return_aarch64:
+ .long 0xd65f03c0 // ret
+
+.globl _sk_seed_shader_aarch64
+_sk_seed_shader_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x3dc00046 // ldr q6, [x2]
+ .long 0x4e040c00 // dup v0.4s, w0
+ .long 0x4f0167e7 // movi v7.4s, #0x3f, lsl #24
+ .long 0x4d40c901 // ld1r {v1.4s}, [x8]
+ .long 0x4e21d800 // scvtf v0.4s, v0.4s
+ .long 0x4e27d400 // fadd v0.4s, v0.4s, v7.4s
+ .long 0x4f03f602 // fmov v2.4s, #1.000000000000000000e+00
+ .long 0x4e21d821 // scvtf v1.4s, v1.4s
+ .long 0x6f00e403 // movi v3.2d, #0x0
+ .long 0x6f00e404 // movi v4.2d, #0x0
+ .long 0x6f00e405 // movi v5.2d, #0x0
+ .long 0x4e26d400 // fadd v0.4s, v0.4s, v6.4s
+ .long 0x6f00e406 // movi v6.2d, #0x0
+ .long 0x4e27d421 // fadd v1.4s, v1.4s, v7.4s
+ .long 0x6f00e407 // movi v7.2d, #0x0
+ .long 0xd61f0060 // br x3
+
+.globl _sk_constant_color_aarch64
+_sk_constant_color_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x3dc00103 // ldr q3, [x8]
+ .long 0x4e040460 // dup v0.4s, v3.s[0]
+ .long 0x4e0c0461 // dup v1.4s, v3.s[1]
+ .long 0x4e140462 // dup v2.4s, v3.s[2]
+ .long 0x4e1c0463 // dup v3.4s, v3.s[3]
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clear_aarch64
+_sk_clear_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x6f00e400 // movi v0.2d, #0x0
+ .long 0x6f00e401 // movi v1.2d, #0x0
+ .long 0x6f00e402 // movi v2.2d, #0x0
+ .long 0x6f00e403 // movi v3.2d, #0x0
+ .long 0xd61f0060 // br x3
+
+.globl _sk_plus__aarch64
+_sk_plus__aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4e24d400 // fadd v0.4s, v0.4s, v4.4s
+ .long 0x4e25d421 // fadd v1.4s, v1.4s, v5.4s
+ .long 0x4e26d442 // fadd v2.4s, v2.4s, v6.4s
+ .long 0x4e27d463 // fadd v3.4s, v3.4s, v7.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_srcover_aarch64
+_sk_srcover_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4f03f610 // fmov v16.4s, #1.000000000000000000e+00
+ .long 0x4ea3d610 // fsub v16.4s, v16.4s, v3.4s
+ .long 0x4e24ce00 // fmla v0.4s, v16.4s, v4.4s
+ .long 0x4e25ce01 // fmla v1.4s, v16.4s, v5.4s
+ .long 0x4e26ce02 // fmla v2.4s, v16.4s, v6.4s
+ .long 0x4e27ce03 // fmla v3.4s, v16.4s, v7.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_dstover_aarch64
+_sk_dstover_aarch64:
+ .long 0x4f03f611 // fmov v17.4s, #1.000000000000000000e+00
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea41c90 // mov v16.16b, v4.16b
+ .long 0x4ea7d634 // fsub v20.4s, v17.4s, v7.4s
+ .long 0x4ea51cb1 // mov v17.16b, v5.16b
+ .long 0x4ea61cd2 // mov v18.16b, v6.16b
+ .long 0x4ea71cf3 // mov v19.16b, v7.16b
+ .long 0x4e20ce90 // fmla v16.4s, v20.4s, v0.4s
+ .long 0x4e21ce91 // fmla v17.4s, v20.4s, v1.4s
+ .long 0x4e22ce92 // fmla v18.4s, v20.4s, v2.4s
+ .long 0x4e23ce93 // fmla v19.4s, v20.4s, v3.4s
+ .long 0x4eb01e00 // mov v0.16b, v16.16b
+ .long 0x4eb11e21 // mov v1.16b, v17.16b
+ .long 0x4eb21e42 // mov v2.16b, v18.16b
+ .long 0x4eb31e63 // mov v3.16b, v19.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clamp_0_aarch64
+_sk_clamp_0_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x6f00e410 // movi v16.2d, #0x0
+ .long 0x4e30f400 // fmax v0.4s, v0.4s, v16.4s
+ .long 0x4e30f421 // fmax v1.4s, v1.4s, v16.4s
+ .long 0x4e30f442 // fmax v2.4s, v2.4s, v16.4s
+ .long 0x4e30f463 // fmax v3.4s, v3.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clamp_1_aarch64
+_sk_clamp_1_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4f03f610 // fmov v16.4s, #1.000000000000000000e+00
+ .long 0x4eb0f400 // fmin v0.4s, v0.4s, v16.4s
+ .long 0x4eb0f421 // fmin v1.4s, v1.4s, v16.4s
+ .long 0x4eb0f442 // fmin v2.4s, v2.4s, v16.4s
+ .long 0x4eb0f463 // fmin v3.4s, v3.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clamp_a_aarch64
+_sk_clamp_a_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4f03f610 // fmov v16.4s, #1.000000000000000000e+00
+ .long 0x4eb0f463 // fmin v3.4s, v3.4s, v16.4s
+ .long 0x4ea3f400 // fmin v0.4s, v0.4s, v3.4s
+ .long 0x4ea3f421 // fmin v1.4s, v1.4s, v3.4s
+ .long 0x4ea3f442 // fmin v2.4s, v2.4s, v3.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_set_rgb_aarch64
+_sk_set_rgb_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xaa0803e9 // mov x9, x8
+ .long 0x4ddfc920 // ld1r {v0.4s}, [x9], #4
+ .long 0x91002108 // add x8, x8, #0x8
+ .long 0x4d40c902 // ld1r {v2.4s}, [x8]
+ .long 0x4d40c921 // ld1r {v1.4s}, [x9]
+ .long 0xd61f0060 // br x3
+
+.globl _sk_swap_rb_aarch64
+_sk_swap_rb_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea01c10 // mov v16.16b, v0.16b
+ .long 0x4ea21c40 // mov v0.16b, v2.16b
+ .long 0x4eb01e02 // mov v2.16b, v16.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_swap_aarch64
+_sk_swap_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea31c70 // mov v16.16b, v3.16b
+ .long 0x4ea21c51 // mov v17.16b, v2.16b
+ .long 0x4ea11c32 // mov v18.16b, v1.16b
+ .long 0x4ea01c13 // mov v19.16b, v0.16b
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
+ .long 0x4ea51ca1 // mov v1.16b, v5.16b
+ .long 0x4ea61cc2 // mov v2.16b, v6.16b
+ .long 0x4ea71ce3 // mov v3.16b, v7.16b
+ .long 0x4eb31e64 // mov v4.16b, v19.16b
+ .long 0x4eb21e45 // mov v5.16b, v18.16b
+ .long 0x4eb11e26 // mov v6.16b, v17.16b
+ .long 0x4eb01e07 // mov v7.16b, v16.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_move_src_dst_aarch64
+_sk_move_src_dst_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea01c04 // mov v4.16b, v0.16b
+ .long 0x4ea11c25 // mov v5.16b, v1.16b
+ .long 0x4ea21c46 // mov v6.16b, v2.16b
+ .long 0x4ea31c67 // mov v7.16b, v3.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_move_dst_src_aarch64
+_sk_move_dst_src_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
+ .long 0x4ea51ca1 // mov v1.16b, v5.16b
+ .long 0x4ea61cc2 // mov v2.16b, v6.16b
+ .long 0x4ea71ce3 // mov v3.16b, v7.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_premul_aarch64
+_sk_premul_aarch64:
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x6e23dc00 // fmul v0.4s, v0.4s, v3.4s
+ .long 0x6e23dc21 // fmul v1.4s, v1.4s, v3.4s
+ .long 0x6e23dc42 // fmul v2.4s, v2.4s, v3.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_unpremul_aarch64
+_sk_unpremul_aarch64:
+ .long 0x4f03f611 // fmov v17.4s, #1.000000000000000000e+00
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4ea0d870 // fcmeq v16.4s, v3.4s, #0.0
+ .long 0x6e23fe31 // fdiv v17.4s, v17.4s, v3.4s
+ .long 0x4e701e30 // bic v16.16b, v17.16b, v16.16b
+ .long 0x6e20de00 // fmul v0.4s, v16.4s, v0.4s
+ .long 0x6e21de01 // fmul v1.4s, v16.4s, v1.4s
+ .long 0x6e22de02 // fmul v2.4s, v16.4s, v2.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_from_srgb_aarch64
+_sk_from_srgb_aarch64:
+ .long 0x52a7d328 // mov w8, #0x3e990000
+ .long 0x72933348 // movk w8, #0x999a
+ .long 0x4e040d10 // dup v16.4s, w8
+ .long 0x52a7e648 // mov w8, #0x3f320000
+ .long 0x7291eb88 // movk w8, #0x8f5c
+ .long 0x4e040d11 // dup v17.4s, w8
+ .long 0x52a76468 // mov w8, #0x3b230000
+ .long 0x729ae148 // movk w8, #0xd70a
+ .long 0x4e040d12 // dup v18.4s, w8
+ .long 0x52a7b3c8 // mov w8, #0x3d9e0000
+ .long 0x72907228 // movk w8, #0x8391
+ .long 0x6e22dc54 // fmul v20.4s, v2.4s, v2.4s
+ .long 0x4eb11e35 // mov v21.16b, v17.16b
+ .long 0x4eb11e37 // mov v23.16b, v17.16b
+ .long 0x4e22ce11 // fmla v17.4s, v16.4s, v2.4s
+ .long 0x4eb21e56 // mov v22.16b, v18.16b
+ .long 0x4eb21e58 // mov v24.16b, v18.16b
+ .long 0x4e34ce32 // fmla v18.4s, v17.4s, v20.4s
+ .long 0x4e040d11 // dup v17.4s, w8
+ .long 0x52a7ac28 // mov w8, #0x3d610000
+ .long 0x6e20dc13 // fmul v19.4s, v0.4s, v0.4s
+ .long 0x7288f5c8 // movk w8, #0x47ae
+ .long 0x4e20ce15 // fmla v21.4s, v16.4s, v0.4s
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x6e21dc34 // fmul v20.4s, v1.4s, v1.4s
+ .long 0x4e33ceb6 // fmla v22.4s, v21.4s, v19.4s
+ .long 0x4e040d13 // dup v19.4s, w8
+ .long 0x4e21ce17 // fmla v23.4s, v16.4s, v1.4s
+ .long 0x6e31dc15 // fmul v21.4s, v0.4s, v17.4s
+ .long 0x6ea0e660 // fcmgt v0.4s, v19.4s, v0.4s
+ .long 0x6e31dc30 // fmul v16.4s, v1.4s, v17.4s
+ .long 0x6ea1e661 // fcmgt v1.4s, v19.4s, v1.4s
+ .long 0x6e31dc51 // fmul v17.4s, v2.4s, v17.4s
+ .long 0x6ea2e662 // fcmgt v2.4s, v19.4s, v2.4s
+ .long 0x4e34cef8 // fmla v24.4s, v23.4s, v20.4s
+ .long 0x6e761ea0 // bsl v0.16b, v21.16b, v22.16b
+ .long 0x6e781e01 // bsl v1.16b, v16.16b, v24.16b
+ .long 0x6e721e22 // bsl v2.16b, v17.16b, v18.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_to_srgb_aarch64
+_sk_to_srgb_aarch64:
+ .long 0x52a828e8 // mov w8, #0x41470000
+ .long 0x728b8528 // movk w8, #0x5c29
+ .long 0x4e040d12 // dup v18.4s, w8
+ .long 0x52a7e608 // mov w8, #0x3f300000
+ .long 0x728df9c8 // movk w8, #0x6fce
+ .long 0x6ea1d811 // frsqrte v17.4s, v0.4s
+ .long 0x4e040d13 // dup v19.4s, w8
+ .long 0x52b7b948 // mov w8, #0xbdca0000
+ .long 0x728af508 // movk w8, #0x57a8
+ .long 0x6ea1d834 // frsqrte v20.4s, v1.4s
+ .long 0x6e31de36 // fmul v22.4s, v17.4s, v17.4s
+ .long 0x4e040d10 // dup v16.4s, w8
+ .long 0x52a77188 // mov w8, #0x3b8c0000
+ .long 0x6ea1d855 // frsqrte v21.4s, v2.4s
+ .long 0x6e34de98 // fmul v24.4s, v20.4s, v20.4s
+ .long 0x4eb6fc16 // frsqrts v22.4s, v0.4s, v22.4s
+ .long 0x729ce088 // movk w8, #0xe704
+ .long 0x6e35deb9 // fmul v25.4s, v21.4s, v21.4s
+ .long 0x4eb8fc38 // frsqrts v24.4s, v1.4s, v24.4s
+ .long 0x6e36de31 // fmul v17.4s, v17.4s, v22.4s
+ .long 0x4e040d17 // dup v23.4s, w8
+ .long 0x4eb9fc59 // frsqrts v25.4s, v2.4s, v25.4s
+ .long 0x6e38de94 // fmul v20.4s, v20.4s, v24.4s
+ .long 0x4ea1da36 // frecpe v22.4s, v17.4s
+ .long 0x6e32dc1a // fmul v26.4s, v0.4s, v18.4s
+ .long 0x6ea0e6e0 // fcmgt v0.4s, v23.4s, v0.4s
+ .long 0x6e32dc3c // fmul v28.4s, v1.4s, v18.4s
+ .long 0x6ea1e6e1 // fcmgt v1.4s, v23.4s, v1.4s
+ .long 0x6e32dc52 // fmul v18.4s, v2.4s, v18.4s
+ .long 0x6ea2e6e2 // fcmgt v2.4s, v23.4s, v2.4s
+ .long 0x6e39deb5 // fmul v21.4s, v21.4s, v25.4s
+ .long 0x4ea1da97 // frecpe v23.4s, v20.4s
+ .long 0x4e36fe39 // frecps v25.4s, v17.4s, v22.4s
+ .long 0x4ea1dab8 // frecpe v24.4s, v21.4s
+ .long 0x6e39ded6 // fmul v22.4s, v22.4s, v25.4s
+ .long 0x4e37fe99 // frecps v25.4s, v20.4s, v23.4s
+ .long 0x4eb01e1b // mov v27.16b, v16.16b
+ .long 0x6e39def7 // fmul v23.4s, v23.4s, v25.4s
+ .long 0x4e38feb9 // frecps v25.4s, v21.4s, v24.4s
+ .long 0x6e39df18 // fmul v24.4s, v24.4s, v25.4s
+ .long 0x4eb01e19 // mov v25.16b, v16.16b
+ .long 0x4e36ce7b // fmla v27.4s, v19.4s, v22.4s
+ .long 0x6ea1da36 // frsqrte v22.4s, v17.4s
+ .long 0x4e37ce79 // fmla v25.4s, v19.4s, v23.4s
+ .long 0x6ea1da97 // frsqrte v23.4s, v20.4s
+ .long 0x4e38ce70 // fmla v16.4s, v19.4s, v24.4s
+ .long 0x6e36ded8 // fmul v24.4s, v22.4s, v22.4s
+ .long 0x6ea1dab3 // frsqrte v19.4s, v21.4s
+ .long 0x4eb8fe31 // frsqrts v17.4s, v17.4s, v24.4s
+ .long 0x6e37def8 // fmul v24.4s, v23.4s, v23.4s
+ .long 0x4eb8fe94 // frsqrts v20.4s, v20.4s, v24.4s
+ .long 0x6e33de78 // fmul v24.4s, v19.4s, v19.4s
+ .long 0x52a7da48 // mov w8, #0x3ed20000
+ .long 0x4eb8feb5 // frsqrts v21.4s, v21.4s, v24.4s
+ .long 0x7290f848 // movk w8, #0x87c2
+ .long 0x6e31ded1 // fmul v17.4s, v22.4s, v17.4s
+ .long 0x6e34def4 // fmul v20.4s, v23.4s, v20.4s
+ .long 0x6e35de73 // fmul v19.4s, v19.4s, v21.4s
+ .long 0x4e040d15 // dup v21.4s, w8
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x4e31cebb // fmla v27.4s, v21.4s, v17.4s
+ .long 0x4f03f611 // fmov v17.4s, #1.000000000000000000e+00
+ .long 0x4e34ceb9 // fmla v25.4s, v21.4s, v20.4s
+ .long 0x4e33ceb0 // fmla v16.4s, v21.4s, v19.4s
+ .long 0x4ebbf633 // fmin v19.4s, v17.4s, v27.4s
+ .long 0x4eb9f634 // fmin v20.4s, v17.4s, v25.4s
+ .long 0x4eb0f630 // fmin v16.4s, v17.4s, v16.4s
+ .long 0x6e731f40 // bsl v0.16b, v26.16b, v19.16b
+ .long 0x6e741f81 // bsl v1.16b, v28.16b, v20.16b
+ .long 0x6e701e42 // bsl v2.16b, v18.16b, v16.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_scale_1_float_aarch64
+_sk_scale_1_float_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4f909000 // fmul v0.4s, v0.4s, v16.s[0]
+ .long 0x4f909021 // fmul v1.4s, v1.4s, v16.s[0]
+ .long 0x4f909042 // fmul v2.4s, v2.4s, v16.s[0]
+ .long 0x4f909063 // fmul v3.4s, v3.4s, v16.s[0]
+ .long 0xd61f0060 // br x3
+
+.globl _sk_scale_u8_aarch64
+_sk_scale_u8_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x52a77009 // mov w9, #0x3b800000
+ .long 0x72901029 // movk w9, #0x8081
+ .long 0x4e040d30 // dup v16.4s, w9
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x8b000108 // add x8, x8, x0
+ .long 0x39400109 // ldrb w9, [x8]
+ .long 0x3940050a // ldrb w10, [x8, #1]
+ .long 0x3940090b // ldrb w11, [x8, #2]
+ .long 0x39400d08 // ldrb w8, [x8, #3]
+ .long 0x4e021d31 // mov v17.h[0], w9
+ .long 0x4e061d51 // mov v17.h[1], w10
+ .long 0x4e0a1d71 // mov v17.h[2], w11
+ .long 0x4e0e1d11 // mov v17.h[3], w8
+ .long 0x2f10a631 // uxtl v17.4s, v17.4h
+ .long 0x6e21da31 // ucvtf v17.4s, v17.4s
+ .long 0x6e30de30 // fmul v16.4s, v17.4s, v16.4s
+ .long 0x6e20de00 // fmul v0.4s, v16.4s, v0.4s
+ .long 0x6e21de01 // fmul v1.4s, v16.4s, v1.4s
+ .long 0x6e22de02 // fmul v2.4s, v16.4s, v2.4s
+ .long 0x6e23de03 // fmul v3.4s, v16.4s, v3.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_lerp_1_float_aarch64
+_sk_lerp_1_float_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x4ea4d411 // fsub v17.4s, v0.4s, v4.4s
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
+ .long 0x4ea5d432 // fsub v18.4s, v1.4s, v5.4s
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4ea51ca1 // mov v1.16b, v5.16b
+ .long 0x4f901220 // fmla v0.4s, v17.4s, v16.s[0]
+ .long 0x4ea6d451 // fsub v17.4s, v2.4s, v6.4s
+ .long 0x4f901241 // fmla v1.4s, v18.4s, v16.s[0]
+ .long 0x4ea61cc2 // mov v2.16b, v6.16b
+ .long 0x4ea7d472 // fsub v18.4s, v3.4s, v7.4s
+ .long 0x4ea71ce3 // mov v3.16b, v7.16b
+ .long 0x4f901222 // fmla v2.4s, v17.4s, v16.s[0]
+ .long 0x4f901243 // fmla v3.4s, v18.4s, v16.s[0]
+ .long 0xd61f0060 // br x3
+
+.globl _sk_lerp_u8_aarch64
+_sk_lerp_u8_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x52a77009 // mov w9, #0x3b800000
+ .long 0x72901029 // movk w9, #0x8081
+ .long 0x4e040d30 // dup v16.4s, w9
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x4ea4d412 // fsub v18.4s, v0.4s, v4.4s
+ .long 0x8b000108 // add x8, x8, x0
+ .long 0x3940010a // ldrb w10, [x8]
+ .long 0x39400509 // ldrb w9, [x8, #1]
+ .long 0x3940090b // ldrb w11, [x8, #2]
+ .long 0x39400d08 // ldrb w8, [x8, #3]
+ .long 0x4e021d51 // mov v17.h[0], w10
+ .long 0x4e061d31 // mov v17.h[1], w9
+ .long 0x4e0a1d71 // mov v17.h[2], w11
+ .long 0x4e0e1d11 // mov v17.h[3], w8
+ .long 0x2f10a620 // uxtl v0.4s, v17.4h
+ .long 0x6e21d800 // ucvtf v0.4s, v0.4s
+ .long 0x6e30dc10 // fmul v16.4s, v0.4s, v16.4s
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
+ .long 0x4ea5d431 // fsub v17.4s, v1.4s, v5.4s
+ .long 0x4ea51ca1 // mov v1.16b, v5.16b
+ .long 0x4e32ce00 // fmla v0.4s, v16.4s, v18.4s
+ .long 0x4ea6d452 // fsub v18.4s, v2.4s, v6.4s
+ .long 0x4e31ce01 // fmla v1.4s, v16.4s, v17.4s
+ .long 0x4ea61cc2 // mov v2.16b, v6.16b
+ .long 0x4ea7d471 // fsub v17.4s, v3.4s, v7.4s
+ .long 0x4ea71ce3 // mov v3.16b, v7.16b
+ .long 0x4e32ce02 // fmla v2.4s, v16.4s, v18.4s
+ .long 0x4e31ce03 // fmla v3.4s, v16.4s, v17.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_lerp_565_aarch64
+_sk_lerp_565_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xd37ff809 // lsl x9, x0, #1
+ .long 0x4f072710 // movi v16.4s, #0xf8, lsl #8
+ .long 0x4ea4d413 // fsub v19.4s, v0.4s, v4.4s
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0xfc696903 // ldr d3, [x8, x9]
+ .long 0x52a6f088 // mov w8, #0x37840000
+ .long 0x72842108 // movk w8, #0x2108
+ .long 0x4e040d11 // dup v17.4s, w8
+ .long 0x2f10a463 // uxtl v3.4s, v3.4h
+ .long 0x321b17e8 // orr w8, wzr, #0x7e0
+ .long 0x4e301c60 // and v0.16b, v3.16b, v16.16b
+ .long 0x4e040d12 // dup v18.4s, w8
+ .long 0x52a74048 // mov w8, #0x3a020000
+ .long 0x4e21d800 // scvtf v0.4s, v0.4s
+ .long 0x72810428 // movk w8, #0x821
+ .long 0x6e31dc10 // fmul v16.4s, v0.4s, v17.4s
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
+ .long 0x4e33ce00 // fmla v0.4s, v16.4s, v19.4s
+ .long 0x4f0007f0 // movi v16.4s, #0x1f
+ .long 0x4e040d11 // dup v17.4s, w8
+ .long 0x52a7a088 // mov w8, #0x3d040000
+ .long 0x4e321c72 // and v18.16b, v3.16b, v18.16b
+ .long 0x72842108 // movk w8, #0x2108
+ .long 0x4e301c63 // and v3.16b, v3.16b, v16.16b
+ .long 0x4ea6d450 // fsub v16.4s, v2.4s, v6.4s
+ .long 0x4e21da42 // scvtf v2.4s, v18.4s
+ .long 0x6e31dc51 // fmul v17.4s, v2.4s, v17.4s
+ .long 0x4e040d02 // dup v2.4s, w8
+ .long 0x4e21d863 // scvtf v3.4s, v3.4s
+ .long 0x4ea5d433 // fsub v19.4s, v1.4s, v5.4s
+ .long 0x4ea51ca1 // mov v1.16b, v5.16b
+ .long 0x6e22dc63 // fmul v3.4s, v3.4s, v2.4s
+ .long 0x4ea61cc2 // mov v2.16b, v6.16b
+ .long 0x4e33ce21 // fmla v1.4s, v17.4s, v19.4s
+ .long 0x4e30cc62 // fmla v2.4s, v3.4s, v16.4s
+ .long 0x4f03f603 // fmov v3.4s, #1.000000000000000000e+00
+ .long 0xd61f0060 // br x3
+
+.globl _sk_load_tables_aarch64
+_sk_load_tables_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xd37ef409 // lsl x9, x0, #2
+ .long 0x6f00e620 // movi v0.2d, #0xff000000ff
+ .long 0x52a7700b // mov w11, #0x3b800000
+ .long 0xa940310a // ldp x10, x12, [x8]
+ .long 0x7290102b // movk w11, #0x8081
+ .long 0x4e040d63 // dup v3.4s, w11
+ .long 0x3ce96942 // ldr q2, [x10, x9]
+ .long 0xa9412109 // ldp x9, x8, [x8, #16]
+ .long 0x4e201c41 // and v1.16b, v2.16b, v0.16b
+ .long 0x1e26002e // fmov w14, s1
+ .long 0x6f380450 // ushr v16.4s, v2.4s, #8
+ .long 0x6f300451 // ushr v17.4s, v2.4s, #16
+ .long 0x8b2e498e // add x14, x12, w14, uxtw #2
+ .long 0x0e0c3c2a // mov w10, v1.s[1]
+ .long 0x0e143c2b // mov w11, v1.s[2]
+ .long 0x0e1c3c2d // mov w13, v1.s[3]
+ .long 0x4e201e01 // and v1.16b, v16.16b, v0.16b
+ .long 0x4e201e30 // and v16.16b, v17.16b, v0.16b
+ .long 0x0d4081c0 // ld1 {v0.s}[0], [x14]
+ .long 0x8b2a498a // add x10, x12, w10, uxtw #2
+ .long 0xbc6b5991 // ldr s17, [x12, w11, uxtw #2]
+ .long 0xbc6d5992 // ldr s18, [x12, w13, uxtw #2]
+ .long 0x0e0c3c2b // mov w11, v1.s[1]
+ .long 0x0e143c2c // mov w12, v1.s[2]
+ .long 0x0e1c3c2d // mov w13, v1.s[3]
+ .long 0x1e26002e // fmov w14, s1
+ .long 0x8b2e492e // add x14, x9, w14, uxtw #2
+ .long 0xbc6c5933 // ldr s19, [x9, w12, uxtw #2]
+ .long 0xbc6d5934 // ldr s20, [x9, w13, uxtw #2]
+ .long 0x8b2b4929 // add x9, x9, w11, uxtw #2
+ .long 0x1e26020b // fmov w11, s16
+ .long 0x6f280442 // ushr v2.4s, v2.4s, #24
+ .long 0x0d409140 // ld1 {v0.s}[1], [x10]
+ .long 0x4e21d842 // scvtf v2.4s, v2.4s
+ .long 0x8b2b490a // add x10, x8, w11, uxtw #2
+ .long 0x0d4081c1 // ld1 {v1.s}[0], [x14]
+ .long 0x6e23dc43 // fmul v3.4s, v2.4s, v3.4s
+ .long 0x0d408142 // ld1 {v2.s}[0], [x10]
+ .long 0x0e0c3e0f // mov w15, v16.s[1]
+ .long 0x0e143e0c // mov w12, v16.s[2]
+ .long 0x8b2f490a // add x10, x8, w15, uxtw #2
+ .long 0x0e1c3e0d // mov w13, v16.s[3]
+ .long 0xbc6c5910 // ldr s16, [x8, w12, uxtw #2]
+ .long 0x0d409121 // ld1 {v1.s}[1], [x9]
+ .long 0x0d409142 // ld1 {v2.s}[1], [x10]
+ .long 0x6e140620 // mov v0.s[2], v17.s[0]
+ .long 0xbc6d5911 // ldr s17, [x8, w13, uxtw #2]
+ .long 0x6e140661 // mov v1.s[2], v19.s[0]
+ .long 0x6e140602 // mov v2.s[2], v16.s[0]
+ .long 0x6e1c0640 // mov v0.s[3], v18.s[0]
+ .long 0x6e1c0681 // mov v1.s[3], v20.s[0]
+ .long 0x6e1c0622 // mov v2.s[3], v17.s[0]
+ .long 0xd61f0060 // br x3
+
+.globl _sk_load_a8_aarch64
+_sk_load_a8_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x52a77009 // mov w9, #0x3b800000
+ .long 0x72901029 // movk w9, #0x8081
+ .long 0x4e040d22 // dup v2.4s, w9
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x6f00e400 // movi v0.2d, #0x0
+ .long 0x6f00e401 // movi v1.2d, #0x0
+ .long 0x8b000108 // add x8, x8, x0
+ .long 0x3940010a // ldrb w10, [x8]
+ .long 0x3940050b // ldrb w11, [x8, #1]
+ .long 0x3940090c // ldrb w12, [x8, #2]
+ .long 0x39400d08 // ldrb w8, [x8, #3]
+ .long 0x4e021d43 // mov v3.h[0], w10
+ .long 0x4e061d63 // mov v3.h[1], w11
+ .long 0x4e0a1d83 // mov v3.h[2], w12
+ .long 0x4e0e1d03 // mov v3.h[3], w8
+ .long 0x2f10a463 // uxtl v3.4s, v3.4h
+ .long 0x6e21d863 // ucvtf v3.4s, v3.4s
+ .long 0x6e22dc63 // fmul v3.4s, v3.4s, v2.4s
+ .long 0x6f00e402 // movi v2.2d, #0x0
+ .long 0xd61f0060 // br x3
+
+.globl _sk_store_a8_aarch64
+_sk_store_a8_aarch64:
+ .long 0xf9400028 // ldr x8, [x1]
+ .long 0x52a86fe9 // mov w9, #0x437f0000
+ .long 0x4e040d30 // dup v16.4s, w9
+ .long 0x6e30dc70 // fmul v16.4s, v3.4s, v16.4s
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x6e21aa10 // fcvtnu v16.4s, v16.4s
+ .long 0x0e612a10 // xtn v16.4h, v16.4s
+ .long 0x0e0e3e09 // umov w9, v16.h[3]
+ .long 0x8b000108 // add x8, x8, x0
+ .long 0x39000d09 // strb w9, [x8, #3]
+ .long 0x0e0a3e09 // umov w9, v16.h[2]
+ .long 0x39000909 // strb w9, [x8, #2]
+ .long 0x0e063e09 // umov w9, v16.h[1]
+ .long 0x39000509 // strb w9, [x8, #1]
+ .long 0x0e023e09 // umov w9, v16.h[0]
+ .long 0x39000109 // strb w9, [x8]
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0xd61f0060 // br x3
+
+.globl _sk_load_565_aarch64
+_sk_load_565_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xd37ff809 // lsl x9, x0, #1
+ .long 0x4f072701 // movi v1.4s, #0xf8, lsl #8
+ .long 0x4f0007e3 // movi v3.4s, #0x1f
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0xfc696900 // ldr d0, [x8, x9]
+ .long 0x321b17e8 // orr w8, wzr, #0x7e0
+ .long 0x4e040d02 // dup v2.4s, w8
+ .long 0x52a6f088 // mov w8, #0x37840000
+ .long 0x72842108 // movk w8, #0x2108
+ .long 0x2f10a400 // uxtl v0.4s, v0.4h
+ .long 0x4e211c01 // and v1.16b, v0.16b, v1.16b
+ .long 0x4e221c02 // and v2.16b, v0.16b, v2.16b
+ .long 0x4e231c03 // and v3.16b, v0.16b, v3.16b
+ .long 0x4e040d00 // dup v0.4s, w8
+ .long 0x52a74048 // mov w8, #0x3a020000
+ .long 0x72810428 // movk w8, #0x821
+ .long 0x4e21d821 // scvtf v1.4s, v1.4s
+ .long 0x6e20dc20 // fmul v0.4s, v1.4s, v0.4s
+ .long 0x4e040d01 // dup v1.4s, w8
+ .long 0x52a7a088 // mov w8, #0x3d040000
+ .long 0x72842108 // movk w8, #0x2108
+ .long 0x4e21d842 // scvtf v2.4s, v2.4s
+ .long 0x6e21dc41 // fmul v1.4s, v2.4s, v1.4s
+ .long 0x4e040d02 // dup v2.4s, w8
+ .long 0x4e21d863 // scvtf v3.4s, v3.4s
+ .long 0x6e22dc62 // fmul v2.4s, v3.4s, v2.4s
+ .long 0x4f03f603 // fmov v3.4s, #1.000000000000000000e+00
+ .long 0xd61f0060 // br x3
+
+.globl _sk_store_565_aarch64
+_sk_store_565_aarch64:
+ .long 0xf9400028 // ldr x8, [x1]
+ .long 0x52a84f8a // mov w10, #0x427c0000
+ .long 0x4f01f7f0 // fmov v16.4s, #3.100000000000000000e+01
+ .long 0x4e040d52 // dup v18.4s, w10
+ .long 0x6e30dc11 // fmul v17.4s, v0.4s, v16.4s
+ .long 0x6e32dc32 // fmul v18.4s, v1.4s, v18.4s
+ .long 0x6e21aa31 // fcvtnu v17.4s, v17.4s
+ .long 0x6e21aa52 // fcvtnu v18.4s, v18.4s
+ .long 0x6e30dc50 // fmul v16.4s, v2.4s, v16.4s
+ .long 0x4f2b5631 // shl v17.4s, v17.4s, #11
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x4f255652 // shl v18.4s, v18.4s, #5
+ .long 0x4eb11e51 // orr v17.16b, v18.16b, v17.16b
+ .long 0x6e21aa10 // fcvtnu v16.4s, v16.4s
+ .long 0x4eb01e30 // orr v16.16b, v17.16b, v16.16b
+ .long 0xd37ff809 // lsl x9, x0, #1
+ .long 0x0e612a10 // xtn v16.4h, v16.4s
+ .long 0xfc296910 // str d16, [x8, x9]
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0xd61f0060 // br x3
+
+.globl _sk_load_8888_aarch64
+_sk_load_8888_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xd37ef409 // lsl x9, x0, #2
+ .long 0x6f00e621 // movi v1.2d, #0xff000000ff
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x3ce96900 // ldr q0, [x8, x9]
+ .long 0x52a77008 // mov w8, #0x3b800000
+ .long 0x72901028 // movk w8, #0x8081
+ .long 0x4e040d02 // dup v2.4s, w8
+ .long 0x6f380410 // ushr v16.4s, v0.4s, #8
+ .long 0x6f300411 // ushr v17.4s, v0.4s, #16
+ .long 0x4e211c03 // and v3.16b, v0.16b, v1.16b
+ .long 0x6f280400 // ushr v0.4s, v0.4s, #24
+ .long 0x4e211e10 // and v16.16b, v16.16b, v1.16b
+ .long 0x4e211e21 // and v1.16b, v17.16b, v1.16b
+ .long 0x4e21d863 // scvtf v3.4s, v3.4s
+ .long 0x4e21d811 // scvtf v17.4s, v0.4s
+ .long 0x4e21da10 // scvtf v16.4s, v16.4s
+ .long 0x4e21d832 // scvtf v18.4s, v1.4s
+ .long 0x6e22dc60 // fmul v0.4s, v3.4s, v2.4s
+ .long 0x6e22de23 // fmul v3.4s, v17.4s, v2.4s
+ .long 0x6e22de01 // fmul v1.4s, v16.4s, v2.4s
+ .long 0x6e22de42 // fmul v2.4s, v18.4s, v2.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_store_8888_aarch64
+_sk_store_8888_aarch64:
+ .long 0x52a86fea // mov w10, #0x437f0000
+ .long 0x4e040d50 // dup v16.4s, w10
+ .long 0xf9400028 // ldr x8, [x1]
+ .long 0x6e30dc32 // fmul v18.4s, v1.4s, v16.4s
+ .long 0x6e30dc11 // fmul v17.4s, v0.4s, v16.4s
+ .long 0x6e21aa52 // fcvtnu v18.4s, v18.4s
+ .long 0x6e21aa31 // fcvtnu v17.4s, v17.4s
+ .long 0x4f285652 // shl v18.4s, v18.4s, #8
+ .long 0x4eb11e51 // orr v17.16b, v18.16b, v17.16b
+ .long 0x6e30dc52 // fmul v18.4s, v2.4s, v16.4s
+ .long 0x6e30dc70 // fmul v16.4s, v3.4s, v16.4s
+ .long 0x6e21aa52 // fcvtnu v18.4s, v18.4s
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x6e21aa10 // fcvtnu v16.4s, v16.4s
+ .long 0x4f305652 // shl v18.4s, v18.4s, #16
+ .long 0x4eb21e31 // orr v17.16b, v17.16b, v18.16b
+ .long 0x4f385610 // shl v16.4s, v16.4s, #24
+ .long 0xd37ef409 // lsl x9, x0, #2
+ .long 0x4eb01e30 // orr v16.16b, v17.16b, v16.16b
+ .long 0x3ca96910 // str q16, [x8, x9]
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0xd61f0060 // br x3
+
+.globl _sk_load_f16_aarch64
+_sk_load_f16_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x8b000d08 // add x8, x8, x0, lsl #3
+ .long 0x0c400510 // ld4 {v16.4h-v19.4h}, [x8]
+ .long 0x0e217a00 // fcvtl v0.4s, v16.4h
+ .long 0x0e217a21 // fcvtl v1.4s, v17.4h
+ .long 0x0e217a42 // fcvtl v2.4s, v18.4h
+ .long 0x0e217a63 // fcvtl v3.4s, v19.4h
+ .long 0xd61f0060 // br x3
+
+.globl _sk_store_f16_aarch64
+_sk_store_f16_aarch64:
+ .long 0xf9400028 // ldr x8, [x1]
+ .long 0x0e216810 // fcvtn v16.4h, v0.4s
+ .long 0x0e216831 // fcvtn v17.4h, v1.4s
+ .long 0x0e216852 // fcvtn v18.4h, v2.4s
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x0e216873 // fcvtn v19.4h, v3.4s
+ .long 0x8b000d08 // add x8, x8, x0, lsl #3
+ .long 0x0c000510 // st4 {v16.4h-v19.4h}, [x8]
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0xd61f0060 // br x3
+
+.globl _sk_store_f32_aarch64
+_sk_store_f32_aarch64:
+ .long 0xf9400028 // ldr x8, [x1]
+ .long 0xf9400108 // ldr x8, [x8]
+ .long 0x8b001108 // add x8, x8, x0, lsl #4
+ .long 0x4c000900 // st4 {v0.4s-v3.4s}, [x8]
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clamp_x_aarch64
+_sk_clamp_x_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x6f00e411 // movi v17.2d, #0x0
+ .long 0x4e20f620 // fmax v0.4s, v17.4s, v0.4s
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0x4d40c910 // ld1r {v16.4s}, [x8]
+ .long 0x4eb18610 // add v16.4s, v16.4s, v17.4s
+ .long 0x4eb0f400 // fmin v0.4s, v0.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_clamp_y_aarch64
+_sk_clamp_y_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x6f00e411 // movi v17.2d, #0x0
+ .long 0x4e21f621 // fmax v1.4s, v17.4s, v1.4s
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0x4d40c910 // ld1r {v16.4s}, [x8]
+ .long 0x4eb18610 // add v16.4s, v16.4s, v17.4s
+ .long 0x4eb0f421 // fmin v1.4s, v1.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_repeat_x_aarch64
+_sk_repeat_x_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x4eb18651 // add v17.4s, v18.4s, v17.4s
+ .long 0x6e32fc12 // fdiv v18.4s, v0.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f905240 // fmls v0.4s, v18.4s, v16.s[0]
+ .long 0x4eb1f400 // fmin v0.4s, v0.4s, v17.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_repeat_y_aarch64
+_sk_repeat_y_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x4eb18651 // add v17.4s, v18.4s, v17.4s
+ .long 0x6e32fc32 // fdiv v18.4s, v1.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f905241 // fmls v1.4s, v18.4s, v16.s[0]
+ .long 0x4eb1f421 // fmin v1.4s, v1.4s, v17.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_mirror_x_aarch64
+_sk_mirror_x_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040611 // dup v17.4s, v16.s[0]
+ .long 0x1e302a10 // fadd s16, s16, s16
+ .long 0x4eb1d400 // fsub v0.4s, v0.4s, v17.4s
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x6e32fc12 // fdiv v18.4s, v0.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f905240 // fmls v0.4s, v18.4s, v16.s[0]
+ .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
+ .long 0x4eb1d400 // fsub v0.4s, v0.4s, v17.4s
+ .long 0x4eb08630 // add v16.4s, v17.4s, v16.4s
+ .long 0x4ea0f800 // fabs v0.4s, v0.4s
+ .long 0x4eb0f400 // fmin v0.4s, v0.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_mirror_y_aarch64
+_sk_mirror_y_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040611 // dup v17.4s, v16.s[0]
+ .long 0x1e302a10 // fadd s16, s16, s16
+ .long 0x4eb1d421 // fsub v1.4s, v1.4s, v17.4s
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x6e32fc32 // fdiv v18.4s, v1.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f905241 // fmls v1.4s, v18.4s, v16.s[0]
+ .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
+ .long 0x4eb1d421 // fsub v1.4s, v1.4s, v17.4s
+ .long 0x4eb08630 // add v16.4s, v17.4s, v16.4s
+ .long 0x4ea0f821 // fabs v1.4s, v1.4s
+ .long 0x4eb0f421 // fmin v1.4s, v1.4s, v16.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_luminance_to_alpha_aarch64
+_sk_luminance_to_alpha_aarch64:
+ .long 0x52a7cb28 // mov w8, #0x3e590000
+ .long 0x72967a08 // movk w8, #0xb3d0
+ .long 0x4e040d11 // dup v17.4s, w8
+ .long 0x52a7e6e8 // mov w8, #0x3f370000
+ .long 0x7282eb28 // movk w8, #0x1759
+ .long 0x4ea01c10 // mov v16.16b, v0.16b
+ .long 0x4e040d00 // dup v0.4s, w8
+ .long 0x52a7b268 // mov w8, #0x3d930000
+ .long 0xf8408423 // ldr x3, [x1], #8
+ .long 0x729bb308 // movk w8, #0xdd98
+ .long 0x6e20dc23 // fmul v3.4s, v1.4s, v0.4s
+ .long 0x4e30ce23 // fmla v3.4s, v17.4s, v16.4s
+ .long 0x4e040d10 // dup v16.4s, w8
+ .long 0x6f00e400 // movi v0.2d, #0x0
+ .long 0x6f00e401 // movi v1.2d, #0x0
+ .long 0x4e22ce03 // fmla v3.4s, v16.4s, v2.4s
+ .long 0x6f00e402 // movi v2.2d, #0x0
+ .long 0xd61f0060 // br x3
+
+.globl _sk_matrix_2x3_aarch64
+_sk_matrix_2x3_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xaa0803e9 // mov x9, x8
+ .long 0x9100410a // add x10, x8, #0x10
+ .long 0x4ddfc932 // ld1r {v18.4s}, [x9], #4
+ .long 0x4d40c950 // ld1r {v16.4s}, [x10]
+ .long 0x2d415113 // ldp s19, s20, [x8, #8]
+ .long 0x9100510a // add x10, x8, #0x14
+ .long 0x4d40c951 // ld1r {v17.4s}, [x10]
+ .long 0x4f931030 // fmla v16.4s, v1.4s, v19.s[0]
+ .long 0xbd400133 // ldr s19, [x9]
+ .long 0x4f941031 // fmla v17.4s, v1.4s, v20.s[0]
+ .long 0x4e20ce50 // fmla v16.4s, v18.4s, v0.4s
+ .long 0x4f931011 // fmla v17.4s, v0.4s, v19.s[0]
+ .long 0x4eb01e00 // mov v0.16b, v16.16b
+ .long 0x4eb11e21 // mov v1.16b, v17.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_matrix_3x4_aarch64
+_sk_matrix_3x4_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xaa0803e9 // mov x9, x8
+ .long 0x9100910a // add x10, x8, #0x24
+ .long 0x4ddfc933 // ld1r {v19.4s}, [x9], #4
+ .long 0x4d40c950 // ld1r {v16.4s}, [x10]
+ .long 0x9100a10a // add x10, x8, #0x28
+ .long 0x4d40c951 // ld1r {v17.4s}, [x10]
+ .long 0x9100b10a // add x10, x8, #0x2c
+ .long 0x2d435514 // ldp s20, s21, [x8, #24]
+ .long 0xbd402116 // ldr s22, [x8, #32]
+ .long 0x4d40c952 // ld1r {v18.4s}, [x10]
+ .long 0x4f941050 // fmla v16.4s, v2.4s, v20.s[0]
+ .long 0x4f951051 // fmla v17.4s, v2.4s, v21.s[0]
+ .long 0x4f961052 // fmla v18.4s, v2.4s, v22.s[0]
+ .long 0x2d425502 // ldp s2, s21, [x8, #16]
+ .long 0x2d415d14 // ldp s20, s23, [x8, #8]
+ .long 0x4f821031 // fmla v17.4s, v1.4s, v2.s[0]
+ .long 0xbd400122 // ldr s2, [x9]
+ .long 0x4f971030 // fmla v16.4s, v1.4s, v23.s[0]
+ .long 0x4f951032 // fmla v18.4s, v1.4s, v21.s[0]
+ .long 0x4e20ce70 // fmla v16.4s, v19.4s, v0.4s
+ .long 0x4f941012 // fmla v18.4s, v0.4s, v20.s[0]
+ .long 0x4f821011 // fmla v17.4s, v0.4s, v2.s[0]
+ .long 0x4eb01e00 // mov v0.16b, v16.16b
+ .long 0x4eb11e21 // mov v1.16b, v17.16b
+ .long 0x4eb21e42 // mov v2.16b, v18.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_matrix_4x5_aarch64
+_sk_matrix_4x5_aarch64:
+ .long 0xf9400029 // ldr x9, [x1]
+ .long 0xaa0903e8 // mov x8, x9
+ .long 0x9101012a // add x10, x9, #0x40
+ .long 0x4ddfc914 // ld1r {v20.4s}, [x8], #4
+ .long 0x4d40c950 // ld1r {v16.4s}, [x10]
+ .long 0x9101112a // add x10, x9, #0x44
+ .long 0x4d40c951 // ld1r {v17.4s}, [x10]
+ .long 0x9101212a // add x10, x9, #0x48
+ .long 0x4d40c952 // ld1r {v18.4s}, [x10]
+ .long 0x2d465533 // ldp s19, s21, [x9, #48]
+ .long 0x2d475d36 // ldp s22, s23, [x9, #56]
+ .long 0x9101312a // add x10, x9, #0x4c
+ .long 0xf9400423 // ldr x3, [x1, #8]
+ .long 0x4f931070 // fmla v16.4s, v3.4s, v19.s[0]
+ .long 0x4d40c953 // ld1r {v19.4s}, [x10]
+ .long 0x4f951071 // fmla v17.4s, v3.4s, v21.s[0]
+ .long 0x4f961072 // fmla v18.4s, v3.4s, v22.s[0]
+ .long 0x2d445935 // ldp s21, s22, [x9, #32]
+ .long 0x4f971073 // fmla v19.4s, v3.4s, v23.s[0]
+ .long 0x2d455d23 // ldp s3, s23, [x9, #40]
+ .long 0x91004021 // add x1, x1, #0x10
+ .long 0x4f951050 // fmla v16.4s, v2.4s, v21.s[0]
+ .long 0x4f961051 // fmla v17.4s, v2.4s, v22.s[0]
+ .long 0x2d425935 // ldp s21, s22, [x9, #16]
+ .long 0x4f971053 // fmla v19.4s, v2.4s, v23.s[0]
+ .long 0x4f831052 // fmla v18.4s, v2.4s, v3.s[0]
+ .long 0x2d410d22 // ldp s2, s3, [x9, #8]
+ .long 0x4f951030 // fmla v16.4s, v1.4s, v21.s[0]
+ .long 0x2d435d35 // ldp s21, s23, [x9, #24]
+ .long 0x4f961031 // fmla v17.4s, v1.4s, v22.s[0]
+ .long 0xbd400116 // ldr s22, [x8]
+ .long 0x4e20ce90 // fmla v16.4s, v20.4s, v0.4s
+ .long 0x4f951032 // fmla v18.4s, v1.4s, v21.s[0]
+ .long 0x4f971033 // fmla v19.4s, v1.4s, v23.s[0]
+ .long 0x4f821012 // fmla v18.4s, v0.4s, v2.s[0]
+ .long 0x4f831013 // fmla v19.4s, v0.4s, v3.s[0]
+ .long 0x4f961011 // fmla v17.4s, v0.4s, v22.s[0]
+ .long 0x4eb01e00 // mov v0.16b, v16.16b
+ .long 0x4eb11e21 // mov v1.16b, v17.16b
+ .long 0x4eb21e42 // mov v2.16b, v18.16b
+ .long 0x4eb31e63 // mov v3.16b, v19.16b
+ .long 0xd61f0060 // br x3
+
+.globl _sk_matrix_perspective_aarch64
+_sk_matrix_perspective_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xaa0803e9 // mov x9, x8
+ .long 0x9100510a // add x10, x8, #0x14
+ .long 0x4ddfc930 // ld1r {v16.4s}, [x9], #4
+ .long 0x4d40c951 // ld1r {v17.4s}, [x10]
+ .long 0x9100810a // add x10, x8, #0x20
+ .long 0x4d40c952 // ld1r {v18.4s}, [x10]
+ .long 0x2d41d113 // ldp s19, s20, [x8, #12]
+ .long 0x2d435915 // ldp s21, s22, [x8, #24]
+ .long 0x91002108 // add x8, x8, #0x8
+ .long 0x4f941031 // fmla v17.4s, v1.4s, v20.s[0]
+ .long 0x4d40c914 // ld1r {v20.4s}, [x8]
+ .long 0x4f961032 // fmla v18.4s, v1.4s, v22.s[0]
+ .long 0xbd400136 // ldr s22, [x9]
+ .long 0x4f951012 // fmla v18.4s, v0.4s, v21.s[0]
+ .long 0x4f931011 // fmla v17.4s, v0.4s, v19.s[0]
+ .long 0x4f961034 // fmla v20.4s, v1.4s, v22.s[0]
+ .long 0x4ea1da41 // frecpe v1.4s, v18.4s
+ .long 0x4e21fe52 // frecps v18.4s, v18.4s, v1.4s
+ .long 0x6e32dc32 // fmul v18.4s, v1.4s, v18.4s
+ .long 0x4e20ce14 // fmla v20.4s, v16.4s, v0.4s
+ .long 0x6e32de21 // fmul v1.4s, v17.4s, v18.4s
+ .long 0x6e32de80 // fmul v0.4s, v20.4s, v18.4s
+ .long 0xd61f0060 // br x3
+
+.globl _sk_linear_gradient_2stops_aarch64
+_sk_linear_gradient_2stops_aarch64:
+ .long 0xa8c10c28 // ldp x8, x3, [x1], #16
+ .long 0xad404503 // ldp q3, q17, [x8]
+ .long 0x4e040470 // dup v16.4s, v3.s[0]
+ .long 0x4e0c0461 // dup v1.4s, v3.s[1]
+ .long 0x4e140462 // dup v2.4s, v3.s[2]
+ .long 0x4e1c0463 // dup v3.4s, v3.s[3]
+ .long 0x4f911010 // fmla v16.4s, v0.4s, v17.s[0]
+ .long 0x4fb11001 // fmla v1.4s, v0.4s, v17.s[1]
+ .long 0x4f911802 // fmla v2.4s, v0.4s, v17.s[2]
+ .long 0x4fb11803 // fmla v3.4s, v0.4s, v17.s[3]
+ .long 0x4eb01e00 // mov v0.16b, v16.16b
+ .long 0xd61f0060 // br x3
+#elif defined(__arm__)
+.balign 4
+
+.globl _sk_start_pipeline_vfp4
+_sk_start_pipeline_vfp4:
+ .long 0xe92d41f0 // push {r4, r5, r6, r7, r8, lr}
+ .long 0xe1a04000 // mov r4, r0
+ .long 0xe2840002 // add r0, r4, #2
+ .long 0xe1a05003 // mov r5, r3
+ .long 0xe1a08002 // mov r8, r2
+ .long 0xe1a07001 // mov r7, r1
+ .long 0xe1500005 // cmp r0, r5
+ .long 0x8a000010 // bhi 64 <sk_start_pipeline_vfp4+0x64>
+ .long 0xe4976004 // ldr r6, [r7], #4
+ .long 0xf2800010 // vmov.i32 d0, #0
+ .long 0xe1a00004 // mov r0, r4
+ .long 0xf2801010 // vmov.i32 d1, #0
+ .long 0xe1a01007 // mov r1, r7
+ .long 0xf2802010 // vmov.i32 d2, #0
+ .long 0xe1a02008 // mov r2, r8
+ .long 0xf2803010 // vmov.i32 d3, #0
+ .long 0xf2804010 // vmov.i32 d4, #0
+ .long 0xf2805010 // vmov.i32 d5, #0
+ .long 0xf2806010 // vmov.i32 d6, #0
+ .long 0xf2807010 // vmov.i32 d7, #0
+ .long 0xe12fff36 // blx r6
+ .long 0xe2840004 // add r0, r4, #4
+ .long 0xe2844002 // add r4, r4, #2
+ .long 0xe1500005 // cmp r0, r5
+ .long 0x9affffef // bls 24 <sk_start_pipeline_vfp4+0x24>
+ .long 0xe1a00004 // mov r0, r4
+ .long 0xe8bd81f0 // pop {r4, r5, r6, r7, r8, pc}
+
+.globl _sk_just_return_vfp4
+_sk_just_return_vfp4:
+ .long 0xe12fff1e // bx lr
+
+.globl _sk_seed_shader_vfp4
+_sk_seed_shader_vfp4:
+ .long 0xee800b90 // vdup.32 d16, r0
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf2c3161f // vmov.i32 d17, #1056964608
+ .long 0xedd23b00 // vldr d19, [r2]
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xf2872f10 // vmov.f32 d2, #1
+ .long 0xf3fb2622 // vcvt.f32.s32 d18, d18
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf2400da1 // vadd.f32 d16, d16, d17
+ .long 0xf2803010 // vmov.i32 d3, #0
+ .long 0xf2804010 // vmov.i32 d4, #0
+ .long 0xf2021da1 // vadd.f32 d1, d18, d17
+ .long 0xf2000da3 // vadd.f32 d0, d16, d19
+ .long 0xf2805010 // vmov.i32 d5, #0
+ .long 0xf2806010 // vmov.i32 d6, #0
+ .long 0xf2807010 // vmov.i32 d7, #0
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_constant_color_vfp4
+_sk_constant_color_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf4630a0f // vld1.8 {d16-d17}, [r3]
+ .long 0xf3b40c20 // vdup.32 d0, d16[0]
+ .long 0xf3bc1c20 // vdup.32 d1, d16[1]
+ .long 0xf3b42c21 // vdup.32 d2, d17[0]
+ .long 0xf3bc3c21 // vdup.32 d3, d17[1]
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_clear_vfp4
+_sk_clear_vfp4:
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2800010 // vmov.i32 d0, #0
+ .long 0xf2801010 // vmov.i32 d1, #0
+ .long 0xf2802010 // vmov.i32 d2, #0
+ .long 0xf2803010 // vmov.i32 d3, #0
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_plus__vfp4
+_sk_plus__vfp4:
+ .long 0xf2000d04 // vadd.f32 d0, d0, d4
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2011d05 // vadd.f32 d1, d1, d5
+ .long 0xf2022d06 // vadd.f32 d2, d2, d6
+ .long 0xf2033d07 // vadd.f32 d3, d3, d7
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_srcover_vfp4
+_sk_srcover_vfp4:
+ .long 0xf2c70f10 // vmov.f32 d16, #1
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2600d83 // vsub.f32 d16, d16, d3
+ .long 0xf2040c30 // vfma.f32 d0, d4, d16
+ .long 0xf2051c30 // vfma.f32 d1, d5, d16
+ .long 0xf2062c30 // vfma.f32 d2, d6, d16
+ .long 0xf2073c30 // vfma.f32 d3, d7, d16
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_dstover_vfp4
+_sk_dstover_vfp4:
+ .long 0xf2c70f10 // vmov.f32 d16, #1
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2651115 // vorr d17, d5, d5
+ .long 0xf2604d87 // vsub.f32 d20, d16, d7
+ .long 0xf2640114 // vorr d16, d4, d4
+ .long 0xf2662116 // vorr d18, d6, d6
+ .long 0xf2673117 // vorr d19, d7, d7
+ .long 0xf2400c34 // vfma.f32 d16, d0, d20
+ .long 0xf2411c34 // vfma.f32 d17, d1, d20
+ .long 0xf2422c34 // vfma.f32 d18, d2, d20
+ .long 0xf2433c34 // vfma.f32 d19, d3, d20
+ .long 0xf22001b0 // vorr d0, d16, d16
+ .long 0xf22111b1 // vorr d1, d17, d17
+ .long 0xf22221b2 // vorr d2, d18, d18
+ .long 0xf22331b3 // vorr d3, d19, d19
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_clamp_0_vfp4
+_sk_clamp_0_vfp4:
+ .long 0xf2c00010 // vmov.i32 d16, #0
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2000f20 // vmax.f32 d0, d0, d16
+ .long 0xf2011f20 // vmax.f32 d1, d1, d16
+ .long 0xf2022f20 // vmax.f32 d2, d2, d16
+ .long 0xf2033f20 // vmax.f32 d3, d3, d16
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_clamp_1_vfp4
+_sk_clamp_1_vfp4:
+ .long 0xf2c70f10 // vmov.f32 d16, #1
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2200f20 // vmin.f32 d0, d0, d16
+ .long 0xf2211f20 // vmin.f32 d1, d1, d16
+ .long 0xf2222f20 // vmin.f32 d2, d2, d16
+ .long 0xf2233f20 // vmin.f32 d3, d3, d16
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_clamp_a_vfp4
+_sk_clamp_a_vfp4:
+ .long 0xf2c70f10 // vmov.f32 d16, #1
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf2233f20 // vmin.f32 d3, d3, d16
+ .long 0xf2200f03 // vmin.f32 d0, d0, d3
+ .long 0xf2211f03 // vmin.f32 d1, d1, d3
+ .long 0xf2222f03 // vmin.f32 d2, d2, d3
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_set_rgb_vfp4
+_sk_set_rgb_vfp4:
+ .long 0xe92d4800 // push {fp, lr}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe283e008 // add lr, r3, #8
+ .long 0xf4a30c9d // vld1.32 {d0[]}, [r3 :32]!
+ .long 0xf4ae2c9f // vld1.32 {d2[]}, [lr :32]
+ .long 0xf4a31c9f // vld1.32 {d1[]}, [r3 :32]
+ .long 0xe8bd4800 // pop {fp, lr}
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_swap_rb_vfp4
+_sk_swap_rb_vfp4:
+ .long 0xeef00b40 // vmov.f64 d16, d0
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xeeb00b42 // vmov.f64 d0, d2
+ .long 0xeeb02b60 // vmov.f64 d2, d16
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_swap_vfp4
+_sk_swap_vfp4:
+ .long 0xeef00b43 // vmov.f64 d16, d3
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xeef01b42 // vmov.f64 d17, d2
+ .long 0xeef02b41 // vmov.f64 d18, d1
+ .long 0xeef03b40 // vmov.f64 d19, d0
+ .long 0xeeb00b44 // vmov.f64 d0, d4
+ .long 0xeeb01b45 // vmov.f64 d1, d5
+ .long 0xeeb02b46 // vmov.f64 d2, d6
+ .long 0xeeb03b47 // vmov.f64 d3, d7
+ .long 0xeeb04b63 // vmov.f64 d4, d19
+ .long 0xeeb05b62 // vmov.f64 d5, d18
+ .long 0xeeb06b61 // vmov.f64 d6, d17
+ .long 0xeeb07b60 // vmov.f64 d7, d16
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_move_src_dst_vfp4
+_sk_move_src_dst_vfp4:
+ .long 0xeeb04b40 // vmov.f64 d4, d0
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xeeb05b41 // vmov.f64 d5, d1
+ .long 0xeeb06b42 // vmov.f64 d6, d2
+ .long 0xeeb07b43 // vmov.f64 d7, d3
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_move_dst_src_vfp4
+_sk_move_dst_src_vfp4:
+ .long 0xeeb00b44 // vmov.f64 d0, d4
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xeeb01b45 // vmov.f64 d1, d5
+ .long 0xeeb02b46 // vmov.f64 d2, d6
+ .long 0xeeb03b47 // vmov.f64 d3, d7
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_premul_vfp4
+_sk_premul_vfp4:
+ .long 0xf3000d13 // vmul.f32 d0, d0, d3
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf3011d13 // vmul.f32 d1, d1, d3
+ .long 0xf3022d13 // vmul.f32 d2, d2, d3
+ .long 0xe12fff13 // bx r3
+
+.globl _sk_unpremul_vfp4
+_sk_unpremul_vfp4:
+ .long 0xed2d8b04 // vpush {d8-d9}
+ .long 0xeeb78a00 // vmov.f32 s16, #112
+ .long 0xf3f91503 // vceq.f32 d17, d3, #0
+ .long 0xf2c00010 // vmov.i32 d16, #0
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xeec89a23 // vdiv.f32 s19, s16, s7
+ .long 0xee889a03 // vdiv.f32 s18, s16, s6
+ .long 0xf3501199 // vbsl d17, d16, d9
+ .long 0xf3010d90 // vmul.f32 d0, d17, d0
+ .long 0xf3011d91 // vmul.f32 d1, d17, d1
+ .long 0xf3012d92 // vmul.f32 d2, d17, d2
+ .long 0xecbd8b04 // vpop {d8-d9}
+ .long 0xe12fff13 // bx r3
+ .long 0xe320f000 // nop {0}
+
+.globl _sk_from_srgb_vfp4
+_sk_from_srgb_vfp4:
+ .long 0xeddf3b20 // vldr d19, [pc, #128]
+ .long 0xf3408d10 // vmul.f32 d24, d0, d0
+ .long 0xeddf0b1c // vldr d16, [pc, #112]
+ .long 0xf26341b3 // vorr d20, d19, d19
+ .long 0xf26351b3 // vorr d21, d19, d19
+ .long 0xeddf9b1f // vldr d25, [pc, #124]
+ .long 0xf2404c30 // vfma.f32 d20, d0, d16
+ .long 0xeddf2b1b // vldr d18, [pc, #108]
+ .long 0xf2415c30 // vfma.f32 d21, d1, d16
+ .long 0xeddfcb1d // vldr d28, [pc, #116]
+ .long 0xf2423c30 // vfma.f32 d19, d2, d16
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf3426d12 // vmul.f32 d22, d2, d2
+ .long 0xf3417d11 // vmul.f32 d23, d1, d1
+ .long 0xf3620e80 // vcgt.f32 d16, d18, d0
+ .long 0xf3621e81 // vcgt.f32 d17, d18, d1
+ .long 0xf341ad39 // vmul.f32 d26, d1, d25
+ .long 0xf342bd39 // vmul.f32 d27, d2, d25
+ .long 0xf3622e82 // vcgt.f32 d18, d18, d2
+ .long 0xf3409d39 // vmul.f32 d25, d0, d25
+ .long 0xf26cd1bc // vorr d29, d28, d28
+ .long 0xf248dcb4 // vfma.f32 d29, d24, d20
+ .long 0xf26c41bc // vorr d20, d28, d28
+ .long 0xf2474cb5 // vfma.f32 d20, d23, d21
+ .long 0xf246ccb3 // vfma.f32 d28, d22, d19
+ .long 0xf35901bd // vbsl d16, d25, d29
+ .long 0xf35a11b4 // vbsl d17, d26, d20
+ .long 0xf35b21bc // vbsl d18, d27, d28
+ .long 0xf22001b0 // vorr d0, d16, d16
+ .long 0xf22111b1 // vorr d1, d17, d17
+ .long 0xf22221b2 // vorr d2, d18, d18
+ .long 0xe12fff13 // bx r3
+ .long 0x3e99999a // .word 0x3e99999a
+ .long 0x3e99999a // .word 0x3e99999a
+ .long 0x3f328f5c // .word 0x3f328f5c
+ .long 0x3f328f5c // .word 0x3f328f5c
+ .long 0x3d6147ae // .word 0x3d6147ae
+ .long 0x3d6147ae // .word 0x3d6147ae
+ .long 0x3d9e8391 // .word 0x3d9e8391
+ .long 0x3d9e8391 // .word 0x3d9e8391
+ .long 0x3b23d70a // .word 0x3b23d70a
+ .long 0x3b23d70a // .word 0x3b23d70a
+
+.globl _sk_to_srgb_vfp4
+_sk_to_srgb_vfp4:
+ .long 0xf3fb0582 // vrsqrte.f32 d16, d2
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf3fb1581 // vrsqrte.f32 d17, d1
+ .long 0xf3fb2580 // vrsqrte.f32 d18, d0
+ .long 0xf3403db0 // vmul.f32 d19, d16, d16
+ .long 0xf3414db1 // vmul.f32 d20, d17, d17
+ .long 0xf3425db2 // vmul.f32 d21, d18, d18
+ .long 0xf2623f33 // vrsqrts.f32 d19, d2, d19
+ .long 0xf2614f34 // vrsqrts.f32 d20, d1, d20
+ .long 0xf2605f35 // vrsqrts.f32 d21, d0, d21
+ .long 0xf3400db3 // vmul.f32 d16, d16, d19
+ .long 0xf3411db4 // vmul.f32 d17, d17, d20
+ .long 0xf3422db5 // vmul.f32 d18, d18, d21
+ .long 0xf3fb3520 // vrecpe.f32 d19, d16
+ .long 0xf3fb4521 // vrecpe.f32 d20, d17
+ .long 0xf3fb6522 // vrecpe.f32 d22, d18
+ .long 0xf3fb55a0 // vrsqrte.f32 d21, d16
+ .long 0xf3fb75a1 // vrsqrte.f32 d23, d17
+ .long 0xf3fb85a2 // vrsqrte.f32 d24, d18
+ .long 0xf2409fb3 // vrecps.f32 d25, d16, d19
+ .long 0xf241afb4 // vrecps.f32 d26, d17, d20
+ .long 0xf242bfb6 // vrecps.f32 d27, d18, d22
+ .long 0xf345cdb5 // vmul.f32 d28, d21, d21
+ .long 0xf347ddb7 // vmul.f32 d29, d23, d23
+ .long 0xf348edb8 // vmul.f32 d30, d24, d24
+ .long 0xf2600fbc // vrsqrts.f32 d16, d16, d28
+ .long 0xf2611fbd // vrsqrts.f32 d17, d17, d29
+ .long 0xf2622fbe // vrsqrts.f32 d18, d18, d30
+ .long 0xf3433db9 // vmul.f32 d19, d19, d25
+ .long 0xeddf9b21 // vldr d25, [pc, #132]
+ .long 0xf3444dba // vmul.f32 d20, d20, d26
+ .long 0xeddfab21 // vldr d26, [pc, #132]
+ .long 0xf3466dbb // vmul.f32 d22, d22, d27
+ .long 0xf26ab1ba // vorr d27, d26, d26
+ .long 0xf243bcb9 // vfma.f32 d27, d19, d25
+ .long 0xf26a31ba // vorr d19, d26, d26
+ .long 0xf2443cb9 // vfma.f32 d19, d20, d25
+ .long 0xeddf4b1d // vldr d20, [pc, #116]
+ .long 0xf246acb9 // vfma.f32 d26, d22, d25
+ .long 0xf3450db0 // vmul.f32 d16, d21, d16
+ .long 0xeddf5b1c // vldr d21, [pc, #112]
+ .long 0xf3471db1 // vmul.f32 d17, d23, d17
+ .long 0xf3482db2 // vmul.f32 d18, d24, d18
+ .long 0xf3406d35 // vmul.f32 d22, d0, d21
+ .long 0xf240bcb4 // vfma.f32 d27, d16, d20
+ .long 0xf2413cb4 // vfma.f32 d19, d17, d20
+ .long 0xf242acb4 // vfma.f32 d26, d18, d20
+ .long 0xeddf2b17 // vldr d18, [pc, #92]
+ .long 0xf3417d35 // vmul.f32 d23, d1, d21
+ .long 0xf3620e80 // vcgt.f32 d16, d18, d0
+ .long 0xf3621e81 // vcgt.f32 d17, d18, d1
+ .long 0xf3622e82 // vcgt.f32 d18, d18, d2
+ .long 0xf3425d35 // vmul.f32 d21, d2, d21
+ .long 0xf2c74f10 // vmov.f32 d20, #1
+ .long 0xf2648faa // vmin.f32 d24, d20, d26
+ .long 0xf2643fa3 // vmin.f32 d19, d20, d19
+ .long 0xf2644fab // vmin.f32 d20, d20, d27
+ .long 0xf35601b8 // vbsl d16, d22, d24
+ .long 0xf35711b3 // vbsl d17, d23, d19
+ .long 0xf35521b4 // vbsl d18, d21, d20
+ .long 0xf22001b0 // vorr d0, d16, d16
+ .long 0xf22111b1 // vorr d1, d17, d17
+ .long 0xf22221b2 // vorr d2, d18, d18
+ .long 0xe12fff13 // bx r3
+ .long 0x3f306fce // .word 0x3f306fce
+ .long 0x3f306fce // .word 0x3f306fce
+ .long 0xbdca57a8 // .word 0xbdca57a8
+ .long 0xbdca57a8 // .word 0xbdca57a8
+ .long 0x3ed287c2 // .word 0x3ed287c2
+ .long 0x3ed287c2 // .word 0x3ed287c2
+ .long 0x41475c29 // .word 0x41475c29
+ .long 0x41475c29 // .word 0x41475c29
+ .long 0x3b8ce704 // .word 0x3b8ce704
+ .long 0x3b8ce704 // .word 0x3b8ce704
+
+.globl _sk_scale_1_float_vfp4
+_sk_scale_1_float_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
+ .long 0xf3000d90 // vmul.f32 d0, d16, d0
+ .long 0xf3001d91 // vmul.f32 d1, d16, d1
+ .long 0xf3002d92 // vmul.f32 d2, d16, d2
+ .long 0xf3003d93 // vmul.f32 d3, d16, d3
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_scale_u8_vfp4
+_sk_scale_u8_vfp4:
+ .long 0xe24dd004 // sub sp, sp, #4
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xe0833000 // add r3, r3, r0
+ .long 0xe1d330b0 // ldrh r3, [r3]
+ .long 0xe1cd30b0 // strh r3, [sp]
+ .long 0xe1a0300d // mov r3, sp
+ .long 0xf4e3041f // vld1.16 {d16[0]}, [r3 :16]
+ .long 0xf3c80a30 // vmovl.u8 q8, d16
+ .long 0xf3d00a30 // vmovl.u16 q8, d16
+ .long 0xf3fb06a0 // vcvt.f32.u32 d16, d16
+ .long 0xeddf1b06 // vldr d17, [pc, #24]
+ .long 0xf3400db1 // vmul.f32 d16, d16, d17
+ .long 0xf3000d90 // vmul.f32 d0, d16, d0
+ .long 0xf3001d91 // vmul.f32 d1, d16, d1
+ .long 0xf3002d92 // vmul.f32 d2, d16, d2
+ .long 0xf3003d93 // vmul.f32 d3, d16, d3
+ .long 0xe28dd004 // add sp, sp, #4
+ .long 0xe12fff1c // bx ip
+ .long 0x3b808081 // .word 0x3b808081
+ .long 0x3b808081 // .word 0x3b808081
+
+.globl _sk_lerp_1_float_vfp4
+_sk_lerp_1_float_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2600d04 // vsub.f32 d16, d0, d4
+ .long 0xf2611d05 // vsub.f32 d17, d1, d5
+ .long 0xf2622d06 // vsub.f32 d18, d2, d6
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf2633d07 // vsub.f32 d19, d3, d7
+ .long 0xf4e34c9f // vld1.32 {d20[]}, [r3 :32]
+ .long 0xf2240114 // vorr d0, d4, d4
+ .long 0xf2251115 // vorr d1, d5, d5
+ .long 0xf2262116 // vorr d2, d6, d6
+ .long 0xf2273117 // vorr d3, d7, d7
+ .long 0xf2000cb4 // vfma.f32 d0, d16, d20
+ .long 0xf2011cb4 // vfma.f32 d1, d17, d20
+ .long 0xf2022cb4 // vfma.f32 d2, d18, d20
+ .long 0xf2033cb4 // vfma.f32 d3, d19, d20
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_lerp_u8_vfp4
+_sk_lerp_u8_vfp4:
+ .long 0xe24dd004 // sub sp, sp, #4
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2602d04 // vsub.f32 d18, d0, d4
+ .long 0xf2623d06 // vsub.f32 d19, d2, d6
+ .long 0xf2634d07 // vsub.f32 d20, d3, d7
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2240114 // vorr d0, d4, d4
+ .long 0xf2262116 // vorr d2, d6, d6
+ .long 0xe0833000 // add r3, r3, r0
+ .long 0xf2273117 // vorr d3, d7, d7
+ .long 0xe1d330b0 // ldrh r3, [r3]
+ .long 0xe1cd30b0 // strh r3, [sp]
+ .long 0xe1a0300d // mov r3, sp
+ .long 0xf4e3041f // vld1.16 {d16[0]}, [r3 :16]
+ .long 0xf3c80a30 // vmovl.u8 q8, d16
+ .long 0xf3d00a30 // vmovl.u16 q8, d16
+ .long 0xf3fb06a0 // vcvt.f32.u32 d16, d16
+ .long 0xeddf1b08 // vldr d17, [pc, #32]
+ .long 0xf3400db1 // vmul.f32 d16, d16, d17
+ .long 0xf2611d05 // vsub.f32 d17, d1, d5
+ .long 0xf2251115 // vorr d1, d5, d5
+ .long 0xf2020cb0 // vfma.f32 d0, d18, d16
+ .long 0xf2011cb0 // vfma.f32 d1, d17, d16
+ .long 0xf2032cb0 // vfma.f32 d2, d19, d16
+ .long 0xf2043cb0 // vfma.f32 d3, d20, d16
+ .long 0xe28dd004 // add sp, sp, #4
+ .long 0xe12fff1c // bx ip
+ .long 0x3b808081 // .word 0x3b808081
+ .long 0x3b808081 // .word 0x3b808081
+
+.globl _sk_lerp_565_vfp4
+_sk_lerp_565_vfp4:
+ .long 0xe24dd004 // sub sp, sp, #4
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf3c72218 // vmov.i32 d18, #63488
+ .long 0xf2c1101f // vmov.i32 d17, #31
+ .long 0xf2603d04 // vsub.f32 d19, d0, d4
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2616d05 // vsub.f32 d22, d1, d5
+ .long 0xf2240114 // vorr d0, d4, d4
+ .long 0xf2251115 // vorr d1, d5, d5
+ .long 0xe7933080 // ldr r3, [r3, r0, lsl #1]
+ .long 0xf2873f10 // vmov.f32 d3, #1
+ .long 0xe58d3000 // str r3, [sp]
+ .long 0xe1a0300d // mov r3, sp
+ .long 0xf4e3083f // vld1.32 {d16[0]}, [r3 :32]
+ .long 0xe3a03e7e // mov r3, #2016
+ .long 0xf3d04a30 // vmovl.u16 q10, d16
+ .long 0xee803b90 // vdup.32 d16, r3
+ .long 0xf24421b2 // vand d18, d20, d18
+ .long 0xf24411b1 // vand d17, d20, d17
+ .long 0xeddf5b12 // vldr d21, [pc, #72]
+ .long 0xf24401b0 // vand d16, d20, d16
+ .long 0xeddf4b0e // vldr d20, [pc, #56]
+ .long 0xf3fb2622 // vcvt.f32.s32 d18, d18
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3fb1621 // vcvt.f32.s32 d17, d17
+ .long 0xf3422db4 // vmul.f32 d18, d18, d20
+ .long 0xeddf4b0d // vldr d20, [pc, #52]
+ .long 0xf3400db5 // vmul.f32 d16, d16, d21
+ .long 0xf2625d06 // vsub.f32 d21, d2, d6
+ .long 0xf3411db4 // vmul.f32 d17, d17, d20
+ .long 0xf2262116 // vorr d2, d6, d6
+ .long 0xf2030cb2 // vfma.f32 d0, d19, d18
+ .long 0xf2061cb0 // vfma.f32 d1, d22, d16
+ .long 0xf2052cb1 // vfma.f32 d2, d21, d17
+ .long 0xe28dd004 // add sp, sp, #4
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x37842108 // .word 0x37842108
+ .long 0x37842108 // .word 0x37842108
+ .long 0x3a020821 // .word 0x3a020821
+ .long 0x3a020821 // .word 0x3a020821
+ .long 0x3d042108 // .word 0x3d042108
+ .long 0x3d042108 // .word 0x3d042108
+
+.globl _sk_load_tables_vfp4
+_sk_load_tables_vfp4:
+ .long 0xe92d48f0 // push {r4, r5, r6, r7, fp, lr}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf3c7001f // vmov.i32 d16, #255
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe593e000 // ldr lr, [r3]
+ .long 0xe99300b0 // ldmib r3, {r4, r5, r7}
+ .long 0xe08e3100 // add r3, lr, r0, lsl #2
+ .long 0xedd31b00 // vldr d17, [r3]
+ .long 0xf24121b0 // vand d18, d17, d16
+ .long 0xf3f83031 // vshr.u32 d19, d17, #8
+ .long 0xee323b90 // vmov.32 r3, d18[1]
+ .long 0xee126b90 // vmov.32 r6, d18[0]
+ .long 0xf3f02031 // vshr.u32 d18, d17, #16
+ .long 0xf24221b0 // vand d18, d18, d16
+ .long 0xf24301b0 // vand d16, d19, d16
+ .long 0xe0843103 // add r3, r4, r3, lsl #2
+ .long 0xedd30a00 // vldr s1, [r3]
+ .long 0xe0843106 // add r3, r4, r6, lsl #2
+ .long 0xee326b90 // vmov.32 r6, d18[1]
+ .long 0xed930a00 // vldr s0, [r3]
+ .long 0xee303b90 // vmov.32 r3, d16[1]
+ .long 0xee104b90 // vmov.32 r4, d16[0]
+ .long 0xf3e80031 // vshr.u32 d16, d17, #24
+ .long 0xeddf1b0d // vldr d17, [pc, #52]
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3003db1 // vmul.f32 d3, d16, d17
+ .long 0xe087e106 // add lr, r7, r6, lsl #2
+ .long 0xee126b90 // vmov.32 r6, d18[0]
+ .long 0xe0853103 // add r3, r5, r3, lsl #2
+ .long 0xedde2a00 // vldr s5, [lr]
+ .long 0xedd31a00 // vldr s3, [r3]
+ .long 0xe0853104 // add r3, r5, r4, lsl #2
+ .long 0xed931a00 // vldr s2, [r3]
+ .long 0xe0873106 // add r3, r7, r6, lsl #2
+ .long 0xed932a00 // vldr s4, [r3]
+ .long 0xe8bd48f0 // pop {r4, r5, r6, r7, fp, lr}
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3b808081 // .word 0x3b808081
+ .long 0x3b808081 // .word 0x3b808081
+
+.globl _sk_load_a8_vfp4
+_sk_load_a8_vfp4:
+ .long 0xe24dd004 // sub sp, sp, #4
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf2800010 // vmov.i32 d0, #0
+ .long 0xf2801010 // vmov.i32 d1, #0
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2802010 // vmov.i32 d2, #0
+ .long 0xe0833000 // add r3, r3, r0
+ .long 0xe1d330b0 // ldrh r3, [r3]
+ .long 0xe1cd30b0 // strh r3, [sp]
+ .long 0xe1a0300d // mov r3, sp
+ .long 0xf4e3041f // vld1.16 {d16[0]}, [r3 :16]
+ .long 0xf3c80a30 // vmovl.u8 q8, d16
+ .long 0xf3d00a30 // vmovl.u16 q8, d16
+ .long 0xf3fb06a0 // vcvt.f32.u32 d16, d16
+ .long 0xeddf1b03 // vldr d17, [pc, #12]
+ .long 0xf3003db1 // vmul.f32 d3, d16, d17
+ .long 0xe28dd004 // add sp, sp, #4
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3b808081 // .word 0x3b808081
+ .long 0x3b808081 // .word 0x3b808081
+
+.globl _sk_store_a8_vfp4
+_sk_store_a8_vfp4:
+ .long 0xe92d4800 // push {fp, lr}
+ .long 0xeddf0b0d // vldr d16, [pc, #52]
+ .long 0xf2c3161f // vmov.i32 d17, #1056964608
+ .long 0xf2431c30 // vfma.f32 d17, d3, d16
+ .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf3fb07a1 // vcvt.u32.f32 d16, d17
+ .long 0xee10eb90 // vmov.32 lr, d16[0]
+ .long 0xee30cb90 // vmov.32 ip, d16[1]
+ .long 0xe7e3e000 // strb lr, [r3, r0]!
+ .long 0xe5c3c001 // strb ip, [r3, #1]
+ .long 0xe2813008 // add r3, r1, #8
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe1a01003 // mov r1, r3
+ .long 0xe8bd4800 // pop {fp, lr}
+ .long 0xe12fff1c // bx ip
+ .long 0x437f0000 // .word 0x437f0000
+ .long 0x437f0000 // .word 0x437f0000
+
+.globl _sk_load_565_vfp4
+_sk_load_565_vfp4:
+ .long 0xe24dd004 // sub sp, sp, #4
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c1101f // vmov.i32 d17, #31
+ .long 0xf3c72218 // vmov.i32 d18, #63488
+ .long 0xeddf3b16 // vldr d19, [pc, #88]
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2873f10 // vmov.f32 d3, #1
+ .long 0xe7933080 // ldr r3, [r3, r0, lsl #1]
+ .long 0xe58d3000 // str r3, [sp]
+ .long 0xe1a0300d // mov r3, sp
+ .long 0xf4e3083f // vld1.32 {d16[0]}, [r3 :32]
+ .long 0xe3a03e7e // mov r3, #2016
+ .long 0xf3d04a30 // vmovl.u16 q10, d16
+ .long 0xee803b90 // vdup.32 d16, r3
+ .long 0xf24411b1 // vand d17, d20, d17
+ .long 0xeddf5b0e // vldr d21, [pc, #56]
+ .long 0xf24421b2 // vand d18, d20, d18
+ .long 0xf24401b0 // vand d16, d20, d16
+ .long 0xeddf4b09 // vldr d20, [pc, #36]
+ .long 0xf3fb2622 // vcvt.f32.s32 d18, d18
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3fb1621 // vcvt.f32.s32 d17, d17
+ .long 0xf3020db3 // vmul.f32 d0, d18, d19
+ .long 0xf3001db4 // vmul.f32 d1, d16, d20
+ .long 0xf3012db5 // vmul.f32 d2, d17, d21
+ .long 0xe28dd004 // add sp, sp, #4
+ .long 0xe12fff1c // bx ip
+ .long 0x37842108 // .word 0x37842108
+ .long 0x37842108 // .word 0x37842108
+ .long 0x3a020821 // .word 0x3a020821
+ .long 0x3a020821 // .word 0x3a020821
+ .long 0x3d042108 // .word 0x3d042108
+ .long 0x3d042108 // .word 0x3d042108
+
+.globl _sk_store_565_vfp4
+_sk_store_565_vfp4:
+ .long 0xf2c30f1f // vmov.f32 d16, #31
+ .long 0xeddf1b15 // vldr d17, [pc, #84]
+ .long 0xf2c3361f // vmov.i32 d19, #1056964608
+ .long 0xe5913000 // ldr r3, [r1]
+ .long 0xf2413c31 // vfma.f32 d19, d1, d17
+ .long 0xf2c3161f // vmov.i32 d17, #1056964608
+ .long 0xf2401c30 // vfma.f32 d17, d0, d16
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2c3261f // vmov.i32 d18, #1056964608
+ .long 0xf2422c30 // vfma.f32 d18, d2, d16
+ .long 0xe0833080 // add r3, r3, r0, lsl #1
+ .long 0xf3fb07a3 // vcvt.u32.f32 d16, d19
+ .long 0xf3fb17a1 // vcvt.u32.f32 d17, d17
+ .long 0xf3fb27a2 // vcvt.u32.f32 d18, d18
+ .long 0xf2e50530 // vshl.s32 d16, d16, #5
+ .long 0xf2eb1531 // vshl.s32 d17, d17, #11
+ .long 0xf26001b1 // vorr d16, d16, d17
+ .long 0xf26001b2 // vorr d16, d16, d18
+ .long 0xf3f60121 // vuzp.16 d16, d17
+ .long 0xf4c3080f // vst1.32 {d16[0]}, [r3]
+ .long 0xe2813008 // add r3, r1, #8
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe1a01003 // mov r1, r3
+ .long 0xe12fff1c // bx ip
+ .long 0x427c0000 // .word 0x427c0000
+ .long 0x427c0000 // .word 0x427c0000
+
+.globl _sk_load_8888_vfp4
+_sk_load_8888_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf3c7001f // vmov.i32 d16, #255
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xe0833100 // add r3, r3, r0, lsl #2
+ .long 0xedd31b00 // vldr d17, [r3]
+ .long 0xf24121b0 // vand d18, d17, d16
+ .long 0xf3f83031 // vshr.u32 d19, d17, #8
+ .long 0xf3e84031 // vshr.u32 d20, d17, #24
+ .long 0xf3f01031 // vshr.u32 d17, d17, #16
+ .long 0xf24331b0 // vand d19, d19, d16
+ .long 0xf24101b0 // vand d16, d17, d16
+ .long 0xeddf1b08 // vldr d17, [pc, #32]
+ .long 0xf3fb2622 // vcvt.f32.s32 d18, d18
+ .long 0xf3fb4624 // vcvt.f32.s32 d20, d20
+ .long 0xf3fb3623 // vcvt.f32.s32 d19, d19
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3020db1 // vmul.f32 d0, d18, d17
+ .long 0xf3043db1 // vmul.f32 d3, d20, d17
+ .long 0xf3031db1 // vmul.f32 d1, d19, d17
+ .long 0xf3002db1 // vmul.f32 d2, d16, d17
+ .long 0xe12fff1c // bx ip
+ .long 0x3b808081 // .word 0x3b808081
+ .long 0x3b808081 // .word 0x3b808081
+
+.globl _sk_store_8888_vfp4
+_sk_store_8888_vfp4:
+ .long 0xeddf0b1a // vldr d16, [pc, #104]
+ .long 0xf2c3261f // vmov.i32 d18, #1056964608
+ .long 0xf2412c30 // vfma.f32 d18, d1, d16
+ .long 0xe5913000 // ldr r3, [r1]
+ .long 0xf2c3361f // vmov.i32 d19, #1056964608
+ .long 0xf2c3161f // vmov.i32 d17, #1056964608
+ .long 0xf2423c30 // vfma.f32 d19, d2, d16
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xf2c3461f // vmov.i32 d20, #1056964608
+ .long 0xf2401c30 // vfma.f32 d17, d0, d16
+ .long 0xe0833100 // add r3, r3, r0, lsl #2
+ .long 0xf2434c30 // vfma.f32 d20, d3, d16
+ .long 0xf3fb07a2 // vcvt.u32.f32 d16, d18
+ .long 0xf3fb27a3 // vcvt.u32.f32 d18, d19
+ .long 0xf3fb17a1 // vcvt.u32.f32 d17, d17
+ .long 0xf3fb37a4 // vcvt.u32.f32 d19, d20
+ .long 0xf2e80530 // vshl.s32 d16, d16, #8
+ .long 0xf2f02532 // vshl.s32 d18, d18, #16
+ .long 0xf26001b1 // vorr d16, d16, d17
+ .long 0xf2f81533 // vshl.s32 d17, d19, #24
+ .long 0xf26001b2 // vorr d16, d16, d18
+ .long 0xf26001b1 // vorr d16, d16, d17
+ .long 0xedc30b00 // vstr d16, [r3]
+ .long 0xe2813008 // add r3, r1, #8
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe1a01003 // mov r1, r3
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x437f0000 // .word 0x437f0000
+ .long 0x437f0000 // .word 0x437f0000
+
+.globl _sk_load_f16_vfp4
+_sk_load_f16_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xe0833180 // add r3, r3, r0, lsl #3
+ .long 0xf463084f // vld2.16 {d16-d17}, [r3]
+ .long 0xf3f62720 // vcvt.f32.f16 q9, d16
+ .long 0xf3f60721 // vcvt.f32.f16 q8, d17
+ .long 0xf22201b2 // vorr d0, d18, d18
+ .long 0xf22011b0 // vorr d1, d16, d16
+ .long 0xf3ba00a3 // vtrn.32 d0, d19
+ .long 0xf22321b3 // vorr d2, d19, d19
+ .long 0xf3ba10a1 // vtrn.32 d1, d17
+ .long 0xf22131b1 // vorr d3, d17, d17
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_store_f16_vfp4
+_sk_store_f16_vfp4:
+ .long 0xeef00b41 // vmov.f64 d16, d1
+ .long 0xeef03b42 // vmov.f64 d19, d2
+ .long 0xf2631113 // vorr d17, d3, d3
+ .long 0xf2602110 // vorr d18, d0, d0
+ .long 0xf3fa00a1 // vtrn.32 d16, d17
+ .long 0xf3f61620 // vcvt.f16.f32 d17, q8
+ .long 0xf3fa20a3 // vtrn.32 d18, d19
+ .long 0xe5913000 // ldr r3, [r1]
+ .long 0xf3f60622 // vcvt.f16.f32 d16, q9
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xe0833180 // add r3, r3, r0, lsl #3
+ .long 0xf443084f // vst2.16 {d16-d17}, [r3]
+ .long 0xe2813008 // add r3, r1, #8
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe1a01003 // mov r1, r3
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_store_f32_vfp4
+_sk_store_f32_vfp4:
+ .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe5933000 // ldr r3, [r3]
+ .long 0xe0833200 // add r3, r3, r0, lsl #4
+ .long 0xf403008f // vst4.32 {d0-d3}, [r3]
+ .long 0xe2813008 // add r3, r1, #8
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe1a01003 // mov r1, r3
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_clamp_x_vfp4
+_sk_clamp_x_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c00010 // vmov.i32 d16, #0
+ .long 0xf3c71e1f // vmov.i8 d17, #255
+ .long 0xf2400f80 // vmax.f32 d16, d16, d0
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xf26218a1 // vadd.i32 d17, d18, d17
+ .long 0xf2200fa1 // vmin.f32 d0, d16, d17
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_clamp_y_vfp4
+_sk_clamp_y_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c00010 // vmov.i32 d16, #0
+ .long 0xf3c71e1f // vmov.i8 d17, #255
+ .long 0xf2400f81 // vmax.f32 d16, d16, d1
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xf26218a1 // vadd.i32 d17, d18, d17
+ .long 0xf2201fa1 // vmin.f32 d1, d16, d17
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_repeat_x_vfp4
+_sk_repeat_x_vfp4:
+ .long 0xed2d8b04 // vpush {d8-d9}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c02010 // vmov.i32 d18, #0
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xeddf3b10 // vldr d19, [pc, #64]
+ .long 0xed938a00 // vldr s16, [r3]
+ .long 0xeec09a88 // vdiv.f32 s19, s1, s16
+ .long 0xee809a08 // vdiv.f32 s18, s0, s16
+ .long 0xf3fb0709 // vcvt.s32.f32 d16, d9
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3601e89 // vcgt.f32 d17, d16, d9
+ .long 0xf35311b2 // vbsl d17, d19, d18
+ .long 0xf3f42c08 // vdup.32 d18, d8[0]
+ .long 0xf2600da1 // vsub.f32 d16, d16, d17
+ .long 0xf3c71e1f // vmov.i8 d17, #255
+ .long 0xf26218a1 // vadd.i32 d17, d18, d17
+ .long 0xf2e009c8 // vmul.f32 d16, d16, d8[0]
+ .long 0xf2600d20 // vsub.f32 d16, d0, d16
+ .long 0xf2200fa1 // vmin.f32 d0, d16, d17
+ .long 0xecbd8b04 // vpop {d8-d9}
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3f800000 // .word 0x3f800000
+ .long 0x3f800000 // .word 0x3f800000
+
+.globl _sk_repeat_y_vfp4
+_sk_repeat_y_vfp4:
+ .long 0xed2d8b04 // vpush {d8-d9}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c02010 // vmov.i32 d18, #0
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xeddf3b10 // vldr d19, [pc, #64]
+ .long 0xed938a00 // vldr s16, [r3]
+ .long 0xeec19a88 // vdiv.f32 s19, s3, s16
+ .long 0xee819a08 // vdiv.f32 s18, s2, s16
+ .long 0xf3fb0709 // vcvt.s32.f32 d16, d9
+ .long 0xf3fb0620 // vcvt.f32.s32 d16, d16
+ .long 0xf3601e89 // vcgt.f32 d17, d16, d9
+ .long 0xf35311b2 // vbsl d17, d19, d18
+ .long 0xf3f42c08 // vdup.32 d18, d8[0]
+ .long 0xf2600da1 // vsub.f32 d16, d16, d17
+ .long 0xf3c71e1f // vmov.i8 d17, #255
+ .long 0xf26218a1 // vadd.i32 d17, d18, d17
+ .long 0xf2e009c8 // vmul.f32 d16, d16, d8[0]
+ .long 0xf2610d20 // vsub.f32 d16, d1, d16
+ .long 0xf2201fa1 // vmin.f32 d1, d16, d17
+ .long 0xecbd8b04 // vpop {d8-d9}
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3f800000 // .word 0x3f800000
+ .long 0x3f800000 // .word 0x3f800000
+
+.globl _sk_mirror_x_vfp4
+_sk_mirror_x_vfp4:
+ .long 0xed2d8b04 // vpush {d8-d9}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c03010 // vmov.i32 d19, #0
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xeddf4b14 // vldr d20, [pc, #80]
+ .long 0xed938a00 // vldr s16, [r3]
+ .long 0xee389a08 // vadd.f32 s18, s16, s16
+ .long 0xf3f40c08 // vdup.32 d16, d8[0]
+ .long 0xf2200d20 // vsub.f32 d0, d0, d16
+ .long 0xeec08a89 // vdiv.f32 s17, s1, s18
+ .long 0xee808a09 // vdiv.f32 s16, s0, s18
+ .long 0xf3fb1708 // vcvt.s32.f32 d17, d8
+ .long 0xf3fb1621 // vcvt.f32.s32 d17, d17
+ .long 0xf3612e88 // vcgt.f32 d18, d17, d8
+ .long 0xf35421b3 // vbsl d18, d20, d19
+ .long 0xf2611da2 // vsub.f32 d17, d17, d18
+ .long 0xf3c72e1f // vmov.i8 d18, #255
+ .long 0xf2e119c9 // vmul.f32 d17, d17, d9[0]
+ .long 0xf2601d21 // vsub.f32 d17, d0, d17
+ .long 0xf2611da0 // vsub.f32 d17, d17, d16
+ .long 0xf26008a2 // vadd.i32 d16, d16, d18
+ .long 0xf3f91721 // vabs.f32 d17, d17
+ .long 0xf2210fa0 // vmin.f32 d0, d17, d16
+ .long 0xecbd8b04 // vpop {d8-d9}
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3f800000 // .word 0x3f800000
+ .long 0x3f800000 // .word 0x3f800000
+
+.globl _sk_mirror_y_vfp4
+_sk_mirror_y_vfp4:
+ .long 0xed2d8b04 // vpush {d8-d9}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2c03010 // vmov.i32 d19, #0
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xeddf4b14 // vldr d20, [pc, #80]
+ .long 0xed938a00 // vldr s16, [r3]
+ .long 0xee389a08 // vadd.f32 s18, s16, s16
+ .long 0xf3f40c08 // vdup.32 d16, d8[0]
+ .long 0xf2211d20 // vsub.f32 d1, d1, d16
+ .long 0xeec18a89 // vdiv.f32 s17, s3, s18
+ .long 0xee818a09 // vdiv.f32 s16, s2, s18
+ .long 0xf3fb1708 // vcvt.s32.f32 d17, d8
+ .long 0xf3fb1621 // vcvt.f32.s32 d17, d17
+ .long 0xf3612e88 // vcgt.f32 d18, d17, d8
+ .long 0xf35421b3 // vbsl d18, d20, d19
+ .long 0xf2611da2 // vsub.f32 d17, d17, d18
+ .long 0xf3c72e1f // vmov.i8 d18, #255
+ .long 0xf2e119c9 // vmul.f32 d17, d17, d9[0]
+ .long 0xf2611d21 // vsub.f32 d17, d1, d17
+ .long 0xf2611da0 // vsub.f32 d17, d17, d16
+ .long 0xf26008a2 // vadd.i32 d16, d16, d18
+ .long 0xf3f91721 // vabs.f32 d17, d17
+ .long 0xf2211fa0 // vmin.f32 d1, d17, d16
+ .long 0xecbd8b04 // vpop {d8-d9}
+ .long 0xe12fff1c // bx ip
+ .long 0xe320f000 // nop {0}
+ .long 0x3f800000 // .word 0x3f800000
+ .long 0x3f800000 // .word 0x3f800000
+
+.globl _sk_luminance_to_alpha_vfp4
+_sk_luminance_to_alpha_vfp4:
+ .long 0xeddf0b0a // vldr d16, [pc, #40]
+ .long 0xeddf1b0b // vldr d17, [pc, #44]
+ .long 0xf3410d30 // vmul.f32 d16, d1, d16
+ .long 0xe4913004 // ldr r3, [r1], #4
+ .long 0xf3401d31 // vmul.f32 d17, d0, d17
+ .long 0xf2800010 // vmov.i32 d0, #0
+ .long 0xf2801010 // vmov.i32 d1, #0
+ .long 0xf2013da0 // vadd.f32 d3, d17, d16
+ .long 0xeddf0b06 // vldr d16, [pc, #24]
+ .long 0xf2023c30 // vfma.f32 d3, d2, d16
+ .long 0xf2802010 // vmov.i32 d2, #0
+ .long 0xe12fff13 // bx r3
+ .long 0x3f371759 // .word 0x3f371759
+ .long 0x3f371759 // .word 0x3f371759
+ .long 0x3e59b3d0 // .word 0x3e59b3d0
+ .long 0x3e59b3d0 // .word 0x3e59b3d0
+ .long 0x3d93dd98 // .word 0x3d93dd98
+ .long 0x3d93dd98 // .word 0x3d93dd98
+
+.globl _sk_matrix_2x3_vfp4
+_sk_matrix_2x3_vfp4:
+ .long 0xe92d4800 // push {fp, lr}
+ .long 0xe591e000 // ldr lr, [r1]
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe28e300c // add r3, lr, #12
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xe28e3008 // add r3, lr, #8
+ .long 0xf4e31c9f // vld1.32 {d17[]}, [r3 :32]
+ .long 0xe28e3010 // add r3, lr, #16
+ .long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
+ .long 0xe28e3014 // add r3, lr, #20
+ .long 0xf2410c31 // vfma.f32 d16, d1, d17
+ .long 0xf4e31c9f // vld1.32 {d17[]}, [r3 :32]
+ .long 0xf2411c32 // vfma.f32 d17, d1, d18
+ .long 0xf4ee2c9d // vld1.32 {d18[]}, [lr :32]!
+ .long 0xf4ee3c9f // vld1.32 {d19[]}, [lr :32]
+ .long 0xf2400c32 // vfma.f32 d16, d0, d18
+ .long 0xf2401c33 // vfma.f32 d17, d0, d19
+ .long 0xf22001b0 // vorr d0, d16, d16
+ .long 0xf22111b1 // vorr d1, d17, d17
+ .long 0xe8bd4800 // pop {fp, lr}
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_matrix_3x4_vfp4
+_sk_matrix_3x4_vfp4:
+ .long 0xe92d4800 // push {fp, lr}
+ .long 0xe591e000 // ldr lr, [r1]
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe28e3020 // add r3, lr, #32
+ .long 0xf4e33c9f // vld1.32 {d19[]}, [r3 :32]
+ .long 0xe28e302c // add r3, lr, #44
+ .long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
+ .long 0xe28e301c // add r3, lr, #28
+ .long 0xf2420c33 // vfma.f32 d16, d2, d19
+ .long 0xf4e34c9f // vld1.32 {d20[]}, [r3 :32]
+ .long 0xe28e3018 // add r3, lr, #24
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xe28e3024 // add r3, lr, #36
+ .long 0xf4e31c9f // vld1.32 {d17[]}, [r3 :32]
+ .long 0xe28e3028 // add r3, lr, #40
+ .long 0xf2421c32 // vfma.f32 d17, d2, d18
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xe28e3010 // add r3, lr, #16
+ .long 0xf2422c34 // vfma.f32 d18, d2, d20
+ .long 0xf4e33c9f // vld1.32 {d19[]}, [r3 :32]
+ .long 0xe28e300c // add r3, lr, #12
+ .long 0xf4e34c9f // vld1.32 {d20[]}, [r3 :32]
+ .long 0xe28e3014 // add r3, lr, #20
+ .long 0xf2411c34 // vfma.f32 d17, d1, d20
+ .long 0xf4e34c9f // vld1.32 {d20[]}, [r3 :32]
+ .long 0xf2410c34 // vfma.f32 d16, d1, d20
+ .long 0xe28e3008 // add r3, lr, #8
+ .long 0xf2412c33 // vfma.f32 d18, d1, d19
+ .long 0xf4ee3c9d // vld1.32 {d19[]}, [lr :32]!
+ .long 0xf4ee4c9f // vld1.32 {d20[]}, [lr :32]
+ .long 0xf2401c33 // vfma.f32 d17, d0, d19
+ .long 0xf4e33c9f // vld1.32 {d19[]}, [r3 :32]
+ .long 0xf2400c33 // vfma.f32 d16, d0, d19
+ .long 0xf2402c34 // vfma.f32 d18, d0, d20
+ .long 0xf22101b1 // vorr d0, d17, d17
+ .long 0xf22021b0 // vorr d2, d16, d16
+ .long 0xf22211b2 // vorr d1, d18, d18
+ .long 0xe8bd4800 // pop {fp, lr}
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_matrix_4x5_vfp4
+_sk_matrix_4x5_vfp4:
+ .long 0xe92d4010 // push {r4, lr}
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xf2620112 // vorr d16, d2, d2
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe2834014 // add r4, r3, #20
+ .long 0xe1a0e003 // mov lr, r3
+ .long 0xf4e45c9f // vld1.32 {d21[]}, [r4 :32]
+ .long 0xe2834028 // add r4, r3, #40
+ .long 0xf4e46c9f // vld1.32 {d22[]}, [r4 :32]
+ .long 0xe2834038 // add r4, r3, #56
+ .long 0xf4e47c9f // vld1.32 {d23[]}, [r4 :32]
+ .long 0xe2834048 // add r4, r3, #72
+ .long 0xf4a42c9f // vld1.32 {d2[]}, [r4 :32]
+ .long 0xe2834034 // add r4, r3, #52
+ .long 0xf2032c37 // vfma.f32 d2, d3, d23
+ .long 0xf4e48c9f // vld1.32 {d24[]}, [r4 :32]
+ .long 0xe2834044 // add r4, r3, #68
+ .long 0xf4e41c9f // vld1.32 {d17[]}, [r4 :32]
+ .long 0xe2834030 // add r4, r3, #48
+ .long 0xf2431c38 // vfma.f32 d17, d3, d24
+ .long 0xf4e49c9f // vld1.32 {d25[]}, [r4 :32]
+ .long 0xe283403c // add r4, r3, #60
+ .long 0xf4e43c9f // vld1.32 {d19[]}, [r4 :32]
+ .long 0xe283404c // add r4, r3, #76
+ .long 0xf2002cb6 // vfma.f32 d2, d16, d22
+ .long 0xf4e42c9f // vld1.32 {d18[]}, [r4 :32]
+ .long 0xe2834040 // add r4, r3, #64
+ .long 0xf2432c33 // vfma.f32 d18, d3, d19
+ .long 0xf4e43c9f // vld1.32 {d19[]}, [r4 :32]
+ .long 0xe2834020 // add r4, r3, #32
+ .long 0xf2433c39 // vfma.f32 d19, d3, d25
+ .long 0xf4e47c9f // vld1.32 {d23[]}, [r4 :32]
+ .long 0xe283402c // add r4, r3, #44
+ .long 0xf4e48c9f // vld1.32 {d24[]}, [r4 :32]
+ .long 0xe2834024 // add r4, r3, #36
+ .long 0xf2402cb8 // vfma.f32 d18, d16, d24
+ .long 0xf4e48c9f // vld1.32 {d24[]}, [r4 :32]
+ .long 0xf2401cb8 // vfma.f32 d17, d16, d24
+ .long 0xe2834010 // add r4, r3, #16
+ .long 0xf2403cb7 // vfma.f32 d19, d16, d23
+ .long 0xf4ee4c9d // vld1.32 {d20[]}, [lr :32]!
+ .long 0xf4e40c9f // vld1.32 {d16[]}, [r4 :32]
+ .long 0xe283401c // add r4, r3, #28
+ .long 0xf4e46c9f // vld1.32 {d22[]}, [r4 :32]
+ .long 0xe2834018 // add r4, r3, #24
+ .long 0xf2412c36 // vfma.f32 d18, d1, d22
+ .long 0xf2411c35 // vfma.f32 d17, d1, d21
+ .long 0xf4ee5c9f // vld1.32 {d21[]}, [lr :32]
+ .long 0xf2413c30 // vfma.f32 d19, d1, d16
+ .long 0xf4e40c9f // vld1.32 {d16[]}, [r4 :32]
+ .long 0xe2834008 // add r4, r3, #8
+ .long 0xe283300c // add r3, r3, #12
+ .long 0xf2012c30 // vfma.f32 d2, d1, d16
+ .long 0xf4e40c9f // vld1.32 {d16[]}, [r4 :32]
+ .long 0xf2401c35 // vfma.f32 d17, d0, d21
+ .long 0xf2403c34 // vfma.f32 d19, d0, d20
+ .long 0xf4e34c9f // vld1.32 {d20[]}, [r3 :32]
+ .long 0xf2402c34 // vfma.f32 d18, d0, d20
+ .long 0xf2002c30 // vfma.f32 d2, d0, d16
+ .long 0xf22111b1 // vorr d1, d17, d17
+ .long 0xf22301b3 // vorr d0, d19, d19
+ .long 0xf22231b2 // vorr d3, d18, d18
+ .long 0xe8bd4010 // pop {r4, lr}
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_matrix_perspective_vfp4
+_sk_matrix_perspective_vfp4:
+ .long 0xe92d4010 // push {r4, lr}
+ .long 0xe591e000 // ldr lr, [r1]
+ .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xe28e301c // add r3, lr, #28
+ .long 0xe28e4010 // add r4, lr, #16
+ .long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
+ .long 0xe28e3020 // add r3, lr, #32
+ .long 0xf4e31c9f // vld1.32 {d17[]}, [r3 :32]
+ .long 0xe28e3018 // add r3, lr, #24
+ .long 0xf2411c30 // vfma.f32 d17, d1, d16
+ .long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
+ .long 0xe1a0300e // mov r3, lr
+ .long 0xf4e42c9f // vld1.32 {d18[]}, [r4 :32]
+ .long 0xe28e4008 // add r4, lr, #8
+ .long 0xf4e43c9f // vld1.32 {d19[]}, [r4 :32]
+ .long 0xf2401c30 // vfma.f32 d17, d0, d16
+ .long 0xf4e30c9d // vld1.32 {d16[]}, [r3 :32]!
+ .long 0xf4e35c9f // vld1.32 {d21[]}, [r3 :32]
+ .long 0xe28e3014 // add r3, lr, #20
+ .long 0xf2413c35 // vfma.f32 d19, d1, d21
+ .long 0xf4e35c9f // vld1.32 {d21[]}, [r3 :32]
+ .long 0xe28e300c // add r3, lr, #12
+ .long 0xf2415c32 // vfma.f32 d21, d1, d18
+ .long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
+ .long 0xf3fb4521 // vrecpe.f32 d20, d17
+ .long 0xf2403c30 // vfma.f32 d19, d0, d16
+ .long 0xf2411fb4 // vrecps.f32 d17, d17, d20
+ .long 0xf2405c32 // vfma.f32 d21, d0, d18
+ .long 0xf3440db1 // vmul.f32 d16, d20, d17
+ .long 0xf3030db0 // vmul.f32 d0, d19, d16
+ .long 0xf3051db0 // vmul.f32 d1, d21, d16
+ .long 0xe8bd4010 // pop {r4, lr}
+ .long 0xe12fff1c // bx ip
+
+.globl _sk_linear_gradient_2stops_vfp4
+_sk_linear_gradient_2stops_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
+ .long 0xe2811008 // add r1, r1, #8
+ .long 0xf4632a0d // vld1.8 {d18-d19}, [r3]!
+ .long 0xf4634a0f // vld1.8 {d20-d21}, [r3]
+ .long 0xf3f40c22 // vdup.32 d16, d18[0]
+ .long 0xf3f41c24 // vdup.32 d17, d20[0]
+ .long 0xf2400c31 // vfma.f32 d16, d0, d17
+ .long 0xf3fc6c24 // vdup.32 d22, d20[1]
+ .long 0xf3bc1c22 // vdup.32 d1, d18[1]
+ .long 0xf3b42c23 // vdup.32 d2, d19[0]
+ .long 0xf2001c36 // vfma.f32 d1, d0, d22
+ .long 0xf3f41c25 // vdup.32 d17, d21[0]
+ .long 0xf3fc4c25 // vdup.32 d20, d21[1]
+ .long 0xf2002c31 // vfma.f32 d2, d0, d17
+ .long 0xf3bc3c23 // vdup.32 d3, d19[1]
+ .long 0xf2003c34 // vfma.f32 d3, d0, d20
+ .long 0xf22001b0 // vorr d0, d16, d16
+ .long 0xe12fff1c // bx ip
+#elif defined(__x86_64__)
+
+.globl _sk_start_pipeline_hsw
+_sk_start_pipeline_hsw:
+ .byte 65,87 // push %r15
+ .byte 65,86 // push %r14
+ .byte 65,85 // push %r13
+ .byte 65,84 // push %r12
+ .byte 83 // push %rbx
+ .byte 73,137,205 // mov %rcx,%r13
+ .byte 73,137,214 // mov %rdx,%r14
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 73,137,199 // mov %rax,%r15
+ .byte 73,137,244 // mov %rsi,%r12
+ .byte 72,141,67,8 // lea 0x8(%rbx),%rax
+ .byte 76,57,232 // cmp %r13,%rax
+ .byte 118,5 // jbe 28 <_sk_start_pipeline_hsw+0x28>
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 235,65 // jmp 69 <_sk_start_pipeline_hsw+0x69>
+ .byte 185,0,0,0,0 // mov $0x0,%ecx
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 76,137,230 // mov %r12,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,215 // callq *%r15
+ .byte 72,141,123,8 // lea 0x8(%rbx),%rdi
+ .byte 72,131,195,16 // add $0x10,%rbx
+ .byte 76,57,235 // cmp %r13,%rbx
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 118,191 // jbe 28 <_sk_start_pipeline_hsw+0x28>
+ .byte 76,137,233 // mov %r13,%rcx
+ .byte 72,41,249 // sub %rdi,%rcx
+ .byte 116,41 // je 9a <_sk_start_pipeline_hsw+0x9a>
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 76,137,230 // mov %r12,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,215 // callq *%r15
+ .byte 76,137,232 // mov %r13,%rax
+ .byte 91 // pop %rbx
+ .byte 65,92 // pop %r12
+ .byte 65,93 // pop %r13
+ .byte 65,94 // pop %r14
+ .byte 65,95 // pop %r15
+ .byte 197,248,119 // vzeroupper
+ .byte 195 // retq
+
+.globl _sk_just_return_hsw
+_sk_just_return_hsw:
+ .byte 195 // retq
+
+.globl _sk_seed_shader_hsw
+_sk_seed_shader_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,249,110,199 // vmovd %edi,%xmm0
+ .byte 196,226,125,88,192 // vpbroadcastd %xmm0,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 65,184,0,0,0,63 // mov $0x3f000000,%r8d
+ .byte 196,193,121,110,200 // vmovd %r8d,%xmm1
+ .byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
+ .byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
+ .byte 197,252,88,2 // vaddps (%rdx),%ymm0,%ymm0
+ .byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 197,236,88,201 // vaddps %ymm1,%ymm2,%ymm1
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,208 // vmovd %eax,%xmm2
+ .byte 196,226,125,88,210 // vpbroadcastd %xmm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_constant_color_hsw
+_sk_constant_color_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
+ .byte 196,226,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm1
+ .byte 196,226,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm2
+ .byte 196,226,125,24,88,12 // vbroadcastss 0xc(%rax),%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clear_hsw
+_sk_clear_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_plus__hsw
+_sk_plus__hsw:
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 197,228,88,223 // vaddps %ymm7,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_srcover_hsw
+_sk_srcover_hsw:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
+ .byte 196,194,93,184,192 // vfmadd231ps %ymm8,%ymm4,%ymm0
+ .byte 196,194,85,184,200 // vfmadd231ps %ymm8,%ymm5,%ymm1
+ .byte 196,194,77,184,208 // vfmadd231ps %ymm8,%ymm6,%ymm2
+ .byte 196,194,69,184,216 // vfmadd231ps %ymm8,%ymm7,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_dstover_hsw
+_sk_dstover_hsw:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
+ .byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
+ .byte 196,226,61,168,205 // vfmadd213ps %ymm5,%ymm8,%ymm1
+ .byte 196,226,61,168,214 // vfmadd213ps %ymm6,%ymm8,%ymm2
+ .byte 196,226,61,168,223 // vfmadd213ps %ymm7,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_0_hsw
+_sk_clamp_0_hsw:
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
+ .byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
+ .byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_1_hsw
+_sk_clamp_1_hsw:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_a_hsw
+_sk_clamp_a_hsw:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
+ .byte 197,252,93,195 // vminps %ymm3,%ymm0,%ymm0
+ .byte 197,244,93,203 // vminps %ymm3,%ymm1,%ymm1
+ .byte 197,236,93,211 // vminps %ymm3,%ymm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_set_rgb_hsw
+_sk_set_rgb_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
+ .byte 196,226,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm1
+ .byte 196,226,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_rb_hsw
+_sk_swap_rb_hsw:
+ .byte 197,124,40,192 // vmovaps %ymm0,%ymm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,194 // vmovaps %ymm2,%ymm0
+ .byte 197,124,41,194 // vmovaps %ymm8,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_hsw
+_sk_swap_hsw:
+ .byte 197,124,40,195 // vmovaps %ymm3,%ymm8
+ .byte 197,124,40,202 // vmovaps %ymm2,%ymm9
+ .byte 197,124,40,209 // vmovaps %ymm1,%ymm10
+ .byte 197,124,40,216 // vmovaps %ymm0,%ymm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,196 // vmovaps %ymm4,%ymm0
+ .byte 197,252,40,205 // vmovaps %ymm5,%ymm1
+ .byte 197,252,40,214 // vmovaps %ymm6,%ymm2
+ .byte 197,252,40,223 // vmovaps %ymm7,%ymm3
+ .byte 197,124,41,220 // vmovaps %ymm11,%ymm4
+ .byte 197,124,41,213 // vmovaps %ymm10,%ymm5
+ .byte 197,124,41,206 // vmovaps %ymm9,%ymm6
+ .byte 197,124,41,199 // vmovaps %ymm8,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_src_dst_hsw
+_sk_move_src_dst_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,224 // vmovaps %ymm0,%ymm4
+ .byte 197,252,40,233 // vmovaps %ymm1,%ymm5
+ .byte 197,252,40,242 // vmovaps %ymm2,%ymm6
+ .byte 197,252,40,251 // vmovaps %ymm3,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_dst_src_hsw
+_sk_move_dst_src_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,196 // vmovaps %ymm4,%ymm0
+ .byte 197,252,40,205 // vmovaps %ymm5,%ymm1
+ .byte 197,252,40,214 // vmovaps %ymm6,%ymm2
+ .byte 197,252,40,223 // vmovaps %ymm7,%ymm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_premul_hsw
+_sk_premul_hsw:
+ .byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
+ .byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
+ .byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_unpremul_hsw
+_sk_unpremul_hsw:
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 196,65,100,194,200,0 // vcmpeqps %ymm8,%ymm3,%ymm9
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,66,125,88,210 // vpbroadcastd %xmm10,%ymm10
+ .byte 197,44,94,211 // vdivps %ymm3,%ymm10,%ymm10
+ .byte 196,67,45,74,192,144 // vblendvps %ymm9,%ymm8,%ymm10,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_from_srgb_hsw
+_sk_from_srgb_hsw:
+ .byte 184,145,131,158,61 // mov $0x3d9e8391,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 197,124,89,208 // vmulps %ymm0,%ymm0,%ymm10
+ .byte 184,154,153,153,62 // mov $0x3e99999a,%eax
+ .byte 197,121,110,216 // vmovd %eax,%xmm11
+ .byte 196,66,125,88,219 // vpbroadcastd %xmm11,%ymm11
+ .byte 184,92,143,50,63 // mov $0x3f328f5c,%eax
+ .byte 197,121,110,224 // vmovd %eax,%xmm12
+ .byte 196,66,125,88,228 // vpbroadcastd %xmm12,%ymm12
+ .byte 196,65,125,111,235 // vmovdqa %ymm11,%ymm13
+ .byte 196,66,125,168,236 // vfmadd213ps %ymm12,%ymm0,%ymm13
+ .byte 184,10,215,35,59 // mov $0x3b23d70a,%eax
+ .byte 197,121,110,240 // vmovd %eax,%xmm14
+ .byte 196,66,125,88,246 // vpbroadcastd %xmm14,%ymm14
+ .byte 196,66,45,168,238 // vfmadd213ps %ymm14,%ymm10,%ymm13
+ .byte 184,174,71,97,61 // mov $0x3d6147ae,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,66,125,88,210 // vpbroadcastd %xmm10,%ymm10
+ .byte 196,193,124,194,194,1 // vcmpltps %ymm10,%ymm0,%ymm0
+ .byte 196,195,21,74,193,0 // vblendvps %ymm0,%ymm9,%ymm13,%ymm0
+ .byte 197,60,89,201 // vmulps %ymm1,%ymm8,%ymm9
+ .byte 197,116,89,233 // vmulps %ymm1,%ymm1,%ymm13
+ .byte 196,65,125,111,251 // vmovdqa %ymm11,%ymm15
+ .byte 196,66,117,168,252 // vfmadd213ps %ymm12,%ymm1,%ymm15
+ .byte 196,66,21,168,254 // vfmadd213ps %ymm14,%ymm13,%ymm15
+ .byte 196,193,116,194,202,1 // vcmpltps %ymm10,%ymm1,%ymm1
+ .byte 196,195,5,74,201,16 // vblendvps %ymm1,%ymm9,%ymm15,%ymm1
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 197,108,89,202 // vmulps %ymm2,%ymm2,%ymm9
+ .byte 196,66,109,168,220 // vfmadd213ps %ymm12,%ymm2,%ymm11
+ .byte 196,66,53,168,222 // vfmadd213ps %ymm14,%ymm9,%ymm11
+ .byte 196,193,108,194,210,1 // vcmpltps %ymm10,%ymm2,%ymm2
+ .byte 196,195,37,74,208,32 // vblendvps %ymm2,%ymm8,%ymm11,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_to_srgb_hsw
+_sk_to_srgb_hsw:
+ .byte 197,124,82,192 // vrsqrtps %ymm0,%ymm8
+ .byte 196,65,124,83,216 // vrcpps %ymm8,%ymm11
+ .byte 196,65,124,82,224 // vrsqrtps %ymm8,%ymm12
+ .byte 184,41,92,71,65 // mov $0x41475c29,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,89,232 // vmulps %ymm0,%ymm8,%ymm13
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,66,125,88,201 // vpbroadcastd %xmm9,%ymm9
+ .byte 184,194,135,210,62 // mov $0x3ed287c2,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,66,125,88,210 // vpbroadcastd %xmm10,%ymm10
+ .byte 184,206,111,48,63 // mov $0x3f306fce,%eax
+ .byte 197,121,110,240 // vmovd %eax,%xmm14
+ .byte 196,66,125,88,246 // vpbroadcastd %xmm14,%ymm14
+ .byte 184,168,87,202,61 // mov $0x3dca57a8,%eax
+ .byte 53,0,0,0,128 // xor $0x80000000,%eax
+ .byte 197,121,110,248 // vmovd %eax,%xmm15
+ .byte 196,66,125,88,255 // vpbroadcastd %xmm15,%ymm15
+ .byte 196,66,13,168,223 // vfmadd213ps %ymm15,%ymm14,%ymm11
+ .byte 196,66,45,184,220 // vfmadd231ps %ymm12,%ymm10,%ymm11
+ .byte 196,65,52,93,219 // vminps %ymm11,%ymm9,%ymm11
+ .byte 184,4,231,140,59 // mov $0x3b8ce704,%eax
+ .byte 197,121,110,224 // vmovd %eax,%xmm12
+ .byte 196,66,125,88,228 // vpbroadcastd %xmm12,%ymm12
+ .byte 196,193,124,194,196,1 // vcmpltps %ymm12,%ymm0,%ymm0
+ .byte 196,195,37,74,197,0 // vblendvps %ymm0,%ymm13,%ymm11,%ymm0
+ .byte 197,124,82,217 // vrsqrtps %ymm1,%ymm11
+ .byte 196,65,124,83,235 // vrcpps %ymm11,%ymm13
+ .byte 196,65,124,82,219 // vrsqrtps %ymm11,%ymm11
+ .byte 196,66,13,168,239 // vfmadd213ps %ymm15,%ymm14,%ymm13
+ .byte 196,66,45,184,235 // vfmadd231ps %ymm11,%ymm10,%ymm13
+ .byte 197,60,89,217 // vmulps %ymm1,%ymm8,%ymm11
+ .byte 196,65,52,93,237 // vminps %ymm13,%ymm9,%ymm13
+ .byte 196,193,116,194,204,1 // vcmpltps %ymm12,%ymm1,%ymm1
+ .byte 196,195,21,74,203,16 // vblendvps %ymm1,%ymm11,%ymm13,%ymm1
+ .byte 197,124,82,218 // vrsqrtps %ymm2,%ymm11
+ .byte 196,65,124,83,235 // vrcpps %ymm11,%ymm13
+ .byte 196,66,13,168,239 // vfmadd213ps %ymm15,%ymm14,%ymm13
+ .byte 196,65,124,82,219 // vrsqrtps %ymm11,%ymm11
+ .byte 196,66,45,184,235 // vfmadd231ps %ymm11,%ymm10,%ymm13
+ .byte 196,65,52,93,205 // vminps %ymm13,%ymm9,%ymm9
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 196,193,108,194,212,1 // vcmpltps %ymm12,%ymm2,%ymm2
+ .byte 196,195,53,74,208,32 // vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_1_float_hsw
+_sk_scale_1_float_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_u8_hsw
+_sk_scale_u8_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,56 // jne 4bf <_sk_scale_u8_hsw+0x48>
+ .byte 197,122,126,0 // vmovq (%rax),%xmm8
+ .byte 196,66,125,49,192 // vpmovzxbd %xmm8,%ymm8
+ .byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,66,125,88,201 // vpbroadcastd %xmm9,%ymm9
+ .byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne 4c7 <_sk_scale_u8_hsw+0x50>
+ .byte 196,65,249,110,193 // vmovq %r9,%xmm8
+ .byte 235,167 // jmp 48b <_sk_scale_u8_hsw+0x14>
+
+.globl _sk_lerp_1_float_hsw
+_sk_lerp_1_float_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,226,61,168,205 // vfmadd213ps %ymm5,%ymm8,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 196,226,61,168,214 // vfmadd213ps %ymm6,%ymm8,%ymm2
+ .byte 197,228,92,223 // vsubps %ymm7,%ymm3,%ymm3
+ .byte 196,226,61,168,223 // vfmadd213ps %ymm7,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_u8_hsw
+_sk_lerp_u8_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,76 // jne 56f <_sk_lerp_u8_hsw+0x5c>
+ .byte 197,122,126,0 // vmovq (%rax),%xmm8
+ .byte 196,66,125,49,192 // vpmovzxbd %xmm8,%ymm8
+ .byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,66,125,88,201 // vpbroadcastd %xmm9,%ymm9
+ .byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,226,61,168,196 // vfmadd213ps %ymm4,%ymm8,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,226,61,168,205 // vfmadd213ps %ymm5,%ymm8,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 196,226,61,168,214 // vfmadd213ps %ymm6,%ymm8,%ymm2
+ .byte 197,228,92,223 // vsubps %ymm7,%ymm3,%ymm3
+ .byte 196,226,61,168,223 // vfmadd213ps %ymm7,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne 577 <_sk_lerp_u8_hsw+0x64>
+ .byte 196,65,249,110,193 // vmovq %r9,%xmm8
+ .byte 235,147 // jmp 527 <_sk_lerp_u8_hsw+0x14>
+
+.globl _sk_lerp_565_hsw
+_sk_lerp_565_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,16 // mov (%rax),%r10
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,179,0,0,0 // jne 655 <_sk_lerp_565_hsw+0xc1>
+ .byte 196,193,122,111,28,122 // vmovdqu (%r10,%rdi,2),%xmm3
+ .byte 196,98,125,51,195 // vpmovzxwd %xmm3,%ymm8
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 196,193,101,219,216 // vpand %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,203 // vcvtdq2ps %ymm3,%ymm9
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,52,89,203 // vmulps %ymm3,%ymm9,%ymm9
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 196,193,101,219,216 // vpand %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,211 // vcvtdq2ps %ymm3,%ymm10
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,44,89,211 // vmulps %ymm3,%ymm10,%ymm10
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 196,193,101,219,216 // vpand %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,195 // vcvtdq2ps %ymm3,%ymm8
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,226,53,168,196 // vfmadd213ps %ymm4,%ymm9,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,226,45,168,205 // vfmadd213ps %ymm5,%ymm10,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 196,226,101,168,214 // vfmadd213ps %ymm6,%ymm3,%ymm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 197,225,239,219 // vpxor %xmm3,%xmm3,%xmm3
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 15,135,59,255,255,255 // ja 5a8 <_sk_lerp_565_hsw+0x14>
+ .byte 69,15,182,192 // movzbl %r8b,%r8d
+ .byte 76,141,13,76,0,0,0 // lea 0x4c(%rip),%r9 # 6c4 <_sk_lerp_565_hsw+0x130>
+ .byte 75,99,4,129 // movslq (%r9,%r8,4),%rax
+ .byte 76,1,200 // add %r9,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,225,239,219 // vpxor %xmm3,%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,12,6 // vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,10,5 // vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,8,4 // vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,6,3 // vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,4,2 // vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,92,122,2,1 // vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 196,193,97,196,28,122,0 // vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
+ .byte 233,231,254,255,255 // jmpq 5a8 <_sk_lerp_565_hsw+0x14>
+ .byte 15,31,0 // nopl (%rax)
+ .byte 241 // icebp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 233,255,255,255,225 // jmpq ffffffffe20006cc <_sk_linear_gradient_2stops_hsw+0xffffffffe1fff4f0>
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 217,255 // fcos
+ .byte 255 // (bad)
+ .byte 255,209 // callq *%rcx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,201 // dec %ecx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 189 // .byte 0xbd
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_tables_hsw
+_sk_load_tables_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,141,12,189,0,0,0,0 // lea 0x0(,%rdi,4),%r9
+ .byte 76,3,8 // add (%rax),%r9
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,121 // jne 76e <_sk_load_tables_hsw+0x8e>
+ .byte 196,193,126,111,25 // vmovdqu (%r9),%ymm3
+ .byte 185,255,0,0,0 // mov $0xff,%ecx
+ .byte 197,249,110,193 // vmovd %ecx,%xmm0
+ .byte 196,226,125,88,208 // vpbroadcastd %xmm0,%ymm2
+ .byte 197,237,219,203 // vpand %ymm3,%ymm2,%ymm1
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 72,139,72,8 // mov 0x8(%rax),%rcx
+ .byte 76,139,72,16 // mov 0x10(%rax),%r9
+ .byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
+ .byte 196,226,53,146,4,137 // vgatherdps %ymm9,(%rcx,%ymm1,4),%ymm0
+ .byte 197,245,114,211,8 // vpsrld $0x8,%ymm3,%ymm1
+ .byte 197,109,219,201 // vpand %ymm1,%ymm2,%ymm9
+ .byte 196,65,45,118,210 // vpcmpeqd %ymm10,%ymm10,%ymm10
+ .byte 196,130,45,146,12,137 // vgatherdps %ymm10,(%r9,%ymm9,4),%ymm1
+ .byte 72,139,64,24 // mov 0x18(%rax),%rax
+ .byte 197,181,114,211,16 // vpsrld $0x10,%ymm3,%ymm9
+ .byte 196,65,109,219,201 // vpand %ymm9,%ymm2,%ymm9
+ .byte 196,162,61,146,20,136 // vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
+ .byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
+ .byte 197,124,91,195 // vcvtdq2ps %ymm3,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 185,8,0,0,0 // mov $0x8,%ecx
+ .byte 68,41,193 // sub %r8d,%ecx
+ .byte 192,225,3 // shl $0x3,%cl
+ .byte 73,199,194,255,255,255,255 // mov $0xffffffffffffffff,%r10
+ .byte 73,211,234 // shr %cl,%r10
+ .byte 196,193,249,110,194 // vmovq %r10,%xmm0
+ .byte 196,226,125,33,192 // vpmovsxbd %xmm0,%ymm0
+ .byte 196,194,125,140,25 // vpmaskmovd (%r9),%ymm0,%ymm3
+ .byte 233,99,255,255,255 // jmpq 6fa <_sk_load_tables_hsw+0x1a>
+
+.globl _sk_load_a8_hsw
+_sk_load_a8_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,50 // jne 7d9 <_sk_load_a8_hsw+0x42>
+ .byte 197,250,126,0 // vmovq (%rax),%xmm0
+ .byte 196,226,125,49,192 // vpmovzxbd %xmm0,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
+ .byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne 7e1 <_sk_load_a8_hsw+0x4a>
+ .byte 196,193,249,110,193 // vmovq %r9,%xmm0
+ .byte 235,173 // jmp 7ab <_sk_load_a8_hsw+0x14>
+
+.globl _sk_store_a8_hsw
+_sk_store_a8_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 184,0,0,127,67 // mov $0x437f0000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,66,57,43,193 // vpackusdw %xmm9,%xmm8,%xmm8
+ .byte 196,65,57,103,192 // vpackuswb %xmm8,%xmm8,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,10 // jne 839 <_sk_store_a8_hsw+0x3b>
+ .byte 196,65,123,17,4,57 // vmovsd %xmm8,(%r9,%rdi,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 119,236 // ja 835 <_sk_store_a8_hsw+0x37>
+ .byte 196,66,121,48,192 // vpmovzxbw %xmm8,%xmm8
+ .byte 65,15,182,192 // movzbl %r8b,%eax
+ .byte 76,141,5,67,0,0,0 // lea 0x43(%rip),%r8 # 89c <_sk_store_a8_hsw+0x9e>
+ .byte 73,99,4,128 // movslq (%r8,%rax,4),%rax
+ .byte 76,1,192 // add %r8,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,67,121,20,68,57,6,12 // vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,5,10 // vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,4,8 // vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,3,6 // vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,2,4 // vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,1,2 // vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
+ .byte 196,67,121,20,4,57,0 // vpextrb $0x0,%xmm8,(%r9,%rdi,1)
+ .byte 235,154 // jmp 835 <_sk_store_a8_hsw+0x37>
+ .byte 144 // nop
+ .byte 246,255 // idiv %bh
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 238 // out %al,(%dx)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,230 // jmpq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 222,255 // fdivrp %st,%st(7)
+ .byte 255 // (bad)
+ .byte 255,214 // callq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,206 // dec %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,198 // inc %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_565_hsw
+_sk_load_565_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,16 // mov (%rax),%r10
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,149,0,0,0 // jne 95b <_sk_load_565_hsw+0xa3>
+ .byte 196,193,122,111,4,122 // vmovdqu (%r10,%rdi,2),%xmm0
+ .byte 196,226,125,51,208 // vpmovzxwd %xmm0,%ymm2
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 197,249,110,192 // vmovd %eax,%xmm0
+ .byte 196,226,125,88,192 // vpbroadcastd %xmm0,%ymm0
+ .byte 197,253,219,194 // vpand %ymm2,%ymm0,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
+ .byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,226,125,88,201 // vpbroadcastd %xmm1,%ymm1
+ .byte 197,245,219,202 // vpand %ymm2,%ymm1,%ymm1
+ .byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,229,219,210 // vpand %ymm2,%ymm3,%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 15,135,89,255,255,255 // ja 8cc <_sk_load_565_hsw+0x14>
+ .byte 69,15,182,192 // movzbl %r8b,%r8d
+ .byte 76,141,13,74,0,0,0 // lea 0x4a(%rip),%r9 # 9c8 <_sk_load_565_hsw+0x110>
+ .byte 75,99,4,129 // movslq (%r9,%r8,4),%rax
+ .byte 76,1,200 // add %r9,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,12,6 // vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,10,5 // vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,8,4 // vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,6,3 // vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,4,2 // vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,2,1 // vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,4,122,0 // vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 233,5,255,255,255 // jmpq 8cc <_sk_load_565_hsw+0x14>
+ .byte 144 // nop
+ .byte 243,255 // repz (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 235,255 // jmp 9cd <_sk_load_565_hsw+0x115>
+ .byte 255 // (bad)
+ .byte 255,227 // jmpq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 219,255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,211 // callq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,203 // dec %ebx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 191 // .byte 0xbf
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_store_565_hsw
+_sk_store_565_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 184,0,0,248,65 // mov $0x41f80000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
+ .byte 196,193,53,114,241,11 // vpslld $0xb,%ymm9,%ymm9
+ .byte 184,0,0,124,66 // mov $0x427c0000,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,66,125,88,210 // vpbroadcastd %xmm10,%ymm10
+ .byte 197,44,89,209 // vmulps %ymm1,%ymm10,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,45,114,242,5 // vpslld $0x5,%ymm10,%ymm10
+ .byte 196,65,45,235,201 // vpor %ymm9,%ymm10,%ymm9
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
+ .byte 196,67,125,57,193,1 // vextracti128 $0x1,%ymm8,%xmm9
+ .byte 196,66,57,43,193 // vpackusdw %xmm9,%xmm8,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,10 // jne a50 <_sk_store_565_hsw+0x6c>
+ .byte 196,65,122,127,4,121 // vmovdqu %xmm8,(%r9,%rdi,2)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 119,236 // ja a4c <_sk_store_565_hsw+0x68>
+ .byte 65,15,182,192 // movzbl %r8b,%eax
+ .byte 76,141,5,69,0,0,0 // lea 0x45(%rip),%r8 # ab0 <_sk_store_565_hsw+0xcc>
+ .byte 73,99,4,128 // movslq (%r8,%rax,4),%rax
+ .byte 76,1,192 // add %r8,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,67,121,21,68,121,12,6 // vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,10,5 // vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,8,4 // vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,6,3 // vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,4,2 // vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,2,1 // vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
+ .byte 196,67,121,21,4,121,0 // vpextrw $0x0,%xmm8,(%r9,%rdi,2)
+ .byte 235,159 // jmp a4c <_sk_store_565_hsw+0x68>
+ .byte 15,31,0 // nopl (%rax)
+ .byte 244 // hlt
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 236 // in (%dx),%al
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,228 // jmpq *%rsp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 220,255 // fdivr %st,%st(7)
+ .byte 255 // (bad)
+ .byte 255,212 // callq *%rsp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,204 // dec %esp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,196 // inc %esp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_8888_hsw
+_sk_load_8888_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,141,12,189,0,0,0,0 // lea 0x0(,%rdi,4),%r9
+ .byte 76,3,8 // add (%rax),%r9
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,104 // jne b49 <_sk_load_8888_hsw+0x7d>
+ .byte 196,193,126,111,25 // vmovdqu (%r9),%ymm3
+ .byte 184,255,0,0,0 // mov $0xff,%eax
+ .byte 197,249,110,192 // vmovd %eax,%xmm0
+ .byte 196,226,125,88,208 // vpbroadcastd %xmm0,%ymm2
+ .byte 197,237,219,195 // vpand %ymm3,%ymm2,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,98,125,88,193 // vpbroadcastd %xmm1,%ymm8
+ .byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
+ .byte 197,245,114,211,8 // vpsrld $0x8,%ymm3,%ymm1
+ .byte 197,237,219,201 // vpand %ymm1,%ymm2,%ymm1
+ .byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
+ .byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
+ .byte 197,181,114,211,16 // vpsrld $0x10,%ymm3,%ymm9
+ .byte 196,193,109,219,209 // vpand %ymm9,%ymm2,%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
+ .byte 197,229,114,211,24 // vpsrld $0x18,%ymm3,%ymm3
+ .byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
+ .byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 185,8,0,0,0 // mov $0x8,%ecx
+ .byte 68,41,193 // sub %r8d,%ecx
+ .byte 192,225,3 // shl $0x3,%cl
+ .byte 72,199,192,255,255,255,255 // mov $0xffffffffffffffff,%rax
+ .byte 72,211,232 // shr %cl,%rax
+ .byte 196,225,249,110,192 // vmovq %rax,%xmm0
+ .byte 196,226,125,33,192 // vpmovsxbd %xmm0,%ymm0
+ .byte 196,194,125,140,25 // vpmaskmovd (%r9),%ymm0,%ymm3
+ .byte 233,116,255,255,255 // jmpq ae6 <_sk_load_8888_hsw+0x1a>
+
+.globl _sk_store_8888_hsw
+_sk_store_8888_hsw:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,141,12,189,0,0,0,0 // lea 0x0(,%rdi,4),%r9
+ .byte 76,3,8 // add (%rax),%r9
+ .byte 184,0,0,127,67 // mov $0x437f0000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,66,125,88,192 // vpbroadcastd %xmm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
+ .byte 197,60,89,209 // vmulps %ymm1,%ymm8,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,45,114,242,8 // vpslld $0x8,%ymm10,%ymm10
+ .byte 196,65,45,235,201 // vpor %ymm9,%ymm10,%ymm9
+ .byte 197,60,89,210 // vmulps %ymm2,%ymm8,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,45,114,242,16 // vpslld $0x10,%ymm10,%ymm10
+ .byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,193,61,114,240,24 // vpslld $0x18,%ymm8,%ymm8
+ .byte 196,65,45,235,192 // vpor %ymm8,%ymm10,%ymm8
+ .byte 196,65,53,235,192 // vpor %ymm8,%ymm9,%ymm8
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,12 // jne be6 <_sk_store_8888_hsw+0x74>
+ .byte 196,65,126,127,1 // vmovdqu %ymm8,(%r9)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 185,8,0,0,0 // mov $0x8,%ecx
+ .byte 68,41,193 // sub %r8d,%ecx
+ .byte 192,225,3 // shl $0x3,%cl
+ .byte 72,199,192,255,255,255,255 // mov $0xffffffffffffffff,%rax
+ .byte 72,211,232 // shr %cl,%rax
+ .byte 196,97,249,110,200 // vmovq %rax,%xmm9
+ .byte 196,66,125,33,201 // vpmovsxbd %xmm9,%ymm9
+ .byte 196,66,53,142,1 // vpmaskmovd %ymm8,%ymm9,(%r9)
+ .byte 235,211 // jmp bdf <_sk_store_8888_hsw+0x6d>
+
+.globl _sk_load_f16_hsw
+_sk_load_f16_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,97 // jne c77 <_sk_load_f16_hsw+0x6b>
+ .byte 197,121,16,4,248 // vmovupd (%rax,%rdi,8),%xmm8
+ .byte 197,249,16,84,248,16 // vmovupd 0x10(%rax,%rdi,8),%xmm2
+ .byte 197,249,16,92,248,32 // vmovupd 0x20(%rax,%rdi,8),%xmm3
+ .byte 197,122,111,76,248,48 // vmovdqu 0x30(%rax,%rdi,8),%xmm9
+ .byte 197,185,97,194 // vpunpcklwd %xmm2,%xmm8,%xmm0
+ .byte 197,185,105,210 // vpunpckhwd %xmm2,%xmm8,%xmm2
+ .byte 196,193,97,97,201 // vpunpcklwd %xmm9,%xmm3,%xmm1
+ .byte 196,193,97,105,217 // vpunpckhwd %xmm9,%xmm3,%xmm3
+ .byte 197,121,97,194 // vpunpcklwd %xmm2,%xmm0,%xmm8
+ .byte 197,121,105,202 // vpunpckhwd %xmm2,%xmm0,%xmm9
+ .byte 197,241,97,211 // vpunpcklwd %xmm3,%xmm1,%xmm2
+ .byte 197,241,105,219 // vpunpckhwd %xmm3,%xmm1,%xmm3
+ .byte 197,185,108,194 // vpunpcklqdq %xmm2,%xmm8,%xmm0
+ .byte 196,226,125,19,192 // vcvtph2ps %xmm0,%ymm0
+ .byte 197,185,109,202 // vpunpckhqdq %xmm2,%xmm8,%xmm1
+ .byte 196,226,125,19,201 // vcvtph2ps %xmm1,%ymm1
+ .byte 197,177,108,211 // vpunpcklqdq %xmm3,%xmm9,%xmm2
+ .byte 196,226,125,19,210 // vcvtph2ps %xmm2,%ymm2
+ .byte 197,177,109,219 // vpunpckhqdq %xmm3,%xmm9,%xmm3
+ .byte 196,226,125,19,219 // vcvtph2ps %xmm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,123,16,4,248 // vmovsd (%rax,%rdi,8),%xmm8
+ .byte 196,65,49,239,201 // vpxor %xmm9,%xmm9,%xmm9
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,79 // je cd6 <_sk_load_f16_hsw+0xca>
+ .byte 197,57,22,68,248,8 // vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,67 // jb cd6 <_sk_load_f16_hsw+0xca>
+ .byte 197,251,16,84,248,16 // vmovsd 0x10(%rax,%rdi,8),%xmm2
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 116,68 // je ce3 <_sk_load_f16_hsw+0xd7>
+ .byte 197,233,22,84,248,24 // vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,56 // jb ce3 <_sk_load_f16_hsw+0xd7>
+ .byte 197,251,16,92,248,32 // vmovsd 0x20(%rax,%rdi,8),%xmm3
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 15,132,114,255,255,255 // je c2d <_sk_load_f16_hsw+0x21>
+ .byte 197,225,22,92,248,40 // vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 15,130,98,255,255,255 // jb c2d <_sk_load_f16_hsw+0x21>
+ .byte 197,122,126,76,248,48 // vmovq 0x30(%rax,%rdi,8),%xmm9
+ .byte 233,87,255,255,255 // jmpq c2d <_sk_load_f16_hsw+0x21>
+ .byte 197,225,87,219 // vxorpd %xmm3,%xmm3,%xmm3
+ .byte 197,233,87,210 // vxorpd %xmm2,%xmm2,%xmm2
+ .byte 233,74,255,255,255 // jmpq c2d <_sk_load_f16_hsw+0x21>
+ .byte 197,225,87,219 // vxorpd %xmm3,%xmm3,%xmm3
+ .byte 233,65,255,255,255 // jmpq c2d <_sk_load_f16_hsw+0x21>
+
+.globl _sk_store_f16_hsw
+_sk_store_f16_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 196,195,125,29,192,4 // vcvtps2ph $0x4,%ymm0,%xmm8
+ .byte 196,195,125,29,201,4 // vcvtps2ph $0x4,%ymm1,%xmm9
+ .byte 196,195,125,29,210,4 // vcvtps2ph $0x4,%ymm2,%xmm10
+ .byte 196,195,125,29,219,4 // vcvtps2ph $0x4,%ymm3,%xmm11
+ .byte 196,65,57,97,225 // vpunpcklwd %xmm9,%xmm8,%xmm12
+ .byte 196,65,57,105,193 // vpunpckhwd %xmm9,%xmm8,%xmm8
+ .byte 196,65,41,97,203 // vpunpcklwd %xmm11,%xmm10,%xmm9
+ .byte 196,65,41,105,235 // vpunpckhwd %xmm11,%xmm10,%xmm13
+ .byte 196,65,25,98,217 // vpunpckldq %xmm9,%xmm12,%xmm11
+ .byte 196,65,25,106,209 // vpunpckhdq %xmm9,%xmm12,%xmm10
+ .byte 196,65,57,98,205 // vpunpckldq %xmm13,%xmm8,%xmm9
+ .byte 196,65,57,106,197 // vpunpckhdq %xmm13,%xmm8,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,27 // jne d51 <_sk_store_f16_hsw+0x65>
+ .byte 197,120,17,28,248 // vmovups %xmm11,(%rax,%rdi,8)
+ .byte 197,120,17,84,248,16 // vmovups %xmm10,0x10(%rax,%rdi,8)
+ .byte 197,120,17,76,248,32 // vmovups %xmm9,0x20(%rax,%rdi,8)
+ .byte 197,122,127,68,248,48 // vmovdqu %xmm8,0x30(%rax,%rdi,8)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,121,214,28,248 // vmovq %xmm11,(%rax,%rdi,8)
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,241 // je d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,23,92,248,8 // vmovhpd %xmm11,0x8(%rax,%rdi,8)
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,229 // jb d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,214,84,248,16 // vmovq %xmm10,0x10(%rax,%rdi,8)
+ .byte 116,221 // je d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,23,84,248,24 // vmovhpd %xmm10,0x18(%rax,%rdi,8)
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,209 // jb d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,214,76,248,32 // vmovq %xmm9,0x20(%rax,%rdi,8)
+ .byte 116,201 // je d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,23,76,248,40 // vmovhpd %xmm9,0x28(%rax,%rdi,8)
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 114,189 // jb d4d <_sk_store_f16_hsw+0x61>
+ .byte 197,121,214,68,248,48 // vmovq %xmm8,0x30(%rax,%rdi,8)
+ .byte 235,181 // jmp d4d <_sk_store_f16_hsw+0x61>
+
+.globl _sk_store_f32_hsw
+_sk_store_f32_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,0 // mov (%rax),%r8
+ .byte 72,141,4,189,0,0,0,0 // lea 0x0(,%rdi,4),%rax
+ .byte 197,124,20,193 // vunpcklps %ymm1,%ymm0,%ymm8
+ .byte 197,124,21,217 // vunpckhps %ymm1,%ymm0,%ymm11
+ .byte 197,108,20,203 // vunpcklps %ymm3,%ymm2,%ymm9
+ .byte 197,108,21,227 // vunpckhps %ymm3,%ymm2,%ymm12
+ .byte 196,65,61,20,209 // vunpcklpd %ymm9,%ymm8,%ymm10
+ .byte 196,65,61,21,201 // vunpckhpd %ymm9,%ymm8,%ymm9
+ .byte 196,65,37,20,196 // vunpcklpd %ymm12,%ymm11,%ymm8
+ .byte 196,65,37,21,220 // vunpckhpd %ymm12,%ymm11,%ymm11
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,55 // jne e05 <_sk_store_f32_hsw+0x6d>
+ .byte 196,67,45,24,225,1 // vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
+ .byte 196,67,61,24,235,1 // vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
+ .byte 196,67,45,6,201,49 // vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
+ .byte 196,67,61,6,195,49 // vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
+ .byte 196,65,125,17,36,128 // vmovupd %ymm12,(%r8,%rax,4)
+ .byte 196,65,125,17,108,128,32 // vmovupd %ymm13,0x20(%r8,%rax,4)
+ .byte 196,65,125,17,76,128,64 // vmovupd %ymm9,0x40(%r8,%rax,4)
+ .byte 196,65,125,17,68,128,96 // vmovupd %ymm8,0x60(%r8,%rax,4)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,65,121,17,20,128 // vmovupd %xmm10,(%r8,%rax,4)
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,240 // je e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,65,121,17,76,128,16 // vmovupd %xmm9,0x10(%r8,%rax,4)
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,227 // jb e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,65,121,17,68,128,32 // vmovupd %xmm8,0x20(%r8,%rax,4)
+ .byte 116,218 // je e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,65,121,17,92,128,48 // vmovupd %xmm11,0x30(%r8,%rax,4)
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,205 // jb e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,67,125,25,84,128,64,1 // vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
+ .byte 116,195 // je e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,67,125,25,76,128,80,1 // vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 114,181 // jb e01 <_sk_store_f32_hsw+0x69>
+ .byte 196,67,125,25,68,128,96,1 // vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
+ .byte 235,171 // jmp e01 <_sk_store_f32_hsw+0x69>
+
+.globl _sk_clamp_x_hsw
+_sk_clamp_x_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,188,95,192 // vmaxps %ymm0,%ymm8,%ymm0
+ .byte 196,98,125,88,0 // vpbroadcastd (%rax),%ymm8
+ .byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_y_hsw
+_sk_clamp_y_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,188,95,201 // vmaxps %ymm1,%ymm8,%ymm1
+ .byte 196,98,125,88,0 // vpbroadcastd (%rax),%ymm8
+ .byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
+ .byte 196,65,61,254,193 // vpaddd %ymm9,%ymm8,%ymm8
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_x_hsw
+_sk_repeat_x_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,65,124,94,200 // vdivps %ymm8,%ymm0,%ymm9
+ .byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
+ .byte 196,98,61,172,200 // vfnmadd213ps %ymm0,%ymm8,%ymm9
+ .byte 197,253,118,192 // vpcmpeqd %ymm0,%ymm0,%ymm0
+ .byte 197,189,254,192 // vpaddd %ymm0,%ymm8,%ymm0
+ .byte 197,180,93,192 // vminps %ymm0,%ymm9,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_y_hsw
+_sk_repeat_y_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,65,116,94,200 // vdivps %ymm8,%ymm1,%ymm9
+ .byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
+ .byte 196,98,61,172,201 // vfnmadd213ps %ymm1,%ymm8,%ymm9
+ .byte 197,245,118,201 // vpcmpeqd %ymm1,%ymm1,%ymm1
+ .byte 197,189,254,201 // vpaddd %ymm1,%ymm8,%ymm1
+ .byte 197,180,93,201 // vminps %ymm1,%ymm9,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_x_hsw
+_sk_mirror_x_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,122,16,0 // vmovss (%rax),%xmm8
+ .byte 196,66,125,24,200 // vbroadcastss %xmm8,%ymm9
+ .byte 196,65,124,92,209 // vsubps %ymm9,%ymm0,%ymm10
+ .byte 196,193,58,88,192 // vaddss %xmm8,%xmm8,%xmm0
+ .byte 196,226,125,24,192 // vbroadcastss %xmm0,%ymm0
+ .byte 197,44,94,192 // vdivps %ymm0,%ymm10,%ymm8
+ .byte 196,67,125,8,192,1 // vroundps $0x1,%ymm8,%ymm8
+ .byte 196,66,125,172,194 // vfnmadd213ps %ymm10,%ymm0,%ymm8
+ .byte 196,193,60,92,193 // vsubps %ymm9,%ymm8,%ymm0
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,92,192 // vsubps %ymm0,%ymm8,%ymm8
+ .byte 197,188,84,192 // vandps %ymm0,%ymm8,%ymm0
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 196,65,53,254,192 // vpaddd %ymm8,%ymm9,%ymm8
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_y_hsw
+_sk_mirror_y_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,122,16,0 // vmovss (%rax),%xmm8
+ .byte 196,66,125,24,200 // vbroadcastss %xmm8,%ymm9
+ .byte 196,65,116,92,209 // vsubps %ymm9,%ymm1,%ymm10
+ .byte 196,193,58,88,200 // vaddss %xmm8,%xmm8,%xmm1
+ .byte 196,226,125,24,201 // vbroadcastss %xmm1,%ymm1
+ .byte 197,44,94,193 // vdivps %ymm1,%ymm10,%ymm8
+ .byte 196,67,125,8,192,1 // vroundps $0x1,%ymm8,%ymm8
+ .byte 196,66,117,172,194 // vfnmadd213ps %ymm10,%ymm1,%ymm8
+ .byte 196,193,60,92,201 // vsubps %ymm9,%ymm8,%ymm1
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,92,193 // vsubps %ymm1,%ymm8,%ymm8
+ .byte 197,188,84,201 // vandps %ymm1,%ymm8,%ymm1
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 196,65,53,254,192 // vpaddd %ymm8,%ymm9,%ymm8
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_luminance_to_alpha_hsw
+_sk_luminance_to_alpha_hsw:
+ .byte 184,208,179,89,62 // mov $0x3e59b3d0,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,98,125,88,195 // vpbroadcastd %xmm3,%ymm8
+ .byte 184,89,23,55,63 // mov $0x3f371759,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,226,125,88,219 // vpbroadcastd %xmm3,%ymm3
+ .byte 197,228,89,201 // vmulps %ymm1,%ymm3,%ymm1
+ .byte 196,98,125,168,193 // vfmadd213ps %ymm1,%ymm0,%ymm8
+ .byte 184,152,221,147,61 // mov $0x3d93dd98,%eax
+ .byte 197,249,110,192 // vmovd %eax,%xmm0
+ .byte 196,226,125,88,216 // vpbroadcastd %xmm0,%ymm3
+ .byte 196,194,109,168,216 // vfmadd213ps %ymm8,%ymm2,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,253,239,192 // vpxor %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_2x3_hsw
+_sk_matrix_2x3_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,8 // vbroadcastss (%rax),%ymm9
+ .byte 196,98,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm10
+ .byte 196,98,125,24,64,16 // vbroadcastss 0x10(%rax),%ymm8
+ .byte 196,66,117,184,194 // vfmadd231ps %ymm10,%ymm1,%ymm8
+ .byte 196,66,125,184,193 // vfmadd231ps %ymm9,%ymm0,%ymm8
+ .byte 196,98,125,24,80,4 // vbroadcastss 0x4(%rax),%ymm10
+ .byte 196,98,125,24,88,12 // vbroadcastss 0xc(%rax),%ymm11
+ .byte 196,98,125,24,72,20 // vbroadcastss 0x14(%rax),%ymm9
+ .byte 196,66,117,184,203 // vfmadd231ps %ymm11,%ymm1,%ymm9
+ .byte 196,66,125,184,202 // vfmadd231ps %ymm10,%ymm0,%ymm9
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 197,124,41,201 // vmovaps %ymm9,%ymm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_3x4_hsw
+_sk_matrix_3x4_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,8 // vbroadcastss (%rax),%ymm9
+ .byte 196,98,125,24,80,12 // vbroadcastss 0xc(%rax),%ymm10
+ .byte 196,98,125,24,88,24 // vbroadcastss 0x18(%rax),%ymm11
+ .byte 196,98,125,24,64,36 // vbroadcastss 0x24(%rax),%ymm8
+ .byte 196,66,109,184,195 // vfmadd231ps %ymm11,%ymm2,%ymm8
+ .byte 196,66,117,184,194 // vfmadd231ps %ymm10,%ymm1,%ymm8
+ .byte 196,66,125,184,193 // vfmadd231ps %ymm9,%ymm0,%ymm8
+ .byte 196,98,125,24,80,4 // vbroadcastss 0x4(%rax),%ymm10
+ .byte 196,98,125,24,88,16 // vbroadcastss 0x10(%rax),%ymm11
+ .byte 196,98,125,24,96,28 // vbroadcastss 0x1c(%rax),%ymm12
+ .byte 196,98,125,24,72,40 // vbroadcastss 0x28(%rax),%ymm9
+ .byte 196,66,109,184,204 // vfmadd231ps %ymm12,%ymm2,%ymm9
+ .byte 196,66,117,184,203 // vfmadd231ps %ymm11,%ymm1,%ymm9
+ .byte 196,66,125,184,202 // vfmadd231ps %ymm10,%ymm0,%ymm9
+ .byte 196,98,125,24,88,8 // vbroadcastss 0x8(%rax),%ymm11
+ .byte 196,98,125,24,96,20 // vbroadcastss 0x14(%rax),%ymm12
+ .byte 196,98,125,24,104,32 // vbroadcastss 0x20(%rax),%ymm13
+ .byte 196,98,125,24,80,44 // vbroadcastss 0x2c(%rax),%ymm10
+ .byte 196,66,109,184,213 // vfmadd231ps %ymm13,%ymm2,%ymm10
+ .byte 196,66,117,184,212 // vfmadd231ps %ymm12,%ymm1,%ymm10
+ .byte 196,66,125,184,211 // vfmadd231ps %ymm11,%ymm0,%ymm10
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 197,124,41,201 // vmovaps %ymm9,%ymm1
+ .byte 197,124,41,210 // vmovaps %ymm10,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_4x5_hsw
+_sk_matrix_4x5_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,8 // vbroadcastss (%rax),%ymm9
+ .byte 196,98,125,24,80,16 // vbroadcastss 0x10(%rax),%ymm10
+ .byte 196,98,125,24,88,32 // vbroadcastss 0x20(%rax),%ymm11
+ .byte 196,98,125,24,96,48 // vbroadcastss 0x30(%rax),%ymm12
+ .byte 196,98,125,24,64,64 // vbroadcastss 0x40(%rax),%ymm8
+ .byte 196,66,101,184,196 // vfmadd231ps %ymm12,%ymm3,%ymm8
+ .byte 196,66,109,184,195 // vfmadd231ps %ymm11,%ymm2,%ymm8
+ .byte 196,66,117,184,194 // vfmadd231ps %ymm10,%ymm1,%ymm8
+ .byte 196,66,125,184,193 // vfmadd231ps %ymm9,%ymm0,%ymm8
+ .byte 196,98,125,24,80,4 // vbroadcastss 0x4(%rax),%ymm10
+ .byte 196,98,125,24,88,20 // vbroadcastss 0x14(%rax),%ymm11
+ .byte 196,98,125,24,96,36 // vbroadcastss 0x24(%rax),%ymm12
+ .byte 196,98,125,24,104,52 // vbroadcastss 0x34(%rax),%ymm13
+ .byte 196,98,125,24,72,68 // vbroadcastss 0x44(%rax),%ymm9
+ .byte 196,66,101,184,205 // vfmadd231ps %ymm13,%ymm3,%ymm9
+ .byte 196,66,109,184,204 // vfmadd231ps %ymm12,%ymm2,%ymm9
+ .byte 196,66,117,184,203 // vfmadd231ps %ymm11,%ymm1,%ymm9
+ .byte 196,66,125,184,202 // vfmadd231ps %ymm10,%ymm0,%ymm9
+ .byte 196,98,125,24,88,8 // vbroadcastss 0x8(%rax),%ymm11
+ .byte 196,98,125,24,96,24 // vbroadcastss 0x18(%rax),%ymm12
+ .byte 196,98,125,24,104,40 // vbroadcastss 0x28(%rax),%ymm13
+ .byte 196,98,125,24,112,56 // vbroadcastss 0x38(%rax),%ymm14
+ .byte 196,98,125,24,80,72 // vbroadcastss 0x48(%rax),%ymm10
+ .byte 196,66,101,184,214 // vfmadd231ps %ymm14,%ymm3,%ymm10
+ .byte 196,66,109,184,213 // vfmadd231ps %ymm13,%ymm2,%ymm10
+ .byte 196,66,117,184,212 // vfmadd231ps %ymm12,%ymm1,%ymm10
+ .byte 196,66,125,184,211 // vfmadd231ps %ymm11,%ymm0,%ymm10
+ .byte 196,98,125,24,96,12 // vbroadcastss 0xc(%rax),%ymm12
+ .byte 196,98,125,24,104,28 // vbroadcastss 0x1c(%rax),%ymm13
+ .byte 196,98,125,24,112,44 // vbroadcastss 0x2c(%rax),%ymm14
+ .byte 196,98,125,24,120,60 // vbroadcastss 0x3c(%rax),%ymm15
+ .byte 196,98,125,24,88,76 // vbroadcastss 0x4c(%rax),%ymm11
+ .byte 196,66,101,184,223 // vfmadd231ps %ymm15,%ymm3,%ymm11
+ .byte 196,66,109,184,222 // vfmadd231ps %ymm14,%ymm2,%ymm11
+ .byte 196,66,117,184,221 // vfmadd231ps %ymm13,%ymm1,%ymm11
+ .byte 196,66,125,184,220 // vfmadd231ps %ymm12,%ymm0,%ymm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 197,124,41,201 // vmovaps %ymm9,%ymm1
+ .byte 197,124,41,210 // vmovaps %ymm10,%ymm2
+ .byte 197,124,41,219 // vmovaps %ymm11,%ymm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_perspective_hsw
+_sk_matrix_perspective_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm9
+ .byte 196,98,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm10
+ .byte 196,66,117,184,209 // vfmadd231ps %ymm9,%ymm1,%ymm10
+ .byte 196,66,125,184,208 // vfmadd231ps %ymm8,%ymm0,%ymm10
+ .byte 196,98,125,24,64,12 // vbroadcastss 0xc(%rax),%ymm8
+ .byte 196,98,125,24,72,16 // vbroadcastss 0x10(%rax),%ymm9
+ .byte 196,98,125,24,88,20 // vbroadcastss 0x14(%rax),%ymm11
+ .byte 196,66,117,184,217 // vfmadd231ps %ymm9,%ymm1,%ymm11
+ .byte 196,66,125,184,216 // vfmadd231ps %ymm8,%ymm0,%ymm11
+ .byte 196,98,125,24,64,24 // vbroadcastss 0x18(%rax),%ymm8
+ .byte 196,98,125,24,72,28 // vbroadcastss 0x1c(%rax),%ymm9
+ .byte 196,98,125,24,96,32 // vbroadcastss 0x20(%rax),%ymm12
+ .byte 196,66,117,184,225 // vfmadd231ps %ymm9,%ymm1,%ymm12
+ .byte 196,66,125,184,224 // vfmadd231ps %ymm8,%ymm0,%ymm12
+ .byte 196,193,124,83,204 // vrcpps %ymm12,%ymm1
+ .byte 197,172,89,193 // vmulps %ymm1,%ymm10,%ymm0
+ .byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_linear_gradient_2stops_hsw
+_sk_linear_gradient_2stops_hsw:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,72,16 // vbroadcastss 0x10(%rax),%ymm1
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,184,193 // vfmadd231ps %ymm1,%ymm0,%ymm8
+ .byte 196,226,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm2
+ .byte 196,226,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm1
+ .byte 196,226,125,184,202 // vfmadd231ps %ymm2,%ymm0,%ymm1
+ .byte 196,226,125,24,88,24 // vbroadcastss 0x18(%rax),%ymm3
+ .byte 196,226,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm2
+ .byte 196,226,125,184,211 // vfmadd231ps %ymm3,%ymm0,%ymm2
+ .byte 196,98,125,24,72,28 // vbroadcastss 0x1c(%rax),%ymm9
+ .byte 196,226,125,24,88,12 // vbroadcastss 0xc(%rax),%ymm3
+ .byte 196,194,125,184,217 // vfmadd231ps %ymm9,%ymm0,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_start_pipeline_avx
+_sk_start_pipeline_avx:
+ .byte 65,87 // push %r15
+ .byte 65,86 // push %r14
+ .byte 65,85 // push %r13
+ .byte 65,84 // push %r12
+ .byte 83 // push %rbx
+ .byte 73,137,205 // mov %rcx,%r13
+ .byte 73,137,214 // mov %rdx,%r14
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 73,137,199 // mov %rax,%r15
+ .byte 73,137,244 // mov %rsi,%r12
+ .byte 72,141,67,8 // lea 0x8(%rbx),%rax
+ .byte 76,57,232 // cmp %r13,%rax
+ .byte 118,5 // jbe 28 <_sk_start_pipeline_avx+0x28>
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 235,65 // jmp 69 <_sk_start_pipeline_avx+0x69>
+ .byte 185,0,0,0,0 // mov $0x0,%ecx
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 76,137,230 // mov %r12,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,215 // callq *%r15
+ .byte 72,141,123,8 // lea 0x8(%rbx),%rdi
+ .byte 72,131,195,16 // add $0x10,%rbx
+ .byte 76,57,235 // cmp %r13,%rbx
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 118,191 // jbe 28 <_sk_start_pipeline_avx+0x28>
+ .byte 76,137,233 // mov %r13,%rcx
+ .byte 72,41,249 // sub %rdi,%rcx
+ .byte 116,41 // je 9a <_sk_start_pipeline_avx+0x9a>
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 76,137,230 // mov %r12,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,215 // callq *%r15
+ .byte 76,137,232 // mov %r13,%rax
+ .byte 91 // pop %rbx
+ .byte 65,92 // pop %r12
+ .byte 65,93 // pop %r13
+ .byte 65,94 // pop %r14
+ .byte 65,95 // pop %r15
+ .byte 197,248,119 // vzeroupper
+ .byte 195 // retq
+
+.globl _sk_just_return_avx
+_sk_just_return_avx:
+ .byte 195 // retq
+
+.globl _sk_seed_shader_avx
+_sk_seed_shader_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,249,110,199 // vmovd %edi,%xmm0
+ .byte 197,249,112,192,0 // vpshufd $0x0,%xmm0,%xmm0
+ .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 65,184,0,0,0,63 // mov $0x3f000000,%r8d
+ .byte 196,193,121,110,200 // vmovd %r8d,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
+ .byte 197,252,88,2 // vaddps (%rdx),%ymm0,%ymm0
+ .byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 197,236,88,201 // vaddps %ymm1,%ymm2,%ymm1
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,208 // vmovd %eax,%xmm2
+ .byte 196,227,121,4,210,0 // vpermilps $0x0,%xmm2,%xmm2
+ .byte 196,227,109,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 197,220,87,228 // vxorps %ymm4,%ymm4,%ymm4
+ .byte 197,212,87,237 // vxorps %ymm5,%ymm5,%ymm5
+ .byte 197,204,87,246 // vxorps %ymm6,%ymm6,%ymm6
+ .byte 197,196,87,255 // vxorps %ymm7,%ymm7,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_constant_color_avx
+_sk_constant_color_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
+ .byte 196,226,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm1
+ .byte 196,226,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm2
+ .byte 196,226,125,24,88,12 // vbroadcastss 0xc(%rax),%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clear_avx
+_sk_clear_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 197,228,87,219 // vxorps %ymm3,%ymm3,%ymm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_plus__avx
+_sk_plus__avx:
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 197,228,88,223 // vaddps %ymm7,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_srcover_avx
+_sk_srcover_avx:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,92,195 // vsubps %ymm3,%ymm8,%ymm8
+ .byte 197,60,89,204 // vmulps %ymm4,%ymm8,%ymm9
+ .byte 197,180,88,192 // vaddps %ymm0,%ymm9,%ymm0
+ .byte 197,60,89,205 // vmulps %ymm5,%ymm8,%ymm9
+ .byte 197,180,88,201 // vaddps %ymm1,%ymm9,%ymm1
+ .byte 197,60,89,206 // vmulps %ymm6,%ymm8,%ymm9
+ .byte 197,180,88,210 // vaddps %ymm2,%ymm9,%ymm2
+ .byte 197,60,89,199 // vmulps %ymm7,%ymm8,%ymm8
+ .byte 197,188,88,219 // vaddps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_dstover_avx
+_sk_dstover_avx:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,92,199 // vsubps %ymm7,%ymm8,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 197,228,88,223 // vaddps %ymm7,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_0_avx
+_sk_clamp_0_avx:
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 196,193,124,95,192 // vmaxps %ymm8,%ymm0,%ymm0
+ .byte 196,193,116,95,200 // vmaxps %ymm8,%ymm1,%ymm1
+ .byte 196,193,108,95,208 // vmaxps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,95,216 // vmaxps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_1_avx
+_sk_clamp_1_avx:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 196,193,124,93,192 // vminps %ymm8,%ymm0,%ymm0
+ .byte 196,193,116,93,200 // vminps %ymm8,%ymm1,%ymm1
+ .byte 196,193,108,93,208 // vminps %ymm8,%ymm2,%ymm2
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_a_avx
+_sk_clamp_a_avx:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 196,193,100,93,216 // vminps %ymm8,%ymm3,%ymm3
+ .byte 197,252,93,195 // vminps %ymm3,%ymm0,%ymm0
+ .byte 197,244,93,203 // vminps %ymm3,%ymm1,%ymm1
+ .byte 197,236,93,211 // vminps %ymm3,%ymm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_set_rgb_avx
+_sk_set_rgb_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,0 // vbroadcastss (%rax),%ymm0
+ .byte 196,226,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm1
+ .byte 196,226,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_rb_avx
+_sk_swap_rb_avx:
+ .byte 197,124,40,192 // vmovaps %ymm0,%ymm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,194 // vmovaps %ymm2,%ymm0
+ .byte 197,124,41,194 // vmovaps %ymm8,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_avx
+_sk_swap_avx:
+ .byte 197,124,40,195 // vmovaps %ymm3,%ymm8
+ .byte 197,124,40,202 // vmovaps %ymm2,%ymm9
+ .byte 197,124,40,209 // vmovaps %ymm1,%ymm10
+ .byte 197,124,40,216 // vmovaps %ymm0,%ymm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,196 // vmovaps %ymm4,%ymm0
+ .byte 197,252,40,205 // vmovaps %ymm5,%ymm1
+ .byte 197,252,40,214 // vmovaps %ymm6,%ymm2
+ .byte 197,252,40,223 // vmovaps %ymm7,%ymm3
+ .byte 197,124,41,220 // vmovaps %ymm11,%ymm4
+ .byte 197,124,41,213 // vmovaps %ymm10,%ymm5
+ .byte 197,124,41,206 // vmovaps %ymm9,%ymm6
+ .byte 197,124,41,199 // vmovaps %ymm8,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_src_dst_avx
+_sk_move_src_dst_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,224 // vmovaps %ymm0,%ymm4
+ .byte 197,252,40,233 // vmovaps %ymm1,%ymm5
+ .byte 197,252,40,242 // vmovaps %ymm2,%ymm6
+ .byte 197,252,40,251 // vmovaps %ymm3,%ymm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_dst_src_avx
+_sk_move_dst_src_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,40,196 // vmovaps %ymm4,%ymm0
+ .byte 197,252,40,205 // vmovaps %ymm5,%ymm1
+ .byte 197,252,40,214 // vmovaps %ymm6,%ymm2
+ .byte 197,252,40,223 // vmovaps %ymm7,%ymm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_premul_avx
+_sk_premul_avx:
+ .byte 197,252,89,195 // vmulps %ymm3,%ymm0,%ymm0
+ .byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
+ .byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_unpremul_avx
+_sk_unpremul_avx:
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 196,65,100,194,200,0 // vcmpeqps %ymm8,%ymm3,%ymm9
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,67,121,4,210,0 // vpermilps $0x0,%xmm10,%xmm10
+ .byte 196,67,45,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ .byte 197,44,94,211 // vdivps %ymm3,%ymm10,%ymm10
+ .byte 196,67,45,74,192,144 // vblendvps %ymm9,%ymm8,%ymm10,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_from_srgb_avx
+_sk_from_srgb_avx:
+ .byte 184,145,131,158,61 // mov $0x3d9e8391,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 197,124,89,208 // vmulps %ymm0,%ymm0,%ymm10
+ .byte 184,154,153,153,62 // mov $0x3e99999a,%eax
+ .byte 197,121,110,216 // vmovd %eax,%xmm11
+ .byte 196,67,121,4,219,0 // vpermilps $0x0,%xmm11,%xmm11
+ .byte 196,67,37,24,219,1 // vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ .byte 184,92,143,50,63 // mov $0x3f328f5c,%eax
+ .byte 197,121,110,224 // vmovd %eax,%xmm12
+ .byte 196,67,121,4,228,0 // vpermilps $0x0,%xmm12,%xmm12
+ .byte 196,67,29,24,228,1 // vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
+ .byte 197,36,89,232 // vmulps %ymm0,%ymm11,%ymm13
+ .byte 196,65,20,88,236 // vaddps %ymm12,%ymm13,%ymm13
+ .byte 184,10,215,35,59 // mov $0x3b23d70a,%eax
+ .byte 197,121,110,240 // vmovd %eax,%xmm14
+ .byte 196,67,121,4,246,0 // vpermilps $0x0,%xmm14,%xmm14
+ .byte 196,67,13,24,246,1 // vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ .byte 196,65,44,89,213 // vmulps %ymm13,%ymm10,%ymm10
+ .byte 196,65,12,88,210 // vaddps %ymm10,%ymm14,%ymm10
+ .byte 184,174,71,97,61 // mov $0x3d6147ae,%eax
+ .byte 197,121,110,232 // vmovd %eax,%xmm13
+ .byte 196,67,121,4,237,0 // vpermilps $0x0,%xmm13,%xmm13
+ .byte 196,67,21,24,237,1 // vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
+ .byte 196,193,124,194,197,1 // vcmpltps %ymm13,%ymm0,%ymm0
+ .byte 196,195,45,74,193,0 // vblendvps %ymm0,%ymm9,%ymm10,%ymm0
+ .byte 197,60,89,201 // vmulps %ymm1,%ymm8,%ymm9
+ .byte 197,116,89,209 // vmulps %ymm1,%ymm1,%ymm10
+ .byte 197,36,89,249 // vmulps %ymm1,%ymm11,%ymm15
+ .byte 196,65,28,88,255 // vaddps %ymm15,%ymm12,%ymm15
+ .byte 196,65,44,89,215 // vmulps %ymm15,%ymm10,%ymm10
+ .byte 196,65,12,88,210 // vaddps %ymm10,%ymm14,%ymm10
+ .byte 196,193,116,194,205,1 // vcmpltps %ymm13,%ymm1,%ymm1
+ .byte 196,195,45,74,201,16 // vblendvps %ymm1,%ymm9,%ymm10,%ymm1
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 197,108,89,202 // vmulps %ymm2,%ymm2,%ymm9
+ .byte 197,36,89,210 // vmulps %ymm2,%ymm11,%ymm10
+ .byte 196,65,28,88,210 // vaddps %ymm10,%ymm12,%ymm10
+ .byte 196,65,52,89,202 // vmulps %ymm10,%ymm9,%ymm9
+ .byte 196,65,12,88,201 // vaddps %ymm9,%ymm14,%ymm9
+ .byte 196,193,108,194,213,1 // vcmpltps %ymm13,%ymm2,%ymm2
+ .byte 196,195,53,74,208,32 // vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_to_srgb_avx
+_sk_to_srgb_avx:
+ .byte 197,124,82,192 // vrsqrtps %ymm0,%ymm8
+ .byte 196,65,124,83,232 // vrcpps %ymm8,%ymm13
+ .byte 196,65,124,82,240 // vrsqrtps %ymm8,%ymm14
+ .byte 184,41,92,71,65 // mov $0x41475c29,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,224 // vmulps %ymm0,%ymm8,%ymm12
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,67,121,4,201,0 // vpermilps $0x0,%xmm9,%xmm9
+ .byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ .byte 184,194,135,210,62 // mov $0x3ed287c2,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,67,121,4,210,0 // vpermilps $0x0,%xmm10,%xmm10
+ .byte 196,67,45,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ .byte 184,206,111,48,63 // mov $0x3f306fce,%eax
+ .byte 197,121,110,216 // vmovd %eax,%xmm11
+ .byte 196,67,121,4,219,0 // vpermilps $0x0,%xmm11,%xmm11
+ .byte 196,67,37,24,219,1 // vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ .byte 184,168,87,202,61 // mov $0x3dca57a8,%eax
+ .byte 53,0,0,0,128 // xor $0x80000000,%eax
+ .byte 197,121,110,248 // vmovd %eax,%xmm15
+ .byte 196,67,121,4,255,0 // vpermilps $0x0,%xmm15,%xmm15
+ .byte 196,67,5,24,255,1 // vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
+ .byte 196,65,20,89,235 // vmulps %ymm11,%ymm13,%ymm13
+ .byte 196,65,20,88,239 // vaddps %ymm15,%ymm13,%ymm13
+ .byte 196,65,12,89,242 // vmulps %ymm10,%ymm14,%ymm14
+ .byte 196,65,12,88,237 // vaddps %ymm13,%ymm14,%ymm13
+ .byte 196,65,52,93,237 // vminps %ymm13,%ymm9,%ymm13
+ .byte 184,4,231,140,59 // mov $0x3b8ce704,%eax
+ .byte 197,121,110,240 // vmovd %eax,%xmm14
+ .byte 196,67,121,4,246,0 // vpermilps $0x0,%xmm14,%xmm14
+ .byte 196,67,13,24,246,1 // vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ .byte 196,193,124,194,198,1 // vcmpltps %ymm14,%ymm0,%ymm0
+ .byte 196,195,21,74,196,0 // vblendvps %ymm0,%ymm12,%ymm13,%ymm0
+ .byte 197,124,82,225 // vrsqrtps %ymm1,%ymm12
+ .byte 196,65,124,83,236 // vrcpps %ymm12,%ymm13
+ .byte 196,65,124,82,228 // vrsqrtps %ymm12,%ymm12
+ .byte 196,65,36,89,237 // vmulps %ymm13,%ymm11,%ymm13
+ .byte 196,65,4,88,237 // vaddps %ymm13,%ymm15,%ymm13
+ .byte 196,65,44,89,228 // vmulps %ymm12,%ymm10,%ymm12
+ .byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
+ .byte 197,60,89,233 // vmulps %ymm1,%ymm8,%ymm13
+ .byte 196,65,52,93,228 // vminps %ymm12,%ymm9,%ymm12
+ .byte 196,193,116,194,206,1 // vcmpltps %ymm14,%ymm1,%ymm1
+ .byte 196,195,29,74,205,16 // vblendvps %ymm1,%ymm13,%ymm12,%ymm1
+ .byte 197,124,82,226 // vrsqrtps %ymm2,%ymm12
+ .byte 196,65,124,83,236 // vrcpps %ymm12,%ymm13
+ .byte 196,65,36,89,221 // vmulps %ymm13,%ymm11,%ymm11
+ .byte 196,65,4,88,219 // vaddps %ymm11,%ymm15,%ymm11
+ .byte 196,65,124,82,228 // vrsqrtps %ymm12,%ymm12
+ .byte 196,65,44,89,212 // vmulps %ymm12,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 196,65,52,93,202 // vminps %ymm10,%ymm9,%ymm9
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 196,193,108,194,214,1 // vcmpltps %ymm14,%ymm2,%ymm2
+ .byte 196,195,53,74,208,32 // vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_1_float_avx
+_sk_scale_1_float_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_u8_avx
+_sk_scale_u8_avx:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,80 // jne 5a2 <_sk_scale_u8_avx+0x60>
+ .byte 197,122,126,0 // vmovq (%rax),%xmm8
+ .byte 196,66,121,49,200 // vpmovzxbd %xmm8,%xmm9
+ .byte 196,67,121,4,192,229 // vpermilps $0xe5,%xmm8,%xmm8
+ .byte 196,66,121,49,192 // vpmovzxbd %xmm8,%xmm8
+ .byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
+ .byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,67,121,4,201,0 // vpermilps $0x0,%xmm9,%xmm9
+ .byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ .byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,188,89,210 // vmulps %ymm2,%ymm8,%ymm2
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne 5aa <_sk_scale_u8_avx+0x68>
+ .byte 196,65,249,110,193 // vmovq %r9,%xmm8
+ .byte 235,143 // jmp 556 <_sk_scale_u8_avx+0x14>
+
+.globl _sk_lerp_1_float_avx
+_sk_lerp_1_float_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 197,228,92,223 // vsubps %ymm7,%ymm3,%ymm3
+ .byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
+ .byte 197,228,88,223 // vaddps %ymm7,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_u8_avx
+_sk_lerp_u8_avx:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,116 // jne 68a <_sk_lerp_u8_avx+0x84>
+ .byte 197,122,126,0 // vmovq (%rax),%xmm8
+ .byte 196,66,121,49,200 // vpmovzxbd %xmm8,%xmm9
+ .byte 196,67,121,4,192,229 // vpermilps $0xe5,%xmm8,%xmm8
+ .byte 196,66,121,49,192 // vpmovzxbd %xmm8,%xmm8
+ .byte 196,67,53,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
+ .byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,121,110,200 // vmovd %eax,%xmm9
+ .byte 196,67,121,4,201,0 // vpermilps $0x0,%xmm9,%xmm9
+ .byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ .byte 196,65,60,89,193 // vmulps %ymm9,%ymm8,%ymm8
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 197,228,92,223 // vsubps %ymm7,%ymm3,%ymm3
+ .byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
+ .byte 197,228,88,223 // vaddps %ymm7,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne 692 <_sk_lerp_u8_avx+0x8c>
+ .byte 196,65,249,110,193 // vmovq %r9,%xmm8
+ .byte 233,104,255,255,255 // jmpq 61a <_sk_lerp_u8_avx+0x14>
+
+.globl _sk_lerp_565_avx
+_sk_lerp_565_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,16 // mov (%rax),%r10
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,250,0,0,0 // jne 7ba <_sk_lerp_565_avx+0x108>
+ .byte 196,65,122,111,4,122 // vmovdqu (%r10,%rdi,2),%xmm8
+ .byte 197,225,239,219 // vpxor %xmm3,%xmm3,%xmm3
+ .byte 197,185,105,219 // vpunpckhwd %xmm3,%xmm8,%xmm3
+ .byte 196,66,121,51,192 // vpmovzxwd %xmm8,%xmm8
+ .byte 196,99,61,24,195,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 197,249,112,219,0 // vpshufd $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,203 // vcvtdq2ps %ymm3,%ymm9
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,52,89,203 // vmulps %ymm3,%ymm9,%ymm9
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 197,249,112,219,0 // vpshufd $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,211 // vcvtdq2ps %ymm3,%ymm10
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,44,89,211 // vmulps %ymm3,%ymm10,%ymm10
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 197,249,112,219,0 // vpshufd $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
+ .byte 197,124,91,195 // vcvtdq2ps %ymm3,%ymm8
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
+ .byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
+ .byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
+ .byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
+ .byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
+ .byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
+ .byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
+ .byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
+ .byte 197,236,88,214 // vaddps %ymm6,%ymm2,%ymm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 196,65,57,239,192 // vpxor %xmm8,%xmm8,%xmm8
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 15,135,243,254,255,255 // ja 6c6 <_sk_lerp_565_avx+0x14>
+ .byte 69,15,182,192 // movzbl %r8b,%r8d
+ .byte 76,141,13,74,0,0,0 // lea 0x4a(%rip),%r9 # 828 <_sk_lerp_565_avx+0x176>
+ .byte 75,99,4,129 // movslq (%r9,%r8,4),%rax
+ .byte 76,1,200 // add %r9,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,225,239,219 // vpxor %xmm3,%xmm3,%xmm3
+ .byte 196,65,97,196,68,122,12,6 // vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm8
+ .byte 196,65,57,196,68,122,10,5 // vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 196,65,57,196,68,122,8,4 // vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 196,65,57,196,68,122,6,3 // vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 196,65,57,196,68,122,4,2 // vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 196,65,57,196,68,122,2,1 // vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 196,65,57,196,4,122,0 // vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
+ .byte 233,159,254,255,255 // jmpq 6c6 <_sk_lerp_565_avx+0x14>
+ .byte 144 // nop
+ .byte 243,255 // repz (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 235,255 // jmp 82d <_sk_lerp_565_avx+0x17b>
+ .byte 255 // (bad)
+ .byte 255,227 // jmpq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 219,255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,211 // callq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,203 // dec %ebx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 191 // .byte 0xbf
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_tables_avx
+_sk_load_tables_avx:
+ .byte 85 // push %rbp
+ .byte 65,87 // push %r15
+ .byte 65,86 // push %r14
+ .byte 65,85 // push %r13
+ .byte 65,84 // push %r12
+ .byte 83 // push %rbx
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,0 // mov (%rax),%r8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,56,2,0,0 // jne a94 <_sk_load_tables_avx+0x250>
+ .byte 196,65,124,16,4,184 // vmovups (%r8,%rdi,4),%ymm8
+ .byte 187,255,0,0,0 // mov $0xff,%ebx
+ .byte 197,249,110,195 // vmovd %ebx,%xmm0
+ .byte 197,249,112,192,0 // vpshufd $0x0,%xmm0,%xmm0
+ .byte 196,99,125,24,200,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
+ .byte 196,193,52,84,192 // vandps %ymm8,%ymm9,%ymm0
+ .byte 196,193,249,126,193 // vmovq %xmm0,%r9
+ .byte 69,137,203 // mov %r9d,%r11d
+ .byte 196,195,249,22,194,1 // vpextrq $0x1,%xmm0,%r10
+ .byte 69,137,214 // mov %r10d,%r14d
+ .byte 73,193,234,32 // shr $0x20,%r10
+ .byte 73,193,233,32 // shr $0x20,%r9
+ .byte 196,227,125,25,192,1 // vextractf128 $0x1,%ymm0,%xmm0
+ .byte 196,193,249,126,196 // vmovq %xmm0,%r12
+ .byte 69,137,231 // mov %r12d,%r15d
+ .byte 196,227,249,22,195,1 // vpextrq $0x1,%xmm0,%rbx
+ .byte 65,137,221 // mov %ebx,%r13d
+ .byte 72,193,235,32 // shr $0x20,%rbx
+ .byte 73,193,236,32 // shr $0x20,%r12
+ .byte 72,139,104,8 // mov 0x8(%rax),%rbp
+ .byte 76,139,64,16 // mov 0x10(%rax),%r8
+ .byte 196,161,122,16,68,189,0 // vmovss 0x0(%rbp,%r15,4),%xmm0
+ .byte 196,163,121,33,68,165,0,16 // vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
+ .byte 196,161,122,16,76,173,0 // vmovss 0x0(%rbp,%r13,4),%xmm1
+ .byte 196,227,121,33,193,32 // vinsertps $0x20,%xmm1,%xmm0,%xmm0
+ .byte 197,250,16,76,157,0 // vmovss 0x0(%rbp,%rbx,4),%xmm1
+ .byte 196,227,121,33,193,48 // vinsertps $0x30,%xmm1,%xmm0,%xmm0
+ .byte 196,161,122,16,76,157,0 // vmovss 0x0(%rbp,%r11,4),%xmm1
+ .byte 196,163,113,33,76,141,0,16 // vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
+ .byte 196,161,122,16,92,181,0 // vmovss 0x0(%rbp,%r14,4),%xmm3
+ .byte 196,227,113,33,203,32 // vinsertps $0x20,%xmm3,%xmm1,%xmm1
+ .byte 196,161,122,16,92,149,0 // vmovss 0x0(%rbp,%r10,4),%xmm3
+ .byte 196,227,113,33,203,48 // vinsertps $0x30,%xmm3,%xmm1,%xmm1
+ .byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
+ .byte 196,193,113,114,208,8 // vpsrld $0x8,%xmm8,%xmm1
+ .byte 196,67,125,25,194,1 // vextractf128 $0x1,%ymm8,%xmm10
+ .byte 196,193,105,114,210,8 // vpsrld $0x8,%xmm10,%xmm2
+ .byte 196,227,117,24,202,1 // vinsertf128 $0x1,%xmm2,%ymm1,%ymm1
+ .byte 197,180,84,201 // vandps %ymm1,%ymm9,%ymm1
+ .byte 196,193,249,126,201 // vmovq %xmm1,%r9
+ .byte 69,137,203 // mov %r9d,%r11d
+ .byte 196,195,249,22,202,1 // vpextrq $0x1,%xmm1,%r10
+ .byte 69,137,214 // mov %r10d,%r14d
+ .byte 73,193,234,32 // shr $0x20,%r10
+ .byte 73,193,233,32 // shr $0x20,%r9
+ .byte 196,227,125,25,201,1 // vextractf128 $0x1,%ymm1,%xmm1
+ .byte 196,225,249,126,205 // vmovq %xmm1,%rbp
+ .byte 65,137,239 // mov %ebp,%r15d
+ .byte 196,227,249,22,203,1 // vpextrq $0x1,%xmm1,%rbx
+ .byte 65,137,220 // mov %ebx,%r12d
+ .byte 72,193,235,32 // shr $0x20,%rbx
+ .byte 72,193,237,32 // shr $0x20,%rbp
+ .byte 196,129,122,16,12,184 // vmovss (%r8,%r15,4),%xmm1
+ .byte 196,195,113,33,12,168,16 // vinsertps $0x10,(%r8,%rbp,4),%xmm1,%xmm1
+ .byte 196,129,122,16,20,160 // vmovss (%r8,%r12,4),%xmm2
+ .byte 196,227,113,33,202,32 // vinsertps $0x20,%xmm2,%xmm1,%xmm1
+ .byte 196,193,122,16,20,152 // vmovss (%r8,%rbx,4),%xmm2
+ .byte 196,227,113,33,202,48 // vinsertps $0x30,%xmm2,%xmm1,%xmm1
+ .byte 196,129,122,16,20,152 // vmovss (%r8,%r11,4),%xmm2
+ .byte 196,131,105,33,20,136,16 // vinsertps $0x10,(%r8,%r9,4),%xmm2,%xmm2
+ .byte 196,129,122,16,28,176 // vmovss (%r8,%r14,4),%xmm3
+ .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
+ .byte 196,129,122,16,28,144 // vmovss (%r8,%r10,4),%xmm3
+ .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
+ .byte 196,227,109,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
+ .byte 72,139,64,24 // mov 0x18(%rax),%rax
+ .byte 196,193,105,114,208,16 // vpsrld $0x10,%xmm8,%xmm2
+ .byte 196,193,97,114,210,16 // vpsrld $0x10,%xmm10,%xmm3
+ .byte 196,227,109,24,211,1 // vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ .byte 197,180,84,210 // vandps %ymm2,%ymm9,%ymm2
+ .byte 196,193,249,126,208 // vmovq %xmm2,%r8
+ .byte 69,137,194 // mov %r8d,%r10d
+ .byte 196,195,249,22,209,1 // vpextrq $0x1,%xmm2,%r9
+ .byte 69,137,203 // mov %r9d,%r11d
+ .byte 73,193,233,32 // shr $0x20,%r9
+ .byte 73,193,232,32 // shr $0x20,%r8
+ .byte 196,227,125,25,210,1 // vextractf128 $0x1,%ymm2,%xmm2
+ .byte 196,225,249,126,213 // vmovq %xmm2,%rbp
+ .byte 65,137,238 // mov %ebp,%r14d
+ .byte 196,227,249,22,211,1 // vpextrq $0x1,%xmm2,%rbx
+ .byte 65,137,223 // mov %ebx,%r15d
+ .byte 72,193,235,32 // shr $0x20,%rbx
+ .byte 72,193,237,32 // shr $0x20,%rbp
+ .byte 196,161,122,16,20,176 // vmovss (%rax,%r14,4),%xmm2
+ .byte 196,227,105,33,20,168,16 // vinsertps $0x10,(%rax,%rbp,4),%xmm2,%xmm2
+ .byte 196,161,122,16,28,184 // vmovss (%rax,%r15,4),%xmm3
+ .byte 196,227,105,33,211,32 // vinsertps $0x20,%xmm3,%xmm2,%xmm2
+ .byte 197,250,16,28,152 // vmovss (%rax,%rbx,4),%xmm3
+ .byte 196,99,105,33,203,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm9
+ .byte 196,161,122,16,28,144 // vmovss (%rax,%r10,4),%xmm3
+ .byte 196,163,97,33,28,128,16 // vinsertps $0x10,(%rax,%r8,4),%xmm3,%xmm3
+ .byte 196,161,122,16,20,152 // vmovss (%rax,%r11,4),%xmm2
+ .byte 196,227,97,33,210,32 // vinsertps $0x20,%xmm2,%xmm3,%xmm2
+ .byte 196,161,122,16,28,136 // vmovss (%rax,%r9,4),%xmm3
+ .byte 196,227,105,33,211,48 // vinsertps $0x30,%xmm3,%xmm2,%xmm2
+ .byte 196,195,109,24,209,1 // vinsertf128 $0x1,%xmm9,%ymm2,%ymm2
+ .byte 196,193,57,114,208,24 // vpsrld $0x18,%xmm8,%xmm8
+ .byte 196,193,97,114,210,24 // vpsrld $0x18,%xmm10,%xmm3
+ .byte 196,227,61,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ .byte 197,124,91,195 // vcvtdq2ps %ymm3,%ymm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 91 // pop %rbx
+ .byte 65,92 // pop %r12
+ .byte 65,93 // pop %r13
+ .byte 65,94 // pop %r14
+ .byte 65,95 // pop %r15
+ .byte 93 // pop %rbp
+ .byte 255,224 // jmpq *%rax
+ .byte 137,203 // mov %ecx,%ebx
+ .byte 128,227,7 // and $0x7,%bl
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 254,203 // dec %bl
+ .byte 128,251,6 // cmp $0x6,%bl
+ .byte 15,135,185,253,255,255 // ja 862 <_sk_load_tables_avx+0x1e>
+ .byte 15,182,219 // movzbl %bl,%ebx
+ .byte 76,141,13,137,0,0,0 // lea 0x89(%rip),%r9 # b3c <_sk_load_tables_avx+0x2f8>
+ .byte 73,99,28,153 // movslq (%r9,%rbx,4),%rbx
+ .byte 76,1,203 // add %r9,%rbx
+ .byte 255,227 // jmpq *%rbx
+ .byte 196,193,121,110,68,184,24 // vmovd 0x18(%r8,%rdi,4),%xmm0
+ .byte 197,249,112,192,68 // vpshufd $0x44,%xmm0,%xmm0
+ .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 196,99,117,12,192,64 // vblendps $0x40,%ymm0,%ymm1,%ymm8
+ .byte 196,99,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm0
+ .byte 196,195,121,34,68,184,20,1 // vpinsrd $0x1,0x14(%r8,%rdi,4),%xmm0,%xmm0
+ .byte 196,99,61,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
+ .byte 196,99,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm0
+ .byte 196,195,121,34,68,184,16,0 // vpinsrd $0x0,0x10(%r8,%rdi,4),%xmm0,%xmm0
+ .byte 196,99,61,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
+ .byte 196,195,57,34,68,184,12,3 // vpinsrd $0x3,0xc(%r8,%rdi,4),%xmm8,%xmm0
+ .byte 196,99,61,12,192,15 // vblendps $0xf,%ymm0,%ymm8,%ymm8
+ .byte 196,195,57,34,68,184,8,2 // vpinsrd $0x2,0x8(%r8,%rdi,4),%xmm8,%xmm0
+ .byte 196,99,61,12,192,15 // vblendps $0xf,%ymm0,%ymm8,%ymm8
+ .byte 196,195,57,34,68,184,4,1 // vpinsrd $0x1,0x4(%r8,%rdi,4),%xmm8,%xmm0
+ .byte 196,99,61,12,192,15 // vblendps $0xf,%ymm0,%ymm8,%ymm8
+ .byte 196,195,57,34,4,184,0 // vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
+ .byte 196,99,61,12,192,15 // vblendps $0xf,%ymm0,%ymm8,%ymm8
+ .byte 233,38,253,255,255 // jmpq 862 <_sk_load_tables_avx+0x1e>
+ .byte 238 // out %al,(%dx)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,224 // jmpq *%rax
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,210 // callq *%rdx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,196 // inc %esp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,176,255,255,255,156 // pushq -0x63000001(%rax)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+ .byte 128,255,255 // cmp $0xff,%bh
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_a8_avx
+_sk_load_a8_avx:
+ .byte 73,137,200 // mov %rcx,%r8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,1,248 // add %rdi,%rax
+ .byte 77,133,192 // test %r8,%r8
+ .byte 117,74 // jne bb2 <_sk_load_a8_avx+0x5a>
+ .byte 197,250,126,0 // vmovq (%rax),%xmm0
+ .byte 196,226,121,49,200 // vpmovzxbd %xmm0,%xmm1
+ .byte 196,227,121,4,192,229 // vpermilps $0xe5,%xmm0,%xmm0
+ .byte 196,226,121,49,192 // vpmovzxbd %xmm0,%xmm0
+ .byte 196,227,117,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,252,89,217 // vmulps %ymm1,%ymm0,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 76,137,193 // mov %r8,%rcx
+ .byte 255,224 // jmpq *%rax
+ .byte 49,201 // xor %ecx,%ecx
+ .byte 77,137,194 // mov %r8,%r10
+ .byte 69,49,201 // xor %r9d,%r9d
+ .byte 68,15,182,24 // movzbl (%rax),%r11d
+ .byte 72,255,192 // inc %rax
+ .byte 73,211,227 // shl %cl,%r11
+ .byte 77,9,217 // or %r11,%r9
+ .byte 72,131,193,8 // add $0x8,%rcx
+ .byte 73,255,202 // dec %r10
+ .byte 117,234 // jne bba <_sk_load_a8_avx+0x62>
+ .byte 196,193,249,110,193 // vmovq %r9,%xmm0
+ .byte 235,149 // jmp b6c <_sk_load_a8_avx+0x14>
+
+.globl _sk_store_a8_avx
+_sk_store_a8_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 184,0,0,127,67 // mov $0x437f0000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,66,57,43,193 // vpackusdw %xmm9,%xmm8,%xmm8
+ .byte 196,65,57,103,192 // vpackuswb %xmm8,%xmm8,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,10 // jne c19 <_sk_store_a8_avx+0x42>
+ .byte 196,65,123,17,4,57 // vmovsd %xmm8,(%r9,%rdi,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 119,236 // ja c15 <_sk_store_a8_avx+0x3e>
+ .byte 196,66,121,48,192 // vpmovzxbw %xmm8,%xmm8
+ .byte 65,15,182,192 // movzbl %r8b,%eax
+ .byte 76,141,5,67,0,0,0 // lea 0x43(%rip),%r8 # c7c <_sk_store_a8_avx+0xa5>
+ .byte 73,99,4,128 // movslq (%r8,%rax,4),%rax
+ .byte 76,1,192 // add %r8,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,67,121,20,68,57,6,12 // vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,5,10 // vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,4,8 // vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,3,6 // vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,2,4 // vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
+ .byte 196,67,121,20,68,57,1,2 // vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
+ .byte 196,67,121,20,4,57,0 // vpextrb $0x0,%xmm8,(%r9,%rdi,1)
+ .byte 235,154 // jmp c15 <_sk_store_a8_avx+0x3e>
+ .byte 144 // nop
+ .byte 246,255 // idiv %bh
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 238 // out %al,(%dx)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,230 // jmpq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 222,255 // fdivrp %st,%st(7)
+ .byte 255 // (bad)
+ .byte 255,214 // callq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,206 // dec %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,198 // inc %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_565_avx
+_sk_load_565_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,16 // mov (%rax),%r10
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,209,0,0,0 // jne d77 <_sk_load_565_avx+0xdf>
+ .byte 196,193,122,111,4,122 // vmovdqu (%r10,%rdi,2),%xmm0
+ .byte 197,241,239,201 // vpxor %xmm1,%xmm1,%xmm1
+ .byte 197,249,105,201 // vpunpckhwd %xmm1,%xmm0,%xmm1
+ .byte 196,226,121,51,192 // vpmovzxwd %xmm0,%xmm0
+ .byte 196,227,125,24,209,1 // vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 197,249,110,192 // vmovd %eax,%xmm0
+ .byte 197,249,112,192,0 // vpshufd $0x0,%xmm0,%xmm0
+ .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 197,252,84,194 // vandps %ymm2,%ymm0,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,252,89,193 // vmulps %ymm1,%ymm0,%ymm0
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 197,249,112,201,0 // vpshufd $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,244,84,202 // vandps %ymm2,%ymm1,%ymm1
+ .byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,244,89,203 // vmulps %ymm3,%ymm1,%ymm1
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 197,249,112,219,0 // vpshufd $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,228,84,210 // vandps %ymm2,%ymm3,%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 15,135,29,255,255,255 // ja cac <_sk_load_565_avx+0x14>
+ .byte 69,15,182,192 // movzbl %r8b,%r8d
+ .byte 76,141,13,74,0,0,0 // lea 0x4a(%rip),%r9 # de4 <_sk_load_565_avx+0x14c>
+ .byte 75,99,4,129 // movslq (%r9,%r8,4),%rax
+ .byte 76,1,200 // add %r9,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,249,239,192 // vpxor %xmm0,%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,12,6 // vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,10,5 // vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,8,4 // vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,6,3 // vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,4,2 // vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,68,122,2,1 // vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 196,193,121,196,4,122,0 // vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
+ .byte 233,201,254,255,255 // jmpq cac <_sk_load_565_avx+0x14>
+ .byte 144 // nop
+ .byte 243,255 // repz (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 235,255 // jmp de9 <_sk_load_565_avx+0x151>
+ .byte 255 // (bad)
+ .byte 255,227 // jmpq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 219,255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,211 // callq *%rbx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,203 // dec %ebx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 191 // .byte 0xbf
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_store_565_avx
+_sk_store_565_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 184,0,0,248,65 // mov $0x41f80000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
+ .byte 196,193,41,114,241,11 // vpslld $0xb,%xmm9,%xmm10
+ .byte 196,67,125,25,201,1 // vextractf128 $0x1,%ymm9,%xmm9
+ .byte 196,193,49,114,241,11 // vpslld $0xb,%xmm9,%xmm9
+ .byte 196,67,45,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
+ .byte 184,0,0,124,66 // mov $0x427c0000,%eax
+ .byte 197,121,110,208 // vmovd %eax,%xmm10
+ .byte 196,67,121,4,210,0 // vpermilps $0x0,%xmm10,%xmm10
+ .byte 196,67,45,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ .byte 197,44,89,209 // vmulps %ymm1,%ymm10,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,33,114,242,5 // vpslld $0x5,%xmm10,%xmm11
+ .byte 196,67,125,25,210,1 // vextractf128 $0x1,%ymm10,%xmm10
+ .byte 196,193,41,114,242,5 // vpslld $0x5,%xmm10,%xmm10
+ .byte 196,67,37,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ .byte 196,65,45,86,201 // vorpd %ymm9,%ymm10,%ymm9
+ .byte 197,60,89,194 // vmulps %ymm2,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,65,53,86,192 // vorpd %ymm8,%ymm9,%ymm8
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,66,57,43,193 // vpackusdw %xmm9,%xmm8,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,10 // jne e9e <_sk_store_565_avx+0x9e>
+ .byte 196,65,122,127,4,121 // vmovdqu %xmm8,(%r9,%rdi,2)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 119,236 // ja e9a <_sk_store_565_avx+0x9a>
+ .byte 65,15,182,192 // movzbl %r8b,%eax
+ .byte 76,141,5,67,0,0,0 // lea 0x43(%rip),%r8 # efc <_sk_store_565_avx+0xfc>
+ .byte 73,99,4,128 // movslq (%r8,%rax,4),%rax
+ .byte 76,1,192 // add %r8,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,67,121,21,68,121,12,6 // vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,10,5 // vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,8,4 // vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,6,3 // vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,4,2 // vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
+ .byte 196,67,121,21,68,121,2,1 // vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
+ .byte 196,67,121,21,4,121,0 // vpextrw $0x0,%xmm8,(%r9,%rdi,2)
+ .byte 235,159 // jmp e9a <_sk_store_565_avx+0x9a>
+ .byte 144 // nop
+ .byte 246,255 // idiv %bh
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 238 // out %al,(%dx)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,230 // jmpq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 222,255 // fdivrp %st,%st(7)
+ .byte 255 // (bad)
+ .byte 255,214 // callq *%rsi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,206 // dec %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,198 // inc %esi
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_8888_avx
+_sk_load_8888_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,16 // mov (%rax),%r10
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,157,0,0,0 // jne fc3 <_sk_load_8888_avx+0xab>
+ .byte 196,65,124,16,12,186 // vmovups (%r10,%rdi,4),%ymm9
+ .byte 184,255,0,0,0 // mov $0xff,%eax
+ .byte 197,249,110,192 // vmovd %eax,%xmm0
+ .byte 197,249,112,192,0 // vpshufd $0x0,%xmm0,%xmm0
+ .byte 196,99,125,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
+ .byte 196,193,36,84,193 // vandps %ymm9,%ymm11,%ymm0
+ .byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,99,117,24,193,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm8
+ .byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
+ .byte 196,193,41,114,209,8 // vpsrld $0x8,%xmm9,%xmm10
+ .byte 196,99,125,25,203,1 // vextractf128 $0x1,%ymm9,%xmm3
+ .byte 197,241,114,211,8 // vpsrld $0x8,%xmm3,%xmm1
+ .byte 196,227,45,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm10,%ymm1
+ .byte 197,164,84,201 // vandps %ymm1,%ymm11,%ymm1
+ .byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
+ .byte 196,193,116,89,200 // vmulps %ymm8,%ymm1,%ymm1
+ .byte 196,193,41,114,209,16 // vpsrld $0x10,%xmm9,%xmm10
+ .byte 197,233,114,211,16 // vpsrld $0x10,%xmm3,%xmm2
+ .byte 196,227,45,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm10,%ymm2
+ .byte 197,164,84,210 // vandps %ymm2,%ymm11,%ymm2
+ .byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
+ .byte 196,193,108,89,208 // vmulps %ymm8,%ymm2,%ymm2
+ .byte 196,193,49,114,209,24 // vpsrld $0x18,%xmm9,%xmm9
+ .byte 197,225,114,211,24 // vpsrld $0x18,%xmm3,%xmm3
+ .byte 196,227,53,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm9,%ymm3
+ .byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
+ .byte 196,193,100,89,216 // vmulps %ymm8,%ymm3,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 196,65,52,87,201 // vxorps %ymm9,%ymm9,%ymm9
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 15,135,80,255,255,255 // ja f2c <_sk_load_8888_avx+0x14>
+ .byte 69,15,182,192 // movzbl %r8b,%r8d
+ .byte 76,141,13,137,0,0,0 // lea 0x89(%rip),%r9 # 1070 <_sk_load_8888_avx+0x158>
+ .byte 75,99,4,129 // movslq (%r9,%r8,4),%rax
+ .byte 76,1,200 // add %r9,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,193,121,110,68,186,24 // vmovd 0x18(%r10,%rdi,4),%xmm0
+ .byte 197,249,112,192,68 // vpshufd $0x44,%xmm0,%xmm0
+ .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 196,99,117,12,200,64 // vblendps $0x40,%ymm0,%ymm1,%ymm9
+ .byte 196,99,125,25,200,1 // vextractf128 $0x1,%ymm9,%xmm0
+ .byte 196,195,121,34,68,186,20,1 // vpinsrd $0x1,0x14(%r10,%rdi,4),%xmm0,%xmm0
+ .byte 196,99,53,24,200,1 // vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
+ .byte 196,99,125,25,200,1 // vextractf128 $0x1,%ymm9,%xmm0
+ .byte 196,195,121,34,68,186,16,0 // vpinsrd $0x0,0x10(%r10,%rdi,4),%xmm0,%xmm0
+ .byte 196,99,53,24,200,1 // vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
+ .byte 196,195,49,34,68,186,12,3 // vpinsrd $0x3,0xc(%r10,%rdi,4),%xmm9,%xmm0
+ .byte 196,99,53,12,200,15 // vblendps $0xf,%ymm0,%ymm9,%ymm9
+ .byte 196,195,49,34,68,186,8,2 // vpinsrd $0x2,0x8(%r10,%rdi,4),%xmm9,%xmm0
+ .byte 196,99,53,12,200,15 // vblendps $0xf,%ymm0,%ymm9,%ymm9
+ .byte 196,195,49,34,68,186,4,1 // vpinsrd $0x1,0x4(%r10,%rdi,4),%xmm9,%xmm0
+ .byte 196,99,53,12,200,15 // vblendps $0xf,%ymm0,%ymm9,%ymm9
+ .byte 196,195,49,34,4,186,0 // vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
+ .byte 196,99,53,12,200,15 // vblendps $0xf,%ymm0,%ymm9,%ymm9
+ .byte 233,188,254,255,255 // jmpq f2c <_sk_load_8888_avx+0x14>
+ .byte 238 // out %al,(%dx)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,224 // jmpq *%rax
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,210 // callq *%rdx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,196 // inc %esp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,176,255,255,255,156 // pushq -0x63000001(%rax)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+ .byte 128,255,255 // cmp $0xff,%bh
+ .byte 255 // .byte 0xff
+
+.globl _sk_store_8888_avx
+_sk_store_8888_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,8 // mov (%rax),%r9
+ .byte 184,0,0,127,67 // mov $0x437f0000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 196,65,125,91,201 // vcvtps2dq %ymm9,%ymm9
+ .byte 197,60,89,209 // vmulps %ymm1,%ymm8,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,33,114,242,8 // vpslld $0x8,%xmm10,%xmm11
+ .byte 196,67,125,25,210,1 // vextractf128 $0x1,%ymm10,%xmm10
+ .byte 196,193,41,114,242,8 // vpslld $0x8,%xmm10,%xmm10
+ .byte 196,67,37,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ .byte 196,65,45,86,201 // vorpd %ymm9,%ymm10,%ymm9
+ .byte 197,60,89,210 // vmulps %ymm2,%ymm8,%ymm10
+ .byte 196,65,125,91,210 // vcvtps2dq %ymm10,%ymm10
+ .byte 196,193,33,114,242,16 // vpslld $0x10,%xmm10,%xmm11
+ .byte 196,67,125,25,210,1 // vextractf128 $0x1,%ymm10,%xmm10
+ .byte 196,193,41,114,242,16 // vpslld $0x10,%xmm10,%xmm10
+ .byte 196,67,37,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ .byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
+ .byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
+ .byte 196,193,33,114,240,24 // vpslld $0x18,%xmm8,%xmm11
+ .byte 196,67,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm8
+ .byte 196,193,57,114,240,24 // vpslld $0x18,%xmm8,%xmm8
+ .byte 196,67,37,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
+ .byte 196,65,45,86,192 // vorpd %ymm8,%ymm10,%ymm8
+ .byte 196,65,53,86,192 // vorpd %ymm8,%ymm9,%ymm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,10 // jne 1130 <_sk_store_8888_avx+0xa4>
+ .byte 196,65,124,17,4,185 // vmovups %ymm8,(%r9,%rdi,4)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 65,137,200 // mov %ecx,%r8d
+ .byte 65,128,224,7 // and $0x7,%r8b
+ .byte 65,254,200 // dec %r8b
+ .byte 65,128,248,6 // cmp $0x6,%r8b
+ .byte 119,236 // ja 112c <_sk_store_8888_avx+0xa0>
+ .byte 65,15,182,192 // movzbl %r8b,%eax
+ .byte 76,141,5,85,0,0,0 // lea 0x55(%rip),%r8 # 11a0 <_sk_store_8888_avx+0x114>
+ .byte 73,99,4,128 // movslq (%r8,%rax,4),%rax
+ .byte 76,1,192 // add %r8,%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,67,121,22,76,185,24,2 // vpextrd $0x2,%xmm9,0x18(%r9,%rdi,4)
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,67,121,22,76,185,20,1 // vpextrd $0x1,%xmm9,0x14(%r9,%rdi,4)
+ .byte 196,67,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm9
+ .byte 196,65,122,17,76,185,16 // vmovss %xmm9,0x10(%r9,%rdi,4)
+ .byte 196,67,121,22,68,185,12,3 // vpextrd $0x3,%xmm8,0xc(%r9,%rdi,4)
+ .byte 196,67,121,22,68,185,8,2 // vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
+ .byte 196,67,121,22,68,185,4,1 // vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
+ .byte 196,65,121,126,4,185 // vmovd %xmm8,(%r9,%rdi,4)
+ .byte 235,143 // jmp 112c <_sk_store_8888_avx+0xa0>
+ .byte 15,31,0 // nopl (%rax)
+ .byte 245 // cmc
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 237 // in (%dx),%eax
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,229 // jmpq *%rbp
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 221,255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,208 // callq *%rax
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255,194 // inc %edx
+ .byte 255 // (bad)
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+ .byte 180,255 // mov $0xff,%ah
+ .byte 255 // (bad)
+ .byte 255 // .byte 0xff
+
+.globl _sk_load_f16_avx
+_sk_load_f16_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 15,133,2,1,0,0 // jne 12cc <_sk_load_f16_avx+0x110>
+ .byte 197,121,16,4,248 // vmovupd (%rax,%rdi,8),%xmm8
+ .byte 197,249,16,84,248,16 // vmovupd 0x10(%rax,%rdi,8),%xmm2
+ .byte 197,249,16,92,248,32 // vmovupd 0x20(%rax,%rdi,8),%xmm3
+ .byte 197,122,111,76,248,48 // vmovdqu 0x30(%rax,%rdi,8),%xmm9
+ .byte 197,185,97,194 // vpunpcklwd %xmm2,%xmm8,%xmm0
+ .byte 197,185,105,210 // vpunpckhwd %xmm2,%xmm8,%xmm2
+ .byte 196,193,97,97,201 // vpunpcklwd %xmm9,%xmm3,%xmm1
+ .byte 196,193,97,105,217 // vpunpckhwd %xmm9,%xmm3,%xmm3
+ .byte 197,121,97,194 // vpunpcklwd %xmm2,%xmm0,%xmm8
+ .byte 197,249,105,194 // vpunpckhwd %xmm2,%xmm0,%xmm0
+ .byte 197,241,97,211 // vpunpcklwd %xmm3,%xmm1,%xmm2
+ .byte 197,113,105,203 // vpunpckhwd %xmm3,%xmm1,%xmm9
+ .byte 184,0,4,0,4 // mov $0x4000400,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 197,249,112,219,0 // vpshufd $0x0,%xmm3,%xmm3
+ .byte 196,193,97,101,200 // vpcmpgtw %xmm8,%xmm3,%xmm1
+ .byte 196,65,113,223,192 // vpandn %xmm8,%xmm1,%xmm8
+ .byte 197,225,101,200 // vpcmpgtw %xmm0,%xmm3,%xmm1
+ .byte 197,241,223,192 // vpandn %xmm0,%xmm1,%xmm0
+ .byte 197,225,101,202 // vpcmpgtw %xmm2,%xmm3,%xmm1
+ .byte 197,241,223,202 // vpandn %xmm2,%xmm1,%xmm1
+ .byte 196,193,97,101,209 // vpcmpgtw %xmm9,%xmm3,%xmm2
+ .byte 196,193,105,223,209 // vpandn %xmm9,%xmm2,%xmm2
+ .byte 196,66,121,51,208 // vpmovzxwd %xmm8,%xmm10
+ .byte 196,98,121,51,201 // vpmovzxwd %xmm1,%xmm9
+ .byte 197,225,239,219 // vpxor %xmm3,%xmm3,%xmm3
+ .byte 197,57,105,195 // vpunpckhwd %xmm3,%xmm8,%xmm8
+ .byte 197,241,105,203 // vpunpckhwd %xmm3,%xmm1,%xmm1
+ .byte 196,98,121,51,216 // vpmovzxwd %xmm0,%xmm11
+ .byte 196,98,121,51,226 // vpmovzxwd %xmm2,%xmm12
+ .byte 197,121,105,235 // vpunpckhwd %xmm3,%xmm0,%xmm13
+ .byte 197,105,105,243 // vpunpckhwd %xmm3,%xmm2,%xmm14
+ .byte 196,193,121,114,242,13 // vpslld $0xd,%xmm10,%xmm0
+ .byte 196,193,105,114,241,13 // vpslld $0xd,%xmm9,%xmm2
+ .byte 196,227,125,24,194,1 // vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
+ .byte 184,0,0,128,119 // mov $0x77800000,%eax
+ .byte 197,249,110,208 // vmovd %eax,%xmm2
+ .byte 197,249,112,210,0 // vpshufd $0x0,%xmm2,%xmm2
+ .byte 196,99,109,24,202,1 // vinsertf128 $0x1,%xmm2,%ymm2,%ymm9
+ .byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
+ .byte 196,193,105,114,240,13 // vpslld $0xd,%xmm8,%xmm2
+ .byte 197,241,114,241,13 // vpslld $0xd,%xmm1,%xmm1
+ .byte 196,227,109,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
+ .byte 197,180,89,201 // vmulps %ymm1,%ymm9,%ymm1
+ .byte 196,193,105,114,243,13 // vpslld $0xd,%xmm11,%xmm2
+ .byte 196,193,97,114,244,13 // vpslld $0xd,%xmm12,%xmm3
+ .byte 196,227,109,24,211,1 // vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ .byte 197,180,89,210 // vmulps %ymm2,%ymm9,%ymm2
+ .byte 196,193,57,114,245,13 // vpslld $0xd,%xmm13,%xmm8
+ .byte 196,193,97,114,246,13 // vpslld $0xd,%xmm14,%xmm3
+ .byte 196,227,61,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ .byte 197,180,89,219 // vmulps %ymm3,%ymm9,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 197,123,16,4,248 // vmovsd (%rax,%rdi,8),%xmm8
+ .byte 196,65,49,239,201 // vpxor %xmm9,%xmm9,%xmm9
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,79 // je 132b <_sk_load_f16_avx+0x16f>
+ .byte 197,57,22,68,248,8 // vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,67 // jb 132b <_sk_load_f16_avx+0x16f>
+ .byte 197,251,16,84,248,16 // vmovsd 0x10(%rax,%rdi,8),%xmm2
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 116,68 // je 1338 <_sk_load_f16_avx+0x17c>
+ .byte 197,233,22,84,248,24 // vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,56 // jb 1338 <_sk_load_f16_avx+0x17c>
+ .byte 197,251,16,92,248,32 // vmovsd 0x20(%rax,%rdi,8),%xmm3
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 15,132,209,254,255,255 // je 11e1 <_sk_load_f16_avx+0x25>
+ .byte 197,225,22,92,248,40 // vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 15,130,193,254,255,255 // jb 11e1 <_sk_load_f16_avx+0x25>
+ .byte 197,122,126,76,248,48 // vmovq 0x30(%rax,%rdi,8),%xmm9
+ .byte 233,182,254,255,255 // jmpq 11e1 <_sk_load_f16_avx+0x25>
+ .byte 197,225,87,219 // vxorpd %xmm3,%xmm3,%xmm3
+ .byte 197,233,87,210 // vxorpd %xmm2,%xmm2,%xmm2
+ .byte 233,169,254,255,255 // jmpq 11e1 <_sk_load_f16_avx+0x25>
+ .byte 197,225,87,219 // vxorpd %xmm3,%xmm3,%xmm3
+ .byte 233,160,254,255,255 // jmpq 11e1 <_sk_load_f16_avx+0x25>
+
+.globl _sk_store_f16_avx
+_sk_store_f16_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,0 // mov (%rax),%r8
+ .byte 184,0,0,128,7 // mov $0x7800000,%eax
+ .byte 197,121,110,192 // vmovd %eax,%xmm8
+ .byte 196,65,121,112,192,0 // vpshufd $0x0,%xmm8,%xmm8
+ .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
+ .byte 196,67,125,25,202,1 // vextractf128 $0x1,%ymm9,%xmm10
+ .byte 196,193,41,114,210,13 // vpsrld $0xd,%xmm10,%xmm10
+ .byte 196,193,49,114,209,13 // vpsrld $0xd,%xmm9,%xmm9
+ .byte 197,60,89,217 // vmulps %ymm1,%ymm8,%ymm11
+ .byte 196,67,125,25,220,1 // vextractf128 $0x1,%ymm11,%xmm12
+ .byte 196,193,25,114,212,13 // vpsrld $0xd,%xmm12,%xmm12
+ .byte 196,193,33,114,211,13 // vpsrld $0xd,%xmm11,%xmm11
+ .byte 197,60,89,234 // vmulps %ymm2,%ymm8,%ymm13
+ .byte 196,67,125,25,238,1 // vextractf128 $0x1,%ymm13,%xmm14
+ .byte 196,193,9,114,214,13 // vpsrld $0xd,%xmm14,%xmm14
+ .byte 196,193,17,114,213,13 // vpsrld $0xd,%xmm13,%xmm13
+ .byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
+ .byte 196,67,125,25,199,1 // vextractf128 $0x1,%ymm8,%xmm15
+ .byte 196,193,1,114,215,13 // vpsrld $0xd,%xmm15,%xmm15
+ .byte 196,193,57,114,208,13 // vpsrld $0xd,%xmm8,%xmm8
+ .byte 196,193,33,115,251,2 // vpslldq $0x2,%xmm11,%xmm11
+ .byte 196,65,33,235,201 // vpor %xmm9,%xmm11,%xmm9
+ .byte 196,193,33,115,252,2 // vpslldq $0x2,%xmm12,%xmm11
+ .byte 196,65,33,235,226 // vpor %xmm10,%xmm11,%xmm12
+ .byte 196,193,57,115,248,2 // vpslldq $0x2,%xmm8,%xmm8
+ .byte 196,65,57,235,197 // vpor %xmm13,%xmm8,%xmm8
+ .byte 196,193,41,115,255,2 // vpslldq $0x2,%xmm15,%xmm10
+ .byte 196,65,41,235,238 // vpor %xmm14,%xmm10,%xmm13
+ .byte 196,65,49,98,216 // vpunpckldq %xmm8,%xmm9,%xmm11
+ .byte 196,65,49,106,208 // vpunpckhdq %xmm8,%xmm9,%xmm10
+ .byte 196,65,25,98,205 // vpunpckldq %xmm13,%xmm12,%xmm9
+ .byte 196,65,25,106,197 // vpunpckhdq %xmm13,%xmm12,%xmm8
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,31 // jne 1417 <_sk_store_f16_avx+0xd6>
+ .byte 196,65,120,17,28,248 // vmovups %xmm11,(%r8,%rdi,8)
+ .byte 196,65,120,17,84,248,16 // vmovups %xmm10,0x10(%r8,%rdi,8)
+ .byte 196,65,120,17,76,248,32 // vmovups %xmm9,0x20(%r8,%rdi,8)
+ .byte 196,65,122,127,68,248,48 // vmovdqu %xmm8,0x30(%r8,%rdi,8)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,65,121,214,28,248 // vmovq %xmm11,(%r8,%rdi,8)
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,240 // je 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,23,92,248,8 // vmovhpd %xmm11,0x8(%r8,%rdi,8)
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,227 // jb 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,214,84,248,16 // vmovq %xmm10,0x10(%r8,%rdi,8)
+ .byte 116,218 // je 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,23,84,248,24 // vmovhpd %xmm10,0x18(%r8,%rdi,8)
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,205 // jb 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,214,76,248,32 // vmovq %xmm9,0x20(%r8,%rdi,8)
+ .byte 116,196 // je 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,23,76,248,40 // vmovhpd %xmm9,0x28(%r8,%rdi,8)
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 114,183 // jb 1413 <_sk_store_f16_avx+0xd2>
+ .byte 196,65,121,214,68,248,48 // vmovq %xmm8,0x30(%r8,%rdi,8)
+ .byte 235,174 // jmp 1413 <_sk_store_f16_avx+0xd2>
+
+.globl _sk_store_f32_avx
+_sk_store_f32_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 76,139,0 // mov (%rax),%r8
+ .byte 72,141,4,189,0,0,0,0 // lea 0x0(,%rdi,4),%rax
+ .byte 197,124,20,193 // vunpcklps %ymm1,%ymm0,%ymm8
+ .byte 197,124,21,217 // vunpckhps %ymm1,%ymm0,%ymm11
+ .byte 197,108,20,203 // vunpcklps %ymm3,%ymm2,%ymm9
+ .byte 197,108,21,227 // vunpckhps %ymm3,%ymm2,%ymm12
+ .byte 196,65,61,20,209 // vunpcklpd %ymm9,%ymm8,%ymm10
+ .byte 196,65,61,21,201 // vunpckhpd %ymm9,%ymm8,%ymm9
+ .byte 196,65,37,20,196 // vunpcklpd %ymm12,%ymm11,%ymm8
+ .byte 196,65,37,21,220 // vunpckhpd %ymm12,%ymm11,%ymm11
+ .byte 72,133,201 // test %rcx,%rcx
+ .byte 117,55 // jne 14d2 <_sk_store_f32_avx+0x6d>
+ .byte 196,67,45,24,225,1 // vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
+ .byte 196,67,61,24,235,1 // vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
+ .byte 196,67,45,6,201,49 // vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
+ .byte 196,67,61,6,195,49 // vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
+ .byte 196,65,125,17,36,128 // vmovupd %ymm12,(%r8,%rax,4)
+ .byte 196,65,125,17,108,128,32 // vmovupd %ymm13,0x20(%r8,%rax,4)
+ .byte 196,65,125,17,76,128,64 // vmovupd %ymm9,0x40(%r8,%rax,4)
+ .byte 196,65,125,17,68,128,96 // vmovupd %ymm8,0x60(%r8,%rax,4)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+ .byte 196,65,121,17,20,128 // vmovupd %xmm10,(%r8,%rax,4)
+ .byte 72,131,249,1 // cmp $0x1,%rcx
+ .byte 116,240 // je 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,65,121,17,76,128,16 // vmovupd %xmm9,0x10(%r8,%rax,4)
+ .byte 72,131,249,3 // cmp $0x3,%rcx
+ .byte 114,227 // jb 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,65,121,17,68,128,32 // vmovupd %xmm8,0x20(%r8,%rax,4)
+ .byte 116,218 // je 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,65,121,17,92,128,48 // vmovupd %xmm11,0x30(%r8,%rax,4)
+ .byte 72,131,249,5 // cmp $0x5,%rcx
+ .byte 114,205 // jb 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,67,125,25,84,128,64,1 // vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
+ .byte 116,195 // je 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,67,125,25,76,128,80,1 // vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
+ .byte 72,131,249,7 // cmp $0x7,%rcx
+ .byte 114,181 // jb 14ce <_sk_store_f32_avx+0x69>
+ .byte 196,67,125,25,68,128,96,1 // vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
+ .byte 235,171 // jmp 14ce <_sk_store_f32_avx+0x69>
+
+.globl _sk_clamp_x_avx
+_sk_clamp_x_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,95,200 // vmaxps %ymm0,%ymm8,%ymm9
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,99,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm0
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,121,254,194 // vpaddd %xmm10,%xmm0,%xmm0
+ .byte 196,65,57,254,194 // vpaddd %xmm10,%xmm8,%xmm8
+ .byte 196,227,61,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
+ .byte 197,180,93,192 // vminps %ymm0,%ymm9,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_y_avx
+_sk_clamp_y_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,95,201 // vmaxps %ymm1,%ymm8,%ymm9
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,99,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm1
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,113,254,202 // vpaddd %xmm10,%xmm1,%xmm1
+ .byte 196,65,57,254,194 // vpaddd %xmm10,%xmm8,%xmm8
+ .byte 196,227,61,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
+ .byte 197,180,93,201 // vminps %ymm1,%ymm9,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_x_avx
+_sk_repeat_x_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,65,124,94,200 // vdivps %ymm8,%ymm0,%ymm9
+ .byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
+ .byte 196,65,52,89,200 // vmulps %ymm8,%ymm9,%ymm9
+ .byte 196,65,124,92,201 // vsubps %ymm9,%ymm0,%ymm9
+ .byte 196,99,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm0
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,121,254,194 // vpaddd %xmm10,%xmm0,%xmm0
+ .byte 196,65,57,254,194 // vpaddd %xmm10,%xmm8,%xmm8
+ .byte 196,227,61,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
+ .byte 197,180,93,192 // vminps %ymm0,%ymm9,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_y_avx
+_sk_repeat_y_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,65,116,94,200 // vdivps %ymm8,%ymm1,%ymm9
+ .byte 196,67,125,8,201,1 // vroundps $0x1,%ymm9,%ymm9
+ .byte 196,65,52,89,200 // vmulps %ymm8,%ymm9,%ymm9
+ .byte 196,65,116,92,201 // vsubps %ymm9,%ymm1,%ymm9
+ .byte 196,99,125,25,193,1 // vextractf128 $0x1,%ymm8,%xmm1
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,113,254,202 // vpaddd %xmm10,%xmm1,%xmm1
+ .byte 196,65,57,254,194 // vpaddd %xmm10,%xmm8,%xmm8
+ .byte 196,227,61,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
+ .byte 197,180,93,201 // vminps %ymm1,%ymm9,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_x_avx
+_sk_mirror_x_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,121,110,0 // vmovd (%rax),%xmm8
+ .byte 196,65,121,112,200,0 // vpshufd $0x0,%xmm8,%xmm9
+ .byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ .byte 196,65,124,92,209 // vsubps %ymm9,%ymm0,%ymm10
+ .byte 196,193,58,88,192 // vaddss %xmm8,%xmm8,%xmm0
+ .byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
+ .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 197,44,94,192 // vdivps %ymm0,%ymm10,%ymm8
+ .byte 196,67,125,8,192,1 // vroundps $0x1,%ymm8,%ymm8
+ .byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
+ .byte 197,172,92,192 // vsubps %ymm0,%ymm10,%ymm0
+ .byte 196,193,124,92,193 // vsubps %ymm9,%ymm0,%ymm0
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,92,192 // vsubps %ymm0,%ymm8,%ymm8
+ .byte 197,60,84,192 // vandps %ymm0,%ymm8,%ymm8
+ .byte 196,99,125,25,200,1 // vextractf128 $0x1,%ymm9,%xmm0
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,121,254,194 // vpaddd %xmm10,%xmm0,%xmm0
+ .byte 196,65,49,254,202 // vpaddd %xmm10,%xmm9,%xmm9
+ .byte 196,227,53,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm9,%ymm0
+ .byte 197,188,93,192 // vminps %ymm0,%ymm8,%ymm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_y_avx
+_sk_mirror_y_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,121,110,0 // vmovd (%rax),%xmm8
+ .byte 196,65,121,112,200,0 // vpshufd $0x0,%xmm8,%xmm9
+ .byte 196,67,53,24,201,1 // vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ .byte 196,65,116,92,209 // vsubps %ymm9,%ymm1,%ymm10
+ .byte 196,193,58,88,200 // vaddss %xmm8,%xmm8,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,44,94,193 // vdivps %ymm1,%ymm10,%ymm8
+ .byte 196,67,125,8,192,1 // vroundps $0x1,%ymm8,%ymm8
+ .byte 197,188,89,201 // vmulps %ymm1,%ymm8,%ymm1
+ .byte 197,172,92,201 // vsubps %ymm1,%ymm10,%ymm1
+ .byte 196,193,116,92,201 // vsubps %ymm9,%ymm1,%ymm1
+ .byte 196,65,60,87,192 // vxorps %ymm8,%ymm8,%ymm8
+ .byte 197,60,92,193 // vsubps %ymm1,%ymm8,%ymm8
+ .byte 197,60,84,193 // vandps %ymm1,%ymm8,%ymm8
+ .byte 196,99,125,25,201,1 // vextractf128 $0x1,%ymm9,%xmm1
+ .byte 196,65,41,118,210 // vpcmpeqd %xmm10,%xmm10,%xmm10
+ .byte 196,193,113,254,202 // vpaddd %xmm10,%xmm1,%xmm1
+ .byte 196,65,49,254,202 // vpaddd %xmm10,%xmm9,%xmm9
+ .byte 196,227,53,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm9,%ymm1
+ .byte 197,188,93,201 // vminps %ymm1,%ymm8,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_luminance_to_alpha_avx
+_sk_luminance_to_alpha_avx:
+ .byte 184,208,179,89,62 // mov $0x3e59b3d0,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,228,89,192 // vmulps %ymm0,%ymm3,%ymm0
+ .byte 184,89,23,55,63 // mov $0x3f371759,%eax
+ .byte 197,249,110,216 // vmovd %eax,%xmm3
+ .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
+ .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 197,228,89,201 // vmulps %ymm1,%ymm3,%ymm1
+ .byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
+ .byte 184,152,221,147,61 // mov $0x3d93dd98,%eax
+ .byte 197,249,110,200 // vmovd %eax,%xmm1
+ .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
+ .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 197,244,89,202 // vmulps %ymm2,%ymm1,%ymm1
+ .byte 197,252,88,217 // vaddps %ymm1,%ymm0,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
+ .byte 197,244,87,201 // vxorps %ymm1,%ymm1,%ymm1
+ .byte 197,236,87,210 // vxorps %ymm2,%ymm2,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_2x3_avx
+_sk_matrix_2x3_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,24,72,8 // vbroadcastss 0x8(%rax),%ymm9
+ .byte 196,98,125,24,80,16 // vbroadcastss 0x10(%rax),%ymm10
+ .byte 197,52,89,201 // vmulps %ymm1,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 197,60,89,192 // vmulps %ymm0,%ymm8,%ymm8
+ .byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
+ .byte 196,98,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm9
+ .byte 196,98,125,24,80,12 // vbroadcastss 0xc(%rax),%ymm10
+ .byte 196,98,125,24,88,20 // vbroadcastss 0x14(%rax),%ymm11
+ .byte 197,172,89,201 // vmulps %ymm1,%ymm10,%ymm1
+ .byte 196,193,116,88,203 // vaddps %ymm11,%ymm1,%ymm1
+ .byte 197,180,89,192 // vmulps %ymm0,%ymm9,%ymm0
+ .byte 197,252,88,201 // vaddps %ymm1,%ymm0,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_3x4_avx
+_sk_matrix_3x4_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,24,72,12 // vbroadcastss 0xc(%rax),%ymm9
+ .byte 196,98,125,24,80,24 // vbroadcastss 0x18(%rax),%ymm10
+ .byte 196,98,125,24,88,36 // vbroadcastss 0x24(%rax),%ymm11
+ .byte 197,44,89,210 // vmulps %ymm2,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 197,52,89,201 // vmulps %ymm1,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 197,60,89,192 // vmulps %ymm0,%ymm8,%ymm8
+ .byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
+ .byte 196,98,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm9
+ .byte 196,98,125,24,80,16 // vbroadcastss 0x10(%rax),%ymm10
+ .byte 196,98,125,24,88,28 // vbroadcastss 0x1c(%rax),%ymm11
+ .byte 196,98,125,24,96,40 // vbroadcastss 0x28(%rax),%ymm12
+ .byte 197,36,89,218 // vmulps %ymm2,%ymm11,%ymm11
+ .byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
+ .byte 197,44,89,209 // vmulps %ymm1,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 197,52,89,200 // vmulps %ymm0,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 196,98,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm10
+ .byte 196,98,125,24,88,20 // vbroadcastss 0x14(%rax),%ymm11
+ .byte 196,98,125,24,96,32 // vbroadcastss 0x20(%rax),%ymm12
+ .byte 196,98,125,24,104,44 // vbroadcastss 0x2c(%rax),%ymm13
+ .byte 197,156,89,210 // vmulps %ymm2,%ymm12,%ymm2
+ .byte 196,193,108,88,213 // vaddps %ymm13,%ymm2,%ymm2
+ .byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
+ .byte 197,244,88,202 // vaddps %ymm2,%ymm1,%ymm1
+ .byte 197,172,89,192 // vmulps %ymm0,%ymm10,%ymm0
+ .byte 197,252,88,209 // vaddps %ymm1,%ymm0,%ymm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 197,124,41,201 // vmovaps %ymm9,%ymm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_4x5_avx
+_sk_matrix_4x5_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,24,72,16 // vbroadcastss 0x10(%rax),%ymm9
+ .byte 196,98,125,24,80,32 // vbroadcastss 0x20(%rax),%ymm10
+ .byte 196,98,125,24,88,48 // vbroadcastss 0x30(%rax),%ymm11
+ .byte 196,98,125,24,96,64 // vbroadcastss 0x40(%rax),%ymm12
+ .byte 197,36,89,219 // vmulps %ymm3,%ymm11,%ymm11
+ .byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
+ .byte 197,44,89,210 // vmulps %ymm2,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 197,52,89,201 // vmulps %ymm1,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 197,60,89,192 // vmulps %ymm0,%ymm8,%ymm8
+ .byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
+ .byte 196,98,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm9
+ .byte 196,98,125,24,80,20 // vbroadcastss 0x14(%rax),%ymm10
+ .byte 196,98,125,24,88,36 // vbroadcastss 0x24(%rax),%ymm11
+ .byte 196,98,125,24,96,52 // vbroadcastss 0x34(%rax),%ymm12
+ .byte 196,98,125,24,104,68 // vbroadcastss 0x44(%rax),%ymm13
+ .byte 197,28,89,227 // vmulps %ymm3,%ymm12,%ymm12
+ .byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
+ .byte 197,36,89,218 // vmulps %ymm2,%ymm11,%ymm11
+ .byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
+ .byte 197,44,89,209 // vmulps %ymm1,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 197,52,89,200 // vmulps %ymm0,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 196,98,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm10
+ .byte 196,98,125,24,88,24 // vbroadcastss 0x18(%rax),%ymm11
+ .byte 196,98,125,24,96,40 // vbroadcastss 0x28(%rax),%ymm12
+ .byte 196,98,125,24,104,56 // vbroadcastss 0x38(%rax),%ymm13
+ .byte 196,98,125,24,112,72 // vbroadcastss 0x48(%rax),%ymm14
+ .byte 197,20,89,235 // vmulps %ymm3,%ymm13,%ymm13
+ .byte 196,65,20,88,238 // vaddps %ymm14,%ymm13,%ymm13
+ .byte 197,28,89,226 // vmulps %ymm2,%ymm12,%ymm12
+ .byte 196,65,28,88,229 // vaddps %ymm13,%ymm12,%ymm12
+ .byte 197,36,89,217 // vmulps %ymm1,%ymm11,%ymm11
+ .byte 196,65,36,88,220 // vaddps %ymm12,%ymm11,%ymm11
+ .byte 197,44,89,208 // vmulps %ymm0,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 196,98,125,24,88,12 // vbroadcastss 0xc(%rax),%ymm11
+ .byte 196,98,125,24,96,28 // vbroadcastss 0x1c(%rax),%ymm12
+ .byte 196,98,125,24,104,44 // vbroadcastss 0x2c(%rax),%ymm13
+ .byte 196,98,125,24,112,60 // vbroadcastss 0x3c(%rax),%ymm14
+ .byte 196,98,125,24,120,76 // vbroadcastss 0x4c(%rax),%ymm15
+ .byte 197,140,89,219 // vmulps %ymm3,%ymm14,%ymm3
+ .byte 196,193,100,88,223 // vaddps %ymm15,%ymm3,%ymm3
+ .byte 197,148,89,210 // vmulps %ymm2,%ymm13,%ymm2
+ .byte 197,236,88,211 // vaddps %ymm3,%ymm2,%ymm2
+ .byte 197,156,89,201 // vmulps %ymm1,%ymm12,%ymm1
+ .byte 197,244,88,202 // vaddps %ymm2,%ymm1,%ymm1
+ .byte 197,164,89,192 // vmulps %ymm0,%ymm11,%ymm0
+ .byte 197,252,88,217 // vaddps %ymm1,%ymm0,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 197,124,41,201 // vmovaps %ymm9,%ymm1
+ .byte 197,124,41,210 // vmovaps %ymm10,%ymm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_perspective_avx
+_sk_matrix_perspective_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,98,125,24,0 // vbroadcastss (%rax),%ymm8
+ .byte 196,98,125,24,72,4 // vbroadcastss 0x4(%rax),%ymm9
+ .byte 196,98,125,24,80,8 // vbroadcastss 0x8(%rax),%ymm10
+ .byte 197,52,89,201 // vmulps %ymm1,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 197,60,89,192 // vmulps %ymm0,%ymm8,%ymm8
+ .byte 196,65,60,88,193 // vaddps %ymm9,%ymm8,%ymm8
+ .byte 196,98,125,24,72,12 // vbroadcastss 0xc(%rax),%ymm9
+ .byte 196,98,125,24,80,16 // vbroadcastss 0x10(%rax),%ymm10
+ .byte 196,98,125,24,88,20 // vbroadcastss 0x14(%rax),%ymm11
+ .byte 197,44,89,209 // vmulps %ymm1,%ymm10,%ymm10
+ .byte 196,65,44,88,211 // vaddps %ymm11,%ymm10,%ymm10
+ .byte 197,52,89,200 // vmulps %ymm0,%ymm9,%ymm9
+ .byte 196,65,52,88,202 // vaddps %ymm10,%ymm9,%ymm9
+ .byte 196,98,125,24,80,24 // vbroadcastss 0x18(%rax),%ymm10
+ .byte 196,98,125,24,88,28 // vbroadcastss 0x1c(%rax),%ymm11
+ .byte 196,98,125,24,96,32 // vbroadcastss 0x20(%rax),%ymm12
+ .byte 197,164,89,201 // vmulps %ymm1,%ymm11,%ymm1
+ .byte 196,193,116,88,204 // vaddps %ymm12,%ymm1,%ymm1
+ .byte 197,172,89,192 // vmulps %ymm0,%ymm10,%ymm0
+ .byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
+ .byte 197,252,83,200 // vrcpps %ymm0,%ymm1
+ .byte 197,188,89,193 // vmulps %ymm1,%ymm8,%ymm0
+ .byte 197,180,89,201 // vmulps %ymm1,%ymm9,%ymm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_linear_gradient_2stops_avx
+_sk_linear_gradient_2stops_avx:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 196,226,125,24,72,16 // vbroadcastss 0x10(%rax),%ymm1
+ .byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
+ .byte 197,244,89,200 // vmulps %ymm0,%ymm1,%ymm1
+ .byte 197,108,88,193 // vaddps %ymm1,%ymm2,%ymm8
+ .byte 196,226,125,24,72,20 // vbroadcastss 0x14(%rax),%ymm1
+ .byte 196,226,125,24,80,4 // vbroadcastss 0x4(%rax),%ymm2
+ .byte 197,244,89,200 // vmulps %ymm0,%ymm1,%ymm1
+ .byte 197,236,88,201 // vaddps %ymm1,%ymm2,%ymm1
+ .byte 196,226,125,24,80,24 // vbroadcastss 0x18(%rax),%ymm2
+ .byte 196,226,125,24,88,8 // vbroadcastss 0x8(%rax),%ymm3
+ .byte 197,236,89,208 // vmulps %ymm0,%ymm2,%ymm2
+ .byte 197,228,88,210 // vaddps %ymm2,%ymm3,%ymm2
+ .byte 196,226,125,24,88,28 // vbroadcastss 0x1c(%rax),%ymm3
+ .byte 196,98,125,24,72,12 // vbroadcastss 0xc(%rax),%ymm9
+ .byte 197,228,89,192 // vmulps %ymm0,%ymm3,%ymm0
+ .byte 197,180,88,216 // vaddps %ymm0,%ymm9,%ymm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 197,124,41,192 // vmovaps %ymm8,%ymm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_start_pipeline_sse41
+_sk_start_pipeline_sse41:
+ .byte 65,87 // push %r15
+ .byte 65,86 // push %r14
+ .byte 65,85 // push %r13
+ .byte 65,84 // push %r12
+ .byte 83 // push %rbx
+ .byte 73,137,207 // mov %rcx,%r15
+ .byte 73,137,214 // mov %rdx,%r14
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 73,137,196 // mov %rax,%r12
+ .byte 73,137,245 // mov %rsi,%r13
+ .byte 72,141,67,4 // lea 0x4(%rbx),%rax
+ .byte 76,57,248 // cmp %r15,%rax
+ .byte 118,5 // jbe 28 <_sk_start_pipeline_sse41+0x28>
+ .byte 72,137,216 // mov %rbx,%rax
+ .byte 235,52 // jmp 5c <_sk_start_pipeline_sse41+0x5c>
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 15,87,228 // xorps %xmm4,%xmm4
+ .byte 15,87,237 // xorps %xmm5,%xmm5
+ .byte 15,87,246 // xorps %xmm6,%xmm6
+ .byte 15,87,255 // xorps %xmm7,%xmm7
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 76,137,238 // mov %r13,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,212 // callq *%r12
+ .byte 72,141,67,4 // lea 0x4(%rbx),%rax
+ .byte 72,131,195,8 // add $0x8,%rbx
+ .byte 76,57,251 // cmp %r15,%rbx
+ .byte 72,137,195 // mov %rax,%rbx
+ .byte 118,204 // jbe 28 <_sk_start_pipeline_sse41+0x28>
+ .byte 91 // pop %rbx
+ .byte 65,92 // pop %r12
+ .byte 65,93 // pop %r13
+ .byte 65,94 // pop %r14
+ .byte 65,95 // pop %r15
+ .byte 195 // retq
+
+.globl _sk_just_return_sse41
+_sk_just_return_sse41:
+ .byte 195 // retq
+
+.globl _sk_seed_shader_sse41
+_sk_seed_shader_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 102,15,110,199 // movd %edi,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
+ .byte 185,0,0,0,63 // mov $0x3f000000,%ecx
+ .byte 102,15,110,209 // movd %ecx,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 15,16,2 // movups (%rdx),%xmm0
+ .byte 15,88,193 // addps %xmm1,%xmm0
+ .byte 102,15,110,8 // movd (%rax),%xmm1
+ .byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
+ .byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 15,87,228 // xorps %xmm4,%xmm4
+ .byte 15,87,237 // xorps %xmm5,%xmm5
+ .byte 15,87,246 // xorps %xmm6,%xmm6
+ .byte 15,87,255 // xorps %xmm7,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_constant_color_sse41
+_sk_constant_color_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,16,24 // movups (%rax),%xmm3
+ .byte 15,40,195 // movaps %xmm3,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,40,203 // movaps %xmm3,%xmm1
+ .byte 15,198,201,85 // shufps $0x55,%xmm1,%xmm1
+ .byte 15,40,211 // movaps %xmm3,%xmm2
+ .byte 15,198,210,170 // shufps $0xaa,%xmm2,%xmm2
+ .byte 15,198,219,255 // shufps $0xff,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clear_sse41
+_sk_clear_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_plus__sse41
+_sk_plus__sse41:
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_srcover_sse41
+_sk_srcover_sse41:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,92,195 // subps %xmm3,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,204 // mulps %xmm4,%xmm9
+ .byte 65,15,88,193 // addps %xmm9,%xmm0
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,205 // mulps %xmm5,%xmm9
+ .byte 65,15,88,201 // addps %xmm9,%xmm1
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,206 // mulps %xmm6,%xmm9
+ .byte 65,15,88,209 // addps %xmm9,%xmm2
+ .byte 68,15,89,199 // mulps %xmm7,%xmm8
+ .byte 65,15,88,216 // addps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_dstover_sse41
+_sk_dstover_sse41:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,92,199 // subps %xmm7,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_0_sse41
+_sk_clamp_0_sse41:
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 65,15,95,192 // maxps %xmm8,%xmm0
+ .byte 65,15,95,200 // maxps %xmm8,%xmm1
+ .byte 65,15,95,208 // maxps %xmm8,%xmm2
+ .byte 65,15,95,216 // maxps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_1_sse41
+_sk_clamp_1_sse41:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,93,192 // minps %xmm8,%xmm0
+ .byte 65,15,93,200 // minps %xmm8,%xmm1
+ .byte 65,15,93,208 // minps %xmm8,%xmm2
+ .byte 65,15,93,216 // minps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_a_sse41
+_sk_clamp_a_sse41:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,93,216 // minps %xmm8,%xmm3
+ .byte 15,93,195 // minps %xmm3,%xmm0
+ .byte 15,93,203 // minps %xmm3,%xmm1
+ .byte 15,93,211 // minps %xmm3,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_set_rgb_sse41
+_sk_set_rgb_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,15,16,80,8 // movss 0x8(%rax),%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_rb_sse41
+_sk_swap_rb_sse41:
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,194 // movaps %xmm2,%xmm0
+ .byte 65,15,40,208 // movaps %xmm8,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_sse41
+_sk_swap_sse41:
+ .byte 68,15,40,195 // movaps %xmm3,%xmm8
+ .byte 68,15,40,202 // movaps %xmm2,%xmm9
+ .byte 68,15,40,209 // movaps %xmm1,%xmm10
+ .byte 68,15,40,216 // movaps %xmm0,%xmm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,196 // movaps %xmm4,%xmm0
+ .byte 15,40,205 // movaps %xmm5,%xmm1
+ .byte 15,40,214 // movaps %xmm6,%xmm2
+ .byte 15,40,223 // movaps %xmm7,%xmm3
+ .byte 65,15,40,227 // movaps %xmm11,%xmm4
+ .byte 65,15,40,234 // movaps %xmm10,%xmm5
+ .byte 65,15,40,241 // movaps %xmm9,%xmm6
+ .byte 65,15,40,248 // movaps %xmm8,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_src_dst_sse41
+_sk_move_src_dst_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,224 // movaps %xmm0,%xmm4
+ .byte 15,40,233 // movaps %xmm1,%xmm5
+ .byte 15,40,242 // movaps %xmm2,%xmm6
+ .byte 15,40,251 // movaps %xmm3,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_dst_src_sse41
+_sk_move_dst_src_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,196 // movaps %xmm4,%xmm0
+ .byte 15,40,205 // movaps %xmm5,%xmm1
+ .byte 15,40,214 // movaps %xmm6,%xmm2
+ .byte 15,40,223 // movaps %xmm7,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_premul_sse41
+_sk_premul_sse41:
+ .byte 15,89,195 // mulps %xmm3,%xmm0
+ .byte 15,89,203 // mulps %xmm3,%xmm1
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_unpremul_sse41
+_sk_unpremul_sse41:
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 68,15,94,203 // divps %xmm3,%xmm9
+ .byte 68,15,194,195,4 // cmpneqps %xmm3,%xmm8
+ .byte 69,15,84,193 // andps %xmm9,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_from_srgb_sse41
+_sk_from_srgb_sse41:
+ .byte 184,145,131,158,61 // mov $0x3d9e8391,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,40,211 // movaps %xmm11,%xmm10
+ .byte 68,15,89,208 // mulps %xmm0,%xmm10
+ .byte 68,15,40,240 // movaps %xmm0,%xmm14
+ .byte 69,15,89,246 // mulps %xmm14,%xmm14
+ .byte 184,154,153,153,62 // mov $0x3e99999a,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 184,92,143,50,63 // mov $0x3f328f5c,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 69,15,88,204 // addps %xmm12,%xmm9
+ .byte 184,10,215,35,59 // mov $0x3b23d70a,%eax
+ .byte 102,68,15,110,232 // movd %eax,%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 69,15,89,206 // mulps %xmm14,%xmm9
+ .byte 69,15,88,205 // addps %xmm13,%xmm9
+ .byte 184,174,71,97,61 // mov $0x3d6147ae,%eax
+ .byte 102,68,15,110,240 // movd %eax,%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 65,15,194,198,1 // cmpltps %xmm14,%xmm0
+ .byte 102,69,15,56,20,202 // blendvps %xmm0,%xmm10,%xmm9
+ .byte 69,15,40,251 // movaps %xmm11,%xmm15
+ .byte 68,15,89,249 // mulps %xmm1,%xmm15
+ .byte 15,40,193 // movaps %xmm1,%xmm0
+ .byte 15,89,192 // mulps %xmm0,%xmm0
+ .byte 69,15,40,208 // movaps %xmm8,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 69,15,88,212 // addps %xmm12,%xmm10
+ .byte 68,15,89,208 // mulps %xmm0,%xmm10
+ .byte 69,15,88,213 // addps %xmm13,%xmm10
+ .byte 65,15,194,206,1 // cmpltps %xmm14,%xmm1
+ .byte 15,40,193 // movaps %xmm1,%xmm0
+ .byte 102,69,15,56,20,215 // blendvps %xmm0,%xmm15,%xmm10
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 15,40,194 // movaps %xmm2,%xmm0
+ .byte 15,89,192 // mulps %xmm0,%xmm0
+ .byte 68,15,89,194 // mulps %xmm2,%xmm8
+ .byte 69,15,88,196 // addps %xmm12,%xmm8
+ .byte 68,15,89,192 // mulps %xmm0,%xmm8
+ .byte 69,15,88,197 // addps %xmm13,%xmm8
+ .byte 65,15,194,214,1 // cmpltps %xmm14,%xmm2
+ .byte 15,40,194 // movaps %xmm2,%xmm0
+ .byte 102,69,15,56,20,195 // blendvps %xmm0,%xmm11,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,193 // movaps %xmm9,%xmm0
+ .byte 65,15,40,202 // movaps %xmm10,%xmm1
+ .byte 65,15,40,208 // movaps %xmm8,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_to_srgb_sse41
+_sk_to_srgb_sse41:
+ .byte 72,131,236,24 // sub $0x18,%rsp
+ .byte 15,41,60,36 // movaps %xmm7,(%rsp)
+ .byte 15,40,254 // movaps %xmm6,%xmm7
+ .byte 15,40,245 // movaps %xmm5,%xmm6
+ .byte 15,40,236 // movaps %xmm4,%xmm5
+ .byte 15,40,227 // movaps %xmm3,%xmm4
+ .byte 15,40,218 // movaps %xmm2,%xmm3
+ .byte 15,40,209 // movaps %xmm1,%xmm2
+ .byte 68,15,82,192 // rsqrtps %xmm0,%xmm8
+ .byte 69,15,83,200 // rcpps %xmm8,%xmm9
+ .byte 69,15,82,248 // rsqrtps %xmm8,%xmm15
+ .byte 184,41,92,71,65 // mov $0x41475c29,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,40,211 // movaps %xmm11,%xmm10
+ .byte 68,15,89,208 // mulps %xmm0,%xmm10
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 184,194,135,210,62 // mov $0x3ed287c2,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 184,206,111,48,63 // mov $0x3f306fce,%eax
+ .byte 102,68,15,110,232 // movd %eax,%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 184,168,87,202,61 // mov $0x3dca57a8,%eax
+ .byte 53,0,0,0,128 // xor $0x80000000,%eax
+ .byte 102,68,15,110,240 // movd %eax,%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 69,15,89,205 // mulps %xmm13,%xmm9
+ .byte 69,15,88,206 // addps %xmm14,%xmm9
+ .byte 69,15,89,252 // mulps %xmm12,%xmm15
+ .byte 69,15,88,249 // addps %xmm9,%xmm15
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 69,15,93,207 // minps %xmm15,%xmm9
+ .byte 184,4,231,140,59 // mov $0x3b8ce704,%eax
+ .byte 102,68,15,110,248 // movd %eax,%xmm15
+ .byte 69,15,198,255,0 // shufps $0x0,%xmm15,%xmm15
+ .byte 65,15,194,199,1 // cmpltps %xmm15,%xmm0
+ .byte 102,69,15,56,20,202 // blendvps %xmm0,%xmm10,%xmm9
+ .byte 68,15,82,210 // rsqrtps %xmm2,%xmm10
+ .byte 65,15,83,194 // rcpps %xmm10,%xmm0
+ .byte 69,15,82,210 // rsqrtps %xmm10,%xmm10
+ .byte 65,15,89,197 // mulps %xmm13,%xmm0
+ .byte 65,15,88,198 // addps %xmm14,%xmm0
+ .byte 69,15,89,212 // mulps %xmm12,%xmm10
+ .byte 68,15,88,208 // addps %xmm0,%xmm10
+ .byte 65,15,40,200 // movaps %xmm8,%xmm1
+ .byte 65,15,93,202 // minps %xmm10,%xmm1
+ .byte 69,15,40,211 // movaps %xmm11,%xmm10
+ .byte 68,15,89,210 // mulps %xmm2,%xmm10
+ .byte 65,15,194,215,1 // cmpltps %xmm15,%xmm2
+ .byte 15,40,194 // movaps %xmm2,%xmm0
+ .byte 102,65,15,56,20,202 // blendvps %xmm0,%xmm10,%xmm1
+ .byte 15,82,195 // rsqrtps %xmm3,%xmm0
+ .byte 15,83,208 // rcpps %xmm0,%xmm2
+ .byte 65,15,89,213 // mulps %xmm13,%xmm2
+ .byte 65,15,88,214 // addps %xmm14,%xmm2
+ .byte 15,82,192 // rsqrtps %xmm0,%xmm0
+ .byte 65,15,89,196 // mulps %xmm12,%xmm0
+ .byte 15,88,194 // addps %xmm2,%xmm0
+ .byte 68,15,93,192 // minps %xmm0,%xmm8
+ .byte 68,15,89,219 // mulps %xmm3,%xmm11
+ .byte 65,15,194,223,1 // cmpltps %xmm15,%xmm3
+ .byte 15,40,195 // movaps %xmm3,%xmm0
+ .byte 102,69,15,56,20,195 // blendvps %xmm0,%xmm11,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,193 // movaps %xmm9,%xmm0
+ .byte 65,15,40,208 // movaps %xmm8,%xmm2
+ .byte 15,40,220 // movaps %xmm4,%xmm3
+ .byte 15,40,229 // movaps %xmm5,%xmm4
+ .byte 15,40,238 // movaps %xmm6,%xmm5
+ .byte 15,40,247 // movaps %xmm7,%xmm6
+ .byte 15,40,60,36 // movaps (%rsp),%xmm7
+ .byte 72,131,196,24 // add $0x18,%rsp
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_1_float_sse41
+_sk_scale_1_float_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_u8_sse41
+_sk_scale_u8_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,68,15,56,49,4,56 // pmovzxbd (%rax,%rdi,1),%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 65,15,89,193 // mulps %xmm9,%xmm0
+ .byte 65,15,89,201 // mulps %xmm9,%xmm1
+ .byte 65,15,89,209 // mulps %xmm9,%xmm2
+ .byte 65,15,89,217 // mulps %xmm9,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_1_float_sse41
+_sk_lerp_1_float_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,92,223 // subps %xmm7,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_u8_sse41
+_sk_lerp_u8_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,68,15,56,49,4,56 // pmovzxbd (%rax,%rdi,1),%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,193 // mulps %xmm9,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,201 // mulps %xmm9,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 65,15,89,209 // mulps %xmm9,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,92,223 // subps %xmm7,%xmm3
+ .byte 65,15,89,217 // mulps %xmm9,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_565_sse41
+_sk_lerp_565_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,68,15,56,51,4,120 // pmovzxwd (%rax,%rdi,2),%xmm8
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,203 // cvtdq2ps %xmm3,%xmm9
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 102,68,15,110,208 // movd %eax,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,203 // cvtdq2ps %xmm3,%xmm9
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,195 // cvtdq2ps %xmm3,%xmm8
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,194 // mulps %xmm10,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,203 // mulps %xmm11,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_tables_sse41
+_sk_load_tables_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,8 // mov (%rax),%rcx
+ .byte 76,139,64,8 // mov 0x8(%rax),%r8
+ .byte 243,68,15,111,4,185 // movdqu (%rcx,%rdi,4),%xmm8
+ .byte 185,255,0,0,0 // mov $0xff,%ecx
+ .byte 102,15,110,193 // movd %ecx,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,65,15,111,200 // movdqa %xmm8,%xmm1
+ .byte 102,15,114,209,8 // psrld $0x8,%xmm1
+ .byte 102,15,219,200 // pand %xmm0,%xmm1
+ .byte 102,65,15,111,208 // movdqa %xmm8,%xmm2
+ .byte 102,15,114,210,16 // psrld $0x10,%xmm2
+ .byte 102,15,219,208 // pand %xmm0,%xmm2
+ .byte 102,65,15,219,192 // pand %xmm8,%xmm0
+ .byte 102,72,15,58,22,193,1 // pextrq $0x1,%xmm0,%rcx
+ .byte 65,137,201 // mov %ecx,%r9d
+ .byte 72,193,233,32 // shr $0x20,%rcx
+ .byte 102,73,15,126,194 // movq %xmm0,%r10
+ .byte 69,137,211 // mov %r10d,%r11d
+ .byte 73,193,234,32 // shr $0x20,%r10
+ .byte 243,67,15,16,4,152 // movss (%r8,%r11,4),%xmm0
+ .byte 102,67,15,58,33,4,144,16 // insertps $0x10,(%r8,%r10,4),%xmm0
+ .byte 102,67,15,58,33,4,136,32 // insertps $0x20,(%r8,%r9,4),%xmm0
+ .byte 102,65,15,58,33,4,136,48 // insertps $0x30,(%r8,%rcx,4),%xmm0
+ .byte 76,139,64,16 // mov 0x10(%rax),%r8
+ .byte 102,73,15,58,22,202,1 // pextrq $0x1,%xmm1,%r10
+ .byte 77,137,209 // mov %r10,%r9
+ .byte 73,193,233,32 // shr $0x20,%r9
+ .byte 102,72,15,126,201 // movq %xmm1,%rcx
+ .byte 65,137,203 // mov %ecx,%r11d
+ .byte 65,129,227,255,255,255,0 // and $0xffffff,%r11d
+ .byte 72,193,233,30 // shr $0x1e,%rcx
+ .byte 65,129,226,255,255,255,0 // and $0xffffff,%r10d
+ .byte 243,67,15,16,12,152 // movss (%r8,%r11,4),%xmm1
+ .byte 102,65,15,58,33,12,8,16 // insertps $0x10,(%r8,%rcx,1),%xmm1
+ .byte 243,67,15,16,28,144 // movss (%r8,%r10,4),%xmm3
+ .byte 102,15,58,33,203,32 // insertps $0x20,%xmm3,%xmm1
+ .byte 243,67,15,16,28,136 // movss (%r8,%r9,4),%xmm3
+ .byte 102,15,58,33,203,48 // insertps $0x30,%xmm3,%xmm1
+ .byte 76,139,72,24 // mov 0x18(%rax),%r9
+ .byte 102,72,15,58,22,209,1 // pextrq $0x1,%xmm2,%rcx
+ .byte 68,15,183,193 // movzwl %cx,%r8d
+ .byte 72,193,233,32 // shr $0x20,%rcx
+ .byte 102,72,15,126,208 // movq %xmm2,%rax
+ .byte 68,15,183,208 // movzwl %ax,%r10d
+ .byte 72,193,232,30 // shr $0x1e,%rax
+ .byte 243,67,15,16,20,145 // movss (%r9,%r10,4),%xmm2
+ .byte 102,65,15,58,33,20,1,16 // insertps $0x10,(%r9,%rax,1),%xmm2
+ .byte 243,67,15,16,28,129 // movss (%r9,%r8,4),%xmm3
+ .byte 102,15,58,33,211,32 // insertps $0x20,%xmm3,%xmm2
+ .byte 243,65,15,16,28,137 // movss (%r9,%rcx,4),%xmm3
+ .byte 102,15,58,33,211,48 // insertps $0x30,%xmm3,%xmm2
+ .byte 102,65,15,114,208,24 // psrld $0x18,%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_a8_sse41
+_sk_load_a8_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,15,56,49,4,56 // pmovzxbd (%rax,%rdi,1),%xmm0
+ .byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_a8_sse41
+_sk_store_a8_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,127,67 // mov $0x437f0000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,69,15,56,43,192 // packusdw %xmm8,%xmm8
+ .byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
+ .byte 102,68,15,126,4,56 // movd %xmm8,(%rax,%rdi,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_565_sse41
+_sk_load_565_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,15,56,51,20,120 // pmovzxwd (%rax,%rdi,2),%xmm2
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,15,219,194 // pand %xmm2,%xmm0
+ .byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 102,15,110,200 // movd %eax,%xmm1
+ .byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
+ .byte 102,15,219,202 // pand %xmm2,%xmm1
+ .byte 15,91,217 // cvtdq2ps %xmm1,%xmm3
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 102,15,110,200 // movd %eax,%xmm1
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 15,89,203 // mulps %xmm3,%xmm1
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,15,219,218 // pand %xmm2,%xmm3
+ .byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_565_sse41
+_sk_store_565_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,248,65 // mov $0x41f80000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 102,65,15,114,241,11 // pslld $0xb,%xmm9
+ .byte 185,0,0,124,66 // mov $0x427c0000,%ecx
+ .byte 102,68,15,110,209 // movd %ecx,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
+ .byte 102,65,15,114,242,5 // pslld $0x5,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 68,15,89,194 // mulps %xmm2,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,69,15,86,194 // orpd %xmm10,%xmm8
+ .byte 102,69,15,56,43,192 // packusdw %xmm8,%xmm8
+ .byte 102,68,15,214,4,120 // movq %xmm8,(%rax,%rdi,2)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_8888_sse41
+_sk_load_8888_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,15,111,28,184 // movdqu (%rax,%rdi,4),%xmm3
+ .byte 184,255,0,0,0 // mov $0xff,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,15,111,203 // movdqa %xmm3,%xmm1
+ .byte 102,15,114,209,8 // psrld $0x8,%xmm1
+ .byte 102,15,219,200 // pand %xmm0,%xmm1
+ .byte 102,15,111,211 // movdqa %xmm3,%xmm2
+ .byte 102,15,114,210,16 // psrld $0x10,%xmm2
+ .byte 102,15,219,208 // pand %xmm0,%xmm2
+ .byte 102,15,219,195 // pand %xmm3,%xmm0
+ .byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 102,15,114,211,24 // psrld $0x18,%xmm3
+ .byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_8888_sse41
+_sk_store_8888_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,127,67 // mov $0x437f0000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 69,15,40,208 // movaps %xmm8,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
+ .byte 102,65,15,114,242,8 // pslld $0x8,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,202 // mulps %xmm2,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 102,65,15,114,241,16 // pslld $0x10,%xmm9
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,65,15,114,240,24 // pslld $0x18,%xmm8
+ .byte 102,69,15,235,193 // por %xmm9,%xmm8
+ .byte 102,69,15,235,194 // por %xmm10,%xmm8
+ .byte 243,68,15,127,4,184 // movdqu %xmm8,(%rax,%rdi,4)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_f16_sse41
+_sk_load_f16_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,15,111,4,248 // movdqu (%rax,%rdi,8),%xmm0
+ .byte 243,15,111,76,248,16 // movdqu 0x10(%rax,%rdi,8),%xmm1
+ .byte 102,15,111,208 // movdqa %xmm0,%xmm2
+ .byte 102,15,97,209 // punpcklwd %xmm1,%xmm2
+ .byte 102,15,105,193 // punpckhwd %xmm1,%xmm0
+ .byte 102,68,15,111,194 // movdqa %xmm2,%xmm8
+ .byte 102,68,15,97,192 // punpcklwd %xmm0,%xmm8
+ .byte 102,15,105,208 // punpckhwd %xmm0,%xmm2
+ .byte 184,0,4,0,4 // mov $0x4000400,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,216,0 // pshufd $0x0,%xmm0,%xmm3
+ .byte 102,15,111,203 // movdqa %xmm3,%xmm1
+ .byte 102,65,15,101,200 // pcmpgtw %xmm8,%xmm1
+ .byte 102,65,15,223,200 // pandn %xmm8,%xmm1
+ .byte 102,15,101,218 // pcmpgtw %xmm2,%xmm3
+ .byte 102,15,223,218 // pandn %xmm2,%xmm3
+ .byte 102,15,56,51,193 // pmovzxwd %xmm1,%xmm0
+ .byte 102,15,114,240,13 // pslld $0xd,%xmm0
+ .byte 184,0,0,128,119 // mov $0x77800000,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 102,68,15,112,194,0 // pshufd $0x0,%xmm2,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 102,69,15,239,201 // pxor %xmm9,%xmm9
+ .byte 102,65,15,105,201 // punpckhwd %xmm9,%xmm1
+ .byte 102,15,114,241,13 // pslld $0xd,%xmm1
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 102,15,56,51,211 // pmovzxwd %xmm3,%xmm2
+ .byte 102,15,114,242,13 // pslld $0xd,%xmm2
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 102,65,15,105,217 // punpckhwd %xmm9,%xmm3
+ .byte 102,15,114,243,13 // pslld $0xd,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_f16_sse41
+_sk_store_f16_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,128,7 // mov $0x7800000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 102,69,15,112,192,0 // pshufd $0x0,%xmm8,%xmm8
+ .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,65,15,114,209,13 // psrld $0xd,%xmm9
+ .byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,65,15,114,210,13 // psrld $0xd,%xmm10
+ .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 102,65,15,114,211,13 // psrld $0xd,%xmm11
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,65,15,114,208,13 // psrld $0xd,%xmm8
+ .byte 102,65,15,115,250,2 // pslldq $0x2,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 102,65,15,115,248,2 // pslldq $0x2,%xmm8
+ .byte 102,69,15,235,195 // por %xmm11,%xmm8
+ .byte 102,69,15,111,202 // movdqa %xmm10,%xmm9
+ .byte 102,69,15,98,200 // punpckldq %xmm8,%xmm9
+ .byte 243,68,15,127,12,248 // movdqu %xmm9,(%rax,%rdi,8)
+ .byte 102,69,15,106,208 // punpckhdq %xmm8,%xmm10
+ .byte 243,68,15,127,84,248,16 // movdqu %xmm10,0x10(%rax,%rdi,8)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_f32_sse41
+_sk_store_f32_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,137,249 // mov %rdi,%rcx
+ .byte 72,193,225,4 // shl $0x4,%rcx
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 68,15,40,200 // movaps %xmm0,%xmm9
+ .byte 68,15,20,201 // unpcklps %xmm1,%xmm9
+ .byte 68,15,40,210 // movaps %xmm2,%xmm10
+ .byte 68,15,40,218 // movaps %xmm2,%xmm11
+ .byte 68,15,20,219 // unpcklps %xmm3,%xmm11
+ .byte 68,15,21,193 // unpckhps %xmm1,%xmm8
+ .byte 68,15,21,211 // unpckhps %xmm3,%xmm10
+ .byte 69,15,40,225 // movaps %xmm9,%xmm12
+ .byte 102,69,15,20,227 // unpcklpd %xmm11,%xmm12
+ .byte 69,15,18,217 // movhlps %xmm9,%xmm11
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 102,69,15,20,202 // unpcklpd %xmm10,%xmm9
+ .byte 69,15,18,208 // movhlps %xmm8,%xmm10
+ .byte 102,68,15,17,36,8 // movupd %xmm12,(%rax,%rcx,1)
+ .byte 68,15,17,92,8,16 // movups %xmm11,0x10(%rax,%rcx,1)
+ .byte 102,68,15,17,76,8,32 // movupd %xmm9,0x20(%rax,%rcx,1)
+ .byte 68,15,17,84,8,48 // movups %xmm10,0x30(%rax,%rcx,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_x_sse41
+_sk_clamp_x_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,95,192 // maxps %xmm0,%xmm8
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 102,15,118,192 // pcmpeqd %xmm0,%xmm0
+ .byte 102,65,15,254,193 // paddd %xmm9,%xmm0
+ .byte 68,15,93,192 // minps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,192 // movaps %xmm8,%xmm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_y_sse41
+_sk_clamp_y_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,95,193 // maxps %xmm1,%xmm8
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 102,15,118,201 // pcmpeqd %xmm1,%xmm1
+ .byte 102,65,15,254,201 // paddd %xmm9,%xmm1
+ .byte 68,15,93,193 // minps %xmm1,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,200 // movaps %xmm8,%xmm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_x_sse41
+_sk_repeat_x_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,200 // movaps %xmm0,%xmm9
+ .byte 69,15,94,200 // divps %xmm8,%xmm9
+ .byte 102,69,15,58,8,201,1 // roundps $0x1,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 65,15,92,193 // subps %xmm9,%xmm0
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,193 // minps %xmm9,%xmm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_y_sse41
+_sk_repeat_y_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 69,15,94,200 // divps %xmm8,%xmm9
+ .byte 102,69,15,58,8,201,1 // roundps $0x1,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 65,15,92,201 // subps %xmm9,%xmm1
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,201 // minps %xmm9,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_x_sse41
+_sk_mirror_x_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 65,15,92,193 // subps %xmm9,%xmm0
+ .byte 243,69,15,88,192 // addss %xmm8,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,208 // movaps %xmm0,%xmm10
+ .byte 69,15,94,208 // divps %xmm8,%xmm10
+ .byte 102,69,15,58,8,210,1 // roundps $0x1,%xmm10,%xmm10
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 65,15,92,194 // subps %xmm10,%xmm0
+ .byte 65,15,92,193 // subps %xmm9,%xmm0
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,92,192 // subps %xmm0,%xmm8
+ .byte 65,15,84,192 // andps %xmm8,%xmm0
+ .byte 102,69,15,118,192 // pcmpeqd %xmm8,%xmm8
+ .byte 102,69,15,254,193 // paddd %xmm9,%xmm8
+ .byte 65,15,93,192 // minps %xmm8,%xmm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_y_sse41
+_sk_mirror_y_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 65,15,92,201 // subps %xmm9,%xmm1
+ .byte 243,69,15,88,192 // addss %xmm8,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,209 // movaps %xmm1,%xmm10
+ .byte 69,15,94,208 // divps %xmm8,%xmm10
+ .byte 102,69,15,58,8,210,1 // roundps $0x1,%xmm10,%xmm10
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 65,15,92,202 // subps %xmm10,%xmm1
+ .byte 65,15,92,201 // subps %xmm9,%xmm1
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,92,193 // subps %xmm1,%xmm8
+ .byte 65,15,84,200 // andps %xmm8,%xmm1
+ .byte 102,69,15,118,192 // pcmpeqd %xmm8,%xmm8
+ .byte 102,69,15,254,193 // paddd %xmm9,%xmm8
+ .byte 65,15,93,200 // minps %xmm8,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_luminance_to_alpha_sse41
+_sk_luminance_to_alpha_sse41:
+ .byte 184,208,179,89,62 // mov $0x3e59b3d0,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 184,89,23,55,63 // mov $0x3f371759,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 15,88,195 // addps %xmm3,%xmm0
+ .byte 184,152,221,147,61 // mov $0x3d93dd98,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,218 // mulps %xmm2,%xmm3
+ .byte 15,88,216 // addps %xmm0,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_2x3_sse41
+_sk_matrix_2x3_sse41:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,16 // movss 0x10(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,12 // movss 0xc(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_3x4_sse41
+_sk_matrix_3x4_sse41:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,12 // movss 0xc(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,24 // movss 0x18(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,36 // movss 0x24(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,28 // movss 0x1c(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,40 // movss 0x28(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,32 // movss 0x20(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,44 // movss 0x2c(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,226 // mulps %xmm2,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,210 // movaps %xmm10,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_4x5_sse41
+_sk_matrix_4x5_sse41:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,32 // movss 0x20(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,48 // movss 0x30(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,64 // movss 0x40(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,227 // mulps %xmm3,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,20 // movss 0x14(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,36 // movss 0x24(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,52 // movss 0x34(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,68 // movss 0x44(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,227 // mulps %xmm3,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,24 // movss 0x18(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,40 // movss 0x28(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,56 // movss 0x38(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 243,68,15,16,112,72 // movss 0x48(%rax),%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 68,15,89,235 // mulps %xmm3,%xmm13
+ .byte 69,15,88,238 // addps %xmm14,%xmm13
+ .byte 68,15,89,226 // mulps %xmm2,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 243,68,15,16,88,12 // movss 0xc(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,28 // movss 0x1c(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,44 // movss 0x2c(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 243,68,15,16,112,60 // movss 0x3c(%rax),%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 243,68,15,16,120,76 // movss 0x4c(%rax),%xmm15
+ .byte 69,15,198,255,0 // shufps $0x0,%xmm15,%xmm15
+ .byte 68,15,89,243 // mulps %xmm3,%xmm14
+ .byte 69,15,88,247 // addps %xmm15,%xmm14
+ .byte 68,15,89,234 // mulps %xmm2,%xmm13
+ .byte 69,15,88,238 // addps %xmm14,%xmm13
+ .byte 69,15,89,225 // mulps %xmm9,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,216 // mulps %xmm8,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,210 // movaps %xmm10,%xmm2
+ .byte 65,15,40,219 // movaps %xmm11,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_perspective_sse41
+_sk_matrix_perspective_sse41:
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,68,15,16,72,4 // movss 0x4(%rax),%xmm9
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 68,15,89,201 // mulps %xmm1,%xmm9
+ .byte 69,15,88,202 // addps %xmm10,%xmm9
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,193 // addps %xmm9,%xmm0
+ .byte 243,68,15,16,72,12 // movss 0xc(%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 69,15,88,202 // addps %xmm10,%xmm9
+ .byte 243,68,15,16,80,24 // movss 0x18(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,28 // movss 0x1c(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,32 // movss 0x20(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,217 // mulps %xmm1,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,83,202 // rcpps %xmm10,%xmm1
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 68,15,89,201 // mulps %xmm1,%xmm9
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,201 // movaps %xmm9,%xmm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_linear_gradient_2stops_sse41
+_sk_linear_gradient_2stops_sse41:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 68,15,16,8 // movups (%rax),%xmm9
+ .byte 15,16,88,16 // movups 0x10(%rax),%xmm3
+ .byte 68,15,40,195 // movaps %xmm3,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,40,201 // movaps %xmm9,%xmm1
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 68,15,89,192 // mulps %xmm0,%xmm8
+ .byte 68,15,88,193 // addps %xmm1,%xmm8
+ .byte 15,40,203 // movaps %xmm3,%xmm1
+ .byte 15,198,201,85 // shufps $0x55,%xmm1,%xmm1
+ .byte 65,15,40,209 // movaps %xmm9,%xmm2
+ .byte 15,198,210,85 // shufps $0x55,%xmm2,%xmm2
+ .byte 15,89,200 // mulps %xmm0,%xmm1
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 15,40,211 // movaps %xmm3,%xmm2
+ .byte 15,198,210,170 // shufps $0xaa,%xmm2,%xmm2
+ .byte 69,15,40,209 // movaps %xmm9,%xmm10
+ .byte 69,15,198,210,170 // shufps $0xaa,%xmm10,%xmm10
+ .byte 15,89,208 // mulps %xmm0,%xmm2
+ .byte 65,15,88,210 // addps %xmm10,%xmm2
+ .byte 15,198,219,255 // shufps $0xff,%xmm3,%xmm3
+ .byte 69,15,198,201,255 // shufps $0xff,%xmm9,%xmm9
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 65,15,88,217 // addps %xmm9,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,192 // movaps %xmm8,%xmm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_start_pipeline_sse2
+_sk_start_pipeline_sse2:
+ .byte 65,87 // push %r15
+ .byte 65,86 // push %r14
+ .byte 65,85 // push %r13
+ .byte 65,84 // push %r12
+ .byte 83 // push %rbx
+ .byte 73,137,207 // mov %rcx,%r15
+ .byte 73,137,214 // mov %rdx,%r14
+ .byte 72,137,251 // mov %rdi,%rbx
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 73,137,196 // mov %rax,%r12
+ .byte 73,137,245 // mov %rsi,%r13
+ .byte 72,141,67,4 // lea 0x4(%rbx),%rax
+ .byte 76,57,248 // cmp %r15,%rax
+ .byte 118,5 // jbe 28 <_sk_start_pipeline_sse2+0x28>
+ .byte 72,137,216 // mov %rbx,%rax
+ .byte 235,52 // jmp 5c <_sk_start_pipeline_sse2+0x5c>
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 15,87,228 // xorps %xmm4,%xmm4
+ .byte 15,87,237 // xorps %xmm5,%xmm5
+ .byte 15,87,246 // xorps %xmm6,%xmm6
+ .byte 15,87,255 // xorps %xmm7,%xmm7
+ .byte 72,137,223 // mov %rbx,%rdi
+ .byte 76,137,238 // mov %r13,%rsi
+ .byte 76,137,242 // mov %r14,%rdx
+ .byte 65,255,212 // callq *%r12
+ .byte 72,141,67,4 // lea 0x4(%rbx),%rax
+ .byte 72,131,195,8 // add $0x8,%rbx
+ .byte 76,57,251 // cmp %r15,%rbx
+ .byte 72,137,195 // mov %rax,%rbx
+ .byte 118,204 // jbe 28 <_sk_start_pipeline_sse2+0x28>
+ .byte 91 // pop %rbx
+ .byte 65,92 // pop %r12
+ .byte 65,93 // pop %r13
+ .byte 65,94 // pop %r14
+ .byte 65,95 // pop %r15
+ .byte 195 // retq
+
+.globl _sk_just_return_sse2
+_sk_just_return_sse2:
+ .byte 195 // retq
+
+.globl _sk_seed_shader_sse2
+_sk_seed_shader_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 102,15,110,199 // movd %edi,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
+ .byte 185,0,0,0,63 // mov $0x3f000000,%ecx
+ .byte 102,15,110,209 // movd %ecx,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 15,16,2 // movups (%rdx),%xmm0
+ .byte 15,88,193 // addps %xmm1,%xmm0
+ .byte 102,15,110,8 // movd (%rax),%xmm1
+ .byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
+ .byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 15,87,228 // xorps %xmm4,%xmm4
+ .byte 15,87,237 // xorps %xmm5,%xmm5
+ .byte 15,87,246 // xorps %xmm6,%xmm6
+ .byte 15,87,255 // xorps %xmm7,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_constant_color_sse2
+_sk_constant_color_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,16,24 // movups (%rax),%xmm3
+ .byte 15,40,195 // movaps %xmm3,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,40,203 // movaps %xmm3,%xmm1
+ .byte 15,198,201,85 // shufps $0x55,%xmm1,%xmm1
+ .byte 15,40,211 // movaps %xmm3,%xmm2
+ .byte 15,198,210,170 // shufps $0xaa,%xmm2,%xmm2
+ .byte 15,198,219,255 // shufps $0xff,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clear_sse2
+_sk_clear_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 15,87,219 // xorps %xmm3,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_plus__sse2
+_sk_plus__sse2:
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_srcover_sse2
+_sk_srcover_sse2:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,92,195 // subps %xmm3,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,204 // mulps %xmm4,%xmm9
+ .byte 65,15,88,193 // addps %xmm9,%xmm0
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,205 // mulps %xmm5,%xmm9
+ .byte 65,15,88,201 // addps %xmm9,%xmm1
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,206 // mulps %xmm6,%xmm9
+ .byte 65,15,88,209 // addps %xmm9,%xmm2
+ .byte 68,15,89,199 // mulps %xmm7,%xmm8
+ .byte 65,15,88,216 // addps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_dstover_sse2
+_sk_dstover_sse2:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,92,199 // subps %xmm7,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_0_sse2
+_sk_clamp_0_sse2:
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 65,15,95,192 // maxps %xmm8,%xmm0
+ .byte 65,15,95,200 // maxps %xmm8,%xmm1
+ .byte 65,15,95,208 // maxps %xmm8,%xmm2
+ .byte 65,15,95,216 // maxps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_1_sse2
+_sk_clamp_1_sse2:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,93,192 // minps %xmm8,%xmm0
+ .byte 65,15,93,200 // minps %xmm8,%xmm1
+ .byte 65,15,93,208 // minps %xmm8,%xmm2
+ .byte 65,15,93,216 // minps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_a_sse2
+_sk_clamp_a_sse2:
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,93,216 // minps %xmm8,%xmm3
+ .byte 15,93,195 // minps %xmm3,%xmm0
+ .byte 15,93,203 // minps %xmm3,%xmm1
+ .byte 15,93,211 // minps %xmm3,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_set_rgb_sse2
+_sk_set_rgb_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,15,16,80,8 // movss 0x8(%rax),%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_rb_sse2
+_sk_swap_rb_sse2:
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,194 // movaps %xmm2,%xmm0
+ .byte 65,15,40,208 // movaps %xmm8,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_swap_sse2
+_sk_swap_sse2:
+ .byte 68,15,40,195 // movaps %xmm3,%xmm8
+ .byte 68,15,40,202 // movaps %xmm2,%xmm9
+ .byte 68,15,40,209 // movaps %xmm1,%xmm10
+ .byte 68,15,40,216 // movaps %xmm0,%xmm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,196 // movaps %xmm4,%xmm0
+ .byte 15,40,205 // movaps %xmm5,%xmm1
+ .byte 15,40,214 // movaps %xmm6,%xmm2
+ .byte 15,40,223 // movaps %xmm7,%xmm3
+ .byte 65,15,40,227 // movaps %xmm11,%xmm4
+ .byte 65,15,40,234 // movaps %xmm10,%xmm5
+ .byte 65,15,40,241 // movaps %xmm9,%xmm6
+ .byte 65,15,40,248 // movaps %xmm8,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_src_dst_sse2
+_sk_move_src_dst_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,224 // movaps %xmm0,%xmm4
+ .byte 15,40,233 // movaps %xmm1,%xmm5
+ .byte 15,40,242 // movaps %xmm2,%xmm6
+ .byte 15,40,251 // movaps %xmm3,%xmm7
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_move_dst_src_sse2
+_sk_move_dst_src_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,40,196 // movaps %xmm4,%xmm0
+ .byte 15,40,205 // movaps %xmm5,%xmm1
+ .byte 15,40,214 // movaps %xmm6,%xmm2
+ .byte 15,40,223 // movaps %xmm7,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_premul_sse2
+_sk_premul_sse2:
+ .byte 15,89,195 // mulps %xmm3,%xmm0
+ .byte 15,89,203 // mulps %xmm3,%xmm1
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_unpremul_sse2
+_sk_unpremul_sse2:
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 68,15,94,203 // divps %xmm3,%xmm9
+ .byte 68,15,194,195,4 // cmpneqps %xmm3,%xmm8
+ .byte 69,15,84,193 // andps %xmm9,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_from_srgb_sse2
+_sk_from_srgb_sse2:
+ .byte 184,145,131,158,61 // mov $0x3d9e8391,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,232 // movaps %xmm8,%xmm13
+ .byte 68,15,89,232 // mulps %xmm0,%xmm13
+ .byte 68,15,40,224 // movaps %xmm0,%xmm12
+ .byte 69,15,89,228 // mulps %xmm12,%xmm12
+ .byte 184,154,153,153,62 // mov $0x3e99999a,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 184,92,143,50,63 // mov $0x3f328f5c,%eax
+ .byte 102,68,15,110,208 // movd %eax,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 69,15,40,241 // movaps %xmm9,%xmm14
+ .byte 68,15,89,240 // mulps %xmm0,%xmm14
+ .byte 69,15,88,242 // addps %xmm10,%xmm14
+ .byte 184,10,215,35,59 // mov $0x3b23d70a,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,244 // mulps %xmm12,%xmm14
+ .byte 69,15,88,243 // addps %xmm11,%xmm14
+ .byte 184,174,71,97,61 // mov $0x3d6147ae,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 65,15,194,196,1 // cmpltps %xmm12,%xmm0
+ .byte 68,15,84,232 // andps %xmm0,%xmm13
+ .byte 65,15,85,198 // andnps %xmm14,%xmm0
+ .byte 65,15,86,197 // orps %xmm13,%xmm0
+ .byte 69,15,40,232 // movaps %xmm8,%xmm13
+ .byte 68,15,89,233 // mulps %xmm1,%xmm13
+ .byte 68,15,40,241 // movaps %xmm1,%xmm14
+ .byte 69,15,89,246 // mulps %xmm14,%xmm14
+ .byte 69,15,40,249 // movaps %xmm9,%xmm15
+ .byte 68,15,89,249 // mulps %xmm1,%xmm15
+ .byte 69,15,88,250 // addps %xmm10,%xmm15
+ .byte 69,15,89,254 // mulps %xmm14,%xmm15
+ .byte 69,15,88,251 // addps %xmm11,%xmm15
+ .byte 65,15,194,204,1 // cmpltps %xmm12,%xmm1
+ .byte 68,15,84,233 // andps %xmm1,%xmm13
+ .byte 65,15,85,207 // andnps %xmm15,%xmm1
+ .byte 65,15,86,205 // orps %xmm13,%xmm1
+ .byte 68,15,89,194 // mulps %xmm2,%xmm8
+ .byte 68,15,40,234 // movaps %xmm2,%xmm13
+ .byte 69,15,89,237 // mulps %xmm13,%xmm13
+ .byte 68,15,89,202 // mulps %xmm2,%xmm9
+ .byte 69,15,88,202 // addps %xmm10,%xmm9
+ .byte 69,15,89,205 // mulps %xmm13,%xmm9
+ .byte 69,15,88,203 // addps %xmm11,%xmm9
+ .byte 65,15,194,212,1 // cmpltps %xmm12,%xmm2
+ .byte 68,15,84,194 // andps %xmm2,%xmm8
+ .byte 65,15,85,209 // andnps %xmm9,%xmm2
+ .byte 65,15,86,208 // orps %xmm8,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_to_srgb_sse2
+_sk_to_srgb_sse2:
+ .byte 68,15,82,192 // rsqrtps %xmm0,%xmm8
+ .byte 69,15,83,248 // rcpps %xmm8,%xmm15
+ .byte 69,15,82,232 // rsqrtps %xmm8,%xmm13
+ .byte 184,41,92,71,65 // mov $0x41475c29,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,240 // movaps %xmm8,%xmm14
+ .byte 68,15,89,240 // mulps %xmm0,%xmm14
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 184,194,135,210,62 // mov $0x3ed287c2,%eax
+ .byte 102,68,15,110,208 // movd %eax,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 184,206,111,48,63 // mov $0x3f306fce,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 184,168,87,202,61 // mov $0x3dca57a8,%eax
+ .byte 53,0,0,0,128 // xor $0x80000000,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 69,15,89,251 // mulps %xmm11,%xmm15
+ .byte 69,15,88,252 // addps %xmm12,%xmm15
+ .byte 69,15,89,234 // mulps %xmm10,%xmm13
+ .byte 69,15,88,239 // addps %xmm15,%xmm13
+ .byte 69,15,40,249 // movaps %xmm9,%xmm15
+ .byte 69,15,93,253 // minps %xmm13,%xmm15
+ .byte 184,4,231,140,59 // mov $0x3b8ce704,%eax
+ .byte 102,68,15,110,232 // movd %eax,%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 65,15,194,197,1 // cmpltps %xmm13,%xmm0
+ .byte 68,15,84,240 // andps %xmm0,%xmm14
+ .byte 65,15,85,199 // andnps %xmm15,%xmm0
+ .byte 65,15,86,198 // orps %xmm14,%xmm0
+ .byte 68,15,82,241 // rsqrtps %xmm1,%xmm14
+ .byte 69,15,83,254 // rcpps %xmm14,%xmm15
+ .byte 69,15,82,246 // rsqrtps %xmm14,%xmm14
+ .byte 69,15,89,251 // mulps %xmm11,%xmm15
+ .byte 69,15,88,252 // addps %xmm12,%xmm15
+ .byte 69,15,89,242 // mulps %xmm10,%xmm14
+ .byte 69,15,88,247 // addps %xmm15,%xmm14
+ .byte 69,15,40,249 // movaps %xmm9,%xmm15
+ .byte 69,15,93,254 // minps %xmm14,%xmm15
+ .byte 69,15,40,240 // movaps %xmm8,%xmm14
+ .byte 68,15,89,241 // mulps %xmm1,%xmm14
+ .byte 65,15,194,205,1 // cmpltps %xmm13,%xmm1
+ .byte 68,15,84,241 // andps %xmm1,%xmm14
+ .byte 65,15,85,207 // andnps %xmm15,%xmm1
+ .byte 65,15,86,206 // orps %xmm14,%xmm1
+ .byte 68,15,82,242 // rsqrtps %xmm2,%xmm14
+ .byte 69,15,83,254 // rcpps %xmm14,%xmm15
+ .byte 69,15,89,251 // mulps %xmm11,%xmm15
+ .byte 69,15,88,252 // addps %xmm12,%xmm15
+ .byte 69,15,82,222 // rsqrtps %xmm14,%xmm11
+ .byte 69,15,89,218 // mulps %xmm10,%xmm11
+ .byte 69,15,88,223 // addps %xmm15,%xmm11
+ .byte 69,15,93,203 // minps %xmm11,%xmm9
+ .byte 68,15,89,194 // mulps %xmm2,%xmm8
+ .byte 65,15,194,213,1 // cmpltps %xmm13,%xmm2
+ .byte 68,15,84,194 // andps %xmm2,%xmm8
+ .byte 65,15,85,209 // andnps %xmm9,%xmm2
+ .byte 65,15,86,208 // orps %xmm8,%xmm2
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_1_float_sse2
+_sk_scale_1_float_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_scale_u8_sse2
+_sk_scale_u8_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,68,15,110,4,56 // movd (%rax,%rdi,1),%xmm8
+ .byte 102,69,15,239,201 // pxor %xmm9,%xmm9
+ .byte 102,69,15,96,193 // punpcklbw %xmm9,%xmm8
+ .byte 102,69,15,97,193 // punpcklwd %xmm9,%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 65,15,89,193 // mulps %xmm9,%xmm0
+ .byte 65,15,89,201 // mulps %xmm9,%xmm1
+ .byte 65,15,89,209 // mulps %xmm9,%xmm2
+ .byte 65,15,89,217 // mulps %xmm9,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_1_float_sse2
+_sk_lerp_1_float_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,92,223 // subps %xmm7,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_u8_sse2
+_sk_lerp_u8_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,68,15,110,4,56 // movd (%rax,%rdi,1),%xmm8
+ .byte 102,69,15,239,201 // pxor %xmm9,%xmm9
+ .byte 102,69,15,96,193 // punpcklbw %xmm9,%xmm8
+ .byte 102,69,15,97,193 // punpcklwd %xmm9,%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,200 // movd %eax,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,193 // mulps %xmm9,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,201 // mulps %xmm9,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 65,15,89,209 // mulps %xmm9,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 15,92,223 // subps %xmm7,%xmm3
+ .byte 65,15,89,217 // mulps %xmm9,%xmm3
+ .byte 15,88,223 // addps %xmm7,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_lerp_565_sse2
+_sk_lerp_565_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,68,15,126,4,120 // movq (%rax,%rdi,2),%xmm8
+ .byte 102,15,239,219 // pxor %xmm3,%xmm3
+ .byte 102,68,15,97,195 // punpcklwd %xmm3,%xmm8
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,203 // cvtdq2ps %xmm3,%xmm9
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 102,68,15,110,208 // movd %eax,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,203 // cvtdq2ps %xmm3,%xmm9
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,65,15,219,216 // pand %xmm8,%xmm3
+ .byte 68,15,91,195 // cvtdq2ps %xmm3,%xmm8
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 15,92,196 // subps %xmm4,%xmm0
+ .byte 65,15,89,194 // mulps %xmm10,%xmm0
+ .byte 15,88,196 // addps %xmm4,%xmm0
+ .byte 15,92,205 // subps %xmm5,%xmm1
+ .byte 65,15,89,203 // mulps %xmm11,%xmm1
+ .byte 15,88,205 // addps %xmm5,%xmm1
+ .byte 15,92,214 // subps %xmm6,%xmm2
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 15,88,214 // addps %xmm6,%xmm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_tables_sse2
+_sk_load_tables_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,8 // mov (%rax),%rcx
+ .byte 76,139,64,8 // mov 0x8(%rax),%r8
+ .byte 243,68,15,111,4,185 // movdqu (%rcx,%rdi,4),%xmm8
+ .byte 185,255,0,0,0 // mov $0xff,%ecx
+ .byte 102,15,110,193 // movd %ecx,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
+ .byte 102,65,15,114,209,8 // psrld $0x8,%xmm9
+ .byte 102,68,15,219,200 // pand %xmm0,%xmm9
+ .byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
+ .byte 102,65,15,114,210,16 // psrld $0x10,%xmm10
+ .byte 102,68,15,219,208 // pand %xmm0,%xmm10
+ .byte 102,65,15,219,192 // pand %xmm8,%xmm0
+ .byte 102,15,112,216,78 // pshufd $0x4e,%xmm0,%xmm3
+ .byte 102,72,15,126,217 // movq %xmm3,%rcx
+ .byte 65,137,201 // mov %ecx,%r9d
+ .byte 72,193,233,32 // shr $0x20,%rcx
+ .byte 102,73,15,126,194 // movq %xmm0,%r10
+ .byte 69,137,211 // mov %r10d,%r11d
+ .byte 73,193,234,32 // shr $0x20,%r10
+ .byte 243,67,15,16,28,144 // movss (%r8,%r10,4),%xmm3
+ .byte 243,65,15,16,4,136 // movss (%r8,%rcx,4),%xmm0
+ .byte 15,20,216 // unpcklps %xmm0,%xmm3
+ .byte 243,67,15,16,4,152 // movss (%r8,%r11,4),%xmm0
+ .byte 243,67,15,16,12,136 // movss (%r8,%r9,4),%xmm1
+ .byte 15,20,193 // unpcklps %xmm1,%xmm0
+ .byte 15,20,195 // unpcklps %xmm3,%xmm0
+ .byte 76,139,64,16 // mov 0x10(%rax),%r8
+ .byte 102,65,15,112,201,78 // pshufd $0x4e,%xmm9,%xmm1
+ .byte 102,73,15,126,202 // movq %xmm1,%r10
+ .byte 77,137,209 // mov %r10,%r9
+ .byte 73,193,233,32 // shr $0x20,%r9
+ .byte 102,76,15,126,201 // movq %xmm9,%rcx
+ .byte 65,137,203 // mov %ecx,%r11d
+ .byte 65,129,227,255,255,255,0 // and $0xffffff,%r11d
+ .byte 72,193,233,30 // shr $0x1e,%rcx
+ .byte 65,129,226,255,255,255,0 // and $0xffffff,%r10d
+ .byte 243,65,15,16,28,8 // movss (%r8,%rcx,1),%xmm3
+ .byte 243,67,15,16,12,136 // movss (%r8,%r9,4),%xmm1
+ .byte 15,20,217 // unpcklps %xmm1,%xmm3
+ .byte 243,67,15,16,12,152 // movss (%r8,%r11,4),%xmm1
+ .byte 243,67,15,16,20,144 // movss (%r8,%r10,4),%xmm2
+ .byte 15,20,202 // unpcklps %xmm2,%xmm1
+ .byte 15,20,203 // unpcklps %xmm3,%xmm1
+ .byte 76,139,72,24 // mov 0x18(%rax),%r9
+ .byte 102,65,15,112,210,78 // pshufd $0x4e,%xmm10,%xmm2
+ .byte 102,72,15,126,209 // movq %xmm2,%rcx
+ .byte 68,15,183,193 // movzwl %cx,%r8d
+ .byte 72,193,233,32 // shr $0x20,%rcx
+ .byte 102,76,15,126,208 // movq %xmm10,%rax
+ .byte 68,15,183,208 // movzwl %ax,%r10d
+ .byte 72,193,232,30 // shr $0x1e,%rax
+ .byte 243,69,15,16,12,1 // movss (%r9,%rax,1),%xmm9
+ .byte 243,65,15,16,20,137 // movss (%r9,%rcx,4),%xmm2
+ .byte 68,15,20,202 // unpcklps %xmm2,%xmm9
+ .byte 243,67,15,16,20,145 // movss (%r9,%r10,4),%xmm2
+ .byte 243,67,15,16,28,129 // movss (%r9,%r8,4),%xmm3
+ .byte 15,20,211 // unpcklps %xmm3,%xmm2
+ .byte 65,15,20,209 // unpcklps %xmm9,%xmm2
+ .byte 102,65,15,114,208,24 // psrld $0x18,%xmm8
+ .byte 69,15,91,192 // cvtdq2ps %xmm8,%xmm8
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_a8_sse2
+_sk_load_a8_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 102,15,110,4,56 // movd (%rax,%rdi,1),%xmm0
+ .byte 102,15,239,201 // pxor %xmm1,%xmm1
+ .byte 102,15,96,193 // punpcklbw %xmm1,%xmm0
+ .byte 102,15,97,193 // punpcklwd %xmm1,%xmm0
+ .byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 102,15,239,201 // pxor %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_a8_sse2
+_sk_store_a8_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,127,67 // mov $0x437f0000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,65,15,114,240,16 // pslld $0x10,%xmm8
+ .byte 102,65,15,114,224,16 // psrad $0x10,%xmm8
+ .byte 102,69,15,107,192 // packssdw %xmm8,%xmm8
+ .byte 102,69,15,103,192 // packuswb %xmm8,%xmm8
+ .byte 102,68,15,126,4,56 // movd %xmm8,(%rax,%rdi,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_565_sse2
+_sk_load_565_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,15,126,20,120 // movq (%rax,%rdi,2),%xmm2
+ .byte 102,15,239,192 // pxor %xmm0,%xmm0
+ .byte 102,15,97,208 // punpcklwd %xmm0,%xmm2
+ .byte 184,0,248,0,0 // mov $0xf800,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,15,219,194 // pand %xmm2,%xmm0
+ .byte 15,91,200 // cvtdq2ps %xmm0,%xmm1
+ .byte 184,8,33,132,55 // mov $0x37842108,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 184,224,7,0,0 // mov $0x7e0,%eax
+ .byte 102,15,110,200 // movd %eax,%xmm1
+ .byte 102,15,112,201,0 // pshufd $0x0,%xmm1,%xmm1
+ .byte 102,15,219,202 // pand %xmm2,%xmm1
+ .byte 15,91,217 // cvtdq2ps %xmm1,%xmm3
+ .byte 184,33,8,2,58 // mov $0x3a020821,%eax
+ .byte 102,15,110,200 // movd %eax,%xmm1
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 15,89,203 // mulps %xmm3,%xmm1
+ .byte 184,31,0,0,0 // mov $0x1f,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 102,15,112,219,0 // pshufd $0x0,%xmm3,%xmm3
+ .byte 102,15,219,218 // pand %xmm2,%xmm3
+ .byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
+ .byte 184,8,33,4,61 // mov $0x3d042108,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 15,198,210,0 // shufps $0x0,%xmm2,%xmm2
+ .byte 15,89,211 // mulps %xmm3,%xmm2
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_565_sse2
+_sk_store_565_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,248,65 // mov $0x41f80000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 102,65,15,114,241,11 // pslld $0xb,%xmm9
+ .byte 185,0,0,124,66 // mov $0x427c0000,%ecx
+ .byte 102,68,15,110,209 // movd %ecx,%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
+ .byte 102,65,15,114,242,5 // pslld $0x5,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 68,15,89,194 // mulps %xmm2,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,69,15,86,194 // orpd %xmm10,%xmm8
+ .byte 102,65,15,114,240,16 // pslld $0x10,%xmm8
+ .byte 102,65,15,114,224,16 // psrad $0x10,%xmm8
+ .byte 102,69,15,107,192 // packssdw %xmm8,%xmm8
+ .byte 102,68,15,214,4,120 // movq %xmm8,(%rax,%rdi,2)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_8888_sse2
+_sk_load_8888_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,15,111,28,184 // movdqu (%rax,%rdi,4),%xmm3
+ .byte 184,255,0,0,0 // mov $0xff,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,192,0 // pshufd $0x0,%xmm0,%xmm0
+ .byte 102,15,111,203 // movdqa %xmm3,%xmm1
+ .byte 102,15,114,209,8 // psrld $0x8,%xmm1
+ .byte 102,15,219,200 // pand %xmm0,%xmm1
+ .byte 102,15,111,211 // movdqa %xmm3,%xmm2
+ .byte 102,15,114,210,16 // psrld $0x10,%xmm2
+ .byte 102,15,219,208 // pand %xmm0,%xmm2
+ .byte 102,15,219,195 // pand %xmm3,%xmm0
+ .byte 15,91,192 // cvtdq2ps %xmm0,%xmm0
+ .byte 184,129,128,128,59 // mov $0x3b808081,%eax
+ .byte 102,68,15,110,192 // movd %eax,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 15,91,201 // cvtdq2ps %xmm1,%xmm1
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 15,91,210 // cvtdq2ps %xmm2,%xmm2
+ .byte 65,15,89,208 // mulps %xmm8,%xmm2
+ .byte 102,15,114,211,24 // psrld $0x18,%xmm3
+ .byte 15,91,219 // cvtdq2ps %xmm3,%xmm3
+ .byte 65,15,89,216 // mulps %xmm8,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_8888_sse2
+_sk_store_8888_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,127,67 // mov $0x437f0000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 69,15,40,208 // movaps %xmm8,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,69,15,91,210 // cvtps2dq %xmm10,%xmm10
+ .byte 102,65,15,114,242,8 // pslld $0x8,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 68,15,89,202 // mulps %xmm2,%xmm9
+ .byte 102,69,15,91,201 // cvtps2dq %xmm9,%xmm9
+ .byte 102,65,15,114,241,16 // pslld $0x10,%xmm9
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,69,15,91,192 // cvtps2dq %xmm8,%xmm8
+ .byte 102,65,15,114,240,24 // pslld $0x18,%xmm8
+ .byte 102,69,15,235,193 // por %xmm9,%xmm8
+ .byte 102,69,15,235,194 // por %xmm10,%xmm8
+ .byte 243,68,15,127,4,184 // movdqu %xmm8,(%rax,%rdi,4)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_load_f16_sse2
+_sk_load_f16_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 243,15,111,4,248 // movdqu (%rax,%rdi,8),%xmm0
+ .byte 243,15,111,76,248,16 // movdqu 0x10(%rax,%rdi,8),%xmm1
+ .byte 102,15,111,208 // movdqa %xmm0,%xmm2
+ .byte 102,15,97,209 // punpcklwd %xmm1,%xmm2
+ .byte 102,15,105,193 // punpckhwd %xmm1,%xmm0
+ .byte 102,68,15,111,194 // movdqa %xmm2,%xmm8
+ .byte 102,68,15,97,192 // punpcklwd %xmm0,%xmm8
+ .byte 102,15,105,208 // punpckhwd %xmm0,%xmm2
+ .byte 184,0,4,0,4 // mov $0x4000400,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 102,15,112,216,0 // pshufd $0x0,%xmm0,%xmm3
+ .byte 102,15,111,203 // movdqa %xmm3,%xmm1
+ .byte 102,65,15,101,200 // pcmpgtw %xmm8,%xmm1
+ .byte 102,65,15,223,200 // pandn %xmm8,%xmm1
+ .byte 102,15,101,218 // pcmpgtw %xmm2,%xmm3
+ .byte 102,15,223,218 // pandn %xmm2,%xmm3
+ .byte 102,69,15,239,192 // pxor %xmm8,%xmm8
+ .byte 102,15,111,193 // movdqa %xmm1,%xmm0
+ .byte 102,65,15,97,192 // punpcklwd %xmm8,%xmm0
+ .byte 102,15,114,240,13 // pslld $0xd,%xmm0
+ .byte 184,0,0,128,119 // mov $0x77800000,%eax
+ .byte 102,15,110,208 // movd %eax,%xmm2
+ .byte 102,68,15,112,202,0 // pshufd $0x0,%xmm2,%xmm9
+ .byte 65,15,89,193 // mulps %xmm9,%xmm0
+ .byte 102,65,15,105,200 // punpckhwd %xmm8,%xmm1
+ .byte 102,15,114,241,13 // pslld $0xd,%xmm1
+ .byte 65,15,89,201 // mulps %xmm9,%xmm1
+ .byte 102,15,111,211 // movdqa %xmm3,%xmm2
+ .byte 102,65,15,97,208 // punpcklwd %xmm8,%xmm2
+ .byte 102,15,114,242,13 // pslld $0xd,%xmm2
+ .byte 65,15,89,209 // mulps %xmm9,%xmm2
+ .byte 102,65,15,105,216 // punpckhwd %xmm8,%xmm3
+ .byte 102,15,114,243,13 // pslld $0xd,%xmm3
+ .byte 65,15,89,217 // mulps %xmm9,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_f16_sse2
+_sk_store_f16_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 185,0,0,128,7 // mov $0x7800000,%ecx
+ .byte 102,68,15,110,193 // movd %ecx,%xmm8
+ .byte 102,69,15,112,192,0 // pshufd $0x0,%xmm8,%xmm8
+ .byte 102,69,15,111,200 // movdqa %xmm8,%xmm9
+ .byte 68,15,89,200 // mulps %xmm0,%xmm9
+ .byte 102,65,15,114,209,13 // psrld $0xd,%xmm9
+ .byte 102,69,15,111,208 // movdqa %xmm8,%xmm10
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 102,65,15,114,210,13 // psrld $0xd,%xmm10
+ .byte 102,69,15,111,216 // movdqa %xmm8,%xmm11
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 102,65,15,114,211,13 // psrld $0xd,%xmm11
+ .byte 68,15,89,195 // mulps %xmm3,%xmm8
+ .byte 102,65,15,114,208,13 // psrld $0xd,%xmm8
+ .byte 102,65,15,115,250,2 // pslldq $0x2,%xmm10
+ .byte 102,69,15,235,209 // por %xmm9,%xmm10
+ .byte 102,65,15,115,248,2 // pslldq $0x2,%xmm8
+ .byte 102,69,15,235,195 // por %xmm11,%xmm8
+ .byte 102,69,15,111,202 // movdqa %xmm10,%xmm9
+ .byte 102,69,15,98,200 // punpckldq %xmm8,%xmm9
+ .byte 243,68,15,127,12,248 // movdqu %xmm9,(%rax,%rdi,8)
+ .byte 102,69,15,106,208 // punpckhdq %xmm8,%xmm10
+ .byte 243,68,15,127,84,248,16 // movdqu %xmm10,0x10(%rax,%rdi,8)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_store_f32_sse2
+_sk_store_f32_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 72,139,0 // mov (%rax),%rax
+ .byte 72,137,249 // mov %rdi,%rcx
+ .byte 72,193,225,4 // shl $0x4,%rcx
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 68,15,40,200 // movaps %xmm0,%xmm9
+ .byte 68,15,20,201 // unpcklps %xmm1,%xmm9
+ .byte 68,15,40,210 // movaps %xmm2,%xmm10
+ .byte 68,15,40,218 // movaps %xmm2,%xmm11
+ .byte 68,15,20,219 // unpcklps %xmm3,%xmm11
+ .byte 68,15,21,193 // unpckhps %xmm1,%xmm8
+ .byte 68,15,21,211 // unpckhps %xmm3,%xmm10
+ .byte 69,15,40,225 // movaps %xmm9,%xmm12
+ .byte 102,69,15,20,227 // unpcklpd %xmm11,%xmm12
+ .byte 69,15,18,217 // movhlps %xmm9,%xmm11
+ .byte 69,15,40,200 // movaps %xmm8,%xmm9
+ .byte 102,69,15,20,202 // unpcklpd %xmm10,%xmm9
+ .byte 69,15,18,208 // movhlps %xmm8,%xmm10
+ .byte 102,68,15,17,36,8 // movupd %xmm12,(%rax,%rcx,1)
+ .byte 68,15,17,92,8,16 // movups %xmm11,0x10(%rax,%rcx,1)
+ .byte 102,68,15,17,76,8,32 // movupd %xmm9,0x20(%rax,%rcx,1)
+ .byte 68,15,17,84,8,48 // movups %xmm10,0x30(%rax,%rcx,1)
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_x_sse2
+_sk_clamp_x_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,95,192 // maxps %xmm0,%xmm8
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 102,15,118,192 // pcmpeqd %xmm0,%xmm0
+ .byte 102,65,15,254,193 // paddd %xmm9,%xmm0
+ .byte 68,15,93,192 // minps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,192 // movaps %xmm8,%xmm0
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_clamp_y_sse2
+_sk_clamp_y_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 69,15,87,192 // xorps %xmm8,%xmm8
+ .byte 68,15,95,193 // maxps %xmm1,%xmm8
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 102,15,118,201 // pcmpeqd %xmm1,%xmm1
+ .byte 102,65,15,254,201 // paddd %xmm9,%xmm1
+ .byte 68,15,93,193 // minps %xmm1,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,200 // movaps %xmm8,%xmm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_x_sse2
+_sk_repeat_x_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,200 // movaps %xmm0,%xmm9
+ .byte 69,15,94,200 // divps %xmm8,%xmm9
+ .byte 243,69,15,91,209 // cvttps2dq %xmm9,%xmm10
+ .byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
+ .byte 69,15,194,202,1 // cmpltps %xmm10,%xmm9
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,84,217 // andps %xmm9,%xmm11
+ .byte 69,15,92,211 // subps %xmm11,%xmm10
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 65,15,92,194 // subps %xmm10,%xmm0
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,193 // minps %xmm9,%xmm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_repeat_y_sse2
+_sk_repeat_y_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,0 // movss (%rax),%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 69,15,94,200 // divps %xmm8,%xmm9
+ .byte 243,69,15,91,209 // cvttps2dq %xmm9,%xmm10
+ .byte 69,15,91,210 // cvtdq2ps %xmm10,%xmm10
+ .byte 69,15,194,202,1 // cmpltps %xmm10,%xmm9
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,216 // movd %eax,%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,84,217 // andps %xmm9,%xmm11
+ .byte 69,15,92,211 // subps %xmm11,%xmm10
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 65,15,92,202 // subps %xmm10,%xmm1
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,201 // minps %xmm9,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_x_sse2
+_sk_mirror_x_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,40,193 // movaps %xmm9,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,92,192 // subps %xmm8,%xmm0
+ .byte 243,69,15,88,201 // addss %xmm9,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 68,15,40,208 // movaps %xmm0,%xmm10
+ .byte 69,15,94,209 // divps %xmm9,%xmm10
+ .byte 243,69,15,91,218 // cvttps2dq %xmm10,%xmm11
+ .byte 69,15,91,219 // cvtdq2ps %xmm11,%xmm11
+ .byte 69,15,194,211,1 // cmpltps %xmm11,%xmm10
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 69,15,84,226 // andps %xmm10,%xmm12
+ .byte 69,15,87,210 // xorps %xmm10,%xmm10
+ .byte 69,15,92,220 // subps %xmm12,%xmm11
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 65,15,92,195 // subps %xmm11,%xmm0
+ .byte 65,15,92,192 // subps %xmm8,%xmm0
+ .byte 68,15,92,208 // subps %xmm0,%xmm10
+ .byte 65,15,84,194 // andps %xmm10,%xmm0
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,193 // minps %xmm9,%xmm0
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_mirror_y_sse2
+_sk_mirror_y_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,68,15,16,8 // movss (%rax),%xmm9
+ .byte 69,15,40,193 // movaps %xmm9,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,92,200 // subps %xmm8,%xmm1
+ .byte 243,69,15,88,201 // addss %xmm9,%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 68,15,40,209 // movaps %xmm1,%xmm10
+ .byte 69,15,94,209 // divps %xmm9,%xmm10
+ .byte 243,69,15,91,218 // cvttps2dq %xmm10,%xmm11
+ .byte 69,15,91,219 // cvtdq2ps %xmm11,%xmm11
+ .byte 69,15,194,211,1 // cmpltps %xmm11,%xmm10
+ .byte 184,0,0,128,63 // mov $0x3f800000,%eax
+ .byte 102,68,15,110,224 // movd %eax,%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 69,15,84,226 // andps %xmm10,%xmm12
+ .byte 69,15,87,210 // xorps %xmm10,%xmm10
+ .byte 69,15,92,220 // subps %xmm12,%xmm11
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 65,15,92,203 // subps %xmm11,%xmm1
+ .byte 65,15,92,200 // subps %xmm8,%xmm1
+ .byte 68,15,92,209 // subps %xmm1,%xmm10
+ .byte 65,15,84,202 // andps %xmm10,%xmm1
+ .byte 102,69,15,118,201 // pcmpeqd %xmm9,%xmm9
+ .byte 102,69,15,254,200 // paddd %xmm8,%xmm9
+ .byte 65,15,93,201 // minps %xmm9,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_luminance_to_alpha_sse2
+_sk_luminance_to_alpha_sse2:
+ .byte 184,208,179,89,62 // mov $0x3e59b3d0,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 184,89,23,55,63 // mov $0x3f371759,%eax
+ .byte 102,15,110,192 // movd %eax,%xmm0
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 15,88,195 // addps %xmm3,%xmm0
+ .byte 184,152,221,147,61 // mov $0x3d93dd98,%eax
+ .byte 102,15,110,216 // movd %eax,%xmm3
+ .byte 15,198,219,0 // shufps $0x0,%xmm3,%xmm3
+ .byte 15,89,218 // mulps %xmm2,%xmm3
+ .byte 15,88,216 // addps %xmm0,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 15,87,192 // xorps %xmm0,%xmm0
+ .byte 15,87,201 // xorps %xmm1,%xmm1
+ .byte 15,87,210 // xorps %xmm2,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_2x3_sse2
+_sk_matrix_2x3_sse2:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,16 // movss 0x10(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,12 // movss 0xc(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_3x4_sse2
+_sk_matrix_3x4_sse2:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,12 // movss 0xc(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,24 // movss 0x18(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,36 // movss 0x24(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,28 // movss 0x1c(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,40 // movss 0x28(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,32 // movss 0x20(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,44 // movss 0x2c(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,226 // mulps %xmm2,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,210 // movaps %xmm10,%xmm2
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_4x5_sse2
+_sk_matrix_4x5_sse2:
+ .byte 68,15,40,201 // movaps %xmm1,%xmm9
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,15,16,72,4 // movss 0x4(%rax),%xmm1
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,32 // movss 0x20(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,48 // movss 0x30(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,64 // movss 0x40(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,227 // mulps %xmm3,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,194 // addps %xmm10,%xmm0
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 243,68,15,16,80,20 // movss 0x14(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,36 // movss 0x24(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,52 // movss 0x34(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,68 // movss 0x44(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 68,15,89,227 // mulps %xmm3,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 68,15,89,218 // mulps %xmm2,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,209 // mulps %xmm9,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,89,200 // mulps %xmm8,%xmm1
+ .byte 65,15,88,202 // addps %xmm10,%xmm1
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,24 // movss 0x18(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,40 // movss 0x28(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,56 // movss 0x38(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 243,68,15,16,112,72 // movss 0x48(%rax),%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 68,15,89,235 // mulps %xmm3,%xmm13
+ .byte 69,15,88,238 // addps %xmm14,%xmm13
+ .byte 68,15,89,226 // mulps %xmm2,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,217 // mulps %xmm9,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 243,68,15,16,88,12 // movss 0xc(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,28 // movss 0x1c(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 243,68,15,16,104,44 // movss 0x2c(%rax),%xmm13
+ .byte 69,15,198,237,0 // shufps $0x0,%xmm13,%xmm13
+ .byte 243,68,15,16,112,60 // movss 0x3c(%rax),%xmm14
+ .byte 69,15,198,246,0 // shufps $0x0,%xmm14,%xmm14
+ .byte 243,68,15,16,120,76 // movss 0x4c(%rax),%xmm15
+ .byte 69,15,198,255,0 // shufps $0x0,%xmm15,%xmm15
+ .byte 68,15,89,243 // mulps %xmm3,%xmm14
+ .byte 69,15,88,247 // addps %xmm15,%xmm14
+ .byte 68,15,89,234 // mulps %xmm2,%xmm13
+ .byte 69,15,88,238 // addps %xmm14,%xmm13
+ .byte 69,15,89,225 // mulps %xmm9,%xmm12
+ .byte 69,15,88,229 // addps %xmm13,%xmm12
+ .byte 69,15,89,216 // mulps %xmm8,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,210 // movaps %xmm10,%xmm2
+ .byte 65,15,40,219 // movaps %xmm11,%xmm3
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_matrix_perspective_sse2
+_sk_matrix_perspective_sse2:
+ .byte 68,15,40,192 // movaps %xmm0,%xmm8
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 243,15,16,0 // movss (%rax),%xmm0
+ .byte 243,68,15,16,72,4 // movss 0x4(%rax),%xmm9
+ .byte 15,198,192,0 // shufps $0x0,%xmm0,%xmm0
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 243,68,15,16,80,8 // movss 0x8(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 68,15,89,201 // mulps %xmm1,%xmm9
+ .byte 69,15,88,202 // addps %xmm10,%xmm9
+ .byte 65,15,89,192 // mulps %xmm8,%xmm0
+ .byte 65,15,88,193 // addps %xmm9,%xmm0
+ .byte 243,68,15,16,72,12 // movss 0xc(%rax),%xmm9
+ .byte 69,15,198,201,0 // shufps $0x0,%xmm9,%xmm9
+ .byte 243,68,15,16,80,16 // movss 0x10(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,20 // movss 0x14(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 68,15,89,209 // mulps %xmm1,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 69,15,89,200 // mulps %xmm8,%xmm9
+ .byte 69,15,88,202 // addps %xmm10,%xmm9
+ .byte 243,68,15,16,80,24 // movss 0x18(%rax),%xmm10
+ .byte 69,15,198,210,0 // shufps $0x0,%xmm10,%xmm10
+ .byte 243,68,15,16,88,28 // movss 0x1c(%rax),%xmm11
+ .byte 69,15,198,219,0 // shufps $0x0,%xmm11,%xmm11
+ .byte 243,68,15,16,96,32 // movss 0x20(%rax),%xmm12
+ .byte 69,15,198,228,0 // shufps $0x0,%xmm12,%xmm12
+ .byte 68,15,89,217 // mulps %xmm1,%xmm11
+ .byte 69,15,88,220 // addps %xmm12,%xmm11
+ .byte 69,15,89,208 // mulps %xmm8,%xmm10
+ .byte 69,15,88,211 // addps %xmm11,%xmm10
+ .byte 65,15,83,202 // rcpps %xmm10,%xmm1
+ .byte 15,89,193 // mulps %xmm1,%xmm0
+ .byte 68,15,89,201 // mulps %xmm1,%xmm9
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,201 // movaps %xmm9,%xmm1
+ .byte 255,224 // jmpq *%rax
+
+.globl _sk_linear_gradient_2stops_sse2
+_sk_linear_gradient_2stops_sse2:
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 68,15,16,8 // movups (%rax),%xmm9
+ .byte 15,16,88,16 // movups 0x10(%rax),%xmm3
+ .byte 68,15,40,195 // movaps %xmm3,%xmm8
+ .byte 69,15,198,192,0 // shufps $0x0,%xmm8,%xmm8
+ .byte 65,15,40,201 // movaps %xmm9,%xmm1
+ .byte 15,198,201,0 // shufps $0x0,%xmm1,%xmm1
+ .byte 68,15,89,192 // mulps %xmm0,%xmm8
+ .byte 68,15,88,193 // addps %xmm1,%xmm8
+ .byte 15,40,203 // movaps %xmm3,%xmm1
+ .byte 15,198,201,85 // shufps $0x55,%xmm1,%xmm1
+ .byte 65,15,40,209 // movaps %xmm9,%xmm2
+ .byte 15,198,210,85 // shufps $0x55,%xmm2,%xmm2
+ .byte 15,89,200 // mulps %xmm0,%xmm1
+ .byte 15,88,202 // addps %xmm2,%xmm1
+ .byte 15,40,211 // movaps %xmm3,%xmm2
+ .byte 15,198,210,170 // shufps $0xaa,%xmm2,%xmm2
+ .byte 69,15,40,209 // movaps %xmm9,%xmm10
+ .byte 69,15,198,210,170 // shufps $0xaa,%xmm10,%xmm10
+ .byte 15,89,208 // mulps %xmm0,%xmm2
+ .byte 65,15,88,210 // addps %xmm10,%xmm2
+ .byte 15,198,219,255 // shufps $0xff,%xmm3,%xmm3
+ .byte 69,15,198,201,255 // shufps $0xff,%xmm9,%xmm9
+ .byte 15,89,216 // mulps %xmm0,%xmm3
+ .byte 65,15,88,217 // addps %xmm9,%xmm3
+ .byte 72,173 // lods %ds:(%rsi),%rax
+ .byte 65,15,40,192 // movaps %xmm8,%xmm0
+ .byte 255,224 // jmpq *%rax
+#endif
diff --git a/src/jumper/SkJumper_generated.cpp b/src/jumper/SkJumper_generated.cpp
deleted file mode 100644
index 2494c1eb05..0000000000
--- a/src/jumper/SkJumper_generated.cpp
+++ /dev/null
@@ -1,13052 +0,0 @@
-/*
- * Copyright 2017 Google Inc.
- *
- * Use of this source code is governed by a BSD-style license that can be
- * found in the LICENSE file.
- */
-
-// This file is generated semi-automatically with this command:
-// $ src/jumper/build_stages.py
-
-#include <stdint.h>
-
-#if defined(_MSC_VER)
- #pragma section("code", read,execute)
- #define CODE extern "C" __declspec(allocate("code"))
-#elif defined(__MACH__)
- #define CODE extern "C" __attribute__((section("__TEXT,__text")))
-#else
- #define CODE extern "C" __attribute__((section(".text")))
-#endif
-
-#if defined(__aarch64__)
-
-CODE const uint32_t sk_start_pipeline_aarch64[] = {
- 0xa9bd5bf7, //stp x23, x22, [sp, #-48]!
- 0xa90153f5, //stp x21, x20, [sp, #16]
- 0xa9027bf3, //stp x19, x30, [sp, #32]
- 0xaa0103f4, //mov x20, x1
- 0xf8408697, //ldr x23, [x20], #8
- 0xaa0003f5, //mov x21, x0
- 0xaa0303f3, //mov x19, x3
- 0x910012a8, //add x8, x21, #0x4
- 0xeb13011f, //cmp x8, x19
- 0xaa0203f6, //mov x22, x2
- 0x54000069, //b.ls 34 <sk_start_pipeline_aarch64+0x34> // b.plast
- 0xaa1503e0, //mov x0, x21
- 0x14000012, //b 78 <sk_start_pipeline_aarch64+0x78>
- 0x6f00e400, //movi v0.2d, #0x0
- 0x6f00e401, //movi v1.2d, #0x0
- 0x6f00e402, //movi v2.2d, #0x0
- 0x6f00e403, //movi v3.2d, #0x0
- 0x6f00e404, //movi v4.2d, #0x0
- 0x6f00e405, //movi v5.2d, #0x0
- 0x6f00e406, //movi v6.2d, #0x0
- 0x6f00e407, //movi v7.2d, #0x0
- 0xaa1503e0, //mov x0, x21
- 0xaa1403e1, //mov x1, x20
- 0xaa1603e2, //mov x2, x22
- 0xd63f02e0, //blr x23
- 0x910012a0, //add x0, x21, #0x4
- 0x910022a8, //add x8, x21, #0x8
- 0xeb13011f, //cmp x8, x19
- 0xaa0003f5, //mov x21, x0
- 0x54fffe09, //b.ls 34 <sk_start_pipeline_aarch64+0x34> // b.plast
- 0xa9427bf3, //ldp x19, x30, [sp, #32]
- 0xa94153f5, //ldp x21, x20, [sp, #16]
- 0xa8c35bf7, //ldp x23, x22, [sp], #48
- 0xd65f03c0, //ret
-};
-
-CODE const uint32_t sk_just_return_aarch64[] = {
- 0xd65f03c0, //ret
-};
-
-CODE const uint32_t sk_seed_shader_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x3dc00046, //ldr q6, [x2]
- 0x4e040c00, //dup v0.4s, w0
- 0x4f0167e7, //movi v7.4s, #0x3f, lsl #24
- 0x4d40c901, //ld1r {v1.4s}, [x8]
- 0x4e21d800, //scvtf v0.4s, v0.4s
- 0x4e27d400, //fadd v0.4s, v0.4s, v7.4s
- 0x4f03f602, //fmov v2.4s, #1.000000000000000000e+00
- 0x4e21d821, //scvtf v1.4s, v1.4s
- 0x6f00e403, //movi v3.2d, #0x0
- 0x6f00e404, //movi v4.2d, #0x0
- 0x6f00e405, //movi v5.2d, #0x0
- 0x4e26d400, //fadd v0.4s, v0.4s, v6.4s
- 0x6f00e406, //movi v6.2d, #0x0
- 0x4e27d421, //fadd v1.4s, v1.4s, v7.4s
- 0x6f00e407, //movi v7.2d, #0x0
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_constant_color_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x3dc00103, //ldr q3, [x8]
- 0x4e040460, //dup v0.4s, v3.s[0]
- 0x4e0c0461, //dup v1.4s, v3.s[1]
- 0x4e140462, //dup v2.4s, v3.s[2]
- 0x4e1c0463, //dup v3.4s, v3.s[3]
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clear_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x6f00e400, //movi v0.2d, #0x0
- 0x6f00e401, //movi v1.2d, #0x0
- 0x6f00e402, //movi v2.2d, #0x0
- 0x6f00e403, //movi v3.2d, #0x0
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_plus__aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4e24d400, //fadd v0.4s, v0.4s, v4.4s
- 0x4e25d421, //fadd v1.4s, v1.4s, v5.4s
- 0x4e26d442, //fadd v2.4s, v2.4s, v6.4s
- 0x4e27d463, //fadd v3.4s, v3.4s, v7.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_srcover_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4f03f610, //fmov v16.4s, #1.000000000000000000e+00
- 0x4ea3d610, //fsub v16.4s, v16.4s, v3.4s
- 0x4e24ce00, //fmla v0.4s, v16.4s, v4.4s
- 0x4e25ce01, //fmla v1.4s, v16.4s, v5.4s
- 0x4e26ce02, //fmla v2.4s, v16.4s, v6.4s
- 0x4e27ce03, //fmla v3.4s, v16.4s, v7.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_dstover_aarch64[] = {
- 0x4f03f611, //fmov v17.4s, #1.000000000000000000e+00
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea41c90, //mov v16.16b, v4.16b
- 0x4ea7d634, //fsub v20.4s, v17.4s, v7.4s
- 0x4ea51cb1, //mov v17.16b, v5.16b
- 0x4ea61cd2, //mov v18.16b, v6.16b
- 0x4ea71cf3, //mov v19.16b, v7.16b
- 0x4e20ce90, //fmla v16.4s, v20.4s, v0.4s
- 0x4e21ce91, //fmla v17.4s, v20.4s, v1.4s
- 0x4e22ce92, //fmla v18.4s, v20.4s, v2.4s
- 0x4e23ce93, //fmla v19.4s, v20.4s, v3.4s
- 0x4eb01e00, //mov v0.16b, v16.16b
- 0x4eb11e21, //mov v1.16b, v17.16b
- 0x4eb21e42, //mov v2.16b, v18.16b
- 0x4eb31e63, //mov v3.16b, v19.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clamp_0_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x6f00e410, //movi v16.2d, #0x0
- 0x4e30f400, //fmax v0.4s, v0.4s, v16.4s
- 0x4e30f421, //fmax v1.4s, v1.4s, v16.4s
- 0x4e30f442, //fmax v2.4s, v2.4s, v16.4s
- 0x4e30f463, //fmax v3.4s, v3.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clamp_1_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4f03f610, //fmov v16.4s, #1.000000000000000000e+00
- 0x4eb0f400, //fmin v0.4s, v0.4s, v16.4s
- 0x4eb0f421, //fmin v1.4s, v1.4s, v16.4s
- 0x4eb0f442, //fmin v2.4s, v2.4s, v16.4s
- 0x4eb0f463, //fmin v3.4s, v3.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clamp_a_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4f03f610, //fmov v16.4s, #1.000000000000000000e+00
- 0x4eb0f463, //fmin v3.4s, v3.4s, v16.4s
- 0x4ea3f400, //fmin v0.4s, v0.4s, v3.4s
- 0x4ea3f421, //fmin v1.4s, v1.4s, v3.4s
- 0x4ea3f442, //fmin v2.4s, v2.4s, v3.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_set_rgb_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xaa0803e9, //mov x9, x8
- 0x4ddfc920, //ld1r {v0.4s}, [x9], #4
- 0x91002108, //add x8, x8, #0x8
- 0x4d40c902, //ld1r {v2.4s}, [x8]
- 0x4d40c921, //ld1r {v1.4s}, [x9]
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_swap_rb_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea01c10, //mov v16.16b, v0.16b
- 0x4ea21c40, //mov v0.16b, v2.16b
- 0x4eb01e02, //mov v2.16b, v16.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_swap_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea31c70, //mov v16.16b, v3.16b
- 0x4ea21c51, //mov v17.16b, v2.16b
- 0x4ea11c32, //mov v18.16b, v1.16b
- 0x4ea01c13, //mov v19.16b, v0.16b
- 0x4ea41c80, //mov v0.16b, v4.16b
- 0x4ea51ca1, //mov v1.16b, v5.16b
- 0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4ea71ce3, //mov v3.16b, v7.16b
- 0x4eb31e64, //mov v4.16b, v19.16b
- 0x4eb21e45, //mov v5.16b, v18.16b
- 0x4eb11e26, //mov v6.16b, v17.16b
- 0x4eb01e07, //mov v7.16b, v16.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_move_src_dst_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea01c04, //mov v4.16b, v0.16b
- 0x4ea11c25, //mov v5.16b, v1.16b
- 0x4ea21c46, //mov v6.16b, v2.16b
- 0x4ea31c67, //mov v7.16b, v3.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_move_dst_src_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea41c80, //mov v0.16b, v4.16b
- 0x4ea51ca1, //mov v1.16b, v5.16b
- 0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4ea71ce3, //mov v3.16b, v7.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_premul_aarch64[] = {
- 0xf8408423, //ldr x3, [x1], #8
- 0x6e23dc00, //fmul v0.4s, v0.4s, v3.4s
- 0x6e23dc21, //fmul v1.4s, v1.4s, v3.4s
- 0x6e23dc42, //fmul v2.4s, v2.4s, v3.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_unpremul_aarch64[] = {
- 0x4f03f611, //fmov v17.4s, #1.000000000000000000e+00
- 0xf8408423, //ldr x3, [x1], #8
- 0x4ea0d870, //fcmeq v16.4s, v3.4s, #0.0
- 0x6e23fe31, //fdiv v17.4s, v17.4s, v3.4s
- 0x4e701e30, //bic v16.16b, v17.16b, v16.16b
- 0x6e20de00, //fmul v0.4s, v16.4s, v0.4s
- 0x6e21de01, //fmul v1.4s, v16.4s, v1.4s
- 0x6e22de02, //fmul v2.4s, v16.4s, v2.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_from_srgb_aarch64[] = {
- 0x52a7d328, //mov w8, #0x3e990000
- 0x72933348, //movk w8, #0x999a
- 0x4e040d10, //dup v16.4s, w8
- 0x52a7e648, //mov w8, #0x3f320000
- 0x7291eb88, //movk w8, #0x8f5c
- 0x4e040d11, //dup v17.4s, w8
- 0x52a76468, //mov w8, #0x3b230000
- 0x729ae148, //movk w8, #0xd70a
- 0x4e040d12, //dup v18.4s, w8
- 0x52a7b3c8, //mov w8, #0x3d9e0000
- 0x72907228, //movk w8, #0x8391
- 0x6e22dc54, //fmul v20.4s, v2.4s, v2.4s
- 0x4eb11e35, //mov v21.16b, v17.16b
- 0x4eb11e37, //mov v23.16b, v17.16b
- 0x4e22ce11, //fmla v17.4s, v16.4s, v2.4s
- 0x4eb21e56, //mov v22.16b, v18.16b
- 0x4eb21e58, //mov v24.16b, v18.16b
- 0x4e34ce32, //fmla v18.4s, v17.4s, v20.4s
- 0x4e040d11, //dup v17.4s, w8
- 0x52a7ac28, //mov w8, #0x3d610000
- 0x6e20dc13, //fmul v19.4s, v0.4s, v0.4s
- 0x7288f5c8, //movk w8, #0x47ae
- 0x4e20ce15, //fmla v21.4s, v16.4s, v0.4s
- 0xf8408423, //ldr x3, [x1], #8
- 0x6e21dc34, //fmul v20.4s, v1.4s, v1.4s
- 0x4e33ceb6, //fmla v22.4s, v21.4s, v19.4s
- 0x4e040d13, //dup v19.4s, w8
- 0x4e21ce17, //fmla v23.4s, v16.4s, v1.4s
- 0x6e31dc15, //fmul v21.4s, v0.4s, v17.4s
- 0x6ea0e660, //fcmgt v0.4s, v19.4s, v0.4s
- 0x6e31dc30, //fmul v16.4s, v1.4s, v17.4s
- 0x6ea1e661, //fcmgt v1.4s, v19.4s, v1.4s
- 0x6e31dc51, //fmul v17.4s, v2.4s, v17.4s
- 0x6ea2e662, //fcmgt v2.4s, v19.4s, v2.4s
- 0x4e34cef8, //fmla v24.4s, v23.4s, v20.4s
- 0x6e761ea0, //bsl v0.16b, v21.16b, v22.16b
- 0x6e781e01, //bsl v1.16b, v16.16b, v24.16b
- 0x6e721e22, //bsl v2.16b, v17.16b, v18.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_to_srgb_aarch64[] = {
- 0x52a828e8, //mov w8, #0x41470000
- 0x728b8528, //movk w8, #0x5c29
- 0x4e040d12, //dup v18.4s, w8
- 0x52a7e608, //mov w8, #0x3f300000
- 0x728df9c8, //movk w8, #0x6fce
- 0x6ea1d811, //frsqrte v17.4s, v0.4s
- 0x4e040d13, //dup v19.4s, w8
- 0x52b7b948, //mov w8, #0xbdca0000
- 0x728af508, //movk w8, #0x57a8
- 0x6ea1d834, //frsqrte v20.4s, v1.4s
- 0x6e31de36, //fmul v22.4s, v17.4s, v17.4s
- 0x4e040d10, //dup v16.4s, w8
- 0x52a77188, //mov w8, #0x3b8c0000
- 0x6ea1d855, //frsqrte v21.4s, v2.4s
- 0x6e34de98, //fmul v24.4s, v20.4s, v20.4s
- 0x4eb6fc16, //frsqrts v22.4s, v0.4s, v22.4s
- 0x729ce088, //movk w8, #0xe704
- 0x6e35deb9, //fmul v25.4s, v21.4s, v21.4s
- 0x4eb8fc38, //frsqrts v24.4s, v1.4s, v24.4s
- 0x6e36de31, //fmul v17.4s, v17.4s, v22.4s
- 0x4e040d17, //dup v23.4s, w8
- 0x4eb9fc59, //frsqrts v25.4s, v2.4s, v25.4s
- 0x6e38de94, //fmul v20.4s, v20.4s, v24.4s
- 0x4ea1da36, //frecpe v22.4s, v17.4s
- 0x6e32dc1a, //fmul v26.4s, v0.4s, v18.4s
- 0x6ea0e6e0, //fcmgt v0.4s, v23.4s, v0.4s
- 0x6e32dc3c, //fmul v28.4s, v1.4s, v18.4s
- 0x6ea1e6e1, //fcmgt v1.4s, v23.4s, v1.4s
- 0x6e32dc52, //fmul v18.4s, v2.4s, v18.4s
- 0x6ea2e6e2, //fcmgt v2.4s, v23.4s, v2.4s
- 0x6e39deb5, //fmul v21.4s, v21.4s, v25.4s
- 0x4ea1da97, //frecpe v23.4s, v20.4s
- 0x4e36fe39, //frecps v25.4s, v17.4s, v22.4s
- 0x4ea1dab8, //frecpe v24.4s, v21.4s
- 0x6e39ded6, //fmul v22.4s, v22.4s, v25.4s
- 0x4e37fe99, //frecps v25.4s, v20.4s, v23.4s
- 0x4eb01e1b, //mov v27.16b, v16.16b
- 0x6e39def7, //fmul v23.4s, v23.4s, v25.4s
- 0x4e38feb9, //frecps v25.4s, v21.4s, v24.4s
- 0x6e39df18, //fmul v24.4s, v24.4s, v25.4s
- 0x4eb01e19, //mov v25.16b, v16.16b
- 0x4e36ce7b, //fmla v27.4s, v19.4s, v22.4s
- 0x6ea1da36, //frsqrte v22.4s, v17.4s
- 0x4e37ce79, //fmla v25.4s, v19.4s, v23.4s
- 0x6ea1da97, //frsqrte v23.4s, v20.4s
- 0x4e38ce70, //fmla v16.4s, v19.4s, v24.4s
- 0x6e36ded8, //fmul v24.4s, v22.4s, v22.4s
- 0x6ea1dab3, //frsqrte v19.4s, v21.4s
- 0x4eb8fe31, //frsqrts v17.4s, v17.4s, v24.4s
- 0x6e37def8, //fmul v24.4s, v23.4s, v23.4s
- 0x4eb8fe94, //frsqrts v20.4s, v20.4s, v24.4s
- 0x6e33de78, //fmul v24.4s, v19.4s, v19.4s
- 0x52a7da48, //mov w8, #0x3ed20000
- 0x4eb8feb5, //frsqrts v21.4s, v21.4s, v24.4s
- 0x7290f848, //movk w8, #0x87c2
- 0x6e31ded1, //fmul v17.4s, v22.4s, v17.4s
- 0x6e34def4, //fmul v20.4s, v23.4s, v20.4s
- 0x6e35de73, //fmul v19.4s, v19.4s, v21.4s
- 0x4e040d15, //dup v21.4s, w8
- 0xf8408423, //ldr x3, [x1], #8
- 0x4e31cebb, //fmla v27.4s, v21.4s, v17.4s
- 0x4f03f611, //fmov v17.4s, #1.000000000000000000e+00
- 0x4e34ceb9, //fmla v25.4s, v21.4s, v20.4s
- 0x4e33ceb0, //fmla v16.4s, v21.4s, v19.4s
- 0x4ebbf633, //fmin v19.4s, v17.4s, v27.4s
- 0x4eb9f634, //fmin v20.4s, v17.4s, v25.4s
- 0x4eb0f630, //fmin v16.4s, v17.4s, v16.4s
- 0x6e731f40, //bsl v0.16b, v26.16b, v19.16b
- 0x6e741f81, //bsl v1.16b, v28.16b, v20.16b
- 0x6e701e42, //bsl v2.16b, v18.16b, v16.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_scale_1_float_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xbd400110, //ldr s16, [x8]
- 0x4f909000, //fmul v0.4s, v0.4s, v16.s[0]
- 0x4f909021, //fmul v1.4s, v1.4s, v16.s[0]
- 0x4f909042, //fmul v2.4s, v2.4s, v16.s[0]
- 0x4f909063, //fmul v3.4s, v3.4s, v16.s[0]
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_scale_u8_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x52a77009, //mov w9, #0x3b800000
- 0x72901029, //movk w9, #0x8081
- 0x4e040d30, //dup v16.4s, w9
- 0xf9400108, //ldr x8, [x8]
- 0x8b000108, //add x8, x8, x0
- 0x39400109, //ldrb w9, [x8]
- 0x3940050a, //ldrb w10, [x8, #1]
- 0x3940090b, //ldrb w11, [x8, #2]
- 0x39400d08, //ldrb w8, [x8, #3]
- 0x4e021d31, //mov v17.h[0], w9
- 0x4e061d51, //mov v17.h[1], w10
- 0x4e0a1d71, //mov v17.h[2], w11
- 0x4e0e1d11, //mov v17.h[3], w8
- 0x2f10a631, //uxtl v17.4s, v17.4h
- 0x6e21da31, //ucvtf v17.4s, v17.4s
- 0x6e30de30, //fmul v16.4s, v17.4s, v16.4s
- 0x6e20de00, //fmul v0.4s, v16.4s, v0.4s
- 0x6e21de01, //fmul v1.4s, v16.4s, v1.4s
- 0x6e22de02, //fmul v2.4s, v16.4s, v2.4s
- 0x6e23de03, //fmul v3.4s, v16.4s, v3.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_lerp_1_float_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x4ea4d411, //fsub v17.4s, v0.4s, v4.4s
- 0x4ea41c80, //mov v0.16b, v4.16b
- 0x4ea5d432, //fsub v18.4s, v1.4s, v5.4s
- 0xbd400110, //ldr s16, [x8]
- 0x4ea51ca1, //mov v1.16b, v5.16b
- 0x4f901220, //fmla v0.4s, v17.4s, v16.s[0]
- 0x4ea6d451, //fsub v17.4s, v2.4s, v6.4s
- 0x4f901241, //fmla v1.4s, v18.4s, v16.s[0]
- 0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4ea7d472, //fsub v18.4s, v3.4s, v7.4s
- 0x4ea71ce3, //mov v3.16b, v7.16b
- 0x4f901222, //fmla v2.4s, v17.4s, v16.s[0]
- 0x4f901243, //fmla v3.4s, v18.4s, v16.s[0]
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_lerp_u8_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x52a77009, //mov w9, #0x3b800000
- 0x72901029, //movk w9, #0x8081
- 0x4e040d30, //dup v16.4s, w9
- 0xf9400108, //ldr x8, [x8]
- 0x4ea4d412, //fsub v18.4s, v0.4s, v4.4s
- 0x8b000108, //add x8, x8, x0
- 0x3940010a, //ldrb w10, [x8]
- 0x39400509, //ldrb w9, [x8, #1]
- 0x3940090b, //ldrb w11, [x8, #2]
- 0x39400d08, //ldrb w8, [x8, #3]
- 0x4e021d51, //mov v17.h[0], w10
- 0x4e061d31, //mov v17.h[1], w9
- 0x4e0a1d71, //mov v17.h[2], w11
- 0x4e0e1d11, //mov v17.h[3], w8
- 0x2f10a620, //uxtl v0.4s, v17.4h
- 0x6e21d800, //ucvtf v0.4s, v0.4s
- 0x6e30dc10, //fmul v16.4s, v0.4s, v16.4s
- 0x4ea41c80, //mov v0.16b, v4.16b
- 0x4ea5d431, //fsub v17.4s, v1.4s, v5.4s
- 0x4ea51ca1, //mov v1.16b, v5.16b
- 0x4e32ce00, //fmla v0.4s, v16.4s, v18.4s
- 0x4ea6d452, //fsub v18.4s, v2.4s, v6.4s
- 0x4e31ce01, //fmla v1.4s, v16.4s, v17.4s
- 0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4ea7d471, //fsub v17.4s, v3.4s, v7.4s
- 0x4ea71ce3, //mov v3.16b, v7.16b
- 0x4e32ce02, //fmla v2.4s, v16.4s, v18.4s
- 0x4e31ce03, //fmla v3.4s, v16.4s, v17.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_lerp_565_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xd37ff809, //lsl x9, x0, #1
- 0x4f072710, //movi v16.4s, #0xf8, lsl #8
- 0x4ea4d413, //fsub v19.4s, v0.4s, v4.4s
- 0xf9400108, //ldr x8, [x8]
- 0xfc696903, //ldr d3, [x8, x9]
- 0x52a6f088, //mov w8, #0x37840000
- 0x72842108, //movk w8, #0x2108
- 0x4e040d11, //dup v17.4s, w8
- 0x2f10a463, //uxtl v3.4s, v3.4h
- 0x321b17e8, //orr w8, wzr, #0x7e0
- 0x4e301c60, //and v0.16b, v3.16b, v16.16b
- 0x4e040d12, //dup v18.4s, w8
- 0x52a74048, //mov w8, #0x3a020000
- 0x4e21d800, //scvtf v0.4s, v0.4s
- 0x72810428, //movk w8, #0x821
- 0x6e31dc10, //fmul v16.4s, v0.4s, v17.4s
- 0x4ea41c80, //mov v0.16b, v4.16b
- 0x4e33ce00, //fmla v0.4s, v16.4s, v19.4s
- 0x4f0007f0, //movi v16.4s, #0x1f
- 0x4e040d11, //dup v17.4s, w8
- 0x52a7a088, //mov w8, #0x3d040000
- 0x4e321c72, //and v18.16b, v3.16b, v18.16b
- 0x72842108, //movk w8, #0x2108
- 0x4e301c63, //and v3.16b, v3.16b, v16.16b
- 0x4ea6d450, //fsub v16.4s, v2.4s, v6.4s
- 0x4e21da42, //scvtf v2.4s, v18.4s
- 0x6e31dc51, //fmul v17.4s, v2.4s, v17.4s
- 0x4e040d02, //dup v2.4s, w8
- 0x4e21d863, //scvtf v3.4s, v3.4s
- 0x4ea5d433, //fsub v19.4s, v1.4s, v5.4s
- 0x4ea51ca1, //mov v1.16b, v5.16b
- 0x6e22dc63, //fmul v3.4s, v3.4s, v2.4s
- 0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4e33ce21, //fmla v1.4s, v17.4s, v19.4s
- 0x4e30cc62, //fmla v2.4s, v3.4s, v16.4s
- 0x4f03f603, //fmov v3.4s, #1.000000000000000000e+00
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_load_tables_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xd37ef409, //lsl x9, x0, #2
- 0x6f00e620, //movi v0.2d, #0xff000000ff
- 0x52a7700b, //mov w11, #0x3b800000
- 0xa940310a, //ldp x10, x12, [x8]
- 0x7290102b, //movk w11, #0x8081
- 0x4e040d63, //dup v3.4s, w11
- 0x3ce96942, //ldr q2, [x10, x9]
- 0xa9412109, //ldp x9, x8, [x8, #16]
- 0x4e201c41, //and v1.16b, v2.16b, v0.16b
- 0x1e26002e, //fmov w14, s1
- 0x6f380450, //ushr v16.4s, v2.4s, #8
- 0x6f300451, //ushr v17.4s, v2.4s, #16
- 0x8b2e498e, //add x14, x12, w14, uxtw #2
- 0x0e0c3c2a, //mov w10, v1.s[1]
- 0x0e143c2b, //mov w11, v1.s[2]
- 0x0e1c3c2d, //mov w13, v1.s[3]
- 0x4e201e01, //and v1.16b, v16.16b, v0.16b
- 0x4e201e30, //and v16.16b, v17.16b, v0.16b
- 0x0d4081c0, //ld1 {v0.s}[0], [x14]
- 0x8b2a498a, //add x10, x12, w10, uxtw #2
- 0xbc6b5991, //ldr s17, [x12, w11, uxtw #2]
- 0xbc6d5992, //ldr s18, [x12, w13, uxtw #2]
- 0x0e0c3c2b, //mov w11, v1.s[1]
- 0x0e143c2c, //mov w12, v1.s[2]
- 0x0e1c3c2d, //mov w13, v1.s[3]
- 0x1e26002e, //fmov w14, s1
- 0x8b2e492e, //add x14, x9, w14, uxtw #2
- 0xbc6c5933, //ldr s19, [x9, w12, uxtw #2]
- 0xbc6d5934, //ldr s20, [x9, w13, uxtw #2]
- 0x8b2b4929, //add x9, x9, w11, uxtw #2
- 0x1e26020b, //fmov w11, s16
- 0x6f280442, //ushr v2.4s, v2.4s, #24
- 0x0d409140, //ld1 {v0.s}[1], [x10]
- 0x4e21d842, //scvtf v2.4s, v2.4s
- 0x8b2b490a, //add x10, x8, w11, uxtw #2
- 0x0d4081c1, //ld1 {v1.s}[0], [x14]
- 0x6e23dc43, //fmul v3.4s, v2.4s, v3.4s
- 0x0d408142, //ld1 {v2.s}[0], [x10]
- 0x0e0c3e0f, //mov w15, v16.s[1]
- 0x0e143e0c, //mov w12, v16.s[2]
- 0x8b2f490a, //add x10, x8, w15, uxtw #2
- 0x0e1c3e0d, //mov w13, v16.s[3]
- 0xbc6c5910, //ldr s16, [x8, w12, uxtw #2]
- 0x0d409121, //ld1 {v1.s}[1], [x9]
- 0x0d409142, //ld1 {v2.s}[1], [x10]
- 0x6e140620, //mov v0.s[2], v17.s[0]
- 0xbc6d5911, //ldr s17, [x8, w13, uxtw #2]
- 0x6e140661, //mov v1.s[2], v19.s[0]
- 0x6e140602, //mov v2.s[2], v16.s[0]
- 0x6e1c0640, //mov v0.s[3], v18.s[0]
- 0x6e1c0681, //mov v1.s[3], v20.s[0]
- 0x6e1c0622, //mov v2.s[3], v17.s[0]
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_load_a8_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x52a77009, //mov w9, #0x3b800000
- 0x72901029, //movk w9, #0x8081
- 0x4e040d22, //dup v2.4s, w9
- 0xf9400108, //ldr x8, [x8]
- 0x6f00e400, //movi v0.2d, #0x0
- 0x6f00e401, //movi v1.2d, #0x0
- 0x8b000108, //add x8, x8, x0
- 0x3940010a, //ldrb w10, [x8]
- 0x3940050b, //ldrb w11, [x8, #1]
- 0x3940090c, //ldrb w12, [x8, #2]
- 0x39400d08, //ldrb w8, [x8, #3]
- 0x4e021d43, //mov v3.h[0], w10
- 0x4e061d63, //mov v3.h[1], w11
- 0x4e0a1d83, //mov v3.h[2], w12
- 0x4e0e1d03, //mov v3.h[3], w8
- 0x2f10a463, //uxtl v3.4s, v3.4h
- 0x6e21d863, //ucvtf v3.4s, v3.4s
- 0x6e22dc63, //fmul v3.4s, v3.4s, v2.4s
- 0x6f00e402, //movi v2.2d, #0x0
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_store_a8_aarch64[] = {
- 0xf9400028, //ldr x8, [x1]
- 0x52a86fe9, //mov w9, #0x437f0000
- 0x4e040d30, //dup v16.4s, w9
- 0x6e30dc70, //fmul v16.4s, v3.4s, v16.4s
- 0xf9400108, //ldr x8, [x8]
- 0x6e21aa10, //fcvtnu v16.4s, v16.4s
- 0x0e612a10, //xtn v16.4h, v16.4s
- 0x0e0e3e09, //umov w9, v16.h[3]
- 0x8b000108, //add x8, x8, x0
- 0x39000d09, //strb w9, [x8, #3]
- 0x0e0a3e09, //umov w9, v16.h[2]
- 0x39000909, //strb w9, [x8, #2]
- 0x0e063e09, //umov w9, v16.h[1]
- 0x39000509, //strb w9, [x8, #1]
- 0x0e023e09, //umov w9, v16.h[0]
- 0x39000109, //strb w9, [x8]
- 0xf9400423, //ldr x3, [x1, #8]
- 0x91004021, //add x1, x1, #0x10
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_load_565_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xd37ff809, //lsl x9, x0, #1
- 0x4f072701, //movi v1.4s, #0xf8, lsl #8
- 0x4f0007e3, //movi v3.4s, #0x1f
- 0xf9400108, //ldr x8, [x8]
- 0xfc696900, //ldr d0, [x8, x9]
- 0x321b17e8, //orr w8, wzr, #0x7e0
- 0x4e040d02, //dup v2.4s, w8
- 0x52a6f088, //mov w8, #0x37840000
- 0x72842108, //movk w8, #0x2108
- 0x2f10a400, //uxtl v0.4s, v0.4h
- 0x4e211c01, //and v1.16b, v0.16b, v1.16b
- 0x4e221c02, //and v2.16b, v0.16b, v2.16b
- 0x4e231c03, //and v3.16b, v0.16b, v3.16b
- 0x4e040d00, //dup v0.4s, w8
- 0x52a74048, //mov w8, #0x3a020000
- 0x72810428, //movk w8, #0x821
- 0x4e21d821, //scvtf v1.4s, v1.4s
- 0x6e20dc20, //fmul v0.4s, v1.4s, v0.4s
- 0x4e040d01, //dup v1.4s, w8
- 0x52a7a088, //mov w8, #0x3d040000
- 0x72842108, //movk w8, #0x2108
- 0x4e21d842, //scvtf v2.4s, v2.4s
- 0x6e21dc41, //fmul v1.4s, v2.4s, v1.4s
- 0x4e040d02, //dup v2.4s, w8
- 0x4e21d863, //scvtf v3.4s, v3.4s
- 0x6e22dc62, //fmul v2.4s, v3.4s, v2.4s
- 0x4f03f603, //fmov v3.4s, #1.000000000000000000e+00
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_store_565_aarch64[] = {
- 0xf9400028, //ldr x8, [x1]
- 0x52a84f8a, //mov w10, #0x427c0000
- 0x4f01f7f0, //fmov v16.4s, #3.100000000000000000e+01
- 0x4e040d52, //dup v18.4s, w10
- 0x6e30dc11, //fmul v17.4s, v0.4s, v16.4s
- 0x6e32dc32, //fmul v18.4s, v1.4s, v18.4s
- 0x6e21aa31, //fcvtnu v17.4s, v17.4s
- 0x6e21aa52, //fcvtnu v18.4s, v18.4s
- 0x6e30dc50, //fmul v16.4s, v2.4s, v16.4s
- 0x4f2b5631, //shl v17.4s, v17.4s, #11
- 0xf9400108, //ldr x8, [x8]
- 0x4f255652, //shl v18.4s, v18.4s, #5
- 0x4eb11e51, //orr v17.16b, v18.16b, v17.16b
- 0x6e21aa10, //fcvtnu v16.4s, v16.4s
- 0x4eb01e30, //orr v16.16b, v17.16b, v16.16b
- 0xd37ff809, //lsl x9, x0, #1
- 0x0e612a10, //xtn v16.4h, v16.4s
- 0xfc296910, //str d16, [x8, x9]
- 0xf9400423, //ldr x3, [x1, #8]
- 0x91004021, //add x1, x1, #0x10
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_load_8888_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xd37ef409, //lsl x9, x0, #2
- 0x6f00e621, //movi v1.2d, #0xff000000ff
- 0xf9400108, //ldr x8, [x8]
- 0x3ce96900, //ldr q0, [x8, x9]
- 0x52a77008, //mov w8, #0x3b800000
- 0x72901028, //movk w8, #0x8081
- 0x4e040d02, //dup v2.4s, w8
- 0x6f380410, //ushr v16.4s, v0.4s, #8
- 0x6f300411, //ushr v17.4s, v0.4s, #16
- 0x4e211c03, //and v3.16b, v0.16b, v1.16b
- 0x6f280400, //ushr v0.4s, v0.4s, #24
- 0x4e211e10, //and v16.16b, v16.16b, v1.16b
- 0x4e211e21, //and v1.16b, v17.16b, v1.16b
- 0x4e21d863, //scvtf v3.4s, v3.4s
- 0x4e21d811, //scvtf v17.4s, v0.4s
- 0x4e21da10, //scvtf v16.4s, v16.4s
- 0x4e21d832, //scvtf v18.4s, v1.4s
- 0x6e22dc60, //fmul v0.4s, v3.4s, v2.4s
- 0x6e22de23, //fmul v3.4s, v17.4s, v2.4s
- 0x6e22de01, //fmul v1.4s, v16.4s, v2.4s
- 0x6e22de42, //fmul v2.4s, v18.4s, v2.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_store_8888_aarch64[] = {
- 0x52a86fea, //mov w10, #0x437f0000
- 0x4e040d50, //dup v16.4s, w10
- 0xf9400028, //ldr x8, [x1]
- 0x6e30dc32, //fmul v18.4s, v1.4s, v16.4s
- 0x6e30dc11, //fmul v17.4s, v0.4s, v16.4s
- 0x6e21aa52, //fcvtnu v18.4s, v18.4s
- 0x6e21aa31, //fcvtnu v17.4s, v17.4s
- 0x4f285652, //shl v18.4s, v18.4s, #8
- 0x4eb11e51, //orr v17.16b, v18.16b, v17.16b
- 0x6e30dc52, //fmul v18.4s, v2.4s, v16.4s
- 0x6e30dc70, //fmul v16.4s, v3.4s, v16.4s
- 0x6e21aa52, //fcvtnu v18.4s, v18.4s
- 0xf9400108, //ldr x8, [x8]
- 0x6e21aa10, //fcvtnu v16.4s, v16.4s
- 0x4f305652, //shl v18.4s, v18.4s, #16
- 0x4eb21e31, //orr v17.16b, v17.16b, v18.16b
- 0x4f385610, //shl v16.4s, v16.4s, #24
- 0xd37ef409, //lsl x9, x0, #2
- 0x4eb01e30, //orr v16.16b, v17.16b, v16.16b
- 0x3ca96910, //str q16, [x8, x9]
- 0xf9400423, //ldr x3, [x1, #8]
- 0x91004021, //add x1, x1, #0x10
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_load_f16_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xf9400108, //ldr x8, [x8]
- 0x8b000d08, //add x8, x8, x0, lsl #3
- 0x0c400510, //ld4 {v16.4h-v19.4h}, [x8]
- 0x0e217a00, //fcvtl v0.4s, v16.4h
- 0x0e217a21, //fcvtl v1.4s, v17.4h
- 0x0e217a42, //fcvtl v2.4s, v18.4h
- 0x0e217a63, //fcvtl v3.4s, v19.4h
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_store_f16_aarch64[] = {
- 0xf9400028, //ldr x8, [x1]
- 0x0e216810, //fcvtn v16.4h, v0.4s
- 0x0e216831, //fcvtn v17.4h, v1.4s
- 0x0e216852, //fcvtn v18.4h, v2.4s
- 0xf9400108, //ldr x8, [x8]
- 0x0e216873, //fcvtn v19.4h, v3.4s
- 0x8b000d08, //add x8, x8, x0, lsl #3
- 0x0c000510, //st4 {v16.4h-v19.4h}, [x8]
- 0xf9400423, //ldr x3, [x1, #8]
- 0x91004021, //add x1, x1, #0x10
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_store_f32_aarch64[] = {
- 0xf9400028, //ldr x8, [x1]
- 0xf9400108, //ldr x8, [x8]
- 0x8b001108, //add x8, x8, x0, lsl #4
- 0x4c000900, //st4 {v0.4s-v3.4s}, [x8]
- 0xf9400423, //ldr x3, [x1, #8]
- 0x91004021, //add x1, x1, #0x10
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clamp_x_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x6f00e411, //movi v17.2d, #0x0
- 0x4e20f620, //fmax v0.4s, v17.4s, v0.4s
- 0x6f07e7f1, //movi v17.2d, #0xffffffffffffffff
- 0x4d40c910, //ld1r {v16.4s}, [x8]
- 0x4eb18610, //add v16.4s, v16.4s, v17.4s
- 0x4eb0f400, //fmin v0.4s, v0.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_clamp_y_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x6f00e411, //movi v17.2d, #0x0
- 0x4e21f621, //fmax v1.4s, v17.4s, v1.4s
- 0x6f07e7f1, //movi v17.2d, #0xffffffffffffffff
- 0x4d40c910, //ld1r {v16.4s}, [x8]
- 0x4eb18610, //add v16.4s, v16.4s, v17.4s
- 0x4eb0f421, //fmin v1.4s, v1.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_repeat_x_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x6f07e7f1, //movi v17.2d, #0xffffffffffffffff
- 0xbd400110, //ldr s16, [x8]
- 0x4e040612, //dup v18.4s, v16.s[0]
- 0x4eb18651, //add v17.4s, v18.4s, v17.4s
- 0x6e32fc12, //fdiv v18.4s, v0.4s, v18.4s
- 0x4e219a52, //frintm v18.4s, v18.4s
- 0x4f905240, //fmls v0.4s, v18.4s, v16.s[0]
- 0x4eb1f400, //fmin v0.4s, v0.4s, v17.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_repeat_y_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x6f07e7f1, //movi v17.2d, #0xffffffffffffffff
- 0xbd400110, //ldr s16, [x8]
- 0x4e040612, //dup v18.4s, v16.s[0]
- 0x4eb18651, //add v17.4s, v18.4s, v17.4s
- 0x6e32fc32, //fdiv v18.4s, v1.4s, v18.4s
- 0x4e219a52, //frintm v18.4s, v18.4s
- 0x4f905241, //fmls v1.4s, v18.4s, v16.s[0]
- 0x4eb1f421, //fmin v1.4s, v1.4s, v17.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_mirror_x_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xbd400110, //ldr s16, [x8]
- 0x4e040611, //dup v17.4s, v16.s[0]
- 0x1e302a10, //fadd s16, s16, s16
- 0x4eb1d400, //fsub v0.4s, v0.4s, v17.4s
- 0x4e040612, //dup v18.4s, v16.s[0]
- 0x6e32fc12, //fdiv v18.4s, v0.4s, v18.4s
- 0x4e219a52, //frintm v18.4s, v18.4s
- 0x4f905240, //fmls v0.4s, v18.4s, v16.s[0]
- 0x6f07e7f0, //movi v16.2d, #0xffffffffffffffff
- 0x4eb1d400, //fsub v0.4s, v0.4s, v17.4s
- 0x4eb08630, //add v16.4s, v17.4s, v16.4s
- 0x4ea0f800, //fabs v0.4s, v0.4s
- 0x4eb0f400, //fmin v0.4s, v0.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_mirror_y_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xbd400110, //ldr s16, [x8]
- 0x4e040611, //dup v17.4s, v16.s[0]
- 0x1e302a10, //fadd s16, s16, s16
- 0x4eb1d421, //fsub v1.4s, v1.4s, v17.4s
- 0x4e040612, //dup v18.4s, v16.s[0]
- 0x6e32fc32, //fdiv v18.4s, v1.4s, v18.4s
- 0x4e219a52, //frintm v18.4s, v18.4s
- 0x4f905241, //fmls v1.4s, v18.4s, v16.s[0]
- 0x6f07e7f0, //movi v16.2d, #0xffffffffffffffff
- 0x4eb1d421, //fsub v1.4s, v1.4s, v17.4s
- 0x4eb08630, //add v16.4s, v17.4s, v16.4s
- 0x4ea0f821, //fabs v1.4s, v1.4s
- 0x4eb0f421, //fmin v1.4s, v1.4s, v16.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_luminance_to_alpha_aarch64[] = {
- 0x52a7cb28, //mov w8, #0x3e590000
- 0x72967a08, //movk w8, #0xb3d0
- 0x4e040d11, //dup v17.4s, w8
- 0x52a7e6e8, //mov w8, #0x3f370000
- 0x7282eb28, //movk w8, #0x1759
- 0x4ea01c10, //mov v16.16b, v0.16b
- 0x4e040d00, //dup v0.4s, w8
- 0x52a7b268, //mov w8, #0x3d930000
- 0xf8408423, //ldr x3, [x1], #8
- 0x729bb308, //movk w8, #0xdd98
- 0x6e20dc23, //fmul v3.4s, v1.4s, v0.4s
- 0x4e30ce23, //fmla v3.4s, v17.4s, v16.4s
- 0x4e040d10, //dup v16.4s, w8
- 0x6f00e400, //movi v0.2d, #0x0
- 0x6f00e401, //movi v1.2d, #0x0
- 0x4e22ce03, //fmla v3.4s, v16.4s, v2.4s
- 0x6f00e402, //movi v2.2d, #0x0
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_matrix_2x3_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xaa0803e9, //mov x9, x8
- 0x9100410a, //add x10, x8, #0x10
- 0x4ddfc932, //ld1r {v18.4s}, [x9], #4
- 0x4d40c950, //ld1r {v16.4s}, [x10]
- 0x2d415113, //ldp s19, s20, [x8, #8]
- 0x9100510a, //add x10, x8, #0x14
- 0x4d40c951, //ld1r {v17.4s}, [x10]
- 0x4f931030, //fmla v16.4s, v1.4s, v19.s[0]
- 0xbd400133, //ldr s19, [x9]
- 0x4f941031, //fmla v17.4s, v1.4s, v20.s[0]
- 0x4e20ce50, //fmla v16.4s, v18.4s, v0.4s
- 0x4f931011, //fmla v17.4s, v0.4s, v19.s[0]
- 0x4eb01e00, //mov v0.16b, v16.16b
- 0x4eb11e21, //mov v1.16b, v17.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_matrix_3x4_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xaa0803e9, //mov x9, x8
- 0x9100910a, //add x10, x8, #0x24
- 0x4ddfc933, //ld1r {v19.4s}, [x9], #4
- 0x4d40c950, //ld1r {v16.4s}, [x10]
- 0x9100a10a, //add x10, x8, #0x28
- 0x4d40c951, //ld1r {v17.4s}, [x10]
- 0x9100b10a, //add x10, x8, #0x2c
- 0x2d435514, //ldp s20, s21, [x8, #24]
- 0xbd402116, //ldr s22, [x8, #32]
- 0x4d40c952, //ld1r {v18.4s}, [x10]
- 0x4f941050, //fmla v16.4s, v2.4s, v20.s[0]
- 0x4f951051, //fmla v17.4s, v2.4s, v21.s[0]
- 0x4f961052, //fmla v18.4s, v2.4s, v22.s[0]
- 0x2d425502, //ldp s2, s21, [x8, #16]
- 0x2d415d14, //ldp s20, s23, [x8, #8]
- 0x4f821031, //fmla v17.4s, v1.4s, v2.s[0]
- 0xbd400122, //ldr s2, [x9]
- 0x4f971030, //fmla v16.4s, v1.4s, v23.s[0]
- 0x4f951032, //fmla v18.4s, v1.4s, v21.s[0]
- 0x4e20ce70, //fmla v16.4s, v19.4s, v0.4s
- 0x4f941012, //fmla v18.4s, v0.4s, v20.s[0]
- 0x4f821011, //fmla v17.4s, v0.4s, v2.s[0]
- 0x4eb01e00, //mov v0.16b, v16.16b
- 0x4eb11e21, //mov v1.16b, v17.16b
- 0x4eb21e42, //mov v2.16b, v18.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_matrix_4x5_aarch64[] = {
- 0xf9400029, //ldr x9, [x1]
- 0xaa0903e8, //mov x8, x9
- 0x9101012a, //add x10, x9, #0x40
- 0x4ddfc914, //ld1r {v20.4s}, [x8], #4
- 0x4d40c950, //ld1r {v16.4s}, [x10]
- 0x9101112a, //add x10, x9, #0x44
- 0x4d40c951, //ld1r {v17.4s}, [x10]
- 0x9101212a, //add x10, x9, #0x48
- 0x4d40c952, //ld1r {v18.4s}, [x10]
- 0x2d465533, //ldp s19, s21, [x9, #48]
- 0x2d475d36, //ldp s22, s23, [x9, #56]
- 0x9101312a, //add x10, x9, #0x4c
- 0xf9400423, //ldr x3, [x1, #8]
- 0x4f931070, //fmla v16.4s, v3.4s, v19.s[0]
- 0x4d40c953, //ld1r {v19.4s}, [x10]
- 0x4f951071, //fmla v17.4s, v3.4s, v21.s[0]
- 0x4f961072, //fmla v18.4s, v3.4s, v22.s[0]
- 0x2d445935, //ldp s21, s22, [x9, #32]
- 0x4f971073, //fmla v19.4s, v3.4s, v23.s[0]
- 0x2d455d23, //ldp s3, s23, [x9, #40]
- 0x91004021, //add x1, x1, #0x10
- 0x4f951050, //fmla v16.4s, v2.4s, v21.s[0]
- 0x4f961051, //fmla v17.4s, v2.4s, v22.s[0]
- 0x2d425935, //ldp s21, s22, [x9, #16]
- 0x4f971053, //fmla v19.4s, v2.4s, v23.s[0]
- 0x4f831052, //fmla v18.4s, v2.4s, v3.s[0]
- 0x2d410d22, //ldp s2, s3, [x9, #8]
- 0x4f951030, //fmla v16.4s, v1.4s, v21.s[0]
- 0x2d435d35, //ldp s21, s23, [x9, #24]
- 0x4f961031, //fmla v17.4s, v1.4s, v22.s[0]
- 0xbd400116, //ldr s22, [x8]
- 0x4e20ce90, //fmla v16.4s, v20.4s, v0.4s
- 0x4f951032, //fmla v18.4s, v1.4s, v21.s[0]
- 0x4f971033, //fmla v19.4s, v1.4s, v23.s[0]
- 0x4f821012, //fmla v18.4s, v0.4s, v2.s[0]
- 0x4f831013, //fmla v19.4s, v0.4s, v3.s[0]
- 0x4f961011, //fmla v17.4s, v0.4s, v22.s[0]
- 0x4eb01e00, //mov v0.16b, v16.16b
- 0x4eb11e21, //mov v1.16b, v17.16b
- 0x4eb21e42, //mov v2.16b, v18.16b
- 0x4eb31e63, //mov v3.16b, v19.16b
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_matrix_perspective_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xaa0803e9, //mov x9, x8
- 0x9100510a, //add x10, x8, #0x14
- 0x4ddfc930, //ld1r {v16.4s}, [x9], #4
- 0x4d40c951, //ld1r {v17.4s}, [x10]
- 0x9100810a, //add x10, x8, #0x20
- 0x4d40c952, //ld1r {v18.4s}, [x10]
- 0x2d41d113, //ldp s19, s20, [x8, #12]
- 0x2d435915, //ldp s21, s22, [x8, #24]
- 0x91002108, //add x8, x8, #0x8
- 0x4f941031, //fmla v17.4s, v1.4s, v20.s[0]
- 0x4d40c914, //ld1r {v20.4s}, [x8]
- 0x4f961032, //fmla v18.4s, v1.4s, v22.s[0]
- 0xbd400136, //ldr s22, [x9]
- 0x4f951012, //fmla v18.4s, v0.4s, v21.s[0]
- 0x4f931011, //fmla v17.4s, v0.4s, v19.s[0]
- 0x4f961034, //fmla v20.4s, v1.4s, v22.s[0]
- 0x4ea1da41, //frecpe v1.4s, v18.4s
- 0x4e21fe52, //frecps v18.4s, v18.4s, v1.4s
- 0x6e32dc32, //fmul v18.4s, v1.4s, v18.4s
- 0x4e20ce14, //fmla v20.4s, v16.4s, v0.4s
- 0x6e32de21, //fmul v1.4s, v17.4s, v18.4s
- 0x6e32de80, //fmul v0.4s, v20.4s, v18.4s
- 0xd61f0060, //br x3
-};
-
-CODE const uint32_t sk_linear_gradient_2stops_aarch64[] = {
- 0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xad404503, //ldp q3, q17, [x8]
- 0x4e040470, //dup v16.4s, v3.s[0]
- 0x4e0c0461, //dup v1.4s, v3.s[1]
- 0x4e140462, //dup v2.4s, v3.s[2]
- 0x4e1c0463, //dup v3.4s, v3.s[3]
- 0x4f911010, //fmla v16.4s, v0.4s, v17.s[0]
- 0x4fb11001, //fmla v1.4s, v0.4s, v17.s[1]
- 0x4f911802, //fmla v2.4s, v0.4s, v17.s[2]
- 0x4fb11803, //fmla v3.4s, v0.4s, v17.s[3]
- 0x4eb01e00, //mov v0.16b, v16.16b
- 0xd61f0060, //br x3
-};
-#elif defined(__arm__)
-
-CODE const uint32_t sk_start_pipeline_vfp4[] = {
- 0xe92d41f0, //push {r4, r5, r6, r7, r8, lr}
- 0xe1a04000, //mov r4, r0
- 0xe2840002, //add r0, r4, #2
- 0xe1a05003, //mov r5, r3
- 0xe1a08002, //mov r8, r2
- 0xe1a07001, //mov r7, r1
- 0xe1500005, //cmp r0, r5
- 0x8a000010, //bhi 64 <sk_start_pipeline_vfp4+0x64>
- 0xe4976004, //ldr r6, [r7], #4
- 0xf2800010, //vmov.i32 d0, #0
- 0xe1a00004, //mov r0, r4
- 0xf2801010, //vmov.i32 d1, #0
- 0xe1a01007, //mov r1, r7
- 0xf2802010, //vmov.i32 d2, #0
- 0xe1a02008, //mov r2, r8
- 0xf2803010, //vmov.i32 d3, #0
- 0xf2804010, //vmov.i32 d4, #0
- 0xf2805010, //vmov.i32 d5, #0
- 0xf2806010, //vmov.i32 d6, #0
- 0xf2807010, //vmov.i32 d7, #0
- 0xe12fff36, //blx r6
- 0xe2840004, //add r0, r4, #4
- 0xe2844002, //add r4, r4, #2
- 0xe1500005, //cmp r0, r5
- 0x9affffef, //bls 24 <sk_start_pipeline_vfp4+0x24>
- 0xe1a00004, //mov r0, r4
- 0xe8bd81f0, //pop {r4, r5, r6, r7, r8, pc}
-};
-
-CODE const uint32_t sk_just_return_vfp4[] = {
- 0xe12fff1e, //bx lr
-};
-
-CODE const uint32_t sk_seed_shader_vfp4[] = {
- 0xee800b90, //vdup.32 d16, r0
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf2c3161f, //vmov.i32 d17, #1056964608
- 0xedd23b00, //vldr d19, [r2]
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf2872f10, //vmov.f32 d2, #1
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xe2811008, //add r1, r1, #8
- 0xf2400da1, //vadd.f32 d16, d16, d17
- 0xf2803010, //vmov.i32 d3, #0
- 0xf2804010, //vmov.i32 d4, #0
- 0xf2021da1, //vadd.f32 d1, d18, d17
- 0xf2000da3, //vadd.f32 d0, d16, d19
- 0xf2805010, //vmov.i32 d5, #0
- 0xf2806010, //vmov.i32 d6, #0
- 0xf2807010, //vmov.i32 d7, #0
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_constant_color_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xf4630a0f, //vld1.8 {d16-d17}, [r3]
- 0xf3b40c20, //vdup.32 d0, d16[0]
- 0xf3bc1c20, //vdup.32 d1, d16[1]
- 0xf3b42c21, //vdup.32 d2, d17[0]
- 0xf3bc3c21, //vdup.32 d3, d17[1]
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_clear_vfp4[] = {
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2800010, //vmov.i32 d0, #0
- 0xf2801010, //vmov.i32 d1, #0
- 0xf2802010, //vmov.i32 d2, #0
- 0xf2803010, //vmov.i32 d3, #0
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_plus__vfp4[] = {
- 0xf2000d04, //vadd.f32 d0, d0, d4
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2011d05, //vadd.f32 d1, d1, d5
- 0xf2022d06, //vadd.f32 d2, d2, d6
- 0xf2033d07, //vadd.f32 d3, d3, d7
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_srcover_vfp4[] = {
- 0xf2c70f10, //vmov.f32 d16, #1
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2600d83, //vsub.f32 d16, d16, d3
- 0xf2040c30, //vfma.f32 d0, d4, d16
- 0xf2051c30, //vfma.f32 d1, d5, d16
- 0xf2062c30, //vfma.f32 d2, d6, d16
- 0xf2073c30, //vfma.f32 d3, d7, d16
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_dstover_vfp4[] = {
- 0xf2c70f10, //vmov.f32 d16, #1
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2651115, //vorr d17, d5, d5
- 0xf2604d87, //vsub.f32 d20, d16, d7
- 0xf2640114, //vorr d16, d4, d4
- 0xf2662116, //vorr d18, d6, d6
- 0xf2673117, //vorr d19, d7, d7
- 0xf2400c34, //vfma.f32 d16, d0, d20
- 0xf2411c34, //vfma.f32 d17, d1, d20
- 0xf2422c34, //vfma.f32 d18, d2, d20
- 0xf2433c34, //vfma.f32 d19, d3, d20
- 0xf22001b0, //vorr d0, d16, d16
- 0xf22111b1, //vorr d1, d17, d17
- 0xf22221b2, //vorr d2, d18, d18
- 0xf22331b3, //vorr d3, d19, d19
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_clamp_0_vfp4[] = {
- 0xf2c00010, //vmov.i32 d16, #0
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2000f20, //vmax.f32 d0, d0, d16
- 0xf2011f20, //vmax.f32 d1, d1, d16
- 0xf2022f20, //vmax.f32 d2, d2, d16
- 0xf2033f20, //vmax.f32 d3, d3, d16
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_clamp_1_vfp4[] = {
- 0xf2c70f10, //vmov.f32 d16, #1
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2200f20, //vmin.f32 d0, d0, d16
- 0xf2211f20, //vmin.f32 d1, d1, d16
- 0xf2222f20, //vmin.f32 d2, d2, d16
- 0xf2233f20, //vmin.f32 d3, d3, d16
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_clamp_a_vfp4[] = {
- 0xf2c70f10, //vmov.f32 d16, #1
- 0xe4913004, //ldr r3, [r1], #4
- 0xf2233f20, //vmin.f32 d3, d3, d16
- 0xf2200f03, //vmin.f32 d0, d0, d3
- 0xf2211f03, //vmin.f32 d1, d1, d3
- 0xf2222f03, //vmin.f32 d2, d2, d3
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_set_rgb_vfp4[] = {
- 0xe92d4800, //push {fp, lr}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xe283e008, //add lr, r3, #8
- 0xf4a30c9d, //vld1.32 {d0[]}, [r3 :32]!
- 0xf4ae2c9f, //vld1.32 {d2[]}, [lr :32]
- 0xf4a31c9f, //vld1.32 {d1[]}, [r3 :32]
- 0xe8bd4800, //pop {fp, lr}
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_swap_rb_vfp4[] = {
- 0xeef00b40, //vmov.f64 d16, d0
- 0xe4913004, //ldr r3, [r1], #4
- 0xeeb00b42, //vmov.f64 d0, d2
- 0xeeb02b60, //vmov.f64 d2, d16
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_swap_vfp4[] = {
- 0xeef00b43, //vmov.f64 d16, d3
- 0xe4913004, //ldr r3, [r1], #4
- 0xeef01b42, //vmov.f64 d17, d2
- 0xeef02b41, //vmov.f64 d18, d1
- 0xeef03b40, //vmov.f64 d19, d0
- 0xeeb00b44, //vmov.f64 d0, d4
- 0xeeb01b45, //vmov.f64 d1, d5
- 0xeeb02b46, //vmov.f64 d2, d6
- 0xeeb03b47, //vmov.f64 d3, d7
- 0xeeb04b63, //vmov.f64 d4, d19
- 0xeeb05b62, //vmov.f64 d5, d18
- 0xeeb06b61, //vmov.f64 d6, d17
- 0xeeb07b60, //vmov.f64 d7, d16
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_move_src_dst_vfp4[] = {
- 0xeeb04b40, //vmov.f64 d4, d0
- 0xe4913004, //ldr r3, [r1], #4
- 0xeeb05b41, //vmov.f64 d5, d1
- 0xeeb06b42, //vmov.f64 d6, d2
- 0xeeb07b43, //vmov.f64 d7, d3
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_move_dst_src_vfp4[] = {
- 0xeeb00b44, //vmov.f64 d0, d4
- 0xe4913004, //ldr r3, [r1], #4
- 0xeeb01b45, //vmov.f64 d1, d5
- 0xeeb02b46, //vmov.f64 d2, d6
- 0xeeb03b47, //vmov.f64 d3, d7
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_premul_vfp4[] = {
- 0xf3000d13, //vmul.f32 d0, d0, d3
- 0xe4913004, //ldr r3, [r1], #4
- 0xf3011d13, //vmul.f32 d1, d1, d3
- 0xf3022d13, //vmul.f32 d2, d2, d3
- 0xe12fff13, //bx r3
-};
-
-CODE const uint32_t sk_unpremul_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xeeb78a00, //vmov.f32 s16, #112
- 0xf3f91503, //vceq.f32 d17, d3, #0
- 0xf2c00010, //vmov.i32 d16, #0
- 0xe4913004, //ldr r3, [r1], #4
- 0xeec89a23, //vdiv.f32 s19, s16, s7
- 0xee889a03, //vdiv.f32 s18, s16, s6
- 0xf3501199, //vbsl d17, d16, d9
- 0xf3010d90, //vmul.f32 d0, d17, d0
- 0xf3011d91, //vmul.f32 d1, d17, d1
- 0xf3012d92, //vmul.f32 d2, d17, d2
- 0xecbd8b04, //vpop {d8-d9}
- 0xe12fff13, //bx r3
- 0xe320f000, //nop {0}
-};
-
-CODE const uint32_t sk_from_srgb_vfp4[] = {
- 0xeddf3b20, //vldr d19, [pc, #128]
- 0xf3408d10, //vmul.f32 d24, d0, d0
- 0xeddf0b1c, //vldr d16, [pc, #112]
- 0xf26341b3, //vorr d20, d19, d19
- 0xf26351b3, //vorr d21, d19, d19
- 0xeddf9b1f, //vldr d25, [pc, #124]
- 0xf2404c30, //vfma.f32 d20, d0, d16
- 0xeddf2b1b, //vldr d18, [pc, #108]
- 0xf2415c30, //vfma.f32 d21, d1, d16
- 0xeddfcb1d, //vldr d28, [pc, #116]
- 0xf2423c30, //vfma.f32 d19, d2, d16
- 0xe4913004, //ldr r3, [r1], #4
- 0xf3426d12, //vmul.f32 d22, d2, d2
- 0xf3417d11, //vmul.f32 d23, d1, d1
- 0xf3620e80, //vcgt.f32 d16, d18, d0
- 0xf3621e81, //vcgt.f32 d17, d18, d1
- 0xf341ad39, //vmul.f32 d26, d1, d25
- 0xf342bd39, //vmul.f32 d27, d2, d25
- 0xf3622e82, //vcgt.f32 d18, d18, d2
- 0xf3409d39, //vmul.f32 d25, d0, d25
- 0xf26cd1bc, //vorr d29, d28, d28
- 0xf248dcb4, //vfma.f32 d29, d24, d20
- 0xf26c41bc, //vorr d20, d28, d28
- 0xf2474cb5, //vfma.f32 d20, d23, d21
- 0xf246ccb3, //vfma.f32 d28, d22, d19
- 0xf35901bd, //vbsl d16, d25, d29
- 0xf35a11b4, //vbsl d17, d26, d20
- 0xf35b21bc, //vbsl d18, d27, d28
- 0xf22001b0, //vorr d0, d16, d16
- 0xf22111b1, //vorr d1, d17, d17
- 0xf22221b2, //vorr d2, d18, d18
- 0xe12fff13, //bx r3
- 0x3e99999a, //.word 0x3e99999a
- 0x3e99999a, //.word 0x3e99999a
- 0x3f328f5c, //.word 0x3f328f5c
- 0x3f328f5c, //.word 0x3f328f5c
- 0x3d6147ae, //.word 0x3d6147ae
- 0x3d6147ae, //.word 0x3d6147ae
- 0x3d9e8391, //.word 0x3d9e8391
- 0x3d9e8391, //.word 0x3d9e8391
- 0x3b23d70a, //.word 0x3b23d70a
- 0x3b23d70a, //.word 0x3b23d70a
-};
-
-CODE const uint32_t sk_to_srgb_vfp4[] = {
- 0xf3fb0582, //vrsqrte.f32 d16, d2
- 0xe4913004, //ldr r3, [r1], #4
- 0xf3fb1581, //vrsqrte.f32 d17, d1
- 0xf3fb2580, //vrsqrte.f32 d18, d0
- 0xf3403db0, //vmul.f32 d19, d16, d16
- 0xf3414db1, //vmul.f32 d20, d17, d17
- 0xf3425db2, //vmul.f32 d21, d18, d18
- 0xf2623f33, //vrsqrts.f32 d19, d2, d19
- 0xf2614f34, //vrsqrts.f32 d20, d1, d20
- 0xf2605f35, //vrsqrts.f32 d21, d0, d21
- 0xf3400db3, //vmul.f32 d16, d16, d19
- 0xf3411db4, //vmul.f32 d17, d17, d20
- 0xf3422db5, //vmul.f32 d18, d18, d21
- 0xf3fb3520, //vrecpe.f32 d19, d16
- 0xf3fb4521, //vrecpe.f32 d20, d17
- 0xf3fb6522, //vrecpe.f32 d22, d18
- 0xf3fb55a0, //vrsqrte.f32 d21, d16
- 0xf3fb75a1, //vrsqrte.f32 d23, d17
- 0xf3fb85a2, //vrsqrte.f32 d24, d18
- 0xf2409fb3, //vrecps.f32 d25, d16, d19
- 0xf241afb4, //vrecps.f32 d26, d17, d20
- 0xf242bfb6, //vrecps.f32 d27, d18, d22
- 0xf345cdb5, //vmul.f32 d28, d21, d21
- 0xf347ddb7, //vmul.f32 d29, d23, d23
- 0xf348edb8, //vmul.f32 d30, d24, d24
- 0xf2600fbc, //vrsqrts.f32 d16, d16, d28
- 0xf2611fbd, //vrsqrts.f32 d17, d17, d29
- 0xf2622fbe, //vrsqrts.f32 d18, d18, d30
- 0xf3433db9, //vmul.f32 d19, d19, d25
- 0xeddf9b21, //vldr d25, [pc, #132]
- 0xf3444dba, //vmul.f32 d20, d20, d26
- 0xeddfab21, //vldr d26, [pc, #132]
- 0xf3466dbb, //vmul.f32 d22, d22, d27
- 0xf26ab1ba, //vorr d27, d26, d26
- 0xf243bcb9, //vfma.f32 d27, d19, d25
- 0xf26a31ba, //vorr d19, d26, d26
- 0xf2443cb9, //vfma.f32 d19, d20, d25
- 0xeddf4b1d, //vldr d20, [pc, #116]
- 0xf246acb9, //vfma.f32 d26, d22, d25
- 0xf3450db0, //vmul.f32 d16, d21, d16
- 0xeddf5b1c, //vldr d21, [pc, #112]
- 0xf3471db1, //vmul.f32 d17, d23, d17
- 0xf3482db2, //vmul.f32 d18, d24, d18
- 0xf3406d35, //vmul.f32 d22, d0, d21
- 0xf240bcb4, //vfma.f32 d27, d16, d20
- 0xf2413cb4, //vfma.f32 d19, d17, d20
- 0xf242acb4, //vfma.f32 d26, d18, d20
- 0xeddf2b17, //vldr d18, [pc, #92]
- 0xf3417d35, //vmul.f32 d23, d1, d21
- 0xf3620e80, //vcgt.f32 d16, d18, d0
- 0xf3621e81, //vcgt.f32 d17, d18, d1
- 0xf3622e82, //vcgt.f32 d18, d18, d2
- 0xf3425d35, //vmul.f32 d21, d2, d21
- 0xf2c74f10, //vmov.f32 d20, #1
- 0xf2648faa, //vmin.f32 d24, d20, d26
- 0xf2643fa3, //vmin.f32 d19, d20, d19
- 0xf2644fab, //vmin.f32 d20, d20, d27
- 0xf35601b8, //vbsl d16, d22, d24
- 0xf35711b3, //vbsl d17, d23, d19
- 0xf35521b4, //vbsl d18, d21, d20
- 0xf22001b0, //vorr d0, d16, d16
- 0xf22111b1, //vorr d1, d17, d17
- 0xf22221b2, //vorr d2, d18, d18
- 0xe12fff13, //bx r3
- 0x3f306fce, //.word 0x3f306fce
- 0x3f306fce, //.word 0x3f306fce
- 0xbdca57a8, //.word 0xbdca57a8
- 0xbdca57a8, //.word 0xbdca57a8
- 0x3ed287c2, //.word 0x3ed287c2
- 0x3ed287c2, //.word 0x3ed287c2
- 0x41475c29, //.word 0x41475c29
- 0x41475c29, //.word 0x41475c29
- 0x3b8ce704, //.word 0x3b8ce704
- 0x3b8ce704, //.word 0x3b8ce704
-};
-
-CODE const uint32_t sk_scale_1_float_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xf3000d90, //vmul.f32 d0, d16, d0
- 0xf3001d91, //vmul.f32 d1, d16, d1
- 0xf3002d92, //vmul.f32 d2, d16, d2
- 0xf3003d93, //vmul.f32 d3, d16, d3
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_scale_u8_vfp4[] = {
- 0xe24dd004, //sub sp, sp, #4
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xe0833000, //add r3, r3, r0
- 0xe1d330b0, //ldrh r3, [r3]
- 0xe1cd30b0, //strh r3, [sp]
- 0xe1a0300d, //mov r3, sp
- 0xf4e3041f, //vld1.16 {d16[0]}, [r3 :16]
- 0xf3c80a30, //vmovl.u8 q8, d16
- 0xf3d00a30, //vmovl.u16 q8, d16
- 0xf3fb06a0, //vcvt.f32.u32 d16, d16
- 0xeddf1b06, //vldr d17, [pc, #24]
- 0xf3400db1, //vmul.f32 d16, d16, d17
- 0xf3000d90, //vmul.f32 d0, d16, d0
- 0xf3001d91, //vmul.f32 d1, d16, d1
- 0xf3002d92, //vmul.f32 d2, d16, d2
- 0xf3003d93, //vmul.f32 d3, d16, d3
- 0xe28dd004, //add sp, sp, #4
- 0xe12fff1c, //bx ip
- 0x3b808081, //.word 0x3b808081
- 0x3b808081, //.word 0x3b808081
-};
-
-CODE const uint32_t sk_lerp_1_float_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2600d04, //vsub.f32 d16, d0, d4
- 0xf2611d05, //vsub.f32 d17, d1, d5
- 0xf2622d06, //vsub.f32 d18, d2, d6
- 0xe2811008, //add r1, r1, #8
- 0xf2633d07, //vsub.f32 d19, d3, d7
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xf2240114, //vorr d0, d4, d4
- 0xf2251115, //vorr d1, d5, d5
- 0xf2262116, //vorr d2, d6, d6
- 0xf2273117, //vorr d3, d7, d7
- 0xf2000cb4, //vfma.f32 d0, d16, d20
- 0xf2011cb4, //vfma.f32 d1, d17, d20
- 0xf2022cb4, //vfma.f32 d2, d18, d20
- 0xf2033cb4, //vfma.f32 d3, d19, d20
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_lerp_u8_vfp4[] = {
- 0xe24dd004, //sub sp, sp, #4
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2602d04, //vsub.f32 d18, d0, d4
- 0xf2623d06, //vsub.f32 d19, d2, d6
- 0xf2634d07, //vsub.f32 d20, d3, d7
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xf2240114, //vorr d0, d4, d4
- 0xf2262116, //vorr d2, d6, d6
- 0xe0833000, //add r3, r3, r0
- 0xf2273117, //vorr d3, d7, d7
- 0xe1d330b0, //ldrh r3, [r3]
- 0xe1cd30b0, //strh r3, [sp]
- 0xe1a0300d, //mov r3, sp
- 0xf4e3041f, //vld1.16 {d16[0]}, [r3 :16]
- 0xf3c80a30, //vmovl.u8 q8, d16
- 0xf3d00a30, //vmovl.u16 q8, d16
- 0xf3fb06a0, //vcvt.f32.u32 d16, d16
- 0xeddf1b08, //vldr d17, [pc, #32]
- 0xf3400db1, //vmul.f32 d16, d16, d17
- 0xf2611d05, //vsub.f32 d17, d1, d5
- 0xf2251115, //vorr d1, d5, d5
- 0xf2020cb0, //vfma.f32 d0, d18, d16
- 0xf2011cb0, //vfma.f32 d1, d17, d16
- 0xf2032cb0, //vfma.f32 d2, d19, d16
- 0xf2043cb0, //vfma.f32 d3, d20, d16
- 0xe28dd004, //add sp, sp, #4
- 0xe12fff1c, //bx ip
- 0x3b808081, //.word 0x3b808081
- 0x3b808081, //.word 0x3b808081
-};
-
-CODE const uint32_t sk_lerp_565_vfp4[] = {
- 0xe24dd004, //sub sp, sp, #4
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf3c72218, //vmov.i32 d18, #63488
- 0xf2c1101f, //vmov.i32 d17, #31
- 0xf2603d04, //vsub.f32 d19, d0, d4
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xf2616d05, //vsub.f32 d22, d1, d5
- 0xf2240114, //vorr d0, d4, d4
- 0xf2251115, //vorr d1, d5, d5
- 0xe7933080, //ldr r3, [r3, r0, lsl #1]
- 0xf2873f10, //vmov.f32 d3, #1
- 0xe58d3000, //str r3, [sp]
- 0xe1a0300d, //mov r3, sp
- 0xf4e3083f, //vld1.32 {d16[0]}, [r3 :32]
- 0xe3a03e7e, //mov r3, #2016
- 0xf3d04a30, //vmovl.u16 q10, d16
- 0xee803b90, //vdup.32 d16, r3
- 0xf24421b2, //vand d18, d20, d18
- 0xf24411b1, //vand d17, d20, d17
- 0xeddf5b12, //vldr d21, [pc, #72]
- 0xf24401b0, //vand d16, d20, d16
- 0xeddf4b0e, //vldr d20, [pc, #56]
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xf3422db4, //vmul.f32 d18, d18, d20
- 0xeddf4b0d, //vldr d20, [pc, #52]
- 0xf3400db5, //vmul.f32 d16, d16, d21
- 0xf2625d06, //vsub.f32 d21, d2, d6
- 0xf3411db4, //vmul.f32 d17, d17, d20
- 0xf2262116, //vorr d2, d6, d6
- 0xf2030cb2, //vfma.f32 d0, d19, d18
- 0xf2061cb0, //vfma.f32 d1, d22, d16
- 0xf2052cb1, //vfma.f32 d2, d21, d17
- 0xe28dd004, //add sp, sp, #4
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x37842108, //.word 0x37842108
- 0x37842108, //.word 0x37842108
- 0x3a020821, //.word 0x3a020821
- 0x3a020821, //.word 0x3a020821
- 0x3d042108, //.word 0x3d042108
- 0x3d042108, //.word 0x3d042108
-};
-
-CODE const uint32_t sk_load_tables_vfp4[] = {
- 0xe92d48f0, //push {r4, r5, r6, r7, fp, lr}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf3c7001f, //vmov.i32 d16, #255
- 0xe2811008, //add r1, r1, #8
- 0xe593e000, //ldr lr, [r3]
- 0xe99300b0, //ldmib r3, {r4, r5, r7}
- 0xe08e3100, //add r3, lr, r0, lsl #2
- 0xedd31b00, //vldr d17, [r3]
- 0xf24121b0, //vand d18, d17, d16
- 0xf3f83031, //vshr.u32 d19, d17, #8
- 0xee323b90, //vmov.32 r3, d18[1]
- 0xee126b90, //vmov.32 r6, d18[0]
- 0xf3f02031, //vshr.u32 d18, d17, #16
- 0xf24221b0, //vand d18, d18, d16
- 0xf24301b0, //vand d16, d19, d16
- 0xe0843103, //add r3, r4, r3, lsl #2
- 0xedd30a00, //vldr s1, [r3]
- 0xe0843106, //add r3, r4, r6, lsl #2
- 0xee326b90, //vmov.32 r6, d18[1]
- 0xed930a00, //vldr s0, [r3]
- 0xee303b90, //vmov.32 r3, d16[1]
- 0xee104b90, //vmov.32 r4, d16[0]
- 0xf3e80031, //vshr.u32 d16, d17, #24
- 0xeddf1b0d, //vldr d17, [pc, #52]
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3003db1, //vmul.f32 d3, d16, d17
- 0xe087e106, //add lr, r7, r6, lsl #2
- 0xee126b90, //vmov.32 r6, d18[0]
- 0xe0853103, //add r3, r5, r3, lsl #2
- 0xedde2a00, //vldr s5, [lr]
- 0xedd31a00, //vldr s3, [r3]
- 0xe0853104, //add r3, r5, r4, lsl #2
- 0xed931a00, //vldr s2, [r3]
- 0xe0873106, //add r3, r7, r6, lsl #2
- 0xed932a00, //vldr s4, [r3]
- 0xe8bd48f0, //pop {r4, r5, r6, r7, fp, lr}
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3b808081, //.word 0x3b808081
- 0x3b808081, //.word 0x3b808081
-};
-
-CODE const uint32_t sk_load_a8_vfp4[] = {
- 0xe24dd004, //sub sp, sp, #4
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xf2800010, //vmov.i32 d0, #0
- 0xf2801010, //vmov.i32 d1, #0
- 0xe5933000, //ldr r3, [r3]
- 0xf2802010, //vmov.i32 d2, #0
- 0xe0833000, //add r3, r3, r0
- 0xe1d330b0, //ldrh r3, [r3]
- 0xe1cd30b0, //strh r3, [sp]
- 0xe1a0300d, //mov r3, sp
- 0xf4e3041f, //vld1.16 {d16[0]}, [r3 :16]
- 0xf3c80a30, //vmovl.u8 q8, d16
- 0xf3d00a30, //vmovl.u16 q8, d16
- 0xf3fb06a0, //vcvt.f32.u32 d16, d16
- 0xeddf1b03, //vldr d17, [pc, #12]
- 0xf3003db1, //vmul.f32 d3, d16, d17
- 0xe28dd004, //add sp, sp, #4
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3b808081, //.word 0x3b808081
- 0x3b808081, //.word 0x3b808081
-};
-
-CODE const uint32_t sk_store_a8_vfp4[] = {
- 0xe92d4800, //push {fp, lr}
- 0xeddf0b0d, //vldr d16, [pc, #52]
- 0xf2c3161f, //vmov.i32 d17, #1056964608
- 0xf2431c30, //vfma.f32 d17, d3, d16
- 0xe5913000, //ldr r3, [r1]
- 0xe5933000, //ldr r3, [r3]
- 0xf3fb07a1, //vcvt.u32.f32 d16, d17
- 0xee10eb90, //vmov.32 lr, d16[0]
- 0xee30cb90, //vmov.32 ip, d16[1]
- 0xe7e3e000, //strb lr, [r3, r0]!
- 0xe5c3c001, //strb ip, [r3, #1]
- 0xe2813008, //add r3, r1, #8
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe1a01003, //mov r1, r3
- 0xe8bd4800, //pop {fp, lr}
- 0xe12fff1c, //bx ip
- 0x437f0000, //.word 0x437f0000
- 0x437f0000, //.word 0x437f0000
-};
-
-CODE const uint32_t sk_load_565_vfp4[] = {
- 0xe24dd004, //sub sp, sp, #4
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c1101f, //vmov.i32 d17, #31
- 0xf3c72218, //vmov.i32 d18, #63488
- 0xeddf3b16, //vldr d19, [pc, #88]
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xf2873f10, //vmov.f32 d3, #1
- 0xe7933080, //ldr r3, [r3, r0, lsl #1]
- 0xe58d3000, //str r3, [sp]
- 0xe1a0300d, //mov r3, sp
- 0xf4e3083f, //vld1.32 {d16[0]}, [r3 :32]
- 0xe3a03e7e, //mov r3, #2016
- 0xf3d04a30, //vmovl.u16 q10, d16
- 0xee803b90, //vdup.32 d16, r3
- 0xf24411b1, //vand d17, d20, d17
- 0xeddf5b0e, //vldr d21, [pc, #56]
- 0xf24421b2, //vand d18, d20, d18
- 0xf24401b0, //vand d16, d20, d16
- 0xeddf4b09, //vldr d20, [pc, #36]
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xf3020db3, //vmul.f32 d0, d18, d19
- 0xf3001db4, //vmul.f32 d1, d16, d20
- 0xf3012db5, //vmul.f32 d2, d17, d21
- 0xe28dd004, //add sp, sp, #4
- 0xe12fff1c, //bx ip
- 0x37842108, //.word 0x37842108
- 0x37842108, //.word 0x37842108
- 0x3a020821, //.word 0x3a020821
- 0x3a020821, //.word 0x3a020821
- 0x3d042108, //.word 0x3d042108
- 0x3d042108, //.word 0x3d042108
-};
-
-CODE const uint32_t sk_store_565_vfp4[] = {
- 0xf2c30f1f, //vmov.f32 d16, #31
- 0xeddf1b15, //vldr d17, [pc, #84]
- 0xf2c3361f, //vmov.i32 d19, #1056964608
- 0xe5913000, //ldr r3, [r1]
- 0xf2413c31, //vfma.f32 d19, d1, d17
- 0xf2c3161f, //vmov.i32 d17, #1056964608
- 0xf2401c30, //vfma.f32 d17, d0, d16
- 0xe5933000, //ldr r3, [r3]
- 0xf2c3261f, //vmov.i32 d18, #1056964608
- 0xf2422c30, //vfma.f32 d18, d2, d16
- 0xe0833080, //add r3, r3, r0, lsl #1
- 0xf3fb07a3, //vcvt.u32.f32 d16, d19
- 0xf3fb17a1, //vcvt.u32.f32 d17, d17
- 0xf3fb27a2, //vcvt.u32.f32 d18, d18
- 0xf2e50530, //vshl.s32 d16, d16, #5
- 0xf2eb1531, //vshl.s32 d17, d17, #11
- 0xf26001b1, //vorr d16, d16, d17
- 0xf26001b2, //vorr d16, d16, d18
- 0xf3f60121, //vuzp.16 d16, d17
- 0xf4c3080f, //vst1.32 {d16[0]}, [r3]
- 0xe2813008, //add r3, r1, #8
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe1a01003, //mov r1, r3
- 0xe12fff1c, //bx ip
- 0x427c0000, //.word 0x427c0000
- 0x427c0000, //.word 0x427c0000
-};
-
-CODE const uint32_t sk_load_8888_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf3c7001f, //vmov.i32 d16, #255
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xe0833100, //add r3, r3, r0, lsl #2
- 0xedd31b00, //vldr d17, [r3]
- 0xf24121b0, //vand d18, d17, d16
- 0xf3f83031, //vshr.u32 d19, d17, #8
- 0xf3e84031, //vshr.u32 d20, d17, #24
- 0xf3f01031, //vshr.u32 d17, d17, #16
- 0xf24331b0, //vand d19, d19, d16
- 0xf24101b0, //vand d16, d17, d16
- 0xeddf1b08, //vldr d17, [pc, #32]
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xf3fb4624, //vcvt.f32.s32 d20, d20
- 0xf3fb3623, //vcvt.f32.s32 d19, d19
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3020db1, //vmul.f32 d0, d18, d17
- 0xf3043db1, //vmul.f32 d3, d20, d17
- 0xf3031db1, //vmul.f32 d1, d19, d17
- 0xf3002db1, //vmul.f32 d2, d16, d17
- 0xe12fff1c, //bx ip
- 0x3b808081, //.word 0x3b808081
- 0x3b808081, //.word 0x3b808081
-};
-
-CODE const uint32_t sk_store_8888_vfp4[] = {
- 0xeddf0b1a, //vldr d16, [pc, #104]
- 0xf2c3261f, //vmov.i32 d18, #1056964608
- 0xf2412c30, //vfma.f32 d18, d1, d16
- 0xe5913000, //ldr r3, [r1]
- 0xf2c3361f, //vmov.i32 d19, #1056964608
- 0xf2c3161f, //vmov.i32 d17, #1056964608
- 0xf2423c30, //vfma.f32 d19, d2, d16
- 0xe5933000, //ldr r3, [r3]
- 0xf2c3461f, //vmov.i32 d20, #1056964608
- 0xf2401c30, //vfma.f32 d17, d0, d16
- 0xe0833100, //add r3, r3, r0, lsl #2
- 0xf2434c30, //vfma.f32 d20, d3, d16
- 0xf3fb07a2, //vcvt.u32.f32 d16, d18
- 0xf3fb27a3, //vcvt.u32.f32 d18, d19
- 0xf3fb17a1, //vcvt.u32.f32 d17, d17
- 0xf3fb37a4, //vcvt.u32.f32 d19, d20
- 0xf2e80530, //vshl.s32 d16, d16, #8
- 0xf2f02532, //vshl.s32 d18, d18, #16
- 0xf26001b1, //vorr d16, d16, d17
- 0xf2f81533, //vshl.s32 d17, d19, #24
- 0xf26001b2, //vorr d16, d16, d18
- 0xf26001b1, //vorr d16, d16, d17
- 0xedc30b00, //vstr d16, [r3]
- 0xe2813008, //add r3, r1, #8
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe1a01003, //mov r1, r3
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x437f0000, //.word 0x437f0000
- 0x437f0000, //.word 0x437f0000
-};
-
-CODE const uint32_t sk_load_f16_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xe5933000, //ldr r3, [r3]
- 0xe0833180, //add r3, r3, r0, lsl #3
- 0xf463084f, //vld2.16 {d16-d17}, [r3]
- 0xf3f62720, //vcvt.f32.f16 q9, d16
- 0xf3f60721, //vcvt.f32.f16 q8, d17
- 0xf22201b2, //vorr d0, d18, d18
- 0xf22011b0, //vorr d1, d16, d16
- 0xf3ba00a3, //vtrn.32 d0, d19
- 0xf22321b3, //vorr d2, d19, d19
- 0xf3ba10a1, //vtrn.32 d1, d17
- 0xf22131b1, //vorr d3, d17, d17
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_store_f16_vfp4[] = {
- 0xeef00b41, //vmov.f64 d16, d1
- 0xeef03b42, //vmov.f64 d19, d2
- 0xf2631113, //vorr d17, d3, d3
- 0xf2602110, //vorr d18, d0, d0
- 0xf3fa00a1, //vtrn.32 d16, d17
- 0xf3f61620, //vcvt.f16.f32 d17, q8
- 0xf3fa20a3, //vtrn.32 d18, d19
- 0xe5913000, //ldr r3, [r1]
- 0xf3f60622, //vcvt.f16.f32 d16, q9
- 0xe5933000, //ldr r3, [r3]
- 0xe0833180, //add r3, r3, r0, lsl #3
- 0xf443084f, //vst2.16 {d16-d17}, [r3]
- 0xe2813008, //add r3, r1, #8
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe1a01003, //mov r1, r3
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_store_f32_vfp4[] = {
- 0xe5913000, //ldr r3, [r1]
- 0xe5933000, //ldr r3, [r3]
- 0xe0833200, //add r3, r3, r0, lsl #4
- 0xf403008f, //vst4.32 {d0-d3}, [r3]
- 0xe2813008, //add r3, r1, #8
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe1a01003, //mov r1, r3
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_clamp_x_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c00010, //vmov.i32 d16, #0
- 0xf3c71e1f, //vmov.i8 d17, #255
- 0xf2400f80, //vmax.f32 d16, d16, d0
- 0xe2811008, //add r1, r1, #8
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf26218a1, //vadd.i32 d17, d18, d17
- 0xf2200fa1, //vmin.f32 d0, d16, d17
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_clamp_y_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c00010, //vmov.i32 d16, #0
- 0xf3c71e1f, //vmov.i8 d17, #255
- 0xf2400f81, //vmax.f32 d16, d16, d1
- 0xe2811008, //add r1, r1, #8
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf26218a1, //vadd.i32 d17, d18, d17
- 0xf2201fa1, //vmin.f32 d1, d16, d17
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_repeat_x_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c02010, //vmov.i32 d18, #0
- 0xe2811008, //add r1, r1, #8
- 0xeddf3b10, //vldr d19, [pc, #64]
- 0xed938a00, //vldr s16, [r3]
- 0xeec09a88, //vdiv.f32 s19, s1, s16
- 0xee809a08, //vdiv.f32 s18, s0, s16
- 0xf3fb0709, //vcvt.s32.f32 d16, d9
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3601e89, //vcgt.f32 d17, d16, d9
- 0xf35311b2, //vbsl d17, d19, d18
- 0xf3f42c08, //vdup.32 d18, d8[0]
- 0xf2600da1, //vsub.f32 d16, d16, d17
- 0xf3c71e1f, //vmov.i8 d17, #255
- 0xf26218a1, //vadd.i32 d17, d18, d17
- 0xf2e009c8, //vmul.f32 d16, d16, d8[0]
- 0xf2600d20, //vsub.f32 d16, d0, d16
- 0xf2200fa1, //vmin.f32 d0, d16, d17
- 0xecbd8b04, //vpop {d8-d9}
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3f800000, //.word 0x3f800000
- 0x3f800000, //.word 0x3f800000
-};
-
-CODE const uint32_t sk_repeat_y_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c02010, //vmov.i32 d18, #0
- 0xe2811008, //add r1, r1, #8
- 0xeddf3b10, //vldr d19, [pc, #64]
- 0xed938a00, //vldr s16, [r3]
- 0xeec19a88, //vdiv.f32 s19, s3, s16
- 0xee819a08, //vdiv.f32 s18, s2, s16
- 0xf3fb0709, //vcvt.s32.f32 d16, d9
- 0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xf3601e89, //vcgt.f32 d17, d16, d9
- 0xf35311b2, //vbsl d17, d19, d18
- 0xf3f42c08, //vdup.32 d18, d8[0]
- 0xf2600da1, //vsub.f32 d16, d16, d17
- 0xf3c71e1f, //vmov.i8 d17, #255
- 0xf26218a1, //vadd.i32 d17, d18, d17
- 0xf2e009c8, //vmul.f32 d16, d16, d8[0]
- 0xf2610d20, //vsub.f32 d16, d1, d16
- 0xf2201fa1, //vmin.f32 d1, d16, d17
- 0xecbd8b04, //vpop {d8-d9}
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3f800000, //.word 0x3f800000
- 0x3f800000, //.word 0x3f800000
-};
-
-CODE const uint32_t sk_mirror_x_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c03010, //vmov.i32 d19, #0
- 0xe2811008, //add r1, r1, #8
- 0xeddf4b14, //vldr d20, [pc, #80]
- 0xed938a00, //vldr s16, [r3]
- 0xee389a08, //vadd.f32 s18, s16, s16
- 0xf3f40c08, //vdup.32 d16, d8[0]
- 0xf2200d20, //vsub.f32 d0, d0, d16
- 0xeec08a89, //vdiv.f32 s17, s1, s18
- 0xee808a09, //vdiv.f32 s16, s0, s18
- 0xf3fb1708, //vcvt.s32.f32 d17, d8
- 0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xf3612e88, //vcgt.f32 d18, d17, d8
- 0xf35421b3, //vbsl d18, d20, d19
- 0xf2611da2, //vsub.f32 d17, d17, d18
- 0xf3c72e1f, //vmov.i8 d18, #255
- 0xf2e119c9, //vmul.f32 d17, d17, d9[0]
- 0xf2601d21, //vsub.f32 d17, d0, d17
- 0xf2611da0, //vsub.f32 d17, d17, d16
- 0xf26008a2, //vadd.i32 d16, d16, d18
- 0xf3f91721, //vabs.f32 d17, d17
- 0xf2210fa0, //vmin.f32 d0, d17, d16
- 0xecbd8b04, //vpop {d8-d9}
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3f800000, //.word 0x3f800000
- 0x3f800000, //.word 0x3f800000
-};
-
-CODE const uint32_t sk_mirror_y_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2c03010, //vmov.i32 d19, #0
- 0xe2811008, //add r1, r1, #8
- 0xeddf4b14, //vldr d20, [pc, #80]
- 0xed938a00, //vldr s16, [r3]
- 0xee389a08, //vadd.f32 s18, s16, s16
- 0xf3f40c08, //vdup.32 d16, d8[0]
- 0xf2211d20, //vsub.f32 d1, d1, d16
- 0xeec18a89, //vdiv.f32 s17, s3, s18
- 0xee818a09, //vdiv.f32 s16, s2, s18
- 0xf3fb1708, //vcvt.s32.f32 d17, d8
- 0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xf3612e88, //vcgt.f32 d18, d17, d8
- 0xf35421b3, //vbsl d18, d20, d19
- 0xf2611da2, //vsub.f32 d17, d17, d18
- 0xf3c72e1f, //vmov.i8 d18, #255
- 0xf2e119c9, //vmul.f32 d17, d17, d9[0]
- 0xf2611d21, //vsub.f32 d17, d1, d17
- 0xf2611da0, //vsub.f32 d17, d17, d16
- 0xf26008a2, //vadd.i32 d16, d16, d18
- 0xf3f91721, //vabs.f32 d17, d17
- 0xf2211fa0, //vmin.f32 d1, d17, d16
- 0xecbd8b04, //vpop {d8-d9}
- 0xe12fff1c, //bx ip
- 0xe320f000, //nop {0}
- 0x3f800000, //.word 0x3f800000
- 0x3f800000, //.word 0x3f800000
-};
-
-CODE const uint32_t sk_luminance_to_alpha_vfp4[] = {
- 0xeddf0b0a, //vldr d16, [pc, #40]
- 0xeddf1b0b, //vldr d17, [pc, #44]
- 0xf3410d30, //vmul.f32 d16, d1, d16
- 0xe4913004, //ldr r3, [r1], #4
- 0xf3401d31, //vmul.f32 d17, d0, d17
- 0xf2800010, //vmov.i32 d0, #0
- 0xf2801010, //vmov.i32 d1, #0
- 0xf2013da0, //vadd.f32 d3, d17, d16
- 0xeddf0b06, //vldr d16, [pc, #24]
- 0xf2023c30, //vfma.f32 d3, d2, d16
- 0xf2802010, //vmov.i32 d2, #0
- 0xe12fff13, //bx r3
- 0x3f371759, //.word 0x3f371759
- 0x3f371759, //.word 0x3f371759
- 0x3e59b3d0, //.word 0x3e59b3d0
- 0x3e59b3d0, //.word 0x3e59b3d0
- 0x3d93dd98, //.word 0x3d93dd98
- 0x3d93dd98, //.word 0x3d93dd98
-};
-
-CODE const uint32_t sk_matrix_2x3_vfp4[] = {
- 0xe92d4800, //push {fp, lr}
- 0xe591e000, //ldr lr, [r1]
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe2811008, //add r1, r1, #8
- 0xe28e300c, //add r3, lr, #12
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xe28e3008, //add r3, lr, #8
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe28e3010, //add r3, lr, #16
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xe28e3014, //add r3, lr, #20
- 0xf2410c31, //vfma.f32 d16, d1, d17
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xf2411c32, //vfma.f32 d17, d1, d18
- 0xf4ee2c9d, //vld1.32 {d18[]}, [lr :32]!
- 0xf4ee3c9f, //vld1.32 {d19[]}, [lr :32]
- 0xf2400c32, //vfma.f32 d16, d0, d18
- 0xf2401c33, //vfma.f32 d17, d0, d19
- 0xf22001b0, //vorr d0, d16, d16
- 0xf22111b1, //vorr d1, d17, d17
- 0xe8bd4800, //pop {fp, lr}
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_matrix_3x4_vfp4[] = {
- 0xe92d4800, //push {fp, lr}
- 0xe591e000, //ldr lr, [r1]
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe2811008, //add r1, r1, #8
- 0xe28e3020, //add r3, lr, #32
- 0xf4e33c9f, //vld1.32 {d19[]}, [r3 :32]
- 0xe28e302c, //add r3, lr, #44
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xe28e301c, //add r3, lr, #28
- 0xf2420c33, //vfma.f32 d16, d2, d19
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xe28e3018, //add r3, lr, #24
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xe28e3024, //add r3, lr, #36
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe28e3028, //add r3, lr, #40
- 0xf2421c32, //vfma.f32 d17, d2, d18
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xe28e3010, //add r3, lr, #16
- 0xf2422c34, //vfma.f32 d18, d2, d20
- 0xf4e33c9f, //vld1.32 {d19[]}, [r3 :32]
- 0xe28e300c, //add r3, lr, #12
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xe28e3014, //add r3, lr, #20
- 0xf2411c34, //vfma.f32 d17, d1, d20
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xf2410c34, //vfma.f32 d16, d1, d20
- 0xe28e3008, //add r3, lr, #8
- 0xf2412c33, //vfma.f32 d18, d1, d19
- 0xf4ee3c9d, //vld1.32 {d19[]}, [lr :32]!
- 0xf4ee4c9f, //vld1.32 {d20[]}, [lr :32]
- 0xf2401c33, //vfma.f32 d17, d0, d19
- 0xf4e33c9f, //vld1.32 {d19[]}, [r3 :32]
- 0xf2400c33, //vfma.f32 d16, d0, d19
- 0xf2402c34, //vfma.f32 d18, d0, d20
- 0xf22101b1, //vorr d0, d17, d17
- 0xf22021b0, //vorr d2, d16, d16
- 0xf22211b2, //vorr d1, d18, d18
- 0xe8bd4800, //pop {fp, lr}
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_matrix_4x5_vfp4[] = {
- 0xe92d4010, //push {r4, lr}
- 0xe8911008, //ldm r1, {r3, ip}
- 0xf2620112, //vorr d16, d2, d2
- 0xe2811008, //add r1, r1, #8
- 0xe2834014, //add r4, r3, #20
- 0xe1a0e003, //mov lr, r3
- 0xf4e45c9f, //vld1.32 {d21[]}, [r4 :32]
- 0xe2834028, //add r4, r3, #40
- 0xf4e46c9f, //vld1.32 {d22[]}, [r4 :32]
- 0xe2834038, //add r4, r3, #56
- 0xf4e47c9f, //vld1.32 {d23[]}, [r4 :32]
- 0xe2834048, //add r4, r3, #72
- 0xf4a42c9f, //vld1.32 {d2[]}, [r4 :32]
- 0xe2834034, //add r4, r3, #52
- 0xf2032c37, //vfma.f32 d2, d3, d23
- 0xf4e48c9f, //vld1.32 {d24[]}, [r4 :32]
- 0xe2834044, //add r4, r3, #68
- 0xf4e41c9f, //vld1.32 {d17[]}, [r4 :32]
- 0xe2834030, //add r4, r3, #48
- 0xf2431c38, //vfma.f32 d17, d3, d24
- 0xf4e49c9f, //vld1.32 {d25[]}, [r4 :32]
- 0xe283403c, //add r4, r3, #60
- 0xf4e43c9f, //vld1.32 {d19[]}, [r4 :32]
- 0xe283404c, //add r4, r3, #76
- 0xf2002cb6, //vfma.f32 d2, d16, d22
- 0xf4e42c9f, //vld1.32 {d18[]}, [r4 :32]
- 0xe2834040, //add r4, r3, #64
- 0xf2432c33, //vfma.f32 d18, d3, d19
- 0xf4e43c9f, //vld1.32 {d19[]}, [r4 :32]
- 0xe2834020, //add r4, r3, #32
- 0xf2433c39, //vfma.f32 d19, d3, d25
- 0xf4e47c9f, //vld1.32 {d23[]}, [r4 :32]
- 0xe283402c, //add r4, r3, #44
- 0xf4e48c9f, //vld1.32 {d24[]}, [r4 :32]
- 0xe2834024, //add r4, r3, #36
- 0xf2402cb8, //vfma.f32 d18, d16, d24
- 0xf4e48c9f, //vld1.32 {d24[]}, [r4 :32]
- 0xf2401cb8, //vfma.f32 d17, d16, d24
- 0xe2834010, //add r4, r3, #16
- 0xf2403cb7, //vfma.f32 d19, d16, d23
- 0xf4ee4c9d, //vld1.32 {d20[]}, [lr :32]!
- 0xf4e40c9f, //vld1.32 {d16[]}, [r4 :32]
- 0xe283401c, //add r4, r3, #28
- 0xf4e46c9f, //vld1.32 {d22[]}, [r4 :32]
- 0xe2834018, //add r4, r3, #24
- 0xf2412c36, //vfma.f32 d18, d1, d22
- 0xf2411c35, //vfma.f32 d17, d1, d21
- 0xf4ee5c9f, //vld1.32 {d21[]}, [lr :32]
- 0xf2413c30, //vfma.f32 d19, d1, d16
- 0xf4e40c9f, //vld1.32 {d16[]}, [r4 :32]
- 0xe2834008, //add r4, r3, #8
- 0xe283300c, //add r3, r3, #12
- 0xf2012c30, //vfma.f32 d2, d1, d16
- 0xf4e40c9f, //vld1.32 {d16[]}, [r4 :32]
- 0xf2401c35, //vfma.f32 d17, d0, d21
- 0xf2403c34, //vfma.f32 d19, d0, d20
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xf2402c34, //vfma.f32 d18, d0, d20
- 0xf2002c30, //vfma.f32 d2, d0, d16
- 0xf22111b1, //vorr d1, d17, d17
- 0xf22301b3, //vorr d0, d19, d19
- 0xf22231b2, //vorr d3, d18, d18
- 0xe8bd4010, //pop {r4, lr}
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_matrix_perspective_vfp4[] = {
- 0xe92d4010, //push {r4, lr}
- 0xe591e000, //ldr lr, [r1]
- 0xe591c004, //ldr ip, [r1, #4]
- 0xe2811008, //add r1, r1, #8
- 0xe28e301c, //add r3, lr, #28
- 0xe28e4010, //add r4, lr, #16
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xe28e3020, //add r3, lr, #32
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe28e3018, //add r3, lr, #24
- 0xf2411c30, //vfma.f32 d17, d1, d16
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xe1a0300e, //mov r3, lr
- 0xf4e42c9f, //vld1.32 {d18[]}, [r4 :32]
- 0xe28e4008, //add r4, lr, #8
- 0xf4e43c9f, //vld1.32 {d19[]}, [r4 :32]
- 0xf2401c30, //vfma.f32 d17, d0, d16
- 0xf4e30c9d, //vld1.32 {d16[]}, [r3 :32]!
- 0xf4e35c9f, //vld1.32 {d21[]}, [r3 :32]
- 0xe28e3014, //add r3, lr, #20
- 0xf2413c35, //vfma.f32 d19, d1, d21
- 0xf4e35c9f, //vld1.32 {d21[]}, [r3 :32]
- 0xe28e300c, //add r3, lr, #12
- 0xf2415c32, //vfma.f32 d21, d1, d18
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf3fb4521, //vrecpe.f32 d20, d17
- 0xf2403c30, //vfma.f32 d19, d0, d16
- 0xf2411fb4, //vrecps.f32 d17, d17, d20
- 0xf2405c32, //vfma.f32 d21, d0, d18
- 0xf3440db1, //vmul.f32 d16, d20, d17
- 0xf3030db0, //vmul.f32 d0, d19, d16
- 0xf3051db0, //vmul.f32 d1, d21, d16
- 0xe8bd4010, //pop {r4, lr}
- 0xe12fff1c, //bx ip
-};
-
-CODE const uint32_t sk_linear_gradient_2stops_vfp4[] = {
- 0xe8911008, //ldm r1, {r3, ip}
- 0xe2811008, //add r1, r1, #8
- 0xf4632a0d, //vld1.8 {d18-d19}, [r3]!
- 0xf4634a0f, //vld1.8 {d20-d21}, [r3]
- 0xf3f40c22, //vdup.32 d16, d18[0]
- 0xf3f41c24, //vdup.32 d17, d20[0]
- 0xf2400c31, //vfma.f32 d16, d0, d17
- 0xf3fc6c24, //vdup.32 d22, d20[1]
- 0xf3bc1c22, //vdup.32 d1, d18[1]
- 0xf3b42c23, //vdup.32 d2, d19[0]
- 0xf2001c36, //vfma.f32 d1, d0, d22
- 0xf3f41c25, //vdup.32 d17, d21[0]
- 0xf3fc4c25, //vdup.32 d20, d21[1]
- 0xf2002c31, //vfma.f32 d2, d0, d17
- 0xf3bc3c23, //vdup.32 d3, d19[1]
- 0xf2003c34, //vfma.f32 d3, d0, d20
- 0xf22001b0, //vorr d0, d16, d16
- 0xe12fff1c, //bx ip
-};
-#elif defined(__x86_64__)
-
-CODE const uint8_t sk_start_pipeline_hsw[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 73,137,205, //mov %rcx,%r13
- 73,137,214, //mov %rdx,%r14
- 72,137,251, //mov %rdi,%rbx
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,199, //mov %rax,%r15
- 73,137,244, //mov %rsi,%r12
- 72,141,67,8, //lea 0x8(%rbx),%rax
- 76,57,232, //cmp %r13,%rax
- 118,5, //jbe 28 <_sk_start_pipeline_hsw+0x28>
- 72,137,223, //mov %rbx,%rdi
- 235,65, //jmp 69 <_sk_start_pipeline_hsw+0x69>
- 185,0,0,0,0, //mov $0x0,%ecx
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 72,141,123,8, //lea 0x8(%rbx),%rdi
- 72,131,195,16, //add $0x10,%rbx
- 76,57,235, //cmp %r13,%rbx
- 72,137,251, //mov %rdi,%rbx
- 118,191, //jbe 28 <_sk_start_pipeline_hsw+0x28>
- 76,137,233, //mov %r13,%rcx
- 72,41,249, //sub %rdi,%rcx
- 116,41, //je 9a <_sk_start_pipeline_hsw+0x9a>
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 76,137,232, //mov %r13,%rax
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 197,248,119, //vzeroupper
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_hsw[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,249,110,199, //vmovd %edi,%xmm0
- 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 65,184,0,0,0,63, //mov $0x3f000000,%r8d
- 196,193,121,110,200, //vmovd %r8d,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 196,226,125,88,210, //vpbroadcastd %xmm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__hsw[] = {
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,92,195, //vsubps %ymm3,%ymm8,%ymm8
- 196,194,93,184,192, //vfmadd231ps %ymm8,%ymm4,%ymm0
- 196,194,85,184,200, //vfmadd231ps %ymm8,%ymm5,%ymm1
- 196,194,77,184,208, //vfmadd231ps %ymm8,%ymm6,%ymm2
- 196,194,69,184,216, //vfmadd231ps %ymm8,%ymm7,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,92,199, //vsubps %ymm7,%ymm8,%ymm8
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_hsw[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,193,124,95,192, //vmaxps %ymm8,%ymm0,%ymm0
- 196,193,116,95,200, //vmaxps %ymm8,%ymm1,%ymm1
- 196,193,108,95,208, //vmaxps %ymm8,%ymm2,%ymm2
- 196,193,100,95,216, //vmaxps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 196,193,108,93,208, //vminps %ymm8,%ymm2,%ymm2
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 197,252,93,195, //vminps %ymm3,%ymm0,%ymm0
- 197,244,93,203, //vminps %ymm3,%ymm1,%ymm1
- 197,236,93,211, //vminps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_hsw[] = {
- 197,124,40,192, //vmovaps %ymm0,%ymm8
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,194, //vmovaps %ymm2,%ymm0
- 197,124,41,194, //vmovaps %ymm8,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_hsw[] = {
- 197,124,40,195, //vmovaps %ymm3,%ymm8
- 197,124,40,202, //vmovaps %ymm2,%ymm9
- 197,124,40,209, //vmovaps %ymm1,%ymm10
- 197,124,40,216, //vmovaps %ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 197,124,41,220, //vmovaps %ymm11,%ymm4
- 197,124,41,213, //vmovaps %ymm10,%ymm5
- 197,124,41,206, //vmovaps %ymm9,%ymm6
- 197,124,41,199, //vmovaps %ymm8,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,224, //vmovaps %ymm0,%ymm4
- 197,252,40,233, //vmovaps %ymm1,%ymm5
- 197,252,40,242, //vmovaps %ymm2,%ymm6
- 197,252,40,251, //vmovaps %ymm3,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_hsw[] = {
- 197,252,89,195, //vmulps %ymm3,%ymm0,%ymm0
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_hsw[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,65,100,194,200,0, //vcmpeqps %ymm8,%ymm3,%ymm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 197,44,94,211, //vdivps %ymm3,%ymm10,%ymm10
- 196,67,45,74,192,144, //vblendvps %ymm9,%ymm8,%ymm10,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_hsw[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,66,125,88,219, //vpbroadcastd %xmm11,%ymm11
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,66,125,88,228, //vpbroadcastd %xmm12,%ymm12
- 196,65,125,111,235, //vmovdqa %ymm11,%ymm13
- 196,66,125,168,236, //vfmadd213ps %ymm12,%ymm0,%ymm13
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,66,125,88,246, //vpbroadcastd %xmm14,%ymm14
- 196,66,45,168,238, //vfmadd213ps %ymm14,%ymm10,%ymm13
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,21,74,193,0, //vblendvps %ymm0,%ymm9,%ymm13,%ymm0
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 197,116,89,233, //vmulps %ymm1,%ymm1,%ymm13
- 196,65,125,111,251, //vmovdqa %ymm11,%ymm15
- 196,66,117,168,252, //vfmadd213ps %ymm12,%ymm1,%ymm15
- 196,66,21,168,254, //vfmadd213ps %ymm14,%ymm13,%ymm15
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,5,74,201,16, //vblendvps %ymm1,%ymm9,%ymm15,%ymm1
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
- 196,66,109,168,220, //vfmadd213ps %ymm12,%ymm2,%ymm11
- 196,66,53,168,222, //vfmadd213ps %ymm14,%ymm9,%ymm11
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
- 196,195,37,74,208,32, //vblendvps %ymm2,%ymm8,%ymm11,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_hsw[] = {
- 197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,216, //vrcpps %ymm8,%ymm11
- 196,65,124,82,224, //vrsqrtps %ymm8,%ymm12
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,232, //vmulps %ymm0,%ymm8,%ymm13
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,66,125,88,246, //vpbroadcastd %xmm14,%ymm14
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 197,121,110,248, //vmovd %eax,%xmm15
- 196,66,125,88,255, //vpbroadcastd %xmm15,%ymm15
- 196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,66,45,184,220, //vfmadd231ps %ymm12,%ymm10,%ymm11
- 196,65,52,93,219, //vminps %ymm11,%ymm9,%ymm11
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,66,125,88,228, //vpbroadcastd %xmm12,%ymm12
- 196,193,124,194,196,1, //vcmpltps %ymm12,%ymm0,%ymm0
- 196,195,37,74,197,0, //vblendvps %ymm0,%ymm13,%ymm11,%ymm0
- 197,124,82,217, //vrsqrtps %ymm1,%ymm11
- 196,65,124,83,235, //vrcpps %ymm11,%ymm13
- 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
- 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
- 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
- 196,193,116,194,204,1, //vcmpltps %ymm12,%ymm1,%ymm1
- 196,195,21,74,203,16, //vblendvps %ymm1,%ymm11,%ymm13,%ymm1
- 197,124,82,218, //vrsqrtps %ymm2,%ymm11
- 196,65,124,83,235, //vrcpps %ymm11,%ymm13
- 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
- 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
- 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
- 196,65,52,93,205, //vminps %ymm13,%ymm9,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,212,1, //vcmpltps %ymm12,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,56, //jne 4bf <_sk_scale_u8_hsw+0x48>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 4c7 <_sk_scale_u8_hsw+0x50>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,167, //jmp 48b <_sk_scale_u8_hsw+0x14>
-};
-
-CODE const uint8_t sk_lerp_1_float_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,76, //jne 56f <_sk_lerp_u8_hsw+0x5c>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 577 <_sk_lerp_u8_hsw+0x64>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,147, //jmp 527 <_sk_lerp_u8_hsw+0x14>
-};
-
-CODE const uint8_t sk_lerp_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,179,0,0,0, //jne 655 <_sk_lerp_565_hsw+0xc1>
- 196,193,122,111,28,122, //vmovdqu (%r10,%rdi,2),%xmm3
- 196,98,125,51,195, //vpmovzxwd %xmm3,%ymm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,53,168,196, //vfmadd213ps %ymm4,%ymm9,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,45,168,205, //vfmadd213ps %ymm5,%ymm10,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,101,168,214, //vfmadd213ps %ymm6,%ymm3,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,59,255,255,255, //ja 5a8 <_sk_lerp_565_hsw+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,76,0,0,0, //lea 0x4c(%rip),%r9 # 6c4 <_sk_lerp_565_hsw+0x130>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 196,193,97,196,92,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,28,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
- 233,231,254,255,255, //jmpq 5a8 <_sk_lerp_565_hsw+0x14>
- 15,31,0, //nopl (%rax)
- 241, //icebp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 233,255,255,255,225, //jmpq ffffffffe20006cc <_sk_linear_gradient_2stops_hsw+0xffffffffe1fff4f0>
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 217,255, //fcos
- 255, //(bad)
- 255,209, //callq *%rcx
- 255, //(bad)
- 255, //(bad)
- 255,201, //dec %ecx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 189, //.byte 0xbd
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_tables_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 77,133,192, //test %r8,%r8
- 117,121, //jne 76e <_sk_load_tables_hsw+0x8e>
- 196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 185,255,0,0,0, //mov $0xff,%ecx
- 197,249,110,193, //vmovd %ecx,%xmm0
- 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
- 197,237,219,203, //vpand %ymm3,%ymm2,%ymm1
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 72,139,72,8, //mov 0x8(%rax),%rcx
- 76,139,72,16, //mov 0x10(%rax),%r9
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,226,53,146,4,137, //vgatherdps %ymm9,(%rcx,%ymm1,4),%ymm0
- 197,245,114,211,8, //vpsrld $0x8,%ymm3,%ymm1
- 197,109,219,201, //vpand %ymm1,%ymm2,%ymm9
- 196,65,45,118,210, //vpcmpeqd %ymm10,%ymm10,%ymm10
- 196,130,45,146,12,137, //vgatherdps %ymm10,(%r9,%ymm9,4),%ymm1
- 72,139,64,24, //mov 0x18(%rax),%rax
- 197,181,114,211,16, //vpsrld $0x10,%ymm3,%ymm9
- 196,65,109,219,201, //vpand %ymm9,%ymm2,%ymm9
- 196,162,61,146,20,136, //vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
- 197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 73,199,194,255,255,255,255, //mov $0xffffffffffffffff,%r10
- 73,211,234, //shr %cl,%r10
- 196,193,249,110,194, //vmovq %r10,%xmm0
- 196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
- 196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,99,255,255,255, //jmpq 6fa <_sk_load_tables_hsw+0x1a>
-};
-
-CODE const uint8_t sk_load_a8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,50, //jne 7d9 <_sk_load_a8_hsw+0x42>
- 197,250,126,0, //vmovq (%rax),%xmm0
- 196,226,125,49,192, //vpmovzxbd %xmm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 7e1 <_sk_load_a8_hsw+0x4a>
- 196,193,249,110,193, //vmovq %r9,%xmm0
- 235,173, //jmp 7ab <_sk_load_a8_hsw+0x14>
-};
-
-CODE const uint8_t sk_store_a8_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne 839 <_sk_store_a8_hsw+0x3b>
- 196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 835 <_sk_store_a8_hsw+0x37>
- 196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # 89c <_sk_store_a8_hsw+0x9e>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,20,68,57,6,12, //vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
- 196,67,121,20,68,57,5,10, //vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
- 196,67,121,20,68,57,4,8, //vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
- 196,67,121,20,68,57,3,6, //vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
- 196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
- 196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
- 196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,154, //jmp 835 <_sk_store_a8_hsw+0x37>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,149,0,0,0, //jne 95b <_sk_load_565_hsw+0xa3>
- 196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
- 196,226,125,51,208, //vpmovzxwd %xmm0,%ymm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
- 197,253,219,194, //vpand %ymm2,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,245,219,202, //vpand %ymm2,%ymm1,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,229,219,210, //vpand %ymm2,%ymm3,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,89,255,255,255, //ja 8cc <_sk_load_565_hsw+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # 9c8 <_sk_load_565_hsw+0x110>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 196,193,121,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,5,255,255,255, //jmpq 8cc <_sk_load_565_hsw+0x14>
- 144, //nop
- 243,255, //repz (bad)
- 255, //(bad)
- 255, //(bad)
- 235,255, //jmp 9cd <_sk_load_565_hsw+0x115>
- 255, //(bad)
- 255,227, //jmpq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 219,255, //(bad)
- 255, //(bad)
- 255,211, //callq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255,203, //dec %ebx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 191, //.byte 0xbf
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,248,65, //mov $0x41f80000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 196,193,53,114,241,11, //vpslld $0xb,%ymm9,%ymm9
- 184,0,0,124,66, //mov $0x427c0000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,5, //vpslld $0x5,%ymm10,%ymm10
- 196,65,45,235,201, //vpor %ymm9,%ymm10,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
- 196,67,125,57,193,1, //vextracti128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne a50 <_sk_store_565_hsw+0x6c>
- 196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja a4c <_sk_store_565_hsw+0x68>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,69,0,0,0, //lea 0x45(%rip),%r8 # ab0 <_sk_store_565_hsw+0xcc>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,21,68,121,12,6, //vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
- 196,67,121,21,68,121,10,5, //vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
- 196,67,121,21,68,121,8,4, //vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
- 196,67,121,21,68,121,6,3, //vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
- 196,67,121,21,68,121,4,2, //vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
- 196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
- 196,67,121,21,4,121,0, //vpextrw $0x0,%xmm8,(%r9,%rdi,2)
- 235,159, //jmp a4c <_sk_store_565_hsw+0x68>
- 15,31,0, //nopl (%rax)
- 244, //hlt
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 236, //in (%dx),%al
- 255, //(bad)
- 255, //(bad)
- 255,228, //jmpq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 220,255, //fdivr %st,%st(7)
- 255, //(bad)
- 255,212, //callq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255,204, //dec %esp
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_8888_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 77,133,192, //test %r8,%r8
- 117,104, //jne b49 <_sk_load_8888_hsw+0x7d>
- 196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 184,255,0,0,0, //mov $0xff,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
- 197,237,219,195, //vpand %ymm3,%ymm2,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,98,125,88,193, //vpbroadcastd %xmm1,%ymm8
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,245,114,211,8, //vpsrld $0x8,%ymm3,%ymm1
- 197,237,219,201, //vpand %ymm1,%ymm2,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,181,114,211,16, //vpsrld $0x10,%ymm3,%ymm9
- 196,193,109,219,209, //vpand %ymm9,%ymm2,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 72,199,192,255,255,255,255, //mov $0xffffffffffffffff,%rax
- 72,211,232, //shr %cl,%rax
- 196,225,249,110,192, //vmovq %rax,%xmm0
- 196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
- 196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,116,255,255,255, //jmpq ae6 <_sk_load_8888_hsw+0x1a>
-};
-
-CODE const uint8_t sk_store_8888_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 197,60,89,209, //vmulps %ymm1,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,8, //vpslld $0x8,%ymm10,%ymm10
- 196,65,45,235,201, //vpor %ymm9,%ymm10,%ymm9
- 197,60,89,210, //vmulps %ymm2,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,16, //vpslld $0x10,%ymm10,%ymm10
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,193,61,114,240,24, //vpslld $0x18,%ymm8,%ymm8
- 196,65,45,235,192, //vpor %ymm8,%ymm10,%ymm8
- 196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
- 77,133,192, //test %r8,%r8
- 117,12, //jne be6 <_sk_store_8888_hsw+0x74>
- 196,65,126,127,1, //vmovdqu %ymm8,(%r9)
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 72,199,192,255,255,255,255, //mov $0xffffffffffffffff,%rax
- 72,211,232, //shr %cl,%rax
- 196,97,249,110,200, //vmovq %rax,%xmm9
- 196,66,125,33,201, //vpmovsxbd %xmm9,%ymm9
- 196,66,53,142,1, //vpmaskmovd %ymm8,%ymm9,(%r9)
- 235,211, //jmp bdf <_sk_store_8888_hsw+0x6d>
-};
-
-CODE const uint8_t sk_load_f16_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,133,201, //test %rcx,%rcx
- 117,97, //jne c77 <_sk_load_f16_hsw+0x6b>
- 197,121,16,4,248, //vmovupd (%rax,%rdi,8),%xmm8
- 197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
- 197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
- 197,122,111,76,248,48, //vmovdqu 0x30(%rax,%rdi,8),%xmm9
- 197,185,97,194, //vpunpcklwd %xmm2,%xmm8,%xmm0
- 197,185,105,210, //vpunpckhwd %xmm2,%xmm8,%xmm2
- 196,193,97,97,201, //vpunpcklwd %xmm9,%xmm3,%xmm1
- 196,193,97,105,217, //vpunpckhwd %xmm9,%xmm3,%xmm3
- 197,121,97,194, //vpunpcklwd %xmm2,%xmm0,%xmm8
- 197,121,105,202, //vpunpckhwd %xmm2,%xmm0,%xmm9
- 197,241,97,211, //vpunpcklwd %xmm3,%xmm1,%xmm2
- 197,241,105,219, //vpunpckhwd %xmm3,%xmm1,%xmm3
- 197,185,108,194, //vpunpcklqdq %xmm2,%xmm8,%xmm0
- 196,226,125,19,192, //vcvtph2ps %xmm0,%ymm0
- 197,185,109,202, //vpunpckhqdq %xmm2,%xmm8,%xmm1
- 196,226,125,19,201, //vcvtph2ps %xmm1,%ymm1
- 197,177,108,211, //vpunpcklqdq %xmm3,%xmm9,%xmm2
- 196,226,125,19,210, //vcvtph2ps %xmm2,%ymm2
- 197,177,109,219, //vpunpckhqdq %xmm3,%xmm9,%xmm3
- 196,226,125,19,219, //vcvtph2ps %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,123,16,4,248, //vmovsd (%rax,%rdi,8),%xmm8
- 196,65,49,239,201, //vpxor %xmm9,%xmm9,%xmm9
- 72,131,249,1, //cmp $0x1,%rcx
- 116,79, //je cd6 <_sk_load_f16_hsw+0xca>
- 197,57,22,68,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
- 72,131,249,3, //cmp $0x3,%rcx
- 114,67, //jb cd6 <_sk_load_f16_hsw+0xca>
- 197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
- 72,131,249,3, //cmp $0x3,%rcx
- 116,68, //je ce3 <_sk_load_f16_hsw+0xd7>
- 197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
- 72,131,249,5, //cmp $0x5,%rcx
- 114,56, //jb ce3 <_sk_load_f16_hsw+0xd7>
- 197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
- 72,131,249,5, //cmp $0x5,%rcx
- 15,132,114,255,255,255, //je c2d <_sk_load_f16_hsw+0x21>
- 197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
- 72,131,249,7, //cmp $0x7,%rcx
- 15,130,98,255,255,255, //jb c2d <_sk_load_f16_hsw+0x21>
- 197,122,126,76,248,48, //vmovq 0x30(%rax,%rdi,8),%xmm9
- 233,87,255,255,255, //jmpq c2d <_sk_load_f16_hsw+0x21>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,74,255,255,255, //jmpq c2d <_sk_load_f16_hsw+0x21>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,65,255,255,255, //jmpq c2d <_sk_load_f16_hsw+0x21>
-};
-
-CODE const uint8_t sk_store_f16_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 196,195,125,29,192,4, //vcvtps2ph $0x4,%ymm0,%xmm8
- 196,195,125,29,201,4, //vcvtps2ph $0x4,%ymm1,%xmm9
- 196,195,125,29,210,4, //vcvtps2ph $0x4,%ymm2,%xmm10
- 196,195,125,29,219,4, //vcvtps2ph $0x4,%ymm3,%xmm11
- 196,65,57,97,225, //vpunpcklwd %xmm9,%xmm8,%xmm12
- 196,65,57,105,193, //vpunpckhwd %xmm9,%xmm8,%xmm8
- 196,65,41,97,203, //vpunpcklwd %xmm11,%xmm10,%xmm9
- 196,65,41,105,235, //vpunpckhwd %xmm11,%xmm10,%xmm13
- 196,65,25,98,217, //vpunpckldq %xmm9,%xmm12,%xmm11
- 196,65,25,106,209, //vpunpckhdq %xmm9,%xmm12,%xmm10
- 196,65,57,98,205, //vpunpckldq %xmm13,%xmm8,%xmm9
- 196,65,57,106,197, //vpunpckhdq %xmm13,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,27, //jne d51 <_sk_store_f16_hsw+0x65>
- 197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
- 197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
- 197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
- 197,122,127,68,248,48, //vmovdqu %xmm8,0x30(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je d4d <_sk_store_f16_hsw+0x61>
- 197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb d4d <_sk_store_f16_hsw+0x61>
- 197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je d4d <_sk_store_f16_hsw+0x61>
- 197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb d4d <_sk_store_f16_hsw+0x61>
- 197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je d4d <_sk_store_f16_hsw+0x61>
- 197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb d4d <_sk_store_f16_hsw+0x61>
- 197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp d4d <_sk_store_f16_hsw+0x61>
-};
-
-CODE const uint8_t sk_store_f32_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,141,4,189,0,0,0,0, //lea 0x0(,%rdi,4),%rax
- 197,124,20,193, //vunpcklps %ymm1,%ymm0,%ymm8
- 197,124,21,217, //vunpckhps %ymm1,%ymm0,%ymm11
- 197,108,20,203, //vunpcklps %ymm3,%ymm2,%ymm9
- 197,108,21,227, //vunpckhps %ymm3,%ymm2,%ymm12
- 196,65,61,20,209, //vunpcklpd %ymm9,%ymm8,%ymm10
- 196,65,61,21,201, //vunpckhpd %ymm9,%ymm8,%ymm9
- 196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
- 196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
- 72,133,201, //test %rcx,%rcx
- 117,55, //jne e05 <_sk_store_f32_hsw+0x6d>
- 196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
- 196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
- 196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
- 196,67,61,6,195,49, //vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
- 196,65,125,17,36,128, //vmovupd %ymm12,(%r8,%rax,4)
- 196,65,125,17,108,128,32, //vmovupd %ymm13,0x20(%r8,%rax,4)
- 196,65,125,17,76,128,64, //vmovupd %ymm9,0x40(%r8,%rax,4)
- 196,65,125,17,68,128,96, //vmovupd %ymm8,0x60(%r8,%rax,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je e01 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb e01 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je e01 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb e01 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je e01 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb e01 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp e01 <_sk_store_f32_hsw+0x69>
-};
-
-CODE const uint8_t sk_clamp_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,188,95,192, //vmaxps %ymm0,%ymm8,%ymm0
- 196,98,125,88,0, //vpbroadcastd (%rax),%ymm8
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,65,61,254,193, //vpaddd %ymm9,%ymm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,188,95,201, //vmaxps %ymm1,%ymm8,%ymm1
- 196,98,125,88,0, //vpbroadcastd (%rax),%ymm8
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,65,61,254,193, //vpaddd %ymm9,%ymm8,%ymm8
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,124,94,200, //vdivps %ymm8,%ymm0,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,98,61,172,200, //vfnmadd213ps %ymm0,%ymm8,%ymm9
- 197,253,118,192, //vpcmpeqd %ymm0,%ymm0,%ymm0
- 197,189,254,192, //vpaddd %ymm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,116,94,200, //vdivps %ymm8,%ymm1,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,98,61,172,201, //vfnmadd213ps %ymm1,%ymm8,%ymm9
- 197,245,118,201, //vpcmpeqd %ymm1,%ymm1,%ymm1
- 197,189,254,201, //vpaddd %ymm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,122,16,0, //vmovss (%rax),%xmm8
- 196,66,125,24,200, //vbroadcastss %xmm8,%ymm9
- 196,65,124,92,209, //vsubps %ymm9,%ymm0,%ymm10
- 196,193,58,88,192, //vaddss %xmm8,%xmm8,%xmm0
- 196,226,125,24,192, //vbroadcastss %xmm0,%ymm0
- 197,44,94,192, //vdivps %ymm0,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 196,66,125,172,194, //vfnmadd213ps %ymm10,%ymm0,%ymm8
- 196,193,60,92,193, //vsubps %ymm9,%ymm8,%ymm0
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,192, //vsubps %ymm0,%ymm8,%ymm8
- 197,188,84,192, //vandps %ymm0,%ymm8,%ymm0
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 196,65,53,254,192, //vpaddd %ymm8,%ymm9,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,122,16,0, //vmovss (%rax),%xmm8
- 196,66,125,24,200, //vbroadcastss %xmm8,%ymm9
- 196,65,116,92,209, //vsubps %ymm9,%ymm1,%ymm10
- 196,193,58,88,200, //vaddss %xmm8,%xmm8,%xmm1
- 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
- 197,44,94,193, //vdivps %ymm1,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 196,66,117,172,194, //vfnmadd213ps %ymm10,%ymm1,%ymm8
- 196,193,60,92,201, //vsubps %ymm9,%ymm8,%ymm1
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,193, //vsubps %ymm1,%ymm8,%ymm8
- 197,188,84,201, //vandps %ymm1,%ymm8,%ymm1
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 196,65,53,254,192, //vpaddd %ymm8,%ymm9,%ymm8
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_hsw[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,98,125,88,195, //vpbroadcastd %xmm3,%ymm8
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,98,125,168,193, //vfmadd213ps %ymm1,%ymm0,%ymm8
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,216, //vpbroadcastd %xmm0,%ymm3
- 196,194,109,168,216, //vfmadd213ps %ymm8,%ymm2,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,253,239,192, //vpxor %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,64,16, //vbroadcastss 0x10(%rax),%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm11
- 196,98,125,24,72,20, //vbroadcastss 0x14(%rax),%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,12, //vbroadcastss 0xc(%rax),%ymm10
- 196,98,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm11
- 196,98,125,24,64,36, //vbroadcastss 0x24(%rax),%ymm8
- 196,66,109,184,195, //vfmadd231ps %ymm11,%ymm2,%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,16, //vbroadcastss 0x10(%rax),%ymm11
- 196,98,125,24,96,28, //vbroadcastss 0x1c(%rax),%ymm12
- 196,98,125,24,72,40, //vbroadcastss 0x28(%rax),%ymm9
- 196,66,109,184,204, //vfmadd231ps %ymm12,%ymm2,%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 196,98,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm11
- 196,98,125,24,96,20, //vbroadcastss 0x14(%rax),%ymm12
- 196,98,125,24,104,32, //vbroadcastss 0x20(%rax),%ymm13
- 196,98,125,24,80,44, //vbroadcastss 0x2c(%rax),%ymm10
- 196,66,109,184,213, //vfmadd231ps %ymm13,%ymm2,%ymm10
- 196,66,117,184,212, //vfmadd231ps %ymm12,%ymm1,%ymm10
- 196,66,125,184,211, //vfmadd231ps %ymm11,%ymm0,%ymm10
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,32, //vbroadcastss 0x20(%rax),%ymm11
- 196,98,125,24,96,48, //vbroadcastss 0x30(%rax),%ymm12
- 196,98,125,24,64,64, //vbroadcastss 0x40(%rax),%ymm8
- 196,66,101,184,196, //vfmadd231ps %ymm12,%ymm3,%ymm8
- 196,66,109,184,195, //vfmadd231ps %ymm11,%ymm2,%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,98,125,24,96,36, //vbroadcastss 0x24(%rax),%ymm12
- 196,98,125,24,104,52, //vbroadcastss 0x34(%rax),%ymm13
- 196,98,125,24,72,68, //vbroadcastss 0x44(%rax),%ymm9
- 196,66,101,184,205, //vfmadd231ps %ymm13,%ymm3,%ymm9
- 196,66,109,184,204, //vfmadd231ps %ymm12,%ymm2,%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 196,98,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm11
- 196,98,125,24,96,24, //vbroadcastss 0x18(%rax),%ymm12
- 196,98,125,24,104,40, //vbroadcastss 0x28(%rax),%ymm13
- 196,98,125,24,112,56, //vbroadcastss 0x38(%rax),%ymm14
- 196,98,125,24,80,72, //vbroadcastss 0x48(%rax),%ymm10
- 196,66,101,184,214, //vfmadd231ps %ymm14,%ymm3,%ymm10
- 196,66,109,184,213, //vfmadd231ps %ymm13,%ymm2,%ymm10
- 196,66,117,184,212, //vfmadd231ps %ymm12,%ymm1,%ymm10
- 196,66,125,184,211, //vfmadd231ps %ymm11,%ymm0,%ymm10
- 196,98,125,24,96,12, //vbroadcastss 0xc(%rax),%ymm12
- 196,98,125,24,104,28, //vbroadcastss 0x1c(%rax),%ymm13
- 196,98,125,24,112,44, //vbroadcastss 0x2c(%rax),%ymm14
- 196,98,125,24,120,60, //vbroadcastss 0x3c(%rax),%ymm15
- 196,98,125,24,88,76, //vbroadcastss 0x4c(%rax),%ymm11
- 196,66,101,184,223, //vfmadd231ps %ymm15,%ymm3,%ymm11
- 196,66,109,184,222, //vfmadd231ps %ymm14,%ymm2,%ymm11
- 196,66,117,184,221, //vfmadd231ps %ymm13,%ymm1,%ymm11
- 196,66,125,184,220, //vfmadd231ps %ymm12,%ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 197,124,41,219, //vmovaps %ymm11,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,66,117,184,209, //vfmadd231ps %ymm9,%ymm1,%ymm10
- 196,66,125,184,208, //vfmadd231ps %ymm8,%ymm0,%ymm10
- 196,98,125,24,64,12, //vbroadcastss 0xc(%rax),%ymm8
- 196,98,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm9
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,66,117,184,217, //vfmadd231ps %ymm9,%ymm1,%ymm11
- 196,66,125,184,216, //vfmadd231ps %ymm8,%ymm0,%ymm11
- 196,98,125,24,64,24, //vbroadcastss 0x18(%rax),%ymm8
- 196,98,125,24,72,28, //vbroadcastss 0x1c(%rax),%ymm9
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 196,66,117,184,225, //vfmadd231ps %ymm9,%ymm1,%ymm12
- 196,66,125,184,224, //vfmadd231ps %ymm8,%ymm0,%ymm12
- 196,193,124,83,204, //vrcpps %ymm12,%ymm1
- 197,172,89,193, //vmulps %ymm1,%ymm10,%ymm0
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm1
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,184,193, //vfmadd231ps %ymm1,%ymm0,%ymm8
- 196,226,125,24,80,20, //vbroadcastss 0x14(%rax),%ymm2
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,184,202, //vfmadd231ps %ymm2,%ymm0,%ymm1
- 196,226,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm3
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,184,211, //vfmadd231ps %ymm3,%ymm0,%ymm2
- 196,98,125,24,72,28, //vbroadcastss 0x1c(%rax),%ymm9
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 196,194,125,184,217, //vfmadd231ps %ymm9,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_avx[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 73,137,205, //mov %rcx,%r13
- 73,137,214, //mov %rdx,%r14
- 72,137,251, //mov %rdi,%rbx
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,199, //mov %rax,%r15
- 73,137,244, //mov %rsi,%r12
- 72,141,67,8, //lea 0x8(%rbx),%rax
- 76,57,232, //cmp %r13,%rax
- 118,5, //jbe 28 <_sk_start_pipeline_avx+0x28>
- 72,137,223, //mov %rbx,%rdi
- 235,65, //jmp 69 <_sk_start_pipeline_avx+0x69>
- 185,0,0,0,0, //mov $0x0,%ecx
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 72,141,123,8, //lea 0x8(%rbx),%rdi
- 72,131,195,16, //add $0x10,%rbx
- 76,57,235, //cmp %r13,%rbx
- 72,137,251, //mov %rdi,%rbx
- 118,191, //jbe 28 <_sk_start_pipeline_avx+0x28>
- 76,137,233, //mov %r13,%rcx
- 72,41,249, //sub %rdi,%rcx
- 116,41, //je 9a <_sk_start_pipeline_avx+0x9a>
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 76,137,232, //mov %r13,%rax
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 197,248,119, //vzeroupper
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_avx[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,249,110,199, //vmovd %edi,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 65,184,0,0,0,63, //mov $0x3f000000,%r8d
- 196,193,121,110,200, //vmovd %r8d,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 196,227,121,4,210,0, //vpermilps $0x0,%xmm2,%xmm2
- 196,227,109,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__avx[] = {
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,92,195, //vsubps %ymm3,%ymm8,%ymm8
- 197,60,89,204, //vmulps %ymm4,%ymm8,%ymm9
- 197,180,88,192, //vaddps %ymm0,%ymm9,%ymm0
- 197,60,89,205, //vmulps %ymm5,%ymm8,%ymm9
- 197,180,88,201, //vaddps %ymm1,%ymm9,%ymm1
- 197,60,89,206, //vmulps %ymm6,%ymm8,%ymm9
- 197,180,88,210, //vaddps %ymm2,%ymm9,%ymm2
- 197,60,89,199, //vmulps %ymm7,%ymm8,%ymm8
- 197,188,88,219, //vaddps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,92,199, //vsubps %ymm7,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_avx[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,193,124,95,192, //vmaxps %ymm8,%ymm0,%ymm0
- 196,193,116,95,200, //vmaxps %ymm8,%ymm1,%ymm1
- 196,193,108,95,208, //vmaxps %ymm8,%ymm2,%ymm2
- 196,193,100,95,216, //vmaxps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 196,193,108,93,208, //vminps %ymm8,%ymm2,%ymm2
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 197,252,93,195, //vminps %ymm3,%ymm0,%ymm0
- 197,244,93,203, //vminps %ymm3,%ymm1,%ymm1
- 197,236,93,211, //vminps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_avx[] = {
- 197,124,40,192, //vmovaps %ymm0,%ymm8
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,194, //vmovaps %ymm2,%ymm0
- 197,124,41,194, //vmovaps %ymm8,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_avx[] = {
- 197,124,40,195, //vmovaps %ymm3,%ymm8
- 197,124,40,202, //vmovaps %ymm2,%ymm9
- 197,124,40,209, //vmovaps %ymm1,%ymm10
- 197,124,40,216, //vmovaps %ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 197,124,41,220, //vmovaps %ymm11,%ymm4
- 197,124,41,213, //vmovaps %ymm10,%ymm5
- 197,124,41,206, //vmovaps %ymm9,%ymm6
- 197,124,41,199, //vmovaps %ymm8,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,224, //vmovaps %ymm0,%ymm4
- 197,252,40,233, //vmovaps %ymm1,%ymm5
- 197,252,40,242, //vmovaps %ymm2,%ymm6
- 197,252,40,251, //vmovaps %ymm3,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_avx[] = {
- 197,252,89,195, //vmulps %ymm3,%ymm0,%ymm0
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_avx[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,65,100,194,200,0, //vcmpeqps %ymm8,%ymm3,%ymm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 197,44,94,211, //vdivps %ymm3,%ymm10,%ymm10
- 196,67,45,74,192,144, //vblendvps %ymm9,%ymm8,%ymm10,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_avx[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
- 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,67,121,4,228,0, //vpermilps $0x0,%xmm12,%xmm12
- 196,67,29,24,228,1, //vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
- 197,36,89,232, //vmulps %ymm0,%ymm11,%ymm13
- 196,65,20,88,236, //vaddps %ymm12,%ymm13,%ymm13
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
- 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
- 196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
- 196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 197,121,110,232, //vmovd %eax,%xmm13
- 196,67,121,4,237,0, //vpermilps $0x0,%xmm13,%xmm13
- 196,67,21,24,237,1, //vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
- 196,193,124,194,197,1, //vcmpltps %ymm13,%ymm0,%ymm0
- 196,195,45,74,193,0, //vblendvps %ymm0,%ymm9,%ymm10,%ymm0
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 197,116,89,209, //vmulps %ymm1,%ymm1,%ymm10
- 197,36,89,249, //vmulps %ymm1,%ymm11,%ymm15
- 196,65,28,88,255, //vaddps %ymm15,%ymm12,%ymm15
- 196,65,44,89,215, //vmulps %ymm15,%ymm10,%ymm10
- 196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 196,193,116,194,205,1, //vcmpltps %ymm13,%ymm1,%ymm1
- 196,195,45,74,201,16, //vblendvps %ymm1,%ymm9,%ymm10,%ymm1
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
- 197,36,89,210, //vmulps %ymm2,%ymm11,%ymm10
- 196,65,28,88,210, //vaddps %ymm10,%ymm12,%ymm10
- 196,65,52,89,202, //vmulps %ymm10,%ymm9,%ymm9
- 196,65,12,88,201, //vaddps %ymm9,%ymm14,%ymm9
- 196,193,108,194,213,1, //vcmpltps %ymm13,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_avx[] = {
- 197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,232, //vrcpps %ymm8,%ymm13
- 196,65,124,82,240, //vrsqrtps %ymm8,%ymm14
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,224, //vmulps %ymm0,%ymm8,%ymm12
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
- 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 197,121,110,248, //vmovd %eax,%xmm15
- 196,67,121,4,255,0, //vpermilps $0x0,%xmm15,%xmm15
- 196,67,5,24,255,1, //vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
- 196,65,20,89,235, //vmulps %ymm11,%ymm13,%ymm13
- 196,65,20,88,239, //vaddps %ymm15,%ymm13,%ymm13
- 196,65,12,89,242, //vmulps %ymm10,%ymm14,%ymm14
- 196,65,12,88,237, //vaddps %ymm13,%ymm14,%ymm13
- 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
- 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
- 196,193,124,194,198,1, //vcmpltps %ymm14,%ymm0,%ymm0
- 196,195,21,74,196,0, //vblendvps %ymm0,%ymm12,%ymm13,%ymm0
- 197,124,82,225, //vrsqrtps %ymm1,%ymm12
- 196,65,124,83,236, //vrcpps %ymm12,%ymm13
- 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
- 196,65,36,89,237, //vmulps %ymm13,%ymm11,%ymm13
- 196,65,4,88,237, //vaddps %ymm13,%ymm15,%ymm13
- 196,65,44,89,228, //vmulps %ymm12,%ymm10,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,60,89,233, //vmulps %ymm1,%ymm8,%ymm13
- 196,65,52,93,228, //vminps %ymm12,%ymm9,%ymm12
- 196,193,116,194,206,1, //vcmpltps %ymm14,%ymm1,%ymm1
- 196,195,29,74,205,16, //vblendvps %ymm1,%ymm13,%ymm12,%ymm1
- 197,124,82,226, //vrsqrtps %ymm2,%ymm12
- 196,65,124,83,236, //vrcpps %ymm12,%ymm13
- 196,65,36,89,221, //vmulps %ymm13,%ymm11,%ymm11
- 196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
- 196,65,44,89,212, //vmulps %ymm12,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 196,65,52,93,202, //vminps %ymm10,%ymm9,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,214,1, //vcmpltps %ymm14,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,80, //jne 5a2 <_sk_scale_u8_avx+0x60>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
- 196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
- 196,66,121,49,192, //vpmovzxbd %xmm8,%xmm8
- 196,67,53,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 5aa <_sk_scale_u8_avx+0x68>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,143, //jmp 556 <_sk_scale_u8_avx+0x14>
-};
-
-CODE const uint8_t sk_lerp_1_float_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,116, //jne 68a <_sk_lerp_u8_avx+0x84>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
- 196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
- 196,66,121,49,192, //vpmovzxbd %xmm8,%xmm8
- 196,67,53,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 692 <_sk_lerp_u8_avx+0x8c>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 233,104,255,255,255, //jmpq 61a <_sk_lerp_u8_avx+0x14>
-};
-
-CODE const uint8_t sk_lerp_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,250,0,0,0, //jne 7ba <_sk_lerp_565_avx+0x108>
- 196,65,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm8
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,185,105,219, //vpunpckhwd %xmm3,%xmm8,%xmm3
- 196,66,121,51,192, //vpmovzxwd %xmm8,%xmm8
- 196,99,61,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,193, //vmulps %ymm9,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,202, //vmulps %ymm10,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 196,65,57,239,192, //vpxor %xmm8,%xmm8,%xmm8
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,243,254,255,255, //ja 6c6 <_sk_lerp_565_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # 828 <_sk_lerp_565_avx+0x176>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 196,65,97,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm8
- 196,65,57,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
- 233,159,254,255,255, //jmpq 6c6 <_sk_lerp_565_avx+0x14>
- 144, //nop
- 243,255, //repz (bad)
- 255, //(bad)
- 255, //(bad)
- 235,255, //jmp 82d <_sk_lerp_565_avx+0x17b>
- 255, //(bad)
- 255,227, //jmpq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 219,255, //(bad)
- 255, //(bad)
- 255,211, //callq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255,203, //dec %ebx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 191, //.byte 0xbf
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_tables_avx[] = {
- 85, //push %rbp
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,133,201, //test %rcx,%rcx
- 15,133,56,2,0,0, //jne a94 <_sk_load_tables_avx+0x250>
- 196,65,124,16,4,184, //vmovups (%r8,%rdi,4),%ymm8
- 187,255,0,0,0, //mov $0xff,%ebx
- 197,249,110,195, //vmovd %ebx,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,99,125,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
- 196,193,52,84,192, //vandps %ymm8,%ymm9,%ymm0
- 196,193,249,126,193, //vmovq %xmm0,%r9
- 69,137,203, //mov %r9d,%r11d
- 196,195,249,22,194,1, //vpextrq $0x1,%xmm0,%r10
- 69,137,214, //mov %r10d,%r14d
- 73,193,234,32, //shr $0x20,%r10
- 73,193,233,32, //shr $0x20,%r9
- 196,227,125,25,192,1, //vextractf128 $0x1,%ymm0,%xmm0
- 196,193,249,126,196, //vmovq %xmm0,%r12
- 69,137,231, //mov %r12d,%r15d
- 196,227,249,22,195,1, //vpextrq $0x1,%xmm0,%rbx
- 65,137,221, //mov %ebx,%r13d
- 72,193,235,32, //shr $0x20,%rbx
- 73,193,236,32, //shr $0x20,%r12
- 72,139,104,8, //mov 0x8(%rax),%rbp
- 76,139,64,16, //mov 0x10(%rax),%r8
- 196,161,122,16,68,189,0, //vmovss 0x0(%rbp,%r15,4),%xmm0
- 196,163,121,33,68,165,0,16, //vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
- 196,161,122,16,76,173,0, //vmovss 0x0(%rbp,%r13,4),%xmm1
- 196,227,121,33,193,32, //vinsertps $0x20,%xmm1,%xmm0,%xmm0
- 197,250,16,76,157,0, //vmovss 0x0(%rbp,%rbx,4),%xmm1
- 196,227,121,33,193,48, //vinsertps $0x30,%xmm1,%xmm0,%xmm0
- 196,161,122,16,76,157,0, //vmovss 0x0(%rbp,%r11,4),%xmm1
- 196,163,113,33,76,141,0,16, //vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
- 196,161,122,16,92,181,0, //vmovss 0x0(%rbp,%r14,4),%xmm3
- 196,227,113,33,203,32, //vinsertps $0x20,%xmm3,%xmm1,%xmm1
- 196,161,122,16,92,149,0, //vmovss 0x0(%rbp,%r10,4),%xmm3
- 196,227,113,33,203,48, //vinsertps $0x30,%xmm3,%xmm1,%xmm1
- 196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
- 196,193,113,114,208,8, //vpsrld $0x8,%xmm8,%xmm1
- 196,67,125,25,194,1, //vextractf128 $0x1,%ymm8,%xmm10
- 196,193,105,114,210,8, //vpsrld $0x8,%xmm10,%xmm2
- 196,227,117,24,202,1, //vinsertf128 $0x1,%xmm2,%ymm1,%ymm1
- 197,180,84,201, //vandps %ymm1,%ymm9,%ymm1
- 196,193,249,126,201, //vmovq %xmm1,%r9
- 69,137,203, //mov %r9d,%r11d
- 196,195,249,22,202,1, //vpextrq $0x1,%xmm1,%r10
- 69,137,214, //mov %r10d,%r14d
- 73,193,234,32, //shr $0x20,%r10
- 73,193,233,32, //shr $0x20,%r9
- 196,227,125,25,201,1, //vextractf128 $0x1,%ymm1,%xmm1
- 196,225,249,126,205, //vmovq %xmm1,%rbp
- 65,137,239, //mov %ebp,%r15d
- 196,227,249,22,203,1, //vpextrq $0x1,%xmm1,%rbx
- 65,137,220, //mov %ebx,%r12d
- 72,193,235,32, //shr $0x20,%rbx
- 72,193,237,32, //shr $0x20,%rbp
- 196,129,122,16,12,184, //vmovss (%r8,%r15,4),%xmm1
- 196,195,113,33,12,168,16, //vinsertps $0x10,(%r8,%rbp,4),%xmm1,%xmm1
- 196,129,122,16,20,160, //vmovss (%r8,%r12,4),%xmm2
- 196,227,113,33,202,32, //vinsertps $0x20,%xmm2,%xmm1,%xmm1
- 196,193,122,16,20,152, //vmovss (%r8,%rbx,4),%xmm2
- 196,227,113,33,202,48, //vinsertps $0x30,%xmm2,%xmm1,%xmm1
- 196,129,122,16,20,152, //vmovss (%r8,%r11,4),%xmm2
- 196,131,105,33,20,136,16, //vinsertps $0x10,(%r8,%r9,4),%xmm2,%xmm2
- 196,129,122,16,28,176, //vmovss (%r8,%r14,4),%xmm3
- 196,227,105,33,211,32, //vinsertps $0x20,%xmm3,%xmm2,%xmm2
- 196,129,122,16,28,144, //vmovss (%r8,%r10,4),%xmm3
- 196,227,105,33,211,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm2
- 196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
- 72,139,64,24, //mov 0x18(%rax),%rax
- 196,193,105,114,208,16, //vpsrld $0x10,%xmm8,%xmm2
- 196,193,97,114,210,16, //vpsrld $0x10,%xmm10,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
- 197,180,84,210, //vandps %ymm2,%ymm9,%ymm2
- 196,193,249,126,208, //vmovq %xmm2,%r8
- 69,137,194, //mov %r8d,%r10d
- 196,195,249,22,209,1, //vpextrq $0x1,%xmm2,%r9
- 69,137,203, //mov %r9d,%r11d
- 73,193,233,32, //shr $0x20,%r9
- 73,193,232,32, //shr $0x20,%r8
- 196,227,125,25,210,1, //vextractf128 $0x1,%ymm2,%xmm2
- 196,225,249,126,213, //vmovq %xmm2,%rbp
- 65,137,238, //mov %ebp,%r14d
- 196,227,249,22,211,1, //vpextrq $0x1,%xmm2,%rbx
- 65,137,223, //mov %ebx,%r15d
- 72,193,235,32, //shr $0x20,%rbx
- 72,193,237,32, //shr $0x20,%rbp
- 196,161,122,16,20,176, //vmovss (%rax,%r14,4),%xmm2
- 196,227,105,33,20,168,16, //vinsertps $0x10,(%rax,%rbp,4),%xmm2,%xmm2
- 196,161,122,16,28,184, //vmovss (%rax,%r15,4),%xmm3
- 196,227,105,33,211,32, //vinsertps $0x20,%xmm3,%xmm2,%xmm2
- 197,250,16,28,152, //vmovss (%rax,%rbx,4),%xmm3
- 196,99,105,33,203,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm9
- 196,161,122,16,28,144, //vmovss (%rax,%r10,4),%xmm3
- 196,163,97,33,28,128,16, //vinsertps $0x10,(%rax,%r8,4),%xmm3,%xmm3
- 196,161,122,16,20,152, //vmovss (%rax,%r11,4),%xmm2
- 196,227,97,33,210,32, //vinsertps $0x20,%xmm2,%xmm3,%xmm2
- 196,161,122,16,28,136, //vmovss (%rax,%r9,4),%xmm3
- 196,227,105,33,211,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm2
- 196,195,109,24,209,1, //vinsertf128 $0x1,%xmm9,%ymm2,%ymm2
- 196,193,57,114,208,24, //vpsrld $0x18,%xmm8,%xmm8
- 196,193,97,114,210,24, //vpsrld $0x18,%xmm10,%xmm3
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 93, //pop %rbp
- 255,224, //jmpq *%rax
- 137,203, //mov %ecx,%ebx
- 128,227,7, //and $0x7,%bl
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 254,203, //dec %bl
- 128,251,6, //cmp $0x6,%bl
- 15,135,185,253,255,255, //ja 862 <_sk_load_tables_avx+0x1e>
- 15,182,219, //movzbl %bl,%ebx
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # b3c <_sk_load_tables_avx+0x2f8>
- 73,99,28,153, //movslq (%r9,%rbx,4),%rbx
- 76,1,203, //add %r9,%rbx
- 255,227, //jmpq *%rbx
- 196,193,121,110,68,184,24, //vmovd 0x18(%r8,%rdi,4),%xmm0
- 197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 196,99,117,12,192,64, //vblendps $0x40,%ymm0,%ymm1,%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,195,121,34,68,184,20,1, //vpinsrd $0x1,0x14(%r8,%rdi,4),%xmm0,%xmm0
- 196,99,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,195,121,34,68,184,16,0, //vpinsrd $0x0,0x10(%r8,%rdi,4),%xmm0,%xmm0
- 196,99,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
- 196,195,57,34,68,184,12,3, //vpinsrd $0x3,0xc(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,68,184,8,2, //vpinsrd $0x2,0x8(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,68,184,4,1, //vpinsrd $0x1,0x4(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,4,184,0, //vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 233,38,253,255,255, //jmpq 862 <_sk_load_tables_avx+0x1e>
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,224, //jmpq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,210, //callq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255,176,255,255,255,156, //pushq -0x63000001(%rax)
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 128,255,255, //cmp $0xff,%bh
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_a8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,74, //jne bb2 <_sk_load_a8_avx+0x5a>
- 197,250,126,0, //vmovq (%rax),%xmm0
- 196,226,121,49,200, //vpmovzxbd %xmm0,%xmm1
- 196,227,121,4,192,229, //vpermilps $0xe5,%xmm0,%xmm0
- 196,226,121,49,192, //vpmovzxbd %xmm0,%xmm0
- 196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne bba <_sk_load_a8_avx+0x62>
- 196,193,249,110,193, //vmovq %r9,%xmm0
- 235,149, //jmp b6c <_sk_load_a8_avx+0x14>
-};
-
-CODE const uint8_t sk_store_a8_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne c19 <_sk_store_a8_avx+0x42>
- 196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja c15 <_sk_store_a8_avx+0x3e>
- 196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # c7c <_sk_store_a8_avx+0xa5>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,20,68,57,6,12, //vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
- 196,67,121,20,68,57,5,10, //vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
- 196,67,121,20,68,57,4,8, //vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
- 196,67,121,20,68,57,3,6, //vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
- 196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
- 196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
- 196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,154, //jmp c15 <_sk_store_a8_avx+0x3e>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,209,0,0,0, //jne d77 <_sk_load_565_avx+0xdf>
- 196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
- 197,241,239,201, //vpxor %xmm1,%xmm1,%xmm1
- 197,249,105,201, //vpunpckhwd %xmm1,%xmm0,%xmm1
- 196,226,121,51,192, //vpmovzxwd %xmm0,%xmm0
- 196,227,125,24,209,1, //vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,252,84,194, //vandps %ymm2,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 197,249,112,201,0, //vpshufd $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,244,84,202, //vandps %ymm2,%ymm1,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,84,210, //vandps %ymm2,%ymm3,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,29,255,255,255, //ja cac <_sk_load_565_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # de4 <_sk_load_565_avx+0x14c>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 196,193,121,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,201,254,255,255, //jmpq cac <_sk_load_565_avx+0x14>
- 144, //nop
- 243,255, //repz (bad)
- 255, //(bad)
- 255, //(bad)
- 235,255, //jmp de9 <_sk_load_565_avx+0x151>
- 255, //(bad)
- 255,227, //jmpq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 219,255, //(bad)
- 255, //(bad)
- 255,211, //callq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255,203, //dec %ebx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 191, //.byte 0xbf
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,248,65, //mov $0x41f80000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 196,193,41,114,241,11, //vpslld $0xb,%xmm9,%xmm10
- 196,67,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm9
- 196,193,49,114,241,11, //vpslld $0xb,%xmm9,%xmm9
- 196,67,45,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
- 184,0,0,124,66, //mov $0x427c0000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,5, //vpslld $0x5,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,5, //vpslld $0x5,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 196,65,45,86,201, //vorpd %ymm9,%ymm10,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne e9e <_sk_store_565_avx+0x9e>
- 196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja e9a <_sk_store_565_avx+0x9a>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # efc <_sk_store_565_avx+0xfc>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,21,68,121,12,6, //vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
- 196,67,121,21,68,121,10,5, //vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
- 196,67,121,21,68,121,8,4, //vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
- 196,67,121,21,68,121,6,3, //vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
- 196,67,121,21,68,121,4,2, //vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
- 196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
- 196,67,121,21,4,121,0, //vpextrw $0x0,%xmm8,(%r9,%rdi,2)
- 235,159, //jmp e9a <_sk_store_565_avx+0x9a>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_8888_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,157,0,0,0, //jne fc3 <_sk_load_8888_avx+0xab>
- 196,65,124,16,12,186, //vmovups (%r10,%rdi,4),%ymm9
- 184,255,0,0,0, //mov $0xff,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,99,125,24,216,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
- 196,193,36,84,193, //vandps %ymm9,%ymm11,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,99,117,24,193,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm8
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 196,193,41,114,209,8, //vpsrld $0x8,%xmm9,%xmm10
- 196,99,125,25,203,1, //vextractf128 $0x1,%ymm9,%xmm3
- 197,241,114,211,8, //vpsrld $0x8,%xmm3,%xmm1
- 196,227,45,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm10,%ymm1
- 197,164,84,201, //vandps %ymm1,%ymm11,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 196,193,41,114,209,16, //vpsrld $0x10,%xmm9,%xmm10
- 197,233,114,211,16, //vpsrld $0x10,%xmm3,%xmm2
- 196,227,45,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm10,%ymm2
- 197,164,84,210, //vandps %ymm2,%ymm11,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 196,193,49,114,209,24, //vpsrld $0x18,%xmm9,%xmm9
- 197,225,114,211,24, //vpsrld $0x18,%xmm3,%xmm3
- 196,227,53,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm9,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 196,65,52,87,201, //vxorps %ymm9,%ymm9,%ymm9
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,80,255,255,255, //ja f2c <_sk_load_8888_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # 1070 <_sk_load_8888_avx+0x158>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 196,193,121,110,68,186,24, //vmovd 0x18(%r10,%rdi,4),%xmm0
- 197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 196,99,117,12,200,64, //vblendps $0x40,%ymm0,%ymm1,%ymm9
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,195,121,34,68,186,20,1, //vpinsrd $0x1,0x14(%r10,%rdi,4),%xmm0,%xmm0
- 196,99,53,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,195,121,34,68,186,16,0, //vpinsrd $0x0,0x10(%r10,%rdi,4),%xmm0,%xmm0
- 196,99,53,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
- 196,195,49,34,68,186,12,3, //vpinsrd $0x3,0xc(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,68,186,8,2, //vpinsrd $0x2,0x8(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,68,186,4,1, //vpinsrd $0x1,0x4(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,4,186,0, //vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 233,188,254,255,255, //jmpq f2c <_sk_load_8888_avx+0x14>
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,224, //jmpq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,210, //callq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255,176,255,255,255,156, //pushq -0x63000001(%rax)
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 128,255,255, //cmp $0xff,%bh
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_8888_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 197,60,89,209, //vmulps %ymm1,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,8, //vpslld $0x8,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,8, //vpslld $0x8,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 196,65,45,86,201, //vorpd %ymm9,%ymm10,%ymm9
- 197,60,89,210, //vmulps %ymm2,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,16, //vpslld $0x10,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,16, //vpslld $0x10,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,193,33,114,240,24, //vpslld $0x18,%xmm8,%xmm11
- 196,67,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm8
- 196,193,57,114,240,24, //vpslld $0x18,%xmm8,%xmm8
- 196,67,37,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
- 196,65,45,86,192, //vorpd %ymm8,%ymm10,%ymm8
- 196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne 1130 <_sk_store_8888_avx+0xa4>
- 196,65,124,17,4,185, //vmovups %ymm8,(%r9,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 112c <_sk_store_8888_avx+0xa0>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,85,0,0,0, //lea 0x55(%rip),%r8 # 11a0 <_sk_store_8888_avx+0x114>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,67,121,22,76,185,24,2, //vpextrd $0x2,%xmm9,0x18(%r9,%rdi,4)
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,67,121,22,76,185,20,1, //vpextrd $0x1,%xmm9,0x14(%r9,%rdi,4)
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,65,122,17,76,185,16, //vmovss %xmm9,0x10(%r9,%rdi,4)
- 196,67,121,22,68,185,12,3, //vpextrd $0x3,%xmm8,0xc(%r9,%rdi,4)
- 196,67,121,22,68,185,8,2, //vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
- 196,67,121,22,68,185,4,1, //vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
- 196,65,121,126,4,185, //vmovd %xmm8,(%r9,%rdi,4)
- 235,143, //jmp 112c <_sk_store_8888_avx+0xa0>
- 15,31,0, //nopl (%rax)
- 245, //cmc
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 237, //in (%dx),%eax
- 255, //(bad)
- 255, //(bad)
- 255,229, //jmpq *%rbp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 221,255, //(bad)
- 255, //(bad)
- 255,208, //callq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,194, //inc %edx
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 180,255, //mov $0xff,%ah
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_f16_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,133,201, //test %rcx,%rcx
- 15,133,2,1,0,0, //jne 12cc <_sk_load_f16_avx+0x110>
- 197,121,16,4,248, //vmovupd (%rax,%rdi,8),%xmm8
- 197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
- 197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
- 197,122,111,76,248,48, //vmovdqu 0x30(%rax,%rdi,8),%xmm9
- 197,185,97,194, //vpunpcklwd %xmm2,%xmm8,%xmm0
- 197,185,105,210, //vpunpckhwd %xmm2,%xmm8,%xmm2
- 196,193,97,97,201, //vpunpcklwd %xmm9,%xmm3,%xmm1
- 196,193,97,105,217, //vpunpckhwd %xmm9,%xmm3,%xmm3
- 197,121,97,194, //vpunpcklwd %xmm2,%xmm0,%xmm8
- 197,249,105,194, //vpunpckhwd %xmm2,%xmm0,%xmm0
- 197,241,97,211, //vpunpcklwd %xmm3,%xmm1,%xmm2
- 197,113,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm9
- 184,0,4,0,4, //mov $0x4000400,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,193,97,101,200, //vpcmpgtw %xmm8,%xmm3,%xmm1
- 196,65,113,223,192, //vpandn %xmm8,%xmm1,%xmm8
- 197,225,101,200, //vpcmpgtw %xmm0,%xmm3,%xmm1
- 197,241,223,192, //vpandn %xmm0,%xmm1,%xmm0
- 197,225,101,202, //vpcmpgtw %xmm2,%xmm3,%xmm1
- 197,241,223,202, //vpandn %xmm2,%xmm1,%xmm1
- 196,193,97,101,209, //vpcmpgtw %xmm9,%xmm3,%xmm2
- 196,193,105,223,209, //vpandn %xmm9,%xmm2,%xmm2
- 196,66,121,51,208, //vpmovzxwd %xmm8,%xmm10
- 196,98,121,51,201, //vpmovzxwd %xmm1,%xmm9
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,57,105,195, //vpunpckhwd %xmm3,%xmm8,%xmm8
- 197,241,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm1
- 196,98,121,51,216, //vpmovzxwd %xmm0,%xmm11
- 196,98,121,51,226, //vpmovzxwd %xmm2,%xmm12
- 197,121,105,235, //vpunpckhwd %xmm3,%xmm0,%xmm13
- 197,105,105,243, //vpunpckhwd %xmm3,%xmm2,%xmm14
- 196,193,121,114,242,13, //vpslld $0xd,%xmm10,%xmm0
- 196,193,105,114,241,13, //vpslld $0xd,%xmm9,%xmm2
- 196,227,125,24,194,1, //vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 197,249,112,210,0, //vpshufd $0x0,%xmm2,%xmm2
- 196,99,109,24,202,1, //vinsertf128 $0x1,%xmm2,%ymm2,%ymm9
- 197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 196,193,105,114,240,13, //vpslld $0xd,%xmm8,%xmm2
- 197,241,114,241,13, //vpslld $0xd,%xmm1,%xmm1
- 196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
- 197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 196,193,105,114,243,13, //vpslld $0xd,%xmm11,%xmm2
- 196,193,97,114,244,13, //vpslld $0xd,%xmm12,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
- 197,180,89,210, //vmulps %ymm2,%ymm9,%ymm2
- 196,193,57,114,245,13, //vpslld $0xd,%xmm13,%xmm8
- 196,193,97,114,246,13, //vpslld $0xd,%xmm14,%xmm3
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,180,89,219, //vmulps %ymm3,%ymm9,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,123,16,4,248, //vmovsd (%rax,%rdi,8),%xmm8
- 196,65,49,239,201, //vpxor %xmm9,%xmm9,%xmm9
- 72,131,249,1, //cmp $0x1,%rcx
- 116,79, //je 132b <_sk_load_f16_avx+0x16f>
- 197,57,22,68,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
- 72,131,249,3, //cmp $0x3,%rcx
- 114,67, //jb 132b <_sk_load_f16_avx+0x16f>
- 197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
- 72,131,249,3, //cmp $0x3,%rcx
- 116,68, //je 1338 <_sk_load_f16_avx+0x17c>
- 197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
- 72,131,249,5, //cmp $0x5,%rcx
- 114,56, //jb 1338 <_sk_load_f16_avx+0x17c>
- 197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
- 72,131,249,5, //cmp $0x5,%rcx
- 15,132,209,254,255,255, //je 11e1 <_sk_load_f16_avx+0x25>
- 197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
- 72,131,249,7, //cmp $0x7,%rcx
- 15,130,193,254,255,255, //jb 11e1 <_sk_load_f16_avx+0x25>
- 197,122,126,76,248,48, //vmovq 0x30(%rax,%rdi,8),%xmm9
- 233,182,254,255,255, //jmpq 11e1 <_sk_load_f16_avx+0x25>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,169,254,255,255, //jmpq 11e1 <_sk_load_f16_avx+0x25>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,160,254,255,255, //jmpq 11e1 <_sk_load_f16_avx+0x25>
-};
-
-CODE const uint8_t sk_store_f16_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 184,0,0,128,7, //mov $0x7800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,65,121,112,192,0, //vpshufd $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,67,125,25,202,1, //vextractf128 $0x1,%ymm9,%xmm10
- 196,193,41,114,210,13, //vpsrld $0xd,%xmm10,%xmm10
- 196,193,49,114,209,13, //vpsrld $0xd,%xmm9,%xmm9
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,67,125,25,220,1, //vextractf128 $0x1,%ymm11,%xmm12
- 196,193,25,114,212,13, //vpsrld $0xd,%xmm12,%xmm12
- 196,193,33,114,211,13, //vpsrld $0xd,%xmm11,%xmm11
- 197,60,89,234, //vmulps %ymm2,%ymm8,%ymm13
- 196,67,125,25,238,1, //vextractf128 $0x1,%ymm13,%xmm14
- 196,193,9,114,214,13, //vpsrld $0xd,%xmm14,%xmm14
- 196,193,17,114,213,13, //vpsrld $0xd,%xmm13,%xmm13
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,67,125,25,199,1, //vextractf128 $0x1,%ymm8,%xmm15
- 196,193,1,114,215,13, //vpsrld $0xd,%xmm15,%xmm15
- 196,193,57,114,208,13, //vpsrld $0xd,%xmm8,%xmm8
- 196,193,33,115,251,2, //vpslldq $0x2,%xmm11,%xmm11
- 196,65,33,235,201, //vpor %xmm9,%xmm11,%xmm9
- 196,193,33,115,252,2, //vpslldq $0x2,%xmm12,%xmm11
- 196,65,33,235,226, //vpor %xmm10,%xmm11,%xmm12
- 196,193,57,115,248,2, //vpslldq $0x2,%xmm8,%xmm8
- 196,65,57,235,197, //vpor %xmm13,%xmm8,%xmm8
- 196,193,41,115,255,2, //vpslldq $0x2,%xmm15,%xmm10
- 196,65,41,235,238, //vpor %xmm14,%xmm10,%xmm13
- 196,65,49,98,216, //vpunpckldq %xmm8,%xmm9,%xmm11
- 196,65,49,106,208, //vpunpckhdq %xmm8,%xmm9,%xmm10
- 196,65,25,98,205, //vpunpckldq %xmm13,%xmm12,%xmm9
- 196,65,25,106,197, //vpunpckhdq %xmm13,%xmm12,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,31, //jne 1417 <_sk_store_f16_avx+0xd6>
- 196,65,120,17,28,248, //vmovups %xmm11,(%r8,%rdi,8)
- 196,65,120,17,84,248,16, //vmovups %xmm10,0x10(%r8,%rdi,8)
- 196,65,120,17,76,248,32, //vmovups %xmm9,0x20(%r8,%rdi,8)
- 196,65,122,127,68,248,48, //vmovdqu %xmm8,0x30(%r8,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,214,28,248, //vmovq %xmm11,(%r8,%rdi,8)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,23,92,248,8, //vmovhpd %xmm11,0x8(%r8,%rdi,8)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,214,84,248,16, //vmovq %xmm10,0x10(%r8,%rdi,8)
- 116,218, //je 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,23,84,248,24, //vmovhpd %xmm10,0x18(%r8,%rdi,8)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,214,76,248,32, //vmovq %xmm9,0x20(%r8,%rdi,8)
- 116,196, //je 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,23,76,248,40, //vmovhpd %xmm9,0x28(%r8,%rdi,8)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,183, //jb 1413 <_sk_store_f16_avx+0xd2>
- 196,65,121,214,68,248,48, //vmovq %xmm8,0x30(%r8,%rdi,8)
- 235,174, //jmp 1413 <_sk_store_f16_avx+0xd2>
-};
-
-CODE const uint8_t sk_store_f32_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,141,4,189,0,0,0,0, //lea 0x0(,%rdi,4),%rax
- 197,124,20,193, //vunpcklps %ymm1,%ymm0,%ymm8
- 197,124,21,217, //vunpckhps %ymm1,%ymm0,%ymm11
- 197,108,20,203, //vunpcklps %ymm3,%ymm2,%ymm9
- 197,108,21,227, //vunpckhps %ymm3,%ymm2,%ymm12
- 196,65,61,20,209, //vunpcklpd %ymm9,%ymm8,%ymm10
- 196,65,61,21,201, //vunpckhpd %ymm9,%ymm8,%ymm9
- 196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
- 196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
- 72,133,201, //test %rcx,%rcx
- 117,55, //jne 14d2 <_sk_store_f32_avx+0x6d>
- 196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
- 196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
- 196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
- 196,67,61,6,195,49, //vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
- 196,65,125,17,36,128, //vmovupd %ymm12,(%r8,%rax,4)
- 196,65,125,17,108,128,32, //vmovupd %ymm13,0x20(%r8,%rax,4)
- 196,65,125,17,76,128,64, //vmovupd %ymm9,0x40(%r8,%rax,4)
- 196,65,125,17,68,128,96, //vmovupd %ymm8,0x60(%r8,%rax,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 14ce <_sk_store_f32_avx+0x69>
- 196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 14ce <_sk_store_f32_avx+0x69>
- 196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je 14ce <_sk_store_f32_avx+0x69>
- 196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 14ce <_sk_store_f32_avx+0x69>
- 196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je 14ce <_sk_store_f32_avx+0x69>
- 196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb 14ce <_sk_store_f32_avx+0x69>
- 196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp 14ce <_sk_store_f32_avx+0x69>
-};
-
-CODE const uint8_t sk_clamp_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,95,200, //vmaxps %ymm0,%ymm8,%ymm9
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,95,201, //vmaxps %ymm1,%ymm8,%ymm9
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,99,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,124,94,200, //vdivps %ymm8,%ymm0,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,65,52,89,200, //vmulps %ymm8,%ymm9,%ymm9
- 196,65,124,92,201, //vsubps %ymm9,%ymm0,%ymm9
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,116,94,200, //vdivps %ymm8,%ymm1,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,65,52,89,200, //vmulps %ymm8,%ymm9,%ymm9
- 196,65,116,92,201, //vsubps %ymm9,%ymm1,%ymm9
- 196,99,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,121,110,0, //vmovd (%rax),%xmm8
- 196,65,121,112,200,0, //vpshufd $0x0,%xmm8,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,124,92,209, //vsubps %ymm9,%ymm0,%ymm10
- 196,193,58,88,192, //vaddss %xmm8,%xmm8,%xmm0
- 196,227,121,4,192,0, //vpermilps $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,44,94,192, //vdivps %ymm0,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,172,92,192, //vsubps %ymm0,%ymm10,%ymm0
- 196,193,124,92,193, //vsubps %ymm9,%ymm0,%ymm0
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,192, //vsubps %ymm0,%ymm8,%ymm8
- 197,60,84,192, //vandps %ymm0,%ymm8,%ymm8
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,49,254,202, //vpaddd %xmm10,%xmm9,%xmm9
- 196,227,53,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm0
- 197,188,93,192, //vminps %ymm0,%ymm8,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,121,110,0, //vmovd (%rax),%xmm8
- 196,65,121,112,200,0, //vpshufd $0x0,%xmm8,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,116,92,209, //vsubps %ymm9,%ymm1,%ymm10
- 196,193,58,88,200, //vaddss %xmm8,%xmm8,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,44,94,193, //vdivps %ymm1,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,172,92,201, //vsubps %ymm1,%ymm10,%ymm1
- 196,193,116,92,201, //vsubps %ymm9,%ymm1,%ymm1
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,193, //vsubps %ymm1,%ymm8,%ymm8
- 197,60,84,193, //vandps %ymm1,%ymm8,%ymm8
- 196,99,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,49,254,202, //vpaddd %xmm10,%xmm9,%xmm9
- 196,227,53,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm9,%ymm1
- 197,188,93,201, //vminps %ymm1,%ymm8,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_avx[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,244,89,202, //vmulps %ymm2,%ymm1,%ymm1
- 197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,8, //vbroadcastss 0x8(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,12, //vbroadcastss 0xc(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 197,172,89,201, //vmulps %ymm1,%ymm10,%ymm1
- 196,193,116,88,203, //vaddps %ymm11,%ymm1,%ymm1
- 197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 197,252,88,201, //vaddps %ymm1,%ymm0,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 196,98,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm10
- 196,98,125,24,88,36, //vbroadcastss 0x24(%rax),%ymm11
- 197,44,89,210, //vmulps %ymm2,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm11
- 196,98,125,24,96,40, //vbroadcastss 0x28(%rax),%ymm12
- 197,36,89,218, //vmulps %ymm2,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 196,98,125,24,104,44, //vbroadcastss 0x2c(%rax),%ymm13
- 197,156,89,210, //vmulps %ymm2,%ymm12,%ymm2
- 196,193,108,88,213, //vaddps %ymm13,%ymm2,%ymm2
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 197,244,88,202, //vaddps %ymm2,%ymm1,%ymm1
- 197,172,89,192, //vmulps %ymm0,%ymm10,%ymm0
- 197,252,88,209, //vaddps %ymm1,%ymm0,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm9
- 196,98,125,24,80,32, //vbroadcastss 0x20(%rax),%ymm10
- 196,98,125,24,88,48, //vbroadcastss 0x30(%rax),%ymm11
- 196,98,125,24,96,64, //vbroadcastss 0x40(%rax),%ymm12
- 197,36,89,219, //vmulps %ymm3,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,210, //vmulps %ymm2,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,20, //vbroadcastss 0x14(%rax),%ymm10
- 196,98,125,24,88,36, //vbroadcastss 0x24(%rax),%ymm11
- 196,98,125,24,96,52, //vbroadcastss 0x34(%rax),%ymm12
- 196,98,125,24,104,68, //vbroadcastss 0x44(%rax),%ymm13
- 197,28,89,227, //vmulps %ymm3,%ymm12,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,36,89,218, //vmulps %ymm2,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm11
- 196,98,125,24,96,40, //vbroadcastss 0x28(%rax),%ymm12
- 196,98,125,24,104,56, //vbroadcastss 0x38(%rax),%ymm13
- 196,98,125,24,112,72, //vbroadcastss 0x48(%rax),%ymm14
- 197,20,89,235, //vmulps %ymm3,%ymm13,%ymm13
- 196,65,20,88,238, //vaddps %ymm14,%ymm13,%ymm13
- 197,28,89,226, //vmulps %ymm2,%ymm12,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,36,89,217, //vmulps %ymm1,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,208, //vmulps %ymm0,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 196,98,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm11
- 196,98,125,24,96,28, //vbroadcastss 0x1c(%rax),%ymm12
- 196,98,125,24,104,44, //vbroadcastss 0x2c(%rax),%ymm13
- 196,98,125,24,112,60, //vbroadcastss 0x3c(%rax),%ymm14
- 196,98,125,24,120,76, //vbroadcastss 0x4c(%rax),%ymm15
- 197,140,89,219, //vmulps %ymm3,%ymm14,%ymm3
- 196,193,100,88,223, //vaddps %ymm15,%ymm3,%ymm3
- 197,148,89,210, //vmulps %ymm2,%ymm13,%ymm2
- 197,236,88,211, //vaddps %ymm3,%ymm2,%ymm2
- 197,156,89,201, //vmulps %ymm1,%ymm12,%ymm1
- 197,244,88,202, //vaddps %ymm2,%ymm1,%ymm1
- 197,164,89,192, //vmulps %ymm0,%ymm11,%ymm0
- 197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm10
- 196,98,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm11
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 196,193,116,88,204, //vaddps %ymm12,%ymm1,%ymm1
- 197,172,89,192, //vmulps %ymm0,%ymm10,%ymm0
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,83,200, //vrcpps %ymm0,%ymm1
- 197,188,89,193, //vmulps %ymm1,%ymm8,%ymm0
- 197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm1
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,244,89,200, //vmulps %ymm0,%ymm1,%ymm1
- 197,108,88,193, //vaddps %ymm1,%ymm2,%ymm8
- 196,226,125,24,72,20, //vbroadcastss 0x14(%rax),%ymm1
- 196,226,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm2
- 197,244,89,200, //vmulps %ymm0,%ymm1,%ymm1
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 196,226,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm2
- 196,226,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm3
- 197,236,89,208, //vmulps %ymm0,%ymm2,%ymm2
- 197,228,88,210, //vaddps %ymm2,%ymm3,%ymm2
- 196,226,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm3
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 197,180,88,216, //vaddps %ymm0,%ymm9,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_sse41[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 73,137,207, //mov %rcx,%r15
- 73,137,214, //mov %rdx,%r14
- 72,137,251, //mov %rdi,%rbx
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,196, //mov %rax,%r12
- 73,137,245, //mov %rsi,%r13
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 76,57,248, //cmp %r15,%rax
- 118,5, //jbe 28 <_sk_start_pipeline_sse41+0x28>
- 72,137,216, //mov %rbx,%rax
- 235,52, //jmp 5c <_sk_start_pipeline_sse41+0x5c>
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,238, //mov %r13,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,212, //callq *%r12
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 72,131,195,8, //add $0x8,%rbx
- 76,57,251, //cmp %r15,%rbx
- 72,137,195, //mov %rax,%rbx
- 118,204, //jbe 28 <_sk_start_pipeline_sse41+0x28>
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_sse41[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 102,15,110,199, //movd %edi,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 185,0,0,0,63, //mov $0x3f000000,%ecx
- 102,15,110,209, //movd %ecx,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,88,202, //addps %xmm2,%xmm1
- 15,16,2, //movups (%rdx),%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 102,15,110,8, //movd (%rax),%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,16,24, //movups (%rax),%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__sse41[] = {
- 15,88,196, //addps %xmm4,%xmm0
- 15,88,205, //addps %xmm5,%xmm1
- 15,88,214, //addps %xmm6,%xmm2
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,195, //subps %xmm3,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,204, //mulps %xmm4,%xmm9
- 65,15,88,193, //addps %xmm9,%xmm0
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,205, //mulps %xmm5,%xmm9
- 65,15,88,201, //addps %xmm9,%xmm1
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,206, //mulps %xmm6,%xmm9
- 65,15,88,209, //addps %xmm9,%xmm2
- 68,15,89,199, //mulps %xmm7,%xmm8
- 65,15,88,216, //addps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,199, //subps %xmm7,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_sse41[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 65,15,95,192, //maxps %xmm8,%xmm0
- 65,15,95,200, //maxps %xmm8,%xmm1
- 65,15,95,208, //maxps %xmm8,%xmm2
- 65,15,95,216, //maxps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 65,15,93,200, //minps %xmm8,%xmm1
- 65,15,93,208, //minps %xmm8,%xmm2
- 65,15,93,216, //minps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,216, //minps %xmm8,%xmm3
- 15,93,195, //minps %xmm3,%xmm0
- 15,93,203, //minps %xmm3,%xmm1
- 15,93,211, //minps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,15,16,80,8, //movss 0x8(%rax),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_sse41[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,194, //movaps %xmm2,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_sse41[] = {
- 68,15,40,195, //movaps %xmm3,%xmm8
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 68,15,40,216, //movaps %xmm0,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 65,15,40,227, //movaps %xmm11,%xmm4
- 65,15,40,234, //movaps %xmm10,%xmm5
- 65,15,40,241, //movaps %xmm9,%xmm6
- 65,15,40,248, //movaps %xmm8,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,224, //movaps %xmm0,%xmm4
- 15,40,233, //movaps %xmm1,%xmm5
- 15,40,242, //movaps %xmm2,%xmm6
- 15,40,251, //movaps %xmm3,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_sse41[] = {
- 15,89,195, //mulps %xmm3,%xmm0
- 15,89,203, //mulps %xmm3,%xmm1
- 15,89,211, //mulps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_sse41[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,94,203, //divps %xmm3,%xmm9
- 68,15,194,195,4, //cmpneqps %xmm3,%xmm8
- 69,15,84,193, //andps %xmm9,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_sse41[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 68,15,40,240, //movaps %xmm0,%xmm14
- 69,15,89,246, //mulps %xmm14,%xmm14
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 69,15,88,204, //addps %xmm12,%xmm9
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 69,15,89,206, //mulps %xmm14,%xmm9
- 69,15,88,205, //addps %xmm13,%xmm9
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 102,68,15,110,240, //movd %eax,%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 65,15,194,198,1, //cmpltps %xmm14,%xmm0
- 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
- 69,15,40,251, //movaps %xmm11,%xmm15
- 68,15,89,249, //mulps %xmm1,%xmm15
- 15,40,193, //movaps %xmm1,%xmm0
- 15,89,192, //mulps %xmm0,%xmm0
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,212, //addps %xmm12,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 69,15,88,213, //addps %xmm13,%xmm10
- 65,15,194,206,1, //cmpltps %xmm14,%xmm1
- 15,40,193, //movaps %xmm1,%xmm0
- 102,69,15,56,20,215, //blendvps %xmm0,%xmm15,%xmm10
- 68,15,89,218, //mulps %xmm2,%xmm11
- 15,40,194, //movaps %xmm2,%xmm0
- 15,89,192, //mulps %xmm0,%xmm0
- 68,15,89,194, //mulps %xmm2,%xmm8
- 69,15,88,196, //addps %xmm12,%xmm8
- 68,15,89,192, //mulps %xmm0,%xmm8
- 69,15,88,197, //addps %xmm13,%xmm8
- 65,15,194,214,1, //cmpltps %xmm14,%xmm2
- 15,40,194, //movaps %xmm2,%xmm0
- 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,202, //movaps %xmm10,%xmm1
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_sse41[] = {
- 72,131,236,24, //sub $0x18,%rsp
- 15,41,60,36, //movaps %xmm7,(%rsp)
- 15,40,254, //movaps %xmm6,%xmm7
- 15,40,245, //movaps %xmm5,%xmm6
- 15,40,236, //movaps %xmm4,%xmm5
- 15,40,227, //movaps %xmm3,%xmm4
- 15,40,218, //movaps %xmm2,%xmm3
- 15,40,209, //movaps %xmm1,%xmm2
- 68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,200, //rcpps %xmm8,%xmm9
- 69,15,82,248, //rsqrtps %xmm8,%xmm15
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 102,68,15,110,240, //movd %eax,%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 69,15,89,205, //mulps %xmm13,%xmm9
- 69,15,88,206, //addps %xmm14,%xmm9
- 69,15,89,252, //mulps %xmm12,%xmm15
- 69,15,88,249, //addps %xmm9,%xmm15
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,93,207, //minps %xmm15,%xmm9
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 102,68,15,110,248, //movd %eax,%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 65,15,194,199,1, //cmpltps %xmm15,%xmm0
- 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
- 68,15,82,210, //rsqrtps %xmm2,%xmm10
- 65,15,83,194, //rcpps %xmm10,%xmm0
- 69,15,82,210, //rsqrtps %xmm10,%xmm10
- 65,15,89,197, //mulps %xmm13,%xmm0
- 65,15,88,198, //addps %xmm14,%xmm0
- 69,15,89,212, //mulps %xmm12,%xmm10
- 68,15,88,208, //addps %xmm0,%xmm10
- 65,15,40,200, //movaps %xmm8,%xmm1
- 65,15,93,202, //minps %xmm10,%xmm1
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,210, //mulps %xmm2,%xmm10
- 65,15,194,215,1, //cmpltps %xmm15,%xmm2
- 15,40,194, //movaps %xmm2,%xmm0
- 102,65,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm1
- 15,82,195, //rsqrtps %xmm3,%xmm0
- 15,83,208, //rcpps %xmm0,%xmm2
- 65,15,89,213, //mulps %xmm13,%xmm2
- 65,15,88,214, //addps %xmm14,%xmm2
- 15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,194, //addps %xmm2,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 68,15,89,219, //mulps %xmm3,%xmm11
- 65,15,194,223,1, //cmpltps %xmm15,%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 15,40,220, //movaps %xmm4,%xmm3
- 15,40,229, //movaps %xmm5,%xmm4
- 15,40,238, //movaps %xmm6,%xmm5
- 15,40,247, //movaps %xmm7,%xmm6
- 15,40,60,36, //movaps (%rsp),%xmm7
- 72,131,196,24, //add $0x18,%rsp
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 65,15,89,201, //mulps %xmm9,%xmm1
- 65,15,89,209, //mulps %xmm9,%xmm2
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_1_float_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,193, //mulps %xmm9,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,51,4,120, //pmovzxwd (%rax,%rdi,2),%xmm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,209, //mulps %xmm9,%xmm10
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,194, //mulps %xmm10,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,203, //mulps %xmm11,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_tables_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,8, //mov (%rax),%rcx
- 76,139,64,8, //mov 0x8(%rax),%r8
- 243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 185,255,0,0,0, //mov $0xff,%ecx
- 102,15,110,193, //movd %ecx,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,111,200, //movdqa %xmm8,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,65,15,111,208, //movdqa %xmm8,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,65,15,219,192, //pand %xmm8,%xmm0
- 102,72,15,58,22,193,1, //pextrq $0x1,%xmm0,%rcx
- 65,137,201, //mov %ecx,%r9d
- 72,193,233,32, //shr $0x20,%rcx
- 102,73,15,126,194, //movq %xmm0,%r10
- 69,137,211, //mov %r10d,%r11d
- 73,193,234,32, //shr $0x20,%r10
- 243,67,15,16,4,152, //movss (%r8,%r11,4),%xmm0
- 102,67,15,58,33,4,144,16, //insertps $0x10,(%r8,%r10,4),%xmm0
- 102,67,15,58,33,4,136,32, //insertps $0x20,(%r8,%r9,4),%xmm0
- 102,65,15,58,33,4,136,48, //insertps $0x30,(%r8,%rcx,4),%xmm0
- 76,139,64,16, //mov 0x10(%rax),%r8
- 102,73,15,58,22,202,1, //pextrq $0x1,%xmm1,%r10
- 77,137,209, //mov %r10,%r9
- 73,193,233,32, //shr $0x20,%r9
- 102,72,15,126,201, //movq %xmm1,%rcx
- 65,137,203, //mov %ecx,%r11d
- 65,129,227,255,255,255,0, //and $0xffffff,%r11d
- 72,193,233,30, //shr $0x1e,%rcx
- 65,129,226,255,255,255,0, //and $0xffffff,%r10d
- 243,67,15,16,12,152, //movss (%r8,%r11,4),%xmm1
- 102,65,15,58,33,12,8,16, //insertps $0x10,(%r8,%rcx,1),%xmm1
- 243,67,15,16,28,144, //movss (%r8,%r10,4),%xmm3
- 102,15,58,33,203,32, //insertps $0x20,%xmm3,%xmm1
- 243,67,15,16,28,136, //movss (%r8,%r9,4),%xmm3
- 102,15,58,33,203,48, //insertps $0x30,%xmm3,%xmm1
- 76,139,72,24, //mov 0x18(%rax),%r9
- 102,72,15,58,22,209,1, //pextrq $0x1,%xmm2,%rcx
- 68,15,183,193, //movzwl %cx,%r8d
- 72,193,233,32, //shr $0x20,%rcx
- 102,72,15,126,208, //movq %xmm2,%rax
- 68,15,183,208, //movzwl %ax,%r10d
- 72,193,232,30, //shr $0x1e,%rax
- 243,67,15,16,20,145, //movss (%r9,%r10,4),%xmm2
- 102,65,15,58,33,20,1,16, //insertps $0x10,(%r9,%rax,1),%xmm2
- 243,67,15,16,28,129, //movss (%r9,%r8,4),%xmm3
- 102,15,58,33,211,32, //insertps $0x20,%xmm3,%xmm2
- 243,65,15,16,28,137, //movss (%r9,%rcx,4),%xmm3
- 102,15,58,33,211,48, //insertps $0x30,%xmm3,%xmm2
- 102,65,15,114,208,24, //psrld $0x18,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_a8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_a8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,56,43,192, //packusdw %xmm8,%xmm8
- 102,69,15,103,192, //packuswb %xmm8,%xmm8
- 102,68,15,126,4,56, //movd %xmm8,(%rax,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,56,51,20,120, //pmovzxwd (%rax,%rdi,2),%xmm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,219,194, //pand %xmm2,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,15,219,202, //pand %xmm2,%xmm1
- 15,91,217, //cvtdq2ps %xmm1,%xmm3
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,15,219,218, //pand %xmm2,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,248,65, //mov $0x41f80000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,11, //pslld $0xb,%xmm9
- 185,0,0,124,66, //mov $0x427c0000,%ecx
- 102,68,15,110,209, //movd %ecx,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,5, //pslld $0x5,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 68,15,89,194, //mulps %xmm2,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,194, //orpd %xmm10,%xmm8
- 102,69,15,56,43,192, //packusdw %xmm8,%xmm8
- 102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_8888_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,28,184, //movdqu (%rax,%rdi,4),%xmm3
- 184,255,0,0,0, //mov $0xff,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,15,219,195, //pand %xmm3,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,91,210, //cvtdq2ps %xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,15,114,211,24, //psrld $0x18,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_8888_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,8, //pslld $0x8,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,202, //mulps %xmm2,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,16, //pslld $0x10,%xmm9
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,24, //pslld $0x18,%xmm8
- 102,69,15,235,193, //por %xmm9,%xmm8
- 102,69,15,235,194, //por %xmm10,%xmm8
- 243,68,15,127,4,184, //movdqu %xmm8,(%rax,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_f16_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,4,248, //movdqu (%rax,%rdi,8),%xmm0
- 243,15,111,76,248,16, //movdqu 0x10(%rax,%rdi,8),%xmm1
- 102,15,111,208, //movdqa %xmm0,%xmm2
- 102,15,97,209, //punpcklwd %xmm1,%xmm2
- 102,15,105,193, //punpckhwd %xmm1,%xmm0
- 102,68,15,111,194, //movdqa %xmm2,%xmm8
- 102,68,15,97,192, //punpcklwd %xmm0,%xmm8
- 102,15,105,208, //punpckhwd %xmm0,%xmm2
- 184,0,4,0,4, //mov $0x4000400,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
- 102,65,15,223,200, //pandn %xmm8,%xmm1
- 102,15,101,218, //pcmpgtw %xmm2,%xmm3
- 102,15,223,218, //pandn %xmm2,%xmm3
- 102,15,56,51,193, //pmovzxwd %xmm1,%xmm0
- 102,15,114,240,13, //pslld $0xd,%xmm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 102,68,15,112,194,0, //pshufd $0x0,%xmm2,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,65,15,105,201, //punpckhwd %xmm9,%xmm1
- 102,15,114,241,13, //pslld $0xd,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,56,51,211, //pmovzxwd %xmm3,%xmm2
- 102,15,114,242,13, //pslld $0xd,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,65,15,105,217, //punpckhwd %xmm9,%xmm3
- 102,15,114,243,13, //pslld $0xd,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f16_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,128,7, //mov $0x7800000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,65,15,114,209,13, //psrld $0xd,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,65,15,114,210,13, //psrld $0xd,%xmm10
- 102,69,15,111,216, //movdqa %xmm8,%xmm11
- 68,15,89,218, //mulps %xmm2,%xmm11
- 102,65,15,114,211,13, //psrld $0xd,%xmm11
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,65,15,114,208,13, //psrld $0xd,%xmm8
- 102,65,15,115,250,2, //pslldq $0x2,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 102,65,15,115,248,2, //pslldq $0x2,%xmm8
- 102,69,15,235,195, //por %xmm11,%xmm8
- 102,69,15,111,202, //movdqa %xmm10,%xmm9
- 102,69,15,98,200, //punpckldq %xmm8,%xmm9
- 243,68,15,127,12,248, //movdqu %xmm9,(%rax,%rdi,8)
- 102,69,15,106,208, //punpckhdq %xmm8,%xmm10
- 243,68,15,127,84,248,16, //movdqu %xmm10,0x10(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f32_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,137,249, //mov %rdi,%rcx
- 72,193,225,4, //shl $0x4,%rcx
- 68,15,40,192, //movaps %xmm0,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 68,15,20,201, //unpcklps %xmm1,%xmm9
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,40,218, //movaps %xmm2,%xmm11
- 68,15,20,219, //unpcklps %xmm3,%xmm11
- 68,15,21,193, //unpckhps %xmm1,%xmm8
- 68,15,21,211, //unpckhps %xmm3,%xmm10
- 69,15,40,225, //movaps %xmm9,%xmm12
- 102,69,15,20,227, //unpcklpd %xmm11,%xmm12
- 69,15,18,217, //movhlps %xmm9,%xmm11
- 69,15,40,200, //movaps %xmm8,%xmm9
- 102,69,15,20,202, //unpcklpd %xmm10,%xmm9
- 69,15,18,208, //movhlps %xmm8,%xmm10
- 102,68,15,17,36,8, //movupd %xmm12,(%rax,%rcx,1)
- 68,15,17,92,8,16, //movups %xmm11,0x10(%rax,%rcx,1)
- 102,68,15,17,76,8,32, //movupd %xmm9,0x20(%rax,%rcx,1)
- 68,15,17,84,8,48, //movups %xmm10,0x30(%rax,%rcx,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,192, //maxps %xmm0,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,192, //pcmpeqd %xmm0,%xmm0
- 102,65,15,254,193, //paddd %xmm9,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,193, //maxps %xmm1,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,201, //pcmpeqd %xmm1,%xmm1
- 102,65,15,254,201, //paddd %xmm9,%xmm1
- 68,15,93,193, //minps %xmm1,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,200, //movaps %xmm8,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 102,69,15,58,8,201,1, //roundps $0x1,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,92,193, //subps %xmm9,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,201, //movaps %xmm1,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 102,69,15,58,8,201,1, //roundps $0x1,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,92,201, //subps %xmm9,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 65,15,92,193, //subps %xmm9,%xmm0
- 243,69,15,88,192, //addss %xmm8,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,208, //movaps %xmm0,%xmm10
- 69,15,94,208, //divps %xmm8,%xmm10
- 102,69,15,58,8,210,1, //roundps $0x1,%xmm10,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,194, //subps %xmm10,%xmm0
- 65,15,92,193, //subps %xmm9,%xmm0
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,92,192, //subps %xmm0,%xmm8
- 65,15,84,192, //andps %xmm8,%xmm0
- 102,69,15,118,192, //pcmpeqd %xmm8,%xmm8
- 102,69,15,254,193, //paddd %xmm9,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 65,15,92,201, //subps %xmm9,%xmm1
- 243,69,15,88,192, //addss %xmm8,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,209, //movaps %xmm1,%xmm10
- 69,15,94,208, //divps %xmm8,%xmm10
- 102,69,15,58,8,210,1, //roundps $0x1,%xmm10,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,202, //subps %xmm10,%xmm1
- 65,15,92,201, //subps %xmm9,%xmm1
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,92,193, //subps %xmm1,%xmm8
- 65,15,84,200, //andps %xmm8,%xmm1
- 102,69,15,118,192, //pcmpeqd %xmm8,%xmm8
- 102,69,15,254,193, //paddd %xmm9,%xmm8
- 65,15,93,200, //minps %xmm8,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_sse41[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 15,88,195, //addps %xmm3,%xmm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,218, //mulps %xmm2,%xmm3
- 15,88,216, //addps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,16, //movss 0x10(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,36, //movss 0x24(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,32, //movss 0x20(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,48, //movss 0x30(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,64, //movss 0x40(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,20, //movss 0x14(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,36, //movss 0x24(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,52, //movss 0x34(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,68, //movss 0x44(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,56, //movss 0x38(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,72, //movss 0x48(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 68,15,89,235, //mulps %xmm3,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 243,68,15,16,88,12, //movss 0xc(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,28, //movss 0x1c(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,60, //movss 0x3c(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 243,68,15,16,120,76, //movss 0x4c(%rax),%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 68,15,89,243, //mulps %xmm3,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 68,15,89,234, //mulps %xmm2,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 69,15,89,225, //mulps %xmm9,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 65,15,40,219, //movaps %xmm11,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_sse41[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,68,15,16,72,4, //movss 0x4(%rax),%xmm9
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,201, //mulps %xmm1,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,193, //addps %xmm9,%xmm0
- 243,68,15,16,72,12, //movss 0xc(%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 69,15,89,200, //mulps %xmm8,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 243,68,15,16,80,24, //movss 0x18(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,217, //mulps %xmm1,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,83,202, //rcpps %xmm10,%xmm1
- 15,89,193, //mulps %xmm1,%xmm0
- 68,15,89,201, //mulps %xmm1,%xmm9
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,201, //movaps %xmm9,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 68,15,16,8, //movups (%rax),%xmm9
- 15,16,88,16, //movups 0x10(%rax),%xmm3
- 68,15,40,195, //movaps %xmm3,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,40,201, //movaps %xmm9,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 68,15,89,192, //mulps %xmm0,%xmm8
- 68,15,88,193, //addps %xmm1,%xmm8
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 65,15,40,209, //movaps %xmm9,%xmm2
- 15,198,210,85, //shufps $0x55,%xmm2,%xmm2
- 15,89,200, //mulps %xmm0,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 69,15,40,209, //movaps %xmm9,%xmm10
- 69,15,198,210,170, //shufps $0xaa,%xmm10,%xmm10
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,210, //addps %xmm10,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 69,15,198,201,255, //shufps $0xff,%xmm9,%xmm9
- 15,89,216, //mulps %xmm0,%xmm3
- 65,15,88,217, //addps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_sse2[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 73,137,207, //mov %rcx,%r15
- 73,137,214, //mov %rdx,%r14
- 72,137,251, //mov %rdi,%rbx
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,196, //mov %rax,%r12
- 73,137,245, //mov %rsi,%r13
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 76,57,248, //cmp %r15,%rax
- 118,5, //jbe 28 <_sk_start_pipeline_sse2+0x28>
- 72,137,216, //mov %rbx,%rax
- 235,52, //jmp 5c <_sk_start_pipeline_sse2+0x5c>
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,238, //mov %r13,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,212, //callq *%r12
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 72,131,195,8, //add $0x8,%rbx
- 76,57,251, //cmp %r15,%rbx
- 72,137,195, //mov %rax,%rbx
- 118,204, //jbe 28 <_sk_start_pipeline_sse2+0x28>
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_sse2[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 102,15,110,199, //movd %edi,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 185,0,0,0,63, //mov $0x3f000000,%ecx
- 102,15,110,209, //movd %ecx,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,88,202, //addps %xmm2,%xmm1
- 15,16,2, //movups (%rdx),%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 102,15,110,8, //movd (%rax),%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,16,24, //movups (%rax),%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__sse2[] = {
- 15,88,196, //addps %xmm4,%xmm0
- 15,88,205, //addps %xmm5,%xmm1
- 15,88,214, //addps %xmm6,%xmm2
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,195, //subps %xmm3,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,204, //mulps %xmm4,%xmm9
- 65,15,88,193, //addps %xmm9,%xmm0
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,205, //mulps %xmm5,%xmm9
- 65,15,88,201, //addps %xmm9,%xmm1
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,206, //mulps %xmm6,%xmm9
- 65,15,88,209, //addps %xmm9,%xmm2
- 68,15,89,199, //mulps %xmm7,%xmm8
- 65,15,88,216, //addps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,199, //subps %xmm7,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_sse2[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 65,15,95,192, //maxps %xmm8,%xmm0
- 65,15,95,200, //maxps %xmm8,%xmm1
- 65,15,95,208, //maxps %xmm8,%xmm2
- 65,15,95,216, //maxps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 65,15,93,200, //minps %xmm8,%xmm1
- 65,15,93,208, //minps %xmm8,%xmm2
- 65,15,93,216, //minps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,216, //minps %xmm8,%xmm3
- 15,93,195, //minps %xmm3,%xmm0
- 15,93,203, //minps %xmm3,%xmm1
- 15,93,211, //minps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,15,16,80,8, //movss 0x8(%rax),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_sse2[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,194, //movaps %xmm2,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_sse2[] = {
- 68,15,40,195, //movaps %xmm3,%xmm8
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 68,15,40,216, //movaps %xmm0,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 65,15,40,227, //movaps %xmm11,%xmm4
- 65,15,40,234, //movaps %xmm10,%xmm5
- 65,15,40,241, //movaps %xmm9,%xmm6
- 65,15,40,248, //movaps %xmm8,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,224, //movaps %xmm0,%xmm4
- 15,40,233, //movaps %xmm1,%xmm5
- 15,40,242, //movaps %xmm2,%xmm6
- 15,40,251, //movaps %xmm3,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_sse2[] = {
- 15,89,195, //mulps %xmm3,%xmm0
- 15,89,203, //mulps %xmm3,%xmm1
- 15,89,211, //mulps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_sse2[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,94,203, //divps %xmm3,%xmm9
- 68,15,194,195,4, //cmpneqps %xmm3,%xmm8
- 69,15,84,193, //andps %xmm9,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_sse2[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,232, //movaps %xmm8,%xmm13
- 68,15,89,232, //mulps %xmm0,%xmm13
- 68,15,40,224, //movaps %xmm0,%xmm12
- 69,15,89,228, //mulps %xmm12,%xmm12
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,40,241, //movaps %xmm9,%xmm14
- 68,15,89,240, //mulps %xmm0,%xmm14
- 69,15,88,242, //addps %xmm10,%xmm14
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,244, //mulps %xmm12,%xmm14
- 69,15,88,243, //addps %xmm11,%xmm14
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 65,15,194,196,1, //cmpltps %xmm12,%xmm0
- 68,15,84,232, //andps %xmm0,%xmm13
- 65,15,85,198, //andnps %xmm14,%xmm0
- 65,15,86,197, //orps %xmm13,%xmm0
- 69,15,40,232, //movaps %xmm8,%xmm13
- 68,15,89,233, //mulps %xmm1,%xmm13
- 68,15,40,241, //movaps %xmm1,%xmm14
- 69,15,89,246, //mulps %xmm14,%xmm14
- 69,15,40,249, //movaps %xmm9,%xmm15
- 68,15,89,249, //mulps %xmm1,%xmm15
- 69,15,88,250, //addps %xmm10,%xmm15
- 69,15,89,254, //mulps %xmm14,%xmm15
- 69,15,88,251, //addps %xmm11,%xmm15
- 65,15,194,204,1, //cmpltps %xmm12,%xmm1
- 68,15,84,233, //andps %xmm1,%xmm13
- 65,15,85,207, //andnps %xmm15,%xmm1
- 65,15,86,205, //orps %xmm13,%xmm1
- 68,15,89,194, //mulps %xmm2,%xmm8
- 68,15,40,234, //movaps %xmm2,%xmm13
- 69,15,89,237, //mulps %xmm13,%xmm13
- 68,15,89,202, //mulps %xmm2,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 69,15,89,205, //mulps %xmm13,%xmm9
- 69,15,88,203, //addps %xmm11,%xmm9
- 65,15,194,212,1, //cmpltps %xmm12,%xmm2
- 68,15,84,194, //andps %xmm2,%xmm8
- 65,15,85,209, //andnps %xmm9,%xmm2
- 65,15,86,208, //orps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_sse2[] = {
- 68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,248, //rcpps %xmm8,%xmm15
- 69,15,82,232, //rsqrtps %xmm8,%xmm13
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,240, //movaps %xmm8,%xmm14
- 68,15,89,240, //mulps %xmm0,%xmm14
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,89,234, //mulps %xmm10,%xmm13
- 69,15,88,239, //addps %xmm15,%xmm13
- 69,15,40,249, //movaps %xmm9,%xmm15
- 69,15,93,253, //minps %xmm13,%xmm15
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 65,15,194,197,1, //cmpltps %xmm13,%xmm0
- 68,15,84,240, //andps %xmm0,%xmm14
- 65,15,85,199, //andnps %xmm15,%xmm0
- 65,15,86,198, //orps %xmm14,%xmm0
- 68,15,82,241, //rsqrtps %xmm1,%xmm14
- 69,15,83,254, //rcpps %xmm14,%xmm15
- 69,15,82,246, //rsqrtps %xmm14,%xmm14
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,89,242, //mulps %xmm10,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 69,15,40,249, //movaps %xmm9,%xmm15
- 69,15,93,254, //minps %xmm14,%xmm15
- 69,15,40,240, //movaps %xmm8,%xmm14
- 68,15,89,241, //mulps %xmm1,%xmm14
- 65,15,194,205,1, //cmpltps %xmm13,%xmm1
- 68,15,84,241, //andps %xmm1,%xmm14
- 65,15,85,207, //andnps %xmm15,%xmm1
- 65,15,86,206, //orps %xmm14,%xmm1
- 68,15,82,242, //rsqrtps %xmm2,%xmm14
- 69,15,83,254, //rcpps %xmm14,%xmm15
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,82,222, //rsqrtps %xmm14,%xmm11
- 69,15,89,218, //mulps %xmm10,%xmm11
- 69,15,88,223, //addps %xmm15,%xmm11
- 69,15,93,203, //minps %xmm11,%xmm9
- 68,15,89,194, //mulps %xmm2,%xmm8
- 65,15,194,213,1, //cmpltps %xmm13,%xmm2
- 68,15,84,194, //andps %xmm2,%xmm8
- 65,15,85,209, //andnps %xmm9,%xmm2
- 65,15,86,208, //orps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,110,4,56, //movd (%rax,%rdi,1),%xmm8
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,69,15,96,193, //punpcklbw %xmm9,%xmm8
- 102,69,15,97,193, //punpcklwd %xmm9,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 65,15,89,201, //mulps %xmm9,%xmm1
- 65,15,89,209, //mulps %xmm9,%xmm2
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_1_float_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,110,4,56, //movd (%rax,%rdi,1),%xmm8
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,69,15,96,193, //punpcklbw %xmm9,%xmm8
- 102,69,15,97,193, //punpcklwd %xmm9,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,193, //mulps %xmm9,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,68,15,126,4,120, //movq (%rax,%rdi,2),%xmm8
- 102,15,239,219, //pxor %xmm3,%xmm3
- 102,68,15,97,195, //punpcklwd %xmm3,%xmm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,209, //mulps %xmm9,%xmm10
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,194, //mulps %xmm10,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,203, //mulps %xmm11,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_tables_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,8, //mov (%rax),%rcx
- 76,139,64,8, //mov 0x8(%rax),%r8
- 243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 185,255,0,0,0, //mov $0xff,%ecx
- 102,15,110,193, //movd %ecx,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 102,65,15,114,209,8, //psrld $0x8,%xmm9
- 102,68,15,219,200, //pand %xmm0,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 102,65,15,114,210,16, //psrld $0x10,%xmm10
- 102,68,15,219,208, //pand %xmm0,%xmm10
- 102,65,15,219,192, //pand %xmm8,%xmm0
- 102,15,112,216,78, //pshufd $0x4e,%xmm0,%xmm3
- 102,72,15,126,217, //movq %xmm3,%rcx
- 65,137,201, //mov %ecx,%r9d
- 72,193,233,32, //shr $0x20,%rcx
- 102,73,15,126,194, //movq %xmm0,%r10
- 69,137,211, //mov %r10d,%r11d
- 73,193,234,32, //shr $0x20,%r10
- 243,67,15,16,28,144, //movss (%r8,%r10,4),%xmm3
- 243,65,15,16,4,136, //movss (%r8,%rcx,4),%xmm0
- 15,20,216, //unpcklps %xmm0,%xmm3
- 243,67,15,16,4,152, //movss (%r8,%r11,4),%xmm0
- 243,67,15,16,12,136, //movss (%r8,%r9,4),%xmm1
- 15,20,193, //unpcklps %xmm1,%xmm0
- 15,20,195, //unpcklps %xmm3,%xmm0
- 76,139,64,16, //mov 0x10(%rax),%r8
- 102,65,15,112,201,78, //pshufd $0x4e,%xmm9,%xmm1
- 102,73,15,126,202, //movq %xmm1,%r10
- 77,137,209, //mov %r10,%r9
- 73,193,233,32, //shr $0x20,%r9
- 102,76,15,126,201, //movq %xmm9,%rcx
- 65,137,203, //mov %ecx,%r11d
- 65,129,227,255,255,255,0, //and $0xffffff,%r11d
- 72,193,233,30, //shr $0x1e,%rcx
- 65,129,226,255,255,255,0, //and $0xffffff,%r10d
- 243,65,15,16,28,8, //movss (%r8,%rcx,1),%xmm3
- 243,67,15,16,12,136, //movss (%r8,%r9,4),%xmm1
- 15,20,217, //unpcklps %xmm1,%xmm3
- 243,67,15,16,12,152, //movss (%r8,%r11,4),%xmm1
- 243,67,15,16,20,144, //movss (%r8,%r10,4),%xmm2
- 15,20,202, //unpcklps %xmm2,%xmm1
- 15,20,203, //unpcklps %xmm3,%xmm1
- 76,139,72,24, //mov 0x18(%rax),%r9
- 102,65,15,112,210,78, //pshufd $0x4e,%xmm10,%xmm2
- 102,72,15,126,209, //movq %xmm2,%rcx
- 68,15,183,193, //movzwl %cx,%r8d
- 72,193,233,32, //shr $0x20,%rcx
- 102,76,15,126,208, //movq %xmm10,%rax
- 68,15,183,208, //movzwl %ax,%r10d
- 72,193,232,30, //shr $0x1e,%rax
- 243,69,15,16,12,1, //movss (%r9,%rax,1),%xmm9
- 243,65,15,16,20,137, //movss (%r9,%rcx,4),%xmm2
- 68,15,20,202, //unpcklps %xmm2,%xmm9
- 243,67,15,16,20,145, //movss (%r9,%r10,4),%xmm2
- 243,67,15,16,28,129, //movss (%r9,%r8,4),%xmm3
- 15,20,211, //unpcklps %xmm3,%xmm2
- 65,15,20,209, //unpcklps %xmm9,%xmm2
- 102,65,15,114,208,24, //psrld $0x18,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_a8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,110,4,56, //movd (%rax,%rdi,1),%xmm0
- 102,15,239,201, //pxor %xmm1,%xmm1
- 102,15,96,193, //punpcklbw %xmm1,%xmm0
- 102,15,97,193, //punpcklwd %xmm1,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 102,15,239,201, //pxor %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_a8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,16, //pslld $0x10,%xmm8
- 102,65,15,114,224,16, //psrad $0x10,%xmm8
- 102,69,15,107,192, //packssdw %xmm8,%xmm8
- 102,69,15,103,192, //packuswb %xmm8,%xmm8
- 102,68,15,126,4,56, //movd %xmm8,(%rax,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,126,20,120, //movq (%rax,%rdi,2),%xmm2
- 102,15,239,192, //pxor %xmm0,%xmm0
- 102,15,97,208, //punpcklwd %xmm0,%xmm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,219,194, //pand %xmm2,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,15,219,202, //pand %xmm2,%xmm1
- 15,91,217, //cvtdq2ps %xmm1,%xmm3
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,15,219,218, //pand %xmm2,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,248,65, //mov $0x41f80000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,11, //pslld $0xb,%xmm9
- 185,0,0,124,66, //mov $0x427c0000,%ecx
- 102,68,15,110,209, //movd %ecx,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,5, //pslld $0x5,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 68,15,89,194, //mulps %xmm2,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,194, //orpd %xmm10,%xmm8
- 102,65,15,114,240,16, //pslld $0x10,%xmm8
- 102,65,15,114,224,16, //psrad $0x10,%xmm8
- 102,69,15,107,192, //packssdw %xmm8,%xmm8
- 102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_8888_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,28,184, //movdqu (%rax,%rdi,4),%xmm3
- 184,255,0,0,0, //mov $0xff,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,15,219,195, //pand %xmm3,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,91,210, //cvtdq2ps %xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,15,114,211,24, //psrld $0x18,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_8888_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,8, //pslld $0x8,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,202, //mulps %xmm2,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,16, //pslld $0x10,%xmm9
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,24, //pslld $0x18,%xmm8
- 102,69,15,235,193, //por %xmm9,%xmm8
- 102,69,15,235,194, //por %xmm10,%xmm8
- 243,68,15,127,4,184, //movdqu %xmm8,(%rax,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_f16_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,4,248, //movdqu (%rax,%rdi,8),%xmm0
- 243,15,111,76,248,16, //movdqu 0x10(%rax,%rdi,8),%xmm1
- 102,15,111,208, //movdqa %xmm0,%xmm2
- 102,15,97,209, //punpcklwd %xmm1,%xmm2
- 102,15,105,193, //punpckhwd %xmm1,%xmm0
- 102,68,15,111,194, //movdqa %xmm2,%xmm8
- 102,68,15,97,192, //punpcklwd %xmm0,%xmm8
- 102,15,105,208, //punpckhwd %xmm0,%xmm2
- 184,0,4,0,4, //mov $0x4000400,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
- 102,65,15,223,200, //pandn %xmm8,%xmm1
- 102,15,101,218, //pcmpgtw %xmm2,%xmm3
- 102,15,223,218, //pandn %xmm2,%xmm3
- 102,69,15,239,192, //pxor %xmm8,%xmm8
- 102,15,111,193, //movdqa %xmm1,%xmm0
- 102,65,15,97,192, //punpcklwd %xmm8,%xmm0
- 102,15,114,240,13, //pslld $0xd,%xmm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 102,68,15,112,202,0, //pshufd $0x0,%xmm2,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 102,65,15,105,200, //punpckhwd %xmm8,%xmm1
- 102,15,114,241,13, //pslld $0xd,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,65,15,97,208, //punpcklwd %xmm8,%xmm2
- 102,15,114,242,13, //pslld $0xd,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 102,65,15,105,216, //punpckhwd %xmm8,%xmm3
- 102,15,114,243,13, //pslld $0xd,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f16_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,128,7, //mov $0x7800000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,65,15,114,209,13, //psrld $0xd,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,65,15,114,210,13, //psrld $0xd,%xmm10
- 102,69,15,111,216, //movdqa %xmm8,%xmm11
- 68,15,89,218, //mulps %xmm2,%xmm11
- 102,65,15,114,211,13, //psrld $0xd,%xmm11
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,65,15,114,208,13, //psrld $0xd,%xmm8
- 102,65,15,115,250,2, //pslldq $0x2,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 102,65,15,115,248,2, //pslldq $0x2,%xmm8
- 102,69,15,235,195, //por %xmm11,%xmm8
- 102,69,15,111,202, //movdqa %xmm10,%xmm9
- 102,69,15,98,200, //punpckldq %xmm8,%xmm9
- 243,68,15,127,12,248, //movdqu %xmm9,(%rax,%rdi,8)
- 102,69,15,106,208, //punpckhdq %xmm8,%xmm10
- 243,68,15,127,84,248,16, //movdqu %xmm10,0x10(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f32_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,137,249, //mov %rdi,%rcx
- 72,193,225,4, //shl $0x4,%rcx
- 68,15,40,192, //movaps %xmm0,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 68,15,20,201, //unpcklps %xmm1,%xmm9
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,40,218, //movaps %xmm2,%xmm11
- 68,15,20,219, //unpcklps %xmm3,%xmm11
- 68,15,21,193, //unpckhps %xmm1,%xmm8
- 68,15,21,211, //unpckhps %xmm3,%xmm10
- 69,15,40,225, //movaps %xmm9,%xmm12
- 102,69,15,20,227, //unpcklpd %xmm11,%xmm12
- 69,15,18,217, //movhlps %xmm9,%xmm11
- 69,15,40,200, //movaps %xmm8,%xmm9
- 102,69,15,20,202, //unpcklpd %xmm10,%xmm9
- 69,15,18,208, //movhlps %xmm8,%xmm10
- 102,68,15,17,36,8, //movupd %xmm12,(%rax,%rcx,1)
- 68,15,17,92,8,16, //movups %xmm11,0x10(%rax,%rcx,1)
- 102,68,15,17,76,8,32, //movupd %xmm9,0x20(%rax,%rcx,1)
- 68,15,17,84,8,48, //movups %xmm10,0x30(%rax,%rcx,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,192, //maxps %xmm0,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,192, //pcmpeqd %xmm0,%xmm0
- 102,65,15,254,193, //paddd %xmm9,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,193, //maxps %xmm1,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,201, //pcmpeqd %xmm1,%xmm1
- 102,65,15,254,201, //paddd %xmm9,%xmm1
- 68,15,93,193, //minps %xmm1,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,200, //movaps %xmm8,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 243,69,15,91,209, //cvttps2dq %xmm9,%xmm10
- 69,15,91,210, //cvtdq2ps %xmm10,%xmm10
- 69,15,194,202,1, //cmpltps %xmm10,%xmm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,84,217, //andps %xmm9,%xmm11
- 69,15,92,211, //subps %xmm11,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,194, //subps %xmm10,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,201, //movaps %xmm1,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 243,69,15,91,209, //cvttps2dq %xmm9,%xmm10
- 69,15,91,210, //cvtdq2ps %xmm10,%xmm10
- 69,15,194,202,1, //cmpltps %xmm10,%xmm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,84,217, //andps %xmm9,%xmm11
- 69,15,92,211, //subps %xmm11,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,202, //subps %xmm10,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,40,193, //movaps %xmm9,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,92,192, //subps %xmm8,%xmm0
- 243,69,15,88,201, //addss %xmm9,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,40,208, //movaps %xmm0,%xmm10
- 69,15,94,209, //divps %xmm9,%xmm10
- 243,69,15,91,218, //cvttps2dq %xmm10,%xmm11
- 69,15,91,219, //cvtdq2ps %xmm11,%xmm11
- 69,15,194,211,1, //cmpltps %xmm11,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,84,226, //andps %xmm10,%xmm12
- 69,15,87,210, //xorps %xmm10,%xmm10
- 69,15,92,220, //subps %xmm12,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 65,15,92,195, //subps %xmm11,%xmm0
- 65,15,92,192, //subps %xmm8,%xmm0
- 68,15,92,208, //subps %xmm0,%xmm10
- 65,15,84,194, //andps %xmm10,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,40,193, //movaps %xmm9,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,92,200, //subps %xmm8,%xmm1
- 243,69,15,88,201, //addss %xmm9,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 69,15,94,209, //divps %xmm9,%xmm10
- 243,69,15,91,218, //cvttps2dq %xmm10,%xmm11
- 69,15,91,219, //cvtdq2ps %xmm11,%xmm11
- 69,15,194,211,1, //cmpltps %xmm11,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,84,226, //andps %xmm10,%xmm12
- 69,15,87,210, //xorps %xmm10,%xmm10
- 69,15,92,220, //subps %xmm12,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 65,15,92,203, //subps %xmm11,%xmm1
- 65,15,92,200, //subps %xmm8,%xmm1
- 68,15,92,209, //subps %xmm1,%xmm10
- 65,15,84,202, //andps %xmm10,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_sse2[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 15,88,195, //addps %xmm3,%xmm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,218, //mulps %xmm2,%xmm3
- 15,88,216, //addps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,16, //movss 0x10(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,36, //movss 0x24(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,32, //movss 0x20(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,48, //movss 0x30(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,64, //movss 0x40(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,20, //movss 0x14(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,36, //movss 0x24(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,52, //movss 0x34(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,68, //movss 0x44(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,56, //movss 0x38(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,72, //movss 0x48(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 68,15,89,235, //mulps %xmm3,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 243,68,15,16,88,12, //movss 0xc(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,28, //movss 0x1c(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,60, //movss 0x3c(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 243,68,15,16,120,76, //movss 0x4c(%rax),%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 68,15,89,243, //mulps %xmm3,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 68,15,89,234, //mulps %xmm2,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 69,15,89,225, //mulps %xmm9,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 65,15,40,219, //movaps %xmm11,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_sse2[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,68,15,16,72,4, //movss 0x4(%rax),%xmm9
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,201, //mulps %xmm1,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,193, //addps %xmm9,%xmm0
- 243,68,15,16,72,12, //movss 0xc(%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 69,15,89,200, //mulps %xmm8,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 243,68,15,16,80,24, //movss 0x18(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,217, //mulps %xmm1,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,83,202, //rcpps %xmm10,%xmm1
- 15,89,193, //mulps %xmm1,%xmm0
- 68,15,89,201, //mulps %xmm1,%xmm9
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,201, //movaps %xmm9,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 68,15,16,8, //movups (%rax),%xmm9
- 15,16,88,16, //movups 0x10(%rax),%xmm3
- 68,15,40,195, //movaps %xmm3,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,40,201, //movaps %xmm9,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 68,15,89,192, //mulps %xmm0,%xmm8
- 68,15,88,193, //addps %xmm1,%xmm8
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 65,15,40,209, //movaps %xmm9,%xmm2
- 15,198,210,85, //shufps $0x55,%xmm2,%xmm2
- 15,89,200, //mulps %xmm0,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 69,15,40,209, //movaps %xmm9,%xmm10
- 69,15,198,210,170, //shufps $0xaa,%xmm10,%xmm10
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,210, //addps %xmm10,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 69,15,198,201,255, //shufps $0xff,%xmm9,%xmm9
- 15,89,216, //mulps %xmm0,%xmm3
- 65,15,88,217, //addps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-#elif defined(_M_X64)
-
-CODE const uint8_t sk_start_pipeline_hsw[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 86, //push %rsi
- 87, //push %rdi
- 83, //push %rbx
- 72,129,236,160,0,0,0, //sub $0xa0,%rsp
- 197,120,41,188,36,144,0,0,0, //vmovaps %xmm15,0x90(%rsp)
- 197,120,41,180,36,128,0,0,0, //vmovaps %xmm14,0x80(%rsp)
- 197,120,41,108,36,112, //vmovaps %xmm13,0x70(%rsp)
- 197,120,41,100,36,96, //vmovaps %xmm12,0x60(%rsp)
- 197,120,41,92,36,80, //vmovaps %xmm11,0x50(%rsp)
- 197,120,41,84,36,64, //vmovaps %xmm10,0x40(%rsp)
- 197,120,41,76,36,48, //vmovaps %xmm9,0x30(%rsp)
- 197,120,41,68,36,32, //vmovaps %xmm8,0x20(%rsp)
- 197,248,41,124,36,16, //vmovaps %xmm7,0x10(%rsp)
- 197,248,41,52,36, //vmovaps %xmm6,(%rsp)
- 77,137,205, //mov %r9,%r13
- 77,137,198, //mov %r8,%r14
- 72,137,203, //mov %rcx,%rbx
- 72,137,214, //mov %rdx,%rsi
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,199, //mov %rax,%r15
- 73,137,244, //mov %rsi,%r12
- 72,141,67,8, //lea 0x8(%rbx),%rax
- 76,57,232, //cmp %r13,%rax
- 118,5, //jbe 75 <_sk_start_pipeline_hsw+0x75>
- 72,137,223, //mov %rbx,%rdi
- 235,65, //jmp b6 <_sk_start_pipeline_hsw+0xb6>
- 185,0,0,0,0, //mov $0x0,%ecx
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 72,141,123,8, //lea 0x8(%rbx),%rdi
- 72,131,195,16, //add $0x10,%rbx
- 76,57,235, //cmp %r13,%rbx
- 72,137,251, //mov %rdi,%rbx
- 118,191, //jbe 75 <_sk_start_pipeline_hsw+0x75>
- 76,137,233, //mov %r13,%rcx
- 72,41,249, //sub %rdi,%rcx
- 116,41, //je e7 <_sk_start_pipeline_hsw+0xe7>
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 76,137,232, //mov %r13,%rax
- 197,248,40,52,36, //vmovaps (%rsp),%xmm6
- 197,248,40,124,36,16, //vmovaps 0x10(%rsp),%xmm7
- 197,120,40,68,36,32, //vmovaps 0x20(%rsp),%xmm8
- 197,120,40,76,36,48, //vmovaps 0x30(%rsp),%xmm9
- 197,120,40,84,36,64, //vmovaps 0x40(%rsp),%xmm10
- 197,120,40,92,36,80, //vmovaps 0x50(%rsp),%xmm11
- 197,120,40,100,36,96, //vmovaps 0x60(%rsp),%xmm12
- 197,120,40,108,36,112, //vmovaps 0x70(%rsp),%xmm13
- 197,120,40,180,36,128,0,0,0, //vmovaps 0x80(%rsp),%xmm14
- 197,120,40,188,36,144,0,0,0, //vmovaps 0x90(%rsp),%xmm15
- 72,129,196,160,0,0,0, //add $0xa0,%rsp
- 91, //pop %rbx
- 95, //pop %rdi
- 94, //pop %rsi
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 197,248,119, //vzeroupper
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_hsw[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,249,110,199, //vmovd %edi,%xmm0
- 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 65,184,0,0,0,63, //mov $0x3f000000,%r8d
- 196,193,121,110,200, //vmovd %r8d,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 196,226,125,88,210, //vpbroadcastd %xmm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__hsw[] = {
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,92,195, //vsubps %ymm3,%ymm8,%ymm8
- 196,194,93,184,192, //vfmadd231ps %ymm8,%ymm4,%ymm0
- 196,194,85,184,200, //vfmadd231ps %ymm8,%ymm5,%ymm1
- 196,194,77,184,208, //vfmadd231ps %ymm8,%ymm6,%ymm2
- 196,194,69,184,216, //vfmadd231ps %ymm8,%ymm7,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,92,199, //vsubps %ymm7,%ymm8,%ymm8
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_hsw[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,193,124,95,192, //vmaxps %ymm8,%ymm0,%ymm0
- 196,193,116,95,200, //vmaxps %ymm8,%ymm1,%ymm1
- 196,193,108,95,208, //vmaxps %ymm8,%ymm2,%ymm2
- 196,193,100,95,216, //vmaxps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 196,193,108,93,208, //vminps %ymm8,%ymm2,%ymm2
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_hsw[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 197,252,93,195, //vminps %ymm3,%ymm0,%ymm0
- 197,244,93,203, //vminps %ymm3,%ymm1,%ymm1
- 197,236,93,211, //vminps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_hsw[] = {
- 197,124,40,192, //vmovaps %ymm0,%ymm8
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,194, //vmovaps %ymm2,%ymm0
- 197,124,41,194, //vmovaps %ymm8,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_hsw[] = {
- 197,124,40,195, //vmovaps %ymm3,%ymm8
- 197,124,40,202, //vmovaps %ymm2,%ymm9
- 197,124,40,209, //vmovaps %ymm1,%ymm10
- 197,124,40,216, //vmovaps %ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 197,124,41,220, //vmovaps %ymm11,%ymm4
- 197,124,41,213, //vmovaps %ymm10,%ymm5
- 197,124,41,206, //vmovaps %ymm9,%ymm6
- 197,124,41,199, //vmovaps %ymm8,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,224, //vmovaps %ymm0,%ymm4
- 197,252,40,233, //vmovaps %ymm1,%ymm5
- 197,252,40,242, //vmovaps %ymm2,%ymm6
- 197,252,40,251, //vmovaps %ymm3,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_hsw[] = {
- 197,252,89,195, //vmulps %ymm3,%ymm0,%ymm0
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_hsw[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,65,100,194,200,0, //vcmpeqps %ymm8,%ymm3,%ymm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 197,44,94,211, //vdivps %ymm3,%ymm10,%ymm10
- 196,67,45,74,192,144, //vblendvps %ymm9,%ymm8,%ymm10,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_hsw[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,66,125,88,219, //vpbroadcastd %xmm11,%ymm11
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,66,125,88,228, //vpbroadcastd %xmm12,%ymm12
- 196,65,125,111,235, //vmovdqa %ymm11,%ymm13
- 196,66,125,168,236, //vfmadd213ps %ymm12,%ymm0,%ymm13
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,66,125,88,246, //vpbroadcastd %xmm14,%ymm14
- 196,66,45,168,238, //vfmadd213ps %ymm14,%ymm10,%ymm13
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,21,74,193,0, //vblendvps %ymm0,%ymm9,%ymm13,%ymm0
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 197,116,89,233, //vmulps %ymm1,%ymm1,%ymm13
- 196,65,125,111,251, //vmovdqa %ymm11,%ymm15
- 196,66,117,168,252, //vfmadd213ps %ymm12,%ymm1,%ymm15
- 196,66,21,168,254, //vfmadd213ps %ymm14,%ymm13,%ymm15
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,5,74,201,16, //vblendvps %ymm1,%ymm9,%ymm15,%ymm1
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
- 196,66,109,168,220, //vfmadd213ps %ymm12,%ymm2,%ymm11
- 196,66,53,168,222, //vfmadd213ps %ymm14,%ymm9,%ymm11
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
- 196,195,37,74,208,32, //vblendvps %ymm2,%ymm8,%ymm11,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_hsw[] = {
- 197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,216, //vrcpps %ymm8,%ymm11
- 196,65,124,82,224, //vrsqrtps %ymm8,%ymm12
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,232, //vmulps %ymm0,%ymm8,%ymm13
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,66,125,88,246, //vpbroadcastd %xmm14,%ymm14
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 197,121,110,248, //vmovd %eax,%xmm15
- 196,66,125,88,255, //vpbroadcastd %xmm15,%ymm15
- 196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,66,45,184,220, //vfmadd231ps %ymm12,%ymm10,%ymm11
- 196,65,52,93,219, //vminps %ymm11,%ymm9,%ymm11
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,66,125,88,228, //vpbroadcastd %xmm12,%ymm12
- 196,193,124,194,196,1, //vcmpltps %ymm12,%ymm0,%ymm0
- 196,195,37,74,197,0, //vblendvps %ymm0,%ymm13,%ymm11,%ymm0
- 197,124,82,217, //vrsqrtps %ymm1,%ymm11
- 196,65,124,83,235, //vrcpps %ymm11,%ymm13
- 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
- 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
- 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
- 196,193,116,194,204,1, //vcmpltps %ymm12,%ymm1,%ymm1
- 196,195,21,74,203,16, //vblendvps %ymm1,%ymm11,%ymm13,%ymm1
- 197,124,82,218, //vrsqrtps %ymm2,%ymm11
- 196,65,124,83,235, //vrcpps %ymm11,%ymm13
- 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
- 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
- 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
- 196,65,52,93,205, //vminps %ymm13,%ymm9,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,212,1, //vcmpltps %ymm12,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,56, //jne 556 <_sk_scale_u8_hsw+0x48>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 55e <_sk_scale_u8_hsw+0x50>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,167, //jmp 522 <_sk_scale_u8_hsw+0x14>
-};
-
-CODE const uint8_t sk_lerp_1_float_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,76, //jne 606 <_sk_lerp_u8_hsw+0x5c>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,66,125,88,201, //vpbroadcastd %xmm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,61,168,205, //vfmadd213ps %ymm5,%ymm8,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,61,168,214, //vfmadd213ps %ymm6,%ymm8,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,226,61,168,223, //vfmadd213ps %ymm7,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 60e <_sk_lerp_u8_hsw+0x64>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,147, //jmp 5be <_sk_lerp_u8_hsw+0x14>
-};
-
-CODE const uint8_t sk_lerp_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,179,0,0,0, //jne 6ec <_sk_lerp_565_hsw+0xc1>
- 196,193,122,111,28,122, //vmovdqu (%r10,%rdi,2),%xmm3
- 196,98,125,51,195, //vpmovzxwd %xmm3,%ymm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,53,168,196, //vfmadd213ps %ymm4,%ymm9,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,45,168,205, //vfmadd213ps %ymm5,%ymm10,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,226,101,168,214, //vfmadd213ps %ymm6,%ymm3,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,59,255,255,255, //ja 63f <_sk_lerp_565_hsw+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,73,0,0,0, //lea 0x49(%rip),%r9 # 758 <_sk_lerp_565_hsw+0x12d>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 196,193,97,196,92,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,92,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
- 196,193,97,196,28,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
- 233,231,254,255,255, //jmpq 63f <_sk_lerp_565_hsw+0x14>
- 244, //hlt
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 236, //in (%dx),%al
- 255, //(bad)
- 255, //(bad)
- 255,228, //jmpq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 220,255, //fdivr %st,%st(7)
- 255, //(bad)
- 255,212, //callq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255,204, //dec %esp
- 255, //(bad)
- 255, //(bad)
- 255,192, //inc %eax
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_tables_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 77,133,192, //test %r8,%r8
- 117,121, //jne 802 <_sk_load_tables_hsw+0x8e>
- 196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 185,255,0,0,0, //mov $0xff,%ecx
- 197,249,110,193, //vmovd %ecx,%xmm0
- 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
- 197,237,219,203, //vpand %ymm3,%ymm2,%ymm1
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 72,139,72,8, //mov 0x8(%rax),%rcx
- 76,139,72,16, //mov 0x10(%rax),%r9
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,226,53,146,4,137, //vgatherdps %ymm9,(%rcx,%ymm1,4),%ymm0
- 197,245,114,211,8, //vpsrld $0x8,%ymm3,%ymm1
- 197,109,219,201, //vpand %ymm1,%ymm2,%ymm9
- 196,65,45,118,210, //vpcmpeqd %ymm10,%ymm10,%ymm10
- 196,130,45,146,12,137, //vgatherdps %ymm10,(%r9,%ymm9,4),%ymm1
- 72,139,64,24, //mov 0x18(%rax),%rax
- 197,181,114,211,16, //vpsrld $0x10,%ymm3,%ymm9
- 196,65,109,219,201, //vpand %ymm9,%ymm2,%ymm9
- 196,162,61,146,20,136, //vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
- 197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 73,199,194,255,255,255,255, //mov $0xffffffffffffffff,%r10
- 73,211,234, //shr %cl,%r10
- 196,193,249,110,194, //vmovq %r10,%xmm0
- 196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
- 196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,99,255,255,255, //jmpq 78e <_sk_load_tables_hsw+0x1a>
-};
-
-CODE const uint8_t sk_load_a8_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,50, //jne 86d <_sk_load_a8_hsw+0x42>
- 197,250,126,0, //vmovq (%rax),%xmm0
- 196,226,125,49,192, //vpmovzxbd %xmm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 875 <_sk_load_a8_hsw+0x4a>
- 196,193,249,110,193, //vmovq %r9,%xmm0
- 235,173, //jmp 83f <_sk_load_a8_hsw+0x14>
-};
-
-CODE const uint8_t sk_store_a8_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne 8cd <_sk_store_a8_hsw+0x3b>
- 196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 8c9 <_sk_store_a8_hsw+0x37>
- 196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # 930 <_sk_store_a8_hsw+0x9e>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,20,68,57,6,12, //vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
- 196,67,121,20,68,57,5,10, //vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
- 196,67,121,20,68,57,4,8, //vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
- 196,67,121,20,68,57,3,6, //vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
- 196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
- 196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
- 196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,154, //jmp 8c9 <_sk_store_a8_hsw+0x37>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,149,0,0,0, //jne 9ef <_sk_load_565_hsw+0xa3>
- 196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
- 196,226,125,51,208, //vpmovzxwd %xmm0,%ymm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
- 197,253,219,194, //vpand %ymm2,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
- 197,245,219,202, //vpand %ymm2,%ymm1,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,229,219,210, //vpand %ymm2,%ymm3,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,89,255,255,255, //ja 960 <_sk_load_565_hsw+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # a5c <_sk_load_565_hsw+0x110>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 196,193,121,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,5,255,255,255, //jmpq 960 <_sk_load_565_hsw+0x14>
- 144, //nop
- 243,255, //repz (bad)
- 255, //(bad)
- 255, //(bad)
- 235,255, //jmp a61 <_sk_load_565_hsw+0x115>
- 255, //(bad)
- 255,227, //jmpq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 219,255, //(bad)
- 255, //(bad)
- 255,211, //callq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255,203, //dec %ebx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 191, //.byte 0xbf
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_565_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,248,65, //mov $0x41f80000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 196,193,53,114,241,11, //vpslld $0xb,%ymm9,%ymm9
- 184,0,0,124,66, //mov $0x427c0000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,66,125,88,210, //vpbroadcastd %xmm10,%ymm10
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,5, //vpslld $0x5,%ymm10,%ymm10
- 196,65,45,235,201, //vpor %ymm9,%ymm10,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
- 196,67,125,57,193,1, //vextracti128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne ae4 <_sk_store_565_hsw+0x6c>
- 196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja ae0 <_sk_store_565_hsw+0x68>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,69,0,0,0, //lea 0x45(%rip),%r8 # b44 <_sk_store_565_hsw+0xcc>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,21,68,121,12,6, //vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
- 196,67,121,21,68,121,10,5, //vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
- 196,67,121,21,68,121,8,4, //vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
- 196,67,121,21,68,121,6,3, //vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
- 196,67,121,21,68,121,4,2, //vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
- 196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
- 196,67,121,21,4,121,0, //vpextrw $0x0,%xmm8,(%r9,%rdi,2)
- 235,159, //jmp ae0 <_sk_store_565_hsw+0x68>
- 15,31,0, //nopl (%rax)
- 244, //hlt
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 236, //in (%dx),%al
- 255, //(bad)
- 255, //(bad)
- 255,228, //jmpq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 220,255, //fdivr %st,%st(7)
- 255, //(bad)
- 255,212, //callq *%rsp
- 255, //(bad)
- 255, //(bad)
- 255,204, //dec %esp
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_8888_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 77,133,192, //test %r8,%r8
- 117,104, //jne bdd <_sk_load_8888_hsw+0x7d>
- 196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 184,255,0,0,0, //mov $0xff,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
- 197,237,219,195, //vpand %ymm3,%ymm2,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,98,125,88,193, //vpbroadcastd %xmm1,%ymm8
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,245,114,211,8, //vpsrld $0x8,%ymm3,%ymm1
- 197,237,219,201, //vpand %ymm1,%ymm2,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,181,114,211,16, //vpsrld $0x10,%ymm3,%ymm9
- 196,193,109,219,209, //vpand %ymm9,%ymm2,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 72,199,192,255,255,255,255, //mov $0xffffffffffffffff,%rax
- 72,211,232, //shr %cl,%rax
- 196,225,249,110,192, //vmovq %rax,%xmm0
- 196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
- 196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,116,255,255,255, //jmpq b7a <_sk_load_8888_hsw+0x1a>
-};
-
-CODE const uint8_t sk_store_8888_hsw[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
- 76,3,8, //add (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,66,125,88,192, //vpbroadcastd %xmm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 197,60,89,209, //vmulps %ymm1,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,8, //vpslld $0x8,%ymm10,%ymm10
- 196,65,45,235,201, //vpor %ymm9,%ymm10,%ymm9
- 197,60,89,210, //vmulps %ymm2,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,45,114,242,16, //vpslld $0x10,%ymm10,%ymm10
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,193,61,114,240,24, //vpslld $0x18,%ymm8,%ymm8
- 196,65,45,235,192, //vpor %ymm8,%ymm10,%ymm8
- 196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
- 77,133,192, //test %r8,%r8
- 117,12, //jne c7a <_sk_store_8888_hsw+0x74>
- 196,65,126,127,1, //vmovdqu %ymm8,(%r9)
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 185,8,0,0,0, //mov $0x8,%ecx
- 68,41,193, //sub %r8d,%ecx
- 192,225,3, //shl $0x3,%cl
- 72,199,192,255,255,255,255, //mov $0xffffffffffffffff,%rax
- 72,211,232, //shr %cl,%rax
- 196,97,249,110,200, //vmovq %rax,%xmm9
- 196,66,125,33,201, //vpmovsxbd %xmm9,%ymm9
- 196,66,53,142,1, //vpmaskmovd %ymm8,%ymm9,(%r9)
- 235,211, //jmp c73 <_sk_store_8888_hsw+0x6d>
-};
-
-CODE const uint8_t sk_load_f16_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,133,201, //test %rcx,%rcx
- 117,97, //jne d0b <_sk_load_f16_hsw+0x6b>
- 197,121,16,4,248, //vmovupd (%rax,%rdi,8),%xmm8
- 197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
- 197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
- 197,122,111,76,248,48, //vmovdqu 0x30(%rax,%rdi,8),%xmm9
- 197,185,97,194, //vpunpcklwd %xmm2,%xmm8,%xmm0
- 197,185,105,210, //vpunpckhwd %xmm2,%xmm8,%xmm2
- 196,193,97,97,201, //vpunpcklwd %xmm9,%xmm3,%xmm1
- 196,193,97,105,217, //vpunpckhwd %xmm9,%xmm3,%xmm3
- 197,121,97,194, //vpunpcklwd %xmm2,%xmm0,%xmm8
- 197,121,105,202, //vpunpckhwd %xmm2,%xmm0,%xmm9
- 197,241,97,211, //vpunpcklwd %xmm3,%xmm1,%xmm2
- 197,241,105,219, //vpunpckhwd %xmm3,%xmm1,%xmm3
- 197,185,108,194, //vpunpcklqdq %xmm2,%xmm8,%xmm0
- 196,226,125,19,192, //vcvtph2ps %xmm0,%ymm0
- 197,185,109,202, //vpunpckhqdq %xmm2,%xmm8,%xmm1
- 196,226,125,19,201, //vcvtph2ps %xmm1,%ymm1
- 197,177,108,211, //vpunpcklqdq %xmm3,%xmm9,%xmm2
- 196,226,125,19,210, //vcvtph2ps %xmm2,%ymm2
- 197,177,109,219, //vpunpckhqdq %xmm3,%xmm9,%xmm3
- 196,226,125,19,219, //vcvtph2ps %xmm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,123,16,4,248, //vmovsd (%rax,%rdi,8),%xmm8
- 196,65,49,239,201, //vpxor %xmm9,%xmm9,%xmm9
- 72,131,249,1, //cmp $0x1,%rcx
- 116,79, //je d6a <_sk_load_f16_hsw+0xca>
- 197,57,22,68,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
- 72,131,249,3, //cmp $0x3,%rcx
- 114,67, //jb d6a <_sk_load_f16_hsw+0xca>
- 197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
- 72,131,249,3, //cmp $0x3,%rcx
- 116,68, //je d77 <_sk_load_f16_hsw+0xd7>
- 197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
- 72,131,249,5, //cmp $0x5,%rcx
- 114,56, //jb d77 <_sk_load_f16_hsw+0xd7>
- 197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
- 72,131,249,5, //cmp $0x5,%rcx
- 15,132,114,255,255,255, //je cc1 <_sk_load_f16_hsw+0x21>
- 197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
- 72,131,249,7, //cmp $0x7,%rcx
- 15,130,98,255,255,255, //jb cc1 <_sk_load_f16_hsw+0x21>
- 197,122,126,76,248,48, //vmovq 0x30(%rax,%rdi,8),%xmm9
- 233,87,255,255,255, //jmpq cc1 <_sk_load_f16_hsw+0x21>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,74,255,255,255, //jmpq cc1 <_sk_load_f16_hsw+0x21>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,65,255,255,255, //jmpq cc1 <_sk_load_f16_hsw+0x21>
-};
-
-CODE const uint8_t sk_store_f16_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 196,195,125,29,192,4, //vcvtps2ph $0x4,%ymm0,%xmm8
- 196,195,125,29,201,4, //vcvtps2ph $0x4,%ymm1,%xmm9
- 196,195,125,29,210,4, //vcvtps2ph $0x4,%ymm2,%xmm10
- 196,195,125,29,219,4, //vcvtps2ph $0x4,%ymm3,%xmm11
- 196,65,57,97,225, //vpunpcklwd %xmm9,%xmm8,%xmm12
- 196,65,57,105,193, //vpunpckhwd %xmm9,%xmm8,%xmm8
- 196,65,41,97,203, //vpunpcklwd %xmm11,%xmm10,%xmm9
- 196,65,41,105,235, //vpunpckhwd %xmm11,%xmm10,%xmm13
- 196,65,25,98,217, //vpunpckldq %xmm9,%xmm12,%xmm11
- 196,65,25,106,209, //vpunpckhdq %xmm9,%xmm12,%xmm10
- 196,65,57,98,205, //vpunpckldq %xmm13,%xmm8,%xmm9
- 196,65,57,106,197, //vpunpckhdq %xmm13,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,27, //jne de5 <_sk_store_f16_hsw+0x65>
- 197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
- 197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
- 197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
- 197,122,127,68,248,48, //vmovdqu %xmm8,0x30(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je de1 <_sk_store_f16_hsw+0x61>
- 197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb de1 <_sk_store_f16_hsw+0x61>
- 197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je de1 <_sk_store_f16_hsw+0x61>
- 197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb de1 <_sk_store_f16_hsw+0x61>
- 197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je de1 <_sk_store_f16_hsw+0x61>
- 197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb de1 <_sk_store_f16_hsw+0x61>
- 197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp de1 <_sk_store_f16_hsw+0x61>
-};
-
-CODE const uint8_t sk_store_f32_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,141,4,189,0,0,0,0, //lea 0x0(,%rdi,4),%rax
- 197,124,20,193, //vunpcklps %ymm1,%ymm0,%ymm8
- 197,124,21,217, //vunpckhps %ymm1,%ymm0,%ymm11
- 197,108,20,203, //vunpcklps %ymm3,%ymm2,%ymm9
- 197,108,21,227, //vunpckhps %ymm3,%ymm2,%ymm12
- 196,65,61,20,209, //vunpcklpd %ymm9,%ymm8,%ymm10
- 196,65,61,21,201, //vunpckhpd %ymm9,%ymm8,%ymm9
- 196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
- 196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
- 72,133,201, //test %rcx,%rcx
- 117,55, //jne e99 <_sk_store_f32_hsw+0x6d>
- 196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
- 196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
- 196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
- 196,67,61,6,195,49, //vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
- 196,65,125,17,36,128, //vmovupd %ymm12,(%r8,%rax,4)
- 196,65,125,17,108,128,32, //vmovupd %ymm13,0x20(%r8,%rax,4)
- 196,65,125,17,76,128,64, //vmovupd %ymm9,0x40(%r8,%rax,4)
- 196,65,125,17,68,128,96, //vmovupd %ymm8,0x60(%r8,%rax,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je e95 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb e95 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je e95 <_sk_store_f32_hsw+0x69>
- 196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb e95 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je e95 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb e95 <_sk_store_f32_hsw+0x69>
- 196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp e95 <_sk_store_f32_hsw+0x69>
-};
-
-CODE const uint8_t sk_clamp_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,188,95,192, //vmaxps %ymm0,%ymm8,%ymm0
- 196,98,125,88,0, //vpbroadcastd (%rax),%ymm8
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,65,61,254,193, //vpaddd %ymm9,%ymm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,188,95,201, //vmaxps %ymm1,%ymm8,%ymm1
- 196,98,125,88,0, //vpbroadcastd (%rax),%ymm8
- 196,65,53,118,201, //vpcmpeqd %ymm9,%ymm9,%ymm9
- 196,65,61,254,193, //vpaddd %ymm9,%ymm8,%ymm8
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,124,94,200, //vdivps %ymm8,%ymm0,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,98,61,172,200, //vfnmadd213ps %ymm0,%ymm8,%ymm9
- 197,253,118,192, //vpcmpeqd %ymm0,%ymm0,%ymm0
- 197,189,254,192, //vpaddd %ymm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,116,94,200, //vdivps %ymm8,%ymm1,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,98,61,172,201, //vfnmadd213ps %ymm1,%ymm8,%ymm9
- 197,245,118,201, //vpcmpeqd %ymm1,%ymm1,%ymm1
- 197,189,254,201, //vpaddd %ymm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,122,16,0, //vmovss (%rax),%xmm8
- 196,66,125,24,200, //vbroadcastss %xmm8,%ymm9
- 196,65,124,92,209, //vsubps %ymm9,%ymm0,%ymm10
- 196,193,58,88,192, //vaddss %xmm8,%xmm8,%xmm0
- 196,226,125,24,192, //vbroadcastss %xmm0,%ymm0
- 197,44,94,192, //vdivps %ymm0,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 196,66,125,172,194, //vfnmadd213ps %ymm10,%ymm0,%ymm8
- 196,193,60,92,193, //vsubps %ymm9,%ymm8,%ymm0
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,192, //vsubps %ymm0,%ymm8,%ymm8
- 197,188,84,192, //vandps %ymm0,%ymm8,%ymm0
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 196,65,53,254,192, //vpaddd %ymm8,%ymm9,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,122,16,0, //vmovss (%rax),%xmm8
- 196,66,125,24,200, //vbroadcastss %xmm8,%ymm9
- 196,65,116,92,209, //vsubps %ymm9,%ymm1,%ymm10
- 196,193,58,88,200, //vaddss %xmm8,%xmm8,%xmm1
- 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
- 197,44,94,193, //vdivps %ymm1,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 196,66,117,172,194, //vfnmadd213ps %ymm10,%ymm1,%ymm8
- 196,193,60,92,201, //vsubps %ymm9,%ymm8,%ymm1
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,193, //vsubps %ymm1,%ymm8,%ymm8
- 197,188,84,201, //vandps %ymm1,%ymm8,%ymm1
- 196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
- 196,65,53,254,192, //vpaddd %ymm8,%ymm9,%ymm8
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_hsw[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,98,125,88,195, //vpbroadcastd %xmm3,%ymm8
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,98,125,168,193, //vfmadd213ps %ymm1,%ymm0,%ymm8
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 196,226,125,88,216, //vpbroadcastd %xmm0,%ymm3
- 196,194,109,168,216, //vfmadd213ps %ymm8,%ymm2,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,253,239,192, //vpxor %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,64,16, //vbroadcastss 0x10(%rax),%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm11
- 196,98,125,24,72,20, //vbroadcastss 0x14(%rax),%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,12, //vbroadcastss 0xc(%rax),%ymm10
- 196,98,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm11
- 196,98,125,24,64,36, //vbroadcastss 0x24(%rax),%ymm8
- 196,66,109,184,195, //vfmadd231ps %ymm11,%ymm2,%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,16, //vbroadcastss 0x10(%rax),%ymm11
- 196,98,125,24,96,28, //vbroadcastss 0x1c(%rax),%ymm12
- 196,98,125,24,72,40, //vbroadcastss 0x28(%rax),%ymm9
- 196,66,109,184,204, //vfmadd231ps %ymm12,%ymm2,%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 196,98,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm11
- 196,98,125,24,96,20, //vbroadcastss 0x14(%rax),%ymm12
- 196,98,125,24,104,32, //vbroadcastss 0x20(%rax),%ymm13
- 196,98,125,24,80,44, //vbroadcastss 0x2c(%rax),%ymm10
- 196,66,109,184,213, //vfmadd231ps %ymm13,%ymm2,%ymm10
- 196,66,117,184,212, //vfmadd231ps %ymm12,%ymm1,%ymm10
- 196,66,125,184,211, //vfmadd231ps %ymm11,%ymm0,%ymm10
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,8, //vbroadcastss (%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,32, //vbroadcastss 0x20(%rax),%ymm11
- 196,98,125,24,96,48, //vbroadcastss 0x30(%rax),%ymm12
- 196,98,125,24,64,64, //vbroadcastss 0x40(%rax),%ymm8
- 196,66,101,184,196, //vfmadd231ps %ymm12,%ymm3,%ymm8
- 196,66,109,184,195, //vfmadd231ps %ymm11,%ymm2,%ymm8
- 196,66,117,184,194, //vfmadd231ps %ymm10,%ymm1,%ymm8
- 196,66,125,184,193, //vfmadd231ps %ymm9,%ymm0,%ymm8
- 196,98,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,98,125,24,96,36, //vbroadcastss 0x24(%rax),%ymm12
- 196,98,125,24,104,52, //vbroadcastss 0x34(%rax),%ymm13
- 196,98,125,24,72,68, //vbroadcastss 0x44(%rax),%ymm9
- 196,66,101,184,205, //vfmadd231ps %ymm13,%ymm3,%ymm9
- 196,66,109,184,204, //vfmadd231ps %ymm12,%ymm2,%ymm9
- 196,66,117,184,203, //vfmadd231ps %ymm11,%ymm1,%ymm9
- 196,66,125,184,202, //vfmadd231ps %ymm10,%ymm0,%ymm9
- 196,98,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm11
- 196,98,125,24,96,24, //vbroadcastss 0x18(%rax),%ymm12
- 196,98,125,24,104,40, //vbroadcastss 0x28(%rax),%ymm13
- 196,98,125,24,112,56, //vbroadcastss 0x38(%rax),%ymm14
- 196,98,125,24,80,72, //vbroadcastss 0x48(%rax),%ymm10
- 196,66,101,184,214, //vfmadd231ps %ymm14,%ymm3,%ymm10
- 196,66,109,184,213, //vfmadd231ps %ymm13,%ymm2,%ymm10
- 196,66,117,184,212, //vfmadd231ps %ymm12,%ymm1,%ymm10
- 196,66,125,184,211, //vfmadd231ps %ymm11,%ymm0,%ymm10
- 196,98,125,24,96,12, //vbroadcastss 0xc(%rax),%ymm12
- 196,98,125,24,104,28, //vbroadcastss 0x1c(%rax),%ymm13
- 196,98,125,24,112,44, //vbroadcastss 0x2c(%rax),%ymm14
- 196,98,125,24,120,60, //vbroadcastss 0x3c(%rax),%ymm15
- 196,98,125,24,88,76, //vbroadcastss 0x4c(%rax),%ymm11
- 196,66,101,184,223, //vfmadd231ps %ymm15,%ymm3,%ymm11
- 196,66,109,184,222, //vfmadd231ps %ymm14,%ymm2,%ymm11
- 196,66,117,184,221, //vfmadd231ps %ymm13,%ymm1,%ymm11
- 196,66,125,184,220, //vfmadd231ps %ymm12,%ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 197,124,41,219, //vmovaps %ymm11,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,66,117,184,209, //vfmadd231ps %ymm9,%ymm1,%ymm10
- 196,66,125,184,208, //vfmadd231ps %ymm8,%ymm0,%ymm10
- 196,98,125,24,64,12, //vbroadcastss 0xc(%rax),%ymm8
- 196,98,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm9
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,66,117,184,217, //vfmadd231ps %ymm9,%ymm1,%ymm11
- 196,66,125,184,216, //vfmadd231ps %ymm8,%ymm0,%ymm11
- 196,98,125,24,64,24, //vbroadcastss 0x18(%rax),%ymm8
- 196,98,125,24,72,28, //vbroadcastss 0x1c(%rax),%ymm9
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 196,66,117,184,225, //vfmadd231ps %ymm9,%ymm1,%ymm12
- 196,66,125,184,224, //vfmadd231ps %ymm8,%ymm0,%ymm12
- 196,193,124,83,204, //vrcpps %ymm12,%ymm1
- 197,172,89,193, //vmulps %ymm1,%ymm10,%ymm0
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_hsw[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm1
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,184,193, //vfmadd231ps %ymm1,%ymm0,%ymm8
- 196,226,125,24,80,20, //vbroadcastss 0x14(%rax),%ymm2
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,184,202, //vfmadd231ps %ymm2,%ymm0,%ymm1
- 196,226,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm3
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,184,211, //vfmadd231ps %ymm3,%ymm0,%ymm2
- 196,98,125,24,72,28, //vbroadcastss 0x1c(%rax),%ymm9
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 196,194,125,184,217, //vfmadd231ps %ymm9,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_avx[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 86, //push %rsi
- 87, //push %rdi
- 83, //push %rbx
- 72,129,236,160,0,0,0, //sub $0xa0,%rsp
- 197,120,41,188,36,144,0,0,0, //vmovaps %xmm15,0x90(%rsp)
- 197,120,41,180,36,128,0,0,0, //vmovaps %xmm14,0x80(%rsp)
- 197,120,41,108,36,112, //vmovaps %xmm13,0x70(%rsp)
- 197,120,41,100,36,96, //vmovaps %xmm12,0x60(%rsp)
- 197,120,41,92,36,80, //vmovaps %xmm11,0x50(%rsp)
- 197,120,41,84,36,64, //vmovaps %xmm10,0x40(%rsp)
- 197,120,41,76,36,48, //vmovaps %xmm9,0x30(%rsp)
- 197,120,41,68,36,32, //vmovaps %xmm8,0x20(%rsp)
- 197,248,41,124,36,16, //vmovaps %xmm7,0x10(%rsp)
- 197,248,41,52,36, //vmovaps %xmm6,(%rsp)
- 77,137,205, //mov %r9,%r13
- 77,137,198, //mov %r8,%r14
- 72,137,203, //mov %rcx,%rbx
- 72,137,214, //mov %rdx,%rsi
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,199, //mov %rax,%r15
- 73,137,244, //mov %rsi,%r12
- 72,141,67,8, //lea 0x8(%rbx),%rax
- 76,57,232, //cmp %r13,%rax
- 118,5, //jbe 75 <_sk_start_pipeline_avx+0x75>
- 72,137,223, //mov %rbx,%rdi
- 235,65, //jmp b6 <_sk_start_pipeline_avx+0xb6>
- 185,0,0,0,0, //mov $0x0,%ecx
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 72,141,123,8, //lea 0x8(%rbx),%rdi
- 72,131,195,16, //add $0x10,%rbx
- 76,57,235, //cmp %r13,%rbx
- 72,137,251, //mov %rdi,%rbx
- 118,191, //jbe 75 <_sk_start_pipeline_avx+0x75>
- 76,137,233, //mov %r13,%rcx
- 72,41,249, //sub %rdi,%rcx
- 116,41, //je e7 <_sk_start_pipeline_avx+0xe7>
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 76,137,230, //mov %r12,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,215, //callq *%r15
- 76,137,232, //mov %r13,%rax
- 197,248,40,52,36, //vmovaps (%rsp),%xmm6
- 197,248,40,124,36,16, //vmovaps 0x10(%rsp),%xmm7
- 197,120,40,68,36,32, //vmovaps 0x20(%rsp),%xmm8
- 197,120,40,76,36,48, //vmovaps 0x30(%rsp),%xmm9
- 197,120,40,84,36,64, //vmovaps 0x40(%rsp),%xmm10
- 197,120,40,92,36,80, //vmovaps 0x50(%rsp),%xmm11
- 197,120,40,100,36,96, //vmovaps 0x60(%rsp),%xmm12
- 197,120,40,108,36,112, //vmovaps 0x70(%rsp),%xmm13
- 197,120,40,180,36,128,0,0,0, //vmovaps 0x80(%rsp),%xmm14
- 197,120,40,188,36,144,0,0,0, //vmovaps 0x90(%rsp),%xmm15
- 72,129,196,160,0,0,0, //add $0xa0,%rsp
- 91, //pop %rbx
- 95, //pop %rdi
- 94, //pop %rsi
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 197,248,119, //vzeroupper
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_avx[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,249,110,199, //vmovd %edi,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 65,184,0,0,0,63, //mov $0x3f000000,%r8d
- 196,193,121,110,200, //vmovd %r8d,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 196,227,121,4,210,0, //vpermilps $0x0,%xmm2,%xmm2
- 196,227,109,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 197,220,87,228, //vxorps %ymm4,%ymm4,%ymm4
- 197,212,87,237, //vxorps %ymm5,%ymm5,%ymm5
- 197,204,87,246, //vxorps %ymm6,%ymm6,%ymm6
- 197,196,87,255, //vxorps %ymm7,%ymm7,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 196,226,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 197,228,87,219, //vxorps %ymm3,%ymm3,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__avx[] = {
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,92,195, //vsubps %ymm3,%ymm8,%ymm8
- 197,60,89,204, //vmulps %ymm4,%ymm8,%ymm9
- 197,180,88,192, //vaddps %ymm0,%ymm9,%ymm0
- 197,60,89,205, //vmulps %ymm5,%ymm8,%ymm9
- 197,180,88,201, //vaddps %ymm1,%ymm9,%ymm1
- 197,60,89,206, //vmulps %ymm6,%ymm8,%ymm9
- 197,180,88,210, //vaddps %ymm2,%ymm9,%ymm2
- 197,60,89,199, //vmulps %ymm7,%ymm8,%ymm8
- 197,188,88,219, //vaddps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,92,199, //vsubps %ymm7,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_avx[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,193,124,95,192, //vmaxps %ymm8,%ymm0,%ymm0
- 196,193,116,95,200, //vmaxps %ymm8,%ymm1,%ymm1
- 196,193,108,95,208, //vmaxps %ymm8,%ymm2,%ymm2
- 196,193,100,95,216, //vmaxps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 196,193,124,93,192, //vminps %ymm8,%ymm0,%ymm0
- 196,193,116,93,200, //vminps %ymm8,%ymm1,%ymm1
- 196,193,108,93,208, //vminps %ymm8,%ymm2,%ymm2
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_avx[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 196,193,100,93,216, //vminps %ymm8,%ymm3,%ymm3
- 197,252,93,195, //vminps %ymm3,%ymm0,%ymm0
- 197,244,93,203, //vminps %ymm3,%ymm1,%ymm1
- 197,236,93,211, //vminps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,0, //vbroadcastss (%rax),%ymm0
- 196,226,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm1
- 196,226,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_avx[] = {
- 197,124,40,192, //vmovaps %ymm0,%ymm8
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,194, //vmovaps %ymm2,%ymm0
- 197,124,41,194, //vmovaps %ymm8,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_avx[] = {
- 197,124,40,195, //vmovaps %ymm3,%ymm8
- 197,124,40,202, //vmovaps %ymm2,%ymm9
- 197,124,40,209, //vmovaps %ymm1,%ymm10
- 197,124,40,216, //vmovaps %ymm0,%ymm11
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 197,124,41,220, //vmovaps %ymm11,%ymm4
- 197,124,41,213, //vmovaps %ymm10,%ymm5
- 197,124,41,206, //vmovaps %ymm9,%ymm6
- 197,124,41,199, //vmovaps %ymm8,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,224, //vmovaps %ymm0,%ymm4
- 197,252,40,233, //vmovaps %ymm1,%ymm5
- 197,252,40,242, //vmovaps %ymm2,%ymm6
- 197,252,40,251, //vmovaps %ymm3,%ymm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,40,196, //vmovaps %ymm4,%ymm0
- 197,252,40,205, //vmovaps %ymm5,%ymm1
- 197,252,40,214, //vmovaps %ymm6,%ymm2
- 197,252,40,223, //vmovaps %ymm7,%ymm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_avx[] = {
- 197,252,89,195, //vmulps %ymm3,%ymm0,%ymm0
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_avx[] = {
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 196,65,100,194,200,0, //vcmpeqps %ymm8,%ymm3,%ymm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 197,44,94,211, //vdivps %ymm3,%ymm10,%ymm10
- 196,67,45,74,192,144, //vblendvps %ymm9,%ymm8,%ymm10,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_avx[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
- 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 197,121,110,224, //vmovd %eax,%xmm12
- 196,67,121,4,228,0, //vpermilps $0x0,%xmm12,%xmm12
- 196,67,29,24,228,1, //vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
- 197,36,89,232, //vmulps %ymm0,%ymm11,%ymm13
- 196,65,20,88,236, //vaddps %ymm12,%ymm13,%ymm13
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
- 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
- 196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
- 196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 197,121,110,232, //vmovd %eax,%xmm13
- 196,67,121,4,237,0, //vpermilps $0x0,%xmm13,%xmm13
- 196,67,21,24,237,1, //vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
- 196,193,124,194,197,1, //vcmpltps %ymm13,%ymm0,%ymm0
- 196,195,45,74,193,0, //vblendvps %ymm0,%ymm9,%ymm10,%ymm0
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 197,116,89,209, //vmulps %ymm1,%ymm1,%ymm10
- 197,36,89,249, //vmulps %ymm1,%ymm11,%ymm15
- 196,65,28,88,255, //vaddps %ymm15,%ymm12,%ymm15
- 196,65,44,89,215, //vmulps %ymm15,%ymm10,%ymm10
- 196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 196,193,116,194,205,1, //vcmpltps %ymm13,%ymm1,%ymm1
- 196,195,45,74,201,16, //vblendvps %ymm1,%ymm9,%ymm10,%ymm1
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
- 197,36,89,210, //vmulps %ymm2,%ymm11,%ymm10
- 196,65,28,88,210, //vaddps %ymm10,%ymm12,%ymm10
- 196,65,52,89,202, //vmulps %ymm10,%ymm9,%ymm9
- 196,65,12,88,201, //vaddps %ymm9,%ymm14,%ymm9
- 196,193,108,194,213,1, //vcmpltps %ymm13,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_avx[] = {
- 197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,232, //vrcpps %ymm8,%ymm13
- 196,65,124,82,240, //vrsqrtps %ymm8,%ymm14
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,224, //vmulps %ymm0,%ymm8,%ymm12
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 197,121,110,216, //vmovd %eax,%xmm11
- 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
- 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 197,121,110,248, //vmovd %eax,%xmm15
- 196,67,121,4,255,0, //vpermilps $0x0,%xmm15,%xmm15
- 196,67,5,24,255,1, //vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
- 196,65,20,89,235, //vmulps %ymm11,%ymm13,%ymm13
- 196,65,20,88,239, //vaddps %ymm15,%ymm13,%ymm13
- 196,65,12,89,242, //vmulps %ymm10,%ymm14,%ymm14
- 196,65,12,88,237, //vaddps %ymm13,%ymm14,%ymm13
- 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 197,121,110,240, //vmovd %eax,%xmm14
- 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
- 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
- 196,193,124,194,198,1, //vcmpltps %ymm14,%ymm0,%ymm0
- 196,195,21,74,196,0, //vblendvps %ymm0,%ymm12,%ymm13,%ymm0
- 197,124,82,225, //vrsqrtps %ymm1,%ymm12
- 196,65,124,83,236, //vrcpps %ymm12,%ymm13
- 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
- 196,65,36,89,237, //vmulps %ymm13,%ymm11,%ymm13
- 196,65,4,88,237, //vaddps %ymm13,%ymm15,%ymm13
- 196,65,44,89,228, //vmulps %ymm12,%ymm10,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,60,89,233, //vmulps %ymm1,%ymm8,%ymm13
- 196,65,52,93,228, //vminps %ymm12,%ymm9,%ymm12
- 196,193,116,194,206,1, //vcmpltps %ymm14,%ymm1,%ymm1
- 196,195,29,74,205,16, //vblendvps %ymm1,%ymm13,%ymm12,%ymm1
- 197,124,82,226, //vrsqrtps %ymm2,%ymm12
- 196,65,124,83,236, //vrcpps %ymm12,%ymm13
- 196,65,36,89,221, //vmulps %ymm13,%ymm11,%ymm11
- 196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
- 196,65,44,89,212, //vmulps %ymm12,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 196,65,52,93,202, //vminps %ymm10,%ymm9,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,214,1, //vcmpltps %ymm14,%ymm2,%ymm2
- 196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,80, //jne 639 <_sk_scale_u8_avx+0x60>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
- 196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
- 196,66,121,49,192, //vpmovzxbd %xmm8,%xmm8
- 196,67,53,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,188,89,210, //vmulps %ymm2,%ymm8,%ymm2
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 641 <_sk_scale_u8_avx+0x68>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 235,143, //jmp 5ed <_sk_scale_u8_avx+0x14>
-};
-
-CODE const uint8_t sk_lerp_1_float_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,116, //jne 721 <_sk_lerp_u8_avx+0x84>
- 197,122,126,0, //vmovq (%rax),%xmm8
- 196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
- 196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
- 196,66,121,49,192, //vpmovzxbd %xmm8,%xmm8
- 196,67,53,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,121,110,200, //vmovd %eax,%xmm9
- 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,60,89,193, //vmulps %ymm9,%ymm8,%ymm8
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 197,228,92,223, //vsubps %ymm7,%ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 197,228,88,223, //vaddps %ymm7,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne 729 <_sk_lerp_u8_avx+0x8c>
- 196,65,249,110,193, //vmovq %r9,%xmm8
- 233,104,255,255,255, //jmpq 6b1 <_sk_lerp_u8_avx+0x14>
-};
-
-CODE const uint8_t sk_lerp_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,250,0,0,0, //jne 851 <_sk_lerp_565_avx+0x108>
- 196,65,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm8
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,185,105,219, //vpunpckhwd %xmm3,%xmm8,%xmm3
- 196,66,121,51,192, //vpmovzxwd %xmm8,%xmm8
- 196,99,61,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,193, //vmulps %ymm9,%ymm0,%ymm0
- 197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
- 197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,202, //vmulps %ymm10,%ymm1,%ymm1
- 197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
- 197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 197,236,88,214, //vaddps %ymm6,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 196,65,57,239,192, //vpxor %xmm8,%xmm8,%xmm8
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,243,254,255,255, //ja 75d <_sk_lerp_565_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,75,0,0,0, //lea 0x4b(%rip),%r9 # 8c0 <_sk_lerp_565_avx+0x177>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 196,65,97,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm8
- 196,65,57,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
- 196,65,57,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
- 233,159,254,255,255, //jmpq 75d <_sk_lerp_565_avx+0x14>
- 102,144, //xchg %ax,%ax
- 242,255, //repnz (bad)
- 255, //(bad)
- 255, //(bad)
- 234, //(bad)
- 255, //(bad)
- 255, //(bad)
- 255,226, //jmpq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 218,255, //(bad)
- 255, //(bad)
- 255,210, //callq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255,202, //dec %edx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 190, //.byte 0xbe
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_tables_avx[] = {
- 85, //push %rbp
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 83, //push %rbx
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,133,201, //test %rcx,%rcx
- 15,133,56,2,0,0, //jne b2c <_sk_load_tables_avx+0x250>
- 196,65,124,16,4,184, //vmovups (%r8,%rdi,4),%ymm8
- 187,255,0,0,0, //mov $0xff,%ebx
- 197,249,110,195, //vmovd %ebx,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,99,125,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
- 196,193,52,84,192, //vandps %ymm8,%ymm9,%ymm0
- 196,193,249,126,193, //vmovq %xmm0,%r9
- 69,137,203, //mov %r9d,%r11d
- 196,195,249,22,194,1, //vpextrq $0x1,%xmm0,%r10
- 69,137,214, //mov %r10d,%r14d
- 73,193,234,32, //shr $0x20,%r10
- 73,193,233,32, //shr $0x20,%r9
- 196,227,125,25,192,1, //vextractf128 $0x1,%ymm0,%xmm0
- 196,193,249,126,196, //vmovq %xmm0,%r12
- 69,137,231, //mov %r12d,%r15d
- 196,227,249,22,195,1, //vpextrq $0x1,%xmm0,%rbx
- 65,137,221, //mov %ebx,%r13d
- 72,193,235,32, //shr $0x20,%rbx
- 73,193,236,32, //shr $0x20,%r12
- 72,139,104,8, //mov 0x8(%rax),%rbp
- 76,139,64,16, //mov 0x10(%rax),%r8
- 196,161,122,16,68,189,0, //vmovss 0x0(%rbp,%r15,4),%xmm0
- 196,163,121,33,68,165,0,16, //vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
- 196,161,122,16,76,173,0, //vmovss 0x0(%rbp,%r13,4),%xmm1
- 196,227,121,33,193,32, //vinsertps $0x20,%xmm1,%xmm0,%xmm0
- 197,250,16,76,157,0, //vmovss 0x0(%rbp,%rbx,4),%xmm1
- 196,227,121,33,193,48, //vinsertps $0x30,%xmm1,%xmm0,%xmm0
- 196,161,122,16,76,157,0, //vmovss 0x0(%rbp,%r11,4),%xmm1
- 196,163,113,33,76,141,0,16, //vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
- 196,161,122,16,92,181,0, //vmovss 0x0(%rbp,%r14,4),%xmm3
- 196,227,113,33,203,32, //vinsertps $0x20,%xmm3,%xmm1,%xmm1
- 196,161,122,16,92,149,0, //vmovss 0x0(%rbp,%r10,4),%xmm3
- 196,227,113,33,203,48, //vinsertps $0x30,%xmm3,%xmm1,%xmm1
- 196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
- 196,193,113,114,208,8, //vpsrld $0x8,%xmm8,%xmm1
- 196,67,125,25,194,1, //vextractf128 $0x1,%ymm8,%xmm10
- 196,193,105,114,210,8, //vpsrld $0x8,%xmm10,%xmm2
- 196,227,117,24,202,1, //vinsertf128 $0x1,%xmm2,%ymm1,%ymm1
- 197,180,84,201, //vandps %ymm1,%ymm9,%ymm1
- 196,193,249,126,201, //vmovq %xmm1,%r9
- 69,137,203, //mov %r9d,%r11d
- 196,195,249,22,202,1, //vpextrq $0x1,%xmm1,%r10
- 69,137,214, //mov %r10d,%r14d
- 73,193,234,32, //shr $0x20,%r10
- 73,193,233,32, //shr $0x20,%r9
- 196,227,125,25,201,1, //vextractf128 $0x1,%ymm1,%xmm1
- 196,225,249,126,205, //vmovq %xmm1,%rbp
- 65,137,239, //mov %ebp,%r15d
- 196,227,249,22,203,1, //vpextrq $0x1,%xmm1,%rbx
- 65,137,220, //mov %ebx,%r12d
- 72,193,235,32, //shr $0x20,%rbx
- 72,193,237,32, //shr $0x20,%rbp
- 196,129,122,16,12,184, //vmovss (%r8,%r15,4),%xmm1
- 196,195,113,33,12,168,16, //vinsertps $0x10,(%r8,%rbp,4),%xmm1,%xmm1
- 196,129,122,16,20,160, //vmovss (%r8,%r12,4),%xmm2
- 196,227,113,33,202,32, //vinsertps $0x20,%xmm2,%xmm1,%xmm1
- 196,193,122,16,20,152, //vmovss (%r8,%rbx,4),%xmm2
- 196,227,113,33,202,48, //vinsertps $0x30,%xmm2,%xmm1,%xmm1
- 196,129,122,16,20,152, //vmovss (%r8,%r11,4),%xmm2
- 196,131,105,33,20,136,16, //vinsertps $0x10,(%r8,%r9,4),%xmm2,%xmm2
- 196,129,122,16,28,176, //vmovss (%r8,%r14,4),%xmm3
- 196,227,105,33,211,32, //vinsertps $0x20,%xmm3,%xmm2,%xmm2
- 196,129,122,16,28,144, //vmovss (%r8,%r10,4),%xmm3
- 196,227,105,33,211,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm2
- 196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
- 72,139,64,24, //mov 0x18(%rax),%rax
- 196,193,105,114,208,16, //vpsrld $0x10,%xmm8,%xmm2
- 196,193,97,114,210,16, //vpsrld $0x10,%xmm10,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
- 197,180,84,210, //vandps %ymm2,%ymm9,%ymm2
- 196,193,249,126,208, //vmovq %xmm2,%r8
- 69,137,194, //mov %r8d,%r10d
- 196,195,249,22,209,1, //vpextrq $0x1,%xmm2,%r9
- 69,137,203, //mov %r9d,%r11d
- 73,193,233,32, //shr $0x20,%r9
- 73,193,232,32, //shr $0x20,%r8
- 196,227,125,25,210,1, //vextractf128 $0x1,%ymm2,%xmm2
- 196,225,249,126,213, //vmovq %xmm2,%rbp
- 65,137,238, //mov %ebp,%r14d
- 196,227,249,22,211,1, //vpextrq $0x1,%xmm2,%rbx
- 65,137,223, //mov %ebx,%r15d
- 72,193,235,32, //shr $0x20,%rbx
- 72,193,237,32, //shr $0x20,%rbp
- 196,161,122,16,20,176, //vmovss (%rax,%r14,4),%xmm2
- 196,227,105,33,20,168,16, //vinsertps $0x10,(%rax,%rbp,4),%xmm2,%xmm2
- 196,161,122,16,28,184, //vmovss (%rax,%r15,4),%xmm3
- 196,227,105,33,211,32, //vinsertps $0x20,%xmm3,%xmm2,%xmm2
- 197,250,16,28,152, //vmovss (%rax,%rbx,4),%xmm3
- 196,99,105,33,203,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm9
- 196,161,122,16,28,144, //vmovss (%rax,%r10,4),%xmm3
- 196,163,97,33,28,128,16, //vinsertps $0x10,(%rax,%r8,4),%xmm3,%xmm3
- 196,161,122,16,20,152, //vmovss (%rax,%r11,4),%xmm2
- 196,227,97,33,210,32, //vinsertps $0x20,%xmm2,%xmm3,%xmm2
- 196,161,122,16,28,136, //vmovss (%rax,%r9,4),%xmm3
- 196,227,105,33,211,48, //vinsertps $0x30,%xmm3,%xmm2,%xmm2
- 196,195,109,24,209,1, //vinsertf128 $0x1,%xmm9,%ymm2,%ymm2
- 196,193,57,114,208,24, //vpsrld $0x18,%xmm8,%xmm8
- 196,193,97,114,210,24, //vpsrld $0x18,%xmm10,%xmm3
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 91, //pop %rbx
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 93, //pop %rbp
- 255,224, //jmpq *%rax
- 137,203, //mov %ecx,%ebx
- 128,227,7, //and $0x7,%bl
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 254,203, //dec %bl
- 128,251,6, //cmp $0x6,%bl
- 15,135,185,253,255,255, //ja 8fa <_sk_load_tables_avx+0x1e>
- 15,182,219, //movzbl %bl,%ebx
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # bd4 <_sk_load_tables_avx+0x2f8>
- 73,99,28,153, //movslq (%r9,%rbx,4),%rbx
- 76,1,203, //add %r9,%rbx
- 255,227, //jmpq *%rbx
- 196,193,121,110,68,184,24, //vmovd 0x18(%r8,%rdi,4),%xmm0
- 197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 196,99,117,12,192,64, //vblendps $0x40,%ymm0,%ymm1,%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,195,121,34,68,184,20,1, //vpinsrd $0x1,0x14(%r8,%rdi,4),%xmm0,%xmm0
- 196,99,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,195,121,34,68,184,16,0, //vpinsrd $0x0,0x10(%r8,%rdi,4),%xmm0,%xmm0
- 196,99,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
- 196,195,57,34,68,184,12,3, //vpinsrd $0x3,0xc(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,68,184,8,2, //vpinsrd $0x2,0x8(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,68,184,4,1, //vpinsrd $0x1,0x4(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 196,195,57,34,4,184,0, //vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
- 196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 233,38,253,255,255, //jmpq 8fa <_sk_load_tables_avx+0x1e>
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,224, //jmpq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,210, //callq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255,176,255,255,255,156, //pushq -0x63000001(%rax)
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 128,255,255, //cmp $0xff,%bh
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_a8_avx[] = {
- 73,137,200, //mov %rcx,%r8
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,1,248, //add %rdi,%rax
- 77,133,192, //test %r8,%r8
- 117,74, //jne c4a <_sk_load_a8_avx+0x5a>
- 197,250,126,0, //vmovq (%rax),%xmm0
- 196,226,121,49,200, //vpmovzxbd %xmm0,%xmm1
- 196,227,121,4,192,229, //vpermilps $0xe5,%xmm0,%xmm0
- 196,226,121,49,192, //vpmovzxbd %xmm0,%xmm0
- 196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 76,137,193, //mov %r8,%rcx
- 255,224, //jmpq *%rax
- 49,201, //xor %ecx,%ecx
- 77,137,194, //mov %r8,%r10
- 69,49,201, //xor %r9d,%r9d
- 68,15,182,24, //movzbl (%rax),%r11d
- 72,255,192, //inc %rax
- 73,211,227, //shl %cl,%r11
- 77,9,217, //or %r11,%r9
- 72,131,193,8, //add $0x8,%rcx
- 73,255,202, //dec %r10
- 117,234, //jne c52 <_sk_load_a8_avx+0x62>
- 196,193,249,110,193, //vmovq %r9,%xmm0
- 235,149, //jmp c04 <_sk_load_a8_avx+0x14>
-};
-
-CODE const uint8_t sk_store_a8_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne cb1 <_sk_store_a8_avx+0x42>
- 196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja cad <_sk_store_a8_avx+0x3e>
- 196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # d14 <_sk_store_a8_avx+0xa5>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,20,68,57,6,12, //vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
- 196,67,121,20,68,57,5,10, //vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
- 196,67,121,20,68,57,4,8, //vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
- 196,67,121,20,68,57,3,6, //vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
- 196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
- 196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
- 196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,154, //jmp cad <_sk_store_a8_avx+0x3e>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,209,0,0,0, //jne e0f <_sk_load_565_avx+0xdf>
- 196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
- 197,241,239,201, //vpxor %xmm1,%xmm1,%xmm1
- 197,249,105,201, //vpunpckhwd %xmm1,%xmm0,%xmm1
- 196,226,121,51,192, //vpmovzxwd %xmm0,%xmm0
- 196,227,125,24,209,1, //vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,252,84,194, //vandps %ymm2,%ymm0,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,8,33,132,55, //mov $0x37842108,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 197,249,112,201,0, //vpshufd $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,244,84,202, //vandps %ymm2,%ymm1,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,84,210, //vandps %ymm2,%ymm3,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,29,255,255,255, //ja d44 <_sk_load_565_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # e7c <_sk_load_565_avx+0x14c>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 197,249,239,192, //vpxor %xmm0,%xmm0,%xmm0
- 196,193,121,196,68,122,12,6, //vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,10,5, //vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,8,4, //vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,6,3, //vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
- 196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,201,254,255,255, //jmpq d44 <_sk_load_565_avx+0x14>
- 144, //nop
- 243,255, //repz (bad)
- 255, //(bad)
- 255, //(bad)
- 235,255, //jmp e81 <_sk_load_565_avx+0x151>
- 255, //(bad)
- 255,227, //jmpq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 219,255, //(bad)
- 255, //(bad)
- 255,211, //callq *%rbx
- 255, //(bad)
- 255, //(bad)
- 255,203, //dec %ebx
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 191, //.byte 0xbf
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_565_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,248,65, //mov $0x41f80000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 196,193,41,114,241,11, //vpslld $0xb,%xmm9,%xmm10
- 196,67,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm9
- 196,193,49,114,241,11, //vpslld $0xb,%xmm9,%xmm9
- 196,67,45,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
- 184,0,0,124,66, //mov $0x427c0000,%eax
- 197,121,110,208, //vmovd %eax,%xmm10
- 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
- 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,5, //vpslld $0x5,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,5, //vpslld $0x5,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 196,65,45,86,201, //vorpd %ymm9,%ymm10,%ymm9
- 197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne f36 <_sk_store_565_avx+0x9e>
- 196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja f32 <_sk_store_565_avx+0x9a>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,67,0,0,0, //lea 0x43(%rip),%r8 # f94 <_sk_store_565_avx+0xfc>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,121,21,68,121,12,6, //vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
- 196,67,121,21,68,121,10,5, //vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
- 196,67,121,21,68,121,8,4, //vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
- 196,67,121,21,68,121,6,3, //vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
- 196,67,121,21,68,121,4,2, //vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
- 196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
- 196,67,121,21,4,121,0, //vpextrw $0x0,%xmm8,(%r9,%rdi,2)
- 235,159, //jmp f32 <_sk_store_565_avx+0x9a>
- 144, //nop
- 246,255, //idiv %bh
- 255, //(bad)
- 255, //(bad)
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,230, //jmpq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 222,255, //fdivrp %st,%st(7)
- 255, //(bad)
- 255,214, //callq *%rsi
- 255, //(bad)
- 255, //(bad)
- 255,206, //dec %esi
- 255, //(bad)
- 255, //(bad)
- 255,198, //inc %esi
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_8888_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,16, //mov (%rax),%r10
- 72,133,201, //test %rcx,%rcx
- 15,133,157,0,0,0, //jne 105b <_sk_load_8888_avx+0xab>
- 196,65,124,16,12,186, //vmovups (%r10,%rdi,4),%ymm9
- 184,255,0,0,0, //mov $0xff,%eax
- 197,249,110,192, //vmovd %eax,%xmm0
- 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
- 196,99,125,24,216,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
- 196,193,36,84,193, //vandps %ymm9,%ymm11,%ymm0
- 197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,99,117,24,193,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm8
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
- 196,193,41,114,209,8, //vpsrld $0x8,%xmm9,%xmm10
- 196,99,125,25,203,1, //vextractf128 $0x1,%ymm9,%xmm3
- 197,241,114,211,8, //vpsrld $0x8,%xmm3,%xmm1
- 196,227,45,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm10,%ymm1
- 197,164,84,201, //vandps %ymm1,%ymm11,%ymm1
- 197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,193,116,89,200, //vmulps %ymm8,%ymm1,%ymm1
- 196,193,41,114,209,16, //vpsrld $0x10,%xmm9,%xmm10
- 197,233,114,211,16, //vpsrld $0x10,%xmm3,%xmm2
- 196,227,45,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm10,%ymm2
- 197,164,84,210, //vandps %ymm2,%ymm11,%ymm2
- 197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,193,108,89,208, //vmulps %ymm8,%ymm2,%ymm2
- 196,193,49,114,209,24, //vpsrld $0x18,%xmm9,%xmm9
- 197,225,114,211,24, //vpsrld $0x18,%xmm3,%xmm3
- 196,227,53,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm9,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 196,65,52,87,201, //vxorps %ymm9,%ymm9,%ymm9
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 15,135,80,255,255,255, //ja fc4 <_sk_load_8888_avx+0x14>
- 69,15,182,192, //movzbl %r8b,%r8d
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # 1108 <_sk_load_8888_avx+0x158>
- 75,99,4,129, //movslq (%r9,%r8,4),%rax
- 76,1,200, //add %r9,%rax
- 255,224, //jmpq *%rax
- 196,193,121,110,68,186,24, //vmovd 0x18(%r10,%rdi,4),%xmm0
- 197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 196,99,117,12,200,64, //vblendps $0x40,%ymm0,%ymm1,%ymm9
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,195,121,34,68,186,20,1, //vpinsrd $0x1,0x14(%r10,%rdi,4),%xmm0,%xmm0
- 196,99,53,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,195,121,34,68,186,16,0, //vpinsrd $0x0,0x10(%r10,%rdi,4),%xmm0,%xmm0
- 196,99,53,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
- 196,195,49,34,68,186,12,3, //vpinsrd $0x3,0xc(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,68,186,8,2, //vpinsrd $0x2,0x8(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,68,186,4,1, //vpinsrd $0x1,0x4(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 196,195,49,34,4,186,0, //vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
- 196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 233,188,254,255,255, //jmpq fc4 <_sk_load_8888_avx+0x14>
- 238, //out %al,(%dx)
- 255, //(bad)
- 255, //(bad)
- 255,224, //jmpq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,210, //callq *%rdx
- 255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
- 255, //(bad)
- 255, //(bad)
- 255,176,255,255,255,156, //pushq -0x63000001(%rax)
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 128,255,255, //cmp $0xff,%bh
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_store_8888_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,8, //mov (%rax),%r9
- 184,0,0,127,67, //mov $0x437f0000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
- 197,60,89,209, //vmulps %ymm1,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,8, //vpslld $0x8,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,8, //vpslld $0x8,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 196,65,45,86,201, //vorpd %ymm9,%ymm10,%ymm9
- 197,60,89,210, //vmulps %ymm2,%ymm8,%ymm10
- 196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
- 196,193,33,114,242,16, //vpslld $0x10,%xmm10,%xmm11
- 196,67,125,25,210,1, //vextractf128 $0x1,%ymm10,%xmm10
- 196,193,41,114,242,16, //vpslld $0x10,%xmm10,%xmm10
- 196,67,37,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
- 196,193,33,114,240,24, //vpslld $0x18,%xmm8,%xmm11
- 196,67,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm8
- 196,193,57,114,240,24, //vpslld $0x18,%xmm8,%xmm8
- 196,67,37,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
- 196,65,45,86,192, //vorpd %ymm8,%ymm10,%ymm8
- 196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
- 72,133,201, //test %rcx,%rcx
- 117,10, //jne 11c8 <_sk_store_8888_avx+0xa4>
- 196,65,124,17,4,185, //vmovups %ymm8,(%r9,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 65,137,200, //mov %ecx,%r8d
- 65,128,224,7, //and $0x7,%r8b
- 65,254,200, //dec %r8b
- 65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 11c4 <_sk_store_8888_avx+0xa0>
- 65,15,182,192, //movzbl %r8b,%eax
- 76,141,5,85,0,0,0, //lea 0x55(%rip),%r8 # 1238 <_sk_store_8888_avx+0x114>
- 73,99,4,128, //movslq (%r8,%rax,4),%rax
- 76,1,192, //add %r8,%rax
- 255,224, //jmpq *%rax
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,67,121,22,76,185,24,2, //vpextrd $0x2,%xmm9,0x18(%r9,%rdi,4)
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,67,121,22,76,185,20,1, //vpextrd $0x1,%xmm9,0x14(%r9,%rdi,4)
- 196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
- 196,65,122,17,76,185,16, //vmovss %xmm9,0x10(%r9,%rdi,4)
- 196,67,121,22,68,185,12,3, //vpextrd $0x3,%xmm8,0xc(%r9,%rdi,4)
- 196,67,121,22,68,185,8,2, //vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
- 196,67,121,22,68,185,4,1, //vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
- 196,65,121,126,4,185, //vmovd %xmm8,(%r9,%rdi,4)
- 235,143, //jmp 11c4 <_sk_store_8888_avx+0xa0>
- 15,31,0, //nopl (%rax)
- 245, //cmc
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 237, //in (%dx),%eax
- 255, //(bad)
- 255, //(bad)
- 255,229, //jmpq *%rbp
- 255, //(bad)
- 255, //(bad)
- 255, //(bad)
- 221,255, //(bad)
- 255, //(bad)
- 255,208, //callq *%rax
- 255, //(bad)
- 255, //(bad)
- 255,194, //inc %edx
- 255, //(bad)
- 255, //(bad)
- 255, //.byte 0xff
- 180,255, //mov $0xff,%ah
- 255, //(bad)
- 255, //.byte 0xff
-};
-
-CODE const uint8_t sk_load_f16_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,133,201, //test %rcx,%rcx
- 15,133,2,1,0,0, //jne 1364 <_sk_load_f16_avx+0x110>
- 197,121,16,4,248, //vmovupd (%rax,%rdi,8),%xmm8
- 197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
- 197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
- 197,122,111,76,248,48, //vmovdqu 0x30(%rax,%rdi,8),%xmm9
- 197,185,97,194, //vpunpcklwd %xmm2,%xmm8,%xmm0
- 197,185,105,210, //vpunpckhwd %xmm2,%xmm8,%xmm2
- 196,193,97,97,201, //vpunpcklwd %xmm9,%xmm3,%xmm1
- 196,193,97,105,217, //vpunpckhwd %xmm9,%xmm3,%xmm3
- 197,121,97,194, //vpunpcklwd %xmm2,%xmm0,%xmm8
- 197,249,105,194, //vpunpckhwd %xmm2,%xmm0,%xmm0
- 197,241,97,211, //vpunpcklwd %xmm3,%xmm1,%xmm2
- 197,113,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm9
- 184,0,4,0,4, //mov $0x4000400,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
- 196,193,97,101,200, //vpcmpgtw %xmm8,%xmm3,%xmm1
- 196,65,113,223,192, //vpandn %xmm8,%xmm1,%xmm8
- 197,225,101,200, //vpcmpgtw %xmm0,%xmm3,%xmm1
- 197,241,223,192, //vpandn %xmm0,%xmm1,%xmm0
- 197,225,101,202, //vpcmpgtw %xmm2,%xmm3,%xmm1
- 197,241,223,202, //vpandn %xmm2,%xmm1,%xmm1
- 196,193,97,101,209, //vpcmpgtw %xmm9,%xmm3,%xmm2
- 196,193,105,223,209, //vpandn %xmm9,%xmm2,%xmm2
- 196,66,121,51,208, //vpmovzxwd %xmm8,%xmm10
- 196,98,121,51,201, //vpmovzxwd %xmm1,%xmm9
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,57,105,195, //vpunpckhwd %xmm3,%xmm8,%xmm8
- 197,241,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm1
- 196,98,121,51,216, //vpmovzxwd %xmm0,%xmm11
- 196,98,121,51,226, //vpmovzxwd %xmm2,%xmm12
- 197,121,105,235, //vpunpckhwd %xmm3,%xmm0,%xmm13
- 197,105,105,243, //vpunpckhwd %xmm3,%xmm2,%xmm14
- 196,193,121,114,242,13, //vpslld $0xd,%xmm10,%xmm0
- 196,193,105,114,241,13, //vpslld $0xd,%xmm9,%xmm2
- 196,227,125,24,194,1, //vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 197,249,110,208, //vmovd %eax,%xmm2
- 197,249,112,210,0, //vpshufd $0x0,%xmm2,%xmm2
- 196,99,109,24,202,1, //vinsertf128 $0x1,%xmm2,%ymm2,%ymm9
- 197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 196,193,105,114,240,13, //vpslld $0xd,%xmm8,%xmm2
- 197,241,114,241,13, //vpslld $0xd,%xmm1,%xmm1
- 196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
- 197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 196,193,105,114,243,13, //vpslld $0xd,%xmm11,%xmm2
- 196,193,97,114,244,13, //vpslld $0xd,%xmm12,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
- 197,180,89,210, //vmulps %ymm2,%ymm9,%ymm2
- 196,193,57,114,245,13, //vpslld $0xd,%xmm13,%xmm8
- 196,193,97,114,246,13, //vpslld $0xd,%xmm14,%xmm3
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,180,89,219, //vmulps %ymm3,%ymm9,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 197,123,16,4,248, //vmovsd (%rax,%rdi,8),%xmm8
- 196,65,49,239,201, //vpxor %xmm9,%xmm9,%xmm9
- 72,131,249,1, //cmp $0x1,%rcx
- 116,79, //je 13c3 <_sk_load_f16_avx+0x16f>
- 197,57,22,68,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
- 72,131,249,3, //cmp $0x3,%rcx
- 114,67, //jb 13c3 <_sk_load_f16_avx+0x16f>
- 197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
- 72,131,249,3, //cmp $0x3,%rcx
- 116,68, //je 13d0 <_sk_load_f16_avx+0x17c>
- 197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
- 72,131,249,5, //cmp $0x5,%rcx
- 114,56, //jb 13d0 <_sk_load_f16_avx+0x17c>
- 197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
- 72,131,249,5, //cmp $0x5,%rcx
- 15,132,209,254,255,255, //je 1279 <_sk_load_f16_avx+0x25>
- 197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
- 72,131,249,7, //cmp $0x7,%rcx
- 15,130,193,254,255,255, //jb 1279 <_sk_load_f16_avx+0x25>
- 197,122,126,76,248,48, //vmovq 0x30(%rax,%rdi,8),%xmm9
- 233,182,254,255,255, //jmpq 1279 <_sk_load_f16_avx+0x25>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,169,254,255,255, //jmpq 1279 <_sk_load_f16_avx+0x25>
- 197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,160,254,255,255, //jmpq 1279 <_sk_load_f16_avx+0x25>
-};
-
-CODE const uint8_t sk_store_f16_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 184,0,0,128,7, //mov $0x7800000,%eax
- 197,121,110,192, //vmovd %eax,%xmm8
- 196,65,121,112,192,0, //vpshufd $0x0,%xmm8,%xmm8
- 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
- 197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
- 196,67,125,25,202,1, //vextractf128 $0x1,%ymm9,%xmm10
- 196,193,41,114,210,13, //vpsrld $0xd,%xmm10,%xmm10
- 196,193,49,114,209,13, //vpsrld $0xd,%xmm9,%xmm9
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,67,125,25,220,1, //vextractf128 $0x1,%ymm11,%xmm12
- 196,193,25,114,212,13, //vpsrld $0xd,%xmm12,%xmm12
- 196,193,33,114,211,13, //vpsrld $0xd,%xmm11,%xmm11
- 197,60,89,234, //vmulps %ymm2,%ymm8,%ymm13
- 196,67,125,25,238,1, //vextractf128 $0x1,%ymm13,%xmm14
- 196,193,9,114,214,13, //vpsrld $0xd,%xmm14,%xmm14
- 196,193,17,114,213,13, //vpsrld $0xd,%xmm13,%xmm13
- 197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
- 196,67,125,25,199,1, //vextractf128 $0x1,%ymm8,%xmm15
- 196,193,1,114,215,13, //vpsrld $0xd,%xmm15,%xmm15
- 196,193,57,114,208,13, //vpsrld $0xd,%xmm8,%xmm8
- 196,193,33,115,251,2, //vpslldq $0x2,%xmm11,%xmm11
- 196,65,33,235,201, //vpor %xmm9,%xmm11,%xmm9
- 196,193,33,115,252,2, //vpslldq $0x2,%xmm12,%xmm11
- 196,65,33,235,226, //vpor %xmm10,%xmm11,%xmm12
- 196,193,57,115,248,2, //vpslldq $0x2,%xmm8,%xmm8
- 196,65,57,235,197, //vpor %xmm13,%xmm8,%xmm8
- 196,193,41,115,255,2, //vpslldq $0x2,%xmm15,%xmm10
- 196,65,41,235,238, //vpor %xmm14,%xmm10,%xmm13
- 196,65,49,98,216, //vpunpckldq %xmm8,%xmm9,%xmm11
- 196,65,49,106,208, //vpunpckhdq %xmm8,%xmm9,%xmm10
- 196,65,25,98,205, //vpunpckldq %xmm13,%xmm12,%xmm9
- 196,65,25,106,197, //vpunpckhdq %xmm13,%xmm12,%xmm8
- 72,133,201, //test %rcx,%rcx
- 117,31, //jne 14af <_sk_store_f16_avx+0xd6>
- 196,65,120,17,28,248, //vmovups %xmm11,(%r8,%rdi,8)
- 196,65,120,17,84,248,16, //vmovups %xmm10,0x10(%r8,%rdi,8)
- 196,65,120,17,76,248,32, //vmovups %xmm9,0x20(%r8,%rdi,8)
- 196,65,122,127,68,248,48, //vmovdqu %xmm8,0x30(%r8,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,214,28,248, //vmovq %xmm11,(%r8,%rdi,8)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,23,92,248,8, //vmovhpd %xmm11,0x8(%r8,%rdi,8)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,214,84,248,16, //vmovq %xmm10,0x10(%r8,%rdi,8)
- 116,218, //je 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,23,84,248,24, //vmovhpd %xmm10,0x18(%r8,%rdi,8)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,214,76,248,32, //vmovq %xmm9,0x20(%r8,%rdi,8)
- 116,196, //je 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,23,76,248,40, //vmovhpd %xmm9,0x28(%r8,%rdi,8)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,183, //jb 14ab <_sk_store_f16_avx+0xd2>
- 196,65,121,214,68,248,48, //vmovq %xmm8,0x30(%r8,%rdi,8)
- 235,174, //jmp 14ab <_sk_store_f16_avx+0xd2>
-};
-
-CODE const uint8_t sk_store_f32_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 76,139,0, //mov (%rax),%r8
- 72,141,4,189,0,0,0,0, //lea 0x0(,%rdi,4),%rax
- 197,124,20,193, //vunpcklps %ymm1,%ymm0,%ymm8
- 197,124,21,217, //vunpckhps %ymm1,%ymm0,%ymm11
- 197,108,20,203, //vunpcklps %ymm3,%ymm2,%ymm9
- 197,108,21,227, //vunpckhps %ymm3,%ymm2,%ymm12
- 196,65,61,20,209, //vunpcklpd %ymm9,%ymm8,%ymm10
- 196,65,61,21,201, //vunpckhpd %ymm9,%ymm8,%ymm9
- 196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
- 196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
- 72,133,201, //test %rcx,%rcx
- 117,55, //jne 156a <_sk_store_f32_avx+0x6d>
- 196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
- 196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
- 196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
- 196,67,61,6,195,49, //vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
- 196,65,125,17,36,128, //vmovupd %ymm12,(%r8,%rax,4)
- 196,65,125,17,108,128,32, //vmovupd %ymm13,0x20(%r8,%rax,4)
- 196,65,125,17,76,128,64, //vmovupd %ymm9,0x40(%r8,%rax,4)
- 196,65,125,17,68,128,96, //vmovupd %ymm8,0x60(%r8,%rax,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
- 196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
- 72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 1566 <_sk_store_f32_avx+0x69>
- 196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
- 72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 1566 <_sk_store_f32_avx+0x69>
- 196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je 1566 <_sk_store_f32_avx+0x69>
- 196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
- 72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 1566 <_sk_store_f32_avx+0x69>
- 196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je 1566 <_sk_store_f32_avx+0x69>
- 196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
- 72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb 1566 <_sk_store_f32_avx+0x69>
- 196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp 1566 <_sk_store_f32_avx+0x69>
-};
-
-CODE const uint8_t sk_clamp_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,95,200, //vmaxps %ymm0,%ymm8,%ymm9
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,95,201, //vmaxps %ymm1,%ymm8,%ymm9
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,99,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,124,94,200, //vdivps %ymm8,%ymm0,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,65,52,89,200, //vmulps %ymm8,%ymm9,%ymm9
- 196,65,124,92,201, //vsubps %ymm9,%ymm0,%ymm9
- 196,99,125,25,192,1, //vextractf128 $0x1,%ymm8,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
- 197,180,93,192, //vminps %ymm0,%ymm9,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,65,116,94,200, //vdivps %ymm8,%ymm1,%ymm9
- 196,67,125,8,201,1, //vroundps $0x1,%ymm9,%ymm9
- 196,65,52,89,200, //vmulps %ymm8,%ymm9,%ymm9
- 196,65,116,92,201, //vsubps %ymm9,%ymm1,%ymm9
- 196,99,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,57,254,194, //vpaddd %xmm10,%xmm8,%xmm8
- 196,227,61,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
- 197,180,93,201, //vminps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,121,110,0, //vmovd (%rax),%xmm8
- 196,65,121,112,200,0, //vpshufd $0x0,%xmm8,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,124,92,209, //vsubps %ymm9,%ymm0,%ymm10
- 196,193,58,88,192, //vaddss %xmm8,%xmm8,%xmm0
- 196,227,121,4,192,0, //vpermilps $0x0,%xmm0,%xmm0
- 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
- 197,44,94,192, //vdivps %ymm0,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 197,188,89,192, //vmulps %ymm0,%ymm8,%ymm0
- 197,172,92,192, //vsubps %ymm0,%ymm10,%ymm0
- 196,193,124,92,193, //vsubps %ymm9,%ymm0,%ymm0
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,192, //vsubps %ymm0,%ymm8,%ymm8
- 197,60,84,192, //vandps %ymm0,%ymm8,%ymm8
- 196,99,125,25,200,1, //vextractf128 $0x1,%ymm9,%xmm0
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,121,254,194, //vpaddd %xmm10,%xmm0,%xmm0
- 196,65,49,254,202, //vpaddd %xmm10,%xmm9,%xmm9
- 196,227,53,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm9,%ymm0
- 197,188,93,192, //vminps %ymm0,%ymm8,%ymm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 197,121,110,0, //vmovd (%rax),%xmm8
- 196,65,121,112,200,0, //vpshufd $0x0,%xmm8,%xmm9
- 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
- 196,65,116,92,209, //vsubps %ymm9,%ymm1,%ymm10
- 196,193,58,88,200, //vaddss %xmm8,%xmm8,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,44,94,193, //vdivps %ymm1,%ymm10,%ymm8
- 196,67,125,8,192,1, //vroundps $0x1,%ymm8,%ymm8
- 197,188,89,201, //vmulps %ymm1,%ymm8,%ymm1
- 197,172,92,201, //vsubps %ymm1,%ymm10,%ymm1
- 196,193,116,92,201, //vsubps %ymm9,%ymm1,%ymm1
- 196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 197,60,92,193, //vsubps %ymm1,%ymm8,%ymm8
- 197,60,84,193, //vandps %ymm1,%ymm8,%ymm8
- 196,99,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm1
- 196,65,41,118,210, //vpcmpeqd %xmm10,%xmm10,%xmm10
- 196,193,113,254,202, //vpaddd %xmm10,%xmm1,%xmm1
- 196,65,49,254,202, //vpaddd %xmm10,%xmm9,%xmm9
- 196,227,53,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm9,%ymm1
- 197,188,93,201, //vminps %ymm1,%ymm8,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_avx[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 197,249,110,216, //vmovd %eax,%xmm3
- 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
- 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 197,249,110,200, //vmovd %eax,%xmm1
- 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
- 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
- 197,244,89,202, //vmulps %ymm2,%ymm1,%ymm1
- 197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
- 197,244,87,201, //vxorps %ymm1,%ymm1,%ymm1
- 197,236,87,210, //vxorps %ymm2,%ymm2,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,8, //vbroadcastss 0x8(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,12, //vbroadcastss 0xc(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 197,172,89,201, //vmulps %ymm1,%ymm10,%ymm1
- 196,193,116,88,203, //vaddps %ymm11,%ymm1,%ymm1
- 197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 197,252,88,201, //vaddps %ymm1,%ymm0,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 196,98,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm10
- 196,98,125,24,88,36, //vbroadcastss 0x24(%rax),%ymm11
- 197,44,89,210, //vmulps %ymm2,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm11
- 196,98,125,24,96,40, //vbroadcastss 0x28(%rax),%ymm12
- 197,36,89,218, //vmulps %ymm2,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 196,98,125,24,104,44, //vbroadcastss 0x2c(%rax),%ymm13
- 197,156,89,210, //vmulps %ymm2,%ymm12,%ymm2
- 196,193,108,88,213, //vaddps %ymm13,%ymm2,%ymm2
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 197,244,88,202, //vaddps %ymm2,%ymm1,%ymm1
- 197,172,89,192, //vmulps %ymm0,%ymm10,%ymm0
- 197,252,88,209, //vaddps %ymm1,%ymm0,%ymm2
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm9
- 196,98,125,24,80,32, //vbroadcastss 0x20(%rax),%ymm10
- 196,98,125,24,88,48, //vbroadcastss 0x30(%rax),%ymm11
- 196,98,125,24,96,64, //vbroadcastss 0x40(%rax),%ymm12
- 197,36,89,219, //vmulps %ymm3,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,210, //vmulps %ymm2,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,20, //vbroadcastss 0x14(%rax),%ymm10
- 196,98,125,24,88,36, //vbroadcastss 0x24(%rax),%ymm11
- 196,98,125,24,96,52, //vbroadcastss 0x34(%rax),%ymm12
- 196,98,125,24,104,68, //vbroadcastss 0x44(%rax),%ymm13
- 197,28,89,227, //vmulps %ymm3,%ymm12,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,36,89,218, //vmulps %ymm2,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 196,98,125,24,88,24, //vbroadcastss 0x18(%rax),%ymm11
- 196,98,125,24,96,40, //vbroadcastss 0x28(%rax),%ymm12
- 196,98,125,24,104,56, //vbroadcastss 0x38(%rax),%ymm13
- 196,98,125,24,112,72, //vbroadcastss 0x48(%rax),%ymm14
- 197,20,89,235, //vmulps %ymm3,%ymm13,%ymm13
- 196,65,20,88,238, //vaddps %ymm14,%ymm13,%ymm13
- 197,28,89,226, //vmulps %ymm2,%ymm12,%ymm12
- 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
- 197,36,89,217, //vmulps %ymm1,%ymm11,%ymm11
- 196,65,36,88,220, //vaddps %ymm12,%ymm11,%ymm11
- 197,44,89,208, //vmulps %ymm0,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 196,98,125,24,88,12, //vbroadcastss 0xc(%rax),%ymm11
- 196,98,125,24,96,28, //vbroadcastss 0x1c(%rax),%ymm12
- 196,98,125,24,104,44, //vbroadcastss 0x2c(%rax),%ymm13
- 196,98,125,24,112,60, //vbroadcastss 0x3c(%rax),%ymm14
- 196,98,125,24,120,76, //vbroadcastss 0x4c(%rax),%ymm15
- 197,140,89,219, //vmulps %ymm3,%ymm14,%ymm3
- 196,193,100,88,223, //vaddps %ymm15,%ymm3,%ymm3
- 197,148,89,210, //vmulps %ymm2,%ymm13,%ymm2
- 197,236,88,211, //vaddps %ymm3,%ymm2,%ymm2
- 197,156,89,201, //vmulps %ymm1,%ymm12,%ymm1
- 197,244,88,202, //vaddps %ymm2,%ymm1,%ymm1
- 197,164,89,192, //vmulps %ymm0,%ymm11,%ymm0
- 197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 197,124,41,201, //vmovaps %ymm9,%ymm1
- 197,124,41,210, //vmovaps %ymm10,%ymm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,98,125,24,0, //vbroadcastss (%rax),%ymm8
- 196,98,125,24,72,4, //vbroadcastss 0x4(%rax),%ymm9
- 196,98,125,24,80,8, //vbroadcastss 0x8(%rax),%ymm10
- 197,52,89,201, //vmulps %ymm1,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 197,60,89,192, //vmulps %ymm0,%ymm8,%ymm8
- 196,65,60,88,193, //vaddps %ymm9,%ymm8,%ymm8
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 196,98,125,24,80,16, //vbroadcastss 0x10(%rax),%ymm10
- 196,98,125,24,88,20, //vbroadcastss 0x14(%rax),%ymm11
- 197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
- 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
- 197,52,89,200, //vmulps %ymm0,%ymm9,%ymm9
- 196,65,52,88,202, //vaddps %ymm10,%ymm9,%ymm9
- 196,98,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm10
- 196,98,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm11
- 196,98,125,24,96,32, //vbroadcastss 0x20(%rax),%ymm12
- 197,164,89,201, //vmulps %ymm1,%ymm11,%ymm1
- 196,193,116,88,204, //vaddps %ymm12,%ymm1,%ymm1
- 197,172,89,192, //vmulps %ymm0,%ymm10,%ymm0
- 197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,83,200, //vrcpps %ymm0,%ymm1
- 197,188,89,193, //vmulps %ymm1,%ymm8,%ymm0
- 197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_avx[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 196,226,125,24,72,16, //vbroadcastss 0x10(%rax),%ymm1
- 196,226,125,24,16, //vbroadcastss (%rax),%ymm2
- 197,244,89,200, //vmulps %ymm0,%ymm1,%ymm1
- 197,108,88,193, //vaddps %ymm1,%ymm2,%ymm8
- 196,226,125,24,72,20, //vbroadcastss 0x14(%rax),%ymm1
- 196,226,125,24,80,4, //vbroadcastss 0x4(%rax),%ymm2
- 197,244,89,200, //vmulps %ymm0,%ymm1,%ymm1
- 197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
- 196,226,125,24,80,24, //vbroadcastss 0x18(%rax),%ymm2
- 196,226,125,24,88,8, //vbroadcastss 0x8(%rax),%ymm3
- 197,236,89,208, //vmulps %ymm0,%ymm2,%ymm2
- 197,228,88,210, //vaddps %ymm2,%ymm3,%ymm2
- 196,226,125,24,88,28, //vbroadcastss 0x1c(%rax),%ymm3
- 196,98,125,24,72,12, //vbroadcastss 0xc(%rax),%ymm9
- 197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 197,180,88,216, //vaddps %ymm0,%ymm9,%ymm3
- 72,173, //lods %ds:(%rsi),%rax
- 197,124,41,192, //vmovaps %ymm8,%ymm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_sse41[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 86, //push %rsi
- 87, //push %rdi
- 83, //push %rbx
- 72,129,236,160,0,0,0, //sub $0xa0,%rsp
- 68,15,41,188,36,144,0,0,0, //movaps %xmm15,0x90(%rsp)
- 68,15,41,180,36,128,0,0,0, //movaps %xmm14,0x80(%rsp)
- 68,15,41,108,36,112, //movaps %xmm13,0x70(%rsp)
- 68,15,41,100,36,96, //movaps %xmm12,0x60(%rsp)
- 68,15,41,92,36,80, //movaps %xmm11,0x50(%rsp)
- 68,15,41,84,36,64, //movaps %xmm10,0x40(%rsp)
- 68,15,41,76,36,48, //movaps %xmm9,0x30(%rsp)
- 68,15,41,68,36,32, //movaps %xmm8,0x20(%rsp)
- 15,41,124,36,16, //movaps %xmm7,0x10(%rsp)
- 15,41,52,36, //movaps %xmm6,(%rsp)
- 77,137,207, //mov %r9,%r15
- 77,137,198, //mov %r8,%r14
- 72,137,203, //mov %rcx,%rbx
- 72,137,214, //mov %rdx,%rsi
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,196, //mov %rax,%r12
- 73,137,245, //mov %rsi,%r13
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 76,57,248, //cmp %r15,%rax
- 118,5, //jbe 73 <_sk_start_pipeline_sse41+0x73>
- 72,137,216, //mov %rbx,%rax
- 235,52, //jmp a7 <_sk_start_pipeline_sse41+0xa7>
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,238, //mov %r13,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,212, //callq *%r12
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 72,131,195,8, //add $0x8,%rbx
- 76,57,251, //cmp %r15,%rbx
- 72,137,195, //mov %rax,%rbx
- 118,204, //jbe 73 <_sk_start_pipeline_sse41+0x73>
- 15,40,52,36, //movaps (%rsp),%xmm6
- 15,40,124,36,16, //movaps 0x10(%rsp),%xmm7
- 68,15,40,68,36,32, //movaps 0x20(%rsp),%xmm8
- 68,15,40,76,36,48, //movaps 0x30(%rsp),%xmm9
- 68,15,40,84,36,64, //movaps 0x40(%rsp),%xmm10
- 68,15,40,92,36,80, //movaps 0x50(%rsp),%xmm11
- 68,15,40,100,36,96, //movaps 0x60(%rsp),%xmm12
- 68,15,40,108,36,112, //movaps 0x70(%rsp),%xmm13
- 68,15,40,180,36,128,0,0,0, //movaps 0x80(%rsp),%xmm14
- 68,15,40,188,36,144,0,0,0, //movaps 0x90(%rsp),%xmm15
- 72,129,196,160,0,0,0, //add $0xa0,%rsp
- 91, //pop %rbx
- 95, //pop %rdi
- 94, //pop %rsi
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_sse41[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 102,15,110,199, //movd %edi,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 185,0,0,0,63, //mov $0x3f000000,%ecx
- 102,15,110,209, //movd %ecx,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,88,202, //addps %xmm2,%xmm1
- 15,16,2, //movups (%rdx),%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 102,15,110,8, //movd (%rax),%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,16,24, //movups (%rax),%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__sse41[] = {
- 15,88,196, //addps %xmm4,%xmm0
- 15,88,205, //addps %xmm5,%xmm1
- 15,88,214, //addps %xmm6,%xmm2
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,195, //subps %xmm3,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,204, //mulps %xmm4,%xmm9
- 65,15,88,193, //addps %xmm9,%xmm0
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,205, //mulps %xmm5,%xmm9
- 65,15,88,201, //addps %xmm9,%xmm1
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,206, //mulps %xmm6,%xmm9
- 65,15,88,209, //addps %xmm9,%xmm2
- 68,15,89,199, //mulps %xmm7,%xmm8
- 65,15,88,216, //addps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,199, //subps %xmm7,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_sse41[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 65,15,95,192, //maxps %xmm8,%xmm0
- 65,15,95,200, //maxps %xmm8,%xmm1
- 65,15,95,208, //maxps %xmm8,%xmm2
- 65,15,95,216, //maxps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 65,15,93,200, //minps %xmm8,%xmm1
- 65,15,93,208, //minps %xmm8,%xmm2
- 65,15,93,216, //minps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_sse41[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,216, //minps %xmm8,%xmm3
- 15,93,195, //minps %xmm3,%xmm0
- 15,93,203, //minps %xmm3,%xmm1
- 15,93,211, //minps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,15,16,80,8, //movss 0x8(%rax),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_sse41[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,194, //movaps %xmm2,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_sse41[] = {
- 68,15,40,195, //movaps %xmm3,%xmm8
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 68,15,40,216, //movaps %xmm0,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 65,15,40,227, //movaps %xmm11,%xmm4
- 65,15,40,234, //movaps %xmm10,%xmm5
- 65,15,40,241, //movaps %xmm9,%xmm6
- 65,15,40,248, //movaps %xmm8,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,224, //movaps %xmm0,%xmm4
- 15,40,233, //movaps %xmm1,%xmm5
- 15,40,242, //movaps %xmm2,%xmm6
- 15,40,251, //movaps %xmm3,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_sse41[] = {
- 15,89,195, //mulps %xmm3,%xmm0
- 15,89,203, //mulps %xmm3,%xmm1
- 15,89,211, //mulps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_sse41[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,94,203, //divps %xmm3,%xmm9
- 68,15,194,195,4, //cmpneqps %xmm3,%xmm8
- 69,15,84,193, //andps %xmm9,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_sse41[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 68,15,40,240, //movaps %xmm0,%xmm14
- 69,15,89,246, //mulps %xmm14,%xmm14
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 69,15,88,204, //addps %xmm12,%xmm9
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 69,15,89,206, //mulps %xmm14,%xmm9
- 69,15,88,205, //addps %xmm13,%xmm9
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 102,68,15,110,240, //movd %eax,%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 65,15,194,198,1, //cmpltps %xmm14,%xmm0
- 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
- 69,15,40,251, //movaps %xmm11,%xmm15
- 68,15,89,249, //mulps %xmm1,%xmm15
- 15,40,193, //movaps %xmm1,%xmm0
- 15,89,192, //mulps %xmm0,%xmm0
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,212, //addps %xmm12,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 69,15,88,213, //addps %xmm13,%xmm10
- 65,15,194,206,1, //cmpltps %xmm14,%xmm1
- 15,40,193, //movaps %xmm1,%xmm0
- 102,69,15,56,20,215, //blendvps %xmm0,%xmm15,%xmm10
- 68,15,89,218, //mulps %xmm2,%xmm11
- 15,40,194, //movaps %xmm2,%xmm0
- 15,89,192, //mulps %xmm0,%xmm0
- 68,15,89,194, //mulps %xmm2,%xmm8
- 69,15,88,196, //addps %xmm12,%xmm8
- 68,15,89,192, //mulps %xmm0,%xmm8
- 69,15,88,197, //addps %xmm13,%xmm8
- 65,15,194,214,1, //cmpltps %xmm14,%xmm2
- 15,40,194, //movaps %xmm2,%xmm0
- 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,202, //movaps %xmm10,%xmm1
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_sse41[] = {
- 72,131,236,24, //sub $0x18,%rsp
- 15,41,60,36, //movaps %xmm7,(%rsp)
- 15,40,254, //movaps %xmm6,%xmm7
- 15,40,245, //movaps %xmm5,%xmm6
- 15,40,236, //movaps %xmm4,%xmm5
- 15,40,227, //movaps %xmm3,%xmm4
- 15,40,218, //movaps %xmm2,%xmm3
- 15,40,209, //movaps %xmm1,%xmm2
- 68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,200, //rcpps %xmm8,%xmm9
- 69,15,82,248, //rsqrtps %xmm8,%xmm15
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 102,68,15,110,240, //movd %eax,%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 69,15,89,205, //mulps %xmm13,%xmm9
- 69,15,88,206, //addps %xmm14,%xmm9
- 69,15,89,252, //mulps %xmm12,%xmm15
- 69,15,88,249, //addps %xmm9,%xmm15
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,93,207, //minps %xmm15,%xmm9
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 102,68,15,110,248, //movd %eax,%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 65,15,194,199,1, //cmpltps %xmm15,%xmm0
- 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
- 68,15,82,210, //rsqrtps %xmm2,%xmm10
- 65,15,83,194, //rcpps %xmm10,%xmm0
- 69,15,82,210, //rsqrtps %xmm10,%xmm10
- 65,15,89,197, //mulps %xmm13,%xmm0
- 65,15,88,198, //addps %xmm14,%xmm0
- 69,15,89,212, //mulps %xmm12,%xmm10
- 68,15,88,208, //addps %xmm0,%xmm10
- 65,15,40,200, //movaps %xmm8,%xmm1
- 65,15,93,202, //minps %xmm10,%xmm1
- 69,15,40,211, //movaps %xmm11,%xmm10
- 68,15,89,210, //mulps %xmm2,%xmm10
- 65,15,194,215,1, //cmpltps %xmm15,%xmm2
- 15,40,194, //movaps %xmm2,%xmm0
- 102,65,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm1
- 15,82,195, //rsqrtps %xmm3,%xmm0
- 15,83,208, //rcpps %xmm0,%xmm2
- 65,15,89,213, //mulps %xmm13,%xmm2
- 65,15,88,214, //addps %xmm14,%xmm2
- 15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,194, //addps %xmm2,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 68,15,89,219, //mulps %xmm3,%xmm11
- 65,15,194,223,1, //cmpltps %xmm15,%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 15,40,220, //movaps %xmm4,%xmm3
- 15,40,229, //movaps %xmm5,%xmm4
- 15,40,238, //movaps %xmm6,%xmm5
- 15,40,247, //movaps %xmm7,%xmm6
- 15,40,60,36, //movaps (%rsp),%xmm7
- 72,131,196,24, //add $0x18,%rsp
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 65,15,89,201, //mulps %xmm9,%xmm1
- 65,15,89,209, //mulps %xmm9,%xmm2
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_1_float_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,193, //mulps %xmm9,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,56,51,4,120, //pmovzxwd (%rax,%rdi,2),%xmm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,209, //mulps %xmm9,%xmm10
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,194, //mulps %xmm10,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,203, //mulps %xmm11,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_tables_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,8, //mov (%rax),%rcx
- 76,139,64,8, //mov 0x8(%rax),%r8
- 243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 185,255,0,0,0, //mov $0xff,%ecx
- 102,15,110,193, //movd %ecx,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,111,200, //movdqa %xmm8,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,65,15,111,208, //movdqa %xmm8,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,65,15,219,192, //pand %xmm8,%xmm0
- 102,72,15,58,22,193,1, //pextrq $0x1,%xmm0,%rcx
- 65,137,201, //mov %ecx,%r9d
- 72,193,233,32, //shr $0x20,%rcx
- 102,73,15,126,194, //movq %xmm0,%r10
- 69,137,211, //mov %r10d,%r11d
- 73,193,234,32, //shr $0x20,%r10
- 243,67,15,16,4,152, //movss (%r8,%r11,4),%xmm0
- 102,67,15,58,33,4,144,16, //insertps $0x10,(%r8,%r10,4),%xmm0
- 102,67,15,58,33,4,136,32, //insertps $0x20,(%r8,%r9,4),%xmm0
- 102,65,15,58,33,4,136,48, //insertps $0x30,(%r8,%rcx,4),%xmm0
- 76,139,64,16, //mov 0x10(%rax),%r8
- 102,73,15,58,22,202,1, //pextrq $0x1,%xmm1,%r10
- 77,137,209, //mov %r10,%r9
- 73,193,233,32, //shr $0x20,%r9
- 102,72,15,126,201, //movq %xmm1,%rcx
- 65,137,203, //mov %ecx,%r11d
- 65,129,227,255,255,255,0, //and $0xffffff,%r11d
- 72,193,233,30, //shr $0x1e,%rcx
- 65,129,226,255,255,255,0, //and $0xffffff,%r10d
- 243,67,15,16,12,152, //movss (%r8,%r11,4),%xmm1
- 102,65,15,58,33,12,8,16, //insertps $0x10,(%r8,%rcx,1),%xmm1
- 243,67,15,16,28,144, //movss (%r8,%r10,4),%xmm3
- 102,15,58,33,203,32, //insertps $0x20,%xmm3,%xmm1
- 243,67,15,16,28,136, //movss (%r8,%r9,4),%xmm3
- 102,15,58,33,203,48, //insertps $0x30,%xmm3,%xmm1
- 76,139,72,24, //mov 0x18(%rax),%r9
- 102,72,15,58,22,209,1, //pextrq $0x1,%xmm2,%rcx
- 68,15,183,193, //movzwl %cx,%r8d
- 72,193,233,32, //shr $0x20,%rcx
- 102,72,15,126,208, //movq %xmm2,%rax
- 68,15,183,208, //movzwl %ax,%r10d
- 72,193,232,30, //shr $0x1e,%rax
- 243,67,15,16,20,145, //movss (%r9,%r10,4),%xmm2
- 102,65,15,58,33,20,1,16, //insertps $0x10,(%r9,%rax,1),%xmm2
- 243,67,15,16,28,129, //movss (%r9,%r8,4),%xmm3
- 102,15,58,33,211,32, //insertps $0x20,%xmm3,%xmm2
- 243,65,15,16,28,137, //movss (%r9,%rcx,4),%xmm3
- 102,15,58,33,211,48, //insertps $0x30,%xmm3,%xmm2
- 102,65,15,114,208,24, //psrld $0x18,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_a8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_a8_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,56,43,192, //packusdw %xmm8,%xmm8
- 102,69,15,103,192, //packuswb %xmm8,%xmm8
- 102,68,15,126,4,56, //movd %xmm8,(%rax,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,56,51,20,120, //pmovzxwd (%rax,%rdi,2),%xmm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,219,194, //pand %xmm2,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,15,219,202, //pand %xmm2,%xmm1
- 15,91,217, //cvtdq2ps %xmm1,%xmm3
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,15,219,218, //pand %xmm2,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_565_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,248,65, //mov $0x41f80000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,11, //pslld $0xb,%xmm9
- 185,0,0,124,66, //mov $0x427c0000,%ecx
- 102,68,15,110,209, //movd %ecx,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,5, //pslld $0x5,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 68,15,89,194, //mulps %xmm2,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,194, //orpd %xmm10,%xmm8
- 102,69,15,56,43,192, //packusdw %xmm8,%xmm8
- 102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_8888_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,28,184, //movdqu (%rax,%rdi,4),%xmm3
- 184,255,0,0,0, //mov $0xff,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,15,219,195, //pand %xmm3,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,91,210, //cvtdq2ps %xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,15,114,211,24, //psrld $0x18,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_8888_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,8, //pslld $0x8,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,202, //mulps %xmm2,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,16, //pslld $0x10,%xmm9
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,24, //pslld $0x18,%xmm8
- 102,69,15,235,193, //por %xmm9,%xmm8
- 102,69,15,235,194, //por %xmm10,%xmm8
- 243,68,15,127,4,184, //movdqu %xmm8,(%rax,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_f16_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,4,248, //movdqu (%rax,%rdi,8),%xmm0
- 243,15,111,76,248,16, //movdqu 0x10(%rax,%rdi,8),%xmm1
- 102,15,111,208, //movdqa %xmm0,%xmm2
- 102,15,97,209, //punpcklwd %xmm1,%xmm2
- 102,15,105,193, //punpckhwd %xmm1,%xmm0
- 102,68,15,111,194, //movdqa %xmm2,%xmm8
- 102,68,15,97,192, //punpcklwd %xmm0,%xmm8
- 102,15,105,208, //punpckhwd %xmm0,%xmm2
- 184,0,4,0,4, //mov $0x4000400,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
- 102,65,15,223,200, //pandn %xmm8,%xmm1
- 102,15,101,218, //pcmpgtw %xmm2,%xmm3
- 102,15,223,218, //pandn %xmm2,%xmm3
- 102,15,56,51,193, //pmovzxwd %xmm1,%xmm0
- 102,15,114,240,13, //pslld $0xd,%xmm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 102,68,15,112,194,0, //pshufd $0x0,%xmm2,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,65,15,105,201, //punpckhwd %xmm9,%xmm1
- 102,15,114,241,13, //pslld $0xd,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,56,51,211, //pmovzxwd %xmm3,%xmm2
- 102,15,114,242,13, //pslld $0xd,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,65,15,105,217, //punpckhwd %xmm9,%xmm3
- 102,15,114,243,13, //pslld $0xd,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f16_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,128,7, //mov $0x7800000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,65,15,114,209,13, //psrld $0xd,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,65,15,114,210,13, //psrld $0xd,%xmm10
- 102,69,15,111,216, //movdqa %xmm8,%xmm11
- 68,15,89,218, //mulps %xmm2,%xmm11
- 102,65,15,114,211,13, //psrld $0xd,%xmm11
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,65,15,114,208,13, //psrld $0xd,%xmm8
- 102,65,15,115,250,2, //pslldq $0x2,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 102,65,15,115,248,2, //pslldq $0x2,%xmm8
- 102,69,15,235,195, //por %xmm11,%xmm8
- 102,69,15,111,202, //movdqa %xmm10,%xmm9
- 102,69,15,98,200, //punpckldq %xmm8,%xmm9
- 243,68,15,127,12,248, //movdqu %xmm9,(%rax,%rdi,8)
- 102,69,15,106,208, //punpckhdq %xmm8,%xmm10
- 243,68,15,127,84,248,16, //movdqu %xmm10,0x10(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f32_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,137,249, //mov %rdi,%rcx
- 72,193,225,4, //shl $0x4,%rcx
- 68,15,40,192, //movaps %xmm0,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 68,15,20,201, //unpcklps %xmm1,%xmm9
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,40,218, //movaps %xmm2,%xmm11
- 68,15,20,219, //unpcklps %xmm3,%xmm11
- 68,15,21,193, //unpckhps %xmm1,%xmm8
- 68,15,21,211, //unpckhps %xmm3,%xmm10
- 69,15,40,225, //movaps %xmm9,%xmm12
- 102,69,15,20,227, //unpcklpd %xmm11,%xmm12
- 69,15,18,217, //movhlps %xmm9,%xmm11
- 69,15,40,200, //movaps %xmm8,%xmm9
- 102,69,15,20,202, //unpcklpd %xmm10,%xmm9
- 69,15,18,208, //movhlps %xmm8,%xmm10
- 102,68,15,17,36,8, //movupd %xmm12,(%rax,%rcx,1)
- 68,15,17,92,8,16, //movups %xmm11,0x10(%rax,%rcx,1)
- 102,68,15,17,76,8,32, //movupd %xmm9,0x20(%rax,%rcx,1)
- 68,15,17,84,8,48, //movups %xmm10,0x30(%rax,%rcx,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,192, //maxps %xmm0,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,192, //pcmpeqd %xmm0,%xmm0
- 102,65,15,254,193, //paddd %xmm9,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,193, //maxps %xmm1,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,201, //pcmpeqd %xmm1,%xmm1
- 102,65,15,254,201, //paddd %xmm9,%xmm1
- 68,15,93,193, //minps %xmm1,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,200, //movaps %xmm8,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 102,69,15,58,8,201,1, //roundps $0x1,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,92,193, //subps %xmm9,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,201, //movaps %xmm1,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 102,69,15,58,8,201,1, //roundps $0x1,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,92,201, //subps %xmm9,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 65,15,92,193, //subps %xmm9,%xmm0
- 243,69,15,88,192, //addss %xmm8,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,208, //movaps %xmm0,%xmm10
- 69,15,94,208, //divps %xmm8,%xmm10
- 102,69,15,58,8,210,1, //roundps $0x1,%xmm10,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,194, //subps %xmm10,%xmm0
- 65,15,92,193, //subps %xmm9,%xmm0
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,92,192, //subps %xmm0,%xmm8
- 65,15,84,192, //andps %xmm8,%xmm0
- 102,69,15,118,192, //pcmpeqd %xmm8,%xmm8
- 102,69,15,254,193, //paddd %xmm9,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 65,15,92,201, //subps %xmm9,%xmm1
- 243,69,15,88,192, //addss %xmm8,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,209, //movaps %xmm1,%xmm10
- 69,15,94,208, //divps %xmm8,%xmm10
- 102,69,15,58,8,210,1, //roundps $0x1,%xmm10,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,202, //subps %xmm10,%xmm1
- 65,15,92,201, //subps %xmm9,%xmm1
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,92,193, //subps %xmm1,%xmm8
- 65,15,84,200, //andps %xmm8,%xmm1
- 102,69,15,118,192, //pcmpeqd %xmm8,%xmm8
- 102,69,15,254,193, //paddd %xmm9,%xmm8
- 65,15,93,200, //minps %xmm8,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_sse41[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 15,88,195, //addps %xmm3,%xmm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,218, //mulps %xmm2,%xmm3
- 15,88,216, //addps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,16, //movss 0x10(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,36, //movss 0x24(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_sse41[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,32, //movss 0x20(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,48, //movss 0x30(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,64, //movss 0x40(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,20, //movss 0x14(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,36, //movss 0x24(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,52, //movss 0x34(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,68, //movss 0x44(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,56, //movss 0x38(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,72, //movss 0x48(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 68,15,89,235, //mulps %xmm3,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 243,68,15,16,88,12, //movss 0xc(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,28, //movss 0x1c(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,60, //movss 0x3c(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 243,68,15,16,120,76, //movss 0x4c(%rax),%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 68,15,89,243, //mulps %xmm3,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 68,15,89,234, //mulps %xmm2,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 69,15,89,225, //mulps %xmm9,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 65,15,40,219, //movaps %xmm11,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_sse41[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,68,15,16,72,4, //movss 0x4(%rax),%xmm9
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,201, //mulps %xmm1,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,193, //addps %xmm9,%xmm0
- 243,68,15,16,72,12, //movss 0xc(%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 69,15,89,200, //mulps %xmm8,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 243,68,15,16,80,24, //movss 0x18(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,217, //mulps %xmm1,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,83,202, //rcpps %xmm10,%xmm1
- 15,89,193, //mulps %xmm1,%xmm0
- 68,15,89,201, //mulps %xmm1,%xmm9
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,201, //movaps %xmm9,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_sse41[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 68,15,16,8, //movups (%rax),%xmm9
- 15,16,88,16, //movups 0x10(%rax),%xmm3
- 68,15,40,195, //movaps %xmm3,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,40,201, //movaps %xmm9,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 68,15,89,192, //mulps %xmm0,%xmm8
- 68,15,88,193, //addps %xmm1,%xmm8
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 65,15,40,209, //movaps %xmm9,%xmm2
- 15,198,210,85, //shufps $0x55,%xmm2,%xmm2
- 15,89,200, //mulps %xmm0,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 69,15,40,209, //movaps %xmm9,%xmm10
- 69,15,198,210,170, //shufps $0xaa,%xmm10,%xmm10
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,210, //addps %xmm10,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 69,15,198,201,255, //shufps $0xff,%xmm9,%xmm9
- 15,89,216, //mulps %xmm0,%xmm3
- 65,15,88,217, //addps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_start_pipeline_sse2[] = {
- 65,87, //push %r15
- 65,86, //push %r14
- 65,85, //push %r13
- 65,84, //push %r12
- 86, //push %rsi
- 87, //push %rdi
- 83, //push %rbx
- 72,129,236,160,0,0,0, //sub $0xa0,%rsp
- 68,15,41,188,36,144,0,0,0, //movaps %xmm15,0x90(%rsp)
- 68,15,41,180,36,128,0,0,0, //movaps %xmm14,0x80(%rsp)
- 68,15,41,108,36,112, //movaps %xmm13,0x70(%rsp)
- 68,15,41,100,36,96, //movaps %xmm12,0x60(%rsp)
- 68,15,41,92,36,80, //movaps %xmm11,0x50(%rsp)
- 68,15,41,84,36,64, //movaps %xmm10,0x40(%rsp)
- 68,15,41,76,36,48, //movaps %xmm9,0x30(%rsp)
- 68,15,41,68,36,32, //movaps %xmm8,0x20(%rsp)
- 15,41,124,36,16, //movaps %xmm7,0x10(%rsp)
- 15,41,52,36, //movaps %xmm6,(%rsp)
- 77,137,207, //mov %r9,%r15
- 77,137,198, //mov %r8,%r14
- 72,137,203, //mov %rcx,%rbx
- 72,137,214, //mov %rdx,%rsi
- 72,173, //lods %ds:(%rsi),%rax
- 73,137,196, //mov %rax,%r12
- 73,137,245, //mov %rsi,%r13
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 76,57,248, //cmp %r15,%rax
- 118,5, //jbe 73 <_sk_start_pipeline_sse2+0x73>
- 72,137,216, //mov %rbx,%rax
- 235,52, //jmp a7 <_sk_start_pipeline_sse2+0xa7>
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 72,137,223, //mov %rbx,%rdi
- 76,137,238, //mov %r13,%rsi
- 76,137,242, //mov %r14,%rdx
- 65,255,212, //callq *%r12
- 72,141,67,4, //lea 0x4(%rbx),%rax
- 72,131,195,8, //add $0x8,%rbx
- 76,57,251, //cmp %r15,%rbx
- 72,137,195, //mov %rax,%rbx
- 118,204, //jbe 73 <_sk_start_pipeline_sse2+0x73>
- 15,40,52,36, //movaps (%rsp),%xmm6
- 15,40,124,36,16, //movaps 0x10(%rsp),%xmm7
- 68,15,40,68,36,32, //movaps 0x20(%rsp),%xmm8
- 68,15,40,76,36,48, //movaps 0x30(%rsp),%xmm9
- 68,15,40,84,36,64, //movaps 0x40(%rsp),%xmm10
- 68,15,40,92,36,80, //movaps 0x50(%rsp),%xmm11
- 68,15,40,100,36,96, //movaps 0x60(%rsp),%xmm12
- 68,15,40,108,36,112, //movaps 0x70(%rsp),%xmm13
- 68,15,40,180,36,128,0,0,0, //movaps 0x80(%rsp),%xmm14
- 68,15,40,188,36,144,0,0,0, //movaps 0x90(%rsp),%xmm15
- 72,129,196,160,0,0,0, //add $0xa0,%rsp
- 91, //pop %rbx
- 95, //pop %rdi
- 94, //pop %rsi
- 65,92, //pop %r12
- 65,93, //pop %r13
- 65,94, //pop %r14
- 65,95, //pop %r15
- 195, //retq
-};
-
-CODE const uint8_t sk_just_return_sse2[] = {
- 195, //retq
-};
-
-CODE const uint8_t sk_seed_shader_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 102,15,110,199, //movd %edi,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 185,0,0,0,63, //mov $0x3f000000,%ecx
- 102,15,110,209, //movd %ecx,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,88,202, //addps %xmm2,%xmm1
- 15,16,2, //movups (%rdx),%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 102,15,110,8, //movd (%rax),%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,219, //xorps %xmm3,%xmm3
- 15,87,228, //xorps %xmm4,%xmm4
- 15,87,237, //xorps %xmm5,%xmm5
- 15,87,246, //xorps %xmm6,%xmm6
- 15,87,255, //xorps %xmm7,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_constant_color_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,16,24, //movups (%rax),%xmm3
- 15,40,195, //movaps %xmm3,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clear_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 15,87,219, //xorps %xmm3,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_plus__sse2[] = {
- 15,88,196, //addps %xmm4,%xmm0
- 15,88,205, //addps %xmm5,%xmm1
- 15,88,214, //addps %xmm6,%xmm2
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_srcover_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,195, //subps %xmm3,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,204, //mulps %xmm4,%xmm9
- 65,15,88,193, //addps %xmm9,%xmm0
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,205, //mulps %xmm5,%xmm9
- 65,15,88,201, //addps %xmm9,%xmm1
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,206, //mulps %xmm6,%xmm9
- 65,15,88,209, //addps %xmm9,%xmm2
- 68,15,89,199, //mulps %xmm7,%xmm8
- 65,15,88,216, //addps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_dstover_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,92,199, //subps %xmm7,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_0_sse2[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 65,15,95,192, //maxps %xmm8,%xmm0
- 65,15,95,200, //maxps %xmm8,%xmm1
- 65,15,95,208, //maxps %xmm8,%xmm2
- 65,15,95,216, //maxps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_1_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,192, //minps %xmm8,%xmm0
- 65,15,93,200, //minps %xmm8,%xmm1
- 65,15,93,208, //minps %xmm8,%xmm2
- 65,15,93,216, //minps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_a_sse2[] = {
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,93,216, //minps %xmm8,%xmm3
- 15,93,195, //minps %xmm3,%xmm0
- 15,93,203, //minps %xmm3,%xmm1
- 15,93,211, //minps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_set_rgb_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,15,16,80,8, //movss 0x8(%rax),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_rb_sse2[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,194, //movaps %xmm2,%xmm0
- 65,15,40,208, //movaps %xmm8,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_swap_sse2[] = {
- 68,15,40,195, //movaps %xmm3,%xmm8
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 68,15,40,216, //movaps %xmm0,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 65,15,40,227, //movaps %xmm11,%xmm4
- 65,15,40,234, //movaps %xmm10,%xmm5
- 65,15,40,241, //movaps %xmm9,%xmm6
- 65,15,40,248, //movaps %xmm8,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_src_dst_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,224, //movaps %xmm0,%xmm4
- 15,40,233, //movaps %xmm1,%xmm5
- 15,40,242, //movaps %xmm2,%xmm6
- 15,40,251, //movaps %xmm3,%xmm7
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_move_dst_src_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 15,40,196, //movaps %xmm4,%xmm0
- 15,40,205, //movaps %xmm5,%xmm1
- 15,40,214, //movaps %xmm6,%xmm2
- 15,40,223, //movaps %xmm7,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_premul_sse2[] = {
- 15,89,195, //mulps %xmm3,%xmm0
- 15,89,203, //mulps %xmm3,%xmm1
- 15,89,211, //mulps %xmm3,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_unpremul_sse2[] = {
- 69,15,87,192, //xorps %xmm8,%xmm8
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,94,203, //divps %xmm3,%xmm9
- 68,15,194,195,4, //cmpneqps %xmm3,%xmm8
- 69,15,84,193, //andps %xmm9,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_from_srgb_sse2[] = {
- 184,145,131,158,61, //mov $0x3d9e8391,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,232, //movaps %xmm8,%xmm13
- 68,15,89,232, //mulps %xmm0,%xmm13
- 68,15,40,224, //movaps %xmm0,%xmm12
- 69,15,89,228, //mulps %xmm12,%xmm12
- 184,154,153,153,62, //mov $0x3e99999a,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 184,92,143,50,63, //mov $0x3f328f5c,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,40,241, //movaps %xmm9,%xmm14
- 68,15,89,240, //mulps %xmm0,%xmm14
- 69,15,88,242, //addps %xmm10,%xmm14
- 184,10,215,35,59, //mov $0x3b23d70a,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,244, //mulps %xmm12,%xmm14
- 69,15,88,243, //addps %xmm11,%xmm14
- 184,174,71,97,61, //mov $0x3d6147ae,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 65,15,194,196,1, //cmpltps %xmm12,%xmm0
- 68,15,84,232, //andps %xmm0,%xmm13
- 65,15,85,198, //andnps %xmm14,%xmm0
- 65,15,86,197, //orps %xmm13,%xmm0
- 69,15,40,232, //movaps %xmm8,%xmm13
- 68,15,89,233, //mulps %xmm1,%xmm13
- 68,15,40,241, //movaps %xmm1,%xmm14
- 69,15,89,246, //mulps %xmm14,%xmm14
- 69,15,40,249, //movaps %xmm9,%xmm15
- 68,15,89,249, //mulps %xmm1,%xmm15
- 69,15,88,250, //addps %xmm10,%xmm15
- 69,15,89,254, //mulps %xmm14,%xmm15
- 69,15,88,251, //addps %xmm11,%xmm15
- 65,15,194,204,1, //cmpltps %xmm12,%xmm1
- 68,15,84,233, //andps %xmm1,%xmm13
- 65,15,85,207, //andnps %xmm15,%xmm1
- 65,15,86,205, //orps %xmm13,%xmm1
- 68,15,89,194, //mulps %xmm2,%xmm8
- 68,15,40,234, //movaps %xmm2,%xmm13
- 69,15,89,237, //mulps %xmm13,%xmm13
- 68,15,89,202, //mulps %xmm2,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 69,15,89,205, //mulps %xmm13,%xmm9
- 69,15,88,203, //addps %xmm11,%xmm9
- 65,15,194,212,1, //cmpltps %xmm12,%xmm2
- 68,15,84,194, //andps %xmm2,%xmm8
- 65,15,85,209, //andnps %xmm9,%xmm2
- 65,15,86,208, //orps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_to_srgb_sse2[] = {
- 68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,248, //rcpps %xmm8,%xmm15
- 69,15,82,232, //rsqrtps %xmm8,%xmm13
- 184,41,92,71,65, //mov $0x41475c29,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,240, //movaps %xmm8,%xmm14
- 68,15,89,240, //mulps %xmm0,%xmm14
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 184,194,135,210,62, //mov $0x3ed287c2,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 184,206,111,48,63, //mov $0x3f306fce,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 184,168,87,202,61, //mov $0x3dca57a8,%eax
- 53,0,0,0,128, //xor $0x80000000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,89,234, //mulps %xmm10,%xmm13
- 69,15,88,239, //addps %xmm15,%xmm13
- 69,15,40,249, //movaps %xmm9,%xmm15
- 69,15,93,253, //minps %xmm13,%xmm15
- 184,4,231,140,59, //mov $0x3b8ce704,%eax
- 102,68,15,110,232, //movd %eax,%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 65,15,194,197,1, //cmpltps %xmm13,%xmm0
- 68,15,84,240, //andps %xmm0,%xmm14
- 65,15,85,199, //andnps %xmm15,%xmm0
- 65,15,86,198, //orps %xmm14,%xmm0
- 68,15,82,241, //rsqrtps %xmm1,%xmm14
- 69,15,83,254, //rcpps %xmm14,%xmm15
- 69,15,82,246, //rsqrtps %xmm14,%xmm14
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,89,242, //mulps %xmm10,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 69,15,40,249, //movaps %xmm9,%xmm15
- 69,15,93,254, //minps %xmm14,%xmm15
- 69,15,40,240, //movaps %xmm8,%xmm14
- 68,15,89,241, //mulps %xmm1,%xmm14
- 65,15,194,205,1, //cmpltps %xmm13,%xmm1
- 68,15,84,241, //andps %xmm1,%xmm14
- 65,15,85,207, //andnps %xmm15,%xmm1
- 65,15,86,206, //orps %xmm14,%xmm1
- 68,15,82,242, //rsqrtps %xmm2,%xmm14
- 69,15,83,254, //rcpps %xmm14,%xmm15
- 69,15,89,251, //mulps %xmm11,%xmm15
- 69,15,88,252, //addps %xmm12,%xmm15
- 69,15,82,222, //rsqrtps %xmm14,%xmm11
- 69,15,89,218, //mulps %xmm10,%xmm11
- 69,15,88,223, //addps %xmm15,%xmm11
- 69,15,93,203, //minps %xmm11,%xmm9
- 68,15,89,194, //mulps %xmm2,%xmm8
- 65,15,194,213,1, //cmpltps %xmm13,%xmm2
- 68,15,84,194, //andps %xmm2,%xmm8
- 65,15,85,209, //andnps %xmm9,%xmm2
- 65,15,86,208, //orps %xmm8,%xmm2
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_1_float_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_scale_u8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,110,4,56, //movd (%rax,%rdi,1),%xmm8
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,69,15,96,193, //punpcklbw %xmm9,%xmm8
- 102,69,15,97,193, //punpcklwd %xmm9,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 65,15,89,201, //mulps %xmm9,%xmm1
- 65,15,89,209, //mulps %xmm9,%xmm2
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_1_float_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_u8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,68,15,110,4,56, //movd (%rax,%rdi,1),%xmm8
- 102,69,15,239,201, //pxor %xmm9,%xmm9
- 102,69,15,96,193, //punpcklbw %xmm9,%xmm8
- 102,69,15,97,193, //punpcklwd %xmm9,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,200, //movd %eax,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 69,15,89,200, //mulps %xmm8,%xmm9
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,193, //mulps %xmm9,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 15,92,223, //subps %xmm7,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 15,88,223, //addps %xmm7,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_lerp_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,68,15,126,4,120, //movq (%rax,%rdi,2),%xmm8
- 102,15,239,219, //pxor %xmm3,%xmm3
- 102,68,15,97,195, //punpcklwd %xmm3,%xmm8
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,68,15,110,208, //movd %eax,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,209, //mulps %xmm9,%xmm10
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,216, //pand %xmm8,%xmm3
- 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 15,92,196, //subps %xmm4,%xmm0
- 65,15,89,194, //mulps %xmm10,%xmm0
- 15,88,196, //addps %xmm4,%xmm0
- 15,92,205, //subps %xmm5,%xmm1
- 65,15,89,203, //mulps %xmm11,%xmm1
- 15,88,205, //addps %xmm5,%xmm1
- 15,92,214, //subps %xmm6,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 15,88,214, //addps %xmm6,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_tables_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,8, //mov (%rax),%rcx
- 76,139,64,8, //mov 0x8(%rax),%r8
- 243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 185,255,0,0,0, //mov $0xff,%ecx
- 102,15,110,193, //movd %ecx,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 102,65,15,114,209,8, //psrld $0x8,%xmm9
- 102,68,15,219,200, //pand %xmm0,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 102,65,15,114,210,16, //psrld $0x10,%xmm10
- 102,68,15,219,208, //pand %xmm0,%xmm10
- 102,65,15,219,192, //pand %xmm8,%xmm0
- 102,15,112,216,78, //pshufd $0x4e,%xmm0,%xmm3
- 102,72,15,126,217, //movq %xmm3,%rcx
- 65,137,201, //mov %ecx,%r9d
- 72,193,233,32, //shr $0x20,%rcx
- 102,73,15,126,194, //movq %xmm0,%r10
- 69,137,211, //mov %r10d,%r11d
- 73,193,234,32, //shr $0x20,%r10
- 243,67,15,16,28,144, //movss (%r8,%r10,4),%xmm3
- 243,65,15,16,4,136, //movss (%r8,%rcx,4),%xmm0
- 15,20,216, //unpcklps %xmm0,%xmm3
- 243,67,15,16,4,152, //movss (%r8,%r11,4),%xmm0
- 243,67,15,16,12,136, //movss (%r8,%r9,4),%xmm1
- 15,20,193, //unpcklps %xmm1,%xmm0
- 15,20,195, //unpcklps %xmm3,%xmm0
- 76,139,64,16, //mov 0x10(%rax),%r8
- 102,65,15,112,201,78, //pshufd $0x4e,%xmm9,%xmm1
- 102,73,15,126,202, //movq %xmm1,%r10
- 77,137,209, //mov %r10,%r9
- 73,193,233,32, //shr $0x20,%r9
- 102,76,15,126,201, //movq %xmm9,%rcx
- 65,137,203, //mov %ecx,%r11d
- 65,129,227,255,255,255,0, //and $0xffffff,%r11d
- 72,193,233,30, //shr $0x1e,%rcx
- 65,129,226,255,255,255,0, //and $0xffffff,%r10d
- 243,65,15,16,28,8, //movss (%r8,%rcx,1),%xmm3
- 243,67,15,16,12,136, //movss (%r8,%r9,4),%xmm1
- 15,20,217, //unpcklps %xmm1,%xmm3
- 243,67,15,16,12,152, //movss (%r8,%r11,4),%xmm1
- 243,67,15,16,20,144, //movss (%r8,%r10,4),%xmm2
- 15,20,202, //unpcklps %xmm2,%xmm1
- 15,20,203, //unpcklps %xmm3,%xmm1
- 76,139,72,24, //mov 0x18(%rax),%r9
- 102,65,15,112,210,78, //pshufd $0x4e,%xmm10,%xmm2
- 102,72,15,126,209, //movq %xmm2,%rcx
- 68,15,183,193, //movzwl %cx,%r8d
- 72,193,233,32, //shr $0x20,%rcx
- 102,76,15,126,208, //movq %xmm10,%rax
- 68,15,183,208, //movzwl %ax,%r10d
- 72,193,232,30, //shr $0x1e,%rax
- 243,69,15,16,12,1, //movss (%r9,%rax,1),%xmm9
- 243,65,15,16,20,137, //movss (%r9,%rcx,4),%xmm2
- 68,15,20,202, //unpcklps %xmm2,%xmm9
- 243,67,15,16,20,145, //movss (%r9,%r10,4),%xmm2
- 243,67,15,16,28,129, //movss (%r9,%r8,4),%xmm3
- 15,20,211, //unpcklps %xmm3,%xmm2
- 65,15,20,209, //unpcklps %xmm9,%xmm2
- 102,65,15,114,208,24, //psrld $0x18,%xmm8
- 69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_a8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 102,15,110,4,56, //movd (%rax,%rdi,1),%xmm0
- 102,15,239,201, //pxor %xmm1,%xmm1
- 102,15,96,193, //punpcklbw %xmm1,%xmm0
- 102,15,97,193, //punpcklwd %xmm1,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 102,15,239,201, //pxor %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_a8_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,16, //pslld $0x10,%xmm8
- 102,65,15,114,224,16, //psrad $0x10,%xmm8
- 102,69,15,107,192, //packssdw %xmm8,%xmm8
- 102,69,15,103,192, //packuswb %xmm8,%xmm8
- 102,68,15,126,4,56, //movd %xmm8,(%rax,%rdi,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,126,20,120, //movq (%rax,%rdi,2),%xmm2
- 102,15,239,192, //pxor %xmm0,%xmm0
- 102,15,97,208, //punpcklwd %xmm0,%xmm2
- 184,0,248,0,0, //mov $0xf800,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,219,194, //pand %xmm2,%xmm0
- 15,91,200, //cvtdq2ps %xmm0,%xmm1
- 184,8,33,132,55, //mov $0x37842108,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 184,224,7,0,0, //mov $0x7e0,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,15,219,202, //pand %xmm2,%xmm1
- 15,91,217, //cvtdq2ps %xmm1,%xmm3
- 184,33,8,2,58, //mov $0x3a020821,%eax
- 102,15,110,200, //movd %eax,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
- 184,31,0,0,0, //mov $0x1f,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,15,219,218, //pand %xmm2,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 184,8,33,4,61, //mov $0x3d042108,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 15,89,211, //mulps %xmm3,%xmm2
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_565_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,248,65, //mov $0x41f80000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,11, //pslld $0xb,%xmm9
- 185,0,0,124,66, //mov $0x427c0000,%ecx
- 102,68,15,110,209, //movd %ecx,%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,5, //pslld $0x5,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 68,15,89,194, //mulps %xmm2,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,194, //orpd %xmm10,%xmm8
- 102,65,15,114,240,16, //pslld $0x10,%xmm8
- 102,65,15,114,224,16, //psrad $0x10,%xmm8
- 102,69,15,107,192, //packssdw %xmm8,%xmm8
- 102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_8888_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,28,184, //movdqu (%rax,%rdi,4),%xmm3
- 184,255,0,0,0, //mov $0xff,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,15,114,209,8, //psrld $0x8,%xmm1
- 102,15,219,200, //pand %xmm0,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,15,114,210,16, //psrld $0x10,%xmm2
- 102,15,219,208, //pand %xmm0,%xmm2
- 102,15,219,195, //pand %xmm3,%xmm0
- 15,91,192, //cvtdq2ps %xmm0,%xmm0
- 184,129,128,128,59, //mov $0x3b808081,%eax
- 102,68,15,110,192, //movd %eax,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,89,192, //mulps %xmm8,%xmm0
- 15,91,201, //cvtdq2ps %xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 15,91,210, //cvtdq2ps %xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
- 102,15,114,211,24, //psrld $0x18,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 65,15,89,216, //mulps %xmm8,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_8888_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,127,67, //mov $0x437f0000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,8, //pslld $0x8,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,202, //mulps %xmm2,%xmm9
- 102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,16, //pslld $0x10,%xmm9
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,65,15,114,240,24, //pslld $0x18,%xmm8
- 102,69,15,235,193, //por %xmm9,%xmm8
- 102,69,15,235,194, //por %xmm10,%xmm8
- 243,68,15,127,4,184, //movdqu %xmm8,(%rax,%rdi,4)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_load_f16_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 243,15,111,4,248, //movdqu (%rax,%rdi,8),%xmm0
- 243,15,111,76,248,16, //movdqu 0x10(%rax,%rdi,8),%xmm1
- 102,15,111,208, //movdqa %xmm0,%xmm2
- 102,15,97,209, //punpcklwd %xmm1,%xmm2
- 102,15,105,193, //punpckhwd %xmm1,%xmm0
- 102,68,15,111,194, //movdqa %xmm2,%xmm8
- 102,68,15,97,192, //punpcklwd %xmm0,%xmm8
- 102,15,105,208, //punpckhwd %xmm0,%xmm2
- 184,0,4,0,4, //mov $0x4000400,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
- 102,15,111,203, //movdqa %xmm3,%xmm1
- 102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
- 102,65,15,223,200, //pandn %xmm8,%xmm1
- 102,15,101,218, //pcmpgtw %xmm2,%xmm3
- 102,15,223,218, //pandn %xmm2,%xmm3
- 102,69,15,239,192, //pxor %xmm8,%xmm8
- 102,15,111,193, //movdqa %xmm1,%xmm0
- 102,65,15,97,192, //punpcklwd %xmm8,%xmm0
- 102,15,114,240,13, //pslld $0xd,%xmm0
- 184,0,0,128,119, //mov $0x77800000,%eax
- 102,15,110,208, //movd %eax,%xmm2
- 102,68,15,112,202,0, //pshufd $0x0,%xmm2,%xmm9
- 65,15,89,193, //mulps %xmm9,%xmm0
- 102,65,15,105,200, //punpckhwd %xmm8,%xmm1
- 102,15,114,241,13, //pslld $0xd,%xmm1
- 65,15,89,201, //mulps %xmm9,%xmm1
- 102,15,111,211, //movdqa %xmm3,%xmm2
- 102,65,15,97,208, //punpcklwd %xmm8,%xmm2
- 102,15,114,242,13, //pslld $0xd,%xmm2
- 65,15,89,209, //mulps %xmm9,%xmm2
- 102,65,15,105,216, //punpckhwd %xmm8,%xmm3
- 102,15,114,243,13, //pslld $0xd,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f16_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 185,0,0,128,7, //mov $0x7800000,%ecx
- 102,68,15,110,193, //movd %ecx,%xmm8
- 102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
- 102,69,15,111,200, //movdqa %xmm8,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 102,65,15,114,209,13, //psrld $0xd,%xmm9
- 102,69,15,111,208, //movdqa %xmm8,%xmm10
- 68,15,89,209, //mulps %xmm1,%xmm10
- 102,65,15,114,210,13, //psrld $0xd,%xmm10
- 102,69,15,111,216, //movdqa %xmm8,%xmm11
- 68,15,89,218, //mulps %xmm2,%xmm11
- 102,65,15,114,211,13, //psrld $0xd,%xmm11
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,65,15,114,208,13, //psrld $0xd,%xmm8
- 102,65,15,115,250,2, //pslldq $0x2,%xmm10
- 102,69,15,235,209, //por %xmm9,%xmm10
- 102,65,15,115,248,2, //pslldq $0x2,%xmm8
- 102,69,15,235,195, //por %xmm11,%xmm8
- 102,69,15,111,202, //movdqa %xmm10,%xmm9
- 102,69,15,98,200, //punpckldq %xmm8,%xmm9
- 243,68,15,127,12,248, //movdqu %xmm9,(%rax,%rdi,8)
- 102,69,15,106,208, //punpckhdq %xmm8,%xmm10
- 243,68,15,127,84,248,16, //movdqu %xmm10,0x10(%rax,%rdi,8)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_store_f32_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 72,137,249, //mov %rdi,%rcx
- 72,193,225,4, //shl $0x4,%rcx
- 68,15,40,192, //movaps %xmm0,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 68,15,20,201, //unpcklps %xmm1,%xmm9
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,40,218, //movaps %xmm2,%xmm11
- 68,15,20,219, //unpcklps %xmm3,%xmm11
- 68,15,21,193, //unpckhps %xmm1,%xmm8
- 68,15,21,211, //unpckhps %xmm3,%xmm10
- 69,15,40,225, //movaps %xmm9,%xmm12
- 102,69,15,20,227, //unpcklpd %xmm11,%xmm12
- 69,15,18,217, //movhlps %xmm9,%xmm11
- 69,15,40,200, //movaps %xmm8,%xmm9
- 102,69,15,20,202, //unpcklpd %xmm10,%xmm9
- 69,15,18,208, //movhlps %xmm8,%xmm10
- 102,68,15,17,36,8, //movupd %xmm12,(%rax,%rcx,1)
- 68,15,17,92,8,16, //movups %xmm11,0x10(%rax,%rcx,1)
- 102,68,15,17,76,8,32, //movupd %xmm9,0x20(%rax,%rcx,1)
- 68,15,17,84,8,48, //movups %xmm10,0x30(%rax,%rcx,1)
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,192, //maxps %xmm0,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,192, //pcmpeqd %xmm0,%xmm0
- 102,65,15,254,193, //paddd %xmm9,%xmm0
- 68,15,93,192, //minps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_clamp_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 69,15,87,192, //xorps %xmm8,%xmm8
- 68,15,95,193, //maxps %xmm1,%xmm8
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 102,15,118,201, //pcmpeqd %xmm1,%xmm1
- 102,65,15,254,201, //paddd %xmm9,%xmm1
- 68,15,93,193, //minps %xmm1,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,200, //movaps %xmm8,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,200, //movaps %xmm0,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 243,69,15,91,209, //cvttps2dq %xmm9,%xmm10
- 69,15,91,210, //cvtdq2ps %xmm10,%xmm10
- 69,15,194,202,1, //cmpltps %xmm10,%xmm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,84,217, //andps %xmm9,%xmm11
- 69,15,92,211, //subps %xmm11,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,194, //subps %xmm10,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_repeat_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,0, //movss (%rax),%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,40,201, //movaps %xmm1,%xmm9
- 69,15,94,200, //divps %xmm8,%xmm9
- 243,69,15,91,209, //cvttps2dq %xmm9,%xmm10
- 69,15,91,210, //cvtdq2ps %xmm10,%xmm10
- 69,15,194,202,1, //cmpltps %xmm10,%xmm9
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,216, //movd %eax,%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,84,217, //andps %xmm9,%xmm11
- 69,15,92,211, //subps %xmm11,%xmm10
- 69,15,89,208, //mulps %xmm8,%xmm10
- 65,15,92,202, //subps %xmm10,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_x_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,40,193, //movaps %xmm9,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,92,192, //subps %xmm8,%xmm0
- 243,69,15,88,201, //addss %xmm9,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,40,208, //movaps %xmm0,%xmm10
- 69,15,94,209, //divps %xmm9,%xmm10
- 243,69,15,91,218, //cvttps2dq %xmm10,%xmm11
- 69,15,91,219, //cvtdq2ps %xmm11,%xmm11
- 69,15,194,211,1, //cmpltps %xmm11,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,84,226, //andps %xmm10,%xmm12
- 69,15,87,210, //xorps %xmm10,%xmm10
- 69,15,92,220, //subps %xmm12,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 65,15,92,195, //subps %xmm11,%xmm0
- 65,15,92,192, //subps %xmm8,%xmm0
- 68,15,92,208, //subps %xmm0,%xmm10
- 65,15,84,194, //andps %xmm10,%xmm0
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,193, //minps %xmm9,%xmm0
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_mirror_y_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 243,68,15,16,8, //movss (%rax),%xmm9
- 69,15,40,193, //movaps %xmm9,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,92,200, //subps %xmm8,%xmm1
- 243,69,15,88,201, //addss %xmm9,%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,40,209, //movaps %xmm1,%xmm10
- 69,15,94,209, //divps %xmm9,%xmm10
- 243,69,15,91,218, //cvttps2dq %xmm10,%xmm11
- 69,15,91,219, //cvtdq2ps %xmm11,%xmm11
- 69,15,194,211,1, //cmpltps %xmm11,%xmm10
- 184,0,0,128,63, //mov $0x3f800000,%eax
- 102,68,15,110,224, //movd %eax,%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,84,226, //andps %xmm10,%xmm12
- 69,15,87,210, //xorps %xmm10,%xmm10
- 69,15,92,220, //subps %xmm12,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 65,15,92,203, //subps %xmm11,%xmm1
- 65,15,92,200, //subps %xmm8,%xmm1
- 68,15,92,209, //subps %xmm1,%xmm10
- 65,15,84,202, //andps %xmm10,%xmm1
- 102,69,15,118,201, //pcmpeqd %xmm9,%xmm9
- 102,69,15,254,200, //paddd %xmm8,%xmm9
- 65,15,93,201, //minps %xmm9,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_luminance_to_alpha_sse2[] = {
- 184,208,179,89,62, //mov $0x3e59b3d0,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,216, //mulps %xmm0,%xmm3
- 184,89,23,55,63, //mov $0x3f371759,%eax
- 102,15,110,192, //movd %eax,%xmm0
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 15,89,193, //mulps %xmm1,%xmm0
- 15,88,195, //addps %xmm3,%xmm0
- 184,152,221,147,61, //mov $0x3d93dd98,%eax
- 102,15,110,216, //movd %eax,%xmm3
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 15,89,218, //mulps %xmm2,%xmm3
- 15,88,216, //addps %xmm0,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 15,87,192, //xorps %xmm0,%xmm0
- 15,87,201, //xorps %xmm1,%xmm1
- 15,87,210, //xorps %xmm2,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_2x3_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,16, //movss 0x10(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 72,173, //lods %ds:(%rsi),%rax
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_3x4_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,12, //movss 0xc(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,36, //movss 0x24(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_4x5_sse2[] = {
- 68,15,40,201, //movaps %xmm1,%xmm9
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,15,16,72,4, //movss 0x4(%rax),%xmm1
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,32, //movss 0x20(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,48, //movss 0x30(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,64, //movss 0x40(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,194, //addps %xmm10,%xmm0
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 243,68,15,16,80,20, //movss 0x14(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,36, //movss 0x24(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,52, //movss 0x34(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,68, //movss 0x44(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,89,227, //mulps %xmm3,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 68,15,89,218, //mulps %xmm2,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,209, //mulps %xmm9,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,89,200, //mulps %xmm8,%xmm1
- 65,15,88,202, //addps %xmm10,%xmm1
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,24, //movss 0x18(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,40, //movss 0x28(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,56, //movss 0x38(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,72, //movss 0x48(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 68,15,89,235, //mulps %xmm3,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 68,15,89,226, //mulps %xmm2,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,217, //mulps %xmm9,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 243,68,15,16,88,12, //movss 0xc(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,28, //movss 0x1c(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,104,44, //movss 0x2c(%rax),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,112,60, //movss 0x3c(%rax),%xmm14
- 69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
- 243,68,15,16,120,76, //movss 0x4c(%rax),%xmm15
- 69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
- 68,15,89,243, //mulps %xmm3,%xmm14
- 69,15,88,247, //addps %xmm15,%xmm14
- 68,15,89,234, //mulps %xmm2,%xmm13
- 69,15,88,238, //addps %xmm14,%xmm13
- 69,15,89,225, //mulps %xmm9,%xmm12
- 69,15,88,229, //addps %xmm13,%xmm12
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,210, //movaps %xmm10,%xmm2
- 65,15,40,219, //movaps %xmm11,%xmm3
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_matrix_perspective_sse2[] = {
- 68,15,40,192, //movaps %xmm0,%xmm8
- 72,173, //lods %ds:(%rsi),%rax
- 243,15,16,0, //movss (%rax),%xmm0
- 243,68,15,16,72,4, //movss 0x4(%rax),%xmm9
- 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,8, //movss 0x8(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,201, //mulps %xmm1,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 65,15,89,192, //mulps %xmm8,%xmm0
- 65,15,88,193, //addps %xmm9,%xmm0
- 243,68,15,16,72,12, //movss 0xc(%rax),%xmm9
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,80,16, //movss 0x10(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,20, //movss 0x14(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 69,15,89,200, //mulps %xmm8,%xmm9
- 69,15,88,202, //addps %xmm10,%xmm9
- 243,68,15,16,80,24, //movss 0x18(%rax),%xmm10
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,88,28, //movss 0x1c(%rax),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,96,32, //movss 0x20(%rax),%xmm12
- 69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 68,15,89,217, //mulps %xmm1,%xmm11
- 69,15,88,220, //addps %xmm12,%xmm11
- 69,15,89,208, //mulps %xmm8,%xmm10
- 69,15,88,211, //addps %xmm11,%xmm10
- 65,15,83,202, //rcpps %xmm10,%xmm1
- 15,89,193, //mulps %xmm1,%xmm0
- 68,15,89,201, //mulps %xmm1,%xmm9
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,201, //movaps %xmm9,%xmm1
- 255,224, //jmpq *%rax
-};
-
-CODE const uint8_t sk_linear_gradient_2stops_sse2[] = {
- 72,173, //lods %ds:(%rsi),%rax
- 68,15,16,8, //movups (%rax),%xmm9
- 15,16,88,16, //movups 0x10(%rax),%xmm3
- 68,15,40,195, //movaps %xmm3,%xmm8
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 65,15,40,201, //movaps %xmm9,%xmm1
- 15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 68,15,89,192, //mulps %xmm0,%xmm8
- 68,15,88,193, //addps %xmm1,%xmm8
- 15,40,203, //movaps %xmm3,%xmm1
- 15,198,201,85, //shufps $0x55,%xmm1,%xmm1
- 65,15,40,209, //movaps %xmm9,%xmm2
- 15,198,210,85, //shufps $0x55,%xmm2,%xmm2
- 15,89,200, //mulps %xmm0,%xmm1
- 15,88,202, //addps %xmm2,%xmm1
- 15,40,211, //movaps %xmm3,%xmm2
- 15,198,210,170, //shufps $0xaa,%xmm2,%xmm2
- 69,15,40,209, //movaps %xmm9,%xmm10
- 69,15,198,210,170, //shufps $0xaa,%xmm10,%xmm10
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,210, //addps %xmm10,%xmm2
- 15,198,219,255, //shufps $0xff,%xmm3,%xmm3
- 69,15,198,201,255, //shufps $0xff,%xmm9,%xmm9
- 15,89,216, //mulps %xmm0,%xmm3
- 65,15,88,217, //addps %xmm9,%xmm3
- 72,173, //lods %ds:(%rsi),%rax
- 65,15,40,192, //movaps %xmm8,%xmm0
- 255,224, //jmpq *%rax
-};
-#endif
diff --git a/src/jumper/SkJumper_generated_win.S b/src/jumper/SkJumper_generated_win.S
new file mode 100644
index 0000000000..984169ebed
--- /dev/null
+++ b/src/jumper/SkJumper_generated_win.S
@@ -0,0 +1,5553 @@
+; Copyright 2017 Google Inc.
+;
+; Use of this source code is governed by a BSD-style license that can be
+; found in the LICENSE file.
+
+; This file is generated semi-automatically with this command:
+; $ src/jumper/build_stages.py
+
+_text SEGMENT
+
+PUBLIC _sk_start_pipeline_hsw
+_sk_start_pipeline_hsw LABEL PROC
+ DB 65,87 ; push %r15
+ DB 65,86 ; push %r14
+ DB 65,85 ; push %r13
+ DB 65,84 ; push %r12
+ DB 86 ; push %rsi
+ DB 87 ; push %rdi
+ DB 83 ; push %rbx
+ DB 72,129,236,160,0,0,0 ; sub $0xa0,%rsp
+ DB 197,120,41,188,36,144,0,0,0 ; vmovaps %xmm15,0x90(%rsp)
+ DB 197,120,41,180,36,128,0,0,0 ; vmovaps %xmm14,0x80(%rsp)
+ DB 197,120,41,108,36,112 ; vmovaps %xmm13,0x70(%rsp)
+ DB 197,120,41,100,36,96 ; vmovaps %xmm12,0x60(%rsp)
+ DB 197,120,41,92,36,80 ; vmovaps %xmm11,0x50(%rsp)
+ DB 197,120,41,84,36,64 ; vmovaps %xmm10,0x40(%rsp)
+ DB 197,120,41,76,36,48 ; vmovaps %xmm9,0x30(%rsp)
+ DB 197,120,41,68,36,32 ; vmovaps %xmm8,0x20(%rsp)
+ DB 197,248,41,124,36,16 ; vmovaps %xmm7,0x10(%rsp)
+ DB 197,248,41,52,36 ; vmovaps %xmm6,(%rsp)
+ DB 77,137,205 ; mov %r9,%r13
+ DB 77,137,198 ; mov %r8,%r14
+ DB 72,137,203 ; mov %rcx,%rbx
+ DB 72,137,214 ; mov %rdx,%rsi
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 73,137,199 ; mov %rax,%r15
+ DB 73,137,244 ; mov %rsi,%r12
+ DB 72,141,67,8 ; lea 0x8(%rbx),%rax
+ DB 76,57,232 ; cmp %r13,%rax
+ DB 118,5 ; jbe 75 <_sk_start_pipeline_hsw+0x75>
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 235,65 ; jmp b6 <_sk_start_pipeline_hsw+0xb6>
+ DB 185,0,0,0,0 ; mov $0x0,%ecx
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 76,137,230 ; mov %r12,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,215 ; callq *%r15
+ DB 72,141,123,8 ; lea 0x8(%rbx),%rdi
+ DB 72,131,195,16 ; add $0x10,%rbx
+ DB 76,57,235 ; cmp %r13,%rbx
+ DB 72,137,251 ; mov %rdi,%rbx
+ DB 118,191 ; jbe 75 <_sk_start_pipeline_hsw+0x75>
+ DB 76,137,233 ; mov %r13,%rcx
+ DB 72,41,249 ; sub %rdi,%rcx
+ DB 116,41 ; je e7 <_sk_start_pipeline_hsw+0xe7>
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 76,137,230 ; mov %r12,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,215 ; callq *%r15
+ DB 76,137,232 ; mov %r13,%rax
+ DB 197,248,40,52,36 ; vmovaps (%rsp),%xmm6
+ DB 197,248,40,124,36,16 ; vmovaps 0x10(%rsp),%xmm7
+ DB 197,120,40,68,36,32 ; vmovaps 0x20(%rsp),%xmm8
+ DB 197,120,40,76,36,48 ; vmovaps 0x30(%rsp),%xmm9
+ DB 197,120,40,84,36,64 ; vmovaps 0x40(%rsp),%xmm10
+ DB 197,120,40,92,36,80 ; vmovaps 0x50(%rsp),%xmm11
+ DB 197,120,40,100,36,96 ; vmovaps 0x60(%rsp),%xmm12
+ DB 197,120,40,108,36,112 ; vmovaps 0x70(%rsp),%xmm13
+ DB 197,120,40,180,36,128,0,0,0 ; vmovaps 0x80(%rsp),%xmm14
+ DB 197,120,40,188,36,144,0,0,0 ; vmovaps 0x90(%rsp),%xmm15
+ DB 72,129,196,160,0,0,0 ; add $0xa0,%rsp
+ DB 91 ; pop %rbx
+ DB 95 ; pop %rdi
+ DB 94 ; pop %rsi
+ DB 65,92 ; pop %r12
+ DB 65,93 ; pop %r13
+ DB 65,94 ; pop %r14
+ DB 65,95 ; pop %r15
+ DB 197,248,119 ; vzeroupper
+ DB 195 ; retq
+
+PUBLIC _sk_just_return_hsw
+_sk_just_return_hsw LABEL PROC
+ DB 195 ; retq
+
+PUBLIC _sk_seed_shader_hsw
+_sk_seed_shader_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,249,110,199 ; vmovd %edi,%xmm0
+ DB 196,226,125,88,192 ; vpbroadcastd %xmm0,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 65,184,0,0,0,63 ; mov $0x3f000000,%r8d
+ DB 196,193,121,110,200 ; vmovd %r8d,%xmm1
+ DB 196,226,125,88,201 ; vpbroadcastd %xmm1,%ymm1
+ DB 197,252,88,193 ; vaddps %ymm1,%ymm0,%ymm0
+ DB 197,252,88,2 ; vaddps (%rdx),%ymm0,%ymm0
+ DB 196,226,125,24,16 ; vbroadcastss (%rax),%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 197,236,88,201 ; vaddps %ymm1,%ymm2,%ymm1
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,208 ; vmovd %eax,%xmm2
+ DB 196,226,125,88,210 ; vpbroadcastd %xmm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_constant_color_hsw
+_sk_constant_color_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,0 ; vbroadcastss (%rax),%ymm0
+ DB 196,226,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm1
+ DB 196,226,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm2
+ DB 196,226,125,24,88,12 ; vbroadcastss 0xc(%rax),%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clear_hsw
+_sk_clear_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_plus__hsw
+_sk_plus__hsw LABEL PROC
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 197,228,88,223 ; vaddps %ymm7,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_srcover_hsw
+_sk_srcover_hsw LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,92,195 ; vsubps %ymm3,%ymm8,%ymm8
+ DB 196,194,93,184,192 ; vfmadd231ps %ymm8,%ymm4,%ymm0
+ DB 196,194,85,184,200 ; vfmadd231ps %ymm8,%ymm5,%ymm1
+ DB 196,194,77,184,208 ; vfmadd231ps %ymm8,%ymm6,%ymm2
+ DB 196,194,69,184,216 ; vfmadd231ps %ymm8,%ymm7,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_dstover_hsw
+_sk_dstover_hsw LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,92,199 ; vsubps %ymm7,%ymm8,%ymm8
+ DB 196,226,61,168,196 ; vfmadd213ps %ymm4,%ymm8,%ymm0
+ DB 196,226,61,168,205 ; vfmadd213ps %ymm5,%ymm8,%ymm1
+ DB 196,226,61,168,214 ; vfmadd213ps %ymm6,%ymm8,%ymm2
+ DB 196,226,61,168,223 ; vfmadd213ps %ymm7,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_0_hsw
+_sk_clamp_0_hsw LABEL PROC
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 196,193,124,95,192 ; vmaxps %ymm8,%ymm0,%ymm0
+ DB 196,193,116,95,200 ; vmaxps %ymm8,%ymm1,%ymm1
+ DB 196,193,108,95,208 ; vmaxps %ymm8,%ymm2,%ymm2
+ DB 196,193,100,95,216 ; vmaxps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_1_hsw
+_sk_clamp_1_hsw LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 196,193,124,93,192 ; vminps %ymm8,%ymm0,%ymm0
+ DB 196,193,116,93,200 ; vminps %ymm8,%ymm1,%ymm1
+ DB 196,193,108,93,208 ; vminps %ymm8,%ymm2,%ymm2
+ DB 196,193,100,93,216 ; vminps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_a_hsw
+_sk_clamp_a_hsw LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 196,193,100,93,216 ; vminps %ymm8,%ymm3,%ymm3
+ DB 197,252,93,195 ; vminps %ymm3,%ymm0,%ymm0
+ DB 197,244,93,203 ; vminps %ymm3,%ymm1,%ymm1
+ DB 197,236,93,211 ; vminps %ymm3,%ymm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_set_rgb_hsw
+_sk_set_rgb_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,0 ; vbroadcastss (%rax),%ymm0
+ DB 196,226,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm1
+ DB 196,226,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_rb_hsw
+_sk_swap_rb_hsw LABEL PROC
+ DB 197,124,40,192 ; vmovaps %ymm0,%ymm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,194 ; vmovaps %ymm2,%ymm0
+ DB 197,124,41,194 ; vmovaps %ymm8,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_hsw
+_sk_swap_hsw LABEL PROC
+ DB 197,124,40,195 ; vmovaps %ymm3,%ymm8
+ DB 197,124,40,202 ; vmovaps %ymm2,%ymm9
+ DB 197,124,40,209 ; vmovaps %ymm1,%ymm10
+ DB 197,124,40,216 ; vmovaps %ymm0,%ymm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,196 ; vmovaps %ymm4,%ymm0
+ DB 197,252,40,205 ; vmovaps %ymm5,%ymm1
+ DB 197,252,40,214 ; vmovaps %ymm6,%ymm2
+ DB 197,252,40,223 ; vmovaps %ymm7,%ymm3
+ DB 197,124,41,220 ; vmovaps %ymm11,%ymm4
+ DB 197,124,41,213 ; vmovaps %ymm10,%ymm5
+ DB 197,124,41,206 ; vmovaps %ymm9,%ymm6
+ DB 197,124,41,199 ; vmovaps %ymm8,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_src_dst_hsw
+_sk_move_src_dst_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,224 ; vmovaps %ymm0,%ymm4
+ DB 197,252,40,233 ; vmovaps %ymm1,%ymm5
+ DB 197,252,40,242 ; vmovaps %ymm2,%ymm6
+ DB 197,252,40,251 ; vmovaps %ymm3,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_dst_src_hsw
+_sk_move_dst_src_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,196 ; vmovaps %ymm4,%ymm0
+ DB 197,252,40,205 ; vmovaps %ymm5,%ymm1
+ DB 197,252,40,214 ; vmovaps %ymm6,%ymm2
+ DB 197,252,40,223 ; vmovaps %ymm7,%ymm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_premul_hsw
+_sk_premul_hsw LABEL PROC
+ DB 197,252,89,195 ; vmulps %ymm3,%ymm0,%ymm0
+ DB 197,244,89,203 ; vmulps %ymm3,%ymm1,%ymm1
+ DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_unpremul_hsw
+_sk_unpremul_hsw LABEL PROC
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 196,65,100,194,200,0 ; vcmpeqps %ymm8,%ymm3,%ymm9
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,66,125,88,210 ; vpbroadcastd %xmm10,%ymm10
+ DB 197,44,94,211 ; vdivps %ymm3,%ymm10,%ymm10
+ DB 196,67,45,74,192,144 ; vblendvps %ymm9,%ymm8,%ymm10,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_from_srgb_hsw
+_sk_from_srgb_hsw LABEL PROC
+ DB 184,145,131,158,61 ; mov $0x3d9e8391,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 197,124,89,208 ; vmulps %ymm0,%ymm0,%ymm10
+ DB 184,154,153,153,62 ; mov $0x3e99999a,%eax
+ DB 197,121,110,216 ; vmovd %eax,%xmm11
+ DB 196,66,125,88,219 ; vpbroadcastd %xmm11,%ymm11
+ DB 184,92,143,50,63 ; mov $0x3f328f5c,%eax
+ DB 197,121,110,224 ; vmovd %eax,%xmm12
+ DB 196,66,125,88,228 ; vpbroadcastd %xmm12,%ymm12
+ DB 196,65,125,111,235 ; vmovdqa %ymm11,%ymm13
+ DB 196,66,125,168,236 ; vfmadd213ps %ymm12,%ymm0,%ymm13
+ DB 184,10,215,35,59 ; mov $0x3b23d70a,%eax
+ DB 197,121,110,240 ; vmovd %eax,%xmm14
+ DB 196,66,125,88,246 ; vpbroadcastd %xmm14,%ymm14
+ DB 196,66,45,168,238 ; vfmadd213ps %ymm14,%ymm10,%ymm13
+ DB 184,174,71,97,61 ; mov $0x3d6147ae,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,66,125,88,210 ; vpbroadcastd %xmm10,%ymm10
+ DB 196,193,124,194,194,1 ; vcmpltps %ymm10,%ymm0,%ymm0
+ DB 196,195,21,74,193,0 ; vblendvps %ymm0,%ymm9,%ymm13,%ymm0
+ DB 197,60,89,201 ; vmulps %ymm1,%ymm8,%ymm9
+ DB 197,116,89,233 ; vmulps %ymm1,%ymm1,%ymm13
+ DB 196,65,125,111,251 ; vmovdqa %ymm11,%ymm15
+ DB 196,66,117,168,252 ; vfmadd213ps %ymm12,%ymm1,%ymm15
+ DB 196,66,21,168,254 ; vfmadd213ps %ymm14,%ymm13,%ymm15
+ DB 196,193,116,194,202,1 ; vcmpltps %ymm10,%ymm1,%ymm1
+ DB 196,195,5,74,201,16 ; vblendvps %ymm1,%ymm9,%ymm15,%ymm1
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 197,108,89,202 ; vmulps %ymm2,%ymm2,%ymm9
+ DB 196,66,109,168,220 ; vfmadd213ps %ymm12,%ymm2,%ymm11
+ DB 196,66,53,168,222 ; vfmadd213ps %ymm14,%ymm9,%ymm11
+ DB 196,193,108,194,210,1 ; vcmpltps %ymm10,%ymm2,%ymm2
+ DB 196,195,37,74,208,32 ; vblendvps %ymm2,%ymm8,%ymm11,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_to_srgb_hsw
+_sk_to_srgb_hsw LABEL PROC
+ DB 197,124,82,192 ; vrsqrtps %ymm0,%ymm8
+ DB 196,65,124,83,216 ; vrcpps %ymm8,%ymm11
+ DB 196,65,124,82,224 ; vrsqrtps %ymm8,%ymm12
+ DB 184,41,92,71,65 ; mov $0x41475c29,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,89,232 ; vmulps %ymm0,%ymm8,%ymm13
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,66,125,88,201 ; vpbroadcastd %xmm9,%ymm9
+ DB 184,194,135,210,62 ; mov $0x3ed287c2,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,66,125,88,210 ; vpbroadcastd %xmm10,%ymm10
+ DB 184,206,111,48,63 ; mov $0x3f306fce,%eax
+ DB 197,121,110,240 ; vmovd %eax,%xmm14
+ DB 196,66,125,88,246 ; vpbroadcastd %xmm14,%ymm14
+ DB 184,168,87,202,61 ; mov $0x3dca57a8,%eax
+ DB 53,0,0,0,128 ; xor $0x80000000,%eax
+ DB 197,121,110,248 ; vmovd %eax,%xmm15
+ DB 196,66,125,88,255 ; vpbroadcastd %xmm15,%ymm15
+ DB 196,66,13,168,223 ; vfmadd213ps %ymm15,%ymm14,%ymm11
+ DB 196,66,45,184,220 ; vfmadd231ps %ymm12,%ymm10,%ymm11
+ DB 196,65,52,93,219 ; vminps %ymm11,%ymm9,%ymm11
+ DB 184,4,231,140,59 ; mov $0x3b8ce704,%eax
+ DB 197,121,110,224 ; vmovd %eax,%xmm12
+ DB 196,66,125,88,228 ; vpbroadcastd %xmm12,%ymm12
+ DB 196,193,124,194,196,1 ; vcmpltps %ymm12,%ymm0,%ymm0
+ DB 196,195,37,74,197,0 ; vblendvps %ymm0,%ymm13,%ymm11,%ymm0
+ DB 197,124,82,217 ; vrsqrtps %ymm1,%ymm11
+ DB 196,65,124,83,235 ; vrcpps %ymm11,%ymm13
+ DB 196,65,124,82,219 ; vrsqrtps %ymm11,%ymm11
+ DB 196,66,13,168,239 ; vfmadd213ps %ymm15,%ymm14,%ymm13
+ DB 196,66,45,184,235 ; vfmadd231ps %ymm11,%ymm10,%ymm13
+ DB 197,60,89,217 ; vmulps %ymm1,%ymm8,%ymm11
+ DB 196,65,52,93,237 ; vminps %ymm13,%ymm9,%ymm13
+ DB 196,193,116,194,204,1 ; vcmpltps %ymm12,%ymm1,%ymm1
+ DB 196,195,21,74,203,16 ; vblendvps %ymm1,%ymm11,%ymm13,%ymm1
+ DB 197,124,82,218 ; vrsqrtps %ymm2,%ymm11
+ DB 196,65,124,83,235 ; vrcpps %ymm11,%ymm13
+ DB 196,66,13,168,239 ; vfmadd213ps %ymm15,%ymm14,%ymm13
+ DB 196,65,124,82,219 ; vrsqrtps %ymm11,%ymm11
+ DB 196,66,45,184,235 ; vfmadd231ps %ymm11,%ymm10,%ymm13
+ DB 196,65,52,93,205 ; vminps %ymm13,%ymm9,%ymm9
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 196,193,108,194,212,1 ; vcmpltps %ymm12,%ymm2,%ymm2
+ DB 196,195,53,74,208,32 ; vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_1_float_hsw
+_sk_scale_1_float_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_u8_hsw
+_sk_scale_u8_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,56 ; jne 556 <_sk_scale_u8_hsw+0x48>
+ DB 197,122,126,0 ; vmovq (%rax),%xmm8
+ DB 196,66,125,49,192 ; vpmovzxbd %xmm8,%ymm8
+ DB 196,65,124,91,192 ; vcvtdq2ps %ymm8,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,66,125,88,201 ; vpbroadcastd %xmm9,%ymm9
+ DB 196,65,60,89,193 ; vmulps %ymm9,%ymm8,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne 55e <_sk_scale_u8_hsw+0x50>
+ DB 196,65,249,110,193 ; vmovq %r9,%xmm8
+ DB 235,167 ; jmp 522 <_sk_scale_u8_hsw+0x14>
+
+PUBLIC _sk_lerp_1_float_hsw
+_sk_lerp_1_float_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,226,61,168,196 ; vfmadd213ps %ymm4,%ymm8,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,226,61,168,205 ; vfmadd213ps %ymm5,%ymm8,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 196,226,61,168,214 ; vfmadd213ps %ymm6,%ymm8,%ymm2
+ DB 197,228,92,223 ; vsubps %ymm7,%ymm3,%ymm3
+ DB 196,226,61,168,223 ; vfmadd213ps %ymm7,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_u8_hsw
+_sk_lerp_u8_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,76 ; jne 606 <_sk_lerp_u8_hsw+0x5c>
+ DB 197,122,126,0 ; vmovq (%rax),%xmm8
+ DB 196,66,125,49,192 ; vpmovzxbd %xmm8,%ymm8
+ DB 196,65,124,91,192 ; vcvtdq2ps %ymm8,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,66,125,88,201 ; vpbroadcastd %xmm9,%ymm9
+ DB 196,65,60,89,193 ; vmulps %ymm9,%ymm8,%ymm8
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,226,61,168,196 ; vfmadd213ps %ymm4,%ymm8,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,226,61,168,205 ; vfmadd213ps %ymm5,%ymm8,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 196,226,61,168,214 ; vfmadd213ps %ymm6,%ymm8,%ymm2
+ DB 197,228,92,223 ; vsubps %ymm7,%ymm3,%ymm3
+ DB 196,226,61,168,223 ; vfmadd213ps %ymm7,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne 60e <_sk_lerp_u8_hsw+0x64>
+ DB 196,65,249,110,193 ; vmovq %r9,%xmm8
+ DB 235,147 ; jmp 5be <_sk_lerp_u8_hsw+0x14>
+
+PUBLIC _sk_lerp_565_hsw
+_sk_lerp_565_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,16 ; mov (%rax),%r10
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,179,0,0,0 ; jne 6ec <_sk_lerp_565_hsw+0xc1>
+ DB 196,193,122,111,28,122 ; vmovdqu (%r10,%rdi,2),%xmm3
+ DB 196,98,125,51,195 ; vpmovzxwd %xmm3,%ymm8
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 196,193,101,219,216 ; vpand %ymm8,%ymm3,%ymm3
+ DB 197,124,91,203 ; vcvtdq2ps %ymm3,%ymm9
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,52,89,203 ; vmulps %ymm3,%ymm9,%ymm9
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 196,193,101,219,216 ; vpand %ymm8,%ymm3,%ymm3
+ DB 197,124,91,211 ; vcvtdq2ps %ymm3,%ymm10
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,44,89,211 ; vmulps %ymm3,%ymm10,%ymm10
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 196,193,101,219,216 ; vpand %ymm8,%ymm3,%ymm3
+ DB 197,124,91,195 ; vcvtdq2ps %ymm3,%ymm8
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,226,53,168,196 ; vfmadd213ps %ymm4,%ymm9,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,226,45,168,205 ; vfmadd213ps %ymm5,%ymm10,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 196,226,101,168,214 ; vfmadd213ps %ymm6,%ymm3,%ymm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 197,225,239,219 ; vpxor %xmm3,%xmm3,%xmm3
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 15,135,59,255,255,255 ; ja 63f <_sk_lerp_565_hsw+0x14>
+ DB 69,15,182,192 ; movzbl %r8b,%r8d
+ DB 76,141,13,73,0,0,0 ; lea 0x49(%rip),%r9 # 758 <_sk_lerp_565_hsw+0x12d>
+ DB 75,99,4,129 ; movslq (%r9,%r8,4),%rax
+ DB 76,1,200 ; add %r9,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,225,239,219 ; vpxor %xmm3,%xmm3,%xmm3
+ DB 196,193,97,196,92,122,12,6 ; vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,92,122,10,5 ; vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,92,122,8,4 ; vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,92,122,6,3 ; vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,92,122,4,2 ; vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,92,122,2,1 ; vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
+ DB 196,193,97,196,28,122,0 ; vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
+ DB 233,231,254,255,255 ; jmpq 63f <_sk_lerp_565_hsw+0x14>
+ DB 244 ; hlt
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 236 ; in (%dx),%al
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,228 ; jmpq *%rsp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 220,255 ; fdivr %st,%st(7)
+ DB 255 ; (bad)
+ DB 255,212 ; callq *%rsp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,204 ; dec %esp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,192 ; inc %eax
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_tables_hsw
+_sk_load_tables_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,141,12,189,0,0,0,0 ; lea 0x0(,%rdi,4),%r9
+ DB 76,3,8 ; add (%rax),%r9
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,121 ; jne 802 <_sk_load_tables_hsw+0x8e>
+ DB 196,193,126,111,25 ; vmovdqu (%r9),%ymm3
+ DB 185,255,0,0,0 ; mov $0xff,%ecx
+ DB 197,249,110,193 ; vmovd %ecx,%xmm0
+ DB 196,226,125,88,208 ; vpbroadcastd %xmm0,%ymm2
+ DB 197,237,219,203 ; vpand %ymm3,%ymm2,%ymm1
+ DB 196,65,61,118,192 ; vpcmpeqd %ymm8,%ymm8,%ymm8
+ DB 72,139,72,8 ; mov 0x8(%rax),%rcx
+ DB 76,139,72,16 ; mov 0x10(%rax),%r9
+ DB 196,65,53,118,201 ; vpcmpeqd %ymm9,%ymm9,%ymm9
+ DB 196,226,53,146,4,137 ; vgatherdps %ymm9,(%rcx,%ymm1,4),%ymm0
+ DB 197,245,114,211,8 ; vpsrld $0x8,%ymm3,%ymm1
+ DB 197,109,219,201 ; vpand %ymm1,%ymm2,%ymm9
+ DB 196,65,45,118,210 ; vpcmpeqd %ymm10,%ymm10,%ymm10
+ DB 196,130,45,146,12,137 ; vgatherdps %ymm10,(%r9,%ymm9,4),%ymm1
+ DB 72,139,64,24 ; mov 0x18(%rax),%rax
+ DB 197,181,114,211,16 ; vpsrld $0x10,%ymm3,%ymm9
+ DB 196,65,109,219,201 ; vpand %ymm9,%ymm2,%ymm9
+ DB 196,162,61,146,20,136 ; vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
+ DB 197,229,114,211,24 ; vpsrld $0x18,%ymm3,%ymm3
+ DB 197,124,91,195 ; vcvtdq2ps %ymm3,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 185,8,0,0,0 ; mov $0x8,%ecx
+ DB 68,41,193 ; sub %r8d,%ecx
+ DB 192,225,3 ; shl $0x3,%cl
+ DB 73,199,194,255,255,255,255 ; mov $0xffffffffffffffff,%r10
+ DB 73,211,234 ; shr %cl,%r10
+ DB 196,193,249,110,194 ; vmovq %r10,%xmm0
+ DB 196,226,125,33,192 ; vpmovsxbd %xmm0,%ymm0
+ DB 196,194,125,140,25 ; vpmaskmovd (%r9),%ymm0,%ymm3
+ DB 233,99,255,255,255 ; jmpq 78e <_sk_load_tables_hsw+0x1a>
+
+PUBLIC _sk_load_a8_hsw
+_sk_load_a8_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,50 ; jne 86d <_sk_load_a8_hsw+0x42>
+ DB 197,250,126,0 ; vmovq (%rax),%xmm0
+ DB 196,226,125,49,192 ; vpmovzxbd %xmm0,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,226,125,88,201 ; vpbroadcastd %xmm1,%ymm1
+ DB 197,252,89,217 ; vmulps %ymm1,%ymm0,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne 875 <_sk_load_a8_hsw+0x4a>
+ DB 196,193,249,110,193 ; vmovq %r9,%xmm0
+ DB 235,173 ; jmp 83f <_sk_load_a8_hsw+0x14>
+
+PUBLIC _sk_store_a8_hsw
+_sk_store_a8_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,8 ; mov (%rax),%r9
+ DB 184,0,0,127,67 ; mov $0x437f0000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,66,57,43,193 ; vpackusdw %xmm9,%xmm8,%xmm8
+ DB 196,65,57,103,192 ; vpackuswb %xmm8,%xmm8,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,10 ; jne 8cd <_sk_store_a8_hsw+0x3b>
+ DB 196,65,123,17,4,57 ; vmovsd %xmm8,(%r9,%rdi,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 119,236 ; ja 8c9 <_sk_store_a8_hsw+0x37>
+ DB 196,66,121,48,192 ; vpmovzxbw %xmm8,%xmm8
+ DB 65,15,182,192 ; movzbl %r8b,%eax
+ DB 76,141,5,67,0,0,0 ; lea 0x43(%rip),%r8 # 930 <_sk_store_a8_hsw+0x9e>
+ DB 73,99,4,128 ; movslq (%r8,%rax,4),%rax
+ DB 76,1,192 ; add %r8,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,67,121,20,68,57,6,12 ; vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,5,10 ; vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,4,8 ; vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,3,6 ; vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,2,4 ; vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,1,2 ; vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
+ DB 196,67,121,20,4,57,0 ; vpextrb $0x0,%xmm8,(%r9,%rdi,1)
+ DB 235,154 ; jmp 8c9 <_sk_store_a8_hsw+0x37>
+ DB 144 ; nop
+ DB 246,255 ; idiv %bh
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 238 ; out %al,(%dx)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,230 ; jmpq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 222,255 ; fdivrp %st,%st(7)
+ DB 255 ; (bad)
+ DB 255,214 ; callq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,206 ; dec %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,198 ; inc %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_565_hsw
+_sk_load_565_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,16 ; mov (%rax),%r10
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,149,0,0,0 ; jne 9ef <_sk_load_565_hsw+0xa3>
+ DB 196,193,122,111,4,122 ; vmovdqu (%r10,%rdi,2),%xmm0
+ DB 196,226,125,51,208 ; vpmovzxwd %xmm0,%ymm2
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 197,249,110,192 ; vmovd %eax,%xmm0
+ DB 196,226,125,88,192 ; vpbroadcastd %xmm0,%ymm0
+ DB 197,253,219,194 ; vpand %ymm2,%ymm0,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,226,125,88,201 ; vpbroadcastd %xmm1,%ymm1
+ DB 197,252,89,193 ; vmulps %ymm1,%ymm0,%ymm0
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,226,125,88,201 ; vpbroadcastd %xmm1,%ymm1
+ DB 197,245,219,202 ; vpand %ymm2,%ymm1,%ymm1
+ DB 197,252,91,201 ; vcvtdq2ps %ymm1,%ymm1
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,244,89,203 ; vmulps %ymm3,%ymm1,%ymm1
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,229,219,210 ; vpand %ymm2,%ymm3,%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 15,135,89,255,255,255 ; ja 960 <_sk_load_565_hsw+0x14>
+ DB 69,15,182,192 ; movzbl %r8b,%r8d
+ DB 76,141,13,74,0,0,0 ; lea 0x4a(%rip),%r9 # a5c <_sk_load_565_hsw+0x110>
+ DB 75,99,4,129 ; movslq (%r9,%r8,4),%rax
+ DB 76,1,200 ; add %r9,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
+ DB 196,193,121,196,68,122,12,6 ; vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,10,5 ; vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,8,4 ; vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,6,3 ; vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,4,2 ; vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,2,1 ; vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,4,122,0 ; vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
+ DB 233,5,255,255,255 ; jmpq 960 <_sk_load_565_hsw+0x14>
+ DB 144 ; nop
+ DB 243,255 ; repz (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 235,255 ; jmp a61 <_sk_load_565_hsw+0x115>
+ DB 255 ; (bad)
+ DB 255,227 ; jmpq *%rbx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 219,255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,211 ; callq *%rbx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,203 ; dec %ebx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 191 ; .byte 0xbf
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_store_565_hsw
+_sk_store_565_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,8 ; mov (%rax),%r9
+ DB 184,0,0,248,65 ; mov $0x41f80000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 196,65,125,91,201 ; vcvtps2dq %ymm9,%ymm9
+ DB 196,193,53,114,241,11 ; vpslld $0xb,%ymm9,%ymm9
+ DB 184,0,0,124,66 ; mov $0x427c0000,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,66,125,88,210 ; vpbroadcastd %xmm10,%ymm10
+ DB 197,44,89,209 ; vmulps %ymm1,%ymm10,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,45,114,242,5 ; vpslld $0x5,%ymm10,%ymm10
+ DB 196,65,45,235,201 ; vpor %ymm9,%ymm10,%ymm9
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,65,53,235,192 ; vpor %ymm8,%ymm9,%ymm8
+ DB 196,67,125,57,193,1 ; vextracti128 $0x1,%ymm8,%xmm9
+ DB 196,66,57,43,193 ; vpackusdw %xmm9,%xmm8,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,10 ; jne ae4 <_sk_store_565_hsw+0x6c>
+ DB 196,65,122,127,4,121 ; vmovdqu %xmm8,(%r9,%rdi,2)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 119,236 ; ja ae0 <_sk_store_565_hsw+0x68>
+ DB 65,15,182,192 ; movzbl %r8b,%eax
+ DB 76,141,5,69,0,0,0 ; lea 0x45(%rip),%r8 # b44 <_sk_store_565_hsw+0xcc>
+ DB 73,99,4,128 ; movslq (%r8,%rax,4),%rax
+ DB 76,1,192 ; add %r8,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,67,121,21,68,121,12,6 ; vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,10,5 ; vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,8,4 ; vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,6,3 ; vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,4,2 ; vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,2,1 ; vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
+ DB 196,67,121,21,4,121,0 ; vpextrw $0x0,%xmm8,(%r9,%rdi,2)
+ DB 235,159 ; jmp ae0 <_sk_store_565_hsw+0x68>
+ DB 15,31,0 ; nopl (%rax)
+ DB 244 ; hlt
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 236 ; in (%dx),%al
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,228 ; jmpq *%rsp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 220,255 ; fdivr %st,%st(7)
+ DB 255 ; (bad)
+ DB 255,212 ; callq *%rsp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,204 ; dec %esp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,196 ; inc %esp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_8888_hsw
+_sk_load_8888_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,141,12,189,0,0,0,0 ; lea 0x0(,%rdi,4),%r9
+ DB 76,3,8 ; add (%rax),%r9
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,104 ; jne bdd <_sk_load_8888_hsw+0x7d>
+ DB 196,193,126,111,25 ; vmovdqu (%r9),%ymm3
+ DB 184,255,0,0,0 ; mov $0xff,%eax
+ DB 197,249,110,192 ; vmovd %eax,%xmm0
+ DB 196,226,125,88,208 ; vpbroadcastd %xmm0,%ymm2
+ DB 197,237,219,195 ; vpand %ymm3,%ymm2,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,98,125,88,193 ; vpbroadcastd %xmm1,%ymm8
+ DB 196,193,124,89,192 ; vmulps %ymm8,%ymm0,%ymm0
+ DB 197,245,114,211,8 ; vpsrld $0x8,%ymm3,%ymm1
+ DB 197,237,219,201 ; vpand %ymm1,%ymm2,%ymm1
+ DB 197,252,91,201 ; vcvtdq2ps %ymm1,%ymm1
+ DB 196,193,116,89,200 ; vmulps %ymm8,%ymm1,%ymm1
+ DB 197,181,114,211,16 ; vpsrld $0x10,%ymm3,%ymm9
+ DB 196,193,109,219,209 ; vpand %ymm9,%ymm2,%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 196,193,108,89,208 ; vmulps %ymm8,%ymm2,%ymm2
+ DB 197,229,114,211,24 ; vpsrld $0x18,%ymm3,%ymm3
+ DB 197,252,91,219 ; vcvtdq2ps %ymm3,%ymm3
+ DB 196,193,100,89,216 ; vmulps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 185,8,0,0,0 ; mov $0x8,%ecx
+ DB 68,41,193 ; sub %r8d,%ecx
+ DB 192,225,3 ; shl $0x3,%cl
+ DB 72,199,192,255,255,255,255 ; mov $0xffffffffffffffff,%rax
+ DB 72,211,232 ; shr %cl,%rax
+ DB 196,225,249,110,192 ; vmovq %rax,%xmm0
+ DB 196,226,125,33,192 ; vpmovsxbd %xmm0,%ymm0
+ DB 196,194,125,140,25 ; vpmaskmovd (%r9),%ymm0,%ymm3
+ DB 233,116,255,255,255 ; jmpq b7a <_sk_load_8888_hsw+0x1a>
+
+PUBLIC _sk_store_8888_hsw
+_sk_store_8888_hsw LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,141,12,189,0,0,0,0 ; lea 0x0(,%rdi,4),%r9
+ DB 76,3,8 ; add (%rax),%r9
+ DB 184,0,0,127,67 ; mov $0x437f0000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,66,125,88,192 ; vpbroadcastd %xmm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 196,65,125,91,201 ; vcvtps2dq %ymm9,%ymm9
+ DB 197,60,89,209 ; vmulps %ymm1,%ymm8,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,45,114,242,8 ; vpslld $0x8,%ymm10,%ymm10
+ DB 196,65,45,235,201 ; vpor %ymm9,%ymm10,%ymm9
+ DB 197,60,89,210 ; vmulps %ymm2,%ymm8,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,45,114,242,16 ; vpslld $0x10,%ymm10,%ymm10
+ DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,193,61,114,240,24 ; vpslld $0x18,%ymm8,%ymm8
+ DB 196,65,45,235,192 ; vpor %ymm8,%ymm10,%ymm8
+ DB 196,65,53,235,192 ; vpor %ymm8,%ymm9,%ymm8
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,12 ; jne c7a <_sk_store_8888_hsw+0x74>
+ DB 196,65,126,127,1 ; vmovdqu %ymm8,(%r9)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 185,8,0,0,0 ; mov $0x8,%ecx
+ DB 68,41,193 ; sub %r8d,%ecx
+ DB 192,225,3 ; shl $0x3,%cl
+ DB 72,199,192,255,255,255,255 ; mov $0xffffffffffffffff,%rax
+ DB 72,211,232 ; shr %cl,%rax
+ DB 196,97,249,110,200 ; vmovq %rax,%xmm9
+ DB 196,66,125,33,201 ; vpmovsxbd %xmm9,%ymm9
+ DB 196,66,53,142,1 ; vpmaskmovd %ymm8,%ymm9,(%r9)
+ DB 235,211 ; jmp c73 <_sk_store_8888_hsw+0x6d>
+
+PUBLIC _sk_load_f16_hsw
+_sk_load_f16_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,97 ; jne d0b <_sk_load_f16_hsw+0x6b>
+ DB 197,121,16,4,248 ; vmovupd (%rax,%rdi,8),%xmm8
+ DB 197,249,16,84,248,16 ; vmovupd 0x10(%rax,%rdi,8),%xmm2
+ DB 197,249,16,92,248,32 ; vmovupd 0x20(%rax,%rdi,8),%xmm3
+ DB 197,122,111,76,248,48 ; vmovdqu 0x30(%rax,%rdi,8),%xmm9
+ DB 197,185,97,194 ; vpunpcklwd %xmm2,%xmm8,%xmm0
+ DB 197,185,105,210 ; vpunpckhwd %xmm2,%xmm8,%xmm2
+ DB 196,193,97,97,201 ; vpunpcklwd %xmm9,%xmm3,%xmm1
+ DB 196,193,97,105,217 ; vpunpckhwd %xmm9,%xmm3,%xmm3
+ DB 197,121,97,194 ; vpunpcklwd %xmm2,%xmm0,%xmm8
+ DB 197,121,105,202 ; vpunpckhwd %xmm2,%xmm0,%xmm9
+ DB 197,241,97,211 ; vpunpcklwd %xmm3,%xmm1,%xmm2
+ DB 197,241,105,219 ; vpunpckhwd %xmm3,%xmm1,%xmm3
+ DB 197,185,108,194 ; vpunpcklqdq %xmm2,%xmm8,%xmm0
+ DB 196,226,125,19,192 ; vcvtph2ps %xmm0,%ymm0
+ DB 197,185,109,202 ; vpunpckhqdq %xmm2,%xmm8,%xmm1
+ DB 196,226,125,19,201 ; vcvtph2ps %xmm1,%ymm1
+ DB 197,177,108,211 ; vpunpcklqdq %xmm3,%xmm9,%xmm2
+ DB 196,226,125,19,210 ; vcvtph2ps %xmm2,%ymm2
+ DB 197,177,109,219 ; vpunpckhqdq %xmm3,%xmm9,%xmm3
+ DB 196,226,125,19,219 ; vcvtph2ps %xmm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,123,16,4,248 ; vmovsd (%rax,%rdi,8),%xmm8
+ DB 196,65,49,239,201 ; vpxor %xmm9,%xmm9,%xmm9
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,79 ; je d6a <_sk_load_f16_hsw+0xca>
+ DB 197,57,22,68,248,8 ; vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,67 ; jb d6a <_sk_load_f16_hsw+0xca>
+ DB 197,251,16,84,248,16 ; vmovsd 0x10(%rax,%rdi,8),%xmm2
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 116,68 ; je d77 <_sk_load_f16_hsw+0xd7>
+ DB 197,233,22,84,248,24 ; vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,56 ; jb d77 <_sk_load_f16_hsw+0xd7>
+ DB 197,251,16,92,248,32 ; vmovsd 0x20(%rax,%rdi,8),%xmm3
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 15,132,114,255,255,255 ; je cc1 <_sk_load_f16_hsw+0x21>
+ DB 197,225,22,92,248,40 ; vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 15,130,98,255,255,255 ; jb cc1 <_sk_load_f16_hsw+0x21>
+ DB 197,122,126,76,248,48 ; vmovq 0x30(%rax,%rdi,8),%xmm9
+ DB 233,87,255,255,255 ; jmpq cc1 <_sk_load_f16_hsw+0x21>
+ DB 197,225,87,219 ; vxorpd %xmm3,%xmm3,%xmm3
+ DB 197,233,87,210 ; vxorpd %xmm2,%xmm2,%xmm2
+ DB 233,74,255,255,255 ; jmpq cc1 <_sk_load_f16_hsw+0x21>
+ DB 197,225,87,219 ; vxorpd %xmm3,%xmm3,%xmm3
+ DB 233,65,255,255,255 ; jmpq cc1 <_sk_load_f16_hsw+0x21>
+
+PUBLIC _sk_store_f16_hsw
+_sk_store_f16_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 196,195,125,29,192,4 ; vcvtps2ph $0x4,%ymm0,%xmm8
+ DB 196,195,125,29,201,4 ; vcvtps2ph $0x4,%ymm1,%xmm9
+ DB 196,195,125,29,210,4 ; vcvtps2ph $0x4,%ymm2,%xmm10
+ DB 196,195,125,29,219,4 ; vcvtps2ph $0x4,%ymm3,%xmm11
+ DB 196,65,57,97,225 ; vpunpcklwd %xmm9,%xmm8,%xmm12
+ DB 196,65,57,105,193 ; vpunpckhwd %xmm9,%xmm8,%xmm8
+ DB 196,65,41,97,203 ; vpunpcklwd %xmm11,%xmm10,%xmm9
+ DB 196,65,41,105,235 ; vpunpckhwd %xmm11,%xmm10,%xmm13
+ DB 196,65,25,98,217 ; vpunpckldq %xmm9,%xmm12,%xmm11
+ DB 196,65,25,106,209 ; vpunpckhdq %xmm9,%xmm12,%xmm10
+ DB 196,65,57,98,205 ; vpunpckldq %xmm13,%xmm8,%xmm9
+ DB 196,65,57,106,197 ; vpunpckhdq %xmm13,%xmm8,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,27 ; jne de5 <_sk_store_f16_hsw+0x65>
+ DB 197,120,17,28,248 ; vmovups %xmm11,(%rax,%rdi,8)
+ DB 197,120,17,84,248,16 ; vmovups %xmm10,0x10(%rax,%rdi,8)
+ DB 197,120,17,76,248,32 ; vmovups %xmm9,0x20(%rax,%rdi,8)
+ DB 197,122,127,68,248,48 ; vmovdqu %xmm8,0x30(%rax,%rdi,8)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,121,214,28,248 ; vmovq %xmm11,(%rax,%rdi,8)
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,241 ; je de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,23,92,248,8 ; vmovhpd %xmm11,0x8(%rax,%rdi,8)
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,229 ; jb de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,214,84,248,16 ; vmovq %xmm10,0x10(%rax,%rdi,8)
+ DB 116,221 ; je de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,23,84,248,24 ; vmovhpd %xmm10,0x18(%rax,%rdi,8)
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,209 ; jb de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,214,76,248,32 ; vmovq %xmm9,0x20(%rax,%rdi,8)
+ DB 116,201 ; je de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,23,76,248,40 ; vmovhpd %xmm9,0x28(%rax,%rdi,8)
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 114,189 ; jb de1 <_sk_store_f16_hsw+0x61>
+ DB 197,121,214,68,248,48 ; vmovq %xmm8,0x30(%rax,%rdi,8)
+ DB 235,181 ; jmp de1 <_sk_store_f16_hsw+0x61>
+
+PUBLIC _sk_store_f32_hsw
+_sk_store_f32_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,0 ; mov (%rax),%r8
+ DB 72,141,4,189,0,0,0,0 ; lea 0x0(,%rdi,4),%rax
+ DB 197,124,20,193 ; vunpcklps %ymm1,%ymm0,%ymm8
+ DB 197,124,21,217 ; vunpckhps %ymm1,%ymm0,%ymm11
+ DB 197,108,20,203 ; vunpcklps %ymm3,%ymm2,%ymm9
+ DB 197,108,21,227 ; vunpckhps %ymm3,%ymm2,%ymm12
+ DB 196,65,61,20,209 ; vunpcklpd %ymm9,%ymm8,%ymm10
+ DB 196,65,61,21,201 ; vunpckhpd %ymm9,%ymm8,%ymm9
+ DB 196,65,37,20,196 ; vunpcklpd %ymm12,%ymm11,%ymm8
+ DB 196,65,37,21,220 ; vunpckhpd %ymm12,%ymm11,%ymm11
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,55 ; jne e99 <_sk_store_f32_hsw+0x6d>
+ DB 196,67,45,24,225,1 ; vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
+ DB 196,67,61,24,235,1 ; vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
+ DB 196,67,45,6,201,49 ; vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
+ DB 196,67,61,6,195,49 ; vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
+ DB 196,65,125,17,36,128 ; vmovupd %ymm12,(%r8,%rax,4)
+ DB 196,65,125,17,108,128,32 ; vmovupd %ymm13,0x20(%r8,%rax,4)
+ DB 196,65,125,17,76,128,64 ; vmovupd %ymm9,0x40(%r8,%rax,4)
+ DB 196,65,125,17,68,128,96 ; vmovupd %ymm8,0x60(%r8,%rax,4)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,65,121,17,20,128 ; vmovupd %xmm10,(%r8,%rax,4)
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,240 ; je e95 <_sk_store_f32_hsw+0x69>
+ DB 196,65,121,17,76,128,16 ; vmovupd %xmm9,0x10(%r8,%rax,4)
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,227 ; jb e95 <_sk_store_f32_hsw+0x69>
+ DB 196,65,121,17,68,128,32 ; vmovupd %xmm8,0x20(%r8,%rax,4)
+ DB 116,218 ; je e95 <_sk_store_f32_hsw+0x69>
+ DB 196,65,121,17,92,128,48 ; vmovupd %xmm11,0x30(%r8,%rax,4)
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,205 ; jb e95 <_sk_store_f32_hsw+0x69>
+ DB 196,67,125,25,84,128,64,1 ; vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
+ DB 116,195 ; je e95 <_sk_store_f32_hsw+0x69>
+ DB 196,67,125,25,76,128,80,1 ; vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 114,181 ; jb e95 <_sk_store_f32_hsw+0x69>
+ DB 196,67,125,25,68,128,96,1 ; vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
+ DB 235,171 ; jmp e95 <_sk_store_f32_hsw+0x69>
+
+PUBLIC _sk_clamp_x_hsw
+_sk_clamp_x_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,188,95,192 ; vmaxps %ymm0,%ymm8,%ymm0
+ DB 196,98,125,88,0 ; vpbroadcastd (%rax),%ymm8
+ DB 196,65,53,118,201 ; vpcmpeqd %ymm9,%ymm9,%ymm9
+ DB 196,65,61,254,193 ; vpaddd %ymm9,%ymm8,%ymm8
+ DB 196,193,124,93,192 ; vminps %ymm8,%ymm0,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_y_hsw
+_sk_clamp_y_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,188,95,201 ; vmaxps %ymm1,%ymm8,%ymm1
+ DB 196,98,125,88,0 ; vpbroadcastd (%rax),%ymm8
+ DB 196,65,53,118,201 ; vpcmpeqd %ymm9,%ymm9,%ymm9
+ DB 196,65,61,254,193 ; vpaddd %ymm9,%ymm8,%ymm8
+ DB 196,193,116,93,200 ; vminps %ymm8,%ymm1,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_x_hsw
+_sk_repeat_x_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,65,124,94,200 ; vdivps %ymm8,%ymm0,%ymm9
+ DB 196,67,125,8,201,1 ; vroundps $0x1,%ymm9,%ymm9
+ DB 196,98,61,172,200 ; vfnmadd213ps %ymm0,%ymm8,%ymm9
+ DB 197,253,118,192 ; vpcmpeqd %ymm0,%ymm0,%ymm0
+ DB 197,189,254,192 ; vpaddd %ymm0,%ymm8,%ymm0
+ DB 197,180,93,192 ; vminps %ymm0,%ymm9,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_y_hsw
+_sk_repeat_y_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,65,116,94,200 ; vdivps %ymm8,%ymm1,%ymm9
+ DB 196,67,125,8,201,1 ; vroundps $0x1,%ymm9,%ymm9
+ DB 196,98,61,172,201 ; vfnmadd213ps %ymm1,%ymm8,%ymm9
+ DB 197,245,118,201 ; vpcmpeqd %ymm1,%ymm1,%ymm1
+ DB 197,189,254,201 ; vpaddd %ymm1,%ymm8,%ymm1
+ DB 197,180,93,201 ; vminps %ymm1,%ymm9,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_x_hsw
+_sk_mirror_x_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,122,16,0 ; vmovss (%rax),%xmm8
+ DB 196,66,125,24,200 ; vbroadcastss %xmm8,%ymm9
+ DB 196,65,124,92,209 ; vsubps %ymm9,%ymm0,%ymm10
+ DB 196,193,58,88,192 ; vaddss %xmm8,%xmm8,%xmm0
+ DB 196,226,125,24,192 ; vbroadcastss %xmm0,%ymm0
+ DB 197,44,94,192 ; vdivps %ymm0,%ymm10,%ymm8
+ DB 196,67,125,8,192,1 ; vroundps $0x1,%ymm8,%ymm8
+ DB 196,66,125,172,194 ; vfnmadd213ps %ymm10,%ymm0,%ymm8
+ DB 196,193,60,92,193 ; vsubps %ymm9,%ymm8,%ymm0
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,92,192 ; vsubps %ymm0,%ymm8,%ymm8
+ DB 197,188,84,192 ; vandps %ymm0,%ymm8,%ymm0
+ DB 196,65,61,118,192 ; vpcmpeqd %ymm8,%ymm8,%ymm8
+ DB 196,65,53,254,192 ; vpaddd %ymm8,%ymm9,%ymm8
+ DB 196,193,124,93,192 ; vminps %ymm8,%ymm0,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_y_hsw
+_sk_mirror_y_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,122,16,0 ; vmovss (%rax),%xmm8
+ DB 196,66,125,24,200 ; vbroadcastss %xmm8,%ymm9
+ DB 196,65,116,92,209 ; vsubps %ymm9,%ymm1,%ymm10
+ DB 196,193,58,88,200 ; vaddss %xmm8,%xmm8,%xmm1
+ DB 196,226,125,24,201 ; vbroadcastss %xmm1,%ymm1
+ DB 197,44,94,193 ; vdivps %ymm1,%ymm10,%ymm8
+ DB 196,67,125,8,192,1 ; vroundps $0x1,%ymm8,%ymm8
+ DB 196,66,117,172,194 ; vfnmadd213ps %ymm10,%ymm1,%ymm8
+ DB 196,193,60,92,201 ; vsubps %ymm9,%ymm8,%ymm1
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,92,193 ; vsubps %ymm1,%ymm8,%ymm8
+ DB 197,188,84,201 ; vandps %ymm1,%ymm8,%ymm1
+ DB 196,65,61,118,192 ; vpcmpeqd %ymm8,%ymm8,%ymm8
+ DB 196,65,53,254,192 ; vpaddd %ymm8,%ymm9,%ymm8
+ DB 196,193,116,93,200 ; vminps %ymm8,%ymm1,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_luminance_to_alpha_hsw
+_sk_luminance_to_alpha_hsw LABEL PROC
+ DB 184,208,179,89,62 ; mov $0x3e59b3d0,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,98,125,88,195 ; vpbroadcastd %xmm3,%ymm8
+ DB 184,89,23,55,63 ; mov $0x3f371759,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,226,125,88,219 ; vpbroadcastd %xmm3,%ymm3
+ DB 197,228,89,201 ; vmulps %ymm1,%ymm3,%ymm1
+ DB 196,98,125,168,193 ; vfmadd213ps %ymm1,%ymm0,%ymm8
+ DB 184,152,221,147,61 ; mov $0x3d93dd98,%eax
+ DB 197,249,110,192 ; vmovd %eax,%xmm0
+ DB 196,226,125,88,216 ; vpbroadcastd %xmm0,%ymm3
+ DB 196,194,109,168,216 ; vfmadd213ps %ymm8,%ymm2,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,253,239,192 ; vpxor %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_2x3_hsw
+_sk_matrix_2x3_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,8 ; vbroadcastss (%rax),%ymm9
+ DB 196,98,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm10
+ DB 196,98,125,24,64,16 ; vbroadcastss 0x10(%rax),%ymm8
+ DB 196,66,117,184,194 ; vfmadd231ps %ymm10,%ymm1,%ymm8
+ DB 196,66,125,184,193 ; vfmadd231ps %ymm9,%ymm0,%ymm8
+ DB 196,98,125,24,80,4 ; vbroadcastss 0x4(%rax),%ymm10
+ DB 196,98,125,24,88,12 ; vbroadcastss 0xc(%rax),%ymm11
+ DB 196,98,125,24,72,20 ; vbroadcastss 0x14(%rax),%ymm9
+ DB 196,66,117,184,203 ; vfmadd231ps %ymm11,%ymm1,%ymm9
+ DB 196,66,125,184,202 ; vfmadd231ps %ymm10,%ymm0,%ymm9
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 197,124,41,201 ; vmovaps %ymm9,%ymm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_3x4_hsw
+_sk_matrix_3x4_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,8 ; vbroadcastss (%rax),%ymm9
+ DB 196,98,125,24,80,12 ; vbroadcastss 0xc(%rax),%ymm10
+ DB 196,98,125,24,88,24 ; vbroadcastss 0x18(%rax),%ymm11
+ DB 196,98,125,24,64,36 ; vbroadcastss 0x24(%rax),%ymm8
+ DB 196,66,109,184,195 ; vfmadd231ps %ymm11,%ymm2,%ymm8
+ DB 196,66,117,184,194 ; vfmadd231ps %ymm10,%ymm1,%ymm8
+ DB 196,66,125,184,193 ; vfmadd231ps %ymm9,%ymm0,%ymm8
+ DB 196,98,125,24,80,4 ; vbroadcastss 0x4(%rax),%ymm10
+ DB 196,98,125,24,88,16 ; vbroadcastss 0x10(%rax),%ymm11
+ DB 196,98,125,24,96,28 ; vbroadcastss 0x1c(%rax),%ymm12
+ DB 196,98,125,24,72,40 ; vbroadcastss 0x28(%rax),%ymm9
+ DB 196,66,109,184,204 ; vfmadd231ps %ymm12,%ymm2,%ymm9
+ DB 196,66,117,184,203 ; vfmadd231ps %ymm11,%ymm1,%ymm9
+ DB 196,66,125,184,202 ; vfmadd231ps %ymm10,%ymm0,%ymm9
+ DB 196,98,125,24,88,8 ; vbroadcastss 0x8(%rax),%ymm11
+ DB 196,98,125,24,96,20 ; vbroadcastss 0x14(%rax),%ymm12
+ DB 196,98,125,24,104,32 ; vbroadcastss 0x20(%rax),%ymm13
+ DB 196,98,125,24,80,44 ; vbroadcastss 0x2c(%rax),%ymm10
+ DB 196,66,109,184,213 ; vfmadd231ps %ymm13,%ymm2,%ymm10
+ DB 196,66,117,184,212 ; vfmadd231ps %ymm12,%ymm1,%ymm10
+ DB 196,66,125,184,211 ; vfmadd231ps %ymm11,%ymm0,%ymm10
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 197,124,41,201 ; vmovaps %ymm9,%ymm1
+ DB 197,124,41,210 ; vmovaps %ymm10,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_4x5_hsw
+_sk_matrix_4x5_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,8 ; vbroadcastss (%rax),%ymm9
+ DB 196,98,125,24,80,16 ; vbroadcastss 0x10(%rax),%ymm10
+ DB 196,98,125,24,88,32 ; vbroadcastss 0x20(%rax),%ymm11
+ DB 196,98,125,24,96,48 ; vbroadcastss 0x30(%rax),%ymm12
+ DB 196,98,125,24,64,64 ; vbroadcastss 0x40(%rax),%ymm8
+ DB 196,66,101,184,196 ; vfmadd231ps %ymm12,%ymm3,%ymm8
+ DB 196,66,109,184,195 ; vfmadd231ps %ymm11,%ymm2,%ymm8
+ DB 196,66,117,184,194 ; vfmadd231ps %ymm10,%ymm1,%ymm8
+ DB 196,66,125,184,193 ; vfmadd231ps %ymm9,%ymm0,%ymm8
+ DB 196,98,125,24,80,4 ; vbroadcastss 0x4(%rax),%ymm10
+ DB 196,98,125,24,88,20 ; vbroadcastss 0x14(%rax),%ymm11
+ DB 196,98,125,24,96,36 ; vbroadcastss 0x24(%rax),%ymm12
+ DB 196,98,125,24,104,52 ; vbroadcastss 0x34(%rax),%ymm13
+ DB 196,98,125,24,72,68 ; vbroadcastss 0x44(%rax),%ymm9
+ DB 196,66,101,184,205 ; vfmadd231ps %ymm13,%ymm3,%ymm9
+ DB 196,66,109,184,204 ; vfmadd231ps %ymm12,%ymm2,%ymm9
+ DB 196,66,117,184,203 ; vfmadd231ps %ymm11,%ymm1,%ymm9
+ DB 196,66,125,184,202 ; vfmadd231ps %ymm10,%ymm0,%ymm9
+ DB 196,98,125,24,88,8 ; vbroadcastss 0x8(%rax),%ymm11
+ DB 196,98,125,24,96,24 ; vbroadcastss 0x18(%rax),%ymm12
+ DB 196,98,125,24,104,40 ; vbroadcastss 0x28(%rax),%ymm13
+ DB 196,98,125,24,112,56 ; vbroadcastss 0x38(%rax),%ymm14
+ DB 196,98,125,24,80,72 ; vbroadcastss 0x48(%rax),%ymm10
+ DB 196,66,101,184,214 ; vfmadd231ps %ymm14,%ymm3,%ymm10
+ DB 196,66,109,184,213 ; vfmadd231ps %ymm13,%ymm2,%ymm10
+ DB 196,66,117,184,212 ; vfmadd231ps %ymm12,%ymm1,%ymm10
+ DB 196,66,125,184,211 ; vfmadd231ps %ymm11,%ymm0,%ymm10
+ DB 196,98,125,24,96,12 ; vbroadcastss 0xc(%rax),%ymm12
+ DB 196,98,125,24,104,28 ; vbroadcastss 0x1c(%rax),%ymm13
+ DB 196,98,125,24,112,44 ; vbroadcastss 0x2c(%rax),%ymm14
+ DB 196,98,125,24,120,60 ; vbroadcastss 0x3c(%rax),%ymm15
+ DB 196,98,125,24,88,76 ; vbroadcastss 0x4c(%rax),%ymm11
+ DB 196,66,101,184,223 ; vfmadd231ps %ymm15,%ymm3,%ymm11
+ DB 196,66,109,184,222 ; vfmadd231ps %ymm14,%ymm2,%ymm11
+ DB 196,66,117,184,221 ; vfmadd231ps %ymm13,%ymm1,%ymm11
+ DB 196,66,125,184,220 ; vfmadd231ps %ymm12,%ymm0,%ymm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 197,124,41,201 ; vmovaps %ymm9,%ymm1
+ DB 197,124,41,210 ; vmovaps %ymm10,%ymm2
+ DB 197,124,41,219 ; vmovaps %ymm11,%ymm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_perspective_hsw
+_sk_matrix_perspective_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm9
+ DB 196,98,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm10
+ DB 196,66,117,184,209 ; vfmadd231ps %ymm9,%ymm1,%ymm10
+ DB 196,66,125,184,208 ; vfmadd231ps %ymm8,%ymm0,%ymm10
+ DB 196,98,125,24,64,12 ; vbroadcastss 0xc(%rax),%ymm8
+ DB 196,98,125,24,72,16 ; vbroadcastss 0x10(%rax),%ymm9
+ DB 196,98,125,24,88,20 ; vbroadcastss 0x14(%rax),%ymm11
+ DB 196,66,117,184,217 ; vfmadd231ps %ymm9,%ymm1,%ymm11
+ DB 196,66,125,184,216 ; vfmadd231ps %ymm8,%ymm0,%ymm11
+ DB 196,98,125,24,64,24 ; vbroadcastss 0x18(%rax),%ymm8
+ DB 196,98,125,24,72,28 ; vbroadcastss 0x1c(%rax),%ymm9
+ DB 196,98,125,24,96,32 ; vbroadcastss 0x20(%rax),%ymm12
+ DB 196,66,117,184,225 ; vfmadd231ps %ymm9,%ymm1,%ymm12
+ DB 196,66,125,184,224 ; vfmadd231ps %ymm8,%ymm0,%ymm12
+ DB 196,193,124,83,204 ; vrcpps %ymm12,%ymm1
+ DB 197,172,89,193 ; vmulps %ymm1,%ymm10,%ymm0
+ DB 197,164,89,201 ; vmulps %ymm1,%ymm11,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_linear_gradient_2stops_hsw
+_sk_linear_gradient_2stops_hsw LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,72,16 ; vbroadcastss 0x10(%rax),%ymm1
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,184,193 ; vfmadd231ps %ymm1,%ymm0,%ymm8
+ DB 196,226,125,24,80,20 ; vbroadcastss 0x14(%rax),%ymm2
+ DB 196,226,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm1
+ DB 196,226,125,184,202 ; vfmadd231ps %ymm2,%ymm0,%ymm1
+ DB 196,226,125,24,88,24 ; vbroadcastss 0x18(%rax),%ymm3
+ DB 196,226,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm2
+ DB 196,226,125,184,211 ; vfmadd231ps %ymm3,%ymm0,%ymm2
+ DB 196,98,125,24,72,28 ; vbroadcastss 0x1c(%rax),%ymm9
+ DB 196,226,125,24,88,12 ; vbroadcastss 0xc(%rax),%ymm3
+ DB 196,194,125,184,217 ; vfmadd231ps %ymm9,%ymm0,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_start_pipeline_avx
+_sk_start_pipeline_avx LABEL PROC
+ DB 65,87 ; push %r15
+ DB 65,86 ; push %r14
+ DB 65,85 ; push %r13
+ DB 65,84 ; push %r12
+ DB 86 ; push %rsi
+ DB 87 ; push %rdi
+ DB 83 ; push %rbx
+ DB 72,129,236,160,0,0,0 ; sub $0xa0,%rsp
+ DB 197,120,41,188,36,144,0,0,0 ; vmovaps %xmm15,0x90(%rsp)
+ DB 197,120,41,180,36,128,0,0,0 ; vmovaps %xmm14,0x80(%rsp)
+ DB 197,120,41,108,36,112 ; vmovaps %xmm13,0x70(%rsp)
+ DB 197,120,41,100,36,96 ; vmovaps %xmm12,0x60(%rsp)
+ DB 197,120,41,92,36,80 ; vmovaps %xmm11,0x50(%rsp)
+ DB 197,120,41,84,36,64 ; vmovaps %xmm10,0x40(%rsp)
+ DB 197,120,41,76,36,48 ; vmovaps %xmm9,0x30(%rsp)
+ DB 197,120,41,68,36,32 ; vmovaps %xmm8,0x20(%rsp)
+ DB 197,248,41,124,36,16 ; vmovaps %xmm7,0x10(%rsp)
+ DB 197,248,41,52,36 ; vmovaps %xmm6,(%rsp)
+ DB 77,137,205 ; mov %r9,%r13
+ DB 77,137,198 ; mov %r8,%r14
+ DB 72,137,203 ; mov %rcx,%rbx
+ DB 72,137,214 ; mov %rdx,%rsi
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 73,137,199 ; mov %rax,%r15
+ DB 73,137,244 ; mov %rsi,%r12
+ DB 72,141,67,8 ; lea 0x8(%rbx),%rax
+ DB 76,57,232 ; cmp %r13,%rax
+ DB 118,5 ; jbe 75 <_sk_start_pipeline_avx+0x75>
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 235,65 ; jmp b6 <_sk_start_pipeline_avx+0xb6>
+ DB 185,0,0,0,0 ; mov $0x0,%ecx
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 76,137,230 ; mov %r12,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,215 ; callq *%r15
+ DB 72,141,123,8 ; lea 0x8(%rbx),%rdi
+ DB 72,131,195,16 ; add $0x10,%rbx
+ DB 76,57,235 ; cmp %r13,%rbx
+ DB 72,137,251 ; mov %rdi,%rbx
+ DB 118,191 ; jbe 75 <_sk_start_pipeline_avx+0x75>
+ DB 76,137,233 ; mov %r13,%rcx
+ DB 72,41,249 ; sub %rdi,%rcx
+ DB 116,41 ; je e7 <_sk_start_pipeline_avx+0xe7>
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 76,137,230 ; mov %r12,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,215 ; callq *%r15
+ DB 76,137,232 ; mov %r13,%rax
+ DB 197,248,40,52,36 ; vmovaps (%rsp),%xmm6
+ DB 197,248,40,124,36,16 ; vmovaps 0x10(%rsp),%xmm7
+ DB 197,120,40,68,36,32 ; vmovaps 0x20(%rsp),%xmm8
+ DB 197,120,40,76,36,48 ; vmovaps 0x30(%rsp),%xmm9
+ DB 197,120,40,84,36,64 ; vmovaps 0x40(%rsp),%xmm10
+ DB 197,120,40,92,36,80 ; vmovaps 0x50(%rsp),%xmm11
+ DB 197,120,40,100,36,96 ; vmovaps 0x60(%rsp),%xmm12
+ DB 197,120,40,108,36,112 ; vmovaps 0x70(%rsp),%xmm13
+ DB 197,120,40,180,36,128,0,0,0 ; vmovaps 0x80(%rsp),%xmm14
+ DB 197,120,40,188,36,144,0,0,0 ; vmovaps 0x90(%rsp),%xmm15
+ DB 72,129,196,160,0,0,0 ; add $0xa0,%rsp
+ DB 91 ; pop %rbx
+ DB 95 ; pop %rdi
+ DB 94 ; pop %rsi
+ DB 65,92 ; pop %r12
+ DB 65,93 ; pop %r13
+ DB 65,94 ; pop %r14
+ DB 65,95 ; pop %r15
+ DB 197,248,119 ; vzeroupper
+ DB 195 ; retq
+
+PUBLIC _sk_just_return_avx
+_sk_just_return_avx LABEL PROC
+ DB 195 ; retq
+
+PUBLIC _sk_seed_shader_avx
+_sk_seed_shader_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,249,110,199 ; vmovd %edi,%xmm0
+ DB 197,249,112,192,0 ; vpshufd $0x0,%xmm0,%xmm0
+ DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 65,184,0,0,0,63 ; mov $0x3f000000,%r8d
+ DB 196,193,121,110,200 ; vmovd %r8d,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,252,88,193 ; vaddps %ymm1,%ymm0,%ymm0
+ DB 197,252,88,2 ; vaddps (%rdx),%ymm0,%ymm0
+ DB 196,226,125,24,16 ; vbroadcastss (%rax),%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 197,236,88,201 ; vaddps %ymm1,%ymm2,%ymm1
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,208 ; vmovd %eax,%xmm2
+ DB 196,227,121,4,210,0 ; vpermilps $0x0,%xmm2,%xmm2
+ DB 196,227,109,24,210,1 ; vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 197,220,87,228 ; vxorps %ymm4,%ymm4,%ymm4
+ DB 197,212,87,237 ; vxorps %ymm5,%ymm5,%ymm5
+ DB 197,204,87,246 ; vxorps %ymm6,%ymm6,%ymm6
+ DB 197,196,87,255 ; vxorps %ymm7,%ymm7,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_constant_color_avx
+_sk_constant_color_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,0 ; vbroadcastss (%rax),%ymm0
+ DB 196,226,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm1
+ DB 196,226,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm2
+ DB 196,226,125,24,88,12 ; vbroadcastss 0xc(%rax),%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clear_avx
+_sk_clear_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 197,228,87,219 ; vxorps %ymm3,%ymm3,%ymm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_plus__avx
+_sk_plus__avx LABEL PROC
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 197,228,88,223 ; vaddps %ymm7,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_srcover_avx
+_sk_srcover_avx LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,92,195 ; vsubps %ymm3,%ymm8,%ymm8
+ DB 197,60,89,204 ; vmulps %ymm4,%ymm8,%ymm9
+ DB 197,180,88,192 ; vaddps %ymm0,%ymm9,%ymm0
+ DB 197,60,89,205 ; vmulps %ymm5,%ymm8,%ymm9
+ DB 197,180,88,201 ; vaddps %ymm1,%ymm9,%ymm1
+ DB 197,60,89,206 ; vmulps %ymm6,%ymm8,%ymm9
+ DB 197,180,88,210 ; vaddps %ymm2,%ymm9,%ymm2
+ DB 197,60,89,199 ; vmulps %ymm7,%ymm8,%ymm8
+ DB 197,188,88,219 ; vaddps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_dstover_avx
+_sk_dstover_avx LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,92,199 ; vsubps %ymm7,%ymm8,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 197,228,88,223 ; vaddps %ymm7,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_0_avx
+_sk_clamp_0_avx LABEL PROC
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 196,193,124,95,192 ; vmaxps %ymm8,%ymm0,%ymm0
+ DB 196,193,116,95,200 ; vmaxps %ymm8,%ymm1,%ymm1
+ DB 196,193,108,95,208 ; vmaxps %ymm8,%ymm2,%ymm2
+ DB 196,193,100,95,216 ; vmaxps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_1_avx
+_sk_clamp_1_avx LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 196,193,124,93,192 ; vminps %ymm8,%ymm0,%ymm0
+ DB 196,193,116,93,200 ; vminps %ymm8,%ymm1,%ymm1
+ DB 196,193,108,93,208 ; vminps %ymm8,%ymm2,%ymm2
+ DB 196,193,100,93,216 ; vminps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_a_avx
+_sk_clamp_a_avx LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 196,193,100,93,216 ; vminps %ymm8,%ymm3,%ymm3
+ DB 197,252,93,195 ; vminps %ymm3,%ymm0,%ymm0
+ DB 197,244,93,203 ; vminps %ymm3,%ymm1,%ymm1
+ DB 197,236,93,211 ; vminps %ymm3,%ymm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_set_rgb_avx
+_sk_set_rgb_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,0 ; vbroadcastss (%rax),%ymm0
+ DB 196,226,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm1
+ DB 196,226,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_rb_avx
+_sk_swap_rb_avx LABEL PROC
+ DB 197,124,40,192 ; vmovaps %ymm0,%ymm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,194 ; vmovaps %ymm2,%ymm0
+ DB 197,124,41,194 ; vmovaps %ymm8,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_avx
+_sk_swap_avx LABEL PROC
+ DB 197,124,40,195 ; vmovaps %ymm3,%ymm8
+ DB 197,124,40,202 ; vmovaps %ymm2,%ymm9
+ DB 197,124,40,209 ; vmovaps %ymm1,%ymm10
+ DB 197,124,40,216 ; vmovaps %ymm0,%ymm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,196 ; vmovaps %ymm4,%ymm0
+ DB 197,252,40,205 ; vmovaps %ymm5,%ymm1
+ DB 197,252,40,214 ; vmovaps %ymm6,%ymm2
+ DB 197,252,40,223 ; vmovaps %ymm7,%ymm3
+ DB 197,124,41,220 ; vmovaps %ymm11,%ymm4
+ DB 197,124,41,213 ; vmovaps %ymm10,%ymm5
+ DB 197,124,41,206 ; vmovaps %ymm9,%ymm6
+ DB 197,124,41,199 ; vmovaps %ymm8,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_src_dst_avx
+_sk_move_src_dst_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,224 ; vmovaps %ymm0,%ymm4
+ DB 197,252,40,233 ; vmovaps %ymm1,%ymm5
+ DB 197,252,40,242 ; vmovaps %ymm2,%ymm6
+ DB 197,252,40,251 ; vmovaps %ymm3,%ymm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_dst_src_avx
+_sk_move_dst_src_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,40,196 ; vmovaps %ymm4,%ymm0
+ DB 197,252,40,205 ; vmovaps %ymm5,%ymm1
+ DB 197,252,40,214 ; vmovaps %ymm6,%ymm2
+ DB 197,252,40,223 ; vmovaps %ymm7,%ymm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_premul_avx
+_sk_premul_avx LABEL PROC
+ DB 197,252,89,195 ; vmulps %ymm3,%ymm0,%ymm0
+ DB 197,244,89,203 ; vmulps %ymm3,%ymm1,%ymm1
+ DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_unpremul_avx
+_sk_unpremul_avx LABEL PROC
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 196,65,100,194,200,0 ; vcmpeqps %ymm8,%ymm3,%ymm9
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,67,121,4,210,0 ; vpermilps $0x0,%xmm10,%xmm10
+ DB 196,67,45,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ DB 197,44,94,211 ; vdivps %ymm3,%ymm10,%ymm10
+ DB 196,67,45,74,192,144 ; vblendvps %ymm9,%ymm8,%ymm10,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_from_srgb_avx
+_sk_from_srgb_avx LABEL PROC
+ DB 184,145,131,158,61 ; mov $0x3d9e8391,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 197,124,89,208 ; vmulps %ymm0,%ymm0,%ymm10
+ DB 184,154,153,153,62 ; mov $0x3e99999a,%eax
+ DB 197,121,110,216 ; vmovd %eax,%xmm11
+ DB 196,67,121,4,219,0 ; vpermilps $0x0,%xmm11,%xmm11
+ DB 196,67,37,24,219,1 ; vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ DB 184,92,143,50,63 ; mov $0x3f328f5c,%eax
+ DB 197,121,110,224 ; vmovd %eax,%xmm12
+ DB 196,67,121,4,228,0 ; vpermilps $0x0,%xmm12,%xmm12
+ DB 196,67,29,24,228,1 ; vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
+ DB 197,36,89,232 ; vmulps %ymm0,%ymm11,%ymm13
+ DB 196,65,20,88,236 ; vaddps %ymm12,%ymm13,%ymm13
+ DB 184,10,215,35,59 ; mov $0x3b23d70a,%eax
+ DB 197,121,110,240 ; vmovd %eax,%xmm14
+ DB 196,67,121,4,246,0 ; vpermilps $0x0,%xmm14,%xmm14
+ DB 196,67,13,24,246,1 ; vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ DB 196,65,44,89,213 ; vmulps %ymm13,%ymm10,%ymm10
+ DB 196,65,12,88,210 ; vaddps %ymm10,%ymm14,%ymm10
+ DB 184,174,71,97,61 ; mov $0x3d6147ae,%eax
+ DB 197,121,110,232 ; vmovd %eax,%xmm13
+ DB 196,67,121,4,237,0 ; vpermilps $0x0,%xmm13,%xmm13
+ DB 196,67,21,24,237,1 ; vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
+ DB 196,193,124,194,197,1 ; vcmpltps %ymm13,%ymm0,%ymm0
+ DB 196,195,45,74,193,0 ; vblendvps %ymm0,%ymm9,%ymm10,%ymm0
+ DB 197,60,89,201 ; vmulps %ymm1,%ymm8,%ymm9
+ DB 197,116,89,209 ; vmulps %ymm1,%ymm1,%ymm10
+ DB 197,36,89,249 ; vmulps %ymm1,%ymm11,%ymm15
+ DB 196,65,28,88,255 ; vaddps %ymm15,%ymm12,%ymm15
+ DB 196,65,44,89,215 ; vmulps %ymm15,%ymm10,%ymm10
+ DB 196,65,12,88,210 ; vaddps %ymm10,%ymm14,%ymm10
+ DB 196,193,116,194,205,1 ; vcmpltps %ymm13,%ymm1,%ymm1
+ DB 196,195,45,74,201,16 ; vblendvps %ymm1,%ymm9,%ymm10,%ymm1
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 197,108,89,202 ; vmulps %ymm2,%ymm2,%ymm9
+ DB 197,36,89,210 ; vmulps %ymm2,%ymm11,%ymm10
+ DB 196,65,28,88,210 ; vaddps %ymm10,%ymm12,%ymm10
+ DB 196,65,52,89,202 ; vmulps %ymm10,%ymm9,%ymm9
+ DB 196,65,12,88,201 ; vaddps %ymm9,%ymm14,%ymm9
+ DB 196,193,108,194,213,1 ; vcmpltps %ymm13,%ymm2,%ymm2
+ DB 196,195,53,74,208,32 ; vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_to_srgb_avx
+_sk_to_srgb_avx LABEL PROC
+ DB 197,124,82,192 ; vrsqrtps %ymm0,%ymm8
+ DB 196,65,124,83,232 ; vrcpps %ymm8,%ymm13
+ DB 196,65,124,82,240 ; vrsqrtps %ymm8,%ymm14
+ DB 184,41,92,71,65 ; mov $0x41475c29,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,224 ; vmulps %ymm0,%ymm8,%ymm12
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,67,121,4,201,0 ; vpermilps $0x0,%xmm9,%xmm9
+ DB 196,67,53,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ DB 184,194,135,210,62 ; mov $0x3ed287c2,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,67,121,4,210,0 ; vpermilps $0x0,%xmm10,%xmm10
+ DB 196,67,45,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ DB 184,206,111,48,63 ; mov $0x3f306fce,%eax
+ DB 197,121,110,216 ; vmovd %eax,%xmm11
+ DB 196,67,121,4,219,0 ; vpermilps $0x0,%xmm11,%xmm11
+ DB 196,67,37,24,219,1 ; vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ DB 184,168,87,202,61 ; mov $0x3dca57a8,%eax
+ DB 53,0,0,0,128 ; xor $0x80000000,%eax
+ DB 197,121,110,248 ; vmovd %eax,%xmm15
+ DB 196,67,121,4,255,0 ; vpermilps $0x0,%xmm15,%xmm15
+ DB 196,67,5,24,255,1 ; vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
+ DB 196,65,20,89,235 ; vmulps %ymm11,%ymm13,%ymm13
+ DB 196,65,20,88,239 ; vaddps %ymm15,%ymm13,%ymm13
+ DB 196,65,12,89,242 ; vmulps %ymm10,%ymm14,%ymm14
+ DB 196,65,12,88,237 ; vaddps %ymm13,%ymm14,%ymm13
+ DB 196,65,52,93,237 ; vminps %ymm13,%ymm9,%ymm13
+ DB 184,4,231,140,59 ; mov $0x3b8ce704,%eax
+ DB 197,121,110,240 ; vmovd %eax,%xmm14
+ DB 196,67,121,4,246,0 ; vpermilps $0x0,%xmm14,%xmm14
+ DB 196,67,13,24,246,1 ; vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ DB 196,193,124,194,198,1 ; vcmpltps %ymm14,%ymm0,%ymm0
+ DB 196,195,21,74,196,0 ; vblendvps %ymm0,%ymm12,%ymm13,%ymm0
+ DB 197,124,82,225 ; vrsqrtps %ymm1,%ymm12
+ DB 196,65,124,83,236 ; vrcpps %ymm12,%ymm13
+ DB 196,65,124,82,228 ; vrsqrtps %ymm12,%ymm12
+ DB 196,65,36,89,237 ; vmulps %ymm13,%ymm11,%ymm13
+ DB 196,65,4,88,237 ; vaddps %ymm13,%ymm15,%ymm13
+ DB 196,65,44,89,228 ; vmulps %ymm12,%ymm10,%ymm12
+ DB 196,65,28,88,229 ; vaddps %ymm13,%ymm12,%ymm12
+ DB 197,60,89,233 ; vmulps %ymm1,%ymm8,%ymm13
+ DB 196,65,52,93,228 ; vminps %ymm12,%ymm9,%ymm12
+ DB 196,193,116,194,206,1 ; vcmpltps %ymm14,%ymm1,%ymm1
+ DB 196,195,29,74,205,16 ; vblendvps %ymm1,%ymm13,%ymm12,%ymm1
+ DB 197,124,82,226 ; vrsqrtps %ymm2,%ymm12
+ DB 196,65,124,83,236 ; vrcpps %ymm12,%ymm13
+ DB 196,65,36,89,221 ; vmulps %ymm13,%ymm11,%ymm11
+ DB 196,65,4,88,219 ; vaddps %ymm11,%ymm15,%ymm11
+ DB 196,65,124,82,228 ; vrsqrtps %ymm12,%ymm12
+ DB 196,65,44,89,212 ; vmulps %ymm12,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 196,65,52,93,202 ; vminps %ymm10,%ymm9,%ymm9
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 196,193,108,194,214,1 ; vcmpltps %ymm14,%ymm2,%ymm2
+ DB 196,195,53,74,208,32 ; vblendvps %ymm2,%ymm8,%ymm9,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_1_float_avx
+_sk_scale_1_float_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_u8_avx
+_sk_scale_u8_avx LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,80 ; jne 639 <_sk_scale_u8_avx+0x60>
+ DB 197,122,126,0 ; vmovq (%rax),%xmm8
+ DB 196,66,121,49,200 ; vpmovzxbd %xmm8,%xmm9
+ DB 196,67,121,4,192,229 ; vpermilps $0xe5,%xmm8,%xmm8
+ DB 196,66,121,49,192 ; vpmovzxbd %xmm8,%xmm8
+ DB 196,67,53,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
+ DB 196,65,124,91,192 ; vcvtdq2ps %ymm8,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,67,121,4,201,0 ; vpermilps $0x0,%xmm9,%xmm9
+ DB 196,67,53,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ DB 196,65,60,89,193 ; vmulps %ymm9,%ymm8,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,188,89,210 ; vmulps %ymm2,%ymm8,%ymm2
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne 641 <_sk_scale_u8_avx+0x68>
+ DB 196,65,249,110,193 ; vmovq %r9,%xmm8
+ DB 235,143 ; jmp 5ed <_sk_scale_u8_avx+0x14>
+
+PUBLIC _sk_lerp_1_float_avx
+_sk_lerp_1_float_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,193,124,89,192 ; vmulps %ymm8,%ymm0,%ymm0
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,193,116,89,200 ; vmulps %ymm8,%ymm1,%ymm1
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 196,193,108,89,208 ; vmulps %ymm8,%ymm2,%ymm2
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 197,228,92,223 ; vsubps %ymm7,%ymm3,%ymm3
+ DB 196,193,100,89,216 ; vmulps %ymm8,%ymm3,%ymm3
+ DB 197,228,88,223 ; vaddps %ymm7,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_u8_avx
+_sk_lerp_u8_avx LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,116 ; jne 721 <_sk_lerp_u8_avx+0x84>
+ DB 197,122,126,0 ; vmovq (%rax),%xmm8
+ DB 196,66,121,49,200 ; vpmovzxbd %xmm8,%xmm9
+ DB 196,67,121,4,192,229 ; vpermilps $0xe5,%xmm8,%xmm8
+ DB 196,66,121,49,192 ; vpmovzxbd %xmm8,%xmm8
+ DB 196,67,53,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm9,%ymm8
+ DB 196,65,124,91,192 ; vcvtdq2ps %ymm8,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,121,110,200 ; vmovd %eax,%xmm9
+ DB 196,67,121,4,201,0 ; vpermilps $0x0,%xmm9,%xmm9
+ DB 196,67,53,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ DB 196,65,60,89,193 ; vmulps %ymm9,%ymm8,%ymm8
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,193,124,89,192 ; vmulps %ymm8,%ymm0,%ymm0
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,193,116,89,200 ; vmulps %ymm8,%ymm1,%ymm1
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 196,193,108,89,208 ; vmulps %ymm8,%ymm2,%ymm2
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 197,228,92,223 ; vsubps %ymm7,%ymm3,%ymm3
+ DB 196,193,100,89,216 ; vmulps %ymm8,%ymm3,%ymm3
+ DB 197,228,88,223 ; vaddps %ymm7,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne 729 <_sk_lerp_u8_avx+0x8c>
+ DB 196,65,249,110,193 ; vmovq %r9,%xmm8
+ DB 233,104,255,255,255 ; jmpq 6b1 <_sk_lerp_u8_avx+0x14>
+
+PUBLIC _sk_lerp_565_avx
+_sk_lerp_565_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,16 ; mov (%rax),%r10
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,250,0,0,0 ; jne 851 <_sk_lerp_565_avx+0x108>
+ DB 196,65,122,111,4,122 ; vmovdqu (%r10,%rdi,2),%xmm8
+ DB 197,225,239,219 ; vpxor %xmm3,%xmm3,%xmm3
+ DB 197,185,105,219 ; vpunpckhwd %xmm3,%xmm8,%xmm3
+ DB 196,66,121,51,192 ; vpmovzxwd %xmm8,%xmm8
+ DB 196,99,61,24,195,1 ; vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 197,249,112,219,0 ; vpshufd $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
+ DB 197,124,91,203 ; vcvtdq2ps %ymm3,%ymm9
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,52,89,203 ; vmulps %ymm3,%ymm9,%ymm9
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 197,249,112,219,0 ; vpshufd $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
+ DB 197,124,91,211 ; vcvtdq2ps %ymm3,%ymm10
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,44,89,211 ; vmulps %ymm3,%ymm10,%ymm10
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 197,249,112,219,0 ; vpshufd $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
+ DB 197,124,91,195 ; vcvtdq2ps %ymm3,%ymm8
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
+ DB 196,193,124,89,193 ; vmulps %ymm9,%ymm0,%ymm0
+ DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
+ DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
+ DB 196,193,116,89,202 ; vmulps %ymm10,%ymm1,%ymm1
+ DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
+ DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
+ DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
+ DB 197,236,88,214 ; vaddps %ymm6,%ymm2,%ymm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 196,65,57,239,192 ; vpxor %xmm8,%xmm8,%xmm8
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 15,135,243,254,255,255 ; ja 75d <_sk_lerp_565_avx+0x14>
+ DB 69,15,182,192 ; movzbl %r8b,%r8d
+ DB 76,141,13,75,0,0,0 ; lea 0x4b(%rip),%r9 # 8c0 <_sk_lerp_565_avx+0x177>
+ DB 75,99,4,129 ; movslq (%r9,%r8,4),%rax
+ DB 76,1,200 ; add %r9,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,225,239,219 ; vpxor %xmm3,%xmm3,%xmm3
+ DB 196,65,97,196,68,122,12,6 ; vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm3,%xmm8
+ DB 196,65,57,196,68,122,10,5 ; vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm8,%xmm8
+ DB 196,65,57,196,68,122,8,4 ; vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm8,%xmm8
+ DB 196,65,57,196,68,122,6,3 ; vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm8,%xmm8
+ DB 196,65,57,196,68,122,4,2 ; vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
+ DB 196,65,57,196,68,122,2,1 ; vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
+ DB 196,65,57,196,4,122,0 ; vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
+ DB 233,159,254,255,255 ; jmpq 75d <_sk_lerp_565_avx+0x14>
+ DB 102,144 ; xchg %ax,%ax
+ DB 242,255 ; repnz (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 234 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,226 ; jmpq *%rdx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 218,255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,210 ; callq *%rdx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,202 ; dec %edx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 190 ; .byte 0xbe
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_tables_avx
+_sk_load_tables_avx LABEL PROC
+ DB 85 ; push %rbp
+ DB 65,87 ; push %r15
+ DB 65,86 ; push %r14
+ DB 65,85 ; push %r13
+ DB 65,84 ; push %r12
+ DB 83 ; push %rbx
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,0 ; mov (%rax),%r8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,56,2,0,0 ; jne b2c <_sk_load_tables_avx+0x250>
+ DB 196,65,124,16,4,184 ; vmovups (%r8,%rdi,4),%ymm8
+ DB 187,255,0,0,0 ; mov $0xff,%ebx
+ DB 197,249,110,195 ; vmovd %ebx,%xmm0
+ DB 197,249,112,192,0 ; vpshufd $0x0,%xmm0,%xmm0
+ DB 196,99,125,24,200,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
+ DB 196,193,52,84,192 ; vandps %ymm8,%ymm9,%ymm0
+ DB 196,193,249,126,193 ; vmovq %xmm0,%r9
+ DB 69,137,203 ; mov %r9d,%r11d
+ DB 196,195,249,22,194,1 ; vpextrq $0x1,%xmm0,%r10
+ DB 69,137,214 ; mov %r10d,%r14d
+ DB 73,193,234,32 ; shr $0x20,%r10
+ DB 73,193,233,32 ; shr $0x20,%r9
+ DB 196,227,125,25,192,1 ; vextractf128 $0x1,%ymm0,%xmm0
+ DB 196,193,249,126,196 ; vmovq %xmm0,%r12
+ DB 69,137,231 ; mov %r12d,%r15d
+ DB 196,227,249,22,195,1 ; vpextrq $0x1,%xmm0,%rbx
+ DB 65,137,221 ; mov %ebx,%r13d
+ DB 72,193,235,32 ; shr $0x20,%rbx
+ DB 73,193,236,32 ; shr $0x20,%r12
+ DB 72,139,104,8 ; mov 0x8(%rax),%rbp
+ DB 76,139,64,16 ; mov 0x10(%rax),%r8
+ DB 196,161,122,16,68,189,0 ; vmovss 0x0(%rbp,%r15,4),%xmm0
+ DB 196,163,121,33,68,165,0,16 ; vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
+ DB 196,161,122,16,76,173,0 ; vmovss 0x0(%rbp,%r13,4),%xmm1
+ DB 196,227,121,33,193,32 ; vinsertps $0x20,%xmm1,%xmm0,%xmm0
+ DB 197,250,16,76,157,0 ; vmovss 0x0(%rbp,%rbx,4),%xmm1
+ DB 196,227,121,33,193,48 ; vinsertps $0x30,%xmm1,%xmm0,%xmm0
+ DB 196,161,122,16,76,157,0 ; vmovss 0x0(%rbp,%r11,4),%xmm1
+ DB 196,163,113,33,76,141,0,16 ; vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
+ DB 196,161,122,16,92,181,0 ; vmovss 0x0(%rbp,%r14,4),%xmm3
+ DB 196,227,113,33,203,32 ; vinsertps $0x20,%xmm3,%xmm1,%xmm1
+ DB 196,161,122,16,92,149,0 ; vmovss 0x0(%rbp,%r10,4),%xmm3
+ DB 196,227,113,33,203,48 ; vinsertps $0x30,%xmm3,%xmm1,%xmm1
+ DB 196,227,117,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
+ DB 196,193,113,114,208,8 ; vpsrld $0x8,%xmm8,%xmm1
+ DB 196,67,125,25,194,1 ; vextractf128 $0x1,%ymm8,%xmm10
+ DB 196,193,105,114,210,8 ; vpsrld $0x8,%xmm10,%xmm2
+ DB 196,227,117,24,202,1 ; vinsertf128 $0x1,%xmm2,%ymm1,%ymm1
+ DB 197,180,84,201 ; vandps %ymm1,%ymm9,%ymm1
+ DB 196,193,249,126,201 ; vmovq %xmm1,%r9
+ DB 69,137,203 ; mov %r9d,%r11d
+ DB 196,195,249,22,202,1 ; vpextrq $0x1,%xmm1,%r10
+ DB 69,137,214 ; mov %r10d,%r14d
+ DB 73,193,234,32 ; shr $0x20,%r10
+ DB 73,193,233,32 ; shr $0x20,%r9
+ DB 196,227,125,25,201,1 ; vextractf128 $0x1,%ymm1,%xmm1
+ DB 196,225,249,126,205 ; vmovq %xmm1,%rbp
+ DB 65,137,239 ; mov %ebp,%r15d
+ DB 196,227,249,22,203,1 ; vpextrq $0x1,%xmm1,%rbx
+ DB 65,137,220 ; mov %ebx,%r12d
+ DB 72,193,235,32 ; shr $0x20,%rbx
+ DB 72,193,237,32 ; shr $0x20,%rbp
+ DB 196,129,122,16,12,184 ; vmovss (%r8,%r15,4),%xmm1
+ DB 196,195,113,33,12,168,16 ; vinsertps $0x10,(%r8,%rbp,4),%xmm1,%xmm1
+ DB 196,129,122,16,20,160 ; vmovss (%r8,%r12,4),%xmm2
+ DB 196,227,113,33,202,32 ; vinsertps $0x20,%xmm2,%xmm1,%xmm1
+ DB 196,193,122,16,20,152 ; vmovss (%r8,%rbx,4),%xmm2
+ DB 196,227,113,33,202,48 ; vinsertps $0x30,%xmm2,%xmm1,%xmm1
+ DB 196,129,122,16,20,152 ; vmovss (%r8,%r11,4),%xmm2
+ DB 196,131,105,33,20,136,16 ; vinsertps $0x10,(%r8,%r9,4),%xmm2,%xmm2
+ DB 196,129,122,16,28,176 ; vmovss (%r8,%r14,4),%xmm3
+ DB 196,227,105,33,211,32 ; vinsertps $0x20,%xmm3,%xmm2,%xmm2
+ DB 196,129,122,16,28,144 ; vmovss (%r8,%r10,4),%xmm3
+ DB 196,227,105,33,211,48 ; vinsertps $0x30,%xmm3,%xmm2,%xmm2
+ DB 196,227,109,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
+ DB 72,139,64,24 ; mov 0x18(%rax),%rax
+ DB 196,193,105,114,208,16 ; vpsrld $0x10,%xmm8,%xmm2
+ DB 196,193,97,114,210,16 ; vpsrld $0x10,%xmm10,%xmm3
+ DB 196,227,109,24,211,1 ; vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ DB 197,180,84,210 ; vandps %ymm2,%ymm9,%ymm2
+ DB 196,193,249,126,208 ; vmovq %xmm2,%r8
+ DB 69,137,194 ; mov %r8d,%r10d
+ DB 196,195,249,22,209,1 ; vpextrq $0x1,%xmm2,%r9
+ DB 69,137,203 ; mov %r9d,%r11d
+ DB 73,193,233,32 ; shr $0x20,%r9
+ DB 73,193,232,32 ; shr $0x20,%r8
+ DB 196,227,125,25,210,1 ; vextractf128 $0x1,%ymm2,%xmm2
+ DB 196,225,249,126,213 ; vmovq %xmm2,%rbp
+ DB 65,137,238 ; mov %ebp,%r14d
+ DB 196,227,249,22,211,1 ; vpextrq $0x1,%xmm2,%rbx
+ DB 65,137,223 ; mov %ebx,%r15d
+ DB 72,193,235,32 ; shr $0x20,%rbx
+ DB 72,193,237,32 ; shr $0x20,%rbp
+ DB 196,161,122,16,20,176 ; vmovss (%rax,%r14,4),%xmm2
+ DB 196,227,105,33,20,168,16 ; vinsertps $0x10,(%rax,%rbp,4),%xmm2,%xmm2
+ DB 196,161,122,16,28,184 ; vmovss (%rax,%r15,4),%xmm3
+ DB 196,227,105,33,211,32 ; vinsertps $0x20,%xmm3,%xmm2,%xmm2
+ DB 197,250,16,28,152 ; vmovss (%rax,%rbx,4),%xmm3
+ DB 196,99,105,33,203,48 ; vinsertps $0x30,%xmm3,%xmm2,%xmm9
+ DB 196,161,122,16,28,144 ; vmovss (%rax,%r10,4),%xmm3
+ DB 196,163,97,33,28,128,16 ; vinsertps $0x10,(%rax,%r8,4),%xmm3,%xmm3
+ DB 196,161,122,16,20,152 ; vmovss (%rax,%r11,4),%xmm2
+ DB 196,227,97,33,210,32 ; vinsertps $0x20,%xmm2,%xmm3,%xmm2
+ DB 196,161,122,16,28,136 ; vmovss (%rax,%r9,4),%xmm3
+ DB 196,227,105,33,211,48 ; vinsertps $0x30,%xmm3,%xmm2,%xmm2
+ DB 196,195,109,24,209,1 ; vinsertf128 $0x1,%xmm9,%ymm2,%ymm2
+ DB 196,193,57,114,208,24 ; vpsrld $0x18,%xmm8,%xmm8
+ DB 196,193,97,114,210,24 ; vpsrld $0x18,%xmm10,%xmm3
+ DB 196,227,61,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ DB 197,124,91,195 ; vcvtdq2ps %ymm3,%ymm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 91 ; pop %rbx
+ DB 65,92 ; pop %r12
+ DB 65,93 ; pop %r13
+ DB 65,94 ; pop %r14
+ DB 65,95 ; pop %r15
+ DB 93 ; pop %rbp
+ DB 255,224 ; jmpq *%rax
+ DB 137,203 ; mov %ecx,%ebx
+ DB 128,227,7 ; and $0x7,%bl
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 254,203 ; dec %bl
+ DB 128,251,6 ; cmp $0x6,%bl
+ DB 15,135,185,253,255,255 ; ja 8fa <_sk_load_tables_avx+0x1e>
+ DB 15,182,219 ; movzbl %bl,%ebx
+ DB 76,141,13,137,0,0,0 ; lea 0x89(%rip),%r9 # bd4 <_sk_load_tables_avx+0x2f8>
+ DB 73,99,28,153 ; movslq (%r9,%rbx,4),%rbx
+ DB 76,1,203 ; add %r9,%rbx
+ DB 255,227 ; jmpq *%rbx
+ DB 196,193,121,110,68,184,24 ; vmovd 0x18(%r8,%rdi,4),%xmm0
+ DB 197,249,112,192,68 ; vpshufd $0x44,%xmm0,%xmm0
+ DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 196,99,117,12,192,64 ; vblendps $0x40,%ymm0,%ymm1,%ymm8
+ DB 196,99,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm0
+ DB 196,195,121,34,68,184,20,1 ; vpinsrd $0x1,0x14(%r8,%rdi,4),%xmm0,%xmm0
+ DB 196,99,61,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
+ DB 196,99,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm0
+ DB 196,195,121,34,68,184,16,0 ; vpinsrd $0x0,0x10(%r8,%rdi,4),%xmm0,%xmm0
+ DB 196,99,61,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm8,%ymm8
+ DB 196,195,57,34,68,184,12,3 ; vpinsrd $0x3,0xc(%r8,%rdi,4),%xmm8,%xmm0
+ DB 196,99,61,12,192,15 ; vblendps $0xf,%ymm0,%ymm8,%ymm8
+ DB 196,195,57,34,68,184,8,2 ; vpinsrd $0x2,0x8(%r8,%rdi,4),%xmm8,%xmm0
+ DB 196,99,61,12,192,15 ; vblendps $0xf,%ymm0,%ymm8,%ymm8
+ DB 196,195,57,34,68,184,4,1 ; vpinsrd $0x1,0x4(%r8,%rdi,4),%xmm8,%xmm0
+ DB 196,99,61,12,192,15 ; vblendps $0xf,%ymm0,%ymm8,%ymm8
+ DB 196,195,57,34,4,184,0 ; vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
+ DB 196,99,61,12,192,15 ; vblendps $0xf,%ymm0,%ymm8,%ymm8
+ DB 233,38,253,255,255 ; jmpq 8fa <_sk_load_tables_avx+0x1e>
+ DB 238 ; out %al,(%dx)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,224 ; jmpq *%rax
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,210 ; callq *%rdx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,196 ; inc %esp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,176,255,255,255,156 ; pushq -0x63000001(%rax)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+ DB 128,255,255 ; cmp $0xff,%bh
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_a8_avx
+_sk_load_a8_avx LABEL PROC
+ DB 73,137,200 ; mov %rcx,%r8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,1,248 ; add %rdi,%rax
+ DB 77,133,192 ; test %r8,%r8
+ DB 117,74 ; jne c4a <_sk_load_a8_avx+0x5a>
+ DB 197,250,126,0 ; vmovq (%rax),%xmm0
+ DB 196,226,121,49,200 ; vpmovzxbd %xmm0,%xmm1
+ DB 196,227,121,4,192,229 ; vpermilps $0xe5,%xmm0,%xmm0
+ DB 196,226,121,49,192 ; vpmovzxbd %xmm0,%xmm0
+ DB 196,227,117,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,252,89,217 ; vmulps %ymm1,%ymm0,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 76,137,193 ; mov %r8,%rcx
+ DB 255,224 ; jmpq *%rax
+ DB 49,201 ; xor %ecx,%ecx
+ DB 77,137,194 ; mov %r8,%r10
+ DB 69,49,201 ; xor %r9d,%r9d
+ DB 68,15,182,24 ; movzbl (%rax),%r11d
+ DB 72,255,192 ; inc %rax
+ DB 73,211,227 ; shl %cl,%r11
+ DB 77,9,217 ; or %r11,%r9
+ DB 72,131,193,8 ; add $0x8,%rcx
+ DB 73,255,202 ; dec %r10
+ DB 117,234 ; jne c52 <_sk_load_a8_avx+0x62>
+ DB 196,193,249,110,193 ; vmovq %r9,%xmm0
+ DB 235,149 ; jmp c04 <_sk_load_a8_avx+0x14>
+
+PUBLIC _sk_store_a8_avx
+_sk_store_a8_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,8 ; mov (%rax),%r9
+ DB 184,0,0,127,67 ; mov $0x437f0000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,66,57,43,193 ; vpackusdw %xmm9,%xmm8,%xmm8
+ DB 196,65,57,103,192 ; vpackuswb %xmm8,%xmm8,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,10 ; jne cb1 <_sk_store_a8_avx+0x42>
+ DB 196,65,123,17,4,57 ; vmovsd %xmm8,(%r9,%rdi,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 119,236 ; ja cad <_sk_store_a8_avx+0x3e>
+ DB 196,66,121,48,192 ; vpmovzxbw %xmm8,%xmm8
+ DB 65,15,182,192 ; movzbl %r8b,%eax
+ DB 76,141,5,67,0,0,0 ; lea 0x43(%rip),%r8 # d14 <_sk_store_a8_avx+0xa5>
+ DB 73,99,4,128 ; movslq (%r8,%rax,4),%rax
+ DB 76,1,192 ; add %r8,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,67,121,20,68,57,6,12 ; vpextrb $0xc,%xmm8,0x6(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,5,10 ; vpextrb $0xa,%xmm8,0x5(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,4,8 ; vpextrb $0x8,%xmm8,0x4(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,3,6 ; vpextrb $0x6,%xmm8,0x3(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,2,4 ; vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
+ DB 196,67,121,20,68,57,1,2 ; vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
+ DB 196,67,121,20,4,57,0 ; vpextrb $0x0,%xmm8,(%r9,%rdi,1)
+ DB 235,154 ; jmp cad <_sk_store_a8_avx+0x3e>
+ DB 144 ; nop
+ DB 246,255 ; idiv %bh
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 238 ; out %al,(%dx)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,230 ; jmpq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 222,255 ; fdivrp %st,%st(7)
+ DB 255 ; (bad)
+ DB 255,214 ; callq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,206 ; dec %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,198 ; inc %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_565_avx
+_sk_load_565_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,16 ; mov (%rax),%r10
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,209,0,0,0 ; jne e0f <_sk_load_565_avx+0xdf>
+ DB 196,193,122,111,4,122 ; vmovdqu (%r10,%rdi,2),%xmm0
+ DB 197,241,239,201 ; vpxor %xmm1,%xmm1,%xmm1
+ DB 197,249,105,201 ; vpunpckhwd %xmm1,%xmm0,%xmm1
+ DB 196,226,121,51,192 ; vpmovzxwd %xmm0,%xmm0
+ DB 196,227,125,24,209,1 ; vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 197,249,110,192 ; vmovd %eax,%xmm0
+ DB 197,249,112,192,0 ; vpshufd $0x0,%xmm0,%xmm0
+ DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 197,252,84,194 ; vandps %ymm2,%ymm0,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,252,89,193 ; vmulps %ymm1,%ymm0,%ymm0
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 197,249,112,201,0 ; vpshufd $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,244,84,202 ; vandps %ymm2,%ymm1,%ymm1
+ DB 197,252,91,201 ; vcvtdq2ps %ymm1,%ymm1
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,244,89,203 ; vmulps %ymm3,%ymm1,%ymm1
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 197,249,112,219,0 ; vpshufd $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,228,84,210 ; vandps %ymm2,%ymm3,%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 15,135,29,255,255,255 ; ja d44 <_sk_load_565_avx+0x14>
+ DB 69,15,182,192 ; movzbl %r8b,%r8d
+ DB 76,141,13,74,0,0,0 ; lea 0x4a(%rip),%r9 # e7c <_sk_load_565_avx+0x14c>
+ DB 75,99,4,129 ; movslq (%r9,%r8,4),%rax
+ DB 76,1,200 ; add %r9,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,249,239,192 ; vpxor %xmm0,%xmm0,%xmm0
+ DB 196,193,121,196,68,122,12,6 ; vpinsrw $0x6,0xc(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,10,5 ; vpinsrw $0x5,0xa(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,8,4 ; vpinsrw $0x4,0x8(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,6,3 ; vpinsrw $0x3,0x6(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,4,2 ; vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,68,122,2,1 ; vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
+ DB 196,193,121,196,4,122,0 ; vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
+ DB 233,201,254,255,255 ; jmpq d44 <_sk_load_565_avx+0x14>
+ DB 144 ; nop
+ DB 243,255 ; repz (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 235,255 ; jmp e81 <_sk_load_565_avx+0x151>
+ DB 255 ; (bad)
+ DB 255,227 ; jmpq *%rbx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 219,255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,211 ; callq *%rbx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,203 ; dec %ebx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 191 ; .byte 0xbf
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_store_565_avx
+_sk_store_565_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,8 ; mov (%rax),%r9
+ DB 184,0,0,248,65 ; mov $0x41f80000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 196,65,125,91,201 ; vcvtps2dq %ymm9,%ymm9
+ DB 196,193,41,114,241,11 ; vpslld $0xb,%xmm9,%xmm10
+ DB 196,67,125,25,201,1 ; vextractf128 $0x1,%ymm9,%xmm9
+ DB 196,193,49,114,241,11 ; vpslld $0xb,%xmm9,%xmm9
+ DB 196,67,45,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
+ DB 184,0,0,124,66 ; mov $0x427c0000,%eax
+ DB 197,121,110,208 ; vmovd %eax,%xmm10
+ DB 196,67,121,4,210,0 ; vpermilps $0x0,%xmm10,%xmm10
+ DB 196,67,45,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ DB 197,44,89,209 ; vmulps %ymm1,%ymm10,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,33,114,242,5 ; vpslld $0x5,%xmm10,%xmm11
+ DB 196,67,125,25,210,1 ; vextractf128 $0x1,%ymm10,%xmm10
+ DB 196,193,41,114,242,5 ; vpslld $0x5,%xmm10,%xmm10
+ DB 196,67,37,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ DB 196,65,45,86,201 ; vorpd %ymm9,%ymm10,%ymm9
+ DB 197,60,89,194 ; vmulps %ymm2,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,65,53,86,192 ; vorpd %ymm8,%ymm9,%ymm8
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,66,57,43,193 ; vpackusdw %xmm9,%xmm8,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,10 ; jne f36 <_sk_store_565_avx+0x9e>
+ DB 196,65,122,127,4,121 ; vmovdqu %xmm8,(%r9,%rdi,2)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 119,236 ; ja f32 <_sk_store_565_avx+0x9a>
+ DB 65,15,182,192 ; movzbl %r8b,%eax
+ DB 76,141,5,67,0,0,0 ; lea 0x43(%rip),%r8 # f94 <_sk_store_565_avx+0xfc>
+ DB 73,99,4,128 ; movslq (%r8,%rax,4),%rax
+ DB 76,1,192 ; add %r8,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,67,121,21,68,121,12,6 ; vpextrw $0x6,%xmm8,0xc(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,10,5 ; vpextrw $0x5,%xmm8,0xa(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,8,4 ; vpextrw $0x4,%xmm8,0x8(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,6,3 ; vpextrw $0x3,%xmm8,0x6(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,4,2 ; vpextrw $0x2,%xmm8,0x4(%r9,%rdi,2)
+ DB 196,67,121,21,68,121,2,1 ; vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
+ DB 196,67,121,21,4,121,0 ; vpextrw $0x0,%xmm8,(%r9,%rdi,2)
+ DB 235,159 ; jmp f32 <_sk_store_565_avx+0x9a>
+ DB 144 ; nop
+ DB 246,255 ; idiv %bh
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 238 ; out %al,(%dx)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,230 ; jmpq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 222,255 ; fdivrp %st,%st(7)
+ DB 255 ; (bad)
+ DB 255,214 ; callq *%rsi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,206 ; dec %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,198 ; inc %esi
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_8888_avx
+_sk_load_8888_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,16 ; mov (%rax),%r10
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,157,0,0,0 ; jne 105b <_sk_load_8888_avx+0xab>
+ DB 196,65,124,16,12,186 ; vmovups (%r10,%rdi,4),%ymm9
+ DB 184,255,0,0,0 ; mov $0xff,%eax
+ DB 197,249,110,192 ; vmovd %eax,%xmm0
+ DB 197,249,112,192,0 ; vpshufd $0x0,%xmm0,%xmm0
+ DB 196,99,125,24,216,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
+ DB 196,193,36,84,193 ; vandps %ymm9,%ymm11,%ymm0
+ DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,99,117,24,193,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm8
+ DB 196,193,124,89,192 ; vmulps %ymm8,%ymm0,%ymm0
+ DB 196,193,41,114,209,8 ; vpsrld $0x8,%xmm9,%xmm10
+ DB 196,99,125,25,203,1 ; vextractf128 $0x1,%ymm9,%xmm3
+ DB 197,241,114,211,8 ; vpsrld $0x8,%xmm3,%xmm1
+ DB 196,227,45,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm10,%ymm1
+ DB 197,164,84,201 ; vandps %ymm1,%ymm11,%ymm1
+ DB 197,252,91,201 ; vcvtdq2ps %ymm1,%ymm1
+ DB 196,193,116,89,200 ; vmulps %ymm8,%ymm1,%ymm1
+ DB 196,193,41,114,209,16 ; vpsrld $0x10,%xmm9,%xmm10
+ DB 197,233,114,211,16 ; vpsrld $0x10,%xmm3,%xmm2
+ DB 196,227,45,24,210,1 ; vinsertf128 $0x1,%xmm2,%ymm10,%ymm2
+ DB 197,164,84,210 ; vandps %ymm2,%ymm11,%ymm2
+ DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
+ DB 196,193,108,89,208 ; vmulps %ymm8,%ymm2,%ymm2
+ DB 196,193,49,114,209,24 ; vpsrld $0x18,%xmm9,%xmm9
+ DB 197,225,114,211,24 ; vpsrld $0x18,%xmm3,%xmm3
+ DB 196,227,53,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm9,%ymm3
+ DB 197,252,91,219 ; vcvtdq2ps %ymm3,%ymm3
+ DB 196,193,100,89,216 ; vmulps %ymm8,%ymm3,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 196,65,52,87,201 ; vxorps %ymm9,%ymm9,%ymm9
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 15,135,80,255,255,255 ; ja fc4 <_sk_load_8888_avx+0x14>
+ DB 69,15,182,192 ; movzbl %r8b,%r8d
+ DB 76,141,13,137,0,0,0 ; lea 0x89(%rip),%r9 # 1108 <_sk_load_8888_avx+0x158>
+ DB 75,99,4,129 ; movslq (%r9,%r8,4),%rax
+ DB 76,1,200 ; add %r9,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,193,121,110,68,186,24 ; vmovd 0x18(%r10,%rdi,4),%xmm0
+ DB 197,249,112,192,68 ; vpshufd $0x44,%xmm0,%xmm0
+ DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 196,99,117,12,200,64 ; vblendps $0x40,%ymm0,%ymm1,%ymm9
+ DB 196,99,125,25,200,1 ; vextractf128 $0x1,%ymm9,%xmm0
+ DB 196,195,121,34,68,186,20,1 ; vpinsrd $0x1,0x14(%r10,%rdi,4),%xmm0,%xmm0
+ DB 196,99,53,24,200,1 ; vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
+ DB 196,99,125,25,200,1 ; vextractf128 $0x1,%ymm9,%xmm0
+ DB 196,195,121,34,68,186,16,0 ; vpinsrd $0x0,0x10(%r10,%rdi,4),%xmm0,%xmm0
+ DB 196,99,53,24,200,1 ; vinsertf128 $0x1,%xmm0,%ymm9,%ymm9
+ DB 196,195,49,34,68,186,12,3 ; vpinsrd $0x3,0xc(%r10,%rdi,4),%xmm9,%xmm0
+ DB 196,99,53,12,200,15 ; vblendps $0xf,%ymm0,%ymm9,%ymm9
+ DB 196,195,49,34,68,186,8,2 ; vpinsrd $0x2,0x8(%r10,%rdi,4),%xmm9,%xmm0
+ DB 196,99,53,12,200,15 ; vblendps $0xf,%ymm0,%ymm9,%ymm9
+ DB 196,195,49,34,68,186,4,1 ; vpinsrd $0x1,0x4(%r10,%rdi,4),%xmm9,%xmm0
+ DB 196,99,53,12,200,15 ; vblendps $0xf,%ymm0,%ymm9,%ymm9
+ DB 196,195,49,34,4,186,0 ; vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
+ DB 196,99,53,12,200,15 ; vblendps $0xf,%ymm0,%ymm9,%ymm9
+ DB 233,188,254,255,255 ; jmpq fc4 <_sk_load_8888_avx+0x14>
+ DB 238 ; out %al,(%dx)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,224 ; jmpq *%rax
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,210 ; callq *%rdx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,196 ; inc %esp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,176,255,255,255,156 ; pushq -0x63000001(%rax)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+ DB 128,255,255 ; cmp $0xff,%bh
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_store_8888_avx
+_sk_store_8888_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,8 ; mov (%rax),%r9
+ DB 184,0,0,127,67 ; mov $0x437f0000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 196,65,125,91,201 ; vcvtps2dq %ymm9,%ymm9
+ DB 197,60,89,209 ; vmulps %ymm1,%ymm8,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,33,114,242,8 ; vpslld $0x8,%xmm10,%xmm11
+ DB 196,67,125,25,210,1 ; vextractf128 $0x1,%ymm10,%xmm10
+ DB 196,193,41,114,242,8 ; vpslld $0x8,%xmm10,%xmm10
+ DB 196,67,37,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ DB 196,65,45,86,201 ; vorpd %ymm9,%ymm10,%ymm9
+ DB 197,60,89,210 ; vmulps %ymm2,%ymm8,%ymm10
+ DB 196,65,125,91,210 ; vcvtps2dq %ymm10,%ymm10
+ DB 196,193,33,114,242,16 ; vpslld $0x10,%xmm10,%xmm11
+ DB 196,67,125,25,210,1 ; vextractf128 $0x1,%ymm10,%xmm10
+ DB 196,193,41,114,242,16 ; vpslld $0x10,%xmm10,%xmm10
+ DB 196,67,37,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
+ DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
+ DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
+ DB 196,193,33,114,240,24 ; vpslld $0x18,%xmm8,%xmm11
+ DB 196,67,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm8
+ DB 196,193,57,114,240,24 ; vpslld $0x18,%xmm8,%xmm8
+ DB 196,67,37,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
+ DB 196,65,45,86,192 ; vorpd %ymm8,%ymm10,%ymm8
+ DB 196,65,53,86,192 ; vorpd %ymm8,%ymm9,%ymm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,10 ; jne 11c8 <_sk_store_8888_avx+0xa4>
+ DB 196,65,124,17,4,185 ; vmovups %ymm8,(%r9,%rdi,4)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 65,137,200 ; mov %ecx,%r8d
+ DB 65,128,224,7 ; and $0x7,%r8b
+ DB 65,254,200 ; dec %r8b
+ DB 65,128,248,6 ; cmp $0x6,%r8b
+ DB 119,236 ; ja 11c4 <_sk_store_8888_avx+0xa0>
+ DB 65,15,182,192 ; movzbl %r8b,%eax
+ DB 76,141,5,85,0,0,0 ; lea 0x55(%rip),%r8 # 1238 <_sk_store_8888_avx+0x114>
+ DB 73,99,4,128 ; movslq (%r8,%rax,4),%rax
+ DB 76,1,192 ; add %r8,%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,67,121,22,76,185,24,2 ; vpextrd $0x2,%xmm9,0x18(%r9,%rdi,4)
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,67,121,22,76,185,20,1 ; vpextrd $0x1,%xmm9,0x14(%r9,%rdi,4)
+ DB 196,67,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm9
+ DB 196,65,122,17,76,185,16 ; vmovss %xmm9,0x10(%r9,%rdi,4)
+ DB 196,67,121,22,68,185,12,3 ; vpextrd $0x3,%xmm8,0xc(%r9,%rdi,4)
+ DB 196,67,121,22,68,185,8,2 ; vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
+ DB 196,67,121,22,68,185,4,1 ; vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
+ DB 196,65,121,126,4,185 ; vmovd %xmm8,(%r9,%rdi,4)
+ DB 235,143 ; jmp 11c4 <_sk_store_8888_avx+0xa0>
+ DB 15,31,0 ; nopl (%rax)
+ DB 245 ; cmc
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 237 ; in (%dx),%eax
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,229 ; jmpq *%rbp
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 221,255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,208 ; callq *%rax
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255,194 ; inc %edx
+ DB 255 ; (bad)
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+ DB 180,255 ; mov $0xff,%ah
+ DB 255 ; (bad)
+ DB 255 ; .byte 0xff
+
+PUBLIC _sk_load_f16_avx
+_sk_load_f16_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 15,133,2,1,0,0 ; jne 1364 <_sk_load_f16_avx+0x110>
+ DB 197,121,16,4,248 ; vmovupd (%rax,%rdi,8),%xmm8
+ DB 197,249,16,84,248,16 ; vmovupd 0x10(%rax,%rdi,8),%xmm2
+ DB 197,249,16,92,248,32 ; vmovupd 0x20(%rax,%rdi,8),%xmm3
+ DB 197,122,111,76,248,48 ; vmovdqu 0x30(%rax,%rdi,8),%xmm9
+ DB 197,185,97,194 ; vpunpcklwd %xmm2,%xmm8,%xmm0
+ DB 197,185,105,210 ; vpunpckhwd %xmm2,%xmm8,%xmm2
+ DB 196,193,97,97,201 ; vpunpcklwd %xmm9,%xmm3,%xmm1
+ DB 196,193,97,105,217 ; vpunpckhwd %xmm9,%xmm3,%xmm3
+ DB 197,121,97,194 ; vpunpcklwd %xmm2,%xmm0,%xmm8
+ DB 197,249,105,194 ; vpunpckhwd %xmm2,%xmm0,%xmm0
+ DB 197,241,97,211 ; vpunpcklwd %xmm3,%xmm1,%xmm2
+ DB 197,113,105,203 ; vpunpckhwd %xmm3,%xmm1,%xmm9
+ DB 184,0,4,0,4 ; mov $0x4000400,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 197,249,112,219,0 ; vpshufd $0x0,%xmm3,%xmm3
+ DB 196,193,97,101,200 ; vpcmpgtw %xmm8,%xmm3,%xmm1
+ DB 196,65,113,223,192 ; vpandn %xmm8,%xmm1,%xmm8
+ DB 197,225,101,200 ; vpcmpgtw %xmm0,%xmm3,%xmm1
+ DB 197,241,223,192 ; vpandn %xmm0,%xmm1,%xmm0
+ DB 197,225,101,202 ; vpcmpgtw %xmm2,%xmm3,%xmm1
+ DB 197,241,223,202 ; vpandn %xmm2,%xmm1,%xmm1
+ DB 196,193,97,101,209 ; vpcmpgtw %xmm9,%xmm3,%xmm2
+ DB 196,193,105,223,209 ; vpandn %xmm9,%xmm2,%xmm2
+ DB 196,66,121,51,208 ; vpmovzxwd %xmm8,%xmm10
+ DB 196,98,121,51,201 ; vpmovzxwd %xmm1,%xmm9
+ DB 197,225,239,219 ; vpxor %xmm3,%xmm3,%xmm3
+ DB 197,57,105,195 ; vpunpckhwd %xmm3,%xmm8,%xmm8
+ DB 197,241,105,203 ; vpunpckhwd %xmm3,%xmm1,%xmm1
+ DB 196,98,121,51,216 ; vpmovzxwd %xmm0,%xmm11
+ DB 196,98,121,51,226 ; vpmovzxwd %xmm2,%xmm12
+ DB 197,121,105,235 ; vpunpckhwd %xmm3,%xmm0,%xmm13
+ DB 197,105,105,243 ; vpunpckhwd %xmm3,%xmm2,%xmm14
+ DB 196,193,121,114,242,13 ; vpslld $0xd,%xmm10,%xmm0
+ DB 196,193,105,114,241,13 ; vpslld $0xd,%xmm9,%xmm2
+ DB 196,227,125,24,194,1 ; vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
+ DB 184,0,0,128,119 ; mov $0x77800000,%eax
+ DB 197,249,110,208 ; vmovd %eax,%xmm2
+ DB 197,249,112,210,0 ; vpshufd $0x0,%xmm2,%xmm2
+ DB 196,99,109,24,202,1 ; vinsertf128 $0x1,%xmm2,%ymm2,%ymm9
+ DB 197,180,89,192 ; vmulps %ymm0,%ymm9,%ymm0
+ DB 196,193,105,114,240,13 ; vpslld $0xd,%xmm8,%xmm2
+ DB 197,241,114,241,13 ; vpslld $0xd,%xmm1,%xmm1
+ DB 196,227,109,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
+ DB 197,180,89,201 ; vmulps %ymm1,%ymm9,%ymm1
+ DB 196,193,105,114,243,13 ; vpslld $0xd,%xmm11,%xmm2
+ DB 196,193,97,114,244,13 ; vpslld $0xd,%xmm12,%xmm3
+ DB 196,227,109,24,211,1 ; vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ DB 197,180,89,210 ; vmulps %ymm2,%ymm9,%ymm2
+ DB 196,193,57,114,245,13 ; vpslld $0xd,%xmm13,%xmm8
+ DB 196,193,97,114,246,13 ; vpslld $0xd,%xmm14,%xmm3
+ DB 196,227,61,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ DB 197,180,89,219 ; vmulps %ymm3,%ymm9,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 197,123,16,4,248 ; vmovsd (%rax,%rdi,8),%xmm8
+ DB 196,65,49,239,201 ; vpxor %xmm9,%xmm9,%xmm9
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,79 ; je 13c3 <_sk_load_f16_avx+0x16f>
+ DB 197,57,22,68,248,8 ; vmovhpd 0x8(%rax,%rdi,8),%xmm8,%xmm8
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,67 ; jb 13c3 <_sk_load_f16_avx+0x16f>
+ DB 197,251,16,84,248,16 ; vmovsd 0x10(%rax,%rdi,8),%xmm2
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 116,68 ; je 13d0 <_sk_load_f16_avx+0x17c>
+ DB 197,233,22,84,248,24 ; vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,56 ; jb 13d0 <_sk_load_f16_avx+0x17c>
+ DB 197,251,16,92,248,32 ; vmovsd 0x20(%rax,%rdi,8),%xmm3
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 15,132,209,254,255,255 ; je 1279 <_sk_load_f16_avx+0x25>
+ DB 197,225,22,92,248,40 ; vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 15,130,193,254,255,255 ; jb 1279 <_sk_load_f16_avx+0x25>
+ DB 197,122,126,76,248,48 ; vmovq 0x30(%rax,%rdi,8),%xmm9
+ DB 233,182,254,255,255 ; jmpq 1279 <_sk_load_f16_avx+0x25>
+ DB 197,225,87,219 ; vxorpd %xmm3,%xmm3,%xmm3
+ DB 197,233,87,210 ; vxorpd %xmm2,%xmm2,%xmm2
+ DB 233,169,254,255,255 ; jmpq 1279 <_sk_load_f16_avx+0x25>
+ DB 197,225,87,219 ; vxorpd %xmm3,%xmm3,%xmm3
+ DB 233,160,254,255,255 ; jmpq 1279 <_sk_load_f16_avx+0x25>
+
+PUBLIC _sk_store_f16_avx
+_sk_store_f16_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,0 ; mov (%rax),%r8
+ DB 184,0,0,128,7 ; mov $0x7800000,%eax
+ DB 197,121,110,192 ; vmovd %eax,%xmm8
+ DB 196,65,121,112,192,0 ; vpshufd $0x0,%xmm8,%xmm8
+ DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
+ DB 196,67,125,25,202,1 ; vextractf128 $0x1,%ymm9,%xmm10
+ DB 196,193,41,114,210,13 ; vpsrld $0xd,%xmm10,%xmm10
+ DB 196,193,49,114,209,13 ; vpsrld $0xd,%xmm9,%xmm9
+ DB 197,60,89,217 ; vmulps %ymm1,%ymm8,%ymm11
+ DB 196,67,125,25,220,1 ; vextractf128 $0x1,%ymm11,%xmm12
+ DB 196,193,25,114,212,13 ; vpsrld $0xd,%xmm12,%xmm12
+ DB 196,193,33,114,211,13 ; vpsrld $0xd,%xmm11,%xmm11
+ DB 197,60,89,234 ; vmulps %ymm2,%ymm8,%ymm13
+ DB 196,67,125,25,238,1 ; vextractf128 $0x1,%ymm13,%xmm14
+ DB 196,193,9,114,214,13 ; vpsrld $0xd,%xmm14,%xmm14
+ DB 196,193,17,114,213,13 ; vpsrld $0xd,%xmm13,%xmm13
+ DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
+ DB 196,67,125,25,199,1 ; vextractf128 $0x1,%ymm8,%xmm15
+ DB 196,193,1,114,215,13 ; vpsrld $0xd,%xmm15,%xmm15
+ DB 196,193,57,114,208,13 ; vpsrld $0xd,%xmm8,%xmm8
+ DB 196,193,33,115,251,2 ; vpslldq $0x2,%xmm11,%xmm11
+ DB 196,65,33,235,201 ; vpor %xmm9,%xmm11,%xmm9
+ DB 196,193,33,115,252,2 ; vpslldq $0x2,%xmm12,%xmm11
+ DB 196,65,33,235,226 ; vpor %xmm10,%xmm11,%xmm12
+ DB 196,193,57,115,248,2 ; vpslldq $0x2,%xmm8,%xmm8
+ DB 196,65,57,235,197 ; vpor %xmm13,%xmm8,%xmm8
+ DB 196,193,41,115,255,2 ; vpslldq $0x2,%xmm15,%xmm10
+ DB 196,65,41,235,238 ; vpor %xmm14,%xmm10,%xmm13
+ DB 196,65,49,98,216 ; vpunpckldq %xmm8,%xmm9,%xmm11
+ DB 196,65,49,106,208 ; vpunpckhdq %xmm8,%xmm9,%xmm10
+ DB 196,65,25,98,205 ; vpunpckldq %xmm13,%xmm12,%xmm9
+ DB 196,65,25,106,197 ; vpunpckhdq %xmm13,%xmm12,%xmm8
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,31 ; jne 14af <_sk_store_f16_avx+0xd6>
+ DB 196,65,120,17,28,248 ; vmovups %xmm11,(%r8,%rdi,8)
+ DB 196,65,120,17,84,248,16 ; vmovups %xmm10,0x10(%r8,%rdi,8)
+ DB 196,65,120,17,76,248,32 ; vmovups %xmm9,0x20(%r8,%rdi,8)
+ DB 196,65,122,127,68,248,48 ; vmovdqu %xmm8,0x30(%r8,%rdi,8)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,65,121,214,28,248 ; vmovq %xmm11,(%r8,%rdi,8)
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,240 ; je 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,23,92,248,8 ; vmovhpd %xmm11,0x8(%r8,%rdi,8)
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,227 ; jb 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,214,84,248,16 ; vmovq %xmm10,0x10(%r8,%rdi,8)
+ DB 116,218 ; je 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,23,84,248,24 ; vmovhpd %xmm10,0x18(%r8,%rdi,8)
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,205 ; jb 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,214,76,248,32 ; vmovq %xmm9,0x20(%r8,%rdi,8)
+ DB 116,196 ; je 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,23,76,248,40 ; vmovhpd %xmm9,0x28(%r8,%rdi,8)
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 114,183 ; jb 14ab <_sk_store_f16_avx+0xd2>
+ DB 196,65,121,214,68,248,48 ; vmovq %xmm8,0x30(%r8,%rdi,8)
+ DB 235,174 ; jmp 14ab <_sk_store_f16_avx+0xd2>
+
+PUBLIC _sk_store_f32_avx
+_sk_store_f32_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 76,139,0 ; mov (%rax),%r8
+ DB 72,141,4,189,0,0,0,0 ; lea 0x0(,%rdi,4),%rax
+ DB 197,124,20,193 ; vunpcklps %ymm1,%ymm0,%ymm8
+ DB 197,124,21,217 ; vunpckhps %ymm1,%ymm0,%ymm11
+ DB 197,108,20,203 ; vunpcklps %ymm3,%ymm2,%ymm9
+ DB 197,108,21,227 ; vunpckhps %ymm3,%ymm2,%ymm12
+ DB 196,65,61,20,209 ; vunpcklpd %ymm9,%ymm8,%ymm10
+ DB 196,65,61,21,201 ; vunpckhpd %ymm9,%ymm8,%ymm9
+ DB 196,65,37,20,196 ; vunpcklpd %ymm12,%ymm11,%ymm8
+ DB 196,65,37,21,220 ; vunpckhpd %ymm12,%ymm11,%ymm11
+ DB 72,133,201 ; test %rcx,%rcx
+ DB 117,55 ; jne 156a <_sk_store_f32_avx+0x6d>
+ DB 196,67,45,24,225,1 ; vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
+ DB 196,67,61,24,235,1 ; vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
+ DB 196,67,45,6,201,49 ; vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
+ DB 196,67,61,6,195,49 ; vperm2f128 $0x31,%ymm11,%ymm8,%ymm8
+ DB 196,65,125,17,36,128 ; vmovupd %ymm12,(%r8,%rax,4)
+ DB 196,65,125,17,108,128,32 ; vmovupd %ymm13,0x20(%r8,%rax,4)
+ DB 196,65,125,17,76,128,64 ; vmovupd %ymm9,0x40(%r8,%rax,4)
+ DB 196,65,125,17,68,128,96 ; vmovupd %ymm8,0x60(%r8,%rax,4)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+ DB 196,65,121,17,20,128 ; vmovupd %xmm10,(%r8,%rax,4)
+ DB 72,131,249,1 ; cmp $0x1,%rcx
+ DB 116,240 ; je 1566 <_sk_store_f32_avx+0x69>
+ DB 196,65,121,17,76,128,16 ; vmovupd %xmm9,0x10(%r8,%rax,4)
+ DB 72,131,249,3 ; cmp $0x3,%rcx
+ DB 114,227 ; jb 1566 <_sk_store_f32_avx+0x69>
+ DB 196,65,121,17,68,128,32 ; vmovupd %xmm8,0x20(%r8,%rax,4)
+ DB 116,218 ; je 1566 <_sk_store_f32_avx+0x69>
+ DB 196,65,121,17,92,128,48 ; vmovupd %xmm11,0x30(%r8,%rax,4)
+ DB 72,131,249,5 ; cmp $0x5,%rcx
+ DB 114,205 ; jb 1566 <_sk_store_f32_avx+0x69>
+ DB 196,67,125,25,84,128,64,1 ; vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
+ DB 116,195 ; je 1566 <_sk_store_f32_avx+0x69>
+ DB 196,67,125,25,76,128,80,1 ; vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
+ DB 72,131,249,7 ; cmp $0x7,%rcx
+ DB 114,181 ; jb 1566 <_sk_store_f32_avx+0x69>
+ DB 196,67,125,25,68,128,96,1 ; vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
+ DB 235,171 ; jmp 1566 <_sk_store_f32_avx+0x69>
+
+PUBLIC _sk_clamp_x_avx
+_sk_clamp_x_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,95,200 ; vmaxps %ymm0,%ymm8,%ymm9
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,99,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm0
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,121,254,194 ; vpaddd %xmm10,%xmm0,%xmm0
+ DB 196,65,57,254,194 ; vpaddd %xmm10,%xmm8,%xmm8
+ DB 196,227,61,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
+ DB 197,180,93,192 ; vminps %ymm0,%ymm9,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_y_avx
+_sk_clamp_y_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,95,201 ; vmaxps %ymm1,%ymm8,%ymm9
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,99,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm1
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,113,254,202 ; vpaddd %xmm10,%xmm1,%xmm1
+ DB 196,65,57,254,194 ; vpaddd %xmm10,%xmm8,%xmm8
+ DB 196,227,61,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
+ DB 197,180,93,201 ; vminps %ymm1,%ymm9,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_x_avx
+_sk_repeat_x_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,65,124,94,200 ; vdivps %ymm8,%ymm0,%ymm9
+ DB 196,67,125,8,201,1 ; vroundps $0x1,%ymm9,%ymm9
+ DB 196,65,52,89,200 ; vmulps %ymm8,%ymm9,%ymm9
+ DB 196,65,124,92,201 ; vsubps %ymm9,%ymm0,%ymm9
+ DB 196,99,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm0
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,121,254,194 ; vpaddd %xmm10,%xmm0,%xmm0
+ DB 196,65,57,254,194 ; vpaddd %xmm10,%xmm8,%xmm8
+ DB 196,227,61,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm8,%ymm0
+ DB 197,180,93,192 ; vminps %ymm0,%ymm9,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_y_avx
+_sk_repeat_y_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,65,116,94,200 ; vdivps %ymm8,%ymm1,%ymm9
+ DB 196,67,125,8,201,1 ; vroundps $0x1,%ymm9,%ymm9
+ DB 196,65,52,89,200 ; vmulps %ymm8,%ymm9,%ymm9
+ DB 196,65,116,92,201 ; vsubps %ymm9,%ymm1,%ymm9
+ DB 196,99,125,25,193,1 ; vextractf128 $0x1,%ymm8,%xmm1
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,113,254,202 ; vpaddd %xmm10,%xmm1,%xmm1
+ DB 196,65,57,254,194 ; vpaddd %xmm10,%xmm8,%xmm8
+ DB 196,227,61,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm8,%ymm1
+ DB 197,180,93,201 ; vminps %ymm1,%ymm9,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_x_avx
+_sk_mirror_x_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,121,110,0 ; vmovd (%rax),%xmm8
+ DB 196,65,121,112,200,0 ; vpshufd $0x0,%xmm8,%xmm9
+ DB 196,67,53,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ DB 196,65,124,92,209 ; vsubps %ymm9,%ymm0,%ymm10
+ DB 196,193,58,88,192 ; vaddss %xmm8,%xmm8,%xmm0
+ DB 196,227,121,4,192,0 ; vpermilps $0x0,%xmm0,%xmm0
+ DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 197,44,94,192 ; vdivps %ymm0,%ymm10,%ymm8
+ DB 196,67,125,8,192,1 ; vroundps $0x1,%ymm8,%ymm8
+ DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
+ DB 197,172,92,192 ; vsubps %ymm0,%ymm10,%ymm0
+ DB 196,193,124,92,193 ; vsubps %ymm9,%ymm0,%ymm0
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,92,192 ; vsubps %ymm0,%ymm8,%ymm8
+ DB 197,60,84,192 ; vandps %ymm0,%ymm8,%ymm8
+ DB 196,99,125,25,200,1 ; vextractf128 $0x1,%ymm9,%xmm0
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,121,254,194 ; vpaddd %xmm10,%xmm0,%xmm0
+ DB 196,65,49,254,202 ; vpaddd %xmm10,%xmm9,%xmm9
+ DB 196,227,53,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm9,%ymm0
+ DB 197,188,93,192 ; vminps %ymm0,%ymm8,%ymm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_y_avx
+_sk_mirror_y_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,121,110,0 ; vmovd (%rax),%xmm8
+ DB 196,65,121,112,200,0 ; vpshufd $0x0,%xmm8,%xmm9
+ DB 196,67,53,24,201,1 ; vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ DB 196,65,116,92,209 ; vsubps %ymm9,%ymm1,%ymm10
+ DB 196,193,58,88,200 ; vaddss %xmm8,%xmm8,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,44,94,193 ; vdivps %ymm1,%ymm10,%ymm8
+ DB 196,67,125,8,192,1 ; vroundps $0x1,%ymm8,%ymm8
+ DB 197,188,89,201 ; vmulps %ymm1,%ymm8,%ymm1
+ DB 197,172,92,201 ; vsubps %ymm1,%ymm10,%ymm1
+ DB 196,193,116,92,201 ; vsubps %ymm9,%ymm1,%ymm1
+ DB 196,65,60,87,192 ; vxorps %ymm8,%ymm8,%ymm8
+ DB 197,60,92,193 ; vsubps %ymm1,%ymm8,%ymm8
+ DB 197,60,84,193 ; vandps %ymm1,%ymm8,%ymm8
+ DB 196,99,125,25,201,1 ; vextractf128 $0x1,%ymm9,%xmm1
+ DB 196,65,41,118,210 ; vpcmpeqd %xmm10,%xmm10,%xmm10
+ DB 196,193,113,254,202 ; vpaddd %xmm10,%xmm1,%xmm1
+ DB 196,65,49,254,202 ; vpaddd %xmm10,%xmm9,%xmm9
+ DB 196,227,53,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm9,%ymm1
+ DB 197,188,93,201 ; vminps %ymm1,%ymm8,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_luminance_to_alpha_avx
+_sk_luminance_to_alpha_avx LABEL PROC
+ DB 184,208,179,89,62 ; mov $0x3e59b3d0,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,228,89,192 ; vmulps %ymm0,%ymm3,%ymm0
+ DB 184,89,23,55,63 ; mov $0x3f371759,%eax
+ DB 197,249,110,216 ; vmovd %eax,%xmm3
+ DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
+ DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 197,228,89,201 ; vmulps %ymm1,%ymm3,%ymm1
+ DB 197,252,88,193 ; vaddps %ymm1,%ymm0,%ymm0
+ DB 184,152,221,147,61 ; mov $0x3d93dd98,%eax
+ DB 197,249,110,200 ; vmovd %eax,%xmm1
+ DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
+ DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 197,244,89,202 ; vmulps %ymm2,%ymm1,%ymm1
+ DB 197,252,88,217 ; vaddps %ymm1,%ymm0,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
+ DB 197,244,87,201 ; vxorps %ymm1,%ymm1,%ymm1
+ DB 197,236,87,210 ; vxorps %ymm2,%ymm2,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_2x3_avx
+_sk_matrix_2x3_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,24,72,8 ; vbroadcastss 0x8(%rax),%ymm9
+ DB 196,98,125,24,80,16 ; vbroadcastss 0x10(%rax),%ymm10
+ DB 197,52,89,201 ; vmulps %ymm1,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 197,60,89,192 ; vmulps %ymm0,%ymm8,%ymm8
+ DB 196,65,60,88,193 ; vaddps %ymm9,%ymm8,%ymm8
+ DB 196,98,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm9
+ DB 196,98,125,24,80,12 ; vbroadcastss 0xc(%rax),%ymm10
+ DB 196,98,125,24,88,20 ; vbroadcastss 0x14(%rax),%ymm11
+ DB 197,172,89,201 ; vmulps %ymm1,%ymm10,%ymm1
+ DB 196,193,116,88,203 ; vaddps %ymm11,%ymm1,%ymm1
+ DB 197,180,89,192 ; vmulps %ymm0,%ymm9,%ymm0
+ DB 197,252,88,201 ; vaddps %ymm1,%ymm0,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_3x4_avx
+_sk_matrix_3x4_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,24,72,12 ; vbroadcastss 0xc(%rax),%ymm9
+ DB 196,98,125,24,80,24 ; vbroadcastss 0x18(%rax),%ymm10
+ DB 196,98,125,24,88,36 ; vbroadcastss 0x24(%rax),%ymm11
+ DB 197,44,89,210 ; vmulps %ymm2,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 197,52,89,201 ; vmulps %ymm1,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 197,60,89,192 ; vmulps %ymm0,%ymm8,%ymm8
+ DB 196,65,60,88,193 ; vaddps %ymm9,%ymm8,%ymm8
+ DB 196,98,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm9
+ DB 196,98,125,24,80,16 ; vbroadcastss 0x10(%rax),%ymm10
+ DB 196,98,125,24,88,28 ; vbroadcastss 0x1c(%rax),%ymm11
+ DB 196,98,125,24,96,40 ; vbroadcastss 0x28(%rax),%ymm12
+ DB 197,36,89,218 ; vmulps %ymm2,%ymm11,%ymm11
+ DB 196,65,36,88,220 ; vaddps %ymm12,%ymm11,%ymm11
+ DB 197,44,89,209 ; vmulps %ymm1,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 197,52,89,200 ; vmulps %ymm0,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 196,98,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm10
+ DB 196,98,125,24,88,20 ; vbroadcastss 0x14(%rax),%ymm11
+ DB 196,98,125,24,96,32 ; vbroadcastss 0x20(%rax),%ymm12
+ DB 196,98,125,24,104,44 ; vbroadcastss 0x2c(%rax),%ymm13
+ DB 197,156,89,210 ; vmulps %ymm2,%ymm12,%ymm2
+ DB 196,193,108,88,213 ; vaddps %ymm13,%ymm2,%ymm2
+ DB 197,164,89,201 ; vmulps %ymm1,%ymm11,%ymm1
+ DB 197,244,88,202 ; vaddps %ymm2,%ymm1,%ymm1
+ DB 197,172,89,192 ; vmulps %ymm0,%ymm10,%ymm0
+ DB 197,252,88,209 ; vaddps %ymm1,%ymm0,%ymm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 197,124,41,201 ; vmovaps %ymm9,%ymm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_4x5_avx
+_sk_matrix_4x5_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,24,72,16 ; vbroadcastss 0x10(%rax),%ymm9
+ DB 196,98,125,24,80,32 ; vbroadcastss 0x20(%rax),%ymm10
+ DB 196,98,125,24,88,48 ; vbroadcastss 0x30(%rax),%ymm11
+ DB 196,98,125,24,96,64 ; vbroadcastss 0x40(%rax),%ymm12
+ DB 197,36,89,219 ; vmulps %ymm3,%ymm11,%ymm11
+ DB 196,65,36,88,220 ; vaddps %ymm12,%ymm11,%ymm11
+ DB 197,44,89,210 ; vmulps %ymm2,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 197,52,89,201 ; vmulps %ymm1,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 197,60,89,192 ; vmulps %ymm0,%ymm8,%ymm8
+ DB 196,65,60,88,193 ; vaddps %ymm9,%ymm8,%ymm8
+ DB 196,98,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm9
+ DB 196,98,125,24,80,20 ; vbroadcastss 0x14(%rax),%ymm10
+ DB 196,98,125,24,88,36 ; vbroadcastss 0x24(%rax),%ymm11
+ DB 196,98,125,24,96,52 ; vbroadcastss 0x34(%rax),%ymm12
+ DB 196,98,125,24,104,68 ; vbroadcastss 0x44(%rax),%ymm13
+ DB 197,28,89,227 ; vmulps %ymm3,%ymm12,%ymm12
+ DB 196,65,28,88,229 ; vaddps %ymm13,%ymm12,%ymm12
+ DB 197,36,89,218 ; vmulps %ymm2,%ymm11,%ymm11
+ DB 196,65,36,88,220 ; vaddps %ymm12,%ymm11,%ymm11
+ DB 197,44,89,209 ; vmulps %ymm1,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 197,52,89,200 ; vmulps %ymm0,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 196,98,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm10
+ DB 196,98,125,24,88,24 ; vbroadcastss 0x18(%rax),%ymm11
+ DB 196,98,125,24,96,40 ; vbroadcastss 0x28(%rax),%ymm12
+ DB 196,98,125,24,104,56 ; vbroadcastss 0x38(%rax),%ymm13
+ DB 196,98,125,24,112,72 ; vbroadcastss 0x48(%rax),%ymm14
+ DB 197,20,89,235 ; vmulps %ymm3,%ymm13,%ymm13
+ DB 196,65,20,88,238 ; vaddps %ymm14,%ymm13,%ymm13
+ DB 197,28,89,226 ; vmulps %ymm2,%ymm12,%ymm12
+ DB 196,65,28,88,229 ; vaddps %ymm13,%ymm12,%ymm12
+ DB 197,36,89,217 ; vmulps %ymm1,%ymm11,%ymm11
+ DB 196,65,36,88,220 ; vaddps %ymm12,%ymm11,%ymm11
+ DB 197,44,89,208 ; vmulps %ymm0,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 196,98,125,24,88,12 ; vbroadcastss 0xc(%rax),%ymm11
+ DB 196,98,125,24,96,28 ; vbroadcastss 0x1c(%rax),%ymm12
+ DB 196,98,125,24,104,44 ; vbroadcastss 0x2c(%rax),%ymm13
+ DB 196,98,125,24,112,60 ; vbroadcastss 0x3c(%rax),%ymm14
+ DB 196,98,125,24,120,76 ; vbroadcastss 0x4c(%rax),%ymm15
+ DB 197,140,89,219 ; vmulps %ymm3,%ymm14,%ymm3
+ DB 196,193,100,88,223 ; vaddps %ymm15,%ymm3,%ymm3
+ DB 197,148,89,210 ; vmulps %ymm2,%ymm13,%ymm2
+ DB 197,236,88,211 ; vaddps %ymm3,%ymm2,%ymm2
+ DB 197,156,89,201 ; vmulps %ymm1,%ymm12,%ymm1
+ DB 197,244,88,202 ; vaddps %ymm2,%ymm1,%ymm1
+ DB 197,164,89,192 ; vmulps %ymm0,%ymm11,%ymm0
+ DB 197,252,88,217 ; vaddps %ymm1,%ymm0,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 197,124,41,201 ; vmovaps %ymm9,%ymm1
+ DB 197,124,41,210 ; vmovaps %ymm10,%ymm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_perspective_avx
+_sk_matrix_perspective_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,98,125,24,0 ; vbroadcastss (%rax),%ymm8
+ DB 196,98,125,24,72,4 ; vbroadcastss 0x4(%rax),%ymm9
+ DB 196,98,125,24,80,8 ; vbroadcastss 0x8(%rax),%ymm10
+ DB 197,52,89,201 ; vmulps %ymm1,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 197,60,89,192 ; vmulps %ymm0,%ymm8,%ymm8
+ DB 196,65,60,88,193 ; vaddps %ymm9,%ymm8,%ymm8
+ DB 196,98,125,24,72,12 ; vbroadcastss 0xc(%rax),%ymm9
+ DB 196,98,125,24,80,16 ; vbroadcastss 0x10(%rax),%ymm10
+ DB 196,98,125,24,88,20 ; vbroadcastss 0x14(%rax),%ymm11
+ DB 197,44,89,209 ; vmulps %ymm1,%ymm10,%ymm10
+ DB 196,65,44,88,211 ; vaddps %ymm11,%ymm10,%ymm10
+ DB 197,52,89,200 ; vmulps %ymm0,%ymm9,%ymm9
+ DB 196,65,52,88,202 ; vaddps %ymm10,%ymm9,%ymm9
+ DB 196,98,125,24,80,24 ; vbroadcastss 0x18(%rax),%ymm10
+ DB 196,98,125,24,88,28 ; vbroadcastss 0x1c(%rax),%ymm11
+ DB 196,98,125,24,96,32 ; vbroadcastss 0x20(%rax),%ymm12
+ DB 197,164,89,201 ; vmulps %ymm1,%ymm11,%ymm1
+ DB 196,193,116,88,204 ; vaddps %ymm12,%ymm1,%ymm1
+ DB 197,172,89,192 ; vmulps %ymm0,%ymm10,%ymm0
+ DB 197,252,88,193 ; vaddps %ymm1,%ymm0,%ymm0
+ DB 197,252,83,200 ; vrcpps %ymm0,%ymm1
+ DB 197,188,89,193 ; vmulps %ymm1,%ymm8,%ymm0
+ DB 197,180,89,201 ; vmulps %ymm1,%ymm9,%ymm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_linear_gradient_2stops_avx
+_sk_linear_gradient_2stops_avx LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 196,226,125,24,72,16 ; vbroadcastss 0x10(%rax),%ymm1
+ DB 196,226,125,24,16 ; vbroadcastss (%rax),%ymm2
+ DB 197,244,89,200 ; vmulps %ymm0,%ymm1,%ymm1
+ DB 197,108,88,193 ; vaddps %ymm1,%ymm2,%ymm8
+ DB 196,226,125,24,72,20 ; vbroadcastss 0x14(%rax),%ymm1
+ DB 196,226,125,24,80,4 ; vbroadcastss 0x4(%rax),%ymm2
+ DB 197,244,89,200 ; vmulps %ymm0,%ymm1,%ymm1
+ DB 197,236,88,201 ; vaddps %ymm1,%ymm2,%ymm1
+ DB 196,226,125,24,80,24 ; vbroadcastss 0x18(%rax),%ymm2
+ DB 196,226,125,24,88,8 ; vbroadcastss 0x8(%rax),%ymm3
+ DB 197,236,89,208 ; vmulps %ymm0,%ymm2,%ymm2
+ DB 197,228,88,210 ; vaddps %ymm2,%ymm3,%ymm2
+ DB 196,226,125,24,88,28 ; vbroadcastss 0x1c(%rax),%ymm3
+ DB 196,98,125,24,72,12 ; vbroadcastss 0xc(%rax),%ymm9
+ DB 197,228,89,192 ; vmulps %ymm0,%ymm3,%ymm0
+ DB 197,180,88,216 ; vaddps %ymm0,%ymm9,%ymm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 197,124,41,192 ; vmovaps %ymm8,%ymm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_start_pipeline_sse41
+_sk_start_pipeline_sse41 LABEL PROC
+ DB 65,87 ; push %r15
+ DB 65,86 ; push %r14
+ DB 65,85 ; push %r13
+ DB 65,84 ; push %r12
+ DB 86 ; push %rsi
+ DB 87 ; push %rdi
+ DB 83 ; push %rbx
+ DB 72,129,236,160,0,0,0 ; sub $0xa0,%rsp
+ DB 68,15,41,188,36,144,0,0,0 ; movaps %xmm15,0x90(%rsp)
+ DB 68,15,41,180,36,128,0,0,0 ; movaps %xmm14,0x80(%rsp)
+ DB 68,15,41,108,36,112 ; movaps %xmm13,0x70(%rsp)
+ DB 68,15,41,100,36,96 ; movaps %xmm12,0x60(%rsp)
+ DB 68,15,41,92,36,80 ; movaps %xmm11,0x50(%rsp)
+ DB 68,15,41,84,36,64 ; movaps %xmm10,0x40(%rsp)
+ DB 68,15,41,76,36,48 ; movaps %xmm9,0x30(%rsp)
+ DB 68,15,41,68,36,32 ; movaps %xmm8,0x20(%rsp)
+ DB 15,41,124,36,16 ; movaps %xmm7,0x10(%rsp)
+ DB 15,41,52,36 ; movaps %xmm6,(%rsp)
+ DB 77,137,207 ; mov %r9,%r15
+ DB 77,137,198 ; mov %r8,%r14
+ DB 72,137,203 ; mov %rcx,%rbx
+ DB 72,137,214 ; mov %rdx,%rsi
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 73,137,196 ; mov %rax,%r12
+ DB 73,137,245 ; mov %rsi,%r13
+ DB 72,141,67,4 ; lea 0x4(%rbx),%rax
+ DB 76,57,248 ; cmp %r15,%rax
+ DB 118,5 ; jbe 73 <_sk_start_pipeline_sse41+0x73>
+ DB 72,137,216 ; mov %rbx,%rax
+ DB 235,52 ; jmp a7 <_sk_start_pipeline_sse41+0xa7>
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 15,87,228 ; xorps %xmm4,%xmm4
+ DB 15,87,237 ; xorps %xmm5,%xmm5
+ DB 15,87,246 ; xorps %xmm6,%xmm6
+ DB 15,87,255 ; xorps %xmm7,%xmm7
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 76,137,238 ; mov %r13,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,212 ; callq *%r12
+ DB 72,141,67,4 ; lea 0x4(%rbx),%rax
+ DB 72,131,195,8 ; add $0x8,%rbx
+ DB 76,57,251 ; cmp %r15,%rbx
+ DB 72,137,195 ; mov %rax,%rbx
+ DB 118,204 ; jbe 73 <_sk_start_pipeline_sse41+0x73>
+ DB 15,40,52,36 ; movaps (%rsp),%xmm6
+ DB 15,40,124,36,16 ; movaps 0x10(%rsp),%xmm7
+ DB 68,15,40,68,36,32 ; movaps 0x20(%rsp),%xmm8
+ DB 68,15,40,76,36,48 ; movaps 0x30(%rsp),%xmm9
+ DB 68,15,40,84,36,64 ; movaps 0x40(%rsp),%xmm10
+ DB 68,15,40,92,36,80 ; movaps 0x50(%rsp),%xmm11
+ DB 68,15,40,100,36,96 ; movaps 0x60(%rsp),%xmm12
+ DB 68,15,40,108,36,112 ; movaps 0x70(%rsp),%xmm13
+ DB 68,15,40,180,36,128,0,0,0 ; movaps 0x80(%rsp),%xmm14
+ DB 68,15,40,188,36,144,0,0,0 ; movaps 0x90(%rsp),%xmm15
+ DB 72,129,196,160,0,0,0 ; add $0xa0,%rsp
+ DB 91 ; pop %rbx
+ DB 95 ; pop %rdi
+ DB 94 ; pop %rsi
+ DB 65,92 ; pop %r12
+ DB 65,93 ; pop %r13
+ DB 65,94 ; pop %r14
+ DB 65,95 ; pop %r15
+ DB 195 ; retq
+
+PUBLIC _sk_just_return_sse41
+_sk_just_return_sse41 LABEL PROC
+ DB 195 ; retq
+
+PUBLIC _sk_seed_shader_sse41
+_sk_seed_shader_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 102,15,110,199 ; movd %edi,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 15,91,200 ; cvtdq2ps %xmm0,%xmm1
+ DB 185,0,0,0,63 ; mov $0x3f000000,%ecx
+ DB 102,15,110,209 ; movd %ecx,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 15,16,2 ; movups (%rdx),%xmm0
+ DB 15,88,193 ; addps %xmm1,%xmm0
+ DB 102,15,110,8 ; movd (%rax),%xmm1
+ DB 102,15,112,201,0 ; pshufd $0x0,%xmm1,%xmm1
+ DB 15,91,201 ; cvtdq2ps %xmm1,%xmm1
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 15,87,228 ; xorps %xmm4,%xmm4
+ DB 15,87,237 ; xorps %xmm5,%xmm5
+ DB 15,87,246 ; xorps %xmm6,%xmm6
+ DB 15,87,255 ; xorps %xmm7,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_constant_color_sse41
+_sk_constant_color_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,16,24 ; movups (%rax),%xmm3
+ DB 15,40,195 ; movaps %xmm3,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,40,203 ; movaps %xmm3,%xmm1
+ DB 15,198,201,85 ; shufps $0x55,%xmm1,%xmm1
+ DB 15,40,211 ; movaps %xmm3,%xmm2
+ DB 15,198,210,170 ; shufps $0xaa,%xmm2,%xmm2
+ DB 15,198,219,255 ; shufps $0xff,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clear_sse41
+_sk_clear_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_plus__sse41
+_sk_plus__sse41 LABEL PROC
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_srcover_sse41
+_sk_srcover_sse41 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,92,195 ; subps %xmm3,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,204 ; mulps %xmm4,%xmm9
+ DB 65,15,88,193 ; addps %xmm9,%xmm0
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,205 ; mulps %xmm5,%xmm9
+ DB 65,15,88,201 ; addps %xmm9,%xmm1
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,206 ; mulps %xmm6,%xmm9
+ DB 65,15,88,209 ; addps %xmm9,%xmm2
+ DB 68,15,89,199 ; mulps %xmm7,%xmm8
+ DB 65,15,88,216 ; addps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_dstover_sse41
+_sk_dstover_sse41 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,92,199 ; subps %xmm7,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_0_sse41
+_sk_clamp_0_sse41 LABEL PROC
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 65,15,95,192 ; maxps %xmm8,%xmm0
+ DB 65,15,95,200 ; maxps %xmm8,%xmm1
+ DB 65,15,95,208 ; maxps %xmm8,%xmm2
+ DB 65,15,95,216 ; maxps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_1_sse41
+_sk_clamp_1_sse41 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,93,192 ; minps %xmm8,%xmm0
+ DB 65,15,93,200 ; minps %xmm8,%xmm1
+ DB 65,15,93,208 ; minps %xmm8,%xmm2
+ DB 65,15,93,216 ; minps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_a_sse41
+_sk_clamp_a_sse41 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,93,216 ; minps %xmm8,%xmm3
+ DB 15,93,195 ; minps %xmm3,%xmm0
+ DB 15,93,203 ; minps %xmm3,%xmm1
+ DB 15,93,211 ; minps %xmm3,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_set_rgb_sse41
+_sk_set_rgb_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,15,16,80,8 ; movss 0x8(%rax),%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_rb_sse41
+_sk_swap_rb_sse41 LABEL PROC
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,194 ; movaps %xmm2,%xmm0
+ DB 65,15,40,208 ; movaps %xmm8,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_sse41
+_sk_swap_sse41 LABEL PROC
+ DB 68,15,40,195 ; movaps %xmm3,%xmm8
+ DB 68,15,40,202 ; movaps %xmm2,%xmm9
+ DB 68,15,40,209 ; movaps %xmm1,%xmm10
+ DB 68,15,40,216 ; movaps %xmm0,%xmm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,196 ; movaps %xmm4,%xmm0
+ DB 15,40,205 ; movaps %xmm5,%xmm1
+ DB 15,40,214 ; movaps %xmm6,%xmm2
+ DB 15,40,223 ; movaps %xmm7,%xmm3
+ DB 65,15,40,227 ; movaps %xmm11,%xmm4
+ DB 65,15,40,234 ; movaps %xmm10,%xmm5
+ DB 65,15,40,241 ; movaps %xmm9,%xmm6
+ DB 65,15,40,248 ; movaps %xmm8,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_src_dst_sse41
+_sk_move_src_dst_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,224 ; movaps %xmm0,%xmm4
+ DB 15,40,233 ; movaps %xmm1,%xmm5
+ DB 15,40,242 ; movaps %xmm2,%xmm6
+ DB 15,40,251 ; movaps %xmm3,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_dst_src_sse41
+_sk_move_dst_src_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,196 ; movaps %xmm4,%xmm0
+ DB 15,40,205 ; movaps %xmm5,%xmm1
+ DB 15,40,214 ; movaps %xmm6,%xmm2
+ DB 15,40,223 ; movaps %xmm7,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_premul_sse41
+_sk_premul_sse41 LABEL PROC
+ DB 15,89,195 ; mulps %xmm3,%xmm0
+ DB 15,89,203 ; mulps %xmm3,%xmm1
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_unpremul_sse41
+_sk_unpremul_sse41 LABEL PROC
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 68,15,94,203 ; divps %xmm3,%xmm9
+ DB 68,15,194,195,4 ; cmpneqps %xmm3,%xmm8
+ DB 69,15,84,193 ; andps %xmm9,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_from_srgb_sse41
+_sk_from_srgb_sse41 LABEL PROC
+ DB 184,145,131,158,61 ; mov $0x3d9e8391,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,40,211 ; movaps %xmm11,%xmm10
+ DB 68,15,89,208 ; mulps %xmm0,%xmm10
+ DB 68,15,40,240 ; movaps %xmm0,%xmm14
+ DB 69,15,89,246 ; mulps %xmm14,%xmm14
+ DB 184,154,153,153,62 ; mov $0x3e99999a,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 184,92,143,50,63 ; mov $0x3f328f5c,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 69,15,88,204 ; addps %xmm12,%xmm9
+ DB 184,10,215,35,59 ; mov $0x3b23d70a,%eax
+ DB 102,68,15,110,232 ; movd %eax,%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 69,15,89,206 ; mulps %xmm14,%xmm9
+ DB 69,15,88,205 ; addps %xmm13,%xmm9
+ DB 184,174,71,97,61 ; mov $0x3d6147ae,%eax
+ DB 102,68,15,110,240 ; movd %eax,%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 65,15,194,198,1 ; cmpltps %xmm14,%xmm0
+ DB 102,69,15,56,20,202 ; blendvps %xmm0,%xmm10,%xmm9
+ DB 69,15,40,251 ; movaps %xmm11,%xmm15
+ DB 68,15,89,249 ; mulps %xmm1,%xmm15
+ DB 15,40,193 ; movaps %xmm1,%xmm0
+ DB 15,89,192 ; mulps %xmm0,%xmm0
+ DB 69,15,40,208 ; movaps %xmm8,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 69,15,88,212 ; addps %xmm12,%xmm10
+ DB 68,15,89,208 ; mulps %xmm0,%xmm10
+ DB 69,15,88,213 ; addps %xmm13,%xmm10
+ DB 65,15,194,206,1 ; cmpltps %xmm14,%xmm1
+ DB 15,40,193 ; movaps %xmm1,%xmm0
+ DB 102,69,15,56,20,215 ; blendvps %xmm0,%xmm15,%xmm10
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 15,40,194 ; movaps %xmm2,%xmm0
+ DB 15,89,192 ; mulps %xmm0,%xmm0
+ DB 68,15,89,194 ; mulps %xmm2,%xmm8
+ DB 69,15,88,196 ; addps %xmm12,%xmm8
+ DB 68,15,89,192 ; mulps %xmm0,%xmm8
+ DB 69,15,88,197 ; addps %xmm13,%xmm8
+ DB 65,15,194,214,1 ; cmpltps %xmm14,%xmm2
+ DB 15,40,194 ; movaps %xmm2,%xmm0
+ DB 102,69,15,56,20,195 ; blendvps %xmm0,%xmm11,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,193 ; movaps %xmm9,%xmm0
+ DB 65,15,40,202 ; movaps %xmm10,%xmm1
+ DB 65,15,40,208 ; movaps %xmm8,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_to_srgb_sse41
+_sk_to_srgb_sse41 LABEL PROC
+ DB 72,131,236,24 ; sub $0x18,%rsp
+ DB 15,41,60,36 ; movaps %xmm7,(%rsp)
+ DB 15,40,254 ; movaps %xmm6,%xmm7
+ DB 15,40,245 ; movaps %xmm5,%xmm6
+ DB 15,40,236 ; movaps %xmm4,%xmm5
+ DB 15,40,227 ; movaps %xmm3,%xmm4
+ DB 15,40,218 ; movaps %xmm2,%xmm3
+ DB 15,40,209 ; movaps %xmm1,%xmm2
+ DB 68,15,82,192 ; rsqrtps %xmm0,%xmm8
+ DB 69,15,83,200 ; rcpps %xmm8,%xmm9
+ DB 69,15,82,248 ; rsqrtps %xmm8,%xmm15
+ DB 184,41,92,71,65 ; mov $0x41475c29,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,40,211 ; movaps %xmm11,%xmm10
+ DB 68,15,89,208 ; mulps %xmm0,%xmm10
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 184,194,135,210,62 ; mov $0x3ed287c2,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 184,206,111,48,63 ; mov $0x3f306fce,%eax
+ DB 102,68,15,110,232 ; movd %eax,%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 184,168,87,202,61 ; mov $0x3dca57a8,%eax
+ DB 53,0,0,0,128 ; xor $0x80000000,%eax
+ DB 102,68,15,110,240 ; movd %eax,%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 69,15,89,205 ; mulps %xmm13,%xmm9
+ DB 69,15,88,206 ; addps %xmm14,%xmm9
+ DB 69,15,89,252 ; mulps %xmm12,%xmm15
+ DB 69,15,88,249 ; addps %xmm9,%xmm15
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 69,15,93,207 ; minps %xmm15,%xmm9
+ DB 184,4,231,140,59 ; mov $0x3b8ce704,%eax
+ DB 102,68,15,110,248 ; movd %eax,%xmm15
+ DB 69,15,198,255,0 ; shufps $0x0,%xmm15,%xmm15
+ DB 65,15,194,199,1 ; cmpltps %xmm15,%xmm0
+ DB 102,69,15,56,20,202 ; blendvps %xmm0,%xmm10,%xmm9
+ DB 68,15,82,210 ; rsqrtps %xmm2,%xmm10
+ DB 65,15,83,194 ; rcpps %xmm10,%xmm0
+ DB 69,15,82,210 ; rsqrtps %xmm10,%xmm10
+ DB 65,15,89,197 ; mulps %xmm13,%xmm0
+ DB 65,15,88,198 ; addps %xmm14,%xmm0
+ DB 69,15,89,212 ; mulps %xmm12,%xmm10
+ DB 68,15,88,208 ; addps %xmm0,%xmm10
+ DB 65,15,40,200 ; movaps %xmm8,%xmm1
+ DB 65,15,93,202 ; minps %xmm10,%xmm1
+ DB 69,15,40,211 ; movaps %xmm11,%xmm10
+ DB 68,15,89,210 ; mulps %xmm2,%xmm10
+ DB 65,15,194,215,1 ; cmpltps %xmm15,%xmm2
+ DB 15,40,194 ; movaps %xmm2,%xmm0
+ DB 102,65,15,56,20,202 ; blendvps %xmm0,%xmm10,%xmm1
+ DB 15,82,195 ; rsqrtps %xmm3,%xmm0
+ DB 15,83,208 ; rcpps %xmm0,%xmm2
+ DB 65,15,89,213 ; mulps %xmm13,%xmm2
+ DB 65,15,88,214 ; addps %xmm14,%xmm2
+ DB 15,82,192 ; rsqrtps %xmm0,%xmm0
+ DB 65,15,89,196 ; mulps %xmm12,%xmm0
+ DB 15,88,194 ; addps %xmm2,%xmm0
+ DB 68,15,93,192 ; minps %xmm0,%xmm8
+ DB 68,15,89,219 ; mulps %xmm3,%xmm11
+ DB 65,15,194,223,1 ; cmpltps %xmm15,%xmm3
+ DB 15,40,195 ; movaps %xmm3,%xmm0
+ DB 102,69,15,56,20,195 ; blendvps %xmm0,%xmm11,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,193 ; movaps %xmm9,%xmm0
+ DB 65,15,40,208 ; movaps %xmm8,%xmm2
+ DB 15,40,220 ; movaps %xmm4,%xmm3
+ DB 15,40,229 ; movaps %xmm5,%xmm4
+ DB 15,40,238 ; movaps %xmm6,%xmm5
+ DB 15,40,247 ; movaps %xmm7,%xmm6
+ DB 15,40,60,36 ; movaps (%rsp),%xmm7
+ DB 72,131,196,24 ; add $0x18,%rsp
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_1_float_sse41
+_sk_scale_1_float_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_u8_sse41
+_sk_scale_u8_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,68,15,56,49,4,56 ; pmovzxbd (%rax,%rdi,1),%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 65,15,89,193 ; mulps %xmm9,%xmm0
+ DB 65,15,89,201 ; mulps %xmm9,%xmm1
+ DB 65,15,89,209 ; mulps %xmm9,%xmm2
+ DB 65,15,89,217 ; mulps %xmm9,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_1_float_sse41
+_sk_lerp_1_float_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,92,223 ; subps %xmm7,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_u8_sse41
+_sk_lerp_u8_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,68,15,56,49,4,56 ; pmovzxbd (%rax,%rdi,1),%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,193 ; mulps %xmm9,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,201 ; mulps %xmm9,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 65,15,89,209 ; mulps %xmm9,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,92,223 ; subps %xmm7,%xmm3
+ DB 65,15,89,217 ; mulps %xmm9,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_565_sse41
+_sk_lerp_565_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,68,15,56,51,4,120 ; pmovzxwd (%rax,%rdi,2),%xmm8
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,203 ; cvtdq2ps %xmm3,%xmm9
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 102,68,15,110,208 ; movd %eax,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,203 ; cvtdq2ps %xmm3,%xmm9
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,195 ; cvtdq2ps %xmm3,%xmm8
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,194 ; mulps %xmm10,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,203 ; mulps %xmm11,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_tables_sse41
+_sk_load_tables_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,8 ; mov (%rax),%rcx
+ DB 76,139,64,8 ; mov 0x8(%rax),%r8
+ DB 243,68,15,111,4,185 ; movdqu (%rcx,%rdi,4),%xmm8
+ DB 185,255,0,0,0 ; mov $0xff,%ecx
+ DB 102,15,110,193 ; movd %ecx,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,65,15,111,200 ; movdqa %xmm8,%xmm1
+ DB 102,15,114,209,8 ; psrld $0x8,%xmm1
+ DB 102,15,219,200 ; pand %xmm0,%xmm1
+ DB 102,65,15,111,208 ; movdqa %xmm8,%xmm2
+ DB 102,15,114,210,16 ; psrld $0x10,%xmm2
+ DB 102,15,219,208 ; pand %xmm0,%xmm2
+ DB 102,65,15,219,192 ; pand %xmm8,%xmm0
+ DB 102,72,15,58,22,193,1 ; pextrq $0x1,%xmm0,%rcx
+ DB 65,137,201 ; mov %ecx,%r9d
+ DB 72,193,233,32 ; shr $0x20,%rcx
+ DB 102,73,15,126,194 ; movq %xmm0,%r10
+ DB 69,137,211 ; mov %r10d,%r11d
+ DB 73,193,234,32 ; shr $0x20,%r10
+ DB 243,67,15,16,4,152 ; movss (%r8,%r11,4),%xmm0
+ DB 102,67,15,58,33,4,144,16 ; insertps $0x10,(%r8,%r10,4),%xmm0
+ DB 102,67,15,58,33,4,136,32 ; insertps $0x20,(%r8,%r9,4),%xmm0
+ DB 102,65,15,58,33,4,136,48 ; insertps $0x30,(%r8,%rcx,4),%xmm0
+ DB 76,139,64,16 ; mov 0x10(%rax),%r8
+ DB 102,73,15,58,22,202,1 ; pextrq $0x1,%xmm1,%r10
+ DB 77,137,209 ; mov %r10,%r9
+ DB 73,193,233,32 ; shr $0x20,%r9
+ DB 102,72,15,126,201 ; movq %xmm1,%rcx
+ DB 65,137,203 ; mov %ecx,%r11d
+ DB 65,129,227,255,255,255,0 ; and $0xffffff,%r11d
+ DB 72,193,233,30 ; shr $0x1e,%rcx
+ DB 65,129,226,255,255,255,0 ; and $0xffffff,%r10d
+ DB 243,67,15,16,12,152 ; movss (%r8,%r11,4),%xmm1
+ DB 102,65,15,58,33,12,8,16 ; insertps $0x10,(%r8,%rcx,1),%xmm1
+ DB 243,67,15,16,28,144 ; movss (%r8,%r10,4),%xmm3
+ DB 102,15,58,33,203,32 ; insertps $0x20,%xmm3,%xmm1
+ DB 243,67,15,16,28,136 ; movss (%r8,%r9,4),%xmm3
+ DB 102,15,58,33,203,48 ; insertps $0x30,%xmm3,%xmm1
+ DB 76,139,72,24 ; mov 0x18(%rax),%r9
+ DB 102,72,15,58,22,209,1 ; pextrq $0x1,%xmm2,%rcx
+ DB 68,15,183,193 ; movzwl %cx,%r8d
+ DB 72,193,233,32 ; shr $0x20,%rcx
+ DB 102,72,15,126,208 ; movq %xmm2,%rax
+ DB 68,15,183,208 ; movzwl %ax,%r10d
+ DB 72,193,232,30 ; shr $0x1e,%rax
+ DB 243,67,15,16,20,145 ; movss (%r9,%r10,4),%xmm2
+ DB 102,65,15,58,33,20,1,16 ; insertps $0x10,(%r9,%rax,1),%xmm2
+ DB 243,67,15,16,28,129 ; movss (%r9,%r8,4),%xmm3
+ DB 102,15,58,33,211,32 ; insertps $0x20,%xmm3,%xmm2
+ DB 243,65,15,16,28,137 ; movss (%r9,%rcx,4),%xmm3
+ DB 102,15,58,33,211,48 ; insertps $0x30,%xmm3,%xmm2
+ DB 102,65,15,114,208,24 ; psrld $0x18,%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_a8_sse41
+_sk_load_a8_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,15,56,49,4,56 ; pmovzxbd (%rax,%rdi,1),%xmm0
+ DB 15,91,192 ; cvtdq2ps %xmm0,%xmm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_a8_sse41
+_sk_store_a8_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,127,67 ; mov $0x437f0000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,69,15,56,43,192 ; packusdw %xmm8,%xmm8
+ DB 102,69,15,103,192 ; packuswb %xmm8,%xmm8
+ DB 102,68,15,126,4,56 ; movd %xmm8,(%rax,%rdi,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_565_sse41
+_sk_load_565_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,15,56,51,20,120 ; pmovzxwd (%rax,%rdi,2),%xmm2
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,15,219,194 ; pand %xmm2,%xmm0
+ DB 15,91,200 ; cvtdq2ps %xmm0,%xmm1
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 102,15,110,200 ; movd %eax,%xmm1
+ DB 102,15,112,201,0 ; pshufd $0x0,%xmm1,%xmm1
+ DB 102,15,219,202 ; pand %xmm2,%xmm1
+ DB 15,91,217 ; cvtdq2ps %xmm1,%xmm3
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 102,15,110,200 ; movd %eax,%xmm1
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 15,89,203 ; mulps %xmm3,%xmm1
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,15,219,218 ; pand %xmm2,%xmm3
+ DB 15,91,219 ; cvtdq2ps %xmm3,%xmm3
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_565_sse41
+_sk_store_565_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,248,65 ; mov $0x41f80000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 102,65,15,114,241,11 ; pslld $0xb,%xmm9
+ DB 185,0,0,124,66 ; mov $0x427c0000,%ecx
+ DB 102,68,15,110,209 ; movd %ecx,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,69,15,91,210 ; cvtps2dq %xmm10,%xmm10
+ DB 102,65,15,114,242,5 ; pslld $0x5,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 68,15,89,194 ; mulps %xmm2,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,69,15,86,194 ; orpd %xmm10,%xmm8
+ DB 102,69,15,56,43,192 ; packusdw %xmm8,%xmm8
+ DB 102,68,15,214,4,120 ; movq %xmm8,(%rax,%rdi,2)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_8888_sse41
+_sk_load_8888_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,15,111,28,184 ; movdqu (%rax,%rdi,4),%xmm3
+ DB 184,255,0,0,0 ; mov $0xff,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,15,111,203 ; movdqa %xmm3,%xmm1
+ DB 102,15,114,209,8 ; psrld $0x8,%xmm1
+ DB 102,15,219,200 ; pand %xmm0,%xmm1
+ DB 102,15,111,211 ; movdqa %xmm3,%xmm2
+ DB 102,15,114,210,16 ; psrld $0x10,%xmm2
+ DB 102,15,219,208 ; pand %xmm0,%xmm2
+ DB 102,15,219,195 ; pand %xmm3,%xmm0
+ DB 15,91,192 ; cvtdq2ps %xmm0,%xmm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,91,201 ; cvtdq2ps %xmm1,%xmm1
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,91,210 ; cvtdq2ps %xmm2,%xmm2
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 102,15,114,211,24 ; psrld $0x18,%xmm3
+ DB 15,91,219 ; cvtdq2ps %xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_8888_sse41
+_sk_store_8888_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,127,67 ; mov $0x437f0000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 69,15,40,208 ; movaps %xmm8,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,69,15,91,210 ; cvtps2dq %xmm10,%xmm10
+ DB 102,65,15,114,242,8 ; pslld $0x8,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,202 ; mulps %xmm2,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 102,65,15,114,241,16 ; pslld $0x10,%xmm9
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,65,15,114,240,24 ; pslld $0x18,%xmm8
+ DB 102,69,15,235,193 ; por %xmm9,%xmm8
+ DB 102,69,15,235,194 ; por %xmm10,%xmm8
+ DB 243,68,15,127,4,184 ; movdqu %xmm8,(%rax,%rdi,4)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_f16_sse41
+_sk_load_f16_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,15,111,4,248 ; movdqu (%rax,%rdi,8),%xmm0
+ DB 243,15,111,76,248,16 ; movdqu 0x10(%rax,%rdi,8),%xmm1
+ DB 102,15,111,208 ; movdqa %xmm0,%xmm2
+ DB 102,15,97,209 ; punpcklwd %xmm1,%xmm2
+ DB 102,15,105,193 ; punpckhwd %xmm1,%xmm0
+ DB 102,68,15,111,194 ; movdqa %xmm2,%xmm8
+ DB 102,68,15,97,192 ; punpcklwd %xmm0,%xmm8
+ DB 102,15,105,208 ; punpckhwd %xmm0,%xmm2
+ DB 184,0,4,0,4 ; mov $0x4000400,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,216,0 ; pshufd $0x0,%xmm0,%xmm3
+ DB 102,15,111,203 ; movdqa %xmm3,%xmm1
+ DB 102,65,15,101,200 ; pcmpgtw %xmm8,%xmm1
+ DB 102,65,15,223,200 ; pandn %xmm8,%xmm1
+ DB 102,15,101,218 ; pcmpgtw %xmm2,%xmm3
+ DB 102,15,223,218 ; pandn %xmm2,%xmm3
+ DB 102,15,56,51,193 ; pmovzxwd %xmm1,%xmm0
+ DB 102,15,114,240,13 ; pslld $0xd,%xmm0
+ DB 184,0,0,128,119 ; mov $0x77800000,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 102,68,15,112,194,0 ; pshufd $0x0,%xmm2,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 102,69,15,239,201 ; pxor %xmm9,%xmm9
+ DB 102,65,15,105,201 ; punpckhwd %xmm9,%xmm1
+ DB 102,15,114,241,13 ; pslld $0xd,%xmm1
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 102,15,56,51,211 ; pmovzxwd %xmm3,%xmm2
+ DB 102,15,114,242,13 ; pslld $0xd,%xmm2
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 102,65,15,105,217 ; punpckhwd %xmm9,%xmm3
+ DB 102,15,114,243,13 ; pslld $0xd,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_f16_sse41
+_sk_store_f16_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,128,7 ; mov $0x7800000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 102,69,15,112,192,0 ; pshufd $0x0,%xmm8,%xmm8
+ DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,65,15,114,209,13 ; psrld $0xd,%xmm9
+ DB 102,69,15,111,208 ; movdqa %xmm8,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,65,15,114,210,13 ; psrld $0xd,%xmm10
+ DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 102,65,15,114,211,13 ; psrld $0xd,%xmm11
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,65,15,114,208,13 ; psrld $0xd,%xmm8
+ DB 102,65,15,115,250,2 ; pslldq $0x2,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 102,65,15,115,248,2 ; pslldq $0x2,%xmm8
+ DB 102,69,15,235,195 ; por %xmm11,%xmm8
+ DB 102,69,15,111,202 ; movdqa %xmm10,%xmm9
+ DB 102,69,15,98,200 ; punpckldq %xmm8,%xmm9
+ DB 243,68,15,127,12,248 ; movdqu %xmm9,(%rax,%rdi,8)
+ DB 102,69,15,106,208 ; punpckhdq %xmm8,%xmm10
+ DB 243,68,15,127,84,248,16 ; movdqu %xmm10,0x10(%rax,%rdi,8)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_f32_sse41
+_sk_store_f32_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,137,249 ; mov %rdi,%rcx
+ DB 72,193,225,4 ; shl $0x4,%rcx
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 68,15,40,200 ; movaps %xmm0,%xmm9
+ DB 68,15,20,201 ; unpcklps %xmm1,%xmm9
+ DB 68,15,40,210 ; movaps %xmm2,%xmm10
+ DB 68,15,40,218 ; movaps %xmm2,%xmm11
+ DB 68,15,20,219 ; unpcklps %xmm3,%xmm11
+ DB 68,15,21,193 ; unpckhps %xmm1,%xmm8
+ DB 68,15,21,211 ; unpckhps %xmm3,%xmm10
+ DB 69,15,40,225 ; movaps %xmm9,%xmm12
+ DB 102,69,15,20,227 ; unpcklpd %xmm11,%xmm12
+ DB 69,15,18,217 ; movhlps %xmm9,%xmm11
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 102,69,15,20,202 ; unpcklpd %xmm10,%xmm9
+ DB 69,15,18,208 ; movhlps %xmm8,%xmm10
+ DB 102,68,15,17,36,8 ; movupd %xmm12,(%rax,%rcx,1)
+ DB 68,15,17,92,8,16 ; movups %xmm11,0x10(%rax,%rcx,1)
+ DB 102,68,15,17,76,8,32 ; movupd %xmm9,0x20(%rax,%rcx,1)
+ DB 68,15,17,84,8,48 ; movups %xmm10,0x30(%rax,%rcx,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_x_sse41
+_sk_clamp_x_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,95,192 ; maxps %xmm0,%xmm8
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 102,15,118,192 ; pcmpeqd %xmm0,%xmm0
+ DB 102,65,15,254,193 ; paddd %xmm9,%xmm0
+ DB 68,15,93,192 ; minps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,192 ; movaps %xmm8,%xmm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_y_sse41
+_sk_clamp_y_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,95,193 ; maxps %xmm1,%xmm8
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 102,15,118,201 ; pcmpeqd %xmm1,%xmm1
+ DB 102,65,15,254,201 ; paddd %xmm9,%xmm1
+ DB 68,15,93,193 ; minps %xmm1,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,200 ; movaps %xmm8,%xmm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_x_sse41
+_sk_repeat_x_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,200 ; movaps %xmm0,%xmm9
+ DB 69,15,94,200 ; divps %xmm8,%xmm9
+ DB 102,69,15,58,8,201,1 ; roundps $0x1,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 65,15,92,193 ; subps %xmm9,%xmm0
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,193 ; minps %xmm9,%xmm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_y_sse41
+_sk_repeat_y_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 69,15,94,200 ; divps %xmm8,%xmm9
+ DB 102,69,15,58,8,201,1 ; roundps $0x1,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 65,15,92,201 ; subps %xmm9,%xmm1
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,201 ; minps %xmm9,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_x_sse41
+_sk_mirror_x_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 65,15,92,193 ; subps %xmm9,%xmm0
+ DB 243,69,15,88,192 ; addss %xmm8,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,208 ; movaps %xmm0,%xmm10
+ DB 69,15,94,208 ; divps %xmm8,%xmm10
+ DB 102,69,15,58,8,210,1 ; roundps $0x1,%xmm10,%xmm10
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 65,15,92,194 ; subps %xmm10,%xmm0
+ DB 65,15,92,193 ; subps %xmm9,%xmm0
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,92,192 ; subps %xmm0,%xmm8
+ DB 65,15,84,192 ; andps %xmm8,%xmm0
+ DB 102,69,15,118,192 ; pcmpeqd %xmm8,%xmm8
+ DB 102,69,15,254,193 ; paddd %xmm9,%xmm8
+ DB 65,15,93,192 ; minps %xmm8,%xmm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_y_sse41
+_sk_mirror_y_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 65,15,92,201 ; subps %xmm9,%xmm1
+ DB 243,69,15,88,192 ; addss %xmm8,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,209 ; movaps %xmm1,%xmm10
+ DB 69,15,94,208 ; divps %xmm8,%xmm10
+ DB 102,69,15,58,8,210,1 ; roundps $0x1,%xmm10,%xmm10
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 65,15,92,202 ; subps %xmm10,%xmm1
+ DB 65,15,92,201 ; subps %xmm9,%xmm1
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,92,193 ; subps %xmm1,%xmm8
+ DB 65,15,84,200 ; andps %xmm8,%xmm1
+ DB 102,69,15,118,192 ; pcmpeqd %xmm8,%xmm8
+ DB 102,69,15,254,193 ; paddd %xmm9,%xmm8
+ DB 65,15,93,200 ; minps %xmm8,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_luminance_to_alpha_sse41
+_sk_luminance_to_alpha_sse41 LABEL PROC
+ DB 184,208,179,89,62 ; mov $0x3e59b3d0,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 184,89,23,55,63 ; mov $0x3f371759,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 15,88,195 ; addps %xmm3,%xmm0
+ DB 184,152,221,147,61 ; mov $0x3d93dd98,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,218 ; mulps %xmm2,%xmm3
+ DB 15,88,216 ; addps %xmm0,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_2x3_sse41
+_sk_matrix_2x3_sse41 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,16 ; movss 0x10(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,12 ; movss 0xc(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_3x4_sse41
+_sk_matrix_3x4_sse41 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,12 ; movss 0xc(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,24 ; movss 0x18(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,36 ; movss 0x24(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,28 ; movss 0x1c(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,40 ; movss 0x28(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,32 ; movss 0x20(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,44 ; movss 0x2c(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,226 ; mulps %xmm2,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,210 ; movaps %xmm10,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_4x5_sse41
+_sk_matrix_4x5_sse41 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,32 ; movss 0x20(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,48 ; movss 0x30(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,64 ; movss 0x40(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,227 ; mulps %xmm3,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,20 ; movss 0x14(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,36 ; movss 0x24(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,52 ; movss 0x34(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,68 ; movss 0x44(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,227 ; mulps %xmm3,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,24 ; movss 0x18(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,40 ; movss 0x28(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,56 ; movss 0x38(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 243,68,15,16,112,72 ; movss 0x48(%rax),%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 68,15,89,235 ; mulps %xmm3,%xmm13
+ DB 69,15,88,238 ; addps %xmm14,%xmm13
+ DB 68,15,89,226 ; mulps %xmm2,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 243,68,15,16,88,12 ; movss 0xc(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,28 ; movss 0x1c(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,44 ; movss 0x2c(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 243,68,15,16,112,60 ; movss 0x3c(%rax),%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 243,68,15,16,120,76 ; movss 0x4c(%rax),%xmm15
+ DB 69,15,198,255,0 ; shufps $0x0,%xmm15,%xmm15
+ DB 68,15,89,243 ; mulps %xmm3,%xmm14
+ DB 69,15,88,247 ; addps %xmm15,%xmm14
+ DB 68,15,89,234 ; mulps %xmm2,%xmm13
+ DB 69,15,88,238 ; addps %xmm14,%xmm13
+ DB 69,15,89,225 ; mulps %xmm9,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,216 ; mulps %xmm8,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,210 ; movaps %xmm10,%xmm2
+ DB 65,15,40,219 ; movaps %xmm11,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_perspective_sse41
+_sk_matrix_perspective_sse41 LABEL PROC
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,68,15,16,72,4 ; movss 0x4(%rax),%xmm9
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 68,15,89,201 ; mulps %xmm1,%xmm9
+ DB 69,15,88,202 ; addps %xmm10,%xmm9
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,193 ; addps %xmm9,%xmm0
+ DB 243,68,15,16,72,12 ; movss 0xc(%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 69,15,88,202 ; addps %xmm10,%xmm9
+ DB 243,68,15,16,80,24 ; movss 0x18(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,28 ; movss 0x1c(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,32 ; movss 0x20(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,217 ; mulps %xmm1,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,83,202 ; rcpps %xmm10,%xmm1
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 68,15,89,201 ; mulps %xmm1,%xmm9
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,201 ; movaps %xmm9,%xmm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_linear_gradient_2stops_sse41
+_sk_linear_gradient_2stops_sse41 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 68,15,16,8 ; movups (%rax),%xmm9
+ DB 15,16,88,16 ; movups 0x10(%rax),%xmm3
+ DB 68,15,40,195 ; movaps %xmm3,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,40,201 ; movaps %xmm9,%xmm1
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 68,15,89,192 ; mulps %xmm0,%xmm8
+ DB 68,15,88,193 ; addps %xmm1,%xmm8
+ DB 15,40,203 ; movaps %xmm3,%xmm1
+ DB 15,198,201,85 ; shufps $0x55,%xmm1,%xmm1
+ DB 65,15,40,209 ; movaps %xmm9,%xmm2
+ DB 15,198,210,85 ; shufps $0x55,%xmm2,%xmm2
+ DB 15,89,200 ; mulps %xmm0,%xmm1
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 15,40,211 ; movaps %xmm3,%xmm2
+ DB 15,198,210,170 ; shufps $0xaa,%xmm2,%xmm2
+ DB 69,15,40,209 ; movaps %xmm9,%xmm10
+ DB 69,15,198,210,170 ; shufps $0xaa,%xmm10,%xmm10
+ DB 15,89,208 ; mulps %xmm0,%xmm2
+ DB 65,15,88,210 ; addps %xmm10,%xmm2
+ DB 15,198,219,255 ; shufps $0xff,%xmm3,%xmm3
+ DB 69,15,198,201,255 ; shufps $0xff,%xmm9,%xmm9
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 65,15,88,217 ; addps %xmm9,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,192 ; movaps %xmm8,%xmm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_start_pipeline_sse2
+_sk_start_pipeline_sse2 LABEL PROC
+ DB 65,87 ; push %r15
+ DB 65,86 ; push %r14
+ DB 65,85 ; push %r13
+ DB 65,84 ; push %r12
+ DB 86 ; push %rsi
+ DB 87 ; push %rdi
+ DB 83 ; push %rbx
+ DB 72,129,236,160,0,0,0 ; sub $0xa0,%rsp
+ DB 68,15,41,188,36,144,0,0,0 ; movaps %xmm15,0x90(%rsp)
+ DB 68,15,41,180,36,128,0,0,0 ; movaps %xmm14,0x80(%rsp)
+ DB 68,15,41,108,36,112 ; movaps %xmm13,0x70(%rsp)
+ DB 68,15,41,100,36,96 ; movaps %xmm12,0x60(%rsp)
+ DB 68,15,41,92,36,80 ; movaps %xmm11,0x50(%rsp)
+ DB 68,15,41,84,36,64 ; movaps %xmm10,0x40(%rsp)
+ DB 68,15,41,76,36,48 ; movaps %xmm9,0x30(%rsp)
+ DB 68,15,41,68,36,32 ; movaps %xmm8,0x20(%rsp)
+ DB 15,41,124,36,16 ; movaps %xmm7,0x10(%rsp)
+ DB 15,41,52,36 ; movaps %xmm6,(%rsp)
+ DB 77,137,207 ; mov %r9,%r15
+ DB 77,137,198 ; mov %r8,%r14
+ DB 72,137,203 ; mov %rcx,%rbx
+ DB 72,137,214 ; mov %rdx,%rsi
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 73,137,196 ; mov %rax,%r12
+ DB 73,137,245 ; mov %rsi,%r13
+ DB 72,141,67,4 ; lea 0x4(%rbx),%rax
+ DB 76,57,248 ; cmp %r15,%rax
+ DB 118,5 ; jbe 73 <_sk_start_pipeline_sse2+0x73>
+ DB 72,137,216 ; mov %rbx,%rax
+ DB 235,52 ; jmp a7 <_sk_start_pipeline_sse2+0xa7>
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 15,87,228 ; xorps %xmm4,%xmm4
+ DB 15,87,237 ; xorps %xmm5,%xmm5
+ DB 15,87,246 ; xorps %xmm6,%xmm6
+ DB 15,87,255 ; xorps %xmm7,%xmm7
+ DB 72,137,223 ; mov %rbx,%rdi
+ DB 76,137,238 ; mov %r13,%rsi
+ DB 76,137,242 ; mov %r14,%rdx
+ DB 65,255,212 ; callq *%r12
+ DB 72,141,67,4 ; lea 0x4(%rbx),%rax
+ DB 72,131,195,8 ; add $0x8,%rbx
+ DB 76,57,251 ; cmp %r15,%rbx
+ DB 72,137,195 ; mov %rax,%rbx
+ DB 118,204 ; jbe 73 <_sk_start_pipeline_sse2+0x73>
+ DB 15,40,52,36 ; movaps (%rsp),%xmm6
+ DB 15,40,124,36,16 ; movaps 0x10(%rsp),%xmm7
+ DB 68,15,40,68,36,32 ; movaps 0x20(%rsp),%xmm8
+ DB 68,15,40,76,36,48 ; movaps 0x30(%rsp),%xmm9
+ DB 68,15,40,84,36,64 ; movaps 0x40(%rsp),%xmm10
+ DB 68,15,40,92,36,80 ; movaps 0x50(%rsp),%xmm11
+ DB 68,15,40,100,36,96 ; movaps 0x60(%rsp),%xmm12
+ DB 68,15,40,108,36,112 ; movaps 0x70(%rsp),%xmm13
+ DB 68,15,40,180,36,128,0,0,0 ; movaps 0x80(%rsp),%xmm14
+ DB 68,15,40,188,36,144,0,0,0 ; movaps 0x90(%rsp),%xmm15
+ DB 72,129,196,160,0,0,0 ; add $0xa0,%rsp
+ DB 91 ; pop %rbx
+ DB 95 ; pop %rdi
+ DB 94 ; pop %rsi
+ DB 65,92 ; pop %r12
+ DB 65,93 ; pop %r13
+ DB 65,94 ; pop %r14
+ DB 65,95 ; pop %r15
+ DB 195 ; retq
+
+PUBLIC _sk_just_return_sse2
+_sk_just_return_sse2 LABEL PROC
+ DB 195 ; retq
+
+PUBLIC _sk_seed_shader_sse2
+_sk_seed_shader_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 102,15,110,199 ; movd %edi,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 15,91,200 ; cvtdq2ps %xmm0,%xmm1
+ DB 185,0,0,0,63 ; mov $0x3f000000,%ecx
+ DB 102,15,110,209 ; movd %ecx,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 15,16,2 ; movups (%rdx),%xmm0
+ DB 15,88,193 ; addps %xmm1,%xmm0
+ DB 102,15,110,8 ; movd (%rax),%xmm1
+ DB 102,15,112,201,0 ; pshufd $0x0,%xmm1,%xmm1
+ DB 15,91,201 ; cvtdq2ps %xmm1,%xmm1
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 15,87,228 ; xorps %xmm4,%xmm4
+ DB 15,87,237 ; xorps %xmm5,%xmm5
+ DB 15,87,246 ; xorps %xmm6,%xmm6
+ DB 15,87,255 ; xorps %xmm7,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_constant_color_sse2
+_sk_constant_color_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,16,24 ; movups (%rax),%xmm3
+ DB 15,40,195 ; movaps %xmm3,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,40,203 ; movaps %xmm3,%xmm1
+ DB 15,198,201,85 ; shufps $0x55,%xmm1,%xmm1
+ DB 15,40,211 ; movaps %xmm3,%xmm2
+ DB 15,198,210,170 ; shufps $0xaa,%xmm2,%xmm2
+ DB 15,198,219,255 ; shufps $0xff,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clear_sse2
+_sk_clear_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 15,87,219 ; xorps %xmm3,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_plus__sse2
+_sk_plus__sse2 LABEL PROC
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_srcover_sse2
+_sk_srcover_sse2 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,92,195 ; subps %xmm3,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,204 ; mulps %xmm4,%xmm9
+ DB 65,15,88,193 ; addps %xmm9,%xmm0
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,205 ; mulps %xmm5,%xmm9
+ DB 65,15,88,201 ; addps %xmm9,%xmm1
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,206 ; mulps %xmm6,%xmm9
+ DB 65,15,88,209 ; addps %xmm9,%xmm2
+ DB 68,15,89,199 ; mulps %xmm7,%xmm8
+ DB 65,15,88,216 ; addps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_dstover_sse2
+_sk_dstover_sse2 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,92,199 ; subps %xmm7,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_0_sse2
+_sk_clamp_0_sse2 LABEL PROC
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 65,15,95,192 ; maxps %xmm8,%xmm0
+ DB 65,15,95,200 ; maxps %xmm8,%xmm1
+ DB 65,15,95,208 ; maxps %xmm8,%xmm2
+ DB 65,15,95,216 ; maxps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_1_sse2
+_sk_clamp_1_sse2 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,93,192 ; minps %xmm8,%xmm0
+ DB 65,15,93,200 ; minps %xmm8,%xmm1
+ DB 65,15,93,208 ; minps %xmm8,%xmm2
+ DB 65,15,93,216 ; minps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_a_sse2
+_sk_clamp_a_sse2 LABEL PROC
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,93,216 ; minps %xmm8,%xmm3
+ DB 15,93,195 ; minps %xmm3,%xmm0
+ DB 15,93,203 ; minps %xmm3,%xmm1
+ DB 15,93,211 ; minps %xmm3,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_set_rgb_sse2
+_sk_set_rgb_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,15,16,80,8 ; movss 0x8(%rax),%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_rb_sse2
+_sk_swap_rb_sse2 LABEL PROC
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,194 ; movaps %xmm2,%xmm0
+ DB 65,15,40,208 ; movaps %xmm8,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_swap_sse2
+_sk_swap_sse2 LABEL PROC
+ DB 68,15,40,195 ; movaps %xmm3,%xmm8
+ DB 68,15,40,202 ; movaps %xmm2,%xmm9
+ DB 68,15,40,209 ; movaps %xmm1,%xmm10
+ DB 68,15,40,216 ; movaps %xmm0,%xmm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,196 ; movaps %xmm4,%xmm0
+ DB 15,40,205 ; movaps %xmm5,%xmm1
+ DB 15,40,214 ; movaps %xmm6,%xmm2
+ DB 15,40,223 ; movaps %xmm7,%xmm3
+ DB 65,15,40,227 ; movaps %xmm11,%xmm4
+ DB 65,15,40,234 ; movaps %xmm10,%xmm5
+ DB 65,15,40,241 ; movaps %xmm9,%xmm6
+ DB 65,15,40,248 ; movaps %xmm8,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_src_dst_sse2
+_sk_move_src_dst_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,224 ; movaps %xmm0,%xmm4
+ DB 15,40,233 ; movaps %xmm1,%xmm5
+ DB 15,40,242 ; movaps %xmm2,%xmm6
+ DB 15,40,251 ; movaps %xmm3,%xmm7
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_move_dst_src_sse2
+_sk_move_dst_src_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,40,196 ; movaps %xmm4,%xmm0
+ DB 15,40,205 ; movaps %xmm5,%xmm1
+ DB 15,40,214 ; movaps %xmm6,%xmm2
+ DB 15,40,223 ; movaps %xmm7,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_premul_sse2
+_sk_premul_sse2 LABEL PROC
+ DB 15,89,195 ; mulps %xmm3,%xmm0
+ DB 15,89,203 ; mulps %xmm3,%xmm1
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_unpremul_sse2
+_sk_unpremul_sse2 LABEL PROC
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 68,15,94,203 ; divps %xmm3,%xmm9
+ DB 68,15,194,195,4 ; cmpneqps %xmm3,%xmm8
+ DB 69,15,84,193 ; andps %xmm9,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_from_srgb_sse2
+_sk_from_srgb_sse2 LABEL PROC
+ DB 184,145,131,158,61 ; mov $0x3d9e8391,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,232 ; movaps %xmm8,%xmm13
+ DB 68,15,89,232 ; mulps %xmm0,%xmm13
+ DB 68,15,40,224 ; movaps %xmm0,%xmm12
+ DB 69,15,89,228 ; mulps %xmm12,%xmm12
+ DB 184,154,153,153,62 ; mov $0x3e99999a,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 184,92,143,50,63 ; mov $0x3f328f5c,%eax
+ DB 102,68,15,110,208 ; movd %eax,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 69,15,40,241 ; movaps %xmm9,%xmm14
+ DB 68,15,89,240 ; mulps %xmm0,%xmm14
+ DB 69,15,88,242 ; addps %xmm10,%xmm14
+ DB 184,10,215,35,59 ; mov $0x3b23d70a,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,244 ; mulps %xmm12,%xmm14
+ DB 69,15,88,243 ; addps %xmm11,%xmm14
+ DB 184,174,71,97,61 ; mov $0x3d6147ae,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 65,15,194,196,1 ; cmpltps %xmm12,%xmm0
+ DB 68,15,84,232 ; andps %xmm0,%xmm13
+ DB 65,15,85,198 ; andnps %xmm14,%xmm0
+ DB 65,15,86,197 ; orps %xmm13,%xmm0
+ DB 69,15,40,232 ; movaps %xmm8,%xmm13
+ DB 68,15,89,233 ; mulps %xmm1,%xmm13
+ DB 68,15,40,241 ; movaps %xmm1,%xmm14
+ DB 69,15,89,246 ; mulps %xmm14,%xmm14
+ DB 69,15,40,249 ; movaps %xmm9,%xmm15
+ DB 68,15,89,249 ; mulps %xmm1,%xmm15
+ DB 69,15,88,250 ; addps %xmm10,%xmm15
+ DB 69,15,89,254 ; mulps %xmm14,%xmm15
+ DB 69,15,88,251 ; addps %xmm11,%xmm15
+ DB 65,15,194,204,1 ; cmpltps %xmm12,%xmm1
+ DB 68,15,84,233 ; andps %xmm1,%xmm13
+ DB 65,15,85,207 ; andnps %xmm15,%xmm1
+ DB 65,15,86,205 ; orps %xmm13,%xmm1
+ DB 68,15,89,194 ; mulps %xmm2,%xmm8
+ DB 68,15,40,234 ; movaps %xmm2,%xmm13
+ DB 69,15,89,237 ; mulps %xmm13,%xmm13
+ DB 68,15,89,202 ; mulps %xmm2,%xmm9
+ DB 69,15,88,202 ; addps %xmm10,%xmm9
+ DB 69,15,89,205 ; mulps %xmm13,%xmm9
+ DB 69,15,88,203 ; addps %xmm11,%xmm9
+ DB 65,15,194,212,1 ; cmpltps %xmm12,%xmm2
+ DB 68,15,84,194 ; andps %xmm2,%xmm8
+ DB 65,15,85,209 ; andnps %xmm9,%xmm2
+ DB 65,15,86,208 ; orps %xmm8,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_to_srgb_sse2
+_sk_to_srgb_sse2 LABEL PROC
+ DB 68,15,82,192 ; rsqrtps %xmm0,%xmm8
+ DB 69,15,83,248 ; rcpps %xmm8,%xmm15
+ DB 69,15,82,232 ; rsqrtps %xmm8,%xmm13
+ DB 184,41,92,71,65 ; mov $0x41475c29,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,240 ; movaps %xmm8,%xmm14
+ DB 68,15,89,240 ; mulps %xmm0,%xmm14
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 184,194,135,210,62 ; mov $0x3ed287c2,%eax
+ DB 102,68,15,110,208 ; movd %eax,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 184,206,111,48,63 ; mov $0x3f306fce,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 184,168,87,202,61 ; mov $0x3dca57a8,%eax
+ DB 53,0,0,0,128 ; xor $0x80000000,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 69,15,89,251 ; mulps %xmm11,%xmm15
+ DB 69,15,88,252 ; addps %xmm12,%xmm15
+ DB 69,15,89,234 ; mulps %xmm10,%xmm13
+ DB 69,15,88,239 ; addps %xmm15,%xmm13
+ DB 69,15,40,249 ; movaps %xmm9,%xmm15
+ DB 69,15,93,253 ; minps %xmm13,%xmm15
+ DB 184,4,231,140,59 ; mov $0x3b8ce704,%eax
+ DB 102,68,15,110,232 ; movd %eax,%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 65,15,194,197,1 ; cmpltps %xmm13,%xmm0
+ DB 68,15,84,240 ; andps %xmm0,%xmm14
+ DB 65,15,85,199 ; andnps %xmm15,%xmm0
+ DB 65,15,86,198 ; orps %xmm14,%xmm0
+ DB 68,15,82,241 ; rsqrtps %xmm1,%xmm14
+ DB 69,15,83,254 ; rcpps %xmm14,%xmm15
+ DB 69,15,82,246 ; rsqrtps %xmm14,%xmm14
+ DB 69,15,89,251 ; mulps %xmm11,%xmm15
+ DB 69,15,88,252 ; addps %xmm12,%xmm15
+ DB 69,15,89,242 ; mulps %xmm10,%xmm14
+ DB 69,15,88,247 ; addps %xmm15,%xmm14
+ DB 69,15,40,249 ; movaps %xmm9,%xmm15
+ DB 69,15,93,254 ; minps %xmm14,%xmm15
+ DB 69,15,40,240 ; movaps %xmm8,%xmm14
+ DB 68,15,89,241 ; mulps %xmm1,%xmm14
+ DB 65,15,194,205,1 ; cmpltps %xmm13,%xmm1
+ DB 68,15,84,241 ; andps %xmm1,%xmm14
+ DB 65,15,85,207 ; andnps %xmm15,%xmm1
+ DB 65,15,86,206 ; orps %xmm14,%xmm1
+ DB 68,15,82,242 ; rsqrtps %xmm2,%xmm14
+ DB 69,15,83,254 ; rcpps %xmm14,%xmm15
+ DB 69,15,89,251 ; mulps %xmm11,%xmm15
+ DB 69,15,88,252 ; addps %xmm12,%xmm15
+ DB 69,15,82,222 ; rsqrtps %xmm14,%xmm11
+ DB 69,15,89,218 ; mulps %xmm10,%xmm11
+ DB 69,15,88,223 ; addps %xmm15,%xmm11
+ DB 69,15,93,203 ; minps %xmm11,%xmm9
+ DB 68,15,89,194 ; mulps %xmm2,%xmm8
+ DB 65,15,194,213,1 ; cmpltps %xmm13,%xmm2
+ DB 68,15,84,194 ; andps %xmm2,%xmm8
+ DB 65,15,85,209 ; andnps %xmm9,%xmm2
+ DB 65,15,86,208 ; orps %xmm8,%xmm2
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_1_float_sse2
+_sk_scale_1_float_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_scale_u8_sse2
+_sk_scale_u8_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,68,15,110,4,56 ; movd (%rax,%rdi,1),%xmm8
+ DB 102,69,15,239,201 ; pxor %xmm9,%xmm9
+ DB 102,69,15,96,193 ; punpcklbw %xmm9,%xmm8
+ DB 102,69,15,97,193 ; punpcklwd %xmm9,%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 65,15,89,193 ; mulps %xmm9,%xmm0
+ DB 65,15,89,201 ; mulps %xmm9,%xmm1
+ DB 65,15,89,209 ; mulps %xmm9,%xmm2
+ DB 65,15,89,217 ; mulps %xmm9,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_1_float_sse2
+_sk_lerp_1_float_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,92,223 ; subps %xmm7,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_u8_sse2
+_sk_lerp_u8_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,68,15,110,4,56 ; movd (%rax,%rdi,1),%xmm8
+ DB 102,69,15,239,201 ; pxor %xmm9,%xmm9
+ DB 102,69,15,96,193 ; punpcklbw %xmm9,%xmm8
+ DB 102,69,15,97,193 ; punpcklwd %xmm9,%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,200 ; movd %eax,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,193 ; mulps %xmm9,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,201 ; mulps %xmm9,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 65,15,89,209 ; mulps %xmm9,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 15,92,223 ; subps %xmm7,%xmm3
+ DB 65,15,89,217 ; mulps %xmm9,%xmm3
+ DB 15,88,223 ; addps %xmm7,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_lerp_565_sse2
+_sk_lerp_565_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,68,15,126,4,120 ; movq (%rax,%rdi,2),%xmm8
+ DB 102,15,239,219 ; pxor %xmm3,%xmm3
+ DB 102,68,15,97,195 ; punpcklwd %xmm3,%xmm8
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,203 ; cvtdq2ps %xmm3,%xmm9
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 102,68,15,110,208 ; movd %eax,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,203 ; cvtdq2ps %xmm3,%xmm9
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,65,15,219,216 ; pand %xmm8,%xmm3
+ DB 68,15,91,195 ; cvtdq2ps %xmm3,%xmm8
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 15,92,196 ; subps %xmm4,%xmm0
+ DB 65,15,89,194 ; mulps %xmm10,%xmm0
+ DB 15,88,196 ; addps %xmm4,%xmm0
+ DB 15,92,205 ; subps %xmm5,%xmm1
+ DB 65,15,89,203 ; mulps %xmm11,%xmm1
+ DB 15,88,205 ; addps %xmm5,%xmm1
+ DB 15,92,214 ; subps %xmm6,%xmm2
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 15,88,214 ; addps %xmm6,%xmm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_tables_sse2
+_sk_load_tables_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,8 ; mov (%rax),%rcx
+ DB 76,139,64,8 ; mov 0x8(%rax),%r8
+ DB 243,68,15,111,4,185 ; movdqu (%rcx,%rdi,4),%xmm8
+ DB 185,255,0,0,0 ; mov $0xff,%ecx
+ DB 102,15,110,193 ; movd %ecx,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
+ DB 102,65,15,114,209,8 ; psrld $0x8,%xmm9
+ DB 102,68,15,219,200 ; pand %xmm0,%xmm9
+ DB 102,69,15,111,208 ; movdqa %xmm8,%xmm10
+ DB 102,65,15,114,210,16 ; psrld $0x10,%xmm10
+ DB 102,68,15,219,208 ; pand %xmm0,%xmm10
+ DB 102,65,15,219,192 ; pand %xmm8,%xmm0
+ DB 102,15,112,216,78 ; pshufd $0x4e,%xmm0,%xmm3
+ DB 102,72,15,126,217 ; movq %xmm3,%rcx
+ DB 65,137,201 ; mov %ecx,%r9d
+ DB 72,193,233,32 ; shr $0x20,%rcx
+ DB 102,73,15,126,194 ; movq %xmm0,%r10
+ DB 69,137,211 ; mov %r10d,%r11d
+ DB 73,193,234,32 ; shr $0x20,%r10
+ DB 243,67,15,16,28,144 ; movss (%r8,%r10,4),%xmm3
+ DB 243,65,15,16,4,136 ; movss (%r8,%rcx,4),%xmm0
+ DB 15,20,216 ; unpcklps %xmm0,%xmm3
+ DB 243,67,15,16,4,152 ; movss (%r8,%r11,4),%xmm0
+ DB 243,67,15,16,12,136 ; movss (%r8,%r9,4),%xmm1
+ DB 15,20,193 ; unpcklps %xmm1,%xmm0
+ DB 15,20,195 ; unpcklps %xmm3,%xmm0
+ DB 76,139,64,16 ; mov 0x10(%rax),%r8
+ DB 102,65,15,112,201,78 ; pshufd $0x4e,%xmm9,%xmm1
+ DB 102,73,15,126,202 ; movq %xmm1,%r10
+ DB 77,137,209 ; mov %r10,%r9
+ DB 73,193,233,32 ; shr $0x20,%r9
+ DB 102,76,15,126,201 ; movq %xmm9,%rcx
+ DB 65,137,203 ; mov %ecx,%r11d
+ DB 65,129,227,255,255,255,0 ; and $0xffffff,%r11d
+ DB 72,193,233,30 ; shr $0x1e,%rcx
+ DB 65,129,226,255,255,255,0 ; and $0xffffff,%r10d
+ DB 243,65,15,16,28,8 ; movss (%r8,%rcx,1),%xmm3
+ DB 243,67,15,16,12,136 ; movss (%r8,%r9,4),%xmm1
+ DB 15,20,217 ; unpcklps %xmm1,%xmm3
+ DB 243,67,15,16,12,152 ; movss (%r8,%r11,4),%xmm1
+ DB 243,67,15,16,20,144 ; movss (%r8,%r10,4),%xmm2
+ DB 15,20,202 ; unpcklps %xmm2,%xmm1
+ DB 15,20,203 ; unpcklps %xmm3,%xmm1
+ DB 76,139,72,24 ; mov 0x18(%rax),%r9
+ DB 102,65,15,112,210,78 ; pshufd $0x4e,%xmm10,%xmm2
+ DB 102,72,15,126,209 ; movq %xmm2,%rcx
+ DB 68,15,183,193 ; movzwl %cx,%r8d
+ DB 72,193,233,32 ; shr $0x20,%rcx
+ DB 102,76,15,126,208 ; movq %xmm10,%rax
+ DB 68,15,183,208 ; movzwl %ax,%r10d
+ DB 72,193,232,30 ; shr $0x1e,%rax
+ DB 243,69,15,16,12,1 ; movss (%r9,%rax,1),%xmm9
+ DB 243,65,15,16,20,137 ; movss (%r9,%rcx,4),%xmm2
+ DB 68,15,20,202 ; unpcklps %xmm2,%xmm9
+ DB 243,67,15,16,20,145 ; movss (%r9,%r10,4),%xmm2
+ DB 243,67,15,16,28,129 ; movss (%r9,%r8,4),%xmm3
+ DB 15,20,211 ; unpcklps %xmm3,%xmm2
+ DB 65,15,20,209 ; unpcklps %xmm9,%xmm2
+ DB 102,65,15,114,208,24 ; psrld $0x18,%xmm8
+ DB 69,15,91,192 ; cvtdq2ps %xmm8,%xmm8
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_a8_sse2
+_sk_load_a8_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 102,15,110,4,56 ; movd (%rax,%rdi,1),%xmm0
+ DB 102,15,239,201 ; pxor %xmm1,%xmm1
+ DB 102,15,96,193 ; punpcklbw %xmm1,%xmm0
+ DB 102,15,97,193 ; punpcklwd %xmm1,%xmm0
+ DB 15,91,192 ; cvtdq2ps %xmm0,%xmm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 102,15,239,201 ; pxor %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_a8_sse2
+_sk_store_a8_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,127,67 ; mov $0x437f0000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,65,15,114,240,16 ; pslld $0x10,%xmm8
+ DB 102,65,15,114,224,16 ; psrad $0x10,%xmm8
+ DB 102,69,15,107,192 ; packssdw %xmm8,%xmm8
+ DB 102,69,15,103,192 ; packuswb %xmm8,%xmm8
+ DB 102,68,15,126,4,56 ; movd %xmm8,(%rax,%rdi,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_565_sse2
+_sk_load_565_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,15,126,20,120 ; movq (%rax,%rdi,2),%xmm2
+ DB 102,15,239,192 ; pxor %xmm0,%xmm0
+ DB 102,15,97,208 ; punpcklwd %xmm0,%xmm2
+ DB 184,0,248,0,0 ; mov $0xf800,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,15,219,194 ; pand %xmm2,%xmm0
+ DB 15,91,200 ; cvtdq2ps %xmm0,%xmm1
+ DB 184,8,33,132,55 ; mov $0x37842108,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 184,224,7,0,0 ; mov $0x7e0,%eax
+ DB 102,15,110,200 ; movd %eax,%xmm1
+ DB 102,15,112,201,0 ; pshufd $0x0,%xmm1,%xmm1
+ DB 102,15,219,202 ; pand %xmm2,%xmm1
+ DB 15,91,217 ; cvtdq2ps %xmm1,%xmm3
+ DB 184,33,8,2,58 ; mov $0x3a020821,%eax
+ DB 102,15,110,200 ; movd %eax,%xmm1
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 15,89,203 ; mulps %xmm3,%xmm1
+ DB 184,31,0,0,0 ; mov $0x1f,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 102,15,112,219,0 ; pshufd $0x0,%xmm3,%xmm3
+ DB 102,15,219,218 ; pand %xmm2,%xmm3
+ DB 15,91,219 ; cvtdq2ps %xmm3,%xmm3
+ DB 184,8,33,4,61 ; mov $0x3d042108,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 15,198,210,0 ; shufps $0x0,%xmm2,%xmm2
+ DB 15,89,211 ; mulps %xmm3,%xmm2
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_565_sse2
+_sk_store_565_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,248,65 ; mov $0x41f80000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 102,65,15,114,241,11 ; pslld $0xb,%xmm9
+ DB 185,0,0,124,66 ; mov $0x427c0000,%ecx
+ DB 102,68,15,110,209 ; movd %ecx,%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,69,15,91,210 ; cvtps2dq %xmm10,%xmm10
+ DB 102,65,15,114,242,5 ; pslld $0x5,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 68,15,89,194 ; mulps %xmm2,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,69,15,86,194 ; orpd %xmm10,%xmm8
+ DB 102,65,15,114,240,16 ; pslld $0x10,%xmm8
+ DB 102,65,15,114,224,16 ; psrad $0x10,%xmm8
+ DB 102,69,15,107,192 ; packssdw %xmm8,%xmm8
+ DB 102,68,15,214,4,120 ; movq %xmm8,(%rax,%rdi,2)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_8888_sse2
+_sk_load_8888_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,15,111,28,184 ; movdqu (%rax,%rdi,4),%xmm3
+ DB 184,255,0,0,0 ; mov $0xff,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,192,0 ; pshufd $0x0,%xmm0,%xmm0
+ DB 102,15,111,203 ; movdqa %xmm3,%xmm1
+ DB 102,15,114,209,8 ; psrld $0x8,%xmm1
+ DB 102,15,219,200 ; pand %xmm0,%xmm1
+ DB 102,15,111,211 ; movdqa %xmm3,%xmm2
+ DB 102,15,114,210,16 ; psrld $0x10,%xmm2
+ DB 102,15,219,208 ; pand %xmm0,%xmm2
+ DB 102,15,219,195 ; pand %xmm3,%xmm0
+ DB 15,91,192 ; cvtdq2ps %xmm0,%xmm0
+ DB 184,129,128,128,59 ; mov $0x3b808081,%eax
+ DB 102,68,15,110,192 ; movd %eax,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 15,91,201 ; cvtdq2ps %xmm1,%xmm1
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 15,91,210 ; cvtdq2ps %xmm2,%xmm2
+ DB 65,15,89,208 ; mulps %xmm8,%xmm2
+ DB 102,15,114,211,24 ; psrld $0x18,%xmm3
+ DB 15,91,219 ; cvtdq2ps %xmm3,%xmm3
+ DB 65,15,89,216 ; mulps %xmm8,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_8888_sse2
+_sk_store_8888_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,127,67 ; mov $0x437f0000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 69,15,40,208 ; movaps %xmm8,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,69,15,91,210 ; cvtps2dq %xmm10,%xmm10
+ DB 102,65,15,114,242,8 ; pslld $0x8,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 68,15,89,202 ; mulps %xmm2,%xmm9
+ DB 102,69,15,91,201 ; cvtps2dq %xmm9,%xmm9
+ DB 102,65,15,114,241,16 ; pslld $0x10,%xmm9
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,69,15,91,192 ; cvtps2dq %xmm8,%xmm8
+ DB 102,65,15,114,240,24 ; pslld $0x18,%xmm8
+ DB 102,69,15,235,193 ; por %xmm9,%xmm8
+ DB 102,69,15,235,194 ; por %xmm10,%xmm8
+ DB 243,68,15,127,4,184 ; movdqu %xmm8,(%rax,%rdi,4)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_load_f16_sse2
+_sk_load_f16_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 243,15,111,4,248 ; movdqu (%rax,%rdi,8),%xmm0
+ DB 243,15,111,76,248,16 ; movdqu 0x10(%rax,%rdi,8),%xmm1
+ DB 102,15,111,208 ; movdqa %xmm0,%xmm2
+ DB 102,15,97,209 ; punpcklwd %xmm1,%xmm2
+ DB 102,15,105,193 ; punpckhwd %xmm1,%xmm0
+ DB 102,68,15,111,194 ; movdqa %xmm2,%xmm8
+ DB 102,68,15,97,192 ; punpcklwd %xmm0,%xmm8
+ DB 102,15,105,208 ; punpckhwd %xmm0,%xmm2
+ DB 184,0,4,0,4 ; mov $0x4000400,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 102,15,112,216,0 ; pshufd $0x0,%xmm0,%xmm3
+ DB 102,15,111,203 ; movdqa %xmm3,%xmm1
+ DB 102,65,15,101,200 ; pcmpgtw %xmm8,%xmm1
+ DB 102,65,15,223,200 ; pandn %xmm8,%xmm1
+ DB 102,15,101,218 ; pcmpgtw %xmm2,%xmm3
+ DB 102,15,223,218 ; pandn %xmm2,%xmm3
+ DB 102,69,15,239,192 ; pxor %xmm8,%xmm8
+ DB 102,15,111,193 ; movdqa %xmm1,%xmm0
+ DB 102,65,15,97,192 ; punpcklwd %xmm8,%xmm0
+ DB 102,15,114,240,13 ; pslld $0xd,%xmm0
+ DB 184,0,0,128,119 ; mov $0x77800000,%eax
+ DB 102,15,110,208 ; movd %eax,%xmm2
+ DB 102,68,15,112,202,0 ; pshufd $0x0,%xmm2,%xmm9
+ DB 65,15,89,193 ; mulps %xmm9,%xmm0
+ DB 102,65,15,105,200 ; punpckhwd %xmm8,%xmm1
+ DB 102,15,114,241,13 ; pslld $0xd,%xmm1
+ DB 65,15,89,201 ; mulps %xmm9,%xmm1
+ DB 102,15,111,211 ; movdqa %xmm3,%xmm2
+ DB 102,65,15,97,208 ; punpcklwd %xmm8,%xmm2
+ DB 102,15,114,242,13 ; pslld $0xd,%xmm2
+ DB 65,15,89,209 ; mulps %xmm9,%xmm2
+ DB 102,65,15,105,216 ; punpckhwd %xmm8,%xmm3
+ DB 102,15,114,243,13 ; pslld $0xd,%xmm3
+ DB 65,15,89,217 ; mulps %xmm9,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_f16_sse2
+_sk_store_f16_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 185,0,0,128,7 ; mov $0x7800000,%ecx
+ DB 102,68,15,110,193 ; movd %ecx,%xmm8
+ DB 102,69,15,112,192,0 ; pshufd $0x0,%xmm8,%xmm8
+ DB 102,69,15,111,200 ; movdqa %xmm8,%xmm9
+ DB 68,15,89,200 ; mulps %xmm0,%xmm9
+ DB 102,65,15,114,209,13 ; psrld $0xd,%xmm9
+ DB 102,69,15,111,208 ; movdqa %xmm8,%xmm10
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 102,65,15,114,210,13 ; psrld $0xd,%xmm10
+ DB 102,69,15,111,216 ; movdqa %xmm8,%xmm11
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 102,65,15,114,211,13 ; psrld $0xd,%xmm11
+ DB 68,15,89,195 ; mulps %xmm3,%xmm8
+ DB 102,65,15,114,208,13 ; psrld $0xd,%xmm8
+ DB 102,65,15,115,250,2 ; pslldq $0x2,%xmm10
+ DB 102,69,15,235,209 ; por %xmm9,%xmm10
+ DB 102,65,15,115,248,2 ; pslldq $0x2,%xmm8
+ DB 102,69,15,235,195 ; por %xmm11,%xmm8
+ DB 102,69,15,111,202 ; movdqa %xmm10,%xmm9
+ DB 102,69,15,98,200 ; punpckldq %xmm8,%xmm9
+ DB 243,68,15,127,12,248 ; movdqu %xmm9,(%rax,%rdi,8)
+ DB 102,69,15,106,208 ; punpckhdq %xmm8,%xmm10
+ DB 243,68,15,127,84,248,16 ; movdqu %xmm10,0x10(%rax,%rdi,8)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_store_f32_sse2
+_sk_store_f32_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 72,139,0 ; mov (%rax),%rax
+ DB 72,137,249 ; mov %rdi,%rcx
+ DB 72,193,225,4 ; shl $0x4,%rcx
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 68,15,40,200 ; movaps %xmm0,%xmm9
+ DB 68,15,20,201 ; unpcklps %xmm1,%xmm9
+ DB 68,15,40,210 ; movaps %xmm2,%xmm10
+ DB 68,15,40,218 ; movaps %xmm2,%xmm11
+ DB 68,15,20,219 ; unpcklps %xmm3,%xmm11
+ DB 68,15,21,193 ; unpckhps %xmm1,%xmm8
+ DB 68,15,21,211 ; unpckhps %xmm3,%xmm10
+ DB 69,15,40,225 ; movaps %xmm9,%xmm12
+ DB 102,69,15,20,227 ; unpcklpd %xmm11,%xmm12
+ DB 69,15,18,217 ; movhlps %xmm9,%xmm11
+ DB 69,15,40,200 ; movaps %xmm8,%xmm9
+ DB 102,69,15,20,202 ; unpcklpd %xmm10,%xmm9
+ DB 69,15,18,208 ; movhlps %xmm8,%xmm10
+ DB 102,68,15,17,36,8 ; movupd %xmm12,(%rax,%rcx,1)
+ DB 68,15,17,92,8,16 ; movups %xmm11,0x10(%rax,%rcx,1)
+ DB 102,68,15,17,76,8,32 ; movupd %xmm9,0x20(%rax,%rcx,1)
+ DB 68,15,17,84,8,48 ; movups %xmm10,0x30(%rax,%rcx,1)
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_x_sse2
+_sk_clamp_x_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,95,192 ; maxps %xmm0,%xmm8
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 102,15,118,192 ; pcmpeqd %xmm0,%xmm0
+ DB 102,65,15,254,193 ; paddd %xmm9,%xmm0
+ DB 68,15,93,192 ; minps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,192 ; movaps %xmm8,%xmm0
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_clamp_y_sse2
+_sk_clamp_y_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 69,15,87,192 ; xorps %xmm8,%xmm8
+ DB 68,15,95,193 ; maxps %xmm1,%xmm8
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 102,15,118,201 ; pcmpeqd %xmm1,%xmm1
+ DB 102,65,15,254,201 ; paddd %xmm9,%xmm1
+ DB 68,15,93,193 ; minps %xmm1,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,200 ; movaps %xmm8,%xmm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_x_sse2
+_sk_repeat_x_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,200 ; movaps %xmm0,%xmm9
+ DB 69,15,94,200 ; divps %xmm8,%xmm9
+ DB 243,69,15,91,209 ; cvttps2dq %xmm9,%xmm10
+ DB 69,15,91,210 ; cvtdq2ps %xmm10,%xmm10
+ DB 69,15,194,202,1 ; cmpltps %xmm10,%xmm9
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,84,217 ; andps %xmm9,%xmm11
+ DB 69,15,92,211 ; subps %xmm11,%xmm10
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 65,15,92,194 ; subps %xmm10,%xmm0
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,193 ; minps %xmm9,%xmm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_repeat_y_sse2
+_sk_repeat_y_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,0 ; movss (%rax),%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 69,15,94,200 ; divps %xmm8,%xmm9
+ DB 243,69,15,91,209 ; cvttps2dq %xmm9,%xmm10
+ DB 69,15,91,210 ; cvtdq2ps %xmm10,%xmm10
+ DB 69,15,194,202,1 ; cmpltps %xmm10,%xmm9
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,216 ; movd %eax,%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,84,217 ; andps %xmm9,%xmm11
+ DB 69,15,92,211 ; subps %xmm11,%xmm10
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 65,15,92,202 ; subps %xmm10,%xmm1
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,201 ; minps %xmm9,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_x_sse2
+_sk_mirror_x_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,40,193 ; movaps %xmm9,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,92,192 ; subps %xmm8,%xmm0
+ DB 243,69,15,88,201 ; addss %xmm9,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 68,15,40,208 ; movaps %xmm0,%xmm10
+ DB 69,15,94,209 ; divps %xmm9,%xmm10
+ DB 243,69,15,91,218 ; cvttps2dq %xmm10,%xmm11
+ DB 69,15,91,219 ; cvtdq2ps %xmm11,%xmm11
+ DB 69,15,194,211,1 ; cmpltps %xmm11,%xmm10
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 69,15,84,226 ; andps %xmm10,%xmm12
+ DB 69,15,87,210 ; xorps %xmm10,%xmm10
+ DB 69,15,92,220 ; subps %xmm12,%xmm11
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 65,15,92,195 ; subps %xmm11,%xmm0
+ DB 65,15,92,192 ; subps %xmm8,%xmm0
+ DB 68,15,92,208 ; subps %xmm0,%xmm10
+ DB 65,15,84,194 ; andps %xmm10,%xmm0
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,193 ; minps %xmm9,%xmm0
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_mirror_y_sse2
+_sk_mirror_y_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,68,15,16,8 ; movss (%rax),%xmm9
+ DB 69,15,40,193 ; movaps %xmm9,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,92,200 ; subps %xmm8,%xmm1
+ DB 243,69,15,88,201 ; addss %xmm9,%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 68,15,40,209 ; movaps %xmm1,%xmm10
+ DB 69,15,94,209 ; divps %xmm9,%xmm10
+ DB 243,69,15,91,218 ; cvttps2dq %xmm10,%xmm11
+ DB 69,15,91,219 ; cvtdq2ps %xmm11,%xmm11
+ DB 69,15,194,211,1 ; cmpltps %xmm11,%xmm10
+ DB 184,0,0,128,63 ; mov $0x3f800000,%eax
+ DB 102,68,15,110,224 ; movd %eax,%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 69,15,84,226 ; andps %xmm10,%xmm12
+ DB 69,15,87,210 ; xorps %xmm10,%xmm10
+ DB 69,15,92,220 ; subps %xmm12,%xmm11
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 65,15,92,203 ; subps %xmm11,%xmm1
+ DB 65,15,92,200 ; subps %xmm8,%xmm1
+ DB 68,15,92,209 ; subps %xmm1,%xmm10
+ DB 65,15,84,202 ; andps %xmm10,%xmm1
+ DB 102,69,15,118,201 ; pcmpeqd %xmm9,%xmm9
+ DB 102,69,15,254,200 ; paddd %xmm8,%xmm9
+ DB 65,15,93,201 ; minps %xmm9,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_luminance_to_alpha_sse2
+_sk_luminance_to_alpha_sse2 LABEL PROC
+ DB 184,208,179,89,62 ; mov $0x3e59b3d0,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 184,89,23,55,63 ; mov $0x3f371759,%eax
+ DB 102,15,110,192 ; movd %eax,%xmm0
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 15,88,195 ; addps %xmm3,%xmm0
+ DB 184,152,221,147,61 ; mov $0x3d93dd98,%eax
+ DB 102,15,110,216 ; movd %eax,%xmm3
+ DB 15,198,219,0 ; shufps $0x0,%xmm3,%xmm3
+ DB 15,89,218 ; mulps %xmm2,%xmm3
+ DB 15,88,216 ; addps %xmm0,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 15,87,192 ; xorps %xmm0,%xmm0
+ DB 15,87,201 ; xorps %xmm1,%xmm1
+ DB 15,87,210 ; xorps %xmm2,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_2x3_sse2
+_sk_matrix_2x3_sse2 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,16 ; movss 0x10(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,12 ; movss 0xc(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_3x4_sse2
+_sk_matrix_3x4_sse2 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,12 ; movss 0xc(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,24 ; movss 0x18(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,36 ; movss 0x24(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,28 ; movss 0x1c(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,40 ; movss 0x28(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,32 ; movss 0x20(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,44 ; movss 0x2c(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,226 ; mulps %xmm2,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,210 ; movaps %xmm10,%xmm2
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_4x5_sse2
+_sk_matrix_4x5_sse2 LABEL PROC
+ DB 68,15,40,201 ; movaps %xmm1,%xmm9
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,15,16,72,4 ; movss 0x4(%rax),%xmm1
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,32 ; movss 0x20(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,48 ; movss 0x30(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,64 ; movss 0x40(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,227 ; mulps %xmm3,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,194 ; addps %xmm10,%xmm0
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 243,68,15,16,80,20 ; movss 0x14(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,36 ; movss 0x24(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,52 ; movss 0x34(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,68 ; movss 0x44(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 68,15,89,227 ; mulps %xmm3,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 68,15,89,218 ; mulps %xmm2,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,209 ; mulps %xmm9,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,89,200 ; mulps %xmm8,%xmm1
+ DB 65,15,88,202 ; addps %xmm10,%xmm1
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,24 ; movss 0x18(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,40 ; movss 0x28(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,56 ; movss 0x38(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 243,68,15,16,112,72 ; movss 0x48(%rax),%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 68,15,89,235 ; mulps %xmm3,%xmm13
+ DB 69,15,88,238 ; addps %xmm14,%xmm13
+ DB 68,15,89,226 ; mulps %xmm2,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,217 ; mulps %xmm9,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 243,68,15,16,88,12 ; movss 0xc(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,28 ; movss 0x1c(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 243,68,15,16,104,44 ; movss 0x2c(%rax),%xmm13
+ DB 69,15,198,237,0 ; shufps $0x0,%xmm13,%xmm13
+ DB 243,68,15,16,112,60 ; movss 0x3c(%rax),%xmm14
+ DB 69,15,198,246,0 ; shufps $0x0,%xmm14,%xmm14
+ DB 243,68,15,16,120,76 ; movss 0x4c(%rax),%xmm15
+ DB 69,15,198,255,0 ; shufps $0x0,%xmm15,%xmm15
+ DB 68,15,89,243 ; mulps %xmm3,%xmm14
+ DB 69,15,88,247 ; addps %xmm15,%xmm14
+ DB 68,15,89,234 ; mulps %xmm2,%xmm13
+ DB 69,15,88,238 ; addps %xmm14,%xmm13
+ DB 69,15,89,225 ; mulps %xmm9,%xmm12
+ DB 69,15,88,229 ; addps %xmm13,%xmm12
+ DB 69,15,89,216 ; mulps %xmm8,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,210 ; movaps %xmm10,%xmm2
+ DB 65,15,40,219 ; movaps %xmm11,%xmm3
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_matrix_perspective_sse2
+_sk_matrix_perspective_sse2 LABEL PROC
+ DB 68,15,40,192 ; movaps %xmm0,%xmm8
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 243,15,16,0 ; movss (%rax),%xmm0
+ DB 243,68,15,16,72,4 ; movss 0x4(%rax),%xmm9
+ DB 15,198,192,0 ; shufps $0x0,%xmm0,%xmm0
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 243,68,15,16,80,8 ; movss 0x8(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 68,15,89,201 ; mulps %xmm1,%xmm9
+ DB 69,15,88,202 ; addps %xmm10,%xmm9
+ DB 65,15,89,192 ; mulps %xmm8,%xmm0
+ DB 65,15,88,193 ; addps %xmm9,%xmm0
+ DB 243,68,15,16,72,12 ; movss 0xc(%rax),%xmm9
+ DB 69,15,198,201,0 ; shufps $0x0,%xmm9,%xmm9
+ DB 243,68,15,16,80,16 ; movss 0x10(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,20 ; movss 0x14(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 68,15,89,209 ; mulps %xmm1,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 69,15,89,200 ; mulps %xmm8,%xmm9
+ DB 69,15,88,202 ; addps %xmm10,%xmm9
+ DB 243,68,15,16,80,24 ; movss 0x18(%rax),%xmm10
+ DB 69,15,198,210,0 ; shufps $0x0,%xmm10,%xmm10
+ DB 243,68,15,16,88,28 ; movss 0x1c(%rax),%xmm11
+ DB 69,15,198,219,0 ; shufps $0x0,%xmm11,%xmm11
+ DB 243,68,15,16,96,32 ; movss 0x20(%rax),%xmm12
+ DB 69,15,198,228,0 ; shufps $0x0,%xmm12,%xmm12
+ DB 68,15,89,217 ; mulps %xmm1,%xmm11
+ DB 69,15,88,220 ; addps %xmm12,%xmm11
+ DB 69,15,89,208 ; mulps %xmm8,%xmm10
+ DB 69,15,88,211 ; addps %xmm11,%xmm10
+ DB 65,15,83,202 ; rcpps %xmm10,%xmm1
+ DB 15,89,193 ; mulps %xmm1,%xmm0
+ DB 68,15,89,201 ; mulps %xmm1,%xmm9
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,201 ; movaps %xmm9,%xmm1
+ DB 255,224 ; jmpq *%rax
+
+PUBLIC _sk_linear_gradient_2stops_sse2
+_sk_linear_gradient_2stops_sse2 LABEL PROC
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 68,15,16,8 ; movups (%rax),%xmm9
+ DB 15,16,88,16 ; movups 0x10(%rax),%xmm3
+ DB 68,15,40,195 ; movaps %xmm3,%xmm8
+ DB 69,15,198,192,0 ; shufps $0x0,%xmm8,%xmm8
+ DB 65,15,40,201 ; movaps %xmm9,%xmm1
+ DB 15,198,201,0 ; shufps $0x0,%xmm1,%xmm1
+ DB 68,15,89,192 ; mulps %xmm0,%xmm8
+ DB 68,15,88,193 ; addps %xmm1,%xmm8
+ DB 15,40,203 ; movaps %xmm3,%xmm1
+ DB 15,198,201,85 ; shufps $0x55,%xmm1,%xmm1
+ DB 65,15,40,209 ; movaps %xmm9,%xmm2
+ DB 15,198,210,85 ; shufps $0x55,%xmm2,%xmm2
+ DB 15,89,200 ; mulps %xmm0,%xmm1
+ DB 15,88,202 ; addps %xmm2,%xmm1
+ DB 15,40,211 ; movaps %xmm3,%xmm2
+ DB 15,198,210,170 ; shufps $0xaa,%xmm2,%xmm2
+ DB 69,15,40,209 ; movaps %xmm9,%xmm10
+ DB 69,15,198,210,170 ; shufps $0xaa,%xmm10,%xmm10
+ DB 15,89,208 ; mulps %xmm0,%xmm2
+ DB 65,15,88,210 ; addps %xmm10,%xmm2
+ DB 15,198,219,255 ; shufps $0xff,%xmm3,%xmm3
+ DB 69,15,198,201,255 ; shufps $0xff,%xmm9,%xmm9
+ DB 15,89,216 ; mulps %xmm0,%xmm3
+ DB 65,15,88,217 ; addps %xmm9,%xmm3
+ DB 72,173 ; lods %ds:(%rsi),%rax
+ DB 65,15,40,192 ; movaps %xmm8,%xmm0
+ DB 255,224 ; jmpq *%rax
+END
diff --git a/src/jumper/build_stages.py b/src/jumper/build_stages.py
index fd50f1a328..fd584265d4 100755
--- a/src/jumper/build_stages.py
+++ b/src/jumper/build_stages.py
@@ -71,7 +71,15 @@ subprocess.check_call(clang + cflags + vfp4 +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'vfp4.o'])
-def parse_object_file(dot_o, array_type, target=None):
+def parse_object_file(dot_o, directive, target=None):
+ globl, label, comment = '.globl', ':', '// '
+ if 'win' in dot_o:
+ globl, label, comment = 'PUBLIC', ' LABEL PROC', '; '
+
+ dehex = lambda h: '0x'+h
+ if directive != '.long':
+ dehex = lambda h: str(int(h,16))
+
cmd = [objdump]
if target:
cmd += ['--target', target]
@@ -84,7 +92,6 @@ def parse_object_file(dot_o, array_type, target=None):
assert snippet not in section_headers
# Ok. Let's disassemble.
- active = False
disassemble = ['-d', '--insn-width=10', dot_o]
for line in subprocess.check_output(cmd + disassemble).split('\n'):
line = line.strip()
@@ -95,11 +102,9 @@ def parse_object_file(dot_o, array_type, target=None):
# E.g. 00000000000003a4 <_load_f16>:
m = re.match('''[0-9a-f]+ <_?(.*)>:''', line)
if m:
- if active:
- print '};'
print
- print 'CODE const', array_type, m.group(1) + '[] = {'
- active = True
+ print globl + ' _' + m.group(1)
+ print '_' + m.group(1) + label
continue
columns = line.split('\t')
@@ -113,54 +118,50 @@ def parse_object_file(dot_o, array_type, target=None):
inst, args = columns[2].split(' ', 1)
code, inst, args = code.strip(), inst.strip(), args.strip()
- dehex = lambda x: '0x'+x
- if array_type == 'uint8_t':
- dehex = lambda x: str(int(x, 16))
-
- hexed = ''.join(dehex(x) + ',' for x in code.split(' '))
- print ' ' + hexed + ' '*(40-len(hexed)) + \
- '//' + inst + (' '*(14-len(inst)) + args if args else '')
- print '};'
-
-sys.stdout = open('src/jumper/SkJumper_generated.cpp', 'w')
-
-print '''/*
- * Copyright 2017 Google Inc.
- *
- * Use of this source code is governed by a BSD-style license that can be
- * found in the LICENSE file.
- */
-
-// This file is generated semi-automatically with this command:
-// $ src/jumper/build_stages.py
-
-#include <stdint.h>
-
-#if defined(_MSC_VER)
- #pragma section("code", read,execute)
- #define CODE extern "C" __declspec(allocate("code"))
-#elif defined(__MACH__)
- #define CODE extern "C" __attribute__((section("__TEXT,__text")))
-#else
- #define CODE extern "C" __attribute__((section(".text")))
-#endif
+ hexed = ','.join(dehex(x) for x in code.split(' '))
+ print ' ' + directive + ' ' + hexed + ' '*(36-len(hexed)) + \
+ comment + inst + (' '*(14-len(inst)) + args if args else '')
+
+sys.stdout = open('src/jumper/SkJumper_generated.S', 'w')
+
+print '''# Copyright 2017 Google Inc.
+#
+# Use of this source code is governed by a BSD-style license that can be
+# found in the LICENSE file.
+
+# This file is generated semi-automatically with this command:
+# $ src/jumper/build_stages.py
'''
+print '.text'
print '#if defined(__aarch64__)'
-parse_object_file('aarch64.o', 'uint32_t')
+print '.balign 4'
+parse_object_file('aarch64.o', '.long')
print '#elif defined(__arm__)'
-parse_object_file('vfp4.o', 'uint32_t', target='elf32-littlearm')
+print '.balign 4'
+parse_object_file('vfp4.o', '.long', target='elf32-littlearm')
print '#elif defined(__x86_64__)'
-parse_object_file('hsw.o', 'uint8_t')
-parse_object_file('avx.o', 'uint8_t')
-parse_object_file('sse41.o', 'uint8_t')
-parse_object_file('sse2.o', 'uint8_t')
-
-print '#elif defined(_M_X64)'
-parse_object_file('win_hsw.o', 'uint8_t')
-parse_object_file('win_avx.o', 'uint8_t')
-parse_object_file('win_sse41.o', 'uint8_t')
-parse_object_file('win_sse2.o', 'uint8_t')
+parse_object_file('hsw.o', '.byte')
+parse_object_file('avx.o', '.byte')
+parse_object_file('sse41.o', '.byte')
+parse_object_file('sse2.o', '.byte')
print '#endif'
+
+sys.stdout = open('src/jumper/SkJumper_generated_win.S', 'w')
+print '''; Copyright 2017 Google Inc.
+;
+; Use of this source code is governed by a BSD-style license that can be
+; found in the LICENSE file.
+
+; This file is generated semi-automatically with this command:
+; $ src/jumper/build_stages.py
+'''
+
+print '_text SEGMENT'
+parse_object_file('win_hsw.o', 'DB')
+parse_object_file('win_avx.o', 'DB')
+parse_object_file('win_sse41.o', 'DB')
+parse_object_file('win_sse2.o', 'DB')
+print 'END'