aboutsummaryrefslogtreecommitdiffhomepage
diff options
context:
space:
mode:
authorGravatar Mike Klein <mtklein@chromium.org>2017-03-07 17:29:54 -0500
committerGravatar Skia Commit-Bot <skia-commit-bot@chromium.org>2017-03-14 23:20:53 +0000
commit5224f46b0d688861ed57490bd3540c833d61eeb1 (patch)
tree531c95ed600813c29cc28f69dedba0ae209f2140
parent3150ec6be91a9611064cf3697d7e179d143c2e8d (diff)
SkJumper: more constants, _f and _i literals.
Generalize section types to avoid, adding another type (.rodata). I've kept K for iota only. Maybe one day... Change-Id: Ie5678a2ea00fefe550bc0e6dcab32f98c31d3fae Reviewed-on: https://skia-review.googlesource.com/9403 Commit-Queue: Mike Klein <mtklein@chromium.org> Reviewed-by: Herb Derby <herb@google.com>
-rw-r--r--src/jumper/SkJumper.cpp8
-rw-r--r--src/jumper/SkJumper.h55
-rw-r--r--src/jumper/SkJumper_generated.cpp3780
-rw-r--r--src/jumper/SkJumper_stages.cpp180
-rwxr-xr-xsrc/jumper/build_stages.py10
5 files changed, 2268 insertions, 1765 deletions
diff --git a/src/jumper/SkJumper.cpp b/src/jumper/SkJumper.cpp
index caa9db2fd3..85f1231b1d 100644
--- a/src/jumper/SkJumper.cpp
+++ b/src/jumper/SkJumper.cpp
@@ -37,15 +37,7 @@
// It's fine to rearrange and add new ones if you update SkJumper_constants.
using K = const SkJumper_constants;
static K kConstants = {
- 1.0f, 0.5f, 255.0f, 1/255.0f, 0x000000ff,
{0.0f, 1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f},
- 0.0025f, 0.6975f, 0.3000f, 1/12.92f, 0.055f, // from_srgb
- 12.46f, 0.411192f, 0.689206f, -0.0988f, 0.0043f, // to_srgb
- 0x77800000, 0x07800000, 0x04000400, // fp16 <-> fp32
- 0x0000f800, 0x000007e0, 0x0000001f, // 565
- 1.0f/0x0000f800, 1.0f/0x000007e0, 1.0f/0x0000001f,
- 31.0f, 63.0f,
- SK_LUM_COEFF_R, SK_LUM_COEFF_G, SK_LUM_COEFF_B, // luminance -> alpha
};
#define STAGES(M) \
diff --git a/src/jumper/SkJumper.h b/src/jumper/SkJumper.h
index 215284b84c..712417a7de 100644
--- a/src/jumper/SkJumper.h
+++ b/src/jumper/SkJumper.h
@@ -14,54 +14,19 @@
#include <stdint.h>
-// SkJumper Stages can use constant literals only if they end up baked into the instruction,
-// like bit shifts and rounding modes. Any other constant values must be pulled from this struct
-// (except 0, ~0, and 0.0f, which always end up as some sort of xor or cmpeq instruction).
+// SkJumper_stages.cpp has some unusual constraints on what constants it can use.
//
-// This constraint makes it much easier to move and reorder the code for each Stage.
+// If the constant is baked into the instruction, that's ok.
+// If the constant is synthesized through code, that's ok.
+// If the constant is loaded from memory, that's no good.
+//
+// We offer a couple facilities to get at any other constants you need:
+// - the C() function usually constrains constants to be directly baked into an instruction; or
+// - the _i and _f user-defined literal operators call C() for you in a prettier way; or
+// - you can load values from this struct.
struct SkJumper_constants {
- float _1; // 1.0f
- float _0_5; // 0.5f
- float _255; // 255.0f
- float _1_255; // 1/255.0f
- uint32_t _0x000000ff; // 0x000000ff
-
- float iota[8]; // 0,1,2,3,4,5,6,7
-
- // from_srgb
- float _00025; // 0.0025f
- float _06975; // 0.6975f
- float _03000; // 0.3000f
- float _1_1292; // 1/12.92f
- float _0055; // 0.055f
-
- // to_srgb
- float _1246; // 12.46f
- float _0411192; // 0.411192f
- float _0689206; // 0.689206f
- float n_00988; // -0.0988f
- float _00043; // 0.0043f
-
- // fp16 <-> fp32
- uint32_t _0x77800000;
- uint32_t _0x07800000;
- uint32_t _0x04000400;
-
- // 565
- uint32_t r_565_mask;
- uint32_t g_565_mask;
- uint32_t b_565_mask;
- float r_565_scale;
- float g_565_scale;
- float b_565_scale;
- float _31;
- float _63;
-
- // luminance -> alpha
- float lum_r;
- float lum_g;
- float lum_b;
+ float iota[8]; // 0,1,2,3,4,5,6,7
};
#endif//SkJumper_DEFINED
diff --git a/src/jumper/SkJumper_generated.cpp b/src/jumper/SkJumper_generated.cpp
index 383922d0ce..df6e75234f 100644
--- a/src/jumper/SkJumper_generated.cpp
+++ b/src/jumper/SkJumper_generated.cpp
@@ -64,7 +64,7 @@ CODE const uint32_t sk_just_return_aarch64[] = {
CODE const uint32_t sk_seed_shader_aarch64[] = {
0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x3cc14046, //ldur q6, [x2, #20]
+ 0x3dc00046, //ldr q6, [x2]
0x4e040c00, //dup v0.4s, w0
0x4f0167e7, //movi v7.4s, #0x3f, lsl #24
0x4d40c901, //ld1r {v1.4s}, [x8]
@@ -244,101 +244,118 @@ CODE const uint32_t sk_unpremul_aarch64[] = {
};
CODE const uint32_t sk_from_srgb_aarch64[] = {
- 0x9100e048, //add x8, x2, #0x38
- 0x4d40c910, //ld1r {v16.4s}, [x8]
- 0x9100d048, //add x8, x2, #0x34
- 0x2d47cc52, //ldp s18, s19, [x2, #60]
- 0x4d40c911, //ld1r {v17.4s}, [x8]
+ 0x52a7d328, //mov w8, #0x3e990000
+ 0x72933348, //movk w8, #0x999a
+ 0x4e040d10, //dup v16.4s, w8
+ 0x52a7e648, //mov w8, #0x3f320000
+ 0x7291eb88, //movk w8, #0x8f5c
+ 0x4e040d11, //dup v17.4s, w8
+ 0x52a76468, //mov w8, #0x3b230000
+ 0x729ae148, //movk w8, #0xd70a
+ 0x4e040d12, //dup v18.4s, w8
+ 0x52a7b3c8, //mov w8, #0x3d9e0000
+ 0x72907228, //movk w8, #0x8391
0x6e22dc54, //fmul v20.4s, v2.4s, v2.4s
- 0x4eb01e15, //mov v21.16b, v16.16b
- 0x4eb01e17, //mov v23.16b, v16.16b
- 0x4f921050, //fmla v16.4s, v2.4s, v18.s[0]
- 0x4eb11e36, //mov v22.16b, v17.16b
- 0x4eb11e38, //mov v24.16b, v17.16b
- 0x4e34ce11, //fmla v17.4s, v16.4s, v20.4s
- 0x6e20dc10, //fmul v16.4s, v0.4s, v0.4s
- 0x91011048, //add x8, x2, #0x44
- 0x4f921015, //fmla v21.4s, v0.4s, v18.s[0]
- 0x4e30ceb6, //fmla v22.4s, v21.4s, v16.4s
- 0x4d40c910, //ld1r {v16.4s}, [x8]
+ 0x4eb11e35, //mov v21.16b, v17.16b
+ 0x4eb11e37, //mov v23.16b, v17.16b
+ 0x4e22ce11, //fmla v17.4s, v16.4s, v2.4s
+ 0x4eb21e56, //mov v22.16b, v18.16b
+ 0x4eb21e58, //mov v24.16b, v18.16b
+ 0x4e34ce32, //fmla v18.4s, v17.4s, v20.4s
+ 0x4e040d11, //dup v17.4s, w8
+ 0x52a7ac28, //mov w8, #0x3d610000
+ 0x6e20dc13, //fmul v19.4s, v0.4s, v0.4s
+ 0x7288f5c8, //movk w8, #0x47ae
+ 0x4e20ce15, //fmla v21.4s, v16.4s, v0.4s
0xf8408423, //ldr x3, [x1], #8
0x6e21dc34, //fmul v20.4s, v1.4s, v1.4s
- 0x4f921037, //fmla v23.4s, v1.4s, v18.s[0]
- 0x4f939015, //fmul v21.4s, v0.4s, v19.s[0]
- 0x4f939032, //fmul v18.4s, v1.4s, v19.s[0]
- 0x4f939053, //fmul v19.4s, v2.4s, v19.s[0]
- 0x6ea0e600, //fcmgt v0.4s, v16.4s, v0.4s
- 0x6ea1e601, //fcmgt v1.4s, v16.4s, v1.4s
- 0x6ea2e602, //fcmgt v2.4s, v16.4s, v2.4s
+ 0x4e33ceb6, //fmla v22.4s, v21.4s, v19.4s
+ 0x4e040d13, //dup v19.4s, w8
+ 0x4e21ce17, //fmla v23.4s, v16.4s, v1.4s
+ 0x6e31dc15, //fmul v21.4s, v0.4s, v17.4s
+ 0x6ea0e660, //fcmgt v0.4s, v19.4s, v0.4s
+ 0x6e31dc30, //fmul v16.4s, v1.4s, v17.4s
+ 0x6ea1e661, //fcmgt v1.4s, v19.4s, v1.4s
+ 0x6e31dc51, //fmul v17.4s, v2.4s, v17.4s
+ 0x6ea2e662, //fcmgt v2.4s, v19.4s, v2.4s
0x4e34cef8, //fmla v24.4s, v23.4s, v20.4s
0x6e761ea0, //bsl v0.16b, v21.16b, v22.16b
- 0x6e781e41, //bsl v1.16b, v18.16b, v24.16b
- 0x6e711e62, //bsl v2.16b, v19.16b, v17.16b
+ 0x6e781e01, //bsl v1.16b, v16.16b, v24.16b
+ 0x6e721e22, //bsl v2.16b, v17.16b, v18.16b
0xd61f0060, //br x3
};
CODE const uint32_t sk_to_srgb_aarch64[] = {
+ 0x52a828e8, //mov w8, #0x41470000
+ 0x728b8528, //movk w8, #0x5c29
+ 0x4e040d12, //dup v18.4s, w8
+ 0x52a7e608, //mov w8, #0x3f300000
+ 0x728df9c8, //movk w8, #0x6fce
0x6ea1d811, //frsqrte v17.4s, v0.4s
- 0x6ea1d835, //frsqrte v21.4s, v1.4s
- 0x6e31de37, //fmul v23.4s, v17.4s, v17.4s
- 0x6ea1d856, //frsqrte v22.4s, v2.4s
+ 0x4e040d13, //dup v19.4s, w8
+ 0x52b7b948, //mov w8, #0xbdca0000
+ 0x728af508, //movk w8, #0x57a8
+ 0x6ea1d834, //frsqrte v20.4s, v1.4s
+ 0x6e31de36, //fmul v22.4s, v17.4s, v17.4s
+ 0x4e040d10, //dup v16.4s, w8
+ 0x52a77188, //mov w8, #0x3b8c0000
+ 0x6ea1d855, //frsqrte v21.4s, v2.4s
+ 0x6e34de98, //fmul v24.4s, v20.4s, v20.4s
+ 0x4eb6fc16, //frsqrts v22.4s, v0.4s, v22.4s
+ 0x729ce088, //movk w8, #0xe704
0x6e35deb9, //fmul v25.4s, v21.4s, v21.4s
- 0x4eb7fc17, //frsqrts v23.4s, v0.4s, v23.4s
- 0x91015048, //add x8, x2, #0x54
- 0x6e36deda, //fmul v26.4s, v22.4s, v22.4s
- 0x4eb9fc39, //frsqrts v25.4s, v1.4s, v25.4s
- 0x6e37de31, //fmul v17.4s, v17.4s, v23.4s
- 0x4d40c914, //ld1r {v20.4s}, [x8]
- 0x4ebafc5a, //frsqrts v26.4s, v2.4s, v26.4s
+ 0x4eb8fc38, //frsqrts v24.4s, v1.4s, v24.4s
+ 0x6e36de31, //fmul v17.4s, v17.4s, v22.4s
+ 0x4e040d17, //dup v23.4s, w8
+ 0x4eb9fc59, //frsqrts v25.4s, v2.4s, v25.4s
+ 0x6e38de94, //fmul v20.4s, v20.4s, v24.4s
+ 0x4ea1da36, //frecpe v22.4s, v17.4s
+ 0x6e32dc1a, //fmul v26.4s, v0.4s, v18.4s
+ 0x6ea0e6e0, //fcmgt v0.4s, v23.4s, v0.4s
+ 0x6e32dc3c, //fmul v28.4s, v1.4s, v18.4s
+ 0x6ea1e6e1, //fcmgt v1.4s, v23.4s, v1.4s
+ 0x6e32dc52, //fmul v18.4s, v2.4s, v18.4s
+ 0x6ea2e6e2, //fcmgt v2.4s, v23.4s, v2.4s
0x6e39deb5, //fmul v21.4s, v21.4s, v25.4s
- 0x4ea1da37, //frecpe v23.4s, v17.4s
- 0xbd405053, //ldr s19, [x2, #80]
- 0x91016048, //add x8, x2, #0x58
- 0x6e3aded6, //fmul v22.4s, v22.4s, v26.4s
- 0x4ea1dabb, //frecpe v27.4s, v21.4s
- 0x4e37fe3d, //frecps v29.4s, v17.4s, v23.4s
- 0x2d494052, //ldp s18, s16, [x2, #72]
- 0x4d40c918, //ld1r {v24.4s}, [x8]
- 0x4ea1dadc, //frecpe v28.4s, v22.4s
- 0x6e3ddef7, //fmul v23.4s, v23.4s, v29.4s
- 0x4e3bfebd, //frecps v29.4s, v21.4s, v27.4s
- 0x6e3ddf7b, //fmul v27.4s, v27.4s, v29.4s
- 0x4e3cfedd, //frecps v29.4s, v22.4s, v28.4s
- 0x6e3ddf9c, //fmul v28.4s, v28.4s, v29.4s
- 0x4eb41e9d, //mov v29.16b, v20.16b
- 0x6ea1da39, //frsqrte v25.4s, v17.4s
- 0x4f9312fd, //fmla v29.4s, v23.4s, v19.s[0]
- 0x4eb41e97, //mov v23.16b, v20.16b
- 0x4f92901a, //fmul v26.4s, v0.4s, v18.s[0]
- 0x4f931377, //fmla v23.4s, v27.4s, v19.s[0]
- 0x4f931394, //fmla v20.4s, v28.4s, v19.s[0]
- 0x4f929033, //fmul v19.4s, v1.4s, v18.s[0]
- 0x4f929052, //fmul v18.4s, v2.4s, v18.s[0]
- 0x6ea0e700, //fcmgt v0.4s, v24.4s, v0.4s
- 0x6ea1e701, //fcmgt v1.4s, v24.4s, v1.4s
- 0x6ea2e702, //fcmgt v2.4s, v24.4s, v2.4s
- 0x6e39df38, //fmul v24.4s, v25.4s, v25.4s
- 0x6ea1dabb, //frsqrte v27.4s, v21.4s
+ 0x4ea1da97, //frecpe v23.4s, v20.4s
+ 0x4e36fe39, //frecps v25.4s, v17.4s, v22.4s
+ 0x4ea1dab8, //frecpe v24.4s, v21.4s
+ 0x6e39ded6, //fmul v22.4s, v22.4s, v25.4s
+ 0x4e37fe99, //frecps v25.4s, v20.4s, v23.4s
+ 0x4eb01e1b, //mov v27.16b, v16.16b
+ 0x6e39def7, //fmul v23.4s, v23.4s, v25.4s
+ 0x4e38feb9, //frecps v25.4s, v21.4s, v24.4s
+ 0x6e39df18, //fmul v24.4s, v24.4s, v25.4s
+ 0x4eb01e19, //mov v25.16b, v16.16b
+ 0x4e36ce7b, //fmla v27.4s, v19.4s, v22.4s
+ 0x6ea1da36, //frsqrte v22.4s, v17.4s
+ 0x4e37ce79, //fmla v25.4s, v19.4s, v23.4s
+ 0x6ea1da97, //frsqrte v23.4s, v20.4s
+ 0x4e38ce70, //fmla v16.4s, v19.4s, v24.4s
+ 0x6e36ded8, //fmul v24.4s, v22.4s, v22.4s
+ 0x6ea1dab3, //frsqrte v19.4s, v21.4s
0x4eb8fe31, //frsqrts v17.4s, v17.4s, v24.4s
- 0x6ea1dadc, //frsqrte v28.4s, v22.4s
- 0x6e3bdf78, //fmul v24.4s, v27.4s, v27.4s
- 0x6e31df31, //fmul v17.4s, v25.4s, v17.4s
+ 0x6e37def8, //fmul v24.4s, v23.4s, v23.4s
+ 0x4eb8fe94, //frsqrts v20.4s, v20.4s, v24.4s
+ 0x6e33de78, //fmul v24.4s, v19.4s, v19.4s
+ 0x52a7da48, //mov w8, #0x3ed20000
0x4eb8feb5, //frsqrts v21.4s, v21.4s, v24.4s
- 0x6e3cdf98, //fmul v24.4s, v28.4s, v28.4s
- 0x4f90123d, //fmla v29.4s, v17.4s, v16.s[0]
- 0x4d40c851, //ld1r {v17.4s}, [x2]
- 0x4eb8fed6, //frsqrts v22.4s, v22.4s, v24.4s
- 0x6e35df75, //fmul v21.4s, v27.4s, v21.4s
- 0x6e36df96, //fmul v22.4s, v28.4s, v22.4s
+ 0x7290f848, //movk w8, #0x87c2
+ 0x6e31ded1, //fmul v17.4s, v22.4s, v17.4s
+ 0x6e34def4, //fmul v20.4s, v23.4s, v20.4s
+ 0x6e35de73, //fmul v19.4s, v19.4s, v21.4s
+ 0x4e040d15, //dup v21.4s, w8
0xf8408423, //ldr x3, [x1], #8
- 0x4f9012b7, //fmla v23.4s, v21.4s, v16.s[0]
- 0x4f9012d4, //fmla v20.4s, v22.4s, v16.s[0]
- 0x4ebdf630, //fmin v16.4s, v17.4s, v29.4s
- 0x4eb7f635, //fmin v21.4s, v17.4s, v23.4s
- 0x4eb4f631, //fmin v17.4s, v17.4s, v20.4s
- 0x6e701f40, //bsl v0.16b, v26.16b, v16.16b
- 0x6e751e61, //bsl v1.16b, v19.16b, v21.16b
- 0x6e711e42, //bsl v2.16b, v18.16b, v17.16b
+ 0x4e31cebb, //fmla v27.4s, v21.4s, v17.4s
+ 0x4f03f611, //fmov v17.4s, #1.000000000000000000e+00
+ 0x4e34ceb9, //fmla v25.4s, v21.4s, v20.4s
+ 0x4e33ceb0, //fmla v16.4s, v21.4s, v19.4s
+ 0x4ebbf633, //fmin v19.4s, v17.4s, v27.4s
+ 0x4eb9f634, //fmin v20.4s, v17.4s, v25.4s
+ 0x4eb0f630, //fmin v16.4s, v17.4s, v16.4s
+ 0x6e731f40, //bsl v0.16b, v26.16b, v19.16b
+ 0x6e741f81, //bsl v1.16b, v28.16b, v20.16b
+ 0x6e701e42, //bsl v2.16b, v18.16b, v16.16b
0xd61f0060, //br x3
};
@@ -433,123 +450,132 @@ CODE const uint32_t sk_lerp_u8_aarch64[] = {
CODE const uint32_t sk_lerp_565_aarch64[] = {
0xa8c10c28, //ldp x8, x3, [x1], #16
0xd37ff809, //lsl x9, x0, #1
- 0x2d4ec851, //ldp s17, s18, [x2, #116]
- 0x4ea4d414, //fsub v20.4s, v0.4s, v4.4s
+ 0x4f072710, //movi v16.4s, #0xf8, lsl #8
+ 0x4ea4d413, //fsub v19.4s, v0.4s, v4.4s
0xf9400108, //ldr x8, [x8]
0xfc696903, //ldr d3, [x8, x9]
- 0x9101a048, //add x8, x2, #0x68
- 0x4d40c910, //ld1r {v16.4s}, [x8]
- 0x9101b048, //add x8, x2, #0x6c
- 0x4d40c913, //ld1r {v19.4s}, [x8]
- 0x9101c048, //add x8, x2, #0x70
+ 0x52a6f088, //mov w8, #0x37840000
+ 0x72842108, //movk w8, #0x2108
+ 0x4e040d11, //dup v17.4s, w8
0x2f10a463, //uxtl v3.4s, v3.4h
- 0x4d40c915, //ld1r {v21.4s}, [x8]
- 0x4e231e00, //and v0.16b, v16.16b, v3.16b
+ 0x321b17e8, //orr w8, wzr, #0x7e0
+ 0x4e301c60, //and v0.16b, v3.16b, v16.16b
+ 0x4e040d12, //dup v18.4s, w8
+ 0x52a74048, //mov w8, #0x3a020000
0x4e21d800, //scvtf v0.4s, v0.4s
- 0x4f919010, //fmul v16.4s, v0.4s, v17.s[0]
+ 0x72810428, //movk w8, #0x821
+ 0x6e31dc10, //fmul v16.4s, v0.4s, v17.4s
0x4ea41c80, //mov v0.16b, v4.16b
- 0xbd407c51, //ldr s17, [x2, #124]
- 0x4e34ce00, //fmla v0.4s, v16.4s, v20.4s
- 0x4e231e70, //and v16.16b, v19.16b, v3.16b
- 0x4e231ea3, //and v3.16b, v21.16b, v3.16b
+ 0x4e33ce00, //fmla v0.4s, v16.4s, v19.4s
+ 0x4f0007f0, //movi v16.4s, #0x1f
+ 0x4e040d11, //dup v17.4s, w8
+ 0x52a7a088, //mov w8, #0x3d040000
+ 0x4e321c72, //and v18.16b, v3.16b, v18.16b
+ 0x72842108, //movk w8, #0x2108
+ 0x4e301c63, //and v3.16b, v3.16b, v16.16b
+ 0x4ea6d450, //fsub v16.4s, v2.4s, v6.4s
+ 0x4e21da42, //scvtf v2.4s, v18.4s
+ 0x6e31dc51, //fmul v17.4s, v2.4s, v17.4s
+ 0x4e040d02, //dup v2.4s, w8
+ 0x4e21d863, //scvtf v3.4s, v3.4s
0x4ea5d433, //fsub v19.4s, v1.4s, v5.4s
- 0x4e21da01, //scvtf v1.4s, v16.4s
- 0x4f929030, //fmul v16.4s, v1.4s, v18.s[0]
- 0x4ea6d452, //fsub v18.4s, v2.4s, v6.4s
- 0x4e21d862, //scvtf v2.4s, v3.4s
0x4ea51ca1, //mov v1.16b, v5.16b
- 0x4f919043, //fmul v3.4s, v2.4s, v17.s[0]
+ 0x6e22dc63, //fmul v3.4s, v3.4s, v2.4s
0x4ea61cc2, //mov v2.16b, v6.16b
- 0x4e33ce01, //fmla v1.4s, v16.4s, v19.4s
- 0x4e32cc62, //fmla v2.4s, v3.4s, v18.4s
+ 0x4e33ce21, //fmla v1.4s, v17.4s, v19.4s
+ 0x4e30cc62, //fmla v2.4s, v3.4s, v16.4s
0x4f03f603, //fmov v3.4s, #1.000000000000000000e+00
0xd61f0060, //br x3
};
CODE const uint32_t sk_load_tables_aarch64[] = {
0xa8c10c28, //ldp x8, x3, [x1], #16
- 0x9100404b, //add x11, x2, #0x10
- 0x4d40c960, //ld1r {v0.4s}, [x11]
0xd37ef409, //lsl x9, x0, #2
- 0xa9402d0a, //ldp x10, x11, [x8]
+ 0x6f00e620, //movi v0.2d, #0xff000000ff
+ 0x52a7700b, //mov w11, #0x3b800000
+ 0xa940310a, //ldp x10, x12, [x8]
+ 0x7290102b, //movk w11, #0x8081
+ 0x4e040d63, //dup v3.4s, w11
0x3ce96942, //ldr q2, [x10, x9]
0xa9412109, //ldp x9, x8, [x8, #16]
- 0x4e221c01, //and v1.16b, v0.16b, v2.16b
- 0x0e143c2c, //mov w12, v1.s[2]
- 0xbc6c5971, //ldr s17, [x11, w12, uxtw #2]
- 0x1e26002c, //fmov w12, s1
- 0x6f380443, //ushr v3.4s, v2.4s, #8
- 0x6f300450, //ushr v16.4s, v2.4s, #16
- 0x8b2c496c, //add x12, x11, w12, uxtw #2
+ 0x4e201c41, //and v1.16b, v2.16b, v0.16b
+ 0x1e26002e, //fmov w14, s1
+ 0x6f380450, //ushr v16.4s, v2.4s, #8
+ 0x6f300451, //ushr v17.4s, v2.4s, #16
+ 0x8b2e498e, //add x14, x12, w14, uxtw #2
0x0e0c3c2a, //mov w10, v1.s[1]
+ 0x0e143c2b, //mov w11, v1.s[2]
0x0e1c3c2d, //mov w13, v1.s[3]
- 0x4e231c01, //and v1.16b, v0.16b, v3.16b
- 0x4e301c03, //and v3.16b, v0.16b, v16.16b
- 0x0d408180, //ld1 {v0.s}[0], [x12]
- 0x0e143c2c, //mov w12, v1.s[2]
- 0xbc6c5932, //ldr s18, [x9, w12, uxtw #2]
- 0x1e26002c, //fmov w12, s1
- 0x8b2a496a, //add x10, x11, w10, uxtw #2
- 0xbc6d5970, //ldr s16, [x11, w13, uxtw #2]
+ 0x4e201e01, //and v1.16b, v16.16b, v0.16b
+ 0x4e201e30, //and v16.16b, v17.16b, v0.16b
+ 0x0d4081c0, //ld1 {v0.s}[0], [x14]
+ 0x8b2a498a, //add x10, x12, w10, uxtw #2
+ 0xbc6b5991, //ldr s17, [x12, w11, uxtw #2]
+ 0xbc6d5992, //ldr s18, [x12, w13, uxtw #2]
0x0e0c3c2b, //mov w11, v1.s[1]
+ 0x0e143c2c, //mov w12, v1.s[2]
0x0e1c3c2d, //mov w13, v1.s[3]
- 0x8b2c492c, //add x12, x9, w12, uxtw #2
- 0xbc6d5933, //ldr s19, [x9, w13, uxtw #2]
- 0x0e0c3c6d, //mov w13, v3.s[1]
+ 0x1e26002e, //fmov w14, s1
+ 0x8b2e492e, //add x14, x9, w14, uxtw #2
+ 0xbc6c5933, //ldr s19, [x9, w12, uxtw #2]
+ 0xbc6d5934, //ldr s20, [x9, w13, uxtw #2]
0x8b2b4929, //add x9, x9, w11, uxtw #2
- 0x0e143c6b, //mov w11, v3.s[2]
- 0x0d408181, //ld1 {v1.s}[0], [x12]
- 0x0e1c3c6c, //mov w12, v3.s[3]
- 0x0d409140, //ld1 {v0.s}[1], [x10]
- 0x1e26006a, //fmov w10, s3
- 0xbd400c43, //ldr s3, [x2, #12]
+ 0x1e26020b, //fmov w11, s16
0x6f280442, //ushr v2.4s, v2.4s, #24
+ 0x0d409140, //ld1 {v0.s}[1], [x10]
0x4e21d842, //scvtf v2.4s, v2.4s
- 0x8b2a490a, //add x10, x8, w10, uxtw #2
- 0x4f839043, //fmul v3.4s, v2.4s, v3.s[0]
+ 0x8b2b490a, //add x10, x8, w11, uxtw #2
+ 0x0d4081c1, //ld1 {v1.s}[0], [x14]
+ 0x6e23dc43, //fmul v3.4s, v2.4s, v3.4s
0x0d408142, //ld1 {v2.s}[0], [x10]
- 0x8b2d490a, //add x10, x8, w13, uxtw #2
- 0x6e140620, //mov v0.s[2], v17.s[0]
- 0xbc6b5911, //ldr s17, [x8, w11, uxtw #2]
+ 0x0e0c3e0f, //mov w15, v16.s[1]
+ 0x0e143e0c, //mov w12, v16.s[2]
+ 0x8b2f490a, //add x10, x8, w15, uxtw #2
+ 0x0e1c3e0d, //mov w13, v16.s[3]
+ 0xbc6c5910, //ldr s16, [x8, w12, uxtw #2]
0x0d409121, //ld1 {v1.s}[1], [x9]
0x0d409142, //ld1 {v2.s}[1], [x10]
- 0x6e1c0600, //mov v0.s[3], v16.s[0]
- 0xbc6c5910, //ldr s16, [x8, w12, uxtw #2]
- 0x6e140641, //mov v1.s[2], v18.s[0]
- 0x6e140622, //mov v2.s[2], v17.s[0]
- 0x6e1c0661, //mov v1.s[3], v19.s[0]
- 0x6e1c0602, //mov v2.s[3], v16.s[0]
+ 0x6e140620, //mov v0.s[2], v17.s[0]
+ 0xbc6d5911, //ldr s17, [x8, w13, uxtw #2]
+ 0x6e140661, //mov v1.s[2], v19.s[0]
+ 0x6e140602, //mov v2.s[2], v16.s[0]
+ 0x6e1c0640, //mov v0.s[3], v18.s[0]
+ 0x6e1c0681, //mov v1.s[3], v20.s[0]
+ 0x6e1c0622, //mov v2.s[3], v17.s[0]
0xd61f0060, //br x3
};
CODE const uint32_t sk_load_a8_aarch64[] = {
0xa8c10c28, //ldp x8, x3, [x1], #16
- 0xbd400c43, //ldr s3, [x2, #12]
+ 0x52a77009, //mov w9, #0x3b800000
+ 0x72901029, //movk w9, #0x8081
+ 0x4e040d22, //dup v2.4s, w9
+ 0xf9400108, //ldr x8, [x8]
0x6f00e400, //movi v0.2d, #0x0
0x6f00e401, //movi v1.2d, #0x0
- 0xf9400108, //ldr x8, [x8]
0x8b000108, //add x8, x8, x0
- 0x39400109, //ldrb w9, [x8]
- 0x3940050a, //ldrb w10, [x8, #1]
- 0x3940090b, //ldrb w11, [x8, #2]
+ 0x3940010a, //ldrb w10, [x8]
+ 0x3940050b, //ldrb w11, [x8, #1]
+ 0x3940090c, //ldrb w12, [x8, #2]
0x39400d08, //ldrb w8, [x8, #3]
- 0x4e021d22, //mov v2.h[0], w9
- 0x4e061d42, //mov v2.h[1], w10
- 0x4e0a1d62, //mov v2.h[2], w11
- 0x4e0e1d02, //mov v2.h[3], w8
- 0x2f07b7e2, //bic v2.4h, #0xff, lsl #8
- 0x2f10a442, //uxtl v2.4s, v2.4h
- 0x6e21d842, //ucvtf v2.4s, v2.4s
- 0x4f839043, //fmul v3.4s, v2.4s, v3.s[0]
+ 0x4e021d43, //mov v3.h[0], w10
+ 0x4e061d63, //mov v3.h[1], w11
+ 0x4e0a1d83, //mov v3.h[2], w12
+ 0x4e0e1d03, //mov v3.h[3], w8
+ 0x2f07b7e3, //bic v3.4h, #0xff, lsl #8
+ 0x2f10a463, //uxtl v3.4s, v3.4h
+ 0x6e21d863, //ucvtf v3.4s, v3.4s
+ 0x6e22dc63, //fmul v3.4s, v3.4s, v2.4s
0x6f00e402, //movi v2.2d, #0x0
0xd61f0060, //br x3
};
CODE const uint32_t sk_store_a8_aarch64[] = {
0xf9400028, //ldr x8, [x1]
- 0xbd400850, //ldr s16, [x2, #8]
+ 0x52a86fe9, //mov w9, #0x437f0000
+ 0x4e040d30, //dup v16.4s, w9
+ 0x6e30dc70, //fmul v16.4s, v3.4s, v16.4s
0xf9400108, //ldr x8, [x8]
- 0x4f909070, //fmul v16.4s, v3.4s, v16.s[0]
0x6e21aa10, //fcvtnu v16.4s, v16.4s
0x0e612a10, //xtn v16.4h, v16.4s
0x0e0e3e09, //umov w9, v16.h[3]
@@ -569,45 +595,52 @@ CODE const uint32_t sk_store_a8_aarch64[] = {
CODE const uint32_t sk_load_565_aarch64[] = {
0xa8c10c28, //ldp x8, x3, [x1], #16
0xd37ff809, //lsl x9, x0, #1
+ 0x4f072701, //movi v1.4s, #0xf8, lsl #8
+ 0x4f0007e3, //movi v3.4s, #0x1f
0xf9400108, //ldr x8, [x8]
0xfc696900, //ldr d0, [x8, x9]
- 0x9101a048, //add x8, x2, #0x68
- 0x4d40c901, //ld1r {v1.4s}, [x8]
- 0x9101b048, //add x8, x2, #0x6c
- 0x4d40c902, //ld1r {v2.4s}, [x8]
- 0x9101c048, //add x8, x2, #0x70
- 0x4d40c903, //ld1r {v3.4s}, [x8]
+ 0x321b17e8, //orr w8, wzr, #0x7e0
+ 0x4e040d02, //dup v2.4s, w8
+ 0x52a6f088, //mov w8, #0x37840000
+ 0x72842108, //movk w8, #0x2108
0x2f10a400, //uxtl v0.4s, v0.4h
- 0x4e201c21, //and v1.16b, v1.16b, v0.16b
- 0x4e201c42, //and v2.16b, v2.16b, v0.16b
- 0x4e201c71, //and v17.16b, v3.16b, v0.16b
- 0x2d4e8c50, //ldp s16, s3, [x2, #116]
- 0x4e21d820, //scvtf v0.4s, v1.4s
- 0x4e21d841, //scvtf v1.4s, v2.4s
- 0x4e21da22, //scvtf v2.4s, v17.4s
- 0x4f909000, //fmul v0.4s, v0.4s, v16.s[0]
- 0xbd407c50, //ldr s16, [x2, #124]
- 0x4f839021, //fmul v1.4s, v1.4s, v3.s[0]
- 0x4d40c843, //ld1r {v3.4s}, [x2]
- 0x4f909042, //fmul v2.4s, v2.4s, v16.s[0]
+ 0x4e211c01, //and v1.16b, v0.16b, v1.16b
+ 0x4e221c02, //and v2.16b, v0.16b, v2.16b
+ 0x4e231c03, //and v3.16b, v0.16b, v3.16b
+ 0x4e040d00, //dup v0.4s, w8
+ 0x52a74048, //mov w8, #0x3a020000
+ 0x72810428, //movk w8, #0x821
+ 0x4e21d821, //scvtf v1.4s, v1.4s
+ 0x6e20dc20, //fmul v0.4s, v1.4s, v0.4s
+ 0x4e040d01, //dup v1.4s, w8
+ 0x52a7a088, //mov w8, #0x3d040000
+ 0x72842108, //movk w8, #0x2108
+ 0x4e21d842, //scvtf v2.4s, v2.4s
+ 0x6e21dc41, //fmul v1.4s, v2.4s, v1.4s
+ 0x4e040d02, //dup v2.4s, w8
+ 0x4e21d863, //scvtf v3.4s, v3.4s
+ 0x6e22dc62, //fmul v2.4s, v3.4s, v2.4s
+ 0x4f03f603, //fmov v3.4s, #1.000000000000000000e+00
0xd61f0060, //br x3
};
CODE const uint32_t sk_store_565_aarch64[] = {
- 0x2d504450, //ldp s16, s17, [x2, #128]
0xf9400028, //ldr x8, [x1]
- 0xd37ff809, //lsl x9, x0, #1
- 0x4f909012, //fmul v18.4s, v0.4s, v16.s[0]
- 0x4f919031, //fmul v17.4s, v1.4s, v17.s[0]
- 0x6e21aa52, //fcvtnu v18.4s, v18.4s
+ 0x52a84f8a, //mov w10, #0x427c0000
+ 0x4f01f7f0, //fmov v16.4s, #3.100000000000000000e+01
+ 0x4e040d52, //dup v18.4s, w10
+ 0x6e30dc11, //fmul v17.4s, v0.4s, v16.4s
+ 0x6e32dc32, //fmul v18.4s, v1.4s, v18.4s
0x6e21aa31, //fcvtnu v17.4s, v17.4s
+ 0x6e21aa52, //fcvtnu v18.4s, v18.4s
+ 0x6e30dc50, //fmul v16.4s, v2.4s, v16.4s
+ 0x4f2b5631, //shl v17.4s, v17.4s, #11
0xf9400108, //ldr x8, [x8]
- 0x4f909050, //fmul v16.4s, v2.4s, v16.s[0]
- 0x4f2b5652, //shl v18.4s, v18.4s, #11
- 0x4f255631, //shl v17.4s, v17.4s, #5
- 0x4eb21e31, //orr v17.16b, v17.16b, v18.16b
+ 0x4f255652, //shl v18.4s, v18.4s, #5
+ 0x4eb11e51, //orr v17.16b, v18.16b, v17.16b
0x6e21aa10, //fcvtnu v16.4s, v16.4s
0x4eb01e30, //orr v16.16b, v17.16b, v16.16b
+ 0xd37ff809, //lsl x9, x0, #1
0x0e612a10, //xtn v16.4h, v16.4s
0xfc296910, //str d16, [x8, x9]
0xf9400423, //ldr x3, [x1, #8]
@@ -788,14 +821,22 @@ CODE const uint32_t sk_mirror_y_aarch64[] = {
};
CODE const uint32_t sk_luminance_to_alpha_aarch64[] = {
- 0x2d510c50, //ldp s16, s3, [x2, #136]
- 0xbd409051, //ldr s17, [x2, #144]
+ 0x52a7cb28, //mov w8, #0x3e590000
+ 0x72967a08, //movk w8, #0xb3d0
+ 0x4e040d11, //dup v17.4s, w8
+ 0x52a7e6e8, //mov w8, #0x3f370000
+ 0x7282eb28, //movk w8, #0x1759
+ 0x4ea01c10, //mov v16.16b, v0.16b
+ 0x4e040d00, //dup v0.4s, w8
+ 0x52a7b268, //mov w8, #0x3d930000
0xf8408423, //ldr x3, [x1], #8
- 0x4f839023, //fmul v3.4s, v1.4s, v3.s[0]
- 0x4f901003, //fmla v3.4s, v0.4s, v16.s[0]
+ 0x729bb308, //movk w8, #0xdd98
+ 0x6e20dc23, //fmul v3.4s, v1.4s, v0.4s
+ 0x4e30ce23, //fmla v3.4s, v17.4s, v16.4s
+ 0x4e040d10, //dup v16.4s, w8
0x6f00e400, //movi v0.2d, #0x0
0x6f00e401, //movi v1.2d, #0x0
- 0x4f911043, //fmla v3.4s, v2.4s, v17.s[0]
+ 0x4e22ce03, //fmla v3.4s, v16.4s, v2.4s
0x6f00e402, //movi v2.2d, #0x0
0xd61f0060, //br x3
};
@@ -975,7 +1016,7 @@ CODE const uint32_t sk_seed_shader_vfp4[] = {
0xe8911008, //ldm r1, {r3, ip}
0xf3fb0620, //vcvt.f32.s32 d16, d16
0xf2c3161f, //vmov.i32 d17, #1056964608
- 0xedd23b05, //vldr d19, [r2, #20]
+ 0xedd23b00, //vldr d19, [r2]
0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
0xf2872f10, //vmov.f32 d2, #1
0xf3fb2622, //vcvt.f32.s32 d18, d18
@@ -1162,111 +1203,125 @@ CODE const uint32_t sk_unpremul_vfp4[] = {
};
CODE const uint32_t sk_from_srgb_vfp4[] = {
- 0xed2d8b02, //vpush {d8}
- 0xe282303c, //add r3, r2, #60
- 0xed928a10, //vldr s16, [r2, #64]
- 0xf3402d10, //vmul.f32 d18, d0, d0
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xe2823038, //add r3, r2, #56
- 0xf3413d11, //vmul.f32 d19, d1, d1
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe2823044, //add r3, r2, #68
- 0xf26141b1, //vorr d20, d17, d17
- 0xf26171b1, //vorr d23, d17, d17
- 0xf4e38c9f, //vld1.32 {d24[]}, [r3 :32]
+ 0xeddf3b20, //vldr d19, [pc, #128]
+ 0xf3408d10, //vmul.f32 d24, d0, d0
+ 0xeddf0b1c, //vldr d16, [pc, #112]
+ 0xf26341b3, //vorr d20, d19, d19
+ 0xf26351b3, //vorr d21, d19, d19
+ 0xeddf9b1f, //vldr d25, [pc, #124]
0xf2404c30, //vfma.f32 d20, d0, d16
- 0xe2823034, //add r3, r2, #52
- 0xf2417c30, //vfma.f32 d23, d1, d16
- 0xf2421c30, //vfma.f32 d17, d2, d16
- 0xf3425d12, //vmul.f32 d21, d2, d2
- 0xf2e16948, //vmul.f32 d22, d1, d8[0]
- 0xf2e00948, //vmul.f32 d16, d0, d8[0]
- 0xf2e29948, //vmul.f32 d25, d2, d8[0]
- 0xf3282e82, //vcgt.f32 d2, d24, d2
- 0xf3281e81, //vcgt.f32 d1, d24, d1
- 0xf3280e80, //vcgt.f32 d0, d24, d0
- 0xf4e38c9f, //vld1.32 {d24[]}, [r3 :32]
- 0xf268a1b8, //vorr d26, d24, d24
- 0xf242acb4, //vfma.f32 d26, d18, d20
- 0xf26821b8, //vorr d18, d24, d24
+ 0xeddf2b1b, //vldr d18, [pc, #108]
+ 0xf2415c30, //vfma.f32 d21, d1, d16
+ 0xeddfcb1d, //vldr d28, [pc, #116]
+ 0xf2423c30, //vfma.f32 d19, d2, d16
0xe4913004, //ldr r3, [r1], #4
- 0xf2432cb7, //vfma.f32 d18, d19, d23
- 0xf2458cb1, //vfma.f32 d24, d21, d17
- 0xf31001ba, //vbsl d0, d16, d26
- 0xf31611b2, //vbsl d1, d22, d18
- 0xf31921b8, //vbsl d2, d25, d24
- 0xecbd8b02, //vpop {d8}
+ 0xf3426d12, //vmul.f32 d22, d2, d2
+ 0xf3417d11, //vmul.f32 d23, d1, d1
+ 0xf3620e80, //vcgt.f32 d16, d18, d0
+ 0xf3621e81, //vcgt.f32 d17, d18, d1
+ 0xf341ad39, //vmul.f32 d26, d1, d25
+ 0xf342bd39, //vmul.f32 d27, d2, d25
+ 0xf3622e82, //vcgt.f32 d18, d18, d2
+ 0xf3409d39, //vmul.f32 d25, d0, d25
+ 0xf26cd1bc, //vorr d29, d28, d28
+ 0xf248dcb4, //vfma.f32 d29, d24, d20
+ 0xf26c41bc, //vorr d20, d28, d28
+ 0xf2474cb5, //vfma.f32 d20, d23, d21
+ 0xf246ccb3, //vfma.f32 d28, d22, d19
+ 0xf35901bd, //vbsl d16, d25, d29
+ 0xf35a11b4, //vbsl d17, d26, d20
+ 0xf35b21bc, //vbsl d18, d27, d28
+ 0xf22001b0, //vorr d0, d16, d16
+ 0xf22111b1, //vorr d1, d17, d17
+ 0xf22221b2, //vorr d2, d18, d18
0xe12fff13, //bx r3
+ 0x3e99999a, //.word 0x3e99999a
+ 0x3e99999a, //.word 0x3e99999a
+ 0x3f328f5c, //.word 0x3f328f5c
+ 0x3f328f5c, //.word 0x3f328f5c
+ 0x3d6147ae, //.word 0x3d6147ae
+ 0x3d6147ae, //.word 0x3d6147ae
+ 0x3d9e8391, //.word 0x3d9e8391
+ 0x3d9e8391, //.word 0x3d9e8391
+ 0x3b23d70a, //.word 0x3b23d70a
+ 0x3b23d70a, //.word 0x3b23d70a
};
CODE const uint32_t sk_to_srgb_vfp4[] = {
- 0xed2d8b02, //vpush {d8}
- 0xf3fb0580, //vrsqrte.f32 d16, d0
- 0xe2823050, //add r3, r2, #80
+ 0xf3fb0582, //vrsqrte.f32 d16, d2
+ 0xe4913004, //ldr r3, [r1], #4
0xf3fb1581, //vrsqrte.f32 d17, d1
- 0xed928a12, //vldr s16, [r2, #72]
- 0xf3fb2582, //vrsqrte.f32 d18, d2
+ 0xf3fb2580, //vrsqrte.f32 d18, d0
0xf3403db0, //vmul.f32 d19, d16, d16
0xf3414db1, //vmul.f32 d20, d17, d17
0xf3425db2, //vmul.f32 d21, d18, d18
- 0xf2603f33, //vrsqrts.f32 d19, d0, d19
+ 0xf2623f33, //vrsqrts.f32 d19, d2, d19
0xf2614f34, //vrsqrts.f32 d20, d1, d20
- 0xf2625f35, //vrsqrts.f32 d21, d2, d21
+ 0xf2605f35, //vrsqrts.f32 d21, d0, d21
0xf3400db3, //vmul.f32 d16, d16, d19
0xf3411db4, //vmul.f32 d17, d17, d20
0xf3422db5, //vmul.f32 d18, d18, d21
0xf3fb3520, //vrecpe.f32 d19, d16
0xf3fb4521, //vrecpe.f32 d20, d17
0xf3fb6522, //vrecpe.f32 d22, d18
- 0xf3fb55a2, //vrsqrte.f32 d21, d18
- 0xf3fb75a0, //vrsqrte.f32 d23, d16
- 0xf3fb85a1, //vrsqrte.f32 d24, d17
+ 0xf3fb55a0, //vrsqrte.f32 d21, d16
+ 0xf3fb75a1, //vrsqrte.f32 d23, d17
+ 0xf3fb85a2, //vrsqrte.f32 d24, d18
0xf2409fb3, //vrecps.f32 d25, d16, d19
0xf241afb4, //vrecps.f32 d26, d17, d20
0xf242bfb6, //vrecps.f32 d27, d18, d22
0xf345cdb5, //vmul.f32 d28, d21, d21
0xf347ddb7, //vmul.f32 d29, d23, d23
0xf348edb8, //vmul.f32 d30, d24, d24
- 0xf2622fbc, //vrsqrts.f32 d18, d18, d28
- 0xf2600fbd, //vrsqrts.f32 d16, d16, d29
- 0xf2611fbe, //vrsqrts.f32 d17, d17, d30
+ 0xf2600fbc, //vrsqrts.f32 d16, d16, d28
+ 0xf2611fbd, //vrsqrts.f32 d17, d17, d29
+ 0xf2622fbe, //vrsqrts.f32 d18, d18, d30
0xf3433db9, //vmul.f32 d19, d19, d25
- 0xf4e39c9f, //vld1.32 {d25[]}, [r3 :32]
- 0xe2823054, //add r3, r2, #84
+ 0xeddf9b21, //vldr d25, [pc, #132]
0xf3444dba, //vmul.f32 d20, d20, d26
+ 0xeddfab21, //vldr d26, [pc, #132]
0xf3466dbb, //vmul.f32 d22, d22, d27
- 0xf4e3ac9f, //vld1.32 {d26[]}, [r3 :32]
- 0xe282304c, //add r3, r2, #76
0xf26ab1ba, //vorr d27, d26, d26
- 0xf249bcb3, //vfma.f32 d27, d25, d19
+ 0xf243bcb9, //vfma.f32 d27, d19, d25
0xf26a31ba, //vorr d19, d26, d26
- 0xf2493cb4, //vfma.f32 d19, d25, d20
- 0xf4e34c9f, //vld1.32 {d20[]}, [r3 :32]
- 0xf249acb6, //vfma.f32 d26, d25, d22
- 0xe2823058, //add r3, r2, #88
- 0xf3452db2, //vmul.f32 d18, d21, d18
- 0xf3470db0, //vmul.f32 d16, d23, d16
- 0xf3481db1, //vmul.f32 d17, d24, d17
- 0xf2e05948, //vmul.f32 d21, d0, d8[0]
- 0xf244bcb0, //vfma.f32 d27, d20, d16
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xf2443cb1, //vfma.f32 d19, d20, d17
- 0xf244acb2, //vfma.f32 d26, d20, d18
- 0xf4e24c9f, //vld1.32 {d20[]}, [r2 :32]
- 0xf2e11948, //vmul.f32 d17, d1, d8[0]
- 0xf2e22948, //vmul.f32 d18, d2, d8[0]
- 0xf3201e81, //vcgt.f32 d1, d16, d1
- 0xe4913004, //ldr r3, [r1], #4
- 0xf3200e80, //vcgt.f32 d0, d16, d0
- 0xf3202e82, //vcgt.f32 d2, d16, d2
- 0xf2640fab, //vmin.f32 d16, d20, d27
+ 0xf2443cb9, //vfma.f32 d19, d20, d25
+ 0xeddf4b1d, //vldr d20, [pc, #116]
+ 0xf246acb9, //vfma.f32 d26, d22, d25
+ 0xf3450db0, //vmul.f32 d16, d21, d16
+ 0xeddf5b1c, //vldr d21, [pc, #112]
+ 0xf3471db1, //vmul.f32 d17, d23, d17
+ 0xf3482db2, //vmul.f32 d18, d24, d18
+ 0xf3406d35, //vmul.f32 d22, d0, d21
+ 0xf240bcb4, //vfma.f32 d27, d16, d20
+ 0xf2413cb4, //vfma.f32 d19, d17, d20
+ 0xf242acb4, //vfma.f32 d26, d18, d20
+ 0xeddf2b17, //vldr d18, [pc, #92]
+ 0xf3417d35, //vmul.f32 d23, d1, d21
+ 0xf3620e80, //vcgt.f32 d16, d18, d0
+ 0xf3621e81, //vcgt.f32 d17, d18, d1
+ 0xf3622e82, //vcgt.f32 d18, d18, d2
+ 0xf3425d35, //vmul.f32 d21, d2, d21
+ 0xf2c74f10, //vmov.f32 d20, #1
+ 0xf2648faa, //vmin.f32 d24, d20, d26
0xf2643fa3, //vmin.f32 d19, d20, d19
- 0xf2644faa, //vmin.f32 d20, d20, d26
- 0xf31501b0, //vbsl d0, d21, d16
- 0xf31111b3, //vbsl d1, d17, d19
- 0xf31221b4, //vbsl d2, d18, d20
- 0xecbd8b02, //vpop {d8}
+ 0xf2644fab, //vmin.f32 d20, d20, d27
+ 0xf35601b8, //vbsl d16, d22, d24
+ 0xf35711b3, //vbsl d17, d23, d19
+ 0xf35521b4, //vbsl d18, d21, d20
+ 0xf22001b0, //vorr d0, d16, d16
+ 0xf22111b1, //vorr d1, d17, d17
+ 0xf22221b2, //vorr d2, d18, d18
0xe12fff13, //bx r3
+ 0x3f306fce, //.word 0x3f306fce
+ 0x3f306fce, //.word 0x3f306fce
+ 0xbdca57a8, //.word 0xbdca57a8
+ 0xbdca57a8, //.word 0xbdca57a8
+ 0x3ed287c2, //.word 0x3ed287c2
+ 0x3ed287c2, //.word 0x3ed287c2
+ 0x41475c29, //.word 0x41475c29
+ 0x41475c29, //.word 0x41475c29
+ 0x3b8ce704, //.word 0x3b8ce704
+ 0x3b8ce704, //.word 0x3b8ce704
};
CODE const uint32_t sk_scale_1_float_vfp4[] = {
@@ -1360,191 +1415,205 @@ CODE const uint32_t sk_lerp_u8_vfp4[] = {
};
CODE const uint32_t sk_lerp_565_vfp4[] = {
- 0xed2d8b04, //vpush {d8-d9}
- 0xe24dd008, //sub sp, sp, #8
+ 0xe24dd004, //sub sp, sp, #4
0xe8911008, //ldm r1, {r3, ip}
+ 0xf3c72218, //vmov.i32 d18, #63488
+ 0xf2c1101f, //vmov.i32 d17, #31
0xf2603d04, //vsub.f32 d19, d0, d4
- 0xf2240114, //vorr d0, d4, d4
0xe2811008, //add r1, r1, #8
0xe5933000, //ldr r3, [r3]
+ 0xf2616d05, //vsub.f32 d22, d1, d5
+ 0xf2240114, //vorr d0, d4, d4
+ 0xf2251115, //vorr d1, d5, d5
0xe7933080, //ldr r3, [r3, r0, lsl #1]
- 0xe58d3004, //str r3, [sp, #4]
- 0xe28d3004, //add r3, sp, #4
- 0xed923a1d, //vldr s6, [r2, #116]
+ 0xf2873f10, //vmov.f32 d3, #1
+ 0xe58d3000, //str r3, [sp]
+ 0xe1a0300d, //mov r3, sp
0xf4e3083f, //vld1.32 {d16[0]}, [r3 :32]
- 0xe282306c, //add r3, r2, #108
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe2823068, //add r3, r2, #104
+ 0xe3a03e7e, //mov r3, #2016
0xf3d04a30, //vmovl.u16 q10, d16
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xe2823070, //add r3, r2, #112
- 0xf24201b4, //vand d16, d18, d20
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf24221b4, //vand d18, d18, d20
- 0xf24111b4, //vand d17, d17, d20
+ 0xee803b90, //vdup.32 d16, r3
+ 0xf24421b2, //vand d18, d20, d18
+ 0xf24411b1, //vand d17, d20, d17
+ 0xeddf5b12, //vldr d21, [pc, #72]
+ 0xf24401b0, //vand d16, d20, d16
+ 0xeddf4b0e, //vldr d20, [pc, #56]
+ 0xf3fb2622, //vcvt.f32.s32 d18, d18
0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xed928a1e, //vldr s16, [r2, #120]
0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xed929a1f, //vldr s18, [r2, #124]
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xf2614d05, //vsub.f32 d20, d1, d5
- 0xf2e009c3, //vmul.f32 d16, d16, d3[0]
+ 0xf3422db4, //vmul.f32 d18, d18, d20
+ 0xeddf4b0d, //vldr d20, [pc, #52]
+ 0xf3400db5, //vmul.f32 d16, d16, d21
0xf2625d06, //vsub.f32 d21, d2, d6
- 0xf2e119c8, //vmul.f32 d17, d17, d8[0]
- 0xf2e229c9, //vmul.f32 d18, d18, d9[0]
- 0xf2251115, //vorr d1, d5, d5
+ 0xf3411db4, //vmul.f32 d17, d17, d20
0xf2262116, //vorr d2, d6, d6
- 0xf2030cb0, //vfma.f32 d0, d19, d16
- 0xf2041cb1, //vfma.f32 d1, d20, d17
- 0xf2052cb2, //vfma.f32 d2, d21, d18
- 0xf2873f10, //vmov.f32 d3, #1
- 0xe28dd008, //add sp, sp, #8
- 0xecbd8b04, //vpop {d8-d9}
+ 0xf2030cb2, //vfma.f32 d0, d19, d18
+ 0xf2061cb0, //vfma.f32 d1, d22, d16
+ 0xf2052cb1, //vfma.f32 d2, d21, d17
+ 0xe28dd004, //add sp, sp, #4
0xe12fff1c, //bx ip
+ 0xe320f000, //nop {0}
+ 0x37842108, //.word 0x37842108
+ 0x37842108, //.word 0x37842108
+ 0x3a020821, //.word 0x3a020821
+ 0x3a020821, //.word 0x3a020821
+ 0x3d042108, //.word 0x3d042108
+ 0x3d042108, //.word 0x3d042108
};
CODE const uint32_t sk_load_tables_vfp4[] = {
0xe92d48f0, //push {r4, r5, r6, r7, fp, lr}
0xe8911008, //ldm r1, {r3, ip}
- 0xe2826010, //add r6, r2, #16
+ 0xf3c7001f, //vmov.i32 d16, #255
0xe2811008, //add r1, r1, #8
0xe593e000, //ldr lr, [r3]
0xe99300b0, //ldmib r3, {r4, r5, r7}
- 0xf4e60c9f, //vld1.32 {d16[]}, [r6 :32]
- 0xe08e6100, //add r6, lr, r0, lsl #2
- 0xedd61b00, //vldr d17, [r6]
- 0xf24021b1, //vand d18, d16, d17
- 0xed922a03, //vldr s4, [r2, #12]
- 0xf3f03031, //vshr.u32 d19, d17, #16
- 0xee326b90, //vmov.32 r6, d18[1]
- 0xe0846106, //add r6, r4, r6, lsl #2
- 0xedd60a00, //vldr s1, [r6]
+ 0xe08e3100, //add r3, lr, r0, lsl #2
+ 0xedd31b00, //vldr d17, [r3]
+ 0xf24121b0, //vand d18, d17, d16
+ 0xf3f83031, //vshr.u32 d19, d17, #8
+ 0xee323b90, //vmov.32 r3, d18[1]
0xee126b90, //vmov.32 r6, d18[0]
- 0xf3f82031, //vshr.u32 d18, d17, #8
- 0xf24021b2, //vand d18, d16, d18
- 0xf24001b3, //vand d16, d16, d19
- 0xee103b90, //vmov.32 r3, d16[0]
- 0xe0846106, //add r6, r4, r6, lsl #2
- 0xee304b90, //vmov.32 r4, d16[1]
- 0xf3e80031, //vshr.u32 d16, d17, #24
- 0xed960a00, //vldr s0, [r6]
+ 0xf3f02031, //vshr.u32 d18, d17, #16
+ 0xf24221b0, //vand d18, d18, d16
+ 0xf24301b0, //vand d16, d19, d16
+ 0xe0843103, //add r3, r4, r3, lsl #2
+ 0xedd30a00, //vldr s1, [r3]
+ 0xe0843106, //add r3, r4, r6, lsl #2
0xee326b90, //vmov.32 r6, d18[1]
+ 0xed930a00, //vldr s0, [r3]
+ 0xee303b90, //vmov.32 r3, d16[1]
+ 0xee104b90, //vmov.32 r4, d16[0]
+ 0xf3e80031, //vshr.u32 d16, d17, #24
+ 0xeddf1b0d, //vldr d17, [pc, #52]
0xf3fb0620, //vcvt.f32.s32 d16, d16
- 0xe0873103, //add r3, r7, r3, lsl #2
- 0xf2a039c2, //vmul.f32 d3, d16, d2[0]
- 0xe0874104, //add r4, r7, r4, lsl #2
- 0xedd42a00, //vldr s5, [r4]
- 0xe0856106, //add r6, r5, r6, lsl #2
- 0xed932a00, //vldr s4, [r3]
- 0xedd61a00, //vldr s3, [r6]
+ 0xf3003db1, //vmul.f32 d3, d16, d17
+ 0xe087e106, //add lr, r7, r6, lsl #2
0xee126b90, //vmov.32 r6, d18[0]
- 0xe0856106, //add r6, r5, r6, lsl #2
- 0xed961a00, //vldr s2, [r6]
+ 0xe0853103, //add r3, r5, r3, lsl #2
+ 0xedde2a00, //vldr s5, [lr]
+ 0xedd31a00, //vldr s3, [r3]
+ 0xe0853104, //add r3, r5, r4, lsl #2
+ 0xed931a00, //vldr s2, [r3]
+ 0xe0873106, //add r3, r7, r6, lsl #2
+ 0xed932a00, //vldr s4, [r3]
0xe8bd48f0, //pop {r4, r5, r6, r7, fp, lr}
0xe12fff1c, //bx ip
+ 0xe320f000, //nop {0}
+ 0x3b808081, //.word 0x3b808081
+ 0x3b808081, //.word 0x3b808081
};
CODE const uint32_t sk_load_a8_vfp4[] = {
0xe24dd004, //sub sp, sp, #4
0xe8911008, //ldm r1, {r3, ip}
0xe2811008, //add r1, r1, #8
+ 0xf2800010, //vmov.i32 d0, #0
0xf2801010, //vmov.i32 d1, #0
- 0xf2802010, //vmov.i32 d2, #0
0xe5933000, //ldr r3, [r3]
+ 0xf2802010, //vmov.i32 d2, #0
0xe0833000, //add r3, r3, r0
0xe1d330b0, //ldrh r3, [r3]
0xe1cd30b0, //strh r3, [sp]
0xe1a0300d, //mov r3, sp
0xf4e3041f, //vld1.16 {d16[0]}, [r3 :16]
- 0xed920a03, //vldr s0, [r2, #12]
0xf3c80a30, //vmovl.u8 q8, d16
0xf3d00a30, //vmovl.u16 q8, d16
0xf3fb06a0, //vcvt.f32.u32 d16, d16
- 0xf2a039c0, //vmul.f32 d3, d16, d0[0]
- 0xf2800010, //vmov.i32 d0, #0
+ 0xeddf1b03, //vldr d17, [pc, #12]
+ 0xf3003db1, //vmul.f32 d3, d16, d17
0xe28dd004, //add sp, sp, #4
0xe12fff1c, //bx ip
+ 0xe320f000, //nop {0}
+ 0x3b808081, //.word 0x3b808081
+ 0x3b808081, //.word 0x3b808081
};
CODE const uint32_t sk_store_a8_vfp4[] = {
0xe92d4800, //push {fp, lr}
- 0xe2823008, //add r3, r2, #8
- 0xf2c3061f, //vmov.i32 d16, #1056964608
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
+ 0xeddf0b0d, //vldr d16, [pc, #52]
+ 0xf2c3161f, //vmov.i32 d17, #1056964608
+ 0xf2431c30, //vfma.f32 d17, d3, d16
0xe5913000, //ldr r3, [r1]
- 0xf2430c31, //vfma.f32 d16, d3, d17
0xe5933000, //ldr r3, [r3]
- 0xf3fb07a0, //vcvt.u32.f32 d16, d16
+ 0xf3fb07a1, //vcvt.u32.f32 d16, d17
0xee10eb90, //vmov.32 lr, d16[0]
0xee30cb90, //vmov.32 ip, d16[1]
0xe7e3e000, //strb lr, [r3, r0]!
0xe5c3c001, //strb ip, [r3, #1]
- 0xe5913004, //ldr r3, [r1, #4]
- 0xe2811008, //add r1, r1, #8
+ 0xe2813008, //add r3, r1, #8
+ 0xe591c004, //ldr ip, [r1, #4]
+ 0xe1a01003, //mov r1, r3
0xe8bd4800, //pop {fp, lr}
- 0xe12fff13, //bx r3
+ 0xe12fff1c, //bx ip
+ 0x437f0000, //.word 0x437f0000
+ 0x437f0000, //.word 0x437f0000
};
CODE const uint32_t sk_load_565_vfp4[] = {
0xe24dd004, //sub sp, sp, #4
0xe8911008, //ldm r1, {r3, ip}
+ 0xf2c1101f, //vmov.i32 d17, #31
+ 0xf3c72218, //vmov.i32 d18, #63488
+ 0xeddf3b16, //vldr d19, [pc, #88]
0xe2811008, //add r1, r1, #8
0xe5933000, //ldr r3, [r3]
+ 0xf2873f10, //vmov.f32 d3, #1
0xe7933080, //ldr r3, [r3, r0, lsl #1]
0xe58d3000, //str r3, [sp]
0xe1a0300d, //mov r3, sp
0xf4e3083f, //vld1.32 {d16[0]}, [r3 :32]
- 0xe282306c, //add r3, r2, #108
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe2823068, //add r3, r2, #104
+ 0xe3a03e7e, //mov r3, #2016
0xf3d04a30, //vmovl.u16 q10, d16
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xe2823070, //add r3, r2, #112
- 0xf24201b4, //vand d16, d18, d20
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf24111b4, //vand d17, d17, d20
- 0xf24221b4, //vand d18, d18, d20
- 0xf4a23c9f, //vld1.32 {d3[]}, [r2 :32]
+ 0xee803b90, //vdup.32 d16, r3
+ 0xf24411b1, //vand d17, d20, d17
+ 0xeddf5b0e, //vldr d21, [pc, #56]
+ 0xf24421b2, //vand d18, d20, d18
+ 0xf24401b0, //vand d16, d20, d16
+ 0xeddf4b09, //vldr d20, [pc, #36]
+ 0xf3fb2622, //vcvt.f32.s32 d18, d18
0xf3fb0620, //vcvt.f32.s32 d16, d16
0xf3fb1621, //vcvt.f32.s32 d17, d17
- 0xf3fb2622, //vcvt.f32.s32 d18, d18
- 0xed920a1d, //vldr s0, [r2, #116]
- 0xed921a1e, //vldr s2, [r2, #120]
- 0xed922a1f, //vldr s4, [r2, #124]
- 0xf2a009c0, //vmul.f32 d0, d16, d0[0]
- 0xf2a119c1, //vmul.f32 d1, d17, d1[0]
- 0xf2a229c2, //vmul.f32 d2, d18, d2[0]
+ 0xf3020db3, //vmul.f32 d0, d18, d19
+ 0xf3001db4, //vmul.f32 d1, d16, d20
+ 0xf3012db5, //vmul.f32 d2, d17, d21
0xe28dd004, //add sp, sp, #4
0xe12fff1c, //bx ip
+ 0x37842108, //.word 0x37842108
+ 0x37842108, //.word 0x37842108
+ 0x3a020821, //.word 0x3a020821
+ 0x3a020821, //.word 0x3a020821
+ 0x3d042108, //.word 0x3d042108
+ 0x3d042108, //.word 0x3d042108
};
CODE const uint32_t sk_store_565_vfp4[] = {
- 0xe2823080, //add r3, r2, #128
+ 0xf2c30f1f, //vmov.f32 d16, #31
+ 0xeddf1b15, //vldr d17, [pc, #84]
0xf2c3361f, //vmov.i32 d19, #1056964608
- 0xf2c3461f, //vmov.i32 d20, #1056964608
- 0xf4e31c9f, //vld1.32 {d17[]}, [r3 :32]
- 0xe2823084, //add r3, r2, #132
- 0xf2403c31, //vfma.f32 d19, d0, d17
- 0xf4e32c9f, //vld1.32 {d18[]}, [r3 :32]
- 0xf2c3061f, //vmov.i32 d16, #1056964608
- 0xf2414c32, //vfma.f32 d20, d1, d18
- 0xf2420c31, //vfma.f32 d16, d2, d17
0xe5913000, //ldr r3, [r1]
+ 0xf2413c31, //vfma.f32 d19, d1, d17
+ 0xf2c3161f, //vmov.i32 d17, #1056964608
+ 0xf2401c30, //vfma.f32 d17, d0, d16
0xe5933000, //ldr r3, [r3]
- 0xf3fb17a3, //vcvt.u32.f32 d17, d19
+ 0xf2c3261f, //vmov.i32 d18, #1056964608
+ 0xf2422c30, //vfma.f32 d18, d2, d16
0xe0833080, //add r3, r3, r0, lsl #1
- 0xf3fb27a4, //vcvt.u32.f32 d18, d20
- 0xf3fb07a0, //vcvt.u32.f32 d16, d16
+ 0xf3fb07a3, //vcvt.u32.f32 d16, d19
+ 0xf3fb17a1, //vcvt.u32.f32 d17, d17
+ 0xf3fb27a2, //vcvt.u32.f32 d18, d18
+ 0xf2e50530, //vshl.s32 d16, d16, #5
0xf2eb1531, //vshl.s32 d17, d17, #11
- 0xf2e52532, //vshl.s32 d18, d18, #5
- 0xf26101b0, //vorr d16, d17, d16
+ 0xf26001b1, //vorr d16, d16, d17
0xf26001b2, //vorr d16, d16, d18
0xf3f60121, //vuzp.16 d16, d17
0xf4c3080f, //vst1.32 {d16[0]}, [r3]
- 0xe5913004, //ldr r3, [r1, #4]
- 0xe2811008, //add r1, r1, #8
- 0xe12fff13, //bx r3
- 0xe320f000, //nop {0}
+ 0xe2813008, //add r3, r1, #8
+ 0xe591c004, //ldr ip, [r1, #4]
+ 0xe1a01003, //mov r1, r3
+ 0xe12fff1c, //bx ip
+ 0x427c0000, //.word 0x427c0000
+ 0x427c0000, //.word 0x427c0000
};
CODE const uint32_t sk_load_8888_vfp4[] = {
@@ -1799,21 +1868,24 @@ CODE const uint32_t sk_mirror_y_vfp4[] = {
};
CODE const uint32_t sk_luminance_to_alpha_vfp4[] = {
- 0xed2d8b02, //vpush {d8}
- 0xed923a22, //vldr s6, [r2, #136]
- 0xe2823090, //add r3, r2, #144
- 0xed928a23, //vldr s16, [r2, #140]
- 0xf2e01943, //vmul.f32 d17, d0, d3[0]
- 0xf2e10948, //vmul.f32 d16, d1, d8[0]
+ 0xeddf0b0a, //vldr d16, [pc, #40]
+ 0xeddf1b0b, //vldr d17, [pc, #44]
+ 0xf3410d30, //vmul.f32 d16, d1, d16
+ 0xe4913004, //ldr r3, [r1], #4
+ 0xf3401d31, //vmul.f32 d17, d0, d17
0xf2800010, //vmov.i32 d0, #0
0xf2801010, //vmov.i32 d1, #0
0xf2013da0, //vadd.f32 d3, d17, d16
- 0xf4e30c9f, //vld1.32 {d16[]}, [r3 :32]
- 0xf2003c92, //vfma.f32 d3, d16, d2
- 0xe4913004, //ldr r3, [r1], #4
+ 0xeddf0b06, //vldr d16, [pc, #24]
+ 0xf2023c30, //vfma.f32 d3, d2, d16
0xf2802010, //vmov.i32 d2, #0
- 0xecbd8b02, //vpop {d8}
0xe12fff13, //bx r3
+ 0x3f371759, //.word 0x3f371759
+ 0x3f371759, //.word 0x3f371759
+ 0x3e59b3d0, //.word 0x3e59b3d0
+ 0x3e59b3d0, //.word 0x3e59b3d0
+ 0x3d93dd98, //.word 0x3d93dd98
+ 0x3d93dd98, //.word 0x3d93dd98
};
CODE const uint32_t sk_matrix_2x3_vfp4[] = {
@@ -2085,7 +2157,7 @@ CODE const uint8_t sk_seed_shader_hsw[] = {
196,193,121,110,200, //vmovd %r8d,%xmm1
196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,66,20, //vaddps 0x14(%rdx),%ymm0,%ymm0
+ 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
196,226,125,24,16, //vbroadcastss (%rax),%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
@@ -2265,16 +2337,26 @@ CODE const uint8_t sk_unpremul_hsw[] = {
};
CODE const uint8_t sk_from_srgb_hsw[] = {
- 196,98,125,24,66,64, //vbroadcastss 0x40(%rdx),%ymm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 196,98,125,24,90,60, //vbroadcastss 0x3c(%rdx),%ymm11
- 196,98,125,24,98,56, //vbroadcastss 0x38(%rdx),%ymm12
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,66,125,24,219, //vbroadcastss %xmm11,%ymm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,66,125,24,228, //vbroadcastss %xmm12,%ymm12
196,65,124,40,235, //vmovaps %ymm11,%ymm13
196,66,125,168,236, //vfmadd213ps %ymm12,%ymm0,%ymm13
- 196,98,125,24,114,52, //vbroadcastss 0x34(%rdx),%ymm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,66,125,24,246, //vbroadcastss %xmm14,%ymm14
196,66,45,168,238, //vfmadd213ps %ymm14,%ymm10,%ymm13
- 196,98,125,24,82,68, //vbroadcastss 0x44(%rdx),%ymm10
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
196,195,21,74,193,0, //vblendvps %ymm0,%ymm9,%ymm13,%ymm0
197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
@@ -2296,37 +2378,50 @@ CODE const uint8_t sk_from_srgb_hsw[] = {
CODE const uint8_t sk_to_srgb_hsw[] = {
197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,200, //vrcpps %ymm8,%ymm9
- 196,65,124,82,208, //vrsqrtps %ymm8,%ymm10
- 196,98,125,24,66,72, //vbroadcastss 0x48(%rdx),%ymm8
- 197,60,89,216, //vmulps %ymm0,%ymm8,%ymm11
- 196,98,125,24,34, //vbroadcastss (%rdx),%ymm12
- 196,98,125,24,106,76, //vbroadcastss 0x4c(%rdx),%ymm13
- 196,98,125,24,114,80, //vbroadcastss 0x50(%rdx),%ymm14
- 196,98,125,24,122,84, //vbroadcastss 0x54(%rdx),%ymm15
- 196,66,13,168,207, //vfmadd213ps %ymm15,%ymm14,%ymm9
- 196,66,21,184,202, //vfmadd231ps %ymm10,%ymm13,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,98,125,24,82,88, //vbroadcastss 0x58(%rdx),%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,53,74,195,0, //vblendvps %ymm0,%ymm11,%ymm9,%ymm0
- 197,124,82,201, //vrsqrtps %ymm1,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,66,21,184,217, //vfmadd231ps %ymm9,%ymm13,%ymm11
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 196,65,28,93,219, //vminps %ymm11,%ymm12,%ymm11
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,37,74,201,16, //vblendvps %ymm1,%ymm9,%ymm11,%ymm1
- 197,124,82,202, //vrsqrtps %ymm2,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
+ 196,65,124,83,216, //vrcpps %ymm8,%ymm11
+ 196,65,124,82,224, //vrsqrtps %ymm8,%ymm12
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
+ 197,60,89,232, //vmulps %ymm0,%ymm8,%ymm13
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,121,110,200, //vmovd %eax,%xmm9
+ 196,66,125,24,201, //vbroadcastss %xmm9,%ymm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,66,125,24,246, //vbroadcastss %xmm14,%ymm14
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 197,121,110,248, //vmovd %eax,%xmm15
+ 196,66,125,24,255, //vbroadcastss %xmm15,%ymm15
196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,66,21,184,217, //vfmadd231ps %ymm9,%ymm13,%ymm11
- 196,65,28,93,203, //vminps %ymm11,%ymm12,%ymm9
+ 196,66,45,184,220, //vfmadd231ps %ymm12,%ymm10,%ymm11
+ 196,65,52,93,219, //vminps %ymm11,%ymm9,%ymm11
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,66,125,24,228, //vbroadcastss %xmm12,%ymm12
+ 196,193,124,194,196,1, //vcmpltps %ymm12,%ymm0,%ymm0
+ 196,195,37,74,197,0, //vblendvps %ymm0,%ymm13,%ymm11,%ymm0
+ 197,124,82,217, //vrsqrtps %ymm1,%ymm11
+ 196,65,124,83,235, //vrcpps %ymm11,%ymm13
+ 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
+ 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
+ 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
+ 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
+ 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
+ 196,193,116,194,204,1, //vcmpltps %ymm12,%ymm1,%ymm1
+ 196,195,21,74,203,16, //vblendvps %ymm1,%ymm11,%ymm13,%ymm1
+ 197,124,82,218, //vrsqrtps %ymm2,%ymm11
+ 196,65,124,83,235, //vrcpps %ymm11,%ymm13
+ 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
+ 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
+ 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
+ 196,65,52,93,205, //vminps %ymm13,%ymm9,%ymm9
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
+ 196,193,108,194,212,1, //vcmpltps %ymm12,%ymm2,%ymm2
196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -2349,7 +2444,7 @@ CODE const uint8_t sk_scale_u8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,56, //jne 462 <_sk_scale_u8_hsw+0x48>
+ 117,56, //jne 4bf <_sk_scale_u8_hsw+0x48>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
@@ -2373,9 +2468,9 @@ CODE const uint8_t sk_scale_u8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 46a <_sk_scale_u8_hsw+0x50>
+ 117,234, //jne 4c7 <_sk_scale_u8_hsw+0x50>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,167, //jmp 42e <_sk_scale_u8_hsw+0x14>
+ 235,167, //jmp 48b <_sk_scale_u8_hsw+0x14>
};
CODE const uint8_t sk_lerp_1_float_hsw[] = {
@@ -2399,7 +2494,7 @@ CODE const uint8_t sk_lerp_u8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,76, //jne 512 <_sk_lerp_u8_hsw+0x5c>
+ 117,76, //jne 56f <_sk_lerp_u8_hsw+0x5c>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
@@ -2427,37 +2522,49 @@ CODE const uint8_t sk_lerp_u8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 51a <_sk_lerp_u8_hsw+0x64>
+ 117,234, //jne 577 <_sk_lerp_u8_hsw+0x64>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,147, //jmp 4ca <_sk_lerp_u8_hsw+0x14>
+ 235,147, //jmp 527 <_sk_lerp_u8_hsw+0x14>
};
CODE const uint8_t sk_lerp_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,132,0,0,0, //jne 5c9 <_sk_lerp_565_hsw+0x92>
+ 15,133,179,0,0,0, //jne 655 <_sk_lerp_565_hsw+0xc1>
196,193,122,111,28,122, //vmovdqu (%r10,%rdi,2),%xmm3
- 196,226,125,51,219, //vpmovzxwd %xmm3,%ymm3
- 196,98,125,88,66,104, //vpbroadcastd 0x68(%rdx),%ymm8
- 197,61,219,195, //vpand %ymm3,%ymm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 196,98,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm9
- 196,65,52,89,192, //vmulps %ymm8,%ymm9,%ymm8
- 196,98,125,88,74,108, //vpbroadcastd 0x6c(%rdx),%ymm9
- 197,53,219,203, //vpand %ymm3,%ymm9,%ymm9
- 196,65,124,91,201, //vcvtdq2ps %ymm9,%ymm9
- 196,98,125,24,82,120, //vbroadcastss 0x78(%rdx),%ymm10
- 196,65,44,89,201, //vmulps %ymm9,%ymm10,%ymm9
- 196,98,125,88,82,112, //vpbroadcastd 0x70(%rdx),%ymm10
- 197,173,219,219, //vpand %ymm3,%ymm10,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,82,124, //vbroadcastss 0x7c(%rdx),%ymm10
- 197,172,89,219, //vmulps %ymm3,%ymm10,%ymm3
+ 196,98,125,51,195, //vpmovzxwd %xmm3,%ymm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
+ 196,226,53,168,196, //vfmadd213ps %ymm4,%ymm9,%ymm0
197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,53,168,205, //vfmadd213ps %ymm5,%ymm9,%ymm1
+ 196,226,45,168,205, //vfmadd213ps %ymm5,%ymm10,%ymm1
197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
196,226,101,168,214, //vfmadd213ps %ymm6,%ymm3,%ymm2
184,0,0,128,63, //mov $0x3f800000,%eax
@@ -2471,8 +2578,8 @@ CODE const uint8_t sk_lerp_565_hsw[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,102,255,255,255, //ja 54b <_sk_lerp_565_hsw+0x14>
- 76,141,13,76,0,0,0, //lea 0x4c(%rip),%r9 # 638 <_sk_lerp_565_hsw+0x101>
+ 15,135,55,255,255,255, //ja 5a8 <_sk_lerp_565_hsw+0x14>
+ 76,141,13,76,0,0,0, //lea 0x4c(%rip),%r9 # 6c4 <_sk_lerp_565_hsw+0x130>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -2484,13 +2591,13 @@ CODE const uint8_t sk_lerp_565_hsw[] = {
196,193,97,196,92,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
196,193,97,196,92,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
196,193,97,196,28,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
- 233,22,255,255,255, //jmpq 54b <_sk_lerp_565_hsw+0x14>
+ 233,231,254,255,255, //jmpq 5a8 <_sk_lerp_565_hsw+0x14>
15,31,0, //nopl (%rax)
241, //icebp
255, //(bad)
255, //(bad)
255, //(bad)
- 233,255,255,255,225, //jmpq ffffffffe2000640 <_sk_linear_gradient_2stops_hsw+0xffffffffe1fff55e>
+ 233,255,255,255,225, //jmpq ffffffffe20006cc <_sk_linear_gradient_2stops_hsw+0xffffffffe1fff4e7>
255, //(bad)
255, //(bad)
255, //(bad)
@@ -2515,9 +2622,11 @@ CODE const uint8_t sk_load_tables_hsw[] = {
76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
76,3,8, //add (%rax),%r9
77,133,192, //test %r8,%r8
- 117,106, //jne 6d3 <_sk_load_tables_hsw+0x7f>
+ 117,121, //jne 76e <_sk_load_tables_hsw+0x8e>
196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 196,226,125,88,82,16, //vpbroadcastd 0x10(%rdx),%ymm2
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 197,249,110,193, //vmovd %ecx,%xmm0
+ 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
197,237,219,203, //vpand %ymm3,%ymm2,%ymm1
196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
72,139,72,8, //mov 0x8(%rax),%rcx
@@ -2533,9 +2642,11 @@ CODE const uint8_t sk_load_tables_hsw[] = {
196,65,109,219,201, //vpand %ymm9,%ymm2,%ymm9
196,162,61,146,20,136, //vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,66,12, //vbroadcastss 0xc(%rdx),%ymm8
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
72,173, //lods %ds:(%rsi),%rax
76,137,193, //mov %r8,%rcx
255,224, //jmpq *%rax
@@ -2547,7 +2658,7 @@ CODE const uint8_t sk_load_tables_hsw[] = {
196,193,249,110,194, //vmovq %r10,%xmm0
196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,114,255,255,255, //jmpq 66e <_sk_load_tables_hsw+0x1a>
+ 233,99,255,255,255, //jmpq 6fa <_sk_load_tables_hsw+0x1a>
};
CODE const uint8_t sk_load_a8_hsw[] = {
@@ -2556,11 +2667,13 @@ CODE const uint8_t sk_load_a8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,42, //jne 736 <_sk_load_a8_hsw+0x3a>
+ 117,50, //jne 7d9 <_sk_load_a8_hsw+0x42>
197,251,16,0, //vmovsd (%rax),%xmm0
196,226,125,49,192, //vpmovzxbd %xmm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,12, //vbroadcastss 0xc(%rdx),%ymm1
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -2577,22 +2690,24 @@ CODE const uint8_t sk_load_a8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 73e <_sk_load_a8_hsw+0x42>
+ 117,234, //jne 7e1 <_sk_load_a8_hsw+0x4a>
196,193,249,110,193, //vmovq %r9,%xmm0
- 235,181, //jmp 710 <_sk_load_a8_hsw+0x14>
+ 235,173, //jmp 7ab <_sk_load_a8_hsw+0x14>
};
CODE const uint8_t sk_store_a8_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,66,8, //vbroadcastss 0x8(%rdx),%ymm8
+ 184,0,0,127,67, //mov $0x437f0000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne 78e <_sk_store_a8_hsw+0x33>
+ 117,10, //jne 839 <_sk_store_a8_hsw+0x3b>
196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -2601,9 +2716,9 @@ CODE const uint8_t sk_store_a8_hsw[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 78a <_sk_store_a8_hsw+0x2f>
+ 119,236, //ja 835 <_sk_store_a8_hsw+0x37>
196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 76,141,21,66,0,0,0, //lea 0x42(%rip),%r10 # 7ec <_sk_store_a8_hsw+0x91>
+ 76,141,21,67,0,0,0, //lea 0x43(%rip),%r10 # 898 <_sk_store_a8_hsw+0x9a>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -2614,26 +2729,27 @@ CODE const uint8_t sk_store_a8_hsw[] = {
196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,158, //jmp 78a <_sk_store_a8_hsw+0x2f>
- 247,255, //idiv %edi
+ 235,158, //jmp 835 <_sk_store_a8_hsw+0x37>
+ 144, //nop
+ 246,255, //idiv %bh
255, //(bad)
255, //(bad)
- 239, //out %eax,(%dx)
+ 238, //out %al,(%dx)
255, //(bad)
255, //(bad)
- 255,231, //jmpq *%rdi
+ 255,230, //jmpq *%rsi
255, //(bad)
255, //(bad)
255, //(bad)
- 223,255, //(bad)
+ 222,255, //fdivrp %st,%st(7)
255, //(bad)
- 255,215, //callq *%rdi
+ 255,214, //callq *%rsi
255, //(bad)
255, //(bad)
- 255,207, //dec %edi
+ 255,206, //dec %esi
255, //(bad)
255, //(bad)
- 255,199, //inc %edi
+ 255,198, //inc %esi
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -2643,25 +2759,39 @@ CODE const uint8_t sk_load_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 117,92, //jne 86e <_sk_load_565_hsw+0x66>
+ 15,133,149,0,0,0, //jne 957 <_sk_load_565_hsw+0xa3>
196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
196,226,125,51,208, //vpmovzxwd %xmm0,%ymm2
- 196,226,125,88,66,104, //vpbroadcastd 0x68(%rdx),%ymm0
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
197,253,219,194, //vpand %ymm2,%ymm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm1
- 197,244,89,192, //vmulps %ymm0,%ymm1,%ymm0
- 196,226,125,88,74,108, //vpbroadcastd 0x6c(%rdx),%ymm1
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
+ 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
197,245,219,202, //vpand %ymm2,%ymm1,%ymm1
197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,226,125,24,90,120, //vbroadcastss 0x78(%rdx),%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,226,125,88,90,112, //vpbroadcastd 0x70(%rdx),%ymm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
197,229,219,210, //vpand %ymm2,%ymm3,%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,226,125,24,90,124, //vbroadcastss 0x7c(%rdx),%ymm3
- 197,228,89,210, //vmulps %ymm2,%ymm3,%ymm2
- 196,226,125,24,26, //vbroadcastss (%rdx),%ymm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
65,137,200, //mov %ecx,%r8d
@@ -2670,8 +2800,8 @@ CODE const uint8_t sk_load_565_hsw[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,146, //ja 818 <_sk_load_565_hsw+0x10>
- 76,141,13,75,0,0,0, //lea 0x4b(%rip),%r9 # 8d8 <_sk_load_565_hsw+0xd0>
+ 15,135,85,255,255,255, //ja 8c8 <_sk_load_565_hsw+0x14>
+ 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # 9c4 <_sk_load_565_hsw+0x110>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -2683,28 +2813,27 @@ CODE const uint8_t sk_load_565_hsw[] = {
196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,66,255,255,255, //jmpq 818 <_sk_load_565_hsw+0x10>
- 102,144, //xchg %ax,%ax
- 242,255, //repnz (bad)
- 255, //(bad)
+ 233,5,255,255,255, //jmpq 8c8 <_sk_load_565_hsw+0x14>
+ 144, //nop
+ 243,255, //repz (bad)
255, //(bad)
- 234, //(bad)
255, //(bad)
+ 235,255, //jmp 9c9 <_sk_load_565_hsw+0x115>
255, //(bad)
- 255,226, //jmpq *%rdx
+ 255,227, //jmpq *%rbx
255, //(bad)
255, //(bad)
255, //(bad)
- 218,255, //(bad)
+ 219,255, //(bad)
255, //(bad)
- 255,210, //callq *%rdx
+ 255,211, //callq *%rbx
255, //(bad)
255, //(bad)
- 255,202, //dec %edx
+ 255,203, //dec %ebx
255, //(bad)
255, //(bad)
255, //(bad)
- 190, //.byte 0xbe
+ 191, //.byte 0xbf
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -2713,11 +2842,15 @@ CODE const uint8_t sk_load_565_hsw[] = {
CODE const uint8_t sk_store_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,130,128,0,0,0, //vbroadcastss 0x80(%rdx),%ymm8
+ 184,0,0,248,65, //mov $0x41f80000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
196,193,53,114,241,11, //vpslld $0xb,%ymm9,%ymm9
- 196,98,125,24,146,132,0,0,0, //vbroadcastss 0x84(%rdx),%ymm10
+ 184,0,0,124,66, //mov $0x427c0000,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
196,193,45,114,242,5, //vpslld $0x5,%ymm10,%ymm10
@@ -2728,7 +2861,7 @@ CODE const uint8_t sk_store_565_hsw[] = {
196,67,125,57,193,1, //vextracti128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne 956 <_sk_store_565_hsw+0x62>
+ 117,10, //jne a4c <_sk_store_565_hsw+0x6c>
196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -2737,8 +2870,8 @@ CODE const uint8_t sk_store_565_hsw[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 952 <_sk_store_565_hsw+0x5e>
- 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # 9b4 <_sk_store_565_hsw+0xc0>
+ 119,236, //ja a48 <_sk_store_565_hsw+0x68>
+ 76,141,21,69,0,0,0, //lea 0x45(%rip),%r10 # aa8 <_sk_store_565_hsw+0xc8>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -2750,27 +2883,28 @@ CODE const uint8_t sk_store_565_hsw[] = {
196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
197,121,126,192, //vmovd %xmm8,%eax
102,65,137,4,121, //mov %ax,(%r9,%rdi,2)
- 235,161, //jmp 952 <_sk_store_565_hsw+0x5e>
- 15,31,0, //nopl (%rax)
- 242,255, //repnz (bad)
+ 235,161, //jmp a48 <_sk_store_565_hsw+0x68>
+ 144, //nop
+ 244, //hlt
255, //(bad)
255, //(bad)
- 234, //(bad)
255, //(bad)
+ 236, //in (%dx),%al
255, //(bad)
- 255,226, //jmpq *%rdx
255, //(bad)
+ 255,228, //jmpq *%rsp
255, //(bad)
255, //(bad)
- 218,255, //(bad)
255, //(bad)
- 255,210, //callq *%rdx
+ 220,255, //fdivr %st,%st(7)
255, //(bad)
+ 255,212, //callq *%rsp
255, //(bad)
- 255,202, //dec %edx
+ 255, //(bad)
+ 255,204, //dec %esp
255, //(bad)
255, //(bad)
- 255,194, //inc %edx
+ 255,196, //inc %esp
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -2782,7 +2916,7 @@ CODE const uint8_t sk_load_8888_hsw[] = {
76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
76,3,8, //add (%rax),%r9
77,133,192, //test %r8,%r8
- 117,104, //jne a4d <_sk_load_8888_hsw+0x7d>
+ 117,104, //jne b41 <_sk_load_8888_hsw+0x7d>
196,193,126,111,25, //vmovdqu (%r9),%ymm3
184,255,0,0,0, //mov $0xff,%eax
197,249,110,192, //vmovd %eax,%xmm0
@@ -2815,7 +2949,7 @@ CODE const uint8_t sk_load_8888_hsw[] = {
196,225,249,110,192, //vmovq %rax,%xmm0
196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,116,255,255,255, //jmpq 9ea <_sk_load_8888_hsw+0x1a>
+ 233,116,255,255,255, //jmpq ade <_sk_load_8888_hsw+0x1a>
};
CODE const uint8_t sk_store_8888_hsw[] = {
@@ -2841,7 +2975,7 @@ CODE const uint8_t sk_store_8888_hsw[] = {
196,65,45,235,192, //vpor %ymm8,%ymm10,%ymm8
196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
77,133,192, //test %r8,%r8
- 117,12, //jne aea <_sk_store_8888_hsw+0x74>
+ 117,12, //jne bde <_sk_store_8888_hsw+0x74>
196,65,126,127,1, //vmovdqu %ymm8,(%r9)
72,173, //lods %ds:(%rsi),%rax
76,137,193, //mov %r8,%rcx
@@ -2854,14 +2988,14 @@ CODE const uint8_t sk_store_8888_hsw[] = {
196,97,249,110,200, //vmovq %rax,%xmm9
196,66,125,33,201, //vpmovsxbd %xmm9,%ymm9
196,66,53,142,1, //vpmaskmovd %ymm8,%ymm9,(%r9)
- 235,211, //jmp ae3 <_sk_store_8888_hsw+0x6d>
+ 235,211, //jmp bd7 <_sk_store_8888_hsw+0x6d>
};
CODE const uint8_t sk_load_f16_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
72,133,201, //test %rcx,%rcx
- 117,97, //jne b7b <_sk_load_f16_hsw+0x6b>
+ 117,97, //jne c6f <_sk_load_f16_hsw+0x6b>
197,249,16,12,248, //vmovupd (%rax,%rdi,8),%xmm1
197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
@@ -2887,35 +3021,35 @@ CODE const uint8_t sk_load_f16_hsw[] = {
197,251,16,12,248, //vmovsd (%rax,%rdi,8),%xmm1
196,65,57,87,192, //vxorpd %xmm8,%xmm8,%xmm8
72,131,249,1, //cmp $0x1,%rcx
- 117,6, //jne b91 <_sk_load_f16_hsw+0x81>
+ 117,6, //jne c85 <_sk_load_f16_hsw+0x81>
197,250,126,201, //vmovq %xmm1,%xmm1
- 235,30, //jmp baf <_sk_load_f16_hsw+0x9f>
+ 235,30, //jmp ca3 <_sk_load_f16_hsw+0x9f>
197,241,22,76,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm1,%xmm1
72,131,249,3, //cmp $0x3,%rcx
- 114,18, //jb baf <_sk_load_f16_hsw+0x9f>
+ 114,18, //jb ca3 <_sk_load_f16_hsw+0x9f>
197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
72,131,249,3, //cmp $0x3,%rcx
- 117,19, //jne bbc <_sk_load_f16_hsw+0xac>
+ 117,19, //jne cb0 <_sk_load_f16_hsw+0xac>
197,250,126,210, //vmovq %xmm2,%xmm2
- 235,46, //jmp bdd <_sk_load_f16_hsw+0xcd>
+ 235,46, //jmp cd1 <_sk_load_f16_hsw+0xcd>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,117,255,255,255, //jmpq b31 <_sk_load_f16_hsw+0x21>
+ 233,117,255,255,255, //jmpq c25 <_sk_load_f16_hsw+0x21>
197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
72,131,249,5, //cmp $0x5,%rcx
- 114,21, //jb bdd <_sk_load_f16_hsw+0xcd>
+ 114,21, //jb cd1 <_sk_load_f16_hsw+0xcd>
197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
72,131,249,5, //cmp $0x5,%rcx
- 117,18, //jne be6 <_sk_load_f16_hsw+0xd6>
+ 117,18, //jne cda <_sk_load_f16_hsw+0xd6>
197,250,126,219, //vmovq %xmm3,%xmm3
- 233,84,255,255,255, //jmpq b31 <_sk_load_f16_hsw+0x21>
+ 233,84,255,255,255, //jmpq c25 <_sk_load_f16_hsw+0x21>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,75,255,255,255, //jmpq b31 <_sk_load_f16_hsw+0x21>
+ 233,75,255,255,255, //jmpq c25 <_sk_load_f16_hsw+0x21>
197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
72,131,249,7, //cmp $0x7,%rcx
- 15,130,59,255,255,255, //jb b31 <_sk_load_f16_hsw+0x21>
+ 15,130,59,255,255,255, //jb c25 <_sk_load_f16_hsw+0x21>
197,123,16,68,248,48, //vmovsd 0x30(%rax,%rdi,8),%xmm8
- 233,48,255,255,255, //jmpq b31 <_sk_load_f16_hsw+0x21>
+ 233,48,255,255,255, //jmpq c25 <_sk_load_f16_hsw+0x21>
};
CODE const uint8_t sk_store_f16_hsw[] = {
@@ -2934,7 +3068,7 @@ CODE const uint8_t sk_store_f16_hsw[] = {
196,65,57,98,205, //vpunpckldq %xmm13,%xmm8,%xmm9
196,65,57,106,197, //vpunpckhdq %xmm13,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,27, //jne c66 <_sk_store_f16_hsw+0x65>
+ 117,27, //jne d5a <_sk_store_f16_hsw+0x65>
197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
@@ -2943,22 +3077,22 @@ CODE const uint8_t sk_store_f16_hsw[] = {
255,224, //jmpq *%rax
197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je c62 <_sk_store_f16_hsw+0x61>
+ 116,241, //je d56 <_sk_store_f16_hsw+0x61>
197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb c62 <_sk_store_f16_hsw+0x61>
+ 114,229, //jb d56 <_sk_store_f16_hsw+0x61>
197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je c62 <_sk_store_f16_hsw+0x61>
+ 116,221, //je d56 <_sk_store_f16_hsw+0x61>
197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb c62 <_sk_store_f16_hsw+0x61>
+ 114,209, //jb d56 <_sk_store_f16_hsw+0x61>
197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je c62 <_sk_store_f16_hsw+0x61>
+ 116,201, //je d56 <_sk_store_f16_hsw+0x61>
197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb c62 <_sk_store_f16_hsw+0x61>
+ 114,189, //jb d56 <_sk_store_f16_hsw+0x61>
197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp c62 <_sk_store_f16_hsw+0x61>
+ 235,181, //jmp d56 <_sk_store_f16_hsw+0x61>
};
CODE const uint8_t sk_store_f32_hsw[] = {
@@ -2974,7 +3108,7 @@ CODE const uint8_t sk_store_f32_hsw[] = {
196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
72,133,201, //test %rcx,%rcx
- 117,55, //jne d1a <_sk_store_f32_hsw+0x6d>
+ 117,55, //jne e0e <_sk_store_f32_hsw+0x6d>
196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
@@ -2987,22 +3121,22 @@ CODE const uint8_t sk_store_f32_hsw[] = {
255,224, //jmpq *%rax
196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je d16 <_sk_store_f32_hsw+0x69>
+ 116,240, //je e0a <_sk_store_f32_hsw+0x69>
196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb d16 <_sk_store_f32_hsw+0x69>
+ 114,227, //jb e0a <_sk_store_f32_hsw+0x69>
196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je d16 <_sk_store_f32_hsw+0x69>
+ 116,218, //je e0a <_sk_store_f32_hsw+0x69>
196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb d16 <_sk_store_f32_hsw+0x69>
+ 114,205, //jb e0a <_sk_store_f32_hsw+0x69>
196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je d16 <_sk_store_f32_hsw+0x69>
+ 116,195, //je e0a <_sk_store_f32_hsw+0x69>
196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb d16 <_sk_store_f32_hsw+0x69>
+ 114,181, //jb e0a <_sk_store_f32_hsw+0x69>
196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp d16 <_sk_store_f32_hsw+0x69>
+ 235,171, //jmp e0a <_sk_store_f32_hsw+0x69>
};
CODE const uint8_t sk_clamp_x_hsw[] = {
@@ -3098,11 +3232,17 @@ CODE const uint8_t sk_mirror_y_hsw[] = {
};
CODE const uint8_t sk_luminance_to_alpha_hsw[] = {
- 196,98,125,24,130,136,0,0,0, //vbroadcastss 0x88(%rdx),%ymm8
- 196,226,125,24,154,140,0,0,0, //vbroadcastss 0x8c(%rdx),%ymm3
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,98,125,24,195, //vbroadcastss %xmm3,%ymm8
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
196,98,125,168,193, //vfmadd213ps %ymm1,%ymm0,%ymm8
- 196,226,125,24,154,144,0,0,0, //vbroadcastss 0x90(%rdx),%ymm3
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 196,226,125,24,216, //vbroadcastss %xmm0,%ymm3
196,194,109,168,216, //vfmadd213ps %ymm8,%ymm2,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -3322,7 +3462,7 @@ CODE const uint8_t sk_seed_shader_avx[] = {
196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,66,20, //vaddps 0x14(%rdx),%ymm0,%ymm0
+ 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
196,226,125,24,16, //vbroadcastss (%rax),%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
@@ -3516,23 +3656,38 @@ CODE const uint8_t sk_unpremul_avx[] = {
};
CODE const uint8_t sk_from_srgb_avx[] = {
- 196,98,125,24,66,64, //vbroadcastss 0x40(%rdx),%ymm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 196,98,125,24,90,60, //vbroadcastss 0x3c(%rdx),%ymm11
- 196,98,125,24,98,56, //vbroadcastss 0x38(%rdx),%ymm12
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
+ 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,67,121,4,228,0, //vpermilps $0x0,%xmm12,%xmm12
+ 196,67,29,24,228,1, //vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
197,36,89,232, //vmulps %ymm0,%ymm11,%ymm13
196,65,20,88,236, //vaddps %ymm12,%ymm13,%ymm13
- 196,98,125,24,114,52, //vbroadcastss 0x34(%rdx),%ymm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
+ 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 196,98,125,24,106,68, //vbroadcastss 0x44(%rdx),%ymm13
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 197,121,110,232, //vmovd %eax,%xmm13
+ 196,67,121,4,237,0, //vpermilps $0x0,%xmm13,%xmm13
+ 196,67,21,24,237,1, //vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
196,193,124,194,197,1, //vcmpltps %ymm13,%ymm0,%ymm0
196,195,45,74,193,0, //vblendvps %ymm0,%ymm9,%ymm10,%ymm0
197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
197,116,89,209, //vmulps %ymm1,%ymm1,%ymm10
197,36,89,249, //vmulps %ymm1,%ymm11,%ymm15
- 196,65,4,88,252, //vaddps %ymm12,%ymm15,%ymm15
+ 196,65,28,88,255, //vaddps %ymm15,%ymm12,%ymm15
196,65,44,89,215, //vmulps %ymm15,%ymm10,%ymm10
196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
196,193,116,194,205,1, //vcmpltps %ymm13,%ymm1,%ymm1
@@ -3540,7 +3695,7 @@ CODE const uint8_t sk_from_srgb_avx[] = {
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
197,36,89,210, //vmulps %ymm2,%ymm11,%ymm10
- 196,65,44,88,212, //vaddps %ymm12,%ymm10,%ymm10
+ 196,65,28,88,210, //vaddps %ymm10,%ymm12,%ymm10
196,65,52,89,202, //vmulps %ymm10,%ymm9,%ymm9
196,65,12,88,201, //vaddps %ymm9,%ymm14,%ymm9
196,193,108,194,213,1, //vcmpltps %ymm13,%ymm2,%ymm2
@@ -3551,43 +3706,62 @@ CODE const uint8_t sk_from_srgb_avx[] = {
CODE const uint8_t sk_to_srgb_avx[] = {
197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,200, //vrcpps %ymm8,%ymm9
- 196,65,124,82,208, //vrsqrtps %ymm8,%ymm10
- 196,98,125,24,66,72, //vbroadcastss 0x48(%rdx),%ymm8
- 197,60,89,216, //vmulps %ymm0,%ymm8,%ymm11
- 196,98,125,24,34, //vbroadcastss (%rdx),%ymm12
- 196,98,125,24,106,76, //vbroadcastss 0x4c(%rdx),%ymm13
- 196,98,125,24,114,80, //vbroadcastss 0x50(%rdx),%ymm14
- 196,98,125,24,122,84, //vbroadcastss 0x54(%rdx),%ymm15
- 196,65,52,89,206, //vmulps %ymm14,%ymm9,%ymm9
- 196,65,52,88,207, //vaddps %ymm15,%ymm9,%ymm9
- 196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
- 196,65,44,88,201, //vaddps %ymm9,%ymm10,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,98,125,24,82,88, //vbroadcastss 0x58(%rdx),%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,53,74,195,0, //vblendvps %ymm0,%ymm11,%ymm9,%ymm0
- 197,124,82,201, //vrsqrtps %ymm1,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,65,12,89,219, //vmulps %ymm11,%ymm14,%ymm11
- 196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,20,89,201, //vmulps %ymm9,%ymm13,%ymm9
- 196,65,52,88,203, //vaddps %ymm11,%ymm9,%ymm9
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,53,74,203,16, //vblendvps %ymm1,%ymm11,%ymm9,%ymm1
- 197,124,82,202, //vrsqrtps %ymm2,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,12,89,219, //vmulps %ymm11,%ymm14,%ymm11
+ 196,65,124,83,232, //vrcpps %ymm8,%ymm13
+ 196,65,124,82,240, //vrsqrtps %ymm8,%ymm14
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ 197,60,89,224, //vmulps %ymm0,%ymm8,%ymm12
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,121,110,200, //vmovd %eax,%xmm9
+ 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
+ 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
+ 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
+ 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 197,121,110,248, //vmovd %eax,%xmm15
+ 196,67,121,4,255,0, //vpermilps $0x0,%xmm15,%xmm15
+ 196,67,5,24,255,1, //vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
+ 196,65,20,89,235, //vmulps %ymm11,%ymm13,%ymm13
+ 196,65,20,88,239, //vaddps %ymm15,%ymm13,%ymm13
+ 196,65,12,89,242, //vmulps %ymm10,%ymm14,%ymm14
+ 196,65,12,88,237, //vaddps %ymm13,%ymm14,%ymm13
+ 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
+ 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ 196,193,124,194,198,1, //vcmpltps %ymm14,%ymm0,%ymm0
+ 196,195,21,74,196,0, //vblendvps %ymm0,%ymm12,%ymm13,%ymm0
+ 197,124,82,225, //vrsqrtps %ymm1,%ymm12
+ 196,65,124,83,236, //vrcpps %ymm12,%ymm13
+ 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
+ 196,65,36,89,237, //vmulps %ymm13,%ymm11,%ymm13
+ 196,65,4,88,237, //vaddps %ymm13,%ymm15,%ymm13
+ 196,65,44,89,228, //vmulps %ymm12,%ymm10,%ymm12
+ 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
+ 197,60,89,233, //vmulps %ymm1,%ymm8,%ymm13
+ 196,65,52,93,228, //vminps %ymm12,%ymm9,%ymm12
+ 196,193,116,194,206,1, //vcmpltps %ymm14,%ymm1,%ymm1
+ 196,195,29,74,205,16, //vblendvps %ymm1,%ymm13,%ymm12,%ymm1
+ 197,124,82,226, //vrsqrtps %ymm2,%ymm12
+ 196,65,124,83,236, //vrcpps %ymm12,%ymm13
+ 196,65,36,89,221, //vmulps %ymm13,%ymm11,%ymm11
196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,65,20,89,201, //vmulps %ymm9,%ymm13,%ymm9
- 196,65,52,88,203, //vaddps %ymm11,%ymm9,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
+ 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
+ 196,65,44,89,212, //vmulps %ymm12,%ymm10,%ymm10
+ 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
+ 196,65,52,93,202, //vminps %ymm10,%ymm9,%ymm9
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
+ 196,193,108,194,214,1, //vcmpltps %ymm14,%ymm2,%ymm2
196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -3610,7 +3784,7 @@ CODE const uint8_t sk_scale_u8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,80, //jne 4f8 <_sk_scale_u8_avx+0x60>
+ 117,80, //jne 5a2 <_sk_scale_u8_avx+0x60>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
@@ -3638,9 +3812,9 @@ CODE const uint8_t sk_scale_u8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 500 <_sk_scale_u8_avx+0x68>
+ 117,234, //jne 5aa <_sk_scale_u8_avx+0x68>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,143, //jmp 4ac <_sk_scale_u8_avx+0x14>
+ 235,143, //jmp 556 <_sk_scale_u8_avx+0x14>
};
CODE const uint8_t sk_lerp_1_float_avx[] = {
@@ -3668,7 +3842,7 @@ CODE const uint8_t sk_lerp_u8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,116, //jne 5e0 <_sk_lerp_u8_avx+0x84>
+ 117,116, //jne 68a <_sk_lerp_u8_avx+0x84>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
@@ -3704,41 +3878,59 @@ CODE const uint8_t sk_lerp_u8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 5e8 <_sk_lerp_u8_avx+0x8c>
+ 117,234, //jne 692 <_sk_lerp_u8_avx+0x8c>
196,65,249,110,193, //vmovq %r9,%xmm8
- 233,104,255,255,255, //jmpq 570 <_sk_lerp_u8_avx+0x14>
+ 233,104,255,255,255, //jmpq 61a <_sk_lerp_u8_avx+0x14>
};
CODE const uint8_t sk_lerp_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,164,0,0,0, //jne 6ba <_sk_lerp_565_avx+0xb2>
+ 15,133,250,0,0,0, //jne 7ba <_sk_lerp_565_avx+0x108>
196,65,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm8
197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
197,185,105,219, //vpunpckhwd %xmm3,%xmm8,%xmm3
196,66,121,51,192, //vpmovzxwd %xmm8,%xmm8
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 196,98,125,24,66,104, //vbroadcastss 0x68(%rdx),%ymm8
- 197,60,84,195, //vandps %ymm3,%ymm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 196,98,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm9
- 196,65,52,89,192, //vmulps %ymm8,%ymm9,%ymm8
- 196,98,125,24,74,108, //vbroadcastss 0x6c(%rdx),%ymm9
- 197,52,84,203, //vandps %ymm3,%ymm9,%ymm9
- 196,65,124,91,201, //vcvtdq2ps %ymm9,%ymm9
- 196,98,125,24,82,120, //vbroadcastss 0x78(%rdx),%ymm10
- 196,65,44,89,201, //vmulps %ymm9,%ymm10,%ymm9
- 196,98,125,24,82,112, //vbroadcastss 0x70(%rdx),%ymm10
- 197,172,84,219, //vandps %ymm3,%ymm10,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,82,124, //vbroadcastss 0x7c(%rdx),%ymm10
- 197,172,89,219, //vmulps %ymm3,%ymm10,%ymm3
+ 196,99,61,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
+ 196,193,124,89,193, //vmulps %ymm9,%ymm0,%ymm0
197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,201, //vmulps %ymm9,%ymm1,%ymm1
+ 196,193,116,89,202, //vmulps %ymm10,%ymm1,%ymm1
197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
@@ -3755,8 +3947,8 @@ CODE const uint8_t sk_lerp_565_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,69,255,255,255, //ja 61c <_sk_lerp_565_avx+0x14>
- 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # 728 <_sk_lerp_565_avx+0x120>
+ 15,135,239,254,255,255, //ja 6c6 <_sk_lerp_565_avx+0x14>
+ 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # 828 <_sk_lerp_565_avx+0x176>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -3768,12 +3960,12 @@ CODE const uint8_t sk_lerp_565_avx[] = {
196,65,57,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
196,65,57,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
196,65,57,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
- 233,245,254,255,255, //jmpq 61c <_sk_lerp_565_avx+0x14>
+ 233,159,254,255,255, //jmpq 6c6 <_sk_lerp_565_avx+0x14>
144, //nop
243,255, //repz (bad)
255, //(bad)
255, //(bad)
- 235,255, //jmp 72d <_sk_lerp_565_avx+0x125>
+ 235,255, //jmp 82d <_sk_lerp_565_avx+0x17b>
255, //(bad)
255,227, //jmpq *%rbx
255, //(bad)
@@ -3804,9 +3996,12 @@ CODE const uint8_t sk_load_tables_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,0, //mov (%rax),%r8
72,133,201, //test %rcx,%rcx
- 15,133,18,2,0,0, //jne 96e <_sk_load_tables_avx+0x22a>
+ 15,133,56,2,0,0, //jne a94 <_sk_load_tables_avx+0x250>
196,65,124,16,4,184, //vmovups (%r8,%rdi,4),%ymm8
- 196,98,125,24,74,16, //vbroadcastss 0x10(%rdx),%ymm9
+ 187,255,0,0,0, //mov $0xff,%ebx
+ 197,249,110,195, //vmovd %ebx,%xmm0
+ 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
+ 196,99,125,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
196,193,52,84,192, //vandps %ymm8,%ymm9,%ymm0
196,193,249,126,193, //vmovq %xmm0,%r9
69,137,203, //mov %r9d,%r11d
@@ -3825,12 +4020,14 @@ CODE const uint8_t sk_load_tables_avx[] = {
76,139,64,16, //mov 0x10(%rax),%r8
196,161,122,16,68,189,0, //vmovss 0x0(%rbp,%r15,4),%xmm0
196,163,121,33,68,165,0,16, //vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
- 196,163,121,33,68,173,0,32, //vinsertps $0x20,0x0(%rbp,%r13,4),%xmm0,%xmm0
+ 196,161,122,16,76,173,0, //vmovss 0x0(%rbp,%r13,4),%xmm1
+ 196,227,121,33,193,32, //vinsertps $0x20,%xmm1,%xmm0,%xmm0
197,250,16,76,157,0, //vmovss 0x0(%rbp,%rbx,4),%xmm1
196,227,121,33,193,48, //vinsertps $0x30,%xmm1,%xmm0,%xmm0
196,161,122,16,76,157,0, //vmovss 0x0(%rbp,%r11,4),%xmm1
196,163,113,33,76,141,0,16, //vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
- 196,163,113,33,76,181,0,32, //vinsertps $0x20,0x0(%rbp,%r14,4),%xmm1,%xmm1
+ 196,161,122,16,92,181,0, //vmovss 0x0(%rbp,%r14,4),%xmm3
+ 196,227,113,33,203,32, //vinsertps $0x20,%xmm3,%xmm1,%xmm1
196,161,122,16,92,149,0, //vmovss 0x0(%rbp,%r10,4),%xmm3
196,227,113,33,203,48, //vinsertps $0x30,%xmm3,%xmm1,%xmm1
196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
@@ -3899,9 +4096,12 @@ CODE const uint8_t sk_load_tables_avx[] = {
196,193,57,114,208,24, //vpsrld $0x18,%xmm8,%xmm8
196,193,97,114,210,24, //vpsrld $0x18,%xmm10,%xmm3
196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,66,12, //vbroadcastss 0xc(%rdx),%ymm8
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
72,173, //lods %ds:(%rsi),%rax
91, //pop %rbx
65,92, //pop %r12
@@ -3910,17 +4110,17 @@ CODE const uint8_t sk_load_tables_avx[] = {
65,95, //pop %r15
93, //pop %rbp
255,224, //jmpq *%rax
- 65,137,201, //mov %ecx,%r9d
- 65,128,225,7, //and $0x7,%r9b
+ 137,203, //mov %ecx,%ebx
+ 128,227,7, //and $0x7,%bl
196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 65,254,201, //dec %r9b
- 69,15,182,201, //movzbl %r9b,%r9d
- 65,128,249,6, //cmp $0x6,%r9b
- 15,135,215,253,255,255, //ja 762 <_sk_load_tables_avx+0x1e>
- 76,141,21,138,0,0,0, //lea 0x8a(%rip),%r10 # a1c <_sk_load_tables_avx+0x2d8>
- 79,99,12,138, //movslq (%r10,%r9,4),%r9
- 77,1,209, //add %r10,%r9
- 65,255,225, //jmpq *%r9
+ 254,203, //dec %bl
+ 15,182,219, //movzbl %bl,%ebx
+ 128,251,6, //cmp $0x6,%bl
+ 15,135,182,253,255,255, //ja 862 <_sk_load_tables_avx+0x1e>
+ 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # b3c <_sk_load_tables_avx+0x2f8>
+ 73,99,28,153, //movslq (%r9,%rbx,4),%rbx
+ 76,1,203, //add %r9,%rbx
+ 255,227, //jmpq *%rbx
196,193,121,110,68,184,24, //vmovd 0x18(%r8,%rdi,4),%xmm0
197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
@@ -3940,7 +4140,7 @@ CODE const uint8_t sk_load_tables_avx[] = {
196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
196,195,57,34,4,184,0, //vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 233,70,253,255,255, //jmpq 762 <_sk_load_tables_avx+0x1e>
+ 233,38,253,255,255, //jmpq 862 <_sk_load_tables_avx+0x1e>
238, //out %al,(%dx)
255, //(bad)
255, //(bad)
@@ -3967,14 +4167,17 @@ CODE const uint8_t sk_load_a8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,59, //jne a83 <_sk_load_a8_avx+0x4b>
+ 117,74, //jne bb2 <_sk_load_a8_avx+0x5a>
197,251,16,0, //vmovsd (%rax),%xmm0
196,226,121,49,200, //vpmovzxbd %xmm0,%xmm1
196,227,121,4,192,229, //vpermilps $0xe5,%xmm0,%xmm0
196,226,121,49,192, //vpmovzxbd %xmm0,%xmm0
196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,12, //vbroadcastss 0xc(%rdx),%ymm1
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -3991,22 +4194,25 @@ CODE const uint8_t sk_load_a8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne a8b <_sk_load_a8_avx+0x53>
+ 117,234, //jne bba <_sk_load_a8_avx+0x62>
196,193,249,110,193, //vmovq %r9,%xmm0
- 235,164, //jmp a4c <_sk_load_a8_avx+0x14>
+ 235,149, //jmp b6c <_sk_load_a8_avx+0x14>
};
CODE const uint8_t sk_store_a8_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,66,8, //vbroadcastss 0x8(%rdx),%ymm8
+ 184,0,0,127,67, //mov $0x437f0000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne adb <_sk_store_a8_avx+0x33>
+ 117,10, //jne c19 <_sk_store_a8_avx+0x42>
196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -4015,9 +4221,9 @@ CODE const uint8_t sk_store_a8_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja ad7 <_sk_store_a8_avx+0x2f>
+ 119,236, //ja c15 <_sk_store_a8_avx+0x3e>
196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 76,141,21,69,0,0,0, //lea 0x45(%rip),%r10 # b3c <_sk_store_a8_avx+0x94>
+ 76,141,21,67,0,0,0, //lea 0x43(%rip),%r10 # c78 <_sk_store_a8_avx+0xa1>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -4028,28 +4234,27 @@ CODE const uint8_t sk_store_a8_avx[] = {
196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,158, //jmp ad7 <_sk_store_a8_avx+0x2f>
- 15,31,0, //nopl (%rax)
- 244, //hlt
- 255, //(bad)
+ 235,158, //jmp c15 <_sk_store_a8_avx+0x3e>
+ 144, //nop
+ 246,255, //idiv %bh
255, //(bad)
255, //(bad)
- 236, //in (%dx),%al
+ 238, //out %al,(%dx)
255, //(bad)
255, //(bad)
- 255,228, //jmpq *%rsp
+ 255,230, //jmpq *%rsi
255, //(bad)
255, //(bad)
255, //(bad)
- 220,255, //fdivr %st,%st(7)
+ 222,255, //fdivrp %st,%st(7)
255, //(bad)
- 255,212, //callq *%rsp
+ 255,214, //callq *%rsi
255, //(bad)
255, //(bad)
- 255,204, //dec %esp
+ 255,206, //dec %esi
255, //(bad)
255, //(bad)
- 255,196, //inc %esp
+ 255,198, //inc %esi
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -4059,28 +4264,49 @@ CODE const uint8_t sk_load_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 117,106, //jne bcc <_sk_load_565_avx+0x74>
+ 15,133,209,0,0,0, //jne d73 <_sk_load_565_avx+0xdf>
196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
197,241,239,201, //vpxor %xmm1,%xmm1,%xmm1
197,249,105,201, //vpunpckhwd %xmm1,%xmm0,%xmm1
196,226,121,51,192, //vpmovzxwd %xmm0,%xmm0
196,227,125,24,209,1, //vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
- 196,226,125,24,66,104, //vbroadcastss 0x68(%rdx),%ymm0
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
+ 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
197,252,84,194, //vandps %ymm2,%ymm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm1
- 197,244,89,192, //vmulps %ymm0,%ymm1,%ymm0
- 196,226,125,24,74,108, //vbroadcastss 0x6c(%rdx),%ymm1
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 197,249,112,201,0, //vpshufd $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,244,84,202, //vandps %ymm2,%ymm1,%ymm1
197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,226,125,24,90,120, //vbroadcastss 0x78(%rdx),%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,226,125,24,90,112, //vbroadcastss 0x70(%rdx),%ymm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,84,210, //vandps %ymm2,%ymm3,%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,226,125,24,90,124, //vbroadcastss 0x7c(%rdx),%ymm3
- 197,228,89,210, //vmulps %ymm2,%ymm3,%ymm2
- 196,226,125,24,26, //vbroadcastss (%rdx),%ymm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
65,137,200, //mov %ecx,%r8d
@@ -4089,8 +4315,8 @@ CODE const uint8_t sk_load_565_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,132, //ja b68 <_sk_load_565_avx+0x10>
- 76,141,13,73,0,0,0, //lea 0x49(%rip),%r9 # c34 <_sk_load_565_avx+0xdc>
+ 15,135,25,255,255,255, //ja ca8 <_sk_load_565_avx+0x14>
+ 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # de0 <_sk_load_565_avx+0x14c>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -4102,27 +4328,27 @@ CODE const uint8_t sk_load_565_avx[] = {
196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,52,255,255,255, //jmpq b68 <_sk_load_565_avx+0x10>
- 244, //hlt
- 255, //(bad)
+ 233,201,254,255,255, //jmpq ca8 <_sk_load_565_avx+0x14>
+ 144, //nop
+ 243,255, //repz (bad)
255, //(bad)
255, //(bad)
- 236, //in (%dx),%al
+ 235,255, //jmp de5 <_sk_load_565_avx+0x151>
255, //(bad)
+ 255,227, //jmpq *%rbx
255, //(bad)
- 255,228, //jmpq *%rsp
255, //(bad)
255, //(bad)
+ 219,255, //(bad)
255, //(bad)
- 220,255, //fdivr %st,%st(7)
+ 255,211, //callq *%rbx
255, //(bad)
- 255,212, //callq *%rsp
255, //(bad)
+ 255,203, //dec %ebx
255, //(bad)
- 255,204, //dec %esp
255, //(bad)
255, //(bad)
- 255,192, //inc %eax
+ 191, //.byte 0xbf
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -4131,14 +4357,20 @@ CODE const uint8_t sk_load_565_avx[] = {
CODE const uint8_t sk_store_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,130,128,0,0,0, //vbroadcastss 0x80(%rdx),%ymm8
+ 184,0,0,248,65, //mov $0x41f80000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
196,193,41,114,241,11, //vpslld $0xb,%xmm9,%xmm10
196,67,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm9
196,193,49,114,241,11, //vpslld $0xb,%xmm9,%xmm9
196,67,45,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
- 196,98,125,24,146,132,0,0,0, //vbroadcastss 0x84(%rdx),%ymm10
+ 184,0,0,124,66, //mov $0x427c0000,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
+ 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
196,193,33,114,242,5, //vpslld $0x5,%xmm10,%xmm11
@@ -4152,7 +4384,7 @@ CODE const uint8_t sk_store_565_avx[] = {
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne cd6 <_sk_store_565_avx+0x86>
+ 117,10, //jne e9a <_sk_store_565_avx+0x9e>
196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -4161,8 +4393,8 @@ CODE const uint8_t sk_store_565_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja cd2 <_sk_store_565_avx+0x82>
- 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # d34 <_sk_store_565_avx+0xe4>
+ 119,236, //ja e96 <_sk_store_565_avx+0x9a>
+ 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # ef8 <_sk_store_565_avx+0xfc>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -4174,7 +4406,7 @@ CODE const uint8_t sk_store_565_avx[] = {
196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
197,121,126,192, //vmovd %xmm8,%eax
102,65,137,4,121, //mov %ax,(%r9,%rdi,2)
- 235,161, //jmp cd2 <_sk_store_565_avx+0x82>
+ 235,161, //jmp e96 <_sk_store_565_avx+0x9a>
15,31,0, //nopl (%rax)
242,255, //repnz (bad)
255, //(bad)
@@ -4204,7 +4436,7 @@ CODE const uint8_t sk_load_8888_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,157,0,0,0, //jne dfb <_sk_load_8888_avx+0xab>
+ 15,133,157,0,0,0, //jne fbf <_sk_load_8888_avx+0xab>
196,65,124,16,12,186, //vmovups (%r10,%rdi,4),%ymm9
184,255,0,0,0, //mov $0xff,%eax
197,249,110,192, //vmovd %eax,%xmm0
@@ -4243,8 +4475,8 @@ CODE const uint8_t sk_load_8888_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,76,255,255,255, //ja d64 <_sk_load_8888_avx+0x14>
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # ea8 <_sk_load_8888_avx+0x158>
+ 15,135,76,255,255,255, //ja f28 <_sk_load_8888_avx+0x14>
+ 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # 106c <_sk_load_8888_avx+0x158>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -4267,7 +4499,7 @@ CODE const uint8_t sk_load_8888_avx[] = {
196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
196,195,49,34,4,186,0, //vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 233,188,254,255,255, //jmpq d64 <_sk_load_8888_avx+0x14>
+ 233,188,254,255,255, //jmpq f28 <_sk_load_8888_avx+0x14>
238, //out %al,(%dx)
255, //(bad)
255, //(bad)
@@ -4319,7 +4551,7 @@ CODE const uint8_t sk_store_8888_avx[] = {
196,65,45,86,192, //vorpd %ymm8,%ymm10,%ymm8
196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne f68 <_sk_store_8888_avx+0xa4>
+ 117,10, //jne 112c <_sk_store_8888_avx+0xa4>
196,65,124,17,4,185, //vmovups %ymm8,(%r9,%rdi,4)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -4328,8 +4560,8 @@ CODE const uint8_t sk_store_8888_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja f64 <_sk_store_8888_avx+0xa0>
- 76,141,21,85,0,0,0, //lea 0x55(%rip),%r10 # fd4 <_sk_store_8888_avx+0x110>
+ 119,236, //ja 1128 <_sk_store_8888_avx+0xa0>
+ 76,141,21,85,0,0,0, //lea 0x55(%rip),%r10 # 1198 <_sk_store_8888_avx+0x110>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -4343,7 +4575,7 @@ CODE const uint8_t sk_store_8888_avx[] = {
196,67,121,22,68,185,8,2, //vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
196,67,121,22,68,185,4,1, //vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
196,65,121,126,4,185, //vmovd %xmm8,(%r9,%rdi,4)
- 235,147, //jmp f64 <_sk_store_8888_avx+0xa0>
+ 235,147, //jmp 1128 <_sk_store_8888_avx+0xa0>
15,31,0, //nopl (%rax)
245, //cmc
255, //(bad)
@@ -4374,7 +4606,7 @@ CODE const uint8_t sk_load_f16_avx[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
72,133,201, //test %rcx,%rcx
- 15,133,240,0,0,0, //jne 10ee <_sk_load_f16_avx+0xfe>
+ 15,133,8,1,0,0, //jne 12ca <_sk_load_f16_avx+0x116>
197,249,16,12,248, //vmovupd (%rax,%rdi,8),%xmm1
197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
@@ -4385,42 +4617,46 @@ CODE const uint8_t sk_load_f16_avx[] = {
196,193,97,105,216, //vpunpckhwd %xmm8,%xmm3,%xmm3
197,121,97,193, //vpunpcklwd %xmm1,%xmm0,%xmm8
197,249,105,193, //vpunpckhwd %xmm1,%xmm0,%xmm0
- 197,233,97,203, //vpunpcklwd %xmm3,%xmm2,%xmm1
+ 197,105,97,211, //vpunpcklwd %xmm3,%xmm2,%xmm10
197,105,105,203, //vpunpckhwd %xmm3,%xmm2,%xmm9
- 197,249,110,90,100, //vmovd 0x64(%rdx),%xmm3
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
196,193,97,101,208, //vpcmpgtw %xmm8,%xmm3,%xmm2
- 196,65,105,223,192, //vpandn %xmm8,%xmm2,%xmm8
- 197,225,101,208, //vpcmpgtw %xmm0,%xmm3,%xmm2
- 197,233,223,192, //vpandn %xmm0,%xmm2,%xmm0
- 197,225,101,209, //vpcmpgtw %xmm1,%xmm3,%xmm2
- 197,233,223,201, //vpandn %xmm1,%xmm2,%xmm1
- 196,193,97,101,209, //vpcmpgtw %xmm9,%xmm3,%xmm2
- 196,193,105,223,209, //vpandn %xmm9,%xmm2,%xmm2
- 196,66,121,51,208, //vpmovzxwd %xmm8,%xmm10
+ 196,193,105,223,208, //vpandn %xmm8,%xmm2,%xmm2
+ 197,225,101,200, //vpcmpgtw %xmm0,%xmm3,%xmm1
+ 197,241,223,192, //vpandn %xmm0,%xmm1,%xmm0
+ 196,193,97,101,202, //vpcmpgtw %xmm10,%xmm3,%xmm1
+ 196,193,113,223,202, //vpandn %xmm10,%xmm1,%xmm1
+ 196,193,97,101,217, //vpcmpgtw %xmm9,%xmm3,%xmm3
+ 196,193,97,223,217, //vpandn %xmm9,%xmm3,%xmm3
+ 196,98,121,51,194, //vpmovzxwd %xmm2,%xmm8
196,98,121,51,201, //vpmovzxwd %xmm1,%xmm9
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,57,105,195, //vpunpckhwd %xmm3,%xmm8,%xmm8
- 197,241,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm1
+ 196,65,41,239,210, //vpxor %xmm10,%xmm10,%xmm10
+ 196,193,105,105,210, //vpunpckhwd %xmm10,%xmm2,%xmm2
+ 196,193,113,105,202, //vpunpckhwd %xmm10,%xmm1,%xmm1
196,98,121,51,216, //vpmovzxwd %xmm0,%xmm11
- 196,98,121,51,226, //vpmovzxwd %xmm2,%xmm12
- 197,121,105,235, //vpunpckhwd %xmm3,%xmm0,%xmm13
- 197,105,105,243, //vpunpckhwd %xmm3,%xmm2,%xmm14
- 196,193,121,114,242,13, //vpslld $0xd,%xmm10,%xmm0
- 196,193,105,114,241,13, //vpslld $0xd,%xmm9,%xmm2
- 196,227,125,24,194,1, //vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- 196,98,125,24,74,92, //vbroadcastss 0x5c(%rdx),%ymm9
+ 196,98,121,51,227, //vpmovzxwd %xmm3,%xmm12
+ 196,65,121,105,234, //vpunpckhwd %xmm10,%xmm0,%xmm13
+ 196,65,97,105,210, //vpunpckhwd %xmm10,%xmm3,%xmm10
+ 196,193,121,114,240,13, //vpslld $0xd,%xmm8,%xmm0
+ 196,193,97,114,241,13, //vpslld $0xd,%xmm9,%xmm3
+ 196,227,125,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm0,%ymm0
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,99,101,24,203,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm9
197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 196,193,105,114,240,13, //vpslld $0xd,%xmm8,%xmm2
+ 197,233,114,242,13, //vpslld $0xd,%xmm2,%xmm2
197,241,114,241,13, //vpslld $0xd,%xmm1,%xmm1
196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 196,193,105,114,243,13, //vpslld $0xd,%xmm11,%xmm2
- 196,193,97,114,244,13, //vpslld $0xd,%xmm12,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ 196,193,57,114,243,13, //vpslld $0xd,%xmm11,%xmm8
+ 196,193,105,114,244,13, //vpslld $0xd,%xmm12,%xmm2
+ 196,227,61,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm8,%ymm2
197,180,89,210, //vmulps %ymm2,%ymm9,%ymm2
196,193,57,114,245,13, //vpslld $0xd,%xmm13,%xmm8
- 196,193,97,114,246,13, //vpslld $0xd,%xmm14,%xmm3
+ 196,193,97,114,242,13, //vpslld $0xd,%xmm10,%xmm3
196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
197,180,89,219, //vmulps %ymm3,%ymm9,%ymm3
72,173, //lods %ds:(%rsi),%rax
@@ -4428,41 +4664,44 @@ CODE const uint8_t sk_load_f16_avx[] = {
197,251,16,12,248, //vmovsd (%rax,%rdi,8),%xmm1
196,65,57,87,192, //vxorpd %xmm8,%xmm8,%xmm8
72,131,249,1, //cmp $0x1,%rcx
- 117,6, //jne 1104 <_sk_load_f16_avx+0x114>
+ 117,6, //jne 12e0 <_sk_load_f16_avx+0x12c>
197,250,126,201, //vmovq %xmm1,%xmm1
- 235,30, //jmp 1122 <_sk_load_f16_avx+0x132>
+ 235,30, //jmp 12fe <_sk_load_f16_avx+0x14a>
197,241,22,76,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm1,%xmm1
72,131,249,3, //cmp $0x3,%rcx
- 114,18, //jb 1122 <_sk_load_f16_avx+0x132>
+ 114,18, //jb 12fe <_sk_load_f16_avx+0x14a>
197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
72,131,249,3, //cmp $0x3,%rcx
- 117,19, //jne 112f <_sk_load_f16_avx+0x13f>
+ 117,19, //jne 130b <_sk_load_f16_avx+0x157>
197,250,126,210, //vmovq %xmm2,%xmm2
- 235,46, //jmp 1150 <_sk_load_f16_avx+0x160>
+ 235,46, //jmp 132c <_sk_load_f16_avx+0x178>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,230,254,255,255, //jmpq 1015 <_sk_load_f16_avx+0x25>
+ 233,206,254,255,255, //jmpq 11d9 <_sk_load_f16_avx+0x25>
197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
72,131,249,5, //cmp $0x5,%rcx
- 114,21, //jb 1150 <_sk_load_f16_avx+0x160>
+ 114,21, //jb 132c <_sk_load_f16_avx+0x178>
197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
72,131,249,5, //cmp $0x5,%rcx
- 117,18, //jne 1159 <_sk_load_f16_avx+0x169>
+ 117,18, //jne 1335 <_sk_load_f16_avx+0x181>
197,250,126,219, //vmovq %xmm3,%xmm3
- 233,197,254,255,255, //jmpq 1015 <_sk_load_f16_avx+0x25>
+ 233,173,254,255,255, //jmpq 11d9 <_sk_load_f16_avx+0x25>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,188,254,255,255, //jmpq 1015 <_sk_load_f16_avx+0x25>
+ 233,164,254,255,255, //jmpq 11d9 <_sk_load_f16_avx+0x25>
197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
72,131,249,7, //cmp $0x7,%rcx
- 15,130,172,254,255,255, //jb 1015 <_sk_load_f16_avx+0x25>
+ 15,130,148,254,255,255, //jb 11d9 <_sk_load_f16_avx+0x25>
197,123,16,68,248,48, //vmovsd 0x30(%rax,%rdi,8),%xmm8
- 233,161,254,255,255, //jmpq 1015 <_sk_load_f16_avx+0x25>
+ 233,137,254,255,255, //jmpq 11d9 <_sk_load_f16_avx+0x25>
};
CODE const uint8_t sk_store_f16_avx[] = {
72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 196,98,125,24,66,96, //vbroadcastss 0x60(%rdx),%ymm8
+ 76,139,0, //mov (%rax),%r8
+ 184,0,0,128,7, //mov $0x7800000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,65,121,112,192,0, //vpshufd $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,67,125,25,202,1, //vextractf128 $0x1,%ymm9,%xmm10
196,193,41,114,210,13, //vpsrld $0xd,%xmm10,%xmm10
@@ -4492,31 +4731,31 @@ CODE const uint8_t sk_store_f16_avx[] = {
196,65,25,98,205, //vpunpckldq %xmm13,%xmm12,%xmm9
196,65,25,106,197, //vpunpckhdq %xmm13,%xmm12,%xmm8
72,133,201, //test %rcx,%rcx
- 117,27, //jne 1237 <_sk_store_f16_avx+0xc3>
- 197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
- 197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
- 197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
- 197,122,127,68,248,48, //vmovdqu %xmm8,0x30(%rax,%rdi,8)
+ 117,31, //jne 1426 <_sk_store_f16_avx+0xd6>
+ 196,65,120,17,28,248, //vmovups %xmm11,(%r8,%rdi,8)
+ 196,65,120,17,84,248,16, //vmovups %xmm10,0x10(%r8,%rdi,8)
+ 196,65,120,17,76,248,32, //vmovups %xmm9,0x20(%r8,%rdi,8)
+ 196,65,122,127,68,248,48, //vmovdqu %xmm8,0x30(%r8,%rdi,8)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
- 197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
+ 196,65,121,214,28,248, //vmovq %xmm11,(%r8,%rdi,8)
72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je 1233 <_sk_store_f16_avx+0xbf>
- 197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
+ 116,240, //je 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,92,248,8, //vmovhpd %xmm11,0x8(%r8,%rdi,8)
72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb 1233 <_sk_store_f16_avx+0xbf>
- 197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je 1233 <_sk_store_f16_avx+0xbf>
- 197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
+ 114,227, //jb 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,84,248,16, //vmovq %xmm10,0x10(%r8,%rdi,8)
+ 116,218, //je 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,84,248,24, //vmovhpd %xmm10,0x18(%r8,%rdi,8)
72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb 1233 <_sk_store_f16_avx+0xbf>
- 197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je 1233 <_sk_store_f16_avx+0xbf>
- 197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
+ 114,205, //jb 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,76,248,32, //vmovq %xmm9,0x20(%r8,%rdi,8)
+ 116,196, //je 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,76,248,40, //vmovhpd %xmm9,0x28(%r8,%rdi,8)
72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb 1233 <_sk_store_f16_avx+0xbf>
- 197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp 1233 <_sk_store_f16_avx+0xbf>
+ 114,183, //jb 1422 <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,68,248,48, //vmovq %xmm8,0x30(%r8,%rdi,8)
+ 235,174, //jmp 1422 <_sk_store_f16_avx+0xd2>
};
CODE const uint8_t sk_store_f32_avx[] = {
@@ -4532,7 +4771,7 @@ CODE const uint8_t sk_store_f32_avx[] = {
196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
72,133,201, //test %rcx,%rcx
- 117,55, //jne 12eb <_sk_store_f32_avx+0x6d>
+ 117,55, //jne 14e1 <_sk_store_f32_avx+0x6d>
196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
@@ -4545,22 +4784,22 @@ CODE const uint8_t sk_store_f32_avx[] = {
255,224, //jmpq *%rax
196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 12e7 <_sk_store_f32_avx+0x69>
+ 116,240, //je 14dd <_sk_store_f32_avx+0x69>
196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 12e7 <_sk_store_f32_avx+0x69>
+ 114,227, //jb 14dd <_sk_store_f32_avx+0x69>
196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je 12e7 <_sk_store_f32_avx+0x69>
+ 116,218, //je 14dd <_sk_store_f32_avx+0x69>
196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 12e7 <_sk_store_f32_avx+0x69>
+ 114,205, //jb 14dd <_sk_store_f32_avx+0x69>
196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je 12e7 <_sk_store_f32_avx+0x69>
+ 116,195, //je 14dd <_sk_store_f32_avx+0x69>
196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb 12e7 <_sk_store_f32_avx+0x69>
+ 114,181, //jb 14dd <_sk_store_f32_avx+0x69>
196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp 12e7 <_sk_store_f32_avx+0x69>
+ 235,171, //jmp 14dd <_sk_store_f32_avx+0x69>
};
CODE const uint8_t sk_clamp_x_avx[] = {
@@ -4682,12 +4921,21 @@ CODE const uint8_t sk_mirror_y_avx[] = {
};
CODE const uint8_t sk_luminance_to_alpha_avx[] = {
- 196,226,125,24,154,136,0,0,0, //vbroadcastss 0x88(%rdx),%ymm3
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 196,226,125,24,154,140,0,0,0, //vbroadcastss 0x8c(%rdx),%ymm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 196,226,125,24,138,144,0,0,0, //vbroadcastss 0x90(%rdx),%ymm1
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,244,89,202, //vmulps %ymm2,%ymm1,%ymm1
197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
@@ -4925,7 +5173,7 @@ CODE const uint8_t sk_seed_shader_sse41[] = {
102,15,110,209, //movd %ecx,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
15,88,202, //addps %xmm2,%xmm1
- 15,16,66,20, //movups 0x14(%rdx),%xmm0
+ 15,16,2, //movups (%rdx),%xmm0
15,88,193, //addps %xmm1,%xmm0
102,15,110,8, //movd (%rax),%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
@@ -5128,25 +5376,29 @@ CODE const uint8_t sk_unpremul_sse41[] = {
};
CODE const uint8_t sk_from_srgb_sse41[] = {
- 68,15,40,194, //movaps %xmm2,%xmm8
- 243,68,15,16,90,64, //movss 0x40(%rdx),%xmm11
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
69,15,40,211, //movaps %xmm11,%xmm10
68,15,89,208, //mulps %xmm0,%xmm10
68,15,40,240, //movaps %xmm0,%xmm14
69,15,89,246, //mulps %xmm14,%xmm14
- 243,15,16,82,60, //movss 0x3c(%rdx),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 243,68,15,16,98,52, //movss 0x34(%rdx),%xmm12
- 243,68,15,16,106,56, //movss 0x38(%rdx),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 69,15,88,205, //addps %xmm13,%xmm9
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
+ 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,206, //mulps %xmm14,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
69,15,88,204, //addps %xmm12,%xmm9
- 243,68,15,16,114,68, //movss 0x44(%rdx),%xmm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
+ 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
+ 69,15,89,206, //mulps %xmm14,%xmm9
+ 69,15,88,205, //addps %xmm13,%xmm9
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 102,68,15,110,240, //movd %eax,%xmm14
69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
65,15,194,198,1, //cmpltps %xmm14,%xmm0
102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
@@ -5154,27 +5406,28 @@ CODE const uint8_t sk_from_srgb_sse41[] = {
68,15,89,249, //mulps %xmm1,%xmm15
15,40,193, //movaps %xmm1,%xmm0
15,89,192, //mulps %xmm0,%xmm0
- 68,15,40,210, //movaps %xmm2,%xmm10
+ 69,15,40,208, //movaps %xmm8,%xmm10
68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,213, //addps %xmm13,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
69,15,88,212, //addps %xmm12,%xmm10
+ 68,15,89,208, //mulps %xmm0,%xmm10
+ 69,15,88,213, //addps %xmm13,%xmm10
65,15,194,206,1, //cmpltps %xmm14,%xmm1
15,40,193, //movaps %xmm1,%xmm0
102,69,15,56,20,215, //blendvps %xmm0,%xmm15,%xmm10
- 69,15,89,216, //mulps %xmm8,%xmm11
- 65,15,40,192, //movaps %xmm8,%xmm0
+ 68,15,89,218, //mulps %xmm2,%xmm11
+ 15,40,194, //movaps %xmm2,%xmm0
15,89,192, //mulps %xmm0,%xmm0
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,88,213, //addps %xmm13,%xmm2
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,212, //addps %xmm12,%xmm2
- 69,15,194,198,1, //cmpltps %xmm14,%xmm8
- 65,15,40,192, //movaps %xmm8,%xmm0
- 102,65,15,56,20,211, //blendvps %xmm0,%xmm11,%xmm2
+ 68,15,89,194, //mulps %xmm2,%xmm8
+ 69,15,88,196, //addps %xmm12,%xmm8
+ 68,15,89,192, //mulps %xmm0,%xmm8
+ 69,15,88,197, //addps %xmm13,%xmm8
+ 65,15,194,214,1, //cmpltps %xmm14,%xmm2
+ 15,40,194, //movaps %xmm2,%xmm0
+ 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
72,173, //lods %ds:(%rsi),%rax
65,15,40,193, //movaps %xmm9,%xmm0
65,15,40,202, //movaps %xmm10,%xmm1
+ 65,15,40,208, //movaps %xmm8,%xmm2
255,224, //jmpq *%rax
};
@@ -5185,62 +5438,69 @@ CODE const uint8_t sk_to_srgb_sse41[] = {
15,40,245, //movaps %xmm5,%xmm6
15,40,236, //movaps %xmm4,%xmm5
15,40,227, //movaps %xmm3,%xmm4
- 68,15,40,194, //movaps %xmm2,%xmm8
- 15,40,217, //movaps %xmm1,%xmm3
- 15,82,208, //rsqrtps %xmm0,%xmm2
- 68,15,83,202, //rcpps %xmm2,%xmm9
- 68,15,82,210, //rsqrtps %xmm2,%xmm10
- 243,15,16,18, //movss (%rdx),%xmm2
- 243,68,15,16,90,72, //movss 0x48(%rdx),%xmm11
+ 15,40,218, //movaps %xmm2,%xmm3
+ 15,40,209, //movaps %xmm1,%xmm2
+ 68,15,82,192, //rsqrtps %xmm0,%xmm8
+ 69,15,83,200, //rcpps %xmm8,%xmm9
+ 69,15,82,248, //rsqrtps %xmm8,%xmm15
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 65,15,40,203, //movaps %xmm11,%xmm1
- 15,89,200, //mulps %xmm0,%xmm1
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 243,68,15,16,98,76, //movss 0x4c(%rdx),%xmm12
+ 69,15,40,211, //movaps %xmm11,%xmm10
+ 68,15,89,208, //mulps %xmm0,%xmm10
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
+ 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,106,80, //movss 0x50(%rdx),%xmm13
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,114,84, //movss 0x54(%rdx),%xmm14
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 102,68,15,110,240, //movd %eax,%xmm14
69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
69,15,89,205, //mulps %xmm13,%xmm9
69,15,88,206, //addps %xmm14,%xmm9
- 69,15,89,212, //mulps %xmm12,%xmm10
- 69,15,88,209, //addps %xmm9,%xmm10
- 68,15,40,202, //movaps %xmm2,%xmm9
- 69,15,93,202, //minps %xmm10,%xmm9
- 243,68,15,16,122,88, //movss 0x58(%rdx),%xmm15
+ 69,15,89,252, //mulps %xmm12,%xmm15
+ 69,15,88,249, //addps %xmm9,%xmm15
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 69,15,93,207, //minps %xmm15,%xmm9
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 102,68,15,110,248, //movd %eax,%xmm15
69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
65,15,194,199,1, //cmpltps %xmm15,%xmm0
- 102,68,15,56,20,201, //blendvps %xmm0,%xmm1,%xmm9
+ 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
+ 68,15,82,210, //rsqrtps %xmm2,%xmm10
+ 65,15,83,194, //rcpps %xmm10,%xmm0
+ 69,15,82,210, //rsqrtps %xmm10,%xmm10
+ 65,15,89,197, //mulps %xmm13,%xmm0
+ 65,15,88,198, //addps %xmm14,%xmm0
+ 69,15,89,212, //mulps %xmm12,%xmm10
+ 68,15,88,208, //addps %xmm0,%xmm10
+ 65,15,40,200, //movaps %xmm8,%xmm1
+ 65,15,93,202, //minps %xmm10,%xmm1
+ 69,15,40,211, //movaps %xmm11,%xmm10
+ 68,15,89,210, //mulps %xmm2,%xmm10
+ 65,15,194,215,1, //cmpltps %xmm15,%xmm2
+ 15,40,194, //movaps %xmm2,%xmm0
+ 102,65,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm1
15,82,195, //rsqrtps %xmm3,%xmm0
- 15,83,200, //rcpps %xmm0,%xmm1
+ 15,83,208, //rcpps %xmm0,%xmm2
+ 65,15,89,213, //mulps %xmm13,%xmm2
+ 65,15,88,214, //addps %xmm14,%xmm2
15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,205, //mulps %xmm13,%xmm1
- 65,15,88,206, //addps %xmm14,%xmm1
65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,93,208, //minps %xmm0,%xmm10
- 65,15,40,203, //movaps %xmm11,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
+ 15,88,194, //addps %xmm2,%xmm0
+ 68,15,93,192, //minps %xmm0,%xmm8
+ 68,15,89,219, //mulps %xmm3,%xmm11
65,15,194,223,1, //cmpltps %xmm15,%xmm3
15,40,195, //movaps %xmm3,%xmm0
- 102,68,15,56,20,209, //blendvps %xmm0,%xmm1,%xmm10
- 65,15,82,192, //rsqrtps %xmm8,%xmm0
- 15,83,200, //rcpps %xmm0,%xmm1
- 65,15,89,205, //mulps %xmm13,%xmm1
- 65,15,88,206, //addps %xmm14,%xmm1
- 15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 15,93,208, //minps %xmm0,%xmm2
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,194,199,1, //cmpltps %xmm15,%xmm8
- 65,15,40,192, //movaps %xmm8,%xmm0
- 102,65,15,56,20,211, //blendvps %xmm0,%xmm11,%xmm2
+ 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
72,173, //lods %ds:(%rsi),%rax
65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,202, //movaps %xmm10,%xmm1
+ 65,15,40,208, //movaps %xmm8,%xmm2
15,40,220, //movaps %xmm4,%xmm3
15,40,229, //movaps %xmm5,%xmm4
15,40,238, //movaps %xmm6,%xmm5
@@ -5328,32 +5588,38 @@ CODE const uint8_t sk_lerp_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
102,68,15,56,51,4,120, //pmovzxwd (%rax,%rdi,2),%xmm8
- 102,15,110,90,104, //movd 0x68(%rdx),%xmm3
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 243,68,15,16,90,116, //movss 0x74(%rdx),%xmm11
- 243,68,15,16,82,120, //movss 0x78(%rdx),%xmm10
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 102,15,110,90,108, //movd 0x6c(%rdx),%xmm3
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 69,15,89,209, //mulps %xmm9,%xmm10
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,211, //mulps %xmm3,%xmm10
- 102,15,110,90,112, //movd 0x70(%rdx),%xmm3
+ 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,217, //mulps %xmm9,%xmm11
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 243,15,16,90,124, //movss 0x7c(%rdx),%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
15,92,196, //subps %xmm4,%xmm0
- 65,15,89,195, //mulps %xmm11,%xmm0
+ 65,15,89,194, //mulps %xmm10,%xmm0
15,88,196, //addps %xmm4,%xmm0
15,92,205, //subps %xmm5,%xmm1
- 65,15,89,202, //mulps %xmm10,%xmm1
+ 65,15,89,203, //mulps %xmm11,%xmm1
15,88,205, //addps %xmm5,%xmm1
15,92,214, //subps %xmm6,%xmm2
15,89,211, //mulps %xmm3,%xmm2
@@ -5370,7 +5636,8 @@ CODE const uint8_t sk_load_tables_sse41[] = {
72,139,8, //mov (%rax),%rcx
76,139,64,8, //mov 0x8(%rax),%r8
243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 102,15,110,66,16, //movd 0x10(%rdx),%xmm0
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 102,15,110,193, //movd %ecx,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
102,65,15,111,200, //movdqa %xmm8,%xmm1
102,15,114,209,8, //psrld $0x8,%xmm1
@@ -5417,7 +5684,8 @@ CODE const uint8_t sk_load_tables_sse41[] = {
102,15,58,33,211,48, //insertps $0x30,%xmm3,%xmm2
102,65,15,114,208,24, //psrld $0x18,%xmm8
69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -5429,7 +5697,8 @@ CODE const uint8_t sk_load_a8_sse41[] = {
72,139,0, //mov (%rax),%rax
102,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm0
15,91,192, //cvtdq2ps %xmm0,%xmm0
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -5442,7 +5711,8 @@ CODE const uint8_t sk_load_a8_sse41[] = {
CODE const uint8_t sk_store_a8_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,66,8, //movss 0x8(%rdx),%xmm8
+ 185,0,0,127,67, //mov $0x437f0000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
68,15,89,195, //mulps %xmm3,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
@@ -5456,29 +5726,36 @@ CODE const uint8_t sk_store_a8_sse41[] = {
CODE const uint8_t sk_load_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,56,51,12,120, //pmovzxwd (%rax,%rdi,2),%xmm9
- 102,15,110,66,104, //movd 0x68(%rdx),%xmm0
+ 102,15,56,51,20,120, //pmovzxwd (%rax,%rdi,2),%xmm2
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,219,193, //pand %xmm9,%xmm0
+ 102,15,219,194, //pand %xmm2,%xmm0
15,91,200, //cvtdq2ps %xmm0,%xmm1
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,15,16,66,116, //movss 0x74(%rdx),%xmm0
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,15,110,192, //movd %eax,%xmm0
15,198,192,0, //shufps $0x0,%xmm0,%xmm0
15,89,193, //mulps %xmm1,%xmm0
- 102,15,110,74,108, //movd 0x6c(%rdx),%xmm1
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,200, //movd %eax,%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,65,15,219,201, //pand %xmm9,%xmm1
- 68,15,91,193, //cvtdq2ps %xmm1,%xmm8
- 243,15,16,74,120, //movss 0x78(%rdx),%xmm1
+ 102,15,219,202, //pand %xmm2,%xmm1
+ 15,91,217, //cvtdq2ps %xmm1,%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,15,110,200, //movd %eax,%xmm1
15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,110,82,112, //movd 0x70(%rdx),%xmm2
- 102,15,112,210,0, //pshufd $0x0,%xmm2,%xmm2
- 102,65,15,219,209, //pand %xmm9,%xmm2
- 68,15,91,194, //cvtdq2ps %xmm2,%xmm8
- 243,15,16,82,124, //movss 0x7c(%rdx),%xmm2
+ 15,89,203, //mulps %xmm3,%xmm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,15,219,218, //pand %xmm2,%xmm3
+ 15,91,219, //cvtdq2ps %xmm3,%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,208, //movd %eax,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
+ 15,89,211, //mulps %xmm3,%xmm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -5487,21 +5764,23 @@ CODE const uint8_t sk_load_565_sse41[] = {
CODE const uint8_t sk_store_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,130,128,0,0,0, //movss 0x80(%rdx),%xmm8
- 243,68,15,16,138,132,0,0,0, //movss 0x84(%rdx),%xmm9
+ 185,0,0,248,65, //mov $0x41f80000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,11, //pslld $0xb,%xmm10
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,5, //pslld $0x5,%xmm9
- 102,69,15,235,202, //por %xmm10,%xmm9
+ 102,65,15,114,241,11, //pslld $0xb,%xmm9
+ 185,0,0,124,66, //mov $0x427c0000,%ecx
+ 102,68,15,110,209, //movd %ecx,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 68,15,89,209, //mulps %xmm1,%xmm10
+ 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
+ 102,65,15,114,242,5, //pslld $0x5,%xmm10
+ 102,69,15,235,209, //por %xmm9,%xmm10
68,15,89,194, //mulps %xmm2,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,193, //orpd %xmm9,%xmm8
+ 102,69,15,86,194, //orpd %xmm10,%xmm8
102,69,15,56,43,192, //packusdw %xmm8,%xmm8
102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
@@ -5577,7 +5856,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
102,68,15,111,194, //movdqa %xmm2,%xmm8
102,68,15,97,192, //punpcklwd %xmm0,%xmm8
102,15,105,208, //punpckhwd %xmm0,%xmm2
- 102,15,110,66,100, //movd 0x64(%rdx),%xmm0
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
102,15,111,203, //movdqa %xmm3,%xmm1
102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
@@ -5586,7 +5866,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
102,15,223,218, //pandn %xmm2,%xmm3
102,15,56,51,193, //pmovzxwd %xmm1,%xmm0
102,15,114,240,13, //pslld $0xd,%xmm0
- 102,15,110,82,92, //movd 0x5c(%rdx),%xmm2
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 102,15,110,208, //movd %eax,%xmm2
102,68,15,112,194,0, //pshufd $0x0,%xmm2,%xmm8
65,15,89,192, //mulps %xmm8,%xmm0
102,69,15,239,201, //pxor %xmm9,%xmm9
@@ -5606,7 +5887,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
CODE const uint8_t sk_store_f16_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,110,66,96, //movd 0x60(%rdx),%xmm8
+ 185,0,0,128,7, //mov $0x7800000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
102,69,15,111,200, //movdqa %xmm8,%xmm9
68,15,89,200, //mulps %xmm0,%xmm9
@@ -5768,17 +6050,20 @@ CODE const uint8_t sk_mirror_y_sse41[] = {
};
CODE const uint8_t sk_luminance_to_alpha_sse41[] = {
- 243,15,16,154,136,0,0,0, //movss 0x88(%rdx),%xmm3
- 243,68,15,16,130,140,0,0,0, //movss 0x8c(%rdx),%xmm8
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,193, //mulps %xmm1,%xmm8
- 68,15,88,195, //addps %xmm3,%xmm8
- 243,15,16,154,144,0,0,0, //movss 0x90(%rdx),%xmm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 102,15,110,192, //movd %eax,%xmm0
+ 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
+ 15,89,193, //mulps %xmm1,%xmm0
+ 15,88,195, //addps %xmm3,%xmm0
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,218, //mulps %xmm2,%xmm3
- 65,15,88,216, //addps %xmm8,%xmm3
+ 15,88,216, //addps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
15,87,192, //xorps %xmm0,%xmm0
15,87,201, //xorps %xmm1,%xmm1
@@ -6074,7 +6359,7 @@ CODE const uint8_t sk_seed_shader_sse2[] = {
102,15,110,209, //movd %ecx,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
15,88,202, //addps %xmm2,%xmm1
- 15,16,66,20, //movups 0x14(%rdx),%xmm0
+ 15,16,2, //movups (%rdx),%xmm0
15,88,193, //addps %xmm1,%xmm0
102,15,110,8, //movd (%rax),%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
@@ -6274,24 +6559,29 @@ CODE const uint8_t sk_unpremul_sse2[] = {
};
CODE const uint8_t sk_from_srgb_sse2[] = {
- 243,68,15,16,66,64, //movss 0x40(%rdx),%xmm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
69,15,40,232, //movaps %xmm8,%xmm13
68,15,89,232, //mulps %xmm0,%xmm13
68,15,40,224, //movaps %xmm0,%xmm12
69,15,89,228, //mulps %xmm12,%xmm12
- 243,68,15,16,74,60, //movss 0x3c(%rdx),%xmm9
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 102,68,15,110,200, //movd %eax,%xmm9
69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,82,52, //movss 0x34(%rdx),%xmm10
- 243,68,15,16,90,56, //movss 0x38(%rdx),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
69,15,40,241, //movaps %xmm9,%xmm14
68,15,89,240, //mulps %xmm0,%xmm14
- 69,15,88,243, //addps %xmm11,%xmm14
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,244, //mulps %xmm12,%xmm14
69,15,88,242, //addps %xmm10,%xmm14
- 243,68,15,16,98,68, //movss 0x44(%rdx),%xmm12
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,244, //mulps %xmm12,%xmm14
+ 69,15,88,243, //addps %xmm11,%xmm14
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
65,15,194,196,1, //cmpltps %xmm12,%xmm0
68,15,84,232, //andps %xmm0,%xmm13
@@ -6303,9 +6593,9 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
69,15,89,246, //mulps %xmm14,%xmm14
69,15,40,249, //movaps %xmm9,%xmm15
68,15,89,249, //mulps %xmm1,%xmm15
- 69,15,88,251, //addps %xmm11,%xmm15
- 69,15,89,254, //mulps %xmm14,%xmm15
69,15,88,250, //addps %xmm10,%xmm15
+ 69,15,89,254, //mulps %xmm14,%xmm15
+ 69,15,88,251, //addps %xmm11,%xmm15
65,15,194,204,1, //cmpltps %xmm12,%xmm1
68,15,84,233, //andps %xmm1,%xmm13
65,15,85,207, //andnps %xmm15,%xmm1
@@ -6314,9 +6604,9 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
68,15,40,234, //movaps %xmm2,%xmm13
69,15,89,237, //mulps %xmm13,%xmm13
68,15,89,202, //mulps %xmm2,%xmm9
- 69,15,88,203, //addps %xmm11,%xmm9
- 69,15,89,205, //mulps %xmm13,%xmm9
69,15,88,202, //addps %xmm10,%xmm9
+ 69,15,89,205, //mulps %xmm13,%xmm9
+ 69,15,88,203, //addps %xmm11,%xmm9
65,15,194,212,1, //cmpltps %xmm12,%xmm2
68,15,84,194, //andps %xmm2,%xmm8
65,15,85,209, //andnps %xmm9,%xmm2
@@ -6326,74 +6616,69 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
};
CODE const uint8_t sk_to_srgb_sse2[] = {
- 72,131,236,40, //sub $0x28,%rsp
- 15,41,124,36,16, //movaps %xmm7,0x10(%rsp)
- 15,41,52,36, //movaps %xmm6,(%rsp)
- 15,40,245, //movaps %xmm5,%xmm6
- 15,40,236, //movaps %xmm4,%xmm5
- 15,40,227, //movaps %xmm3,%xmm4
68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,232, //rcpps %xmm8,%xmm13
- 69,15,82,248, //rsqrtps %xmm8,%xmm15
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,68,15,16,66,72, //movss 0x48(%rdx),%xmm8
+ 69,15,83,248, //rcpps %xmm8,%xmm15
+ 69,15,82,232, //rsqrtps %xmm8,%xmm13
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
69,15,40,240, //movaps %xmm8,%xmm14
68,15,89,240, //mulps %xmm0,%xmm14
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 243,68,15,16,82,76, //movss 0x4c(%rdx),%xmm10
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,68,15,110,200, //movd %eax,%xmm9
+ 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,90,80, //movss 0x50(%rdx),%xmm11
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,98,84, //movss 0x54(%rdx),%xmm12
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,235, //mulps %xmm11,%xmm13
- 69,15,88,236, //addps %xmm12,%xmm13
- 69,15,89,250, //mulps %xmm10,%xmm15
- 69,15,88,253, //addps %xmm13,%xmm15
- 68,15,40,203, //movaps %xmm3,%xmm9
- 69,15,93,207, //minps %xmm15,%xmm9
- 243,68,15,16,106,88, //movss 0x58(%rdx),%xmm13
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,89,234, //mulps %xmm10,%xmm13
+ 69,15,88,239, //addps %xmm15,%xmm13
+ 69,15,40,249, //movaps %xmm9,%xmm15
+ 69,15,93,253, //minps %xmm13,%xmm15
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
65,15,194,197,1, //cmpltps %xmm13,%xmm0
68,15,84,240, //andps %xmm0,%xmm14
- 65,15,85,193, //andnps %xmm9,%xmm0
+ 65,15,85,199, //andnps %xmm15,%xmm0
65,15,86,198, //orps %xmm14,%xmm0
- 68,15,82,201, //rsqrtps %xmm1,%xmm9
- 69,15,83,241, //rcpps %xmm9,%xmm14
- 69,15,82,201, //rsqrtps %xmm9,%xmm9
- 69,15,89,243, //mulps %xmm11,%xmm14
- 69,15,88,244, //addps %xmm12,%xmm14
- 69,15,89,202, //mulps %xmm10,%xmm9
- 69,15,88,206, //addps %xmm14,%xmm9
- 68,15,40,243, //movaps %xmm3,%xmm14
- 69,15,93,241, //minps %xmm9,%xmm14
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 68,15,82,241, //rsqrtps %xmm1,%xmm14
+ 69,15,83,254, //rcpps %xmm14,%xmm15
+ 69,15,82,246, //rsqrtps %xmm14,%xmm14
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,89,242, //mulps %xmm10,%xmm14
+ 69,15,88,247, //addps %xmm15,%xmm14
+ 69,15,40,249, //movaps %xmm9,%xmm15
+ 69,15,93,254, //minps %xmm14,%xmm15
+ 69,15,40,240, //movaps %xmm8,%xmm14
+ 68,15,89,241, //mulps %xmm1,%xmm14
65,15,194,205,1, //cmpltps %xmm13,%xmm1
- 68,15,84,201, //andps %xmm1,%xmm9
- 65,15,85,206, //andnps %xmm14,%xmm1
- 65,15,86,201, //orps %xmm9,%xmm1
- 68,15,82,202, //rsqrtps %xmm2,%xmm9
- 69,15,83,241, //rcpps %xmm9,%xmm14
- 69,15,89,243, //mulps %xmm11,%xmm14
- 69,15,88,244, //addps %xmm12,%xmm14
- 65,15,82,249, //rsqrtps %xmm9,%xmm7
- 65,15,89,250, //mulps %xmm10,%xmm7
- 65,15,88,254, //addps %xmm14,%xmm7
- 15,93,223, //minps %xmm7,%xmm3
+ 68,15,84,241, //andps %xmm1,%xmm14
+ 65,15,85,207, //andnps %xmm15,%xmm1
+ 65,15,86,206, //orps %xmm14,%xmm1
+ 68,15,82,242, //rsqrtps %xmm2,%xmm14
+ 69,15,83,254, //rcpps %xmm14,%xmm15
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,82,222, //rsqrtps %xmm14,%xmm11
+ 69,15,89,218, //mulps %xmm10,%xmm11
+ 69,15,88,223, //addps %xmm15,%xmm11
+ 69,15,93,203, //minps %xmm11,%xmm9
68,15,89,194, //mulps %xmm2,%xmm8
65,15,194,213,1, //cmpltps %xmm13,%xmm2
68,15,84,194, //andps %xmm2,%xmm8
- 15,85,211, //andnps %xmm3,%xmm2
+ 65,15,85,209, //andnps %xmm9,%xmm2
65,15,86,208, //orps %xmm8,%xmm2
72,173, //lods %ds:(%rsi),%rax
- 15,40,220, //movaps %xmm4,%xmm3
- 15,40,229, //movaps %xmm5,%xmm4
- 15,40,238, //movaps %xmm6,%xmm5
- 15,40,52,36, //movaps (%rsp),%xmm6
- 15,40,124,36,16, //movaps 0x10(%rsp),%xmm7
- 72,131,196,40, //add $0x28,%rsp
255,224, //jmpq *%rax
};
@@ -6480,35 +6765,41 @@ CODE const uint8_t sk_lerp_u8_sse2[] = {
CODE const uint8_t sk_lerp_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,126,12,120, //movq (%rax,%rdi,2),%xmm9
+ 243,68,15,126,4,120, //movq (%rax,%rdi,2),%xmm8
102,15,239,219, //pxor %xmm3,%xmm3
- 102,68,15,97,203, //punpcklwd %xmm3,%xmm9
- 102,15,110,90,104, //movd 0x68(%rdx),%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
- 68,15,91,211, //cvtdq2ps %xmm3,%xmm10
- 243,68,15,16,90,116, //movss 0x74(%rdx),%xmm11
- 243,68,15,16,66,120, //movss 0x78(%rdx),%xmm8
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,218, //mulps %xmm10,%xmm11
- 102,15,110,90,108, //movd 0x6c(%rdx),%xmm3
+ 102,68,15,97,195, //punpcklwd %xmm3,%xmm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,15,110,90,112, //movd 0x70(%rdx),%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
+ 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 69,15,89,209, //mulps %xmm9,%xmm10
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 243,15,16,90,124, //movss 0x7c(%rdx),%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,217, //mulps %xmm9,%xmm11
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
+ 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
+ 65,15,89,216, //mulps %xmm8,%xmm3
15,92,196, //subps %xmm4,%xmm0
- 65,15,89,195, //mulps %xmm11,%xmm0
+ 65,15,89,194, //mulps %xmm10,%xmm0
15,88,196, //addps %xmm4,%xmm0
15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
+ 65,15,89,203, //mulps %xmm11,%xmm1
15,88,205, //addps %xmm5,%xmm1
15,92,214, //subps %xmm6,%xmm2
15,89,211, //mulps %xmm3,%xmm2
@@ -6525,7 +6816,8 @@ CODE const uint8_t sk_load_tables_sse2[] = {
72,139,8, //mov (%rax),%rcx
76,139,64,8, //mov 0x8(%rax),%r8
243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 102,15,110,66,16, //movd 0x10(%rdx),%xmm0
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 102,15,110,193, //movd %ecx,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
102,69,15,111,200, //movdqa %xmm8,%xmm9
102,65,15,114,209,8, //psrld $0x8,%xmm9
@@ -6580,7 +6872,8 @@ CODE const uint8_t sk_load_tables_sse2[] = {
65,15,20,209, //unpcklps %xmm9,%xmm2
102,65,15,114,208,24, //psrld $0x18,%xmm8
69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -6595,7 +6888,8 @@ CODE const uint8_t sk_load_a8_sse2[] = {
102,15,96,193, //punpcklbw %xmm1,%xmm0
102,15,97,193, //punpcklwd %xmm1,%xmm0
15,91,192, //cvtdq2ps %xmm0,%xmm0
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -6608,7 +6902,8 @@ CODE const uint8_t sk_load_a8_sse2[] = {
CODE const uint8_t sk_store_a8_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,66,8, //movss 0x8(%rdx),%xmm8
+ 185,0,0,127,67, //mov $0x437f0000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
68,15,89,195, //mulps %xmm3,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
@@ -6624,31 +6919,38 @@ CODE const uint8_t sk_store_a8_sse2[] = {
CODE const uint8_t sk_load_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,126,12,120, //movq (%rax,%rdi,2),%xmm9
+ 243,15,126,20,120, //movq (%rax,%rdi,2),%xmm2
102,15,239,192, //pxor %xmm0,%xmm0
- 102,68,15,97,200, //punpcklwd %xmm0,%xmm9
- 102,15,110,66,104, //movd 0x68(%rdx),%xmm0
+ 102,15,97,208, //punpcklwd %xmm0,%xmm2
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,219,193, //pand %xmm9,%xmm0
+ 102,15,219,194, //pand %xmm2,%xmm0
15,91,200, //cvtdq2ps %xmm0,%xmm1
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,15,16,66,116, //movss 0x74(%rdx),%xmm0
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,15,110,192, //movd %eax,%xmm0
15,198,192,0, //shufps $0x0,%xmm0,%xmm0
15,89,193, //mulps %xmm1,%xmm0
- 102,15,110,74,108, //movd 0x6c(%rdx),%xmm1
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,200, //movd %eax,%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,65,15,219,201, //pand %xmm9,%xmm1
- 68,15,91,193, //cvtdq2ps %xmm1,%xmm8
- 243,15,16,74,120, //movss 0x78(%rdx),%xmm1
+ 102,15,219,202, //pand %xmm2,%xmm1
+ 15,91,217, //cvtdq2ps %xmm1,%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,15,110,200, //movd %eax,%xmm1
15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,110,82,112, //movd 0x70(%rdx),%xmm2
- 102,15,112,210,0, //pshufd $0x0,%xmm2,%xmm2
- 102,65,15,219,209, //pand %xmm9,%xmm2
- 68,15,91,194, //cvtdq2ps %xmm2,%xmm8
- 243,15,16,82,124, //movss 0x7c(%rdx),%xmm2
+ 15,89,203, //mulps %xmm3,%xmm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,15,219,218, //pand %xmm2,%xmm3
+ 15,91,219, //cvtdq2ps %xmm3,%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,208, //movd %eax,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
+ 15,89,211, //mulps %xmm3,%xmm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -6657,21 +6959,23 @@ CODE const uint8_t sk_load_565_sse2[] = {
CODE const uint8_t sk_store_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,130,128,0,0,0, //movss 0x80(%rdx),%xmm8
- 243,68,15,16,138,132,0,0,0, //movss 0x84(%rdx),%xmm9
+ 185,0,0,248,65, //mov $0x41f80000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,11, //pslld $0xb,%xmm10
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,5, //pslld $0x5,%xmm9
- 102,69,15,235,202, //por %xmm10,%xmm9
+ 102,65,15,114,241,11, //pslld $0xb,%xmm9
+ 185,0,0,124,66, //mov $0x427c0000,%ecx
+ 102,68,15,110,209, //movd %ecx,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 68,15,89,209, //mulps %xmm1,%xmm10
+ 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
+ 102,65,15,114,242,5, //pslld $0x5,%xmm10
+ 102,69,15,235,209, //por %xmm9,%xmm10
68,15,89,194, //mulps %xmm2,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,193, //orpd %xmm9,%xmm8
+ 102,69,15,86,194, //orpd %xmm10,%xmm8
102,65,15,114,240,16, //pslld $0x10,%xmm8
102,65,15,114,224,16, //psrad $0x10,%xmm8
102,69,15,107,192, //packssdw %xmm8,%xmm8
@@ -6749,7 +7053,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
102,68,15,111,194, //movdqa %xmm2,%xmm8
102,68,15,97,192, //punpcklwd %xmm0,%xmm8
102,15,105,208, //punpckhwd %xmm0,%xmm2
- 102,15,110,66,100, //movd 0x64(%rdx),%xmm0
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
102,15,111,203, //movdqa %xmm3,%xmm1
102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
@@ -6760,7 +7065,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
102,15,111,193, //movdqa %xmm1,%xmm0
102,65,15,97,192, //punpcklwd %xmm8,%xmm0
102,15,114,240,13, //pslld $0xd,%xmm0
- 102,15,110,82,92, //movd 0x5c(%rdx),%xmm2
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 102,15,110,208, //movd %eax,%xmm2
102,68,15,112,202,0, //pshufd $0x0,%xmm2,%xmm9
65,15,89,193, //mulps %xmm9,%xmm0
102,65,15,105,200, //punpckhwd %xmm8,%xmm1
@@ -6780,7 +7086,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
CODE const uint8_t sk_store_f16_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,110,66,96, //movd 0x60(%rdx),%xmm8
+ 185,0,0,128,7, //mov $0x7800000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
102,69,15,111,200, //movdqa %xmm8,%xmm9
68,15,89,200, //mulps %xmm0,%xmm9
@@ -6970,17 +7277,20 @@ CODE const uint8_t sk_mirror_y_sse2[] = {
};
CODE const uint8_t sk_luminance_to_alpha_sse2[] = {
- 243,15,16,154,136,0,0,0, //movss 0x88(%rdx),%xmm3
- 243,68,15,16,130,140,0,0,0, //movss 0x8c(%rdx),%xmm8
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,193, //mulps %xmm1,%xmm8
- 68,15,88,195, //addps %xmm3,%xmm8
- 243,15,16,154,144,0,0,0, //movss 0x90(%rdx),%xmm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 102,15,110,192, //movd %eax,%xmm0
+ 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
+ 15,89,193, //mulps %xmm1,%xmm0
+ 15,88,195, //addps %xmm3,%xmm0
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,218, //mulps %xmm2,%xmm3
- 65,15,88,216, //addps %xmm8,%xmm3
+ 15,88,216, //addps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
15,87,192, //xorps %xmm0,%xmm0
15,87,201, //xorps %xmm1,%xmm1
@@ -7321,7 +7631,7 @@ CODE const uint8_t sk_seed_shader_hsw[] = {
196,193,121,110,200, //vmovd %r8d,%xmm1
196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,66,20, //vaddps 0x14(%rdx),%ymm0,%ymm0
+ 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
196,226,125,24,16, //vbroadcastss (%rax),%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
@@ -7501,16 +7811,26 @@ CODE const uint8_t sk_unpremul_hsw[] = {
};
CODE const uint8_t sk_from_srgb_hsw[] = {
- 196,98,125,24,66,64, //vbroadcastss 0x40(%rdx),%ymm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 196,98,125,24,90,60, //vbroadcastss 0x3c(%rdx),%ymm11
- 196,98,125,24,98,56, //vbroadcastss 0x38(%rdx),%ymm12
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,66,125,24,219, //vbroadcastss %xmm11,%ymm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,66,125,24,228, //vbroadcastss %xmm12,%ymm12
196,65,124,40,235, //vmovaps %ymm11,%ymm13
196,66,125,168,236, //vfmadd213ps %ymm12,%ymm0,%ymm13
- 196,98,125,24,114,52, //vbroadcastss 0x34(%rdx),%ymm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,66,125,24,246, //vbroadcastss %xmm14,%ymm14
196,66,45,168,238, //vfmadd213ps %ymm14,%ymm10,%ymm13
- 196,98,125,24,82,68, //vbroadcastss 0x44(%rdx),%ymm10
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
196,195,21,74,193,0, //vblendvps %ymm0,%ymm9,%ymm13,%ymm0
197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
@@ -7532,37 +7852,50 @@ CODE const uint8_t sk_from_srgb_hsw[] = {
CODE const uint8_t sk_to_srgb_hsw[] = {
197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,200, //vrcpps %ymm8,%ymm9
- 196,65,124,82,208, //vrsqrtps %ymm8,%ymm10
- 196,98,125,24,66,72, //vbroadcastss 0x48(%rdx),%ymm8
- 197,60,89,216, //vmulps %ymm0,%ymm8,%ymm11
- 196,98,125,24,34, //vbroadcastss (%rdx),%ymm12
- 196,98,125,24,106,76, //vbroadcastss 0x4c(%rdx),%ymm13
- 196,98,125,24,114,80, //vbroadcastss 0x50(%rdx),%ymm14
- 196,98,125,24,122,84, //vbroadcastss 0x54(%rdx),%ymm15
- 196,66,13,168,207, //vfmadd213ps %ymm15,%ymm14,%ymm9
- 196,66,21,184,202, //vfmadd231ps %ymm10,%ymm13,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,98,125,24,82,88, //vbroadcastss 0x58(%rdx),%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,53,74,195,0, //vblendvps %ymm0,%ymm11,%ymm9,%ymm0
- 197,124,82,201, //vrsqrtps %ymm1,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,66,21,184,217, //vfmadd231ps %ymm9,%ymm13,%ymm11
- 197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
- 196,65,28,93,219, //vminps %ymm11,%ymm12,%ymm11
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,37,74,201,16, //vblendvps %ymm1,%ymm9,%ymm11,%ymm1
- 197,124,82,202, //vrsqrtps %ymm2,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
+ 196,65,124,83,216, //vrcpps %ymm8,%ymm11
+ 196,65,124,82,224, //vrsqrtps %ymm8,%ymm12
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
+ 197,60,89,232, //vmulps %ymm0,%ymm8,%ymm13
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,121,110,200, //vmovd %eax,%xmm9
+ 196,66,125,24,201, //vbroadcastss %xmm9,%ymm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,66,125,24,246, //vbroadcastss %xmm14,%ymm14
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 197,121,110,248, //vmovd %eax,%xmm15
+ 196,66,125,24,255, //vbroadcastss %xmm15,%ymm15
196,66,13,168,223, //vfmadd213ps %ymm15,%ymm14,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,66,21,184,217, //vfmadd231ps %ymm9,%ymm13,%ymm11
- 196,65,28,93,203, //vminps %ymm11,%ymm12,%ymm9
+ 196,66,45,184,220, //vfmadd231ps %ymm12,%ymm10,%ymm11
+ 196,65,52,93,219, //vminps %ymm11,%ymm9,%ymm11
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,66,125,24,228, //vbroadcastss %xmm12,%ymm12
+ 196,193,124,194,196,1, //vcmpltps %ymm12,%ymm0,%ymm0
+ 196,195,37,74,197,0, //vblendvps %ymm0,%ymm13,%ymm11,%ymm0
+ 197,124,82,217, //vrsqrtps %ymm1,%ymm11
+ 196,65,124,83,235, //vrcpps %ymm11,%ymm13
+ 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
+ 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
+ 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
+ 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
+ 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
+ 196,193,116,194,204,1, //vcmpltps %ymm12,%ymm1,%ymm1
+ 196,195,21,74,203,16, //vblendvps %ymm1,%ymm11,%ymm13,%ymm1
+ 197,124,82,218, //vrsqrtps %ymm2,%ymm11
+ 196,65,124,83,235, //vrcpps %ymm11,%ymm13
+ 196,66,13,168,239, //vfmadd213ps %ymm15,%ymm14,%ymm13
+ 196,65,124,82,219, //vrsqrtps %ymm11,%ymm11
+ 196,66,45,184,235, //vfmadd231ps %ymm11,%ymm10,%ymm13
+ 196,65,52,93,205, //vminps %ymm13,%ymm9,%ymm9
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
+ 196,193,108,194,212,1, //vcmpltps %ymm12,%ymm2,%ymm2
196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -7585,7 +7918,7 @@ CODE const uint8_t sk_scale_u8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,56, //jne 4f9 <_sk_scale_u8_hsw+0x48>
+ 117,56, //jne 556 <_sk_scale_u8_hsw+0x48>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
@@ -7609,9 +7942,9 @@ CODE const uint8_t sk_scale_u8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 501 <_sk_scale_u8_hsw+0x50>
+ 117,234, //jne 55e <_sk_scale_u8_hsw+0x50>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,167, //jmp 4c5 <_sk_scale_u8_hsw+0x14>
+ 235,167, //jmp 522 <_sk_scale_u8_hsw+0x14>
};
CODE const uint8_t sk_lerp_1_float_hsw[] = {
@@ -7635,7 +7968,7 @@ CODE const uint8_t sk_lerp_u8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,76, //jne 5a9 <_sk_lerp_u8_hsw+0x5c>
+ 117,76, //jne 606 <_sk_lerp_u8_hsw+0x5c>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,125,49,192, //vpmovzxbd %xmm8,%ymm8
196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
@@ -7663,37 +7996,49 @@ CODE const uint8_t sk_lerp_u8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 5b1 <_sk_lerp_u8_hsw+0x64>
+ 117,234, //jne 60e <_sk_lerp_u8_hsw+0x64>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,147, //jmp 561 <_sk_lerp_u8_hsw+0x14>
+ 235,147, //jmp 5be <_sk_lerp_u8_hsw+0x14>
};
CODE const uint8_t sk_lerp_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,132,0,0,0, //jne 660 <_sk_lerp_565_hsw+0x92>
+ 15,133,179,0,0,0, //jne 6ec <_sk_lerp_565_hsw+0xc1>
196,193,122,111,28,122, //vmovdqu (%r10,%rdi,2),%xmm3
- 196,226,125,51,219, //vpmovzxwd %xmm3,%ymm3
- 196,98,125,88,66,104, //vpbroadcastd 0x68(%rdx),%ymm8
- 197,61,219,195, //vpand %ymm3,%ymm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 196,98,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm9
- 196,65,52,89,192, //vmulps %ymm8,%ymm9,%ymm8
- 196,98,125,88,74,108, //vpbroadcastd 0x6c(%rdx),%ymm9
- 197,53,219,203, //vpand %ymm3,%ymm9,%ymm9
- 196,65,124,91,201, //vcvtdq2ps %ymm9,%ymm9
- 196,98,125,24,82,120, //vbroadcastss 0x78(%rdx),%ymm10
- 196,65,44,89,201, //vmulps %ymm9,%ymm10,%ymm9
- 196,98,125,88,82,112, //vpbroadcastd 0x70(%rdx),%ymm10
- 197,173,219,219, //vpand %ymm3,%ymm10,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,82,124, //vbroadcastss 0x7c(%rdx),%ymm10
- 197,172,89,219, //vmulps %ymm3,%ymm10,%ymm3
+ 196,98,125,51,195, //vpmovzxwd %xmm3,%ymm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
+ 196,193,101,219,216, //vpand %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,226,61,168,196, //vfmadd213ps %ymm4,%ymm8,%ymm0
+ 196,226,53,168,196, //vfmadd213ps %ymm4,%ymm9,%ymm0
197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,226,53,168,205, //vfmadd213ps %ymm5,%ymm9,%ymm1
+ 196,226,45,168,205, //vfmadd213ps %ymm5,%ymm10,%ymm1
197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
196,226,101,168,214, //vfmadd213ps %ymm6,%ymm3,%ymm2
184,0,0,128,63, //mov $0x3f800000,%eax
@@ -7707,8 +8052,8 @@ CODE const uint8_t sk_lerp_565_hsw[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,102,255,255,255, //ja 5e2 <_sk_lerp_565_hsw+0x14>
- 76,141,13,73,0,0,0, //lea 0x49(%rip),%r9 # 6cc <_sk_lerp_565_hsw+0xfe>
+ 15,135,55,255,255,255, //ja 63f <_sk_lerp_565_hsw+0x14>
+ 76,141,13,73,0,0,0, //lea 0x49(%rip),%r9 # 758 <_sk_lerp_565_hsw+0x12d>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -7720,7 +8065,7 @@ CODE const uint8_t sk_lerp_565_hsw[] = {
196,193,97,196,92,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm3,%xmm3
196,193,97,196,92,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm3,%xmm3
196,193,97,196,28,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm3,%xmm3
- 233,22,255,255,255, //jmpq 5e2 <_sk_lerp_565_hsw+0x14>
+ 233,231,254,255,255, //jmpq 63f <_sk_lerp_565_hsw+0x14>
244, //hlt
255, //(bad)
255, //(bad)
@@ -7752,9 +8097,11 @@ CODE const uint8_t sk_load_tables_hsw[] = {
76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
76,3,8, //add (%rax),%r9
77,133,192, //test %r8,%r8
- 117,106, //jne 767 <_sk_load_tables_hsw+0x7f>
+ 117,121, //jne 802 <_sk_load_tables_hsw+0x8e>
196,193,126,111,25, //vmovdqu (%r9),%ymm3
- 196,226,125,88,82,16, //vpbroadcastd 0x10(%rdx),%ymm2
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 197,249,110,193, //vmovd %ecx,%xmm0
+ 196,226,125,88,208, //vpbroadcastd %xmm0,%ymm2
197,237,219,203, //vpand %ymm3,%ymm2,%ymm1
196,65,61,118,192, //vpcmpeqd %ymm8,%ymm8,%ymm8
72,139,72,8, //mov 0x8(%rax),%rcx
@@ -7770,9 +8117,11 @@ CODE const uint8_t sk_load_tables_hsw[] = {
196,65,109,219,201, //vpand %ymm9,%ymm2,%ymm9
196,162,61,146,20,136, //vgatherdps %ymm8,(%rax,%ymm9,4),%ymm2
197,229,114,211,24, //vpsrld $0x18,%ymm3,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,66,12, //vbroadcastss 0xc(%rdx),%ymm8
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
72,173, //lods %ds:(%rsi),%rax
76,137,193, //mov %r8,%rcx
255,224, //jmpq *%rax
@@ -7784,7 +8133,7 @@ CODE const uint8_t sk_load_tables_hsw[] = {
196,193,249,110,194, //vmovq %r10,%xmm0
196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,114,255,255,255, //jmpq 702 <_sk_load_tables_hsw+0x1a>
+ 233,99,255,255,255, //jmpq 78e <_sk_load_tables_hsw+0x1a>
};
CODE const uint8_t sk_load_a8_hsw[] = {
@@ -7793,11 +8142,13 @@ CODE const uint8_t sk_load_a8_hsw[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,42, //jne 7ca <_sk_load_a8_hsw+0x3a>
+ 117,50, //jne 86d <_sk_load_a8_hsw+0x42>
197,251,16,0, //vmovsd (%rax),%xmm0
196,226,125,49,192, //vpmovzxbd %xmm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,12, //vbroadcastss 0xc(%rdx),%ymm1
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -7814,22 +8165,24 @@ CODE const uint8_t sk_load_a8_hsw[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 7d2 <_sk_load_a8_hsw+0x42>
+ 117,234, //jne 875 <_sk_load_a8_hsw+0x4a>
196,193,249,110,193, //vmovq %r9,%xmm0
- 235,181, //jmp 7a4 <_sk_load_a8_hsw+0x14>
+ 235,173, //jmp 83f <_sk_load_a8_hsw+0x14>
};
CODE const uint8_t sk_store_a8_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,66,8, //vbroadcastss 0x8(%rdx),%ymm8
+ 184,0,0,127,67, //mov $0x437f0000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne 822 <_sk_store_a8_hsw+0x33>
+ 117,10, //jne 8cd <_sk_store_a8_hsw+0x3b>
196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -7838,9 +8191,9 @@ CODE const uint8_t sk_store_a8_hsw[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 81e <_sk_store_a8_hsw+0x2f>
+ 119,236, //ja 8c9 <_sk_store_a8_hsw+0x37>
196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 76,141,21,66,0,0,0, //lea 0x42(%rip),%r10 # 880 <_sk_store_a8_hsw+0x91>
+ 76,141,21,67,0,0,0, //lea 0x43(%rip),%r10 # 92c <_sk_store_a8_hsw+0x9a>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -7851,26 +8204,27 @@ CODE const uint8_t sk_store_a8_hsw[] = {
196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,158, //jmp 81e <_sk_store_a8_hsw+0x2f>
- 247,255, //idiv %edi
+ 235,158, //jmp 8c9 <_sk_store_a8_hsw+0x37>
+ 144, //nop
+ 246,255, //idiv %bh
255, //(bad)
255, //(bad)
- 239, //out %eax,(%dx)
+ 238, //out %al,(%dx)
255, //(bad)
255, //(bad)
- 255,231, //jmpq *%rdi
+ 255,230, //jmpq *%rsi
255, //(bad)
255, //(bad)
255, //(bad)
- 223,255, //(bad)
+ 222,255, //fdivrp %st,%st(7)
255, //(bad)
- 255,215, //callq *%rdi
+ 255,214, //callq *%rsi
255, //(bad)
255, //(bad)
- 255,207, //dec %edi
+ 255,206, //dec %esi
255, //(bad)
255, //(bad)
- 255,199, //inc %edi
+ 255,198, //inc %esi
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -7880,25 +8234,39 @@ CODE const uint8_t sk_load_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 117,92, //jne 902 <_sk_load_565_hsw+0x66>
+ 15,133,149,0,0,0, //jne 9eb <_sk_load_565_hsw+0xa3>
196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
196,226,125,51,208, //vpmovzxwd %xmm0,%ymm2
- 196,226,125,88,66,104, //vpbroadcastd 0x68(%rdx),%ymm0
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 196,226,125,88,192, //vpbroadcastd %xmm0,%ymm0
197,253,219,194, //vpand %ymm2,%ymm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm1
- 197,244,89,192, //vmulps %ymm0,%ymm1,%ymm0
- 196,226,125,88,74,108, //vpbroadcastd 0x6c(%rdx),%ymm1
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,24,201, //vbroadcastss %xmm1,%ymm1
+ 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,226,125,88,201, //vpbroadcastd %xmm1,%ymm1
197,245,219,202, //vpand %ymm2,%ymm1,%ymm1
197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,226,125,24,90,120, //vbroadcastss 0x78(%rdx),%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,226,125,88,90,112, //vpbroadcastd 0x70(%rdx),%ymm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,88,219, //vpbroadcastd %xmm3,%ymm3
197,229,219,210, //vpand %ymm2,%ymm3,%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,226,125,24,90,124, //vbroadcastss 0x7c(%rdx),%ymm3
- 197,228,89,210, //vmulps %ymm2,%ymm3,%ymm2
- 196,226,125,24,26, //vbroadcastss (%rdx),%ymm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
+ 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
65,137,200, //mov %ecx,%r8d
@@ -7907,8 +8275,8 @@ CODE const uint8_t sk_load_565_hsw[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,146, //ja 8ac <_sk_load_565_hsw+0x10>
- 76,141,13,75,0,0,0, //lea 0x4b(%rip),%r9 # 96c <_sk_load_565_hsw+0xd0>
+ 15,135,85,255,255,255, //ja 95c <_sk_load_565_hsw+0x14>
+ 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # a58 <_sk_load_565_hsw+0x110>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -7920,28 +8288,27 @@ CODE const uint8_t sk_load_565_hsw[] = {
196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,66,255,255,255, //jmpq 8ac <_sk_load_565_hsw+0x10>
- 102,144, //xchg %ax,%ax
- 242,255, //repnz (bad)
- 255, //(bad)
+ 233,5,255,255,255, //jmpq 95c <_sk_load_565_hsw+0x14>
+ 144, //nop
+ 243,255, //repz (bad)
255, //(bad)
- 234, //(bad)
255, //(bad)
+ 235,255, //jmp a5d <_sk_load_565_hsw+0x115>
255, //(bad)
- 255,226, //jmpq *%rdx
+ 255,227, //jmpq *%rbx
255, //(bad)
255, //(bad)
255, //(bad)
- 218,255, //(bad)
+ 219,255, //(bad)
255, //(bad)
- 255,210, //callq *%rdx
+ 255,211, //callq *%rbx
255, //(bad)
255, //(bad)
- 255,202, //dec %edx
+ 255,203, //dec %ebx
255, //(bad)
255, //(bad)
255, //(bad)
- 190, //.byte 0xbe
+ 191, //.byte 0xbf
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -7950,11 +8317,15 @@ CODE const uint8_t sk_load_565_hsw[] = {
CODE const uint8_t sk_store_565_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,130,128,0,0,0, //vbroadcastss 0x80(%rdx),%ymm8
+ 184,0,0,248,65, //mov $0x41f80000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,66,125,24,192, //vbroadcastss %xmm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
196,193,53,114,241,11, //vpslld $0xb,%ymm9,%ymm9
- 196,98,125,24,146,132,0,0,0, //vbroadcastss 0x84(%rdx),%ymm10
+ 184,0,0,124,66, //mov $0x427c0000,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,66,125,24,210, //vbroadcastss %xmm10,%ymm10
197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
196,193,45,114,242,5, //vpslld $0x5,%ymm10,%ymm10
@@ -7965,7 +8336,7 @@ CODE const uint8_t sk_store_565_hsw[] = {
196,67,125,57,193,1, //vextracti128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne 9ea <_sk_store_565_hsw+0x62>
+ 117,10, //jne ae0 <_sk_store_565_hsw+0x6c>
196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -7974,8 +8345,8 @@ CODE const uint8_t sk_store_565_hsw[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja 9e6 <_sk_store_565_hsw+0x5e>
- 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # a48 <_sk_store_565_hsw+0xc0>
+ 119,236, //ja adc <_sk_store_565_hsw+0x68>
+ 76,141,21,69,0,0,0, //lea 0x45(%rip),%r10 # b3c <_sk_store_565_hsw+0xc8>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -7987,27 +8358,28 @@ CODE const uint8_t sk_store_565_hsw[] = {
196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
197,121,126,192, //vmovd %xmm8,%eax
102,65,137,4,121, //mov %ax,(%r9,%rdi,2)
- 235,161, //jmp 9e6 <_sk_store_565_hsw+0x5e>
- 15,31,0, //nopl (%rax)
- 242,255, //repnz (bad)
+ 235,161, //jmp adc <_sk_store_565_hsw+0x68>
+ 144, //nop
+ 244, //hlt
255, //(bad)
255, //(bad)
- 234, //(bad)
255, //(bad)
+ 236, //in (%dx),%al
255, //(bad)
- 255,226, //jmpq *%rdx
255, //(bad)
+ 255,228, //jmpq *%rsp
255, //(bad)
255, //(bad)
- 218,255, //(bad)
255, //(bad)
- 255,210, //callq *%rdx
+ 220,255, //fdivr %st,%st(7)
255, //(bad)
+ 255,212, //callq *%rsp
255, //(bad)
- 255,202, //dec %edx
255, //(bad)
+ 255,204, //dec %esp
255, //(bad)
- 255,194, //inc %edx
+ 255, //(bad)
+ 255,196, //inc %esp
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -8019,7 +8391,7 @@ CODE const uint8_t sk_load_8888_hsw[] = {
76,141,12,189,0,0,0,0, //lea 0x0(,%rdi,4),%r9
76,3,8, //add (%rax),%r9
77,133,192, //test %r8,%r8
- 117,104, //jne ae1 <_sk_load_8888_hsw+0x7d>
+ 117,104, //jne bd5 <_sk_load_8888_hsw+0x7d>
196,193,126,111,25, //vmovdqu (%r9),%ymm3
184,255,0,0,0, //mov $0xff,%eax
197,249,110,192, //vmovd %eax,%xmm0
@@ -8052,7 +8424,7 @@ CODE const uint8_t sk_load_8888_hsw[] = {
196,225,249,110,192, //vmovq %rax,%xmm0
196,226,125,33,192, //vpmovsxbd %xmm0,%ymm0
196,194,125,140,25, //vpmaskmovd (%r9),%ymm0,%ymm3
- 233,116,255,255,255, //jmpq a7e <_sk_load_8888_hsw+0x1a>
+ 233,116,255,255,255, //jmpq b72 <_sk_load_8888_hsw+0x1a>
};
CODE const uint8_t sk_store_8888_hsw[] = {
@@ -8078,7 +8450,7 @@ CODE const uint8_t sk_store_8888_hsw[] = {
196,65,45,235,192, //vpor %ymm8,%ymm10,%ymm8
196,65,53,235,192, //vpor %ymm8,%ymm9,%ymm8
77,133,192, //test %r8,%r8
- 117,12, //jne b7e <_sk_store_8888_hsw+0x74>
+ 117,12, //jne c72 <_sk_store_8888_hsw+0x74>
196,65,126,127,1, //vmovdqu %ymm8,(%r9)
72,173, //lods %ds:(%rsi),%rax
76,137,193, //mov %r8,%rcx
@@ -8091,14 +8463,14 @@ CODE const uint8_t sk_store_8888_hsw[] = {
196,97,249,110,200, //vmovq %rax,%xmm9
196,66,125,33,201, //vpmovsxbd %xmm9,%ymm9
196,66,53,142,1, //vpmaskmovd %ymm8,%ymm9,(%r9)
- 235,211, //jmp b77 <_sk_store_8888_hsw+0x6d>
+ 235,211, //jmp c6b <_sk_store_8888_hsw+0x6d>
};
CODE const uint8_t sk_load_f16_hsw[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
72,133,201, //test %rcx,%rcx
- 117,97, //jne c0f <_sk_load_f16_hsw+0x6b>
+ 117,97, //jne d03 <_sk_load_f16_hsw+0x6b>
197,249,16,12,248, //vmovupd (%rax,%rdi,8),%xmm1
197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
@@ -8124,35 +8496,35 @@ CODE const uint8_t sk_load_f16_hsw[] = {
197,251,16,12,248, //vmovsd (%rax,%rdi,8),%xmm1
196,65,57,87,192, //vxorpd %xmm8,%xmm8,%xmm8
72,131,249,1, //cmp $0x1,%rcx
- 117,6, //jne c25 <_sk_load_f16_hsw+0x81>
+ 117,6, //jne d19 <_sk_load_f16_hsw+0x81>
197,250,126,201, //vmovq %xmm1,%xmm1
- 235,30, //jmp c43 <_sk_load_f16_hsw+0x9f>
+ 235,30, //jmp d37 <_sk_load_f16_hsw+0x9f>
197,241,22,76,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm1,%xmm1
72,131,249,3, //cmp $0x3,%rcx
- 114,18, //jb c43 <_sk_load_f16_hsw+0x9f>
+ 114,18, //jb d37 <_sk_load_f16_hsw+0x9f>
197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
72,131,249,3, //cmp $0x3,%rcx
- 117,19, //jne c50 <_sk_load_f16_hsw+0xac>
+ 117,19, //jne d44 <_sk_load_f16_hsw+0xac>
197,250,126,210, //vmovq %xmm2,%xmm2
- 235,46, //jmp c71 <_sk_load_f16_hsw+0xcd>
+ 235,46, //jmp d65 <_sk_load_f16_hsw+0xcd>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,117,255,255,255, //jmpq bc5 <_sk_load_f16_hsw+0x21>
+ 233,117,255,255,255, //jmpq cb9 <_sk_load_f16_hsw+0x21>
197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
72,131,249,5, //cmp $0x5,%rcx
- 114,21, //jb c71 <_sk_load_f16_hsw+0xcd>
+ 114,21, //jb d65 <_sk_load_f16_hsw+0xcd>
197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
72,131,249,5, //cmp $0x5,%rcx
- 117,18, //jne c7a <_sk_load_f16_hsw+0xd6>
+ 117,18, //jne d6e <_sk_load_f16_hsw+0xd6>
197,250,126,219, //vmovq %xmm3,%xmm3
- 233,84,255,255,255, //jmpq bc5 <_sk_load_f16_hsw+0x21>
+ 233,84,255,255,255, //jmpq cb9 <_sk_load_f16_hsw+0x21>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,75,255,255,255, //jmpq bc5 <_sk_load_f16_hsw+0x21>
+ 233,75,255,255,255, //jmpq cb9 <_sk_load_f16_hsw+0x21>
197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
72,131,249,7, //cmp $0x7,%rcx
- 15,130,59,255,255,255, //jb bc5 <_sk_load_f16_hsw+0x21>
+ 15,130,59,255,255,255, //jb cb9 <_sk_load_f16_hsw+0x21>
197,123,16,68,248,48, //vmovsd 0x30(%rax,%rdi,8),%xmm8
- 233,48,255,255,255, //jmpq bc5 <_sk_load_f16_hsw+0x21>
+ 233,48,255,255,255, //jmpq cb9 <_sk_load_f16_hsw+0x21>
};
CODE const uint8_t sk_store_f16_hsw[] = {
@@ -8171,7 +8543,7 @@ CODE const uint8_t sk_store_f16_hsw[] = {
196,65,57,98,205, //vpunpckldq %xmm13,%xmm8,%xmm9
196,65,57,106,197, //vpunpckhdq %xmm13,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,27, //jne cfa <_sk_store_f16_hsw+0x65>
+ 117,27, //jne dee <_sk_store_f16_hsw+0x65>
197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
@@ -8180,22 +8552,22 @@ CODE const uint8_t sk_store_f16_hsw[] = {
255,224, //jmpq *%rax
197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je cf6 <_sk_store_f16_hsw+0x61>
+ 116,241, //je dea <_sk_store_f16_hsw+0x61>
197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb cf6 <_sk_store_f16_hsw+0x61>
+ 114,229, //jb dea <_sk_store_f16_hsw+0x61>
197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je cf6 <_sk_store_f16_hsw+0x61>
+ 116,221, //je dea <_sk_store_f16_hsw+0x61>
197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb cf6 <_sk_store_f16_hsw+0x61>
+ 114,209, //jb dea <_sk_store_f16_hsw+0x61>
197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je cf6 <_sk_store_f16_hsw+0x61>
+ 116,201, //je dea <_sk_store_f16_hsw+0x61>
197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb cf6 <_sk_store_f16_hsw+0x61>
+ 114,189, //jb dea <_sk_store_f16_hsw+0x61>
197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp cf6 <_sk_store_f16_hsw+0x61>
+ 235,181, //jmp dea <_sk_store_f16_hsw+0x61>
};
CODE const uint8_t sk_store_f32_hsw[] = {
@@ -8211,7 +8583,7 @@ CODE const uint8_t sk_store_f32_hsw[] = {
196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
72,133,201, //test %rcx,%rcx
- 117,55, //jne dae <_sk_store_f32_hsw+0x6d>
+ 117,55, //jne ea2 <_sk_store_f32_hsw+0x6d>
196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
@@ -8224,22 +8596,22 @@ CODE const uint8_t sk_store_f32_hsw[] = {
255,224, //jmpq *%rax
196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je daa <_sk_store_f32_hsw+0x69>
+ 116,240, //je e9e <_sk_store_f32_hsw+0x69>
196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb daa <_sk_store_f32_hsw+0x69>
+ 114,227, //jb e9e <_sk_store_f32_hsw+0x69>
196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je daa <_sk_store_f32_hsw+0x69>
+ 116,218, //je e9e <_sk_store_f32_hsw+0x69>
196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb daa <_sk_store_f32_hsw+0x69>
+ 114,205, //jb e9e <_sk_store_f32_hsw+0x69>
196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je daa <_sk_store_f32_hsw+0x69>
+ 116,195, //je e9e <_sk_store_f32_hsw+0x69>
196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb daa <_sk_store_f32_hsw+0x69>
+ 114,181, //jb e9e <_sk_store_f32_hsw+0x69>
196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp daa <_sk_store_f32_hsw+0x69>
+ 235,171, //jmp e9e <_sk_store_f32_hsw+0x69>
};
CODE const uint8_t sk_clamp_x_hsw[] = {
@@ -8335,11 +8707,17 @@ CODE const uint8_t sk_mirror_y_hsw[] = {
};
CODE const uint8_t sk_luminance_to_alpha_hsw[] = {
- 196,98,125,24,130,136,0,0,0, //vbroadcastss 0x88(%rdx),%ymm8
- 196,226,125,24,154,140,0,0,0, //vbroadcastss 0x8c(%rdx),%ymm3
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,98,125,24,195, //vbroadcastss %xmm3,%ymm8
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,226,125,24,219, //vbroadcastss %xmm3,%ymm3
197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
196,98,125,168,193, //vfmadd213ps %ymm1,%ymm0,%ymm8
- 196,226,125,24,154,144,0,0,0, //vbroadcastss 0x90(%rdx),%ymm3
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 196,226,125,24,216, //vbroadcastss %xmm0,%ymm3
196,194,109,168,216, //vfmadd213ps %ymm8,%ymm2,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -8586,7 +8964,7 @@ CODE const uint8_t sk_seed_shader_avx[] = {
196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 197,252,88,66,20, //vaddps 0x14(%rdx),%ymm0,%ymm0
+ 197,252,88,2, //vaddps (%rdx),%ymm0,%ymm0
196,226,125,24,16, //vbroadcastss (%rax),%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
197,236,88,201, //vaddps %ymm1,%ymm2,%ymm1
@@ -8780,23 +9158,38 @@ CODE const uint8_t sk_unpremul_avx[] = {
};
CODE const uint8_t sk_from_srgb_avx[] = {
- 196,98,125,24,66,64, //vbroadcastss 0x40(%rdx),%ymm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
197,124,89,208, //vmulps %ymm0,%ymm0,%ymm10
- 196,98,125,24,90,60, //vbroadcastss 0x3c(%rdx),%ymm11
- 196,98,125,24,98,56, //vbroadcastss 0x38(%rdx),%ymm12
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
+ 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 197,121,110,224, //vmovd %eax,%xmm12
+ 196,67,121,4,228,0, //vpermilps $0x0,%xmm12,%xmm12
+ 196,67,29,24,228,1, //vinsertf128 $0x1,%xmm12,%ymm12,%ymm12
197,36,89,232, //vmulps %ymm0,%ymm11,%ymm13
196,65,20,88,236, //vaddps %ymm12,%ymm13,%ymm13
- 196,98,125,24,114,52, //vbroadcastss 0x34(%rdx),%ymm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
+ 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
- 196,98,125,24,106,68, //vbroadcastss 0x44(%rdx),%ymm13
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 197,121,110,232, //vmovd %eax,%xmm13
+ 196,67,121,4,237,0, //vpermilps $0x0,%xmm13,%xmm13
+ 196,67,21,24,237,1, //vinsertf128 $0x1,%xmm13,%ymm13,%ymm13
196,193,124,194,197,1, //vcmpltps %ymm13,%ymm0,%ymm0
196,195,45,74,193,0, //vblendvps %ymm0,%ymm9,%ymm10,%ymm0
197,60,89,201, //vmulps %ymm1,%ymm8,%ymm9
197,116,89,209, //vmulps %ymm1,%ymm1,%ymm10
197,36,89,249, //vmulps %ymm1,%ymm11,%ymm15
- 196,65,4,88,252, //vaddps %ymm12,%ymm15,%ymm15
+ 196,65,28,88,255, //vaddps %ymm15,%ymm12,%ymm15
196,65,44,89,215, //vmulps %ymm15,%ymm10,%ymm10
196,65,12,88,210, //vaddps %ymm10,%ymm14,%ymm10
196,193,116,194,205,1, //vcmpltps %ymm13,%ymm1,%ymm1
@@ -8804,7 +9197,7 @@ CODE const uint8_t sk_from_srgb_avx[] = {
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
197,108,89,202, //vmulps %ymm2,%ymm2,%ymm9
197,36,89,210, //vmulps %ymm2,%ymm11,%ymm10
- 196,65,44,88,212, //vaddps %ymm12,%ymm10,%ymm10
+ 196,65,28,88,210, //vaddps %ymm10,%ymm12,%ymm10
196,65,52,89,202, //vmulps %ymm10,%ymm9,%ymm9
196,65,12,88,201, //vaddps %ymm9,%ymm14,%ymm9
196,193,108,194,213,1, //vcmpltps %ymm13,%ymm2,%ymm2
@@ -8815,43 +9208,62 @@ CODE const uint8_t sk_from_srgb_avx[] = {
CODE const uint8_t sk_to_srgb_avx[] = {
197,124,82,192, //vrsqrtps %ymm0,%ymm8
- 196,65,124,83,200, //vrcpps %ymm8,%ymm9
- 196,65,124,82,208, //vrsqrtps %ymm8,%ymm10
- 196,98,125,24,66,72, //vbroadcastss 0x48(%rdx),%ymm8
- 197,60,89,216, //vmulps %ymm0,%ymm8,%ymm11
- 196,98,125,24,34, //vbroadcastss (%rdx),%ymm12
- 196,98,125,24,106,76, //vbroadcastss 0x4c(%rdx),%ymm13
- 196,98,125,24,114,80, //vbroadcastss 0x50(%rdx),%ymm14
- 196,98,125,24,122,84, //vbroadcastss 0x54(%rdx),%ymm15
- 196,65,52,89,206, //vmulps %ymm14,%ymm9,%ymm9
- 196,65,52,88,207, //vaddps %ymm15,%ymm9,%ymm9
- 196,65,44,89,213, //vmulps %ymm13,%ymm10,%ymm10
- 196,65,44,88,201, //vaddps %ymm9,%ymm10,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,98,125,24,82,88, //vbroadcastss 0x58(%rdx),%ymm10
- 196,193,124,194,194,1, //vcmpltps %ymm10,%ymm0,%ymm0
- 196,195,53,74,195,0, //vblendvps %ymm0,%ymm11,%ymm9,%ymm0
- 197,124,82,201, //vrsqrtps %ymm1,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,65,12,89,219, //vmulps %ymm11,%ymm14,%ymm11
- 196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,20,89,201, //vmulps %ymm9,%ymm13,%ymm9
- 196,65,52,88,203, //vaddps %ymm11,%ymm9,%ymm9
- 197,60,89,217, //vmulps %ymm1,%ymm8,%ymm11
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
- 196,193,116,194,202,1, //vcmpltps %ymm10,%ymm1,%ymm1
- 196,195,53,74,203,16, //vblendvps %ymm1,%ymm11,%ymm9,%ymm1
- 197,124,82,202, //vrsqrtps %ymm2,%ymm9
- 196,65,124,83,217, //vrcpps %ymm9,%ymm11
- 196,65,12,89,219, //vmulps %ymm11,%ymm14,%ymm11
+ 196,65,124,83,232, //vrcpps %ymm8,%ymm13
+ 196,65,124,82,240, //vrsqrtps %ymm8,%ymm14
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ 197,60,89,224, //vmulps %ymm0,%ymm8,%ymm12
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,121,110,200, //vmovd %eax,%xmm9
+ 196,67,121,4,201,0, //vpermilps $0x0,%xmm9,%xmm9
+ 196,67,53,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm9,%ymm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
+ 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 197,121,110,216, //vmovd %eax,%xmm11
+ 196,67,121,4,219,0, //vpermilps $0x0,%xmm11,%xmm11
+ 196,67,37,24,219,1, //vinsertf128 $0x1,%xmm11,%ymm11,%ymm11
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 197,121,110,248, //vmovd %eax,%xmm15
+ 196,67,121,4,255,0, //vpermilps $0x0,%xmm15,%xmm15
+ 196,67,5,24,255,1, //vinsertf128 $0x1,%xmm15,%ymm15,%ymm15
+ 196,65,20,89,235, //vmulps %ymm11,%ymm13,%ymm13
+ 196,65,20,88,239, //vaddps %ymm15,%ymm13,%ymm13
+ 196,65,12,89,242, //vmulps %ymm10,%ymm14,%ymm14
+ 196,65,12,88,237, //vaddps %ymm13,%ymm14,%ymm13
+ 196,65,52,93,237, //vminps %ymm13,%ymm9,%ymm13
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 197,121,110,240, //vmovd %eax,%xmm14
+ 196,67,121,4,246,0, //vpermilps $0x0,%xmm14,%xmm14
+ 196,67,13,24,246,1, //vinsertf128 $0x1,%xmm14,%ymm14,%ymm14
+ 196,193,124,194,198,1, //vcmpltps %ymm14,%ymm0,%ymm0
+ 196,195,21,74,196,0, //vblendvps %ymm0,%ymm12,%ymm13,%ymm0
+ 197,124,82,225, //vrsqrtps %ymm1,%ymm12
+ 196,65,124,83,236, //vrcpps %ymm12,%ymm13
+ 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
+ 196,65,36,89,237, //vmulps %ymm13,%ymm11,%ymm13
+ 196,65,4,88,237, //vaddps %ymm13,%ymm15,%ymm13
+ 196,65,44,89,228, //vmulps %ymm12,%ymm10,%ymm12
+ 196,65,28,88,229, //vaddps %ymm13,%ymm12,%ymm12
+ 197,60,89,233, //vmulps %ymm1,%ymm8,%ymm13
+ 196,65,52,93,228, //vminps %ymm12,%ymm9,%ymm12
+ 196,193,116,194,206,1, //vcmpltps %ymm14,%ymm1,%ymm1
+ 196,195,29,74,205,16, //vblendvps %ymm1,%ymm13,%ymm12,%ymm1
+ 197,124,82,226, //vrsqrtps %ymm2,%ymm12
+ 196,65,124,83,236, //vrcpps %ymm12,%ymm13
+ 196,65,36,89,221, //vmulps %ymm13,%ymm11,%ymm11
196,65,4,88,219, //vaddps %ymm11,%ymm15,%ymm11
- 196,65,124,82,201, //vrsqrtps %ymm9,%ymm9
- 196,65,20,89,201, //vmulps %ymm9,%ymm13,%ymm9
- 196,65,52,88,203, //vaddps %ymm11,%ymm9,%ymm9
- 196,65,28,93,201, //vminps %ymm9,%ymm12,%ymm9
+ 196,65,124,82,228, //vrsqrtps %ymm12,%ymm12
+ 196,65,44,89,212, //vmulps %ymm12,%ymm10,%ymm10
+ 196,65,44,88,211, //vaddps %ymm11,%ymm10,%ymm10
+ 196,65,52,93,202, //vminps %ymm10,%ymm9,%ymm9
197,60,89,194, //vmulps %ymm2,%ymm8,%ymm8
- 196,193,108,194,210,1, //vcmpltps %ymm10,%ymm2,%ymm2
+ 196,193,108,194,214,1, //vcmpltps %ymm14,%ymm2,%ymm2
196,195,53,74,208,32, //vblendvps %ymm2,%ymm8,%ymm9,%ymm2
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -8874,7 +9286,7 @@ CODE const uint8_t sk_scale_u8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,80, //jne 58f <_sk_scale_u8_avx+0x60>
+ 117,80, //jne 639 <_sk_scale_u8_avx+0x60>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
@@ -8902,9 +9314,9 @@ CODE const uint8_t sk_scale_u8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 597 <_sk_scale_u8_avx+0x68>
+ 117,234, //jne 641 <_sk_scale_u8_avx+0x68>
196,65,249,110,193, //vmovq %r9,%xmm8
- 235,143, //jmp 543 <_sk_scale_u8_avx+0x14>
+ 235,143, //jmp 5ed <_sk_scale_u8_avx+0x14>
};
CODE const uint8_t sk_lerp_1_float_avx[] = {
@@ -8932,7 +9344,7 @@ CODE const uint8_t sk_lerp_u8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,116, //jne 677 <_sk_lerp_u8_avx+0x84>
+ 117,116, //jne 721 <_sk_lerp_u8_avx+0x84>
197,123,16,0, //vmovsd (%rax),%xmm8
196,66,121,49,200, //vpmovzxbd %xmm8,%xmm9
196,67,121,4,192,229, //vpermilps $0xe5,%xmm8,%xmm8
@@ -8968,41 +9380,59 @@ CODE const uint8_t sk_lerp_u8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne 67f <_sk_lerp_u8_avx+0x8c>
+ 117,234, //jne 729 <_sk_lerp_u8_avx+0x8c>
196,65,249,110,193, //vmovq %r9,%xmm8
- 233,104,255,255,255, //jmpq 607 <_sk_lerp_u8_avx+0x14>
+ 233,104,255,255,255, //jmpq 6b1 <_sk_lerp_u8_avx+0x14>
};
CODE const uint8_t sk_lerp_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,164,0,0,0, //jne 751 <_sk_lerp_565_avx+0xb2>
+ 15,133,250,0,0,0, //jne 851 <_sk_lerp_565_avx+0x108>
196,65,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm8
197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
197,185,105,219, //vpunpckhwd %xmm3,%xmm8,%xmm3
196,66,121,51,192, //vpmovzxwd %xmm8,%xmm8
- 196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 196,98,125,24,66,104, //vbroadcastss 0x68(%rdx),%ymm8
- 197,60,84,195, //vandps %ymm3,%ymm8,%ymm8
- 196,65,124,91,192, //vcvtdq2ps %ymm8,%ymm8
- 196,98,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm9
- 196,65,52,89,192, //vmulps %ymm8,%ymm9,%ymm8
- 196,98,125,24,74,108, //vbroadcastss 0x6c(%rdx),%ymm9
- 197,52,84,203, //vandps %ymm3,%ymm9,%ymm9
- 196,65,124,91,201, //vcvtdq2ps %ymm9,%ymm9
- 196,98,125,24,82,120, //vbroadcastss 0x78(%rdx),%ymm10
- 196,65,44,89,201, //vmulps %ymm9,%ymm10,%ymm9
- 196,98,125,24,82,112, //vbroadcastss 0x70(%rdx),%ymm10
- 197,172,84,219, //vandps %ymm3,%ymm10,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,82,124, //vbroadcastss 0x7c(%rdx),%ymm10
- 197,172,89,219, //vmulps %ymm3,%ymm10,%ymm3
+ 196,99,61,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,203, //vcvtdq2ps %ymm3,%ymm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,52,89,203, //vmulps %ymm3,%ymm9,%ymm9
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,211, //vcvtdq2ps %ymm3,%ymm10
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,44,89,211, //vmulps %ymm3,%ymm10,%ymm10
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 196,193,100,84,216, //vandps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
197,252,92,196, //vsubps %ymm4,%ymm0,%ymm0
- 196,193,124,89,192, //vmulps %ymm8,%ymm0,%ymm0
+ 196,193,124,89,193, //vmulps %ymm9,%ymm0,%ymm0
197,252,88,196, //vaddps %ymm4,%ymm0,%ymm0
197,244,92,205, //vsubps %ymm5,%ymm1,%ymm1
- 196,193,116,89,201, //vmulps %ymm9,%ymm1,%ymm1
+ 196,193,116,89,202, //vmulps %ymm10,%ymm1,%ymm1
197,244,88,205, //vaddps %ymm5,%ymm1,%ymm1
197,236,92,214, //vsubps %ymm6,%ymm2,%ymm2
197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
@@ -9019,8 +9449,8 @@ CODE const uint8_t sk_lerp_565_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,69,255,255,255, //ja 6b3 <_sk_lerp_565_avx+0x14>
- 76,141,13,75,0,0,0, //lea 0x4b(%rip),%r9 # 7c0 <_sk_lerp_565_avx+0x121>
+ 15,135,239,254,255,255, //ja 75d <_sk_lerp_565_avx+0x14>
+ 76,141,13,75,0,0,0, //lea 0x4b(%rip),%r9 # 8c0 <_sk_lerp_565_avx+0x177>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -9032,7 +9462,7 @@ CODE const uint8_t sk_lerp_565_avx[] = {
196,65,57,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm8,%xmm8
196,65,57,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm8,%xmm8
196,65,57,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm8,%xmm8
- 233,245,254,255,255, //jmpq 6b3 <_sk_lerp_565_avx+0x14>
+ 233,159,254,255,255, //jmpq 75d <_sk_lerp_565_avx+0x14>
102,144, //xchg %ax,%ax
242,255, //repnz (bad)
255, //(bad)
@@ -9069,9 +9499,12 @@ CODE const uint8_t sk_load_tables_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,0, //mov (%rax),%r8
72,133,201, //test %rcx,%rcx
- 15,133,18,2,0,0, //jne a06 <_sk_load_tables_avx+0x22a>
+ 15,133,56,2,0,0, //jne b2c <_sk_load_tables_avx+0x250>
196,65,124,16,4,184, //vmovups (%r8,%rdi,4),%ymm8
- 196,98,125,24,74,16, //vbroadcastss 0x10(%rdx),%ymm9
+ 187,255,0,0,0, //mov $0xff,%ebx
+ 197,249,110,195, //vmovd %ebx,%xmm0
+ 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
+ 196,99,125,24,200,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
196,193,52,84,192, //vandps %ymm8,%ymm9,%ymm0
196,193,249,126,193, //vmovq %xmm0,%r9
69,137,203, //mov %r9d,%r11d
@@ -9090,12 +9523,14 @@ CODE const uint8_t sk_load_tables_avx[] = {
76,139,64,16, //mov 0x10(%rax),%r8
196,161,122,16,68,189,0, //vmovss 0x0(%rbp,%r15,4),%xmm0
196,163,121,33,68,165,0,16, //vinsertps $0x10,0x0(%rbp,%r12,4),%xmm0,%xmm0
- 196,163,121,33,68,173,0,32, //vinsertps $0x20,0x0(%rbp,%r13,4),%xmm0,%xmm0
+ 196,161,122,16,76,173,0, //vmovss 0x0(%rbp,%r13,4),%xmm1
+ 196,227,121,33,193,32, //vinsertps $0x20,%xmm1,%xmm0,%xmm0
197,250,16,76,157,0, //vmovss 0x0(%rbp,%rbx,4),%xmm1
196,227,121,33,193,48, //vinsertps $0x30,%xmm1,%xmm0,%xmm0
196,161,122,16,76,157,0, //vmovss 0x0(%rbp,%r11,4),%xmm1
196,163,113,33,76,141,0,16, //vinsertps $0x10,0x0(%rbp,%r9,4),%xmm1,%xmm1
- 196,163,113,33,76,181,0,32, //vinsertps $0x20,0x0(%rbp,%r14,4),%xmm1,%xmm1
+ 196,161,122,16,92,181,0, //vmovss 0x0(%rbp,%r14,4),%xmm3
+ 196,227,113,33,203,32, //vinsertps $0x20,%xmm3,%xmm1,%xmm1
196,161,122,16,92,149,0, //vmovss 0x0(%rbp,%r10,4),%xmm3
196,227,113,33,203,48, //vinsertps $0x30,%xmm3,%xmm1,%xmm1
196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
@@ -9164,9 +9599,12 @@ CODE const uint8_t sk_load_tables_avx[] = {
196,193,57,114,208,24, //vpsrld $0x18,%xmm8,%xmm8
196,193,97,114,210,24, //vpsrld $0x18,%xmm10,%xmm3
196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
- 197,252,91,219, //vcvtdq2ps %ymm3,%ymm3
- 196,98,125,24,66,12, //vbroadcastss 0xc(%rdx),%ymm8
- 196,193,100,89,216, //vmulps %ymm8,%ymm3,%ymm3
+ 197,124,91,195, //vcvtdq2ps %ymm3,%ymm8
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,188,89,219, //vmulps %ymm3,%ymm8,%ymm3
72,173, //lods %ds:(%rsi),%rax
91, //pop %rbx
65,92, //pop %r12
@@ -9175,17 +9613,17 @@ CODE const uint8_t sk_load_tables_avx[] = {
65,95, //pop %r15
93, //pop %rbp
255,224, //jmpq *%rax
- 65,137,201, //mov %ecx,%r9d
- 65,128,225,7, //and $0x7,%r9b
+ 137,203, //mov %ecx,%ebx
+ 128,227,7, //and $0x7,%bl
196,65,60,87,192, //vxorps %ymm8,%ymm8,%ymm8
- 65,254,201, //dec %r9b
- 69,15,182,201, //movzbl %r9b,%r9d
- 65,128,249,6, //cmp $0x6,%r9b
- 15,135,215,253,255,255, //ja 7fa <_sk_load_tables_avx+0x1e>
- 76,141,21,138,0,0,0, //lea 0x8a(%rip),%r10 # ab4 <_sk_load_tables_avx+0x2d8>
- 79,99,12,138, //movslq (%r10,%r9,4),%r9
- 77,1,209, //add %r10,%r9
- 65,255,225, //jmpq *%r9
+ 254,203, //dec %bl
+ 15,182,219, //movzbl %bl,%ebx
+ 128,251,6, //cmp $0x6,%bl
+ 15,135,182,253,255,255, //ja 8fa <_sk_load_tables_avx+0x1e>
+ 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # bd4 <_sk_load_tables_avx+0x2f8>
+ 73,99,28,153, //movslq (%r9,%rbx,4),%rbx
+ 76,1,203, //add %r9,%rbx
+ 255,227, //jmpq *%rbx
196,193,121,110,68,184,24, //vmovd 0x18(%r8,%rdi,4),%xmm0
197,249,112,192,68, //vpshufd $0x44,%xmm0,%xmm0
196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
@@ -9205,7 +9643,7 @@ CODE const uint8_t sk_load_tables_avx[] = {
196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
196,195,57,34,4,184,0, //vpinsrd $0x0,(%r8,%rdi,4),%xmm8,%xmm0
196,99,61,12,192,15, //vblendps $0xf,%ymm0,%ymm8,%ymm8
- 233,70,253,255,255, //jmpq 7fa <_sk_load_tables_avx+0x1e>
+ 233,38,253,255,255, //jmpq 8fa <_sk_load_tables_avx+0x1e>
238, //out %al,(%dx)
255, //(bad)
255, //(bad)
@@ -9232,14 +9670,17 @@ CODE const uint8_t sk_load_a8_avx[] = {
72,139,0, //mov (%rax),%rax
72,1,248, //add %rdi,%rax
77,133,192, //test %r8,%r8
- 117,59, //jne b1b <_sk_load_a8_avx+0x4b>
+ 117,74, //jne c4a <_sk_load_a8_avx+0x5a>
197,251,16,0, //vmovsd (%rax),%xmm0
196,226,121,49,200, //vpmovzxbd %xmm0,%xmm1
196,227,121,4,192,229, //vpermilps $0xe5,%xmm0,%xmm0
196,226,121,49,192, //vpmovzxbd %xmm0,%xmm0
196,227,117,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm1,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,12, //vbroadcastss 0xc(%rdx),%ymm1
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,252,89,217, //vmulps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
197,252,87,192, //vxorps %ymm0,%ymm0,%ymm0
@@ -9256,22 +9697,25 @@ CODE const uint8_t sk_load_a8_avx[] = {
77,9,217, //or %r11,%r9
72,131,193,8, //add $0x8,%rcx
73,255,202, //dec %r10
- 117,234, //jne b23 <_sk_load_a8_avx+0x53>
+ 117,234, //jne c52 <_sk_load_a8_avx+0x62>
196,193,249,110,193, //vmovq %r9,%xmm0
- 235,164, //jmp ae4 <_sk_load_a8_avx+0x14>
+ 235,149, //jmp c04 <_sk_load_a8_avx+0x14>
};
CODE const uint8_t sk_store_a8_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,66,8, //vbroadcastss 0x8(%rdx),%ymm8
+ 184,0,0,127,67, //mov $0x437f0000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,195, //vmulps %ymm3,%ymm8,%ymm8
196,65,125,91,192, //vcvtps2dq %ymm8,%ymm8
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
196,65,57,103,192, //vpackuswb %xmm8,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne b73 <_sk_store_a8_avx+0x33>
+ 117,10, //jne cb1 <_sk_store_a8_avx+0x42>
196,65,123,17,4,57, //vmovsd %xmm8,(%r9,%rdi,1)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -9280,9 +9724,9 @@ CODE const uint8_t sk_store_a8_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja b6f <_sk_store_a8_avx+0x2f>
+ 119,236, //ja cad <_sk_store_a8_avx+0x3e>
196,66,121,48,192, //vpmovzxbw %xmm8,%xmm8
- 76,141,21,69,0,0,0, //lea 0x45(%rip),%r10 # bd4 <_sk_store_a8_avx+0x94>
+ 76,141,21,67,0,0,0, //lea 0x43(%rip),%r10 # d10 <_sk_store_a8_avx+0xa1>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -9293,28 +9737,27 @@ CODE const uint8_t sk_store_a8_avx[] = {
196,67,121,20,68,57,2,4, //vpextrb $0x4,%xmm8,0x2(%r9,%rdi,1)
196,67,121,20,68,57,1,2, //vpextrb $0x2,%xmm8,0x1(%r9,%rdi,1)
196,67,121,20,4,57,0, //vpextrb $0x0,%xmm8,(%r9,%rdi,1)
- 235,158, //jmp b6f <_sk_store_a8_avx+0x2f>
- 15,31,0, //nopl (%rax)
- 244, //hlt
+ 235,158, //jmp cad <_sk_store_a8_avx+0x3e>
+ 144, //nop
+ 246,255, //idiv %bh
255, //(bad)
255, //(bad)
+ 238, //out %al,(%dx)
255, //(bad)
- 236, //in (%dx),%al
255, //(bad)
+ 255,230, //jmpq *%rsi
255, //(bad)
- 255,228, //jmpq *%rsp
255, //(bad)
255, //(bad)
+ 222,255, //fdivrp %st,%st(7)
255, //(bad)
- 220,255, //fdivr %st,%st(7)
+ 255,214, //callq *%rsi
255, //(bad)
- 255,212, //callq *%rsp
255, //(bad)
+ 255,206, //dec %esi
255, //(bad)
- 255,204, //dec %esp
255, //(bad)
- 255, //(bad)
- 255,196, //inc %esp
+ 255,198, //inc %esi
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -9324,28 +9767,49 @@ CODE const uint8_t sk_load_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 117,106, //jne c64 <_sk_load_565_avx+0x74>
+ 15,133,209,0,0,0, //jne e0b <_sk_load_565_avx+0xdf>
196,193,122,111,4,122, //vmovdqu (%r10,%rdi,2),%xmm0
197,241,239,201, //vpxor %xmm1,%xmm1,%xmm1
197,249,105,201, //vpunpckhwd %xmm1,%xmm0,%xmm1
196,226,121,51,192, //vpmovzxwd %xmm0,%xmm0
196,227,125,24,209,1, //vinsertf128 $0x1,%xmm1,%ymm0,%ymm2
- 196,226,125,24,66,104, //vbroadcastss 0x68(%rdx),%ymm0
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 197,249,110,192, //vmovd %eax,%xmm0
+ 197,249,112,192,0, //vpshufd $0x0,%xmm0,%xmm0
+ 196,227,125,24,192,1, //vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
197,252,84,194, //vandps %ymm2,%ymm0,%ymm0
197,252,91,192, //vcvtdq2ps %ymm0,%ymm0
- 196,226,125,24,74,116, //vbroadcastss 0x74(%rdx),%ymm1
- 197,244,89,192, //vmulps %ymm0,%ymm1,%ymm0
- 196,226,125,24,74,108, //vbroadcastss 0x6c(%rdx),%ymm1
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ 197,252,89,193, //vmulps %ymm1,%ymm0,%ymm0
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 197,249,112,201,0, //vpshufd $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,244,84,202, //vandps %ymm2,%ymm1,%ymm1
197,252,91,201, //vcvtdq2ps %ymm1,%ymm1
- 196,226,125,24,90,120, //vbroadcastss 0x78(%rdx),%ymm3
- 197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
- 196,226,125,24,90,112, //vbroadcastss 0x70(%rdx),%ymm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,244,89,203, //vmulps %ymm3,%ymm1,%ymm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,84,210, //vandps %ymm2,%ymm3,%ymm2
197,252,91,210, //vcvtdq2ps %ymm2,%ymm2
- 196,226,125,24,90,124, //vbroadcastss 0x7c(%rdx),%ymm3
- 197,228,89,210, //vmulps %ymm2,%ymm3,%ymm2
- 196,226,125,24,26, //vbroadcastss (%rdx),%ymm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ 197,236,89,211, //vmulps %ymm3,%ymm2,%ymm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
65,137,200, //mov %ecx,%r8d
@@ -9354,8 +9818,8 @@ CODE const uint8_t sk_load_565_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,132, //ja c00 <_sk_load_565_avx+0x10>
- 76,141,13,73,0,0,0, //lea 0x49(%rip),%r9 # ccc <_sk_load_565_avx+0xdc>
+ 15,135,25,255,255,255, //ja d40 <_sk_load_565_avx+0x14>
+ 76,141,13,74,0,0,0, //lea 0x4a(%rip),%r9 # e78 <_sk_load_565_avx+0x14c>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -9367,27 +9831,27 @@ CODE const uint8_t sk_load_565_avx[] = {
196,193,121,196,68,122,4,2, //vpinsrw $0x2,0x4(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,68,122,2,1, //vpinsrw $0x1,0x2(%r10,%rdi,2),%xmm0,%xmm0
196,193,121,196,4,122,0, //vpinsrw $0x0,(%r10,%rdi,2),%xmm0,%xmm0
- 233,52,255,255,255, //jmpq c00 <_sk_load_565_avx+0x10>
- 244, //hlt
- 255, //(bad)
+ 233,201,254,255,255, //jmpq d40 <_sk_load_565_avx+0x14>
+ 144, //nop
+ 243,255, //repz (bad)
255, //(bad)
255, //(bad)
- 236, //in (%dx),%al
+ 235,255, //jmp e7d <_sk_load_565_avx+0x151>
255, //(bad)
+ 255,227, //jmpq *%rbx
255, //(bad)
- 255,228, //jmpq *%rsp
255, //(bad)
255, //(bad)
+ 219,255, //(bad)
255, //(bad)
- 220,255, //fdivr %st,%st(7)
+ 255,211, //callq *%rbx
255, //(bad)
- 255,212, //callq *%rsp
255, //(bad)
+ 255,203, //dec %ebx
255, //(bad)
- 255,204, //dec %esp
255, //(bad)
255, //(bad)
- 255,192, //inc %eax
+ 191, //.byte 0xbf
255, //(bad)
255, //(bad)
255, //.byte 0xff
@@ -9396,14 +9860,20 @@ CODE const uint8_t sk_load_565_avx[] = {
CODE const uint8_t sk_store_565_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,8, //mov (%rax),%r9
- 196,98,125,24,130,128,0,0,0, //vbroadcastss 0x80(%rdx),%ymm8
+ 184,0,0,248,65, //mov $0x41f80000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,67,121,4,192,0, //vpermilps $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,65,125,91,201, //vcvtps2dq %ymm9,%ymm9
196,193,41,114,241,11, //vpslld $0xb,%xmm9,%xmm10
196,67,125,25,201,1, //vextractf128 $0x1,%ymm9,%xmm9
196,193,49,114,241,11, //vpslld $0xb,%xmm9,%xmm9
196,67,45,24,201,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm9
- 196,98,125,24,146,132,0,0,0, //vbroadcastss 0x84(%rdx),%ymm10
+ 184,0,0,124,66, //mov $0x427c0000,%eax
+ 197,121,110,208, //vmovd %eax,%xmm10
+ 196,67,121,4,210,0, //vpermilps $0x0,%xmm10,%xmm10
+ 196,67,45,24,210,1, //vinsertf128 $0x1,%xmm10,%ymm10,%ymm10
197,44,89,209, //vmulps %ymm1,%ymm10,%ymm10
196,65,125,91,210, //vcvtps2dq %ymm10,%ymm10
196,193,33,114,242,5, //vpslld $0x5,%xmm10,%xmm11
@@ -9417,7 +9887,7 @@ CODE const uint8_t sk_store_565_avx[] = {
196,67,125,25,193,1, //vextractf128 $0x1,%ymm8,%xmm9
196,66,57,43,193, //vpackusdw %xmm9,%xmm8,%xmm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne d6e <_sk_store_565_avx+0x86>
+ 117,10, //jne f32 <_sk_store_565_avx+0x9e>
196,65,122,127,4,121, //vmovdqu %xmm8,(%r9,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -9426,8 +9896,8 @@ CODE const uint8_t sk_store_565_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja d6a <_sk_store_565_avx+0x82>
- 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # dcc <_sk_store_565_avx+0xe4>
+ 119,236, //ja f2e <_sk_store_565_avx+0x9a>
+ 76,141,21,71,0,0,0, //lea 0x47(%rip),%r10 # f90 <_sk_store_565_avx+0xfc>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -9439,7 +9909,7 @@ CODE const uint8_t sk_store_565_avx[] = {
196,67,121,21,68,121,2,1, //vpextrw $0x1,%xmm8,0x2(%r9,%rdi,2)
197,121,126,192, //vmovd %xmm8,%eax
102,65,137,4,121, //mov %ax,(%r9,%rdi,2)
- 235,161, //jmp d6a <_sk_store_565_avx+0x82>
+ 235,161, //jmp f2e <_sk_store_565_avx+0x9a>
15,31,0, //nopl (%rax)
242,255, //repnz (bad)
255, //(bad)
@@ -9469,7 +9939,7 @@ CODE const uint8_t sk_load_8888_avx[] = {
72,173, //lods %ds:(%rsi),%rax
76,139,16, //mov (%rax),%r10
72,133,201, //test %rcx,%rcx
- 15,133,157,0,0,0, //jne e93 <_sk_load_8888_avx+0xab>
+ 15,133,157,0,0,0, //jne 1057 <_sk_load_8888_avx+0xab>
196,65,124,16,12,186, //vmovups (%r10,%rdi,4),%ymm9
184,255,0,0,0, //mov $0xff,%eax
197,249,110,192, //vmovd %eax,%xmm0
@@ -9508,8 +9978,8 @@ CODE const uint8_t sk_load_8888_avx[] = {
65,254,200, //dec %r8b
69,15,182,192, //movzbl %r8b,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 15,135,76,255,255,255, //ja dfc <_sk_load_8888_avx+0x14>
- 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # f40 <_sk_load_8888_avx+0x158>
+ 15,135,76,255,255,255, //ja fc0 <_sk_load_8888_avx+0x14>
+ 76,141,13,137,0,0,0, //lea 0x89(%rip),%r9 # 1104 <_sk_load_8888_avx+0x158>
75,99,4,129, //movslq (%r9,%r8,4),%rax
76,1,200, //add %r9,%rax
255,224, //jmpq *%rax
@@ -9532,7 +10002,7 @@ CODE const uint8_t sk_load_8888_avx[] = {
196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
196,195,49,34,4,186,0, //vpinsrd $0x0,(%r10,%rdi,4),%xmm9,%xmm0
196,99,53,12,200,15, //vblendps $0xf,%ymm0,%ymm9,%ymm9
- 233,188,254,255,255, //jmpq dfc <_sk_load_8888_avx+0x14>
+ 233,188,254,255,255, //jmpq fc0 <_sk_load_8888_avx+0x14>
238, //out %al,(%dx)
255, //(bad)
255, //(bad)
@@ -9584,7 +10054,7 @@ CODE const uint8_t sk_store_8888_avx[] = {
196,65,45,86,192, //vorpd %ymm8,%ymm10,%ymm8
196,65,53,86,192, //vorpd %ymm8,%ymm9,%ymm8
72,133,201, //test %rcx,%rcx
- 117,10, //jne 1000 <_sk_store_8888_avx+0xa4>
+ 117,10, //jne 11c4 <_sk_store_8888_avx+0xa4>
196,65,124,17,4,185, //vmovups %ymm8,(%r9,%rdi,4)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -9593,8 +10063,8 @@ CODE const uint8_t sk_store_8888_avx[] = {
254,200, //dec %al
68,15,182,192, //movzbl %al,%r8d
65,128,248,6, //cmp $0x6,%r8b
- 119,236, //ja ffc <_sk_store_8888_avx+0xa0>
- 76,141,21,85,0,0,0, //lea 0x55(%rip),%r10 # 106c <_sk_store_8888_avx+0x110>
+ 119,236, //ja 11c0 <_sk_store_8888_avx+0xa0>
+ 76,141,21,85,0,0,0, //lea 0x55(%rip),%r10 # 1230 <_sk_store_8888_avx+0x110>
75,99,4,130, //movslq (%r10,%r8,4),%rax
76,1,208, //add %r10,%rax
255,224, //jmpq *%rax
@@ -9608,7 +10078,7 @@ CODE const uint8_t sk_store_8888_avx[] = {
196,67,121,22,68,185,8,2, //vpextrd $0x2,%xmm8,0x8(%r9,%rdi,4)
196,67,121,22,68,185,4,1, //vpextrd $0x1,%xmm8,0x4(%r9,%rdi,4)
196,65,121,126,4,185, //vmovd %xmm8,(%r9,%rdi,4)
- 235,147, //jmp ffc <_sk_store_8888_avx+0xa0>
+ 235,147, //jmp 11c0 <_sk_store_8888_avx+0xa0>
15,31,0, //nopl (%rax)
245, //cmc
255, //(bad)
@@ -9639,7 +10109,7 @@ CODE const uint8_t sk_load_f16_avx[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
72,133,201, //test %rcx,%rcx
- 15,133,240,0,0,0, //jne 1186 <_sk_load_f16_avx+0xfe>
+ 15,133,8,1,0,0, //jne 1362 <_sk_load_f16_avx+0x116>
197,249,16,12,248, //vmovupd (%rax,%rdi,8),%xmm1
197,249,16,84,248,16, //vmovupd 0x10(%rax,%rdi,8),%xmm2
197,249,16,92,248,32, //vmovupd 0x20(%rax,%rdi,8),%xmm3
@@ -9650,42 +10120,46 @@ CODE const uint8_t sk_load_f16_avx[] = {
196,193,97,105,216, //vpunpckhwd %xmm8,%xmm3,%xmm3
197,121,97,193, //vpunpcklwd %xmm1,%xmm0,%xmm8
197,249,105,193, //vpunpckhwd %xmm1,%xmm0,%xmm0
- 197,233,97,203, //vpunpcklwd %xmm3,%xmm2,%xmm1
+ 197,105,97,211, //vpunpcklwd %xmm3,%xmm2,%xmm10
197,105,105,203, //vpunpckhwd %xmm3,%xmm2,%xmm9
- 197,249,110,90,100, //vmovd 0x64(%rdx),%xmm3
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
196,193,97,101,208, //vpcmpgtw %xmm8,%xmm3,%xmm2
- 196,65,105,223,192, //vpandn %xmm8,%xmm2,%xmm8
- 197,225,101,208, //vpcmpgtw %xmm0,%xmm3,%xmm2
- 197,233,223,192, //vpandn %xmm0,%xmm2,%xmm0
- 197,225,101,209, //vpcmpgtw %xmm1,%xmm3,%xmm2
- 197,233,223,201, //vpandn %xmm1,%xmm2,%xmm1
- 196,193,97,101,209, //vpcmpgtw %xmm9,%xmm3,%xmm2
- 196,193,105,223,209, //vpandn %xmm9,%xmm2,%xmm2
- 196,66,121,51,208, //vpmovzxwd %xmm8,%xmm10
+ 196,193,105,223,208, //vpandn %xmm8,%xmm2,%xmm2
+ 197,225,101,200, //vpcmpgtw %xmm0,%xmm3,%xmm1
+ 197,241,223,192, //vpandn %xmm0,%xmm1,%xmm0
+ 196,193,97,101,202, //vpcmpgtw %xmm10,%xmm3,%xmm1
+ 196,193,113,223,202, //vpandn %xmm10,%xmm1,%xmm1
+ 196,193,97,101,217, //vpcmpgtw %xmm9,%xmm3,%xmm3
+ 196,193,97,223,217, //vpandn %xmm9,%xmm3,%xmm3
+ 196,98,121,51,194, //vpmovzxwd %xmm2,%xmm8
196,98,121,51,201, //vpmovzxwd %xmm1,%xmm9
- 197,225,239,219, //vpxor %xmm3,%xmm3,%xmm3
- 197,57,105,195, //vpunpckhwd %xmm3,%xmm8,%xmm8
- 197,241,105,203, //vpunpckhwd %xmm3,%xmm1,%xmm1
+ 196,65,41,239,210, //vpxor %xmm10,%xmm10,%xmm10
+ 196,193,105,105,210, //vpunpckhwd %xmm10,%xmm2,%xmm2
+ 196,193,113,105,202, //vpunpckhwd %xmm10,%xmm1,%xmm1
196,98,121,51,216, //vpmovzxwd %xmm0,%xmm11
- 196,98,121,51,226, //vpmovzxwd %xmm2,%xmm12
- 197,121,105,235, //vpunpckhwd %xmm3,%xmm0,%xmm13
- 197,105,105,243, //vpunpckhwd %xmm3,%xmm2,%xmm14
- 196,193,121,114,242,13, //vpslld $0xd,%xmm10,%xmm0
- 196,193,105,114,241,13, //vpslld $0xd,%xmm9,%xmm2
- 196,227,125,24,194,1, //vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- 196,98,125,24,74,92, //vbroadcastss 0x5c(%rdx),%ymm9
+ 196,98,121,51,227, //vpmovzxwd %xmm3,%xmm12
+ 196,65,121,105,234, //vpunpckhwd %xmm10,%xmm0,%xmm13
+ 196,65,97,105,210, //vpunpckhwd %xmm10,%xmm3,%xmm10
+ 196,193,121,114,240,13, //vpslld $0xd,%xmm8,%xmm0
+ 196,193,97,114,241,13, //vpslld $0xd,%xmm9,%xmm3
+ 196,227,125,24,195,1, //vinsertf128 $0x1,%xmm3,%ymm0,%ymm0
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 197,249,112,219,0, //vpshufd $0x0,%xmm3,%xmm3
+ 196,99,101,24,203,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm9
197,180,89,192, //vmulps %ymm0,%ymm9,%ymm0
- 196,193,105,114,240,13, //vpslld $0xd,%xmm8,%xmm2
+ 197,233,114,242,13, //vpslld $0xd,%xmm2,%xmm2
197,241,114,241,13, //vpslld $0xd,%xmm1,%xmm1
196,227,109,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm2,%ymm1
197,180,89,201, //vmulps %ymm1,%ymm9,%ymm1
- 196,193,105,114,243,13, //vpslld $0xd,%xmm11,%xmm2
- 196,193,97,114,244,13, //vpslld $0xd,%xmm12,%xmm3
- 196,227,109,24,211,1, //vinsertf128 $0x1,%xmm3,%ymm2,%ymm2
+ 196,193,57,114,243,13, //vpslld $0xd,%xmm11,%xmm8
+ 196,193,105,114,244,13, //vpslld $0xd,%xmm12,%xmm2
+ 196,227,61,24,210,1, //vinsertf128 $0x1,%xmm2,%ymm8,%ymm2
197,180,89,210, //vmulps %ymm2,%ymm9,%ymm2
196,193,57,114,245,13, //vpslld $0xd,%xmm13,%xmm8
- 196,193,97,114,246,13, //vpslld $0xd,%xmm14,%xmm3
+ 196,193,97,114,242,13, //vpslld $0xd,%xmm10,%xmm3
196,227,61,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
197,180,89,219, //vmulps %ymm3,%ymm9,%ymm3
72,173, //lods %ds:(%rsi),%rax
@@ -9693,41 +10167,44 @@ CODE const uint8_t sk_load_f16_avx[] = {
197,251,16,12,248, //vmovsd (%rax,%rdi,8),%xmm1
196,65,57,87,192, //vxorpd %xmm8,%xmm8,%xmm8
72,131,249,1, //cmp $0x1,%rcx
- 117,6, //jne 119c <_sk_load_f16_avx+0x114>
+ 117,6, //jne 1378 <_sk_load_f16_avx+0x12c>
197,250,126,201, //vmovq %xmm1,%xmm1
- 235,30, //jmp 11ba <_sk_load_f16_avx+0x132>
+ 235,30, //jmp 1396 <_sk_load_f16_avx+0x14a>
197,241,22,76,248,8, //vmovhpd 0x8(%rax,%rdi,8),%xmm1,%xmm1
72,131,249,3, //cmp $0x3,%rcx
- 114,18, //jb 11ba <_sk_load_f16_avx+0x132>
+ 114,18, //jb 1396 <_sk_load_f16_avx+0x14a>
197,251,16,84,248,16, //vmovsd 0x10(%rax,%rdi,8),%xmm2
72,131,249,3, //cmp $0x3,%rcx
- 117,19, //jne 11c7 <_sk_load_f16_avx+0x13f>
+ 117,19, //jne 13a3 <_sk_load_f16_avx+0x157>
197,250,126,210, //vmovq %xmm2,%xmm2
- 235,46, //jmp 11e8 <_sk_load_f16_avx+0x160>
+ 235,46, //jmp 13c4 <_sk_load_f16_avx+0x178>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
197,233,87,210, //vxorpd %xmm2,%xmm2,%xmm2
- 233,230,254,255,255, //jmpq 10ad <_sk_load_f16_avx+0x25>
+ 233,206,254,255,255, //jmpq 1271 <_sk_load_f16_avx+0x25>
197,233,22,84,248,24, //vmovhpd 0x18(%rax,%rdi,8),%xmm2,%xmm2
72,131,249,5, //cmp $0x5,%rcx
- 114,21, //jb 11e8 <_sk_load_f16_avx+0x160>
+ 114,21, //jb 13c4 <_sk_load_f16_avx+0x178>
197,251,16,92,248,32, //vmovsd 0x20(%rax,%rdi,8),%xmm3
72,131,249,5, //cmp $0x5,%rcx
- 117,18, //jne 11f1 <_sk_load_f16_avx+0x169>
+ 117,18, //jne 13cd <_sk_load_f16_avx+0x181>
197,250,126,219, //vmovq %xmm3,%xmm3
- 233,197,254,255,255, //jmpq 10ad <_sk_load_f16_avx+0x25>
+ 233,173,254,255,255, //jmpq 1271 <_sk_load_f16_avx+0x25>
197,225,87,219, //vxorpd %xmm3,%xmm3,%xmm3
- 233,188,254,255,255, //jmpq 10ad <_sk_load_f16_avx+0x25>
+ 233,164,254,255,255, //jmpq 1271 <_sk_load_f16_avx+0x25>
197,225,22,92,248,40, //vmovhpd 0x28(%rax,%rdi,8),%xmm3,%xmm3
72,131,249,7, //cmp $0x7,%rcx
- 15,130,172,254,255,255, //jb 10ad <_sk_load_f16_avx+0x25>
+ 15,130,148,254,255,255, //jb 1271 <_sk_load_f16_avx+0x25>
197,123,16,68,248,48, //vmovsd 0x30(%rax,%rdi,8),%xmm8
- 233,161,254,255,255, //jmpq 10ad <_sk_load_f16_avx+0x25>
+ 233,137,254,255,255, //jmpq 1271 <_sk_load_f16_avx+0x25>
};
CODE const uint8_t sk_store_f16_avx[] = {
72,173, //lods %ds:(%rsi),%rax
- 72,139,0, //mov (%rax),%rax
- 196,98,125,24,66,96, //vbroadcastss 0x60(%rdx),%ymm8
+ 76,139,0, //mov (%rax),%r8
+ 184,0,0,128,7, //mov $0x7800000,%eax
+ 197,121,110,192, //vmovd %eax,%xmm8
+ 196,65,121,112,192,0, //vpshufd $0x0,%xmm8,%xmm8
+ 196,67,61,24,192,1, //vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
197,60,89,200, //vmulps %ymm0,%ymm8,%ymm9
196,67,125,25,202,1, //vextractf128 $0x1,%ymm9,%xmm10
196,193,41,114,210,13, //vpsrld $0xd,%xmm10,%xmm10
@@ -9757,31 +10234,31 @@ CODE const uint8_t sk_store_f16_avx[] = {
196,65,25,98,205, //vpunpckldq %xmm13,%xmm12,%xmm9
196,65,25,106,197, //vpunpckhdq %xmm13,%xmm12,%xmm8
72,133,201, //test %rcx,%rcx
- 117,27, //jne 12cf <_sk_store_f16_avx+0xc3>
- 197,120,17,28,248, //vmovups %xmm11,(%rax,%rdi,8)
- 197,120,17,84,248,16, //vmovups %xmm10,0x10(%rax,%rdi,8)
- 197,120,17,76,248,32, //vmovups %xmm9,0x20(%rax,%rdi,8)
- 197,122,127,68,248,48, //vmovdqu %xmm8,0x30(%rax,%rdi,8)
+ 117,31, //jne 14be <_sk_store_f16_avx+0xd6>
+ 196,65,120,17,28,248, //vmovups %xmm11,(%r8,%rdi,8)
+ 196,65,120,17,84,248,16, //vmovups %xmm10,0x10(%r8,%rdi,8)
+ 196,65,120,17,76,248,32, //vmovups %xmm9,0x20(%r8,%rdi,8)
+ 196,65,122,127,68,248,48, //vmovdqu %xmm8,0x30(%r8,%rdi,8)
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
- 197,121,214,28,248, //vmovq %xmm11,(%rax,%rdi,8)
+ 196,65,121,214,28,248, //vmovq %xmm11,(%r8,%rdi,8)
72,131,249,1, //cmp $0x1,%rcx
- 116,241, //je 12cb <_sk_store_f16_avx+0xbf>
- 197,121,23,92,248,8, //vmovhpd %xmm11,0x8(%rax,%rdi,8)
+ 116,240, //je 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,92,248,8, //vmovhpd %xmm11,0x8(%r8,%rdi,8)
72,131,249,3, //cmp $0x3,%rcx
- 114,229, //jb 12cb <_sk_store_f16_avx+0xbf>
- 197,121,214,84,248,16, //vmovq %xmm10,0x10(%rax,%rdi,8)
- 116,221, //je 12cb <_sk_store_f16_avx+0xbf>
- 197,121,23,84,248,24, //vmovhpd %xmm10,0x18(%rax,%rdi,8)
+ 114,227, //jb 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,84,248,16, //vmovq %xmm10,0x10(%r8,%rdi,8)
+ 116,218, //je 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,84,248,24, //vmovhpd %xmm10,0x18(%r8,%rdi,8)
72,131,249,5, //cmp $0x5,%rcx
- 114,209, //jb 12cb <_sk_store_f16_avx+0xbf>
- 197,121,214,76,248,32, //vmovq %xmm9,0x20(%rax,%rdi,8)
- 116,201, //je 12cb <_sk_store_f16_avx+0xbf>
- 197,121,23,76,248,40, //vmovhpd %xmm9,0x28(%rax,%rdi,8)
+ 114,205, //jb 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,76,248,32, //vmovq %xmm9,0x20(%r8,%rdi,8)
+ 116,196, //je 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,23,76,248,40, //vmovhpd %xmm9,0x28(%r8,%rdi,8)
72,131,249,7, //cmp $0x7,%rcx
- 114,189, //jb 12cb <_sk_store_f16_avx+0xbf>
- 197,121,214,68,248,48, //vmovq %xmm8,0x30(%rax,%rdi,8)
- 235,181, //jmp 12cb <_sk_store_f16_avx+0xbf>
+ 114,183, //jb 14ba <_sk_store_f16_avx+0xd2>
+ 196,65,121,214,68,248,48, //vmovq %xmm8,0x30(%r8,%rdi,8)
+ 235,174, //jmp 14ba <_sk_store_f16_avx+0xd2>
};
CODE const uint8_t sk_store_f32_avx[] = {
@@ -9797,7 +10274,7 @@ CODE const uint8_t sk_store_f32_avx[] = {
196,65,37,20,196, //vunpcklpd %ymm12,%ymm11,%ymm8
196,65,37,21,220, //vunpckhpd %ymm12,%ymm11,%ymm11
72,133,201, //test %rcx,%rcx
- 117,55, //jne 1383 <_sk_store_f32_avx+0x6d>
+ 117,55, //jne 1579 <_sk_store_f32_avx+0x6d>
196,67,45,24,225,1, //vinsertf128 $0x1,%xmm9,%ymm10,%ymm12
196,67,61,24,235,1, //vinsertf128 $0x1,%xmm11,%ymm8,%ymm13
196,67,45,6,201,49, //vperm2f128 $0x31,%ymm9,%ymm10,%ymm9
@@ -9810,22 +10287,22 @@ CODE const uint8_t sk_store_f32_avx[] = {
255,224, //jmpq *%rax
196,65,121,17,20,128, //vmovupd %xmm10,(%r8,%rax,4)
72,131,249,1, //cmp $0x1,%rcx
- 116,240, //je 137f <_sk_store_f32_avx+0x69>
+ 116,240, //je 1575 <_sk_store_f32_avx+0x69>
196,65,121,17,76,128,16, //vmovupd %xmm9,0x10(%r8,%rax,4)
72,131,249,3, //cmp $0x3,%rcx
- 114,227, //jb 137f <_sk_store_f32_avx+0x69>
+ 114,227, //jb 1575 <_sk_store_f32_avx+0x69>
196,65,121,17,68,128,32, //vmovupd %xmm8,0x20(%r8,%rax,4)
- 116,218, //je 137f <_sk_store_f32_avx+0x69>
+ 116,218, //je 1575 <_sk_store_f32_avx+0x69>
196,65,121,17,92,128,48, //vmovupd %xmm11,0x30(%r8,%rax,4)
72,131,249,5, //cmp $0x5,%rcx
- 114,205, //jb 137f <_sk_store_f32_avx+0x69>
+ 114,205, //jb 1575 <_sk_store_f32_avx+0x69>
196,67,125,25,84,128,64,1, //vextractf128 $0x1,%ymm10,0x40(%r8,%rax,4)
- 116,195, //je 137f <_sk_store_f32_avx+0x69>
+ 116,195, //je 1575 <_sk_store_f32_avx+0x69>
196,67,125,25,76,128,80,1, //vextractf128 $0x1,%ymm9,0x50(%r8,%rax,4)
72,131,249,7, //cmp $0x7,%rcx
- 114,181, //jb 137f <_sk_store_f32_avx+0x69>
+ 114,181, //jb 1575 <_sk_store_f32_avx+0x69>
196,67,125,25,68,128,96,1, //vextractf128 $0x1,%ymm8,0x60(%r8,%rax,4)
- 235,171, //jmp 137f <_sk_store_f32_avx+0x69>
+ 235,171, //jmp 1575 <_sk_store_f32_avx+0x69>
};
CODE const uint8_t sk_clamp_x_avx[] = {
@@ -9947,12 +10424,21 @@ CODE const uint8_t sk_mirror_y_avx[] = {
};
CODE const uint8_t sk_luminance_to_alpha_avx[] = {
- 196,226,125,24,154,136,0,0,0, //vbroadcastss 0x88(%rdx),%ymm3
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,89,192, //vmulps %ymm0,%ymm3,%ymm0
- 196,226,125,24,154,140,0,0,0, //vbroadcastss 0x8c(%rdx),%ymm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 197,249,110,216, //vmovd %eax,%xmm3
+ 196,227,121,4,219,0, //vpermilps $0x0,%xmm3,%xmm3
+ 196,227,101,24,219,1, //vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
197,228,89,201, //vmulps %ymm1,%ymm3,%ymm1
197,252,88,193, //vaddps %ymm1,%ymm0,%ymm0
- 196,226,125,24,138,144,0,0,0, //vbroadcastss 0x90(%rdx),%ymm1
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 197,249,110,200, //vmovd %eax,%xmm1
+ 196,227,121,4,201,0, //vpermilps $0x0,%xmm1,%xmm1
+ 196,227,117,24,201,1, //vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
197,244,89,202, //vmulps %ymm2,%ymm1,%ymm1
197,252,88,217, //vaddps %ymm1,%ymm0,%ymm3
72,173, //lods %ds:(%rsi),%rax
@@ -10217,7 +10703,7 @@ CODE const uint8_t sk_seed_shader_sse41[] = {
102,15,110,209, //movd %ecx,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
15,88,202, //addps %xmm2,%xmm1
- 15,16,66,20, //movups 0x14(%rdx),%xmm0
+ 15,16,2, //movups (%rdx),%xmm0
15,88,193, //addps %xmm1,%xmm0
102,15,110,8, //movd (%rax),%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
@@ -10420,25 +10906,29 @@ CODE const uint8_t sk_unpremul_sse41[] = {
};
CODE const uint8_t sk_from_srgb_sse41[] = {
- 68,15,40,194, //movaps %xmm2,%xmm8
- 243,68,15,16,90,64, //movss 0x40(%rdx),%xmm11
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
69,15,40,211, //movaps %xmm11,%xmm10
68,15,89,208, //mulps %xmm0,%xmm10
68,15,40,240, //movaps %xmm0,%xmm14
69,15,89,246, //mulps %xmm14,%xmm14
- 243,15,16,82,60, //movss 0x3c(%rdx),%xmm2
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 243,68,15,16,98,52, //movss 0x34(%rdx),%xmm12
- 243,68,15,16,106,56, //movss 0x38(%rdx),%xmm13
- 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 68,15,40,202, //movaps %xmm2,%xmm9
- 68,15,89,200, //mulps %xmm0,%xmm9
- 69,15,88,205, //addps %xmm13,%xmm9
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
+ 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,206, //mulps %xmm14,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
69,15,88,204, //addps %xmm12,%xmm9
- 243,68,15,16,114,68, //movss 0x44(%rdx),%xmm14
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
+ 69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
+ 69,15,89,206, //mulps %xmm14,%xmm9
+ 69,15,88,205, //addps %xmm13,%xmm9
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 102,68,15,110,240, //movd %eax,%xmm14
69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
65,15,194,198,1, //cmpltps %xmm14,%xmm0
102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
@@ -10446,27 +10936,28 @@ CODE const uint8_t sk_from_srgb_sse41[] = {
68,15,89,249, //mulps %xmm1,%xmm15
15,40,193, //movaps %xmm1,%xmm0
15,89,192, //mulps %xmm0,%xmm0
- 68,15,40,210, //movaps %xmm2,%xmm10
+ 69,15,40,208, //movaps %xmm8,%xmm10
68,15,89,209, //mulps %xmm1,%xmm10
- 69,15,88,213, //addps %xmm13,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
69,15,88,212, //addps %xmm12,%xmm10
+ 68,15,89,208, //mulps %xmm0,%xmm10
+ 69,15,88,213, //addps %xmm13,%xmm10
65,15,194,206,1, //cmpltps %xmm14,%xmm1
15,40,193, //movaps %xmm1,%xmm0
102,69,15,56,20,215, //blendvps %xmm0,%xmm15,%xmm10
- 69,15,89,216, //mulps %xmm8,%xmm11
- 65,15,40,192, //movaps %xmm8,%xmm0
+ 68,15,89,218, //mulps %xmm2,%xmm11
+ 15,40,194, //movaps %xmm2,%xmm0
15,89,192, //mulps %xmm0,%xmm0
- 65,15,89,208, //mulps %xmm8,%xmm2
- 65,15,88,213, //addps %xmm13,%xmm2
- 15,89,208, //mulps %xmm0,%xmm2
- 65,15,88,212, //addps %xmm12,%xmm2
- 69,15,194,198,1, //cmpltps %xmm14,%xmm8
- 65,15,40,192, //movaps %xmm8,%xmm0
- 102,65,15,56,20,211, //blendvps %xmm0,%xmm11,%xmm2
+ 68,15,89,194, //mulps %xmm2,%xmm8
+ 69,15,88,196, //addps %xmm12,%xmm8
+ 68,15,89,192, //mulps %xmm0,%xmm8
+ 69,15,88,197, //addps %xmm13,%xmm8
+ 65,15,194,214,1, //cmpltps %xmm14,%xmm2
+ 15,40,194, //movaps %xmm2,%xmm0
+ 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
72,173, //lods %ds:(%rsi),%rax
65,15,40,193, //movaps %xmm9,%xmm0
65,15,40,202, //movaps %xmm10,%xmm1
+ 65,15,40,208, //movaps %xmm8,%xmm2
255,224, //jmpq *%rax
};
@@ -10477,62 +10968,69 @@ CODE const uint8_t sk_to_srgb_sse41[] = {
15,40,245, //movaps %xmm5,%xmm6
15,40,236, //movaps %xmm4,%xmm5
15,40,227, //movaps %xmm3,%xmm4
- 68,15,40,194, //movaps %xmm2,%xmm8
- 15,40,217, //movaps %xmm1,%xmm3
- 15,82,208, //rsqrtps %xmm0,%xmm2
- 68,15,83,202, //rcpps %xmm2,%xmm9
- 68,15,82,210, //rsqrtps %xmm2,%xmm10
- 243,15,16,18, //movss (%rdx),%xmm2
- 243,68,15,16,90,72, //movss 0x48(%rdx),%xmm11
+ 15,40,218, //movaps %xmm2,%xmm3
+ 15,40,209, //movaps %xmm1,%xmm2
+ 68,15,82,192, //rsqrtps %xmm0,%xmm8
+ 69,15,83,200, //rcpps %xmm8,%xmm9
+ 69,15,82,248, //rsqrtps %xmm8,%xmm15
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 65,15,40,203, //movaps %xmm11,%xmm1
- 15,89,200, //mulps %xmm0,%xmm1
- 15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 243,68,15,16,98,76, //movss 0x4c(%rdx),%xmm12
+ 69,15,40,211, //movaps %xmm11,%xmm10
+ 68,15,89,208, //mulps %xmm0,%xmm10
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
+ 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 243,68,15,16,106,80, //movss 0x50(%rdx),%xmm13
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
- 243,68,15,16,114,84, //movss 0x54(%rdx),%xmm14
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 102,68,15,110,240, //movd %eax,%xmm14
69,15,198,246,0, //shufps $0x0,%xmm14,%xmm14
69,15,89,205, //mulps %xmm13,%xmm9
69,15,88,206, //addps %xmm14,%xmm9
- 69,15,89,212, //mulps %xmm12,%xmm10
- 69,15,88,209, //addps %xmm9,%xmm10
- 68,15,40,202, //movaps %xmm2,%xmm9
- 69,15,93,202, //minps %xmm10,%xmm9
- 243,68,15,16,122,88, //movss 0x58(%rdx),%xmm15
+ 69,15,89,252, //mulps %xmm12,%xmm15
+ 69,15,88,249, //addps %xmm9,%xmm15
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 69,15,93,207, //minps %xmm15,%xmm9
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 102,68,15,110,248, //movd %eax,%xmm15
69,15,198,255,0, //shufps $0x0,%xmm15,%xmm15
65,15,194,199,1, //cmpltps %xmm15,%xmm0
- 102,68,15,56,20,201, //blendvps %xmm0,%xmm1,%xmm9
+ 102,69,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm9
+ 68,15,82,210, //rsqrtps %xmm2,%xmm10
+ 65,15,83,194, //rcpps %xmm10,%xmm0
+ 69,15,82,210, //rsqrtps %xmm10,%xmm10
+ 65,15,89,197, //mulps %xmm13,%xmm0
+ 65,15,88,198, //addps %xmm14,%xmm0
+ 69,15,89,212, //mulps %xmm12,%xmm10
+ 68,15,88,208, //addps %xmm0,%xmm10
+ 65,15,40,200, //movaps %xmm8,%xmm1
+ 65,15,93,202, //minps %xmm10,%xmm1
+ 69,15,40,211, //movaps %xmm11,%xmm10
+ 68,15,89,210, //mulps %xmm2,%xmm10
+ 65,15,194,215,1, //cmpltps %xmm15,%xmm2
+ 15,40,194, //movaps %xmm2,%xmm0
+ 102,65,15,56,20,202, //blendvps %xmm0,%xmm10,%xmm1
15,82,195, //rsqrtps %xmm3,%xmm0
- 15,83,200, //rcpps %xmm0,%xmm1
+ 15,83,208, //rcpps %xmm0,%xmm2
+ 65,15,89,213, //mulps %xmm13,%xmm2
+ 65,15,88,214, //addps %xmm14,%xmm2
15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,205, //mulps %xmm13,%xmm1
- 65,15,88,206, //addps %xmm14,%xmm1
65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 68,15,40,210, //movaps %xmm2,%xmm10
- 68,15,93,208, //minps %xmm0,%xmm10
- 65,15,40,203, //movaps %xmm11,%xmm1
- 15,89,203, //mulps %xmm3,%xmm1
+ 15,88,194, //addps %xmm2,%xmm0
+ 68,15,93,192, //minps %xmm0,%xmm8
+ 68,15,89,219, //mulps %xmm3,%xmm11
65,15,194,223,1, //cmpltps %xmm15,%xmm3
15,40,195, //movaps %xmm3,%xmm0
- 102,68,15,56,20,209, //blendvps %xmm0,%xmm1,%xmm10
- 65,15,82,192, //rsqrtps %xmm8,%xmm0
- 15,83,200, //rcpps %xmm0,%xmm1
- 65,15,89,205, //mulps %xmm13,%xmm1
- 65,15,88,206, //addps %xmm14,%xmm1
- 15,82,192, //rsqrtps %xmm0,%xmm0
- 65,15,89,196, //mulps %xmm12,%xmm0
- 15,88,193, //addps %xmm1,%xmm0
- 15,93,208, //minps %xmm0,%xmm2
- 69,15,89,216, //mulps %xmm8,%xmm11
- 69,15,194,199,1, //cmpltps %xmm15,%xmm8
- 65,15,40,192, //movaps %xmm8,%xmm0
- 102,65,15,56,20,211, //blendvps %xmm0,%xmm11,%xmm2
+ 102,69,15,56,20,195, //blendvps %xmm0,%xmm11,%xmm8
72,173, //lods %ds:(%rsi),%rax
65,15,40,193, //movaps %xmm9,%xmm0
- 65,15,40,202, //movaps %xmm10,%xmm1
+ 65,15,40,208, //movaps %xmm8,%xmm2
15,40,220, //movaps %xmm4,%xmm3
15,40,229, //movaps %xmm5,%xmm4
15,40,238, //movaps %xmm6,%xmm5
@@ -10620,32 +11118,38 @@ CODE const uint8_t sk_lerp_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
102,68,15,56,51,4,120, //pmovzxwd (%rax,%rdi,2),%xmm8
- 102,15,110,90,104, //movd 0x68(%rdx),%xmm3
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 243,68,15,16,90,116, //movss 0x74(%rdx),%xmm11
- 243,68,15,16,82,120, //movss 0x78(%rdx),%xmm10
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,217, //mulps %xmm9,%xmm11
- 102,15,110,90,108, //movd 0x6c(%rdx),%xmm3
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 69,15,89,209, //mulps %xmm9,%xmm10
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 68,15,89,211, //mulps %xmm3,%xmm10
- 102,15,110,90,112, //movd 0x70(%rdx),%xmm3
+ 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,217, //mulps %xmm9,%xmm11
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,195, //cvtdq2ps %xmm3,%xmm8
- 243,15,16,90,124, //movss 0x7c(%rdx),%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
15,92,196, //subps %xmm4,%xmm0
- 65,15,89,195, //mulps %xmm11,%xmm0
+ 65,15,89,194, //mulps %xmm10,%xmm0
15,88,196, //addps %xmm4,%xmm0
15,92,205, //subps %xmm5,%xmm1
- 65,15,89,202, //mulps %xmm10,%xmm1
+ 65,15,89,203, //mulps %xmm11,%xmm1
15,88,205, //addps %xmm5,%xmm1
15,92,214, //subps %xmm6,%xmm2
15,89,211, //mulps %xmm3,%xmm2
@@ -10662,7 +11166,8 @@ CODE const uint8_t sk_load_tables_sse41[] = {
72,139,8, //mov (%rax),%rcx
76,139,64,8, //mov 0x8(%rax),%r8
243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 102,15,110,66,16, //movd 0x10(%rdx),%xmm0
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 102,15,110,193, //movd %ecx,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
102,65,15,111,200, //movdqa %xmm8,%xmm1
102,15,114,209,8, //psrld $0x8,%xmm1
@@ -10709,7 +11214,8 @@ CODE const uint8_t sk_load_tables_sse41[] = {
102,15,58,33,211,48, //insertps $0x30,%xmm3,%xmm2
102,65,15,114,208,24, //psrld $0x18,%xmm8
69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -10721,7 +11227,8 @@ CODE const uint8_t sk_load_a8_sse41[] = {
72,139,0, //mov (%rax),%rax
102,15,56,49,4,56, //pmovzxbd (%rax,%rdi,1),%xmm0
15,91,192, //cvtdq2ps %xmm0,%xmm0
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -10734,7 +11241,8 @@ CODE const uint8_t sk_load_a8_sse41[] = {
CODE const uint8_t sk_store_a8_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,66,8, //movss 0x8(%rdx),%xmm8
+ 185,0,0,127,67, //mov $0x437f0000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
68,15,89,195, //mulps %xmm3,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
@@ -10748,29 +11256,36 @@ CODE const uint8_t sk_store_a8_sse41[] = {
CODE const uint8_t sk_load_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,56,51,12,120, //pmovzxwd (%rax,%rdi,2),%xmm9
- 102,15,110,66,104, //movd 0x68(%rdx),%xmm0
+ 102,15,56,51,20,120, //pmovzxwd (%rax,%rdi,2),%xmm2
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,219,193, //pand %xmm9,%xmm0
+ 102,15,219,194, //pand %xmm2,%xmm0
15,91,200, //cvtdq2ps %xmm0,%xmm1
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,15,16,66,116, //movss 0x74(%rdx),%xmm0
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,15,110,192, //movd %eax,%xmm0
15,198,192,0, //shufps $0x0,%xmm0,%xmm0
15,89,193, //mulps %xmm1,%xmm0
- 102,15,110,74,108, //movd 0x6c(%rdx),%xmm1
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,200, //movd %eax,%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,65,15,219,201, //pand %xmm9,%xmm1
- 68,15,91,193, //cvtdq2ps %xmm1,%xmm8
- 243,15,16,74,120, //movss 0x78(%rdx),%xmm1
+ 102,15,219,202, //pand %xmm2,%xmm1
+ 15,91,217, //cvtdq2ps %xmm1,%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,15,110,200, //movd %eax,%xmm1
15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,110,82,112, //movd 0x70(%rdx),%xmm2
- 102,15,112,210,0, //pshufd $0x0,%xmm2,%xmm2
- 102,65,15,219,209, //pand %xmm9,%xmm2
- 68,15,91,194, //cvtdq2ps %xmm2,%xmm8
- 243,15,16,82,124, //movss 0x7c(%rdx),%xmm2
+ 15,89,203, //mulps %xmm3,%xmm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,15,219,218, //pand %xmm2,%xmm3
+ 15,91,219, //cvtdq2ps %xmm3,%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,208, //movd %eax,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
+ 15,89,211, //mulps %xmm3,%xmm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -10779,21 +11294,23 @@ CODE const uint8_t sk_load_565_sse41[] = {
CODE const uint8_t sk_store_565_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,130,128,0,0,0, //movss 0x80(%rdx),%xmm8
- 243,68,15,16,138,132,0,0,0, //movss 0x84(%rdx),%xmm9
+ 185,0,0,248,65, //mov $0x41f80000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,11, //pslld $0xb,%xmm10
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,5, //pslld $0x5,%xmm9
- 102,69,15,235,202, //por %xmm10,%xmm9
+ 102,65,15,114,241,11, //pslld $0xb,%xmm9
+ 185,0,0,124,66, //mov $0x427c0000,%ecx
+ 102,68,15,110,209, //movd %ecx,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 68,15,89,209, //mulps %xmm1,%xmm10
+ 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
+ 102,65,15,114,242,5, //pslld $0x5,%xmm10
+ 102,69,15,235,209, //por %xmm9,%xmm10
68,15,89,194, //mulps %xmm2,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,193, //orpd %xmm9,%xmm8
+ 102,69,15,86,194, //orpd %xmm10,%xmm8
102,69,15,56,43,192, //packusdw %xmm8,%xmm8
102,68,15,214,4,120, //movq %xmm8,(%rax,%rdi,2)
72,173, //lods %ds:(%rsi),%rax
@@ -10869,7 +11386,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
102,68,15,111,194, //movdqa %xmm2,%xmm8
102,68,15,97,192, //punpcklwd %xmm0,%xmm8
102,15,105,208, //punpckhwd %xmm0,%xmm2
- 102,15,110,66,100, //movd 0x64(%rdx),%xmm0
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
102,15,111,203, //movdqa %xmm3,%xmm1
102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
@@ -10878,7 +11396,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
102,15,223,218, //pandn %xmm2,%xmm3
102,15,56,51,193, //pmovzxwd %xmm1,%xmm0
102,15,114,240,13, //pslld $0xd,%xmm0
- 102,15,110,82,92, //movd 0x5c(%rdx),%xmm2
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 102,15,110,208, //movd %eax,%xmm2
102,68,15,112,194,0, //pshufd $0x0,%xmm2,%xmm8
65,15,89,192, //mulps %xmm8,%xmm0
102,69,15,239,201, //pxor %xmm9,%xmm9
@@ -10898,7 +11417,8 @@ CODE const uint8_t sk_load_f16_sse41[] = {
CODE const uint8_t sk_store_f16_sse41[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,110,66,96, //movd 0x60(%rdx),%xmm8
+ 185,0,0,128,7, //mov $0x7800000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
102,69,15,111,200, //movdqa %xmm8,%xmm9
68,15,89,200, //mulps %xmm0,%xmm9
@@ -11060,17 +11580,20 @@ CODE const uint8_t sk_mirror_y_sse41[] = {
};
CODE const uint8_t sk_luminance_to_alpha_sse41[] = {
- 243,15,16,154,136,0,0,0, //movss 0x88(%rdx),%xmm3
- 243,68,15,16,130,140,0,0,0, //movss 0x8c(%rdx),%xmm8
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,193, //mulps %xmm1,%xmm8
- 68,15,88,195, //addps %xmm3,%xmm8
- 243,15,16,154,144,0,0,0, //movss 0x90(%rdx),%xmm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 102,15,110,192, //movd %eax,%xmm0
+ 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
+ 15,89,193, //mulps %xmm1,%xmm0
+ 15,88,195, //addps %xmm3,%xmm0
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,218, //mulps %xmm2,%xmm3
- 65,15,88,216, //addps %xmm8,%xmm3
+ 15,88,216, //addps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
15,87,192, //xorps %xmm0,%xmm0
15,87,201, //xorps %xmm1,%xmm1
@@ -11393,7 +11916,7 @@ CODE const uint8_t sk_seed_shader_sse2[] = {
102,15,110,209, //movd %ecx,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
15,88,202, //addps %xmm2,%xmm1
- 15,16,66,20, //movups 0x14(%rdx),%xmm0
+ 15,16,2, //movups (%rdx),%xmm0
15,88,193, //addps %xmm1,%xmm0
102,15,110,8, //movd (%rax),%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
@@ -11593,24 +12116,29 @@ CODE const uint8_t sk_unpremul_sse2[] = {
};
CODE const uint8_t sk_from_srgb_sse2[] = {
- 243,68,15,16,66,64, //movss 0x40(%rdx),%xmm8
+ 184,145,131,158,61, //mov $0x3d9e8391,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
69,15,40,232, //movaps %xmm8,%xmm13
68,15,89,232, //mulps %xmm0,%xmm13
68,15,40,224, //movaps %xmm0,%xmm12
69,15,89,228, //mulps %xmm12,%xmm12
- 243,68,15,16,74,60, //movss 0x3c(%rdx),%xmm9
+ 184,154,153,153,62, //mov $0x3e99999a,%eax
+ 102,68,15,110,200, //movd %eax,%xmm9
69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 243,68,15,16,82,52, //movss 0x34(%rdx),%xmm10
- 243,68,15,16,90,56, //movss 0x38(%rdx),%xmm11
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 184,92,143,50,63, //mov $0x3f328f5c,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
69,15,40,241, //movaps %xmm9,%xmm14
68,15,89,240, //mulps %xmm0,%xmm14
- 69,15,88,243, //addps %xmm11,%xmm14
- 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 69,15,89,244, //mulps %xmm12,%xmm14
69,15,88,242, //addps %xmm10,%xmm14
- 243,68,15,16,98,68, //movss 0x44(%rdx),%xmm12
+ 184,10,215,35,59, //mov $0x3b23d70a,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,244, //mulps %xmm12,%xmm14
+ 69,15,88,243, //addps %xmm11,%xmm14
+ 184,174,71,97,61, //mov $0x3d6147ae,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
65,15,194,196,1, //cmpltps %xmm12,%xmm0
68,15,84,232, //andps %xmm0,%xmm13
@@ -11622,9 +12150,9 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
69,15,89,246, //mulps %xmm14,%xmm14
69,15,40,249, //movaps %xmm9,%xmm15
68,15,89,249, //mulps %xmm1,%xmm15
- 69,15,88,251, //addps %xmm11,%xmm15
- 69,15,89,254, //mulps %xmm14,%xmm15
69,15,88,250, //addps %xmm10,%xmm15
+ 69,15,89,254, //mulps %xmm14,%xmm15
+ 69,15,88,251, //addps %xmm11,%xmm15
65,15,194,204,1, //cmpltps %xmm12,%xmm1
68,15,84,233, //andps %xmm1,%xmm13
65,15,85,207, //andnps %xmm15,%xmm1
@@ -11633,9 +12161,9 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
68,15,40,234, //movaps %xmm2,%xmm13
69,15,89,237, //mulps %xmm13,%xmm13
68,15,89,202, //mulps %xmm2,%xmm9
- 69,15,88,203, //addps %xmm11,%xmm9
- 69,15,89,205, //mulps %xmm13,%xmm9
69,15,88,202, //addps %xmm10,%xmm9
+ 69,15,89,205, //mulps %xmm13,%xmm9
+ 69,15,88,203, //addps %xmm11,%xmm9
65,15,194,212,1, //cmpltps %xmm12,%xmm2
68,15,84,194, //andps %xmm2,%xmm8
65,15,85,209, //andnps %xmm9,%xmm2
@@ -11645,74 +12173,69 @@ CODE const uint8_t sk_from_srgb_sse2[] = {
};
CODE const uint8_t sk_to_srgb_sse2[] = {
- 72,131,236,40, //sub $0x28,%rsp
- 15,41,124,36,16, //movaps %xmm7,0x10(%rsp)
- 15,41,52,36, //movaps %xmm6,(%rsp)
- 15,40,245, //movaps %xmm5,%xmm6
- 15,40,236, //movaps %xmm4,%xmm5
- 15,40,227, //movaps %xmm3,%xmm4
68,15,82,192, //rsqrtps %xmm0,%xmm8
- 69,15,83,232, //rcpps %xmm8,%xmm13
- 69,15,82,248, //rsqrtps %xmm8,%xmm15
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,68,15,16,66,72, //movss 0x48(%rdx),%xmm8
+ 69,15,83,248, //rcpps %xmm8,%xmm15
+ 69,15,82,232, //rsqrtps %xmm8,%xmm13
+ 184,41,92,71,65, //mov $0x41475c29,%eax
+ 102,68,15,110,192, //movd %eax,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
69,15,40,240, //movaps %xmm8,%xmm14
68,15,89,240, //mulps %xmm0,%xmm14
- 15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 243,68,15,16,82,76, //movss 0x4c(%rdx),%xmm10
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,68,15,110,200, //movd %eax,%xmm9
+ 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
+ 184,194,135,210,62, //mov $0x3ed287c2,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
- 243,68,15,16,90,80, //movss 0x50(%rdx),%xmm11
+ 184,206,111,48,63, //mov $0x3f306fce,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 243,68,15,16,98,84, //movss 0x54(%rdx),%xmm12
+ 184,168,87,202,61, //mov $0x3dca57a8,%eax
+ 53,0,0,0,128, //xor $0x80000000,%eax
+ 102,68,15,110,224, //movd %eax,%xmm12
69,15,198,228,0, //shufps $0x0,%xmm12,%xmm12
- 69,15,89,235, //mulps %xmm11,%xmm13
- 69,15,88,236, //addps %xmm12,%xmm13
- 69,15,89,250, //mulps %xmm10,%xmm15
- 69,15,88,253, //addps %xmm13,%xmm15
- 68,15,40,203, //movaps %xmm3,%xmm9
- 69,15,93,207, //minps %xmm15,%xmm9
- 243,68,15,16,106,88, //movss 0x58(%rdx),%xmm13
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,89,234, //mulps %xmm10,%xmm13
+ 69,15,88,239, //addps %xmm15,%xmm13
+ 69,15,40,249, //movaps %xmm9,%xmm15
+ 69,15,93,253, //minps %xmm13,%xmm15
+ 184,4,231,140,59, //mov $0x3b8ce704,%eax
+ 102,68,15,110,232, //movd %eax,%xmm13
69,15,198,237,0, //shufps $0x0,%xmm13,%xmm13
65,15,194,197,1, //cmpltps %xmm13,%xmm0
68,15,84,240, //andps %xmm0,%xmm14
- 65,15,85,193, //andnps %xmm9,%xmm0
+ 65,15,85,199, //andnps %xmm15,%xmm0
65,15,86,198, //orps %xmm14,%xmm0
- 68,15,82,201, //rsqrtps %xmm1,%xmm9
- 69,15,83,241, //rcpps %xmm9,%xmm14
- 69,15,82,201, //rsqrtps %xmm9,%xmm9
- 69,15,89,243, //mulps %xmm11,%xmm14
- 69,15,88,244, //addps %xmm12,%xmm14
- 69,15,89,202, //mulps %xmm10,%xmm9
- 69,15,88,206, //addps %xmm14,%xmm9
- 68,15,40,243, //movaps %xmm3,%xmm14
- 69,15,93,241, //minps %xmm9,%xmm14
- 69,15,40,200, //movaps %xmm8,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 68,15,82,241, //rsqrtps %xmm1,%xmm14
+ 69,15,83,254, //rcpps %xmm14,%xmm15
+ 69,15,82,246, //rsqrtps %xmm14,%xmm14
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,89,242, //mulps %xmm10,%xmm14
+ 69,15,88,247, //addps %xmm15,%xmm14
+ 69,15,40,249, //movaps %xmm9,%xmm15
+ 69,15,93,254, //minps %xmm14,%xmm15
+ 69,15,40,240, //movaps %xmm8,%xmm14
+ 68,15,89,241, //mulps %xmm1,%xmm14
65,15,194,205,1, //cmpltps %xmm13,%xmm1
- 68,15,84,201, //andps %xmm1,%xmm9
- 65,15,85,206, //andnps %xmm14,%xmm1
- 65,15,86,201, //orps %xmm9,%xmm1
- 68,15,82,202, //rsqrtps %xmm2,%xmm9
- 69,15,83,241, //rcpps %xmm9,%xmm14
- 69,15,89,243, //mulps %xmm11,%xmm14
- 69,15,88,244, //addps %xmm12,%xmm14
- 65,15,82,249, //rsqrtps %xmm9,%xmm7
- 65,15,89,250, //mulps %xmm10,%xmm7
- 65,15,88,254, //addps %xmm14,%xmm7
- 15,93,223, //minps %xmm7,%xmm3
+ 68,15,84,241, //andps %xmm1,%xmm14
+ 65,15,85,207, //andnps %xmm15,%xmm1
+ 65,15,86,206, //orps %xmm14,%xmm1
+ 68,15,82,242, //rsqrtps %xmm2,%xmm14
+ 69,15,83,254, //rcpps %xmm14,%xmm15
+ 69,15,89,251, //mulps %xmm11,%xmm15
+ 69,15,88,252, //addps %xmm12,%xmm15
+ 69,15,82,222, //rsqrtps %xmm14,%xmm11
+ 69,15,89,218, //mulps %xmm10,%xmm11
+ 69,15,88,223, //addps %xmm15,%xmm11
+ 69,15,93,203, //minps %xmm11,%xmm9
68,15,89,194, //mulps %xmm2,%xmm8
65,15,194,213,1, //cmpltps %xmm13,%xmm2
68,15,84,194, //andps %xmm2,%xmm8
- 15,85,211, //andnps %xmm3,%xmm2
+ 65,15,85,209, //andnps %xmm9,%xmm2
65,15,86,208, //orps %xmm8,%xmm2
72,173, //lods %ds:(%rsi),%rax
- 15,40,220, //movaps %xmm4,%xmm3
- 15,40,229, //movaps %xmm5,%xmm4
- 15,40,238, //movaps %xmm6,%xmm5
- 15,40,52,36, //movaps (%rsp),%xmm6
- 15,40,124,36,16, //movaps 0x10(%rsp),%xmm7
- 72,131,196,40, //add $0x28,%rsp
255,224, //jmpq *%rax
};
@@ -11799,35 +12322,41 @@ CODE const uint8_t sk_lerp_u8_sse2[] = {
CODE const uint8_t sk_lerp_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,126,12,120, //movq (%rax,%rdi,2),%xmm9
+ 243,68,15,126,4,120, //movq (%rax,%rdi,2),%xmm8
102,15,239,219, //pxor %xmm3,%xmm3
- 102,68,15,97,203, //punpcklwd %xmm3,%xmm9
- 102,15,110,90,104, //movd 0x68(%rdx),%xmm3
- 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
- 68,15,91,211, //cvtdq2ps %xmm3,%xmm10
- 243,68,15,16,90,116, //movss 0x74(%rdx),%xmm11
- 243,68,15,16,66,120, //movss 0x78(%rdx),%xmm8
- 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
- 69,15,89,218, //mulps %xmm10,%xmm11
- 102,15,110,90,108, //movd 0x6c(%rdx),%xmm3
+ 102,68,15,97,195, //punpcklwd %xmm3,%xmm8
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
- 15,91,219, //cvtdq2ps %xmm3,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,195, //mulps %xmm3,%xmm8
- 102,15,110,90,112, //movd 0x70(%rdx),%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
+ 68,15,91,203, //cvtdq2ps %xmm3,%xmm9
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,68,15,110,208, //movd %eax,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 69,15,89,209, //mulps %xmm9,%xmm10
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
- 102,65,15,219,217, //pand %xmm9,%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
68,15,91,203, //cvtdq2ps %xmm3,%xmm9
- 243,15,16,90,124, //movss 0x7c(%rdx),%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,68,15,110,216, //movd %eax,%xmm11
+ 69,15,198,219,0, //shufps $0x0,%xmm11,%xmm11
+ 69,15,89,217, //mulps %xmm9,%xmm11
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,65,15,219,216, //pand %xmm8,%xmm3
+ 68,15,91,195, //cvtdq2ps %xmm3,%xmm8
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
- 65,15,89,217, //mulps %xmm9,%xmm3
+ 65,15,89,216, //mulps %xmm8,%xmm3
15,92,196, //subps %xmm4,%xmm0
- 65,15,89,195, //mulps %xmm11,%xmm0
+ 65,15,89,194, //mulps %xmm10,%xmm0
15,88,196, //addps %xmm4,%xmm0
15,92,205, //subps %xmm5,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
+ 65,15,89,203, //mulps %xmm11,%xmm1
15,88,205, //addps %xmm5,%xmm1
15,92,214, //subps %xmm6,%xmm2
15,89,211, //mulps %xmm3,%xmm2
@@ -11844,7 +12373,8 @@ CODE const uint8_t sk_load_tables_sse2[] = {
72,139,8, //mov (%rax),%rcx
76,139,64,8, //mov 0x8(%rax),%r8
243,68,15,111,4,185, //movdqu (%rcx,%rdi,4),%xmm8
- 102,15,110,66,16, //movd 0x10(%rdx),%xmm0
+ 185,255,0,0,0, //mov $0xff,%ecx
+ 102,15,110,193, //movd %ecx,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
102,69,15,111,200, //movdqa %xmm8,%xmm9
102,65,15,114,209,8, //psrld $0x8,%xmm9
@@ -11899,7 +12429,8 @@ CODE const uint8_t sk_load_tables_sse2[] = {
65,15,20,209, //unpcklps %xmm9,%xmm2
102,65,15,114,208,24, //psrld $0x18,%xmm8
69,15,91,192, //cvtdq2ps %xmm8,%xmm8
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
65,15,89,216, //mulps %xmm8,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -11914,7 +12445,8 @@ CODE const uint8_t sk_load_a8_sse2[] = {
102,15,96,193, //punpcklbw %xmm1,%xmm0
102,15,97,193, //punpcklwd %xmm1,%xmm0
15,91,192, //cvtdq2ps %xmm0,%xmm0
- 243,15,16,90,12, //movss 0xc(%rdx),%xmm3
+ 184,129,128,128,59, //mov $0x3b808081,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
@@ -11927,7 +12459,8 @@ CODE const uint8_t sk_load_a8_sse2[] = {
CODE const uint8_t sk_store_a8_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,66,8, //movss 0x8(%rdx),%xmm8
+ 185,0,0,127,67, //mov $0x437f0000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
68,15,89,195, //mulps %xmm3,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
@@ -11943,31 +12476,38 @@ CODE const uint8_t sk_store_a8_sse2[] = {
CODE const uint8_t sk_load_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,126,12,120, //movq (%rax,%rdi,2),%xmm9
+ 243,15,126,20,120, //movq (%rax,%rdi,2),%xmm2
102,15,239,192, //pxor %xmm0,%xmm0
- 102,68,15,97,200, //punpcklwd %xmm0,%xmm9
- 102,15,110,66,104, //movd 0x68(%rdx),%xmm0
+ 102,15,97,208, //punpcklwd %xmm0,%xmm2
+ 184,0,248,0,0, //mov $0xf800,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,192,0, //pshufd $0x0,%xmm0,%xmm0
- 102,65,15,219,193, //pand %xmm9,%xmm0
+ 102,15,219,194, //pand %xmm2,%xmm0
15,91,200, //cvtdq2ps %xmm0,%xmm1
- 243,15,16,26, //movss (%rdx),%xmm3
- 243,15,16,66,116, //movss 0x74(%rdx),%xmm0
+ 184,8,33,132,55, //mov $0x37842108,%eax
+ 102,15,110,192, //movd %eax,%xmm0
15,198,192,0, //shufps $0x0,%xmm0,%xmm0
15,89,193, //mulps %xmm1,%xmm0
- 102,15,110,74,108, //movd 0x6c(%rdx),%xmm1
+ 184,224,7,0,0, //mov $0x7e0,%eax
+ 102,15,110,200, //movd %eax,%xmm1
102,15,112,201,0, //pshufd $0x0,%xmm1,%xmm1
- 102,65,15,219,201, //pand %xmm9,%xmm1
- 68,15,91,193, //cvtdq2ps %xmm1,%xmm8
- 243,15,16,74,120, //movss 0x78(%rdx),%xmm1
+ 102,15,219,202, //pand %xmm2,%xmm1
+ 15,91,217, //cvtdq2ps %xmm1,%xmm3
+ 184,33,8,2,58, //mov $0x3a020821,%eax
+ 102,15,110,200, //movd %eax,%xmm1
15,198,201,0, //shufps $0x0,%xmm1,%xmm1
- 65,15,89,200, //mulps %xmm8,%xmm1
- 102,15,110,82,112, //movd 0x70(%rdx),%xmm2
- 102,15,112,210,0, //pshufd $0x0,%xmm2,%xmm2
- 102,65,15,219,209, //pand %xmm9,%xmm2
- 68,15,91,194, //cvtdq2ps %xmm2,%xmm8
- 243,15,16,82,124, //movss 0x7c(%rdx),%xmm2
+ 15,89,203, //mulps %xmm3,%xmm1
+ 184,31,0,0,0, //mov $0x1f,%eax
+ 102,15,110,216, //movd %eax,%xmm3
+ 102,15,112,219,0, //pshufd $0x0,%xmm3,%xmm3
+ 102,15,219,218, //pand %xmm2,%xmm3
+ 15,91,219, //cvtdq2ps %xmm3,%xmm3
+ 184,8,33,4,61, //mov $0x3d042108,%eax
+ 102,15,110,208, //movd %eax,%xmm2
15,198,210,0, //shufps $0x0,%xmm2,%xmm2
- 65,15,89,208, //mulps %xmm8,%xmm2
+ 15,89,211, //mulps %xmm3,%xmm2
+ 184,0,0,128,63, //mov $0x3f800000,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
72,173, //lods %ds:(%rsi),%rax
255,224, //jmpq *%rax
@@ -11976,21 +12516,23 @@ CODE const uint8_t sk_load_565_sse2[] = {
CODE const uint8_t sk_store_565_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 243,68,15,16,130,128,0,0,0, //movss 0x80(%rdx),%xmm8
- 243,68,15,16,138,132,0,0,0, //movss 0x84(%rdx),%xmm9
+ 185,0,0,248,65, //mov $0x41f80000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 69,15,40,208, //movaps %xmm8,%xmm10
- 68,15,89,208, //mulps %xmm0,%xmm10
- 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
- 102,65,15,114,242,11, //pslld $0xb,%xmm10
- 69,15,198,201,0, //shufps $0x0,%xmm9,%xmm9
- 68,15,89,201, //mulps %xmm1,%xmm9
+ 69,15,40,200, //movaps %xmm8,%xmm9
+ 68,15,89,200, //mulps %xmm0,%xmm9
102,69,15,91,201, //cvtps2dq %xmm9,%xmm9
- 102,65,15,114,241,5, //pslld $0x5,%xmm9
- 102,69,15,235,202, //por %xmm10,%xmm9
+ 102,65,15,114,241,11, //pslld $0xb,%xmm9
+ 185,0,0,124,66, //mov $0x427c0000,%ecx
+ 102,68,15,110,209, //movd %ecx,%xmm10
+ 69,15,198,210,0, //shufps $0x0,%xmm10,%xmm10
+ 68,15,89,209, //mulps %xmm1,%xmm10
+ 102,69,15,91,210, //cvtps2dq %xmm10,%xmm10
+ 102,65,15,114,242,5, //pslld $0x5,%xmm10
+ 102,69,15,235,209, //por %xmm9,%xmm10
68,15,89,194, //mulps %xmm2,%xmm8
102,69,15,91,192, //cvtps2dq %xmm8,%xmm8
- 102,69,15,86,193, //orpd %xmm9,%xmm8
+ 102,69,15,86,194, //orpd %xmm10,%xmm8
102,65,15,114,240,16, //pslld $0x10,%xmm8
102,65,15,114,224,16, //psrad $0x10,%xmm8
102,69,15,107,192, //packssdw %xmm8,%xmm8
@@ -12068,7 +12610,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
102,68,15,111,194, //movdqa %xmm2,%xmm8
102,68,15,97,192, //punpcklwd %xmm0,%xmm8
102,15,105,208, //punpckhwd %xmm0,%xmm2
- 102,15,110,66,100, //movd 0x64(%rdx),%xmm0
+ 184,0,4,0,4, //mov $0x4000400,%eax
+ 102,15,110,192, //movd %eax,%xmm0
102,15,112,216,0, //pshufd $0x0,%xmm0,%xmm3
102,15,111,203, //movdqa %xmm3,%xmm1
102,65,15,101,200, //pcmpgtw %xmm8,%xmm1
@@ -12079,7 +12622,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
102,15,111,193, //movdqa %xmm1,%xmm0
102,65,15,97,192, //punpcklwd %xmm8,%xmm0
102,15,114,240,13, //pslld $0xd,%xmm0
- 102,15,110,82,92, //movd 0x5c(%rdx),%xmm2
+ 184,0,0,128,119, //mov $0x77800000,%eax
+ 102,15,110,208, //movd %eax,%xmm2
102,68,15,112,202,0, //pshufd $0x0,%xmm2,%xmm9
65,15,89,193, //mulps %xmm9,%xmm0
102,65,15,105,200, //punpckhwd %xmm8,%xmm1
@@ -12099,7 +12643,8 @@ CODE const uint8_t sk_load_f16_sse2[] = {
CODE const uint8_t sk_store_f16_sse2[] = {
72,173, //lods %ds:(%rsi),%rax
72,139,0, //mov (%rax),%rax
- 102,68,15,110,66,96, //movd 0x60(%rdx),%xmm8
+ 185,0,0,128,7, //mov $0x7800000,%ecx
+ 102,68,15,110,193, //movd %ecx,%xmm8
102,69,15,112,192,0, //pshufd $0x0,%xmm8,%xmm8
102,69,15,111,200, //movdqa %xmm8,%xmm9
68,15,89,200, //mulps %xmm0,%xmm9
@@ -12289,17 +12834,20 @@ CODE const uint8_t sk_mirror_y_sse2[] = {
};
CODE const uint8_t sk_luminance_to_alpha_sse2[] = {
- 243,15,16,154,136,0,0,0, //movss 0x88(%rdx),%xmm3
- 243,68,15,16,130,140,0,0,0, //movss 0x8c(%rdx),%xmm8
+ 184,208,179,89,62, //mov $0x3e59b3d0,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,216, //mulps %xmm0,%xmm3
- 69,15,198,192,0, //shufps $0x0,%xmm8,%xmm8
- 68,15,89,193, //mulps %xmm1,%xmm8
- 68,15,88,195, //addps %xmm3,%xmm8
- 243,15,16,154,144,0,0,0, //movss 0x90(%rdx),%xmm3
+ 184,89,23,55,63, //mov $0x3f371759,%eax
+ 102,15,110,192, //movd %eax,%xmm0
+ 15,198,192,0, //shufps $0x0,%xmm0,%xmm0
+ 15,89,193, //mulps %xmm1,%xmm0
+ 15,88,195, //addps %xmm3,%xmm0
+ 184,152,221,147,61, //mov $0x3d93dd98,%eax
+ 102,15,110,216, //movd %eax,%xmm3
15,198,219,0, //shufps $0x0,%xmm3,%xmm3
15,89,218, //mulps %xmm2,%xmm3
- 65,15,88,216, //addps %xmm8,%xmm3
+ 15,88,216, //addps %xmm0,%xmm3
72,173, //lods %ds:(%rsi),%rax
15,87,192, //xorps %xmm0,%xmm0
15,87,201, //xorps %xmm1,%xmm1
diff --git a/src/jumper/SkJumper_stages.cpp b/src/jumper/SkJumper_stages.cpp
index 13c33cc91e..a4a8975f67 100644
--- a/src/jumper/SkJumper_stages.cpp
+++ b/src/jumper/SkJumper_stages.cpp
@@ -8,9 +8,6 @@
#include "SkJumper.h"
#include <string.h>
-// It's tricky to relocate code referencing ordinary constants, so we read them from this struct.
-using K = const SkJumper_constants;
-
template <typename T, typename P>
static T unaligned_load(const P* p) {
T v;
@@ -26,19 +23,22 @@ static Dst bit_cast(const Src& src) {
// A couple functions for embedding constants directly into code,
// so that no .const or .literal4 section is created.
-
-static inline int constant(int x) {
+static inline int C(int x) {
#if defined(JUMPER) && defined(__x86_64__)
// Move x-the-compile-time-constant as a literal into x-the-register.
asm("mov %1, %0" : "=r"(x) : "i"(x));
#endif
return x;
}
-
-static inline float constant(float f) {
- int x = constant(unaligned_load<int>(&f));
+static inline float C(float f) {
+ int x = C(unaligned_load<int>(&f));
return unaligned_load<float>(&x);
}
+static inline int operator "" _i(unsigned long long int i) { return C( (int)i); }
+static inline float operator "" _f( long double f) { return C((float)f); }
+
+// Not all constants can be generated using C() or _i/_f. We read the rest from this struct.
+using K = const SkJumper_constants;
#if !defined(JUMPER)
// This path should lead to portable code that can be compiled directly into Skia.
@@ -55,7 +55,7 @@ static inline float constant(float f) {
static F min(F a, F b) { return fminf(a,b); }
static F max(F a, F b) { return fmaxf(a,b); }
static F abs_ (F v) { return fabsf(v); }
- static F floor(F v, K*) { return floorf(v); }
+ static F floor(F v) { return floorf(v); }
static F rcp (F v) { return 1.0f / v; }
static F rsqrt(F v) { return 1.0f / sqrtf(v); }
static U32 round(F v, F scale) { return (uint32_t)lrintf(v*scale); }
@@ -83,7 +83,7 @@ static inline float constant(float f) {
static F min(F a, F b) { return vminq_f32(a,b); }
static F max(F a, F b) { return vmaxq_f32(a,b); }
static F abs_ (F v) { return vabsq_f32(v); }
- static F floor(F v, K*) { return vrndmq_f32(v); }
+ static F floor(F v) { return vrndmq_f32(v); }
static F rcp (F v) { auto e = vrecpeq_f32 (v); return vrecpsq_f32 (v,e ) * e; }
static F rsqrt(F v) { auto e = vrsqrteq_f32(v); return vrsqrtsq_f32(v,e*e) * e; }
static U32 round(F v, F scale) { return vcvtnq_u32_f32(v*scale); }
@@ -121,9 +121,9 @@ static inline float constant(float f) {
static F if_then_else(I32 c, F t, F e) { return vbsl_f32((U32)c,t,e); }
- static F floor(F v, K* k) {
+ static F floor(F v) {
F roundtrip = vcvt_f32_s32(vcvt_s32_f32(v));
- return roundtrip - if_then_else(roundtrip > v, constant(1.0f), 0);
+ return roundtrip - if_then_else(roundtrip > v, 1.0_f, 0);
}
static F gather(const float* p, U32 ix) { return {p[ix[0]], p[ix[1]]}; }
@@ -151,7 +151,7 @@ static inline float constant(float f) {
static F min(F a, F b) { return _mm256_min_ps(a,b); }
static F max(F a, F b) { return _mm256_max_ps(a,b); }
static F abs_(F v) { return _mm256_and_ps(v, 0-v); }
- static F floor(F v, K*) { return _mm256_floor_ps(v); }
+ static F floor(F v) { return _mm256_floor_ps(v); }
static F rcp (F v) { return _mm256_rcp_ps (v); }
static F rsqrt(F v) { return _mm256_rsqrt_ps(v); }
static U32 round(F v, F scale) { return _mm256_cvtps_epi32(v*scale); }
@@ -220,12 +220,12 @@ static inline float constant(float f) {
return _mm_or_ps(_mm_and_ps(c, t), _mm_andnot_ps(c, e));
}
- static F floor(F v, K* k) {
+ static F floor(F v) {
#if defined(__SSE4_1__)
return _mm_floor_ps(v);
#else
F roundtrip = _mm_cvtepi32_ps(_mm_cvttps_epi32(v));
- return roundtrip - if_then_else(roundtrip > v, constant(1.0f), 0);
+ return roundtrip - if_then_else(roundtrip > v, 1.0_f, 0);
#endif
}
@@ -345,11 +345,11 @@ static F lerp(F from, F to, F t) {
return mad(to-from, t, from);
}
-static void from_565(U16 _565, F* r, F* g, F* b, K* k) {
+static void from_565(U16 _565, F* r, F* g, F* b) {
U32 wide = expand(_565);
- *r = cast(wide & k->r_565_mask) * k->r_565_scale;
- *g = cast(wide & k->g_565_mask) * k->g_565_scale;
- *b = cast(wide & k->b_565_mask) * k->b_565_scale;
+ *r = cast(wide & C(31<<11)) * C(1.0f / (31<<11));
+ *g = cast(wide & C(63<< 5)) * C(1.0f / (63<< 5));
+ *b = cast(wide & C(31<< 0)) * C(1.0f / (31<< 0));
}
// Sometimes we want to work with 4 floats directly, regardless of the depth of the F vector.
@@ -503,10 +503,9 @@ STAGE(seed_shader) {
// It's important for speed to explicitly cast(x) and cast(y),
// which has the effect of splatting them to vectors before converting to floats.
// On Intel this breaks a data dependency on previous loop iterations' registers.
-
- r = cast(x) + constant(0.5f) + unaligned_load<F>(k->iota);
- g = cast(y) + constant(0.5f);
- b = constant(1.0f);
+ r = cast(x) + 0.5_f + unaligned_load<F>(k->iota);
+ g = cast(y) + 0.5_f;
+ b = 1.0_f;
a = 0;
dr = dg = db = da = 0;
}
@@ -531,14 +530,14 @@ STAGE(plus_) {
}
STAGE(srcover) {
- auto A = constant(1.0f) - a;
+ auto A = C(1.0f) - a;
r = mad(dr, A, r);
g = mad(dg, A, g);
b = mad(db, A, b);
a = mad(da, A, a);
}
STAGE(dstover) {
- auto DA = constant(1.0f) - da;
+ auto DA = 1.0_f - da;
r = mad(r, DA, dr);
g = mad(g, DA, dg);
b = mad(b, DA, db);
@@ -553,14 +552,14 @@ STAGE(clamp_0) {
}
STAGE(clamp_1) {
- r = min(r, constant(1.0f));
- g = min(g, constant(1.0f));
- b = min(b, constant(1.0f));
- a = min(a, constant(1.0f));
+ r = min(r, 1.0_f);
+ g = min(g, 1.0_f);
+ b = min(b, 1.0_f);
+ a = min(a, 1.0_f);
}
STAGE(clamp_a) {
- a = min(a, constant(1.0f));
+ a = min(a, 1.0_f);
r = min(r, a);
g = min(g, a);
b = min(b, a);
@@ -608,7 +607,7 @@ STAGE(premul) {
b = b * a;
}
STAGE(unpremul) {
- auto scale = if_then_else(a == 0, 0, constant(1.0f) / a);
+ auto scale = if_then_else(a == 0, 0, 1.0_f / a);
r = r * scale;
g = g * scale;
b = b * scale;
@@ -616,9 +615,9 @@ STAGE(unpremul) {
STAGE(from_srgb) {
auto fn = [&](F s) {
- auto lo = s * k->_1_1292;
- auto hi = mad(s*s, mad(s, k->_03000, k->_06975), k->_00025);
- return if_then_else(s < k->_0055, lo, hi);
+ auto lo = s * C(1/12.92f);
+ auto hi = mad(s*s, mad(s, 0.3000_f, 0.6975_f), 0.0025_f);
+ return if_then_else(s < 0.055_f, lo, hi);
};
r = fn(r);
g = fn(g);
@@ -628,11 +627,10 @@ STAGE(to_srgb) {
auto fn = [&](F l) {
F sqrt = rcp (rsqrt(l)),
ftrt = rsqrt(rsqrt(l));
- auto lo = l * k->_1246;
- auto hi = min(k->_1, mad(k->_0411192, ftrt,
- mad(k->_0689206, sqrt,
- k->n_00988)));
- return if_then_else(l < k->_00043, lo, hi);
+ auto lo = l * 12.46_f;
+ auto hi = min(1.0_f, mad(0.411192_f, ftrt,
+ mad(0.689206_f, sqrt, -0.0988_f)));
+ return if_then_else(l < 0.0043_f, lo, hi);
};
r = fn(r);
g = fn(g);
@@ -651,7 +649,7 @@ STAGE(scale_u8) {
auto ptr = *(const uint8_t**)ctx + x;
auto scales = load<U8>(ptr, tail);
- auto c = cast(expand(scales)) * constant(1/255.0f);
+ auto c = cast(expand(scales)) * C(1/255.0f);
r = r * c;
g = g * c;
@@ -671,7 +669,7 @@ STAGE(lerp_u8) {
auto ptr = *(const uint8_t**)ctx + x;
auto scales = load<U8>(ptr, tail);
- auto c = cast(expand(scales)) * constant(1/255.0f);
+ auto c = cast(expand(scales)) * C(1/255.0f);
r = lerp(dr, r, c);
g = lerp(dg, g, c);
@@ -682,12 +680,12 @@ STAGE(lerp_565) {
auto ptr = *(const uint16_t**)ctx + x;
F cr,cg,cb;
- from_565(load<U16>(ptr, tail), &cr, &cg, &cb, k);
+ from_565(load<U16>(ptr, tail), &cr, &cg, &cb);
r = lerp(dr, r, cr);
g = lerp(dg, g, cg);
b = lerp(db, b, cb);
- a = constant(1.0f);
+ a = 1.0_f;
}
STAGE(load_tables) {
@@ -698,37 +696,37 @@ STAGE(load_tables) {
auto c = (const Ctx*)ctx;
auto px = load<U32>(c->src + x, tail);
- r = gather(c->r, (px ) & k->_0x000000ff);
- g = gather(c->g, (px >> 8) & k->_0x000000ff);
- b = gather(c->b, (px >> 16) & k->_0x000000ff);
- a = cast( (px >> 24)) * k->_1_255;
+ r = gather(c->r, (px ) & 0xff_i);
+ g = gather(c->g, (px >> 8) & 0xff_i);
+ b = gather(c->b, (px >> 16) & 0xff_i);
+ a = cast( (px >> 24)) * C(1/255.0f);
}
STAGE(load_a8) {
auto ptr = *(const uint8_t**)ctx + x;
r = g = b = 0.0f;
- a = cast(expand(load<U8>(ptr, tail))) * k->_1_255;
+ a = cast(expand(load<U8>(ptr, tail))) * C(1/255.0f);
}
STAGE(store_a8) {
auto ptr = *(uint8_t**)ctx + x;
- U8 packed = pack(pack(round(a, k->_255)));
+ U8 packed = pack(pack(round(a, 255.0_f)));
store(ptr, packed, tail);
}
STAGE(load_565) {
auto ptr = *(const uint16_t**)ctx + x;
- from_565(load<U16>(ptr, tail), &r,&g,&b, k);
- a = k->_1;
+ from_565(load<U16>(ptr, tail), &r,&g,&b);
+ a = 1.0_f;
}
STAGE(store_565) {
auto ptr = *(uint16_t**)ctx + x;
- U16 px = pack( round(r, k->_31) << 11
- | round(g, k->_63) << 5
- | round(b, k->_31) );
+ U16 px = pack( round(r, 31.0_f) << 11
+ | round(g, 63.0_f) << 5
+ | round(b, 31.0_f) );
store(ptr, px, tail);
}
@@ -736,19 +734,19 @@ STAGE(load_8888) {
auto ptr = *(const uint32_t**)ctx + x;
auto px = load<U32>(ptr, tail);
- r = cast((px ) & constant(0xff)) * constant(1/255.0f);
- g = cast((px >> 8) & constant(0xff)) * constant(1/255.0f);
- b = cast((px >> 16) & constant(0xff)) * constant(1/255.0f);
- a = cast((px >> 24) ) * constant(1/255.0f);
+ r = cast((px ) & 0xff_i) * C(1/255.0f);
+ g = cast((px >> 8) & 0xff_i) * C(1/255.0f);
+ b = cast((px >> 16) & 0xff_i) * C(1/255.0f);
+ a = cast((px >> 24) ) * C(1/255.0f);
}
STAGE(store_8888) {
auto ptr = *(uint32_t**)ctx + x;
- U32 px = round(r, constant(255.0f))
- | round(g, constant(255.0f)) << 8
- | round(b, constant(255.0f)) << 16
- | round(a, constant(255.0f)) << 24;
+ U32 px = round(r, 255.0_f)
+ | round(g, 255.0_f) << 8
+ | round(b, 255.0_f) << 16
+ | round(a, 255.0_f) << 24;
store(ptr, px, tail);
}
@@ -757,9 +755,9 @@ STAGE(load_f16) {
#if !defined(JUMPER)
auto half_to_float = [&](int16_t h) {
- if (h < 0x0400) { h = 0; } // Flush denorm and negative to zero.
- return bit_cast<F>(h << 13) // Line up the mantissa,
- * bit_cast<F>(U32(k->_0x77800000)); // then fix up the exponent.
+ if (h < 0x0400) { h = 0; } // Flush denorm and negative to zero.
+ return bit_cast<F>(h << 13) // Line up the mantissa,
+ * bit_cast<F>(U32(0x77800000)); // then fix up the exponent.
};
auto rgba = (const int16_t*)ptr;
r = half_to_float(rgba[0]);
@@ -844,8 +842,8 @@ STAGE(load_f16) {
// half_to_float() slows down ~10x for denorm inputs, so we flush them to zero.
// With a signed comparison this conveniently also flushes negative half floats to zero.
- auto ftz = [k](__m128i v) {
- return _mm_andnot_si128(_mm_cmplt_epi16(v, _mm_set1_epi32(k->_0x04000400)), v);
+ auto ftz = [](__m128i v) {
+ return _mm_andnot_si128(_mm_cmplt_epi16(v, _mm_set1_epi32(0x04000400_i)), v);
};
rg0123 = ftz(rg0123);
ba0123 = ftz(ba0123);
@@ -862,8 +860,8 @@ STAGE(load_f16) {
_mm_unpackhi_epi16(ba4567, _mm_setzero_si128()));
auto half_to_float = [&](U32 h) {
- return bit_cast<F>(h << 13) // Line up the mantissa,
- * bit_cast<F>(U32(k->_0x77800000)); // then fix up the exponent.
+ return bit_cast<F>(h << 13) // Line up the mantissa,
+ * bit_cast<F>(U32(0x77800000_i)); // then fix up the exponent.
};
r = half_to_float(R);
@@ -882,15 +880,15 @@ STAGE(load_f16) {
ba = _mm_unpackhi_epi16(_02, _13); // b0 b1 b2 b3 a0 a1 a2 a3
// Same deal as AVX, flush denorms and negatives to zero.
- auto ftz = [k](__m128i v) {
- return _mm_andnot_si128(_mm_cmplt_epi16(v, _mm_set1_epi32(k->_0x04000400)), v);
+ auto ftz = [](__m128i v) {
+ return _mm_andnot_si128(_mm_cmplt_epi16(v, _mm_set1_epi32(0x04000400_i)), v);
};
rg = ftz(rg);
ba = ftz(ba);
auto half_to_float = [&](U32 h) {
- return bit_cast<F>(h << 13) // Line up the mantissa,
- * bit_cast<F>(U32(k->_0x77800000)); // then fix up the exponent.
+ return bit_cast<F>(h << 13) // Line up the mantissa,
+ * bit_cast<F>(U32(0x77800000_i)); // then fix up the exponent.
};
r = half_to_float(_mm_unpacklo_epi16(rg, _mm_setzero_si128()));
@@ -905,8 +903,8 @@ STAGE(store_f16) {
#if !defined(JUMPER)
auto float_to_half = [&](F f) {
- return bit_cast<U32>(f * bit_cast<F>(U32(k->_0x07800000))) // Fix up the exponent,
- >> 13; // then line up the mantissa.
+ return bit_cast<U32>(f * bit_cast<F>(U32(0x07800000_i))) // Fix up the exponent,
+ >> 13; // then line up the mantissa.
};
auto rgba = (int16_t*)ptr;
rgba[0] = float_to_half(r);
@@ -960,8 +958,8 @@ STAGE(store_f16) {
}
#elif defined(__AVX__)
auto float_to_half = [&](F f) {
- return bit_cast<U32>(f * bit_cast<F>(U32(k->_0x07800000))) // Fix up the exponent,
- >> 13; // then line up the mantissa.
+ return bit_cast<U32>(f * bit_cast<F>(U32(0x07800000_i))) // Fix up the exponent,
+ >> 13; // then line up the mantissa.
};
U32 R = float_to_half(r),
G = float_to_half(g),
@@ -1002,8 +1000,8 @@ STAGE(store_f16) {
}
#elif defined(__SSE2__)
auto float_to_half = [&](F f) {
- return bit_cast<U32>(f * bit_cast<F>(U32(k->_0x07800000))) // Fix up the exponent,
- >> 13; // then line up the mantissa.
+ return bit_cast<U32>(f * bit_cast<F>(U32(0x07800000_i))) // Fix up the exponent,
+ >> 13; // then line up the mantissa.
};
U32 R = float_to_half(r),
G = float_to_half(g),
@@ -1070,27 +1068,27 @@ STAGE(store_f32) {
static F ulp_before(F v) {
return bit_cast<F>(bit_cast<U32>(v) + U32(0xffffffff));
}
-static F clamp(F v, float limit, K*) {
+static F clamp(F v, float limit) {
v = max(0, v);
return min(v, ulp_before(limit));
}
-static F repeat(F v, float limit, K* k) {
- v = v - floor(v/limit, k)*limit;
+static F repeat(F v, float limit) {
+ v = v - floor(v/limit)*limit;
return min(v, ulp_before(limit));
}
-static F mirror(F v, float limit, K* k) {
- v = abs_( (v-limit) - (limit+limit)*floor((v-limit)/(limit+limit),k) - limit );
+static F mirror(F v, float limit) {
+ v = abs_( (v-limit) - (limit+limit)*floor((v-limit)/(limit+limit)) - limit );
return min(v, ulp_before(limit));
}
-STAGE(clamp_x) { r = clamp (r, *(const float*)ctx, k); }
-STAGE(clamp_y) { g = clamp (g, *(const float*)ctx, k); }
-STAGE(repeat_x) { r = repeat(r, *(const float*)ctx, k); }
-STAGE(repeat_y) { g = repeat(g, *(const float*)ctx, k); }
-STAGE(mirror_x) { r = mirror(r, *(const float*)ctx, k); }
-STAGE(mirror_y) { g = mirror(g, *(const float*)ctx, k); }
+STAGE(clamp_x) { r = clamp (r, *(const float*)ctx); }
+STAGE(clamp_y) { g = clamp (g, *(const float*)ctx); }
+STAGE(repeat_x) { r = repeat(r, *(const float*)ctx); }
+STAGE(repeat_y) { g = repeat(g, *(const float*)ctx); }
+STAGE(mirror_x) { r = mirror(r, *(const float*)ctx); }
+STAGE(mirror_y) { g = mirror(g, *(const float*)ctx); }
STAGE(luminance_to_alpha) {
- a = r*k->lum_r + g*k->lum_g + b*k->lum_b;
+ a = r*0.2126_f + g*0.7152_f + b*0.0722_f;
r = g = b = 0;
}
diff --git a/src/jumper/build_stages.py b/src/jumper/build_stages.py
index 118291d16d..7513b0df09 100755
--- a/src/jumper/build_stages.py
+++ b/src/jumper/build_stages.py
@@ -78,14 +78,14 @@ def parse_object_file(dot_o, array_type, target=None):
# Look for sections we know we can't handle.
section_headers = subprocess.check_output(cmd + ['-h', dot_o])
- for section in ['.literal4', '.literal8', '.literal16', '.const']:
- if section in section_headers:
- print >>sys.stderr, 'Found %s section, which we cannot handle.' % section
- assert section not in section_headers
+ for snippet in ['.literal', '.const', '.rodata']:
+ if snippet in section_headers:
+ print >>sys.stderr, 'Found %s in section.' % snippet
+ assert snippet not in section_headers
# Ok. Let's disassemble.
active = False
- disassemble = ['-d', '--insn-width=9', dot_o]
+ disassemble = ['-d', '--insn-width=10', dot_o]
for line in subprocess.check_output(cmd + disassemble).split('\n'):
line = line.strip()