aboutsummaryrefslogtreecommitdiffhomepage
diff options
context:
space:
mode:
-rw-r--r--src/jumper/SkJumper_generated.S342
-rw-r--r--src/jumper/SkJumper_generated_win.S83
-rwxr-xr-xsrc/jumper/build_stages.py34
3 files changed, 198 insertions, 261 deletions
diff --git a/src/jumper/SkJumper_generated.S b/src/jumper/SkJumper_generated.S
index 7da14891ae..06a5e5edf8 100644
--- a/src/jumper/SkJumper_generated.S
+++ b/src/jumper/SkJumper_generated.S
@@ -12,21 +12,19 @@
.globl _sk_start_pipeline_aarch64
_sk_start_pipeline_aarch64:
- .long 0xa9bc5ff8 // stp x24, x23, [sp,#-64]!
- .long 0xa90157f6 // stp x22, x21, [sp,#16]
- .long 0xa9024ff4 // stp x20, x19, [sp,#32]
- .long 0xa9037bfd // stp x29, x30, [sp,#48]
- .long 0x9100c3fd // add x29, sp, #0x30
- .long 0xaa0103f3 // mov x19, x1
- .long 0xf8408677 // ldr x23, [x19],#8
+ .long 0xa9bd5bf7 // stp x23, x22, [sp,#-48]!
+ .long 0xa90153f5 // stp x21, x20, [sp,#16]
+ .long 0xa9027bf3 // stp x19, x30, [sp,#32]
+ .long 0xaa0103f5 // mov x21, x1
+ .long 0xf84086b7 // ldr x23, [x21],#8
.long 0xaa0003f6 // mov x22, x0
- .long 0xaa0303f4 // mov x20, x3
- .long 0xaa0203f5 // mov x21, x2
+ .long 0xaa0303f3 // mov x19, x3
+ .long 0xaa0203f4 // mov x20, x2
.long 0x910012c8 // add x8, x22, #0x4
- .long 0xeb14011f // cmp x8, x20
- .long 0x54000069 // b.ls 3c <sk_start_pipeline_aarch64+0x3c>
+ .long 0xeb13011f // cmp x8, x19
+ .long 0x54000069 // b.ls 34 <sk_start_pipeline_aarch64+0x34>
.long 0xaa1603e0 // mov x0, x22
- .long 0x14000012 // b 80 <sk_start_pipeline_aarch64+0x80>
+ .long 0x14000012 // b 78 <sk_start_pipeline_aarch64+0x78>
.long 0x6f00e400 // movi v0.2d, #0x0
.long 0x6f00e401 // movi v1.2d, #0x0
.long 0x6f00e402 // movi v2.2d, #0x0
@@ -36,18 +34,17 @@ _sk_start_pipeline_aarch64:
.long 0x6f00e406 // movi v6.2d, #0x0
.long 0x6f00e407 // movi v7.2d, #0x0
.long 0xaa1603e0 // mov x0, x22
- .long 0xaa1303e1 // mov x1, x19
- .long 0xaa1503e2 // mov x2, x21
+ .long 0xaa1503e1 // mov x1, x21
+ .long 0xaa1403e2 // mov x2, x20
.long 0xd63f02e0 // blr x23
.long 0x910022c8 // add x8, x22, #0x8
.long 0x910012c0 // add x0, x22, #0x4
- .long 0xeb14011f // cmp x8, x20
+ .long 0xeb13011f // cmp x8, x19
.long 0xaa0003f6 // mov x22, x0
- .long 0x54fffe09 // b.ls 3c <sk_start_pipeline_aarch64+0x3c>
- .long 0xa9437bfd // ldp x29, x30, [sp,#48]
- .long 0xa9424ff4 // ldp x20, x19, [sp,#32]
- .long 0xa94157f6 // ldp x22, x21, [sp,#16]
- .long 0xa8c45ff8 // ldp x24, x23, [sp],#64
+ .long 0x54fffe09 // b.ls 34 <sk_start_pipeline_aarch64+0x34>
+ .long 0xa9427bf3 // ldp x19, x30, [sp,#32]
+ .long 0xa94153f5 // ldp x21, x20, [sp,#16]
+ .long 0xa8c35bf7 // ldp x23, x22, [sp],#48
.long 0xd65f03c0 // ret
.globl _sk_just_return_aarch64
@@ -57,22 +54,24 @@ _sk_just_return_aarch64:
.globl _sk_seed_shader_aarch64
_sk_seed_shader_aarch64:
.long 0xaa0203e9 // mov x9, x2
- .long 0xa8c10c28 // ldp x8, x3, [x1],#16
+ .long 0xa9400c28 // ldp x8, x3, [x1]
.long 0x4ddfc922 // ld1r {v2.4s}, [x9], #4
- .long 0x3cc14041 // ldur q1, [x2,#20]
+ .long 0x3cc14047 // ldur q7, [x2,#20]
.long 0x4e040c00 // dup v0.4s, w0
- .long 0x4d40c903 // ld1r {v3.4s}, [x8]
- .long 0x4d40c924 // ld1r {v4.4s}, [x9]
+ .long 0x4d40c901 // ld1r {v1.4s}, [x8]
+ .long 0x4d40c926 // ld1r {v6.4s}, [x9]
.long 0x4e21d800 // scvtf v0.4s, v0.4s
- .long 0x6f00e405 // movi v5.2d, #0x0
- .long 0x4e21d863 // scvtf v3.4s, v3.4s
- .long 0x4e24d400 // fadd v0.4s, v0.4s, v4.4s
- .long 0x4e20d420 // fadd v0.4s, v1.4s, v0.4s
- .long 0x4e24d461 // fadd v1.4s, v3.4s, v4.4s
+ .long 0x91004028 // add x8, x1, #0x10
+ .long 0x4e21d821 // scvtf v1.4s, v1.4s
+ .long 0x4e26d400 // fadd v0.4s, v0.4s, v6.4s
.long 0x6f00e403 // movi v3.2d, #0x0
.long 0x6f00e404 // movi v4.2d, #0x0
+ .long 0x6f00e405 // movi v5.2d, #0x0
+ .long 0x4e26d421 // fadd v1.4s, v1.4s, v6.4s
.long 0x6f00e406 // movi v6.2d, #0x0
+ .long 0x4e20d4e0 // fadd v0.4s, v7.4s, v0.4s
.long 0x6f00e407 // movi v7.2d, #0x0
+ .long 0xaa0803e1 // mov x1, x8
.long 0xd61f0060 // br x3
.globl _sk_constant_color_aarch64
@@ -174,10 +173,11 @@ _sk_clamp_a_aarch64:
.globl _sk_set_rgb_aarch64
_sk_set_rgb_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x91002109 // add x9, x8, #0x8
- .long 0x4ddfc900 // ld1r {v0.4s}, [x8], #4
- .long 0x4d40c922 // ld1r {v2.4s}, [x9]
- .long 0x4d40c901 // ld1r {v1.4s}, [x8]
+ .long 0xaa0803e9 // mov x9, x8
+ .long 0x4ddfc920 // ld1r {v0.4s}, [x9], #4
+ .long 0x91002108 // add x8, x8, #0x8
+ .long 0x4d40c902 // ld1r {v2.4s}, [x8]
+ .long 0x4d40c921 // ld1r {v1.4s}, [x9]
.long 0xd61f0060 // br x3
.globl _sk_swap_rb_aarch64
@@ -254,8 +254,8 @@ _sk_from_srgb_aarch64:
.long 0x9100e048 // add x8, x2, #0x38
.long 0x4d40c910 // ld1r {v16.4s}, [x8]
.long 0x9100d048 // add x8, x2, #0x34
- .long 0x4d40c911 // ld1r {v17.4s}, [x8]
.long 0x2d47cc52 // ldp s18, s19, [x2,#60]
+ .long 0x4d40c911 // ld1r {v17.4s}, [x8]
.long 0x6e22dc54 // fmul v20.4s, v2.4s, v2.4s
.long 0x4eb01e15 // mov v21.16b, v16.16b
.long 0x4eb01e17 // mov v23.16b, v16.16b
@@ -296,7 +296,6 @@ _sk_to_srgb_aarch64:
.long 0x6e36deda // fmul v26.4s, v22.4s, v22.4s
.long 0x4eb9fc39 // frsqrts v25.4s, v1.4s, v25.4s
.long 0x6e37de31 // fmul v17.4s, v17.4s, v23.4s
- .long 0x2d494052 // ldp s18, s16, [x2,#72]
.long 0x4d40c914 // ld1r {v20.4s}, [x8]
.long 0x4ebafc5a // frsqrts v26.4s, v2.4s, v26.4s
.long 0x6e39deb5 // fmul v21.4s, v21.4s, v25.4s
@@ -306,6 +305,7 @@ _sk_to_srgb_aarch64:
.long 0x6e3aded6 // fmul v22.4s, v22.4s, v26.4s
.long 0x4ea1dabb // frecpe v27.4s, v21.4s
.long 0x4e37fe3d // frecps v29.4s, v17.4s, v23.4s
+ .long 0x2d494052 // ldp s18, s16, [x2,#72]
.long 0x4d40c918 // ld1r {v24.4s}, [x8]
.long 0x4ea1dadc // frecpe v28.4s, v22.4s
.long 0x6e3ddef7 // fmul v23.4s, v23.4s, v29.4s
@@ -438,15 +438,15 @@ _sk_lerp_u8_aarch64:
_sk_lerp_565_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
.long 0xd37ff809 // lsl x9, x0, #1
+ .long 0x2d4ec851 // ldp s17, s18, [x2,#116]
.long 0x4ea4d413 // fsub v19.4s, v0.4s, v4.4s
- .long 0x4ea41c80 // mov v0.16b, v4.16b
.long 0xf9400108 // ldr x8, [x8]
+ .long 0x4ea41c80 // mov v0.16b, v4.16b
.long 0xfc696903 // ldr d3, [x8,x9]
.long 0x9101a048 // add x8, x2, #0x68
.long 0x4d40c910 // ld1r {v16.4s}, [x8]
- .long 0x2d4ec851 // ldp s17, s18, [x2,#116]
- .long 0x2f10a463 // uxtl v3.4s, v3.4h
.long 0x9101b048 // add x8, x2, #0x6c
+ .long 0x2f10a463 // uxtl v3.4s, v3.4h
.long 0x4e231e10 // and v16.16b, v16.16b, v3.16b
.long 0x4e21da10 // scvtf v16.4s, v16.4s
.long 0x4f919210 // fmul v16.4s, v16.4s, v17.s[0]
@@ -533,20 +533,20 @@ _sk_load_a8_aarch64:
.long 0x6f00e400 // movi v0.2d, #0x0
.long 0x6f00e401 // movi v1.2d, #0x0
.long 0xf9400108 // ldr x8, [x8]
- .long 0x6f00e402 // movi v2.2d, #0x0
.long 0x8b000108 // add x8, x8, x0
.long 0x39400109 // ldrb w9, [x8]
.long 0x3940050a // ldrb w10, [x8,#1]
.long 0x3940090b // ldrb w11, [x8,#2]
.long 0x39400d08 // ldrb w8, [x8,#3]
- .long 0x4e021d30 // mov v16.h[0], w9
- .long 0x4e061d50 // mov v16.h[1], w10
- .long 0x4e0a1d70 // mov v16.h[2], w11
- .long 0x4e0e1d10 // mov v16.h[3], w8
- .long 0x2f07b7f0 // bic v16.4h, #0xff, lsl #8
- .long 0x2f10a610 // uxtl v16.4s, v16.4h
- .long 0x6e21da10 // ucvtf v16.4s, v16.4s
- .long 0x4f839203 // fmul v3.4s, v16.4s, v3.s[0]
+ .long 0x4e021d22 // mov v2.h[0], w9
+ .long 0x4e061d42 // mov v2.h[1], w10
+ .long 0x4e0a1d62 // mov v2.h[2], w11
+ .long 0x4e0e1d02 // mov v2.h[3], w8
+ .long 0x2f07b7e2 // bic v2.4h, #0xff, lsl #8
+ .long 0x2f10a442 // uxtl v2.4s, v2.4h
+ .long 0x6e21d842 // ucvtf v2.4s, v2.4s
+ .long 0x4f839043 // fmul v3.4s, v2.4s, v3.s[0]
+ .long 0x6f00e402 // movi v2.2d, #0x0
.long 0xd61f0060 // br x3
.globl _sk_store_a8_aarch64
@@ -599,14 +599,14 @@ _sk_load_565_aarch64:
.globl _sk_store_565_aarch64
_sk_store_565_aarch64:
- .long 0xf9400028 // ldr x8, [x1]
.long 0x2d504450 // ldp s16, s17, [x2,#128]
+ .long 0xf9400028 // ldr x8, [x1]
.long 0xd37ff809 // lsl x9, x0, #1
- .long 0xf9400108 // ldr x8, [x8]
.long 0x4f909012 // fmul v18.4s, v0.4s, v16.s[0]
.long 0x4f919031 // fmul v17.4s, v1.4s, v17.s[0]
.long 0x6e21aa52 // fcvtnu v18.4s, v18.4s
.long 0x6e21aa31 // fcvtnu v17.4s, v17.4s
+ .long 0xf9400108 // ldr x8, [x8]
.long 0x4f909050 // fmul v16.4s, v2.4s, v16.s[0]
.long 0x4f2b5652 // shl v18.4s, v18.4s, #11
.long 0x4f255631 // shl v17.4s, v17.4s, #5
@@ -698,8 +698,8 @@ _sk_store_f16_aarch64:
.globl _sk_clamp_x_aarch64
_sk_clamp_x_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f00e410 // movi v16.2d, #0x0
- .long 0x4e20f600 // fmax v0.4s, v16.4s, v0.4s
+ .long 0x6f00e411 // movi v17.2d, #0x0
+ .long 0x4e20f620 // fmax v0.4s, v17.4s, v0.4s
.long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
.long 0x4d40c910 // ld1r {v16.4s}, [x8]
.long 0x4eb18610 // add v16.4s, v16.4s, v17.4s
@@ -709,8 +709,8 @@ _sk_clamp_x_aarch64:
.globl _sk_clamp_y_aarch64
_sk_clamp_y_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f00e410 // movi v16.2d, #0x0
- .long 0x4e21f601 // fmax v1.4s, v16.4s, v1.4s
+ .long 0x6f00e411 // movi v17.2d, #0x0
+ .long 0x4e21f621 // fmax v1.4s, v17.4s, v1.4s
.long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
.long 0x4d40c910 // ld1r {v16.4s}, [x8]
.long 0x4eb18610 // add v16.4s, v16.4s, v17.4s
@@ -720,67 +720,67 @@ _sk_clamp_y_aarch64:
.globl _sk_repeat_x_aarch64
_sk_repeat_x_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
- .long 0xbd400111 // ldr s17, [x8]
- .long 0x4e040632 // dup v18.4s, v17.s[0]
- .long 0x4eb08650 // add v16.4s, v18.4s, v16.4s
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x4eb18651 // add v17.4s, v18.4s, v17.4s
.long 0x6e32fc12 // fdiv v18.4s, v0.4s, v18.4s
.long 0x4e219a52 // frintm v18.4s, v18.4s
- .long 0x4f919251 // fmul v17.4s, v18.4s, v17.s[0]
- .long 0x4eb1d400 // fsub v0.4s, v0.4s, v17.4s
- .long 0x4eb0f400 // fmin v0.4s, v0.4s, v16.4s
+ .long 0x4f909250 // fmul v16.4s, v18.4s, v16.s[0]
+ .long 0x4eb0d400 // fsub v0.4s, v0.4s, v16.4s
+ .long 0x4eb1f400 // fmin v0.4s, v0.4s, v17.4s
.long 0xd61f0060 // br x3
.globl _sk_repeat_y_aarch64
_sk_repeat_y_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
- .long 0xbd400111 // ldr s17, [x8]
- .long 0x4e040632 // dup v18.4s, v17.s[0]
- .long 0x4eb08650 // add v16.4s, v18.4s, v16.4s
+ .long 0x6f07e7f1 // movi v17.2d, #0xffffffffffffffff
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x4eb18651 // add v17.4s, v18.4s, v17.4s
.long 0x6e32fc32 // fdiv v18.4s, v1.4s, v18.4s
.long 0x4e219a52 // frintm v18.4s, v18.4s
- .long 0x4f919251 // fmul v17.4s, v18.4s, v17.s[0]
- .long 0x4eb1d421 // fsub v1.4s, v1.4s, v17.4s
- .long 0x4eb0f421 // fmin v1.4s, v1.4s, v16.4s
+ .long 0x4f909250 // fmul v16.4s, v18.4s, v16.s[0]
+ .long 0x4eb0d421 // fsub v1.4s, v1.4s, v16.4s
+ .long 0x4eb1f421 // fmin v1.4s, v1.4s, v17.4s
.long 0xd61f0060 // br x3
.globl _sk_mirror_x_aarch64
_sk_mirror_x_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
- .long 0xbd400111 // ldr s17, [x8]
- .long 0x4e040632 // dup v18.4s, v17.s[0]
- .long 0x1e312a31 // fadd s17, s17, s17
- .long 0x4eb2d400 // fsub v0.4s, v0.4s, v18.4s
- .long 0x4e040633 // dup v19.4s, v17.s[0]
- .long 0x6e33fc13 // fdiv v19.4s, v0.4s, v19.4s
- .long 0x4e219a73 // frintm v19.4s, v19.4s
- .long 0x4f919271 // fmul v17.4s, v19.4s, v17.s[0]
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040611 // dup v17.4s, v16.s[0]
+ .long 0x1e302a10 // fadd s16, s16, s16
+ .long 0x4eb1d400 // fsub v0.4s, v0.4s, v17.4s
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x6e32fc12 // fdiv v18.4s, v0.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f909250 // fmul v16.4s, v18.4s, v16.s[0]
+ .long 0x4eb0d400 // fsub v0.4s, v0.4s, v16.4s
+ .long 0x6f07e7f2 // movi v18.2d, #0xffffffffffffffff
.long 0x4eb1d400 // fsub v0.4s, v0.4s, v17.4s
- .long 0x4eb2d400 // fsub v0.4s, v0.4s, v18.4s
+ .long 0x4eb28632 // add v18.4s, v17.4s, v18.4s
.long 0x4ea0f800 // fabs v0.4s, v0.4s
- .long 0x4eb08650 // add v16.4s, v18.4s, v16.4s
- .long 0x4eb0f400 // fmin v0.4s, v0.4s, v16.4s
+ .long 0x4eb2f400 // fmin v0.4s, v0.4s, v18.4s
.long 0xd61f0060 // br x3
.globl _sk_mirror_y_aarch64
_sk_mirror_y_aarch64:
.long 0xa8c10c28 // ldp x8, x3, [x1],#16
- .long 0x6f07e7f0 // movi v16.2d, #0xffffffffffffffff
- .long 0xbd400111 // ldr s17, [x8]
- .long 0x4e040632 // dup v18.4s, v17.s[0]
- .long 0x1e312a31 // fadd s17, s17, s17
- .long 0x4eb2d421 // fsub v1.4s, v1.4s, v18.4s
- .long 0x4e040633 // dup v19.4s, v17.s[0]
- .long 0x6e33fc33 // fdiv v19.4s, v1.4s, v19.4s
- .long 0x4e219a73 // frintm v19.4s, v19.4s
- .long 0x4f919271 // fmul v17.4s, v19.4s, v17.s[0]
+ .long 0xbd400110 // ldr s16, [x8]
+ .long 0x4e040611 // dup v17.4s, v16.s[0]
+ .long 0x1e302a10 // fadd s16, s16, s16
+ .long 0x4eb1d421 // fsub v1.4s, v1.4s, v17.4s
+ .long 0x4e040612 // dup v18.4s, v16.s[0]
+ .long 0x6e32fc32 // fdiv v18.4s, v1.4s, v18.4s
+ .long 0x4e219a52 // frintm v18.4s, v18.4s
+ .long 0x4f909250 // fmul v16.4s, v18.4s, v16.s[0]
+ .long 0x4eb0d421 // fsub v1.4s, v1.4s, v16.4s
+ .long 0x6f07e7f2 // movi v18.2d, #0xffffffffffffffff
.long 0x4eb1d421 // fsub v1.4s, v1.4s, v17.4s
- .long 0x4eb2d421 // fsub v1.4s, v1.4s, v18.4s
+ .long 0x4eb28632 // add v18.4s, v17.4s, v18.4s
.long 0x4ea0f821 // fabs v1.4s, v1.4s
- .long 0x4eb08650 // add v16.4s, v18.4s, v16.4s
- .long 0x4eb0f421 // fmin v1.4s, v1.4s, v16.4s
+ .long 0x4eb2f421 // fmin v1.4s, v1.4s, v18.4s
.long 0xd61f0060 // br x3
.globl _sk_matrix_2x3_aarch64
@@ -816,15 +816,15 @@ _sk_matrix_3x4_aarch64:
.long 0xbd402116 // ldr s22, [x8,#32]
.long 0x4d40c952 // ld1r {v18.4s}, [x10]
.long 0x4f941050 // fmla v16.4s, v2.4s, v20.s[0]
- .long 0x2d415d14 // ldp s20, s23, [x8,#8]
.long 0x4f951051 // fmla v17.4s, v2.4s, v21.s[0]
.long 0x4f961052 // fmla v18.4s, v2.4s, v22.s[0]
.long 0x2d425502 // ldp s2, s21, [x8,#16]
- .long 0x4f971030 // fmla v16.4s, v1.4s, v23.s[0]
- .long 0x4e20ce70 // fmla v16.4s, v19.4s, v0.4s
+ .long 0x2d415d14 // ldp s20, s23, [x8,#8]
.long 0x4f821031 // fmla v17.4s, v1.4s, v2.s[0]
.long 0xbd400122 // ldr s2, [x9]
+ .long 0x4f971030 // fmla v16.4s, v1.4s, v23.s[0]
.long 0x4f951032 // fmla v18.4s, v1.4s, v21.s[0]
+ .long 0x4e20ce70 // fmla v16.4s, v19.4s, v0.4s
.long 0x4f941012 // fmla v18.4s, v0.4s, v20.s[0]
.long 0x4f821011 // fmla v17.4s, v0.4s, v2.s[0]
.long 0x4eb01e00 // mov v0.16b, v16.16b
@@ -911,11 +911,10 @@ _sk_just_return_vfp4:
.globl _sk_seed_shader_vfp4
_sk_seed_shader_vfp4:
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xee800b90 // vdup.32 d16, r0
.long 0xf3fb0620 // vcvt.f32.s32 d16, d16
.long 0xedd23b05 // vldr d19, [r2, #20]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2803010 // vmov.i32 d3, #0
.long 0xf4e31c9f // vld1.32 {d17[]}, [r3 :32]
.long 0xe2823004 // add r3, r2, #4
@@ -934,8 +933,7 @@ _sk_seed_shader_vfp4:
.globl _sk_constant_color_vfp4
_sk_constant_color_vfp4:
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xf4630a0f // vld1.8 {d16-d17}, [r3]
.long 0xf3b40c20 // vdup.32 d0, d16[0]
@@ -1230,8 +1228,7 @@ _sk_to_srgb_vfp4:
.globl _sk_scale_1_float_vfp4
_sk_scale_1_float_vfp4:
.long 0xed2d8b02 // vpush {d8}
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xed938a00 // vldr s16, [r3]
.long 0xf2a00948 // vmul.f32 d0, d0, d8[0]
@@ -1245,8 +1242,7 @@ _sk_scale_1_float_vfp4:
_sk_scale_u8_vfp4:
.long 0xed2d8b02 // vpush {d8}
.long 0xe24dd008 // sub sp, sp, #8
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xe5933000 // ldr r3, [r3]
.long 0xe0833000 // add r3, r3, r0
@@ -1269,10 +1265,9 @@ _sk_scale_u8_vfp4:
.globl _sk_lerp_1_float_vfp4
_sk_lerp_1_float_vfp4:
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2600d04 // vsub.f32 d16, d0, d4
.long 0xf2611d05 // vsub.f32 d17, d1, d5
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2622d06 // vsub.f32 d18, d2, d6
.long 0xe2811008 // add r1, r1, #8
.long 0xf2633d07 // vsub.f32 d19, d3, d7
@@ -1291,9 +1286,8 @@ _sk_lerp_1_float_vfp4:
_sk_lerp_u8_vfp4:
.long 0xed2d8b02 // vpush {d8}
.long 0xe24dd008 // sub sp, sp, #8
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2612d05 // vsub.f32 d18, d1, d5
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2623d06 // vsub.f32 d19, d2, d6
.long 0xf2634d07 // vsub.f32 d20, d3, d7
.long 0xe2811008 // add r1, r1, #8
@@ -1325,9 +1319,8 @@ _sk_lerp_u8_vfp4:
_sk_lerp_565_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
.long 0xe24dd008 // sub sp, sp, #8
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2603d04 // vsub.f32 d19, d0, d4
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2240114 // vorr d0, d4, d4
.long 0xe2811008 // add r1, r1, #8
.long 0xe5933000 // ldr r3, [r3]
@@ -1369,19 +1362,16 @@ _sk_lerp_565_vfp4:
.globl _sk_load_tables_vfp4
_sk_load_tables_vfp4:
.long 0xe92d48f0 // push {r4, r5, r6, r7, fp, lr}
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2826010 // add r6, r2, #16
- .long 0xed922a03 // vldr s4, [r2, #12]
- .long 0xf4e60c9f // vld1.32 {d16[]}, [r6 :32]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xe2811008 // add r1, r1, #8
.long 0xe593e000 // ldr lr, [r3]
- .long 0xe5934004 // ldr r4, [r3, #4]
+ .long 0xe99300b0 // ldmib r3, {r4, r5, r7}
+ .long 0xf4e60c9f // vld1.32 {d16[]}, [r6 :32]
.long 0xe08e6100 // add r6, lr, r0, lsl #2
- .long 0xe5935008 // ldr r5, [r3, #8]
- .long 0xe593700c // ldr r7, [r3, #12]
.long 0xedd61b00 // vldr d17, [r6]
.long 0xf24021b1 // vand d18, d16, d17
+ .long 0xed922a03 // vldr s4, [r2, #12]
.long 0xf3f03031 // vshr.u32 d19, d17, #16
.long 0xee326b90 // vmov.32 r6, d18[1]
.long 0xe0846106 // add r6, r4, r6, lsl #2
@@ -1413,10 +1403,9 @@ _sk_load_tables_vfp4:
.globl _sk_load_a8_vfp4
_sk_load_a8_vfp4:
.long 0xe24dd004 // sub sp, sp, #4
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xf2801010 // vmov.i32 d1, #0
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
+ .long 0xf2801010 // vmov.i32 d1, #0
.long 0xf2802010 // vmov.i32 d2, #0
.long 0xe5933000 // ldr r3, [r3]
.long 0xe0833000 // add r3, r3, r0
@@ -1455,8 +1444,7 @@ _sk_store_a8_vfp4:
.globl _sk_load_565_vfp4
_sk_load_565_vfp4:
.long 0xe24dd004 // sub sp, sp, #4
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xe5933000 // ldr r3, [r3]
.long 0xe7933080 // ldr r3, [r3, r0, lsl #1]
@@ -1517,10 +1505,9 @@ _sk_store_565_vfp4:
.globl _sk_load_8888_vfp4
_sk_load_8888_vfp4:
.long 0xe92d4800 // push {fp, lr}
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xed922a03 // vldr s4, [r2, #12]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
+ .long 0xed922a03 // vldr s4, [r2, #12]
.long 0xe593e000 // ldr lr, [r3]
.long 0xe2823010 // add r3, r2, #16
.long 0xf4e30c9f // vld1.32 {d16[]}, [r3 :32]
@@ -1576,8 +1563,7 @@ _sk_store_8888_vfp4:
.globl _sk_load_f16_vfp4
_sk_load_f16_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xe5933000 // ldr r3, [r3]
.long 0xe0833180 // add r3, r3, r0, lsl #3
@@ -1598,8 +1584,8 @@ _sk_load_f16_vfp4:
.globl _sk_store_f16_vfp4
_sk_store_f16_vfp4:
.long 0xeef00b41 // vmov.f64 d16, d1
- .long 0xf2631113 // vorr d17, d3, d3
.long 0xeef03b42 // vmov.f64 d19, d2
+ .long 0xf2631113 // vorr d17, d3, d3
.long 0xf2602110 // vorr d18, d0, d0
.long 0xf3fa00a1 // vtrn.32 d16, d17
.long 0xf3f61620 // vcvt.f16.f32 d17, q8
@@ -1616,10 +1602,9 @@ _sk_store_f16_vfp4:
.globl _sk_clamp_x_vfp4
_sk_clamp_x_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c00010 // vmov.i32 d16, #0
- .long 0xe5913000 // ldr r3, [r1]
.long 0xf3c71e1f // vmov.i8 d17, #255
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2400f80 // vmax.f32 d16, d16, d0
.long 0xe2811008 // add r1, r1, #8
.long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
@@ -1629,10 +1614,9 @@ _sk_clamp_x_vfp4:
.globl _sk_clamp_y_vfp4
_sk_clamp_y_vfp4:
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c00010 // vmov.i32 d16, #0
- .long 0xe5913000 // ldr r3, [r1]
.long 0xf3c71e1f // vmov.i8 d17, #255
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xf2400f81 // vmax.f32 d16, d16, d1
.long 0xe2811008 // add r1, r1, #8
.long 0xf4e32c9f // vld1.32 {d18[]}, [r3 :32]
@@ -1643,10 +1627,9 @@ _sk_clamp_y_vfp4:
.globl _sk_repeat_x_vfp4
_sk_repeat_x_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c02010 // vmov.i32 d18, #0
.long 0xf4e23c9f // vld1.32 {d19[]}, [r2 :32]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xe2811008 // add r1, r1, #8
.long 0xed938a00 // vldr s16, [r3]
.long 0xeec09a88 // vdiv.f32 s19, s1, s16
@@ -1668,10 +1651,9 @@ _sk_repeat_x_vfp4:
.globl _sk_repeat_y_vfp4
_sk_repeat_y_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c02010 // vmov.i32 d18, #0
.long 0xf4e23c9f // vld1.32 {d19[]}, [r2 :32]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xe2811008 // add r1, r1, #8
.long 0xed938a00 // vldr s16, [r3]
.long 0xeec19a88 // vdiv.f32 s19, s3, s16
@@ -1693,10 +1675,9 @@ _sk_repeat_y_vfp4:
.globl _sk_mirror_x_vfp4
_sk_mirror_x_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c03010 // vmov.i32 d19, #0
.long 0xf4e24c9f // vld1.32 {d20[]}, [r2 :32]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xe2811008 // add r1, r1, #8
.long 0xed938a00 // vldr s16, [r3]
.long 0xee389a08 // vadd.f32 s18, s16, s16
@@ -1722,10 +1703,9 @@ _sk_mirror_x_vfp4:
.globl _sk_mirror_y_vfp4
_sk_mirror_y_vfp4:
.long 0xed2d8b04 // vpush {d8-d9}
- .long 0xe5913000 // ldr r3, [r1]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xf2c03010 // vmov.i32 d19, #0
.long 0xf4e24c9f // vld1.32 {d20[]}, [r2 :32]
- .long 0xe591c004 // ldr ip, [r1, #4]
.long 0xe2811008 // add r1, r1, #8
.long 0xed938a00 // vldr s16, [r3]
.long 0xee389a08 // vadd.f32 s18, s16, s16
@@ -1857,8 +1837,7 @@ _sk_matrix_perspective_vfp4:
.globl _sk_linear_gradient_2stops_vfp4
_sk_linear_gradient_2stops_vfp4:
- .long 0xe5913000 // ldr r3, [r1]
- .long 0xe591c004 // ldr ip, [r1, #4]
+ .long 0xe8911008 // ldm r1, {r3, ip}
.long 0xe2811008 // add r1, r1, #8
.long 0xf4632a0d // vld1.8 {d18-d19}, [r3]!
.long 0xf4634a0f // vld1.8 {d20-d21}, [r3]
@@ -2280,14 +2259,13 @@ _sk_load_tables_hsw:
.byte 197,252,16,28,185 // vmovups (%rcx,%rdi,4),%ymm3
.byte 196,226,125,24,82,16 // vbroadcastss 0x10(%rdx),%ymm2
.byte 197,236,84,203 // vandps %ymm3,%ymm2,%ymm1
- .byte 197,252,87,192 // vxorps %ymm0,%ymm0,%ymm0
- .byte 197,124,194,192,0 // vcmpeqps %ymm0,%ymm0,%ymm8
- .byte 196,65,124,40,200 // vmovaps %ymm8,%ymm9
+ .byte 196,65,61,118,192 // vpcmpeqd %ymm8,%ymm8,%ymm8
+ .byte 196,65,53,118,201 // vpcmpeqd %ymm9,%ymm9,%ymm9
.byte 196,194,53,146,4,136 // vgatherdps %ymm9,(%r8,%ymm1,4),%ymm0
.byte 72,139,72,16 // mov 0x10(%rax),%rcx
.byte 197,245,114,211,8 // vpsrld $0x8,%ymm3,%ymm1
.byte 197,108,84,201 // vandps %ymm1,%ymm2,%ymm9
- .byte 196,65,124,40,208 // vmovaps %ymm8,%ymm10
+ .byte 196,65,45,118,210 // vpcmpeqd %ymm10,%ymm10,%ymm10
.byte 196,162,45,146,12,137 // vgatherdps %ymm10,(%rcx,%ymm9,4),%ymm1
.byte 72,139,64,24 // mov 0x18(%rax),%rax
.byte 197,181,114,211,16 // vpsrld $0x10,%ymm3,%ymm9
@@ -2709,15 +2687,13 @@ _sk_just_return_avx:
_sk_seed_shader_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 197,249,110,199 // vmovd %edi,%xmm0
- .byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
+ .byte 197,249,112,192,0 // vpshufd $0x0,%xmm0,%xmm0
.byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 196,226,125,24,74,4 // vbroadcastss 0x4(%rdx),%ymm1
.byte 197,252,88,193 // vaddps %ymm1,%ymm0,%ymm0
.byte 197,252,88,66,20 // vaddps 0x14(%rdx),%ymm0,%ymm0
- .byte 197,249,110,16 // vmovd (%rax),%xmm2
- .byte 196,227,121,4,210,0 // vpermilps $0x0,%xmm2,%xmm2
- .byte 196,227,109,24,210,1 // vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
+ .byte 196,226,125,24,16 // vbroadcastss (%rax),%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 197,236,88,201 // vaddps %ymm1,%ymm2,%ymm1
.byte 196,226,125,24,18 // vbroadcastss (%rdx),%ymm2
@@ -3061,33 +3037,27 @@ _sk_lerp_565_avx:
.byte 72,139,0 // mov (%rax),%rax
.byte 196,226,121,51,92,120,8 // vpmovzxwd 0x8(%rax,%rdi,2),%xmm3
.byte 196,98,121,51,4,120 // vpmovzxwd (%rax,%rdi,2),%xmm8
- .byte 196,99,61,24,195,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
- .byte 197,249,110,90,104 // vmovd 0x68(%rdx),%xmm3
- .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
- .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
- .byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
+ .byte 196,227,61,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ .byte 196,98,125,24,66,104 // vbroadcastss 0x68(%rdx),%ymm8
+ .byte 197,60,84,195 // vandps %ymm3,%ymm8,%ymm8
+ .byte 196,65,124,91,192 // vcvtdq2ps %ymm8,%ymm8
.byte 196,98,125,24,74,116 // vbroadcastss 0x74(%rdx),%ymm9
- .byte 197,52,89,203 // vmulps %ymm3,%ymm9,%ymm9
- .byte 197,249,110,90,108 // vmovd 0x6c(%rdx),%xmm3
- .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
- .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
- .byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
+ .byte 196,65,52,89,192 // vmulps %ymm8,%ymm9,%ymm8
+ .byte 196,98,125,24,74,108 // vbroadcastss 0x6c(%rdx),%ymm9
+ .byte 197,52,84,203 // vandps %ymm3,%ymm9,%ymm9
+ .byte 196,65,124,91,201 // vcvtdq2ps %ymm9,%ymm9
.byte 196,98,125,24,82,120 // vbroadcastss 0x78(%rdx),%ymm10
- .byte 197,44,89,211 // vmulps %ymm3,%ymm10,%ymm10
- .byte 197,249,110,90,112 // vmovd 0x70(%rdx),%xmm3
- .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
- .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- .byte 196,193,100,84,216 // vandps %ymm8,%ymm3,%ymm3
+ .byte 196,65,44,89,201 // vmulps %ymm9,%ymm10,%ymm9
+ .byte 196,98,125,24,82,112 // vbroadcastss 0x70(%rdx),%ymm10
+ .byte 197,172,84,219 // vandps %ymm3,%ymm10,%ymm3
.byte 197,252,91,219 // vcvtdq2ps %ymm3,%ymm3
- .byte 196,98,125,24,66,124 // vbroadcastss 0x7c(%rdx),%ymm8
- .byte 197,188,89,219 // vmulps %ymm3,%ymm8,%ymm3
+ .byte 196,98,125,24,82,124 // vbroadcastss 0x7c(%rdx),%ymm10
+ .byte 197,172,89,219 // vmulps %ymm3,%ymm10,%ymm3
.byte 197,252,92,196 // vsubps %ymm4,%ymm0,%ymm0
- .byte 196,193,124,89,193 // vmulps %ymm9,%ymm0,%ymm0
+ .byte 196,193,124,89,192 // vmulps %ymm8,%ymm0,%ymm0
.byte 197,252,88,196 // vaddps %ymm4,%ymm0,%ymm0
.byte 197,244,92,205 // vsubps %ymm5,%ymm1,%ymm1
- .byte 196,193,116,89,202 // vmulps %ymm10,%ymm1,%ymm1
+ .byte 196,193,116,89,201 // vmulps %ymm9,%ymm1,%ymm1
.byte 197,244,88,205 // vaddps %ymm5,%ymm1,%ymm1
.byte 197,236,92,214 // vsubps %ymm6,%ymm2,%ymm2
.byte 197,236,89,211 // vmulps %ymm3,%ymm2,%ymm2
@@ -3106,9 +3076,7 @@ _sk_load_tables_avx:
.byte 76,139,0 // mov (%rax),%r8
.byte 72,139,72,8 // mov 0x8(%rax),%rcx
.byte 196,65,124,16,20,184 // vmovups (%r8,%rdi,4),%ymm10
- .byte 197,249,110,66,16 // vmovd 0x10(%rdx),%xmm0
- .byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
- .byte 196,99,125,24,200,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
+ .byte 196,98,125,24,74,16 // vbroadcastss 0x10(%rdx),%ymm9
.byte 196,193,52,84,194 // vandps %ymm10,%ymm9,%ymm0
.byte 196,193,249,126,192 // vmovq %xmm0,%r8
.byte 69,137,193 // mov %r8d,%r9d
@@ -3245,23 +3213,17 @@ _sk_load_565_avx:
.byte 196,226,121,51,68,120,8 // vpmovzxwd 0x8(%rax,%rdi,2),%xmm0
.byte 196,226,121,51,12,120 // vpmovzxwd (%rax,%rdi,2),%xmm1
.byte 196,227,117,24,208,1 // vinsertf128 $0x1,%xmm0,%ymm1,%ymm2
- .byte 197,249,110,66,104 // vmovd 0x68(%rdx),%xmm0
- .byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
- .byte 196,227,125,24,192,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ .byte 196,226,125,24,66,104 // vbroadcastss 0x68(%rdx),%ymm0
.byte 197,252,84,194 // vandps %ymm2,%ymm0,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 196,226,125,24,74,116 // vbroadcastss 0x74(%rdx),%ymm1
.byte 197,244,89,192 // vmulps %ymm0,%ymm1,%ymm0
- .byte 197,249,110,74,108 // vmovd 0x6c(%rdx),%xmm1
- .byte 196,227,121,4,201,0 // vpermilps $0x0,%xmm1,%xmm1
- .byte 196,227,117,24,201,1 // vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ .byte 196,226,125,24,74,108 // vbroadcastss 0x6c(%rdx),%ymm1
.byte 197,244,84,202 // vandps %ymm2,%ymm1,%ymm1
.byte 197,252,91,201 // vcvtdq2ps %ymm1,%ymm1
.byte 196,226,125,24,90,120 // vbroadcastss 0x78(%rdx),%ymm3
.byte 197,228,89,201 // vmulps %ymm1,%ymm3,%ymm1
- .byte 197,249,110,90,112 // vmovd 0x70(%rdx),%xmm3
- .byte 196,227,121,4,219,0 // vpermilps $0x0,%xmm3,%xmm3
- .byte 196,227,101,24,219,1 // vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ .byte 196,226,125,24,90,112 // vbroadcastss 0x70(%rdx),%ymm3
.byte 197,228,84,210 // vandps %ymm2,%ymm3,%ymm2
.byte 197,252,91,210 // vcvtdq2ps %ymm2,%ymm2
.byte 196,226,125,24,90,124 // vbroadcastss 0x7c(%rdx),%ymm3
@@ -3303,9 +3265,7 @@ _sk_load_8888_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 72,139,0 // mov (%rax),%rax
.byte 197,252,16,28,184 // vmovups (%rax,%rdi,4),%ymm3
- .byte 197,249,110,66,16 // vmovd 0x10(%rdx),%xmm0
- .byte 196,227,121,4,192,0 // vpermilps $0x0,%xmm0,%xmm0
- .byte 196,99,125,24,216,1 // vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
+ .byte 196,98,125,24,90,16 // vbroadcastss 0x10(%rdx),%ymm11
.byte 197,164,84,195 // vandps %ymm3,%ymm11,%ymm0
.byte 197,252,91,192 // vcvtdq2ps %ymm0,%ymm0
.byte 196,98,125,24,66,12 // vbroadcastss 0xc(%rdx),%ymm8
@@ -3351,13 +3311,13 @@ _sk_store_8888_avx:
.byte 196,67,125,25,210,1 // vextractf128 $0x1,%ymm10,%xmm10
.byte 196,193,41,114,242,16 // vpslld $0x10,%xmm10,%xmm10
.byte 196,67,37,24,210,1 // vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- .byte 196,65,53,86,202 // vorpd %ymm10,%ymm9,%ymm9
.byte 197,60,89,195 // vmulps %ymm3,%ymm8,%ymm8
.byte 196,65,125,91,192 // vcvtps2dq %ymm8,%ymm8
- .byte 196,193,41,114,240,24 // vpslld $0x18,%xmm8,%xmm10
+ .byte 196,193,33,114,240,24 // vpslld $0x18,%xmm8,%xmm11
.byte 196,67,125,25,192,1 // vextractf128 $0x1,%ymm8,%xmm8
.byte 196,193,57,114,240,24 // vpslld $0x18,%xmm8,%xmm8
- .byte 196,67,45,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm10,%ymm8
+ .byte 196,67,37,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
+ .byte 196,65,45,86,192 // vorpd %ymm8,%ymm10,%ymm8
.byte 196,65,53,86,192 // vorpd %ymm8,%ymm9,%ymm8
.byte 197,125,17,4,184 // vmovupd %ymm8,(%rax,%rdi,4)
.byte 72,173 // lods %ds:(%rsi),%rax
@@ -3401,9 +3361,7 @@ _sk_load_f16_avx:
.byte 196,193,121,114,240,13 // vpslld $0xd,%xmm8,%xmm0
.byte 196,193,105,114,241,13 // vpslld $0xd,%xmm9,%xmm2
.byte 196,227,125,24,194,1 // vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- .byte 197,249,110,82,92 // vmovd 0x5c(%rdx),%xmm2
- .byte 196,227,121,4,210,0 // vpermilps $0x0,%xmm2,%xmm2
- .byte 196,99,109,24,194,1 // vinsertf128 $0x1,%xmm2,%ymm2,%ymm8
+ .byte 196,98,125,24,66,92 // vbroadcastss 0x5c(%rdx),%ymm8
.byte 197,188,89,192 // vmulps %ymm0,%ymm8,%ymm0
.byte 197,241,114,241,13 // vpslld $0xd,%xmm1,%xmm1
.byte 197,233,114,243,13 // vpslld $0xd,%xmm3,%xmm2
@@ -3424,9 +3382,7 @@ _sk_load_f16_avx:
_sk_store_f16_avx:
.byte 72,173 // lods %ds:(%rsi),%rax
.byte 72,139,0 // mov (%rax),%rax
- .byte 197,121,110,66,96 // vmovd 0x60(%rdx),%xmm8
- .byte 196,67,121,4,192,0 // vpermilps $0x0,%xmm8,%xmm8
- .byte 196,67,61,24,192,1 // vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ .byte 196,98,125,24,66,96 // vbroadcastss 0x60(%rdx),%ymm8
.byte 197,60,89,200 // vmulps %ymm0,%ymm8,%ymm9
.byte 196,67,125,25,202,1 // vextractf128 $0x1,%ymm9,%xmm10
.byte 196,193,41,114,210,13 // vpsrld $0xd,%xmm10,%xmm10
diff --git a/src/jumper/SkJumper_generated_win.S b/src/jumper/SkJumper_generated_win.S
index eb7359d17b..1cf5e7eb63 100644
--- a/src/jumper/SkJumper_generated_win.S
+++ b/src/jumper/SkJumper_generated_win.S
@@ -437,14 +437,13 @@ _sk_load_tables_hsw LABEL PROC
DB 197,252,16,28,185 ; vmovups (%rcx,%rdi,4),%ymm3
DB 196,226,125,24,82,16 ; vbroadcastss 0x10(%rdx),%ymm2
DB 197,236,84,203 ; vandps %ymm3,%ymm2,%ymm1
- DB 197,252,87,192 ; vxorps %ymm0,%ymm0,%ymm0
- DB 197,124,194,192,0 ; vcmpeqps %ymm0,%ymm0,%ymm8
- DB 196,65,124,40,200 ; vmovaps %ymm8,%ymm9
+ DB 196,65,61,118,192 ; vpcmpeqd %ymm8,%ymm8,%ymm8
+ DB 196,65,53,118,201 ; vpcmpeqd %ymm9,%ymm9,%ymm9
DB 196,194,53,146,4,136 ; vgatherdps %ymm9,(%r8,%ymm1,4),%ymm0
DB 72,139,72,16 ; mov 0x10(%rax),%rcx
DB 197,245,114,211,8 ; vpsrld $0x8,%ymm3,%ymm1
DB 197,108,84,201 ; vandps %ymm1,%ymm2,%ymm9
- DB 196,65,124,40,208 ; vmovaps %ymm8,%ymm10
+ DB 196,65,45,118,210 ; vpcmpeqd %ymm10,%ymm10,%ymm10
DB 196,162,45,146,12,137 ; vgatherdps %ymm10,(%rcx,%ymm9,4),%ymm1
DB 72,139,64,24 ; mov 0x18(%rax),%rax
DB 197,181,114,211,16 ; vpsrld $0x10,%ymm3,%ymm9
@@ -893,15 +892,13 @@ PUBLIC _sk_seed_shader_avx
_sk_seed_shader_avx LABEL PROC
DB 72,173 ; lods %ds:(%rsi),%rax
DB 197,249,110,199 ; vmovd %edi,%xmm0
- DB 196,227,121,4,192,0 ; vpermilps $0x0,%xmm0,%xmm0
+ DB 197,249,112,192,0 ; vpshufd $0x0,%xmm0,%xmm0
DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
DB 196,226,125,24,74,4 ; vbroadcastss 0x4(%rdx),%ymm1
DB 197,252,88,193 ; vaddps %ymm1,%ymm0,%ymm0
DB 197,252,88,66,20 ; vaddps 0x14(%rdx),%ymm0,%ymm0
- DB 197,249,110,16 ; vmovd (%rax),%xmm2
- DB 196,227,121,4,210,0 ; vpermilps $0x0,%xmm2,%xmm2
- DB 196,227,109,24,210,1 ; vinsertf128 $0x1,%xmm2,%ymm2,%ymm2
+ DB 196,226,125,24,16 ; vbroadcastss (%rax),%ymm2
DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
DB 197,236,88,201 ; vaddps %ymm1,%ymm2,%ymm1
DB 196,226,125,24,18 ; vbroadcastss (%rdx),%ymm2
@@ -1245,33 +1242,27 @@ _sk_lerp_565_avx LABEL PROC
DB 72,139,0 ; mov (%rax),%rax
DB 196,226,121,51,92,120,8 ; vpmovzxwd 0x8(%rax,%rdi,2),%xmm3
DB 196,98,121,51,4,120 ; vpmovzxwd (%rax,%rdi,2),%xmm8
- DB 196,99,61,24,195,1 ; vinsertf128 $0x1,%xmm3,%ymm8,%ymm8
- DB 197,249,110,90,104 ; vmovd 0x68(%rdx),%xmm3
- DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
- DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
- DB 197,252,91,219 ; vcvtdq2ps %ymm3,%ymm3
+ DB 196,227,61,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm8,%ymm3
+ DB 196,98,125,24,66,104 ; vbroadcastss 0x68(%rdx),%ymm8
+ DB 197,60,84,195 ; vandps %ymm3,%ymm8,%ymm8
+ DB 196,65,124,91,192 ; vcvtdq2ps %ymm8,%ymm8
DB 196,98,125,24,74,116 ; vbroadcastss 0x74(%rdx),%ymm9
- DB 197,52,89,203 ; vmulps %ymm3,%ymm9,%ymm9
- DB 197,249,110,90,108 ; vmovd 0x6c(%rdx),%xmm3
- DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
- DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
- DB 197,252,91,219 ; vcvtdq2ps %ymm3,%ymm3
+ DB 196,65,52,89,192 ; vmulps %ymm8,%ymm9,%ymm8
+ DB 196,98,125,24,74,108 ; vbroadcastss 0x6c(%rdx),%ymm9
+ DB 197,52,84,203 ; vandps %ymm3,%ymm9,%ymm9
+ DB 196,65,124,91,201 ; vcvtdq2ps %ymm9,%ymm9
DB 196,98,125,24,82,120 ; vbroadcastss 0x78(%rdx),%ymm10
- DB 197,44,89,211 ; vmulps %ymm3,%ymm10,%ymm10
- DB 197,249,110,90,112 ; vmovd 0x70(%rdx),%xmm3
- DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
- DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
- DB 196,193,100,84,216 ; vandps %ymm8,%ymm3,%ymm3
+ DB 196,65,44,89,201 ; vmulps %ymm9,%ymm10,%ymm9
+ DB 196,98,125,24,82,112 ; vbroadcastss 0x70(%rdx),%ymm10
+ DB 197,172,84,219 ; vandps %ymm3,%ymm10,%ymm3
DB 197,252,91,219 ; vcvtdq2ps %ymm3,%ymm3
- DB 196,98,125,24,66,124 ; vbroadcastss 0x7c(%rdx),%ymm8
- DB 197,188,89,219 ; vmulps %ymm3,%ymm8,%ymm3
+ DB 196,98,125,24,82,124 ; vbroadcastss 0x7c(%rdx),%ymm10
+ DB 197,172,89,219 ; vmulps %ymm3,%ymm10,%ymm3
DB 197,252,92,196 ; vsubps %ymm4,%ymm0,%ymm0
- DB 196,193,124,89,193 ; vmulps %ymm9,%ymm0,%ymm0
+ DB 196,193,124,89,192 ; vmulps %ymm8,%ymm0,%ymm0
DB 197,252,88,196 ; vaddps %ymm4,%ymm0,%ymm0
DB 197,244,92,205 ; vsubps %ymm5,%ymm1,%ymm1
- DB 196,193,116,89,202 ; vmulps %ymm10,%ymm1,%ymm1
+ DB 196,193,116,89,201 ; vmulps %ymm9,%ymm1,%ymm1
DB 197,244,88,205 ; vaddps %ymm5,%ymm1,%ymm1
DB 197,236,92,214 ; vsubps %ymm6,%ymm2,%ymm2
DB 197,236,89,211 ; vmulps %ymm3,%ymm2,%ymm2
@@ -1290,9 +1281,7 @@ _sk_load_tables_avx LABEL PROC
DB 76,139,0 ; mov (%rax),%r8
DB 72,139,72,8 ; mov 0x8(%rax),%rcx
DB 196,65,124,16,20,184 ; vmovups (%r8,%rdi,4),%ymm10
- DB 197,249,110,66,16 ; vmovd 0x10(%rdx),%xmm0
- DB 196,227,121,4,192,0 ; vpermilps $0x0,%xmm0,%xmm0
- DB 196,99,125,24,200,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm9
+ DB 196,98,125,24,74,16 ; vbroadcastss 0x10(%rdx),%ymm9
DB 196,193,52,84,194 ; vandps %ymm10,%ymm9,%ymm0
DB 196,193,249,126,192 ; vmovq %xmm0,%r8
DB 69,137,193 ; mov %r8d,%r9d
@@ -1429,23 +1418,17 @@ _sk_load_565_avx LABEL PROC
DB 196,226,121,51,68,120,8 ; vpmovzxwd 0x8(%rax,%rdi,2),%xmm0
DB 196,226,121,51,12,120 ; vpmovzxwd (%rax,%rdi,2),%xmm1
DB 196,227,117,24,208,1 ; vinsertf128 $0x1,%xmm0,%ymm1,%ymm2
- DB 197,249,110,66,104 ; vmovd 0x68(%rdx),%xmm0
- DB 196,227,121,4,192,0 ; vpermilps $0x0,%xmm0,%xmm0
- DB 196,227,125,24,192,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm0
+ DB 196,226,125,24,66,104 ; vbroadcastss 0x68(%rdx),%ymm0
DB 197,252,84,194 ; vandps %ymm2,%ymm0,%ymm0
DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
DB 196,226,125,24,74,116 ; vbroadcastss 0x74(%rdx),%ymm1
DB 197,244,89,192 ; vmulps %ymm0,%ymm1,%ymm0
- DB 197,249,110,74,108 ; vmovd 0x6c(%rdx),%xmm1
- DB 196,227,121,4,201,0 ; vpermilps $0x0,%xmm1,%xmm1
- DB 196,227,117,24,201,1 ; vinsertf128 $0x1,%xmm1,%ymm1,%ymm1
+ DB 196,226,125,24,74,108 ; vbroadcastss 0x6c(%rdx),%ymm1
DB 197,244,84,202 ; vandps %ymm2,%ymm1,%ymm1
DB 197,252,91,201 ; vcvtdq2ps %ymm1,%ymm1
DB 196,226,125,24,90,120 ; vbroadcastss 0x78(%rdx),%ymm3
DB 197,228,89,201 ; vmulps %ymm1,%ymm3,%ymm1
- DB 197,249,110,90,112 ; vmovd 0x70(%rdx),%xmm3
- DB 196,227,121,4,219,0 ; vpermilps $0x0,%xmm3,%xmm3
- DB 196,227,101,24,219,1 ; vinsertf128 $0x1,%xmm3,%ymm3,%ymm3
+ DB 196,226,125,24,90,112 ; vbroadcastss 0x70(%rdx),%ymm3
DB 197,228,84,210 ; vandps %ymm2,%ymm3,%ymm2
DB 197,252,91,210 ; vcvtdq2ps %ymm2,%ymm2
DB 196,226,125,24,90,124 ; vbroadcastss 0x7c(%rdx),%ymm3
@@ -1487,9 +1470,7 @@ _sk_load_8888_avx LABEL PROC
DB 72,173 ; lods %ds:(%rsi),%rax
DB 72,139,0 ; mov (%rax),%rax
DB 197,252,16,28,184 ; vmovups (%rax,%rdi,4),%ymm3
- DB 197,249,110,66,16 ; vmovd 0x10(%rdx),%xmm0
- DB 196,227,121,4,192,0 ; vpermilps $0x0,%xmm0,%xmm0
- DB 196,99,125,24,216,1 ; vinsertf128 $0x1,%xmm0,%ymm0,%ymm11
+ DB 196,98,125,24,90,16 ; vbroadcastss 0x10(%rdx),%ymm11
DB 197,164,84,195 ; vandps %ymm3,%ymm11,%ymm0
DB 197,252,91,192 ; vcvtdq2ps %ymm0,%ymm0
DB 196,98,125,24,66,12 ; vbroadcastss 0xc(%rdx),%ymm8
@@ -1535,13 +1516,13 @@ _sk_store_8888_avx LABEL PROC
DB 196,67,125,25,210,1 ; vextractf128 $0x1,%ymm10,%xmm10
DB 196,193,41,114,242,16 ; vpslld $0x10,%xmm10,%xmm10
DB 196,67,37,24,210,1 ; vinsertf128 $0x1,%xmm10,%ymm11,%ymm10
- DB 196,65,53,86,202 ; vorpd %ymm10,%ymm9,%ymm9
DB 197,60,89,195 ; vmulps %ymm3,%ymm8,%ymm8
DB 196,65,125,91,192 ; vcvtps2dq %ymm8,%ymm8
- DB 196,193,41,114,240,24 ; vpslld $0x18,%xmm8,%xmm10
+ DB 196,193,33,114,240,24 ; vpslld $0x18,%xmm8,%xmm11
DB 196,67,125,25,192,1 ; vextractf128 $0x1,%ymm8,%xmm8
DB 196,193,57,114,240,24 ; vpslld $0x18,%xmm8,%xmm8
- DB 196,67,45,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm10,%ymm8
+ DB 196,67,37,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm11,%ymm8
+ DB 196,65,45,86,192 ; vorpd %ymm8,%ymm10,%ymm8
DB 196,65,53,86,192 ; vorpd %ymm8,%ymm9,%ymm8
DB 197,125,17,4,184 ; vmovupd %ymm8,(%rax,%rdi,4)
DB 72,173 ; lods %ds:(%rsi),%rax
@@ -1585,9 +1566,7 @@ _sk_load_f16_avx LABEL PROC
DB 196,193,121,114,240,13 ; vpslld $0xd,%xmm8,%xmm0
DB 196,193,105,114,241,13 ; vpslld $0xd,%xmm9,%xmm2
DB 196,227,125,24,194,1 ; vinsertf128 $0x1,%xmm2,%ymm0,%ymm0
- DB 197,249,110,82,92 ; vmovd 0x5c(%rdx),%xmm2
- DB 196,227,121,4,210,0 ; vpermilps $0x0,%xmm2,%xmm2
- DB 196,99,109,24,194,1 ; vinsertf128 $0x1,%xmm2,%ymm2,%ymm8
+ DB 196,98,125,24,66,92 ; vbroadcastss 0x5c(%rdx),%ymm8
DB 197,188,89,192 ; vmulps %ymm0,%ymm8,%ymm0
DB 197,241,114,241,13 ; vpslld $0xd,%xmm1,%xmm1
DB 197,233,114,243,13 ; vpslld $0xd,%xmm3,%xmm2
@@ -1608,9 +1587,7 @@ PUBLIC _sk_store_f16_avx
_sk_store_f16_avx LABEL PROC
DB 72,173 ; lods %ds:(%rsi),%rax
DB 72,139,0 ; mov (%rax),%rax
- DB 197,121,110,66,96 ; vmovd 0x60(%rdx),%xmm8
- DB 196,67,121,4,192,0 ; vpermilps $0x0,%xmm8,%xmm8
- DB 196,67,61,24,192,1 ; vinsertf128 $0x1,%xmm8,%ymm8,%ymm8
+ DB 196,98,125,24,66,96 ; vbroadcastss 0x60(%rdx),%ymm8
DB 197,60,89,200 ; vmulps %ymm0,%ymm8,%ymm9
DB 196,67,125,25,202,1 ; vextractf128 $0x1,%ymm9,%xmm10
DB 196,193,41,114,210,13 ; vpsrld $0xd,%xmm10,%xmm10
diff --git a/src/jumper/build_stages.py b/src/jumper/build_stages.py
index e22ab9b136..b1914f6a29 100755
--- a/src/jumper/build_stages.py
+++ b/src/jumper/build_stages.py
@@ -9,6 +9,9 @@ import re
import subprocess
import sys
+#clang = ['clang++']
+clang = ['clang-3.9', '-x', 'c++']
+
ndk = '/Users/mtklein/brew/opt/android-ndk/'
objdump = 'gobjdump'
@@ -18,34 +21,34 @@ objdump = 'gobjdump'
cflags = '-std=c++11 -Os -fomit-frame-pointer -DJUMPER'.split()
sse2 = '-mno-red-zone -msse2 -mno-sse3 -mno-ssse3 -mno-sse4.1'.split()
-subprocess.check_call(['clang++'] + cflags + sse2 +
+subprocess.check_call(clang + cflags + sse2 +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'sse2.o'])
-subprocess.check_call(['clang++'] + cflags + sse2 + ['-DWIN'] +
+subprocess.check_call(clang + cflags + sse2 + ['-DWIN'] +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'win_sse2.o'])
sse41 = '-mno-red-zone -msse4.1'.split()
-subprocess.check_call(['clang++'] + cflags + sse41 +
+subprocess.check_call(clang + cflags + sse41 +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'sse41.o'])
-subprocess.check_call(['clang++'] + cflags + sse41 + ['-DWIN'] +
+subprocess.check_call(clang + cflags + sse41 + ['-DWIN'] +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'win_sse41.o'])
avx = '-mno-red-zone -mavx'.split()
-subprocess.check_call(['clang++'] + cflags + avx +
+subprocess.check_call(clang + cflags + avx +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'avx.o'])
-subprocess.check_call(['clang++'] + cflags + avx + ['-DWIN'] +
+subprocess.check_call(clang + cflags + avx + ['-DWIN'] +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'win_avx.o'])
hsw = '-mno-red-zone -mavx2 -mfma -mf16c'.split()
-subprocess.check_call(['clang++'] + cflags + hsw +
+subprocess.check_call(clang + cflags + hsw +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'hsw.o'])
-subprocess.check_call(['clang++'] + cflags + hsw + ['-DWIN'] +
+subprocess.check_call(clang + cflags + hsw + ['-DWIN'] +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'win_hsw.o'])
@@ -53,7 +56,7 @@ aarch64 = [
'--target=aarch64-linux-android',
'--sysroot=' + ndk + 'platforms/android-21/arch-arm64',
]
-subprocess.check_call(['clang++'] + cflags + aarch64 +
+subprocess.check_call(clang + cflags + aarch64 +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'aarch64.o'])
@@ -63,7 +66,7 @@ vfp4 = [
'-mfpu=neon-vfpv4',
'-mfloat-abi=hard',
]
-subprocess.check_call(['clang++'] + cflags + vfp4 +
+subprocess.check_call(clang + cflags + vfp4 +
['-c', 'src/jumper/SkJumper_stages.cpp'] +
['-o', 'vfp4.o'])
@@ -94,6 +97,12 @@ def parse_object_file(dot_o, directive, target=None):
print '_' + m.group(1) + label
continue
+ # ip-relative addressing usually means we're loading a constant,
+ # which we don't support.
+ if '%rip' in line:
+ print >>sys.stderr, line
+ assert '%rip' not in line
+
columns = line.split('\t')
code = columns[1]
if len(columns) >= 4:
@@ -105,12 +114,7 @@ def parse_object_file(dot_o, directive, target=None):
inst, args = columns[2].split(' ', 1)
code, inst, args = code.strip(), inst.strip(), args.strip()
- # We can't work with code that uses ip-relative addressing.
- for arg in args:
- assert 'rip' not in arg # TODO: detect on aarch64 too
-
hexed = ','.join(dehex(x) for x in code.split(' '))
-
print ' ' + directive + ' ' + hexed + ' '*(36-len(hexed)) + \
comment + inst + (' '*(14-len(inst)) + args if args else '')