diff options
-rw-r--r-- | src/splicer/SkSplicer.cpp | 68 | ||||
-rw-r--r-- | src/splicer/SkSplicer_generated_lowp.h | 535 | ||||
-rw-r--r-- | src/splicer/SkSplicer_shared.h | 5 | ||||
-rw-r--r-- | src/splicer/SkSplicer_stages_lowp.cpp | 367 | ||||
-rwxr-xr-x | src/splicer/build_stages.py | 36 |
5 files changed, 20 insertions, 991 deletions
diff --git a/src/splicer/SkSplicer.cpp b/src/splicer/SkSplicer.cpp index 0330aff558..a25a578943 100644 --- a/src/splicer/SkSplicer.cpp +++ b/src/splicer/SkSplicer.cpp @@ -16,7 +16,6 @@ #endif #include "SkSplicer_generated.h" -#include "SkSplicer_generated_lowp.h" #include "SkSplicer_shared.h" // Uncomment to dump output JIT'd pipeline. @@ -43,18 +42,11 @@ namespace { 0.0025f, 0.6975f, 0.3000f, 1/12.92f, 0.055f, // from_srgb 12.46f, 0.411192f, 0.689206f, -0.0988f, 0.0043f, // to_srgb }; - static const SkSplicer_constants_lowp kConstants_lowp = { - 0x8000, 0x8081, - }; // We do this a lot, so it's nice to infer the correct size. Works fine with arrays. template <typename T> static void splice(SkWStream* buf, const T& val) { - // This null check makes determining whether we can drop to lowp easier. - // It's always known at compile time.. - if (buf) { - buf->write(&val, sizeof(val)); - } + buf->write(&val, sizeof(val)); } #if defined(__aarch64__) @@ -241,30 +233,7 @@ namespace { } #endif - static bool splice_lowp(SkWStream* buf, SkRasterPipeline::StockStage st) { - switch (st) { - default: return false; - case SkRasterPipeline::clamp_0: break; // lowp can't go below 0. - #define CASE(st) case SkRasterPipeline::st: splice(buf, kSplice_##st##_lowp); break - CASE(clear); - CASE(plus_); - CASE(srcover); - CASE(dstover); - CASE(clamp_1); - CASE(clamp_a); - CASE(swap); - CASE(move_src_dst); - CASE(move_dst_src); - CASE(premul); - CASE(scale_u8); - CASE(load_8888); - CASE(store_8888); - #undef CASE - } - return true; - } - - static bool splice_highp(SkWStream* buf, SkRasterPipeline::StockStage st) { + static bool splice(SkWStream* buf, SkRasterPipeline::StockStage st) { switch (st) { default: return false; #define CASE(st) case SkRasterPipeline::st: splice(buf, kSplice_##st); break @@ -303,7 +272,6 @@ namespace { fBackup = SkOpts::compile_pipeline(stages, nstages); fSplicedLen = 0; fSpliced = nullptr; - fLowp = false; // If we return early anywhere in here, !fSpliced means we'll use fBackup instead. #if defined(__aarch64__) @@ -319,17 +287,6 @@ namespace { } #endif - // See if all the stages can run in lowp mode. If so, we can run at ~2x speed. - bool lowp = true; - for (int i = 0; i < nstages; i++) { - if (!splice_lowp(nullptr, stages[i].stage)) { - //SkDebugf("SkSplicer can't yet handle stage %d in lowp.\n", stages[i].stage); - lowp = false; - break; - } - } - fLowp = lowp; - SkDynamicMemoryWStream buf; // Our loop is the equivalent of this C++ code: @@ -347,18 +304,13 @@ namespace { } // Splice in the code for the Stages, generated offline into SkSplicer_generated.h. - if (lowp) { - SkAssertResult(splice_lowp(&buf, stages[i].stage)); - continue; - } - if (!splice_highp(&buf, stages[i].stage)) { + if (!splice(&buf, stages[i].stage)) { //SkDebugf("SkSplicer can't yet handle stage %d.\n", stages[i].stage); return; } } - lowp ? splice(&buf, kSplice_inc_x_lowp) - : splice(&buf, kSplice_inc_x); + splice(&buf, kSplice_inc_x); loop(&buf, loop_start); // Loop back to handle more pixels if not done. after_loop(&buf); ret(&buf); // We're done. @@ -375,8 +327,7 @@ namespace { // Spliced is stored in a std::function, so it needs to be copyable. Spliced(const Spliced& o) : fBackup (o.fBackup) , fSplicedLen(o.fSplicedLen) - , fSpliced (copy_to_executable_mem(o.fSpliced, &fSplicedLen)) - , fLowp (o.fLowp) {} + , fSpliced (copy_to_executable_mem(o.fSpliced, &fSplicedLen)) {} ~Spliced() { cleanup_executable_mem(fSpliced, fSplicedLen); @@ -384,14 +335,10 @@ namespace { // Here's where we call fSpliced if we created it, fBackup if not. void operator()(size_t x, size_t n) const { - size_t stride = fLowp ? kStride*2 - : kStride; - size_t body = n/stride*stride; // Largest multiple of stride (2, 4, 8, or 16) <= n. + size_t body = n/kStride*kStride; // Largest multiple of kStride (2, 4, 8, or 16) <= n. if (fSpliced && body) { // Can we run fSpliced for at least one stride? using Fn = void(size_t x, size_t limit, void* ctx, const void* k); - auto k = fLowp ? (const void*)&kConstants_lowp - : (const void*)&kConstants; - ((Fn*)fSpliced)(x, x+body, nullptr, k); + ((Fn*)fSpliced)(x, x+body, nullptr, &kConstants); // Fall through to fBackup for any n<stride last pixels. x += body; @@ -403,7 +350,6 @@ namespace { std::function<void(size_t, size_t)> fBackup; size_t fSplicedLen; void* fSpliced; - bool fLowp; }; } diff --git a/src/splicer/SkSplicer_generated_lowp.h b/src/splicer/SkSplicer_generated_lowp.h deleted file mode 100644 index 58e03c75ea..0000000000 --- a/src/splicer/SkSplicer_generated_lowp.h +++ /dev/null @@ -1,535 +0,0 @@ -/* - * Copyright 2017 Google Inc. - * - * Use of this source code is governed by a BSD-style license that can be - * found in the LICENSE file. - */ - -#ifndef SkSplicer_generated_lowp_DEFINED -#define SkSplicer_generated_lowp_DEFINED - -// This file is generated semi-automatically with this command: -// $ src/splicer/build_stages.py - -#if defined(__aarch64__) - -static const unsigned int kSplice_inc_x_lowp[] = { - 0x91002000, // add x0, x0, #0x8 -}; -static const unsigned int kSplice_clear_lowp[] = { - 0x6f00e400, // movi v0.2d, #0x0 - 0x6f00e401, // movi v1.2d, #0x0 - 0x6f00e402, // movi v2.2d, #0x0 - 0x6f00e403, // movi v3.2d, #0x0 -}; -static const unsigned int kSplice_plus__lowp[] = { - 0x6e640c00, // uqadd v0.8h, v0.8h, v4.8h - 0x6e650c21, // uqadd v1.8h, v1.8h, v5.8h - 0x6e660c42, // uqadd v2.8h, v2.8h, v6.8h - 0x6e670c63, // uqadd v3.8h, v3.8h, v7.8h -}; -static const unsigned int kSplice_srcover_lowp[] = { - 0x4d40c470, // ld1r {v16.8h}, [x3] - 0x6e632e10, // uqsub v16.8h, v16.8h, v3.8h - 0x6e70b491, // sqrdmulh v17.8h, v4.8h, v16.8h - 0x4e241e12, // and v18.16b, v16.16b, v4.16b - 0x6e70b4b3, // sqrdmulh v19.8h, v5.8h, v16.8h - 0x4e60ba31, // abs v17.8h, v17.8h - 0x4e251e14, // and v20.16b, v16.16b, v5.16b - 0x6f111651, // usra v17.8h, v18.8h, #15 - 0x6e70b4d2, // sqrdmulh v18.8h, v6.8h, v16.8h - 0x4e60ba73, // abs v19.8h, v19.8h - 0x6f111693, // usra v19.8h, v20.8h, #15 - 0x4e261e14, // and v20.16b, v16.16b, v6.16b - 0x4e60ba52, // abs v18.8h, v18.8h - 0x6f111692, // usra v18.8h, v20.8h, #15 - 0x6e70b4f4, // sqrdmulh v20.8h, v7.8h, v16.8h - 0x4e271e10, // and v16.16b, v16.16b, v7.16b - 0x4e60ba94, // abs v20.8h, v20.8h - 0x6f111614, // usra v20.8h, v16.8h, #15 - 0x6e600e20, // uqadd v0.8h, v17.8h, v0.8h - 0x6e610e61, // uqadd v1.8h, v19.8h, v1.8h - 0x6e620e42, // uqadd v2.8h, v18.8h, v2.8h - 0x6e630e83, // uqadd v3.8h, v20.8h, v3.8h -}; -static const unsigned int kSplice_dstover_lowp[] = { - 0x4d40c470, // ld1r {v16.8h}, [x3] - 0x6e672e10, // uqsub v16.8h, v16.8h, v7.8h - 0x6e70b411, // sqrdmulh v17.8h, v0.8h, v16.8h - 0x4e201e12, // and v18.16b, v16.16b, v0.16b - 0x6e70b433, // sqrdmulh v19.8h, v1.8h, v16.8h - 0x4e60ba31, // abs v17.8h, v17.8h - 0x4e211e14, // and v20.16b, v16.16b, v1.16b - 0x6f111651, // usra v17.8h, v18.8h, #15 - 0x6e70b452, // sqrdmulh v18.8h, v2.8h, v16.8h - 0x4e60ba73, // abs v19.8h, v19.8h - 0x6f111693, // usra v19.8h, v20.8h, #15 - 0x4e221e14, // and v20.16b, v16.16b, v2.16b - 0x4e60ba52, // abs v18.8h, v18.8h - 0x6f111692, // usra v18.8h, v20.8h, #15 - 0x6e70b474, // sqrdmulh v20.8h, v3.8h, v16.8h - 0x4e231e10, // and v16.16b, v16.16b, v3.16b - 0x4e60ba94, // abs v20.8h, v20.8h - 0x6f111614, // usra v20.8h, v16.8h, #15 - 0x6e640e24, // uqadd v4.8h, v17.8h, v4.8h - 0x6e650e65, // uqadd v5.8h, v19.8h, v5.8h - 0x6e660e46, // uqadd v6.8h, v18.8h, v6.8h - 0x6e670e87, // uqadd v7.8h, v20.8h, v7.8h -}; -static const unsigned int kSplice_clamp_1_lowp[] = { - 0x4d40c470, // ld1r {v16.8h}, [x3] - 0x6e706c00, // umin v0.8h, v0.8h, v16.8h - 0x6e706c21, // umin v1.8h, v1.8h, v16.8h - 0x6e706c42, // umin v2.8h, v2.8h, v16.8h - 0x6e706c63, // umin v3.8h, v3.8h, v16.8h -}; -static const unsigned int kSplice_clamp_a_lowp[] = { - 0x4d40c470, // ld1r {v16.8h}, [x3] - 0x6e706c63, // umin v3.8h, v3.8h, v16.8h - 0x6e636c00, // umin v0.8h, v0.8h, v3.8h - 0x6e636c21, // umin v1.8h, v1.8h, v3.8h - 0x6e636c42, // umin v2.8h, v2.8h, v3.8h -}; -static const unsigned int kSplice_swap_lowp[] = { - 0x4ea31c70, // mov v16.16b, v3.16b - 0x4ea21c51, // mov v17.16b, v2.16b - 0x4ea11c32, // mov v18.16b, v1.16b - 0x4ea01c13, // mov v19.16b, v0.16b - 0x4ea41c80, // mov v0.16b, v4.16b - 0x4ea51ca1, // mov v1.16b, v5.16b - 0x4ea61cc2, // mov v2.16b, v6.16b - 0x4ea71ce3, // mov v3.16b, v7.16b - 0x4eb31e64, // mov v4.16b, v19.16b - 0x4eb21e45, // mov v5.16b, v18.16b - 0x4eb11e26, // mov v6.16b, v17.16b - 0x4eb01e07, // mov v7.16b, v16.16b -}; -static const unsigned int kSplice_move_src_dst_lowp[] = { - 0x4ea01c04, // mov v4.16b, v0.16b - 0x4ea11c25, // mov v5.16b, v1.16b - 0x4ea21c46, // mov v6.16b, v2.16b - 0x4ea31c67, // mov v7.16b, v3.16b -}; -static const unsigned int kSplice_move_dst_src_lowp[] = { - 0x4ea41c80, // mov v0.16b, v4.16b - 0x4ea51ca1, // mov v1.16b, v5.16b - 0x4ea61cc2, // mov v2.16b, v6.16b - 0x4ea71ce3, // mov v3.16b, v7.16b -}; -static const unsigned int kSplice_premul_lowp[] = { - 0x6e63b410, // sqrdmulh v16.8h, v0.8h, v3.8h - 0x4e201c71, // and v17.16b, v3.16b, v0.16b - 0x4e60ba00, // abs v0.8h, v16.8h - 0x6e63b430, // sqrdmulh v16.8h, v1.8h, v3.8h - 0x6f111620, // usra v0.8h, v17.8h, #15 - 0x4e211c71, // and v17.16b, v3.16b, v1.16b - 0x4e60ba01, // abs v1.8h, v16.8h - 0x6e63b450, // sqrdmulh v16.8h, v2.8h, v3.8h - 0x6f111621, // usra v1.8h, v17.8h, #15 - 0x4e221c71, // and v17.16b, v3.16b, v2.16b - 0x4e60ba02, // abs v2.8h, v16.8h - 0x6f111622, // usra v2.8h, v17.8h, #15 -}; -static const unsigned int kSplice_scale_u8_lowp[] = { - 0xf9400048, // ldr x8, [x2] - 0xfc606910, // ldr d16, [x8,x0] - 0x2f0fa610, // ushll v16.8h, v16.8b, #7 - 0x6f183610, // ursra v16.8h, v16.8h, #8 - 0x6e70b411, // sqrdmulh v17.8h, v0.8h, v16.8h - 0x6e70b433, // sqrdmulh v19.8h, v1.8h, v16.8h - 0x6e70b455, // sqrdmulh v21.8h, v2.8h, v16.8h - 0x6e70b477, // sqrdmulh v23.8h, v3.8h, v16.8h - 0x4e201e12, // and v18.16b, v16.16b, v0.16b - 0x4e211e14, // and v20.16b, v16.16b, v1.16b - 0x4e221e16, // and v22.16b, v16.16b, v2.16b - 0x4e231e10, // and v16.16b, v16.16b, v3.16b - 0x4e60ba20, // abs v0.8h, v17.8h - 0x4e60ba61, // abs v1.8h, v19.8h - 0x4e60baa2, // abs v2.8h, v21.8h - 0x4e60bae3, // abs v3.8h, v23.8h - 0x6f111640, // usra v0.8h, v18.8h, #15 - 0x6f111681, // usra v1.8h, v20.8h, #15 - 0x6f1116c2, // usra v2.8h, v22.8h, #15 - 0x6f111603, // usra v3.8h, v16.8h, #15 -}; -static const unsigned int kSplice_load_8888_lowp[] = { - 0xf9400048, // ldr x8, [x2] - 0x8b000908, // add x8, x8, x0, lsl #2 - 0x0c400110, // ld4 {v16.8b-v19.8b}, [x8] - 0x2f0fa600, // ushll v0.8h, v16.8b, #7 - 0x2f0fa621, // ushll v1.8h, v17.8b, #7 - 0x2f0fa642, // ushll v2.8h, v18.8b, #7 - 0x2f0fa663, // ushll v3.8h, v19.8b, #7 - 0x6f183400, // ursra v0.8h, v0.8h, #8 - 0x6f183421, // ursra v1.8h, v1.8h, #8 - 0x6f183442, // ursra v2.8h, v2.8h, #8 - 0x6f183463, // ursra v3.8h, v3.8h, #8 -}; -static const unsigned int kSplice_store_8888_lowp[] = { - 0xf9400048, // ldr x8, [x2] - 0x2f099410, // uqshrn v16.8b, v0.8h, #7 - 0x2f099431, // uqshrn v17.8b, v1.8h, #7 - 0x2f099452, // uqshrn v18.8b, v2.8h, #7 - 0x8b000908, // add x8, x8, x0, lsl #2 - 0x2f099473, // uqshrn v19.8b, v3.8h, #7 - 0x0c000110, // st4 {v16.8b-v19.8b}, [x8] -}; - -#elif defined(__ARM_NEON__) - -static const unsigned int kSplice_inc_x_lowp[] = { - 0xe2800004, // add r0, r0, #4 -}; -static const unsigned int kSplice_clear_lowp[] = { - 0xf2800010, // vmov.i32 d0, #0 - 0xf2801010, // vmov.i32 d1, #0 - 0xf2802010, // vmov.i32 d2, #0 - 0xf2803010, // vmov.i32 d3, #0 -}; -static const unsigned int kSplice_plus__lowp[] = { - 0xf3100014, // vqadd.u16 d0, d0, d4 - 0xf3111015, // vqadd.u16 d1, d1, d5 - 0xf3122016, // vqadd.u16 d2, d2, d6 - 0xf3133017, // vqadd.u16 d3, d3, d7 -}; -static const unsigned int kSplice_srcover_lowp[] = { - 0xf4e30c5f, // vld1.16 {d16[]}, [r3 :16] - 0xf3500293, // vqsub.u16 d16, d16, d3 - 0xf3541b20, // vqrdmulh.s16 d17, d4, d16 - 0xf3552b20, // vqrdmulh.s16 d18, d5, d16 - 0xf3563b20, // vqrdmulh.s16 d19, d6, d16 - 0xf3574b20, // vqrdmulh.s16 d20, d7, d16 - 0xf2405194, // vand d21, d16, d4 - 0xf2406195, // vand d22, d16, d5 - 0xf2407196, // vand d23, d16, d6 - 0xf2400197, // vand d16, d16, d7 - 0xf3f51321, // vabs.s16 d17, d17 - 0xf3f52322, // vabs.s16 d18, d18 - 0xf3f53323, // vabs.s16 d19, d19 - 0xf3f54324, // vabs.s16 d20, d20 - 0xf3d11135, // vsra.u16 d17, d21, #15 - 0xf3d12136, // vsra.u16 d18, d22, #15 - 0xf3d13137, // vsra.u16 d19, d23, #15 - 0xf3d14130, // vsra.u16 d20, d16, #15 - 0xf3110090, // vqadd.u16 d0, d17, d0 - 0xf3121091, // vqadd.u16 d1, d18, d1 - 0xf3132092, // vqadd.u16 d2, d19, d2 - 0xf3143093, // vqadd.u16 d3, d20, d3 -}; -static const unsigned int kSplice_dstover_lowp[] = { - 0xf4e30c5f, // vld1.16 {d16[]}, [r3 :16] - 0xf3500297, // vqsub.u16 d16, d16, d7 - 0xf3501b20, // vqrdmulh.s16 d17, d0, d16 - 0xf3512b20, // vqrdmulh.s16 d18, d1, d16 - 0xf3523b20, // vqrdmulh.s16 d19, d2, d16 - 0xf3534b20, // vqrdmulh.s16 d20, d3, d16 - 0xf2405190, // vand d21, d16, d0 - 0xf2406191, // vand d22, d16, d1 - 0xf2407192, // vand d23, d16, d2 - 0xf2400193, // vand d16, d16, d3 - 0xf3f51321, // vabs.s16 d17, d17 - 0xf3f52322, // vabs.s16 d18, d18 - 0xf3f53323, // vabs.s16 d19, d19 - 0xf3f54324, // vabs.s16 d20, d20 - 0xf3d11135, // vsra.u16 d17, d21, #15 - 0xf3d12136, // vsra.u16 d18, d22, #15 - 0xf3d13137, // vsra.u16 d19, d23, #15 - 0xf3d14130, // vsra.u16 d20, d16, #15 - 0xf3114094, // vqadd.u16 d4, d17, d4 - 0xf3125095, // vqadd.u16 d5, d18, d5 - 0xf3136096, // vqadd.u16 d6, d19, d6 - 0xf3147097, // vqadd.u16 d7, d20, d7 -}; -static const unsigned int kSplice_clamp_1_lowp[] = { - 0xf4e30c5f, // vld1.16 {d16[]}, [r3 :16] - 0xf3100630, // vmin.u16 d0, d0, d16 - 0xf3111630, // vmin.u16 d1, d1, d16 - 0xf3122630, // vmin.u16 d2, d2, d16 - 0xf3133630, // vmin.u16 d3, d3, d16 -}; -static const unsigned int kSplice_clamp_a_lowp[] = { - 0xf4e30c5f, // vld1.16 {d16[]}, [r3 :16] - 0xf3133630, // vmin.u16 d3, d3, d16 - 0xf3100613, // vmin.u16 d0, d0, d3 - 0xf3111613, // vmin.u16 d1, d1, d3 - 0xf3122613, // vmin.u16 d2, d2, d3 -}; -static const unsigned int kSplice_swap_lowp[] = { - 0xeef00b43, // vmov.f64 d16, d3 - 0xeef01b42, // vmov.f64 d17, d2 - 0xeef02b41, // vmov.f64 d18, d1 - 0xeef03b40, // vmov.f64 d19, d0 - 0xeeb00b44, // vmov.f64 d0, d4 - 0xeeb01b45, // vmov.f64 d1, d5 - 0xeeb02b46, // vmov.f64 d2, d6 - 0xeeb03b47, // vmov.f64 d3, d7 - 0xeeb04b63, // vmov.f64 d4, d19 - 0xeeb05b62, // vmov.f64 d5, d18 - 0xeeb06b61, // vmov.f64 d6, d17 - 0xeeb07b60, // vmov.f64 d7, d16 -}; -static const unsigned int kSplice_move_src_dst_lowp[] = { - 0xeeb04b40, // vmov.f64 d4, d0 - 0xeeb05b41, // vmov.f64 d5, d1 - 0xeeb06b42, // vmov.f64 d6, d2 - 0xeeb07b43, // vmov.f64 d7, d3 -}; -static const unsigned int kSplice_move_dst_src_lowp[] = { - 0xeeb00b44, // vmov.f64 d0, d4 - 0xeeb01b45, // vmov.f64 d1, d5 - 0xeeb02b46, // vmov.f64 d2, d6 - 0xeeb03b47, // vmov.f64 d3, d7 -}; -static const unsigned int kSplice_premul_lowp[] = { - 0xf3500b03, // vqrdmulh.s16 d16, d0, d3 - 0xf3511b03, // vqrdmulh.s16 d17, d1, d3 - 0xf3522b03, // vqrdmulh.s16 d18, d2, d3 - 0xf2433110, // vand d19, d3, d0 - 0xf2434111, // vand d20, d3, d1 - 0xf3b50320, // vabs.s16 d0, d16 - 0xf2430112, // vand d16, d3, d2 - 0xf3b51321, // vabs.s16 d1, d17 - 0xf3b52322, // vabs.s16 d2, d18 - 0xf3910133, // vsra.u16 d0, d19, #15 - 0xf3911134, // vsra.u16 d1, d20, #15 - 0xf3912130, // vsra.u16 d2, d16, #15 -}; -static const unsigned int kSplice_scale_u8_lowp[] = { - 0xe592c000, // ldr ip, [r2] - 0xe08cc000, // add ip, ip, r0 - 0xf4ec0c8f, // vld1.32 {d16[]}, [ip] - 0xf3cf0a30, // vshll.u8 q8, d16, #7 - 0xf3d80370, // vrsra.u16 q8, q8, #8 - 0xf3502b20, // vqrdmulh.s16 d18, d0, d16 - 0xf3513b20, // vqrdmulh.s16 d19, d1, d16 - 0xf3524b20, // vqrdmulh.s16 d20, d2, d16 - 0xf3535b20, // vqrdmulh.s16 d21, d3, d16 - 0xf2406190, // vand d22, d16, d0 - 0xf3b50322, // vabs.s16 d0, d18 - 0xf2407191, // vand d23, d16, d1 - 0xf2402192, // vand d18, d16, d2 - 0xf2400193, // vand d16, d16, d3 - 0xf3b51323, // vabs.s16 d1, d19 - 0xf3b52324, // vabs.s16 d2, d20 - 0xf3b53325, // vabs.s16 d3, d21 - 0xf3910136, // vsra.u16 d0, d22, #15 - 0xf3911137, // vsra.u16 d1, d23, #15 - 0xf3912132, // vsra.u16 d2, d18, #15 - 0xf3913130, // vsra.u16 d3, d16, #15 -}; -static const unsigned int kSplice_load_8888_lowp[] = { - 0xe592c000, // ldr ip, [r2] - 0xe08cc100, // add ip, ip, r0, lsl #2 - 0xf4ec030d, // vld4.8 {d16[0],d17[0],d18[0],d19[0]}, [ip]! - 0xf4ec032d, // vld4.8 {d16[1],d17[1],d18[1],d19[1]}, [ip]! - 0xf4ec034d, // vld4.8 {d16[2],d17[2],d18[2],d19[2]}, [ip]! - 0xf4ec036d, // vld4.8 {d16[3],d17[3],d18[3],d19[3]}, [ip]! - 0xf38f0a30, // vshll.u8 q0, d16, #7 - 0xf38f2a32, // vshll.u8 q1, d18, #7 - 0xf3cf0a31, // vshll.u8 q8, d17, #7 - 0xf3cf2a33, // vshll.u8 q9, d19, #7 - 0xf3980350, // vrsra.u16 q0, q0, #8 - 0xf3d80370, // vrsra.u16 q8, q8, #8 - 0xf3d82372, // vrsra.u16 q9, q9, #8 - 0xf3982352, // vrsra.u16 q1, q1, #8 - 0xf22011b0, // vorr d1, d16, d16 - 0xf22231b2, // vorr d3, d18, d18 -}; -static const unsigned int kSplice_store_8888_lowp[] = { - 0xf2630113, // vorr d16, d3, d3 - 0xe592c000, // ldr ip, [r2] - 0xf2612111, // vorr d18, d1, d1 - 0xf3c94910, // vqshrn.u16 d20, q0, #7 - 0xe08cc100, // add ip, ip, r0, lsl #2 - 0xf3c96912, // vqshrn.u16 d22, q1, #7 - 0xf3c95932, // vqshrn.u16 d21, q9, #7 - 0xf3c97930, // vqshrn.u16 d23, q8, #7 - 0xf4cc430d, // vst4.8 {d20[0],d21[0],d22[0],d23[0]}, [ip]! - 0xf4cc432d, // vst4.8 {d20[1],d21[1],d22[1],d23[1]}, [ip]! - 0xf4cc434d, // vst4.8 {d20[2],d21[2],d22[2],d23[2]}, [ip]! - 0xf4cc436d, // vst4.8 {d20[3],d21[3],d22[3],d23[3]}, [ip]! -}; - -#else - -static const unsigned char kSplice_inc_x_lowp[] = { - 0x48,0x83,0xc7,0x10, // add $0x10,%rdi -}; -static const unsigned char kSplice_clear_lowp[] = { - 0xc5,0xfc,0x57,0xc0, // vxorps %ymm0,%ymm0,%ymm0 - 0xc5,0xf4,0x57,0xc9, // vxorps %ymm1,%ymm1,%ymm1 - 0xc5,0xec,0x57,0xd2, // vxorps %ymm2,%ymm2,%ymm2 - 0xc5,0xe4,0x57,0xdb, // vxorps %ymm3,%ymm3,%ymm3 -}; -static const unsigned char kSplice_plus__lowp[] = { - 0xc5,0xfd,0xdd,0xc4, // vpaddusw %ymm4,%ymm0,%ymm0 - 0xc5,0xf5,0xdd,0xcd, // vpaddusw %ymm5,%ymm1,%ymm1 - 0xc5,0xed,0xdd,0xd6, // vpaddusw %ymm6,%ymm2,%ymm2 - 0xc5,0xe5,0xdd,0xdf, // vpaddusw %ymm7,%ymm3,%ymm3 -}; -static const unsigned char kSplice_srcover_lowp[] = { - 0xc4,0x62,0x7d,0x79,0x01, // vpbroadcastw (%rcx),%ymm8 - 0xc5,0x3d,0xd9,0xc3, // vpsubusw %ymm3,%ymm8,%ymm8 - 0xc4,0x42,0x5d,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm4,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xc0, // vpaddusw %ymm0,%ymm9,%ymm0 - 0xc4,0x42,0x55,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm5,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xc9, // vpaddusw %ymm1,%ymm9,%ymm1 - 0xc4,0x42,0x4d,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm6,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xd2, // vpaddusw %ymm2,%ymm9,%ymm2 - 0xc4,0x42,0x45,0x0b,0xc0, // vpmulhrsw %ymm8,%ymm7,%ymm8 - 0xc4,0x42,0x7d,0x1d,0xc0, // vpabsw %ymm8,%ymm8 - 0xc5,0xbd,0xdd,0xdb, // vpaddusw %ymm3,%ymm8,%ymm3 -}; -static const unsigned char kSplice_dstover_lowp[] = { - 0xc4,0x62,0x7d,0x79,0x01, // vpbroadcastw (%rcx),%ymm8 - 0xc5,0x3d,0xd9,0xc7, // vpsubusw %ymm7,%ymm8,%ymm8 - 0xc4,0x42,0x7d,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm0,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xe4, // vpaddusw %ymm4,%ymm9,%ymm4 - 0xc4,0x42,0x75,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm1,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xed, // vpaddusw %ymm5,%ymm9,%ymm5 - 0xc4,0x42,0x6d,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm2,%ymm9 - 0xc4,0x42,0x7d,0x1d,0xc9, // vpabsw %ymm9,%ymm9 - 0xc5,0xb5,0xdd,0xf6, // vpaddusw %ymm6,%ymm9,%ymm6 - 0xc4,0x42,0x65,0x0b,0xc0, // vpmulhrsw %ymm8,%ymm3,%ymm8 - 0xc4,0x42,0x7d,0x1d,0xc0, // vpabsw %ymm8,%ymm8 - 0xc5,0xbd,0xdd,0xff, // vpaddusw %ymm7,%ymm8,%ymm7 -}; -static const unsigned char kSplice_clamp_1_lowp[] = { - 0xc4,0x62,0x7d,0x79,0x01, // vpbroadcastw (%rcx),%ymm8 - 0xc4,0xc2,0x7d,0x3a,0xc0, // vpminuw %ymm8,%ymm0,%ymm0 - 0xc4,0xc2,0x75,0x3a,0xc8, // vpminuw %ymm8,%ymm1,%ymm1 - 0xc4,0xc2,0x6d,0x3a,0xd0, // vpminuw %ymm8,%ymm2,%ymm2 - 0xc4,0xc2,0x65,0x3a,0xd8, // vpminuw %ymm8,%ymm3,%ymm3 -}; -static const unsigned char kSplice_clamp_a_lowp[] = { - 0xc4,0x62,0x7d,0x79,0x01, // vpbroadcastw (%rcx),%ymm8 - 0xc4,0xc2,0x65,0x3a,0xd8, // vpminuw %ymm8,%ymm3,%ymm3 - 0xc4,0xe2,0x7d,0x3a,0xc3, // vpminuw %ymm3,%ymm0,%ymm0 - 0xc4,0xe2,0x75,0x3a,0xcb, // vpminuw %ymm3,%ymm1,%ymm1 - 0xc4,0xe2,0x6d,0x3a,0xd3, // vpminuw %ymm3,%ymm2,%ymm2 -}; -static const unsigned char kSplice_swap_lowp[] = { - 0xc5,0x7c,0x28,0xc3, // vmovaps %ymm3,%ymm8 - 0xc5,0x7c,0x28,0xca, // vmovaps %ymm2,%ymm9 - 0xc5,0x7c,0x28,0xd1, // vmovaps %ymm1,%ymm10 - 0xc5,0x7c,0x28,0xd8, // vmovaps %ymm0,%ymm11 - 0xc5,0xfc,0x28,0xc4, // vmovaps %ymm4,%ymm0 - 0xc5,0xfc,0x28,0xcd, // vmovaps %ymm5,%ymm1 - 0xc5,0xfc,0x28,0xd6, // vmovaps %ymm6,%ymm2 - 0xc5,0xfc,0x28,0xdf, // vmovaps %ymm7,%ymm3 - 0xc5,0x7c,0x29,0xdc, // vmovaps %ymm11,%ymm4 - 0xc5,0x7c,0x29,0xd5, // vmovaps %ymm10,%ymm5 - 0xc5,0x7c,0x29,0xce, // vmovaps %ymm9,%ymm6 - 0xc5,0x7c,0x29,0xc7, // vmovaps %ymm8,%ymm7 -}; -static const unsigned char kSplice_move_src_dst_lowp[] = { - 0xc5,0xfc,0x28,0xe0, // vmovaps %ymm0,%ymm4 - 0xc5,0xfc,0x28,0xe9, // vmovaps %ymm1,%ymm5 - 0xc5,0xfc,0x28,0xf2, // vmovaps %ymm2,%ymm6 - 0xc5,0xfc,0x28,0xfb, // vmovaps %ymm3,%ymm7 -}; -static const unsigned char kSplice_move_dst_src_lowp[] = { - 0xc5,0xfc,0x28,0xc4, // vmovaps %ymm4,%ymm0 - 0xc5,0xfc,0x28,0xcd, // vmovaps %ymm5,%ymm1 - 0xc5,0xfc,0x28,0xd6, // vmovaps %ymm6,%ymm2 - 0xc5,0xfc,0x28,0xdf, // vmovaps %ymm7,%ymm3 -}; -static const unsigned char kSplice_premul_lowp[] = { - 0xc4,0xe2,0x7d,0x0b,0xc3, // vpmulhrsw %ymm3,%ymm0,%ymm0 - 0xc4,0xe2,0x7d,0x1d,0xc0, // vpabsw %ymm0,%ymm0 - 0xc4,0xe2,0x75,0x0b,0xcb, // vpmulhrsw %ymm3,%ymm1,%ymm1 - 0xc4,0xe2,0x7d,0x1d,0xc9, // vpabsw %ymm1,%ymm1 - 0xc4,0xe2,0x6d,0x0b,0xd3, // vpmulhrsw %ymm3,%ymm2,%ymm2 - 0xc4,0xe2,0x7d,0x1d,0xd2, // vpabsw %ymm2,%ymm2 -}; -static const unsigned char kSplice_scale_u8_lowp[] = { - 0x48,0x8b,0x02, // mov (%rdx),%rax - 0xc4,0x62,0x7d,0x30,0x04,0x38, // vpmovzxbw (%rax,%rdi,1),%ymm8 - 0xc4,0xc1,0x3d,0x71,0xf0,0x08, // vpsllw $0x8,%ymm8,%ymm8 - 0xc4,0x62,0x7d,0x79,0x49,0x02, // vpbroadcastw 0x2(%rcx),%ymm9 - 0xc4,0x41,0x3d,0xe4,0xc1, // vpmulhuw %ymm9,%ymm8,%ymm8 - 0xc4,0xc2,0x7d,0x0b,0xc0, // vpmulhrsw %ymm8,%ymm0,%ymm0 - 0xc4,0xe2,0x7d,0x1d,0xc0, // vpabsw %ymm0,%ymm0 - 0xc4,0xc2,0x75,0x0b,0xc8, // vpmulhrsw %ymm8,%ymm1,%ymm1 - 0xc4,0xe2,0x7d,0x1d,0xc9, // vpabsw %ymm1,%ymm1 - 0xc4,0xc2,0x6d,0x0b,0xd0, // vpmulhrsw %ymm8,%ymm2,%ymm2 - 0xc4,0xe2,0x7d,0x1d,0xd2, // vpabsw %ymm2,%ymm2 - 0xc4,0xc2,0x65,0x0b,0xd8, // vpmulhrsw %ymm8,%ymm3,%ymm3 - 0xc4,0xe2,0x7d,0x1d,0xdb, // vpabsw %ymm3,%ymm3 -}; -static const unsigned char kSplice_load_8888_lowp[] = { - 0x48,0x8b,0x02, // mov (%rdx),%rax - 0xc5,0xfa,0x6f,0x04,0xb8, // vmovdqu (%rax,%rdi,4),%xmm0 - 0xc5,0xfa,0x6f,0x4c,0xb8,0x10, // vmovdqu 0x10(%rax,%rdi,4),%xmm1 - 0xc5,0xfa,0x6f,0x54,0xb8,0x20, // vmovdqu 0x20(%rax,%rdi,4),%xmm2 - 0xc5,0xfa,0x6f,0x5c,0xb8,0x30, // vmovdqu 0x30(%rax,%rdi,4),%xmm3 - 0xc5,0x79,0x60,0xc1, // vpunpcklbw %xmm1,%xmm0,%xmm8 - 0xc5,0xf9,0x68,0xc1, // vpunpckhbw %xmm1,%xmm0,%xmm0 - 0xc5,0xe9,0x60,0xcb, // vpunpcklbw %xmm3,%xmm2,%xmm1 - 0xc5,0xe9,0x68,0xd3, // vpunpckhbw %xmm3,%xmm2,%xmm2 - 0xc5,0xb9,0x60,0xd8, // vpunpcklbw %xmm0,%xmm8,%xmm3 - 0xc5,0xb9,0x68,0xc0, // vpunpckhbw %xmm0,%xmm8,%xmm0 - 0xc5,0x71,0x60,0xc2, // vpunpcklbw %xmm2,%xmm1,%xmm8 - 0xc5,0xf1,0x68,0xca, // vpunpckhbw %xmm2,%xmm1,%xmm1 - 0xc5,0xe1,0x60,0xd0, // vpunpcklbw %xmm0,%xmm3,%xmm2 - 0xc5,0x61,0x68,0xc8, // vpunpckhbw %xmm0,%xmm3,%xmm9 - 0xc5,0xb9,0x60,0xd9, // vpunpcklbw %xmm1,%xmm8,%xmm3 - 0xc5,0x39,0x68,0xc1, // vpunpckhbw %xmm1,%xmm8,%xmm8 - 0xc5,0xe9,0x6c,0xc3, // vpunpcklqdq %xmm3,%xmm2,%xmm0 - 0xc4,0xe2,0x7d,0x30,0xc0, // vpmovzxbw %xmm0,%ymm0 - 0xc5,0xfd,0x71,0xf0,0x08, // vpsllw $0x8,%ymm0,%ymm0 - 0xc4,0x62,0x7d,0x79,0x51,0x02, // vpbroadcastw 0x2(%rcx),%ymm10 - 0xc4,0xc1,0x7d,0xe4,0xc2, // vpmulhuw %ymm10,%ymm0,%ymm0 - 0xc5,0xe9,0x6d,0xcb, // vpunpckhqdq %xmm3,%xmm2,%xmm1 - 0xc4,0xe2,0x7d,0x30,0xc9, // vpmovzxbw %xmm1,%ymm1 - 0xc5,0xf5,0x71,0xf1,0x08, // vpsllw $0x8,%ymm1,%ymm1 - 0xc4,0xc1,0x75,0xe4,0xca, // vpmulhuw %ymm10,%ymm1,%ymm1 - 0xc4,0xc1,0x31,0x6c,0xd0, // vpunpcklqdq %xmm8,%xmm9,%xmm2 - 0xc4,0xe2,0x7d,0x30,0xd2, // vpmovzxbw %xmm2,%ymm2 - 0xc5,0xed,0x71,0xf2,0x08, // vpsllw $0x8,%ymm2,%ymm2 - 0xc4,0xc1,0x6d,0xe4,0xd2, // vpmulhuw %ymm10,%ymm2,%ymm2 - 0xc4,0xc1,0x31,0x6d,0xd8, // vpunpckhqdq %xmm8,%xmm9,%xmm3 - 0xc4,0xe2,0x7d,0x30,0xdb, // vpmovzxbw %xmm3,%ymm3 - 0xc5,0xe5,0x71,0xf3,0x08, // vpsllw $0x8,%ymm3,%ymm3 - 0xc4,0xc1,0x65,0xe4,0xda, // vpmulhuw %ymm10,%ymm3,%ymm3 -}; -static const unsigned char kSplice_store_8888_lowp[] = { - 0x48,0x8b,0x02, // mov (%rdx),%rax - 0xc5,0x7d,0xdd,0xc0, // vpaddusw %ymm0,%ymm0,%ymm8 - 0xc4,0xc1,0x3d,0x71,0xd0,0x08, // vpsrlw $0x8,%ymm8,%ymm8 - 0xc4,0x43,0x7d,0x39,0xc1,0x01, // vextracti128 $0x1,%ymm8,%xmm9 - 0xc4,0x41,0x39,0x67,0xc1, // vpackuswb %xmm9,%xmm8,%xmm8 - 0xc5,0x75,0xdd,0xc9, // vpaddusw %ymm1,%ymm1,%ymm9 - 0xc4,0xc1,0x35,0x71,0xd1,0x08, // vpsrlw $0x8,%ymm9,%ymm9 - 0xc4,0x43,0x7d,0x39,0xca,0x01, // vextracti128 $0x1,%ymm9,%xmm10 - 0xc4,0x41,0x31,0x67,0xca, // vpackuswb %xmm10,%xmm9,%xmm9 - 0xc5,0x6d,0xdd,0xd2, // vpaddusw %ymm2,%ymm2,%ymm10 - 0xc4,0xc1,0x2d,0x71,0xd2,0x08, // vpsrlw $0x8,%ymm10,%ymm10 - 0xc4,0x43,0x7d,0x39,0xd3,0x01, // vextracti128 $0x1,%ymm10,%xmm11 - 0xc4,0x41,0x29,0x67,0xd3, // vpackuswb %xmm11,%xmm10,%xmm10 - 0xc5,0x65,0xdd,0xdb, // vpaddusw %ymm3,%ymm3,%ymm11 - 0xc4,0xc1,0x25,0x71,0xd3,0x08, // vpsrlw $0x8,%ymm11,%ymm11 - 0xc4,0x43,0x7d,0x39,0xdc,0x01, // vextracti128 $0x1,%ymm11,%xmm12 - 0xc4,0x41,0x21,0x67,0xdc, // vpackuswb %xmm12,%xmm11,%xmm11 - 0xc4,0x41,0x39,0x60,0xe1, // vpunpcklbw %xmm9,%xmm8,%xmm12 - 0xc4,0x41,0x39,0x68,0xc1, // vpunpckhbw %xmm9,%xmm8,%xmm8 - 0xc4,0x41,0x29,0x60,0xcb, // vpunpcklbw %xmm11,%xmm10,%xmm9 - 0xc4,0x41,0x29,0x68,0xd3, // vpunpckhbw %xmm11,%xmm10,%xmm10 - 0xc4,0x41,0x19,0x61,0xd9, // vpunpcklwd %xmm9,%xmm12,%xmm11 - 0xc5,0x7a,0x7f,0x1c,0xb8, // vmovdqu %xmm11,(%rax,%rdi,4) - 0xc4,0x41,0x19,0x69,0xc9, // vpunpckhwd %xmm9,%xmm12,%xmm9 - 0xc5,0x7a,0x7f,0x4c,0xb8,0x10, // vmovdqu %xmm9,0x10(%rax,%rdi,4) - 0xc4,0x41,0x39,0x61,0xca, // vpunpcklwd %xmm10,%xmm8,%xmm9 - 0xc5,0x7a,0x7f,0x4c,0xb8,0x20, // vmovdqu %xmm9,0x20(%rax,%rdi,4) - 0xc4,0x41,0x39,0x69,0xc2, // vpunpckhwd %xmm10,%xmm8,%xmm8 - 0xc5,0x7a,0x7f,0x44,0xb8,0x30, // vmovdqu %xmm8,0x30(%rax,%rdi,4) -}; - -#endif - -#endif//SkSplicer_generated_lowp_DEFINED diff --git a/src/splicer/SkSplicer_shared.h b/src/splicer/SkSplicer_shared.h index 0a4c31a839..6a8f14c54d 100644 --- a/src/splicer/SkSplicer_shared.h +++ b/src/splicer/SkSplicer_shared.h @@ -40,9 +40,4 @@ struct SkSplicer_constants { float _00043; // 0.0043f }; -struct SkSplicer_constants_lowp { - uint16_t _1; // 0x8000 == 32768 == 1.0 - uint16_t _0x8081; // 0x8081 == 32897, closest value to 32768 * (256/255). -}; - #endif//SkSplicer_shared_DEFINED diff --git a/src/splicer/SkSplicer_stages_lowp.cpp b/src/splicer/SkSplicer_stages_lowp.cpp deleted file mode 100644 index 244c733f09..0000000000 --- a/src/splicer/SkSplicer_stages_lowp.cpp +++ /dev/null @@ -1,367 +0,0 @@ -/* - * Copyright 2017 Google Inc. - * - * Use of this source code is governed by a BSD-style license that can be - * found in the LICENSE file. - */ - -// This file is very similar to SkSplicer_stages.cpp, and you will want to read through that file -// first before trying to understand this one. We'll note only key differences here. - -#include "SkSplicer_shared.h" -#include <string.h> - -#if !defined(__clang__) - #error This file is not like the rest of Skia. It must be compiled with clang. -#endif - -// We use a set of constants suitable for SkFixed15 math. -using K = const SkSplicer_constants_lowp; - -#if defined(__aarch64__) - #include <arm_neon.h> - - using U8 = uint8_t __attribute__((ext_vector_type(8))); - - // In this file, F is a vector of SkFixed15. - // See SkFixed15.h for notes on its various operations. - struct F { - using V = uint16_t __attribute__((ext_vector_type(8))); - - V vec; - - F(uint16x8_t v) : vec(v) {} - operator V() const { return vec; } - - F() = default; - F(uint16_t v) : vec(v) {} - - F operator+(F o) const { return vqaddq_u16(vec, o.vec); } - F operator-(F o) const { return vqsubq_u16(vec, o.vec); } - F operator*(F o) const { - return vsraq_n_u16(vabsq_s16(vqrdmulhq_s16(vec, o.vec)), - vandq_s16(vec, o.vec), 15); - } - F operator>>(int k) const { return vec >> k; } - F operator<<(int k) const { return vec << k; } - }; - static F min(F a, F b) { return vminq_u16(a,b); } - static F max(F a, F b) { return vmaxq_u16(a,b); } - - static F from_u8(U8 u8, K*) { - // u8 * (32768/255) == u8 * 128.50196... == u8*128 + u8/2 + (u8+1)>>8 - // - // Here we do (u8*128 <rounding +> u8/2), which is correct for 0 and 255, - // and never off by more than 1 anywhere. It's just 2 instructions in NEON: - auto u16 = vshll_n_u8(u8, 7); // u16 = u8*128 - u16 = vrsraq_n_u16(u16, u16, 8); // u16 += u16/256, with rounding - return u16; - }; - -#elif defined(__ARM_NEON__) - #if defined(__thumb2__) || !defined(__ARM_ARCH_7A__) || !defined(__ARM_VFPV4__) - #error On ARMv7, compile with -march=armv7-a -mfpu=neon-vfp4, without -mthumb. - #endif - #include <arm_neon.h> - - using U8 = uint8_t __attribute__((ext_vector_type(8))); // But, only low 4 lanes active. - - struct F { - using V = uint16_t __attribute__((ext_vector_type(4))); - - V vec; - - F(uint16x4_t v) : vec(v) {} - operator V() const { return vec; } - - F() = default; - F(uint16_t v) : vec(v) {} - - F operator+(F o) const { return vqadd_u16(vec, o.vec); } - F operator-(F o) const { return vqsub_u16(vec, o.vec); } - F operator*(F o) const { - return vsra_n_u16(vabs_s16(vqrdmulh_s16(vec, o.vec)), - vand_s16(vec, o.vec), 15); - } - F operator>>(int k) const { return vec >> k; } - F operator<<(int k) const { return vec << k; } - }; - static F min(F a, F b) { return vmin_u16(a,b); } - static F max(F a, F b) { return vmax_u16(a,b); } - - static F from_u8(U8 u8, K*) { - auto u16 = vshll_n_u8(u8, 7); // Identical to aarch64... - u16 = vrsraq_n_u16(u16, u16, 8); // - return vget_low_u16(u16); // ...but only the low 4 lanes are active. - } - -#else - #if !defined(__AVX2__) || !defined(__FMA__) || !defined(__F16C__) - #error On x86, compile with -mavx2 -mfma -mf16c. - #endif - #include <immintrin.h> - - using U8 = uint8_t __attribute__((ext_vector_type(16))); - - struct F { - using V = uint16_t __attribute__((ext_vector_type(16))); - - V vec; - - F(__m256 v) : vec(v) {} - operator V() const { return vec; } - - F() = default; - F(uint16_t v) : vec(v) {} - - F operator+(F o) const { return _mm256_adds_epu16(vec, o.vec); } - F operator-(F o) const { return _mm256_subs_epu16(vec, o.vec); } - F operator*(F o) const { return _mm256_abs_epi16(_mm256_mulhrs_epi16(vec, o.vec)); } - F operator>>(int k) const { return vec >> k; } - F operator<<(int k) const { return vec << k; } - }; - static F min(F a, F b) { return _mm256_min_epu16(a,b); } - static F max(F a, F b) { return _mm256_max_epu16(a,b); } - - static F from_u8(U8 u8, K* k) { - // Ideally we'd multiply by 32768/255 = 128.50196... - // We can approximate that very cheaply as 256*32897/65536 = 128.50391... - // 0 and 255 map to 0 and 32768 correctly, and the max error is 1 (on about 1/4 of values). - F u16 = _mm256_cvtepu8_epi16(u8); - return _mm256_mulhi_epu16(u16 << 8, F(k->_0x8081)); - } -#endif - -// No platform actually supports FMA for SkFixed15. -// This fma() method just makes it easier to port stages to lowp. -static F fma(F f, F m, F a) { return f*m+a; } - -template <typename T, typename P> -static T unaligned_load(const P* p) { - T v; - memcpy(&v, p, sizeof(v)); - return v; -} - -#define C extern "C" - -using Stage = void(size_t x, size_t limit, void* ctx, K* k, F,F,F,F, F,F,F,F); - -// The armv7 aapcs-vfp calling convention makes us pass F::V instead of F if we want them in -// registers. This shouldn't affect performance or how you write STAGEs in any way. -C void done(size_t, size_t, void*, K*, F::V,F::V,F::V,F::V, F::V,F::V,F::V,F::V); - -#define STAGE(name) \ - static void name##_k(size_t& x, size_t limit, void* ctx, K* k, \ - F& r, F& g, F& b, F& a, F& dr, F& dg, F& db, F& da); \ - C void name##_lowp(size_t x, size_t limit, void* ctx, K* k, \ - F::V R, F::V G, F::V B, F::V A, \ - F::V DR, F::V DG, F::V DB, F::V DA) { \ - F r = R, g = G, b = B, a = A, dr = DR, dg = DG, db = DB, da = DA; \ - name##_k(x,limit,ctx,k, r,g,b,a, dr,dg,db,da); \ - done (x,limit,ctx,k, r,g,b,a, dr,dg,db,da); \ - } \ - static void name##_k(size_t& x, size_t limit, void* ctx, K* k, \ - F& r, F& g, F& b, F& a, F& dr, F& dg, F& db, F& da) - -STAGE(inc_x) { - x += sizeof(F) / sizeof(uint16_t); -} - -STAGE(clear) { - r = g = b = a = 0; -} - -STAGE(plus_) { - r = r + dr; - g = g + dg; - b = b + db; - a = a + da; -} - -STAGE(srcover) { - auto A = F(k->_1) - a; - r = fma(dr, A, r); - g = fma(dg, A, g); - b = fma(db, A, b); - a = fma(da, A, a); -} -STAGE(dstover) { srcover_k(x,limit,ctx,k, dr,dg,db,da, r,g,b,a); } - -STAGE(clamp_1) { - r = min(r, k->_1); - g = min(g, k->_1); - b = min(b, k->_1); - a = min(a, k->_1); -} - -STAGE(clamp_a) { - a = min(a, k->_1); - r = min(r, a); - g = min(g, a); - b = min(b, a); -} - -STAGE(swap) { - auto swap = [](F& v, F& dv) { - auto tmp = v; - v = dv; - dv = tmp; - }; - swap(r, dr); - swap(g, dg); - swap(b, db); - swap(a, da); -} -STAGE(move_src_dst) { - dr = r; - dg = g; - db = b; - da = a; -} -STAGE(move_dst_src) { - r = dr; - g = dg; - b = db; - a = da; -} - -STAGE(premul) { - r = r * a; - g = g * a; - b = b * a; -} - -STAGE(scale_u8) { - auto ptr = *(const uint8_t**)ctx + x; - -#if defined(__ARM_NEON__) - // On armv7, U8 can fit 8 bytes, but we only want to load 4. - U8 scales = vdup_n_u32(unaligned_load<uint32_t>(ptr)); -#else - U8 scales = unaligned_load<U8>(ptr); -#endif - - auto c = from_u8(scales, k); - r = r * c; - g = g * c; - b = b * c; - a = a * c; -} - -STAGE(load_8888) { - auto ptr = *(const uint32_t**)ctx + x; - -#if defined(__aarch64__) - uint8x8x4_t rgba = vld4_u8((const uint8_t*)ptr); - r = from_u8(rgba.val[0], k); - g = from_u8(rgba.val[1], k); - b = from_u8(rgba.val[2], k); - a = from_u8(rgba.val[3], k); - -#elif defined(__ARM_NEON__) - // I can't get quite the code generation I want using vld4_lane_u8(), - // so we're going to drop into assembly to do the loads. :/ - - uint8x8_t R,G,B,A; - asm("vld4.8 {%1[0],%2[0],%3[0],%4[0]}, [%0]!\n" - "vld4.8 {%1[1],%2[1],%3[1],%4[1]}, [%0]!\n" - "vld4.8 {%1[2],%2[2],%3[2],%4[2]}, [%0]!\n" - "vld4.8 {%1[3],%2[3],%3[3],%4[3]}, [%0]!\n" - : "+r"(ptr), "=w"(R), "=w"(G), "=w"(B), "=w"(A)); - r = from_u8(R, k); - g = from_u8(G, k); - b = from_u8(B, k); - a = from_u8(A, k); - -#else - // TODO: shorter, more confusing, faster with 256-bit loads and shuffles - - // Load 16 interplaced pixels. - auto _0123 = _mm_loadu_si128((const __m128i*)ptr + 0), - _4567 = _mm_loadu_si128((const __m128i*)ptr + 1), - _89AB = _mm_loadu_si128((const __m128i*)ptr + 2), - _CDEF = _mm_loadu_si128((const __m128i*)ptr + 3); - - // We've got an awful lot of unpacking to do to transpose this... - auto _0415 = _mm_unpacklo_epi8(_0123, _4567), // r04 g04 b04 a04 r15 g15 b15 a15 - _2637 = _mm_unpackhi_epi8(_0123, _4567), // r26 g26 b26 a26 r37 g37 b37 a37 - _8C9D = _mm_unpacklo_epi8(_89AB, _CDEF), - _AEBF = _mm_unpackhi_epi8(_89AB, _CDEF); - - auto _0246 = _mm_unpacklo_epi8(_0415, _2637), // r0246 g0246 b0246 a0246 - _1357 = _mm_unpackhi_epi8(_0415, _2637), // r1357 g1357 b1357 a1357 - _8ACE = _mm_unpacklo_epi8(_8C9D, _AEBF), - _9BDF = _mm_unpackhi_epi8(_8C9D, _AEBF); - - auto rg_01234567 = _mm_unpacklo_epi8(_0246, _1357), // r01234567 g01234567 - ba_01234567 = _mm_unpackhi_epi8(_0246, _1357), // b01234567 a01234567 - rg_89ABCDEF = _mm_unpacklo_epi8(_8ACE, _9BDF), // r89ABCDEF g89ABCDEF - ba_89ABCDEF = _mm_unpackhi_epi8(_8ACE, _9BDF); // b89ABCDEF a89ABCDEF - - r = from_u8(_mm_unpacklo_epi64(rg_01234567, rg_89ABCDEF), k); - g = from_u8(_mm_unpackhi_epi64(rg_01234567, rg_89ABCDEF), k); - b = from_u8(_mm_unpacklo_epi64(ba_01234567, ba_89ABCDEF), k); - a = from_u8(_mm_unpackhi_epi64(ba_01234567, ba_89ABCDEF), k); -#endif -} - -STAGE(store_8888) { - auto ptr = *(uint32_t**)ctx + x; - -#if defined(__aarch64__) - auto to_u8 = [](F v) { - // The canonical math for this from SkFixed15.h is (v - (v>>8)) >> 7. - // But what's really most important is that all bytes round trip. - - // We can do this in NEON in one instruction, a saturating narrowing right shift: - return vqshrn_n_u16(v, 7); - }; - - uint8x8x4_t rgba = {{ - to_u8(r), - to_u8(g), - to_u8(b), - to_u8(a), - }}; - vst4_u8((uint8_t*)ptr, rgba); -#elif defined(__ARM_NEON__) - auto to_u8 = [](F v) { - // Same as aarch64, but first we need to pad our vectors from 8 to 16 bytes. - F whatever; - return vqshrn_n_u16(vcombine_u8(v, whatever), 7); - }; - - // As in load_8888, I can't get quite the ideal code generation using vst4_lane_u8(). - asm("vst4.8 {%1[0],%2[0],%3[0],%4[0]}, [%0]!\n" - "vst4.8 {%1[1],%2[1],%3[1],%4[1]}, [%0]!\n" - "vst4.8 {%1[2],%2[2],%3[2],%4[2]}, [%0]!\n" - "vst4.8 {%1[3],%2[3],%3[3],%4[3]}, [%0]!\n" - : "+r"(ptr) - : "w"(to_u8(r)), "w"(to_u8(g)), "w"(to_u8(b)), "w"(to_u8(a)) - : "memory"); - -#else - auto to_u8 = [](F v) { - // See the note in aarch64's to_u8(). The same roundtrip goal applies here. - // Here we take a different approach: (v saturated+ v) >> 8. - v = (v+v) >> 8; - return _mm_packus_epi16(_mm256_extracti128_si256(v, 0), - _mm256_extracti128_si256(v, 1)); - }; - - auto R = to_u8(r), - G = to_u8(g), - B = to_u8(b), - A = to_u8(a); - - auto rg_01234567 = _mm_unpacklo_epi8(R,G), // rg0 rg1 rg2 ... rg7 - rg_89ABCDEF = _mm_unpackhi_epi8(R,G), // rg8 rg9 rgA ... rgF - ba_01234567 = _mm_unpacklo_epi8(B,A), - ba_89ABCDEF = _mm_unpackhi_epi8(B,A); - _mm_storeu_si128((__m128i*)ptr + 0, _mm_unpacklo_epi16(rg_01234567, ba_01234567)); - _mm_storeu_si128((__m128i*)ptr + 1, _mm_unpackhi_epi16(rg_01234567, ba_01234567)); - _mm_storeu_si128((__m128i*)ptr + 2, _mm_unpacklo_epi16(rg_89ABCDEF, ba_89ABCDEF)); - _mm_storeu_si128((__m128i*)ptr + 3, _mm_unpackhi_epi16(rg_89ABCDEF, ba_89ABCDEF)); -#endif -} diff --git a/src/splicer/build_stages.py b/src/splicer/build_stages.py index 9a52bec875..e301494f2f 100755 --- a/src/splicer/build_stages.py +++ b/src/splicer/build_stages.py @@ -21,9 +21,6 @@ hsw = '-mavx2 -mfma -mf16c'.split() subprocess.check_call(['clang++'] + cflags + hsw + ['-c', 'src/splicer/SkSplicer_stages.cpp'] + ['-o', 'hsw.o']) -subprocess.check_call(['clang++'] + cflags + hsw + - ['-c', 'src/splicer/SkSplicer_stages_lowp.cpp'] + - ['-o', 'hsw_lowp.o']) aarch64 = [ '--target=aarch64-linux-android', @@ -32,9 +29,6 @@ aarch64 = [ subprocess.check_call(['clang++'] + cflags + aarch64 + ['-c', 'src/splicer/SkSplicer_stages.cpp'] + ['-o', 'aarch64.o']) -subprocess.check_call(['clang++'] + cflags + aarch64 + - ['-c', 'src/splicer/SkSplicer_stages_lowp.cpp'] + - ['-o', 'aarch64_lowp.o']) armv7 = [ '--target=armv7a-linux-android', @@ -45,9 +39,6 @@ armv7 = [ subprocess.check_call(['clang++'] + cflags + armv7 + ['-c', 'src/splicer/SkSplicer_stages.cpp'] + ['-o', 'armv7.o']) -subprocess.check_call(['clang++'] + cflags + armv7 + - ['-c', 'src/splicer/SkSplicer_stages_lowp.cpp'] + - ['-o', 'armv7_lowp.o']) def parse_object_file(dst, dot_o, array_type, done, target=None): cmd = [ objdump, '-d', dot_o] @@ -85,28 +76,27 @@ def parse_object_file(dst, dot_o, array_type, done, target=None): print >>dst,' ' + hexed + ' '*(44-len(hexed)) + \ '// ' + inst + ' '*(14-len(inst)) + args -for suffix in ['', '_lowp']: - with open('src/splicer/SkSplicer_generated%s.h' % suffix, 'w') as f: - print >>f,'''/* +with open('src/splicer/SkSplicer_generated.h', 'w') as f: + print >>f,'''/* * Copyright 2017 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ -#ifndef SkSplicer_generated%s_DEFINED -#define SkSplicer_generated%s_DEFINED +#ifndef SkSplicer_generated_DEFINED +#define SkSplicer_generated_DEFINED // This file is generated semi-automatically with this command: // $ src/splicer/build_stages.py #if defined(__aarch64__) -''' % (suffix, suffix) - parse_object_file(f, 'aarch64%s.o' % suffix, 'unsigned int', '14000000') - print >>f,'\n#elif defined(__ARM_NEON__)\n' - parse_object_file(f, 'armv7%s.o' % suffix, 'unsigned int', 'eafffffe', - target='elf32-littlearm') - print >>f,'\n#else\n' - parse_object_file(f, 'hsw%s.o' % suffix, 'unsigned char', 'e9 00 00 00 00') - print >>f,'\n#endif\n' - print >>f,'#endif//SkSplicer_generated%s_DEFINED' % suffix +''' + parse_object_file(f, 'aarch64.o', 'unsigned int', '14000000') + print >>f,'\n#elif defined(__ARM_NEON__)\n' + parse_object_file(f, 'armv7.o', 'unsigned int', 'eafffffe', + target='elf32-littlearm') + print >>f,'\n#else\n' + parse_object_file(f, 'hsw.o', 'unsigned char', 'e9 00 00 00 00') + print >>f,'\n#endif\n' + print >>f,'#endif//SkSplicer_generated_DEFINED' |