diff options
author | 2015-11-11 11:39:09 -0800 | |
---|---|---|
committer | 2015-11-11 11:39:09 -0800 | |
commit | 084db25d47dbad3ffbd7d15c04b63d344b351f90 (patch) | |
tree | 1152701d1608211322670f1246c7228978d8e32e /src/opts/SkNx_avx.h | |
parent | 9be5ff6f9871ef22740094e7c25dd67329a73d20 (diff) |
float xfermodes (burn, dodge, softlight) in Sk8f, possibly using AVX.
Xfermode_ColorDodge_aa 10.3ms -> 7.85ms 0.76x
Xfermode_SoftLight_aa 13.8ms -> 10.2ms 0.74x
Xfermode_ColorBurn_aa 10.7ms -> 7.82ms 0.73x
Xfermode_SoftLight 33.6ms -> 23.2ms 0.69x
Xfermode_ColorDodge 25ms -> 16.5ms 0.66x
Xfermode_ColorBurn 26.1ms -> 16.6ms 0.63x
Ought to be no pixel diffs:
https://gold.skia.org/search2?issue=1432903002&unt=true&query=source_type%3Dgm&master=false
Incidental stuff:
I made the SkNx(T) constructors implicit to make writing math expressions simpler.
This allows us to write expressions like
Sk4f v;
...
v = v*4;
rather than
Sk4f v;
...
v = v * Sk4f(4);
As written it only works when the constant is on the right-hand side,
so expressions like `(Sk4f(1) - da)` have to stay for now. I plan on
following up with a CL that lets those become `(1 - da)` too.
BUG=skia:4117
CQ_EXTRA_TRYBOTS=client.skia:Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Release-SKNX_NO_SIMD-Trybot
Review URL: https://codereview.chromium.org/1432903002
Diffstat (limited to 'src/opts/SkNx_avx.h')
-rw-r--r-- | src/opts/SkNx_avx.h | 90 |
1 files changed, 90 insertions, 0 deletions
diff --git a/src/opts/SkNx_avx.h b/src/opts/SkNx_avx.h new file mode 100644 index 0000000000..86caac20b6 --- /dev/null +++ b/src/opts/SkNx_avx.h @@ -0,0 +1,90 @@ +/* + * Copyright 2015 Google Inc. + * + * Use of this source code is governed by a BSD-style license that can be + * found in the LICENSE file. + */ + +#ifndef SkNx_avx_DEFINED +#define SkNx_avx_DEFINED + +// This file may assume <= AVX, but must check SK_CPU_SSE_LEVEL for anything more recent. + +// All the SSE specializations are still good ideas. We'll just add Sk8f. +#include "SkNx_sse.h" + +namespace { // See SkNx.h + +template <> +class SkNf<8> { +public: + SkNf(const __m256& vec) : fVec(vec) {} + + SkNf() {} + SkNf(float val) : fVec(_mm256_set1_ps(val)) {} + static SkNf Load(const float vals[8]) { return _mm256_loadu_ps(vals); } + + static SkNf FromBytes(const uint8_t bytes[8]) { + __m128i fix8 = _mm_loadl_epi64((const __m128i*)bytes), + fix16 = _mm_unpacklo_epi8 (fix8 , _mm_setzero_si128()), + lo32 = _mm_unpacklo_epi16(fix16, _mm_setzero_si128()), + hi32 = _mm_unpackhi_epi16(fix16, _mm_setzero_si128()); + __m256i fix32 = _mm256_insertf128_si256(_mm256_castsi128_si256(lo32), hi32, 1); + return _mm256_cvtepi32_ps(fix32); + } + + SkNf(float a, float b, float c, float d, + float e, float f, float g, float h) : fVec(_mm256_setr_ps(a,b,c,d,e,f,g,h)) {} + + void store(float vals[8]) const { _mm256_storeu_ps(vals, fVec); } + void toBytes(uint8_t bytes[8]) const { + __m256i fix32 = _mm256_cvttps_epi32(fVec); + __m128i lo32 = _mm256_extractf128_si256(fix32, 0), + hi32 = _mm256_extractf128_si256(fix32, 1), + fix16 = _mm_packus_epi32(lo32, hi32), + fix8 = _mm_packus_epi16(fix16, fix16); + _mm_storel_epi64((__m128i*)bytes, fix8); + } + + SkNf operator + (const SkNf& o) const { return _mm256_add_ps(fVec, o.fVec); } + SkNf operator - (const SkNf& o) const { return _mm256_sub_ps(fVec, o.fVec); } + SkNf operator * (const SkNf& o) const { return _mm256_mul_ps(fVec, o.fVec); } + SkNf operator / (const SkNf& o) const { return _mm256_div_ps(fVec, o.fVec); } + + SkNf operator == (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_EQ_OQ); } + SkNf operator != (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_NEQ_OQ); } + SkNf operator < (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_LT_OQ); } + SkNf operator > (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_GT_OQ); } + SkNf operator <= (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_LE_OQ); } + SkNf operator >= (const SkNf& o) const { return _mm256_cmp_ps(fVec, o.fVec, _CMP_GE_OQ); } + + static SkNf Min(const SkNf& l, const SkNf& r) { return _mm256_min_ps(l.fVec, r.fVec); } + static SkNf Max(const SkNf& l, const SkNf& r) { return _mm256_max_ps(l.fVec, r.fVec); } + + SkNf sqrt() const { return _mm256_sqrt_ps (fVec); } + SkNf rsqrt0() const { return _mm256_rsqrt_ps(fVec); } + SkNf rsqrt1() const { return this->rsqrt0(); } + SkNf rsqrt2() const { return this->rsqrt1(); } + + SkNf invert() const { return SkNf(1) / *this; } + SkNf approxInvert() const { return _mm256_rcp_ps(fVec); } + + template <int k> float kth() const { + SkASSERT(0 <= k && k < 8); + union { __m256 v; float fs[8]; } pun = {fVec}; + return pun.fs[k&7]; + } + + bool allTrue() const { return 0xff == _mm256_movemask_ps(fVec); } + bool anyTrue() const { return 0x00 != _mm256_movemask_ps(fVec); } + + SkNf thenElse(const SkNf& t, const SkNf& e) const { + return _mm256_blendv_ps(e.fVec, t.fVec, fVec); + } + + __m256 fVec; +}; + +} // namespace + +#endif//SkNx_avx_DEFINED |