From 4e49fd5eb97840d66f6d16208d862200465b7397 Mon Sep 17 00:00:00 2001 From: Benoit Steiner Date: Fri, 4 Mar 2016 14:49:18 -0800 Subject: MSVC uses __uint128 while other compilers use __uint128_t to encode 128bit unsigned integers. Make the cxx11_tensor_uint128.cpp test work in both cases. --- unsupported/test/cxx11_tensor_uint128.cpp | 39 ++++++++++++++++++------------- 1 file changed, 23 insertions(+), 16 deletions(-) (limited to 'unsupported/test/cxx11_tensor_uint128.cpp') diff --git a/unsupported/test/cxx11_tensor_uint128.cpp b/unsupported/test/cxx11_tensor_uint128.cpp index 424c70197..c6766c6c6 100644 --- a/unsupported/test/cxx11_tensor_uint128.cpp +++ b/unsupported/test/cxx11_tensor_uint128.cpp @@ -11,10 +11,17 @@ #include +#if EIGEN_COMP_MSVC +typedef __uint128 uint128_t; +#else +typedef __uint128_t uint128_t; +#endif + + using Eigen::internal::TensorUInt128; using Eigen::internal::static_val; -void VERIFY_EQUAL(TensorUInt128 actual, __uint128_t expected) { +void VERIFY_EQUAL(TensorUInt128 actual, uint128_t expected) { bool matchl = actual.lower() == static_cast(expected); bool matchh = actual.upper() == static_cast(expected >> 64); if (!matchl || !matchh) { @@ -32,13 +39,13 @@ void test_add() { for (uint64_t i1 = 0; i1 < 100; ++i1) { for (uint64_t i2 = 1; i2 < 100 * incr; i2 += incr) { TensorUInt128 i(i1, i2); - __uint128_t a = (static_cast<__uint128_t>(i1) << 64) + static_cast<__uint128_t>(i2); + uint128_t a = (static_cast(i1) << 64) + static_cast(i2); for (uint64_t j1 = 0; j1 < 100; ++j1) { for (uint64_t j2 = 1; j2 < 100 * incr; j2 += incr) { TensorUInt128 j(j1, j2); - __uint128_t b = (static_cast<__uint128_t>(j1) << 64) + static_cast<__uint128_t>(j2); + uint128_t b = (static_cast(j1) << 64) + static_cast(j2); TensorUInt128 actual = i + j; - __uint128_t expected = a + b; + uint128_t expected = a + b; VERIFY_EQUAL(actual, expected); } } @@ -51,13 +58,13 @@ void test_sub() { for (uint64_t i1 = 0; i1 < 100; ++i1) { for (uint64_t i2 = 1; i2 < 100 * incr; i2 += incr) { TensorUInt128 i(i1, i2); - __uint128_t a = (static_cast<__uint128_t>(i1) << 64) + static_cast<__uint128_t>(i2); + uint128_t a = (static_cast(i1) << 64) + static_cast(i2); for (uint64_t j1 = 0; j1 < 100; ++j1) { for (uint64_t j2 = 1; j2 < 100 * incr; j2 += incr) { TensorUInt128 j(j1, j2); - __uint128_t b = (static_cast<__uint128_t>(j1) << 64) + static_cast<__uint128_t>(j2); + uint128_t b = (static_cast(j1) << 64) + static_cast(j2); TensorUInt128 actual = i - j; - __uint128_t expected = a - b; + uint128_t expected = a - b; VERIFY_EQUAL(actual, expected); } } @@ -70,13 +77,13 @@ void test_mul() { for (uint64_t i1 = 0; i1 < 100; ++i1) { for (uint64_t i2 = 1; i2 < 100 * incr; i2 += incr) { TensorUInt128 i(i1, i2); - __uint128_t a = (static_cast<__uint128_t>(i1) << 64) + static_cast<__uint128_t>(i2); + uint128_t a = (static_cast(i1) << 64) + static_cast(i2); for (uint64_t j1 = 0; j1 < 100; ++j1) { for (uint64_t j2 = 1; j2 < 100 * incr; j2 += incr) { TensorUInt128 j(j1, j2); - __uint128_t b = (static_cast<__uint128_t>(j1) << 64) + static_cast<__uint128_t>(j2); + uint128_t b = (static_cast(j1) << 64) + static_cast(j2); TensorUInt128 actual = i * j; - __uint128_t expected = a * b; + uint128_t expected = a * b; VERIFY_EQUAL(actual, expected); } } @@ -89,13 +96,13 @@ void test_div() { for (uint64_t i1 = 0; i1 < 100; ++i1) { for (uint64_t i2 = 1; i2 < 100 * incr; i2 += incr) { TensorUInt128 i(i1, i2); - __uint128_t a = (static_cast<__uint128_t>(i1) << 64) + static_cast<__uint128_t>(i2); + uint128_t a = (static_cast(i1) << 64) + static_cast(i2); for (uint64_t j1 = 0; j1 < 100; ++j1) { for (uint64_t j2 = 1; j2 < 100 * incr; j2 += incr) { TensorUInt128 j(j1, j2); - __uint128_t b = (static_cast<__uint128_t>(j1) << 64) + static_cast<__uint128_t>(j2); + uint128_t b = (static_cast(j1) << 64) + static_cast(j2); TensorUInt128 actual = i / j; - __uint128_t expected = a / b; + uint128_t expected = a / b; VERIFY_EQUAL(actual, expected); } } @@ -107,10 +114,10 @@ void test_misc1() { uint64_t incr = internal::random(1, 9999999999); for (uint64_t i2 = 1; i2 < 100 * incr; i2 += incr) { TensorUInt128, uint64_t> i(0, i2); - __uint128_t a = static_cast<__uint128_t>(i2); + uint128_t a = static_cast(i2); for (uint64_t j2 = 1; j2 < 100 * incr; j2 += incr) { TensorUInt128, uint64_t> j(0, j2); - __uint128_t b = static_cast<__uint128_t>(j2); + uint128_t b = static_cast(j2); uint64_t actual = (i * j).upper(); uint64_t expected = (a * b) >> 64; VERIFY_IS_EQUAL(actual, expected); @@ -122,7 +129,7 @@ void test_misc2() { int64_t incr = internal::random(1, 100); for (int64_t log_div = 0; log_div < 63; ++log_div) { for (int64_t divider = 1; divider <= 1000000 * incr; divider += incr) { - uint64_t expected = (static_cast<__uint128_t>(1) << (64+log_div)) / static_cast<__uint128_t>(divider) - (static_cast<__uint128_t>(1) << 64) + 1; + uint64_t expected = (static_cast(1) << (64+log_div)) / static_cast(divider) - (static_cast(1) << 64) + 1; uint64_t shift = 1ULL << log_div; TensorUInt128 result = (TensorUInt128 >(shift, 0) / TensorUInt128, uint64_t>(divider) - TensorUInt128, static_val<0> >(1, 0) + TensorUInt128, static_val<1> >(1)); -- cgit v1.2.3