// Copyright 2016 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include #include #include #include #include #include #include #include #include #include #include #include "src/base/bits.h" #include "src/base/logging.h" #include "src/base/macros.h" #include "src/base/memory.h" #include "src/base/overflowing-math.h" #include "src/base/safe_conversions.h" #include "src/base/utils/random-number-generator.h" #include "src/base/vector.h" #include "src/codegen/cpu-features.h" #include "src/codegen/machine-type.h" #include "src/common/globals.h" #include "src/flags/flags.h" #include "src/utils/utils.h" #include "src/wasm/compilation-environment.h" #include "src/wasm/value-type.h" #include "src/wasm/wasm-constants.h" #include "src/wasm/wasm-opcodes.h" #include "test/cctest/cctest.h" #include "test/cctest/wasm/wasm-run-utils.h" #include "test/cctest/wasm/wasm-simd-utils.h" #include "test/common/flag-utils.h" #include "test/common/value-helper.h" #include "test/common/wasm/flag-utils.h" #include "test/common/wasm/wasm-macro-gen.h" namespace v8 { namespace internal { namespace wasm { namespace test_run_wasm_simd { namespace { using Shuffle = std::array; // For signed integral types, use base::AddWithWraparound. template ::value>::type> T Add(T a, T b) { return a + b; } // For signed integral types, use base::SubWithWraparound. template ::value>::type> T Sub(T a, T b) { return a - b; } // For signed integral types, use base::MulWithWraparound. template ::value>::type> T Mul(T a, T b) { return a * b; } template T UnsignedMinimum(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) <= static_cast(b) ? a : b; } template T UnsignedMaximum(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >= static_cast(b) ? a : b; } template U Equal(T a, T b) { return a == b ? -1 : 0; } template <> int32_t Equal(float a, float b) { return a == b ? -1 : 0; } template <> int64_t Equal(double a, double b) { return a == b ? -1 : 0; } template U NotEqual(T a, T b) { return a != b ? -1 : 0; } template <> int32_t NotEqual(float a, float b) { return a != b ? -1 : 0; } template <> int64_t NotEqual(double a, double b) { return a != b ? -1 : 0; } template U Less(T a, T b) { return a < b ? -1 : 0; } template <> int32_t Less(float a, float b) { return a < b ? -1 : 0; } template <> int64_t Less(double a, double b) { return a < b ? -1 : 0; } template U LessEqual(T a, T b) { return a <= b ? -1 : 0; } template <> int32_t LessEqual(float a, float b) { return a <= b ? -1 : 0; } template <> int64_t LessEqual(double a, double b) { return a <= b ? -1 : 0; } template U Greater(T a, T b) { return a > b ? -1 : 0; } template <> int32_t Greater(float a, float b) { return a > b ? -1 : 0; } template <> int64_t Greater(double a, double b) { return a > b ? -1 : 0; } template U GreaterEqual(T a, T b) { return a >= b ? -1 : 0; } template <> int32_t GreaterEqual(float a, float b) { return a >= b ? -1 : 0; } template <> int64_t GreaterEqual(double a, double b) { return a >= b ? -1 : 0; } template T UnsignedLess(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) < static_cast(b) ? -1 : 0; } template T UnsignedLessEqual(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) <= static_cast(b) ? -1 : 0; } template T UnsignedGreater(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) > static_cast(b) ? -1 : 0; } template T UnsignedGreaterEqual(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >= static_cast(b) ? -1 : 0; } template T LogicalShiftLeft(T a, int shift) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) << (shift % (sizeof(T) * 8)); } template T LogicalShiftRight(T a, int shift) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >> (shift % (sizeof(T) * 8)); } // Define our own ArithmeticShiftRight instead of using the one from utils.h // because the shift amount needs to be taken modulo lane width. template T ArithmeticShiftRight(T a, int shift) { return a >> (shift % (sizeof(T) * 8)); } template T Abs(T a) { return std::abs(a); } } // namespace #define WASM_SIMD_CHECK_LANE_S(TYPE, value, LANE_TYPE, lane_value, lane_index) \ WASM_IF(WASM_##LANE_TYPE##_NE(WASM_LOCAL_GET(lane_value), \ WASM_SIMD_##TYPE##_EXTRACT_LANE( \ lane_index, WASM_LOCAL_GET(value))), \ WASM_RETURN(WASM_ZERO)) // Unsigned Extracts are only available for I8x16, I16x8 types #define WASM_SIMD_CHECK_LANE_U(TYPE, value, LANE_TYPE, lane_value, lane_index) \ WASM_IF(WASM_##LANE_TYPE##_NE(WASM_LOCAL_GET(lane_value), \ WASM_SIMD_##TYPE##_EXTRACT_LANE_U( \ lane_index, WASM_LOCAL_GET(value))), \ WASM_RETURN(WASM_ZERO)) WASM_EXEC_TEST(S128Globals) { WasmRunner r(execution_tier); // Set up a global to hold input and output vectors. int32_t* g0 = r.builder().AddGlobal(kWasmS128); int32_t* g1 = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(1, WASM_GLOBAL_GET(0)), WASM_ONE}); FOR_INT32_INPUTS(x) { for (int i = 0; i < 4; i++) { LANE(g0, i) = x; } r.Call(); int32_t expected = x; for (int i = 0; i < 4; i++) { int32_t actual = LANE(g1, i); CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(F32x4Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. float* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_FLOAT32_INPUTS(x) { r.Call(x); float expected = x; for (int i = 0; i < 4; i++) { float actual = LANE(g, i); if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } } WASM_EXEC_TEST(F32x4ReplaceLane) { WasmRunner r(execution_tier); // Set up a global to hold input/output vector. float* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its (FP) index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_SPLAT(WASM_F32(3.14159f))), WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_F32(0.0f))), WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_F32(1.0f))), WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 2, WASM_LOCAL_GET(temp1), WASM_F32(2.0f))), WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_REPLACE_LANE( 3, WASM_LOCAL_GET(temp1), WASM_F32(3.0f))), WASM_ONE}); r.Call(); for (int i = 0; i < 4; i++) { CHECK_EQ(static_cast(i), LANE(g, i)); } } // Tests both signed and unsigned conversion. WASM_EXEC_TEST(F32x4ConvertI32x4) { WasmRunner r(execution_tier); // Create two output vectors to hold signed and unsigned results. float* g0 = r.builder().AddGlobal(kWasmS128); float* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprF32x4SConvertI32x4, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_UNOP(kExprF32x4UConvertI32x4, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT32_INPUTS(x) { r.Call(x); float expected_signed = static_cast(x); float expected_unsigned = static_cast(static_cast(x)); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_unsigned, LANE(g1, i)); } } } template void RunF128CompareOpConstImmTest( TestExecutionTier execution_tier, WasmOpcode cmp_opcode, WasmOpcode splat_opcode, ScalarType (*expected_op)(FloatType, FloatType)) { for (FloatType x : compiler::ValueHelper::GetVector()) { if (!PlatformCanRepresent(x)) continue; WasmRunner r(execution_tier); // Set up globals to hold mask output for left and right cases ScalarType* g1 = r.builder().template AddGlobal(kWasmS128); ScalarType* g2 = r.builder().template AddGlobal(kWasmS128); // Build fn to splat test values, perform compare op on both sides, and // write the result. uint8_t value = 0; uint8_t temp = r.AllocateLocal(kWasmS128); uint8_t const_buffer[kSimd128Size]; for (size_t i = 0; i < kSimd128Size / sizeof(FloatType); i++) { WriteLittleEndianValue( base::bit_cast(&const_buffer[0]) + i, x); } r.Build( {WASM_LOCAL_SET(temp, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(cmp_opcode, WASM_SIMD_CONSTANT(const_buffer), WASM_LOCAL_GET(temp))), WASM_GLOBAL_SET(1, WASM_SIMD_BINOP(cmp_opcode, WASM_LOCAL_GET(temp), WASM_SIMD_CONSTANT(const_buffer))), WASM_ONE}); for (FloatType y : compiler::ValueHelper::GetVector()) { if (!PlatformCanRepresent(y)) continue; FloatType diff = x - y; // Model comparison as subtraction. if (!PlatformCanRepresent(diff)) continue; r.Call(y); ScalarType expected1 = expected_op(x, y); ScalarType expected2 = expected_op(y, x); for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { CHECK_EQ(expected1, LANE(g1, i)); CHECK_EQ(expected2, LANE(g2, i)); } } } } WASM_EXEC_TEST(F32x4Abs) { RunF32x4UnOpTest(execution_tier, kExprF32x4Abs, std::abs); } WASM_EXEC_TEST(F32x4Neg) { RunF32x4UnOpTest(execution_tier, kExprF32x4Neg, Negate); } WASM_EXEC_TEST(F32x4Sqrt) { RunF32x4UnOpTest(execution_tier, kExprF32x4Sqrt, std::sqrt); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_F32x8Abs) { RunF32x8UnOpRevecTest(kExprF32x4Abs, std::abs); } TEST(RunWasmTurbofan_F32x8Neg) { RunF32x8UnOpRevecTest(kExprF32x4Neg, Negate); } TEST(RunWasmTurbofan_F32x8Sqrt) { RunF32x8UnOpRevecTest(kExprF32x4Sqrt, std::sqrt); } #endif WASM_EXEC_TEST(F32x4Ceil) { RunF32x4UnOpTest(execution_tier, kExprF32x4Ceil, ceilf, true); } WASM_EXEC_TEST(F32x4Floor) { RunF32x4UnOpTest(execution_tier, kExprF32x4Floor, floorf, true); } WASM_EXEC_TEST(F32x4Trunc) { RunF32x4UnOpTest(execution_tier, kExprF32x4Trunc, truncf, true); } WASM_EXEC_TEST(F32x4NearestInt) { RunF32x4UnOpTest(execution_tier, kExprF32x4NearestInt, nearbyintf, true); } WASM_EXEC_TEST(F32x4Add) { RunF32x4BinOpTest(execution_tier, kExprF32x4Add, Add); } WASM_EXEC_TEST(F32x4Sub) { RunF32x4BinOpTest(execution_tier, kExprF32x4Sub, Sub); } WASM_EXEC_TEST(F32x4Mul) { RunF32x4BinOpTest(execution_tier, kExprF32x4Mul, Mul); } WASM_EXEC_TEST(F32x4Div) { RunF32x4BinOpTest(execution_tier, kExprF32x4Div, base::Divide); } WASM_EXEC_TEST(F32x4Min) { RunF32x4BinOpTest(execution_tier, kExprF32x4Min, JSMin); } WASM_EXEC_TEST(F32x4Max) { RunF32x4BinOpTest(execution_tier, kExprF32x4Max, JSMax); } WASM_EXEC_TEST(F32x4Pmin) { RunF32x4BinOpTest(execution_tier, kExprF32x4Pmin, Minimum); } WASM_EXEC_TEST(F32x4Pmax) { RunF32x4BinOpTest(execution_tier, kExprF32x4Pmax, Maximum); } WASM_EXEC_TEST(F32x4Eq) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Eq, Equal); } WASM_EXEC_TEST(F32x4Ne) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Ne, NotEqual); } WASM_EXEC_TEST(F32x4Gt) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Gt, Greater); } WASM_EXEC_TEST(F32x4Ge) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Ge, GreaterEqual); } WASM_EXEC_TEST(F32x4Lt) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Lt, Less); } WASM_EXEC_TEST(F32x4Le) { RunF32x4CompareOpTest(execution_tier, kExprF32x4Le, LessEqual); } template void RunShiftAddTestSequence(TestExecutionTier execution_tier, WasmOpcode shiftr_opcode, WasmOpcode add_opcode, WasmOpcode splat_opcode, int32_t imm, ScalarType (*shift_fn)(ScalarType, int32_t)) { WasmRunner r(execution_tier); // globals to store results for left and right cases ScalarType* g1 = r.builder().template AddGlobal(kWasmS128); ScalarType* g2 = r.builder().template AddGlobal(kWasmS128); uint8_t param = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); uint8_t temp2 = r.AllocateLocal(kWasmS128); auto expected_fn = [shift_fn](ScalarType x, ScalarType y, uint32_t imm) { return base::AddWithWraparound(x, shift_fn(y, imm)); }; r.Build( {WASM_LOCAL_SET(temp1, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(param))), WASM_LOCAL_SET(temp2, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(param))), WASM_GLOBAL_SET(0, WASM_SIMD_BINOP(add_opcode, WASM_SIMD_BINOP(shiftr_opcode, WASM_LOCAL_GET(temp2), WASM_I32V(imm)), WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_BINOP(add_opcode, WASM_LOCAL_GET(temp1), WASM_SIMD_BINOP(shiftr_opcode, WASM_LOCAL_GET(temp2), WASM_I32V(imm)))), WASM_ONE}); for (ScalarType x : compiler::ValueHelper::GetVector()) { r.Call(x); ScalarType expected = expected_fn(x, x, imm); for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { CHECK_EQ(expected, LANE(g1, i)); CHECK_EQ(expected, LANE(g2, i)); } } } WASM_EXEC_TEST(F32x4EqZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Eq, kExprF32x4Splat, Equal); } WASM_EXEC_TEST(F32x4NeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Ne, kExprF32x4Splat, NotEqual); } WASM_EXEC_TEST(F32x4GtZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Gt, kExprF32x4Splat, Greater); } WASM_EXEC_TEST(F32x4GeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Ge, kExprF32x4Splat, GreaterEqual); } WASM_EXEC_TEST(F32x4LtZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Lt, kExprF32x4Splat, Less); } WASM_EXEC_TEST(F32x4LeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF32x4Le, kExprF32x4Splat, LessEqual); } WASM_EXEC_TEST(I64x2Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. int64_t* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_I64x2_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_INT64_INPUTS(x) { r.Call(x); int64_t expected = x; for (int i = 0; i < 2; i++) { int64_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(I64x2ExtractLane) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmI64); r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET( 0, WASM_SIMD_I64x2_EXTRACT_LANE( 0, WASM_SIMD_I64x2_SPLAT(WASM_I64V(0xFFFFFFFFFF)))), WASM_LOCAL_SET(1, WASM_SIMD_I64x2_SPLAT(WASM_LOCAL_GET(0))), WASM_SIMD_I64x2_EXTRACT_LANE(1, WASM_LOCAL_GET(1))}); CHECK_EQ(0xFFFFFFFFFF, r.Call()); } WASM_EXEC_TEST(I64x2ReplaceLane) { WasmRunner r(execution_tier); // Set up a global to hold input/output vector. int64_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I64x2_SPLAT(WASM_I64V(-1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I64x2_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_I64V(0))), WASM_GLOBAL_SET(0, WASM_SIMD_I64x2_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_I64V(1))), WASM_ONE}); r.Call(); for (int64_t i = 0; i < 2; i++) { CHECK_EQ(i, LANE(g, i)); } } WASM_EXEC_TEST(I64x2Neg) { RunI64x2UnOpTest(execution_tier, kExprI64x2Neg, base::NegateWithWraparound); } WASM_EXEC_TEST(I64x2Abs) { RunI64x2UnOpTest(execution_tier, kExprI64x2Abs, std::abs); } WASM_EXEC_TEST(I64x2Shl) { RunI64x2ShiftOpTest(execution_tier, kExprI64x2Shl, LogicalShiftLeft); } WASM_EXEC_TEST(I64x2ShrS) { RunI64x2ShiftOpTest(execution_tier, kExprI64x2ShrS, ArithmeticShiftRight); } WASM_EXEC_TEST(I64x2ShrU) { RunI64x2ShiftOpTest(execution_tier, kExprI64x2ShrU, LogicalShiftRight); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I64x4Shl) { RunI64x4ShiftOpRevecTest(kExprI64x2Shl, LogicalShiftLeft); } TEST(RunWasmTurbofan_I64x4ShrU) { RunI64x4ShiftOpRevecTest(kExprI64x2ShrU, LogicalShiftRight); } #endif WASM_EXEC_TEST(I64x2ShiftAdd) { for (int imm = 0; imm <= 64; imm++) { RunShiftAddTestSequence(execution_tier, kExprI64x2ShrU, kExprI64x2Add, kExprI64x2Splat, imm, LogicalShiftRight); RunShiftAddTestSequence(execution_tier, kExprI64x2ShrS, kExprI64x2Add, kExprI64x2Splat, imm, ArithmeticShiftRight); } } WASM_EXEC_TEST(I64x2Add) { RunI64x2BinOpTest(execution_tier, kExprI64x2Add, base::AddWithWraparound); } WASM_EXEC_TEST(I64x2Sub) { RunI64x2BinOpTest(execution_tier, kExprI64x2Sub, base::SubWithWraparound); } WASM_EXEC_TEST(I64x2Eq) { RunI64x2BinOpTest(execution_tier, kExprI64x2Eq, Equal); } WASM_EXEC_TEST(I64x2Ne) { RunI64x2BinOpTest(execution_tier, kExprI64x2Ne, NotEqual); } WASM_EXEC_TEST(I64x2LtS) { RunI64x2BinOpTest(execution_tier, kExprI64x2LtS, Less); } WASM_EXEC_TEST(I64x2LeS) { RunI64x2BinOpTest(execution_tier, kExprI64x2LeS, LessEqual); } WASM_EXEC_TEST(I64x2GtS) { RunI64x2BinOpTest(execution_tier, kExprI64x2GtS, Greater); } WASM_EXEC_TEST(I64x2GeS) { RunI64x2BinOpTest(execution_tier, kExprI64x2GeS, GreaterEqual); } namespace { template void RunICompareOpConstImmTest(TestExecutionTier execution_tier, WasmOpcode cmp_opcode, WasmOpcode splat_opcode, ScalarType (*expected_op)(ScalarType, ScalarType)) { for (ScalarType x : compiler::ValueHelper::GetVector()) { WasmRunner r(execution_tier); // Set up global to hold mask output for left and right cases ScalarType* g1 = r.builder().template AddGlobal(kWasmS128); ScalarType* g2 = r.builder().template AddGlobal(kWasmS128); // Build fn to splat test values, perform compare op on both sides, and // write the result. uint8_t value = 0; uint8_t temp = r.AllocateLocal(kWasmS128); uint8_t const_buffer[kSimd128Size]; for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { WriteLittleEndianValue( base::bit_cast(&const_buffer[0]) + i, x); } r.Build( {WASM_LOCAL_SET(temp, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(cmp_opcode, WASM_SIMD_CONSTANT(const_buffer), WASM_LOCAL_GET(temp))), WASM_GLOBAL_SET(1, WASM_SIMD_BINOP(cmp_opcode, WASM_LOCAL_GET(temp), WASM_SIMD_CONSTANT(const_buffer))), WASM_ONE}); for (ScalarType y : compiler::ValueHelper::GetVector()) { r.Call(y); ScalarType expected1 = expected_op(x, y); ScalarType expected2 = expected_op(y, x); for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { CHECK_EQ(expected1, LANE(g1, i)); CHECK_EQ(expected2, LANE(g2, i)); } } } } } // namespace WASM_EXEC_TEST(I64x2EqZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2Eq, kExprI64x2Splat, Equal); } WASM_EXEC_TEST(I64x2NeZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2Ne, kExprI64x2Splat, NotEqual); } WASM_EXEC_TEST(I64x2GtZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2GtS, kExprI64x2Splat, Greater); } WASM_EXEC_TEST(I64x2GeZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2GeS, kExprI64x2Splat, GreaterEqual); } WASM_EXEC_TEST(I64x2LtZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2LtS, kExprI64x2Splat, Less); } WASM_EXEC_TEST(I64x2LeZero) { RunICompareOpConstImmTest(execution_tier, kExprI64x2LeS, kExprI64x2Splat, LessEqual); } WASM_EXEC_TEST(F64x2Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. double* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_F64x2_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_FLOAT64_INPUTS(x) { r.Call(x); double expected = x; for (int i = 0; i < 2; i++) { double actual = LANE(g, i); if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } } WASM_EXEC_TEST(F64x2ExtractLane) { WasmRunner r(execution_tier); uint8_t param1 = 0; uint8_t temp1 = r.AllocateLocal(kWasmF64); uint8_t temp2 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET( temp1, WASM_SIMD_F64x2_EXTRACT_LANE( 0, WASM_SIMD_F64x2_SPLAT(WASM_LOCAL_GET(param1)))), WASM_LOCAL_SET(temp2, WASM_SIMD_F64x2_SPLAT(WASM_LOCAL_GET(temp1))), WASM_SIMD_F64x2_EXTRACT_LANE(1, WASM_LOCAL_GET(temp2))}); FOR_FLOAT64_INPUTS(x) { double actual = r.Call(x); double expected = x; if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(F64x2ReplaceLane) { WasmRunner r(execution_tier); // Set up globals to hold input/output vector. double* g0 = r.builder().AddGlobal(kWasmS128); double* g1 = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its (FP) index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_F64x2_SPLAT(WASM_F64(1e100))), // Replace lane 0. WASM_GLOBAL_SET(0, WASM_SIMD_F64x2_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_F64(0.0f))), // Replace lane 1. WASM_GLOBAL_SET(1, WASM_SIMD_F64x2_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_F64(1.0f))), WASM_ONE}); r.Call(); CHECK_EQ(0., LANE(g0, 0)); CHECK_EQ(1e100, LANE(g0, 1)); CHECK_EQ(1e100, LANE(g1, 0)); CHECK_EQ(1., LANE(g1, 1)); } WASM_EXEC_TEST(F64x2ExtractLaneWithI64x2) { WasmRunner r(execution_tier); r.Build({WASM_IF_ELSE_L( WASM_F64_EQ(WASM_SIMD_F64x2_EXTRACT_LANE( 0, WASM_SIMD_I64x2_SPLAT(WASM_I64V(1e15))), WASM_F64_REINTERPRET_I64(WASM_I64V(1e15))), WASM_I64V(1), WASM_I64V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(I64x2ExtractWithF64x2) { WasmRunner r(execution_tier); r.Build( {WASM_IF_ELSE_L(WASM_I64_EQ(WASM_SIMD_I64x2_EXTRACT_LANE( 0, WASM_SIMD_F64x2_SPLAT(WASM_F64(1e15))), WASM_I64_REINTERPRET_F64(WASM_F64(1e15))), WASM_I64V(1), WASM_I64V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(F64x2Abs) { RunF64x2UnOpTest(execution_tier, kExprF64x2Abs, std::abs); } WASM_EXEC_TEST(F64x2Neg) { RunF64x2UnOpTest(execution_tier, kExprF64x2Neg, Negate); } WASM_EXEC_TEST(F64x2Sqrt) { RunF64x2UnOpTest(execution_tier, kExprF64x2Sqrt, std::sqrt); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_F64x4Sqrt) { RunF64x4UnOpRevecTest(kExprF64x2Sqrt, std::sqrt); } #endif WASM_EXEC_TEST(F64x2Ceil) { RunF64x2UnOpTest(execution_tier, kExprF64x2Ceil, ceil, true); } WASM_EXEC_TEST(F64x2Floor) { RunF64x2UnOpTest(execution_tier, kExprF64x2Floor, floor, true); } WASM_EXEC_TEST(F64x2Trunc) { RunF64x2UnOpTest(execution_tier, kExprF64x2Trunc, trunc, true); } WASM_EXEC_TEST(F64x2NearestInt) { RunF64x2UnOpTest(execution_tier, kExprF64x2NearestInt, nearbyint, true); } template void RunF64x2ConvertLowI32x4Test(TestExecutionTier execution_tier, WasmOpcode opcode) { WasmRunner r(execution_tier); double* g = r.builder().template AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_UNOP( opcode, // Set top lane of i64x2 == set top 2 lanes of i32x4. WASM_SIMD_I64x2_REPLACE_LANE( 1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(0)), WASM_ZERO64))), WASM_ONE}); for (SrcType x : compiler::ValueHelper::GetVector()) { r.Call(x); double expected = static_cast(x); for (int i = 0; i < 2; i++) { double actual = LANE(g, i); CheckDoubleResult(x, x, expected, actual, true); } } } WASM_EXEC_TEST(F64x2ConvertLowI32x4S) { RunF64x2ConvertLowI32x4Test(execution_tier, kExprF64x2ConvertLowI32x4S); } WASM_EXEC_TEST(F64x2ConvertLowI32x4U) { RunF64x2ConvertLowI32x4Test(execution_tier, kExprF64x2ConvertLowI32x4U); } template void RunI32x4TruncSatF64x2Test(TestExecutionTier execution_tier, WasmOpcode opcode) { WasmRunner r(execution_tier); SrcType* g = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(opcode, WASM_SIMD_F64x2_SPLAT( WASM_LOCAL_GET(0)))), WASM_ONE}); FOR_FLOAT64_INPUTS(x) { r.Call(x); SrcType expected = base::saturated_cast(x); for (int i = 0; i < 2; i++) { SrcType actual = LANE(g, i); CHECK_EQ(expected, actual); } // Top lanes are zero-ed. for (int i = 2; i < 4; i++) { CHECK_EQ(0, LANE(g, i)); } } } WASM_EXEC_TEST(I32x4TruncSatF64x2SZero) { RunI32x4TruncSatF64x2Test(execution_tier, kExprI32x4TruncSatF64x2SZero); } WASM_EXEC_TEST(I32x4TruncSatF64x2UZero) { RunI32x4TruncSatF64x2Test(execution_tier, kExprI32x4TruncSatF64x2UZero); } WASM_EXEC_TEST(F32x4DemoteF64x2Zero) { WasmRunner r(execution_tier); float* g = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_UNOP(kExprF32x4DemoteF64x2Zero, WASM_SIMD_F64x2_SPLAT(WASM_LOCAL_GET(0)))), WASM_ONE}); FOR_FLOAT64_INPUTS(x) { r.Call(x); float expected = DoubleToFloat32(x); for (int i = 0; i < 2; i++) { float actual = LANE(g, i); CheckFloatResult(x, x, expected, actual, true); } for (int i = 2; i < 4; i++) { float actual = LANE(g, i); CheckFloatResult(x, x, 0, actual, true); } } } WASM_EXEC_TEST(F64x2PromoteLowF32x4) { WasmRunner r(execution_tier); double* g = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_UNOP(kExprF64x2PromoteLowF32x4, WASM_SIMD_F32x4_SPLAT(WASM_LOCAL_GET(0)))), WASM_ONE}); FOR_FLOAT32_INPUTS(x) { r.Call(x); double expected = static_cast(x); for (int i = 0; i < 2; i++) { double actual = LANE(g, i); CheckDoubleResult(x, x, expected, actual, true); } } } // Test F64x2PromoteLowF32x4 with S128Load64Zero optimization (only on some // architectures). These 2 opcodes should be fused into a single instruction // with memory operands, which is tested in instruction-selector tests. This // test checks that we get correct results. WASM_EXEC_TEST(F64x2PromoteLowF32x4WithS128Load64Zero) { { WasmRunner r(execution_tier); double* g = r.builder().AddGlobal(kWasmS128); float* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(float)); r.builder().RandomizeMemory(); r.builder().WriteMemory(&memory[0], 1.0f); r.builder().WriteMemory(&memory[1], 3.0f); r.builder().WriteMemory(&memory[2], 5.0f); r.builder().WriteMemory(&memory[3], 8.0f); // Load at 4 (index) + 4 (offset) bytes, which is 2 floats. r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_UNOP(kExprF64x2PromoteLowF32x4, WASM_SIMD_LOAD_OP_OFFSET(kExprS128Load64Zero, WASM_I32V(4), 4))), WASM_ONE}); r.Call(); CHECK_EQ(5.0f, LANE(g, 0)); CHECK_EQ(8.0f, LANE(g, 1)); } { // OOB tests. WasmRunner r(execution_tier); r.builder().AddGlobal(kWasmS128); r.builder().AddMemoryElems(kWasmPageSize / sizeof(float)); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprF64x2PromoteLowF32x4, WASM_SIMD_LOAD_OP( kExprS128Load64Zero, WASM_I32V(kWasmPageSize)))), WASM_ONE}); CHECK_TRAP(r.Call()); } } WASM_EXEC_TEST(F64x2Add) { RunF64x2BinOpTest(execution_tier, kExprF64x2Add, Add); } WASM_EXEC_TEST(F64x2Sub) { RunF64x2BinOpTest(execution_tier, kExprF64x2Sub, Sub); } WASM_EXEC_TEST(F64x2Mul) { RunF64x2BinOpTest(execution_tier, kExprF64x2Mul, Mul); } WASM_EXEC_TEST(F64x2Div) { RunF64x2BinOpTest(execution_tier, kExprF64x2Div, base::Divide); } WASM_EXEC_TEST(F64x2Pmin) { RunF64x2BinOpTest(execution_tier, kExprF64x2Pmin, Minimum); } WASM_EXEC_TEST(F64x2Pmax) { RunF64x2BinOpTest(execution_tier, kExprF64x2Pmax, Maximum); } WASM_EXEC_TEST(F64x2Eq) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Eq, Equal); } WASM_EXEC_TEST(F64x2Ne) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Ne, NotEqual); } WASM_EXEC_TEST(F64x2Gt) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Gt, Greater); } WASM_EXEC_TEST(F64x2Ge) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Ge, GreaterEqual); } WASM_EXEC_TEST(F64x2Lt) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Lt, Less); } WASM_EXEC_TEST(F64x2Le) { RunF64x2CompareOpTest(execution_tier, kExprF64x2Le, LessEqual); } WASM_EXEC_TEST(F64x2EqZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Eq, kExprF64x2Splat, Equal); } WASM_EXEC_TEST(F64x2NeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Ne, kExprF64x2Splat, NotEqual); } WASM_EXEC_TEST(F64x2GtZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Gt, kExprF64x2Splat, Greater); } WASM_EXEC_TEST(F64x2GeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Ge, kExprF64x2Splat, GreaterEqual); } WASM_EXEC_TEST(F64x2LtZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Lt, kExprF64x2Splat, Less); } WASM_EXEC_TEST(F64x2LeZero) { RunF128CompareOpConstImmTest(execution_tier, kExprF64x2Le, kExprF64x2Splat, LessEqual); } WASM_EXEC_TEST(F64x2Min) { RunF64x2BinOpTest(execution_tier, kExprF64x2Min, JSMin); } WASM_EXEC_TEST(F64x2Max) { RunF64x2BinOpTest(execution_tier, kExprF64x2Max, JSMax); } WASM_EXEC_TEST(I64x2Mul) { RunI64x2BinOpTest(execution_tier, kExprI64x2Mul, base::MulWithWraparound); } WASM_EXEC_TEST(I32x4Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. int32_t* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_INT32_INPUTS(x) { r.Call(x); int32_t expected = x; for (int i = 0; i < 4; i++) { int32_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(I32x4ReplaceLane) { WasmRunner r(execution_tier); // Set up a global to hold input/output vector. int32_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_SPLAT(WASM_I32V(-1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_I32V(0))), WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_I32V(1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 2, WASM_LOCAL_GET(temp1), WASM_I32V(2))), WASM_GLOBAL_SET(0, WASM_SIMD_I32x4_REPLACE_LANE( 3, WASM_LOCAL_GET(temp1), WASM_I32V(3))), WASM_ONE}); r.Call(); for (int32_t i = 0; i < 4; i++) { CHECK_EQ(i, LANE(g, i)); } } WASM_EXEC_TEST(I16x8Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. int16_t* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_INT16_INPUTS(x) { r.Call(x); int16_t expected = x; for (int i = 0; i < 8; i++) { int16_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } // Test values that do not fit in a int16. FOR_INT32_INPUTS(x) { r.Call(x); int16_t expected = truncate_to_int16(x); for (int i = 0; i < 8; i++) { int16_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(I16x8ReplaceLane) { WasmRunner r(execution_tier); // Set up a global to hold input/output vector. int16_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_SPLAT(WASM_I32V(-1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_I32V(0))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_I32V(1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 2, WASM_LOCAL_GET(temp1), WASM_I32V(2))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 3, WASM_LOCAL_GET(temp1), WASM_I32V(3))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 4, WASM_LOCAL_GET(temp1), WASM_I32V(4))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 5, WASM_LOCAL_GET(temp1), WASM_I32V(5))), WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 6, WASM_LOCAL_GET(temp1), WASM_I32V(6))), WASM_GLOBAL_SET(0, WASM_SIMD_I16x8_REPLACE_LANE( 7, WASM_LOCAL_GET(temp1), WASM_I32V(7))), WASM_ONE}); r.Call(); for (int16_t i = 0; i < 8; i++) { CHECK_EQ(i, LANE(g, i)); } } WASM_EXEC_TEST(I8x16BitMask) { WasmRunner r(execution_tier); uint8_t value1 = r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(value1, WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I8x16_REPLACE_LANE( 0, WASM_LOCAL_GET(value1), WASM_I32V(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I8x16_REPLACE_LANE( 1, WASM_LOCAL_GET(value1), WASM_I32V(-1))), WASM_SIMD_UNOP(kExprI8x16BitMask, WASM_LOCAL_GET(value1))}); FOR_INT8_INPUTS(x) { int32_t actual = r.Call(x); // Lane 0 is always 0 (positive), lane 1 is always -1. int32_t expected = std::signbit(static_cast(x)) ? 0xFFFE : 0x0002; CHECK_EQ(actual, expected); } } WASM_EXEC_TEST(I16x8BitMask) { WasmRunner r(execution_tier); uint8_t value1 = r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(value1, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I16x8_REPLACE_LANE( 0, WASM_LOCAL_GET(value1), WASM_I32V(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I16x8_REPLACE_LANE( 1, WASM_LOCAL_GET(value1), WASM_I32V(-1))), WASM_SIMD_UNOP(kExprI16x8BitMask, WASM_LOCAL_GET(value1))}); FOR_INT16_INPUTS(x) { int32_t actual = r.Call(x); // Lane 0 is always 0 (positive), lane 1 is always -1. int32_t expected = std::signbit(static_cast(x)) ? 0xFE : 2; CHECK_EQ(actual, expected); } } WASM_EXEC_TEST(I32x4BitMask) { WasmRunner r(execution_tier); uint8_t value1 = r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(value1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I32x4_REPLACE_LANE( 0, WASM_LOCAL_GET(value1), WASM_I32V(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I32x4_REPLACE_LANE( 1, WASM_LOCAL_GET(value1), WASM_I32V(-1))), WASM_SIMD_UNOP(kExprI32x4BitMask, WASM_LOCAL_GET(value1))}); FOR_INT32_INPUTS(x) { int32_t actual = r.Call(x); // Lane 0 is always 0 (positive), lane 1 is always -1. int32_t expected = std::signbit(static_cast(x)) ? 0xE : 2; CHECK_EQ(actual, expected); } } WASM_EXEC_TEST(I64x2BitMask) { WasmRunner r(execution_tier); uint8_t value1 = r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(value1, WASM_SIMD_I64x2_SPLAT(WASM_LOCAL_GET(0))), WASM_LOCAL_SET(value1, WASM_SIMD_I64x2_REPLACE_LANE( 0, WASM_LOCAL_GET(value1), WASM_I64V_1(0))), WASM_SIMD_UNOP(kExprI64x2BitMask, WASM_LOCAL_GET(value1))}); for (int64_t x : compiler::ValueHelper::GetVector()) { int32_t actual = r.Call(x); // Lane 0 is always 0 (positive). int32_t expected = std::signbit(static_cast(x)) ? 0x2 : 0x0; CHECK_EQ(actual, expected); } } WASM_EXEC_TEST(I8x16Splat) { WasmRunner r(execution_tier); // Set up a global to hold output vector. int8_t* g = r.builder().AddGlobal(kWasmS128); uint8_t param1 = 0; r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(param1))), WASM_ONE}); FOR_INT8_INPUTS(x) { r.Call(x); int8_t expected = x; for (int i = 0; i < 16; i++) { int8_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } // Test values that do not fit in a int16. FOR_INT16_INPUTS(x) { r.Call(x); int8_t expected = truncate_to_int8(x); for (int i = 0; i < 16; i++) { int8_t actual = LANE(g, i); CHECK_EQ(actual, expected); } } } WASM_EXEC_TEST(I8x16ReplaceLane) { WasmRunner r(execution_tier); // Set up a global to hold input/output vector. int8_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_SPLAT(WASM_I32V(-1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 0, WASM_LOCAL_GET(temp1), WASM_I32V(0))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 1, WASM_LOCAL_GET(temp1), WASM_I32V(1))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 2, WASM_LOCAL_GET(temp1), WASM_I32V(2))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 3, WASM_LOCAL_GET(temp1), WASM_I32V(3))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 4, WASM_LOCAL_GET(temp1), WASM_I32V(4))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 5, WASM_LOCAL_GET(temp1), WASM_I32V(5))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 6, WASM_LOCAL_GET(temp1), WASM_I32V(6))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 7, WASM_LOCAL_GET(temp1), WASM_I32V(7))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 8, WASM_LOCAL_GET(temp1), WASM_I32V(8))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 9, WASM_LOCAL_GET(temp1), WASM_I32V(9))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 10, WASM_LOCAL_GET(temp1), WASM_I32V(10))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 11, WASM_LOCAL_GET(temp1), WASM_I32V(11))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 12, WASM_LOCAL_GET(temp1), WASM_I32V(12))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 13, WASM_LOCAL_GET(temp1), WASM_I32V(13))), WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 14, WASM_LOCAL_GET(temp1), WASM_I32V(14))), WASM_GLOBAL_SET(0, WASM_SIMD_I8x16_REPLACE_LANE( 15, WASM_LOCAL_GET(temp1), WASM_I32V(15))), WASM_ONE}); r.Call(); for (int8_t i = 0; i < 16; i++) { CHECK_EQ(i, LANE(g, i)); } } // Use doubles to ensure exact conversion. int32_t ConvertToInt(double val, bool unsigned_integer) { if (std::isnan(val)) return 0; if (unsigned_integer) { if (val < 0) return 0; if (val > kMaxUInt32) return kMaxUInt32; return static_cast(val); } else { if (val < kMinInt) return kMinInt; if (val > kMaxInt) return kMaxInt; return static_cast(val); } } // Tests both signed and unsigned conversion. WASM_EXEC_TEST(I32x4ConvertF32x4) { WasmRunner r(execution_tier); // Create two output vectors to hold signed and unsigned results. int32_t* g0 = r.builder().AddGlobal(kWasmS128); int32_t* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprI32x4SConvertF32x4, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_UNOP(kExprI32x4UConvertF32x4, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_FLOAT32_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; r.Call(x); int32_t expected_signed = ConvertToInt(x, false); int32_t expected_unsigned = ConvertToInt(x, true); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_unsigned, LANE(g1, i)); } } } // Tests both signed and unsigned conversion from I16x8 (unpacking). WASM_EXEC_TEST(I32x4ConvertI16x8) { WasmRunner r(execution_tier); // Create four output vectors to hold signed and unsigned results. int32_t* g0 = r.builder().AddGlobal(kWasmS128); int32_t* g1 = r.builder().AddGlobal(kWasmS128); int32_t* g2 = r.builder().AddGlobal(kWasmS128); int32_t* g3 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprI32x4SConvertI16x8High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_UNOP(kExprI32x4SConvertI16x8Low, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(2, WASM_SIMD_UNOP(kExprI32x4UConvertI16x8High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(3, WASM_SIMD_UNOP(kExprI32x4UConvertI16x8Low, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT16_INPUTS(x) { r.Call(x); int32_t expected_signed = static_cast(x); int32_t expected_unsigned = static_cast(static_cast(x)); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_signed, LANE(g1, i)); CHECK_EQ(expected_unsigned, LANE(g2, i)); CHECK_EQ(expected_unsigned, LANE(g3, i)); } } } // Tests both signed and unsigned conversion from I32x4 (unpacking). WASM_EXEC_TEST(I64x2ConvertI32x4) { WasmRunner r(execution_tier); // Create four output vectors to hold signed and unsigned results. int64_t* g0 = r.builder().AddGlobal(kWasmS128); int64_t* g1 = r.builder().AddGlobal(kWasmS128); uint64_t* g2 = r.builder().AddGlobal(kWasmS128); uint64_t* g3 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprI64x2SConvertI32x4High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_UNOP(kExprI64x2SConvertI32x4Low, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(2, WASM_SIMD_UNOP(kExprI64x2UConvertI32x4High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(3, WASM_SIMD_UNOP(kExprI64x2UConvertI32x4Low, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT32_INPUTS(x) { r.Call(x); int64_t expected_signed = static_cast(x); uint64_t expected_unsigned = static_cast(static_cast(x)); for (int i = 0; i < 2; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_signed, LANE(g1, i)); CHECK_EQ(expected_unsigned, LANE(g2, i)); CHECK_EQ(expected_unsigned, LANE(g3, i)); } } } WASM_EXEC_TEST(I32x4Neg) { RunI32x4UnOpTest(execution_tier, kExprI32x4Neg, base::NegateWithWraparound); } WASM_EXEC_TEST(I32x4Abs) { RunI32x4UnOpTest(execution_tier, kExprI32x4Abs, std::abs); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I32x8Neg) { RunI32x8UnOpRevecTest(kExprI32x4Neg, base::NegateWithWraparound); } TEST(RunWasmTurbofan_I32x8Abs) { RunI32x8UnOpRevecTest(kExprI32x4Abs, std::abs); } #endif WASM_EXEC_TEST(S128Not) { RunI32x4UnOpTest(execution_tier, kExprS128Not, [](int32_t x) { return ~x; }); } template void RunExtAddPairwiseTest(TestExecutionTier execution_tier, WasmOpcode ext_add_pairwise, WasmOpcode splat, Shuffle interleaving_shuffle) { constexpr int num_lanes = kSimd128Size / sizeof(Wide); WasmRunner r(execution_tier); Wide* g = r.builder().template AddGlobal(kWasmS128); r.Build({WASM_SIMD_I8x16_SHUFFLE_OP(kExprI8x16Shuffle, interleaving_shuffle, WASM_SIMD_UNOP(splat, WASM_LOCAL_GET(0)), WASM_SIMD_UNOP(splat, WASM_LOCAL_GET(1))), WASM_SIMD_OP(ext_add_pairwise), kExprGlobalSet, 0, WASM_ONE}); auto v = compiler::ValueHelper::GetVector(); // Iterate vector from both ends to try and splat two different values. for (auto i = v.begin(), j = v.end() - 1; i < v.end(); i++, j--) { r.Call(*i, *j); Wide expected = AddLong(*i, *j); for (int l = 0; l < num_lanes; l++) { CHECK_EQ(expected, LANE(g, l)); } } } // interleave even lanes from one input and odd lanes from another. constexpr Shuffle interleave_16x8_shuffle = {0, 1, 18, 19, 4, 5, 22, 23, 8, 9, 26, 27, 12, 13, 30, 31}; constexpr Shuffle interleave_8x16_shuffle = {0, 17, 2, 19, 4, 21, 6, 23, 8, 25, 10, 27, 12, 29, 14, 31}; WASM_EXEC_TEST(I32x4ExtAddPairwiseI16x8S) { RunExtAddPairwiseTest( execution_tier, kExprI32x4ExtAddPairwiseI16x8S, kExprI16x8Splat, interleave_16x8_shuffle); } WASM_EXEC_TEST(I32x4ExtAddPairwiseI16x8U) { RunExtAddPairwiseTest( execution_tier, kExprI32x4ExtAddPairwiseI16x8U, kExprI16x8Splat, interleave_16x8_shuffle); } WASM_EXEC_TEST(I16x8ExtAddPairwiseI8x16S) { RunExtAddPairwiseTest( execution_tier, kExprI16x8ExtAddPairwiseI8x16S, kExprI8x16Splat, interleave_8x16_shuffle); } WASM_EXEC_TEST(I16x8ExtAddPairwiseI8x16U) { RunExtAddPairwiseTest( execution_tier, kExprI16x8ExtAddPairwiseI8x16U, kExprI8x16Splat, interleave_8x16_shuffle); } WASM_EXEC_TEST(I32x4Add) { RunI32x4BinOpTest(execution_tier, kExprI32x4Add, base::AddWithWraparound); } WASM_EXEC_TEST(I32x4Sub) { RunI32x4BinOpTest(execution_tier, kExprI32x4Sub, base::SubWithWraparound); } WASM_EXEC_TEST(I32x4Mul) { RunI32x4BinOpTest(execution_tier, kExprI32x4Mul, base::MulWithWraparound); } WASM_EXEC_TEST(I32x4MinS) { RunI32x4BinOpTest(execution_tier, kExprI32x4MinS, Minimum); } WASM_EXEC_TEST(I32x4MaxS) { RunI32x4BinOpTest(execution_tier, kExprI32x4MaxS, Maximum); } WASM_EXEC_TEST(I32x4MinU) { RunI32x4BinOpTest(execution_tier, kExprI32x4MinU, UnsignedMinimum); } WASM_EXEC_TEST(I32x4MaxU) { RunI32x4BinOpTest(execution_tier, kExprI32x4MaxU, UnsignedMaximum); } WASM_EXEC_TEST(S128And) { RunI32x4BinOpTest(execution_tier, kExprS128And, [](int32_t x, int32_t y) { return x & y; }); } enum ConstSide { kConstLeft, kConstRight }; template using BinOp = ScalarType (*)(ScalarType, ScalarType); template void RunS128ConstBinOpTest(TestExecutionTier execution_tier, ConstSide const_side, WasmOpcode binop_opcode, WasmOpcode splat_opcode, BinOp expected_op) { for (ScalarType x : compiler::ValueHelper::GetVector()) { WasmRunner r(execution_tier); // Global to hold output. ScalarType* g = r.builder().template AddGlobal(kWasmS128); // Build a function to splat one argument into a local, // and execute the op with a const as the second argument uint8_t value = 0; uint8_t temp = r.AllocateLocal(kWasmS128); uint8_t const_buffer[16]; for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { WriteLittleEndianValue( base::bit_cast(&const_buffer[0]) + i, x); } switch (const_side) { case kConstLeft: r.Build({WASM_LOCAL_SET( temp, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(binop_opcode, WASM_SIMD_CONSTANT(const_buffer), WASM_LOCAL_GET(temp))), WASM_ONE}); break; case kConstRight: r.Build({WASM_LOCAL_SET( temp, WASM_SIMD_OPN(splat_opcode, WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(binop_opcode, WASM_LOCAL_GET(temp), WASM_SIMD_CONSTANT(const_buffer))), WASM_ONE}); break; } for (ScalarType y : compiler::ValueHelper::GetVector()) { r.Call(y); ScalarType expected = (const_side == kConstLeft) ? expected_op(x, y) : expected_op(y, x); for (size_t i = 0; i < kSimd128Size / sizeof(ScalarType); i++) { CHECK_EQ(expected, LANE(g, i)); } } } } WASM_EXEC_TEST(S128AndImm) { RunS128ConstBinOpTest(execution_tier, kConstLeft, kExprS128And, kExprI32x4Splat, [](int32_t x, int32_t y) { return x & y; }); RunS128ConstBinOpTest(execution_tier, kConstRight, kExprS128And, kExprI32x4Splat, [](int32_t x, int32_t y) { return x & y; }); RunS128ConstBinOpTest( execution_tier, kConstLeft, kExprS128And, kExprI16x8Splat, [](int16_t x, int16_t y) { return static_cast(x & y); }); RunS128ConstBinOpTest( execution_tier, kConstRight, kExprS128And, kExprI16x8Splat, [](int16_t x, int16_t y) { return static_cast(x & y); }); } WASM_EXEC_TEST(S128Or) { RunI32x4BinOpTest(execution_tier, kExprS128Or, [](int32_t x, int32_t y) { return x | y; }); } WASM_EXEC_TEST(S128Xor) { RunI32x4BinOpTest(execution_tier, kExprS128Xor, [](int32_t x, int32_t y) { return x ^ y; }); } // Bitwise operation, doesn't really matter what simd type we test it with. WASM_EXEC_TEST(S128AndNot) { RunI32x4BinOpTest(execution_tier, kExprS128AndNot, [](int32_t x, int32_t y) { return x & ~y; }); } WASM_EXEC_TEST(S128AndNotImm) { RunS128ConstBinOpTest(execution_tier, kConstLeft, kExprS128AndNot, kExprI32x4Splat, [](int32_t x, int32_t y) { return x & ~y; }); RunS128ConstBinOpTest(execution_tier, kConstRight, kExprS128AndNot, kExprI32x4Splat, [](int32_t x, int32_t y) { return x & ~y; }); RunS128ConstBinOpTest( execution_tier, kConstLeft, kExprS128AndNot, kExprI16x8Splat, [](int16_t x, int16_t y) { return static_cast(x & ~y); }); RunS128ConstBinOpTest( execution_tier, kConstRight, kExprS128AndNot, kExprI16x8Splat, [](int16_t x, int16_t y) { return static_cast(x & ~y); }); } WASM_EXEC_TEST(I32x4Eq) { RunI32x4BinOpTest(execution_tier, kExprI32x4Eq, Equal); } WASM_EXEC_TEST(I32x4Ne) { RunI32x4BinOpTest(execution_tier, kExprI32x4Ne, NotEqual); } WASM_EXEC_TEST(I32x4LtS) { RunI32x4BinOpTest(execution_tier, kExprI32x4LtS, Less); } WASM_EXEC_TEST(I32x4LeS) { RunI32x4BinOpTest(execution_tier, kExprI32x4LeS, LessEqual); } WASM_EXEC_TEST(I32x4GtS) { RunI32x4BinOpTest(execution_tier, kExprI32x4GtS, Greater); } WASM_EXEC_TEST(I32x4GeS) { RunI32x4BinOpTest(execution_tier, kExprI32x4GeS, GreaterEqual); } WASM_EXEC_TEST(I32x4LtU) { RunI32x4BinOpTest(execution_tier, kExprI32x4LtU, UnsignedLess); } WASM_EXEC_TEST(I32x4LeU) { RunI32x4BinOpTest(execution_tier, kExprI32x4LeU, UnsignedLessEqual); } WASM_EXEC_TEST(I32x4GtU) { RunI32x4BinOpTest(execution_tier, kExprI32x4GtU, UnsignedGreater); } WASM_EXEC_TEST(I32x4GeU) { RunI32x4BinOpTest(execution_tier, kExprI32x4GeU, UnsignedGreaterEqual); } WASM_EXEC_TEST(I32x4EqZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4Eq, kExprI32x4Splat, Equal); } WASM_EXEC_TEST(I32x4NeZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4Ne, kExprI32x4Splat, NotEqual); } WASM_EXEC_TEST(I32x4GtZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4GtS, kExprI32x4Splat, Greater); } WASM_EXEC_TEST(I32x4GeZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4GeS, kExprI32x4Splat, GreaterEqual); } WASM_EXEC_TEST(I32x4LtZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4LtS, kExprI32x4Splat, Less); } WASM_EXEC_TEST(I32x4LeZero) { RunICompareOpConstImmTest(execution_tier, kExprI32x4LeS, kExprI32x4Splat, LessEqual); } WASM_EXEC_TEST(I32x4Shl) { RunI32x4ShiftOpTest(execution_tier, kExprI32x4Shl, LogicalShiftLeft); } WASM_EXEC_TEST(I32x4ShrS) { RunI32x4ShiftOpTest(execution_tier, kExprI32x4ShrS, ArithmeticShiftRight); } WASM_EXEC_TEST(I32x4ShrU) { RunI32x4ShiftOpTest(execution_tier, kExprI32x4ShrU, LogicalShiftRight); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I32x8Shl) { RunI32x8ShiftOpRevecTest(kExprI32x4Shl, LogicalShiftLeft); } TEST(RunWasmTurbofan_I32x8ShrS) { RunI32x8ShiftOpRevecTest(kExprI32x4ShrS, ArithmeticShiftRight); } TEST(RunWasmTurbofan_I32x8ShrU) { RunI32x8ShiftOpRevecTest(kExprI32x4ShrU, LogicalShiftRight); } #endif WASM_EXEC_TEST(I32x4ShiftAdd) { for (int imm = 0; imm <= 32; imm++) { RunShiftAddTestSequence(execution_tier, kExprI32x4ShrU, kExprI32x4Add, kExprI32x4Splat, imm, LogicalShiftRight); RunShiftAddTestSequence(execution_tier, kExprI32x4ShrS, kExprI32x4Add, kExprI32x4Splat, imm, ArithmeticShiftRight); } } // Tests both signed and unsigned conversion from I8x16 (unpacking). WASM_EXEC_TEST(I16x8ConvertI8x16) { WasmRunner r(execution_tier); // Create four output vectors to hold signed and unsigned results. int16_t* g0 = r.builder().AddGlobal(kWasmS128); int16_t* g1 = r.builder().AddGlobal(kWasmS128); int16_t* g2 = r.builder().AddGlobal(kWasmS128); int16_t* g3 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_UNOP(kExprI16x8SConvertI8x16High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_UNOP(kExprI16x8SConvertI8x16Low, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(2, WASM_SIMD_UNOP(kExprI16x8UConvertI8x16High, WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(3, WASM_SIMD_UNOP(kExprI16x8UConvertI8x16Low, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT8_INPUTS(x) { r.Call(x); int16_t expected_signed = static_cast(x); int16_t expected_unsigned = static_cast(static_cast(x)); for (int i = 0; i < 8; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_signed, LANE(g1, i)); CHECK_EQ(expected_unsigned, LANE(g2, i)); CHECK_EQ(expected_unsigned, LANE(g3, i)); } } } // Tests both signed and unsigned conversion from I32x4 (packing). WASM_EXEC_TEST(I16x8ConvertI32x4) { WasmRunner r(execution_tier); // Create output vectors to hold signed and unsigned results. int16_t* g0 = r.builder().AddGlobal(kWasmS128); int16_t* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_BINOP(kExprI16x8SConvertI32x4, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_BINOP(kExprI16x8UConvertI32x4, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT32_INPUTS(x) { r.Call(x); int16_t expected_signed = base::saturated_cast(x); int16_t expected_unsigned = base::saturated_cast(x); for (int i = 0; i < 8; i++) { CHECK_EQ(expected_signed, LANE(g0, i)); CHECK_EQ(expected_unsigned, LANE(g1, i)); } } } WASM_EXEC_TEST(I16x8Neg) { RunI16x8UnOpTest(execution_tier, kExprI16x8Neg, base::NegateWithWraparound); } WASM_EXEC_TEST(I16x8Abs) { RunI16x8UnOpTest(execution_tier, kExprI16x8Abs, Abs); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I16x16Neg) { RunI16x16UnOpRevecTest(kExprI16x8Neg, base::NegateWithWraparound); } TEST(RunWasmTurbofan_I16x16Abs) { RunI16x16UnOpRevecTest(kExprI16x8Abs, Abs); } #endif WASM_EXEC_TEST(I16x8Add) { RunI16x8BinOpTest(execution_tier, kExprI16x8Add, base::AddWithWraparound); } WASM_EXEC_TEST(I16x8AddSatS) { RunI16x8BinOpTest(execution_tier, kExprI16x8AddSatS, SaturateAdd); } WASM_EXEC_TEST(I16x8Sub) { RunI16x8BinOpTest(execution_tier, kExprI16x8Sub, base::SubWithWraparound); } WASM_EXEC_TEST(I16x8SubSatS) { RunI16x8BinOpTest(execution_tier, kExprI16x8SubSatS, SaturateSub); } WASM_EXEC_TEST(I16x8Mul) { RunI16x8BinOpTest(execution_tier, kExprI16x8Mul, base::MulWithWraparound); } WASM_EXEC_TEST(I16x8MinS) { RunI16x8BinOpTest(execution_tier, kExprI16x8MinS, Minimum); } WASM_EXEC_TEST(I16x8MaxS) { RunI16x8BinOpTest(execution_tier, kExprI16x8MaxS, Maximum); } WASM_EXEC_TEST(I16x8AddSatU) { RunI16x8BinOpTest(execution_tier, kExprI16x8AddSatU, SaturateAdd); } WASM_EXEC_TEST(I16x8SubSatU) { RunI16x8BinOpTest(execution_tier, kExprI16x8SubSatU, SaturateSub); } WASM_EXEC_TEST(I16x8MinU) { RunI16x8BinOpTest(execution_tier, kExprI16x8MinU, UnsignedMinimum); } WASM_EXEC_TEST(I16x8MaxU) { RunI16x8BinOpTest(execution_tier, kExprI16x8MaxU, UnsignedMaximum); } WASM_EXEC_TEST(I16x8Eq) { RunI16x8BinOpTest(execution_tier, kExprI16x8Eq, Equal); } WASM_EXEC_TEST(I16x8Ne) { RunI16x8BinOpTest(execution_tier, kExprI16x8Ne, NotEqual); } WASM_EXEC_TEST(I16x8LtS) { RunI16x8BinOpTest(execution_tier, kExprI16x8LtS, Less); } WASM_EXEC_TEST(I16x8LeS) { RunI16x8BinOpTest(execution_tier, kExprI16x8LeS, LessEqual); } WASM_EXEC_TEST(I16x8GtS) { RunI16x8BinOpTest(execution_tier, kExprI16x8GtS, Greater); } WASM_EXEC_TEST(I16x8GeS) { RunI16x8BinOpTest(execution_tier, kExprI16x8GeS, GreaterEqual); } WASM_EXEC_TEST(I16x8GtU) { RunI16x8BinOpTest(execution_tier, kExprI16x8GtU, UnsignedGreater); } WASM_EXEC_TEST(I16x8GeU) { RunI16x8BinOpTest(execution_tier, kExprI16x8GeU, UnsignedGreaterEqual); } WASM_EXEC_TEST(I16x8LtU) { RunI16x8BinOpTest(execution_tier, kExprI16x8LtU, UnsignedLess); } WASM_EXEC_TEST(I16x8LeU) { RunI16x8BinOpTest(execution_tier, kExprI16x8LeU, UnsignedLessEqual); } WASM_EXEC_TEST(I16x8EqZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8Eq, kExprI16x8Splat, Equal); } WASM_EXEC_TEST(I16x8NeZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8Ne, kExprI16x8Splat, NotEqual); } WASM_EXEC_TEST(I16x8GtZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8GtS, kExprI16x8Splat, Greater); } WASM_EXEC_TEST(I16x8GeZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8GeS, kExprI16x8Splat, GreaterEqual); } WASM_EXEC_TEST(I16x8LtZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8LtS, kExprI16x8Splat, Less); } WASM_EXEC_TEST(I16x8LeZero) { RunICompareOpConstImmTest(execution_tier, kExprI16x8LeS, kExprI16x8Splat, LessEqual); } WASM_EXEC_TEST(I16x8RoundingAverageU) { RunI16x8BinOpTest(execution_tier, kExprI16x8RoundingAverageU, RoundingAverageUnsigned); } WASM_EXEC_TEST(I16x8Q15MulRSatS) { RunI16x8BinOpTest(execution_tier, kExprI16x8Q15MulRSatS, SaturateRoundingQMul); } namespace { enum class MulHalf { kLow, kHigh }; // Helper to run ext mul tests. It will splat 2 input values into 2 v128, call // the mul op on these operands, and set the result into a global. // It will zero the top or bottom half of one of the operands, this will catch // mistakes if we are multiply the incorrect halves. template void RunExtMulTest(TestExecutionTier execution_tier, WasmOpcode opcode, OpType expected_op, WasmOpcode splat, MulHalf half) { WasmRunner r(execution_tier); int lane_to_zero = half == MulHalf::kLow ? 1 : 0; T* g = r.builder().template AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(opcode, WASM_SIMD_I64x2_REPLACE_LANE( lane_to_zero, WASM_SIMD_UNOP(splat, WASM_LOCAL_GET(0)), WASM_I64V_1(0)), WASM_SIMD_UNOP(splat, WASM_LOCAL_GET(1)))), WASM_ONE}); constexpr int lanes = kSimd128Size / sizeof(T); for (S x : compiler::ValueHelper::GetVector()) { for (S y : compiler::ValueHelper::GetVector()) { r.Call(x, y); T expected = expected_op(x, y); for (int i = 0; i < lanes; i++) { CHECK_EQ(expected, LANE(g, i)); } } } } } // namespace WASM_EXEC_TEST(I16x8ExtMulLowI8x16S) { RunExtMulTest(execution_tier, kExprI16x8ExtMulLowI8x16S, MultiplyLong, kExprI8x16Splat, MulHalf::kLow); } WASM_EXEC_TEST(I16x8ExtMulHighI8x16S) { RunExtMulTest(execution_tier, kExprI16x8ExtMulHighI8x16S, MultiplyLong, kExprI8x16Splat, MulHalf::kHigh); } WASM_EXEC_TEST(I16x8ExtMulLowI8x16U) { RunExtMulTest(execution_tier, kExprI16x8ExtMulLowI8x16U, MultiplyLong, kExprI8x16Splat, MulHalf::kLow); } WASM_EXEC_TEST(I16x8ExtMulHighI8x16U) { RunExtMulTest(execution_tier, kExprI16x8ExtMulHighI8x16U, MultiplyLong, kExprI8x16Splat, MulHalf::kHigh); } WASM_EXEC_TEST(I32x4ExtMulLowI16x8S) { RunExtMulTest(execution_tier, kExprI32x4ExtMulLowI16x8S, MultiplyLong, kExprI16x8Splat, MulHalf::kLow); } WASM_EXEC_TEST(I32x4ExtMulHighI16x8S) { RunExtMulTest(execution_tier, kExprI32x4ExtMulHighI16x8S, MultiplyLong, kExprI16x8Splat, MulHalf::kHigh); } WASM_EXEC_TEST(I32x4ExtMulLowI16x8U) { RunExtMulTest(execution_tier, kExprI32x4ExtMulLowI16x8U, MultiplyLong, kExprI16x8Splat, MulHalf::kLow); } WASM_EXEC_TEST(I32x4ExtMulHighI16x8U) { RunExtMulTest(execution_tier, kExprI32x4ExtMulHighI16x8U, MultiplyLong, kExprI16x8Splat, MulHalf::kHigh); } WASM_EXEC_TEST(I64x2ExtMulLowI32x4S) { RunExtMulTest(execution_tier, kExprI64x2ExtMulLowI32x4S, MultiplyLong, kExprI32x4Splat, MulHalf::kLow); } WASM_EXEC_TEST(I64x2ExtMulHighI32x4S) { RunExtMulTest(execution_tier, kExprI64x2ExtMulHighI32x4S, MultiplyLong, kExprI32x4Splat, MulHalf::kHigh); } WASM_EXEC_TEST(I64x2ExtMulLowI32x4U) { RunExtMulTest(execution_tier, kExprI64x2ExtMulLowI32x4U, MultiplyLong, kExprI32x4Splat, MulHalf::kLow); } WASM_EXEC_TEST(I64x2ExtMulHighI32x4U) { RunExtMulTest(execution_tier, kExprI64x2ExtMulHighI32x4U, MultiplyLong, kExprI32x4Splat, MulHalf::kHigh); } namespace { // Test add(mul(x, y, z) optimizations. template void RunExtMulAddOptimizationTest(TestExecutionTier execution_tier, WasmOpcode ext_mul, WasmOpcode narrow_splat, WasmOpcode wide_splat, WasmOpcode wide_add, std::function addop) { WasmRunner r(execution_tier); T* g = r.builder().template AddGlobal(kWasmS128); // global[0] = // add( // splat(local[1]), // extmul(splat(local[0]), splat(local[0]))) r.Build( {WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP( wide_add, WASM_SIMD_UNOP(wide_splat, WASM_LOCAL_GET(1)), WASM_SIMD_BINOP( ext_mul, WASM_SIMD_UNOP(narrow_splat, WASM_LOCAL_GET(0)), WASM_SIMD_UNOP(narrow_splat, WASM_LOCAL_GET(0))))), WASM_ONE}); constexpr int lanes = kSimd128Size / sizeof(T); for (S x : compiler::ValueHelper::GetVector()) { for (T y : compiler::ValueHelper::GetVector()) { r.Call(x, y); T expected = addop(MultiplyLong(x, x), y); for (int i = 0; i < lanes; i++) { CHECK_EQ(expected, LANE(g, i)); } } } } } // namespace // Helper which defines high/low, signed/unsigned test cases for extmul + add // optimization. #define EXTMUL_ADD_OPTIMIZATION_TEST(NarrowType, NarrowShape, WideType, \ WideShape) \ WASM_EXEC_TEST(WideShape##ExtMulLow##NarrowShape##SAddOptimization) { \ RunExtMulAddOptimizationTest( \ execution_tier, kExpr##WideShape##ExtMulLow##NarrowShape##S, \ kExpr##NarrowShape##Splat, kExpr##WideShape##Splat, \ kExpr##WideShape##Add, base::AddWithWraparound); \ } \ WASM_EXEC_TEST(WideShape##ExtMulHigh##NarrowShape##SAddOptimization) { \ RunExtMulAddOptimizationTest( \ execution_tier, kExpr##WideShape##ExtMulHigh##NarrowShape##S, \ kExpr##NarrowShape##Splat, kExpr##WideShape##Splat, \ kExpr##WideShape##Add, base::AddWithWraparound); \ } \ WASM_EXEC_TEST(WideShape##ExtMulLow##NarrowShape##UAddOptimization) { \ RunExtMulAddOptimizationTest( \ execution_tier, kExpr##WideShape##ExtMulLow##NarrowShape##U, \ kExpr##NarrowShape##Splat, kExpr##WideShape##Splat, \ kExpr##WideShape##Add, std::plus()); \ } \ WASM_EXEC_TEST(WideShape##ExtMulHigh##NarrowShape##UAddOptimization) { \ RunExtMulAddOptimizationTest( \ execution_tier, kExpr##WideShape##ExtMulHigh##NarrowShape##U, \ kExpr##NarrowShape##Splat, kExpr##WideShape##Splat, \ kExpr##WideShape##Add, std::plus()); \ } EXTMUL_ADD_OPTIMIZATION_TEST(int8_t, I8x16, int16_t, I16x8) EXTMUL_ADD_OPTIMIZATION_TEST(int16_t, I16x8, int32_t, I32x4) #undef EXTMUL_ADD_OPTIMIZATION_TEST WASM_EXEC_TEST(I32x4DotI16x8S) { WasmRunner r(execution_tier); int32_t* g = r.builder().template AddGlobal(kWasmS128); uint8_t value1 = 0, value2 = 1; uint8_t temp1 = r.AllocateLocal(kWasmS128); uint8_t temp2 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(value1))), WASM_LOCAL_SET(temp2, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(value2))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(kExprI32x4DotI16x8S, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp2))), WASM_ONE}); for (int16_t x : compiler::ValueHelper::GetVector()) { for (int16_t y : compiler::ValueHelper::GetVector()) { r.Call(x, y); // x * y * 2 can overflow (0x8000), the behavior is to wraparound. int32_t expected = base::MulWithWraparound(x * y, 2); for (int i = 0; i < 4; i++) { CHECK_EQ(expected, LANE(g, i)); } } } } WASM_EXEC_TEST(I16x8Shl) { RunI16x8ShiftOpTest(execution_tier, kExprI16x8Shl, LogicalShiftLeft); } WASM_EXEC_TEST(I16x8ShrS) { RunI16x8ShiftOpTest(execution_tier, kExprI16x8ShrS, ArithmeticShiftRight); } WASM_EXEC_TEST(I16x8ShrU) { RunI16x8ShiftOpTest(execution_tier, kExprI16x8ShrU, LogicalShiftRight); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I16x16Shl) { RunI16x16ShiftOpRevecTest(kExprI16x8Shl, LogicalShiftLeft); } TEST(RunWasmTurbofan_I16x16ShrS) { RunI16x16ShiftOpRevecTest(kExprI16x8ShrS, ArithmeticShiftRight); } TEST(RunWasmTurbofan_I16x16ShrU) { RunI16x16ShiftOpRevecTest(kExprI16x8ShrU, LogicalShiftRight); } #endif WASM_EXEC_TEST(I16x8ShiftAdd) { for (int imm = 0; imm <= 16; imm++) { RunShiftAddTestSequence(execution_tier, kExprI16x8ShrU, kExprI16x8Add, kExprI16x8Splat, imm, LogicalShiftRight); RunShiftAddTestSequence(execution_tier, kExprI16x8ShrS, kExprI16x8Add, kExprI16x8Splat, imm, ArithmeticShiftRight); } } WASM_EXEC_TEST(I8x16Neg) { RunI8x16UnOpTest(execution_tier, kExprI8x16Neg, base::NegateWithWraparound); } WASM_EXEC_TEST(I8x16Abs) { RunI8x16UnOpTest(execution_tier, kExprI8x16Abs, Abs); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_I8x32Neg) { RunI8x32UnOpRevecTest(kExprI8x16Neg, base::NegateWithWraparound); } TEST(RunWasmTurbofan_I8x32Abs) { RunI8x32UnOpRevecTest(kExprI8x16Abs, Abs); } #endif WASM_EXEC_TEST(I8x16Popcnt) { WasmRunner r(execution_tier); // Global to hold output. int8_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_UNOP(kExprI8x16Popcnt, WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_UINT8_INPUTS(x) { r.Call(x); unsigned expected = base::bits::CountPopulation(x); for (int i = 0; i < 16; i++) { CHECK_EQ(expected, LANE(g, i)); } } } // Tests both signed and unsigned conversion from I16x8 (packing). WASM_EXEC_TEST(I8x16ConvertI16x8) { WasmRunner r(execution_tier); // Create output vectors to hold signed and unsigned results. int8_t* g_s = r.builder().AddGlobal(kWasmS128); uint8_t* g_u = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. uint8_t value = 0; uint8_t temp1 = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(temp1, WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(value))), WASM_GLOBAL_SET(0, WASM_SIMD_BINOP(kExprI8x16SConvertI16x8, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp1))), WASM_GLOBAL_SET(1, WASM_SIMD_BINOP(kExprI8x16UConvertI16x8, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp1))), WASM_ONE}); FOR_INT16_INPUTS(x) { r.Call(x); int8_t expected_signed = base::saturated_cast(x); uint8_t expected_unsigned = base::saturated_cast(x); for (int i = 0; i < 16; i++) { CHECK_EQ(expected_signed, LANE(g_s, i)); CHECK_EQ(expected_unsigned, LANE(g_u, i)); } } } WASM_EXEC_TEST(I8x16Add) { RunI8x16BinOpTest(execution_tier, kExprI8x16Add, base::AddWithWraparound); } WASM_EXEC_TEST(I8x16AddSatS) { RunI8x16BinOpTest(execution_tier, kExprI8x16AddSatS, SaturateAdd); } WASM_EXEC_TEST(I8x16Sub) { RunI8x16BinOpTest(execution_tier, kExprI8x16Sub, base::SubWithWraparound); } WASM_EXEC_TEST(I8x16SubSatS) { RunI8x16BinOpTest(execution_tier, kExprI8x16SubSatS, SaturateSub); } WASM_EXEC_TEST(I8x16MinS) { RunI8x16BinOpTest(execution_tier, kExprI8x16MinS, Minimum); } WASM_EXEC_TEST(I8x16MaxS) { RunI8x16BinOpTest(execution_tier, kExprI8x16MaxS, Maximum); } WASM_EXEC_TEST(I8x16AddSatU) { RunI8x16BinOpTest(execution_tier, kExprI8x16AddSatU, SaturateAdd); } WASM_EXEC_TEST(I8x16SubSatU) { RunI8x16BinOpTest(execution_tier, kExprI8x16SubSatU, SaturateSub); } WASM_EXEC_TEST(I8x16MinU) { RunI8x16BinOpTest(execution_tier, kExprI8x16MinU, UnsignedMinimum); } WASM_EXEC_TEST(I8x16MaxU) { RunI8x16BinOpTest(execution_tier, kExprI8x16MaxU, UnsignedMaximum); } WASM_EXEC_TEST(I8x16Eq) { RunI8x16BinOpTest(execution_tier, kExprI8x16Eq, Equal); } WASM_EXEC_TEST(I8x16Ne) { RunI8x16BinOpTest(execution_tier, kExprI8x16Ne, NotEqual); } WASM_EXEC_TEST(I8x16GtS) { RunI8x16BinOpTest(execution_tier, kExprI8x16GtS, Greater); } WASM_EXEC_TEST(I8x16GeS) { RunI8x16BinOpTest(execution_tier, kExprI8x16GeS, GreaterEqual); } WASM_EXEC_TEST(I8x16LtS) { RunI8x16BinOpTest(execution_tier, kExprI8x16LtS, Less); } WASM_EXEC_TEST(I8x16LeS) { RunI8x16BinOpTest(execution_tier, kExprI8x16LeS, LessEqual); } WASM_EXEC_TEST(I8x16GtU) { RunI8x16BinOpTest(execution_tier, kExprI8x16GtU, UnsignedGreater); } WASM_EXEC_TEST(I8x16GeU) { RunI8x16BinOpTest(execution_tier, kExprI8x16GeU, UnsignedGreaterEqual); } WASM_EXEC_TEST(I8x16LtU) { RunI8x16BinOpTest(execution_tier, kExprI8x16LtU, UnsignedLess); } WASM_EXEC_TEST(I8x16LeU) { RunI8x16BinOpTest(execution_tier, kExprI8x16LeU, UnsignedLessEqual); } WASM_EXEC_TEST(I8x16EqZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16Eq, kExprI8x16Splat, Equal); } WASM_EXEC_TEST(I8x16NeZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16Ne, kExprI8x16Splat, NotEqual); } WASM_EXEC_TEST(I8x16GtZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16GtS, kExprI8x16Splat, Greater); } WASM_EXEC_TEST(I8x16GeZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16GeS, kExprI8x16Splat, GreaterEqual); } WASM_EXEC_TEST(I8x16LtZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16LtS, kExprI8x16Splat, Less); } WASM_EXEC_TEST(I8x16LeZero) { RunICompareOpConstImmTest(execution_tier, kExprI8x16LeS, kExprI8x16Splat, LessEqual); } WASM_EXEC_TEST(I8x16RoundingAverageU) { RunI8x16BinOpTest(execution_tier, kExprI8x16RoundingAverageU, RoundingAverageUnsigned); } WASM_EXEC_TEST(I8x16Shl) { RunI8x16ShiftOpTest(execution_tier, kExprI8x16Shl, LogicalShiftLeft); } WASM_EXEC_TEST(I8x16ShrS) { RunI8x16ShiftOpTest(execution_tier, kExprI8x16ShrS, ArithmeticShiftRight); } WASM_EXEC_TEST(I8x16ShrU) { RunI8x16ShiftOpTest(execution_tier, kExprI8x16ShrU, LogicalShiftRight); } WASM_EXEC_TEST(I8x16ShiftAdd) { for (int imm = 0; imm <= 8; imm++) { RunShiftAddTestSequence(execution_tier, kExprI8x16ShrU, kExprI8x16Add, kExprI8x16Splat, imm, LogicalShiftRight); RunShiftAddTestSequence(execution_tier, kExprI8x16ShrS, kExprI8x16Add, kExprI8x16Splat, imm, ArithmeticShiftRight); } } // Test Select by making a mask where the 0th and 3rd lanes are true and the // rest false, and comparing for non-equality with zero to convert to a boolean // vector. #define WASM_SIMD_SELECT_TEST(format) \ WASM_EXEC_TEST(S##format##Select) { \ WasmRunner r(execution_tier); \ uint8_t val1 = 0; \ uint8_t val2 = 1; \ uint8_t src1 = r.AllocateLocal(kWasmS128); \ uint8_t src2 = r.AllocateLocal(kWasmS128); \ uint8_t zero = r.AllocateLocal(kWasmS128); \ uint8_t mask = r.AllocateLocal(kWasmS128); \ r.Build( \ {WASM_LOCAL_SET(src1, \ WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(val1))), \ WASM_LOCAL_SET(src2, \ WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(val2))), \ WASM_LOCAL_SET(zero, WASM_SIMD_I##format##_SPLAT(WASM_ZERO)), \ WASM_LOCAL_SET(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 1, WASM_LOCAL_GET(zero), WASM_I32V(-1))), \ WASM_LOCAL_SET(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 2, WASM_LOCAL_GET(mask), WASM_I32V(-1))), \ WASM_LOCAL_SET( \ mask, \ WASM_SIMD_SELECT( \ format, WASM_LOCAL_GET(src1), WASM_LOCAL_GET(src2), \ WASM_SIMD_BINOP(kExprI##format##Ne, WASM_LOCAL_GET(mask), \ WASM_LOCAL_GET(zero)))), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val2, 0), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val1, 1), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val1, 2), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val2, 3), WASM_ONE}); \ \ CHECK_EQ(1, r.Call(0x12, 0x34)); \ } WASM_SIMD_SELECT_TEST(32x4) WASM_SIMD_SELECT_TEST(16x8) WASM_SIMD_SELECT_TEST(8x16) // Test Select by making a mask where the 0th and 3rd lanes are non-zero and the // rest 0. The mask is not the result of a comparison op. #define WASM_SIMD_NON_CANONICAL_SELECT_TEST(format) \ WASM_EXEC_TEST(S##format##NonCanonicalSelect) { \ WasmRunner r(execution_tier); \ uint8_t val1 = 0; \ uint8_t val2 = 1; \ uint8_t combined = 2; \ uint8_t src1 = r.AllocateLocal(kWasmS128); \ uint8_t src2 = r.AllocateLocal(kWasmS128); \ uint8_t zero = r.AllocateLocal(kWasmS128); \ uint8_t mask = r.AllocateLocal(kWasmS128); \ r.Build( \ {WASM_LOCAL_SET(src1, \ WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(val1))), \ WASM_LOCAL_SET(src2, \ WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(val2))), \ WASM_LOCAL_SET(zero, WASM_SIMD_I##format##_SPLAT(WASM_ZERO)), \ WASM_LOCAL_SET(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 1, WASM_LOCAL_GET(zero), WASM_I32V(0xF))), \ WASM_LOCAL_SET(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 2, WASM_LOCAL_GET(mask), WASM_I32V(0xF))), \ WASM_LOCAL_SET(mask, WASM_SIMD_SELECT(format, WASM_LOCAL_GET(src1), \ WASM_LOCAL_GET(src2), \ WASM_LOCAL_GET(mask))), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val2, 0), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, combined, 1), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, combined, 2), \ WASM_SIMD_CHECK_LANE_S(I##format, mask, I32, val2, 3), WASM_ONE}); \ \ CHECK_EQ(1, r.Call(0x12, 0x34, 0x32)); \ } WASM_SIMD_NON_CANONICAL_SELECT_TEST(32x4) WASM_SIMD_NON_CANONICAL_SELECT_TEST(16x8) WASM_SIMD_NON_CANONICAL_SELECT_TEST(8x16) // Test binary ops with two lane test patterns, all lanes distinct. template void RunBinaryLaneOpTest( TestExecutionTier execution_tier, WasmOpcode simd_op, const std::array& expected) { WasmRunner r(execution_tier); // Set up two test patterns as globals, e.g. [0, 1, 2, 3] and [4, 5, 6, 7]. T* src0 = r.builder().AddGlobal(kWasmS128); T* src1 = r.builder().AddGlobal(kWasmS128); static const int kElems = kSimd128Size / sizeof(T); for (int i = 0; i < kElems; i++) { LANE(src0, i) = i; LANE(src1, i) = kElems + i; } if (simd_op == kExprI8x16Shuffle) { r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_I8x16_SHUFFLE_OP(simd_op, expected, WASM_GLOBAL_GET(0), WASM_GLOBAL_GET(1))), WASM_ONE}); } else { r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_BINOP(simd_op, WASM_GLOBAL_GET(0), WASM_GLOBAL_GET(1))), WASM_ONE}); } CHECK_EQ(1, r.Call()); for (size_t i = 0; i < expected.size(); i++) { CHECK_EQ(LANE(src0, i), expected[i]); } } // Test shuffle ops. void RunShuffleOpTest(TestExecutionTier execution_tier, WasmOpcode simd_op, const std::array& shuffle) { // Test the original shuffle. RunBinaryLaneOpTest(execution_tier, simd_op, shuffle); // Test a non-canonical (inputs reversed) version of the shuffle. std::array other_shuffle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) other_shuffle[i] ^= kSimd128Size; RunBinaryLaneOpTest(execution_tier, simd_op, other_shuffle); // Test the swizzle (one-operand) version of the shuffle. std::array swizzle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) swizzle[i] &= (kSimd128Size - 1); RunBinaryLaneOpTest(execution_tier, simd_op, swizzle); // Test the non-canonical swizzle (one-operand) version of the shuffle. std::array other_swizzle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) other_swizzle[i] |= kSimd128Size; RunBinaryLaneOpTest(execution_tier, simd_op, other_swizzle); } #define SHUFFLE_LIST(V) \ V(S128Identity) \ V(S32x4Dup) \ V(S32x4ZipLeft) \ V(S32x4ZipRight) \ V(S32x4UnzipLeft) \ V(S32x4UnzipRight) \ V(S32x4TransposeLeft) \ V(S32x4TransposeRight) \ V(S32x2Reverse) \ V(S32x4Irregular) \ V(S32x4Rotate) \ V(S16x8Dup) \ V(S16x8ZipLeft) \ V(S16x8ZipRight) \ V(S16x8UnzipLeft) \ V(S16x8UnzipRight) \ V(S16x8TransposeLeft) \ V(S16x8TransposeRight) \ V(S16x4Reverse) \ V(S16x2Reverse) \ V(S16x8Irregular) \ V(S8x16Dup) \ V(S8x16ZipLeft) \ V(S8x16ZipRight) \ V(S8x16UnzipLeft) \ V(S8x16UnzipRight) \ V(S8x16TransposeLeft) \ V(S8x16TransposeRight) \ V(S8x8Reverse) \ V(S8x4Reverse) \ V(S8x2Reverse) \ V(S8x16Irregular) enum ShuffleKey { #define SHUFFLE_ENUM_VALUE(Name) k##Name, SHUFFLE_LIST(SHUFFLE_ENUM_VALUE) #undef SHUFFLE_ENUM_VALUE kNumShuffleKeys }; using ShuffleMap = std::map; ShuffleMap test_shuffles = { {kS128Identity, {{16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31}}}, {kS32x4Dup, {{16, 17, 18, 19, 16, 17, 18, 19, 16, 17, 18, 19, 16, 17, 18, 19}}}, {kS32x4ZipLeft, {{0, 1, 2, 3, 16, 17, 18, 19, 4, 5, 6, 7, 20, 21, 22, 23}}}, {kS32x4ZipRight, {{8, 9, 10, 11, 24, 25, 26, 27, 12, 13, 14, 15, 28, 29, 30, 31}}}, {kS32x4UnzipLeft, {{0, 1, 2, 3, 8, 9, 10, 11, 16, 17, 18, 19, 24, 25, 26, 27}}}, {kS32x4UnzipRight, {{4, 5, 6, 7, 12, 13, 14, 15, 20, 21, 22, 23, 28, 29, 30, 31}}}, {kS32x4TransposeLeft, {{0, 1, 2, 3, 16, 17, 18, 19, 8, 9, 10, 11, 24, 25, 26, 27}}}, {kS32x4TransposeRight, {{4, 5, 6, 7, 20, 21, 22, 23, 12, 13, 14, 15, 28, 29, 30, 31}}}, {kS32x2Reverse, // swizzle only {{4, 5, 6, 7, 0, 1, 2, 3, 12, 13, 14, 15, 8, 9, 10, 11}}}, {kS32x4Irregular, {{0, 1, 2, 3, 16, 17, 18, 19, 16, 17, 18, 19, 20, 21, 22, 23}}}, {kS32x4Rotate, {{4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 0, 1, 2, 3}}}, {kS16x8Dup, {{18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19}}}, {kS16x8ZipLeft, {{0, 1, 16, 17, 2, 3, 18, 19, 4, 5, 20, 21, 6, 7, 22, 23}}}, {kS16x8ZipRight, {{8, 9, 24, 25, 10, 11, 26, 27, 12, 13, 28, 29, 14, 15, 30, 31}}}, {kS16x8UnzipLeft, {{0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}}}, {kS16x8UnzipRight, {{2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}}}, {kS16x8TransposeLeft, {{0, 1, 16, 17, 4, 5, 20, 21, 8, 9, 24, 25, 12, 13, 28, 29}}}, {kS16x8TransposeRight, {{2, 3, 18, 19, 6, 7, 22, 23, 10, 11, 26, 27, 14, 15, 30, 31}}}, {kS16x4Reverse, // swizzle only {{6, 7, 4, 5, 2, 3, 0, 1, 14, 15, 12, 13, 10, 11, 8, 9}}}, {kS16x2Reverse, // swizzle only {{2, 3, 0, 1, 6, 7, 4, 5, 10, 11, 8, 9, 14, 15, 12, 13}}}, {kS16x8Irregular, {{0, 1, 16, 17, 16, 17, 0, 1, 4, 5, 20, 21, 6, 7, 22, 23}}}, {kS8x16Dup, {{19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19}}}, {kS8x16ZipLeft, {{0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23}}}, {kS8x16ZipRight, {{8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31}}}, {kS8x16UnzipLeft, {{0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}}}, {kS8x16UnzipRight, {{1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31}}}, {kS8x16TransposeLeft, {{0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30}}}, {kS8x16TransposeRight, {{1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31}}}, {kS8x8Reverse, // swizzle only {{7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8}}}, {kS8x4Reverse, // swizzle only {{3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12}}}, {kS8x2Reverse, // swizzle only {{1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14}}}, {kS8x16Irregular, {{0, 16, 0, 16, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23}}}, }; #define SHUFFLE_TEST(Name) \ WASM_EXEC_TEST(Name) { \ ShuffleMap::const_iterator it = test_shuffles.find(k##Name); \ DCHECK_NE(it, test_shuffles.end()); \ RunShuffleOpTest(execution_tier, kExprI8x16Shuffle, it->second); \ } SHUFFLE_LIST(SHUFFLE_TEST) #undef SHUFFLE_TEST #undef SHUFFLE_LIST // Test shuffles that blend the two vectors (elements remain in their lanes.) WASM_EXEC_TEST(S8x16Blend) { std::array expected; for (int bias = 1; bias < kSimd128Size; bias++) { for (int i = 0; i < bias; i++) expected[i] = i; for (int i = bias; i < kSimd128Size; i++) expected[i] = i + kSimd128Size; RunShuffleOpTest(execution_tier, kExprI8x16Shuffle, expected); } } // Test shuffles that concatenate the two vectors. WASM_EXEC_TEST(S8x16Concat) { std::array expected; // n is offset or bias of concatenation. for (int n = 1; n < kSimd128Size; ++n) { int i = 0; // last kLanes - n bytes of first vector. for (int j = n; j < kSimd128Size; ++j) { expected[i++] = j; } // first n bytes of second vector for (int j = 0; j < n; ++j) { expected[i++] = j + kSimd128Size; } RunShuffleOpTest(execution_tier, kExprI8x16Shuffle, expected); } } WASM_EXEC_TEST(ShuffleShufps) { // We reverse engineer the shufps immediates into 8x16 shuffles. std::array expected; for (int mask = 0; mask < 256; mask++) { // Each iteration of this loop sets byte[i] of the 32x4 lanes. // Low 2 lanes (2-bits each) select from first input. uint8_t index0 = (mask & 3) * 4; uint8_t index1 = ((mask >> 2) & 3) * 4; // Next 2 bits select from src2, so add 16 to the index. uint8_t index2 = ((mask >> 4) & 3) * 4 + 16; uint8_t index3 = ((mask >> 6) & 3) * 4 + 16; for (int i = 0; i < 4; i++) { expected[0 + i] = index0 + i; expected[4 + i] = index1 + i; expected[8 + i] = index2 + i; expected[12 + i] = index3 + i; } RunShuffleOpTest(execution_tier, kExprI8x16Shuffle, expected); } } WASM_EXEC_TEST(I8x16ShuffleWithZeroInput) { WasmRunner r(execution_tier); static const int kElems = kSimd128Size / sizeof(uint8_t); uint8_t* dst = r.builder().AddGlobal(kWasmS128); uint8_t* src1 = r.builder().AddGlobal(kWasmS128); // src0 is zero, it's used to zero extend src1 for (int i = 0; i < kElems; i++) { LANE(src1, i) = i; } // Zero extend first 4 elments of src1 to 32 bit constexpr std::array shuffle = {16, 1, 2, 3, 17, 5, 6, 7, 18, 9, 10, 11, 19, 13, 14, 15}; constexpr std::array expected = {0, 0, 0, 0, 1, 0, 0, 0, 2, 0, 0, 0, 3, 0, 0, 0}; constexpr std::array zeros = {0}; r.Build( {WASM_GLOBAL_SET(0, WASM_SIMD_I8x16_SHUFFLE_OP(kExprI8x16Shuffle, shuffle, WASM_SIMD_CONSTANT(zeros), WASM_GLOBAL_GET(1))), WASM_ONE}); CHECK_EQ(1, r.Call()); for (int i = 0; i < kElems; i++) { CHECK_EQ(LANE(dst, i), expected[i]); } } struct SwizzleTestArgs { const Shuffle input; const Shuffle indices; const Shuffle expected; }; static constexpr SwizzleTestArgs swizzle_test_args[] = { {{15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0}, {15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0}, {0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15}}, {{15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0}, {15, 0, 14, 1, 13, 2, 12, 3, 11, 4, 10, 5, 9, 6, 8, 7}, {0, 15, 1, 14, 2, 13, 3, 12, 4, 11, 5, 10, 6, 9, 7, 8}}, {{15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0}, {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}, {15, 13, 11, 9, 7, 5, 3, 1, 0, 0, 0, 0, 0, 0, 0, 0}}, // all indices are out of range {{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15}, {16, 17, 18, 19, 20, 124, 125, 126, 127, -1, -2, -3, -4, -5, -6, -7}, {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}}}; static constexpr base::Vector swizzle_test_vector = base::ArrayVector(swizzle_test_args); WASM_EXEC_TEST(I8x16Swizzle) { // RunBinaryLaneOpTest set up the two globals to be consecutive integers, // [0-15] and [16-31]. Using [0-15] as the indices will not sufficiently test // swizzle since the expected result is a no-op, using [16-31] will result in // all 0s. { WasmRunner r(execution_tier); static const int kElems = kSimd128Size / sizeof(uint8_t); uint8_t* dst = r.builder().AddGlobal(kWasmS128); uint8_t* src0 = r.builder().AddGlobal(kWasmS128); uint8_t* src1 = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(kExprI8x16Swizzle, WASM_GLOBAL_GET(1), WASM_GLOBAL_GET(2))), WASM_ONE}); for (SwizzleTestArgs si : swizzle_test_vector) { for (int i = 0; i < kElems; i++) { LANE(src0, i) = si.input[i]; LANE(src1, i) = si.indices[i]; } CHECK_EQ(1, r.Call()); for (int i = 0; i < kElems; i++) { CHECK_EQ(LANE(dst, i), si.expected[i]); } } } { // We have an optimization for constant indices, test this case. for (SwizzleTestArgs si : swizzle_test_vector) { WasmRunner r(execution_tier); uint8_t* dst = r.builder().AddGlobal(kWasmS128); uint8_t* src0 = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP(kExprI8x16Swizzle, WASM_GLOBAL_GET(1), WASM_SIMD_CONSTANT(si.indices))), WASM_ONE}); for (int i = 0; i < kSimd128Size; i++) { LANE(src0, i) = si.input[i]; } CHECK_EQ(1, r.Call()); for (int i = 0; i < kSimd128Size; i++) { CHECK_EQ(LANE(dst, i), si.expected[i]); } } } } // Combine 3 shuffles a, b, and c by applying both a and b and then applying c // to those two results. Shuffle Combine(const Shuffle& a, const Shuffle& b, const Shuffle& c) { Shuffle result; for (int i = 0; i < kSimd128Size; ++i) { result[i] = c[i] < kSimd128Size ? a[c[i]] : b[c[i] - kSimd128Size]; } return result; } const Shuffle& GetRandomTestShuffle(v8::base::RandomNumberGenerator* rng) { return test_shuffles[static_cast(rng->NextInt(kNumShuffleKeys))]; } // Test shuffles that are random combinations of 3 test shuffles. Completely // random shuffles almost always generate the slow general shuffle code, so // don't exercise as many code paths. WASM_EXEC_TEST(I8x16ShuffleFuzz) { v8::base::RandomNumberGenerator* rng = CcTest::random_number_generator(); static const int kTests = 100; for (int i = 0; i < kTests; ++i) { auto shuffle = Combine(GetRandomTestShuffle(rng), GetRandomTestShuffle(rng), GetRandomTestShuffle(rng)); RunShuffleOpTest(execution_tier, kExprI8x16Shuffle, shuffle); } } void AppendShuffle(const Shuffle& shuffle, std::vector* buffer) { uint8_t opcode[] = {WASM_SIMD_OP(kExprI8x16Shuffle)}; for (size_t i = 0; i < arraysize(opcode); ++i) buffer->push_back(opcode[i]); for (size_t i = 0; i < kSimd128Size; ++i) buffer->push_back((shuffle[i])); } void BuildShuffle(const std::vector& shuffles, std::vector* buffer) { // Perform the leaf shuffles on globals 0 and 1. size_t row_index = (shuffles.size() - 1) / 2; for (size_t i = row_index; i < shuffles.size(); ++i) { uint8_t operands[] = {WASM_GLOBAL_GET(0), WASM_GLOBAL_GET(1)}; for (size_t j = 0; j < arraysize(operands); ++j) buffer->push_back(operands[j]); AppendShuffle(shuffles[i], buffer); } // Now perform inner shuffles in the correct order on operands on the stack. do { for (size_t i = row_index / 2; i < row_index; ++i) { AppendShuffle(shuffles[i], buffer); } row_index /= 2; } while (row_index != 0); uint8_t epilog[] = {kExprGlobalSet, static_cast(0), WASM_ONE}; for (size_t j = 0; j < arraysize(epilog); ++j) buffer->push_back(epilog[j]); } void RunWasmCode(TestExecutionTier execution_tier, const std::vector& code, std::array* result) { WasmRunner r(execution_tier); // Set up two test patterns as globals, e.g. [0, 1, 2, 3] and [4, 5, 6, 7]. int8_t* src0 = r.builder().AddGlobal(kWasmS128); int8_t* src1 = r.builder().AddGlobal(kWasmS128); for (int i = 0; i < kSimd128Size; ++i) { LANE(src0, i) = i; LANE(src1, i) = kSimd128Size + i; } r.Build(code.data(), code.data() + code.size()); CHECK_EQ(1, r.Call()); for (size_t i = 0; i < kSimd128Size; i++) { (*result)[i] = LANE(src0, i); } } // Boolean unary operations are 'AllTrue' and 'AnyTrue', which return an integer // result. Use relational ops on numeric vectors to create the boolean vector // test inputs. Test inputs with all true, all false, one true, and one false. #define WASM_SIMD_BOOL_REDUCTION_TEST(format, lanes, int_type) \ WASM_EXEC_TEST(ReductionTest##lanes) { \ WasmRunner r(execution_tier); \ if (lanes == 2) return; \ uint8_t zero = r.AllocateLocal(kWasmS128); \ uint8_t one_one = r.AllocateLocal(kWasmS128); \ uint8_t reduced = r.AllocateLocal(kWasmI32); \ r.Build( \ {WASM_LOCAL_SET(zero, WASM_SIMD_I##format##_SPLAT(int_type(0))), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprV128AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_LOCAL_GET(zero), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprV128AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_LOCAL_GET(zero), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_NE(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprI##format##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_LOCAL_GET(zero), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprI##format##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_LOCAL_GET(zero), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_NE(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET(one_one, \ WASM_SIMD_I##format##_REPLACE_LANE( \ lanes - 1, WASM_LOCAL_GET(zero), int_type(1))), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprV128AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_LOCAL_GET(one_one), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprV128AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_LOCAL_GET(one_one), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprI##format##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_LOCAL_GET(one_one), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_NE(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_LOCAL_SET( \ reduced, WASM_SIMD_UNOP(kExprI##format##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_LOCAL_GET(one_one), \ WASM_LOCAL_GET(zero)))), \ WASM_IF(WASM_I32_NE(WASM_LOCAL_GET(reduced), WASM_ZERO), \ WASM_RETURN(WASM_ZERO)), \ WASM_ONE}); \ CHECK_EQ(1, r.Call()); \ } WASM_SIMD_BOOL_REDUCTION_TEST(64x2, 2, WASM_I64V) WASM_SIMD_BOOL_REDUCTION_TEST(32x4, 4, WASM_I32V) WASM_SIMD_BOOL_REDUCTION_TEST(16x8, 8, WASM_I32V) WASM_SIMD_BOOL_REDUCTION_TEST(8x16, 16, WASM_I32V) WASM_EXEC_TEST(SimdI32x4ExtractWithF32x4) { WasmRunner r(execution_tier); r.Build( {WASM_IF_ELSE_I(WASM_I32_EQ(WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_F32x4_SPLAT(WASM_F32(30.5))), WASM_I32_REINTERPRET_F32(WASM_F32(30.5))), WASM_I32V(1), WASM_I32V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(SimdF32x4ExtractWithI32x4) { WasmRunner r(execution_tier); r.Build( {WASM_IF_ELSE_I(WASM_F32_EQ(WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(15))), WASM_F32_REINTERPRET_I32(WASM_I32V(15))), WASM_I32V(1), WASM_I32V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(SimdF32x4ExtractLane) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmF32); r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(0, WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_F32x4_SPLAT(WASM_F32(30.5)))), WASM_LOCAL_SET(1, WASM_SIMD_F32x4_SPLAT(WASM_LOCAL_GET(0))), WASM_SIMD_F32x4_EXTRACT_LANE(1, WASM_LOCAL_GET(1))}); CHECK_EQ(30.5, r.Call()); } WASM_EXEC_TEST(SimdF32x4AddWithI32x4) { // Choose two floating point values whose sum is normal and exactly // representable as a float. const int kOne = 0x3F800000; const int kTwo = 0x40000000; WasmRunner r(execution_tier); r.Build({WASM_IF_ELSE_I( WASM_F32_EQ( WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_BINOP(kExprF32x4Add, WASM_SIMD_I32x4_SPLAT(WASM_I32V(kOne)), WASM_SIMD_I32x4_SPLAT(WASM_I32V(kTwo)))), WASM_F32_ADD(WASM_F32_REINTERPRET_I32(WASM_I32V(kOne)), WASM_F32_REINTERPRET_I32(WASM_I32V(kTwo)))), WASM_I32V(1), WASM_I32V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(SimdI32x4AddWithF32x4) { WasmRunner r(execution_tier); r.Build({WASM_IF_ELSE_I( WASM_I32_EQ( WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_BINOP(kExprI32x4Add, WASM_SIMD_F32x4_SPLAT(WASM_F32(21.25)), WASM_SIMD_F32x4_SPLAT(WASM_F32(31.5)))), WASM_I32_ADD(WASM_I32_REINTERPRET_F32(WASM_F32(21.25)), WASM_I32_REINTERPRET_F32(WASM_F32(31.5)))), WASM_I32V(1), WASM_I32V(0))}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(SimdI32x4Local) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(31))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_LOCAL_GET(0))}); CHECK_EQ(31, r.Call()); } WASM_EXEC_TEST(SimdI32x4SplatFromExtract) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(0, WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(76)))), WASM_LOCAL_SET(1, WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(0))), WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_LOCAL_GET(1))}); CHECK_EQ(76, r.Call()); } WASM_EXEC_TEST(SimdI32x4For) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(1, WASM_SIMD_I32x4_SPLAT(WASM_I32V(31))), WASM_LOCAL_SET(1, WASM_SIMD_I32x4_REPLACE_LANE(1, WASM_LOCAL_GET(1), WASM_I32V(53))), WASM_LOCAL_SET(1, WASM_SIMD_I32x4_REPLACE_LANE(2, WASM_LOCAL_GET(1), WASM_I32V(23))), WASM_LOCAL_SET(0, WASM_I32V(0)), WASM_LOOP( WASM_LOCAL_SET(1, WASM_SIMD_BINOP(kExprI32x4Add, WASM_LOCAL_GET(1), WASM_SIMD_I32x4_SPLAT(WASM_I32V(1)))), WASM_IF(WASM_I32_NE(WASM_INC_LOCAL(0), WASM_I32V(5)), WASM_BR(1))), WASM_LOCAL_SET(0, WASM_I32V(1)), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_LOCAL_GET(1)), WASM_I32V(36)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_LOCAL_GET(1)), WASM_I32V(58)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(2, WASM_LOCAL_GET(1)), WASM_I32V(28)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(3, WASM_LOCAL_GET(1)), WASM_I32V(36)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_LOCAL_GET(0)}); CHECK_EQ(1, r.Call()); } WASM_EXEC_TEST(SimdF32x4For) { WasmRunner r(execution_tier); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); r.Build( {WASM_LOCAL_SET(1, WASM_SIMD_F32x4_SPLAT(WASM_F32(21.25))), WASM_LOCAL_SET(1, WASM_SIMD_F32x4_REPLACE_LANE(3, WASM_LOCAL_GET(1), WASM_F32(19.5))), WASM_LOCAL_SET(0, WASM_I32V(0)), WASM_LOOP( WASM_LOCAL_SET( 1, WASM_SIMD_BINOP(kExprF32x4Add, WASM_LOCAL_GET(1), WASM_SIMD_F32x4_SPLAT(WASM_F32(2.0)))), WASM_IF(WASM_I32_NE(WASM_INC_LOCAL(0), WASM_I32V(3)), WASM_BR(1))), WASM_LOCAL_SET(0, WASM_I32V(1)), WASM_IF(WASM_F32_NE(WASM_SIMD_F32x4_EXTRACT_LANE(0, WASM_LOCAL_GET(1)), WASM_F32(27.25)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_SIMD_F32x4_EXTRACT_LANE(3, WASM_LOCAL_GET(1)), WASM_F32(25.5)), WASM_LOCAL_SET(0, WASM_I32V(0))), WASM_LOCAL_GET(0)}); CHECK_EQ(1, r.Call()); } template void SetVectorByLanes(T* v, const std::array& arr) { for (int lane = 0; lane < numLanes; lane++) { LANE(v, lane) = arr[lane]; } } template const T GetScalar(T* v, int lane) { DCHECK_GE(lane, 0); DCHECK_LT(static_cast(lane), kSimd128Size / sizeof(T)); return LANE(v, lane); } WASM_EXEC_TEST(SimdI32x4GetGlobal) { WasmRunner r(execution_tier); // Pad the globals with a few unused slots to get a non-zero offset. r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused int32_t* global = r.builder().AddGlobal(kWasmS128); SetVectorByLanes(global, {{0, 1, 2, 3}}); r.AllocateLocal(kWasmI32); r.Build( {WASM_LOCAL_SET(1, WASM_I32V(1)), WASM_IF(WASM_I32_NE(WASM_I32V(0), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_GLOBAL_GET(4))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(1), WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_GLOBAL_GET(4))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(2), WASM_SIMD_I32x4_EXTRACT_LANE(2, WASM_GLOBAL_GET(4))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(3), WASM_SIMD_I32x4_EXTRACT_LANE(3, WASM_GLOBAL_GET(4))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_LOCAL_GET(1)}); CHECK_EQ(1, r.Call(0)); } WASM_EXEC_TEST(SimdI32x4SetGlobal) { WasmRunner r(execution_tier); // Pad the globals with a few unused slots to get a non-zero offset. r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused int32_t* global = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(4, WASM_SIMD_I32x4_SPLAT(WASM_I32V(23))), WASM_GLOBAL_SET(4, WASM_SIMD_I32x4_REPLACE_LANE( 1, WASM_GLOBAL_GET(4), WASM_I32V(34))), WASM_GLOBAL_SET(4, WASM_SIMD_I32x4_REPLACE_LANE( 2, WASM_GLOBAL_GET(4), WASM_I32V(45))), WASM_GLOBAL_SET(4, WASM_SIMD_I32x4_REPLACE_LANE( 3, WASM_GLOBAL_GET(4), WASM_I32V(56))), WASM_I32V(1)}); CHECK_EQ(1, r.Call(0)); CHECK_EQ(GetScalar(global, 0), 23); CHECK_EQ(GetScalar(global, 1), 34); CHECK_EQ(GetScalar(global, 2), 45); CHECK_EQ(GetScalar(global, 3), 56); } WASM_EXEC_TEST(SimdF32x4GetGlobal) { WasmRunner r(execution_tier); float* global = r.builder().AddGlobal(kWasmS128); SetVectorByLanes(global, {{0.0, 1.5, 2.25, 3.5}}); r.AllocateLocal(kWasmI32); r.Build( {WASM_LOCAL_SET(1, WASM_I32V(1)), WASM_IF(WASM_F32_NE(WASM_F32(0.0), WASM_SIMD_F32x4_EXTRACT_LANE(0, WASM_GLOBAL_GET(0))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(1.5), WASM_SIMD_F32x4_EXTRACT_LANE(1, WASM_GLOBAL_GET(0))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(2.25), WASM_SIMD_F32x4_EXTRACT_LANE(2, WASM_GLOBAL_GET(0))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(3.5), WASM_SIMD_F32x4_EXTRACT_LANE(3, WASM_GLOBAL_GET(0))), WASM_LOCAL_SET(1, WASM_I32V(0))), WASM_LOCAL_GET(1)}); CHECK_EQ(1, r.Call(0)); } WASM_EXEC_TEST(SimdF32x4SetGlobal) { WasmRunner r(execution_tier); float* global = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_SPLAT(WASM_F32(13.5))), WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_REPLACE_LANE( 1, WASM_GLOBAL_GET(0), WASM_F32(45.5))), WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_REPLACE_LANE( 2, WASM_GLOBAL_GET(0), WASM_F32(32.25))), WASM_GLOBAL_SET(0, WASM_SIMD_F32x4_REPLACE_LANE( 3, WASM_GLOBAL_GET(0), WASM_F32(65.0))), WASM_I32V(1)}); CHECK_EQ(1, r.Call(0)); CHECK_EQ(GetScalar(global, 0), 13.5f); CHECK_EQ(GetScalar(global, 1), 45.5f); CHECK_EQ(GetScalar(global, 2), 32.25f); CHECK_EQ(GetScalar(global, 3), 65.0f); } #ifdef V8_ENABLE_WASM_SIMD256_REVEC TEST(RunWasmTurbofan_F32x4AddRevec) { WasmRunner r(TestExecutionTier::kTurbofan); float* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(float)); uint8_t param1 = 0; uint8_t param2 = 1; uint8_t temp1 = r.AllocateLocal(kWasmS128); uint8_t temp2 = r.AllocateLocal(kWasmS128); uint8_t temp3 = r.AllocateLocal(kWasmS128); uint8_t temp4 = r.AllocateLocal(kWasmS128); uint8_t temp5 = r.AllocateLocal(kWasmF32); uint8_t temp6 = r.AllocateLocal(kWasmF32); constexpr uint8_t offset = 16; // Multiple a vector of F32x8 with a constant and store the result to another // array r.Build( {WASM_LOCAL_SET(temp1, WASM_SIMD_F32x4_SPLAT(WASM_F32(10.0f))), WASM_LOCAL_SET(temp2, WASM_SIMD_LOAD_MEM(WASM_LOCAL_GET(param1))), WASM_LOCAL_SET(temp3, WASM_SIMD_BINOP(kExprF32x4Add, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp2))), WASM_LOCAL_SET( temp2, WASM_SIMD_LOAD_MEM_OFFSET(offset, WASM_LOCAL_GET(param1))), WASM_LOCAL_SET(temp4, WASM_SIMD_BINOP(kExprF32x4Add, WASM_LOCAL_GET(temp1), WASM_LOCAL_GET(temp2))), WASM_SIMD_STORE_MEM(WASM_LOCAL_GET(param2), WASM_LOCAL_GET(temp3)), WASM_SIMD_STORE_MEM_OFFSET(offset, WASM_LOCAL_GET(param2), WASM_LOCAL_GET(temp4)), WASM_LOCAL_SET(temp5, WASM_SIMD_F32x4_EXTRACT_LANE( 1, WASM_SIMD_LOAD_MEM(WASM_LOCAL_GET(param2)))), WASM_LOCAL_SET(temp6, WASM_SIMD_F32x4_EXTRACT_LANE( 2, WASM_SIMD_LOAD_MEM_OFFSET( offset, WASM_LOCAL_GET(param2)))), WASM_BINOP(kExprF32Add, WASM_LOCAL_GET(temp5), WASM_LOCAL_GET(temp6))}); r.builder().WriteMemory(&memory[1], 1.0f); r.builder().WriteMemory(&memory[6], 2.0f); CHECK_EQ(23.0f, r.Call(0, 32)); } #endif WASM_EXEC_TEST(SimdLoadStoreLoad) { { WasmRunner r(execution_tier); int32_t* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); // Load memory, store it, then reload it and extract the first lane. Use a // non-zero offset into the memory of 1 lane (4 bytes) to test indexing. r.Build( {WASM_SIMD_STORE_MEM(WASM_I32V(8), WASM_SIMD_LOAD_MEM(WASM_I32V(4))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_SIMD_LOAD_MEM(WASM_I32V(8)))}); FOR_INT32_INPUTS(i) { int32_t expected = i; r.builder().WriteMemory(&memory[1], expected); CHECK_EQ(expected, r.Call()); } } { // OOB tests for loads. WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); r.Build({WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_LOAD_MEM(WASM_LOCAL_GET(0)))}); for (uint32_t offset = kWasmPageSize - (kSimd128Size - 1); offset < kWasmPageSize; ++offset) { CHECK_TRAP(r.Call(offset)); } } { // OOB tests for stores. WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); r.Build( {WASM_SIMD_STORE_MEM(WASM_LOCAL_GET(0), WASM_SIMD_LOAD_MEM(WASM_ZERO)), WASM_ONE}); for (uint32_t offset = kWasmPageSize - (kSimd128Size - 1); offset < kWasmPageSize; ++offset) { CHECK_TRAP(r.Call(offset)); } } } WASM_EXEC_TEST(SimdLoadStoreLoadMemargOffset) { { WasmRunner r(execution_tier); int32_t* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); constexpr uint8_t offset_1 = 4; constexpr uint8_t offset_2 = 8; // Load from memory at offset_1, store to offset_2, load from offset_2, and // extract first lane. We use non-zero memarg offsets to test offset // decoding. r.Build({WASM_SIMD_STORE_MEM_OFFSET( offset_2, WASM_ZERO, WASM_SIMD_LOAD_MEM_OFFSET(offset_1, WASM_ZERO)), WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_LOAD_MEM_OFFSET(offset_2, WASM_ZERO))}); FOR_INT32_INPUTS(i) { int32_t expected = i; // Index 1 of memory (int32_t) will be bytes 4 to 8. r.builder().WriteMemory(&memory[1], expected); CHECK_EQ(expected, r.Call()); } } { // OOB tests for loads with offsets. for (uint32_t offset = kWasmPageSize - (kSimd128Size - 1); offset < kWasmPageSize; ++offset) { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); r.Build({WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_LOAD_MEM_OFFSET(U32V_3(offset), WASM_ZERO))}); CHECK_TRAP(r.Call()); } } { // OOB tests for stores with offsets for (uint32_t offset = kWasmPageSize - (kSimd128Size - 1); offset < kWasmPageSize; ++offset) { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); r.Build({WASM_SIMD_STORE_MEM_OFFSET(U32V_3(offset), WASM_ZERO, WASM_SIMD_LOAD_MEM(WASM_ZERO)), WASM_ONE}); CHECK_TRAP(r.Call(offset)); } } } // Test a multi-byte opcode with offset values that encode into valid opcodes. // This is to exercise decoding logic and make sure we get the lengths right. WASM_EXEC_TEST(S128Load8SplatOffset) { // This offset is [82, 22] when encoded, which contains valid opcodes. constexpr int offset = 4354; WasmRunner r(execution_tier); int8_t* memory = r.builder().AddMemoryElems(kWasmPageSize); int8_t* global = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_LOAD_OP_OFFSET(kExprS128Load8Splat, WASM_I32V(0), U32V_2(offset))), WASM_ONE}); // We don't really care about all valid values, so just test for 1. int8_t x = 7; r.builder().WriteMemory(&memory[offset], x); r.Call(); for (int i = 0; i < 16; i++) { CHECK_EQ(x, LANE(global, i)); } } template void RunLoadSplatTest(TestExecutionTier execution_tier, WasmOpcode op) { constexpr int lanes = 16 / sizeof(T); constexpr int mem_index = 16; // Load from mem index 16 (bytes). { WasmRunner r(execution_tier); T* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); T* global = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP(op, WASM_I32V(mem_index))), WASM_ONE}); for (T x : compiler::ValueHelper::GetVector()) { // 16-th byte in memory is lanes-th element (size T) of memory. r.builder().WriteMemory(&memory[lanes], x); r.Call(); for (int i = 0; i < lanes; i++) { CHECK_EQ(x, LANE(global, i)); } } } // Test for OOB. { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP(op, WASM_LOCAL_GET(0))), WASM_ONE}); // Load splats load sizeof(T) bytes. for (uint32_t offset = kWasmPageSize - (sizeof(T) - 1); offset < kWasmPageSize; ++offset) { CHECK_TRAP(r.Call(offset)); } } } WASM_EXEC_TEST(S128Load8Splat) { RunLoadSplatTest(execution_tier, kExprS128Load8Splat); } WASM_EXEC_TEST(S128Load16Splat) { RunLoadSplatTest(execution_tier, kExprS128Load16Splat); } WASM_EXEC_TEST(S128Load32Splat) { RunLoadSplatTest(execution_tier, kExprS128Load32Splat); } WASM_EXEC_TEST(S128Load64Splat) { RunLoadSplatTest(execution_tier, kExprS128Load64Splat); } template void RunLoadExtendTest(TestExecutionTier execution_tier, WasmOpcode op) { static_assert(sizeof(S) < sizeof(T), "load extend should go from smaller to larger type"); constexpr int lanes_s = 16 / sizeof(S); constexpr int lanes_t = 16 / sizeof(T); constexpr int mem_index = 16; // Load from mem index 16 (bytes). // Load extends always load 64 bits, so alignment values can be from 0 to 3. for (uint8_t alignment = 0; alignment <= 3; alignment++) { WasmRunner r(execution_tier); S* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(S)); T* global = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP_ALIGNMENT( op, WASM_I32V(mem_index), alignment)), WASM_ONE}); for (S x : compiler::ValueHelper::GetVector()) { for (int i = 0; i < lanes_s; i++) { // 16-th byte in memory is lanes-th element (size T) of memory. r.builder().WriteMemory(&memory[lanes_s + i], x); } r.Call(); for (int i = 0; i < lanes_t; i++) { CHECK_EQ(static_cast(x), LANE(global, i)); } } } // Test for offset. { WasmRunner r(execution_tier); S* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(S)); T* global = r.builder().AddGlobal(kWasmS128); constexpr uint8_t offset = sizeof(S); r.Build( {WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP_OFFSET(op, WASM_ZERO, offset)), WASM_ONE}); // Let max_s be the max_s value for type S, we set up the memory as such: // memory = [max_s, max_s - 1, ... max_s - (lane_s - 1)]. constexpr S max_s = std::numeric_limits::max(); for (int i = 0; i < lanes_s; i++) { // Integer promotion due to -, static_cast to narrow. r.builder().WriteMemory(&memory[i], static_cast(max_s - i)); } r.Call(); // Loads will be offset by sizeof(S), so will always start from (max_s - 1). for (int i = 0; i < lanes_t; i++) { // Integer promotion due to -, static_cast to narrow. T expected = static_cast(max_s - i - 1); CHECK_EQ(expected, LANE(global, i)); } } // Test for OOB. { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(S)); r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP(op, WASM_LOCAL_GET(0))), WASM_ONE}); // Load extends load 8 bytes, so should trap from -7. for (uint32_t offset = kWasmPageSize - 7; offset < kWasmPageSize; ++offset) { CHECK_TRAP(r.Call(offset)); } } } WASM_EXEC_TEST(S128Load8x8U) { RunLoadExtendTest(execution_tier, kExprS128Load8x8U); } WASM_EXEC_TEST(S128Load8x8S) { RunLoadExtendTest(execution_tier, kExprS128Load8x8S); } WASM_EXEC_TEST(S128Load16x4U) { RunLoadExtendTest(execution_tier, kExprS128Load16x4U); } WASM_EXEC_TEST(S128Load16x4S) { RunLoadExtendTest(execution_tier, kExprS128Load16x4S); } WASM_EXEC_TEST(S128Load32x2U) { RunLoadExtendTest(execution_tier, kExprS128Load32x2U); } WASM_EXEC_TEST(S128Load32x2S) { RunLoadExtendTest(execution_tier, kExprS128Load32x2S); } template void RunLoadZeroTest(TestExecutionTier execution_tier, WasmOpcode op) { constexpr int lanes_s = kSimd128Size / sizeof(S); constexpr int mem_index = 16; // Load from mem index 16 (bytes). constexpr S sentinel = S{-1}; S* memory; S* global; auto initialize_builder = [=](WasmRunner* r) -> std::tuple { S* memory = r->builder().AddMemoryElems(kWasmPageSize / sizeof(S)); S* global = r->builder().AddGlobal(kWasmS128); r->builder().RandomizeMemory(); r->builder().WriteMemory(&memory[lanes_s], sentinel); return std::make_tuple(memory, global); }; // Check all supported alignments. constexpr int max_alignment = base::bits::CountTrailingZeros(sizeof(S)); for (uint8_t alignment = 0; alignment <= max_alignment; alignment++) { WasmRunner r(execution_tier); std::tie(memory, global) = initialize_builder(&r); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP(op, WASM_I32V(mem_index))), WASM_ONE}); r.Call(); // Only first lane is set to sentinel. CHECK_EQ(sentinel, LANE(global, 0)); // The other lanes are zero. for (int i = 1; i < lanes_s; i++) { CHECK_EQ(S{0}, LANE(global, i)); } } { // Use memarg to specific offset. WasmRunner r(execution_tier); std::tie(memory, global) = initialize_builder(&r); r.Build( {WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP_OFFSET(op, WASM_ZERO, mem_index)), WASM_ONE}); r.Call(); // Only first lane is set to sentinel. CHECK_EQ(sentinel, LANE(global, 0)); // The other lanes are zero. for (int i = 1; i < lanes_s; i++) { CHECK_EQ(S{0}, LANE(global, i)); } } // Test for OOB. { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(S)); r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(0, WASM_SIMD_LOAD_OP(op, WASM_LOCAL_GET(0))), WASM_ONE}); // Load extends load sizeof(S) bytes. for (uint32_t offset = kWasmPageSize - (sizeof(S) - 1); offset < kWasmPageSize; ++offset) { CHECK_TRAP(r.Call(offset)); } } } WASM_EXEC_TEST(S128Load32Zero) { RunLoadZeroTest(execution_tier, kExprS128Load32Zero); } WASM_EXEC_TEST(S128Load64Zero) { RunLoadZeroTest(execution_tier, kExprS128Load64Zero); } template void RunLoadLaneTest(TestExecutionTier execution_tier, WasmOpcode load_op, WasmOpcode splat_op) { uint8_t const_op = static_cast( splat_op == kExprI64x2Splat ? kExprI64Const : kExprI32Const); constexpr uint8_t lanes_s = kSimd128Size / sizeof(T); constexpr int mem_index = 16; // Load from mem index 16 (bytes). constexpr uint8_t splat_value = 33; T sentinel = T{-1}; T* memory; T* global; auto build_fn = [=, &memory, &global](WasmRunner& r, int mem_index, uint8_t lane, uint8_t alignment, uint8_t offset) { memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); global = r.builder().AddGlobal(kWasmS128); r.builder().WriteMemory(&memory[lanes_s], sentinel); // Splat splat_value, then only load and replace a single lane with the // sentinel value. r.Build({WASM_I32V(mem_index), const_op, splat_value, WASM_SIMD_OP(splat_op), WASM_SIMD_OP(load_op), alignment, offset, lane, kExprGlobalSet, 0, WASM_ONE}); }; auto check_results = [=](T* global, int sentinel_lane = 0) { // Only one lane is loaded, the rest of the lanes are unchanged. for (uint8_t i = 0; i < lanes_s; i++) { T expected = i == sentinel_lane ? sentinel : static_cast(splat_value); CHECK_EQ(expected, LANE(global, i)); } }; for (uint8_t lane_index = 0; lane_index < lanes_s; ++lane_index) { WasmRunner r(execution_tier); build_fn(r, mem_index, lane_index, /*alignment=*/0, /*offset=*/0); r.Call(); check_results(global, lane_index); } // Check all possible alignments. constexpr int max_alignment = base::bits::CountTrailingZeros(sizeof(T)); for (uint8_t alignment = 0; alignment <= max_alignment; ++alignment) { WasmRunner r(execution_tier); build_fn(r, mem_index, /*lane=*/0, alignment, /*offset=*/0); r.Call(); check_results(global); } { // Use memarg to specify offset. int lane_index = 0; WasmRunner r(execution_tier); build_fn(r, /*mem_index=*/0, /*lane=*/0, /*alignment=*/0, /*offset=*/mem_index); r.Call(); check_results(global, lane_index); } // Test for OOB. { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); r.builder().AddGlobal(kWasmS128); r.Build({WASM_LOCAL_GET(0), const_op, splat_value, WASM_SIMD_OP(splat_op), WASM_SIMD_OP(load_op), ZERO_ALIGNMENT, ZERO_OFFSET, 0, kExprGlobalSet, 0, WASM_ONE}); // Load lane load sizeof(T) bytes. for (uint32_t index = kWasmPageSize - (sizeof(T) - 1); index < kWasmPageSize; ++index) { CHECK_TRAP(r.Call(index)); } } } WASM_EXEC_TEST(S128Load8Lane) { RunLoadLaneTest(execution_tier, kExprS128Load8Lane, kExprI8x16Splat); } WASM_EXEC_TEST(S128Load16Lane) { RunLoadLaneTest(execution_tier, kExprS128Load16Lane, kExprI16x8Splat); } WASM_EXEC_TEST(S128Load32Lane) { RunLoadLaneTest(execution_tier, kExprS128Load32Lane, kExprI32x4Splat); } WASM_EXEC_TEST(S128Load64Lane) { RunLoadLaneTest(execution_tier, kExprS128Load64Lane, kExprI64x2Splat); } template void RunStoreLaneTest(TestExecutionTier execution_tier, WasmOpcode store_op, WasmOpcode splat_op) { constexpr uint8_t lanes = kSimd128Size / sizeof(T); constexpr int mem_index = 16; // Store to mem index 16 (bytes). constexpr uint8_t splat_value = 33; uint8_t const_op = static_cast( splat_op == kExprI64x2Splat ? kExprI64Const : kExprI32Const); T* memory; // Will be set by build_fn. auto build_fn = [=, &memory](WasmRunner& r, int mem_index, uint8_t lane_index, uint8_t alignment, uint8_t offset) { memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); // Splat splat_value, then only Store and replace a single lane. r.Build({WASM_I32V(mem_index), const_op, splat_value, WASM_SIMD_OP(splat_op), WASM_SIMD_OP(store_op), alignment, offset, lane_index, WASM_ONE}); r.builder().BlankMemory(); }; auto check_results = [=](WasmRunner& r, T* memory) { for (uint8_t i = 0; i < lanes; i++) { CHECK_EQ(0, r.builder().ReadMemory(&memory[i])); } CHECK_EQ(splat_value, r.builder().ReadMemory(&memory[lanes])); for (uint8_t i = lanes + 1; i < lanes * 2; i++) { CHECK_EQ(0, r.builder().ReadMemory(&memory[i])); } }; for (uint8_t lane_index = 0; lane_index < lanes; lane_index++) { WasmRunner r(execution_tier); build_fn(r, mem_index, lane_index, ZERO_ALIGNMENT, ZERO_OFFSET); r.Call(); check_results(r, memory); } // Check all possible alignments. constexpr int max_alignment = base::bits::CountTrailingZeros(sizeof(T)); for (uint8_t alignment = 0; alignment <= max_alignment; ++alignment) { WasmRunner r(execution_tier); build_fn(r, mem_index, /*lane_index=*/0, alignment, ZERO_OFFSET); r.Call(); check_results(r, memory); } { // Use memarg for offset. WasmRunner r(execution_tier); build_fn(r, /*mem_index=*/0, /*lane_index=*/0, ZERO_ALIGNMENT, mem_index); r.Call(); check_results(r, memory); } // OOB stores { WasmRunner r(execution_tier); r.builder().AddMemoryElems(kWasmPageSize / sizeof(T)); r.Build({WASM_LOCAL_GET(0), const_op, splat_value, WASM_SIMD_OP(splat_op), WASM_SIMD_OP(store_op), ZERO_ALIGNMENT, ZERO_OFFSET, 0, WASM_ONE}); // StoreLane stores sizeof(T) bytes. for (uint32_t index = kWasmPageSize - (sizeof(T) - 1); index < kWasmPageSize; ++index) { CHECK_TRAP(r.Call(index)); } } } WASM_EXEC_TEST(S128Store8Lane) { RunStoreLaneTest(execution_tier, kExprS128Store8Lane, kExprI8x16Splat); } WASM_EXEC_TEST(S128Store16Lane) { RunStoreLaneTest(execution_tier, kExprS128Store16Lane, kExprI16x8Splat); } WASM_EXEC_TEST(S128Store32Lane) { RunStoreLaneTest(execution_tier, kExprS128Store32Lane, kExprI32x4Splat); } WASM_EXEC_TEST(S128Store64Lane) { RunStoreLaneTest(execution_tier, kExprS128Store64Lane, kExprI64x2Splat); } #define WASM_SIMD_ANYTRUE_TEST(format, lanes, max, param_type) \ WASM_EXEC_TEST(S##format##AnyTrue) { \ WasmRunner r(execution_tier); \ if (lanes == 2) return; \ uint8_t simd = r.AllocateLocal(kWasmS128); \ r.Build( \ {WASM_LOCAL_SET(simd, WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(0))), \ WASM_SIMD_UNOP(kExprV128AnyTrue, WASM_LOCAL_GET(simd))}); \ CHECK_EQ(1, r.Call(max)); \ CHECK_EQ(1, r.Call(5)); \ CHECK_EQ(0, r.Call(0)); \ } WASM_SIMD_ANYTRUE_TEST(32x4, 4, 0xffffffff, int32_t) WASM_SIMD_ANYTRUE_TEST(16x8, 8, 0xffff, int32_t) WASM_SIMD_ANYTRUE_TEST(8x16, 16, 0xff, int32_t) // Special any true test cases that splats a -0.0 double into a i64x2. // This is specifically to ensure that our implementation correct handles that // 0.0 and -0.0 will be different in an anytrue (IEEE753 says they are equals). WASM_EXEC_TEST(V128AnytrueWithNegativeZero) { WasmRunner r(execution_tier); uint8_t simd = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(simd, WASM_SIMD_I64x2_SPLAT(WASM_LOCAL_GET(0))), WASM_SIMD_UNOP(kExprV128AnyTrue, WASM_LOCAL_GET(simd))}); CHECK_EQ(1, r.Call(0x8000000000000000)); CHECK_EQ(0, r.Call(0x0000000000000000)); } #define WASM_SIMD_ALLTRUE_TEST(format, lanes, max, param_type) \ WASM_EXEC_TEST(I##format##AllTrue) { \ WasmRunner r(execution_tier); \ if (lanes == 2) return; \ uint8_t simd = r.AllocateLocal(kWasmS128); \ r.Build( \ {WASM_LOCAL_SET(simd, WASM_SIMD_I##format##_SPLAT(WASM_LOCAL_GET(0))), \ WASM_SIMD_UNOP(kExprI##format##AllTrue, WASM_LOCAL_GET(simd))}); \ CHECK_EQ(1, r.Call(max)); \ CHECK_EQ(1, r.Call(0x1)); \ CHECK_EQ(0, r.Call(0)); \ } WASM_SIMD_ALLTRUE_TEST(64x2, 2, 0xffffffffffffffff, int64_t) WASM_SIMD_ALLTRUE_TEST(32x4, 4, 0xffffffff, int32_t) WASM_SIMD_ALLTRUE_TEST(16x8, 8, 0xffff, int32_t) WASM_SIMD_ALLTRUE_TEST(8x16, 16, 0xff, int32_t) WASM_EXEC_TEST(BitSelect) { WasmRunner r(execution_tier); uint8_t simd = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET( simd, WASM_SIMD_SELECT( 32x4, WASM_SIMD_I32x4_SPLAT(WASM_I32V(0x01020304)), WASM_SIMD_I32x4_SPLAT(WASM_I32V(0)), WASM_SIMD_I32x4_SPLAT(WASM_LOCAL_GET(0)))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_LOCAL_GET(simd))}); CHECK_EQ(0x01020304, r.Call(0xFFFFFFFF)); } void RunSimdConstTest(TestExecutionTier execution_tier, const std::array& expected) { WasmRunner r(execution_tier); uint8_t temp1 = r.AllocateLocal(kWasmS128); uint8_t* src0 = r.builder().AddGlobal(kWasmS128); r.Build({WASM_GLOBAL_SET(temp1, WASM_SIMD_CONSTANT(expected)), WASM_ONE}); CHECK_EQ(1, r.Call()); for (size_t i = 0; i < expected.size(); i++) { CHECK_EQ(LANE(src0, i), expected[i]); } } WASM_EXEC_TEST(S128Const) { std::array expected; // Test for generic constant for (int i = 0; i < kSimd128Size; i++) { expected[i] = i; } RunSimdConstTest(execution_tier, expected); // Keep the first 4 lanes as 0, set the remaining ones. for (int i = 0; i < 4; i++) { expected[i] = 0; } for (int i = 4; i < kSimd128Size; i++) { expected[i] = i; } RunSimdConstTest(execution_tier, expected); // Check sign extension logic used to pack int32s into int64. expected = {0}; // Set the top bit of lane 3 (top bit of first int32), the rest can be 0. expected[3] = 0x80; RunSimdConstTest(execution_tier, expected); } WASM_EXEC_TEST(S128ConstAllZero) { std::array expected = {0}; RunSimdConstTest(execution_tier, expected); } WASM_EXEC_TEST(S128ConstAllOnes) { std::array expected; // Test for generic constant for (int i = 0; i < kSimd128Size; i++) { expected[i] = 0xff; } RunSimdConstTest(execution_tier, expected); } WASM_EXEC_TEST(I8x16LeUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, kExprI8x16LeU, UnsignedLessEqual); } WASM_EXEC_TEST(I8x16LtUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, kExprI8x16LtU, UnsignedLess); } WASM_EXEC_TEST(I8x16GeUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, kExprI8x16GeU, UnsignedGreaterEqual); } WASM_EXEC_TEST(I8x16GtUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, kExprI8x16GtU, UnsignedGreater); } WASM_EXEC_TEST(I16x8LeUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, kExprI16x8LeU, UnsignedLessEqual); } WASM_EXEC_TEST(I16x8LtUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, kExprI16x8LtU, UnsignedLess); } WASM_EXEC_TEST(I16x8GeUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, kExprI16x8GeU, UnsignedGreaterEqual); } WASM_EXEC_TEST(I16x8GtUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, kExprI16x8GtU, UnsignedGreater); } WASM_EXEC_TEST(I16x8ExtractLaneU_I8x16Splat) { // Test that we are correctly signed/unsigned extending when extracting. WasmRunner r(execution_tier); uint8_t simd_val = r.AllocateLocal(kWasmS128); r.Build({WASM_LOCAL_SET(simd_val, WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(0))), WASM_SIMD_I16x8_EXTRACT_LANE_U(0, WASM_LOCAL_GET(simd_val))}); CHECK_EQ(0xfafa, r.Call(0xfa)); } enum ExtAddSide { LEFT, RIGHT }; template void RunAddExtAddPairwiseTest( TestExecutionTier execution_tier, ExtAddSide extAddSide, WasmOpcode addOpcode, const std::array addInput, WasmOpcode extAddOpcode, const std::array extAddInput, const std::array expectedOutput) { WasmRunner r(execution_tier); T* x = r.builder().AddGlobal(kWasmS128); for (size_t i = 0; i < addInput.size(); i++) { LANE(x, i) = addInput[i]; } U* y = r.builder().AddGlobal(kWasmS128); for (size_t i = 0; i < extAddInput.size(); i++) { LANE(y, i) = extAddInput[i]; } switch (extAddSide) { case LEFT: // x = add(extadd_pairwise_s(y), x) r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP( addOpcode, WASM_SIMD_UNOP(extAddOpcode, WASM_GLOBAL_GET(1)), WASM_GLOBAL_GET(0))), WASM_ONE}); break; case RIGHT: // x = add(x, extadd_pairwise_s(y)) r.Build({WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP( addOpcode, WASM_GLOBAL_GET(0), WASM_SIMD_UNOP(extAddOpcode, WASM_GLOBAL_GET(1)))), WASM_ONE}); break; } r.Call(); for (size_t i = 0; i < expectedOutput.size(); i++) { CHECK_EQ(expectedOutput[i], LANE(x, i)); } } WASM_EXEC_TEST(AddExtAddPairwiseI32Right) { RunAddExtAddPairwiseTest( execution_tier, RIGHT, kExprI32x4Add, {1, 2, 3, 4}, kExprI32x4ExtAddPairwiseI16x8S, {-1, -2, -3, -4, -5, -6, -7, -8}, {-2, -5, -8, -11}); } WASM_EXEC_TEST(AddExtAddPairwiseI32Left) { RunAddExtAddPairwiseTest( execution_tier, LEFT, kExprI32x4Add, {1, 2, 3, 4}, kExprI32x4ExtAddPairwiseI16x8S, {-1, -2, -3, -4, -5, -6, -7, -8}, {-2, -5, -8, -11}); } WASM_EXEC_TEST(AddExtAddPairwiseI16Right) { RunAddExtAddPairwiseTest( execution_tier, RIGHT, kExprI16x8Add, {1, 2, 3, 4, 5, 6, 7, 8}, kExprI16x8ExtAddPairwiseI8x16S, {-1, -2, -3, -4, -5, -6, -7, -8, -9, -10, -11, -12, -13, -14, -15, -16}, {-2, -5, -8, -11, -14, -17, -20, -23}); } WASM_EXEC_TEST(AddExtAddPairwiseI16Left) { RunAddExtAddPairwiseTest( execution_tier, LEFT, kExprI16x8Add, {1, 2, 3, 4, 5, 6, 7, 8}, kExprI16x8ExtAddPairwiseI8x16S, {1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16}, {4, 9, 14, 19, 24, 29, 34, 39}); } WASM_EXEC_TEST(AddExtAddPairwiseI32RightUnsigned) { RunAddExtAddPairwiseTest( execution_tier, RIGHT, kExprI32x4Add, {1, 2, 3, 4}, kExprI32x4ExtAddPairwiseI16x8U, {1, 2, 3, 4, 5, 6, 7, 8}, {4, 9, 14, 19}); } WASM_EXEC_TEST(AddExtAddPairwiseI32LeftUnsigned) { RunAddExtAddPairwiseTest( execution_tier, LEFT, kExprI32x4Add, {1, 2, 3, 4}, kExprI32x4ExtAddPairwiseI16x8U, {1, 2, 3, 4, 5, 6, 7, 8}, {4, 9, 14, 19}); } // Regression test from https://crbug.com/v8/12237 to exercise a codegen bug // for i64x2.gts which overwrote one of the inputs. WASM_EXEC_TEST(Regress_12237) { WasmRunner r(execution_tier); int64_t* g = r.builder().AddGlobal(kWasmS128); uint8_t value = 0; uint8_t temp = r.AllocateLocal(kWasmS128); int64_t local = 123; r.Build({WASM_LOCAL_SET( temp, WASM_SIMD_OPN(kExprI64x2Splat, WASM_LOCAL_GET(value))), WASM_GLOBAL_SET( 0, WASM_SIMD_BINOP( kExprI64x2GtS, WASM_LOCAL_GET(temp), WASM_SIMD_BINOP(kExprI64x2Sub, WASM_LOCAL_GET(temp), WASM_LOCAL_GET(temp)))), WASM_ONE}); r.Call(local); int64_t expected = Greater(local, local - local); for (size_t i = 0; i < kSimd128Size / sizeof(int64_t); i++) { CHECK_EQ(expected, LANE(g, 0)); } } #define WASM_EXTRACT_I16x8_TEST(Sign, Type) \ WASM_EXEC_TEST(I16X8ExtractLane##Sign) { \ WasmRunner r(execution_tier); \ uint8_t int_val = r.AllocateLocal(kWasmI32); \ uint8_t simd_val = r.AllocateLocal(kWasmS128); \ r.Build({WASM_LOCAL_SET(simd_val, \ WASM_SIMD_I16x8_SPLAT(WASM_LOCAL_GET(int_val))), \ WASM_SIMD_CHECK_LANE_U(I16x8, simd_val, I32, int_val, 0), \ WASM_SIMD_CHECK_LANE_U(I16x8, simd_val, I32, int_val, 2), \ WASM_SIMD_CHECK_LANE_U(I16x8, simd_val, I32, int_val, 4), \ WASM_SIMD_CHECK_LANE_U(I16x8, simd_val, I32, int_val, 6), \ WASM_ONE}); \ FOR_##Type##_INPUTS(x) { CHECK_EQ(1, r.Call(x)); } \ } WASM_EXTRACT_I16x8_TEST(S, UINT16) WASM_EXTRACT_I16x8_TEST(I, INT16) #undef WASM_EXTRACT_I16x8_TEST #define WASM_EXTRACT_I8x16_TEST(Sign, Type) \ WASM_EXEC_TEST(I8x16ExtractLane##Sign) { \ WasmRunner r(execution_tier); \ uint8_t int_val = r.AllocateLocal(kWasmI32); \ uint8_t simd_val = r.AllocateLocal(kWasmS128); \ r.Build({WASM_LOCAL_SET(simd_val, \ WASM_SIMD_I8x16_SPLAT(WASM_LOCAL_GET(int_val))), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 1), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 3), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 5), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 7), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 9), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 10), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 11), \ WASM_SIMD_CHECK_LANE_U(I8x16, simd_val, I32, int_val, 13), \ WASM_ONE}); \ FOR_##Type##_INPUTS(x) { CHECK_EQ(1, r.Call(x)); } \ } WASM_EXTRACT_I8x16_TEST(S, UINT8) WASM_EXTRACT_I8x16_TEST(I, INT8) #undef WASM_EXTRACT_I8x16_TEST #undef WASM_SIMD_CHECK_LANE_S #undef WASM_SIMD_CHECK_LANE_U #undef TO_BYTE #undef WASM_SIMD_OP #undef WASM_SIMD_SPLAT #undef WASM_SIMD_UNOP #undef WASM_SIMD_BINOP #undef WASM_SIMD_SHIFT_OP #undef WASM_SIMD_CONCAT_OP #undef WASM_SIMD_SELECT #undef WASM_SIMD_F64x2_SPLAT #undef WASM_SIMD_F64x2_EXTRACT_LANE #undef WASM_SIMD_F64x2_REPLACE_LANE #undef WASM_SIMD_F32x4_SPLAT #undef WASM_SIMD_F32x4_EXTRACT_LANE #undef WASM_SIMD_F32x4_REPLACE_LANE #undef WASM_SIMD_I64x2_SPLAT #undef WASM_SIMD_I64x2_EXTRACT_LANE #undef WASM_SIMD_I64x2_REPLACE_LANE #undef WASM_SIMD_I32x4_SPLAT #undef WASM_SIMD_I32x4_EXTRACT_LANE #undef WASM_SIMD_I32x4_REPLACE_LANE #undef WASM_SIMD_I16x8_SPLAT #undef WASM_SIMD_I16x8_EXTRACT_LANE #undef WASM_SIMD_I16x8_EXTRACT_LANE_U #undef WASM_SIMD_I16x8_REPLACE_LANE #undef WASM_SIMD_I8x16_SPLAT #undef WASM_SIMD_I8x16_EXTRACT_LANE #undef WASM_SIMD_I8x16_EXTRACT_LANE_U #undef WASM_SIMD_I8x16_REPLACE_LANE #undef WASM_SIMD_I8x16_SHUFFLE_OP #undef WASM_SIMD_LOAD_MEM #undef WASM_SIMD_LOAD_MEM_OFFSET #undef WASM_SIMD_STORE_MEM #undef WASM_SIMD_STORE_MEM_OFFSET #undef WASM_SIMD_SELECT_TEST #undef WASM_SIMD_NON_CANONICAL_SELECT_TEST #undef WASM_SIMD_BOOL_REDUCTION_TEST #undef WASM_SIMD_ANYTRUE_TEST #undef WASM_SIMD_ALLTRUE_TEST #undef WASM_SIMD_F64x2_QFMA #undef WASM_SIMD_F64x2_QFMS #undef WASM_SIMD_F32x4_QFMA #undef WASM_SIMD_F32x4_QFMS #undef WASM_SIMD_LOAD_OP #undef WASM_SIMD_LOAD_OP_OFFSET #undef WASM_SIMD_LOAD_OP_ALIGNMENT } // namespace test_run_wasm_simd } // namespace wasm } // namespace internal } // namespace v8