aboutsummaryrefslogtreecommitdiffstats
path: root/contrib/libs/hyperscan/src/rose/validate_shufti.h
diff options
context:
space:
mode:
authorIvan Blinkov <ivan@blinkov.ru>2022-02-10 16:47:11 +0300
committerDaniil Cherednik <dcherednik@yandex-team.ru>2022-02-10 16:47:11 +0300
commit5b283123c882433dafbaf6b338adeea16c1a0ea0 (patch)
tree339adc63bce23800021202ae4a8328a843dc447a /contrib/libs/hyperscan/src/rose/validate_shufti.h
parent1aeb9a455974457866f78722ad98114bafc84e8a (diff)
downloadydb-5b283123c882433dafbaf6b338adeea16c1a0ea0.tar.gz
Restoring authorship annotation for Ivan Blinkov <ivan@blinkov.ru>. Commit 2 of 2.
Diffstat (limited to 'contrib/libs/hyperscan/src/rose/validate_shufti.h')
-rw-r--r--contrib/libs/hyperscan/src/rose/validate_shufti.h586
1 files changed, 293 insertions, 293 deletions
diff --git a/contrib/libs/hyperscan/src/rose/validate_shufti.h b/contrib/libs/hyperscan/src/rose/validate_shufti.h
index f936140b11..351df36a76 100644
--- a/contrib/libs/hyperscan/src/rose/validate_shufti.h
+++ b/contrib/libs/hyperscan/src/rose/validate_shufti.h
@@ -1,182 +1,182 @@
-/*
+/*
* Copyright (c) 2016-2020, Intel Corporation
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions are met:
- *
- * * Redistributions of source code must retain the above copyright notice,
- * this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- * * Neither the name of Intel Corporation nor the names of its contributors
- * may be used to endorse or promote products derived from this software
- * without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- */
-
-#ifndef VALIDATE_SHUFTI_H
-#define VALIDATE_SHUFTI_H
-
-#include "ue2common.h"
-#include "util/simd_utils.h"
-
-#if defined(DEBUG)
-static
-void dumpMask(const void *mask, int len) {
- const u8 *c = (const u8 *)mask;
- for (int i = 0; i < len; i++) {
- printf("%02x", c[i]);
- }
- printf("\n");
-}
-#endif
-
-static really_inline
-int validateShuftiMask16x16(const m256 data, const m256 hi_mask,
- const m256 lo_mask, const m256 and_mask,
- const u32 neg_mask, const u32 valid_data_mask) {
- m256 low4bits = set32x8(0xf);
- m256 c_lo = pshufb_m256(lo_mask, and256(data, low4bits));
- m256 c_hi = pshufb_m256(hi_mask,
- rshift64_m256(andnot256(low4bits, data), 4));
- m256 t = and256(c_lo, c_hi);
- u32 nresult = movemask256(eq256(and256(t, and_mask), zeroes256()));
-#ifdef DEBUG
- DEBUG_PRINTF("data\n");
- dumpMask(&data, 32);
- DEBUG_PRINTF("hi_mask\n");
- dumpMask(&hi_mask, 32);
- DEBUG_PRINTF("lo_mask\n");
- dumpMask(&lo_mask, 32);
- DEBUG_PRINTF("c_lo\n");
- dumpMask(&c_lo, 32);
- DEBUG_PRINTF("c_hi\n");
- dumpMask(&c_hi, 32);
- DEBUG_PRINTF("and_mask\n");
- dumpMask(&and_mask, 32);
- DEBUG_PRINTF("nresult %x\n", nresult);
- DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
-#endif
- u32 cmp_result = (((nresult >> 16) & nresult) ^ neg_mask) & valid_data_mask;
- return !cmp_result;
-}
-
-static really_inline
-int validateShuftiMask16x8(const m128 data, const m256 nib_mask,
- const m128 and_mask, const u32 neg_mask,
- const u32 valid_data_mask) {
- m256 data_m256 = combine2x128(rshift64_m128(data, 4), data);
- m256 low4bits = set32x8(0xf);
- m256 c_nib = pshufb_m256(nib_mask, and256(data_m256, low4bits));
- m128 t = and128(movdq_hi(c_nib), movdq_lo(c_nib));
- m128 nresult = eq128(and128(t, and_mask), zeroes128());
-#ifdef DEBUG
- DEBUG_PRINTF("data\n");
- dumpMask(&data_m256, 32);
- DEBUG_PRINTF("nib_mask\n");
- dumpMask(&nib_mask, 32);
- DEBUG_PRINTF("c_nib\n");
- dumpMask(&c_nib, 32);
- DEBUG_PRINTF("nresult\n");
- dumpMask(&nresult, 16);
- DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
-#endif
- u32 cmp_result = (movemask128(nresult) ^ neg_mask) & valid_data_mask;
- return !cmp_result;
-}
-
-static really_inline
-int validateShuftiMask32x8(const m256 data, const m256 hi_mask,
- const m256 lo_mask, const m256 and_mask,
- const u32 neg_mask, const u32 valid_data_mask) {
- m256 low4bits = set32x8(0xf);
- m256 c_lo = pshufb_m256(lo_mask, and256(data, low4bits));
- m256 c_hi = pshufb_m256(hi_mask,
- rshift64_m256(andnot256(low4bits, data), 4));
- m256 t = and256(c_lo, c_hi);
- m256 nresult = eq256(and256(t, and_mask), zeroes256());
-#ifdef DEBUG
- DEBUG_PRINTF("data\n");
- dumpMask(&data, 32);
- DEBUG_PRINTF("hi_mask\n");
- dumpMask(&hi_mask, 32);
- DEBUG_PRINTF("lo_mask\n");
- dumpMask(&lo_mask, 32);
- DEBUG_PRINTF("c_lo\n");
- dumpMask(&c_lo, 32);
- DEBUG_PRINTF("c_hi\n");
- dumpMask(&c_hi, 32);
- DEBUG_PRINTF("nresult\n");
- dumpMask(&nresult, 32);
- DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
-#endif
- u32 cmp_result = (movemask256(nresult) ^ neg_mask) & valid_data_mask;
- return !cmp_result;
-}
-
-static really_inline
-int validateShuftiMask32x16(const m256 data,
- const m256 hi_mask_1, const m256 hi_mask_2,
- const m256 lo_mask_1, const m256 lo_mask_2,
- const m256 bucket_mask_hi,
- const m256 bucket_mask_lo, const u32 neg_mask,
- const u32 valid_data_mask) {
- m256 low4bits = set32x8(0xf);
- m256 data_lo = and256(data, low4bits);
- m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
- m256 c_lo_1 = pshufb_m256(lo_mask_1, data_lo);
- m256 c_lo_2 = pshufb_m256(lo_mask_2, data_lo);
- m256 c_hi_1 = pshufb_m256(hi_mask_1, data_hi);
- m256 c_hi_2 = pshufb_m256(hi_mask_2, data_hi);
- m256 t1 = and256(c_lo_1, c_hi_1);
- m256 t2 = and256(c_lo_2, c_hi_2);
- m256 result = or256(and256(t1, bucket_mask_lo), and256(t2, bucket_mask_hi));
- u32 nresult = movemask256(eq256(result, zeroes256()));
-#ifdef DEBUG
- DEBUG_PRINTF("data\n");
- dumpMask(&data, 32);
- DEBUG_PRINTF("data_lo\n");
- dumpMask(&data_lo, 32);
- DEBUG_PRINTF("data_hi\n");
- dumpMask(&data_hi, 32);
- DEBUG_PRINTF("hi_mask_1\n");
- dumpMask(&hi_mask_1, 16);
- DEBUG_PRINTF("hi_mask_2\n");
- dumpMask(&hi_mask_2, 16);
- DEBUG_PRINTF("lo_mask_1\n");
- dumpMask(&lo_mask_1, 16);
- DEBUG_PRINTF("lo_mask_2\n");
- dumpMask(&lo_mask_2, 16);
- DEBUG_PRINTF("c_lo_1\n");
- dumpMask(&c_lo_1, 32);
- DEBUG_PRINTF("c_lo_2\n");
- dumpMask(&c_lo_2, 32);
- DEBUG_PRINTF("c_hi_1\n");
- dumpMask(&c_hi_1, 32);
- DEBUG_PRINTF("c_hi_2\n");
- dumpMask(&c_hi_2, 32);
- DEBUG_PRINTF("result\n");
- dumpMask(&result, 32);
- DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
-#endif
- u32 cmp_result = (nresult ^ neg_mask) & valid_data_mask;
- return !cmp_result;
-}
-
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are met:
+ *
+ * * Redistributions of source code must retain the above copyright notice,
+ * this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * * Neither the name of Intel Corporation nor the names of its contributors
+ * may be used to endorse or promote products derived from this software
+ * without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#ifndef VALIDATE_SHUFTI_H
+#define VALIDATE_SHUFTI_H
+
+#include "ue2common.h"
+#include "util/simd_utils.h"
+
+#if defined(DEBUG)
+static
+void dumpMask(const void *mask, int len) {
+ const u8 *c = (const u8 *)mask;
+ for (int i = 0; i < len; i++) {
+ printf("%02x", c[i]);
+ }
+ printf("\n");
+}
+#endif
+
+static really_inline
+int validateShuftiMask16x16(const m256 data, const m256 hi_mask,
+ const m256 lo_mask, const m256 and_mask,
+ const u32 neg_mask, const u32 valid_data_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 c_lo = pshufb_m256(lo_mask, and256(data, low4bits));
+ m256 c_hi = pshufb_m256(hi_mask,
+ rshift64_m256(andnot256(low4bits, data), 4));
+ m256 t = and256(c_lo, c_hi);
+ u32 nresult = movemask256(eq256(and256(t, and_mask), zeroes256()));
+#ifdef DEBUG
+ DEBUG_PRINTF("data\n");
+ dumpMask(&data, 32);
+ DEBUG_PRINTF("hi_mask\n");
+ dumpMask(&hi_mask, 32);
+ DEBUG_PRINTF("lo_mask\n");
+ dumpMask(&lo_mask, 32);
+ DEBUG_PRINTF("c_lo\n");
+ dumpMask(&c_lo, 32);
+ DEBUG_PRINTF("c_hi\n");
+ dumpMask(&c_hi, 32);
+ DEBUG_PRINTF("and_mask\n");
+ dumpMask(&and_mask, 32);
+ DEBUG_PRINTF("nresult %x\n", nresult);
+ DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
+#endif
+ u32 cmp_result = (((nresult >> 16) & nresult) ^ neg_mask) & valid_data_mask;
+ return !cmp_result;
+}
+
+static really_inline
+int validateShuftiMask16x8(const m128 data, const m256 nib_mask,
+ const m128 and_mask, const u32 neg_mask,
+ const u32 valid_data_mask) {
+ m256 data_m256 = combine2x128(rshift64_m128(data, 4), data);
+ m256 low4bits = set32x8(0xf);
+ m256 c_nib = pshufb_m256(nib_mask, and256(data_m256, low4bits));
+ m128 t = and128(movdq_hi(c_nib), movdq_lo(c_nib));
+ m128 nresult = eq128(and128(t, and_mask), zeroes128());
+#ifdef DEBUG
+ DEBUG_PRINTF("data\n");
+ dumpMask(&data_m256, 32);
+ DEBUG_PRINTF("nib_mask\n");
+ dumpMask(&nib_mask, 32);
+ DEBUG_PRINTF("c_nib\n");
+ dumpMask(&c_nib, 32);
+ DEBUG_PRINTF("nresult\n");
+ dumpMask(&nresult, 16);
+ DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
+#endif
+ u32 cmp_result = (movemask128(nresult) ^ neg_mask) & valid_data_mask;
+ return !cmp_result;
+}
+
+static really_inline
+int validateShuftiMask32x8(const m256 data, const m256 hi_mask,
+ const m256 lo_mask, const m256 and_mask,
+ const u32 neg_mask, const u32 valid_data_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 c_lo = pshufb_m256(lo_mask, and256(data, low4bits));
+ m256 c_hi = pshufb_m256(hi_mask,
+ rshift64_m256(andnot256(low4bits, data), 4));
+ m256 t = and256(c_lo, c_hi);
+ m256 nresult = eq256(and256(t, and_mask), zeroes256());
+#ifdef DEBUG
+ DEBUG_PRINTF("data\n");
+ dumpMask(&data, 32);
+ DEBUG_PRINTF("hi_mask\n");
+ dumpMask(&hi_mask, 32);
+ DEBUG_PRINTF("lo_mask\n");
+ dumpMask(&lo_mask, 32);
+ DEBUG_PRINTF("c_lo\n");
+ dumpMask(&c_lo, 32);
+ DEBUG_PRINTF("c_hi\n");
+ dumpMask(&c_hi, 32);
+ DEBUG_PRINTF("nresult\n");
+ dumpMask(&nresult, 32);
+ DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
+#endif
+ u32 cmp_result = (movemask256(nresult) ^ neg_mask) & valid_data_mask;
+ return !cmp_result;
+}
+
+static really_inline
+int validateShuftiMask32x16(const m256 data,
+ const m256 hi_mask_1, const m256 hi_mask_2,
+ const m256 lo_mask_1, const m256 lo_mask_2,
+ const m256 bucket_mask_hi,
+ const m256 bucket_mask_lo, const u32 neg_mask,
+ const u32 valid_data_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 data_lo = and256(data, low4bits);
+ m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
+ m256 c_lo_1 = pshufb_m256(lo_mask_1, data_lo);
+ m256 c_lo_2 = pshufb_m256(lo_mask_2, data_lo);
+ m256 c_hi_1 = pshufb_m256(hi_mask_1, data_hi);
+ m256 c_hi_2 = pshufb_m256(hi_mask_2, data_hi);
+ m256 t1 = and256(c_lo_1, c_hi_1);
+ m256 t2 = and256(c_lo_2, c_hi_2);
+ m256 result = or256(and256(t1, bucket_mask_lo), and256(t2, bucket_mask_hi));
+ u32 nresult = movemask256(eq256(result, zeroes256()));
+#ifdef DEBUG
+ DEBUG_PRINTF("data\n");
+ dumpMask(&data, 32);
+ DEBUG_PRINTF("data_lo\n");
+ dumpMask(&data_lo, 32);
+ DEBUG_PRINTF("data_hi\n");
+ dumpMask(&data_hi, 32);
+ DEBUG_PRINTF("hi_mask_1\n");
+ dumpMask(&hi_mask_1, 16);
+ DEBUG_PRINTF("hi_mask_2\n");
+ dumpMask(&hi_mask_2, 16);
+ DEBUG_PRINTF("lo_mask_1\n");
+ dumpMask(&lo_mask_1, 16);
+ DEBUG_PRINTF("lo_mask_2\n");
+ dumpMask(&lo_mask_2, 16);
+ DEBUG_PRINTF("c_lo_1\n");
+ dumpMask(&c_lo_1, 32);
+ DEBUG_PRINTF("c_lo_2\n");
+ dumpMask(&c_lo_2, 32);
+ DEBUG_PRINTF("c_hi_1\n");
+ dumpMask(&c_hi_1, 32);
+ DEBUG_PRINTF("c_hi_2\n");
+ dumpMask(&c_hi_2, 32);
+ DEBUG_PRINTF("result\n");
+ dumpMask(&result, 32);
+ DEBUG_PRINTF("valid_data_mask %x\n", valid_data_mask);
+#endif
+ u32 cmp_result = (nresult ^ neg_mask) & valid_data_mask;
+ return !cmp_result;
+}
+
#ifdef HAVE_AVX512
-static really_inline
+static really_inline
int validateShuftiMask64x8(const m512 data, const m512 hi_mask,
const m512 lo_mask, const m512 and_mask,
const u64a neg_mask, const u64a valid_data_mask) {
@@ -254,119 +254,119 @@ int validateShuftiMask64x16(const m512 data,
#endif
static really_inline
-int checkMultipath32(u32 data, u32 hi_bits, u32 lo_bits) {
- u32 t = ~(data | hi_bits);
- t += lo_bits;
- t &= (~data) & hi_bits;
- DEBUG_PRINTF("t %x\n", t);
- return !!t;
-}
-
-static really_inline
-int checkMultipath64(u64a data, u64a hi_bits, u64a lo_bits) {
- u64a t = ~(data | hi_bits);
- t += lo_bits;
- t &= (~data) & hi_bits;
- DEBUG_PRINTF("t %llx\n", t);
- return !!t;
-}
-
-static really_inline
-int validateMultipathShuftiMask16x8(const m128 data,
- const m256 nib_mask,
- const m128 bucket_select_mask,
- const u32 hi_bits, const u32 lo_bits,
- const u32 neg_mask,
- const u32 valid_path_mask) {
- m256 data_256 = combine2x128(rshift64_m128(data, 4), data);
- m256 low4bits = set32x8(0xf);
- m256 c_nib = pshufb_m256(nib_mask, and256(data_256, low4bits));
- m128 t = and128(movdq_hi(c_nib), movdq_lo(c_nib));
- m128 result = and128(t, bucket_select_mask);
- u32 nresult = movemask128(eq128(result, zeroes128()));
- u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
-
- DEBUG_PRINTF("cmp_result %x\n", cmp_result);
-
- return checkMultipath32(cmp_result, hi_bits, lo_bits);
-}
-
-static really_inline
-int validateMultipathShuftiMask32x8(const m256 data,
- const m256 hi_mask, const m256 lo_mask,
- const m256 bucket_select_mask,
- const u32 hi_bits, const u32 lo_bits,
- const u32 neg_mask,
- const u32 valid_path_mask) {
- m256 low4bits = set32x8(0xf);
- m256 data_lo = and256(data, low4bits);
- m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
- m256 c_lo = pshufb_m256(lo_mask, data_lo);
- m256 c_hi = pshufb_m256(hi_mask, data_hi);
- m256 c = and256(c_lo, c_hi);
- m256 result = and256(c, bucket_select_mask);
- u32 nresult = movemask256(eq256(result, zeroes256()));
- u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
-
- DEBUG_PRINTF("cmp_result %x\n", cmp_result);
-
- return checkMultipath32(cmp_result, hi_bits, lo_bits);
-}
-
-static really_inline
-int validateMultipathShuftiMask32x16(const m256 data,
- const m256 hi_mask_1, const m256 hi_mask_2,
- const m256 lo_mask_1, const m256 lo_mask_2,
- const m256 bucket_select_mask_hi,
- const m256 bucket_select_mask_lo,
- const u32 hi_bits, const u32 lo_bits,
- const u32 neg_mask,
- const u32 valid_path_mask) {
- m256 low4bits = set32x8(0xf);
- m256 data_lo = and256(data, low4bits);
- m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
- m256 c_lo_1 = pshufb_m256(lo_mask_1, data_lo);
- m256 c_lo_2 = pshufb_m256(lo_mask_2, data_lo);
- m256 c_hi_1 = pshufb_m256(hi_mask_1, data_hi);
- m256 c_hi_2 = pshufb_m256(hi_mask_2, data_hi);
- m256 t1 = and256(c_lo_1, c_hi_1);
- m256 t2 = and256(c_lo_2, c_hi_2);
- m256 result = or256(and256(t1, bucket_select_mask_lo),
- and256(t2, bucket_select_mask_hi));
- u32 nresult = movemask256(eq256(result, zeroes256()));
- u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
-
- DEBUG_PRINTF("cmp_result %x\n", cmp_result);
-
- return checkMultipath32(cmp_result, hi_bits, lo_bits);
-}
-
-static really_inline
-int validateMultipathShuftiMask64(const m256 data_1, const m256 data_2,
- const m256 hi_mask, const m256 lo_mask,
- const m256 bucket_select_mask_1,
- const m256 bucket_select_mask_2,
- const u64a hi_bits, const u64a lo_bits,
- const u64a neg_mask,
- const u64a valid_path_mask) {
- m256 low4bits = set32x8(0xf);
- m256 c_lo_1 = pshufb_m256(lo_mask, and256(data_1, low4bits));
- m256 c_lo_2 = pshufb_m256(lo_mask, and256(data_2, low4bits));
- m256 c_hi_1 = pshufb_m256(hi_mask,
- rshift64_m256(andnot256(low4bits, data_1), 4));
- m256 c_hi_2 = pshufb_m256(hi_mask,
- rshift64_m256(andnot256(low4bits, data_2), 4));
- m256 t1 = and256(c_lo_1, c_hi_1);
- m256 t2 = and256(c_lo_2, c_hi_2);
- m256 nresult_1 = eq256(and256(t1, bucket_select_mask_1), zeroes256());
- m256 nresult_2 = eq256(and256(t2, bucket_select_mask_2), zeroes256());
- u64a nresult = (u64a)movemask256(nresult_1) |
- (u64a)movemask256(nresult_2) << 32;
- u64a cmp_result = (nresult ^ neg_mask) | valid_path_mask;
-
- DEBUG_PRINTF("cmp_result %llx\n", cmp_result);
-
- return checkMultipath64(cmp_result, hi_bits, lo_bits);
-}
-
-#endif
+int checkMultipath32(u32 data, u32 hi_bits, u32 lo_bits) {
+ u32 t = ~(data | hi_bits);
+ t += lo_bits;
+ t &= (~data) & hi_bits;
+ DEBUG_PRINTF("t %x\n", t);
+ return !!t;
+}
+
+static really_inline
+int checkMultipath64(u64a data, u64a hi_bits, u64a lo_bits) {
+ u64a t = ~(data | hi_bits);
+ t += lo_bits;
+ t &= (~data) & hi_bits;
+ DEBUG_PRINTF("t %llx\n", t);
+ return !!t;
+}
+
+static really_inline
+int validateMultipathShuftiMask16x8(const m128 data,
+ const m256 nib_mask,
+ const m128 bucket_select_mask,
+ const u32 hi_bits, const u32 lo_bits,
+ const u32 neg_mask,
+ const u32 valid_path_mask) {
+ m256 data_256 = combine2x128(rshift64_m128(data, 4), data);
+ m256 low4bits = set32x8(0xf);
+ m256 c_nib = pshufb_m256(nib_mask, and256(data_256, low4bits));
+ m128 t = and128(movdq_hi(c_nib), movdq_lo(c_nib));
+ m128 result = and128(t, bucket_select_mask);
+ u32 nresult = movemask128(eq128(result, zeroes128()));
+ u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
+
+ DEBUG_PRINTF("cmp_result %x\n", cmp_result);
+
+ return checkMultipath32(cmp_result, hi_bits, lo_bits);
+}
+
+static really_inline
+int validateMultipathShuftiMask32x8(const m256 data,
+ const m256 hi_mask, const m256 lo_mask,
+ const m256 bucket_select_mask,
+ const u32 hi_bits, const u32 lo_bits,
+ const u32 neg_mask,
+ const u32 valid_path_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 data_lo = and256(data, low4bits);
+ m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
+ m256 c_lo = pshufb_m256(lo_mask, data_lo);
+ m256 c_hi = pshufb_m256(hi_mask, data_hi);
+ m256 c = and256(c_lo, c_hi);
+ m256 result = and256(c, bucket_select_mask);
+ u32 nresult = movemask256(eq256(result, zeroes256()));
+ u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
+
+ DEBUG_PRINTF("cmp_result %x\n", cmp_result);
+
+ return checkMultipath32(cmp_result, hi_bits, lo_bits);
+}
+
+static really_inline
+int validateMultipathShuftiMask32x16(const m256 data,
+ const m256 hi_mask_1, const m256 hi_mask_2,
+ const m256 lo_mask_1, const m256 lo_mask_2,
+ const m256 bucket_select_mask_hi,
+ const m256 bucket_select_mask_lo,
+ const u32 hi_bits, const u32 lo_bits,
+ const u32 neg_mask,
+ const u32 valid_path_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 data_lo = and256(data, low4bits);
+ m256 data_hi = and256(rshift64_m256(data, 4), low4bits);
+ m256 c_lo_1 = pshufb_m256(lo_mask_1, data_lo);
+ m256 c_lo_2 = pshufb_m256(lo_mask_2, data_lo);
+ m256 c_hi_1 = pshufb_m256(hi_mask_1, data_hi);
+ m256 c_hi_2 = pshufb_m256(hi_mask_2, data_hi);
+ m256 t1 = and256(c_lo_1, c_hi_1);
+ m256 t2 = and256(c_lo_2, c_hi_2);
+ m256 result = or256(and256(t1, bucket_select_mask_lo),
+ and256(t2, bucket_select_mask_hi));
+ u32 nresult = movemask256(eq256(result, zeroes256()));
+ u32 cmp_result = (nresult ^ neg_mask) | valid_path_mask;
+
+ DEBUG_PRINTF("cmp_result %x\n", cmp_result);
+
+ return checkMultipath32(cmp_result, hi_bits, lo_bits);
+}
+
+static really_inline
+int validateMultipathShuftiMask64(const m256 data_1, const m256 data_2,
+ const m256 hi_mask, const m256 lo_mask,
+ const m256 bucket_select_mask_1,
+ const m256 bucket_select_mask_2,
+ const u64a hi_bits, const u64a lo_bits,
+ const u64a neg_mask,
+ const u64a valid_path_mask) {
+ m256 low4bits = set32x8(0xf);
+ m256 c_lo_1 = pshufb_m256(lo_mask, and256(data_1, low4bits));
+ m256 c_lo_2 = pshufb_m256(lo_mask, and256(data_2, low4bits));
+ m256 c_hi_1 = pshufb_m256(hi_mask,
+ rshift64_m256(andnot256(low4bits, data_1), 4));
+ m256 c_hi_2 = pshufb_m256(hi_mask,
+ rshift64_m256(andnot256(low4bits, data_2), 4));
+ m256 t1 = and256(c_lo_1, c_hi_1);
+ m256 t2 = and256(c_lo_2, c_hi_2);
+ m256 nresult_1 = eq256(and256(t1, bucket_select_mask_1), zeroes256());
+ m256 nresult_2 = eq256(and256(t2, bucket_select_mask_2), zeroes256());
+ u64a nresult = (u64a)movemask256(nresult_1) |
+ (u64a)movemask256(nresult_2) << 32;
+ u64a cmp_result = (nresult ^ neg_mask) | valid_path_mask;
+
+ DEBUG_PRINTF("cmp_result %llx\n", cmp_result);
+
+ return checkMultipath64(cmp_result, hi_bits, lo_bits);
+}
+
+#endif