/root/doris/be/src/util/hash_util.hpp
Line | Count | Source |
1 | | // Licensed to the Apache Software Foundation (ASF) under one |
2 | | // or more contributor license agreements. See the NOTICE file |
3 | | // distributed with this work for additional information |
4 | | // regarding copyright ownership. The ASF licenses this file |
5 | | // to you under the Apache License, Version 2.0 (the |
6 | | // "License"); you may not use this file except in compliance |
7 | | // with the License. You may obtain a copy of the License at |
8 | | // |
9 | | // http://www.apache.org/licenses/LICENSE-2.0 |
10 | | // |
11 | | // Unless required by applicable law or agreed to in writing, |
12 | | // software distributed under the License is distributed on an |
13 | | // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
14 | | // KIND, either express or implied. See the License for the |
15 | | // specific language governing permissions and limitations |
16 | | // under the License. |
17 | | // This file is copied from |
18 | | // https://github.com/apache/impala/blob/branch-2.9.0/be/src/util/hash-util.h |
19 | | // and modified by Doris |
20 | | |
21 | | #pragma once |
22 | | |
23 | | #include <crc32c/crc32c.h> |
24 | | #include <gen_cpp/Types_types.h> |
25 | | #include <xxh3.h> |
26 | | #include <xxhash.h> |
27 | | #include <zlib.h> |
28 | | |
29 | | #include <bit> |
30 | | #include <functional> |
31 | | |
32 | | #include "common/compiler_util.h" // IWYU pragma: keep |
33 | | #include "util/cpu_info.h" |
34 | | #include "util/hash/city.h" |
35 | | #include "util/murmur_hash3.h" |
36 | | #include "util/sse_util.hpp" |
37 | | #include "vec/common/endian.h" |
38 | | |
39 | | namespace doris { |
40 | | #include "common/compile_check_begin.h" |
41 | | namespace detail { |
42 | | // Slicing-by-4 table: t[0] is the standard byte-at-a-time table, |
43 | | // t[1..3] are extended tables for parallel 4-byte processing. |
44 | | struct CRC32SliceBy4Table { |
45 | | uint32_t t[4][256] {}; |
46 | 0 | constexpr CRC32SliceBy4Table() { |
47 | 0 | // t[0]: standard CRC32 lookup table |
48 | 0 | for (uint32_t i = 0; i < 256; i++) { |
49 | 0 | uint32_t c = i; |
50 | 0 | for (int j = 0; j < 8; j++) { |
51 | 0 | c = (c & 1) ? ((c >> 1) ^ 0xEDB88320U) : (c >> 1); |
52 | 0 | } |
53 | 0 | t[0][i] = c; |
54 | 0 | } |
55 | 0 | // t[1..3]: each entry is one additional CRC byte-step applied to t[k-1] |
56 | 0 | for (uint32_t i = 0; i < 256; i++) { |
57 | 0 | uint32_t c = t[0][i]; |
58 | 0 | for (int k = 1; k < 4; k++) { |
59 | 0 | c = t[0][c & 0xFF] ^ (c >> 8); |
60 | 0 | t[k][i] = c; |
61 | 0 | } |
62 | 0 | } |
63 | 0 | } |
64 | | }; |
65 | | } // namespace detail |
66 | | |
67 | | // Utility class to compute hash values. |
68 | | class HashUtil { |
69 | | private: |
70 | | static inline constexpr detail::CRC32SliceBy4Table CRC32_TABLE {}; |
71 | | |
72 | | public: |
73 | 641k | static uint32_t zlib_crc_hash(const void* data, uint32_t bytes, uint32_t hash) { |
74 | 641k | return (uint32_t)crc32(hash, (const unsigned char*)data, bytes); |
75 | 641k | } |
76 | | |
77 | | // Inline CRC32 (zlib-compatible, standard CRC32 polynomial) for fixed-size types. |
78 | | // Uses Slicing-by-4 technique for 4/8-byte types: processes 4 bytes at a time using |
79 | | // 4 precomputed lookup tables, reducing serial table lookups from 4 to 1 per 4-byte chunk. |
80 | | // Polynomial: 0xEDB88320 (reflected form of 0x04C11DB7). |
81 | | // Endian note: CRC32 reflected algorithm processes bytes in address order (byte[0] first). |
82 | | // Slicing-by-4 requires byte[0] at LSB of the loaded uint32_t, which is little-endian layout. |
83 | | // LittleEndian::Load32 provides this on ALL platforms: noop on LE, bswap on BE. |
84 | | template <typename T> |
85 | 4.23M | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { |
86 | 4.23M | const auto* p = reinterpret_cast<const uint8_t*>(&value); |
87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF |
88 | 4.23M | uint32_t crc = hash ^ 0xFFFFFFFFU; |
89 | | |
90 | 4.23M | if constexpr (sizeof(T) == 1) { |
91 | | // 1 byte: single table lookup |
92 | 202 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); |
93 | 207 | } else if constexpr (sizeof(T) == 2) { |
94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) |
95 | 207 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); |
96 | 207 | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); |
97 | 4.22M | } else if constexpr (sizeof(T) == 4) { |
98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel |
99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, |
100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. |
101 | 4.22M | uint32_t word = LittleEndian::Load32(p) ^ crc; |
102 | 4.22M | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ |
103 | 4.22M | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; |
104 | 4.22M | } else if constexpr (sizeof(T) == 8) { |
105 | | // 8 bytes: two Slicing-by-4 steps |
106 | 1.34k | uint32_t word = LittleEndian::Load32(p) ^ crc; |
107 | 1.34k | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ |
108 | 1.34k | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; |
109 | | |
110 | 1.34k | word = LittleEndian::Load32(p + 4) ^ crc; |
111 | 1.34k | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ |
112 | 1.34k | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; |
113 | 2.88k | } else { |
114 | | // Fallback to zlib for larger/unusual types |
115 | 2.88k | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); |
116 | 2.88k | } |
117 | 0 | return crc ^ 0xFFFFFFFFU; |
118 | 4.23M | } _ZN5doris8HashUtil16zlib_crc32_fixedIiEEjRKT_j Line | Count | Source | 85 | 4.22M | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 4.22M | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 4.22M | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | 4.22M | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | 4.22M | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | 4.22M | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | 4.22M | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 4.22M | return crc ^ 0xFFFFFFFFU; | 118 | 4.22M | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIhEEjRKT_j Line | Count | Source | 85 | 41 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 41 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 41 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | 41 | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | 41 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 41 | return crc ^ 0xFFFFFFFFU; | 118 | 41 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIsEEjRKT_j Line | Count | Source | 85 | 186 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 186 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 186 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | 186 | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | 186 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | 186 | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 186 | return crc ^ 0xFFFFFFFFU; | 118 | 186 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedItEEjRKT_j Line | Count | Source | 85 | 21 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 21 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 21 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | 21 | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | 21 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | 21 | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 21 | return crc ^ 0xFFFFFFFFU; | 118 | 21 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIjEEjRKT_j Line | Count | Source | 85 | 51 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 51 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 51 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | 51 | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | 51 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | 51 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | 51 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 51 | return crc ^ 0xFFFFFFFFU; | 118 | 51 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIlEEjRKT_j Line | Count | Source | 85 | 206 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 206 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 206 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | 206 | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | 206 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | 206 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | 206 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | 206 | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | 206 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | 206 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 206 | return crc ^ 0xFFFFFFFFU; | 118 | 206 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedImEEjRKT_j Line | Count | Source | 85 | 18 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 18 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 18 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | 18 | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | 18 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | 18 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | 18 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | 18 | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | 18 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | 18 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 18 | return crc ^ 0xFFFFFFFFU; | 118 | 18 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIfEEjRKT_j Line | Count | Source | 85 | 20 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 20 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 20 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | 20 | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | 20 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | 20 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | 20 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 20 | return crc ^ 0xFFFFFFFFU; | 118 | 20 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIdEEjRKT_j Line | Count | Source | 85 | 42 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 42 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 42 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | 42 | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | 42 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | 42 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | 42 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | 42 | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | 42 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | 42 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 42 | return crc ^ 0xFFFFFFFFU; | 118 | 42 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedInEEjRKT_j Line | Count | Source | 85 | 158 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 158 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 158 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | 158 | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | 158 | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | 158 | } | 117 | 0 | return crc ^ 0xFFFFFFFFU; | 118 | 158 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_10vectorized7DecimalIiEEEEjRKT_j Line | Count | Source | 85 | 519 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 519 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 519 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | 519 | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | 519 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | 519 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | 519 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 519 | return crc ^ 0xFFFFFFFFU; | 118 | 519 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_10vectorized7DecimalIlEEEEjRKT_j Line | Count | Source | 85 | 966 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 966 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 966 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | 966 | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | 966 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | 966 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | 966 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | 966 | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | 966 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | 966 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 966 | return crc ^ 0xFFFFFFFFU; | 118 | 966 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_10vectorized12Decimal128V3EEEjRKT_j Line | Count | Source | 85 | 1.05k | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 1.05k | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 1.05k | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | 1.05k | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | 1.05k | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | 1.05k | } | 117 | 0 | return crc ^ 0xFFFFFFFFU; | 118 | 1.05k | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_10vectorized7DecimalIN4wide7integerILm256EiEEEEEEjRKT_j Line | Count | Source | 85 | 1.65k | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 1.65k | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 1.65k | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | 1.65k | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | 1.65k | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | 1.65k | } | 117 | 0 | return crc ^ 0xFFFFFFFFU; | 118 | 1.65k | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIaEEjRKT_j Line | Count | Source | 85 | 161 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 161 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 161 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | 161 | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | 161 | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 161 | return crc ^ 0xFFFFFFFFU; | 118 | 161 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedIoEEjRKT_j Line | Count | Source | 85 | 12 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 12 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 12 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | 12 | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | 12 | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | 12 | } | 117 | 0 | return crc ^ 0xFFFFFFFFU; | 118 | 12 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_11DateV2ValueINS_15DateV2ValueTypeEEEEEjRKT_j Line | Count | Source | 85 | 37 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 37 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 37 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | 37 | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | 37 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | 37 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | 37 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 37 | return crc ^ 0xFFFFFFFFU; | 118 | 37 | } |
_ZN5doris8HashUtil16zlib_crc32_fixedINS_11DateV2ValueINS_19DateTimeV2ValueTypeEEEEEjRKT_j Line | Count | Source | 85 | 111 | static uint32_t zlib_crc32_fixed(const T& value, uint32_t hash) { | 86 | 111 | const auto* p = reinterpret_cast<const uint8_t*>(&value); | 87 | | // zlib convention: pre/post XOR with 0xFFFFFFFF | 88 | 111 | uint32_t crc = hash ^ 0xFFFFFFFFU; | 89 | | | 90 | | if constexpr (sizeof(T) == 1) { | 91 | | // 1 byte: single table lookup | 92 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 93 | | } else if constexpr (sizeof(T) == 2) { | 94 | | // 2 bytes: two sequential table lookups (slicing doesn't help below 4 bytes) | 95 | | crc = CRC32_TABLE.t[0][(crc ^ p[0]) & 0xFF] ^ (crc >> 8); | 96 | | crc = CRC32_TABLE.t[0][(crc ^ p[1]) & 0xFF] ^ (crc >> 8); | 97 | | } else if constexpr (sizeof(T) == 4) { | 98 | | // 4 bytes: one Slicing-by-4 step — 4 independent lookups in parallel | 99 | | // LittleEndian::Load32 handles unaligned load + byte-swap on big-endian, | 100 | | // ensuring byte[0] is always at LSB for correct CRC byte processing order. | 101 | | uint32_t word = LittleEndian::Load32(p) ^ crc; | 102 | | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 103 | | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 104 | 111 | } else if constexpr (sizeof(T) == 8) { | 105 | | // 8 bytes: two Slicing-by-4 steps | 106 | 111 | uint32_t word = LittleEndian::Load32(p) ^ crc; | 107 | 111 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 108 | 111 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 109 | | | 110 | 111 | word = LittleEndian::Load32(p + 4) ^ crc; | 111 | 111 | crc = CRC32_TABLE.t[3][(word)&0xFF] ^ CRC32_TABLE.t[2][(word >> 8) & 0xFF] ^ | 112 | 111 | CRC32_TABLE.t[1][(word >> 16) & 0xFF] ^ CRC32_TABLE.t[0][(word >> 24) & 0xFF]; | 113 | | } else { | 114 | | // Fallback to zlib for larger/unusual types | 115 | | return (uint32_t)crc32(hash, (const unsigned char*)&value, sizeof(T)); | 116 | | } | 117 | 111 | return crc ^ 0xFFFFFFFFU; | 118 | 111 | } |
Unexecuted instantiation: _ZN5doris8HashUtil16zlib_crc32_fixedINS_16TimestampTzValueEEEjRKT_j |
119 | | |
120 | 1.08M | static uint32_t zlib_crc_hash_null(uint32_t hash) { |
121 | | // null is treat as 0 when hash |
122 | 1.08M | static const int INT_VALUE = 0; |
123 | 1.08M | return zlib_crc32_fixed(INT_VALUE, hash); |
124 | 1.08M | } |
125 | | |
126 | | template <typename T> |
127 | 83.7k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { |
128 | 83.7k | if constexpr (sizeof(T) == 1) { |
129 | 5.90k | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); |
130 | 5.90k | } else if constexpr (sizeof(T) == 2) { |
131 | 4.93k | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); |
132 | 26.9k | } else if constexpr (sizeof(T) == 4) { |
133 | 26.9k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); |
134 | 26.9k | } else if constexpr (sizeof(T) == 8) { |
135 | 26.2k | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); |
136 | 26.2k | } else { |
137 | 19.6k | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); |
138 | 19.6k | } |
139 | 83.7k | } _ZN5doris8HashUtil12crc32c_fixedIiEEjRKT_j Line | Count | Source | 127 | 12.0k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | 12.0k | } else if constexpr (sizeof(T) == 4) { | 133 | 12.0k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 12.0k | } |
_ZN5doris8HashUtil12crc32c_fixedINS_10vectorized7DecimalIiEEEEjRKT_j Line | Count | Source | 127 | 2.16k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | 2.16k | } else if constexpr (sizeof(T) == 4) { | 133 | 2.16k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 2.16k | } |
_ZN5doris8HashUtil12crc32c_fixedINS_10vectorized7DecimalIlEEEEjRKT_j Line | Count | Source | 127 | 3.49k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | 3.49k | } else if constexpr (sizeof(T) == 8) { | 135 | 3.49k | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 3.49k | } |
Unexecuted instantiation: _ZN5doris8HashUtil12crc32c_fixedINS_14DecimalV2ValueEEEjRKT_j _ZN5doris8HashUtil12crc32c_fixedINS_10vectorized12Decimal128V3EEEjRKT_j Line | Count | Source | 127 | 3.91k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | 3.91k | } else { | 137 | 3.91k | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | 3.91k | } | 139 | 3.91k | } |
_ZN5doris8HashUtil12crc32c_fixedINS_10vectorized7DecimalIN4wide7integerILm256EiEEEEEEjRKT_j Line | Count | Source | 127 | 7.44k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | 7.44k | } else { | 137 | 7.44k | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | 7.44k | } | 139 | 7.44k | } |
_ZN5doris8HashUtil12crc32c_fixedIhEEjRKT_j Line | Count | Source | 127 | 804 | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | 804 | if constexpr (sizeof(T) == 1) { | 129 | 804 | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 804 | } |
_ZN5doris8HashUtil12crc32c_fixedIaEEjRKT_j Line | Count | Source | 127 | 5.09k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | 5.09k | if constexpr (sizeof(T) == 1) { | 129 | 5.09k | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 5.09k | } |
_ZN5doris8HashUtil12crc32c_fixedIsEEjRKT_j Line | Count | Source | 127 | 4.93k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | 4.93k | } else if constexpr (sizeof(T) == 2) { | 131 | 4.93k | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 4.93k | } |
_ZN5doris8HashUtil12crc32c_fixedIlEEjRKT_j Line | Count | Source | 127 | 6.90k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | 6.90k | } else if constexpr (sizeof(T) == 8) { | 135 | 6.90k | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 6.90k | } |
_ZN5doris8HashUtil12crc32c_fixedInEEjRKT_j Line | Count | Source | 127 | 3.77k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | 3.77k | } else { | 137 | 3.77k | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | 3.77k | } | 139 | 3.77k | } |
_ZN5doris8HashUtil12crc32c_fixedIfEEjRKT_j Line | Count | Source | 127 | 4.32k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | 4.32k | } else if constexpr (sizeof(T) == 4) { | 133 | 4.32k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 4.32k | } |
_ZN5doris8HashUtil12crc32c_fixedIdEEjRKT_j Line | Count | Source | 127 | 4.85k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | 4.85k | } else if constexpr (sizeof(T) == 8) { | 135 | 4.85k | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 4.85k | } |
_ZN5doris8HashUtil12crc32c_fixedIjEEjRKT_j Line | Count | Source | 127 | 6.79k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | 6.79k | } else if constexpr (sizeof(T) == 4) { | 133 | 6.79k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 6.79k | } |
_ZN5doris8HashUtil12crc32c_fixedIoEEjRKT_j Line | Count | Source | 127 | 4.53k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | 4.53k | } else { | 137 | 4.53k | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | 4.53k | } | 139 | 4.53k | } |
_ZN5doris8HashUtil12crc32c_fixedINS_11DateV2ValueINS_15DateV2ValueTypeEEEEEjRKT_j Line | Count | Source | 127 | 1.57k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | 1.57k | } else if constexpr (sizeof(T) == 4) { | 133 | 1.57k | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | | } else if constexpr (sizeof(T) == 8) { | 135 | | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 1.57k | } |
_ZN5doris8HashUtil12crc32c_fixedINS_11DateV2ValueINS_19DateTimeV2ValueTypeEEEEEjRKT_j Line | Count | Source | 127 | 11.0k | static uint32_t crc32c_fixed(const T& value, uint32_t hash) { | 128 | | if constexpr (sizeof(T) == 1) { | 129 | | return _mm_crc32_u8(hash, *reinterpret_cast<const uint8_t*>(&value)); | 130 | | } else if constexpr (sizeof(T) == 2) { | 131 | | return _mm_crc32_u16(hash, *reinterpret_cast<const uint16_t*>(&value)); | 132 | | } else if constexpr (sizeof(T) == 4) { | 133 | | return _mm_crc32_u32(hash, *reinterpret_cast<const uint32_t*>(&value)); | 134 | 11.0k | } else if constexpr (sizeof(T) == 8) { | 135 | 11.0k | return (uint32_t)_mm_crc32_u64(hash, *reinterpret_cast<const uint64_t*>(&value)); | 136 | | } else { | 137 | | return crc32c_extend(hash, (const uint8_t*)&value, sizeof(T)); | 138 | | } | 139 | 11.0k | } |
Unexecuted instantiation: _ZN5doris8HashUtil12crc32c_fixedINS_16TimestampTzValueEEEjRKT_j Unexecuted instantiation: _ZN5doris8HashUtil12crc32c_fixedImEEjRKT_j |
140 | | |
141 | 460 | static uint32_t crc32c_null(uint32_t hash) { |
142 | | // null is treat as 0 when hash |
143 | 460 | static const int INT_VALUE = 0; |
144 | 460 | return crc32c_fixed(INT_VALUE, hash); |
145 | 460 | } |
146 | | |
147 | | // Compute the Crc32 hash for data using SSE4 instructions. The input hash parameter is |
148 | | // the current hash/seed value. |
149 | | // This should only be called if SSE is supported. |
150 | | // This is ~4x faster than Fnv/Boost Hash. |
151 | | // NOTE: DO NOT use this method for checksum! This does not generate the standard CRC32 checksum! |
152 | | // For checksum, use CRC-32C algorithm from crc32c.h |
153 | | // NOTE: Any changes made to this function need to be reflected in Codegen::GetHashFn. |
154 | | // TODO: crc32 hashes with different seeds do not result in different hash functions. |
155 | | // The resulting hashes are correlated. |
156 | | // ATTN: prefer do not use this function anymore, use crc32c::Extend instead |
157 | | // This function is retained because it is not certain whether there are compatibility issues with historical data. |
158 | 181k | static uint32_t crc_hash(const void* data, uint32_t bytes, uint32_t hash) { |
159 | 181k | if (!CpuInfo::is_supported(CpuInfo::SSE4_2)) { |
160 | 0 | return zlib_crc_hash(data, bytes, hash); |
161 | 0 | } |
162 | 181k | uint32_t words = bytes / sizeof(uint32_t); |
163 | 181k | bytes = bytes % sizeof(uint32_t); |
164 | | |
165 | 181k | const uint32_t* p = reinterpret_cast<const uint32_t*>(data); |
166 | | |
167 | 319k | while (words--) { |
168 | 138k | hash = _mm_crc32_u32(hash, *p); |
169 | 138k | ++p; |
170 | 138k | } |
171 | | |
172 | 181k | const uint8_t* s = reinterpret_cast<const uint8_t*>(p); |
173 | | |
174 | 185k | while (bytes--) { |
175 | 4.36k | hash = _mm_crc32_u8(hash, *s); |
176 | 4.36k | ++s; |
177 | 4.36k | } |
178 | | |
179 | | // The lower half of the CRC hash has has poor uniformity, so swap the halves |
180 | | // for anyone who only uses the first several bits of the hash. |
181 | 181k | hash = (hash << 16) | (hash >> 16); |
182 | 181k | return hash; |
183 | 181k | } |
184 | | |
185 | 0 | static uint64_t crc_hash64(const void* data, uint32_t bytes, uint64_t hash) { |
186 | 0 | uint32_t words = bytes / sizeof(uint32_t); |
187 | 0 | bytes = bytes % sizeof(uint32_t); |
188 | |
|
189 | 0 | uint32_t h1 = hash >> 32; |
190 | 0 | uint32_t h2 = (hash << 32) >> 32; |
191 | |
|
192 | 0 | const uint32_t* p = reinterpret_cast<const uint32_t*>(data); |
193 | 0 | while (words--) { |
194 | 0 | (words & 1) ? (h1 = _mm_crc32_u32(h1, *p)) : (h2 = _mm_crc32_u32(h2, *p)); |
195 | 0 | ++p; |
196 | 0 | } |
197 | |
|
198 | 0 | const uint8_t* s = reinterpret_cast<const uint8_t*>(p); |
199 | 0 | while (bytes--) { |
200 | 0 | (bytes & 1) ? (h1 = _mm_crc32_u8(h1, *s)) : (h2 = _mm_crc32_u8(h2, *s)); |
201 | 0 | ++s; |
202 | 0 | } |
203 | 0 | union { |
204 | 0 | uint64_t u64; |
205 | 0 | uint32_t u32[2]; |
206 | 0 | } converter; |
207 | 0 | converter.u64 = hash; |
208 | |
|
209 | 0 | h1 = (h1 << 16) | (h1 >> 16); |
210 | 0 | h2 = (h2 << 16) | (h2 >> 16); |
211 | 0 | converter.u32[0] = h1; |
212 | 0 | converter.u32[1] = h2; |
213 | |
|
214 | 0 | return converter.u64; |
215 | 0 | } |
216 | | |
217 | | // refer to https://github.com/apache/commons-codec/blob/master/src/main/java/org/apache/commons/codec/digest/MurmurHash3.java |
218 | | static const uint32_t MURMUR3_32_SEED = 104729; |
219 | | |
220 | | // modify from https://github.com/aappleby/smhasher/blob/master/src/MurmurHash3.cpp |
221 | 20 | static uint32_t murmur_hash3_32(const void* key, int64_t len, uint32_t seed) { |
222 | 20 | uint32_t out = 0; |
223 | 20 | murmur_hash3_x86_32(key, len, seed, &out); |
224 | 20 | return out; |
225 | 20 | } |
226 | | |
227 | | template <bool is_mmh64_v2> |
228 | 15 | static uint64_t murmur_hash3_64(const void* key, int64_t len, uint64_t seed) { |
229 | 15 | uint64_t out = 0; |
230 | 15 | if constexpr (is_mmh64_v2) { |
231 | 3 | murmur_hash3_x64_64_shared(key, len, seed, &out); |
232 | 12 | } else { |
233 | 12 | murmur_hash3_x64_64(key, len, seed, &out); |
234 | 12 | } |
235 | 15 | return out; |
236 | 15 | } _ZN5doris8HashUtil15murmur_hash3_64ILb0EEEmPKvlm Line | Count | Source | 228 | 12 | static uint64_t murmur_hash3_64(const void* key, int64_t len, uint64_t seed) { | 229 | 12 | uint64_t out = 0; | 230 | | if constexpr (is_mmh64_v2) { | 231 | | murmur_hash3_x64_64_shared(key, len, seed, &out); | 232 | 12 | } else { | 233 | 12 | murmur_hash3_x64_64(key, len, seed, &out); | 234 | 12 | } | 235 | 12 | return out; | 236 | 12 | } |
_ZN5doris8HashUtil15murmur_hash3_64ILb1EEEmPKvlm Line | Count | Source | 228 | 3 | static uint64_t murmur_hash3_64(const void* key, int64_t len, uint64_t seed) { | 229 | 3 | uint64_t out = 0; | 230 | 3 | if constexpr (is_mmh64_v2) { | 231 | 3 | murmur_hash3_x64_64_shared(key, len, seed, &out); | 232 | | } else { | 233 | | murmur_hash3_x64_64(key, len, seed, &out); | 234 | | } | 235 | 3 | return out; | 236 | 3 | } |
|
237 | | |
238 | | static const int MURMUR_R = 47; |
239 | | |
240 | | // Murmur2 hash implementation returning 64-bit hashes. |
241 | 0 | static uint64_t murmur_hash2_64(const void* input, int len, uint64_t seed) { |
242 | 0 | uint64_t h = seed ^ (len * MURMUR_PRIME); |
243 | 0 |
|
244 | 0 | const uint64_t* data = reinterpret_cast<const uint64_t*>(input); |
245 | 0 | const uint64_t* end = data + (len / sizeof(uint64_t)); |
246 | 0 |
|
247 | 0 | while (data != end) { |
248 | 0 | uint64_t k = *data++; |
249 | 0 | k *= MURMUR_PRIME; |
250 | 0 | k ^= k >> MURMUR_R; |
251 | 0 | k *= MURMUR_PRIME; |
252 | 0 | h ^= k; |
253 | 0 | h *= MURMUR_PRIME; |
254 | 0 | } |
255 | 0 |
|
256 | 0 | const uint8_t* data2 = reinterpret_cast<const uint8_t*>(data); |
257 | 0 | switch (len & 7) { |
258 | 0 | case 7: |
259 | 0 | h ^= uint64_t(data2[6]) << 48; |
260 | 0 | [[fallthrough]]; |
261 | 0 | case 6: |
262 | 0 | h ^= uint64_t(data2[5]) << 40; |
263 | 0 | [[fallthrough]]; |
264 | 0 | case 5: |
265 | 0 | h ^= uint64_t(data2[4]) << 32; |
266 | 0 | [[fallthrough]]; |
267 | 0 | case 4: |
268 | 0 | h ^= uint64_t(data2[3]) << 24; |
269 | 0 | [[fallthrough]]; |
270 | 0 | case 3: |
271 | 0 | h ^= uint64_t(data2[2]) << 16; |
272 | 0 | [[fallthrough]]; |
273 | 0 | case 2: |
274 | 0 | h ^= uint64_t(data2[1]) << 8; |
275 | 0 | [[fallthrough]]; |
276 | 0 | case 1: |
277 | 0 | h ^= uint64_t(data2[0]); |
278 | 0 | h *= MURMUR_PRIME; |
279 | 0 | } |
280 | 0 |
|
281 | 0 | h ^= h >> MURMUR_R; |
282 | 0 | h *= MURMUR_PRIME; |
283 | 0 | h ^= h >> MURMUR_R; |
284 | 0 | return h; |
285 | 0 | } |
286 | | |
287 | | // default values recommended by http://isthe.com/chongo/tech/comp/fnv/ |
288 | | static const uint32_t FNV_PRIME = 0x01000193; // 16777619 |
289 | | static const uint32_t FNV_SEED = 0x811C9DC5; // 2166136261 |
290 | | static const uint64_t FNV64_PRIME = 1099511628211UL; |
291 | | static const uint64_t FNV64_SEED = 14695981039346656037UL; |
292 | | static const uint64_t MURMUR_PRIME = 0xc6a4a7935bd1e995ULL; |
293 | | static const uint32_t MURMUR_SEED = 0xadc83b19ULL; |
294 | | // Implementation of the Fowler–Noll–Vo hash function. This is not as performant |
295 | | // as boost's hash on int types (2x slower) but has bit entropy. |
296 | | // For ints, boost just returns the value of the int which can be pathological. |
297 | | // For example, if the data is <1000, 2000, 3000, 4000, ..> and then the mod of 1000 |
298 | | // is taken on the hash, all values will collide to the same bucket. |
299 | | // For string values, Fnv is slightly faster than boost. |
300 | 0 | static uint32_t fnv_hash(const void* data, uint32_t bytes, uint32_t hash) { |
301 | 0 | const uint8_t* ptr = reinterpret_cast<const uint8_t*>(data); |
302 | |
|
303 | 0 | while (bytes--) { |
304 | 0 | hash = (*ptr ^ hash) * FNV_PRIME; |
305 | 0 | ++ptr; |
306 | 0 | } |
307 | |
|
308 | 0 | return hash; |
309 | 0 | } |
310 | | |
311 | 0 | static uint64_t fnv_hash64(const void* data, uint32_t bytes, uint64_t hash) { |
312 | 0 | const uint8_t* ptr = reinterpret_cast<const uint8_t*>(data); |
313 | 0 |
|
314 | 0 | while (bytes--) { |
315 | 0 | hash = (*ptr ^ hash) * FNV64_PRIME; |
316 | 0 | ++ptr; |
317 | 0 | } |
318 | 0 |
|
319 | 0 | return hash; |
320 | 0 | } |
321 | | |
322 | | // Our hash function is MurmurHash2, 64 bit version. |
323 | | // It was modified in order to provide the same result in |
324 | | // big and little endian archs (endian neutral). |
325 | 67.8k | static uint64_t murmur_hash64A(const void* key, int64_t len, unsigned int seed) { |
326 | 67.8k | const uint64_t m = MURMUR_PRIME; |
327 | 67.8k | const int r = 47; |
328 | 67.8k | uint64_t h = seed ^ (len * m); |
329 | 67.8k | const uint8_t* data = (const uint8_t*)key; |
330 | 67.8k | const uint8_t* end = data + (len - (len & 7)); |
331 | | |
332 | 135k | while (data != end) { |
333 | 67.8k | uint64_t k; |
334 | | if constexpr (std::endian::native == std::endian::big) { |
335 | | k = (uint64_t)data[0]; |
336 | | k |= (uint64_t)data[1] << 8; |
337 | | k |= (uint64_t)data[2] << 16; |
338 | | k |= (uint64_t)data[3] << 24; |
339 | | k |= (uint64_t)data[4] << 32; |
340 | | k |= (uint64_t)data[5] << 40; |
341 | | k |= (uint64_t)data[6] << 48; |
342 | | k |= (uint64_t)data[7] << 56; |
343 | 67.8k | } else if constexpr (std::endian::native == std::endian::little) { |
344 | 67.8k | memcpy(&k, data, sizeof(k)); |
345 | | } else { |
346 | | static_assert(std::endian::native == std::endian::big || |
347 | | std::endian::native == std::endian::little, |
348 | | "Unsupported endianness"); |
349 | | } |
350 | | |
351 | 67.8k | k *= m; |
352 | 67.8k | k ^= k >> r; |
353 | 67.8k | k *= m; |
354 | 67.8k | h ^= k; |
355 | 67.8k | h *= m; |
356 | 67.8k | data += 8; |
357 | 67.8k | } |
358 | | |
359 | 67.8k | switch (len & 7) { |
360 | 0 | case 7: |
361 | 0 | h ^= (uint64_t)data[6] << 48; |
362 | 0 | [[fallthrough]]; |
363 | 0 | case 6: |
364 | 0 | h ^= (uint64_t)data[5] << 40; |
365 | 0 | [[fallthrough]]; |
366 | 0 | case 5: |
367 | 0 | h ^= (uint64_t)data[4] << 32; |
368 | 0 | [[fallthrough]]; |
369 | 3 | case 4: |
370 | 3 | h ^= (uint64_t)data[3] << 24; |
371 | 3 | [[fallthrough]]; |
372 | 3 | case 3: |
373 | 3 | h ^= (uint64_t)data[2] << 16; |
374 | 3 | [[fallthrough]]; |
375 | 3 | case 2: |
376 | 3 | h ^= (uint64_t)data[1] << 8; |
377 | 3 | [[fallthrough]]; |
378 | 6 | case 1: |
379 | 6 | h ^= (uint64_t)data[0]; |
380 | 6 | h *= m; |
381 | 67.8k | } |
382 | | |
383 | 67.8k | h ^= h >> r; |
384 | 67.8k | h *= m; |
385 | 67.8k | h ^= h >> r; |
386 | 67.8k | return h; |
387 | 67.8k | } |
388 | | |
389 | | // Computes the hash value for data. Will call either CrcHash or FnvHash |
390 | | // depending on hardware capabilities. |
391 | | // Seed values for different steps of the query execution should use different seeds |
392 | | // to prevent accidental key collisions. (See IMPALA-219 for more details). |
393 | 181k | static uint32_t hash(const void* data, uint32_t bytes, uint32_t seed) { |
394 | 181k | #ifdef __SSE4_2__ |
395 | | |
396 | 181k | if (LIKELY(CpuInfo::is_supported(CpuInfo::SSE4_2))) { |
397 | 181k | return crc_hash(data, bytes, seed); |
398 | 181k | } else { |
399 | 0 | return fnv_hash(data, bytes, seed); |
400 | 0 | } |
401 | | |
402 | | #else |
403 | | return fnv_hash(data, bytes, seed); |
404 | | #endif |
405 | 181k | } |
406 | | |
407 | 88.5k | static uint64_t hash64(const void* data, uint64_t bytes, uint64_t seed) { |
408 | | #ifdef _SSE4_2_ |
409 | | if (LIKELY(CpuInfo::is_supported(CpuInfo::SSE4_2))) { |
410 | | return crc_hash64(data, bytes, seed); |
411 | | |
412 | | } else { |
413 | | uint64_t hash = 0; |
414 | | murmur_hash3_x64_64(data, bytes, seed, &hash); |
415 | | return hash; |
416 | | } |
417 | | #else |
418 | 88.5k | uint64_t hash = 0; |
419 | 88.5k | murmur_hash3_x64_64(data, bytes, seed, &hash); |
420 | 88.5k | return hash; |
421 | 88.5k | #endif |
422 | 88.5k | } |
423 | | // hash_combine is the same with boost hash_combine, |
424 | | // except replace boost::hash with std::hash |
425 | | template <class T> |
426 | 406 | static inline void hash_combine(std::size_t& seed, const T& v) { |
427 | 406 | std::hash<T> hasher; |
428 | 406 | seed ^= hasher(v) + 0x9e3779b9 + (seed << 6) + (seed >> 2); |
429 | 406 | } _ZN5doris8HashUtil12hash_combineINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEEEvRmRKT_ Line | Count | Source | 426 | 404 | static inline void hash_combine(std::size_t& seed, const T& v) { | 427 | 404 | std::hash<T> hasher; | 428 | 404 | seed ^= hasher(v) + 0x9e3779b9 + (seed << 6) + (seed >> 2); | 429 | 404 | } |
_ZN5doris8HashUtil12hash_combineIlEEvRmRKT_ Line | Count | Source | 426 | 2 | static inline void hash_combine(std::size_t& seed, const T& v) { | 427 | 2 | std::hash<T> hasher; | 428 | 2 | seed ^= hasher(v) + 0x9e3779b9 + (seed << 6) + (seed >> 2); | 429 | 2 | } |
|
430 | | |
431 | | #if defined(__clang__) |
432 | | #pragma clang diagnostic push |
433 | | #pragma clang diagnostic ignored "-Wused-but-marked-unused" |
434 | | #endif |
435 | | // xxHash function for a byte array. For convenience, a 64-bit seed is also |
436 | | // hashed into the result. The mapping may change from time to time. |
437 | 24 | static xxh_u32 xxHash32WithSeed(const char* s, size_t len, xxh_u32 seed) { |
438 | 24 | return XXH32(s, len, seed); |
439 | 24 | } |
440 | | |
441 | | // same to the up function, just for null value |
442 | 0 | static xxh_u32 xxHash32NullWithSeed(xxh_u32 seed) { |
443 | 0 | static const int INT_VALUE = 0; |
444 | 0 | return XXH32(reinterpret_cast<const char*>(&INT_VALUE), sizeof(int), seed); |
445 | 0 | } |
446 | | |
447 | 156k | static xxh_u64 xxHash64WithSeed(const char* s, size_t len, xxh_u64 seed) { |
448 | 156k | return XXH3_64bits_withSeed(s, len, seed); |
449 | 156k | } |
450 | | |
451 | | // same to the up function, just for null value |
452 | 1.08M | static xxh_u64 xxHash64NullWithSeed(xxh_u64 seed) { |
453 | 1.08M | static const int INT_VALUE = 0; |
454 | 1.08M | return XXH3_64bits_withSeed(reinterpret_cast<const char*>(&INT_VALUE), sizeof(int), seed); |
455 | 1.08M | } |
456 | | |
457 | 21 | static xxh_u64 xxhash64_compat_with_seed(const char* s, size_t len, xxh_u64 seed) { |
458 | 21 | return XXH64(reinterpret_cast<const void*>(s), len, seed); |
459 | 21 | } |
460 | | |
461 | 0 | static xxh_u64 xxhash64_compat_null_with_seed(xxh_u64 seed) { |
462 | 0 | static const int INT_VALUE = 0; |
463 | 0 | return XXH64(reinterpret_cast<const void*>(&INT_VALUE), sizeof(int), seed); |
464 | 0 | } |
465 | | |
466 | | #if defined(__clang__) |
467 | | #pragma clang diagnostic pop |
468 | | #endif |
469 | | }; |
470 | | |
471 | | } // namespace doris |
472 | | |
473 | | template <> |
474 | | struct std::hash<doris::TUniqueId> { |
475 | 3.54k | size_t operator()(const doris::TUniqueId& id) const { |
476 | 3.54k | uint32_t seed = 0; |
477 | 3.54k | seed = doris::HashUtil::hash(&id.lo, sizeof(id.lo), seed); |
478 | 3.54k | seed = doris::HashUtil::hash(&id.hi, sizeof(id.hi), seed); |
479 | 3.54k | return seed; |
480 | 3.54k | } |
481 | | }; |
482 | | |
483 | | template <> |
484 | | struct std::hash<doris::TNetworkAddress> { |
485 | 0 | size_t operator()(const doris::TNetworkAddress& address) const { |
486 | 0 | uint32_t seed = 0; |
487 | 0 | seed = doris::HashUtil::hash(address.hostname.data(), (uint32_t)address.hostname.size(), |
488 | 0 | seed); |
489 | 0 | seed = doris::HashUtil::hash(&address.port, 4, seed); |
490 | 0 | return seed; |
491 | 0 | } |
492 | | }; |
493 | | |
494 | | template <> |
495 | | struct std::hash<std::pair<doris::TUniqueId, int64_t>> { |
496 | 0 | size_t operator()(const std::pair<doris::TUniqueId, int64_t>& pair) const { |
497 | 0 | uint32_t seed = 0; |
498 | 0 | seed = doris::HashUtil::hash(&pair.first.lo, sizeof(pair.first.lo), seed); |
499 | 0 | seed = doris::HashUtil::hash(&pair.first.hi, sizeof(pair.first.hi), seed); |
500 | 0 | seed = doris::HashUtil::hash(&pair.second, sizeof(pair.second), seed); |
501 | 0 | return seed; |
502 | 0 | } |
503 | | }; |
504 | | |
505 | | template <class First, class Second> |
506 | | struct std::hash<std::pair<First, Second>> { |
507 | 37.1k | size_t operator()(const pair<First, Second>& p) const { |
508 | 37.1k | size_t h1 = std::hash<First>()(p.first); |
509 | 37.1k | size_t h2 = std::hash<Second>()(p.second); |
510 | 37.1k | return doris::util_hash::HashLen16(h1, h2); |
511 | 37.1k | } Unexecuted instantiation: _ZNKSt4hashISt4pairIlN5doris8RowsetIdEEEclERKS3_ _ZNKSt4hashISt4pairINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEElEEclERKS7_ Line | Count | Source | 507 | 32 | size_t operator()(const pair<First, Second>& p) const { | 508 | 32 | size_t h1 = std::hash<First>()(p.first); | 509 | 32 | size_t h2 = std::hash<Second>()(p.second); | 510 | 32 | return doris::util_hash::HashLen16(h1, h2); | 511 | 32 | } |
_ZNKSt4hashISt4pairIiN5doris10vectorized10PathInDataEEEclERKS4_ Line | Count | Source | 507 | 37.0k | size_t operator()(const pair<First, Second>& p) const { | 508 | 37.0k | size_t h1 = std::hash<First>()(p.first); | 509 | 37.0k | size_t h2 = std::hash<Second>()(p.second); | 510 | 37.0k | return doris::util_hash::HashLen16(h1, h2); | 511 | 37.0k | } |
_ZNKSt4hashISt4pairIllEEclERKS1_ Line | Count | Source | 507 | 64 | size_t operator()(const pair<First, Second>& p) const { | 508 | 64 | size_t h1 = std::hash<First>()(p.first); | 509 | 64 | size_t h2 = std::hash<Second>()(p.second); | 510 | 64 | return doris::util_hash::HashLen16(h1, h2); | 511 | 64 | } |
Unexecuted instantiation: _ZNKSt4hashISt4pairIN5doris9TUniqueIdEiEEclERKS3_ |
512 | | }; |
513 | | |
514 | | #include "common/compile_check_end.h" |