be/src/exec/exchange/local_exchanger.cpp
Line | Count | Source |
1 | | // Licensed to the Apache Software Foundation (ASF) under one |
2 | | // or more contributor license agreements. See the NOTICE file |
3 | | // distributed with this work for additional information |
4 | | // regarding copyright ownership. The ASF licenses this file |
5 | | // to you under the Apache License, Version 2.0 (the |
6 | | // "License"); you may not use this file except in compliance |
7 | | // with the License. You may obtain a copy of the License at |
8 | | // |
9 | | // http://www.apache.org/licenses/LICENSE-2.0 |
10 | | // |
11 | | // Unless required by applicable law or agreed to in writing, |
12 | | // software distributed under the License is distributed on an |
13 | | // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
14 | | // KIND, either express or implied. See the License for the |
15 | | // specific language governing permissions and limitations |
16 | | // under the License. |
17 | | |
18 | | #include "exec/exchange/local_exchanger.h" |
19 | | |
20 | | #include "common/cast_set.h" |
21 | | #include "common/status.h" |
22 | | #include "exec/exchange/local_exchange_sink_operator.h" |
23 | | #include "exec/exchange/local_exchange_source_operator.h" |
24 | | #include "exec/partitioner/partitioner.h" |
25 | | |
26 | | namespace doris { |
27 | | template <typename BlockType> |
28 | | void Exchanger<BlockType>::_enqueue_data_and_set_ready(int channel_id, |
29 | | LocalExchangeSinkLocalState* local_state, |
30 | 255k | BlockType&& block) { |
31 | 255k | if (local_state == nullptr) { |
32 | 0 | _enqueue_data_and_set_ready(channel_id, std::move(block)); |
33 | 0 | return; |
34 | 0 | } |
35 | | // PartitionedBlock is used by shuffle exchanger. |
36 | | // PartitionedBlock will be push into multiple queues with different row ranges, so it will be |
37 | | // referenced multiple times. Otherwise, we only ref the block once because it is only push into |
38 | | // one queue. |
39 | 255k | std::unique_lock l(*_m[channel_id]); |
40 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || |
41 | 53.9k | std::is_same_v<BroadcastBlock, BlockType>) { |
42 | 53.9k | block.first->record_channel_id(channel_id); |
43 | 201k | } else { |
44 | 201k | block->record_channel_id(channel_id); |
45 | 201k | } |
46 | | |
47 | 255k | if (_data_queue[channel_id].enqueue(std::move(block))) { |
48 | 255k | local_state->_shared_state->set_ready_to_read(channel_id); |
49 | 255k | } |
50 | 255k | } _ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_18PartitionedRowIdxsEEE27_enqueue_data_and_set_readyEiPNS_27LocalExchangeSinkLocalStateEOS7_ Line | Count | Source | 30 | 36.4k | BlockType&& block) { | 31 | 36.4k | if (local_state == nullptr) { | 32 | 0 | _enqueue_data_and_set_ready(channel_id, std::move(block)); | 33 | 0 | return; | 34 | 0 | } | 35 | | // PartitionedBlock is used by shuffle exchanger. | 36 | | // PartitionedBlock will be push into multiple queues with different row ranges, so it will be | 37 | | // referenced multiple times. Otherwise, we only ref the block once because it is only push into | 38 | | // one queue. | 39 | 36.4k | std::unique_lock l(*_m[channel_id]); | 40 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 41 | 36.4k | std::is_same_v<BroadcastBlock, BlockType>) { | 42 | 36.4k | block.first->record_channel_id(channel_id); | 43 | | } else { | 44 | | block->record_channel_id(channel_id); | 45 | | } | 46 | | | 47 | 36.4k | if (_data_queue[channel_id].enqueue(std::move(block))) { | 48 | 36.4k | local_state->_shared_state->set_ready_to_read(channel_id); | 49 | 36.4k | } | 50 | 36.4k | } |
_ZN5doris9ExchangerISt10shared_ptrINS_13ExchangerBase12BlockWrapperEEE27_enqueue_data_and_set_readyEiPNS_27LocalExchangeSinkLocalStateEOS4_ Line | Count | Source | 30 | 201k | BlockType&& block) { | 31 | 201k | if (local_state == nullptr) { | 32 | 0 | _enqueue_data_and_set_ready(channel_id, std::move(block)); | 33 | 0 | return; | 34 | 0 | } | 35 | | // PartitionedBlock is used by shuffle exchanger. | 36 | | // PartitionedBlock will be push into multiple queues with different row ranges, so it will be | 37 | | // referenced multiple times. Otherwise, we only ref the block once because it is only push into | 38 | | // one queue. | 39 | 201k | std::unique_lock l(*_m[channel_id]); | 40 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 41 | | std::is_same_v<BroadcastBlock, BlockType>) { | 42 | | block.first->record_channel_id(channel_id); | 43 | 201k | } else { | 44 | 201k | block->record_channel_id(channel_id); | 45 | 201k | } | 46 | | | 47 | 201k | if (_data_queue[channel_id].enqueue(std::move(block))) { | 48 | 201k | local_state->_shared_state->set_ready_to_read(channel_id); | 49 | 201k | } | 50 | 201k | } |
_ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_17BroadcastRowRangeEEE27_enqueue_data_and_set_readyEiPNS_27LocalExchangeSinkLocalStateEOS7_ Line | Count | Source | 30 | 17.5k | BlockType&& block) { | 31 | 17.5k | if (local_state == nullptr) { | 32 | 0 | _enqueue_data_and_set_ready(channel_id, std::move(block)); | 33 | 0 | return; | 34 | 0 | } | 35 | | // PartitionedBlock is used by shuffle exchanger. | 36 | | // PartitionedBlock will be push into multiple queues with different row ranges, so it will be | 37 | | // referenced multiple times. Otherwise, we only ref the block once because it is only push into | 38 | | // one queue. | 39 | 17.5k | std::unique_lock l(*_m[channel_id]); | 40 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 41 | 17.5k | std::is_same_v<BroadcastBlock, BlockType>) { | 42 | 17.5k | block.first->record_channel_id(channel_id); | 43 | | } else { | 44 | | block->record_channel_id(channel_id); | 45 | | } | 46 | | | 47 | 17.5k | if (_data_queue[channel_id].enqueue(std::move(block))) { | 48 | 17.5k | local_state->_shared_state->set_ready_to_read(channel_id); | 49 | 17.5k | } | 50 | 17.5k | } |
|
51 | | |
52 | | template <typename BlockType> |
53 | | bool Exchanger<BlockType>::_dequeue_data(LocalExchangeSourceLocalState* local_state, |
54 | | BlockType& block, bool* eos, Block* data_block, |
55 | 1.78M | int channel_id) { |
56 | 1.78M | if (local_state == nullptr) { |
57 | 20 | return _dequeue_data(block, eos, data_block, channel_id); |
58 | 20 | } |
59 | 1.78M | bool all_finished = _running_sink_operators == 0; |
60 | 1.78M | if (_data_queue[channel_id].try_dequeue(block)) { |
61 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || |
62 | 53.8k | std::is_same_v<BroadcastBlock, BlockType>) { |
63 | 53.8k | local_state->_shared_state->sub_mem_usage(channel_id, block.first->_allocated_bytes); |
64 | 201k | } else { |
65 | 201k | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); |
66 | 201k | data_block->swap(block->_data_block); |
67 | 201k | } |
68 | 255k | return true; |
69 | 1.52M | } else if (all_finished) { |
70 | 1.36M | *eos = true; |
71 | 1.36M | } else { |
72 | 158k | std::unique_lock l(*_m[channel_id]); |
73 | 158k | if (_data_queue[channel_id].try_dequeue(block)) { |
74 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || |
75 | 10 | std::is_same_v<BroadcastBlock, BlockType>) { |
76 | 10 | local_state->_shared_state->sub_mem_usage(channel_id, |
77 | 10 | block.first->_allocated_bytes); |
78 | 10 | } else { |
79 | 6 | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); |
80 | 6 | data_block->swap(block->_data_block); |
81 | 6 | } |
82 | 16 | return true; |
83 | 16 | } |
84 | 158k | COUNTER_UPDATE(local_state->_get_block_failed_counter, 1); |
85 | 158k | local_state->_dependency->block(); |
86 | 158k | } |
87 | 1.52M | return false; |
88 | 1.78M | } _ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_18PartitionedRowIdxsEEE13_dequeue_dataEPNS_29LocalExchangeSourceLocalStateERS7_PbPNS_5BlockEi Line | Count | Source | 55 | 204k | int channel_id) { | 56 | 204k | if (local_state == nullptr) { | 57 | 4 | return _dequeue_data(block, eos, data_block, channel_id); | 58 | 4 | } | 59 | 204k | bool all_finished = _running_sink_operators == 0; | 60 | 204k | if (_data_queue[channel_id].try_dequeue(block)) { | 61 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 62 | 36.3k | std::is_same_v<BroadcastBlock, BlockType>) { | 63 | 36.3k | local_state->_shared_state->sub_mem_usage(channel_id, block.first->_allocated_bytes); | 64 | | } else { | 65 | | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 66 | | data_block->swap(block->_data_block); | 67 | | } | 68 | 36.3k | return true; | 69 | 167k | } else if (all_finished) { | 70 | 144k | *eos = true; | 71 | 144k | } else { | 72 | 23.5k | std::unique_lock l(*_m[channel_id]); | 73 | 23.5k | if (_data_queue[channel_id].try_dequeue(block)) { | 74 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 75 | 8 | std::is_same_v<BroadcastBlock, BlockType>) { | 76 | 8 | local_state->_shared_state->sub_mem_usage(channel_id, | 77 | 8 | block.first->_allocated_bytes); | 78 | | } else { | 79 | | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 80 | | data_block->swap(block->_data_block); | 81 | | } | 82 | 8 | return true; | 83 | 8 | } | 84 | 23.5k | COUNTER_UPDATE(local_state->_get_block_failed_counter, 1); | 85 | 23.5k | local_state->_dependency->block(); | 86 | 23.5k | } | 87 | 167k | return false; | 88 | 204k | } |
_ZN5doris9ExchangerISt10shared_ptrINS_13ExchangerBase12BlockWrapperEEE13_dequeue_dataEPNS_29LocalExchangeSourceLocalStateERS4_PbPNS_5BlockEi Line | Count | Source | 55 | 1.53M | int channel_id) { | 56 | 1.53M | if (local_state == nullptr) { | 57 | 12 | return _dequeue_data(block, eos, data_block, channel_id); | 58 | 12 | } | 59 | 1.53M | bool all_finished = _running_sink_operators == 0; | 60 | 1.53M | if (_data_queue[channel_id].try_dequeue(block)) { | 61 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 62 | | std::is_same_v<BroadcastBlock, BlockType>) { | 63 | | local_state->_shared_state->sub_mem_usage(channel_id, block.first->_allocated_bytes); | 64 | 201k | } else { | 65 | 201k | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 66 | 201k | data_block->swap(block->_data_block); | 67 | 201k | } | 68 | 201k | return true; | 69 | 1.33M | } else if (all_finished) { | 70 | 1.20M | *eos = true; | 71 | 1.20M | } else { | 72 | 131k | std::unique_lock l(*_m[channel_id]); | 73 | 131k | if (_data_queue[channel_id].try_dequeue(block)) { | 74 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 75 | | std::is_same_v<BroadcastBlock, BlockType>) { | 76 | | local_state->_shared_state->sub_mem_usage(channel_id, | 77 | | block.first->_allocated_bytes); | 78 | 6 | } else { | 79 | 6 | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 80 | 6 | data_block->swap(block->_data_block); | 81 | 6 | } | 82 | 6 | return true; | 83 | 6 | } | 84 | 131k | COUNTER_UPDATE(local_state->_get_block_failed_counter, 1); | 85 | 131k | local_state->_dependency->block(); | 86 | 131k | } | 87 | 1.33M | return false; | 88 | 1.53M | } |
_ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_17BroadcastRowRangeEEE13_dequeue_dataEPNS_29LocalExchangeSourceLocalStateERS7_PbPNS_5BlockEi Line | Count | Source | 55 | 41.3k | int channel_id) { | 56 | 41.3k | if (local_state == nullptr) { | 57 | 4 | return _dequeue_data(block, eos, data_block, channel_id); | 58 | 4 | } | 59 | 41.3k | bool all_finished = _running_sink_operators == 0; | 60 | 41.3k | if (_data_queue[channel_id].try_dequeue(block)) { | 61 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 62 | 17.5k | std::is_same_v<BroadcastBlock, BlockType>) { | 63 | 17.5k | local_state->_shared_state->sub_mem_usage(channel_id, block.first->_allocated_bytes); | 64 | | } else { | 65 | | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 66 | | data_block->swap(block->_data_block); | 67 | | } | 68 | 17.5k | return true; | 69 | 23.8k | } else if (all_finished) { | 70 | 19.9k | *eos = true; | 71 | 19.9k | } else { | 72 | 3.88k | std::unique_lock l(*_m[channel_id]); | 73 | 3.88k | if (_data_queue[channel_id].try_dequeue(block)) { | 74 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || | 75 | 2 | std::is_same_v<BroadcastBlock, BlockType>) { | 76 | 2 | local_state->_shared_state->sub_mem_usage(channel_id, | 77 | 2 | block.first->_allocated_bytes); | 78 | | } else { | 79 | | local_state->_shared_state->sub_mem_usage(channel_id, block->_allocated_bytes); | 80 | | data_block->swap(block->_data_block); | 81 | | } | 82 | 2 | return true; | 83 | 2 | } | 84 | 3.88k | COUNTER_UPDATE(local_state->_get_block_failed_counter, 1); | 85 | 3.88k | local_state->_dependency->block(); | 86 | 3.88k | } | 87 | 23.8k | return false; | 88 | 41.3k | } |
|
89 | | |
90 | | template <typename BlockType> |
91 | 0 | void Exchanger<BlockType>::_enqueue_data_and_set_ready(int channel_id, BlockType&& block) { |
92 | | if constexpr (std::is_same_v<PartitionedBlock, BlockType> || |
93 | 0 | std::is_same_v<BroadcastBlock, BlockType>) { |
94 | 0 | block.first->record_channel_id(channel_id); |
95 | 0 | } else { |
96 | 0 | block->record_channel_id(channel_id); |
97 | 0 | } |
98 | 0 | _data_queue[channel_id].enqueue(std::move(block)); |
99 | 0 | } Unexecuted instantiation: _ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_18PartitionedRowIdxsEEE27_enqueue_data_and_set_readyEiOS7_ Unexecuted instantiation: _ZN5doris9ExchangerISt10shared_ptrINS_13ExchangerBase12BlockWrapperEEE27_enqueue_data_and_set_readyEiOS4_ Unexecuted instantiation: _ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_17BroadcastRowRangeEEE27_enqueue_data_and_set_readyEiOS7_ |
100 | | |
101 | | template <typename BlockType> |
102 | | bool Exchanger<BlockType>::_dequeue_data(BlockType& block, bool* eos, Block* data_block, |
103 | 20 | int channel_id) { |
104 | 20 | if (_data_queue[channel_id].try_dequeue(block)) { |
105 | | if constexpr (!std::is_same_v<PartitionedBlock, BlockType> && |
106 | 0 | !std::is_same_v<BroadcastBlock, BlockType>) { |
107 | 0 | data_block->swap(block->_data_block); |
108 | 0 | } |
109 | 0 | return true; |
110 | 0 | } |
111 | 20 | return false; |
112 | 20 | } _ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_18PartitionedRowIdxsEEE13_dequeue_dataERS7_PbPNS_5BlockEi Line | Count | Source | 103 | 4 | int channel_id) { | 104 | 4 | if (_data_queue[channel_id].try_dequeue(block)) { | 105 | | if constexpr (!std::is_same_v<PartitionedBlock, BlockType> && | 106 | | !std::is_same_v<BroadcastBlock, BlockType>) { | 107 | | data_block->swap(block->_data_block); | 108 | | } | 109 | 0 | return true; | 110 | 0 | } | 111 | 4 | return false; | 112 | 4 | } |
_ZN5doris9ExchangerISt10shared_ptrINS_13ExchangerBase12BlockWrapperEEE13_dequeue_dataERS4_PbPNS_5BlockEi Line | Count | Source | 103 | 12 | int channel_id) { | 104 | 12 | if (_data_queue[channel_id].try_dequeue(block)) { | 105 | | if constexpr (!std::is_same_v<PartitionedBlock, BlockType> && | 106 | 0 | !std::is_same_v<BroadcastBlock, BlockType>) { | 107 | 0 | data_block->swap(block->_data_block); | 108 | 0 | } | 109 | 0 | return true; | 110 | 0 | } | 111 | 12 | return false; | 112 | 12 | } |
_ZN5doris9ExchangerISt4pairISt10shared_ptrINS_13ExchangerBase12BlockWrapperEENS_17BroadcastRowRangeEEE13_dequeue_dataERS7_PbPNS_5BlockEi Line | Count | Source | 103 | 4 | int channel_id) { | 104 | 4 | if (_data_queue[channel_id].try_dequeue(block)) { | 105 | | if constexpr (!std::is_same_v<PartitionedBlock, BlockType> && | 106 | | !std::is_same_v<BroadcastBlock, BlockType>) { | 107 | | data_block->swap(block->_data_block); | 108 | | } | 109 | 0 | return true; | 110 | 0 | } | 111 | 4 | return false; | 112 | 4 | } |
|
113 | | |
114 | | Status ShuffleExchanger::sink(RuntimeState* state, Block* in_block, bool eos, Profile&& profile, |
115 | 90.6k | SinkInfo& sink_info) { |
116 | 90.6k | if (in_block->empty()) { |
117 | 72.4k | return Status::OK(); |
118 | 72.4k | } |
119 | 18.1k | { |
120 | 18.1k | SCOPED_TIMER(profile.compute_hash_value_timer); |
121 | 18.1k | RETURN_IF_ERROR(sink_info.partitioner->do_partitioning(state, in_block)); |
122 | 18.1k | } |
123 | 18.1k | { |
124 | 18.1k | SCOPED_TIMER(profile.distribute_timer); |
125 | 18.1k | RETURN_IF_ERROR(_split_rows(state, sink_info.partitioner->get_channel_ids(), in_block, |
126 | 18.1k | *sink_info.channel_id, sink_info.local_state, |
127 | 18.1k | sink_info.shuffle_idx_to_instance_idx)); |
128 | 18.1k | } |
129 | | |
130 | 18.1k | sink_info.local_state->_memory_used_counter->set( |
131 | 18.1k | sink_info.local_state->_shared_state->mem_usage); |
132 | 18.1k | return Status::OK(); |
133 | 18.1k | } |
134 | | |
135 | 73.2k | void ShuffleExchanger::close(SourceInfo&& source_info) { |
136 | 73.2k | PartitionedBlock partitioned_block; |
137 | 73.2k | bool eos; |
138 | 73.2k | Block block; |
139 | 73.2k | _data_queue[source_info.channel_id].set_eos(); |
140 | 73.2k | while (_dequeue_data(source_info.local_state, partitioned_block, &eos, &block, |
141 | 73.2k | source_info.channel_id)) { |
142 | | // do nothing |
143 | 14 | } |
144 | 73.2k | } |
145 | | |
146 | | Status ShuffleExchanger::get_block(RuntimeState* state, Block* block, bool* eos, Profile&& profile, |
147 | 95.1k | SourceInfo&& source_info) { |
148 | 95.1k | PartitionedBlock partitioned_block; |
149 | 95.1k | MutableBlock mutable_block; |
150 | | |
151 | 95.1k | auto get_data = [&]() -> Status { |
152 | 36.3k | do { |
153 | 36.3k | const auto* offset_start = partitioned_block.second.row_idxs->data() + |
154 | 36.3k | partitioned_block.second.offset_start; |
155 | 36.3k | auto block_wrapper = partitioned_block.first; |
156 | 36.3k | RETURN_IF_ERROR(mutable_block.add_rows(&block_wrapper->_data_block, offset_start, |
157 | 36.3k | offset_start + partitioned_block.second.length)); |
158 | 36.3k | } while (mutable_block.rows() < state->batch_size() && !*eos && |
159 | 36.3k | _dequeue_data(source_info.local_state, partitioned_block, eos, block, |
160 | 35.8k | source_info.channel_id)); |
161 | 29.1k | return Status::OK(); |
162 | 29.1k | }; |
163 | | |
164 | 95.1k | if (_dequeue_data(source_info.local_state, partitioned_block, eos, block, |
165 | 95.1k | source_info.channel_id)) { |
166 | 29.2k | SCOPED_TIMER(profile.copy_data_timer); |
167 | 29.2k | mutable_block = VectorizedUtils::build_mutable_mem_reuse_block( |
168 | 29.2k | block, partitioned_block.first->_data_block); |
169 | 29.2k | RETURN_IF_ERROR(get_data()); |
170 | 29.2k | } |
171 | 95.1k | return Status::OK(); |
172 | 95.1k | } |
173 | | |
174 | | Status ShuffleExchanger::_split_rows(RuntimeState* state, const std::vector<uint32_t>& channel_ids, |
175 | | Block* block, int channel_id, |
176 | | LocalExchangeSinkLocalState* local_state, |
177 | 18.0k | std::map<int, int>* shuffle_idx_to_instance_idx) { |
178 | 18.0k | if (local_state == nullptr) { |
179 | 0 | return _split_rows(state, channel_ids, block, channel_id); |
180 | 0 | } |
181 | 18.0k | const auto rows = cast_set<int32_t>(block->rows()); |
182 | 18.0k | auto row_idx = std::make_shared<PODArray<uint32_t>>(rows); |
183 | 18.0k | auto& partition_rows_histogram = _partition_rows_histogram[channel_id]; |
184 | 18.0k | { |
185 | 18.0k | partition_rows_histogram.assign(_num_partitions + 1, 0); |
186 | 8.87M | for (int32_t i = 0; i < rows; ++i) { |
187 | 8.85M | partition_rows_histogram[channel_ids[i]]++; |
188 | 8.85M | } |
189 | 217k | for (int32_t i = 1; i <= _num_partitions; ++i) { |
190 | 199k | partition_rows_histogram[i] += partition_rows_histogram[i - 1]; |
191 | 199k | } |
192 | 8.24M | for (int32_t i = rows - 1; i >= 0; --i) { |
193 | 8.23M | (*row_idx)[partition_rows_histogram[channel_ids[i]] - 1] = i; |
194 | 8.23M | partition_rows_histogram[channel_ids[i]]--; |
195 | 8.23M | } |
196 | 18.0k | } |
197 | | |
198 | 18.0k | Block data_block; |
199 | 18.0k | std::shared_ptr<BlockWrapper> new_block_wrapper; |
200 | 18.0k | if (!_free_blocks.try_dequeue(data_block)) { |
201 | 9.14k | data_block = block->clone_empty(); |
202 | 9.14k | } |
203 | 18.0k | data_block.swap(*block); |
204 | 18.0k | new_block_wrapper = |
205 | 18.0k | BlockWrapper::create_shared(std::move(data_block), local_state->_shared_state, -1); |
206 | 18.0k | if (new_block_wrapper->_data_block.empty()) { |
207 | 0 | return Status::OK(); |
208 | 0 | } |
209 | | /** |
210 | | * Data are hash-shuffled and distributed to all instances of |
211 | | * all BEs. So we need a shuffleId-To-InstanceId mapping. |
212 | | * For example, row 1 get a hash value 1 which means we should distribute to instance 1 on |
213 | | * BE 1 and row 2 get a hash value 2 which means we should distribute to instance 1 on BE 3. |
214 | | */ |
215 | 18.0k | DCHECK(shuffle_idx_to_instance_idx && shuffle_idx_to_instance_idx->size() > 0); |
216 | 18.0k | const auto& map = *shuffle_idx_to_instance_idx; |
217 | 18.0k | int32_t enqueue_rows = 0; |
218 | 198k | for (const auto& it : map) { |
219 | 18.4E | DCHECK(it.second >= 0 && it.second < _num_partitions) |
220 | 18.4E | << it.first << " : " << it.second << " " << _num_partitions; |
221 | 198k | uint32_t start = partition_rows_histogram[it.first]; |
222 | 198k | uint32_t size = partition_rows_histogram[it.first + 1] - start; |
223 | 198k | if (size > 0) { |
224 | 36.4k | enqueue_rows += size; |
225 | 36.4k | _enqueue_data_and_set_ready( |
226 | 36.4k | it.second, local_state, |
227 | 36.4k | {new_block_wrapper, |
228 | 36.4k | {.row_idxs = row_idx, .offset_start = start, .length = size}}); |
229 | 36.4k | } |
230 | 198k | } |
231 | 18.0k | if (enqueue_rows != rows) [[unlikely]] { |
232 | 1 | fmt::memory_buffer debug_string_buffer; |
233 | 1 | fmt::format_to(debug_string_buffer, "Type: {}, Local Exchange Id: {}, Shuffled Map: ", |
234 | 1 | get_exchange_type_name(get_type()), local_state->parent()->node_id()); |
235 | 3 | for (const auto& it : map) { |
236 | 3 | fmt::format_to(debug_string_buffer, "[{}:{}], ", it.first, it.second); |
237 | 3 | } |
238 | 1 | return Status::InternalError( |
239 | 1 | "Rows mismatched! Data may be lost. [Expected enqueue rows={}, Real enqueue " |
240 | 1 | "rows={}, Detail: {}]", |
241 | 1 | rows, enqueue_rows, fmt::to_string(debug_string_buffer)); |
242 | 1 | } |
243 | | |
244 | 18.0k | return Status::OK(); |
245 | 18.0k | } |
246 | | |
247 | | Status ShuffleExchanger::_split_rows(RuntimeState* state, const std::vector<uint32_t>& channel_ids, |
248 | 0 | Block* block, int channel_id) { |
249 | 0 | const auto rows = cast_set<int32_t>(block->rows()); |
250 | 0 | auto row_idx = std::make_shared<PODArray<uint32_t>>(rows); |
251 | 0 | auto& partition_rows_histogram = _partition_rows_histogram[channel_id]; |
252 | 0 | { |
253 | 0 | partition_rows_histogram.assign(_num_partitions + 1, 0); |
254 | 0 | for (int32_t i = 0; i < rows; ++i) { |
255 | 0 | partition_rows_histogram[channel_ids[i]]++; |
256 | 0 | } |
257 | 0 | for (int32_t i = 1; i <= _num_partitions; ++i) { |
258 | 0 | partition_rows_histogram[i] += partition_rows_histogram[i - 1]; |
259 | 0 | } |
260 | 0 | for (int32_t i = rows - 1; i >= 0; --i) { |
261 | 0 | (*row_idx)[partition_rows_histogram[channel_ids[i]] - 1] = i; |
262 | 0 | partition_rows_histogram[channel_ids[i]]--; |
263 | 0 | } |
264 | 0 | } |
265 | |
|
266 | 0 | Block data_block; |
267 | 0 | std::shared_ptr<BlockWrapper> new_block_wrapper; |
268 | 0 | if (!_free_blocks.try_dequeue(data_block)) { |
269 | 0 | data_block = block->clone_empty(); |
270 | 0 | } |
271 | 0 | data_block.swap(*block); |
272 | 0 | new_block_wrapper = BlockWrapper::create_shared(std::move(data_block), nullptr, -1); |
273 | 0 | if (new_block_wrapper->_data_block.empty()) { |
274 | 0 | return Status::OK(); |
275 | 0 | } |
276 | 0 | for (int i = 0; i < _num_partitions; i++) { |
277 | 0 | uint32_t start = partition_rows_histogram[i]; |
278 | 0 | uint32_t size = partition_rows_histogram[i + 1] - start; |
279 | 0 | if (size > 0) { |
280 | 0 | _enqueue_data_and_set_ready( |
281 | 0 | i, {new_block_wrapper, |
282 | 0 | {.row_idxs = row_idx, .offset_start = start, .length = size}}); |
283 | 0 | } |
284 | 0 | } |
285 | |
|
286 | 0 | return Status::OK(); |
287 | 0 | } |
288 | | |
289 | | Status PassthroughExchanger::sink(RuntimeState* state, Block* in_block, bool eos, Profile&& profile, |
290 | 302k | SinkInfo& sink_info) { |
291 | 302k | if (in_block->empty()) { |
292 | 109k | return Status::OK(); |
293 | 109k | } |
294 | 193k | Block new_block; |
295 | 193k | if (!_free_blocks.try_dequeue(new_block)) { |
296 | 75.2k | new_block = {in_block->clone_empty()}; |
297 | 75.2k | } |
298 | 193k | new_block.swap(*in_block); |
299 | 193k | auto channel_id = ((*sink_info.channel_id)++) % _num_partitions; |
300 | 193k | BlockWrapperSPtr wrapper = BlockWrapper::create_shared( |
301 | 193k | std::move(new_block), |
302 | 193k | sink_info.local_state ? sink_info.local_state->_shared_state : nullptr, channel_id); |
303 | | |
304 | 193k | _enqueue_data_and_set_ready(channel_id, sink_info.local_state, std::move(wrapper)); |
305 | | |
306 | 193k | sink_info.local_state->_memory_used_counter->set( |
307 | 193k | sink_info.local_state->_shared_state->mem_usage); |
308 | | |
309 | 193k | return Status::OK(); |
310 | 302k | } |
311 | | |
312 | 592k | void PassthroughExchanger::close(SourceInfo&& source_info) { |
313 | 592k | Block next_block; |
314 | 592k | BlockWrapperSPtr wrapper; |
315 | 592k | bool eos; |
316 | 592k | _data_queue[source_info.channel_id].set_eos(); |
317 | 592k | while (_dequeue_data(source_info.local_state, wrapper, &eos, &next_block, |
318 | 592k | source_info.channel_id)) { |
319 | | // do nothing |
320 | 220 | } |
321 | 592k | } |
322 | | |
323 | 10.4k | void PassToOneExchanger::close(SourceInfo&& source_info) { |
324 | 10.4k | Block next_block; |
325 | 10.4k | BlockWrapperSPtr wrapper; |
326 | 10.4k | bool eos; |
327 | 10.4k | _data_queue[source_info.channel_id].set_eos(); |
328 | 10.4k | while (_dequeue_data(source_info.local_state, wrapper, &eos, &next_block, |
329 | 10.4k | source_info.channel_id)) { |
330 | | // do nothing |
331 | 0 | } |
332 | 10.4k | } |
333 | | |
334 | | Status PassthroughExchanger::get_block(RuntimeState* state, Block* block, bool* eos, |
335 | 913k | Profile&& profile, SourceInfo&& source_info) { |
336 | 913k | BlockWrapperSPtr next_block; |
337 | 913k | _dequeue_data(source_info.local_state, next_block, eos, block, source_info.channel_id); |
338 | 913k | return Status::OK(); |
339 | 913k | } |
340 | | |
341 | | Status PassToOneExchanger::sink(RuntimeState* state, Block* in_block, bool eos, Profile&& profile, |
342 | 4.25k | SinkInfo& sink_info) { |
343 | 4.25k | if (in_block->empty()) { |
344 | 1.48k | return Status::OK(); |
345 | 1.48k | } |
346 | 2.77k | Block new_block; |
347 | 2.77k | if (!_free_blocks.try_dequeue(new_block)) { |
348 | 1.72k | new_block = {in_block->clone_empty()}; |
349 | 1.72k | } |
350 | 2.77k | new_block.swap(*in_block); |
351 | | |
352 | 2.77k | BlockWrapperSPtr wrapper = BlockWrapper::create_shared( |
353 | 2.77k | std::move(new_block), |
354 | 2.77k | sink_info.local_state ? sink_info.local_state->_shared_state : nullptr, 0); |
355 | 2.77k | _enqueue_data_and_set_ready(0, sink_info.local_state, std::move(wrapper)); |
356 | | |
357 | 2.77k | sink_info.local_state->_memory_used_counter->set( |
358 | 2.77k | sink_info.local_state->_shared_state->mem_usage); |
359 | | |
360 | 2.77k | return Status::OK(); |
361 | 4.25k | } |
362 | | |
363 | | Status PassToOneExchanger::get_block(RuntimeState* state, Block* block, bool* eos, |
364 | 5.14k | Profile&& profile, SourceInfo&& source_info) { |
365 | 5.14k | if (source_info.channel_id != 0) { |
366 | 3 | *eos = true; |
367 | 3 | return Status::OK(); |
368 | 3 | } |
369 | 5.14k | BlockWrapperSPtr next_block; |
370 | 5.14k | _dequeue_data(source_info.local_state, next_block, eos, block, source_info.channel_id); |
371 | 5.14k | return Status::OK(); |
372 | 5.14k | } |
373 | | |
374 | 96.7k | void ExchangerBase::finalize() { |
375 | 96.7k | DCHECK(_running_source_operators == 0); |
376 | 96.7k | Block block; |
377 | 187k | while (_free_blocks.try_dequeue(block)) { |
378 | | // do nothing |
379 | 90.3k | } |
380 | 96.7k | } |
381 | | |
382 | | Status BroadcastExchanger::sink(RuntimeState* state, Block* in_block, bool eos, Profile&& profile, |
383 | 3.13k | SinkInfo& sink_info) { |
384 | 3.13k | if (in_block->empty()) { |
385 | 1.19k | return Status::OK(); |
386 | 1.19k | } |
387 | 1.94k | Block new_block; |
388 | 1.94k | if (!_free_blocks.try_dequeue(new_block)) { |
389 | 1.55k | new_block = {in_block->clone_empty()}; |
390 | 1.55k | } |
391 | 1.94k | new_block.swap(*in_block); |
392 | 1.94k | auto wrapper = BlockWrapper::create_shared( |
393 | 1.94k | std::move(new_block), |
394 | 1.94k | sink_info.local_state ? sink_info.local_state->_shared_state : nullptr, -1); |
395 | 19.4k | for (int i = 0; i < _num_partitions; i++) { |
396 | 17.5k | _enqueue_data_and_set_ready( |
397 | 17.5k | i, sink_info.local_state, |
398 | 17.5k | {wrapper, {.offset_start = 0, .length = wrapper->_data_block.rows()}}); |
399 | 17.5k | } |
400 | | |
401 | 1.94k | return Status::OK(); |
402 | 3.13k | } |
403 | | |
404 | 9.96k | void BroadcastExchanger::close(SourceInfo&& source_info) { |
405 | 9.96k | BroadcastBlock partitioned_block; |
406 | 9.96k | bool eos; |
407 | 9.96k | Block block; |
408 | 9.96k | _data_queue[source_info.channel_id].set_eos(); |
409 | 9.96k | while (_dequeue_data(source_info.local_state, partitioned_block, &eos, &block, |
410 | 9.96k | source_info.channel_id)) { |
411 | | // do nothing |
412 | 0 | } |
413 | 9.96k | } |
414 | | |
415 | | Status BroadcastExchanger::get_block(RuntimeState* state, Block* block, bool* eos, |
416 | 31.3k | Profile&& profile, SourceInfo&& source_info) { |
417 | 31.3k | BroadcastBlock partitioned_block; |
418 | | |
419 | 31.3k | if (_dequeue_data(source_info.local_state, partitioned_block, eos, block, |
420 | 31.3k | source_info.channel_id)) { |
421 | 17.5k | SCOPED_TIMER(profile.copy_data_timer); |
422 | 17.5k | MutableBlock mutable_block = VectorizedUtils::build_mutable_mem_reuse_block( |
423 | 17.5k | block, partitioned_block.first->_data_block); |
424 | 17.5k | auto block_wrapper = partitioned_block.first; |
425 | 17.5k | RETURN_IF_ERROR(mutable_block.add_rows(&block_wrapper->_data_block, |
426 | 17.5k | partitioned_block.second.offset_start, |
427 | 17.5k | partitioned_block.second.length)); |
428 | 17.5k | } |
429 | | |
430 | 31.3k | return Status::OK(); |
431 | 31.3k | } |
432 | | |
433 | | Status AdaptivePassthroughExchanger::_passthrough_sink(RuntimeState* state, Block* in_block, |
434 | 775 | SinkInfo& sink_info) { |
435 | 775 | Block new_block; |
436 | 775 | if (!_free_blocks.try_dequeue(new_block)) { |
437 | 279 | new_block = {in_block->clone_empty()}; |
438 | 279 | } |
439 | 775 | new_block.swap(*in_block); |
440 | 775 | auto channel_id = ((*sink_info.channel_id)++) % _num_partitions; |
441 | 775 | _enqueue_data_and_set_ready( |
442 | 775 | channel_id, sink_info.local_state, |
443 | 775 | BlockWrapper::create_shared( |
444 | 775 | std::move(new_block), |
445 | 775 | sink_info.local_state ? sink_info.local_state->_shared_state : nullptr, |
446 | 775 | channel_id)); |
447 | | |
448 | 775 | sink_info.local_state->_memory_used_counter->set( |
449 | 775 | sink_info.local_state->_shared_state->mem_usage); |
450 | 775 | return Status::OK(); |
451 | 775 | } |
452 | | |
453 | | Status AdaptivePassthroughExchanger::_shuffle_sink(RuntimeState* state, Block* block, |
454 | 1.52k | SinkInfo& sink_info) { |
455 | 1.52k | std::vector<uint32_t> channel_ids; |
456 | 1.52k | const auto num_rows = block->rows(); |
457 | 1.52k | channel_ids.resize(num_rows, 0); |
458 | 1.52k | if (num_rows <= _num_partitions) { |
459 | 1.09k | std::iota(channel_ids.begin(), channel_ids.end(), 0); |
460 | 1.09k | } else { |
461 | 429 | size_t i = 0; |
462 | 8.15k | for (; i < num_rows - _num_partitions; i += _num_partitions) { |
463 | 7.72k | std::iota(channel_ids.begin() + i, channel_ids.begin() + i + _num_partitions, 0); |
464 | 7.72k | } |
465 | 429 | if (i < num_rows - 1) { |
466 | 347 | std::iota(channel_ids.begin() + i, channel_ids.end(), 0); |
467 | 347 | } |
468 | 429 | } |
469 | | |
470 | 1.52k | sink_info.local_state->_memory_used_counter->set( |
471 | 1.52k | sink_info.local_state->_shared_state->mem_usage); |
472 | 1.52k | RETURN_IF_ERROR(_split_rows(state, channel_ids, block, sink_info)); |
473 | 1.52k | return Status::OK(); |
474 | 1.52k | } |
475 | | |
476 | | Status AdaptivePassthroughExchanger::_split_rows(RuntimeState* state, |
477 | | const std::vector<uint32_t>& channel_ids, |
478 | 1.52k | Block* block, SinkInfo& sink_info) { |
479 | 1.52k | const auto rows = cast_set<int32_t>(block->rows()); |
480 | 1.52k | auto row_idx = std::make_shared<std::vector<uint32_t>>(rows); |
481 | 1.52k | auto& partition_rows_histogram = _partition_rows_histogram[*sink_info.channel_id]; |
482 | 1.52k | { |
483 | 1.52k | partition_rows_histogram.assign(_num_partitions + 1, 0); |
484 | 38.7k | for (int32_t i = 0; i < rows; ++i) { |
485 | 37.2k | partition_rows_histogram[channel_ids[i]]++; |
486 | 37.2k | } |
487 | 10.8k | for (int32_t i = 1; i <= _num_partitions; ++i) { |
488 | 9.32k | partition_rows_histogram[i] += partition_rows_histogram[i - 1]; |
489 | 9.32k | } |
490 | | |
491 | 38.7k | for (int32_t i = rows - 1; i >= 0; --i) { |
492 | 37.2k | (*row_idx)[partition_rows_histogram[channel_ids[i]] - 1] = i; |
493 | 37.2k | partition_rows_histogram[channel_ids[i]]--; |
494 | 37.2k | } |
495 | 1.52k | } |
496 | 10.8k | for (int32_t i = 0; i < _num_partitions; i++) { |
497 | 9.32k | const size_t start = partition_rows_histogram[i]; |
498 | 9.32k | const size_t size = partition_rows_histogram[i + 1] - start; |
499 | 9.32k | if (size > 0) { |
500 | 4.52k | std::unique_ptr<MutableBlock> mutable_block = |
501 | 4.52k | MutableBlock::create_unique(block->clone_empty()); |
502 | 4.52k | RETURN_IF_ERROR(mutable_block->add_rows(block, start, size)); |
503 | 4.52k | auto new_block = mutable_block->to_block(); |
504 | | |
505 | 4.52k | _enqueue_data_and_set_ready( |
506 | 4.52k | i, sink_info.local_state, |
507 | 4.52k | BlockWrapper::create_shared( |
508 | 4.52k | std::move(new_block), |
509 | 4.52k | sink_info.local_state ? sink_info.local_state->_shared_state : nullptr, |
510 | 4.52k | i)); |
511 | 4.52k | } |
512 | 9.32k | } |
513 | 1.52k | return Status::OK(); |
514 | 1.52k | } |
515 | | |
516 | | Status AdaptivePassthroughExchanger::sink(RuntimeState* state, Block* in_block, bool eos, |
517 | 6.35k | Profile&& profile, SinkInfo& sink_info) { |
518 | 6.35k | if (in_block->empty()) { |
519 | 4.04k | return Status::OK(); |
520 | 4.04k | } |
521 | 2.30k | if (_is_pass_through) { |
522 | 775 | return _passthrough_sink(state, in_block, sink_info); |
523 | 1.53k | } else { |
524 | 1.53k | if (++_total_block >= _num_partitions) { |
525 | 284 | _is_pass_through = true; |
526 | 284 | } |
527 | 1.53k | return _shuffle_sink(state, in_block, sink_info); |
528 | 1.53k | } |
529 | 2.30k | } |
530 | | |
531 | | Status AdaptivePassthroughExchanger::get_block(RuntimeState* state, Block* block, bool* eos, |
532 | 10.5k | Profile&& profile, SourceInfo&& source_info) { |
533 | 10.5k | BlockWrapperSPtr next_block; |
534 | 10.5k | _dequeue_data(source_info.local_state, next_block, eos, block, source_info.channel_id); |
535 | 10.5k | return Status::OK(); |
536 | 10.5k | } |
537 | | |
538 | 4.58k | void AdaptivePassthroughExchanger::close(SourceInfo&& source_info) { |
539 | 4.58k | Block next_block; |
540 | 4.58k | bool eos; |
541 | 4.58k | BlockWrapperSPtr wrapper; |
542 | 4.58k | _data_queue[source_info.channel_id].set_eos(); |
543 | 4.59k | while (_dequeue_data(source_info.local_state, wrapper, &eos, &next_block, |
544 | 4.59k | source_info.channel_id)) { |
545 | | // do nothing |
546 | 5 | } |
547 | 4.58k | } |
548 | | |
549 | | } // namespace doris |