be/src/exec/sink/scale_writer_partitioning_exchanger.hpp
Line | Count | Source |
1 | | // Licensed to the Apache Software Foundation (ASF) under one |
2 | | // or more contributor license agreements. See the NOTICE file |
3 | | // distributed with this work for additional information |
4 | | // regarding copyright ownership. The ASF licenses this file |
5 | | // to you under the Apache License, Version 2.0 (the |
6 | | // "License"); you may not use this file except in compliance |
7 | | // with the License. You may obtain a copy of the License at |
8 | | // |
9 | | // http://www.apache.org/licenses/LICENSE-2.0 |
10 | | // |
11 | | // Unless required by applicable law or agreed to in writing, |
12 | | // software distributed under the License is distributed on an |
13 | | // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
14 | | // KIND, either express or implied. See the License for the |
15 | | // specific language governing permissions and limitations |
16 | | // under the License. |
17 | | |
18 | | #pragma once |
19 | | |
20 | | #include <memory> |
21 | | #include <vector> |
22 | | |
23 | | #include "core/block/block.h" |
24 | | #include "exec/connector/skewed_partition_rebalancer.h" |
25 | | #include "exec/partitioner/partitioner.h" |
26 | | |
27 | | namespace doris { |
28 | | #include "common/compile_check_begin.h" |
29 | | class ScaleWriterPartitioner final : public PartitionerBase { |
30 | | public: |
31 | | ScaleWriterPartitioner(int channel_size, int partition_count, int task_count, |
32 | | int task_bucket_count, |
33 | | long min_partition_data_processed_rebalance_threshold, |
34 | | long min_data_processed_rebalance_threshold) |
35 | 0 | : PartitionerBase(partition_count), |
36 | 0 | _channel_size(channel_size), |
37 | 0 | _partition_rebalancer(partition_count, task_count, task_bucket_count, |
38 | 0 | min_partition_data_processed_rebalance_threshold, |
39 | 0 | min_data_processed_rebalance_threshold), |
40 | 0 | _partition_row_counts(partition_count, 0), |
41 | 0 | _partition_writer_ids(partition_count, -1), |
42 | 0 | _partition_writer_indexes(partition_count, 0), |
43 | 0 | _task_count(task_count), |
44 | 0 | _task_bucket_count(task_bucket_count), |
45 | | _min_partition_data_processed_rebalance_threshold( |
46 | 0 | min_partition_data_processed_rebalance_threshold), |
47 | 0 | _min_data_processed_rebalance_threshold(min_data_processed_rebalance_threshold) { |
48 | 0 | _crc_partitioner = |
49 | 0 | std::make_unique<Crc32HashPartitioner<ShuffleChannelIds>>(_partition_count); |
50 | 0 | } |
51 | | |
52 | 0 | ~ScaleWriterPartitioner() override = default; |
53 | | |
54 | 0 | Status init(const std::vector<TExpr>& texprs) override { |
55 | 0 | return _crc_partitioner->init(texprs); |
56 | 0 | } |
57 | | |
58 | 0 | Status prepare(RuntimeState* state, const RowDescriptor& row_desc) override { |
59 | 0 | return _crc_partitioner->prepare(state, row_desc); |
60 | 0 | } |
61 | | |
62 | 0 | Status open(RuntimeState* state) override { return _crc_partitioner->open(state); } |
63 | | |
64 | 0 | Status close(RuntimeState* state) override { return _crc_partitioner->close(state); } |
65 | | |
66 | 0 | Status do_partitioning(RuntimeState* state, Block* block) const override { |
67 | 0 | _hash_vals.resize(block->rows()); |
68 | 0 | for (int partition_id = 0; partition_id < _partition_row_counts.size(); partition_id++) { |
69 | 0 | _partition_row_counts[partition_id] = 0; |
70 | 0 | _partition_writer_ids[partition_id] = -1; |
71 | 0 | } |
72 | |
|
73 | 0 | _partition_rebalancer.rebalance(); |
74 | |
|
75 | 0 | RETURN_IF_ERROR(_crc_partitioner->do_partitioning(state, block)); |
76 | 0 | const auto& channel_ids = _crc_partitioner->get_channel_ids(); |
77 | 0 | for (size_t position = 0; position < block->rows(); position++) { |
78 | 0 | auto partition_id = channel_ids[position]; |
79 | 0 | _partition_row_counts[partition_id] += 1; |
80 | | |
81 | | // Get writer id for this partition by looking at the scaling state |
82 | 0 | int writer_id = _partition_writer_ids[partition_id]; |
83 | 0 | if (writer_id == -1) { |
84 | 0 | writer_id = _get_next_writer_id(partition_id); |
85 | 0 | _partition_writer_ids[partition_id] = writer_id; |
86 | 0 | } |
87 | 0 | _hash_vals[position] = writer_id; |
88 | 0 | } |
89 | |
|
90 | 0 | for (int partition_id = 0; partition_id < _partition_row_counts.size(); partition_id++) { |
91 | 0 | _partition_rebalancer.add_partition_row_count(partition_id, |
92 | 0 | _partition_row_counts[partition_id]); |
93 | 0 | } |
94 | 0 | _partition_rebalancer.add_data_processed(block->bytes()); |
95 | |
|
96 | 0 | return Status::OK(); |
97 | 0 | } |
98 | | |
99 | 0 | const std::vector<HashValType>& get_channel_ids() const override { return _hash_vals; } |
100 | | |
101 | 0 | Status clone(RuntimeState* state, std::unique_ptr<PartitionerBase>& partitioner) override { |
102 | 0 | partitioner = std::make_unique<ScaleWriterPartitioner>( |
103 | 0 | _channel_size, (int)_partition_count, _task_count, _task_bucket_count, |
104 | 0 | _min_partition_data_processed_rebalance_threshold, |
105 | 0 | _min_data_processed_rebalance_threshold); |
106 | 0 | return Status::OK(); |
107 | 0 | } |
108 | | |
109 | | private: |
110 | 0 | int _get_next_writer_id(HashValType partition_id) const { |
111 | 0 | return _partition_rebalancer.get_task_id(partition_id, |
112 | 0 | _partition_writer_indexes[partition_id]++); |
113 | 0 | } |
114 | | |
115 | | int _channel_size; |
116 | | std::unique_ptr<PartitionerBase> _crc_partitioner; |
117 | | mutable SkewedPartitionRebalancer _partition_rebalancer; |
118 | | mutable std::vector<int> _partition_row_counts; |
119 | | mutable std::vector<int> _partition_writer_ids; |
120 | | mutable std::vector<int> _partition_writer_indexes; |
121 | | mutable std::vector<HashValType> _hash_vals; |
122 | | const int _task_count; |
123 | | const int _task_bucket_count; |
124 | | const long _min_partition_data_processed_rebalance_threshold; |
125 | | const long _min_data_processed_rebalance_threshold; |
126 | | }; |
127 | | #include "common/compile_check_end.h" |
128 | | } // namespace doris |