Coverage Report

Created: 2025-06-18 14:41

/root/doris/be/src/olap/merger.cpp
Line
Count
Source (jump to first uncovered line)
1
// Licensed to the Apache Software Foundation (ASF) under one
2
// or more contributor license agreements.  See the NOTICE file
3
// distributed with this work for additional information
4
// regarding copyright ownership.  The ASF licenses this file
5
// to you under the Apache License, Version 2.0 (the
6
// "License"); you may not use this file except in compliance
7
// with the License.  You may obtain a copy of the License at
8
//
9
//   http://www.apache.org/licenses/LICENSE-2.0
10
//
11
// Unless required by applicable law or agreed to in writing,
12
// software distributed under the License is distributed on an
13
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
14
// KIND, either express or implied.  See the License for the
15
// specific language governing permissions and limitations
16
// under the License.
17
18
#include "olap/merger.h"
19
20
#include <gen_cpp/olap_file.pb.h>
21
#include <gen_cpp/types.pb.h>
22
#include <stddef.h>
23
#include <unistd.h>
24
25
#include <algorithm>
26
#include <iterator>
27
#include <memory>
28
#include <mutex>
29
#include <numeric>
30
#include <ostream>
31
#include <shared_mutex>
32
#include <string>
33
#include <utility>
34
#include <vector>
35
36
#include "common/config.h"
37
#include "common/logging.h"
38
#include "common/status.h"
39
#include "olap/base_tablet.h"
40
#include "olap/iterators.h"
41
#include "olap/olap_common.h"
42
#include "olap/olap_define.h"
43
#include "olap/rowid_conversion.h"
44
#include "olap/rowset/rowset.h"
45
#include "olap/rowset/rowset_meta.h"
46
#include "olap/rowset/rowset_writer.h"
47
#include "olap/rowset/segment_v2/segment_writer.h"
48
#include "olap/storage_engine.h"
49
#include "olap/tablet.h"
50
#include "olap/tablet_fwd.h"
51
#include "olap/tablet_meta.h"
52
#include "olap/tablet_reader.h"
53
#include "olap/utils.h"
54
#include "util/slice.h"
55
#include "vec/core/block.h"
56
#include "vec/olap/block_reader.h"
57
#include "vec/olap/vertical_block_reader.h"
58
#include "vec/olap/vertical_merge_iterator.h"
59
60
namespace doris {
61
62
Status Merger::vmerge_rowsets(BaseTabletSPtr tablet, ReaderType reader_type,
63
                              const TabletSchema& cur_tablet_schema,
64
                              const std::vector<RowsetReaderSharedPtr>& src_rowset_readers,
65
48
                              RowsetWriter* dst_rowset_writer, Statistics* stats_output) {
66
48
    vectorized::BlockReader reader;
67
48
    TabletReader::ReaderParams reader_params;
68
48
    reader_params.tablet = tablet;
69
48
    reader_params.reader_type = reader_type;
70
71
48
    TabletReader::ReadSource read_source;
72
48
    read_source.rs_splits.reserve(src_rowset_readers.size());
73
144
    for (const RowsetReaderSharedPtr& rs_reader : src_rowset_readers) {
74
144
        read_source.rs_splits.emplace_back(rs_reader);
75
144
    }
76
48
    read_source.fill_delete_predicates();
77
48
    reader_params.set_read_source(std::move(read_source));
78
79
48
    reader_params.version = dst_rowset_writer->version();
80
81
48
    TabletSchemaSPtr merge_tablet_schema = std::make_shared<TabletSchema>();
82
48
    merge_tablet_schema->copy_from(cur_tablet_schema);
83
84
    // Merge the columns in delete predicate that not in latest schema in to current tablet schema
85
48
    for (auto& del_pred_rs : reader_params.delete_predicates) {
86
24
        merge_tablet_schema->merge_dropped_columns(*del_pred_rs->tablet_schema());
87
24
    }
88
48
    reader_params.tablet_schema = merge_tablet_schema;
89
48
    if (!tablet->tablet_schema()->cluster_key_idxes().empty()) {
90
0
        reader_params.delete_bitmap = &tablet->tablet_meta()->delete_bitmap();
91
0
    }
92
93
48
    if (stats_output && stats_output->rowid_conversion) {
94
48
        reader_params.record_rowids = true;
95
48
        reader_params.rowid_conversion = stats_output->rowid_conversion;
96
48
        stats_output->rowid_conversion->set_dst_rowset_id(dst_rowset_writer->rowset_id());
97
48
    }
98
99
48
    reader_params.return_columns.resize(cur_tablet_schema.num_columns());
100
48
    std::iota(reader_params.return_columns.begin(), reader_params.return_columns.end(), 0);
101
48
    reader_params.origin_return_columns = &reader_params.return_columns;
102
48
    RETURN_IF_ERROR(reader.init(reader_params));
103
104
48
    vectorized::Block block = cur_tablet_schema.create_block(reader_params.return_columns);
105
48
    size_t output_rows = 0;
106
48
    bool eof = false;
107
626
    while (!eof && !ExecEnv::GetInstance()->storage_engine().stopped()) {
108
578
        auto tablet_state = tablet->tablet_state();
109
578
        if (tablet_state != TABLET_RUNNING && tablet_state != TABLET_NOTREADY) {
110
0
            tablet->clear_cache();
111
0
            return Status::Error<INTERNAL_ERROR>("tablet {} is not used any more",
112
0
                                                 tablet->tablet_id());
113
0
        }
114
115
        // Read one block from block reader
116
578
        RETURN_NOT_OK_STATUS_WITH_WARN(reader.next_block_with_aggregation(&block, &eof),
117
578
                                       "failed to read next block when merging rowsets of tablet " +
118
578
                                               std::to_string(tablet->tablet_id()));
119
578
        RETURN_NOT_OK_STATUS_WITH_WARN(dst_rowset_writer->add_block(&block),
120
578
                                       "failed to write block when merging rowsets of tablet " +
121
578
                                               std::to_string(tablet->tablet_id()));
122
123
578
        if (reader_params.record_rowids && block.rows() > 0) {
124
578
            std::vector<uint32_t> segment_num_rows;
125
578
            RETURN_IF_ERROR(dst_rowset_writer->get_segment_num_rows(&segment_num_rows));
126
578
            stats_output->rowid_conversion->add(reader.current_block_row_locations(),
127
578
                                                segment_num_rows);
128
578
        }
129
130
578
        output_rows += block.rows();
131
578
        block.clear_column_data();
132
578
    }
133
48
    if (ExecEnv::GetInstance()->storage_engine().stopped()) {
134
0
        return Status::Error<INTERNAL_ERROR>("tablet {} failed to do compaction, engine stopped",
135
0
                                             tablet->tablet_id());
136
0
    }
137
138
48
    if (stats_output != nullptr) {
139
48
        stats_output->output_rows = output_rows;
140
48
        stats_output->merged_rows = reader.merged_rows();
141
48
        stats_output->filtered_rows = reader.filtered_rows();
142
48
        stats_output->bytes_read_from_local = reader.stats().file_cache_stats.bytes_read_from_local;
143
48
        stats_output->bytes_read_from_remote =
144
48
                reader.stats().file_cache_stats.bytes_read_from_remote;
145
48
        stats_output->cached_bytes_total = reader.stats().file_cache_stats.bytes_write_into_cache;
146
48
    }
147
148
48
    RETURN_NOT_OK_STATUS_WITH_WARN(dst_rowset_writer->flush(),
149
48
                                   "failed to flush rowset when merging rowsets of tablet " +
150
48
                                           std::to_string(tablet->tablet_id()));
151
152
48
    return Status::OK();
153
48
}
154
155
// split columns into several groups, make sure all keys in one group
156
// unique_key should consider sequence&delete column
157
void Merger::vertical_split_columns(const TabletSchema& tablet_schema,
158
                                    std::vector<std::vector<uint32_t>>* column_groups,
159
90
                                    std::vector<uint32_t>* key_group_cluster_key_idxes) {
160
90
    uint32_t num_key_cols = tablet_schema.num_key_columns();
161
90
    uint32_t total_cols = tablet_schema.num_columns();
162
90
    std::vector<uint32_t> key_columns;
163
176
    for (auto i = 0; i < num_key_cols; ++i) {
164
86
        key_columns.emplace_back(i);
165
86
    }
166
    // in unique key, sequence & delete sign column should merge with key columns
167
90
    int32_t sequence_col_idx = -1;
168
90
    int32_t delete_sign_idx = -1;
169
    // in key column compaction, seq_col real index is _num_key_columns
170
    // and delete_sign column is _block->columns() - 1
171
90
    if (tablet_schema.keys_type() == KeysType::UNIQUE_KEYS) {
172
49
        if (tablet_schema.has_sequence_col()) {
173
4
            sequence_col_idx = tablet_schema.sequence_col_idx();
174
4
            key_columns.emplace_back(sequence_col_idx);
175
4
        }
176
49
        delete_sign_idx = tablet_schema.field_index(DELETE_SIGN);
177
49
        if (delete_sign_idx != -1) {
178
43
            key_columns.emplace_back(delete_sign_idx);
179
43
        }
180
49
        if (!tablet_schema.cluster_key_idxes().empty()) {
181
0
            for (const auto& cid : tablet_schema.cluster_key_idxes()) {
182
0
                auto idx = tablet_schema.field_index(cid);
183
0
                DCHECK(idx >= 0) << "could not find cluster key column with unique_id=" << cid
184
0
                                 << " in tablet schema, table_id=" << tablet_schema.table_id();
185
0
                if (idx >= num_key_cols) {
186
0
                    key_columns.emplace_back(idx);
187
0
                }
188
0
            }
189
            // tablet schema unique ids: [1, 2, 5, 3, 6, 4], [1 2] is key columns
190
            // cluster key unique ids: [3, 1, 4]
191
            // the key_columns should be [0, 1, 3, 5]
192
            // the key_group_cluster_key_idxes should be [2, 1, 3]
193
0
            for (const auto& cid : tablet_schema.cluster_key_idxes()) {
194
0
                auto idx = tablet_schema.field_index(cid);
195
0
                for (auto i = 0; i < key_columns.size(); ++i) {
196
0
                    if (idx == key_columns[i]) {
197
0
                        key_group_cluster_key_idxes->emplace_back(i);
198
0
                        break;
199
0
                    }
200
0
                }
201
0
            }
202
0
        }
203
49
    }
204
90
    VLOG_NOTICE << "sequence_col_idx=" << sequence_col_idx
205
0
                << ", delete_sign_idx=" << delete_sign_idx;
206
    // for duplicate no keys
207
90
    if (!key_columns.empty()) {
208
75
        column_groups->emplace_back(key_columns);
209
75
    }
210
211
90
    std::vector<uint32_t> value_columns;
212
213
968
    for (uint32_t i = num_key_cols; i < total_cols; ++i) {
214
878
        if (i == sequence_col_idx || i == delete_sign_idx ||
215
878
            key_columns.end() != std::find(key_columns.begin(), key_columns.end(), i)) {
216
47
            continue;
217
47
        }
218
219
831
        if (!value_columns.empty() &&
220
831
            value_columns.size() % config::vertical_compaction_num_columns_per_group == 0) {
221
140
            column_groups->push_back(value_columns);
222
140
            value_columns.clear();
223
140
        }
224
831
        value_columns.push_back(i);
225
831
    }
226
227
90
    if (!value_columns.empty()) {
228
90
        column_groups->push_back(value_columns);
229
90
    }
230
90
}
231
232
Status Merger::vertical_compact_one_group(
233
        BaseTabletSPtr tablet, ReaderType reader_type, const TabletSchema& tablet_schema,
234
        bool is_key, const std::vector<uint32_t>& column_group,
235
        vectorized::RowSourcesBuffer* row_source_buf,
236
        const std::vector<RowsetReaderSharedPtr>& src_rowset_readers,
237
        RowsetWriter* dst_rowset_writer, int64_t max_rows_per_segment, Statistics* stats_output,
238
        std::vector<uint32_t> key_group_cluster_key_idxes, int64_t batch_size,
239
283
        CompactionSampleInfo* sample_info) {
240
    // build tablet reader
241
283
    VLOG_NOTICE << "vertical compact one group, max_rows_per_segment=" << max_rows_per_segment;
242
283
    vectorized::VerticalBlockReader reader(row_source_buf);
243
283
    TabletReader::ReaderParams reader_params;
244
283
    reader_params.is_key_column_group = is_key;
245
283
    reader_params.key_group_cluster_key_idxes = key_group_cluster_key_idxes;
246
283
    reader_params.tablet = tablet;
247
283
    reader_params.reader_type = reader_type;
248
249
283
    TabletReader::ReadSource read_source;
250
283
    read_source.rs_splits.reserve(src_rowset_readers.size());
251
911
    for (const RowsetReaderSharedPtr& rs_reader : src_rowset_readers) {
252
911
        read_source.rs_splits.emplace_back(rs_reader);
253
911
    }
254
283
    read_source.fill_delete_predicates();
255
283
    reader_params.set_read_source(std::move(read_source));
256
257
283
    reader_params.version = dst_rowset_writer->version();
258
259
283
    TabletSchemaSPtr merge_tablet_schema = std::make_shared<TabletSchema>();
260
283
    merge_tablet_schema->copy_from(tablet_schema);
261
262
283
    for (auto& del_pred_rs : reader_params.delete_predicates) {
263
125
        merge_tablet_schema->merge_dropped_columns(*del_pred_rs->tablet_schema());
264
125
    }
265
266
283
    reader_params.tablet_schema = merge_tablet_schema;
267
283
    if (!tablet->tablet_schema()->cluster_key_idxes().empty()) {
268
0
        reader_params.delete_bitmap = &tablet->tablet_meta()->delete_bitmap();
269
0
    }
270
271
283
    if (is_key && stats_output && stats_output->rowid_conversion) {
272
79
        reader_params.record_rowids = true;
273
79
        reader_params.rowid_conversion = stats_output->rowid_conversion;
274
79
        stats_output->rowid_conversion->set_dst_rowset_id(dst_rowset_writer->rowset_id());
275
79
    }
276
277
283
    reader_params.return_columns = column_group;
278
283
    reader_params.origin_return_columns = &reader_params.return_columns;
279
283
    reader_params.batch_size = batch_size;
280
283
    RETURN_IF_ERROR(reader.init(reader_params, sample_info));
281
282
282
    vectorized::Block block = tablet_schema.create_block(reader_params.return_columns);
283
282
    size_t output_rows = 0;
284
282
    bool eof = false;
285
8.03k
    while (!eof && !ExecEnv::GetInstance()->storage_engine().stopped()) {
286
7.75k
        auto tablet_state = tablet->tablet_state();
287
7.75k
        if (tablet_state != TABLET_RUNNING && tablet_state != TABLET_NOTREADY) {
288
0
            tablet->clear_cache();
289
0
            return Status::Error<INTERNAL_ERROR>("tablet {} is not used any more",
290
0
                                                 tablet->tablet_id());
291
0
        }
292
        // Read one block from block reader
293
7.75k
        RETURN_NOT_OK_STATUS_WITH_WARN(reader.next_block_with_aggregation(&block, &eof),
294
7.75k
                                       "failed to read next block when merging rowsets of tablet " +
295
7.75k
                                               std::to_string(tablet->tablet_id()));
296
7.75k
        RETURN_NOT_OK_STATUS_WITH_WARN(
297
7.75k
                dst_rowset_writer->add_columns(&block, column_group, is_key, max_rows_per_segment),
298
7.75k
                "failed to write block when merging rowsets of tablet " +
299
7.75k
                        std::to_string(tablet->tablet_id()));
300
301
7.75k
        if (is_key && reader_params.record_rowids && block.rows() > 0) {
302
4.60k
            std::vector<uint32_t> segment_num_rows;
303
4.60k
            RETURN_IF_ERROR(dst_rowset_writer->get_segment_num_rows(&segment_num_rows));
304
4.60k
            stats_output->rowid_conversion->add(reader.current_block_row_locations(),
305
4.60k
                                                segment_num_rows);
306
4.60k
        }
307
7.75k
        output_rows += block.rows();
308
7.75k
        block.clear_column_data();
309
7.75k
    }
310
282
    if (ExecEnv::GetInstance()->storage_engine().stopped()) {
311
0
        return Status::Error<INTERNAL_ERROR>("tablet {} failed to do compaction, engine stopped",
312
0
                                             tablet->tablet_id());
313
0
    }
314
315
282
    if (is_key && stats_output != nullptr) {
316
78
        stats_output->output_rows = output_rows;
317
78
        stats_output->merged_rows = reader.merged_rows();
318
78
        stats_output->filtered_rows = reader.filtered_rows();
319
78
        stats_output->bytes_read_from_local = reader.stats().file_cache_stats.bytes_read_from_local;
320
78
        stats_output->bytes_read_from_remote =
321
78
                reader.stats().file_cache_stats.bytes_read_from_remote;
322
78
        stats_output->cached_bytes_total = reader.stats().file_cache_stats.bytes_write_into_cache;
323
78
    }
324
282
    RETURN_IF_ERROR(dst_rowset_writer->flush_columns(is_key));
325
326
282
    return Status::OK();
327
282
}
328
329
// for segcompaction
330
Status Merger::vertical_compact_one_group(
331
        int64_t tablet_id, ReaderType reader_type, const TabletSchema& tablet_schema, bool is_key,
332
        const std::vector<uint32_t>& column_group, vectorized::RowSourcesBuffer* row_source_buf,
333
        vectorized::VerticalBlockReader& src_block_reader,
334
        segment_v2::SegmentWriter& dst_segment_writer, Statistics* stats_output,
335
22
        uint64_t* index_size, KeyBoundsPB& key_bounds, SimpleRowIdConversion* rowid_conversion) {
336
    // TODO: record_rowids
337
22
    vectorized::Block block = tablet_schema.create_block(column_group);
338
22
    size_t output_rows = 0;
339
22
    bool eof = false;
340
138
    while (!eof && !ExecEnv::GetInstance()->storage_engine().stopped()) {
341
        // Read one block from block reader
342
116
        RETURN_NOT_OK_STATUS_WITH_WARN(src_block_reader.next_block_with_aggregation(&block, &eof),
343
116
                                       "failed to read next block when merging rowsets of tablet " +
344
116
                                               std::to_string(tablet_id));
345
116
        if (!block.rows()) {
346
0
            break;
347
0
        }
348
116
        RETURN_NOT_OK_STATUS_WITH_WARN(dst_segment_writer.append_block(&block, 0, block.rows()),
349
116
                                       "failed to write block when merging rowsets of tablet " +
350
116
                                               std::to_string(tablet_id));
351
352
116
        if (is_key && rowid_conversion != nullptr) {
353
30
            rowid_conversion->add(src_block_reader.current_block_row_locations());
354
30
        }
355
116
        output_rows += block.rows();
356
116
        block.clear_column_data();
357
116
    }
358
22
    if (ExecEnv::GetInstance()->storage_engine().stopped()) {
359
0
        return Status::Error<INTERNAL_ERROR>("tablet {} failed to do compaction, engine stopped",
360
0
                                             tablet_id);
361
0
    }
362
363
22
    if (is_key && stats_output != nullptr) {
364
11
        stats_output->output_rows = output_rows;
365
11
        stats_output->merged_rows = src_block_reader.merged_rows();
366
11
        stats_output->filtered_rows = src_block_reader.filtered_rows();
367
11
        stats_output->bytes_read_from_local =
368
11
                src_block_reader.stats().file_cache_stats.bytes_read_from_local;
369
11
        stats_output->bytes_read_from_remote =
370
11
                src_block_reader.stats().file_cache_stats.bytes_read_from_remote;
371
11
        stats_output->cached_bytes_total =
372
11
                src_block_reader.stats().file_cache_stats.bytes_write_into_cache;
373
11
    }
374
375
    // segcompaction produce only one segment at once
376
22
    RETURN_IF_ERROR(dst_segment_writer.finalize_columns_data());
377
22
    RETURN_IF_ERROR(dst_segment_writer.finalize_columns_index(index_size));
378
379
22
    if (is_key) {
380
11
        Slice min_key = dst_segment_writer.min_encoded_key();
381
11
        Slice max_key = dst_segment_writer.max_encoded_key();
382
11
        DCHECK_LE(min_key.compare(max_key), 0);
383
11
        key_bounds.set_min_key(min_key.to_string());
384
11
        key_bounds.set_max_key(max_key.to_string());
385
11
    }
386
387
22
    return Status::OK();
388
22
}
389
390
90
int64_t estimate_batch_size(int group_index, BaseTabletSPtr tablet, int64_t way_cnt) {
391
90
    std::unique_lock<std::mutex> lock(tablet->sample_info_lock);
392
90
    CompactionSampleInfo info = tablet->sample_infos[group_index];
393
90
    if (way_cnt <= 0) {
394
0
        LOG(INFO) << "estimate batch size for vertical compaction, tablet id: "
395
0
                  << tablet->tablet_id() << " way cnt: " << way_cnt;
396
0
        return 4096 - 32;
397
0
    }
398
90
    int64_t block_mem_limit = config::compaction_memory_bytes_limit / way_cnt;
399
90
    if (tablet->last_compaction_status.is<ErrorCode::MEM_LIMIT_EXCEEDED>()) {
400
0
        block_mem_limit /= 4;
401
0
    }
402
403
90
    int64_t group_data_size = 0;
404
90
    if (info.group_data_size > 0 && info.bytes > 0 && info.rows > 0) {
405
0
        float smoothing_factor = 0.5;
406
0
        group_data_size = int64_t(info.group_data_size * (1 - smoothing_factor) +
407
0
                                  info.bytes / info.rows * smoothing_factor);
408
0
        tablet->sample_infos[group_index].group_data_size = group_data_size;
409
90
    } else if (info.group_data_size > 0 && (info.bytes <= 0 || info.rows <= 0)) {
410
0
        group_data_size = info.group_data_size;
411
90
    } else if (info.group_data_size <= 0 && info.bytes > 0 && info.rows > 0) {
412
0
        group_data_size = info.bytes / info.rows;
413
0
        tablet->sample_infos[group_index].group_data_size = group_data_size;
414
90
    } else {
415
90
        LOG(INFO) << "estimate batch size for vertical compaction, tablet id: "
416
90
                  << tablet->tablet_id() << " group data size: " << info.group_data_size
417
90
                  << " row num: " << info.rows << " consume bytes: " << info.bytes;
418
90
        return 1024 - 32;
419
90
    }
420
421
0
    if (group_data_size <= 0) {
422
0
        LOG(WARNING) << "estimate batch size for vertical compaction, tablet id: "
423
0
                     << tablet->tablet_id() << " unexpected group data size: " << group_data_size;
424
0
        return 4096 - 32;
425
0
    }
426
427
0
    tablet->sample_infos[group_index].bytes = 0;
428
0
    tablet->sample_infos[group_index].rows = 0;
429
430
0
    int64_t batch_size = block_mem_limit / group_data_size;
431
0
    int64_t res = std::max(std::min(batch_size, int64_t(4096 - 32)), int64_t(32L));
432
0
    LOG(INFO) << "estimate batch size for vertical compaction, tablet id: " << tablet->tablet_id()
433
0
              << " group data size: " << info.group_data_size << " row num: " << info.rows
434
0
              << " consume bytes: " << info.bytes << " way cnt: " << way_cnt
435
0
              << " batch size: " << res;
436
0
    return res;
437
0
}
438
439
// steps to do vertical merge:
440
// 1. split columns into column groups
441
// 2. compact groups one by one, generate a row_source_buf when compact key group
442
// and use this row_source_buf to compact value column groups
443
// 3. build output rowset
444
Status Merger::vertical_merge_rowsets(BaseTabletSPtr tablet, ReaderType reader_type,
445
                                      const TabletSchema& tablet_schema,
446
                                      const std::vector<RowsetReaderSharedPtr>& src_rowset_readers,
447
                                      RowsetWriter* dst_rowset_writer, int64_t max_rows_per_segment,
448
79
                                      int64_t merge_way_num, Statistics* stats_output) {
449
79
    LOG(INFO) << "Start to do vertical compaction, tablet_id: " << tablet->tablet_id();
450
79
    std::vector<std::vector<uint32_t>> column_groups;
451
79
    std::vector<uint32_t> key_group_cluster_key_idxes;
452
79
    vertical_split_columns(tablet_schema, &column_groups, &key_group_cluster_key_idxes);
453
454
79
    vectorized::RowSourcesBuffer row_sources_buf(
455
79
            tablet->tablet_id(), dst_rowset_writer->context().tablet_path, reader_type);
456
79
    {
457
79
        std::unique_lock<std::mutex> lock(tablet->sample_info_lock);
458
79
        tablet->sample_infos.resize(column_groups.size(), {0, 0, 0});
459
79
    }
460
    // compact group one by one
461
361
    for (auto i = 0; i < column_groups.size(); ++i) {
462
283
        VLOG_NOTICE << "row source size: " << row_sources_buf.total_size();
463
283
        bool is_key = (i == 0);
464
283
        int64_t batch_size = config::compaction_batch_size != -1
465
283
                                     ? config::compaction_batch_size
466
283
                                     : estimate_batch_size(i, tablet, merge_way_num);
467
283
        CompactionSampleInfo sample_info;
468
283
        Status st = vertical_compact_one_group(
469
283
                tablet, reader_type, tablet_schema, is_key, column_groups[i], &row_sources_buf,
470
283
                src_rowset_readers, dst_rowset_writer, max_rows_per_segment, stats_output,
471
283
                key_group_cluster_key_idxes, batch_size, &sample_info);
472
283
        {
473
283
            std::unique_lock<std::mutex> lock(tablet->sample_info_lock);
474
283
            tablet->sample_infos[i] = sample_info;
475
283
        }
476
283
        RETURN_IF_ERROR(st);
477
282
        if (is_key) {
478
78
            RETURN_IF_ERROR(row_sources_buf.flush());
479
78
        }
480
282
        RETURN_IF_ERROR(row_sources_buf.seek_to_begin());
481
282
    }
482
483
    // finish compact, build output rowset
484
78
    VLOG_NOTICE << "finish compact groups";
485
78
    RETURN_IF_ERROR(dst_rowset_writer->final_flush());
486
487
78
    return Status::OK();
488
78
}
489
490
} // namespace doris