Coverage Report

Created: 2026-03-17 00:16

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
be/src/format/parquet/vparquet_column_reader.cpp
Line
Count
Source
1
// Licensed to the Apache Software Foundation (ASF) under one
2
// or more contributor license agreements.  See the NOTICE file
3
// distributed with this work for additional information
4
// regarding copyright ownership.  The ASF licenses this file
5
// to you under the Apache License, Version 2.0 (the
6
// "License"); you may not use this file except in compliance
7
// with the License.  You may obtain a copy of the License at
8
//
9
//   http://www.apache.org/licenses/LICENSE-2.0
10
//
11
// Unless required by applicable law or agreed to in writing,
12
// software distributed under the License is distributed on an
13
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
14
// KIND, either express or implied.  See the License for the
15
// specific language governing permissions and limitations
16
// under the License.
17
18
#include "format/parquet/vparquet_column_reader.h"
19
20
#include <gen_cpp/parquet_types.h>
21
#include <limits.h>
22
#include <sys/types.h>
23
24
#include <algorithm>
25
#include <utility>
26
27
#include "common/status.h"
28
#include "core/column/column.h"
29
#include "core/column/column_array.h"
30
#include "core/column/column_map.h"
31
#include "core/column/column_nullable.h"
32
#include "core/column/column_struct.h"
33
#include "core/data_type/data_type_array.h"
34
#include "core/data_type/data_type_map.h"
35
#include "core/data_type/data_type_nullable.h"
36
#include "core/data_type/data_type_struct.h"
37
#include "core/data_type/define_primitive_type.h"
38
#include "format/parquet/level_decoder.h"
39
#include "format/parquet/schema_desc.h"
40
#include "format/parquet/vparquet_column_chunk_reader.h"
41
#include "io/fs/tracing_file_reader.h"
42
#include "runtime/runtime_profile.h"
43
44
namespace doris {
45
#include "common/compile_check_begin.h"
46
static void fill_struct_null_map(FieldSchema* field, NullMap& null_map,
47
                                 const std::vector<level_t>& rep_levels,
48
10
                                 const std::vector<level_t>& def_levels) {
49
10
    size_t num_levels = def_levels.size();
50
10
    DCHECK_EQ(num_levels, rep_levels.size());
51
10
    size_t origin_size = null_map.size();
52
10
    null_map.resize(origin_size + num_levels);
53
10
    size_t pos = origin_size;
54
24
    for (size_t i = 0; i < num_levels; ++i) {
55
        // skip the levels affect its ancestor or its descendants
56
14
        if (def_levels[i] < field->repeated_parent_def_level ||
57
14
            rep_levels[i] > field->repetition_level) {
58
0
            continue;
59
0
        }
60
14
        if (def_levels[i] >= field->definition_level) {
61
14
            null_map[pos++] = 0;
62
14
        } else {
63
0
            null_map[pos++] = 1;
64
0
        }
65
14
    }
66
10
    null_map.resize(pos);
67
10
}
68
69
static void fill_array_offset(FieldSchema* field, ColumnArray::Offsets64& offsets_data,
70
                              NullMap* null_map_ptr, const std::vector<level_t>& rep_levels,
71
2
                              const std::vector<level_t>& def_levels) {
72
2
    size_t num_levels = rep_levels.size();
73
2
    DCHECK_EQ(num_levels, def_levels.size());
74
2
    size_t origin_size = offsets_data.size();
75
2
    offsets_data.resize(origin_size + num_levels);
76
2
    if (null_map_ptr != nullptr) {
77
2
        null_map_ptr->resize(origin_size + num_levels);
78
2
    }
79
2
    size_t offset_pos = origin_size - 1;
80
8
    for (size_t i = 0; i < num_levels; ++i) {
81
        // skip the levels affect its ancestor or its descendants
82
6
        if (def_levels[i] < field->repeated_parent_def_level ||
83
6
            rep_levels[i] > field->repetition_level) {
84
0
            continue;
85
0
        }
86
6
        if (rep_levels[i] == field->repetition_level) {
87
4
            offsets_data[offset_pos]++;
88
4
            continue;
89
4
        }
90
2
        offset_pos++;
91
2
        offsets_data[offset_pos] = offsets_data[offset_pos - 1];
92
2
        if (def_levels[i] >= field->definition_level) {
93
2
            offsets_data[offset_pos]++;
94
2
        }
95
2
        if (def_levels[i] >= field->definition_level - 1) {
96
2
            (*null_map_ptr)[offset_pos] = 0;
97
2
        } else {
98
0
            (*null_map_ptr)[offset_pos] = 1;
99
0
        }
100
2
    }
101
2
    offsets_data.resize(offset_pos + 1);
102
2
    if (null_map_ptr != nullptr) {
103
2
        null_map_ptr->resize(offset_pos + 1);
104
2
    }
105
2
}
106
107
Status ParquetColumnReader::create(io::FileReaderSPtr file, FieldSchema* field,
108
                                   const tparquet::RowGroup& row_group, const RowRanges& row_ranges,
109
                                   const cctz::time_zone* ctz, io::IOContext* io_ctx,
110
                                   std::unique_ptr<ParquetColumnReader>& reader,
111
                                   size_t max_buf_size,
112
                                   std::unordered_map<int, tparquet::OffsetIndex>& col_offsets,
113
                                   RuntimeState* state, bool in_collection,
114
                                   const std::set<uint64_t>& column_ids,
115
122
                                   const std::set<uint64_t>& filter_column_ids) {
116
122
    size_t total_rows = row_group.num_rows;
117
122
    if (field->data_type->get_primitive_type() == TYPE_ARRAY) {
118
2
        std::unique_ptr<ParquetColumnReader> element_reader;
119
2
        RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx,
120
2
                               element_reader, max_buf_size, col_offsets, state, true, column_ids,
121
2
                               filter_column_ids));
122
2
        auto array_reader = ArrayColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx);
123
2
        element_reader->set_column_in_nested();
124
2
        RETURN_IF_ERROR(array_reader->init(std::move(element_reader), field));
125
2
        array_reader->_filter_column_ids = filter_column_ids;
126
2
        reader.reset(array_reader.release());
127
120
    } else if (field->data_type->get_primitive_type() == TYPE_MAP) {
128
0
        std::unique_ptr<ParquetColumnReader> key_reader;
129
0
        std::unique_ptr<ParquetColumnReader> value_reader;
130
131
0
        if (column_ids.empty() ||
132
0
            column_ids.find(field->children[0].get_column_id()) != column_ids.end()) {
133
            // Create key reader
134
0
            RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx,
135
0
                                   key_reader, max_buf_size, col_offsets, state, true, column_ids,
136
0
                                   filter_column_ids));
137
0
        } else {
138
0
            auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz,
139
0
                                                                   io_ctx, &field->children[0]);
140
0
            key_reader = std::move(skip_reader);
141
0
        }
142
143
0
        if (column_ids.empty() ||
144
0
            column_ids.find(field->children[1].get_column_id()) != column_ids.end()) {
145
            // Create value reader
146
0
            RETURN_IF_ERROR(create(file, &field->children[1], row_group, row_ranges, ctz, io_ctx,
147
0
                                   value_reader, max_buf_size, col_offsets, state, true, column_ids,
148
0
                                   filter_column_ids));
149
0
        } else {
150
0
            auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz,
151
0
                                                                   io_ctx, &field->children[0]);
152
0
            value_reader = std::move(skip_reader);
153
0
        }
154
155
0
        auto map_reader = MapColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx);
156
0
        key_reader->set_column_in_nested();
157
0
        value_reader->set_column_in_nested();
158
0
        RETURN_IF_ERROR(map_reader->init(std::move(key_reader), std::move(value_reader), field));
159
0
        map_reader->_filter_column_ids = filter_column_ids;
160
0
        reader.reset(map_reader.release());
161
120
    } else if (field->data_type->get_primitive_type() == TYPE_STRUCT) {
162
10
        std::unordered_map<std::string, std::unique_ptr<ParquetColumnReader>> child_readers;
163
10
        child_readers.reserve(field->children.size());
164
10
        int non_skip_reader_idx = -1;
165
34
        for (int i = 0; i < field->children.size(); ++i) {
166
24
            auto& child = field->children[i];
167
24
            std::unique_ptr<ParquetColumnReader> child_reader;
168
24
            if (column_ids.empty() || column_ids.find(child.get_column_id()) != column_ids.end()) {
169
16
                RETURN_IF_ERROR(create(file, &child, row_group, row_ranges, ctz, io_ctx,
170
16
                                       child_reader, max_buf_size, col_offsets, state,
171
16
                                       in_collection, column_ids, filter_column_ids));
172
16
                child_readers[child.name] = std::move(child_reader);
173
                // Record the first non-SkippingReader
174
16
                if (non_skip_reader_idx == -1) {
175
10
                    non_skip_reader_idx = i;
176
10
                }
177
16
            } else {
178
8
                auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz,
179
8
                                                                       io_ctx, &child);
180
8
                skip_reader->_filter_column_ids = filter_column_ids;
181
8
                child_readers[child.name] = std::move(skip_reader);
182
8
            }
183
24
            child_readers[child.name]->set_column_in_nested();
184
24
        }
185
        // If all children are SkipReadingReader, force the first child to call create
186
10
        if (non_skip_reader_idx == -1) {
187
0
            std::unique_ptr<ParquetColumnReader> child_reader;
188
0
            RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx,
189
0
                                   child_reader, max_buf_size, col_offsets, state, in_collection,
190
0
                                   column_ids, filter_column_ids));
191
0
            child_reader->set_column_in_nested();
192
0
            child_readers[field->children[0].name] = std::move(child_reader);
193
0
        }
194
10
        auto struct_reader = StructColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx);
195
10
        RETURN_IF_ERROR(struct_reader->init(std::move(child_readers), field));
196
10
        struct_reader->_filter_column_ids = filter_column_ids;
197
10
        reader.reset(struct_reader.release());
198
110
    } else {
199
110
        auto physical_index = field->physical_column_index;
200
110
        const tparquet::OffsetIndex* offset_index =
201
110
                col_offsets.find(physical_index) != col_offsets.end() ? &col_offsets[physical_index]
202
110
                                                                      : nullptr;
203
204
110
        const tparquet::ColumnChunk& chunk = row_group.columns[physical_index];
205
110
        if (in_collection) {
206
2
            if (offset_index == nullptr) {
207
2
                auto scalar_reader = ScalarColumnReader<true, false>::create_unique(
208
2
                        row_ranges, total_rows, chunk, offset_index, ctz, io_ctx);
209
210
2
                RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state));
211
2
                scalar_reader->_filter_column_ids = filter_column_ids;
212
2
                reader.reset(scalar_reader.release());
213
2
            } else {
214
0
                auto scalar_reader = ScalarColumnReader<true, true>::create_unique(
215
0
                        row_ranges, total_rows, chunk, offset_index, ctz, io_ctx);
216
217
0
                RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state));
218
0
                scalar_reader->_filter_column_ids = filter_column_ids;
219
0
                reader.reset(scalar_reader.release());
220
0
            }
221
108
        } else {
222
108
            if (offset_index == nullptr) {
223
108
                auto scalar_reader = ScalarColumnReader<false, false>::create_unique(
224
108
                        row_ranges, total_rows, chunk, offset_index, ctz, io_ctx);
225
226
108
                RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state));
227
108
                scalar_reader->_filter_column_ids = filter_column_ids;
228
108
                reader.reset(scalar_reader.release());
229
108
            } else {
230
0
                auto scalar_reader = ScalarColumnReader<false, true>::create_unique(
231
0
                        row_ranges, total_rows, chunk, offset_index, ctz, io_ctx);
232
233
0
                RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state));
234
0
                scalar_reader->_filter_column_ids = filter_column_ids;
235
0
                reader.reset(scalar_reader.release());
236
0
            }
237
108
        }
238
110
    }
239
122
    return Status::OK();
240
122
}
241
242
void ParquetColumnReader::_generate_read_ranges(RowRange page_row_range,
243
182
                                                RowRanges* result_ranges) const {
244
182
    result_ranges->add(page_row_range);
245
182
    RowRanges::ranges_intersection(*result_ranges, _row_ranges, result_ranges);
246
182
}
247
248
template <bool IN_COLLECTION, bool OFFSET_INDEX>
249
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::init(io::FileReaderSPtr file,
250
                                                             FieldSchema* field,
251
                                                             size_t max_buf_size,
252
110
                                                             RuntimeState* state) {
253
110
    _field_schema = field;
254
110
    auto& chunk_meta = _chunk_meta.meta_data;
255
110
    int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset
256
110
                                                    : chunk_meta.data_page_offset;
257
110
    size_t chunk_len = chunk_meta.total_compressed_size;
258
110
    size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size);
259
110
    if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) &&
260
110
         typeid_cast<io::MergeRangeFileReader*>(
261
51
                 ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) ||
262
110
        typeid_cast<io::MergeRangeFileReader*>(file.get())) {
263
        // turn off prefetch data when using MergeRangeFileReader
264
110
        prefetch_buffer_size = 0;
265
110
    }
266
110
    _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len,
267
110
                                                                    prefetch_buffer_size);
268
110
    ParquetPageReadContext ctx(
269
110
            (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache);
270
271
110
    _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>(
272
110
            _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx);
273
110
    RETURN_IF_ERROR(_chunk_reader->init());
274
110
    return Status::OK();
275
110
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE
_ZN5doris18ScalarColumnReaderILb1ELb0EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE
Line
Count
Source
252
2
                                                             RuntimeState* state) {
253
2
    _field_schema = field;
254
2
    auto& chunk_meta = _chunk_meta.meta_data;
255
2
    int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset
256
2
                                                    : chunk_meta.data_page_offset;
257
2
    size_t chunk_len = chunk_meta.total_compressed_size;
258
2
    size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size);
259
2
    if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) &&
260
2
         typeid_cast<io::MergeRangeFileReader*>(
261
0
                 ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) ||
262
2
        typeid_cast<io::MergeRangeFileReader*>(file.get())) {
263
        // turn off prefetch data when using MergeRangeFileReader
264
2
        prefetch_buffer_size = 0;
265
2
    }
266
2
    _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len,
267
2
                                                                    prefetch_buffer_size);
268
2
    ParquetPageReadContext ctx(
269
2
            (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache);
270
271
2
    _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>(
272
2
            _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx);
273
2
    RETURN_IF_ERROR(_chunk_reader->init());
274
2
    return Status::OK();
275
2
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE
_ZN5doris18ScalarColumnReaderILb0ELb0EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE
Line
Count
Source
252
108
                                                             RuntimeState* state) {
253
108
    _field_schema = field;
254
108
    auto& chunk_meta = _chunk_meta.meta_data;
255
108
    int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset
256
108
                                                    : chunk_meta.data_page_offset;
257
108
    size_t chunk_len = chunk_meta.total_compressed_size;
258
108
    size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size);
259
108
    if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) &&
260
108
         typeid_cast<io::MergeRangeFileReader*>(
261
51
                 ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) ||
262
108
        typeid_cast<io::MergeRangeFileReader*>(file.get())) {
263
        // turn off prefetch data when using MergeRangeFileReader
264
108
        prefetch_buffer_size = 0;
265
108
    }
266
108
    _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len,
267
108
                                                                    prefetch_buffer_size);
268
108
    ParquetPageReadContext ctx(
269
108
            (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache);
270
271
108
    _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>(
272
108
            _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx);
273
108
    RETURN_IF_ERROR(_chunk_reader->init());
274
108
    return Status::OK();
275
108
}
276
277
template <bool IN_COLLECTION, bool OFFSET_INDEX>
278
152
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_skip_values(size_t num_values) {
279
152
    if (num_values == 0) {
280
50
        return Status::OK();
281
50
    }
282
102
    if (_chunk_reader->max_def_level() > 0) {
283
102
        LevelDecoder& def_decoder = _chunk_reader->def_level_decoder();
284
102
        size_t skipped = 0;
285
102
        size_t null_size = 0;
286
102
        size_t nonnull_size = 0;
287
217
        while (skipped < num_values) {
288
115
            level_t def_level = -1;
289
115
            size_t loop_skip = def_decoder.get_next_run(&def_level, num_values - skipped);
290
115
            if (loop_skip == 0) {
291
0
                std::stringstream ss;
292
0
                auto& bit_reader = def_decoder.rle_decoder().bit_reader();
293
0
                ss << "def_decoder buffer (hex): ";
294
0
                for (size_t i = 0; i < bit_reader.max_bytes(); ++i) {
295
0
                    ss << std::hex << std::setw(2) << std::setfill('0')
296
0
                       << static_cast<int>(bit_reader.buffer()[i]) << " ";
297
0
                }
298
0
                LOG(WARNING) << ss.str();
299
0
                return Status::InternalError("Failed to decode definition level.");
300
0
            }
301
115
            if (def_level < _field_schema->definition_level) {
302
8
                null_size += loop_skip;
303
107
            } else {
304
107
                nonnull_size += loop_skip;
305
107
            }
306
115
            skipped += loop_skip;
307
115
        }
308
102
        if (null_size > 0) {
309
5
            RETURN_IF_ERROR(_chunk_reader->skip_values(null_size, false));
310
5
        }
311
102
        if (nonnull_size > 0) {
312
101
            RETURN_IF_ERROR(_chunk_reader->skip_values(nonnull_size, true));
313
101
        }
314
102
    } else {
315
0
        RETURN_IF_ERROR(_chunk_reader->skip_values(num_values));
316
0
    }
317
102
    return Status::OK();
318
102
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE12_skip_valuesEm
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE12_skip_valuesEm
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE12_skip_valuesEm
_ZN5doris18ScalarColumnReaderILb0ELb0EE12_skip_valuesEm
Line
Count
Source
278
152
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_skip_values(size_t num_values) {
279
152
    if (num_values == 0) {
280
50
        return Status::OK();
281
50
    }
282
102
    if (_chunk_reader->max_def_level() > 0) {
283
102
        LevelDecoder& def_decoder = _chunk_reader->def_level_decoder();
284
102
        size_t skipped = 0;
285
102
        size_t null_size = 0;
286
102
        size_t nonnull_size = 0;
287
217
        while (skipped < num_values) {
288
115
            level_t def_level = -1;
289
115
            size_t loop_skip = def_decoder.get_next_run(&def_level, num_values - skipped);
290
115
            if (loop_skip == 0) {
291
0
                std::stringstream ss;
292
0
                auto& bit_reader = def_decoder.rle_decoder().bit_reader();
293
0
                ss << "def_decoder buffer (hex): ";
294
0
                for (size_t i = 0; i < bit_reader.max_bytes(); ++i) {
295
0
                    ss << std::hex << std::setw(2) << std::setfill('0')
296
0
                       << static_cast<int>(bit_reader.buffer()[i]) << " ";
297
0
                }
298
0
                LOG(WARNING) << ss.str();
299
0
                return Status::InternalError("Failed to decode definition level.");
300
0
            }
301
115
            if (def_level < _field_schema->definition_level) {
302
8
                null_size += loop_skip;
303
107
            } else {
304
107
                nonnull_size += loop_skip;
305
107
            }
306
115
            skipped += loop_skip;
307
115
        }
308
102
        if (null_size > 0) {
309
5
            RETURN_IF_ERROR(_chunk_reader->skip_values(null_size, false));
310
5
        }
311
102
        if (nonnull_size > 0) {
312
101
            RETURN_IF_ERROR(_chunk_reader->skip_values(nonnull_size, true));
313
101
        }
314
102
    } else {
315
0
        RETURN_IF_ERROR(_chunk_reader->skip_values(num_values));
316
0
    }
317
102
    return Status::OK();
318
102
}
319
320
template <bool IN_COLLECTION, bool OFFSET_INDEX>
321
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_read_values(size_t num_values,
322
                                                                     ColumnPtr& doris_column,
323
                                                                     DataTypePtr& type,
324
                                                                     FilterMap& filter_map,
325
152
                                                                     bool is_dict_filter) {
326
152
    if (num_values == 0) {
327
0
        return Status::OK();
328
0
    }
329
152
    MutableColumnPtr data_column;
330
152
    std::vector<uint16_t> null_map;
331
152
    NullMap* map_data_column = nullptr;
332
152
    if (doris_column->is_nullable()) {
333
152
        SCOPED_RAW_TIMER(&_decode_null_map_time);
334
        // doris_column either originates from a mutable block in vparquet_group_reader
335
        // or is a newly created ColumnPtr, and therefore can be modified.
336
152
        auto* nullable_column =
337
152
                assert_cast<ColumnNullable*>(const_cast<IColumn*>(doris_column.get()));
338
339
152
        data_column = nullable_column->get_nested_column_ptr();
340
152
        map_data_column = &(nullable_column->get_null_map_data());
341
152
        if (_chunk_reader->max_def_level() > 0) {
342
132
            LevelDecoder& def_decoder = _chunk_reader->def_level_decoder();
343
132
            size_t has_read = 0;
344
132
            bool prev_is_null = true;
345
264
            while (has_read < num_values) {
346
132
                level_t def_level;
347
132
                size_t loop_read = def_decoder.get_next_run(&def_level, num_values - has_read);
348
132
                if (loop_read == 0) {
349
0
                    std::stringstream ss;
350
0
                    auto& bit_reader = def_decoder.rle_decoder().bit_reader();
351
0
                    ss << "def_decoder buffer (hex): ";
352
0
                    for (size_t i = 0; i < bit_reader.max_bytes(); ++i) {
353
0
                        ss << std::hex << std::setw(2) << std::setfill('0')
354
0
                           << static_cast<int>(bit_reader.buffer()[i]) << " ";
355
0
                    }
356
0
                    LOG(WARNING) << ss.str();
357
0
                    return Status::InternalError("Failed to decode definition level.");
358
0
                }
359
360
132
                bool is_null = def_level < _field_schema->definition_level;
361
132
                if (!(prev_is_null ^ is_null)) {
362
17
                    null_map.emplace_back(0);
363
17
                }
364
132
                size_t remaining = loop_read;
365
132
                while (remaining > USHRT_MAX) {
366
0
                    null_map.emplace_back(USHRT_MAX);
367
0
                    null_map.emplace_back(0);
368
0
                    remaining -= USHRT_MAX;
369
0
                }
370
132
                null_map.emplace_back((u_short)remaining);
371
132
                prev_is_null = is_null;
372
132
                has_read += loop_read;
373
132
            }
374
132
        }
375
152
    } else {
376
0
        if (_chunk_reader->max_def_level() > 0) {
377
0
            return Status::Corruption("Not nullable column has null values in parquet file");
378
0
        }
379
0
        data_column = doris_column->assume_mutable();
380
0
    }
381
152
    if (null_map.size() == 0) {
382
20
        size_t remaining = num_values;
383
20
        while (remaining > USHRT_MAX) {
384
0
            null_map.emplace_back(USHRT_MAX);
385
0
            null_map.emplace_back(0);
386
0
            remaining -= USHRT_MAX;
387
0
        }
388
20
        null_map.emplace_back((u_short)remaining);
389
20
    }
390
152
    ColumnSelectVector select_vector;
391
152
    {
392
152
        SCOPED_RAW_TIMER(&_decode_null_map_time);
393
152
        RETURN_IF_ERROR(select_vector.init(null_map, num_values, map_data_column, &filter_map,
394
152
                                           _filter_map_index));
395
152
        _filter_map_index += num_values;
396
152
    }
397
0
    return _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter);
398
152
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb
_ZN5doris18ScalarColumnReaderILb0ELb0EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb
Line
Count
Source
325
152
                                                                     bool is_dict_filter) {
326
152
    if (num_values == 0) {
327
0
        return Status::OK();
328
0
    }
329
152
    MutableColumnPtr data_column;
330
152
    std::vector<uint16_t> null_map;
331
152
    NullMap* map_data_column = nullptr;
332
152
    if (doris_column->is_nullable()) {
333
152
        SCOPED_RAW_TIMER(&_decode_null_map_time);
334
        // doris_column either originates from a mutable block in vparquet_group_reader
335
        // or is a newly created ColumnPtr, and therefore can be modified.
336
152
        auto* nullable_column =
337
152
                assert_cast<ColumnNullable*>(const_cast<IColumn*>(doris_column.get()));
338
339
152
        data_column = nullable_column->get_nested_column_ptr();
340
152
        map_data_column = &(nullable_column->get_null_map_data());
341
152
        if (_chunk_reader->max_def_level() > 0) {
342
132
            LevelDecoder& def_decoder = _chunk_reader->def_level_decoder();
343
132
            size_t has_read = 0;
344
132
            bool prev_is_null = true;
345
264
            while (has_read < num_values) {
346
132
                level_t def_level;
347
132
                size_t loop_read = def_decoder.get_next_run(&def_level, num_values - has_read);
348
132
                if (loop_read == 0) {
349
0
                    std::stringstream ss;
350
0
                    auto& bit_reader = def_decoder.rle_decoder().bit_reader();
351
0
                    ss << "def_decoder buffer (hex): ";
352
0
                    for (size_t i = 0; i < bit_reader.max_bytes(); ++i) {
353
0
                        ss << std::hex << std::setw(2) << std::setfill('0')
354
0
                           << static_cast<int>(bit_reader.buffer()[i]) << " ";
355
0
                    }
356
0
                    LOG(WARNING) << ss.str();
357
0
                    return Status::InternalError("Failed to decode definition level.");
358
0
                }
359
360
132
                bool is_null = def_level < _field_schema->definition_level;
361
132
                if (!(prev_is_null ^ is_null)) {
362
17
                    null_map.emplace_back(0);
363
17
                }
364
132
                size_t remaining = loop_read;
365
132
                while (remaining > USHRT_MAX) {
366
0
                    null_map.emplace_back(USHRT_MAX);
367
0
                    null_map.emplace_back(0);
368
0
                    remaining -= USHRT_MAX;
369
0
                }
370
132
                null_map.emplace_back((u_short)remaining);
371
132
                prev_is_null = is_null;
372
132
                has_read += loop_read;
373
132
            }
374
132
        }
375
152
    } else {
376
0
        if (_chunk_reader->max_def_level() > 0) {
377
0
            return Status::Corruption("Not nullable column has null values in parquet file");
378
0
        }
379
0
        data_column = doris_column->assume_mutable();
380
0
    }
381
152
    if (null_map.size() == 0) {
382
20
        size_t remaining = num_values;
383
20
        while (remaining > USHRT_MAX) {
384
0
            null_map.emplace_back(USHRT_MAX);
385
0
            null_map.emplace_back(0);
386
0
            remaining -= USHRT_MAX;
387
0
        }
388
20
        null_map.emplace_back((u_short)remaining);
389
20
    }
390
152
    ColumnSelectVector select_vector;
391
152
    {
392
152
        SCOPED_RAW_TIMER(&_decode_null_map_time);
393
152
        RETURN_IF_ERROR(select_vector.init(null_map, num_values, map_data_column, &filter_map,
394
152
                                           _filter_map_index));
395
152
        _filter_map_index += num_values;
396
152
    }
397
0
    return _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter);
398
152
}
399
400
/**
401
 * Load the nested column data of complex type.
402
 * A row of complex type may be stored across two(or more) pages, and the parameter `align_rows` indicates that
403
 * whether the reader should read the remaining value of the last row in previous page.
404
 */
405
template <bool IN_COLLECTION, bool OFFSET_INDEX>
406
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_read_nested_column(
407
        ColumnPtr& doris_column, DataTypePtr& type, FilterMap& filter_map, size_t batch_size,
408
8
        size_t* read_rows, bool* eof, bool is_dict_filter) {
409
8
    _rep_levels.clear();
410
8
    _def_levels.clear();
411
412
    // Handle nullable columns
413
8
    MutableColumnPtr data_column;
414
8
    NullMap* map_data_column = nullptr;
415
8
    if (doris_column->is_nullable()) {
416
8
        SCOPED_RAW_TIMER(&_decode_null_map_time);
417
        // doris_column either originates from a mutable block in vparquet_group_reader
418
        // or is a newly created ColumnPtr, and therefore can be modified.
419
8
        auto* nullable_column =
420
8
                const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get()));
421
8
        data_column = nullable_column->get_nested_column_ptr();
422
8
        map_data_column = &(nullable_column->get_null_map_data());
423
8
    } else {
424
0
        if (_field_schema->data_type->is_nullable()) {
425
0
            return Status::Corruption("Not nullable column has null values in parquet file");
426
0
        }
427
0
        data_column = doris_column->assume_mutable();
428
0
    }
429
430
8
    std::vector<uint16_t> null_map;
431
8
    std::unordered_set<size_t> ancestor_null_indices;
432
8
    std::vector<uint8_t> nested_filter_map_data;
433
434
8
    auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) {
435
8
        RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels));
436
8
        std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>();
437
8
        if (filter_map.has_filter()) {
438
0
            RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz,
439
0
                                           _rep_levels.size(), nested_filter_map_data,
440
0
                                           &nested_filter_map));
441
0
        }
442
443
8
        null_map.clear();
444
8
        ancestor_null_indices.clear();
445
8
        RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map,
446
8
                                            ancestor_null_indices));
447
448
8
        ColumnSelectVector select_vector;
449
8
        {
450
8
            SCOPED_RAW_TIMER(&_decode_null_map_time);
451
8
            RETURN_IF_ERROR(select_vector.init(
452
8
                    null_map,
453
8
                    _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(),
454
8
                    map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices));
455
8
        }
456
457
8
        RETURN_IF_ERROR(
458
8
                _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter));
459
8
        if (ancestor_null_indices.size() != 0) {
460
0
            RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false));
461
0
        }
462
8
        if (filter_map.has_filter()) {
463
0
            auto new_rep_sz = before_rep_level_sz;
464
0
            for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) {
465
0
                if (nested_filter_map_data[idx - before_rep_level_sz]) {
466
0
                    _rep_levels[new_rep_sz] = _rep_levels[idx];
467
0
                    _def_levels[new_rep_sz] = _def_levels[idx];
468
0
                    new_rep_sz++;
469
0
                }
470
0
            }
471
0
            _rep_levels.resize(new_rep_sz);
472
0
            _def_levels.resize(new_rep_sz);
473
0
        }
474
8
        return Status::OK();
475
8
    };
Unexecuted instantiation: _ZZN5doris18ScalarColumnReaderILb1ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm
_ZZN5doris18ScalarColumnReaderILb1ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm
Line
Count
Source
434
2
    auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) {
435
2
        RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels));
436
2
        std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>();
437
2
        if (filter_map.has_filter()) {
438
0
            RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz,
439
0
                                           _rep_levels.size(), nested_filter_map_data,
440
0
                                           &nested_filter_map));
441
0
        }
442
443
2
        null_map.clear();
444
2
        ancestor_null_indices.clear();
445
2
        RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map,
446
2
                                            ancestor_null_indices));
447
448
2
        ColumnSelectVector select_vector;
449
2
        {
450
2
            SCOPED_RAW_TIMER(&_decode_null_map_time);
451
2
            RETURN_IF_ERROR(select_vector.init(
452
2
                    null_map,
453
2
                    _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(),
454
2
                    map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices));
455
2
        }
456
457
2
        RETURN_IF_ERROR(
458
2
                _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter));
459
2
        if (ancestor_null_indices.size() != 0) {
460
0
            RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false));
461
0
        }
462
2
        if (filter_map.has_filter()) {
463
0
            auto new_rep_sz = before_rep_level_sz;
464
0
            for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) {
465
0
                if (nested_filter_map_data[idx - before_rep_level_sz]) {
466
0
                    _rep_levels[new_rep_sz] = _rep_levels[idx];
467
0
                    _def_levels[new_rep_sz] = _def_levels[idx];
468
0
                    new_rep_sz++;
469
0
                }
470
0
            }
471
0
            _rep_levels.resize(new_rep_sz);
472
0
            _def_levels.resize(new_rep_sz);
473
0
        }
474
2
        return Status::OK();
475
2
    };
Unexecuted instantiation: _ZZN5doris18ScalarColumnReaderILb0ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm
_ZZN5doris18ScalarColumnReaderILb0ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm
Line
Count
Source
434
6
    auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) {
435
6
        RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels));
436
6
        std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>();
437
6
        if (filter_map.has_filter()) {
438
0
            RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz,
439
0
                                           _rep_levels.size(), nested_filter_map_data,
440
0
                                           &nested_filter_map));
441
0
        }
442
443
6
        null_map.clear();
444
6
        ancestor_null_indices.clear();
445
6
        RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map,
446
6
                                            ancestor_null_indices));
447
448
6
        ColumnSelectVector select_vector;
449
6
        {
450
6
            SCOPED_RAW_TIMER(&_decode_null_map_time);
451
6
            RETURN_IF_ERROR(select_vector.init(
452
6
                    null_map,
453
6
                    _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(),
454
6
                    map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices));
455
6
        }
456
457
6
        RETURN_IF_ERROR(
458
6
                _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter));
459
6
        if (ancestor_null_indices.size() != 0) {
460
0
            RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false));
461
0
        }
462
6
        if (filter_map.has_filter()) {
463
0
            auto new_rep_sz = before_rep_level_sz;
464
0
            for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) {
465
0
                if (nested_filter_map_data[idx - before_rep_level_sz]) {
466
0
                    _rep_levels[new_rep_sz] = _rep_levels[idx];
467
0
                    _def_levels[new_rep_sz] = _def_levels[idx];
468
0
                    new_rep_sz++;
469
0
                }
470
0
            }
471
0
            _rep_levels.resize(new_rep_sz);
472
0
            _def_levels.resize(new_rep_sz);
473
0
        }
474
6
        return Status::OK();
475
6
    };
476
477
10
    while (_current_range_idx < _row_ranges.range_size()) {
478
8
        size_t left_row =
479
8
                std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx));
480
8
        size_t right_row = std::min(left_row + batch_size - *read_rows,
481
8
                                    (size_t)_row_ranges.get_range_to(_current_range_idx));
482
8
        _current_row_index = left_row;
483
8
        RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row));
484
8
        size_t load_rows = 0;
485
8
        bool cross_page = false;
486
8
        size_t before_rep_level_sz = _rep_levels.size();
487
8
        RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row,
488
8
                                                             &load_rows, &cross_page));
489
8
        RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index));
490
8
        _filter_map_index += load_rows;
491
8
        while (cross_page) {
492
0
            before_rep_level_sz = _rep_levels.size();
493
0
            RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page));
494
0
            RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1));
495
0
        }
496
8
        *read_rows += load_rows;
497
8
        _current_row_index += load_rows;
498
8
        _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx));
499
8
        if (*read_rows == batch_size) {
500
6
            break;
501
6
        }
502
8
    }
503
8
    *eof = _current_range_idx == _row_ranges.range_size();
504
8
    return Status::OK();
505
8
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb
_ZN5doris18ScalarColumnReaderILb1ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb
Line
Count
Source
408
2
        size_t* read_rows, bool* eof, bool is_dict_filter) {
409
2
    _rep_levels.clear();
410
2
    _def_levels.clear();
411
412
    // Handle nullable columns
413
2
    MutableColumnPtr data_column;
414
2
    NullMap* map_data_column = nullptr;
415
2
    if (doris_column->is_nullable()) {
416
2
        SCOPED_RAW_TIMER(&_decode_null_map_time);
417
        // doris_column either originates from a mutable block in vparquet_group_reader
418
        // or is a newly created ColumnPtr, and therefore can be modified.
419
2
        auto* nullable_column =
420
2
                const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get()));
421
2
        data_column = nullable_column->get_nested_column_ptr();
422
2
        map_data_column = &(nullable_column->get_null_map_data());
423
2
    } else {
424
0
        if (_field_schema->data_type->is_nullable()) {
425
0
            return Status::Corruption("Not nullable column has null values in parquet file");
426
0
        }
427
0
        data_column = doris_column->assume_mutable();
428
0
    }
429
430
2
    std::vector<uint16_t> null_map;
431
2
    std::unordered_set<size_t> ancestor_null_indices;
432
2
    std::vector<uint8_t> nested_filter_map_data;
433
434
2
    auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) {
435
2
        RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels));
436
2
        std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>();
437
2
        if (filter_map.has_filter()) {
438
2
            RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz,
439
2
                                           _rep_levels.size(), nested_filter_map_data,
440
2
                                           &nested_filter_map));
441
2
        }
442
443
2
        null_map.clear();
444
2
        ancestor_null_indices.clear();
445
2
        RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map,
446
2
                                            ancestor_null_indices));
447
448
2
        ColumnSelectVector select_vector;
449
2
        {
450
2
            SCOPED_RAW_TIMER(&_decode_null_map_time);
451
2
            RETURN_IF_ERROR(select_vector.init(
452
2
                    null_map,
453
2
                    _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(),
454
2
                    map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices));
455
2
        }
456
457
2
        RETURN_IF_ERROR(
458
2
                _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter));
459
2
        if (ancestor_null_indices.size() != 0) {
460
2
            RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false));
461
2
        }
462
2
        if (filter_map.has_filter()) {
463
2
            auto new_rep_sz = before_rep_level_sz;
464
2
            for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) {
465
2
                if (nested_filter_map_data[idx - before_rep_level_sz]) {
466
2
                    _rep_levels[new_rep_sz] = _rep_levels[idx];
467
2
                    _def_levels[new_rep_sz] = _def_levels[idx];
468
2
                    new_rep_sz++;
469
2
                }
470
2
            }
471
2
            _rep_levels.resize(new_rep_sz);
472
2
            _def_levels.resize(new_rep_sz);
473
2
        }
474
2
        return Status::OK();
475
2
    };
476
477
2
    while (_current_range_idx < _row_ranges.range_size()) {
478
2
        size_t left_row =
479
2
                std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx));
480
2
        size_t right_row = std::min(left_row + batch_size - *read_rows,
481
2
                                    (size_t)_row_ranges.get_range_to(_current_range_idx));
482
2
        _current_row_index = left_row;
483
2
        RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row));
484
2
        size_t load_rows = 0;
485
2
        bool cross_page = false;
486
2
        size_t before_rep_level_sz = _rep_levels.size();
487
2
        RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row,
488
2
                                                             &load_rows, &cross_page));
489
2
        RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index));
490
2
        _filter_map_index += load_rows;
491
2
        while (cross_page) {
492
0
            before_rep_level_sz = _rep_levels.size();
493
0
            RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page));
494
0
            RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1));
495
0
        }
496
2
        *read_rows += load_rows;
497
2
        _current_row_index += load_rows;
498
2
        _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx));
499
2
        if (*read_rows == batch_size) {
500
2
            break;
501
2
        }
502
2
    }
503
2
    *eof = _current_range_idx == _row_ranges.range_size();
504
2
    return Status::OK();
505
2
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb
_ZN5doris18ScalarColumnReaderILb0ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb
Line
Count
Source
408
6
        size_t* read_rows, bool* eof, bool is_dict_filter) {
409
6
    _rep_levels.clear();
410
6
    _def_levels.clear();
411
412
    // Handle nullable columns
413
6
    MutableColumnPtr data_column;
414
6
    NullMap* map_data_column = nullptr;
415
6
    if (doris_column->is_nullable()) {
416
6
        SCOPED_RAW_TIMER(&_decode_null_map_time);
417
        // doris_column either originates from a mutable block in vparquet_group_reader
418
        // or is a newly created ColumnPtr, and therefore can be modified.
419
6
        auto* nullable_column =
420
6
                const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get()));
421
6
        data_column = nullable_column->get_nested_column_ptr();
422
6
        map_data_column = &(nullable_column->get_null_map_data());
423
6
    } else {
424
0
        if (_field_schema->data_type->is_nullable()) {
425
0
            return Status::Corruption("Not nullable column has null values in parquet file");
426
0
        }
427
0
        data_column = doris_column->assume_mutable();
428
0
    }
429
430
6
    std::vector<uint16_t> null_map;
431
6
    std::unordered_set<size_t> ancestor_null_indices;
432
6
    std::vector<uint8_t> nested_filter_map_data;
433
434
6
    auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) {
435
6
        RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels));
436
6
        std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>();
437
6
        if (filter_map.has_filter()) {
438
6
            RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz,
439
6
                                           _rep_levels.size(), nested_filter_map_data,
440
6
                                           &nested_filter_map));
441
6
        }
442
443
6
        null_map.clear();
444
6
        ancestor_null_indices.clear();
445
6
        RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map,
446
6
                                            ancestor_null_indices));
447
448
6
        ColumnSelectVector select_vector;
449
6
        {
450
6
            SCOPED_RAW_TIMER(&_decode_null_map_time);
451
6
            RETURN_IF_ERROR(select_vector.init(
452
6
                    null_map,
453
6
                    _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(),
454
6
                    map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices));
455
6
        }
456
457
6
        RETURN_IF_ERROR(
458
6
                _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter));
459
6
        if (ancestor_null_indices.size() != 0) {
460
6
            RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false));
461
6
        }
462
6
        if (filter_map.has_filter()) {
463
6
            auto new_rep_sz = before_rep_level_sz;
464
6
            for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) {
465
6
                if (nested_filter_map_data[idx - before_rep_level_sz]) {
466
6
                    _rep_levels[new_rep_sz] = _rep_levels[idx];
467
6
                    _def_levels[new_rep_sz] = _def_levels[idx];
468
6
                    new_rep_sz++;
469
6
                }
470
6
            }
471
6
            _rep_levels.resize(new_rep_sz);
472
6
            _def_levels.resize(new_rep_sz);
473
6
        }
474
6
        return Status::OK();
475
6
    };
476
477
8
    while (_current_range_idx < _row_ranges.range_size()) {
478
6
        size_t left_row =
479
6
                std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx));
480
6
        size_t right_row = std::min(left_row + batch_size - *read_rows,
481
6
                                    (size_t)_row_ranges.get_range_to(_current_range_idx));
482
6
        _current_row_index = left_row;
483
6
        RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row));
484
6
        size_t load_rows = 0;
485
6
        bool cross_page = false;
486
6
        size_t before_rep_level_sz = _rep_levels.size();
487
6
        RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row,
488
6
                                                             &load_rows, &cross_page));
489
6
        RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index));
490
6
        _filter_map_index += load_rows;
491
6
        while (cross_page) {
492
0
            before_rep_level_sz = _rep_levels.size();
493
0
            RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page));
494
0
            RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1));
495
0
        }
496
6
        *read_rows += load_rows;
497
6
        _current_row_index += load_rows;
498
6
        _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx));
499
6
        if (*read_rows == batch_size) {
500
4
            break;
501
4
        }
502
6
    }
503
6
    *eof = _current_range_idx == _row_ranges.range_size();
504
6
    return Status::OK();
505
6
}
506
507
template <bool IN_COLLECTION, bool OFFSET_INDEX>
508
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::read_dict_values_to_column(
509
2
        MutableColumnPtr& doris_column, bool* has_dict) {
510
2
    bool loaded;
511
2
    RETURN_IF_ERROR(_try_load_dict_page(&loaded, has_dict));
512
2
    if (loaded && *has_dict) {
513
2
        return _chunk_reader->read_dict_values_to_column(doris_column);
514
2
    }
515
0
    return Status::OK();
516
2
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb
_ZN5doris18ScalarColumnReaderILb0ELb0EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb
Line
Count
Source
509
2
        MutableColumnPtr& doris_column, bool* has_dict) {
510
2
    bool loaded;
511
2
    RETURN_IF_ERROR(_try_load_dict_page(&loaded, has_dict));
512
2
    if (loaded && *has_dict) {
513
2
        return _chunk_reader->read_dict_values_to_column(doris_column);
514
2
    }
515
0
    return Status::OK();
516
2
}
517
template <bool IN_COLLECTION, bool OFFSET_INDEX>
518
Result<MutableColumnPtr>
519
ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::convert_dict_column_to_string_column(
520
0
        const ColumnInt32* dict_column) {
521
0
    return _chunk_reader->convert_dict_column_to_string_column(dict_column);
522
0
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb0EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE
523
524
template <bool IN_COLLECTION, bool OFFSET_INDEX>
525
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_try_load_dict_page(bool* loaded,
526
2
                                                                            bool* has_dict) {
527
    // _chunk_reader init will load first page header to check whether has dict page
528
2
    *loaded = true;
529
2
    *has_dict = _chunk_reader->has_dict();
530
2
    return Status::OK();
531
2
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE19_try_load_dict_pageEPbS2_
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE19_try_load_dict_pageEPbS2_
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE19_try_load_dict_pageEPbS2_
_ZN5doris18ScalarColumnReaderILb0ELb0EE19_try_load_dict_pageEPbS2_
Line
Count
Source
526
2
                                                                            bool* has_dict) {
527
    // _chunk_reader init will load first page header to check whether has dict page
528
2
    *loaded = true;
529
2
    *has_dict = _chunk_reader->has_dict();
530
2
    return Status::OK();
531
2
}
532
533
template <bool IN_COLLECTION, bool OFFSET_INDEX>
534
Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::read_column_data(
535
        ColumnPtr& doris_column, const DataTypePtr& type,
536
        const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map,
537
        size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter,
538
190
        int64_t real_column_size) {
539
190
    if (_converter == nullptr) {
540
105
        _converter = parquet::PhysicalToLogicalConverter::get_converter(
541
105
                _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter);
542
105
        if (!_converter->support()) {
543
0
            return Status::InternalError(
544
0
                    "The column type of '{}' is not supported: {}, is_dict_filter: {}, "
545
0
                    "src_logical_type: {}, dst_logical_type: {}",
546
0
                    _field_schema->name, _converter->get_error_msg(), is_dict_filter,
547
0
                    _field_schema->data_type->get_name(), type->get_name());
548
0
        }
549
105
    }
550
    // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column?
551
190
    ColumnPtr resolved_column =
552
190
            _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type,
553
190
                                            doris_column, type, is_dict_filter);
554
190
    DataTypePtr& resolved_type = _converter->get_physical_type();
555
556
190
    _def_levels.clear();
557
190
    _rep_levels.clear();
558
190
    *read_rows = 0;
559
560
190
    if (_in_nested) {
561
8
        RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size,
562
8
                                            read_rows, eof, is_dict_filter));
563
8
        return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
564
8
                                   is_dict_filter);
565
8
    }
566
567
182
    int64_t right_row = 0;
568
182
    if constexpr (OFFSET_INDEX == false) {
569
182
        RETURN_IF_ERROR(_chunk_reader->parse_page_header());
570
182
        right_row = _chunk_reader->page_end_row();
571
182
    } else {
572
0
        right_row = _chunk_reader->page_end_row();
573
0
    }
574
575
182
    do {
576
        // generate the row ranges that should be read
577
182
        RowRanges read_ranges;
578
182
        _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges);
579
182
        if (read_ranges.count() == 0) {
580
            // skip the whole page
581
63
            _current_row_index = right_row;
582
119
        } else {
583
119
            bool skip_whole_batch = false;
584
            // Determining whether to skip page or batch will increase the calculation time.
585
            // When the filtering effect is greater than 60%, it is possible to skip the page or batch.
586
119
            if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) {
587
                // lazy read
588
0
                size_t remaining_num_values = read_ranges.count();
589
0
                if (batch_size >= remaining_num_values &&
590
0
                    filter_map.can_filter_all(remaining_num_values, _filter_map_index)) {
591
                    // We can skip the whole page if the remaining values are filtered by predicate columns
592
0
                    _filter_map_index += remaining_num_values;
593
0
                    _current_row_index = right_row;
594
0
                    *read_rows = remaining_num_values;
595
0
                    break;
596
0
                }
597
0
                skip_whole_batch = batch_size <= remaining_num_values &&
598
0
                                   filter_map.can_filter_all(batch_size, _filter_map_index);
599
0
                if (skip_whole_batch) {
600
0
                    _filter_map_index += batch_size;
601
0
                }
602
0
            }
603
            // load page data to decode or skip values
604
119
            RETURN_IF_ERROR(_chunk_reader->parse_page_header());
605
119
            RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent());
606
119
            size_t has_read = 0;
607
248
            for (size_t idx = 0; idx < read_ranges.range_size(); idx++) {
608
152
                auto range = read_ranges.get_range(idx);
609
                // generate the skipped values
610
152
                size_t skip_values = range.from() - _current_row_index;
611
152
                RETURN_IF_ERROR(_skip_values(skip_values));
612
152
                _current_row_index += skip_values;
613
                // generate the read values
614
152
                size_t read_values =
615
152
                        std::min((size_t)(range.to() - range.from()), batch_size - has_read);
616
152
                if (skip_whole_batch) {
617
0
                    RETURN_IF_ERROR(_skip_values(read_values));
618
152
                } else {
619
152
                    RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type,
620
152
                                                 filter_map, is_dict_filter));
621
152
                }
622
152
                has_read += read_values;
623
152
                *read_rows += read_values;
624
152
                _current_row_index += read_values;
625
152
                if (has_read == batch_size) {
626
23
                    break;
627
23
                }
628
152
            }
629
119
        }
630
182
    } while (false);
631
632
182
    if (right_row == _current_row_index) {
633
97
        if (!_chunk_reader->has_next_page()) {
634
97
            *eof = true;
635
97
        } else {
636
0
            RETURN_IF_ERROR(_chunk_reader->next_page());
637
0
        }
638
97
    }
639
640
182
    return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
641
182
                               is_dict_filter);
642
182
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl
_ZN5doris18ScalarColumnReaderILb1ELb0EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl
Line
Count
Source
538
2
        int64_t real_column_size) {
539
2
    if (_converter == nullptr) {
540
2
        _converter = parquet::PhysicalToLogicalConverter::get_converter(
541
2
                _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter);
542
2
        if (!_converter->support()) {
543
0
            return Status::InternalError(
544
0
                    "The column type of '{}' is not supported: {}, is_dict_filter: {}, "
545
0
                    "src_logical_type: {}, dst_logical_type: {}",
546
0
                    _field_schema->name, _converter->get_error_msg(), is_dict_filter,
547
0
                    _field_schema->data_type->get_name(), type->get_name());
548
0
        }
549
2
    }
550
    // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column?
551
2
    ColumnPtr resolved_column =
552
2
            _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type,
553
2
                                            doris_column, type, is_dict_filter);
554
2
    DataTypePtr& resolved_type = _converter->get_physical_type();
555
556
2
    _def_levels.clear();
557
2
    _rep_levels.clear();
558
2
    *read_rows = 0;
559
560
2
    if (_in_nested) {
561
2
        RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size,
562
2
                                            read_rows, eof, is_dict_filter));
563
2
        return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
564
2
                                   is_dict_filter);
565
2
    }
566
567
0
    int64_t right_row = 0;
568
0
    if constexpr (OFFSET_INDEX == false) {
569
0
        RETURN_IF_ERROR(_chunk_reader->parse_page_header());
570
0
        right_row = _chunk_reader->page_end_row();
571
    } else {
572
        right_row = _chunk_reader->page_end_row();
573
    }
574
575
0
    do {
576
        // generate the row ranges that should be read
577
0
        RowRanges read_ranges;
578
0
        _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges);
579
0
        if (read_ranges.count() == 0) {
580
            // skip the whole page
581
0
            _current_row_index = right_row;
582
0
        } else {
583
0
            bool skip_whole_batch = false;
584
            // Determining whether to skip page or batch will increase the calculation time.
585
            // When the filtering effect is greater than 60%, it is possible to skip the page or batch.
586
0
            if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) {
587
                // lazy read
588
0
                size_t remaining_num_values = read_ranges.count();
589
0
                if (batch_size >= remaining_num_values &&
590
0
                    filter_map.can_filter_all(remaining_num_values, _filter_map_index)) {
591
                    // We can skip the whole page if the remaining values are filtered by predicate columns
592
0
                    _filter_map_index += remaining_num_values;
593
0
                    _current_row_index = right_row;
594
0
                    *read_rows = remaining_num_values;
595
0
                    break;
596
0
                }
597
0
                skip_whole_batch = batch_size <= remaining_num_values &&
598
0
                                   filter_map.can_filter_all(batch_size, _filter_map_index);
599
0
                if (skip_whole_batch) {
600
0
                    _filter_map_index += batch_size;
601
0
                }
602
0
            }
603
            // load page data to decode or skip values
604
0
            RETURN_IF_ERROR(_chunk_reader->parse_page_header());
605
0
            RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent());
606
0
            size_t has_read = 0;
607
0
            for (size_t idx = 0; idx < read_ranges.range_size(); idx++) {
608
0
                auto range = read_ranges.get_range(idx);
609
                // generate the skipped values
610
0
                size_t skip_values = range.from() - _current_row_index;
611
0
                RETURN_IF_ERROR(_skip_values(skip_values));
612
0
                _current_row_index += skip_values;
613
                // generate the read values
614
0
                size_t read_values =
615
0
                        std::min((size_t)(range.to() - range.from()), batch_size - has_read);
616
0
                if (skip_whole_batch) {
617
0
                    RETURN_IF_ERROR(_skip_values(read_values));
618
0
                } else {
619
0
                    RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type,
620
0
                                                 filter_map, is_dict_filter));
621
0
                }
622
0
                has_read += read_values;
623
0
                *read_rows += read_values;
624
0
                _current_row_index += read_values;
625
0
                if (has_read == batch_size) {
626
0
                    break;
627
0
                }
628
0
            }
629
0
        }
630
0
    } while (false);
631
632
0
    if (right_row == _current_row_index) {
633
0
        if (!_chunk_reader->has_next_page()) {
634
0
            *eof = true;
635
0
        } else {
636
0
            RETURN_IF_ERROR(_chunk_reader->next_page());
637
0
        }
638
0
    }
639
640
0
    return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
641
0
                               is_dict_filter);
642
0
}
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl
_ZN5doris18ScalarColumnReaderILb0ELb0EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl
Line
Count
Source
538
188
        int64_t real_column_size) {
539
188
    if (_converter == nullptr) {
540
103
        _converter = parquet::PhysicalToLogicalConverter::get_converter(
541
103
                _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter);
542
103
        if (!_converter->support()) {
543
0
            return Status::InternalError(
544
0
                    "The column type of '{}' is not supported: {}, is_dict_filter: {}, "
545
0
                    "src_logical_type: {}, dst_logical_type: {}",
546
0
                    _field_schema->name, _converter->get_error_msg(), is_dict_filter,
547
0
                    _field_schema->data_type->get_name(), type->get_name());
548
0
        }
549
103
    }
550
    // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column?
551
188
    ColumnPtr resolved_column =
552
188
            _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type,
553
188
                                            doris_column, type, is_dict_filter);
554
188
    DataTypePtr& resolved_type = _converter->get_physical_type();
555
556
188
    _def_levels.clear();
557
188
    _rep_levels.clear();
558
188
    *read_rows = 0;
559
560
188
    if (_in_nested) {
561
6
        RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size,
562
6
                                            read_rows, eof, is_dict_filter));
563
6
        return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
564
6
                                   is_dict_filter);
565
6
    }
566
567
182
    int64_t right_row = 0;
568
182
    if constexpr (OFFSET_INDEX == false) {
569
182
        RETURN_IF_ERROR(_chunk_reader->parse_page_header());
570
182
        right_row = _chunk_reader->page_end_row();
571
    } else {
572
        right_row = _chunk_reader->page_end_row();
573
    }
574
575
182
    do {
576
        // generate the row ranges that should be read
577
182
        RowRanges read_ranges;
578
182
        _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges);
579
182
        if (read_ranges.count() == 0) {
580
            // skip the whole page
581
63
            _current_row_index = right_row;
582
119
        } else {
583
119
            bool skip_whole_batch = false;
584
            // Determining whether to skip page or batch will increase the calculation time.
585
            // When the filtering effect is greater than 60%, it is possible to skip the page or batch.
586
119
            if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) {
587
                // lazy read
588
0
                size_t remaining_num_values = read_ranges.count();
589
0
                if (batch_size >= remaining_num_values &&
590
0
                    filter_map.can_filter_all(remaining_num_values, _filter_map_index)) {
591
                    // We can skip the whole page if the remaining values are filtered by predicate columns
592
0
                    _filter_map_index += remaining_num_values;
593
0
                    _current_row_index = right_row;
594
0
                    *read_rows = remaining_num_values;
595
0
                    break;
596
0
                }
597
0
                skip_whole_batch = batch_size <= remaining_num_values &&
598
0
                                   filter_map.can_filter_all(batch_size, _filter_map_index);
599
0
                if (skip_whole_batch) {
600
0
                    _filter_map_index += batch_size;
601
0
                }
602
0
            }
603
            // load page data to decode or skip values
604
119
            RETURN_IF_ERROR(_chunk_reader->parse_page_header());
605
119
            RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent());
606
119
            size_t has_read = 0;
607
248
            for (size_t idx = 0; idx < read_ranges.range_size(); idx++) {
608
152
                auto range = read_ranges.get_range(idx);
609
                // generate the skipped values
610
152
                size_t skip_values = range.from() - _current_row_index;
611
152
                RETURN_IF_ERROR(_skip_values(skip_values));
612
152
                _current_row_index += skip_values;
613
                // generate the read values
614
152
                size_t read_values =
615
152
                        std::min((size_t)(range.to() - range.from()), batch_size - has_read);
616
152
                if (skip_whole_batch) {
617
0
                    RETURN_IF_ERROR(_skip_values(read_values));
618
152
                } else {
619
152
                    RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type,
620
152
                                                 filter_map, is_dict_filter));
621
152
                }
622
152
                has_read += read_values;
623
152
                *read_rows += read_values;
624
152
                _current_row_index += read_values;
625
152
                if (has_read == batch_size) {
626
23
                    break;
627
23
                }
628
152
            }
629
119
        }
630
182
    } while (false);
631
632
182
    if (right_row == _current_row_index) {
633
97
        if (!_chunk_reader->has_next_page()) {
634
97
            *eof = true;
635
97
        } else {
636
0
            RETURN_IF_ERROR(_chunk_reader->next_page());
637
0
        }
638
97
    }
639
640
182
    return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column,
641
182
                               is_dict_filter);
642
182
}
643
644
Status ArrayColumnReader::init(std::unique_ptr<ParquetColumnReader> element_reader,
645
2
                               FieldSchema* field) {
646
2
    _field_schema = field;
647
2
    _element_reader = std::move(element_reader);
648
2
    return Status::OK();
649
2
}
650
651
Status ArrayColumnReader::read_column_data(
652
        ColumnPtr& doris_column, const DataTypePtr& type,
653
        const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map,
654
        size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter,
655
2
        int64_t real_column_size) {
656
2
    MutableColumnPtr data_column;
657
2
    NullMap* null_map_ptr = nullptr;
658
2
    if (doris_column->is_nullable()) {
659
2
        auto mutable_column = doris_column->assume_mutable();
660
2
        auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get());
661
2
        null_map_ptr = &nullable_column->get_null_map_data();
662
2
        data_column = nullable_column->get_nested_column_ptr();
663
2
    } else {
664
0
        if (_field_schema->data_type->is_nullable()) {
665
0
            return Status::Corruption("Not nullable column has null values in parquet file");
666
0
        }
667
0
        data_column = doris_column->assume_mutable();
668
0
    }
669
2
    if (type->get_primitive_type() != PrimitiveType::TYPE_ARRAY) {
670
0
        return Status::Corruption(
671
0
                "Wrong data type for column '{}', expected Array type, actual type: {}.",
672
0
                _field_schema->name, type->get_name());
673
0
    }
674
675
2
    ColumnPtr& element_column = assert_cast<ColumnArray&>(*data_column).get_data_ptr();
676
2
    const DataTypePtr& element_type =
677
2
            (assert_cast<const DataTypeArray*>(remove_nullable(type).get()))->get_nested_type();
678
    // read nested column
679
2
    RETURN_IF_ERROR(_element_reader->read_column_data(element_column, element_type,
680
2
                                                      root_node->get_element_node(), filter_map,
681
2
                                                      batch_size, read_rows, eof, is_dict_filter));
682
2
    if (*read_rows == 0) {
683
0
        return Status::OK();
684
0
    }
685
686
2
    ColumnArray::Offsets64& offsets_data = assert_cast<ColumnArray&>(*data_column).get_offsets();
687
    // fill offset and null map
688
2
    fill_array_offset(_field_schema, offsets_data, null_map_ptr, _element_reader->get_rep_level(),
689
2
                      _element_reader->get_def_level());
690
2
    DCHECK_EQ(element_column->size(), offsets_data.back());
691
2
#ifndef NDEBUG
692
2
    doris_column->sanity_check();
693
2
#endif
694
2
    return Status::OK();
695
2
}
696
697
Status MapColumnReader::init(std::unique_ptr<ParquetColumnReader> key_reader,
698
                             std::unique_ptr<ParquetColumnReader> value_reader,
699
0
                             FieldSchema* field) {
700
0
    _field_schema = field;
701
0
    _key_reader = std::move(key_reader);
702
0
    _value_reader = std::move(value_reader);
703
0
    return Status::OK();
704
0
}
705
706
Status MapColumnReader::read_column_data(
707
        ColumnPtr& doris_column, const DataTypePtr& type,
708
        const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map,
709
        size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter,
710
0
        int64_t real_column_size) {
711
0
    MutableColumnPtr data_column;
712
0
    NullMap* null_map_ptr = nullptr;
713
0
    if (doris_column->is_nullable()) {
714
0
        auto mutable_column = doris_column->assume_mutable();
715
0
        auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get());
716
0
        null_map_ptr = &nullable_column->get_null_map_data();
717
0
        data_column = nullable_column->get_nested_column_ptr();
718
0
    } else {
719
0
        if (_field_schema->data_type->is_nullable()) {
720
0
            return Status::Corruption("Not nullable column has null values in parquet file");
721
0
        }
722
0
        data_column = doris_column->assume_mutable();
723
0
    }
724
0
    if (remove_nullable(type)->get_primitive_type() != PrimitiveType::TYPE_MAP) {
725
0
        return Status::Corruption(
726
0
                "Wrong data type for column '{}', expected Map type, actual type id {}.",
727
0
                _field_schema->name, type->get_name());
728
0
    }
729
730
0
    auto& map = assert_cast<ColumnMap&>(*data_column);
731
0
    const DataTypePtr& key_type =
732
0
            assert_cast<const DataTypeMap*>(remove_nullable(type).get())->get_key_type();
733
0
    const DataTypePtr& value_type =
734
0
            assert_cast<const DataTypeMap*>(remove_nullable(type).get())->get_value_type();
735
0
    ColumnPtr& key_column = map.get_keys_ptr();
736
0
    ColumnPtr& value_column = map.get_values_ptr();
737
738
0
    size_t key_rows = 0;
739
0
    size_t value_rows = 0;
740
0
    bool key_eof = false;
741
0
    bool value_eof = false;
742
0
    int64_t orig_col_column_size = key_column->size();
743
744
0
    RETURN_IF_ERROR(_key_reader->read_column_data(key_column, key_type, root_node->get_key_node(),
745
0
                                                  filter_map, batch_size, &key_rows, &key_eof,
746
0
                                                  is_dict_filter));
747
748
0
    while (value_rows < key_rows && !value_eof) {
749
0
        size_t loop_rows = 0;
750
0
        RETURN_IF_ERROR(_value_reader->read_column_data(
751
0
                value_column, value_type, root_node->get_value_node(), filter_map,
752
0
                key_rows - value_rows, &loop_rows, &value_eof, is_dict_filter,
753
0
                key_column->size() - orig_col_column_size));
754
0
        value_rows += loop_rows;
755
0
    }
756
0
    DCHECK_EQ(key_rows, value_rows);
757
0
    *read_rows = key_rows;
758
0
    *eof = key_eof;
759
760
0
    if (*read_rows == 0) {
761
0
        return Status::OK();
762
0
    }
763
764
0
    DCHECK_EQ(key_column->size(), value_column->size());
765
    // fill offset and null map
766
0
    fill_array_offset(_field_schema, map.get_offsets(), null_map_ptr, _key_reader->get_rep_level(),
767
0
                      _key_reader->get_def_level());
768
0
    DCHECK_EQ(key_column->size(), map.get_offsets().back());
769
0
#ifndef NDEBUG
770
0
    doris_column->sanity_check();
771
0
#endif
772
0
    return Status::OK();
773
0
}
774
775
Status StructColumnReader::init(
776
        std::unordered_map<std::string, std::unique_ptr<ParquetColumnReader>>&& child_readers,
777
10
        FieldSchema* field) {
778
10
    _field_schema = field;
779
10
    _child_readers = std::move(child_readers);
780
10
    return Status::OK();
781
10
}
782
Status StructColumnReader::read_column_data(
783
        ColumnPtr& doris_column, const DataTypePtr& type,
784
        const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map,
785
        size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter,
786
10
        int64_t real_column_size) {
787
10
    MutableColumnPtr data_column;
788
10
    NullMap* null_map_ptr = nullptr;
789
10
    if (doris_column->is_nullable()) {
790
10
        auto mutable_column = doris_column->assume_mutable();
791
10
        auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get());
792
10
        null_map_ptr = &nullable_column->get_null_map_data();
793
10
        data_column = nullable_column->get_nested_column_ptr();
794
10
    } else {
795
0
        if (_field_schema->data_type->is_nullable()) {
796
0
            return Status::Corruption("Not nullable column has null values in parquet file");
797
0
        }
798
0
        data_column = doris_column->assume_mutable();
799
0
    }
800
10
    if (type->get_primitive_type() != PrimitiveType::TYPE_STRUCT) {
801
0
        return Status::Corruption(
802
0
                "Wrong data type for column '{}', expected Struct type, actual type id {}.",
803
0
                _field_schema->name, type->get_name());
804
0
    }
805
806
10
    auto& doris_struct = assert_cast<ColumnStruct&>(*data_column);
807
10
    const auto* doris_struct_type = assert_cast<const DataTypeStruct*>(remove_nullable(type).get());
808
809
10
    int64_t not_missing_column_id = -1;
810
10
    size_t not_missing_orig_column_size = 0;
811
10
    std::vector<size_t> missing_column_idxs {};
812
10
    std::vector<size_t> skip_reading_column_idxs {};
813
814
10
    _read_column_names.clear();
815
816
34
    for (size_t i = 0; i < doris_struct.tuple_size(); ++i) {
817
24
        ColumnPtr& doris_field = doris_struct.get_column_ptr(i);
818
24
        auto& doris_type = doris_struct_type->get_element(i);
819
24
        auto& doris_name = doris_struct_type->get_element_name(i);
820
24
        if (!root_node->children_column_exists(doris_name)) {
821
0
            missing_column_idxs.push_back(i);
822
0
            VLOG_DEBUG << "[ParquetReader] Missing column in schema: column_idx[" << i
823
0
                       << "], doris_name: " << doris_name << " (column not exists in root node)";
824
0
            continue;
825
0
        }
826
24
        auto file_name = root_node->children_file_column_name(doris_name);
827
828
        // Check if this is a SkipReadingReader - we should skip it when choosing reference column
829
        // because SkipReadingReader doesn't know the actual data size in nested context
830
24
        bool is_skip_reader =
831
24
                dynamic_cast<SkipReadingReader*>(_child_readers[file_name].get()) != nullptr;
832
833
24
        if (is_skip_reader) {
834
            // Store SkipReadingReader columns to fill them later based on reference column size
835
8
            skip_reading_column_idxs.push_back(i);
836
8
            continue;
837
8
        }
838
839
        // Only add non-SkipReadingReader columns to _read_column_names
840
        // This ensures get_rep_level() and get_def_level() return valid levels
841
16
        _read_column_names.emplace_back(file_name);
842
843
16
        size_t field_rows = 0;
844
16
        bool field_eof = false;
845
16
        if (not_missing_column_id == -1) {
846
10
            not_missing_column_id = i;
847
10
            not_missing_orig_column_size = doris_field->size();
848
10
            RETURN_IF_ERROR(_child_readers[file_name]->read_column_data(
849
10
                    doris_field, doris_type, root_node->get_children_node(doris_name), filter_map,
850
10
                    batch_size, &field_rows, &field_eof, is_dict_filter));
851
10
            *read_rows = field_rows;
852
10
            *eof = field_eof;
853
            /*
854
             * Considering the issue in the `_read_nested_column` function where data may span across pages, leading
855
             * to missing definition and repetition levels, when filling the null_map of the struct later, it is
856
             * crucial to use the definition and repetition levels from the first read column
857
             * (since `_read_nested_column` is not called repeatedly).
858
             *
859
             *  It is worth mentioning that, theoretically, any sub-column can be chosen to fill the null_map,
860
             *  and selecting the shortest one will offer better performance
861
             */
862
10
        } else {
863
12
            while (field_rows < *read_rows && !field_eof) {
864
6
                size_t loop_rows = 0;
865
6
                RETURN_IF_ERROR(_child_readers[file_name]->read_column_data(
866
6
                        doris_field, doris_type, root_node->get_children_node(doris_name),
867
6
                        filter_map, *read_rows - field_rows, &loop_rows, &field_eof,
868
6
                        is_dict_filter));
869
6
                field_rows += loop_rows;
870
6
            }
871
6
            DCHECK_EQ(*read_rows, field_rows);
872
            //            DCHECK_EQ(*eof, field_eof);
873
6
        }
874
16
    }
875
876
10
    int64_t missing_column_sz = -1;
877
878
10
    if (not_missing_column_id == -1) {
879
        // All queried columns are missing in the file (e.g., all added after schema change)
880
        // We need to pick a column from _field_schema children that exists in the file for RL/DL reference
881
0
        std::string reference_file_column_name;
882
0
        std::unique_ptr<ParquetColumnReader>* reference_reader = nullptr;
883
884
0
        for (const auto& child : _field_schema->children) {
885
0
            auto it = _child_readers.find(child.name);
886
0
            if (it != _child_readers.end()) {
887
                // Skip SkipReadingReader as they don't have valid RL/DL
888
0
                bool is_skip_reader = dynamic_cast<SkipReadingReader*>(it->second.get()) != nullptr;
889
0
                if (!is_skip_reader) {
890
0
                    reference_file_column_name = child.name;
891
0
                    reference_reader = &(it->second);
892
0
                    break;
893
0
                }
894
0
            }
895
0
        }
896
897
0
        if (reference_reader != nullptr) {
898
            // Read the reference column to get correct RL/DL information
899
            // TODO: Optimize by only reading RL/DL without actual data decoding
900
901
            // We need to find the FieldSchema for the reference column from _field_schema children
902
0
            FieldSchema* ref_field_schema = nullptr;
903
0
            for (auto& child : _field_schema->children) {
904
0
                if (child.name == reference_file_column_name) {
905
0
                    ref_field_schema = &child;
906
0
                    break;
907
0
                }
908
0
            }
909
910
0
            if (ref_field_schema == nullptr) {
911
0
                return Status::InternalError(
912
0
                        "Cannot find field schema for reference column '{}' in struct '{}'",
913
0
                        reference_file_column_name, _field_schema->name);
914
0
            }
915
916
            // Create a temporary column to hold the data (we'll use its size for missing_column_sz)
917
0
            ColumnPtr temp_column = ref_field_schema->data_type->create_column();
918
0
            auto temp_type = ref_field_schema->data_type;
919
920
0
            size_t field_rows = 0;
921
0
            bool field_eof = false;
922
923
            // Use ConstNode for the reference column instead of looking up from root_node.
924
            // The reference column is only used to get RL/DL information for determining the number
925
            // of elements in the struct. It may be a column that has been dropped from the table
926
            // schema (e.g., 'removed' field), but still exists in older parquet files.
927
            // Since we don't need schema mapping for this column (we just need its RL/DL levels),
928
            // using ConstNode is safe and avoids the issue where the reference column doesn't exist
929
            // in root_node (because it was dropped from table schema).
930
0
            auto ref_child_node = TableSchemaChangeHelper::ConstNode::get_instance();
931
0
            not_missing_orig_column_size = temp_column->size();
932
933
0
            RETURN_IF_ERROR((*reference_reader)
934
0
                                    ->read_column_data(temp_column, temp_type, ref_child_node,
935
0
                                                       filter_map, batch_size, &field_rows,
936
0
                                                       &field_eof, is_dict_filter));
937
938
0
            *read_rows = field_rows;
939
0
            *eof = field_eof;
940
941
            // Store this reference column name for get_rep_level/get_def_level to use
942
0
            _read_column_names.emplace_back(reference_file_column_name);
943
944
0
            missing_column_sz = temp_column->size() - not_missing_orig_column_size;
945
0
        } else {
946
0
            return Status::Corruption(
947
0
                    "Cannot read struct '{}': all queried columns are missing and no reference "
948
0
                    "column found in file",
949
0
                    _field_schema->name);
950
0
        }
951
0
    }
952
953
    //  This missing_column_sz is not *read_rows. Because read_rows returns the number of rows.
954
    //  For example: suppose we have a column array<struct<a:int,b:string>>,
955
    //  where b is a newly added column, that is, a missing column.
956
    //  There are two rows of data in this column,
957
    //      [{1,null},{2,null},{3,null}]
958
    //      [{4,null},{5,null}]
959
    //  When you first read subcolumn a, you read 5 data items and the value of *read_rows is 2.
960
    //  You should insert 5 records into subcolumn b instead of 2.
961
10
    if (missing_column_sz == -1) {
962
10
        missing_column_sz = doris_struct.get_column(not_missing_column_id).size() -
963
10
                            not_missing_orig_column_size;
964
10
    }
965
966
    // Fill SkipReadingReader columns with the correct amount of data based on reference column
967
    // Let SkipReadingReader handle the data filling through its read_column_data method
968
10
    for (auto idx : skip_reading_column_idxs) {
969
8
        auto& doris_field = doris_struct.get_column_ptr(idx);
970
8
        auto& doris_type = const_cast<DataTypePtr&>(doris_struct_type->get_element(idx));
971
8
        auto& doris_name = const_cast<String&>(doris_struct_type->get_element_name(idx));
972
8
        auto file_name = root_node->children_file_column_name(doris_name);
973
974
8
        size_t field_rows = 0;
975
8
        bool field_eof = false;
976
8
        RETURN_IF_ERROR(_child_readers[file_name]->read_column_data(
977
8
                doris_field, doris_type, root_node->get_children_node(doris_name), filter_map,
978
8
                missing_column_sz, &field_rows, &field_eof, is_dict_filter, missing_column_sz));
979
8
    }
980
981
    // Fill truly missing columns (not in root_node) with null or default value
982
10
    for (auto idx : missing_column_idxs) {
983
0
        auto& doris_field = doris_struct.get_column_ptr(idx);
984
0
        auto& doris_type = doris_struct_type->get_element(idx);
985
0
        DCHECK(doris_type->is_nullable());
986
0
        auto mutable_column = doris_field->assume_mutable();
987
0
        auto* nullable_column = static_cast<ColumnNullable*>(mutable_column.get());
988
0
        nullable_column->insert_many_defaults(missing_column_sz);
989
0
    }
990
991
10
    if (null_map_ptr != nullptr) {
992
10
        fill_struct_null_map(_field_schema, *null_map_ptr, this->get_rep_level(),
993
10
                             this->get_def_level());
994
10
    }
995
10
#ifndef NDEBUG
996
10
    doris_column->sanity_check();
997
10
#endif
998
10
    return Status::OK();
999
10
}
1000
1001
template class ScalarColumnReader<true, true>;
1002
template class ScalarColumnReader<true, false>;
1003
template class ScalarColumnReader<false, true>;
1004
template class ScalarColumnReader<false, false>;
1005
1006
#include "common/compile_check_end.h"
1007
1008
}; // namespace doris