be/src/format/parquet/vparquet_column_reader.cpp
Line | Count | Source |
1 | | // Licensed to the Apache Software Foundation (ASF) under one |
2 | | // or more contributor license agreements. See the NOTICE file |
3 | | // distributed with this work for additional information |
4 | | // regarding copyright ownership. The ASF licenses this file |
5 | | // to you under the Apache License, Version 2.0 (the |
6 | | // "License"); you may not use this file except in compliance |
7 | | // with the License. You may obtain a copy of the License at |
8 | | // |
9 | | // http://www.apache.org/licenses/LICENSE-2.0 |
10 | | // |
11 | | // Unless required by applicable law or agreed to in writing, |
12 | | // software distributed under the License is distributed on an |
13 | | // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
14 | | // KIND, either express or implied. See the License for the |
15 | | // specific language governing permissions and limitations |
16 | | // under the License. |
17 | | |
18 | | #include "format/parquet/vparquet_column_reader.h" |
19 | | |
20 | | #include <gen_cpp/parquet_types.h> |
21 | | #include <limits.h> |
22 | | #include <sys/types.h> |
23 | | |
24 | | #include <algorithm> |
25 | | #include <utility> |
26 | | |
27 | | #include "common/status.h" |
28 | | #include "core/column/column.h" |
29 | | #include "core/column/column_array.h" |
30 | | #include "core/column/column_map.h" |
31 | | #include "core/column/column_nullable.h" |
32 | | #include "core/column/column_struct.h" |
33 | | #include "core/data_type/data_type_array.h" |
34 | | #include "core/data_type/data_type_map.h" |
35 | | #include "core/data_type/data_type_nullable.h" |
36 | | #include "core/data_type/data_type_struct.h" |
37 | | #include "core/data_type/define_primitive_type.h" |
38 | | #include "format/parquet/level_decoder.h" |
39 | | #include "format/parquet/schema_desc.h" |
40 | | #include "format/parquet/vparquet_column_chunk_reader.h" |
41 | | #include "io/fs/tracing_file_reader.h" |
42 | | #include "runtime/runtime_profile.h" |
43 | | |
44 | | namespace doris { |
45 | | #include "common/compile_check_begin.h" |
46 | | static void fill_struct_null_map(FieldSchema* field, NullMap& null_map, |
47 | | const std::vector<level_t>& rep_levels, |
48 | 10 | const std::vector<level_t>& def_levels) { |
49 | 10 | size_t num_levels = def_levels.size(); |
50 | 10 | DCHECK_EQ(num_levels, rep_levels.size()); |
51 | 10 | size_t origin_size = null_map.size(); |
52 | 10 | null_map.resize(origin_size + num_levels); |
53 | 10 | size_t pos = origin_size; |
54 | 24 | for (size_t i = 0; i < num_levels; ++i) { |
55 | | // skip the levels affect its ancestor or its descendants |
56 | 14 | if (def_levels[i] < field->repeated_parent_def_level || |
57 | 14 | rep_levels[i] > field->repetition_level) { |
58 | 0 | continue; |
59 | 0 | } |
60 | 14 | if (def_levels[i] >= field->definition_level) { |
61 | 14 | null_map[pos++] = 0; |
62 | 14 | } else { |
63 | 0 | null_map[pos++] = 1; |
64 | 0 | } |
65 | 14 | } |
66 | 10 | null_map.resize(pos); |
67 | 10 | } |
68 | | |
69 | | static void fill_array_offset(FieldSchema* field, ColumnArray::Offsets64& offsets_data, |
70 | | NullMap* null_map_ptr, const std::vector<level_t>& rep_levels, |
71 | 2 | const std::vector<level_t>& def_levels) { |
72 | 2 | size_t num_levels = rep_levels.size(); |
73 | 2 | DCHECK_EQ(num_levels, def_levels.size()); |
74 | 2 | size_t origin_size = offsets_data.size(); |
75 | 2 | offsets_data.resize(origin_size + num_levels); |
76 | 2 | if (null_map_ptr != nullptr) { |
77 | 2 | null_map_ptr->resize(origin_size + num_levels); |
78 | 2 | } |
79 | 2 | size_t offset_pos = origin_size - 1; |
80 | 8 | for (size_t i = 0; i < num_levels; ++i) { |
81 | | // skip the levels affect its ancestor or its descendants |
82 | 6 | if (def_levels[i] < field->repeated_parent_def_level || |
83 | 6 | rep_levels[i] > field->repetition_level) { |
84 | 0 | continue; |
85 | 0 | } |
86 | 6 | if (rep_levels[i] == field->repetition_level) { |
87 | 4 | offsets_data[offset_pos]++; |
88 | 4 | continue; |
89 | 4 | } |
90 | 2 | offset_pos++; |
91 | 2 | offsets_data[offset_pos] = offsets_data[offset_pos - 1]; |
92 | 2 | if (def_levels[i] >= field->definition_level) { |
93 | 2 | offsets_data[offset_pos]++; |
94 | 2 | } |
95 | 2 | if (def_levels[i] >= field->definition_level - 1) { |
96 | 2 | (*null_map_ptr)[offset_pos] = 0; |
97 | 2 | } else { |
98 | 0 | (*null_map_ptr)[offset_pos] = 1; |
99 | 0 | } |
100 | 2 | } |
101 | 2 | offsets_data.resize(offset_pos + 1); |
102 | 2 | if (null_map_ptr != nullptr) { |
103 | 2 | null_map_ptr->resize(offset_pos + 1); |
104 | 2 | } |
105 | 2 | } |
106 | | |
107 | | Status ParquetColumnReader::create(io::FileReaderSPtr file, FieldSchema* field, |
108 | | const tparquet::RowGroup& row_group, const RowRanges& row_ranges, |
109 | | const cctz::time_zone* ctz, io::IOContext* io_ctx, |
110 | | std::unique_ptr<ParquetColumnReader>& reader, |
111 | | size_t max_buf_size, |
112 | | std::unordered_map<int, tparquet::OffsetIndex>& col_offsets, |
113 | | RuntimeState* state, bool in_collection, |
114 | | const std::set<uint64_t>& column_ids, |
115 | 122 | const std::set<uint64_t>& filter_column_ids) { |
116 | 122 | size_t total_rows = row_group.num_rows; |
117 | 122 | if (field->data_type->get_primitive_type() == TYPE_ARRAY) { |
118 | 2 | std::unique_ptr<ParquetColumnReader> element_reader; |
119 | 2 | RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx, |
120 | 2 | element_reader, max_buf_size, col_offsets, state, true, column_ids, |
121 | 2 | filter_column_ids)); |
122 | 2 | auto array_reader = ArrayColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx); |
123 | 2 | element_reader->set_column_in_nested(); |
124 | 2 | RETURN_IF_ERROR(array_reader->init(std::move(element_reader), field)); |
125 | 2 | array_reader->_filter_column_ids = filter_column_ids; |
126 | 2 | reader.reset(array_reader.release()); |
127 | 120 | } else if (field->data_type->get_primitive_type() == TYPE_MAP) { |
128 | 0 | std::unique_ptr<ParquetColumnReader> key_reader; |
129 | 0 | std::unique_ptr<ParquetColumnReader> value_reader; |
130 | |
|
131 | 0 | if (column_ids.empty() || |
132 | 0 | column_ids.find(field->children[0].get_column_id()) != column_ids.end()) { |
133 | | // Create key reader |
134 | 0 | RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx, |
135 | 0 | key_reader, max_buf_size, col_offsets, state, true, column_ids, |
136 | 0 | filter_column_ids)); |
137 | 0 | } else { |
138 | 0 | auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz, |
139 | 0 | io_ctx, &field->children[0]); |
140 | 0 | key_reader = std::move(skip_reader); |
141 | 0 | } |
142 | | |
143 | 0 | if (column_ids.empty() || |
144 | 0 | column_ids.find(field->children[1].get_column_id()) != column_ids.end()) { |
145 | | // Create value reader |
146 | 0 | RETURN_IF_ERROR(create(file, &field->children[1], row_group, row_ranges, ctz, io_ctx, |
147 | 0 | value_reader, max_buf_size, col_offsets, state, true, column_ids, |
148 | 0 | filter_column_ids)); |
149 | 0 | } else { |
150 | 0 | auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz, |
151 | 0 | io_ctx, &field->children[0]); |
152 | 0 | value_reader = std::move(skip_reader); |
153 | 0 | } |
154 | | |
155 | 0 | auto map_reader = MapColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx); |
156 | 0 | key_reader->set_column_in_nested(); |
157 | 0 | value_reader->set_column_in_nested(); |
158 | 0 | RETURN_IF_ERROR(map_reader->init(std::move(key_reader), std::move(value_reader), field)); |
159 | 0 | map_reader->_filter_column_ids = filter_column_ids; |
160 | 0 | reader.reset(map_reader.release()); |
161 | 120 | } else if (field->data_type->get_primitive_type() == TYPE_STRUCT) { |
162 | 10 | std::unordered_map<std::string, std::unique_ptr<ParquetColumnReader>> child_readers; |
163 | 10 | child_readers.reserve(field->children.size()); |
164 | 10 | int non_skip_reader_idx = -1; |
165 | 34 | for (int i = 0; i < field->children.size(); ++i) { |
166 | 24 | auto& child = field->children[i]; |
167 | 24 | std::unique_ptr<ParquetColumnReader> child_reader; |
168 | 24 | if (column_ids.empty() || column_ids.find(child.get_column_id()) != column_ids.end()) { |
169 | 16 | RETURN_IF_ERROR(create(file, &child, row_group, row_ranges, ctz, io_ctx, |
170 | 16 | child_reader, max_buf_size, col_offsets, state, |
171 | 16 | in_collection, column_ids, filter_column_ids)); |
172 | 16 | child_readers[child.name] = std::move(child_reader); |
173 | | // Record the first non-SkippingReader |
174 | 16 | if (non_skip_reader_idx == -1) { |
175 | 10 | non_skip_reader_idx = i; |
176 | 10 | } |
177 | 16 | } else { |
178 | 8 | auto skip_reader = std::make_unique<SkipReadingReader>(row_ranges, total_rows, ctz, |
179 | 8 | io_ctx, &child); |
180 | 8 | skip_reader->_filter_column_ids = filter_column_ids; |
181 | 8 | child_readers[child.name] = std::move(skip_reader); |
182 | 8 | } |
183 | 24 | child_readers[child.name]->set_column_in_nested(); |
184 | 24 | } |
185 | | // If all children are SkipReadingReader, force the first child to call create |
186 | 10 | if (non_skip_reader_idx == -1) { |
187 | 0 | std::unique_ptr<ParquetColumnReader> child_reader; |
188 | 0 | RETURN_IF_ERROR(create(file, &field->children[0], row_group, row_ranges, ctz, io_ctx, |
189 | 0 | child_reader, max_buf_size, col_offsets, state, in_collection, |
190 | 0 | column_ids, filter_column_ids)); |
191 | 0 | child_reader->set_column_in_nested(); |
192 | 0 | child_readers[field->children[0].name] = std::move(child_reader); |
193 | 0 | } |
194 | 10 | auto struct_reader = StructColumnReader::create_unique(row_ranges, total_rows, ctz, io_ctx); |
195 | 10 | RETURN_IF_ERROR(struct_reader->init(std::move(child_readers), field)); |
196 | 10 | struct_reader->_filter_column_ids = filter_column_ids; |
197 | 10 | reader.reset(struct_reader.release()); |
198 | 110 | } else { |
199 | 110 | auto physical_index = field->physical_column_index; |
200 | 110 | const tparquet::OffsetIndex* offset_index = |
201 | 110 | col_offsets.find(physical_index) != col_offsets.end() ? &col_offsets[physical_index] |
202 | 110 | : nullptr; |
203 | | |
204 | 110 | const tparquet::ColumnChunk& chunk = row_group.columns[physical_index]; |
205 | 110 | if (in_collection) { |
206 | 2 | if (offset_index == nullptr) { |
207 | 2 | auto scalar_reader = ScalarColumnReader<true, false>::create_unique( |
208 | 2 | row_ranges, total_rows, chunk, offset_index, ctz, io_ctx); |
209 | | |
210 | 2 | RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state)); |
211 | 2 | scalar_reader->_filter_column_ids = filter_column_ids; |
212 | 2 | reader.reset(scalar_reader.release()); |
213 | 2 | } else { |
214 | 0 | auto scalar_reader = ScalarColumnReader<true, true>::create_unique( |
215 | 0 | row_ranges, total_rows, chunk, offset_index, ctz, io_ctx); |
216 | |
|
217 | 0 | RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state)); |
218 | 0 | scalar_reader->_filter_column_ids = filter_column_ids; |
219 | 0 | reader.reset(scalar_reader.release()); |
220 | 0 | } |
221 | 108 | } else { |
222 | 108 | if (offset_index == nullptr) { |
223 | 108 | auto scalar_reader = ScalarColumnReader<false, false>::create_unique( |
224 | 108 | row_ranges, total_rows, chunk, offset_index, ctz, io_ctx); |
225 | | |
226 | 108 | RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state)); |
227 | 108 | scalar_reader->_filter_column_ids = filter_column_ids; |
228 | 108 | reader.reset(scalar_reader.release()); |
229 | 108 | } else { |
230 | 0 | auto scalar_reader = ScalarColumnReader<false, true>::create_unique( |
231 | 0 | row_ranges, total_rows, chunk, offset_index, ctz, io_ctx); |
232 | |
|
233 | 0 | RETURN_IF_ERROR(scalar_reader->init(file, field, max_buf_size, state)); |
234 | 0 | scalar_reader->_filter_column_ids = filter_column_ids; |
235 | 0 | reader.reset(scalar_reader.release()); |
236 | 0 | } |
237 | 108 | } |
238 | 110 | } |
239 | 122 | return Status::OK(); |
240 | 122 | } |
241 | | |
242 | | void ParquetColumnReader::_generate_read_ranges(RowRange page_row_range, |
243 | 182 | RowRanges* result_ranges) const { |
244 | 182 | result_ranges->add(page_row_range); |
245 | 182 | RowRanges::ranges_intersection(*result_ranges, _row_ranges, result_ranges); |
246 | 182 | } |
247 | | |
248 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
249 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::init(io::FileReaderSPtr file, |
250 | | FieldSchema* field, |
251 | | size_t max_buf_size, |
252 | 110 | RuntimeState* state) { |
253 | 110 | _field_schema = field; |
254 | 110 | auto& chunk_meta = _chunk_meta.meta_data; |
255 | 110 | int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset |
256 | 110 | : chunk_meta.data_page_offset; |
257 | 110 | size_t chunk_len = chunk_meta.total_compressed_size; |
258 | 110 | size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size); |
259 | 110 | if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) && |
260 | 110 | typeid_cast<io::MergeRangeFileReader*>( |
261 | 51 | ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) || |
262 | 110 | typeid_cast<io::MergeRangeFileReader*>(file.get())) { |
263 | | // turn off prefetch data when using MergeRangeFileReader |
264 | 110 | prefetch_buffer_size = 0; |
265 | 110 | } |
266 | 110 | _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len, |
267 | 110 | prefetch_buffer_size); |
268 | 110 | ParquetPageReadContext ctx( |
269 | 110 | (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache); |
270 | | |
271 | 110 | _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>( |
272 | 110 | _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx); |
273 | 110 | RETURN_IF_ERROR(_chunk_reader->init()); |
274 | 110 | return Status::OK(); |
275 | 110 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE _ZN5doris18ScalarColumnReaderILb1ELb0EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE Line | Count | Source | 252 | 2 | RuntimeState* state) { | 253 | 2 | _field_schema = field; | 254 | 2 | auto& chunk_meta = _chunk_meta.meta_data; | 255 | 2 | int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset | 256 | 2 | : chunk_meta.data_page_offset; | 257 | 2 | size_t chunk_len = chunk_meta.total_compressed_size; | 258 | 2 | size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size); | 259 | 2 | if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) && | 260 | 2 | typeid_cast<io::MergeRangeFileReader*>( | 261 | 0 | ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) || | 262 | 2 | typeid_cast<io::MergeRangeFileReader*>(file.get())) { | 263 | | // turn off prefetch data when using MergeRangeFileReader | 264 | 2 | prefetch_buffer_size = 0; | 265 | 2 | } | 266 | 2 | _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len, | 267 | 2 | prefetch_buffer_size); | 268 | 2 | ParquetPageReadContext ctx( | 269 | 2 | (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache); | 270 | | | 271 | 2 | _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>( | 272 | 2 | _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx); | 273 | 2 | RETURN_IF_ERROR(_chunk_reader->init()); | 274 | 2 | return Status::OK(); | 275 | 2 | } |
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE _ZN5doris18ScalarColumnReaderILb0ELb0EE4initESt10shared_ptrINS_2io10FileReaderEEPNS_11FieldSchemaEmPNS_12RuntimeStateE Line | Count | Source | 252 | 108 | RuntimeState* state) { | 253 | 108 | _field_schema = field; | 254 | 108 | auto& chunk_meta = _chunk_meta.meta_data; | 255 | 108 | int64_t chunk_start = has_dict_page(chunk_meta) ? chunk_meta.dictionary_page_offset | 256 | 108 | : chunk_meta.data_page_offset; | 257 | 108 | size_t chunk_len = chunk_meta.total_compressed_size; | 258 | 108 | size_t prefetch_buffer_size = std::min(chunk_len, max_buf_size); | 259 | 108 | if ((typeid_cast<doris::io::TracingFileReader*>(file.get()) && | 260 | 108 | typeid_cast<io::MergeRangeFileReader*>( | 261 | 51 | ((doris::io::TracingFileReader*)(file.get()))->inner_reader().get())) || | 262 | 108 | typeid_cast<io::MergeRangeFileReader*>(file.get())) { | 263 | | // turn off prefetch data when using MergeRangeFileReader | 264 | 108 | prefetch_buffer_size = 0; | 265 | 108 | } | 266 | 108 | _stream_reader = std::make_unique<io::BufferedFileStreamReader>(file, chunk_start, chunk_len, | 267 | 108 | prefetch_buffer_size); | 268 | 108 | ParquetPageReadContext ctx( | 269 | 108 | (state == nullptr) ? true : state->query_options().enable_parquet_file_page_cache); | 270 | | | 271 | 108 | _chunk_reader = std::make_unique<ColumnChunkReader<IN_COLLECTION, OFFSET_INDEX>>( | 272 | 108 | _stream_reader.get(), &_chunk_meta, field, _offset_index, _total_rows, _io_ctx, ctx); | 273 | 108 | RETURN_IF_ERROR(_chunk_reader->init()); | 274 | 108 | return Status::OK(); | 275 | 108 | } |
|
276 | | |
277 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
278 | 152 | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_skip_values(size_t num_values) { |
279 | 152 | if (num_values == 0) { |
280 | 50 | return Status::OK(); |
281 | 50 | } |
282 | 102 | if (_chunk_reader->max_def_level() > 0) { |
283 | 102 | LevelDecoder& def_decoder = _chunk_reader->def_level_decoder(); |
284 | 102 | size_t skipped = 0; |
285 | 102 | size_t null_size = 0; |
286 | 102 | size_t nonnull_size = 0; |
287 | 217 | while (skipped < num_values) { |
288 | 115 | level_t def_level = -1; |
289 | 115 | size_t loop_skip = def_decoder.get_next_run(&def_level, num_values - skipped); |
290 | 115 | if (loop_skip == 0) { |
291 | 0 | std::stringstream ss; |
292 | 0 | auto& bit_reader = def_decoder.rle_decoder().bit_reader(); |
293 | 0 | ss << "def_decoder buffer (hex): "; |
294 | 0 | for (size_t i = 0; i < bit_reader.max_bytes(); ++i) { |
295 | 0 | ss << std::hex << std::setw(2) << std::setfill('0') |
296 | 0 | << static_cast<int>(bit_reader.buffer()[i]) << " "; |
297 | 0 | } |
298 | 0 | LOG(WARNING) << ss.str(); |
299 | 0 | return Status::InternalError("Failed to decode definition level."); |
300 | 0 | } |
301 | 115 | if (def_level < _field_schema->definition_level) { |
302 | 8 | null_size += loop_skip; |
303 | 107 | } else { |
304 | 107 | nonnull_size += loop_skip; |
305 | 107 | } |
306 | 115 | skipped += loop_skip; |
307 | 115 | } |
308 | 102 | if (null_size > 0) { |
309 | 5 | RETURN_IF_ERROR(_chunk_reader->skip_values(null_size, false)); |
310 | 5 | } |
311 | 102 | if (nonnull_size > 0) { |
312 | 101 | RETURN_IF_ERROR(_chunk_reader->skip_values(nonnull_size, true)); |
313 | 101 | } |
314 | 102 | } else { |
315 | 0 | RETURN_IF_ERROR(_chunk_reader->skip_values(num_values)); |
316 | 0 | } |
317 | 102 | return Status::OK(); |
318 | 102 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE12_skip_valuesEm Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE12_skip_valuesEm Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE12_skip_valuesEm _ZN5doris18ScalarColumnReaderILb0ELb0EE12_skip_valuesEm Line | Count | Source | 278 | 152 | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_skip_values(size_t num_values) { | 279 | 152 | if (num_values == 0) { | 280 | 50 | return Status::OK(); | 281 | 50 | } | 282 | 102 | if (_chunk_reader->max_def_level() > 0) { | 283 | 102 | LevelDecoder& def_decoder = _chunk_reader->def_level_decoder(); | 284 | 102 | size_t skipped = 0; | 285 | 102 | size_t null_size = 0; | 286 | 102 | size_t nonnull_size = 0; | 287 | 217 | while (skipped < num_values) { | 288 | 115 | level_t def_level = -1; | 289 | 115 | size_t loop_skip = def_decoder.get_next_run(&def_level, num_values - skipped); | 290 | 115 | if (loop_skip == 0) { | 291 | 0 | std::stringstream ss; | 292 | 0 | auto& bit_reader = def_decoder.rle_decoder().bit_reader(); | 293 | 0 | ss << "def_decoder buffer (hex): "; | 294 | 0 | for (size_t i = 0; i < bit_reader.max_bytes(); ++i) { | 295 | 0 | ss << std::hex << std::setw(2) << std::setfill('0') | 296 | 0 | << static_cast<int>(bit_reader.buffer()[i]) << " "; | 297 | 0 | } | 298 | 0 | LOG(WARNING) << ss.str(); | 299 | 0 | return Status::InternalError("Failed to decode definition level."); | 300 | 0 | } | 301 | 115 | if (def_level < _field_schema->definition_level) { | 302 | 8 | null_size += loop_skip; | 303 | 107 | } else { | 304 | 107 | nonnull_size += loop_skip; | 305 | 107 | } | 306 | 115 | skipped += loop_skip; | 307 | 115 | } | 308 | 102 | if (null_size > 0) { | 309 | 5 | RETURN_IF_ERROR(_chunk_reader->skip_values(null_size, false)); | 310 | 5 | } | 311 | 102 | if (nonnull_size > 0) { | 312 | 101 | RETURN_IF_ERROR(_chunk_reader->skip_values(nonnull_size, true)); | 313 | 101 | } | 314 | 102 | } else { | 315 | 0 | RETURN_IF_ERROR(_chunk_reader->skip_values(num_values)); | 316 | 0 | } | 317 | 102 | return Status::OK(); | 318 | 102 | } |
|
319 | | |
320 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
321 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_read_values(size_t num_values, |
322 | | ColumnPtr& doris_column, |
323 | | DataTypePtr& type, |
324 | | FilterMap& filter_map, |
325 | 152 | bool is_dict_filter) { |
326 | 152 | if (num_values == 0) { |
327 | 0 | return Status::OK(); |
328 | 0 | } |
329 | 152 | MutableColumnPtr data_column; |
330 | 152 | std::vector<uint16_t> null_map; |
331 | 152 | NullMap* map_data_column = nullptr; |
332 | 152 | if (doris_column->is_nullable()) { |
333 | 152 | SCOPED_RAW_TIMER(&_decode_null_map_time); |
334 | | // doris_column either originates from a mutable block in vparquet_group_reader |
335 | | // or is a newly created ColumnPtr, and therefore can be modified. |
336 | 152 | auto* nullable_column = |
337 | 152 | assert_cast<ColumnNullable*>(const_cast<IColumn*>(doris_column.get())); |
338 | | |
339 | 152 | data_column = nullable_column->get_nested_column_ptr(); |
340 | 152 | map_data_column = &(nullable_column->get_null_map_data()); |
341 | 152 | if (_chunk_reader->max_def_level() > 0) { |
342 | 132 | LevelDecoder& def_decoder = _chunk_reader->def_level_decoder(); |
343 | 132 | size_t has_read = 0; |
344 | 132 | bool prev_is_null = true; |
345 | 264 | while (has_read < num_values) { |
346 | 132 | level_t def_level; |
347 | 132 | size_t loop_read = def_decoder.get_next_run(&def_level, num_values - has_read); |
348 | 132 | if (loop_read == 0) { |
349 | 0 | std::stringstream ss; |
350 | 0 | auto& bit_reader = def_decoder.rle_decoder().bit_reader(); |
351 | 0 | ss << "def_decoder buffer (hex): "; |
352 | 0 | for (size_t i = 0; i < bit_reader.max_bytes(); ++i) { |
353 | 0 | ss << std::hex << std::setw(2) << std::setfill('0') |
354 | 0 | << static_cast<int>(bit_reader.buffer()[i]) << " "; |
355 | 0 | } |
356 | 0 | LOG(WARNING) << ss.str(); |
357 | 0 | return Status::InternalError("Failed to decode definition level."); |
358 | 0 | } |
359 | | |
360 | 132 | bool is_null = def_level < _field_schema->definition_level; |
361 | 132 | if (!(prev_is_null ^ is_null)) { |
362 | 17 | null_map.emplace_back(0); |
363 | 17 | } |
364 | 132 | size_t remaining = loop_read; |
365 | 132 | while (remaining > USHRT_MAX) { |
366 | 0 | null_map.emplace_back(USHRT_MAX); |
367 | 0 | null_map.emplace_back(0); |
368 | 0 | remaining -= USHRT_MAX; |
369 | 0 | } |
370 | 132 | null_map.emplace_back((u_short)remaining); |
371 | 132 | prev_is_null = is_null; |
372 | 132 | has_read += loop_read; |
373 | 132 | } |
374 | 132 | } |
375 | 152 | } else { |
376 | 0 | if (_chunk_reader->max_def_level() > 0) { |
377 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); |
378 | 0 | } |
379 | 0 | data_column = doris_column->assume_mutable(); |
380 | 0 | } |
381 | 152 | if (null_map.size() == 0) { |
382 | 20 | size_t remaining = num_values; |
383 | 20 | while (remaining > USHRT_MAX) { |
384 | 0 | null_map.emplace_back(USHRT_MAX); |
385 | 0 | null_map.emplace_back(0); |
386 | 0 | remaining -= USHRT_MAX; |
387 | 0 | } |
388 | 20 | null_map.emplace_back((u_short)remaining); |
389 | 20 | } |
390 | 152 | ColumnSelectVector select_vector; |
391 | 152 | { |
392 | 152 | SCOPED_RAW_TIMER(&_decode_null_map_time); |
393 | 152 | RETURN_IF_ERROR(select_vector.init(null_map, num_values, map_data_column, &filter_map, |
394 | 152 | _filter_map_index)); |
395 | 152 | _filter_map_index += num_values; |
396 | 152 | } |
397 | 0 | return _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter); |
398 | 152 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb _ZN5doris18ScalarColumnReaderILb0ELb0EE12_read_valuesEmRNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEb Line | Count | Source | 325 | 152 | bool is_dict_filter) { | 326 | 152 | if (num_values == 0) { | 327 | 0 | return Status::OK(); | 328 | 0 | } | 329 | 152 | MutableColumnPtr data_column; | 330 | 152 | std::vector<uint16_t> null_map; | 331 | 152 | NullMap* map_data_column = nullptr; | 332 | 152 | if (doris_column->is_nullable()) { | 333 | 152 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 334 | | // doris_column either originates from a mutable block in vparquet_group_reader | 335 | | // or is a newly created ColumnPtr, and therefore can be modified. | 336 | 152 | auto* nullable_column = | 337 | 152 | assert_cast<ColumnNullable*>(const_cast<IColumn*>(doris_column.get())); | 338 | | | 339 | 152 | data_column = nullable_column->get_nested_column_ptr(); | 340 | 152 | map_data_column = &(nullable_column->get_null_map_data()); | 341 | 152 | if (_chunk_reader->max_def_level() > 0) { | 342 | 132 | LevelDecoder& def_decoder = _chunk_reader->def_level_decoder(); | 343 | 132 | size_t has_read = 0; | 344 | 132 | bool prev_is_null = true; | 345 | 264 | while (has_read < num_values) { | 346 | 132 | level_t def_level; | 347 | 132 | size_t loop_read = def_decoder.get_next_run(&def_level, num_values - has_read); | 348 | 132 | if (loop_read == 0) { | 349 | 0 | std::stringstream ss; | 350 | 0 | auto& bit_reader = def_decoder.rle_decoder().bit_reader(); | 351 | 0 | ss << "def_decoder buffer (hex): "; | 352 | 0 | for (size_t i = 0; i < bit_reader.max_bytes(); ++i) { | 353 | 0 | ss << std::hex << std::setw(2) << std::setfill('0') | 354 | 0 | << static_cast<int>(bit_reader.buffer()[i]) << " "; | 355 | 0 | } | 356 | 0 | LOG(WARNING) << ss.str(); | 357 | 0 | return Status::InternalError("Failed to decode definition level."); | 358 | 0 | } | 359 | | | 360 | 132 | bool is_null = def_level < _field_schema->definition_level; | 361 | 132 | if (!(prev_is_null ^ is_null)) { | 362 | 17 | null_map.emplace_back(0); | 363 | 17 | } | 364 | 132 | size_t remaining = loop_read; | 365 | 132 | while (remaining > USHRT_MAX) { | 366 | 0 | null_map.emplace_back(USHRT_MAX); | 367 | 0 | null_map.emplace_back(0); | 368 | 0 | remaining -= USHRT_MAX; | 369 | 0 | } | 370 | 132 | null_map.emplace_back((u_short)remaining); | 371 | 132 | prev_is_null = is_null; | 372 | 132 | has_read += loop_read; | 373 | 132 | } | 374 | 132 | } | 375 | 152 | } else { | 376 | 0 | if (_chunk_reader->max_def_level() > 0) { | 377 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); | 378 | 0 | } | 379 | 0 | data_column = doris_column->assume_mutable(); | 380 | 0 | } | 381 | 152 | if (null_map.size() == 0) { | 382 | 20 | size_t remaining = num_values; | 383 | 20 | while (remaining > USHRT_MAX) { | 384 | 0 | null_map.emplace_back(USHRT_MAX); | 385 | 0 | null_map.emplace_back(0); | 386 | 0 | remaining -= USHRT_MAX; | 387 | 0 | } | 388 | 20 | null_map.emplace_back((u_short)remaining); | 389 | 20 | } | 390 | 152 | ColumnSelectVector select_vector; | 391 | 152 | { | 392 | 152 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 393 | 152 | RETURN_IF_ERROR(select_vector.init(null_map, num_values, map_data_column, &filter_map, | 394 | 152 | _filter_map_index)); | 395 | 152 | _filter_map_index += num_values; | 396 | 152 | } | 397 | 0 | return _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter); | 398 | 152 | } |
|
399 | | |
400 | | /** |
401 | | * Load the nested column data of complex type. |
402 | | * A row of complex type may be stored across two(or more) pages, and the parameter `align_rows` indicates that |
403 | | * whether the reader should read the remaining value of the last row in previous page. |
404 | | */ |
405 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
406 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_read_nested_column( |
407 | | ColumnPtr& doris_column, DataTypePtr& type, FilterMap& filter_map, size_t batch_size, |
408 | 8 | size_t* read_rows, bool* eof, bool is_dict_filter) { |
409 | 8 | _rep_levels.clear(); |
410 | 8 | _def_levels.clear(); |
411 | | |
412 | | // Handle nullable columns |
413 | 8 | MutableColumnPtr data_column; |
414 | 8 | NullMap* map_data_column = nullptr; |
415 | 8 | if (doris_column->is_nullable()) { |
416 | 8 | SCOPED_RAW_TIMER(&_decode_null_map_time); |
417 | | // doris_column either originates from a mutable block in vparquet_group_reader |
418 | | // or is a newly created ColumnPtr, and therefore can be modified. |
419 | 8 | auto* nullable_column = |
420 | 8 | const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get())); |
421 | 8 | data_column = nullable_column->get_nested_column_ptr(); |
422 | 8 | map_data_column = &(nullable_column->get_null_map_data()); |
423 | 8 | } else { |
424 | 0 | if (_field_schema->data_type->is_nullable()) { |
425 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); |
426 | 0 | } |
427 | 0 | data_column = doris_column->assume_mutable(); |
428 | 0 | } |
429 | | |
430 | 8 | std::vector<uint16_t> null_map; |
431 | 8 | std::unordered_set<size_t> ancestor_null_indices; |
432 | 8 | std::vector<uint8_t> nested_filter_map_data; |
433 | | |
434 | 8 | auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) { |
435 | 8 | RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels)); |
436 | 8 | std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>(); |
437 | 8 | if (filter_map.has_filter()) { |
438 | 0 | RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz, |
439 | 0 | _rep_levels.size(), nested_filter_map_data, |
440 | 0 | &nested_filter_map)); |
441 | 0 | } |
442 | | |
443 | 8 | null_map.clear(); |
444 | 8 | ancestor_null_indices.clear(); |
445 | 8 | RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map, |
446 | 8 | ancestor_null_indices)); |
447 | | |
448 | 8 | ColumnSelectVector select_vector; |
449 | 8 | { |
450 | 8 | SCOPED_RAW_TIMER(&_decode_null_map_time); |
451 | 8 | RETURN_IF_ERROR(select_vector.init( |
452 | 8 | null_map, |
453 | 8 | _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(), |
454 | 8 | map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices)); |
455 | 8 | } |
456 | | |
457 | 8 | RETURN_IF_ERROR( |
458 | 8 | _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter)); |
459 | 8 | if (ancestor_null_indices.size() != 0) { |
460 | 0 | RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false)); |
461 | 0 | } |
462 | 8 | if (filter_map.has_filter()) { |
463 | 0 | auto new_rep_sz = before_rep_level_sz; |
464 | 0 | for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) { |
465 | 0 | if (nested_filter_map_data[idx - before_rep_level_sz]) { |
466 | 0 | _rep_levels[new_rep_sz] = _rep_levels[idx]; |
467 | 0 | _def_levels[new_rep_sz] = _def_levels[idx]; |
468 | 0 | new_rep_sz++; |
469 | 0 | } |
470 | 0 | } |
471 | 0 | _rep_levels.resize(new_rep_sz); |
472 | 0 | _def_levels.resize(new_rep_sz); |
473 | 0 | } |
474 | 8 | return Status::OK(); |
475 | 8 | }; Unexecuted instantiation: _ZZN5doris18ScalarColumnReaderILb1ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm _ZZN5doris18ScalarColumnReaderILb1ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm Line | Count | Source | 434 | 2 | auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) { | 435 | 2 | RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels)); | 436 | 2 | std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>(); | 437 | 2 | if (filter_map.has_filter()) { | 438 | 0 | RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz, | 439 | 0 | _rep_levels.size(), nested_filter_map_data, | 440 | 0 | &nested_filter_map)); | 441 | 0 | } | 442 | | | 443 | 2 | null_map.clear(); | 444 | 2 | ancestor_null_indices.clear(); | 445 | 2 | RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map, | 446 | 2 | ancestor_null_indices)); | 447 | | | 448 | 2 | ColumnSelectVector select_vector; | 449 | 2 | { | 450 | 2 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 451 | 2 | RETURN_IF_ERROR(select_vector.init( | 452 | 2 | null_map, | 453 | 2 | _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(), | 454 | 2 | map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices)); | 455 | 2 | } | 456 | | | 457 | 2 | RETURN_IF_ERROR( | 458 | 2 | _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter)); | 459 | 2 | if (ancestor_null_indices.size() != 0) { | 460 | 0 | RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false)); | 461 | 0 | } | 462 | 2 | if (filter_map.has_filter()) { | 463 | 0 | auto new_rep_sz = before_rep_level_sz; | 464 | 0 | for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) { | 465 | 0 | if (nested_filter_map_data[idx - before_rep_level_sz]) { | 466 | 0 | _rep_levels[new_rep_sz] = _rep_levels[idx]; | 467 | 0 | _def_levels[new_rep_sz] = _def_levels[idx]; | 468 | 0 | new_rep_sz++; | 469 | 0 | } | 470 | 0 | } | 471 | 0 | _rep_levels.resize(new_rep_sz); | 472 | 0 | _def_levels.resize(new_rep_sz); | 473 | 0 | } | 474 | 2 | return Status::OK(); | 475 | 2 | }; |
Unexecuted instantiation: _ZZN5doris18ScalarColumnReaderILb0ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm _ZZN5doris18ScalarColumnReaderILb0ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbbENKUlmmE_clEmm Line | Count | Source | 434 | 6 | auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) { | 435 | 6 | RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels)); | 436 | 6 | std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>(); | 437 | 6 | if (filter_map.has_filter()) { | 438 | 0 | RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz, | 439 | 0 | _rep_levels.size(), nested_filter_map_data, | 440 | 0 | &nested_filter_map)); | 441 | 0 | } | 442 | | | 443 | 6 | null_map.clear(); | 444 | 6 | ancestor_null_indices.clear(); | 445 | 6 | RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map, | 446 | 6 | ancestor_null_indices)); | 447 | | | 448 | 6 | ColumnSelectVector select_vector; | 449 | 6 | { | 450 | 6 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 451 | 6 | RETURN_IF_ERROR(select_vector.init( | 452 | 6 | null_map, | 453 | 6 | _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(), | 454 | 6 | map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices)); | 455 | 6 | } | 456 | | | 457 | 6 | RETURN_IF_ERROR( | 458 | 6 | _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter)); | 459 | 6 | if (ancestor_null_indices.size() != 0) { | 460 | 0 | RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false)); | 461 | 0 | } | 462 | 6 | if (filter_map.has_filter()) { | 463 | 0 | auto new_rep_sz = before_rep_level_sz; | 464 | 0 | for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) { | 465 | 0 | if (nested_filter_map_data[idx - before_rep_level_sz]) { | 466 | 0 | _rep_levels[new_rep_sz] = _rep_levels[idx]; | 467 | 0 | _def_levels[new_rep_sz] = _def_levels[idx]; | 468 | 0 | new_rep_sz++; | 469 | 0 | } | 470 | 0 | } | 471 | 0 | _rep_levels.resize(new_rep_sz); | 472 | 0 | _def_levels.resize(new_rep_sz); | 473 | 0 | } | 474 | 6 | return Status::OK(); | 475 | 6 | }; |
|
476 | | |
477 | 10 | while (_current_range_idx < _row_ranges.range_size()) { |
478 | 8 | size_t left_row = |
479 | 8 | std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx)); |
480 | 8 | size_t right_row = std::min(left_row + batch_size - *read_rows, |
481 | 8 | (size_t)_row_ranges.get_range_to(_current_range_idx)); |
482 | 8 | _current_row_index = left_row; |
483 | 8 | RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row)); |
484 | 8 | size_t load_rows = 0; |
485 | 8 | bool cross_page = false; |
486 | 8 | size_t before_rep_level_sz = _rep_levels.size(); |
487 | 8 | RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row, |
488 | 8 | &load_rows, &cross_page)); |
489 | 8 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index)); |
490 | 8 | _filter_map_index += load_rows; |
491 | 8 | while (cross_page) { |
492 | 0 | before_rep_level_sz = _rep_levels.size(); |
493 | 0 | RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page)); |
494 | 0 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1)); |
495 | 0 | } |
496 | 8 | *read_rows += load_rows; |
497 | 8 | _current_row_index += load_rows; |
498 | 8 | _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx)); |
499 | 8 | if (*read_rows == batch_size) { |
500 | 6 | break; |
501 | 6 | } |
502 | 8 | } |
503 | 8 | *eof = _current_range_idx == _row_ranges.range_size(); |
504 | 8 | return Status::OK(); |
505 | 8 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb _ZN5doris18ScalarColumnReaderILb1ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb Line | Count | Source | 408 | 2 | size_t* read_rows, bool* eof, bool is_dict_filter) { | 409 | 2 | _rep_levels.clear(); | 410 | 2 | _def_levels.clear(); | 411 | | | 412 | | // Handle nullable columns | 413 | 2 | MutableColumnPtr data_column; | 414 | 2 | NullMap* map_data_column = nullptr; | 415 | 2 | if (doris_column->is_nullable()) { | 416 | 2 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 417 | | // doris_column either originates from a mutable block in vparquet_group_reader | 418 | | // or is a newly created ColumnPtr, and therefore can be modified. | 419 | 2 | auto* nullable_column = | 420 | 2 | const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get())); | 421 | 2 | data_column = nullable_column->get_nested_column_ptr(); | 422 | 2 | map_data_column = &(nullable_column->get_null_map_data()); | 423 | 2 | } else { | 424 | 0 | if (_field_schema->data_type->is_nullable()) { | 425 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); | 426 | 0 | } | 427 | 0 | data_column = doris_column->assume_mutable(); | 428 | 0 | } | 429 | | | 430 | 2 | std::vector<uint16_t> null_map; | 431 | 2 | std::unordered_set<size_t> ancestor_null_indices; | 432 | 2 | std::vector<uint8_t> nested_filter_map_data; | 433 | | | 434 | 2 | auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) { | 435 | 2 | RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels)); | 436 | 2 | std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>(); | 437 | 2 | if (filter_map.has_filter()) { | 438 | 2 | RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz, | 439 | 2 | _rep_levels.size(), nested_filter_map_data, | 440 | 2 | &nested_filter_map)); | 441 | 2 | } | 442 | | | 443 | 2 | null_map.clear(); | 444 | 2 | ancestor_null_indices.clear(); | 445 | 2 | RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map, | 446 | 2 | ancestor_null_indices)); | 447 | | | 448 | 2 | ColumnSelectVector select_vector; | 449 | 2 | { | 450 | 2 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 451 | 2 | RETURN_IF_ERROR(select_vector.init( | 452 | 2 | null_map, | 453 | 2 | _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(), | 454 | 2 | map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices)); | 455 | 2 | } | 456 | | | 457 | 2 | RETURN_IF_ERROR( | 458 | 2 | _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter)); | 459 | 2 | if (ancestor_null_indices.size() != 0) { | 460 | 2 | RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false)); | 461 | 2 | } | 462 | 2 | if (filter_map.has_filter()) { | 463 | 2 | auto new_rep_sz = before_rep_level_sz; | 464 | 2 | for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) { | 465 | 2 | if (nested_filter_map_data[idx - before_rep_level_sz]) { | 466 | 2 | _rep_levels[new_rep_sz] = _rep_levels[idx]; | 467 | 2 | _def_levels[new_rep_sz] = _def_levels[idx]; | 468 | 2 | new_rep_sz++; | 469 | 2 | } | 470 | 2 | } | 471 | 2 | _rep_levels.resize(new_rep_sz); | 472 | 2 | _def_levels.resize(new_rep_sz); | 473 | 2 | } | 474 | 2 | return Status::OK(); | 475 | 2 | }; | 476 | | | 477 | 2 | while (_current_range_idx < _row_ranges.range_size()) { | 478 | 2 | size_t left_row = | 479 | 2 | std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx)); | 480 | 2 | size_t right_row = std::min(left_row + batch_size - *read_rows, | 481 | 2 | (size_t)_row_ranges.get_range_to(_current_range_idx)); | 482 | 2 | _current_row_index = left_row; | 483 | 2 | RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row)); | 484 | 2 | size_t load_rows = 0; | 485 | 2 | bool cross_page = false; | 486 | 2 | size_t before_rep_level_sz = _rep_levels.size(); | 487 | 2 | RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row, | 488 | 2 | &load_rows, &cross_page)); | 489 | 2 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index)); | 490 | 2 | _filter_map_index += load_rows; | 491 | 2 | while (cross_page) { | 492 | 0 | before_rep_level_sz = _rep_levels.size(); | 493 | 0 | RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page)); | 494 | 0 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1)); | 495 | 0 | } | 496 | 2 | *read_rows += load_rows; | 497 | 2 | _current_row_index += load_rows; | 498 | 2 | _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx)); | 499 | 2 | if (*read_rows == batch_size) { | 500 | 2 | break; | 501 | 2 | } | 502 | 2 | } | 503 | 2 | *eof = _current_range_idx == _row_ranges.range_size(); | 504 | 2 | return Status::OK(); | 505 | 2 | } |
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb _ZN5doris18ScalarColumnReaderILb0ELb0EE19_read_nested_columnERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERSt10shared_ptrIKNS_9IDataTypeEERNS_9FilterMapEmPmPbb Line | Count | Source | 408 | 6 | size_t* read_rows, bool* eof, bool is_dict_filter) { | 409 | 6 | _rep_levels.clear(); | 410 | 6 | _def_levels.clear(); | 411 | | | 412 | | // Handle nullable columns | 413 | 6 | MutableColumnPtr data_column; | 414 | 6 | NullMap* map_data_column = nullptr; | 415 | 6 | if (doris_column->is_nullable()) { | 416 | 6 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 417 | | // doris_column either originates from a mutable block in vparquet_group_reader | 418 | | // or is a newly created ColumnPtr, and therefore can be modified. | 419 | 6 | auto* nullable_column = | 420 | 6 | const_cast<ColumnNullable*>(assert_cast<const ColumnNullable*>(doris_column.get())); | 421 | 6 | data_column = nullable_column->get_nested_column_ptr(); | 422 | 6 | map_data_column = &(nullable_column->get_null_map_data()); | 423 | 6 | } else { | 424 | 0 | if (_field_schema->data_type->is_nullable()) { | 425 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); | 426 | 0 | } | 427 | 0 | data_column = doris_column->assume_mutable(); | 428 | 0 | } | 429 | | | 430 | 6 | std::vector<uint16_t> null_map; | 431 | 6 | std::unordered_set<size_t> ancestor_null_indices; | 432 | 6 | std::vector<uint8_t> nested_filter_map_data; | 433 | | | 434 | 6 | auto read_and_fill_data = [&](size_t before_rep_level_sz, size_t filter_map_index) { | 435 | 6 | RETURN_IF_ERROR(_chunk_reader->fill_def(_def_levels)); | 436 | 6 | std::unique_ptr<FilterMap> nested_filter_map = std::make_unique<FilterMap>(); | 437 | 6 | if (filter_map.has_filter()) { | 438 | 6 | RETURN_IF_ERROR(gen_filter_map(filter_map, filter_map_index, before_rep_level_sz, | 439 | 6 | _rep_levels.size(), nested_filter_map_data, | 440 | 6 | &nested_filter_map)); | 441 | 6 | } | 442 | | | 443 | 6 | null_map.clear(); | 444 | 6 | ancestor_null_indices.clear(); | 445 | 6 | RETURN_IF_ERROR(gen_nested_null_map(before_rep_level_sz, _rep_levels.size(), null_map, | 446 | 6 | ancestor_null_indices)); | 447 | | | 448 | 6 | ColumnSelectVector select_vector; | 449 | 6 | { | 450 | 6 | SCOPED_RAW_TIMER(&_decode_null_map_time); | 451 | 6 | RETURN_IF_ERROR(select_vector.init( | 452 | 6 | null_map, | 453 | 6 | _rep_levels.size() - before_rep_level_sz - ancestor_null_indices.size(), | 454 | 6 | map_data_column, nested_filter_map.get(), 0, &ancestor_null_indices)); | 455 | 6 | } | 456 | | | 457 | 6 | RETURN_IF_ERROR( | 458 | 6 | _chunk_reader->decode_values(data_column, type, select_vector, is_dict_filter)); | 459 | 6 | if (ancestor_null_indices.size() != 0) { | 460 | 6 | RETURN_IF_ERROR(_chunk_reader->skip_values(ancestor_null_indices.size(), false)); | 461 | 6 | } | 462 | 6 | if (filter_map.has_filter()) { | 463 | 6 | auto new_rep_sz = before_rep_level_sz; | 464 | 6 | for (size_t idx = before_rep_level_sz; idx < _rep_levels.size(); idx++) { | 465 | 6 | if (nested_filter_map_data[idx - before_rep_level_sz]) { | 466 | 6 | _rep_levels[new_rep_sz] = _rep_levels[idx]; | 467 | 6 | _def_levels[new_rep_sz] = _def_levels[idx]; | 468 | 6 | new_rep_sz++; | 469 | 6 | } | 470 | 6 | } | 471 | 6 | _rep_levels.resize(new_rep_sz); | 472 | 6 | _def_levels.resize(new_rep_sz); | 473 | 6 | } | 474 | 6 | return Status::OK(); | 475 | 6 | }; | 476 | | | 477 | 8 | while (_current_range_idx < _row_ranges.range_size()) { | 478 | 6 | size_t left_row = | 479 | 6 | std::max(_current_row_index, _row_ranges.get_range_from(_current_range_idx)); | 480 | 6 | size_t right_row = std::min(left_row + batch_size - *read_rows, | 481 | 6 | (size_t)_row_ranges.get_range_to(_current_range_idx)); | 482 | 6 | _current_row_index = left_row; | 483 | 6 | RETURN_IF_ERROR(_chunk_reader->seek_to_nested_row(left_row)); | 484 | 6 | size_t load_rows = 0; | 485 | 6 | bool cross_page = false; | 486 | 6 | size_t before_rep_level_sz = _rep_levels.size(); | 487 | 6 | RETURN_IF_ERROR(_chunk_reader->load_page_nested_rows(_rep_levels, right_row - left_row, | 488 | 6 | &load_rows, &cross_page)); | 489 | 6 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index)); | 490 | 6 | _filter_map_index += load_rows; | 491 | 6 | while (cross_page) { | 492 | 0 | before_rep_level_sz = _rep_levels.size(); | 493 | 0 | RETURN_IF_ERROR(_chunk_reader->load_cross_page_nested_row(_rep_levels, &cross_page)); | 494 | 0 | RETURN_IF_ERROR(read_and_fill_data(before_rep_level_sz, _filter_map_index - 1)); | 495 | 0 | } | 496 | 6 | *read_rows += load_rows; | 497 | 6 | _current_row_index += load_rows; | 498 | 6 | _current_range_idx += (_current_row_index == _row_ranges.get_range_to(_current_range_idx)); | 499 | 6 | if (*read_rows == batch_size) { | 500 | 4 | break; | 501 | 4 | } | 502 | 6 | } | 503 | 6 | *eof = _current_range_idx == _row_ranges.range_size(); | 504 | 6 | return Status::OK(); | 505 | 6 | } |
|
506 | | |
507 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
508 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::read_dict_values_to_column( |
509 | 2 | MutableColumnPtr& doris_column, bool* has_dict) { |
510 | 2 | bool loaded; |
511 | 2 | RETURN_IF_ERROR(_try_load_dict_page(&loaded, has_dict)); |
512 | 2 | if (loaded && *has_dict) { |
513 | 2 | return _chunk_reader->read_dict_values_to_column(doris_column); |
514 | 2 | } |
515 | 0 | return Status::OK(); |
516 | 2 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb _ZN5doris18ScalarColumnReaderILb0ELb0EE26read_dict_values_to_columnERNS_3COWINS_7IColumnEE11mutable_ptrIS3_EEPb Line | Count | Source | 509 | 2 | MutableColumnPtr& doris_column, bool* has_dict) { | 510 | 2 | bool loaded; | 511 | 2 | RETURN_IF_ERROR(_try_load_dict_page(&loaded, has_dict)); | 512 | 2 | if (loaded && *has_dict) { | 513 | 2 | return _chunk_reader->read_dict_values_to_column(doris_column); | 514 | 2 | } | 515 | 0 | return Status::OK(); | 516 | 2 | } |
|
517 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
518 | | Result<MutableColumnPtr> |
519 | | ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::convert_dict_column_to_string_column( |
520 | 0 | const ColumnInt32* dict_column) { |
521 | 0 | return _chunk_reader->convert_dict_column_to_string_column(dict_column); |
522 | 0 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb0EE36convert_dict_column_to_string_columnEPKNS_12ColumnVectorILNS_13PrimitiveTypeE5EEE |
523 | | |
524 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
525 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::_try_load_dict_page(bool* loaded, |
526 | 2 | bool* has_dict) { |
527 | | // _chunk_reader init will load first page header to check whether has dict page |
528 | 2 | *loaded = true; |
529 | 2 | *has_dict = _chunk_reader->has_dict(); |
530 | 2 | return Status::OK(); |
531 | 2 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE19_try_load_dict_pageEPbS2_ Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb0EE19_try_load_dict_pageEPbS2_ Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE19_try_load_dict_pageEPbS2_ _ZN5doris18ScalarColumnReaderILb0ELb0EE19_try_load_dict_pageEPbS2_ Line | Count | Source | 526 | 2 | bool* has_dict) { | 527 | | // _chunk_reader init will load first page header to check whether has dict page | 528 | 2 | *loaded = true; | 529 | 2 | *has_dict = _chunk_reader->has_dict(); | 530 | 2 | return Status::OK(); | 531 | 2 | } |
|
532 | | |
533 | | template <bool IN_COLLECTION, bool OFFSET_INDEX> |
534 | | Status ScalarColumnReader<IN_COLLECTION, OFFSET_INDEX>::read_column_data( |
535 | | ColumnPtr& doris_column, const DataTypePtr& type, |
536 | | const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map, |
537 | | size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter, |
538 | 190 | int64_t real_column_size) { |
539 | 190 | if (_converter == nullptr) { |
540 | 105 | _converter = parquet::PhysicalToLogicalConverter::get_converter( |
541 | 105 | _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter); |
542 | 105 | if (!_converter->support()) { |
543 | 0 | return Status::InternalError( |
544 | 0 | "The column type of '{}' is not supported: {}, is_dict_filter: {}, " |
545 | 0 | "src_logical_type: {}, dst_logical_type: {}", |
546 | 0 | _field_schema->name, _converter->get_error_msg(), is_dict_filter, |
547 | 0 | _field_schema->data_type->get_name(), type->get_name()); |
548 | 0 | } |
549 | 105 | } |
550 | | // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column? |
551 | 190 | ColumnPtr resolved_column = |
552 | 190 | _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type, |
553 | 190 | doris_column, type, is_dict_filter); |
554 | 190 | DataTypePtr& resolved_type = _converter->get_physical_type(); |
555 | | |
556 | 190 | _def_levels.clear(); |
557 | 190 | _rep_levels.clear(); |
558 | 190 | *read_rows = 0; |
559 | | |
560 | 190 | if (_in_nested) { |
561 | 8 | RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size, |
562 | 8 | read_rows, eof, is_dict_filter)); |
563 | 8 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, |
564 | 8 | is_dict_filter); |
565 | 8 | } |
566 | | |
567 | 182 | int64_t right_row = 0; |
568 | 182 | if constexpr (OFFSET_INDEX == false) { |
569 | 182 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); |
570 | 182 | right_row = _chunk_reader->page_end_row(); |
571 | 182 | } else { |
572 | 0 | right_row = _chunk_reader->page_end_row(); |
573 | 0 | } |
574 | | |
575 | 182 | do { |
576 | | // generate the row ranges that should be read |
577 | 182 | RowRanges read_ranges; |
578 | 182 | _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges); |
579 | 182 | if (read_ranges.count() == 0) { |
580 | | // skip the whole page |
581 | 63 | _current_row_index = right_row; |
582 | 119 | } else { |
583 | 119 | bool skip_whole_batch = false; |
584 | | // Determining whether to skip page or batch will increase the calculation time. |
585 | | // When the filtering effect is greater than 60%, it is possible to skip the page or batch. |
586 | 119 | if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) { |
587 | | // lazy read |
588 | 0 | size_t remaining_num_values = read_ranges.count(); |
589 | 0 | if (batch_size >= remaining_num_values && |
590 | 0 | filter_map.can_filter_all(remaining_num_values, _filter_map_index)) { |
591 | | // We can skip the whole page if the remaining values are filtered by predicate columns |
592 | 0 | _filter_map_index += remaining_num_values; |
593 | 0 | _current_row_index = right_row; |
594 | 0 | *read_rows = remaining_num_values; |
595 | 0 | break; |
596 | 0 | } |
597 | 0 | skip_whole_batch = batch_size <= remaining_num_values && |
598 | 0 | filter_map.can_filter_all(batch_size, _filter_map_index); |
599 | 0 | if (skip_whole_batch) { |
600 | 0 | _filter_map_index += batch_size; |
601 | 0 | } |
602 | 0 | } |
603 | | // load page data to decode or skip values |
604 | 119 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); |
605 | 119 | RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent()); |
606 | 119 | size_t has_read = 0; |
607 | 248 | for (size_t idx = 0; idx < read_ranges.range_size(); idx++) { |
608 | 152 | auto range = read_ranges.get_range(idx); |
609 | | // generate the skipped values |
610 | 152 | size_t skip_values = range.from() - _current_row_index; |
611 | 152 | RETURN_IF_ERROR(_skip_values(skip_values)); |
612 | 152 | _current_row_index += skip_values; |
613 | | // generate the read values |
614 | 152 | size_t read_values = |
615 | 152 | std::min((size_t)(range.to() - range.from()), batch_size - has_read); |
616 | 152 | if (skip_whole_batch) { |
617 | 0 | RETURN_IF_ERROR(_skip_values(read_values)); |
618 | 152 | } else { |
619 | 152 | RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type, |
620 | 152 | filter_map, is_dict_filter)); |
621 | 152 | } |
622 | 152 | has_read += read_values; |
623 | 152 | *read_rows += read_values; |
624 | 152 | _current_row_index += read_values; |
625 | 152 | if (has_read == batch_size) { |
626 | 23 | break; |
627 | 23 | } |
628 | 152 | } |
629 | 119 | } |
630 | 182 | } while (false); |
631 | | |
632 | 182 | if (right_row == _current_row_index) { |
633 | 97 | if (!_chunk_reader->has_next_page()) { |
634 | 97 | *eof = true; |
635 | 97 | } else { |
636 | 0 | RETURN_IF_ERROR(_chunk_reader->next_page()); |
637 | 0 | } |
638 | 97 | } |
639 | | |
640 | 182 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, |
641 | 182 | is_dict_filter); |
642 | 182 | } Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb1ELb1EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl _ZN5doris18ScalarColumnReaderILb1ELb0EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl Line | Count | Source | 538 | 2 | int64_t real_column_size) { | 539 | 2 | if (_converter == nullptr) { | 540 | 2 | _converter = parquet::PhysicalToLogicalConverter::get_converter( | 541 | 2 | _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter); | 542 | 2 | if (!_converter->support()) { | 543 | 0 | return Status::InternalError( | 544 | 0 | "The column type of '{}' is not supported: {}, is_dict_filter: {}, " | 545 | 0 | "src_logical_type: {}, dst_logical_type: {}", | 546 | 0 | _field_schema->name, _converter->get_error_msg(), is_dict_filter, | 547 | 0 | _field_schema->data_type->get_name(), type->get_name()); | 548 | 0 | } | 549 | 2 | } | 550 | | // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column? | 551 | 2 | ColumnPtr resolved_column = | 552 | 2 | _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type, | 553 | 2 | doris_column, type, is_dict_filter); | 554 | 2 | DataTypePtr& resolved_type = _converter->get_physical_type(); | 555 | | | 556 | 2 | _def_levels.clear(); | 557 | 2 | _rep_levels.clear(); | 558 | 2 | *read_rows = 0; | 559 | | | 560 | 2 | if (_in_nested) { | 561 | 2 | RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size, | 562 | 2 | read_rows, eof, is_dict_filter)); | 563 | 2 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, | 564 | 2 | is_dict_filter); | 565 | 2 | } | 566 | | | 567 | 0 | int64_t right_row = 0; | 568 | 0 | if constexpr (OFFSET_INDEX == false) { | 569 | 0 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); | 570 | 0 | right_row = _chunk_reader->page_end_row(); | 571 | | } else { | 572 | | right_row = _chunk_reader->page_end_row(); | 573 | | } | 574 | | | 575 | 0 | do { | 576 | | // generate the row ranges that should be read | 577 | 0 | RowRanges read_ranges; | 578 | 0 | _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges); | 579 | 0 | if (read_ranges.count() == 0) { | 580 | | // skip the whole page | 581 | 0 | _current_row_index = right_row; | 582 | 0 | } else { | 583 | 0 | bool skip_whole_batch = false; | 584 | | // Determining whether to skip page or batch will increase the calculation time. | 585 | | // When the filtering effect is greater than 60%, it is possible to skip the page or batch. | 586 | 0 | if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) { | 587 | | // lazy read | 588 | 0 | size_t remaining_num_values = read_ranges.count(); | 589 | 0 | if (batch_size >= remaining_num_values && | 590 | 0 | filter_map.can_filter_all(remaining_num_values, _filter_map_index)) { | 591 | | // We can skip the whole page if the remaining values are filtered by predicate columns | 592 | 0 | _filter_map_index += remaining_num_values; | 593 | 0 | _current_row_index = right_row; | 594 | 0 | *read_rows = remaining_num_values; | 595 | 0 | break; | 596 | 0 | } | 597 | 0 | skip_whole_batch = batch_size <= remaining_num_values && | 598 | 0 | filter_map.can_filter_all(batch_size, _filter_map_index); | 599 | 0 | if (skip_whole_batch) { | 600 | 0 | _filter_map_index += batch_size; | 601 | 0 | } | 602 | 0 | } | 603 | | // load page data to decode or skip values | 604 | 0 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); | 605 | 0 | RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent()); | 606 | 0 | size_t has_read = 0; | 607 | 0 | for (size_t idx = 0; idx < read_ranges.range_size(); idx++) { | 608 | 0 | auto range = read_ranges.get_range(idx); | 609 | | // generate the skipped values | 610 | 0 | size_t skip_values = range.from() - _current_row_index; | 611 | 0 | RETURN_IF_ERROR(_skip_values(skip_values)); | 612 | 0 | _current_row_index += skip_values; | 613 | | // generate the read values | 614 | 0 | size_t read_values = | 615 | 0 | std::min((size_t)(range.to() - range.from()), batch_size - has_read); | 616 | 0 | if (skip_whole_batch) { | 617 | 0 | RETURN_IF_ERROR(_skip_values(read_values)); | 618 | 0 | } else { | 619 | 0 | RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type, | 620 | 0 | filter_map, is_dict_filter)); | 621 | 0 | } | 622 | 0 | has_read += read_values; | 623 | 0 | *read_rows += read_values; | 624 | 0 | _current_row_index += read_values; | 625 | 0 | if (has_read == batch_size) { | 626 | 0 | break; | 627 | 0 | } | 628 | 0 | } | 629 | 0 | } | 630 | 0 | } while (false); | 631 | | | 632 | 0 | if (right_row == _current_row_index) { | 633 | 0 | if (!_chunk_reader->has_next_page()) { | 634 | 0 | *eof = true; | 635 | 0 | } else { | 636 | 0 | RETURN_IF_ERROR(_chunk_reader->next_page()); | 637 | 0 | } | 638 | 0 | } | 639 | | | 640 | 0 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, | 641 | 0 | is_dict_filter); | 642 | 0 | } |
Unexecuted instantiation: _ZN5doris18ScalarColumnReaderILb0ELb1EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl _ZN5doris18ScalarColumnReaderILb0ELb0EE16read_column_dataERNS_3COWINS_7IColumnEE13immutable_ptrIS3_EERKSt10shared_ptrIKNS_9IDataTypeEERKS8_INS_23TableSchemaChangeHelper4NodeEERNS_9FilterMapEmPmPbbl Line | Count | Source | 538 | 188 | int64_t real_column_size) { | 539 | 188 | if (_converter == nullptr) { | 540 | 103 | _converter = parquet::PhysicalToLogicalConverter::get_converter( | 541 | 103 | _field_schema, _field_schema->data_type, type, _ctz, is_dict_filter); | 542 | 103 | if (!_converter->support()) { | 543 | 0 | return Status::InternalError( | 544 | 0 | "The column type of '{}' is not supported: {}, is_dict_filter: {}, " | 545 | 0 | "src_logical_type: {}, dst_logical_type: {}", | 546 | 0 | _field_schema->name, _converter->get_error_msg(), is_dict_filter, | 547 | 0 | _field_schema->data_type->get_name(), type->get_name()); | 548 | 0 | } | 549 | 103 | } | 550 | | // !FIXME: We should verify whether the get_physical_column logic is correct, why do we return a doris_column? | 551 | 188 | ColumnPtr resolved_column = | 552 | 188 | _converter->get_physical_column(_field_schema->physical_type, _field_schema->data_type, | 553 | 188 | doris_column, type, is_dict_filter); | 554 | 188 | DataTypePtr& resolved_type = _converter->get_physical_type(); | 555 | | | 556 | 188 | _def_levels.clear(); | 557 | 188 | _rep_levels.clear(); | 558 | 188 | *read_rows = 0; | 559 | | | 560 | 188 | if (_in_nested) { | 561 | 6 | RETURN_IF_ERROR(_read_nested_column(resolved_column, resolved_type, filter_map, batch_size, | 562 | 6 | read_rows, eof, is_dict_filter)); | 563 | 6 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, | 564 | 6 | is_dict_filter); | 565 | 6 | } | 566 | | | 567 | 182 | int64_t right_row = 0; | 568 | 182 | if constexpr (OFFSET_INDEX == false) { | 569 | 182 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); | 570 | 182 | right_row = _chunk_reader->page_end_row(); | 571 | | } else { | 572 | | right_row = _chunk_reader->page_end_row(); | 573 | | } | 574 | | | 575 | 182 | do { | 576 | | // generate the row ranges that should be read | 577 | 182 | RowRanges read_ranges; | 578 | 182 | _generate_read_ranges(RowRange {_current_row_index, right_row}, &read_ranges); | 579 | 182 | if (read_ranges.count() == 0) { | 580 | | // skip the whole page | 581 | 63 | _current_row_index = right_row; | 582 | 119 | } else { | 583 | 119 | bool skip_whole_batch = false; | 584 | | // Determining whether to skip page or batch will increase the calculation time. | 585 | | // When the filtering effect is greater than 60%, it is possible to skip the page or batch. | 586 | 119 | if (filter_map.has_filter() && filter_map.filter_ratio() > 0.6) { | 587 | | // lazy read | 588 | 0 | size_t remaining_num_values = read_ranges.count(); | 589 | 0 | if (batch_size >= remaining_num_values && | 590 | 0 | filter_map.can_filter_all(remaining_num_values, _filter_map_index)) { | 591 | | // We can skip the whole page if the remaining values are filtered by predicate columns | 592 | 0 | _filter_map_index += remaining_num_values; | 593 | 0 | _current_row_index = right_row; | 594 | 0 | *read_rows = remaining_num_values; | 595 | 0 | break; | 596 | 0 | } | 597 | 0 | skip_whole_batch = batch_size <= remaining_num_values && | 598 | 0 | filter_map.can_filter_all(batch_size, _filter_map_index); | 599 | 0 | if (skip_whole_batch) { | 600 | 0 | _filter_map_index += batch_size; | 601 | 0 | } | 602 | 0 | } | 603 | | // load page data to decode or skip values | 604 | 119 | RETURN_IF_ERROR(_chunk_reader->parse_page_header()); | 605 | 119 | RETURN_IF_ERROR(_chunk_reader->load_page_data_idempotent()); | 606 | 119 | size_t has_read = 0; | 607 | 248 | for (size_t idx = 0; idx < read_ranges.range_size(); idx++) { | 608 | 152 | auto range = read_ranges.get_range(idx); | 609 | | // generate the skipped values | 610 | 152 | size_t skip_values = range.from() - _current_row_index; | 611 | 152 | RETURN_IF_ERROR(_skip_values(skip_values)); | 612 | 152 | _current_row_index += skip_values; | 613 | | // generate the read values | 614 | 152 | size_t read_values = | 615 | 152 | std::min((size_t)(range.to() - range.from()), batch_size - has_read); | 616 | 152 | if (skip_whole_batch) { | 617 | 0 | RETURN_IF_ERROR(_skip_values(read_values)); | 618 | 152 | } else { | 619 | 152 | RETURN_IF_ERROR(_read_values(read_values, resolved_column, resolved_type, | 620 | 152 | filter_map, is_dict_filter)); | 621 | 152 | } | 622 | 152 | has_read += read_values; | 623 | 152 | *read_rows += read_values; | 624 | 152 | _current_row_index += read_values; | 625 | 152 | if (has_read == batch_size) { | 626 | 23 | break; | 627 | 23 | } | 628 | 152 | } | 629 | 119 | } | 630 | 182 | } while (false); | 631 | | | 632 | 182 | if (right_row == _current_row_index) { | 633 | 97 | if (!_chunk_reader->has_next_page()) { | 634 | 97 | *eof = true; | 635 | 97 | } else { | 636 | 0 | RETURN_IF_ERROR(_chunk_reader->next_page()); | 637 | 0 | } | 638 | 97 | } | 639 | | | 640 | 182 | return _converter->convert(resolved_column, _field_schema->data_type, type, doris_column, | 641 | 182 | is_dict_filter); | 642 | 182 | } |
|
643 | | |
644 | | Status ArrayColumnReader::init(std::unique_ptr<ParquetColumnReader> element_reader, |
645 | 2 | FieldSchema* field) { |
646 | 2 | _field_schema = field; |
647 | 2 | _element_reader = std::move(element_reader); |
648 | 2 | return Status::OK(); |
649 | 2 | } |
650 | | |
651 | | Status ArrayColumnReader::read_column_data( |
652 | | ColumnPtr& doris_column, const DataTypePtr& type, |
653 | | const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map, |
654 | | size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter, |
655 | 2 | int64_t real_column_size) { |
656 | 2 | MutableColumnPtr data_column; |
657 | 2 | NullMap* null_map_ptr = nullptr; |
658 | 2 | if (doris_column->is_nullable()) { |
659 | 2 | auto mutable_column = doris_column->assume_mutable(); |
660 | 2 | auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get()); |
661 | 2 | null_map_ptr = &nullable_column->get_null_map_data(); |
662 | 2 | data_column = nullable_column->get_nested_column_ptr(); |
663 | 2 | } else { |
664 | 0 | if (_field_schema->data_type->is_nullable()) { |
665 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); |
666 | 0 | } |
667 | 0 | data_column = doris_column->assume_mutable(); |
668 | 0 | } |
669 | 2 | if (type->get_primitive_type() != PrimitiveType::TYPE_ARRAY) { |
670 | 0 | return Status::Corruption( |
671 | 0 | "Wrong data type for column '{}', expected Array type, actual type: {}.", |
672 | 0 | _field_schema->name, type->get_name()); |
673 | 0 | } |
674 | | |
675 | 2 | ColumnPtr& element_column = assert_cast<ColumnArray&>(*data_column).get_data_ptr(); |
676 | 2 | const DataTypePtr& element_type = |
677 | 2 | (assert_cast<const DataTypeArray*>(remove_nullable(type).get()))->get_nested_type(); |
678 | | // read nested column |
679 | 2 | RETURN_IF_ERROR(_element_reader->read_column_data(element_column, element_type, |
680 | 2 | root_node->get_element_node(), filter_map, |
681 | 2 | batch_size, read_rows, eof, is_dict_filter)); |
682 | 2 | if (*read_rows == 0) { |
683 | 0 | return Status::OK(); |
684 | 0 | } |
685 | | |
686 | 2 | ColumnArray::Offsets64& offsets_data = assert_cast<ColumnArray&>(*data_column).get_offsets(); |
687 | | // fill offset and null map |
688 | 2 | fill_array_offset(_field_schema, offsets_data, null_map_ptr, _element_reader->get_rep_level(), |
689 | 2 | _element_reader->get_def_level()); |
690 | 2 | DCHECK_EQ(element_column->size(), offsets_data.back()); |
691 | 2 | #ifndef NDEBUG |
692 | 2 | doris_column->sanity_check(); |
693 | 2 | #endif |
694 | 2 | return Status::OK(); |
695 | 2 | } |
696 | | |
697 | | Status MapColumnReader::init(std::unique_ptr<ParquetColumnReader> key_reader, |
698 | | std::unique_ptr<ParquetColumnReader> value_reader, |
699 | 0 | FieldSchema* field) { |
700 | 0 | _field_schema = field; |
701 | 0 | _key_reader = std::move(key_reader); |
702 | 0 | _value_reader = std::move(value_reader); |
703 | 0 | return Status::OK(); |
704 | 0 | } |
705 | | |
706 | | Status MapColumnReader::read_column_data( |
707 | | ColumnPtr& doris_column, const DataTypePtr& type, |
708 | | const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map, |
709 | | size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter, |
710 | 0 | int64_t real_column_size) { |
711 | 0 | MutableColumnPtr data_column; |
712 | 0 | NullMap* null_map_ptr = nullptr; |
713 | 0 | if (doris_column->is_nullable()) { |
714 | 0 | auto mutable_column = doris_column->assume_mutable(); |
715 | 0 | auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get()); |
716 | 0 | null_map_ptr = &nullable_column->get_null_map_data(); |
717 | 0 | data_column = nullable_column->get_nested_column_ptr(); |
718 | 0 | } else { |
719 | 0 | if (_field_schema->data_type->is_nullable()) { |
720 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); |
721 | 0 | } |
722 | 0 | data_column = doris_column->assume_mutable(); |
723 | 0 | } |
724 | 0 | if (remove_nullable(type)->get_primitive_type() != PrimitiveType::TYPE_MAP) { |
725 | 0 | return Status::Corruption( |
726 | 0 | "Wrong data type for column '{}', expected Map type, actual type id {}.", |
727 | 0 | _field_schema->name, type->get_name()); |
728 | 0 | } |
729 | | |
730 | 0 | auto& map = assert_cast<ColumnMap&>(*data_column); |
731 | 0 | const DataTypePtr& key_type = |
732 | 0 | assert_cast<const DataTypeMap*>(remove_nullable(type).get())->get_key_type(); |
733 | 0 | const DataTypePtr& value_type = |
734 | 0 | assert_cast<const DataTypeMap*>(remove_nullable(type).get())->get_value_type(); |
735 | 0 | ColumnPtr& key_column = map.get_keys_ptr(); |
736 | 0 | ColumnPtr& value_column = map.get_values_ptr(); |
737 | |
|
738 | 0 | size_t key_rows = 0; |
739 | 0 | size_t value_rows = 0; |
740 | 0 | bool key_eof = false; |
741 | 0 | bool value_eof = false; |
742 | 0 | int64_t orig_col_column_size = key_column->size(); |
743 | |
|
744 | 0 | RETURN_IF_ERROR(_key_reader->read_column_data(key_column, key_type, root_node->get_key_node(), |
745 | 0 | filter_map, batch_size, &key_rows, &key_eof, |
746 | 0 | is_dict_filter)); |
747 | | |
748 | 0 | while (value_rows < key_rows && !value_eof) { |
749 | 0 | size_t loop_rows = 0; |
750 | 0 | RETURN_IF_ERROR(_value_reader->read_column_data( |
751 | 0 | value_column, value_type, root_node->get_value_node(), filter_map, |
752 | 0 | key_rows - value_rows, &loop_rows, &value_eof, is_dict_filter, |
753 | 0 | key_column->size() - orig_col_column_size)); |
754 | 0 | value_rows += loop_rows; |
755 | 0 | } |
756 | 0 | DCHECK_EQ(key_rows, value_rows); |
757 | 0 | *read_rows = key_rows; |
758 | 0 | *eof = key_eof; |
759 | |
|
760 | 0 | if (*read_rows == 0) { |
761 | 0 | return Status::OK(); |
762 | 0 | } |
763 | | |
764 | 0 | DCHECK_EQ(key_column->size(), value_column->size()); |
765 | | // fill offset and null map |
766 | 0 | fill_array_offset(_field_schema, map.get_offsets(), null_map_ptr, _key_reader->get_rep_level(), |
767 | 0 | _key_reader->get_def_level()); |
768 | 0 | DCHECK_EQ(key_column->size(), map.get_offsets().back()); |
769 | 0 | #ifndef NDEBUG |
770 | 0 | doris_column->sanity_check(); |
771 | 0 | #endif |
772 | 0 | return Status::OK(); |
773 | 0 | } |
774 | | |
775 | | Status StructColumnReader::init( |
776 | | std::unordered_map<std::string, std::unique_ptr<ParquetColumnReader>>&& child_readers, |
777 | 10 | FieldSchema* field) { |
778 | 10 | _field_schema = field; |
779 | 10 | _child_readers = std::move(child_readers); |
780 | 10 | return Status::OK(); |
781 | 10 | } |
782 | | Status StructColumnReader::read_column_data( |
783 | | ColumnPtr& doris_column, const DataTypePtr& type, |
784 | | const std::shared_ptr<TableSchemaChangeHelper::Node>& root_node, FilterMap& filter_map, |
785 | | size_t batch_size, size_t* read_rows, bool* eof, bool is_dict_filter, |
786 | 10 | int64_t real_column_size) { |
787 | 10 | MutableColumnPtr data_column; |
788 | 10 | NullMap* null_map_ptr = nullptr; |
789 | 10 | if (doris_column->is_nullable()) { |
790 | 10 | auto mutable_column = doris_column->assume_mutable(); |
791 | 10 | auto* nullable_column = assert_cast<ColumnNullable*>(mutable_column.get()); |
792 | 10 | null_map_ptr = &nullable_column->get_null_map_data(); |
793 | 10 | data_column = nullable_column->get_nested_column_ptr(); |
794 | 10 | } else { |
795 | 0 | if (_field_schema->data_type->is_nullable()) { |
796 | 0 | return Status::Corruption("Not nullable column has null values in parquet file"); |
797 | 0 | } |
798 | 0 | data_column = doris_column->assume_mutable(); |
799 | 0 | } |
800 | 10 | if (type->get_primitive_type() != PrimitiveType::TYPE_STRUCT) { |
801 | 0 | return Status::Corruption( |
802 | 0 | "Wrong data type for column '{}', expected Struct type, actual type id {}.", |
803 | 0 | _field_schema->name, type->get_name()); |
804 | 0 | } |
805 | | |
806 | 10 | auto& doris_struct = assert_cast<ColumnStruct&>(*data_column); |
807 | 10 | const auto* doris_struct_type = assert_cast<const DataTypeStruct*>(remove_nullable(type).get()); |
808 | | |
809 | 10 | int64_t not_missing_column_id = -1; |
810 | 10 | size_t not_missing_orig_column_size = 0; |
811 | 10 | std::vector<size_t> missing_column_idxs {}; |
812 | 10 | std::vector<size_t> skip_reading_column_idxs {}; |
813 | | |
814 | 10 | _read_column_names.clear(); |
815 | | |
816 | 34 | for (size_t i = 0; i < doris_struct.tuple_size(); ++i) { |
817 | 24 | ColumnPtr& doris_field = doris_struct.get_column_ptr(i); |
818 | 24 | auto& doris_type = doris_struct_type->get_element(i); |
819 | 24 | auto& doris_name = doris_struct_type->get_element_name(i); |
820 | 24 | if (!root_node->children_column_exists(doris_name)) { |
821 | 0 | missing_column_idxs.push_back(i); |
822 | 0 | VLOG_DEBUG << "[ParquetReader] Missing column in schema: column_idx[" << i |
823 | 0 | << "], doris_name: " << doris_name << " (column not exists in root node)"; |
824 | 0 | continue; |
825 | 0 | } |
826 | 24 | auto file_name = root_node->children_file_column_name(doris_name); |
827 | | |
828 | | // Check if this is a SkipReadingReader - we should skip it when choosing reference column |
829 | | // because SkipReadingReader doesn't know the actual data size in nested context |
830 | 24 | bool is_skip_reader = |
831 | 24 | dynamic_cast<SkipReadingReader*>(_child_readers[file_name].get()) != nullptr; |
832 | | |
833 | 24 | if (is_skip_reader) { |
834 | | // Store SkipReadingReader columns to fill them later based on reference column size |
835 | 8 | skip_reading_column_idxs.push_back(i); |
836 | 8 | continue; |
837 | 8 | } |
838 | | |
839 | | // Only add non-SkipReadingReader columns to _read_column_names |
840 | | // This ensures get_rep_level() and get_def_level() return valid levels |
841 | 16 | _read_column_names.emplace_back(file_name); |
842 | | |
843 | 16 | size_t field_rows = 0; |
844 | 16 | bool field_eof = false; |
845 | 16 | if (not_missing_column_id == -1) { |
846 | 10 | not_missing_column_id = i; |
847 | 10 | not_missing_orig_column_size = doris_field->size(); |
848 | 10 | RETURN_IF_ERROR(_child_readers[file_name]->read_column_data( |
849 | 10 | doris_field, doris_type, root_node->get_children_node(doris_name), filter_map, |
850 | 10 | batch_size, &field_rows, &field_eof, is_dict_filter)); |
851 | 10 | *read_rows = field_rows; |
852 | 10 | *eof = field_eof; |
853 | | /* |
854 | | * Considering the issue in the `_read_nested_column` function where data may span across pages, leading |
855 | | * to missing definition and repetition levels, when filling the null_map of the struct later, it is |
856 | | * crucial to use the definition and repetition levels from the first read column |
857 | | * (since `_read_nested_column` is not called repeatedly). |
858 | | * |
859 | | * It is worth mentioning that, theoretically, any sub-column can be chosen to fill the null_map, |
860 | | * and selecting the shortest one will offer better performance |
861 | | */ |
862 | 10 | } else { |
863 | 12 | while (field_rows < *read_rows && !field_eof) { |
864 | 6 | size_t loop_rows = 0; |
865 | 6 | RETURN_IF_ERROR(_child_readers[file_name]->read_column_data( |
866 | 6 | doris_field, doris_type, root_node->get_children_node(doris_name), |
867 | 6 | filter_map, *read_rows - field_rows, &loop_rows, &field_eof, |
868 | 6 | is_dict_filter)); |
869 | 6 | field_rows += loop_rows; |
870 | 6 | } |
871 | 6 | DCHECK_EQ(*read_rows, field_rows); |
872 | | // DCHECK_EQ(*eof, field_eof); |
873 | 6 | } |
874 | 16 | } |
875 | | |
876 | 10 | int64_t missing_column_sz = -1; |
877 | | |
878 | 10 | if (not_missing_column_id == -1) { |
879 | | // All queried columns are missing in the file (e.g., all added after schema change) |
880 | | // We need to pick a column from _field_schema children that exists in the file for RL/DL reference |
881 | 0 | std::string reference_file_column_name; |
882 | 0 | std::unique_ptr<ParquetColumnReader>* reference_reader = nullptr; |
883 | |
|
884 | 0 | for (const auto& child : _field_schema->children) { |
885 | 0 | auto it = _child_readers.find(child.name); |
886 | 0 | if (it != _child_readers.end()) { |
887 | | // Skip SkipReadingReader as they don't have valid RL/DL |
888 | 0 | bool is_skip_reader = dynamic_cast<SkipReadingReader*>(it->second.get()) != nullptr; |
889 | 0 | if (!is_skip_reader) { |
890 | 0 | reference_file_column_name = child.name; |
891 | 0 | reference_reader = &(it->second); |
892 | 0 | break; |
893 | 0 | } |
894 | 0 | } |
895 | 0 | } |
896 | |
|
897 | 0 | if (reference_reader != nullptr) { |
898 | | // Read the reference column to get correct RL/DL information |
899 | | // TODO: Optimize by only reading RL/DL without actual data decoding |
900 | | |
901 | | // We need to find the FieldSchema for the reference column from _field_schema children |
902 | 0 | FieldSchema* ref_field_schema = nullptr; |
903 | 0 | for (auto& child : _field_schema->children) { |
904 | 0 | if (child.name == reference_file_column_name) { |
905 | 0 | ref_field_schema = &child; |
906 | 0 | break; |
907 | 0 | } |
908 | 0 | } |
909 | |
|
910 | 0 | if (ref_field_schema == nullptr) { |
911 | 0 | return Status::InternalError( |
912 | 0 | "Cannot find field schema for reference column '{}' in struct '{}'", |
913 | 0 | reference_file_column_name, _field_schema->name); |
914 | 0 | } |
915 | | |
916 | | // Create a temporary column to hold the data (we'll use its size for missing_column_sz) |
917 | 0 | ColumnPtr temp_column = ref_field_schema->data_type->create_column(); |
918 | 0 | auto temp_type = ref_field_schema->data_type; |
919 | |
|
920 | 0 | size_t field_rows = 0; |
921 | 0 | bool field_eof = false; |
922 | | |
923 | | // Use ConstNode for the reference column instead of looking up from root_node. |
924 | | // The reference column is only used to get RL/DL information for determining the number |
925 | | // of elements in the struct. It may be a column that has been dropped from the table |
926 | | // schema (e.g., 'removed' field), but still exists in older parquet files. |
927 | | // Since we don't need schema mapping for this column (we just need its RL/DL levels), |
928 | | // using ConstNode is safe and avoids the issue where the reference column doesn't exist |
929 | | // in root_node (because it was dropped from table schema). |
930 | 0 | auto ref_child_node = TableSchemaChangeHelper::ConstNode::get_instance(); |
931 | 0 | not_missing_orig_column_size = temp_column->size(); |
932 | |
|
933 | 0 | RETURN_IF_ERROR((*reference_reader) |
934 | 0 | ->read_column_data(temp_column, temp_type, ref_child_node, |
935 | 0 | filter_map, batch_size, &field_rows, |
936 | 0 | &field_eof, is_dict_filter)); |
937 | | |
938 | 0 | *read_rows = field_rows; |
939 | 0 | *eof = field_eof; |
940 | | |
941 | | // Store this reference column name for get_rep_level/get_def_level to use |
942 | 0 | _read_column_names.emplace_back(reference_file_column_name); |
943 | |
|
944 | 0 | missing_column_sz = temp_column->size() - not_missing_orig_column_size; |
945 | 0 | } else { |
946 | 0 | return Status::Corruption( |
947 | 0 | "Cannot read struct '{}': all queried columns are missing and no reference " |
948 | 0 | "column found in file", |
949 | 0 | _field_schema->name); |
950 | 0 | } |
951 | 0 | } |
952 | | |
953 | | // This missing_column_sz is not *read_rows. Because read_rows returns the number of rows. |
954 | | // For example: suppose we have a column array<struct<a:int,b:string>>, |
955 | | // where b is a newly added column, that is, a missing column. |
956 | | // There are two rows of data in this column, |
957 | | // [{1,null},{2,null},{3,null}] |
958 | | // [{4,null},{5,null}] |
959 | | // When you first read subcolumn a, you read 5 data items and the value of *read_rows is 2. |
960 | | // You should insert 5 records into subcolumn b instead of 2. |
961 | 10 | if (missing_column_sz == -1) { |
962 | 10 | missing_column_sz = doris_struct.get_column(not_missing_column_id).size() - |
963 | 10 | not_missing_orig_column_size; |
964 | 10 | } |
965 | | |
966 | | // Fill SkipReadingReader columns with the correct amount of data based on reference column |
967 | | // Let SkipReadingReader handle the data filling through its read_column_data method |
968 | 10 | for (auto idx : skip_reading_column_idxs) { |
969 | 8 | auto& doris_field = doris_struct.get_column_ptr(idx); |
970 | 8 | auto& doris_type = const_cast<DataTypePtr&>(doris_struct_type->get_element(idx)); |
971 | 8 | auto& doris_name = const_cast<String&>(doris_struct_type->get_element_name(idx)); |
972 | 8 | auto file_name = root_node->children_file_column_name(doris_name); |
973 | | |
974 | 8 | size_t field_rows = 0; |
975 | 8 | bool field_eof = false; |
976 | 8 | RETURN_IF_ERROR(_child_readers[file_name]->read_column_data( |
977 | 8 | doris_field, doris_type, root_node->get_children_node(doris_name), filter_map, |
978 | 8 | missing_column_sz, &field_rows, &field_eof, is_dict_filter, missing_column_sz)); |
979 | 8 | } |
980 | | |
981 | | // Fill truly missing columns (not in root_node) with null or default value |
982 | 10 | for (auto idx : missing_column_idxs) { |
983 | 0 | auto& doris_field = doris_struct.get_column_ptr(idx); |
984 | 0 | auto& doris_type = doris_struct_type->get_element(idx); |
985 | 0 | DCHECK(doris_type->is_nullable()); |
986 | 0 | auto mutable_column = doris_field->assume_mutable(); |
987 | 0 | auto* nullable_column = static_cast<ColumnNullable*>(mutable_column.get()); |
988 | 0 | nullable_column->insert_many_defaults(missing_column_sz); |
989 | 0 | } |
990 | | |
991 | 10 | if (null_map_ptr != nullptr) { |
992 | 10 | fill_struct_null_map(_field_schema, *null_map_ptr, this->get_rep_level(), |
993 | 10 | this->get_def_level()); |
994 | 10 | } |
995 | 10 | #ifndef NDEBUG |
996 | 10 | doris_column->sanity_check(); |
997 | 10 | #endif |
998 | 10 | return Status::OK(); |
999 | 10 | } |
1000 | | |
1001 | | template class ScalarColumnReader<true, true>; |
1002 | | template class ScalarColumnReader<true, false>; |
1003 | | template class ScalarColumnReader<false, true>; |
1004 | | template class ScalarColumnReader<false, false>; |
1005 | | |
1006 | | #include "common/compile_check_end.h" |
1007 | | |
1008 | | }; // namespace doris |