be/src/cloud/cloud_tablet.h
Line | Count | Source |
1 | | // Licensed to the Apache Software Foundation (ASF) under one |
2 | | // or more contributor license agreements. See the NOTICE file |
3 | | // distributed with this work for additional information |
4 | | // regarding copyright ownership. The ASF licenses this file |
5 | | // to you under the Apache License, Version 2.0 (the |
6 | | // "License"); you may not use this file except in compliance |
7 | | // with the License. You may obtain a copy of the License at |
8 | | // |
9 | | // http://www.apache.org/licenses/LICENSE-2.0 |
10 | | // |
11 | | // Unless required by applicable law or agreed to in writing, |
12 | | // software distributed under the License is distributed on an |
13 | | // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
14 | | // KIND, either express or implied. See the License for the |
15 | | // specific language governing permissions and limitations |
16 | | // under the License. |
17 | | |
18 | | #pragma once |
19 | | |
20 | | #include <memory> |
21 | | |
22 | | #include "storage/partial_update_info.h" |
23 | | #include "storage/rowset/rowset.h" |
24 | | #include "storage/tablet/base_tablet.h" |
25 | | |
26 | | namespace doris { |
27 | | |
28 | | class CloudStorageEngine; |
29 | | |
30 | | enum class WarmUpTriggerSource : int { NONE, SYNC_ROWSET, EVENT_DRIVEN, JOB }; |
31 | | |
32 | | enum class WarmUpProgress : int { NONE, DOING, DONE }; |
33 | | |
34 | | struct WarmUpState { |
35 | | WarmUpTriggerSource trigger_source {WarmUpTriggerSource::NONE}; |
36 | | WarmUpProgress progress {WarmUpProgress::NONE}; |
37 | | |
38 | 56 | bool operator==(const WarmUpState& other) const { |
39 | 56 | return trigger_source == other.trigger_source && progress == other.progress; |
40 | 56 | } |
41 | | }; |
42 | | |
43 | | struct SyncRowsetStats { |
44 | | int64_t get_remote_rowsets_num {0}; |
45 | | int64_t get_remote_rowsets_rpc_ns {0}; |
46 | | |
47 | | int64_t get_local_delete_bitmap_rowsets_num {0}; |
48 | | int64_t get_remote_delete_bitmap_rowsets_num {0}; |
49 | | int64_t get_remote_delete_bitmap_key_count {0}; |
50 | | int64_t get_remote_delete_bitmap_bytes {0}; |
51 | | int64_t get_remote_delete_bitmap_rpc_ns {0}; |
52 | | |
53 | | int64_t get_remote_tablet_meta_rpc_ns {0}; |
54 | | int64_t tablet_meta_cache_hit {0}; |
55 | | int64_t tablet_meta_cache_miss {0}; |
56 | | |
57 | | int64_t bthread_schedule_delay_ns {0}; |
58 | | int64_t meta_lock_wait_ns {0}; // _meta_lock (std::shared_mutex) wait across all acquisitions |
59 | | int64_t sync_meta_lock_wait_ns { |
60 | | 0}; // _sync_meta_lock (bthread::Mutex) wait across all acquisitions |
61 | | }; |
62 | | |
63 | | struct SyncOptions { |
64 | | bool warmup_delta_data = false; |
65 | | bool sync_delete_bitmap = true; |
66 | | bool full_sync = false; |
67 | | bool merge_schema = false; |
68 | | int64_t query_version = -1; |
69 | | }; |
70 | | |
71 | | struct RecycledRowsets { |
72 | | RowsetId rowset_id; |
73 | | int64_t num_segments; |
74 | | std::vector<std::string> index_file_names; |
75 | | }; |
76 | | |
77 | | class CloudTablet final : public BaseTablet { |
78 | | public: |
79 | | CloudTablet(CloudStorageEngine& engine, TabletMetaSharedPtr tablet_meta); |
80 | | |
81 | | ~CloudTablet() override; |
82 | | |
83 | | bool exceed_version_limit(int32_t limit) override; |
84 | | |
85 | | Result<std::unique_ptr<RowsetWriter>> create_rowset_writer(RowsetWriterContext& context, |
86 | | bool vertical) override; |
87 | | |
88 | | Status capture_rs_readers(const Version& spec_version, std::vector<RowSetSplits>* rs_splits, |
89 | | const CaptureRowsetOps& opts) override; |
90 | | |
91 | | [[nodiscard]] Result<std::vector<Version>> capture_consistent_versions_unlocked( |
92 | | const Version& version_range, const CaptureRowsetOps& options) const override; |
93 | | |
94 | | // Capture versions with cache preference optimization. |
95 | | // This method prioritizes using cached/warmed-up rowsets when building version paths, |
96 | | // avoiding cold data reads when possible. It uses capture_consistent_versions_prefer_cache |
97 | | // to find a consistent version path that prefers already warmed-up rowsets. |
98 | | Result<std::vector<Version>> capture_versions_prefer_cache(const Version& spec_version) const; |
99 | | |
100 | | // Capture versions with query freshness tolerance. |
101 | | // This method finds a consistent version path where all rowsets are warmed up, |
102 | | // but allows fallback to normal capture if there are newer rowsets that should be |
103 | | // visible (based on freshness tolerance) but haven't been warmed up yet. |
104 | | // For merge-on-write tables, uses special validation to ensure data correctness. |
105 | | // |
106 | | // IMPORTANT: The returned version may be smaller than the requested version if newer |
107 | | // data hasn't been warmed up yet. This can cause different tablets in the same query |
108 | | // to read from different versions, potentially leading to inconsistent query results. |
109 | | // |
110 | | // @param options.query_freshness_tolerance_ms: Time tolerance in milliseconds. Rowsets that |
111 | | // became visible within this time range (after current_time - query_freshness_tolerance_ms) |
112 | | // can be skipped if not warmed up. However, if older rowsets (before this time point) |
113 | | // are not warmed up, the method will fallback to normal capture. |
114 | | Result<std::vector<Version>> capture_versions_with_freshness_tolerance( |
115 | | const Version& spec_version, const CaptureRowsetOps& options) const; |
116 | | |
117 | 0 | size_t tablet_footprint() override { |
118 | 0 | return _approximate_data_size.load(std::memory_order_relaxed); |
119 | 0 | } |
120 | | |
121 | | std::string tablet_path() const override; |
122 | | |
123 | | // clang-format off |
124 | 14 | int64_t fetch_add_approximate_num_rowsets (int64_t x) { return _approximate_num_rowsets .fetch_add(x, std::memory_order_relaxed); } |
125 | 0 | int64_t fetch_add_approximate_num_segments(int64_t x) { return _approximate_num_segments.fetch_add(x, std::memory_order_relaxed); } |
126 | 0 | int64_t fetch_add_approximate_num_rows (int64_t x) { return _approximate_num_rows .fetch_add(x, std::memory_order_relaxed); } |
127 | 0 | int64_t fetch_add_approximate_data_size (int64_t x) { return _approximate_data_size .fetch_add(x, std::memory_order_relaxed); } |
128 | 2 | int64_t fetch_add_approximate_cumu_num_rowsets (int64_t x) { return _approximate_cumu_num_rowsets.fetch_add(x, std::memory_order_relaxed); } |
129 | 0 | int64_t fetch_add_approximate_cumu_num_deltas (int64_t x) { return _approximate_cumu_num_deltas.fetch_add(x, std::memory_order_relaxed); } |
130 | | // clang-format on |
131 | | |
132 | | // meta lock must be held when calling this function |
133 | | void reset_approximate_stats(int64_t num_rowsets, int64_t num_segments, int64_t num_rows, |
134 | | int64_t data_size); |
135 | | |
136 | | // return a json string to show the compaction status of this tablet |
137 | | void get_compaction_status(std::string* json_result); |
138 | | |
139 | | // Synchronize the rowsets from meta service. |
140 | | // If tablet state is not `TABLET_RUNNING`, sync tablet meta and all visible rowsets. |
141 | | // If `query_version` > 0 and local max_version of the tablet >= `query_version`, do nothing. |
142 | | // If 'need_download_data_async' is true, it means that we need to download the new version |
143 | | // rowsets datum async. |
144 | | Status sync_rowsets(const SyncOptions& options = {}, SyncRowsetStats* stats = nullptr); |
145 | | |
146 | | // Synchronize the tablet meta from meta service. |
147 | | Status sync_meta(); |
148 | | |
149 | | // If `version_overlap` is true, function will delete rowsets with overlapped version in this tablet. |
150 | | // If 'warmup_delta_data' is true, download the new version rowset data in background. |
151 | | // MUST hold EXCLUSIVE `_meta_lock`. |
152 | | // If 'need_download_data_async' is true, it means that we need to download the new version |
153 | | // rowsets datum async. |
154 | | void add_rowsets(std::vector<RowsetSharedPtr> to_add, bool version_overlap, |
155 | | std::unique_lock<std::shared_mutex>& meta_lock, |
156 | | bool warmup_delta_data = false); |
157 | | |
158 | | // MUST hold EXCLUSIVE `_meta_lock`. |
159 | | void delete_rowsets(const std::vector<RowsetSharedPtr>& to_delete, |
160 | | std::unique_lock<std::shared_mutex>& meta_lock); |
161 | | |
162 | | // When the tablet is dropped, we need to recycle cached data: |
163 | | // 1. The data in file cache |
164 | | // 2. The memory in tablet cache |
165 | | void clear_cache() override; |
166 | | |
167 | | // Return number of deleted stale rowsets |
168 | | uint64_t delete_expired_stale_rowsets(); |
169 | | |
170 | 0 | bool has_stale_rowsets() const { return !_stale_rs_version_map.empty(); } |
171 | | |
172 | | int64_t get_cloud_base_compaction_score() const; |
173 | | int64_t get_cloud_cumu_compaction_score() const; |
174 | | |
175 | 29 | int64_t max_version_unlocked() const override { return _max_version; } |
176 | 3 | int64_t base_compaction_cnt() const { return _base_compaction_cnt; } |
177 | 3 | int64_t cumulative_compaction_cnt() const { return _cumulative_compaction_cnt; } |
178 | 0 | int64_t full_compaction_cnt() const { return _full_compaction_cnt; } |
179 | 6 | int64_t cumulative_layer_point() const { |
180 | 6 | return _cumulative_point.load(std::memory_order_relaxed); |
181 | 6 | } |
182 | | |
183 | 0 | void set_base_compaction_cnt(int64_t cnt) { _base_compaction_cnt = cnt; } |
184 | 0 | void set_cumulative_compaction_cnt(int64_t cnt) { _cumulative_compaction_cnt = cnt; } |
185 | 0 | void set_full_compaction_cnt(int64_t cnt) { _full_compaction_cnt = cnt; } |
186 | | void set_cumulative_layer_point(int64_t new_point); |
187 | | |
188 | 1 | int64_t last_cumu_compaction_failure_time() { return _last_cumu_compaction_failure_millis; } |
189 | 3 | void set_last_cumu_compaction_failure_time(int64_t millis) { |
190 | 3 | _last_cumu_compaction_failure_millis = millis; |
191 | 3 | } |
192 | | |
193 | 2 | int64_t last_base_compaction_failure_time() { return _last_base_compaction_failure_millis; } |
194 | 3 | void set_last_base_compaction_failure_time(int64_t millis) { |
195 | 3 | _last_base_compaction_failure_millis = millis; |
196 | 3 | } |
197 | | |
198 | 0 | int64_t last_full_compaction_failure_time() { return _last_full_compaction_failure_millis; } |
199 | 0 | void set_last_full_compaction_failure_time(int64_t millis) { |
200 | 0 | _last_full_compaction_failure_millis = millis; |
201 | 0 | } |
202 | | |
203 | 0 | int64_t last_cumu_compaction_success_time() { return _last_cumu_compaction_success_millis; } |
204 | 0 | void set_last_cumu_compaction_success_time(int64_t millis) { |
205 | 0 | _last_cumu_compaction_success_millis = millis; |
206 | 0 | } |
207 | | |
208 | 0 | int64_t last_base_compaction_success_time() { return _last_base_compaction_success_millis; } |
209 | 0 | void set_last_base_compaction_success_time(int64_t millis) { |
210 | 0 | _last_base_compaction_success_millis = millis; |
211 | 0 | } |
212 | | |
213 | 0 | int64_t last_full_compaction_success_time() { return _last_full_compaction_success_millis; } |
214 | 0 | void set_last_full_compaction_success_time(int64_t millis) { |
215 | 0 | _last_full_compaction_success_millis = millis; |
216 | 0 | } |
217 | | |
218 | 0 | int64_t last_cumu_compaction_schedule_time() { return _last_cumu_compaction_schedule_millis; } |
219 | 0 | void set_last_cumu_compaction_schedule_time(int64_t millis) { |
220 | 0 | _last_cumu_compaction_schedule_millis = millis; |
221 | 0 | } |
222 | | |
223 | 0 | int64_t last_base_compaction_schedule_time() { return _last_base_compaction_schedule_millis; } |
224 | 0 | void set_last_base_compaction_schedule_time(int64_t millis) { |
225 | 0 | _last_base_compaction_schedule_millis = millis; |
226 | 0 | } |
227 | | |
228 | 0 | int64_t last_full_compaction_schedule_time() { return _last_full_compaction_schedule_millis; } |
229 | 0 | void set_last_full_compaction_schedule_time(int64_t millis) { |
230 | 0 | _last_full_compaction_schedule_millis = millis; |
231 | 0 | } |
232 | | |
233 | 0 | void set_last_cumu_compaction_status(std::string status) { |
234 | 0 | _last_cumu_compaction_status = std::move(status); |
235 | 0 | } |
236 | | |
237 | 0 | std::string get_last_cumu_compaction_status() { return _last_cumu_compaction_status; } |
238 | | |
239 | 0 | void set_last_base_compaction_status(std::string status) { |
240 | 0 | _last_base_compaction_status = std::move(status); |
241 | 0 | } |
242 | | |
243 | 0 | std::string get_last_base_compaction_status() { return _last_base_compaction_status; } |
244 | | |
245 | 0 | void set_last_full_compaction_status(std::string status) { |
246 | 0 | _last_full_compaction_status = std::move(status); |
247 | 0 | } |
248 | | |
249 | 0 | std::string get_last_full_compaction_status() { return _last_full_compaction_status; } |
250 | | |
251 | 0 | int64_t alter_version() const { return _alter_version; } |
252 | 2 | void set_alter_version(int64_t alter_version) { _alter_version = alter_version; } |
253 | | |
254 | | // Last active cluster info for compaction read-write separation |
255 | 16 | std::string last_active_cluster_id() const { |
256 | 16 | std::shared_lock lock(_cluster_info_mutex); |
257 | 16 | return _last_active_cluster_id; |
258 | 16 | } |
259 | 3 | int64_t last_active_time_ms() const { |
260 | 3 | std::shared_lock lock(_cluster_info_mutex); |
261 | 3 | return _last_active_time_ms; |
262 | 3 | } |
263 | 11 | void set_last_active_cluster_info(const std::string& cluster_id, int64_t time_ms) { |
264 | 11 | std::unique_lock lock(_cluster_info_mutex); |
265 | 11 | _last_active_cluster_id = cluster_id; |
266 | 11 | _last_active_time_ms = time_ms; |
267 | 11 | } |
268 | | |
269 | | std::vector<RowsetSharedPtr> pick_candidate_rowsets_to_base_compaction(); |
270 | | |
271 | 24 | inline Version max_version() const { |
272 | 24 | std::shared_lock rdlock(_meta_lock); |
273 | 24 | return _tablet_meta->max_version(); |
274 | 24 | } |
275 | | |
276 | 9 | int64_t base_size() const { return _base_size; } |
277 | | |
278 | | std::vector<RowsetSharedPtr> pick_candidate_rowsets_to_full_compaction(); |
279 | | Result<RowsetSharedPtr> pick_a_rowset_for_index_change(int schema_version, |
280 | | bool& is_base_rowset); |
281 | | Status check_rowset_schema_for_build_index(std::vector<TColumn>& columns, int schema_version); |
282 | | |
283 | 0 | std::mutex& get_base_compaction_lock() { return _base_compaction_lock; } |
284 | 0 | std::mutex& get_cumulative_compaction_lock() { return _cumulative_compaction_lock; } |
285 | | |
286 | | Result<std::unique_ptr<RowsetWriter>> create_transient_rowset_writer( |
287 | | const Rowset& rowset, std::shared_ptr<PartialUpdateInfo> partial_update_info, |
288 | | int64_t txn_expiration = 0) override; |
289 | | |
290 | | CalcDeleteBitmapExecutor* calc_delete_bitmap_executor() override; |
291 | | |
292 | | Status save_delete_bitmap(const TabletTxnInfo* txn_info, int64_t txn_id, |
293 | | DeleteBitmapPtr delete_bitmap, RowsetWriter* rowset_writer, |
294 | | const RowsetIdUnorderedSet& cur_rowset_ids, int64_t lock_id = -1, |
295 | | int64_t next_visible_version = -1) override; |
296 | | |
297 | | Status save_delete_bitmap_to_ms(int64_t cur_version, int64_t txn_id, |
298 | | DeleteBitmapPtr delete_bitmap, int64_t lock_id, |
299 | | int64_t next_visible_version, RowsetSharedPtr rowset); |
300 | | |
301 | | Status calc_delete_bitmap_for_compaction(const std::vector<RowsetSharedPtr>& input_rowsets, |
302 | | const RowsetSharedPtr& output_rowset, |
303 | | const RowIdConversion& rowid_conversion, |
304 | | ReaderType compaction_type, int64_t merged_rows, |
305 | | int64_t filtered_rows, int64_t initiator, |
306 | | DeleteBitmapPtr& output_rowset_delete_bitmap, |
307 | | bool allow_delete_in_cumu_compaction, |
308 | | int64_t& get_delete_bitmap_lock_start_time); |
309 | | |
310 | | // Find the missed versions until the spec_version. |
311 | | // |
312 | | // for example: |
313 | | // [0-4][5-5][8-8][9-9][14-14] |
314 | | // if spec_version = 12, it will return [6-7],[10-12] |
315 | | Versions calc_missed_versions(int64_t spec_version, Versions existing_versions) const override; |
316 | | |
317 | 0 | std::mutex& get_rowset_update_lock() { return _rowset_update_lock; } |
318 | | |
319 | 1 | bthread::Mutex& get_sync_meta_lock() { return _sync_meta_lock; } |
320 | | |
321 | 119 | const auto& rowset_map() const { return _rs_version_map; } |
322 | | |
323 | | int64_t last_sync_time_s = 0; |
324 | | int64_t last_load_time_ms = 0; |
325 | | int64_t last_base_compaction_success_time_ms = 0; |
326 | | int64_t last_cumu_compaction_success_time_ms = 0; |
327 | | int64_t last_cumu_no_suitable_version_ms = 0; |
328 | | int64_t last_access_time_ms = 0; |
329 | | |
330 | | std::atomic<int64_t> local_read_time_us = 0; |
331 | | std::atomic<int64_t> remote_read_time_us = 0; |
332 | | std::atomic<int64_t> exec_compaction_time_us = 0; |
333 | | |
334 | | void build_tablet_report_info(TTabletInfo* tablet_info); |
335 | | |
336 | | // check that if the delete bitmap in delete bitmap cache has the same cardinality with the expected_delete_bitmap's |
337 | | Status check_delete_bitmap_cache(int64_t txn_id, DeleteBitmap* expected_delete_bitmap) override; |
338 | | |
339 | | void agg_delete_bitmap_for_compaction(int64_t start_version, int64_t end_version, |
340 | | const std::vector<RowsetSharedPtr>& pre_rowsets, |
341 | | DeleteBitmapPtr& new_delete_bitmap, |
342 | | std::map<std::string, int64_t>& pre_rowset_to_versions); |
343 | | |
344 | | bool need_remove_unused_rowsets(); |
345 | | |
346 | | void add_unused_rowsets(const std::vector<RowsetSharedPtr>& rowsets); |
347 | | void remove_unused_rowsets(); |
348 | | |
349 | | // For each given rowset not in active use, clears its file cache and returns its |
350 | | // ID, segment count, and index file names as RecycledRowsets entries. |
351 | | static std::vector<RecycledRowsets> recycle_cached_data( |
352 | | const std::vector<RowsetSharedPtr>& rowsets); |
353 | | |
354 | | // Add warmup state management |
355 | | WarmUpState get_rowset_warmup_state(RowsetId rowset_id); |
356 | | bool add_rowset_warmup_state( |
357 | | const RowsetMeta& rowset, WarmUpTriggerSource source, |
358 | | std::chrono::steady_clock::time_point start_tp = std::chrono::steady_clock::now()); |
359 | | bool update_rowset_warmup_state_inverted_idx_num(WarmUpTriggerSource source, RowsetId rowset_id, |
360 | | int64_t delta); |
361 | | bool update_rowset_warmup_state_inverted_idx_num_unlocked(WarmUpTriggerSource source, |
362 | | RowsetId rowset_id, int64_t delta); |
363 | | WarmUpState complete_rowset_segment_warmup(WarmUpTriggerSource trigger_source, |
364 | | RowsetId rowset_id, Status status, |
365 | | int64_t segment_num, int64_t inverted_idx_num); |
366 | | |
367 | | bool is_rowset_warmed_up(const RowsetId& rowset_id) const; |
368 | | |
369 | | void add_warmed_up_rowset(const RowsetId& rowset_id); |
370 | | // Test helper: add a rowset to the warmup state map with DOING progress, |
371 | | // so that is_rowset_warmed_up() returns false for it. |
372 | | void add_not_warmed_up_rowset(const RowsetId& rowset_id); |
373 | | |
374 | | // Try to apply visible pending rowsets to tablet meta in version order |
375 | | // This should be called after receiving FE notification or when new rowsets are added |
376 | | // @return Status::OK() if successfully applied, error otherwise |
377 | | void apply_visible_pending_rowsets(); |
378 | | |
379 | | void try_make_committed_rs_visible(int64_t txn_id, int64_t visible_version, |
380 | | int64_t version_update_time_ms); |
381 | | void try_make_committed_rs_visible_for_mow(int64_t txn_id, int64_t visible_version, |
382 | | int64_t version_update_time_ms); |
383 | | |
384 | | void clear_unused_visible_pending_rowsets(); |
385 | | |
386 | 5 | std::string rowset_warmup_digest() const { |
387 | 5 | std::string res; |
388 | 104 | auto add_log = [&](const RowsetSharedPtr& rs) { |
389 | 104 | auto tmp = fmt::format("{}{}", rs->rowset_id().to_string(), rs->version().to_string()); |
390 | 104 | if (_rowset_warm_up_states.contains(rs->rowset_id())) { |
391 | 104 | tmp += fmt::format( |
392 | 104 | ", progress={}, segments_warmed_up={}/{}, inverted_idx_warmed_up={}/{}", |
393 | 104 | _rowset_warm_up_states.at(rs->rowset_id()).state.progress, |
394 | 104 | _rowset_warm_up_states.at(rs->rowset_id()).num_segments_warmed_up, |
395 | 104 | _rowset_warm_up_states.at(rs->rowset_id()).num_segments, |
396 | 104 | _rowset_warm_up_states.at(rs->rowset_id()).num_inverted_idx_warmed_up, |
397 | 104 | _rowset_warm_up_states.at(rs->rowset_id()).num_inverted_idx); |
398 | 104 | } |
399 | 104 | res += fmt::format("[{}],", tmp); |
400 | 104 | }; |
401 | 5 | traverse_rowsets_unlocked(add_log, true); |
402 | 5 | return res; |
403 | 5 | } |
404 | | |
405 | | private: |
406 | | // FIXME(plat1ko): No need to record base size if rowsets are ordered by version |
407 | | void update_base_size(const Rowset& rs); |
408 | | |
409 | | Status sync_if_not_running(SyncRowsetStats* stats = nullptr); |
410 | | |
411 | | bool add_rowset_warmup_state_unlocked( |
412 | | const RowsetMeta& rowset, WarmUpTriggerSource source, |
413 | | std::chrono::steady_clock::time_point start_tp = std::chrono::steady_clock::now()); |
414 | | |
415 | | // used by capture_rs_reader_xxx functions |
416 | | bool rowset_is_warmed_up_unlocked(int64_t start_version, int64_t end_version) const; |
417 | | |
418 | | // Check if a rowset should be visible but not warmed up within freshness tolerance |
419 | | bool _check_rowset_should_be_visible_but_not_warmed_up( |
420 | | const RowsetMetaSharedPtr& rs_meta, int64_t path_max_version, |
421 | | std::chrono::system_clock::time_point freshness_limit_tp) const; |
422 | | |
423 | | // Submit a segment download task for warming up |
424 | | void _submit_segment_download_task(const RowsetSharedPtr& rs, |
425 | | const StorageResource* storage_resource, int seg_id, |
426 | | int64_t expiration_time); |
427 | | |
428 | | // Submit an inverted index download task for warming up |
429 | | void _submit_inverted_index_download_task(const RowsetSharedPtr& rs, |
430 | | const StorageResource* storage_resource, |
431 | | const io::Path& idx_path, int64_t idx_size, |
432 | | int64_t expiration_time); |
433 | | |
434 | | // Add rowsets directly with warmup |
435 | | void _add_rowsets_directly(std::vector<RowsetSharedPtr>& rowsets, bool warmup_delta_data); |
436 | | |
437 | | CloudStorageEngine& _engine; |
438 | | |
439 | | // this mutex MUST ONLY be used when sync meta |
440 | | bthread::Mutex _sync_meta_lock; |
441 | | // ATTENTION: lock order should be: _sync_meta_lock -> _meta_lock |
442 | | |
443 | | std::atomic<int64_t> _cumulative_point {-1}; |
444 | | std::atomic<int64_t> _approximate_num_rowsets {-1}; |
445 | | std::atomic<int64_t> _approximate_num_segments {-1}; |
446 | | std::atomic<int64_t> _approximate_num_rows {-1}; |
447 | | std::atomic<int64_t> _approximate_data_size {-1}; |
448 | | std::atomic<int64_t> _approximate_cumu_num_rowsets {-1}; |
449 | | // Number of sorted arrays (e.g. for rowset with N segments, if rowset is overlapping, delta is N, otherwise 1) after cumu point |
450 | | std::atomic<int64_t> _approximate_cumu_num_deltas {-1}; |
451 | | |
452 | | // timestamp of last cumu compaction failure |
453 | | std::atomic<int64_t> _last_cumu_compaction_failure_millis; |
454 | | // timestamp of last base compaction failure |
455 | | std::atomic<int64_t> _last_base_compaction_failure_millis; |
456 | | // timestamp of last full compaction failure |
457 | | std::atomic<int64_t> _last_full_compaction_failure_millis; |
458 | | // timestamp of last cumu compaction success |
459 | | std::atomic<int64_t> _last_cumu_compaction_success_millis; |
460 | | // timestamp of last base compaction success |
461 | | std::atomic<int64_t> _last_base_compaction_success_millis; |
462 | | // timestamp of last full compaction success |
463 | | std::atomic<int64_t> _last_full_compaction_success_millis; |
464 | | // timestamp of last cumu compaction schedule time |
465 | | std::atomic<int64_t> _last_cumu_compaction_schedule_millis; |
466 | | // timestamp of last base compaction schedule time |
467 | | std::atomic<int64_t> _last_base_compaction_schedule_millis; |
468 | | // timestamp of last full compaction schedule time |
469 | | std::atomic<int64_t> _last_full_compaction_schedule_millis; |
470 | | |
471 | | std::string _last_cumu_compaction_status; |
472 | | std::string _last_base_compaction_status; |
473 | | std::string _last_full_compaction_status; |
474 | | |
475 | | int64_t _base_compaction_cnt = 0; |
476 | | int64_t _cumulative_compaction_cnt = 0; |
477 | | int64_t _full_compaction_cnt = 0; |
478 | | int64_t _max_version = -1; |
479 | | int64_t _base_size = 0; |
480 | | int64_t _alter_version = -1; |
481 | | |
482 | | std::mutex _base_compaction_lock; |
483 | | std::mutex _cumulative_compaction_lock; |
484 | | |
485 | | // To avoid multiple calc delete bitmap tasks on same (txn_id, tablet_id) with different |
486 | | // signatures being executed concurrently, we use _rowset_update_lock to serialize them |
487 | | mutable std::mutex _rowset_update_lock; |
488 | | |
489 | | // unused_rowsets, [start_version, end_version] |
490 | | std::mutex _gc_mutex; |
491 | | std::unordered_map<RowsetId, RowsetSharedPtr> _unused_rowsets; |
492 | | std::vector<std::pair<std::vector<RowsetId>, DeleteBitmapKeyRanges>> _unused_delete_bitmap; |
493 | | |
494 | | // for warm up states management |
495 | | struct RowsetWarmUpInfo { |
496 | | WarmUpState state; |
497 | | int64_t num_segments = 0; |
498 | | int64_t num_inverted_idx = 0; |
499 | | int64_t num_segments_warmed_up = 0; |
500 | | int64_t num_inverted_idx_warmed_up = 0; |
501 | | std::chrono::steady_clock::time_point start_tp; |
502 | | |
503 | 18 | void done(int64_t input_num_segments, int64_t input_num_inverted_idx) { |
504 | 18 | num_segments_warmed_up += input_num_segments; |
505 | 18 | num_inverted_idx_warmed_up += input_num_inverted_idx; |
506 | 18 | update_state(); |
507 | 18 | } |
508 | | |
509 | 52 | bool has_finished() const { |
510 | 52 | return (num_segments_warmed_up >= num_segments) && |
511 | 52 | (num_inverted_idx_warmed_up >= num_inverted_idx); |
512 | 52 | } |
513 | | |
514 | | void update_state(); |
515 | | }; |
516 | | std::unordered_map<RowsetId, RowsetWarmUpInfo> _rowset_warm_up_states; |
517 | | |
518 | | mutable std::shared_mutex _warmed_up_rowsets_mutex; |
519 | | std::unordered_set<RowsetId> _warmed_up_rowsets; |
520 | | |
521 | | // Cluster info for compaction read-write separation |
522 | | mutable std::shared_mutex _cluster_info_mutex; |
523 | | std::string _last_active_cluster_id; |
524 | | int64_t _last_active_time_ms {0}; |
525 | | |
526 | | // Map: version -> <rowset_meta, expiration_time> |
527 | | // Stores rowsets that have been notified by FE but not yet added to tablet meta |
528 | | // due to out-of-order notification or version discontinuity |
529 | | struct VisiblePendingRowset { |
530 | | const bool is_empty_rowset; |
531 | | const int64_t expiration_time; // seconds since epoch |
532 | | RowsetMetaSharedPtr rowset_meta; |
533 | | |
534 | | VisiblePendingRowset(RowsetMetaSharedPtr rowset_meta_, int64_t expiration_time_, |
535 | | bool is_empty_rowset_ = false) |
536 | 22 | : is_empty_rowset(is_empty_rowset_), |
537 | 22 | expiration_time(expiration_time_), |
538 | 22 | rowset_meta(std::move(rowset_meta_)) {} |
539 | | }; |
540 | | mutable std::mutex _visible_pending_rs_lock; |
541 | | std::map<int64_t, VisiblePendingRowset> _visible_pending_rs_map; |
542 | | }; |
543 | | |
544 | | using CloudTabletSPtr = std::shared_ptr<CloudTablet>; |
545 | | |
546 | | } // namespace doris |