1 | #include "duckdb/execution/executor.hpp" |
2 | |
3 | #include "duckdb/execution/execution_context.hpp" |
4 | #include "duckdb/execution/operator/helper/physical_result_collector.hpp" |
5 | #include "duckdb/execution/operator/set/physical_recursive_cte.hpp" |
6 | #include "duckdb/execution/physical_operator.hpp" |
7 | #include "duckdb/main/client_context.hpp" |
8 | #include "duckdb/main/client_data.hpp" |
9 | #include "duckdb/parallel/meta_pipeline.hpp" |
10 | #include "duckdb/parallel/pipeline_complete_event.hpp" |
11 | #include "duckdb/parallel/pipeline_event.hpp" |
12 | #include "duckdb/parallel/pipeline_executor.hpp" |
13 | #include "duckdb/parallel/pipeline_finish_event.hpp" |
14 | #include "duckdb/parallel/pipeline_initialize_event.hpp" |
15 | #include "duckdb/parallel/task_scheduler.hpp" |
16 | #include "duckdb/parallel/thread_context.hpp" |
17 | |
18 | #include <algorithm> |
19 | |
20 | namespace duckdb { |
21 | |
22 | Executor::Executor(ClientContext &context) : context(context) { |
23 | } |
24 | |
25 | Executor::~Executor() { |
26 | } |
27 | |
28 | Executor &Executor::Get(ClientContext &context) { |
29 | return context.GetExecutor(); |
30 | } |
31 | |
32 | void Executor::AddEvent(shared_ptr<Event> event) { |
33 | lock_guard<mutex> elock(executor_lock); |
34 | if (cancelled) { |
35 | return; |
36 | } |
37 | events.push_back(x: std::move(event)); |
38 | } |
39 | |
40 | struct PipelineEventStack { |
41 | PipelineEventStack(Event &pipeline_initialize_event, Event &pipeline_event, Event &pipeline_finish_event, |
42 | Event &pipeline_complete_event) |
43 | : pipeline_initialize_event(pipeline_initialize_event), pipeline_event(pipeline_event), |
44 | pipeline_finish_event(pipeline_finish_event), pipeline_complete_event(pipeline_complete_event) { |
45 | } |
46 | |
47 | Event &pipeline_initialize_event; |
48 | Event &pipeline_event; |
49 | Event &pipeline_finish_event; |
50 | Event &pipeline_complete_event; |
51 | }; |
52 | |
53 | using event_map_t = reference_map_t<Pipeline, PipelineEventStack>; |
54 | |
55 | struct ScheduleEventData { |
56 | ScheduleEventData(const vector<shared_ptr<MetaPipeline>> &meta_pipelines, vector<shared_ptr<Event>> &events, |
57 | bool initial_schedule) |
58 | : meta_pipelines(meta_pipelines), events(events), initial_schedule(initial_schedule) { |
59 | } |
60 | |
61 | const vector<shared_ptr<MetaPipeline>> &meta_pipelines; |
62 | vector<shared_ptr<Event>> &events; |
63 | bool initial_schedule; |
64 | event_map_t event_map; |
65 | }; |
66 | |
67 | void Executor::SchedulePipeline(const shared_ptr<MetaPipeline> &meta_pipeline, ScheduleEventData &event_data) { |
68 | D_ASSERT(meta_pipeline); |
69 | auto &events = event_data.events; |
70 | auto &event_map = event_data.event_map; |
71 | |
72 | // create events/stack for the base pipeline |
73 | auto base_pipeline = meta_pipeline->GetBasePipeline(); |
74 | auto base_initialize_event = make_shared<PipelineInitializeEvent>(args&: base_pipeline); |
75 | auto base_event = make_shared<PipelineEvent>(args&: base_pipeline); |
76 | auto base_finish_event = make_shared<PipelineFinishEvent>(args&: base_pipeline); |
77 | auto base_complete_event = make_shared<PipelineCompleteEvent>(args&: base_pipeline->executor, args&: event_data.initial_schedule); |
78 | PipelineEventStack base_stack(*base_initialize_event, *base_event, *base_finish_event, *base_complete_event); |
79 | events.push_back(x: std::move(base_initialize_event)); |
80 | events.push_back(x: std::move(base_event)); |
81 | events.push_back(x: std::move(base_finish_event)); |
82 | events.push_back(x: std::move(base_complete_event)); |
83 | |
84 | // dependencies: initialize -> event -> finish -> complete |
85 | base_stack.pipeline_event.AddDependency(event&: base_stack.pipeline_initialize_event); |
86 | base_stack.pipeline_finish_event.AddDependency(event&: base_stack.pipeline_event); |
87 | base_stack.pipeline_complete_event.AddDependency(event&: base_stack.pipeline_finish_event); |
88 | |
89 | // create an event and stack for all pipelines in the MetaPipeline |
90 | vector<shared_ptr<Pipeline>> pipelines; |
91 | meta_pipeline->GetPipelines(result&: pipelines, recursive: false); |
92 | for (idx_t i = 1; i < pipelines.size(); i++) { // loop starts at 1 because 0 is the base pipeline |
93 | auto &pipeline = pipelines[i]; |
94 | D_ASSERT(pipeline); |
95 | |
96 | // create events/stack for this pipeline |
97 | auto pipeline_event = make_shared<PipelineEvent>(args&: pipeline); |
98 | |
99 | auto finish_group = meta_pipeline->GetFinishGroup(pipeline: pipeline.get()); |
100 | if (finish_group) { |
101 | // this pipeline is part of a finish group |
102 | const auto group_entry = event_map.find(x: *finish_group.get()); |
103 | D_ASSERT(group_entry != event_map.end()); |
104 | auto &group_stack = group_entry->second; |
105 | PipelineEventStack pipeline_stack(base_stack.pipeline_initialize_event, *pipeline_event, |
106 | group_stack.pipeline_finish_event, base_stack.pipeline_complete_event); |
107 | |
108 | // dependencies: base_finish -> pipeline_event -> group_finish |
109 | pipeline_stack.pipeline_event.AddDependency(event&: base_stack.pipeline_event); |
110 | group_stack.pipeline_finish_event.AddDependency(event&: pipeline_stack.pipeline_event); |
111 | |
112 | // add pipeline stack to event map |
113 | event_map.insert(x: make_pair(x: reference<Pipeline>(*pipeline), y&: pipeline_stack)); |
114 | } else if (meta_pipeline->HasFinishEvent(pipeline: pipeline.get())) { |
115 | // this pipeline has its own finish event (despite going into the same sink - Finalize twice!) |
116 | auto pipeline_finish_event = make_shared<PipelineFinishEvent>(args&: pipeline); |
117 | PipelineEventStack pipeline_stack(base_stack.pipeline_initialize_event, *pipeline_event, |
118 | *pipeline_finish_event, base_stack.pipeline_complete_event); |
119 | events.push_back(x: std::move(pipeline_finish_event)); |
120 | |
121 | // dependencies: base_finish -> pipeline_event -> pipeline_finish -> base_complete |
122 | pipeline_stack.pipeline_event.AddDependency(event&: base_stack.pipeline_finish_event); |
123 | pipeline_stack.pipeline_finish_event.AddDependency(event&: pipeline_stack.pipeline_event); |
124 | base_stack.pipeline_complete_event.AddDependency(event&: pipeline_stack.pipeline_finish_event); |
125 | |
126 | // add pipeline stack to event map |
127 | event_map.insert(x: make_pair(x: reference<Pipeline>(*pipeline), y&: pipeline_stack)); |
128 | |
129 | } else { |
130 | // no additional finish event |
131 | PipelineEventStack pipeline_stack(base_stack.pipeline_initialize_event, *pipeline_event, |
132 | base_stack.pipeline_finish_event, base_stack.pipeline_complete_event); |
133 | |
134 | // dependencies: base_initialize -> pipeline_event -> base_finish |
135 | pipeline_stack.pipeline_event.AddDependency(event&: base_stack.pipeline_initialize_event); |
136 | base_stack.pipeline_finish_event.AddDependency(event&: pipeline_stack.pipeline_event); |
137 | |
138 | // add pipeline stack to event map |
139 | event_map.insert(x: make_pair(x: reference<Pipeline>(*pipeline), y&: pipeline_stack)); |
140 | } |
141 | events.push_back(x: std::move(pipeline_event)); |
142 | } |
143 | |
144 | // add base stack to the event data too |
145 | event_map.insert(x: make_pair(x: reference<Pipeline>(*base_pipeline), y&: base_stack)); |
146 | |
147 | // set up the dependencies within this MetaPipeline |
148 | for (auto &pipeline : pipelines) { |
149 | auto source = pipeline->GetSource(); |
150 | if (source->type == PhysicalOperatorType::TABLE_SCAN) { |
151 | // we have to reset the source here (in the main thread), because some of our clients (looking at you, R) |
152 | // do not like it when threads other than the main thread call into R, for e.g., arrow scans |
153 | pipeline->ResetSource(force: true); |
154 | } |
155 | |
156 | auto dependencies = meta_pipeline->GetDependencies(dependant: pipeline.get()); |
157 | if (!dependencies) { |
158 | continue; |
159 | } |
160 | auto root_entry = event_map.find(x: *pipeline); |
161 | D_ASSERT(root_entry != event_map.end()); |
162 | auto &pipeline_stack = root_entry->second; |
163 | for (auto &dependency : *dependencies) { |
164 | auto event_entry = event_map.find(x: *dependency); |
165 | D_ASSERT(event_entry != event_map.end()); |
166 | auto &dependency_stack = event_entry->second; |
167 | pipeline_stack.pipeline_event.AddDependency(event&: dependency_stack.pipeline_event); |
168 | } |
169 | } |
170 | } |
171 | |
172 | void Executor::ScheduleEventsInternal(ScheduleEventData &event_data) { |
173 | auto &events = event_data.events; |
174 | D_ASSERT(events.empty()); |
175 | |
176 | // create all the required pipeline events |
177 | for (auto &pipeline : event_data.meta_pipelines) { |
178 | SchedulePipeline(meta_pipeline: pipeline, event_data); |
179 | } |
180 | |
181 | // set up the dependencies across MetaPipelines |
182 | auto &event_map = event_data.event_map; |
183 | for (auto &entry : event_map) { |
184 | auto &pipeline = entry.first.get(); |
185 | for (auto &dependency : pipeline.dependencies) { |
186 | auto dep = dependency.lock(); |
187 | D_ASSERT(dep); |
188 | auto event_map_entry = event_map.find(x: *dep); |
189 | D_ASSERT(event_map_entry != event_map.end()); |
190 | auto &dep_entry = event_map_entry->second; |
191 | entry.second.pipeline_event.AddDependency(event&: dep_entry.pipeline_complete_event); |
192 | } |
193 | } |
194 | |
195 | // verify that we have no cyclic dependencies |
196 | VerifyScheduledEvents(event_data); |
197 | |
198 | // schedule the pipelines that do not have dependencies |
199 | for (auto &event : events) { |
200 | if (!event->HasDependencies()) { |
201 | event->Schedule(); |
202 | } |
203 | } |
204 | } |
205 | |
206 | void Executor::ScheduleEvents(const vector<shared_ptr<MetaPipeline>> &meta_pipelines) { |
207 | ScheduleEventData event_data(meta_pipelines, events, true); |
208 | ScheduleEventsInternal(event_data); |
209 | } |
210 | |
211 | void Executor::VerifyScheduledEvents(const ScheduleEventData &event_data) { |
212 | #ifdef DEBUG |
213 | const idx_t count = event_data.events.size(); |
214 | vector<Event *> vertices; |
215 | vertices.reserve(count); |
216 | for (const auto &event : event_data.events) { |
217 | vertices.push_back(event.get()); |
218 | } |
219 | vector<bool> visited(count, false); |
220 | vector<bool> recursion_stack(count, false); |
221 | for (idx_t i = 0; i < count; i++) { |
222 | VerifyScheduledEventsInternal(i, vertices, visited, recursion_stack); |
223 | } |
224 | #endif |
225 | } |
226 | |
227 | void Executor::VerifyScheduledEventsInternal(const idx_t vertex, const vector<Event *> &vertices, vector<bool> &visited, |
228 | vector<bool> &recursion_stack) { |
229 | D_ASSERT(!recursion_stack[vertex]); // this vertex is in the recursion stack: circular dependency! |
230 | if (visited[vertex]) { |
231 | return; // early out: we already visited this vertex |
232 | } |
233 | |
234 | auto &parents = vertices[vertex]->GetParentsVerification(); |
235 | if (parents.empty()) { |
236 | return; // early out: outgoing edges |
237 | } |
238 | |
239 | // create a vector the indices of the adjacent events |
240 | vector<idx_t> adjacent; |
241 | const idx_t count = vertices.size(); |
242 | for (auto parent : parents) { |
243 | idx_t i; |
244 | for (i = 0; i < count; i++) { |
245 | if (vertices[i] == parent) { |
246 | adjacent.push_back(x: i); |
247 | break; |
248 | } |
249 | } |
250 | D_ASSERT(i != count); // dependency must be in there somewhere |
251 | } |
252 | |
253 | // mark vertex as visited and add to recursion stack |
254 | visited[vertex] = true; |
255 | recursion_stack[vertex] = true; |
256 | |
257 | // recurse into adjacent vertices |
258 | for (const auto &i : adjacent) { |
259 | VerifyScheduledEventsInternal(vertex: i, vertices, visited, recursion_stack); |
260 | } |
261 | |
262 | // remove vertex from recursion stack |
263 | recursion_stack[vertex] = false; |
264 | } |
265 | |
266 | void Executor::AddRecursiveCTE(PhysicalOperator &rec_cte) { |
267 | recursive_ctes.push_back(x: rec_cte); |
268 | } |
269 | |
270 | void Executor::ReschedulePipelines(const vector<shared_ptr<MetaPipeline>> &pipelines_p, |
271 | vector<shared_ptr<Event>> &events_p) { |
272 | ScheduleEventData event_data(pipelines_p, events_p, false); |
273 | ScheduleEventsInternal(event_data); |
274 | } |
275 | |
276 | bool Executor::NextExecutor() { |
277 | if (root_pipeline_idx >= root_pipelines.size()) { |
278 | return false; |
279 | } |
280 | root_pipelines[root_pipeline_idx]->Reset(); |
281 | root_executor = make_uniq<PipelineExecutor>(args&: context, args&: *root_pipelines[root_pipeline_idx]); |
282 | root_pipeline_idx++; |
283 | return true; |
284 | } |
285 | |
286 | void Executor::VerifyPipeline(Pipeline &pipeline) { |
287 | D_ASSERT(!pipeline.ToString().empty()); |
288 | auto operators = pipeline.GetOperators(); |
289 | for (auto &other_pipeline : pipelines) { |
290 | auto other_operators = other_pipeline->GetOperators(); |
291 | for (idx_t op_idx = 0; op_idx < operators.size(); op_idx++) { |
292 | for (idx_t other_idx = 0; other_idx < other_operators.size(); other_idx++) { |
293 | auto &left = operators[op_idx].get(); |
294 | auto &right = other_operators[other_idx].get(); |
295 | if (left.Equals(other: right)) { |
296 | D_ASSERT(right.Equals(left)); |
297 | } else { |
298 | D_ASSERT(!right.Equals(left)); |
299 | } |
300 | } |
301 | } |
302 | } |
303 | } |
304 | |
305 | void Executor::VerifyPipelines() { |
306 | #ifdef DEBUG |
307 | for (auto &pipeline : pipelines) { |
308 | VerifyPipeline(*pipeline); |
309 | } |
310 | #endif |
311 | } |
312 | |
313 | void Executor::Initialize(unique_ptr<PhysicalOperator> physical_plan) { |
314 | Reset(); |
315 | owned_plan = std::move(physical_plan); |
316 | InitializeInternal(physical_plan&: *owned_plan); |
317 | } |
318 | |
319 | void Executor::Initialize(PhysicalOperator &plan) { |
320 | Reset(); |
321 | InitializeInternal(physical_plan&: plan); |
322 | } |
323 | |
324 | void Executor::InitializeInternal(PhysicalOperator &plan) { |
325 | |
326 | auto &scheduler = TaskScheduler::GetScheduler(context); |
327 | { |
328 | lock_guard<mutex> elock(executor_lock); |
329 | physical_plan = &plan; |
330 | |
331 | this->profiler = ClientData::Get(context).profiler; |
332 | profiler->Initialize(root: plan); |
333 | this->producer = scheduler.CreateProducer(); |
334 | |
335 | // build and ready the pipelines |
336 | PipelineBuildState state; |
337 | auto root_pipeline = make_shared<MetaPipeline>(args&: *this, args&: state, args: nullptr); |
338 | root_pipeline->Build(op&: *physical_plan); |
339 | root_pipeline->Ready(); |
340 | |
341 | // ready recursive cte pipelines too |
342 | for (auto &rec_cte_ref : recursive_ctes) { |
343 | auto &rec_cte = rec_cte_ref.get().Cast<PhysicalRecursiveCTE>(); |
344 | rec_cte.recursive_meta_pipeline->Ready(); |
345 | } |
346 | |
347 | // set root pipelines, i.e., all pipelines that end in the final sink |
348 | root_pipeline->GetPipelines(result&: root_pipelines, recursive: false); |
349 | root_pipeline_idx = 0; |
350 | |
351 | // collect all meta-pipelines from the root pipeline |
352 | vector<shared_ptr<MetaPipeline>> to_schedule; |
353 | root_pipeline->GetMetaPipelines(result&: to_schedule, recursive: true, skip: true); |
354 | |
355 | // number of 'PipelineCompleteEvent's is equal to the number of meta pipelines, so we have to set it here |
356 | total_pipelines = to_schedule.size(); |
357 | |
358 | // collect all pipelines from the root pipelines (recursively) for the progress bar and verify them |
359 | root_pipeline->GetPipelines(result&: pipelines, recursive: true); |
360 | |
361 | // finally, verify and schedule |
362 | VerifyPipelines(); |
363 | ScheduleEvents(meta_pipelines: to_schedule); |
364 | } |
365 | } |
366 | |
367 | void Executor::CancelTasks() { |
368 | task.reset(); |
369 | // we do this by creating weak pointers to all pipelines |
370 | // then clearing our references to the pipelines |
371 | // and waiting until all pipelines have been destroyed |
372 | vector<weak_ptr<Pipeline>> weak_references; |
373 | { |
374 | lock_guard<mutex> elock(executor_lock); |
375 | weak_references.reserve(n: pipelines.size()); |
376 | cancelled = true; |
377 | for (auto &pipeline : pipelines) { |
378 | weak_references.push_back(x: weak_ptr<Pipeline>(pipeline)); |
379 | } |
380 | for (auto &rec_cte_ref : recursive_ctes) { |
381 | auto &rec_cte = rec_cte_ref.get().Cast<PhysicalRecursiveCTE>(); |
382 | rec_cte.recursive_meta_pipeline.reset(); |
383 | } |
384 | pipelines.clear(); |
385 | root_pipelines.clear(); |
386 | to_be_rescheduled_tasks.clear(); |
387 | events.clear(); |
388 | } |
389 | WorkOnTasks(); |
390 | for (auto &weak_ref : weak_references) { |
391 | while (true) { |
392 | auto weak = weak_ref.lock(); |
393 | if (!weak) { |
394 | break; |
395 | } |
396 | } |
397 | } |
398 | } |
399 | |
400 | void Executor::WorkOnTasks() { |
401 | auto &scheduler = TaskScheduler::GetScheduler(context); |
402 | |
403 | shared_ptr<Task> task; |
404 | while (scheduler.GetTaskFromProducer(token&: *producer, task)) { |
405 | auto res = task->Execute(mode: TaskExecutionMode::PROCESS_ALL); |
406 | if (res == TaskExecutionResult::TASK_BLOCKED) { |
407 | task->Deschedule(); |
408 | } |
409 | task.reset(); |
410 | } |
411 | } |
412 | |
413 | void Executor::RescheduleTask(shared_ptr<Task> &task) { |
414 | // This function will spin lock until the task provided is added to the to_be_rescheduled_tasks |
415 | while (true) { |
416 | lock_guard<mutex> l(executor_lock); |
417 | if (cancelled) { |
418 | return; |
419 | } |
420 | auto entry = to_be_rescheduled_tasks.find(x: task.get()); |
421 | if (entry != to_be_rescheduled_tasks.end()) { |
422 | auto &scheduler = TaskScheduler::GetScheduler(context); |
423 | to_be_rescheduled_tasks.erase(x: task.get()); |
424 | scheduler.ScheduleTask(producer&: GetToken(), task); |
425 | break; |
426 | } |
427 | } |
428 | } |
429 | |
430 | void Executor::AddToBeRescheduled(shared_ptr<Task> &task) { |
431 | lock_guard<mutex> l(executor_lock); |
432 | if (cancelled) { |
433 | return; |
434 | } |
435 | if (to_be_rescheduled_tasks.find(x: task.get()) != to_be_rescheduled_tasks.end()) { |
436 | return; |
437 | } |
438 | to_be_rescheduled_tasks[task.get()] = std::move(task); |
439 | } |
440 | |
441 | bool Executor::ExecutionIsFinished() { |
442 | return completed_pipelines >= total_pipelines || HasError(); |
443 | } |
444 | |
445 | PendingExecutionResult Executor::ExecuteTask() { |
446 | if (execution_result != PendingExecutionResult::RESULT_NOT_READY) { |
447 | return execution_result; |
448 | } |
449 | // check if there are any incomplete pipelines |
450 | auto &scheduler = TaskScheduler::GetScheduler(context); |
451 | while (completed_pipelines < total_pipelines) { |
452 | // there are! if we don't already have a task, fetch one |
453 | if (!task) { |
454 | scheduler.GetTaskFromProducer(token&: *producer, task); |
455 | } |
456 | if (task) { |
457 | // if we have a task, partially process it |
458 | auto result = task->Execute(mode: TaskExecutionMode::PROCESS_PARTIAL); |
459 | if (result == TaskExecutionResult::TASK_BLOCKED) { |
460 | task->Deschedule(); |
461 | task.reset(); |
462 | } else if (result == TaskExecutionResult::TASK_FINISHED) { |
463 | // if the task is finished, clean it up |
464 | task.reset(); |
465 | } |
466 | } |
467 | if (!HasError()) { |
468 | // we (partially) processed a task and no exceptions were thrown |
469 | // give back control to the caller |
470 | return PendingExecutionResult::RESULT_NOT_READY; |
471 | } |
472 | execution_result = PendingExecutionResult::EXECUTION_ERROR; |
473 | |
474 | // an exception has occurred executing one of the pipelines |
475 | // we need to cancel all tasks associated with this executor |
476 | CancelTasks(); |
477 | ThrowException(); |
478 | } |
479 | D_ASSERT(!task); |
480 | |
481 | lock_guard<mutex> elock(executor_lock); |
482 | pipelines.clear(); |
483 | NextExecutor(); |
484 | if (HasError()) { // LCOV_EXCL_START |
485 | // an exception has occurred executing one of the pipelines |
486 | execution_result = PendingExecutionResult::EXECUTION_ERROR; |
487 | ThrowException(); |
488 | } // LCOV_EXCL_STOP |
489 | execution_result = PendingExecutionResult::RESULT_READY; |
490 | return execution_result; |
491 | } |
492 | |
493 | void Executor::Reset() { |
494 | lock_guard<mutex> elock(executor_lock); |
495 | physical_plan = nullptr; |
496 | cancelled = false; |
497 | owned_plan.reset(); |
498 | root_executor.reset(); |
499 | root_pipelines.clear(); |
500 | root_pipeline_idx = 0; |
501 | completed_pipelines = 0; |
502 | total_pipelines = 0; |
503 | exceptions.clear(); |
504 | pipelines.clear(); |
505 | events.clear(); |
506 | to_be_rescheduled_tasks.clear(); |
507 | execution_result = PendingExecutionResult::RESULT_NOT_READY; |
508 | } |
509 | |
510 | shared_ptr<Pipeline> Executor::CreateChildPipeline(Pipeline ¤t, PhysicalOperator &op) { |
511 | D_ASSERT(!current.operators.empty()); |
512 | D_ASSERT(op.IsSource()); |
513 | // found another operator that is a source, schedule a child pipeline |
514 | // 'op' is the source, and the sink is the same |
515 | auto child_pipeline = make_shared<Pipeline>(args&: *this); |
516 | child_pipeline->sink = current.sink; |
517 | child_pipeline->source = &op; |
518 | |
519 | // the child pipeline has the same operators up until 'op' |
520 | for (auto current_op : current.operators) { |
521 | if (¤t_op.get() == &op) { |
522 | break; |
523 | } |
524 | child_pipeline->operators.push_back(x: current_op); |
525 | } |
526 | |
527 | return child_pipeline; |
528 | } |
529 | |
530 | vector<LogicalType> Executor::GetTypes() { |
531 | D_ASSERT(physical_plan); |
532 | return physical_plan->GetTypes(); |
533 | } |
534 | |
535 | void Executor::PushError(PreservedError exception) { |
536 | lock_guard<mutex> elock(error_lock); |
537 | // interrupt execution of any other pipelines that belong to this executor |
538 | context.interrupted = true; |
539 | // push the exception onto the stack |
540 | exceptions.push_back(x: std::move(exception)); |
541 | } |
542 | |
543 | bool Executor::HasError() { |
544 | lock_guard<mutex> elock(error_lock); |
545 | return !exceptions.empty(); |
546 | } |
547 | |
548 | void Executor::ThrowException() { |
549 | lock_guard<mutex> elock(error_lock); |
550 | D_ASSERT(!exceptions.empty()); |
551 | auto &entry = exceptions[0]; |
552 | entry.Throw(); |
553 | } |
554 | |
555 | void Executor::Flush(ThreadContext &tcontext) { |
556 | profiler->Flush(profiler&: tcontext.profiler); |
557 | } |
558 | |
559 | bool Executor::GetPipelinesProgress(double ¤t_progress) { // LCOV_EXCL_START |
560 | lock_guard<mutex> elock(executor_lock); |
561 | |
562 | vector<double> progress; |
563 | vector<idx_t> cardinality; |
564 | idx_t total_cardinality = 0; |
565 | for (auto &pipeline : pipelines) { |
566 | double child_percentage; |
567 | idx_t child_cardinality; |
568 | |
569 | if (!pipeline->GetProgress(current_percentage&: child_percentage, estimated_cardinality&: child_cardinality)) { |
570 | return false; |
571 | } |
572 | progress.push_back(x: child_percentage); |
573 | cardinality.push_back(x: child_cardinality); |
574 | total_cardinality += child_cardinality; |
575 | } |
576 | current_progress = 0; |
577 | for (size_t i = 0; i < progress.size(); i++) { |
578 | current_progress += progress[i] * double(cardinality[i]) / double(total_cardinality); |
579 | } |
580 | return true; |
581 | } // LCOV_EXCL_STOP |
582 | |
583 | bool Executor::HasResultCollector() { |
584 | return physical_plan->type == PhysicalOperatorType::RESULT_COLLECTOR; |
585 | } |
586 | |
587 | unique_ptr<QueryResult> Executor::GetResult() { |
588 | D_ASSERT(HasResultCollector()); |
589 | auto &result_collector = physical_plan->Cast<PhysicalResultCollector>(); |
590 | D_ASSERT(result_collector.sink_state); |
591 | return result_collector.GetResult(state&: *result_collector.sink_state); |
592 | } |
593 | |
594 | unique_ptr<DataChunk> Executor::FetchChunk() { |
595 | D_ASSERT(physical_plan); |
596 | |
597 | auto chunk = make_uniq<DataChunk>(); |
598 | root_executor->InitializeChunk(chunk&: *chunk); |
599 | while (true) { |
600 | root_executor->ExecutePull(result&: *chunk); |
601 | if (chunk->size() == 0) { |
602 | root_executor->PullFinalize(); |
603 | if (NextExecutor()) { |
604 | continue; |
605 | } |
606 | break; |
607 | } else { |
608 | break; |
609 | } |
610 | } |
611 | return chunk; |
612 | } |
613 | |
614 | } // namespace duckdb |
615 | |