1 //===-- ThreadPlanStack.cpp -------------------------------------*- C++ -*-===//
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
7 //===----------------------------------------------------------------------===//
9 #include "lldb/Target/ThreadPlanStack.h"
10 #include "lldb/Target/Process.h"
11 #include "lldb/Target/Target.h"
12 #include "lldb/Target/Thread.h"
13 #include "lldb/Target/ThreadPlan.h"
14 #include "lldb/Utility/Log.h"
17 using namespace lldb_private
;
19 static void PrintPlanElement(Stream
&s
, const ThreadPlanSP
&plan
,
20 lldb::DescriptionLevel desc_level
,
24 s
.Printf("Element %d: ", elem_idx
);
25 plan
->GetDescription(&s
, desc_level
);
30 ThreadPlanStack::ThreadPlanStack(const Thread
&thread
, bool make_null
) {
32 // The ThreadPlanNull doesn't do anything to the Thread, so this is actually
33 // still a const operation.
35 ThreadPlanSP(new ThreadPlanNull(const_cast<Thread
&>(thread
))));
39 void ThreadPlanStack::DumpThreadPlans(Stream
&s
,
40 lldb::DescriptionLevel desc_level
,
41 bool include_internal
) const {
42 llvm::sys::ScopedReader
guard(m_stack_mutex
);
44 PrintOneStackNoLock(s
, "Active plan stack", m_plans
, desc_level
,
46 PrintOneStackNoLock(s
, "Completed plan stack", m_completed_plans
, desc_level
,
48 PrintOneStackNoLock(s
, "Discarded plan stack", m_discarded_plans
, desc_level
,
53 void ThreadPlanStack::PrintOneStackNoLock(Stream
&s
, llvm::StringRef stack_name
,
54 const PlanStack
&stack
,
55 lldb::DescriptionLevel desc_level
,
56 bool include_internal
) const {
57 // If the stack is empty, just exit:
61 // Make sure there are public completed plans:
62 bool any_public
= false;
63 if (!include_internal
) {
64 for (auto plan
: stack
) {
65 if (!plan
->GetPrivate()) {
72 if (include_internal
|| any_public
) {
75 s
<< stack_name
<< ":\n";
76 for (auto plan
: stack
) {
77 if (!include_internal
&& plan
->GetPrivate())
79 PrintPlanElement(s
, plan
, desc_level
, print_idx
++);
84 size_t ThreadPlanStack::CheckpointCompletedPlans() {
85 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
86 m_completed_plan_checkpoint
++;
87 m_completed_plan_store
.insert(
88 std::make_pair(m_completed_plan_checkpoint
, m_completed_plans
));
89 return m_completed_plan_checkpoint
;
92 void ThreadPlanStack::RestoreCompletedPlanCheckpoint(size_t checkpoint
) {
93 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
94 auto result
= m_completed_plan_store
.find(checkpoint
);
95 assert(result
!= m_completed_plan_store
.end() &&
96 "Asked for a checkpoint that didn't exist");
97 m_completed_plans
.swap((*result
).second
);
98 m_completed_plan_store
.erase(result
);
101 void ThreadPlanStack::DiscardCompletedPlanCheckpoint(size_t checkpoint
) {
102 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
103 m_completed_plan_store
.erase(checkpoint
);
106 void ThreadPlanStack::ThreadDestroyed(Thread
*thread
) {
107 // Tell the plan stacks that this thread is going away:
108 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
109 for (ThreadPlanSP plan
: m_plans
)
110 plan
->ThreadDestroyed();
112 for (ThreadPlanSP plan
: m_discarded_plans
)
113 plan
->ThreadDestroyed();
115 for (ThreadPlanSP plan
: m_completed_plans
)
116 plan
->ThreadDestroyed();
118 // Now clear the current plan stacks:
120 m_discarded_plans
.clear();
121 m_completed_plans
.clear();
123 // Push a ThreadPlanNull on the plan stack. That way we can continue
124 // assuming that the plan stack is never empty, but if somebody errantly asks
125 // questions of a destroyed thread without checking first whether it is
126 // destroyed, they won't crash.
127 if (thread
!= nullptr) {
128 lldb::ThreadPlanSP
null_plan_sp(new ThreadPlanNull(*thread
));
129 m_plans
.push_back(null_plan_sp
);
133 void ThreadPlanStack::PushPlan(lldb::ThreadPlanSP new_plan_sp
) {
134 // If the thread plan doesn't already have a tracer, give it its parent's
136 // The first plan has to be a base plan:
137 { // Scope for Lock - DidPush often adds plans to the stack:
138 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
139 assert((m_plans
.size() > 0 || new_plan_sp
->IsBasePlan()) &&
140 "Zeroth plan must be a base plan");
142 if (!new_plan_sp
->GetThreadPlanTracer()) {
143 assert(!m_plans
.empty());
144 new_plan_sp
->SetThreadPlanTracer(m_plans
.back()->GetThreadPlanTracer());
146 m_plans
.push_back(new_plan_sp
);
148 new_plan_sp
->DidPush();
151 lldb::ThreadPlanSP
ThreadPlanStack::PopPlan() {
152 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
153 assert(m_plans
.size() > 1 && "Can't pop the base thread plan");
155 // Note that moving the top element of the vector would leave it in an
156 // undefined state, and break the guarantee that the stack's thread plans are
158 lldb::ThreadPlanSP plan_sp
= m_plans
.back();
160 m_completed_plans
.push_back(plan_sp
);
165 lldb::ThreadPlanSP
ThreadPlanStack::DiscardPlan() {
166 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
167 return DiscardPlanNoLock();
170 lldb::ThreadPlanSP
ThreadPlanStack::DiscardPlanNoLock() {
171 assert(m_plans
.size() > 1 && "Can't discard the base thread plan");
173 // Note that moving the top element of the vector would leave it in an
174 // undefined state, and break the guarantee that the stack's thread plans are
176 lldb::ThreadPlanSP plan_sp
= m_plans
.back();
178 m_discarded_plans
.push_back(plan_sp
);
183 // If the input plan is nullptr, discard all plans. Otherwise make sure this
184 // plan is in the stack, and if so discard up to and including it.
185 void ThreadPlanStack::DiscardPlansUpToPlan(ThreadPlan
*up_to_plan_ptr
) {
186 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
187 int stack_size
= m_plans
.size();
189 if (up_to_plan_ptr
== nullptr) {
190 for (int i
= stack_size
- 1; i
> 0; i
--)
195 bool found_it
= false;
196 for (int i
= stack_size
- 1; i
> 0; i
--) {
197 if (m_plans
[i
].get() == up_to_plan_ptr
) {
204 bool last_one
= false;
205 for (int i
= stack_size
- 1; i
> 0 && !last_one
; i
--) {
206 if (GetCurrentPlanNoLock().get() == up_to_plan_ptr
)
213 void ThreadPlanStack::DiscardAllPlans() {
214 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
215 int stack_size
= m_plans
.size();
216 for (int i
= stack_size
- 1; i
> 0; i
--) {
221 void ThreadPlanStack::DiscardConsultingControllingPlans() {
222 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
224 int controlling_plan_idx
;
227 // Find the first controlling plan, see if it wants discarding, and if yes
229 for (controlling_plan_idx
= m_plans
.size() - 1; controlling_plan_idx
>= 0;
230 controlling_plan_idx
--) {
231 if (m_plans
[controlling_plan_idx
]->IsControllingPlan()) {
232 discard
= m_plans
[controlling_plan_idx
]->OkayToDiscard();
237 // If the controlling plan doesn't want to get discarded, then we're done.
241 // First pop all the dependent plans:
242 for (int i
= m_plans
.size() - 1; i
> controlling_plan_idx
; i
--) {
246 // Now discard the controlling plan itself.
247 // The bottom-most plan never gets discarded. "OkayToDiscard" for it
248 // means discard it's dependent plans, but not it...
249 if (controlling_plan_idx
> 0) {
255 lldb::ThreadPlanSP
ThreadPlanStack::GetCurrentPlan() const {
256 llvm::sys::ScopedReader
guard(m_stack_mutex
);
257 return GetCurrentPlanNoLock();
260 lldb::ThreadPlanSP
ThreadPlanStack::GetCurrentPlanNoLock() const {
261 assert(m_plans
.size() != 0 && "There will always be a base plan.");
262 return m_plans
.back();
265 lldb::ThreadPlanSP
ThreadPlanStack::GetCompletedPlan(bool skip_private
) const {
266 llvm::sys::ScopedReader
guard(m_stack_mutex
);
267 if (m_completed_plans
.empty())
271 return m_completed_plans
.back();
273 for (int i
= m_completed_plans
.size() - 1; i
>= 0; i
--) {
274 lldb::ThreadPlanSP completed_plan_sp
;
275 completed_plan_sp
= m_completed_plans
[i
];
276 if (!completed_plan_sp
->GetPrivate())
277 return completed_plan_sp
;
282 lldb::ThreadPlanSP
ThreadPlanStack::GetPlanByIndex(uint32_t plan_idx
,
283 bool skip_private
) const {
284 llvm::sys::ScopedReader
guard(m_stack_mutex
);
287 for (lldb::ThreadPlanSP plan_sp
: m_plans
) {
288 if (skip_private
&& plan_sp
->GetPrivate())
297 lldb::ValueObjectSP
ThreadPlanStack::GetReturnValueObject() const {
298 llvm::sys::ScopedReader
guard(m_stack_mutex
);
299 if (m_completed_plans
.empty())
302 for (int i
= m_completed_plans
.size() - 1; i
>= 0; i
--) {
303 lldb::ValueObjectSP return_valobj_sp
;
304 return_valobj_sp
= m_completed_plans
[i
]->GetReturnValueObject();
305 if (return_valobj_sp
)
306 return return_valobj_sp
;
311 lldb::ExpressionVariableSP
ThreadPlanStack::GetExpressionVariable() const {
312 llvm::sys::ScopedReader
guard(m_stack_mutex
);
313 if (m_completed_plans
.empty())
316 for (int i
= m_completed_plans
.size() - 1; i
>= 0; i
--) {
317 lldb::ExpressionVariableSP expression_variable_sp
;
318 expression_variable_sp
= m_completed_plans
[i
]->GetExpressionVariable();
319 if (expression_variable_sp
)
320 return expression_variable_sp
;
324 bool ThreadPlanStack::AnyPlans() const {
325 llvm::sys::ScopedReader
guard(m_stack_mutex
);
326 // There is always a base plan...
327 return m_plans
.size() > 1;
330 bool ThreadPlanStack::AnyCompletedPlans() const {
331 llvm::sys::ScopedReader
guard(m_stack_mutex
);
332 return !m_completed_plans
.empty();
335 bool ThreadPlanStack::AnyDiscardedPlans() const {
336 llvm::sys::ScopedReader
guard(m_stack_mutex
);
337 return !m_discarded_plans
.empty();
340 bool ThreadPlanStack::IsPlanDone(ThreadPlan
*in_plan
) const {
341 llvm::sys::ScopedReader
guard(m_stack_mutex
);
342 for (auto plan
: m_completed_plans
) {
343 if (plan
.get() == in_plan
)
349 bool ThreadPlanStack::WasPlanDiscarded(ThreadPlan
*in_plan
) const {
350 llvm::sys::ScopedReader
guard(m_stack_mutex
);
351 for (auto plan
: m_discarded_plans
) {
352 if (plan
.get() == in_plan
)
358 ThreadPlan
*ThreadPlanStack::GetPreviousPlan(ThreadPlan
*current_plan
) const {
359 llvm::sys::ScopedReader
guard(m_stack_mutex
);
360 if (current_plan
== nullptr)
363 // Look first in the completed plans, if the plan is here and there is
364 // a completed plan above it, return that.
365 int stack_size
= m_completed_plans
.size();
366 for (int i
= stack_size
- 1; i
> 0; i
--) {
367 if (current_plan
== m_completed_plans
[i
].get())
368 return m_completed_plans
[i
- 1].get();
371 // If this is the first completed plan, the previous one is the
372 // bottom of the regular plan stack.
373 if (stack_size
> 0 && m_completed_plans
[0].get() == current_plan
) {
374 return GetCurrentPlanNoLock().get();
377 // Otherwise look for it in the regular plans.
378 stack_size
= m_plans
.size();
379 for (int i
= stack_size
- 1; i
> 0; i
--) {
380 if (current_plan
== m_plans
[i
].get())
381 return m_plans
[i
- 1].get();
386 ThreadPlan
*ThreadPlanStack::GetInnermostExpression() const {
387 llvm::sys::ScopedReader
guard(m_stack_mutex
);
388 int stack_size
= m_plans
.size();
390 for (int i
= stack_size
- 1; i
> 0; i
--) {
391 if (m_plans
[i
]->GetKind() == ThreadPlan::eKindCallFunction
)
392 return m_plans
[i
].get();
397 void ThreadPlanStack::ClearThreadCache() {
398 llvm::sys::ScopedReader
guard(m_stack_mutex
);
399 for (lldb::ThreadPlanSP thread_plan_sp
: m_plans
)
400 thread_plan_sp
->ClearThreadCache();
403 void ThreadPlanStack::WillResume() {
404 llvm::sys::ScopedWriter
guard(m_stack_mutex
);
405 m_completed_plans
.clear();
406 m_discarded_plans
.clear();
409 void ThreadPlanStackMap::Update(ThreadList
¤t_threads
,
411 bool check_for_new
) {
413 std::lock_guard
<std::recursive_mutex
> guard(m_stack_map_mutex
);
414 // Now find all the new threads and add them to the map:
416 for (auto thread
: current_threads
.Threads()) {
417 lldb::tid_t cur_tid
= thread
->GetID();
418 if (!Find(cur_tid
)) {
420 thread
->QueueBasePlan(true);
425 // If we aren't reaping missing threads at this point,
429 // Otherwise scan for absent TID's.
430 std::vector
<lldb::tid_t
> missing_threads
;
431 // If we are going to delete plans from the plan stack,
432 // then scan for absent TID's:
433 for (auto &thread_plans
: m_plans_list
) {
434 lldb::tid_t cur_tid
= thread_plans
.first
;
435 ThreadSP thread_sp
= current_threads
.FindThreadByID(cur_tid
);
437 missing_threads
.push_back(cur_tid
);
439 for (lldb::tid_t tid
: missing_threads
) {
444 void ThreadPlanStackMap::DumpPlans(Stream
&strm
,
445 lldb::DescriptionLevel desc_level
,
446 bool internal
, bool condense_if_trivial
,
447 bool skip_unreported
) {
448 std::lock_guard
<std::recursive_mutex
> guard(m_stack_map_mutex
);
449 for (auto &elem
: m_plans_list
) {
450 lldb::tid_t tid
= elem
.first
;
451 uint32_t index_id
= 0;
452 ThreadSP thread_sp
= m_process
.GetThreadList().FindThreadByID(tid
);
454 if (skip_unreported
) {
459 index_id
= thread_sp
->GetIndexID();
461 if (condense_if_trivial
) {
462 if (!elem
.second
.AnyPlans() && !elem
.second
.AnyCompletedPlans() &&
463 !elem
.second
.AnyDiscardedPlans()) {
464 strm
.Printf("thread #%u: tid = 0x%4.4" PRIx64
"\n", index_id
, tid
);
467 strm
.Printf("No active thread plans\n");
474 strm
.Printf("thread #%u: tid = 0x%4.4" PRIx64
":\n", index_id
, tid
);
476 elem
.second
.DumpThreadPlans(strm
, desc_level
, internal
);
480 bool ThreadPlanStackMap::DumpPlansForTID(Stream
&strm
, lldb::tid_t tid
,
481 lldb::DescriptionLevel desc_level
,
483 bool condense_if_trivial
,
484 bool skip_unreported
) {
485 std::lock_guard
<std::recursive_mutex
> guard(m_stack_map_mutex
);
486 uint32_t index_id
= 0;
487 ThreadSP thread_sp
= m_process
.GetThreadList().FindThreadByID(tid
);
489 if (skip_unreported
) {
491 strm
.Format("Unknown TID: {0}", tid
);
497 index_id
= thread_sp
->GetIndexID();
498 ThreadPlanStack
*stack
= Find(tid
);
500 strm
.Format("Unknown TID: {0}\n", tid
);
504 if (condense_if_trivial
) {
505 if (!stack
->AnyPlans() && !stack
->AnyCompletedPlans() &&
506 !stack
->AnyDiscardedPlans()) {
507 strm
.Printf("thread #%u: tid = 0x%4.4" PRIx64
"\n", index_id
, tid
);
510 strm
.Printf("No active thread plans\n");
517 strm
.Printf("thread #%u: tid = 0x%4.4" PRIx64
":\n", index_id
, tid
);
519 stack
->DumpThreadPlans(strm
, desc_level
, internal
);
523 bool ThreadPlanStackMap::PrunePlansForTID(lldb::tid_t tid
) {
524 // We only remove the plans for unreported TID's.
525 std::lock_guard
<std::recursive_mutex
> guard(m_stack_map_mutex
);
526 ThreadSP thread_sp
= m_process
.GetThreadList().FindThreadByID(tid
);
530 return RemoveTID(tid
);