1 // Copyright (c) 2012 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
5 #ifndef BASE_TRACKED_OBJECTS_H_
6 #define BASE_TRACKED_OBJECTS_H_
15 #include "base/atomicops.h"
16 #include "base/base_export.h"
17 #include "base/basictypes.h"
18 #include "base/containers/hash_tables.h"
19 #include "base/gtest_prod_util.h"
20 #include "base/lazy_instance.h"
21 #include "base/location.h"
22 #include "base/process/process_handle.h"
23 #include "base/profiler/alternate_timer.h"
24 #include "base/profiler/tracked_time.h"
25 #include "base/synchronization/lock.h"
26 #include "base/threading/thread_checker.h"
27 #include "base/threading/thread_local_storage.h"
33 // TrackedObjects provides a database of stats about objects (generally Tasks)
34 // that are tracked. Tracking means their birth, death, duration, birth thread,
35 // death thread, and birth place are recorded. This data is carefully spread
36 // across a series of objects so that the counts and times can be rapidly
37 // updated without (usually) having to lock the data, and hence there is usually
38 // very little contention caused by the tracking. The data can be viewed via
39 // the about:profiler URL, with a variety of sorting and filtering choices.
41 // These classes serve as the basis of a profiler of sorts for the Tasks system.
42 // As a result, design decisions were made to maximize speed, by minimizing
43 // recurring allocation/deallocation, lock contention and data copying. In the
44 // "stable" state, which is reached relatively quickly, there is no separate
45 // marginal allocation cost associated with construction or destruction of
46 // tracked objects, no locks are generally employed, and probably the largest
47 // computational cost is associated with obtaining start and stop times for
48 // instances as they are created and destroyed.
50 // The following describes the life cycle of tracking an instance.
52 // First off, when the instance is created, the FROM_HERE macro is expanded
53 // to specify the birth place (file, line, function) where the instance was
54 // created. That data is used to create a transient Location instance
55 // encapsulating the above triple of information. The strings (like __FILE__)
56 // are passed around by reference, with the assumption that they are static, and
57 // will never go away. This ensures that the strings can be dealt with as atoms
58 // with great efficiency (i.e., copying of strings is never needed, and
59 // comparisons for equality can be based on pointer comparisons).
61 // Next, a Births instance is created for use ONLY on the thread where this
62 // instance was created. That Births instance records (in a base class
63 // BirthOnThread) references to the static data provided in a Location instance,
64 // as well as a pointer specifying the thread on which the birth takes place.
65 // Hence there is at most one Births instance for each Location on each thread.
66 // The derived Births class contains slots for recording statistics about all
67 // instances born at the same location. Statistics currently include only the
68 // count of instances constructed.
70 // Since the base class BirthOnThread contains only constant data, it can be
71 // freely accessed by any thread at any time (i.e., only the statistic needs to
72 // be handled carefully, and stats are updated exclusively on the birth thread).
74 // For Tasks, having now either constructed or found the Births instance
75 // described above, a pointer to the Births instance is then recorded into the
76 // PendingTask structure in MessageLoop. This fact alone is very useful in
77 // debugging, when there is a question of where an instance came from. In
78 // addition, the birth time is also recorded and used to later evaluate the
79 // lifetime duration of the whole Task. As a result of the above embedding, we
80 // can find out a Task's location of birth, and thread of birth, without using
81 // any locks, as all that data is constant across the life of the process.
83 // The above work *could* also be done for any other object as well by calling
84 // TallyABirthIfActive() and TallyRunOnNamedThreadIfTracking() as appropriate.
86 // The amount of memory used in the above data structures depends on how many
87 // threads there are, and how many Locations of construction there are.
88 // Fortunately, we don't use memory that is the product of those two counts, but
89 // rather we only need one Births instance for each thread that constructs an
90 // instance at a Location. In many cases, instances are only created on one
91 // thread, so the memory utilization is actually fairly restrained.
93 // Lastly, when an instance is deleted, the final tallies of statistics are
94 // carefully accumulated. That tallying writes into slots (members) in a
95 // collection of DeathData instances. For each birth place Location that is
96 // destroyed on a thread, there is a DeathData instance to record the additional
97 // death count, as well as accumulate the run-time and queue-time durations for
98 // the instance as it is destroyed (dies). By maintaining a single place to
99 // aggregate this running sum *only* for the given thread, we avoid the need to
100 // lock such DeathData instances. (i.e., these accumulated stats in a DeathData
101 // instance are exclusively updated by the singular owning thread).
103 // With the above life cycle description complete, the major remaining detail
104 // is explaining how each thread maintains a list of DeathData instances, and
105 // of Births instances, and is able to avoid additional (redundant/unnecessary)
108 // Each thread maintains a list of data items specific to that thread in a
109 // ThreadData instance (for that specific thread only). The two critical items
110 // are lists of DeathData and Births instances. These lists are maintained in
111 // STL maps, which are indexed by Location. As noted earlier, we can compare
112 // locations very efficiently as we consider the underlying data (file,
113 // function, line) to be atoms, and hence pointer comparison is used rather than
114 // (slow) string comparisons.
116 // To provide a mechanism for iterating over all "known threads," which means
117 // threads that have recorded a birth or a death, we create a singly linked list
118 // of ThreadData instances. Each such instance maintains a pointer to the next
119 // one. A static member of ThreadData provides a pointer to the first item on
120 // this global list, and access via that all_thread_data_list_head_ item
121 // requires the use of the list_lock_.
122 // When new ThreadData instances is added to the global list, it is pre-pended,
123 // which ensures that any prior acquisition of the list is valid (i.e., the
124 // holder can iterate over it without fear of it changing, or the necessity of
125 // using an additional lock. Iterations are actually pretty rare (used
126 // primarily for cleanup, or snapshotting data for display), so this lock has
127 // very little global performance impact.
129 // The above description tries to define the high performance (run time)
130 // portions of these classes. After gathering statistics, calls instigated
131 // by visiting about:profiler will assemble and aggregate data for display. The
132 // following data structures are used for producing such displays. They are
133 // not performance critical, and their only major constraint is that they should
134 // be able to run concurrently with ongoing augmentation of the birth and death
137 // This header also exports collection of classes that provide "snapshotted"
138 // representations of the core tracked_objects:: classes. These snapshotted
139 // representations are designed for safe transmission of the tracked_objects::
140 // data across process boundaries. Each consists of:
141 // (1) a default constructor, to support the IPC serialization macros,
142 // (2) a constructor that extracts data from the type being snapshotted, and
143 // (3) the snapshotted data.
145 // For a given birth location, information about births is spread across data
146 // structures that are asynchronously changing on various threads. For
147 // serialization and display purposes, we need to construct TaskSnapshot
148 // instances for each combination of birth thread, death thread, and location,
149 // along with the count of such lifetimes. We gather such data into a
150 // TaskSnapshot instances, so that such instances can be sorted and
151 // aggregated (and remain frozen during our processing).
153 // Profiling consists of phases. The concrete phase in the sequence of phases
154 // is identified by its 0-based index.
156 // The ProcessDataPhaseSnapshot struct is a serialized representation of the
157 // list of ThreadData objects for a process for a concrete profiling phase. It
158 // holds a set of TaskSnapshots. The statistics in a snapshot are gathered
159 // asynhcronously relative to their ongoing updates.
160 // It is possible, though highly unlikely, that stats could be incorrectly
161 // recorded by this process (all data is held in 32 bit ints, but we are not
162 // atomically collecting all data, so we could have count that does not, for
163 // example, match with the number of durations we accumulated). The advantage
164 // to having fast (non-atomic) updates of the data outweighs the minimal risk of
165 // a singular corrupt statistic snapshot (only the snapshot could be corrupt,
166 // not the underlying and ongoing statistic). In contrast, pointer data that
167 // is accessed during snapshotting is completely invariant, and hence is
168 // perfectly acquired (i.e., no potential corruption, and no risk of a bad
169 // memory reference).
171 // TODO(jar): We can implement a Snapshot system that *tries* to grab the
172 // snapshots on the source threads *when* they have MessageLoops available
173 // (worker threads don't have message loops generally, and hence gathering from
174 // them will continue to be asynchronous). We had an implementation of this in
175 // the past, but the difficulty is dealing with message loops being terminated.
176 // We can *try* to spam the available threads via some message loop proxy to
177 // achieve this feat, and it *might* be valuable when we are collecting data
178 // for upload via UMA (where correctness of data may be more significant than
179 // for a single screen of about:profiler).
181 // TODO(jar): We need to store DataCollections, and provide facilities for
182 // taking the difference between two gathered DataCollections. For now, we're
183 // just adding a hack that Reset()s to zero all counts and stats. This is also
184 // done in a slightly thread-unsafe fashion, as the resetting is done
185 // asynchronously relative to ongoing updates (but all data is 32 bit in size).
186 // For basic profiling, this will work "most of the time," and should be
187 // sufficient... but storing away DataCollections is the "right way" to do this.
188 // We'll accomplish this via JavaScript storage of snapshots, and then we'll
189 // remove the Reset() methods. We may also need a short-term-max value in
190 // DeathData that is reset (as synchronously as possible) during each snapshot.
191 // This will facilitate displaying a max value for each snapshot period.
193 namespace tracked_objects
{
195 //------------------------------------------------------------------------------
196 // For a specific thread, and a specific birth place, the collection of all
197 // death info (with tallies for each death thread, to prevent access conflicts).
199 class BASE_EXPORT BirthOnThread
{
201 BirthOnThread(const Location
& location
, const ThreadData
& current
);
203 const Location
& location() const { return location_
; }
204 const ThreadData
* birth_thread() const { return birth_thread_
; }
207 // File/lineno of birth. This defines the essence of the task, as the context
208 // of the birth (construction) often tell what the item is for. This field
209 // is const, and hence safe to access from any thread.
210 const Location location_
;
212 // The thread that records births into this object. Only this thread is
213 // allowed to update birth_count_ (which changes over time).
214 const ThreadData
* const birth_thread_
;
216 DISALLOW_COPY_AND_ASSIGN(BirthOnThread
);
219 //------------------------------------------------------------------------------
220 // A "snapshotted" representation of the BirthOnThread class.
222 struct BASE_EXPORT BirthOnThreadSnapshot
{
223 BirthOnThreadSnapshot();
224 explicit BirthOnThreadSnapshot(const BirthOnThread
& birth
);
225 ~BirthOnThreadSnapshot();
227 LocationSnapshot location
;
228 std::string thread_name
;
231 //------------------------------------------------------------------------------
232 // A class for accumulating counts of births (without bothering with a map<>).
234 class BASE_EXPORT Births
: public BirthOnThread
{
236 Births(const Location
& location
, const ThreadData
& current
);
238 int birth_count() const;
240 // When we have a birth we update the count for this birthplace.
244 // The number of births on this thread for our location_.
247 DISALLOW_COPY_AND_ASSIGN(Births
);
250 //------------------------------------------------------------------------------
251 // A "snapshotted" representation of the DeathData class.
253 struct BASE_EXPORT DeathDataSnapshot
{
256 // Constructs the snapshot from individual values.
257 // The alternative would be taking a DeathData parameter, but this would
258 // create a loop since DeathData indirectly refers DeathDataSnapshot. Passing
259 // a wrapper structure as a param or using an empty constructor for
260 // snapshotting DeathData would be less efficient.
261 DeathDataSnapshot(int count
,
262 int32 run_duration_sum
,
263 int32 run_duration_max
,
264 int32 run_duration_sample
,
265 int32 queue_duration_sum
,
266 int32 queue_duration_max
,
267 int32 queue_duration_sample
);
268 ~DeathDataSnapshot();
270 // Calculates and returns the delta between this snapshot and an earlier
271 // snapshot of the same task |older|.
272 DeathDataSnapshot
Delta(const DeathDataSnapshot
& older
) const;
275 int32 run_duration_sum
;
276 int32 run_duration_max
;
277 int32 run_duration_sample
;
278 int32 queue_duration_sum
;
279 int32 queue_duration_max
;
280 int32 queue_duration_sample
;
283 //------------------------------------------------------------------------------
284 // A "snapshotted" representation of the DeathData for a particular profiling
285 // phase. Used as an element of the list of phase snapshots owned by DeathData.
287 struct DeathDataPhaseSnapshot
{
288 DeathDataPhaseSnapshot(int profiling_phase
,
290 int32 run_duration_sum
,
291 int32 run_duration_max
,
292 int32 run_duration_sample
,
293 int32 queue_duration_sum
,
294 int32 queue_duration_max
,
295 int32 queue_duration_sample
,
296 const DeathDataPhaseSnapshot
* prev
);
298 // Profiling phase at which completion this snapshot was taken.
301 // Death data snapshot.
302 DeathDataSnapshot death_data
;
304 // Pointer to a snapshot from the previous phase.
305 const DeathDataPhaseSnapshot
* prev
;
308 //------------------------------------------------------------------------------
309 // Information about deaths of a task on a given thread, called "death thread".
310 // Access to members of this class is never protected by a lock. The fields
311 // are accessed in such a way that corruptions resulting from race conditions
312 // are not significant, and don't accumulate as a result of multiple accesses.
313 // All invocations of DeathData::OnProfilingPhaseCompleted and
314 // ThreadData::SnapshotMaps (which takes DeathData snapshot) in a given process
315 // must be called from the same thread. It doesn't matter what thread it is, but
316 // it's important the same thread is used as a snapshot thread during the whole
317 // process lifetime. All fields except sample_probability_count_ can be
320 class BASE_EXPORT DeathData
{
323 DeathData(const DeathData
& other
);
326 // Update stats for a task destruction (death) that had a Run() time of
327 // |duration|, and has had a queueing delay of |queue_duration|.
328 void RecordDeath(const int32 queue_duration
,
329 const int32 run_duration
,
330 const uint32 random_number
);
332 // Metrics and past snapshots accessors, used only for serialization and in
334 int count() const { return count_
; }
335 int32
run_duration_sum() const { return run_duration_sum_
; }
336 int32
run_duration_max() const { return run_duration_max_
; }
337 int32
run_duration_sample() const { return run_duration_sample_
; }
338 int32
queue_duration_sum() const { return queue_duration_sum_
; }
339 int32
queue_duration_max() const { return queue_duration_max_
; }
340 int32
queue_duration_sample() const { return queue_duration_sample_
; }
341 const DeathDataPhaseSnapshot
* last_phase_snapshot() const {
342 return last_phase_snapshot_
;
345 // Called when the current profiling phase, identified by |profiling_phase|,
347 // Must be called only on the snapshot thread.
348 void OnProfilingPhaseCompleted(int profiling_phase
);
351 // Members are ordered from most regularly read and updated, to least
352 // frequently used. This might help a bit with cache lines.
353 // Number of runs seen (divisor for calculating averages).
354 // Can be incremented only on the death thread.
357 // Count used in determining probability of selecting exec/queue times from a
358 // recorded death as samples.
359 // Gets incremented only on the death thread, but can be set to 0 by
360 // OnProfilingPhaseCompleted() on the snapshot thread.
361 int sample_probability_count_
;
363 // Basic tallies, used to compute averages. Can be incremented only on the
365 int32 run_duration_sum_
;
366 int32 queue_duration_sum_
;
367 // Max values, used by local visualization routines. These are often read,
368 // but rarely updated. The max values get assigned only on the death thread,
369 // but these fields can be set to 0 by OnProfilingPhaseCompleted() on the
371 int32 run_duration_max_
;
372 int32 queue_duration_max_
;
373 // Samples, used by crowd sourcing gatherers. These are almost never read,
374 // and rarely updated. They can be modified only on the death thread.
375 int32 run_duration_sample_
;
376 int32 queue_duration_sample_
;
378 // Snapshot of this death data made at the last profiling phase completion, if
379 // any. DeathData owns the whole list starting with this pointer.
380 // Can be accessed only on the snapshot thread.
381 const DeathDataPhaseSnapshot
* last_phase_snapshot_
;
383 DISALLOW_ASSIGN(DeathData
);
386 //------------------------------------------------------------------------------
387 // A temporary collection of data that can be sorted and summarized. It is
388 // gathered (carefully) from many threads. Instances are held in arrays and
389 // processed, filtered, and rendered.
390 // The source of this data was collected on many threads, and is asynchronously
391 // changing. The data in this instance is not asynchronously changing.
393 struct BASE_EXPORT TaskSnapshot
{
395 TaskSnapshot(const BirthOnThreadSnapshot
& birth
,
396 const DeathDataSnapshot
& death_data
,
397 const std::string
& death_thread_name
);
400 BirthOnThreadSnapshot birth
;
401 // Delta between death data for a thread for a certain profiling phase and the
402 // snapshot for the pervious phase, if any. Otherwise, just a snapshot.
403 DeathDataSnapshot death_data
;
404 std::string death_thread_name
;
407 //------------------------------------------------------------------------------
408 // For each thread, we have a ThreadData that stores all tracking info generated
409 // on this thread. This prevents the need for locking as data accumulates.
410 // We use ThreadLocalStorage to quickly identfy the current ThreadData context.
411 // We also have a linked list of ThreadData instances, and that list is used to
412 // harvest data from all existing instances.
414 struct ProcessDataPhaseSnapshot
;
415 struct ProcessDataSnapshot
;
416 class BASE_EXPORT TaskStopwatch
;
418 // Map from profiling phase number to the process-wide snapshotted
419 // representation of the list of ThreadData objects that died during the given
421 typedef std::map
<int, ProcessDataPhaseSnapshot
> PhasedProcessDataSnapshotMap
;
423 class BASE_EXPORT ThreadData
{
425 // Current allowable states of the tracking system. The states can vary
426 // between ACTIVE and DEACTIVATED, but can never go back to UNINITIALIZED.
428 UNINITIALIZED
, // Pristine, link-time state before running.
429 DORMANT_DURING_TESTS
, // Only used during testing.
430 DEACTIVATED
, // No longer recording profiling.
431 PROFILING_ACTIVE
, // Recording profiles.
432 STATUS_LAST
= PROFILING_ACTIVE
435 typedef base::hash_map
<Location
, Births
*, Location::Hash
> BirthMap
;
436 typedef std::map
<const Births
*, DeathData
> DeathMap
;
438 // Initialize the current thread context with a new instance of ThreadData.
439 // This is used by all threads that have names, and should be explicitly
440 // set *before* any births on the threads have taken place. It is generally
441 // only used by the message loop, which has a well defined thread name.
442 static void InitializeThreadContext(const std::string
& suggested_name
);
444 // Using Thread Local Store, find the current instance for collecting data.
445 // If an instance does not exist, construct one (and remember it for use on
447 // This may return NULL if the system is disabled for any reason.
448 static ThreadData
* Get();
450 // Fills |process_data_snapshot| with phased snapshots of all profiling
451 // phases, including the current one, identified by |current_profiling_phase|.
452 // |current_profiling_phase| is necessary because a child process can start
453 // after several phase-changing events, so it needs to receive the current
454 // phase number from the browser process to fill the correct entry for the
455 // current phase in the |process_data_snapshot| map.
456 static void Snapshot(int current_profiling_phase
,
457 ProcessDataSnapshot
* process_data_snapshot
);
459 // Called when the current profiling phase, identified by |profiling_phase|,
461 // |profiling_phase| is necessary because a child process can start after
462 // several phase-changing events, so it needs to receive the phase number from
463 // the browser process to fill the correct entry in the
464 // completed_phases_snapshots_ map.
465 static void OnProfilingPhaseCompleted(int profiling_phase
);
467 // Finds (or creates) a place to count births from the given location in this
468 // thread, and increment that tally.
469 // TallyABirthIfActive will returns NULL if the birth cannot be tallied.
470 static Births
* TallyABirthIfActive(const Location
& location
);
472 // Records the end of a timed run of an object. The |completed_task| contains
473 // a pointer to a Births, the time_posted, and a delayed_start_time if any.
474 // The |start_of_run| indicates when we started to perform the run of the
475 // task. The delayed_start_time is non-null for tasks that were posted as
476 // delayed tasks, and it indicates when the task should have run (i.e., when
477 // it should have posted out of the timer queue, and into the work queue.
478 // The |end_of_run| was just obtained by a call to Now() (just after the task
479 // finished). It is provided as an argument to help with testing.
480 static void TallyRunOnNamedThreadIfTracking(
481 const base::TrackingInfo
& completed_task
,
482 const TaskStopwatch
& stopwatch
);
484 // Record the end of a timed run of an object. The |birth| is the record for
485 // the instance, the |time_posted| records that instant, which is presumed to
486 // be when the task was posted into a queue to run on a worker thread.
487 // The |start_of_run| is when the worker thread started to perform the run of
489 // The |end_of_run| was just obtained by a call to Now() (just after the task
491 static void TallyRunOnWorkerThreadIfTracking(const Births
* births
,
492 const TrackedTime
& time_posted
,
493 const TaskStopwatch
& stopwatch
);
495 // Record the end of execution in region, generally corresponding to a scope
497 static void TallyRunInAScopedRegionIfTracking(const Births
* births
,
498 const TaskStopwatch
& stopwatch
);
500 const std::string
& thread_name() const { return thread_name_
; }
502 // Initializes all statics if needed (this initialization call should be made
503 // while we are single threaded).
504 static void Initialize();
506 // Sets internal status_.
507 // If |status| is false, then status_ is set to DEACTIVATED.
508 // If |status| is true, then status_ is set to PROFILING_ACTIVE.
509 static void InitializeAndSetTrackingStatus(Status status
);
511 static Status
status();
513 // Indicate if any sort of profiling is being done (i.e., we are more than
515 static bool TrackingStatus();
517 // Enables profiler timing.
518 static void EnableProfilerTiming();
520 // Provide a time function that does nothing (runs fast) when we don't have
521 // the profiler enabled. It will generally be optimized away when it is
522 // ifdef'ed to be small enough (allowing the profiler to be "compiled out" of
524 static TrackedTime
Now();
526 // Use the function |now| to provide current times, instead of calling the
527 // TrackedTime::Now() function. Since this alternate function is being used,
528 // the other time arguments (used for calculating queueing delay) will be
530 static void SetAlternateTimeSource(NowFunction
* now
);
532 // This function can be called at process termination to validate that thread
533 // cleanup routines have been called for at least some number of named
535 static void EnsureCleanupWasCalled(int major_threads_shutdown_count
);
538 friend class TaskStopwatch
;
539 // Allow only tests to call ShutdownSingleThreadedCleanup. We NEVER call it
540 // in production code.
541 // TODO(jar): Make this a friend in DEBUG only, so that the optimizer has a
542 // better change of optimizing (inlining? etc.) private methods (knowing that
543 // there will be no need for an external entry point).
544 friend class TrackedObjectsTest
;
545 FRIEND_TEST_ALL_PREFIXES(TrackedObjectsTest
, MinimalStartupShutdown
);
546 FRIEND_TEST_ALL_PREFIXES(TrackedObjectsTest
, TinyStartupShutdown
);
548 typedef std::map
<const BirthOnThread
*, int> BirthCountMap
;
550 typedef std::vector
<std::pair
<const Births
*, DeathDataPhaseSnapshot
>>
553 // Worker thread construction creates a name since there is none.
554 explicit ThreadData(int thread_number
);
556 // Message loop based construction should provide a name.
557 explicit ThreadData(const std::string
& suggested_name
);
561 // Push this instance to the head of all_thread_data_list_head_, linking it to
562 // the previous head. This is performed after each construction, and leaves
563 // the instance permanently on that list.
564 void PushToHeadOfList();
566 // (Thread safe) Get start of list of all ThreadData instances using the lock.
567 static ThreadData
* first();
569 // Iterate through the null terminated list of ThreadData instances.
570 ThreadData
* next() const;
573 // In this thread's data, record a new birth.
574 Births
* TallyABirth(const Location
& location
);
576 // Find a place to record a death on this thread.
577 void TallyADeath(const Births
& births
,
578 int32 queue_duration
,
579 const TaskStopwatch
& stopwatch
);
581 // Snapshots (under a lock) the profiled data for the tasks for this thread
582 // and writes all of the executed tasks' data -- i.e. the data for all
583 // profiling phases (including the current one: |current_profiling_phase|) for
584 // the tasks with with entries in the death_map_ -- into |phased_snapshots|.
585 // Also updates the |birth_counts| tally for each task to keep track of the
586 // number of living instances of the task -- that is, each task maps to the
587 // number of births for the task that have not yet been balanced by a death.
588 void SnapshotExecutedTasks(int current_profiling_phase
,
589 PhasedProcessDataSnapshotMap
* phased_snapshots
,
590 BirthCountMap
* birth_counts
);
592 // Using our lock, make a copy of the specified maps. This call may be made
593 // on non-local threads, which necessitate the use of the lock to prevent
594 // the map(s) from being reallocated while they are copied.
595 void SnapshotMaps(int profiling_phase
,
597 DeathsSnapshot
* deaths
);
599 // Called for this thread when the current profiling phase, identified by
600 // |profiling_phase|, ends.
601 void OnProfilingPhaseCompletedOnThread(int profiling_phase
);
603 // This method is called by the TLS system when a thread terminates.
604 // The argument may be NULL if this thread has never tracked a birth or death.
605 static void OnThreadTermination(void* thread_data
);
607 // This method should be called when a worker thread terminates, so that we
608 // can save all the thread data into a cache of reusable ThreadData instances.
609 void OnThreadTerminationCleanup();
611 // Cleans up data structures, and returns statics to near pristine (mostly
612 // uninitialized) state. If there is any chance that other threads are still
613 // using the data structures, then the |leak| argument should be passed in as
614 // true, and the data structures (birth maps, death maps, ThreadData
615 // insntances, etc.) will be leaked and not deleted. If you have joined all
616 // threads since the time that InitializeAndSetTrackingStatus() was called,
617 // then you can pass in a |leak| value of false, and this function will
618 // delete recursively all data structures, starting with the list of
619 // ThreadData instances.
620 static void ShutdownSingleThreadedCleanup(bool leak
);
622 // When non-null, this specifies an external function that supplies monotone
623 // increasing time functcion.
624 static NowFunction
* now_function_
;
626 // If true, now_function_ returns values that can be used to calculate queue
628 static bool now_function_is_time_
;
630 // We use thread local store to identify which ThreadData to interact with.
631 static base::ThreadLocalStorage::StaticSlot tls_index_
;
633 // List of ThreadData instances for use with worker threads. When a worker
634 // thread is done (terminated), we push it onto this list. When a new worker
635 // thread is created, we first try to re-use a ThreadData instance from the
636 // list, and if none are available, construct a new one.
637 // This is only accessed while list_lock_ is held.
638 static ThreadData
* first_retired_worker_
;
640 // Link to the most recently created instance (starts a null terminated list).
641 // The list is traversed by about:profiler when it needs to snapshot data.
642 // This is only accessed while list_lock_ is held.
643 static ThreadData
* all_thread_data_list_head_
;
645 // The next available worker thread number. This should only be accessed when
646 // the list_lock_ is held.
647 static int worker_thread_data_creation_count_
;
649 // The number of times TLS has called us back to cleanup a ThreadData
650 // instance. This is only accessed while list_lock_ is held.
651 static int cleanup_count_
;
653 // Incarnation sequence number, indicating how many times (during unittests)
654 // we've either transitioned out of UNINITIALIZED, or into that state. This
655 // value is only accessed while the list_lock_ is held.
656 static int incarnation_counter_
;
658 // Protection for access to all_thread_data_list_head_, and to
659 // unregistered_thread_data_pool_. This lock is leaked at shutdown.
660 // The lock is very infrequently used, so we can afford to just make a lazy
661 // instance and be safe.
662 static base::LazyInstance
<base::Lock
>::Leaky list_lock_
;
664 // We set status_ to SHUTDOWN when we shut down the tracking service.
665 static base::subtle::Atomic32 status_
;
667 // Link to next instance (null terminated list). Used to globally track all
668 // registered instances (corresponds to all registered threads where we keep
672 // Pointer to another ThreadData instance for a Worker-Thread that has been
673 // retired (its thread was terminated). This value is non-NULL only for a
674 // retired ThreadData associated with a Worker-Thread.
675 ThreadData
* next_retired_worker_
;
677 // The name of the thread that is being recorded. If this thread has no
678 // message_loop, then this is a worker thread, with a sequence number postfix.
679 std::string thread_name_
;
681 // Indicate if this is a worker thread, and the ThreadData contexts should be
682 // stored in the unregistered_thread_data_pool_ when not in use.
683 // Value is zero when it is not a worker thread. Value is a positive integer
684 // corresponding to the created thread name if it is a worker thread.
685 int worker_thread_number_
;
687 // A map used on each thread to keep track of Births on this thread.
688 // This map should only be accessed on the thread it was constructed on.
689 // When a snapshot is needed, this structure can be locked in place for the
690 // duration of the snapshotting activity.
693 // Similar to birth_map_, this records informations about death of tracked
694 // instances (i.e., when a tracked instance was destroyed on this thread).
695 // It is locked before changing, and hence other threads may access it by
696 // locking before reading it.
699 // Lock to protect *some* access to BirthMap and DeathMap. The maps are
700 // regularly read and written on this thread, but may only be read from other
701 // threads. To support this, we acquire this lock if we are writing from this
702 // thread, or reading from another thread. For reading from this thread we
703 // don't need a lock, as there is no potential for a conflict since the
704 // writing is only done from this thread.
705 mutable base::Lock map_lock_
;
707 // A random number that we used to select decide which sample to keep as a
708 // representative sample in each DeathData instance. We can't start off with
709 // much randomness (because we can't call RandInt() on all our threads), so
710 // we stir in more and more as we go.
711 uint32 random_number_
;
713 // Record of what the incarnation_counter_ was when this instance was created.
714 // If the incarnation_counter_ has changed, then we avoid pushing into the
715 // pool (this is only critical in tests which go through multiple
717 int incarnation_count_for_pool_
;
719 // Most recently started (i.e. most nested) stopwatch on the current thread,
720 // if it exists; NULL otherwise.
721 TaskStopwatch
* current_stopwatch_
;
723 DISALLOW_COPY_AND_ASSIGN(ThreadData
);
726 //------------------------------------------------------------------------------
727 // Stopwatch to measure task run time or simply create a time interval that will
728 // be subtracted from the current most nested task's run time. Stopwatches
729 // coordinate with the stopwatches in which they are nested to avoid
730 // double-counting nested tasks run times.
732 class BASE_EXPORT TaskStopwatch
{
734 // Starts the stopwatch.
744 // Returns the start time.
745 TrackedTime
StartTime() const;
747 // Task's duration is calculated as the wallclock duration between starting
748 // and stopping this stopwatch, minus the wallclock durations of any other
749 // instances that are immediately nested in this one, started and stopped on
750 // this thread during that period.
751 int32
RunDurationMs() const;
753 // Returns tracking info for the current thread.
754 ThreadData
* GetThreadData() const;
757 // Time when the stopwatch was started.
758 TrackedTime start_time_
;
760 // Wallclock duration of the task.
761 int32 wallclock_duration_ms_
;
763 // Tracking info for the current thread.
764 ThreadData
* current_thread_data_
;
766 // Sum of wallclock durations of all stopwatches that were directly nested in
768 int32 excluded_duration_ms_
;
770 // Stopwatch which was running on our thread when this stopwatch was started.
771 // That preexisting stopwatch must be adjusted to the exclude the wallclock
772 // duration of this stopwatch.
773 TaskStopwatch
* parent_
;
776 // State of the stopwatch. Stopwatch is first constructed in a created state
777 // state, then is optionally started/stopped, then destructed.
778 enum { CREATED
, RUNNING
, STOPPED
} state_
;
780 // Currently running stopwatch that is directly nested in this one, if such
781 // stopwatch exists. NULL otherwise.
782 TaskStopwatch
* child_
;
786 //------------------------------------------------------------------------------
787 // A snapshotted representation of the list of ThreadData objects for a process,
788 // for a single profiling phase.
790 struct BASE_EXPORT ProcessDataPhaseSnapshot
{
792 ProcessDataPhaseSnapshot();
793 ~ProcessDataPhaseSnapshot();
795 std::vector
<TaskSnapshot
> tasks
;
798 //------------------------------------------------------------------------------
799 // A snapshotted representation of the list of ThreadData objects for a process,
800 // for all profiling phases, including the current one.
802 struct BASE_EXPORT ProcessDataSnapshot
{
804 ProcessDataSnapshot();
805 ~ProcessDataSnapshot();
807 PhasedProcessDataSnapshotMap phased_snapshots
;
808 base::ProcessId process_id
;
811 } // namespace tracked_objects
813 #endif // BASE_TRACKED_OBJECTS_H_