Supervised users: Re-check ManagementPolicy when ProfileIsSupervised changes.
[chromium-blink-merge.git] / base / tracked_objects.h
blob964326511b23355f0bca71b1671919d8cb3e8548
1 // Copyright (c) 2012 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
5 #ifndef BASE_TRACKED_OBJECTS_H_
6 #define BASE_TRACKED_OBJECTS_H_
8 #include <map>
9 #include <set>
10 #include <stack>
11 #include <string>
12 #include <utility>
13 #include <vector>
15 #include "base/base_export.h"
16 #include "base/basictypes.h"
17 #include "base/containers/hash_tables.h"
18 #include "base/gtest_prod_util.h"
19 #include "base/lazy_instance.h"
20 #include "base/location.h"
21 #include "base/process/process_handle.h"
22 #include "base/profiler/alternate_timer.h"
23 #include "base/profiler/tracked_time.h"
24 #include "base/synchronization/lock.h"
25 #include "base/threading/thread_local_storage.h"
27 namespace base {
28 struct TrackingInfo;
31 // TrackedObjects provides a database of stats about objects (generally Tasks)
32 // that are tracked. Tracking means their birth, death, duration, birth thread,
33 // death thread, and birth place are recorded. This data is carefully spread
34 // across a series of objects so that the counts and times can be rapidly
35 // updated without (usually) having to lock the data, and hence there is usually
36 // very little contention caused by the tracking. The data can be viewed via
37 // the about:profiler URL, with a variety of sorting and filtering choices.
39 // These classes serve as the basis of a profiler of sorts for the Tasks system.
40 // As a result, design decisions were made to maximize speed, by minimizing
41 // recurring allocation/deallocation, lock contention and data copying. In the
42 // "stable" state, which is reached relatively quickly, there is no separate
43 // marginal allocation cost associated with construction or destruction of
44 // tracked objects, no locks are generally employed, and probably the largest
45 // computational cost is associated with obtaining start and stop times for
46 // instances as they are created and destroyed.
48 // The following describes the life cycle of tracking an instance.
50 // First off, when the instance is created, the FROM_HERE macro is expanded
51 // to specify the birth place (file, line, function) where the instance was
52 // created. That data is used to create a transient Location instance
53 // encapsulating the above triple of information. The strings (like __FILE__)
54 // are passed around by reference, with the assumption that they are static, and
55 // will never go away. This ensures that the strings can be dealt with as atoms
56 // with great efficiency (i.e., copying of strings is never needed, and
57 // comparisons for equality can be based on pointer comparisons).
59 // Next, a Births instance is created for use ONLY on the thread where this
60 // instance was created. That Births instance records (in a base class
61 // BirthOnThread) references to the static data provided in a Location instance,
62 // as well as a pointer specifying the thread on which the birth takes place.
63 // Hence there is at most one Births instance for each Location on each thread.
64 // The derived Births class contains slots for recording statistics about all
65 // instances born at the same location. Statistics currently include only the
66 // count of instances constructed.
68 // Since the base class BirthOnThread contains only constant data, it can be
69 // freely accessed by any thread at any time (i.e., only the statistic needs to
70 // be handled carefully, and stats are updated exclusively on the birth thread).
72 // For Tasks, having now either constructed or found the Births instance
73 // described above, a pointer to the Births instance is then recorded into the
74 // PendingTask structure in MessageLoop. This fact alone is very useful in
75 // debugging, when there is a question of where an instance came from. In
76 // addition, the birth time is also recorded and used to later evaluate the
77 // lifetime duration of the whole Task. As a result of the above embedding, we
78 // can find out a Task's location of birth, and thread of birth, without using
79 // any locks, as all that data is constant across the life of the process.
81 // The above work *could* also be done for any other object as well by calling
82 // TallyABirthIfActive() and TallyRunOnNamedThreadIfTracking() as appropriate.
84 // The amount of memory used in the above data structures depends on how many
85 // threads there are, and how many Locations of construction there are.
86 // Fortunately, we don't use memory that is the product of those two counts, but
87 // rather we only need one Births instance for each thread that constructs an
88 // instance at a Location. In many cases, instances are only created on one
89 // thread, so the memory utilization is actually fairly restrained.
91 // Lastly, when an instance is deleted, the final tallies of statistics are
92 // carefully accumulated. That tallying writes into slots (members) in a
93 // collection of DeathData instances. For each birth place Location that is
94 // destroyed on a thread, there is a DeathData instance to record the additional
95 // death count, as well as accumulate the run-time and queue-time durations for
96 // the instance as it is destroyed (dies). By maintaining a single place to
97 // aggregate this running sum *only* for the given thread, we avoid the need to
98 // lock such DeathData instances. (i.e., these accumulated stats in a DeathData
99 // instance are exclusively updated by the singular owning thread).
101 // With the above life cycle description complete, the major remaining detail
102 // is explaining how each thread maintains a list of DeathData instances, and
103 // of Births instances, and is able to avoid additional (redundant/unnecessary)
104 // allocations.
106 // Each thread maintains a list of data items specific to that thread in a
107 // ThreadData instance (for that specific thread only). The two critical items
108 // are lists of DeathData and Births instances. These lists are maintained in
109 // STL maps, which are indexed by Location. As noted earlier, we can compare
110 // locations very efficiently as we consider the underlying data (file,
111 // function, line) to be atoms, and hence pointer comparison is used rather than
112 // (slow) string comparisons.
114 // To provide a mechanism for iterating over all "known threads," which means
115 // threads that have recorded a birth or a death, we create a singly linked list
116 // of ThreadData instances. Each such instance maintains a pointer to the next
117 // one. A static member of ThreadData provides a pointer to the first item on
118 // this global list, and access via that all_thread_data_list_head_ item
119 // requires the use of the list_lock_.
120 // When new ThreadData instances is added to the global list, it is pre-pended,
121 // which ensures that any prior acquisition of the list is valid (i.e., the
122 // holder can iterate over it without fear of it changing, or the necessity of
123 // using an additional lock. Iterations are actually pretty rare (used
124 // primarily for cleanup, or snapshotting data for display), so this lock has
125 // very little global performance impact.
127 // The above description tries to define the high performance (run time)
128 // portions of these classes. After gathering statistics, calls instigated
129 // by visiting about:profiler will assemble and aggregate data for display. The
130 // following data structures are used for producing such displays. They are
131 // not performance critical, and their only major constraint is that they should
132 // be able to run concurrently with ongoing augmentation of the birth and death
133 // data.
135 // This header also exports collection of classes that provide "snapshotted"
136 // representations of the core tracked_objects:: classes. These snapshotted
137 // representations are designed for safe transmission of the tracked_objects::
138 // data across process boundaries. Each consists of:
139 // (1) a default constructor, to support the IPC serialization macros,
140 // (2) a constructor that extracts data from the type being snapshotted, and
141 // (3) the snapshotted data.
143 // For a given birth location, information about births is spread across data
144 // structures that are asynchronously changing on various threads. For
145 // serialization and display purposes, we need to construct TaskSnapshot
146 // instances for each combination of birth thread, death thread, and location,
147 // along with the count of such lifetimes. We gather such data into a
148 // TaskSnapshot instances, so that such instances can be sorted and
149 // aggregated (and remain frozen during our processing).
151 // Profiling consists of phases. The concrete phase in the sequence of phases is
152 // identified by its 0-based index.
154 // The ProcessDataPhaseSnapshot struct is a serialized representation of the
155 // list of ThreadData objects for a process for a concrete profiling phase. It
156 // holds a set of TaskSnapshots and tracks parent/child relationships for the
157 // executed tasks. The statistics in a snapshot are gathered asynhcronously
158 // relative to their ongoing updates.
159 // It is possible, though highly unlikely, that stats could be incorrectly
160 // recorded by this process (all data is held in 32 bit ints, but we are not
161 // atomically collecting all data, so we could have count that does not, for
162 // example, match with the number of durations we accumulated). The advantage
163 // to having fast (non-atomic) updates of the data outweighs the minimal risk of
164 // a singular corrupt statistic snapshot (only the snapshot could be corrupt,
165 // not the underlying and ongoing statistic). In contrast, pointer data that
166 // is accessed during snapshotting is completely invariant, and hence is
167 // perfectly acquired (i.e., no potential corruption, and no risk of a bad
168 // memory reference).
170 // TODO(jar): We can implement a Snapshot system that *tries* to grab the
171 // snapshots on the source threads *when* they have MessageLoops available
172 // (worker threads don't have message loops generally, and hence gathering from
173 // them will continue to be asynchronous). We had an implementation of this in
174 // the past, but the difficulty is dealing with message loops being terminated.
175 // We can *try* to spam the available threads via some message loop proxy to
176 // achieve this feat, and it *might* be valuable when we are collecting data
177 // for upload via UMA (where correctness of data may be more significant than
178 // for a single screen of about:profiler).
180 // TODO(jar): We should support (optionally) the recording of parent-child
181 // relationships for tasks. This should be done by detecting what tasks are
182 // Born during the running of a parent task. The resulting data can be used by
183 // a smarter profiler to aggregate the cost of a series of child tasks into
184 // the ancestor task. It can also be used to illuminate what child or parent is
185 // related to each task.
187 // TODO(jar): We need to store DataCollections, and provide facilities for
188 // taking the difference between two gathered DataCollections. For now, we're
189 // just adding a hack that Reset()s to zero all counts and stats. This is also
190 // done in a slightly thread-unsafe fashion, as the resetting is done
191 // asynchronously relative to ongoing updates (but all data is 32 bit in size).
192 // For basic profiling, this will work "most of the time," and should be
193 // sufficient... but storing away DataCollections is the "right way" to do this.
194 // We'll accomplish this via JavaScript storage of snapshots, and then we'll
195 // remove the Reset() methods. We may also need a short-term-max value in
196 // DeathData that is reset (as synchronously as possible) during each snapshot.
197 // This will facilitate displaying a max value for each snapshot period.
199 namespace tracked_objects {
201 //------------------------------------------------------------------------------
202 // For a specific thread, and a specific birth place, the collection of all
203 // death info (with tallies for each death thread, to prevent access conflicts).
204 class ThreadData;
205 class BASE_EXPORT BirthOnThread {
206 public:
207 BirthOnThread(const Location& location, const ThreadData& current);
209 const Location location() const { return location_; }
210 const ThreadData* birth_thread() const { return birth_thread_; }
212 private:
213 // File/lineno of birth. This defines the essence of the task, as the context
214 // of the birth (construction) often tell what the item is for. This field
215 // is const, and hence safe to access from any thread.
216 const Location location_;
218 // The thread that records births into this object. Only this thread is
219 // allowed to update birth_count_ (which changes over time).
220 const ThreadData* const birth_thread_;
222 DISALLOW_COPY_AND_ASSIGN(BirthOnThread);
225 //------------------------------------------------------------------------------
226 // A "snapshotted" representation of the BirthOnThread class.
228 struct BASE_EXPORT BirthOnThreadSnapshot {
229 BirthOnThreadSnapshot();
230 explicit BirthOnThreadSnapshot(const BirthOnThread& birth);
231 ~BirthOnThreadSnapshot();
233 LocationSnapshot location;
234 std::string thread_name;
237 //------------------------------------------------------------------------------
238 // A class for accumulating counts of births (without bothering with a map<>).
240 class BASE_EXPORT Births: public BirthOnThread {
241 public:
242 Births(const Location& location, const ThreadData& current);
244 int birth_count() const;
246 // When we have a birth we update the count for this birthplace.
247 void RecordBirth();
249 private:
250 // The number of births on this thread for our location_.
251 int birth_count_;
253 DISALLOW_COPY_AND_ASSIGN(Births);
256 //------------------------------------------------------------------------------
257 // Basic info summarizing multiple destructions of a tracked object with a
258 // single birthplace (fixed Location). Used both on specific threads, and also
259 // in snapshots when integrating assembled data.
261 class BASE_EXPORT DeathData {
262 public:
263 // Default initializer.
264 DeathData();
266 // When deaths have not yet taken place, and we gather data from all the
267 // threads, we create DeathData stats that tally the number of births without
268 // a corresponding death.
269 explicit DeathData(int count);
271 // Update stats for a task destruction (death) that had a Run() time of
272 // |duration|, and has had a queueing delay of |queue_duration|.
273 void RecordDeath(const int32 queue_duration,
274 const int32 run_duration,
275 const uint32 random_number);
277 // Metrics accessors, used only for serialization and in tests.
278 int count() const;
279 int32 run_duration_sum() const;
280 int32 run_duration_max() const;
281 int32 run_duration_sample() const;
282 int32 queue_duration_sum() const;
283 int32 queue_duration_max() const;
284 int32 queue_duration_sample() const;
286 // Reset all tallies to zero. This is used as a hack on realtime data.
287 void Clear();
289 private:
290 // Members are ordered from most regularly read and updated, to least
291 // frequently used. This might help a bit with cache lines.
292 // Number of runs seen (divisor for calculating averages).
293 int count_;
294 // Basic tallies, used to compute averages.
295 int32 run_duration_sum_;
296 int32 queue_duration_sum_;
297 // Max values, used by local visualization routines. These are often read,
298 // but rarely updated.
299 int32 run_duration_max_;
300 int32 queue_duration_max_;
301 // Samples, used by crowd sourcing gatherers. These are almost never read,
302 // and rarely updated.
303 int32 run_duration_sample_;
304 int32 queue_duration_sample_;
307 //------------------------------------------------------------------------------
308 // A "snapshotted" representation of the DeathData class.
310 struct BASE_EXPORT DeathDataSnapshot {
311 DeathDataSnapshot();
312 explicit DeathDataSnapshot(const DeathData& death_data);
313 ~DeathDataSnapshot();
315 int count;
316 int32 run_duration_sum;
317 int32 run_duration_max;
318 int32 run_duration_sample;
319 int32 queue_duration_sum;
320 int32 queue_duration_max;
321 int32 queue_duration_sample;
324 //------------------------------------------------------------------------------
325 // A temporary collection of data that can be sorted and summarized. It is
326 // gathered (carefully) from many threads. Instances are held in arrays and
327 // processed, filtered, and rendered.
328 // The source of this data was collected on many threads, and is asynchronously
329 // changing. The data in this instance is not asynchronously changing.
331 struct BASE_EXPORT TaskSnapshot {
332 TaskSnapshot();
333 TaskSnapshot(const BirthOnThread& birth,
334 const DeathData& death_data,
335 const std::string& death_thread_name);
336 ~TaskSnapshot();
338 BirthOnThreadSnapshot birth;
339 DeathDataSnapshot death_data;
340 std::string death_thread_name;
343 //------------------------------------------------------------------------------
344 // For each thread, we have a ThreadData that stores all tracking info generated
345 // on this thread. This prevents the need for locking as data accumulates.
346 // We use ThreadLocalStorage to quickly identfy the current ThreadData context.
347 // We also have a linked list of ThreadData instances, and that list is used to
348 // harvest data from all existing instances.
350 struct ProcessDataPhaseSnapshot;
351 struct ProcessDataSnapshot;
352 class BASE_EXPORT TaskStopwatch;
354 // Map from profiling phase number to the process-wide snapshotted
355 // representation of the list of ThreadData objects that died during the given
356 // phase.
357 typedef std::map<int, ProcessDataPhaseSnapshot> PhasedProcessDataSnapshotMap;
359 class BASE_EXPORT ThreadData {
360 public:
361 // Current allowable states of the tracking system. The states can vary
362 // between ACTIVE and DEACTIVATED, but can never go back to UNINITIALIZED.
363 enum Status {
364 UNINITIALIZED, // PRistine, link-time state before running.
365 DORMANT_DURING_TESTS, // Only used during testing.
366 DEACTIVATED, // No longer recording profiling.
367 PROFILING_ACTIVE, // Recording profiles (no parent-child links).
368 PROFILING_CHILDREN_ACTIVE, // Fully active, recording parent-child links.
369 STATUS_LAST = PROFILING_CHILDREN_ACTIVE
372 typedef base::hash_map<Location, Births*, Location::Hash> BirthMap;
373 typedef std::map<const Births*, DeathData> DeathMap;
374 typedef std::pair<const Births*, const Births*> ParentChildPair;
375 typedef std::set<ParentChildPair> ParentChildSet;
376 typedef std::stack<const Births*> ParentStack;
378 // Initialize the current thread context with a new instance of ThreadData.
379 // This is used by all threads that have names, and should be explicitly
380 // set *before* any births on the threads have taken place. It is generally
381 // only used by the message loop, which has a well defined thread name.
382 static void InitializeThreadContext(const std::string& suggested_name);
384 // Using Thread Local Store, find the current instance for collecting data.
385 // If an instance does not exist, construct one (and remember it for use on
386 // this thread.
387 // This may return NULL if the system is disabled for any reason.
388 static ThreadData* Get();
390 // Fills |process_data_snapshot| with phased snapshots of all profiling
391 // phases, including the current one.
392 static void Snapshot(ProcessDataSnapshot* process_data_snapshot);
394 // Finds (or creates) a place to count births from the given location in this
395 // thread, and increment that tally.
396 // TallyABirthIfActive will returns NULL if the birth cannot be tallied.
397 static Births* TallyABirthIfActive(const Location& location);
399 // Records the end of a timed run of an object. The |completed_task| contains
400 // a pointer to a Births, the time_posted, and a delayed_start_time if any.
401 // The |start_of_run| indicates when we started to perform the run of the
402 // task. The delayed_start_time is non-null for tasks that were posted as
403 // delayed tasks, and it indicates when the task should have run (i.e., when
404 // it should have posted out of the timer queue, and into the work queue.
405 // The |end_of_run| was just obtained by a call to Now() (just after the task
406 // finished). It is provided as an argument to help with testing.
407 static void TallyRunOnNamedThreadIfTracking(
408 const base::TrackingInfo& completed_task,
409 const TaskStopwatch& stopwatch);
411 // Record the end of a timed run of an object. The |birth| is the record for
412 // the instance, the |time_posted| records that instant, which is presumed to
413 // be when the task was posted into a queue to run on a worker thread.
414 // The |start_of_run| is when the worker thread started to perform the run of
415 // the task.
416 // The |end_of_run| was just obtained by a call to Now() (just after the task
417 // finished).
418 static void TallyRunOnWorkerThreadIfTracking(const Births* birth,
419 const TrackedTime& time_posted,
420 const TaskStopwatch& stopwatch);
422 // Record the end of execution in region, generally corresponding to a scope
423 // being exited.
424 static void TallyRunInAScopedRegionIfTracking(const Births* birth,
425 const TaskStopwatch& stopwatch);
427 const std::string& thread_name() const { return thread_name_; }
429 // Initializes all statics if needed (this initialization call should be made
430 // while we are single threaded). Returns false if unable to initialize.
431 static bool Initialize();
433 // Sets internal status_.
434 // If |status| is false, then status_ is set to DEACTIVATED.
435 // If |status| is true, then status_ is set to, PROFILING_ACTIVE, or
436 // PROFILING_CHILDREN_ACTIVE.
437 // If tracking is not compiled in, this function will return false.
438 // If parent-child tracking is not compiled in, then an attempt to set the
439 // status to PROFILING_CHILDREN_ACTIVE will only result in a status of
440 // PROFILING_ACTIVE (i.e., it can't be set to a higher level than what is
441 // compiled into the binary, and parent-child tracking at the
442 // PROFILING_CHILDREN_ACTIVE level might not be compiled in).
443 static bool InitializeAndSetTrackingStatus(Status status);
445 static Status status();
447 // Indicate if any sort of profiling is being done (i.e., we are more than
448 // DEACTIVATED).
449 static bool TrackingStatus();
451 // For testing only, indicate if the status of parent-child tracking is turned
452 // on. This is currently a compiled option, atop TrackingStatus().
453 static bool TrackingParentChildStatus();
455 // Marks a start of a tracked run. It's super fast when tracking is disabled,
456 // and has some internal side effects when we are tracking, so that we can
457 // deduce the amount of time accumulated outside of execution of tracked runs.
458 // The task that will be tracked is passed in as |parent| so that parent-child
459 // relationships can be (optionally) calculated.
460 static void PrepareForStartOfRun(const Births* parent);
462 // Enables profiler timing.
463 static void EnableProfilerTiming();
465 // Provide a time function that does nothing (runs fast) when we don't have
466 // the profiler enabled. It will generally be optimized away when it is
467 // ifdef'ed to be small enough (allowing the profiler to be "compiled out" of
468 // the code).
469 static TrackedTime Now();
471 // Use the function |now| to provide current times, instead of calling the
472 // TrackedTime::Now() function. Since this alternate function is being used,
473 // the other time arguments (used for calculating queueing delay) will be
474 // ignored.
475 static void SetAlternateTimeSource(NowFunction* now);
477 // This function can be called at process termination to validate that thread
478 // cleanup routines have been called for at least some number of named
479 // threads.
480 static void EnsureCleanupWasCalled(int major_threads_shutdown_count);
482 private:
483 friend class TaskStopwatch;
484 // Allow only tests to call ShutdownSingleThreadedCleanup. We NEVER call it
485 // in production code.
486 // TODO(jar): Make this a friend in DEBUG only, so that the optimizer has a
487 // better change of optimizing (inlining? etc.) private methods (knowing that
488 // there will be no need for an external entry point).
489 friend class TrackedObjectsTest;
490 FRIEND_TEST_ALL_PREFIXES(TrackedObjectsTest, MinimalStartupShutdown);
491 FRIEND_TEST_ALL_PREFIXES(TrackedObjectsTest, TinyStartupShutdown);
492 FRIEND_TEST_ALL_PREFIXES(TrackedObjectsTest, ParentChildTest);
494 typedef std::map<const BirthOnThread*, int> BirthCountMap;
496 // Worker thread construction creates a name since there is none.
497 explicit ThreadData(int thread_number);
499 // Message loop based construction should provide a name.
500 explicit ThreadData(const std::string& suggested_name);
502 ~ThreadData();
504 // Push this instance to the head of all_thread_data_list_head_, linking it to
505 // the previous head. This is performed after each construction, and leaves
506 // the instance permanently on that list.
507 void PushToHeadOfList();
509 // (Thread safe) Get start of list of all ThreadData instances using the lock.
510 static ThreadData* first();
512 // Iterate through the null terminated list of ThreadData instances.
513 ThreadData* next() const;
516 // In this thread's data, record a new birth.
517 Births* TallyABirth(const Location& location);
519 // Find a place to record a death on this thread.
520 void TallyADeath(const Births& birth,
521 int32 queue_duration,
522 const TaskStopwatch& stopwatch);
524 // Snapshot (under a lock) the profiled data for the tasks in each ThreadData
525 // instance. Also updates the |birth_counts| tally for each task to keep
526 // track of the number of living instances of the task.
527 static void SnapshotAllExecutedTasks(
528 ProcessDataPhaseSnapshot* process_data_phase,
529 BirthCountMap* birth_counts);
531 // Fills |process_data_phase| with all the recursive results in our process.
532 static void SnapshotCurrentPhase(
533 ProcessDataPhaseSnapshot* process_data_phase);
535 // Snapshots (under a lock) the profiled data for the tasks for this thread
536 // and writes all of the executed tasks' data -- i.e. the data for the tasks
537 // with with entries in the death_map_ -- into |process_data_phase|. Also
538 // updates the |birth_counts| tally for each task to keep track of the number
539 // of living instances of the task -- that is, each task maps to the number of
540 // births for the task that have not yet been balanced by a death.
541 void SnapshotExecutedTasks(ProcessDataPhaseSnapshot* process_data_phase,
542 BirthCountMap* birth_counts);
544 // Using our lock, make a copy of the specified maps. This call may be made
545 // on non-local threads, which necessitate the use of the lock to prevent
546 // the map(s) from being reallocated while they are copied.
547 void SnapshotMaps(BirthMap* birth_map,
548 DeathMap* death_map,
549 ParentChildSet* parent_child_set);
551 // This method is called by the TLS system when a thread terminates.
552 // The argument may be NULL if this thread has never tracked a birth or death.
553 static void OnThreadTermination(void* thread_data);
555 // This method should be called when a worker thread terminates, so that we
556 // can save all the thread data into a cache of reusable ThreadData instances.
557 void OnThreadTerminationCleanup();
559 // Cleans up data structures, and returns statics to near pristine (mostly
560 // uninitialized) state. If there is any chance that other threads are still
561 // using the data structures, then the |leak| argument should be passed in as
562 // true, and the data structures (birth maps, death maps, ThreadData
563 // insntances, etc.) will be leaked and not deleted. If you have joined all
564 // threads since the time that InitializeAndSetTrackingStatus() was called,
565 // then you can pass in a |leak| value of false, and this function will
566 // delete recursively all data structures, starting with the list of
567 // ThreadData instances.
568 static void ShutdownSingleThreadedCleanup(bool leak);
570 // When non-null, this specifies an external function that supplies monotone
571 // increasing time functcion.
572 static NowFunction* now_function_;
574 // If true, now_function_ returns values that can be used to calculate queue
575 // time.
576 static bool now_function_is_time_;
578 // We use thread local store to identify which ThreadData to interact with.
579 static base::ThreadLocalStorage::StaticSlot tls_index_;
581 // List of ThreadData instances for use with worker threads. When a worker
582 // thread is done (terminated), we push it onto this list. When a new worker
583 // thread is created, we first try to re-use a ThreadData instance from the
584 // list, and if none are available, construct a new one.
585 // This is only accessed while list_lock_ is held.
586 static ThreadData* first_retired_worker_;
588 // Link to the most recently created instance (starts a null terminated list).
589 // The list is traversed by about:profiler when it needs to snapshot data.
590 // This is only accessed while list_lock_ is held.
591 static ThreadData* all_thread_data_list_head_;
593 // The next available worker thread number. This should only be accessed when
594 // the list_lock_ is held.
595 static int worker_thread_data_creation_count_;
597 // The number of times TLS has called us back to cleanup a ThreadData
598 // instance. This is only accessed while list_lock_ is held.
599 static int cleanup_count_;
601 // Incarnation sequence number, indicating how many times (during unittests)
602 // we've either transitioned out of UNINITIALIZED, or into that state. This
603 // value is only accessed while the list_lock_ is held.
604 static int incarnation_counter_;
606 // Protection for access to all_thread_data_list_head_, and to
607 // unregistered_thread_data_pool_. This lock is leaked at shutdown.
608 // The lock is very infrequently used, so we can afford to just make a lazy
609 // instance and be safe.
610 static base::LazyInstance<base::Lock>::Leaky list_lock_;
612 // We set status_ to SHUTDOWN when we shut down the tracking service.
613 static Status status_;
615 // Link to next instance (null terminated list). Used to globally track all
616 // registered instances (corresponds to all registered threads where we keep
617 // data).
618 ThreadData* next_;
620 // Pointer to another ThreadData instance for a Worker-Thread that has been
621 // retired (its thread was terminated). This value is non-NULL only for a
622 // retired ThreadData associated with a Worker-Thread.
623 ThreadData* next_retired_worker_;
625 // The name of the thread that is being recorded. If this thread has no
626 // message_loop, then this is a worker thread, with a sequence number postfix.
627 std::string thread_name_;
629 // Indicate if this is a worker thread, and the ThreadData contexts should be
630 // stored in the unregistered_thread_data_pool_ when not in use.
631 // Value is zero when it is not a worker thread. Value is a positive integer
632 // corresponding to the created thread name if it is a worker thread.
633 int worker_thread_number_;
635 // A map used on each thread to keep track of Births on this thread.
636 // This map should only be accessed on the thread it was constructed on.
637 // When a snapshot is needed, this structure can be locked in place for the
638 // duration of the snapshotting activity.
639 BirthMap birth_map_;
641 // Similar to birth_map_, this records informations about death of tracked
642 // instances (i.e., when a tracked instance was destroyed on this thread).
643 // It is locked before changing, and hence other threads may access it by
644 // locking before reading it.
645 DeathMap death_map_;
647 // A set of parents that created children tasks on this thread. Each pair
648 // corresponds to potentially non-local Births (location and thread), and a
649 // local Births (that took place on this thread).
650 ParentChildSet parent_child_set_;
652 // Lock to protect *some* access to BirthMap and DeathMap. The maps are
653 // regularly read and written on this thread, but may only be read from other
654 // threads. To support this, we acquire this lock if we are writing from this
655 // thread, or reading from another thread. For reading from this thread we
656 // don't need a lock, as there is no potential for a conflict since the
657 // writing is only done from this thread.
658 mutable base::Lock map_lock_;
660 // The stack of parents that are currently being profiled. This includes only
661 // tasks that have started a timer recently via PrepareForStartOfRun(), but
662 // not yet concluded with a NowForEndOfRun(). Usually this stack is one deep,
663 // but if a scoped region is profiled, or <sigh> a task runs a nested-message
664 // loop, then the stack can grow larger. Note that we don't try to deduct
665 // time in nested profiles, as our current timer is based on wall-clock time,
666 // and not CPU time (and we're hopeful that nested timing won't be a
667 // significant additional cost).
668 ParentStack parent_stack_;
670 // A random number that we used to select decide which sample to keep as a
671 // representative sample in each DeathData instance. We can't start off with
672 // much randomness (because we can't call RandInt() on all our threads), so
673 // we stir in more and more as we go.
674 uint32 random_number_;
676 // Record of what the incarnation_counter_ was when this instance was created.
677 // If the incarnation_counter_ has changed, then we avoid pushing into the
678 // pool (this is only critical in tests which go through multiple
679 // incarnations).
680 int incarnation_count_for_pool_;
682 // Most recently started (i.e. most nested) stopwatch on the current thread,
683 // if it exists; NULL otherwise.
684 TaskStopwatch* current_stopwatch_;
686 DISALLOW_COPY_AND_ASSIGN(ThreadData);
689 //------------------------------------------------------------------------------
690 // Stopwatch to measure task run time or simply create a time interval that will
691 // be subtracted from the current most nested task's run time. Stopwatches
692 // coordinate with the stopwatches in which they are nested to avoid
693 // double-counting nested tasks run times.
695 class BASE_EXPORT TaskStopwatch {
696 public:
697 // Starts the stopwatch.
698 TaskStopwatch();
699 ~TaskStopwatch();
701 // Starts stopwatch.
702 void Start();
704 // Stops stopwatch.
705 void Stop();
707 // Returns the start time.
708 TrackedTime StartTime() const;
710 // Task's duration is calculated as the wallclock duration between starting
711 // and stopping this stopwatch, minus the wallclock durations of any other
712 // instances that are immediately nested in this one, started and stopped on
713 // this thread during that period.
714 int32 RunDurationMs() const;
716 // Returns tracking info for the current thread.
717 ThreadData* GetThreadData() const;
719 private:
720 // Time when the stopwatch was started.
721 TrackedTime start_time_;
723 // Wallclock duration of the task.
724 int32 wallclock_duration_ms_;
726 // Tracking info for the current thread.
727 ThreadData* current_thread_data_;
729 // Sum of wallclock durations of all stopwatches that were directly nested in
730 // this one.
731 int32 excluded_duration_ms_;
733 // Stopwatch which was running on our thread when this stopwatch was started.
734 // That preexisting stopwatch must be adjusted to the exclude the wallclock
735 // duration of this stopwatch.
736 TaskStopwatch* parent_;
738 #if DCHECK_IS_ON()
739 // State of the stopwatch. Stopwatch is first constructed in a created state
740 // state, then is optionally started/stopped, then destructed.
741 enum { CREATED, RUNNING, STOPPED } state_;
743 // Currently running stopwatch that is directly nested in this one, if such
744 // stopwatch exists. NULL otherwise.
745 TaskStopwatch* child_;
746 #endif
749 //------------------------------------------------------------------------------
750 // A snapshotted representation of a (parent, child) task pair, for tracking
751 // hierarchical profiles.
753 struct BASE_EXPORT ParentChildPairSnapshot {
754 public:
755 ParentChildPairSnapshot();
756 explicit ParentChildPairSnapshot(
757 const ThreadData::ParentChildPair& parent_child);
758 ~ParentChildPairSnapshot();
760 BirthOnThreadSnapshot parent;
761 BirthOnThreadSnapshot child;
764 //------------------------------------------------------------------------------
765 // A snapshotted representation of the list of ThreadData objects for a process,
766 // for a single profiling phase.
768 struct BASE_EXPORT ProcessDataPhaseSnapshot {
769 public:
770 ProcessDataPhaseSnapshot();
771 ~ProcessDataPhaseSnapshot();
773 std::vector<TaskSnapshot> tasks;
774 std::vector<ParentChildPairSnapshot> descendants;
777 //------------------------------------------------------------------------------
778 // A snapshotted representation of the list of ThreadData objects for a process,
779 // for all profiling phases, including the current one.
781 struct BASE_EXPORT ProcessDataSnapshot {
782 public:
783 ProcessDataSnapshot();
784 ~ProcessDataSnapshot();
786 PhasedProcessDataSnapshotMap phased_process_data_snapshots;
787 base::ProcessId process_id;
790 } // namespace tracked_objects
792 #endif // BASE_TRACKED_OBJECTS_H_