diff options
Diffstat (limited to 'base/trace_event/heap_profiler_allocation_context_tracker.h')
-rw-r--r-- | base/trace_event/heap_profiler_allocation_context_tracker.h | 140 |
1 files changed, 140 insertions, 0 deletions
diff --git a/base/trace_event/heap_profiler_allocation_context_tracker.h b/base/trace_event/heap_profiler_allocation_context_tracker.h new file mode 100644 index 0000000000..da03b7f6d6 --- /dev/null +++ b/base/trace_event/heap_profiler_allocation_context_tracker.h @@ -0,0 +1,140 @@ +// Copyright 2015 The Chromium Authors. All rights reserved. +// Use of this source code is governed by a BSD-style license that can be +// found in the LICENSE file. + +#ifndef BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_ +#define BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_ + +#include <vector> + +#include "base/atomicops.h" +#include "base/base_export.h" +#include "base/macros.h" +#include "base/trace_event/heap_profiler_allocation_context.h" + +namespace base { +namespace trace_event { + +// AllocationContextTracker is a thread-local object. Its main purpose is to +// keep track of a pseudo stack of trace events. Chrome has been instrumented +// with lots of `TRACE_EVENT` macros. These trace events push their name to a +// thread-local stack when they go into scope, and pop when they go out of +// scope, if all of the following conditions have been met: +// +// * A trace is being recorded. +// * The category of the event is enabled in the trace config. +// * Heap profiling is enabled (with the `--enable-heap-profiling` flag). +// +// This means that allocations that occur before tracing is started will not +// have backtrace information in their context. +// +// AllocationContextTracker also keeps track of some thread state not related to +// trace events. See |AllocationContext|. +// +// A thread-local instance of the context tracker is initialized lazily when it +// is first accessed. This might be because a trace event pushed or popped, or +// because `GetContextSnapshot()` was called when an allocation occurred +class BASE_EXPORT AllocationContextTracker { + public: + enum class CaptureMode : int32_t { + DISABLED, // Don't capture anything + PSEUDO_STACK, // Backtrace has trace events + MIXED_STACK, // Backtrace has trace events + from + // HeapProfilerScopedStackFrame + NATIVE_STACK, // Backtrace has full native backtraces from stack unwinding + }; + + // Stack frame constructed from trace events in codebase. + struct BASE_EXPORT PseudoStackFrame { + const char* trace_event_category; + const char* trace_event_name; + + bool operator==(const PseudoStackFrame& other) const { + return trace_event_category == other.trace_event_category && + trace_event_name == other.trace_event_name; + } + }; + + // Globally sets capturing mode. + // TODO(primiano): How to guard against *_STACK -> DISABLED -> *_STACK? + static void SetCaptureMode(CaptureMode mode); + + // Returns global capturing mode. + inline static CaptureMode capture_mode() { + // A little lag after heap profiling is enabled or disabled is fine, it is + // more important that the check is as cheap as possible when capturing is + // not enabled, so do not issue a memory barrier in the fast path. + if (subtle::NoBarrier_Load(&capture_mode_) == + static_cast<int32_t>(CaptureMode::DISABLED)) + return CaptureMode::DISABLED; + + // In the slow path, an acquire load is required to pair with the release + // store in |SetCaptureMode|. This is to ensure that the TLS slot for + // the thread-local allocation context tracker has been initialized if + // |capture_mode| returns something other than DISABLED. + return static_cast<CaptureMode>(subtle::Acquire_Load(&capture_mode_)); + } + + // Returns the thread-local instance, creating one if necessary. Returns + // always a valid instance, unless it is called re-entrantly, in which case + // returns nullptr in the nested calls. + static AllocationContextTracker* GetInstanceForCurrentThread(); + + // Set the thread name in the AllocationContextTracker of the current thread + // if capture is enabled. + static void SetCurrentThreadName(const char* name); + + // Starts and ends a new ignore scope between which the allocations are + // ignored by the heap profiler. GetContextSnapshot() returns false when + // allocations are ignored. + void begin_ignore_scope() { ignore_scope_depth_++; } + void end_ignore_scope() { + if (ignore_scope_depth_) + ignore_scope_depth_--; + } + + // Pushes and pops a frame onto the thread-local pseudo stack. + // TODO(ssid): Change PseudoStackFrame to const char*. Only event name is + // used. + void PushPseudoStackFrame(PseudoStackFrame stack_frame); + void PopPseudoStackFrame(PseudoStackFrame stack_frame); + + // Pushes and pops a native stack frame onto thread local tracked stack. + void PushNativeStackFrame(const void* pc); + void PopNativeStackFrame(const void* pc); + + // Push and pop current task's context. A stack is used to support nested + // tasks and the top of the stack will be used in allocation context. + void PushCurrentTaskContext(const char* context); + void PopCurrentTaskContext(const char* context); + + // Fills a snapshot of the current thread-local context. Doesn't fill and + // returns false if allocations are being ignored. + bool GetContextSnapshot(AllocationContext* snapshot); + + ~AllocationContextTracker(); + + private: + AllocationContextTracker(); + + static subtle::Atomic32 capture_mode_; + + // The pseudo stack where frames are |TRACE_EVENT| names or inserted PCs. + std::vector<StackFrame> tracked_stack_; + + // The thread name is used as the first entry in the pseudo stack. + const char* thread_name_; + + // Stack of tasks' contexts. Context serves as a different dimension than + // pseudo stack to cluster allocations. + std::vector<const char*> task_contexts_; + + uint32_t ignore_scope_depth_; + + DISALLOW_COPY_AND_ASSIGN(AllocationContextTracker); +}; + +} // namespace trace_event +} // namespace base + +#endif // BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_ |