// Copyright 2015 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_ #define BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_ #include #include "base/atomicops.h" #include "base/base_export.h" #include "base/macros.h" #include "base/trace_event/heap_profiler_allocation_context.h" namespace base { namespace trace_event { // AllocationContextTracker is a thread-local object. Its main purpose is to // keep track of a pseudo stack of trace events. Chrome has been instrumented // with lots of `TRACE_EVENT` macros. These trace events push their name to a // thread-local stack when they go into scope, and pop when they go out of // scope, if all of the following conditions have been met: // // * A trace is being recorded. // * The category of the event is enabled in the trace config. // * Heap profiling is enabled (with the `--enable-heap-profiling` flag). // // This means that allocations that occur before tracing is started will not // have backtrace information in their context. // // AllocationContextTracker also keeps track of some thread state not related to // trace events. See |AllocationContext|. // // A thread-local instance of the context tracker is initialized lazily when it // is first accessed. This might be because a trace event pushed or popped, or // because `GetContextSnapshot()` was called when an allocation occurred class BASE_EXPORT AllocationContextTracker { public: enum class CaptureMode : int32_t { DISABLED, // Don't capture anything PSEUDO_STACK, // Backtrace has trace events MIXED_STACK, // Backtrace has trace events + from // HeapProfilerScopedStackFrame NATIVE_STACK, // Backtrace has full native backtraces from stack unwinding }; // Stack frame constructed from trace events in codebase. struct BASE_EXPORT PseudoStackFrame { const char* trace_event_category; const char* trace_event_name; bool operator==(const PseudoStackFrame& other) const { return trace_event_category == other.trace_event_category && trace_event_name == other.trace_event_name; } }; // Globally sets capturing mode. // TODO(primiano): How to guard against *_STACK -> DISABLED -> *_STACK? static void SetCaptureMode(CaptureMode mode); // Returns global capturing mode. inline static CaptureMode capture_mode() { // A little lag after heap profiling is enabled or disabled is fine, it is // more important that the check is as cheap as possible when capturing is // not enabled, so do not issue a memory barrier in the fast path. if (subtle::NoBarrier_Load(&capture_mode_) == static_cast(CaptureMode::DISABLED)) return CaptureMode::DISABLED; // In the slow path, an acquire load is required to pair with the release // store in |SetCaptureMode|. This is to ensure that the TLS slot for // the thread-local allocation context tracker has been initialized if // |capture_mode| returns something other than DISABLED. return static_cast(subtle::Acquire_Load(&capture_mode_)); } // Returns the thread-local instance, creating one if necessary. Returns // always a valid instance, unless it is called re-entrantly, in which case // returns nullptr in the nested calls. static AllocationContextTracker* GetInstanceForCurrentThread(); // Set the thread name in the AllocationContextTracker of the current thread // if capture is enabled. static void SetCurrentThreadName(const char* name); // Starts and ends a new ignore scope between which the allocations are // ignored by the heap profiler. GetContextSnapshot() returns false when // allocations are ignored. void begin_ignore_scope() { ignore_scope_depth_++; } void end_ignore_scope() { if (ignore_scope_depth_) ignore_scope_depth_--; } // Pushes and pops a frame onto the thread-local pseudo stack. // TODO(ssid): Change PseudoStackFrame to const char*. Only event name is // used. void PushPseudoStackFrame(PseudoStackFrame stack_frame); void PopPseudoStackFrame(PseudoStackFrame stack_frame); // Pushes and pops a native stack frame onto thread local tracked stack. void PushNativeStackFrame(const void* pc); void PopNativeStackFrame(const void* pc); // Push and pop current task's context. A stack is used to support nested // tasks and the top of the stack will be used in allocation context. void PushCurrentTaskContext(const char* context); void PopCurrentTaskContext(const char* context); // Fills a snapshot of the current thread-local context. Doesn't fill and // returns false if allocations are being ignored. bool GetContextSnapshot(AllocationContext* snapshot); ~AllocationContextTracker(); private: AllocationContextTracker(); static subtle::Atomic32 capture_mode_; // The pseudo stack where frames are |TRACE_EVENT| names or inserted PCs. std::vector tracked_stack_; // The thread name is used as the first entry in the pseudo stack. const char* thread_name_; // Stack of tasks' contexts. Context serves as a different dimension than // pseudo stack to cluster allocations. std::vector task_contexts_; uint32_t ignore_scope_depth_; DISALLOW_COPY_AND_ASSIGN(AllocationContextTracker); }; } // namespace trace_event } // namespace base #endif // BASE_TRACE_EVENT_HEAP_PROFILER_ALLOCATION_CONTEXT_TRACKER_H_