Retro68/gcc/libsanitizer/lsan/lsan_common.h

264 lines
8.6 KiB
C
Raw Normal View History

2014-09-21 17:33:12 +00:00
//=-- lsan_common.h -------------------------------------------------------===//
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
//
// This file is a part of LeakSanitizer.
// Private LSan header.
//
//===----------------------------------------------------------------------===//
#ifndef LSAN_COMMON_H
#define LSAN_COMMON_H
#include "sanitizer_common/sanitizer_allocator.h"
#include "sanitizer_common/sanitizer_common.h"
#include "sanitizer_common/sanitizer_internal_defs.h"
#include "sanitizer_common/sanitizer_platform.h"
2017-04-10 11:32:00 +00:00
#include "sanitizer_common/sanitizer_stoptheworld.h"
2014-09-21 17:33:12 +00:00
#include "sanitizer_common/sanitizer_symbolizer.h"
2018-12-28 15:30:48 +00:00
// LeakSanitizer relies on some Glibc's internals (e.g. TLS machinery) thus
// supported for Linux only. Also, LSan doesn't like 32 bit architectures
// because of "small" (4 bytes) pointer size that leads to high false negative
// ratio on large leaks. But we still want to have it for some 32 bit arches
// (e.g. x86), see https://github.com/google/sanitizers/issues/403.
2019-06-02 15:48:37 +00:00
// To enable LeakSanitizer on a new architecture, one needs to implement the
// internal_clone function as well as (probably) adjust the TLS machinery for
// the new architecture inside the sanitizer library.
2018-12-28 15:30:48 +00:00
#if (SANITIZER_LINUX && !SANITIZER_ANDROID || SANITIZER_MAC) && \
(SANITIZER_WORDSIZE == 64) && \
(defined(__x86_64__) || defined(__mips64) || defined(__aarch64__) || \
defined(__powerpc64__))
#define CAN_SANITIZE_LEAKS 1
#elif defined(__i386__) && \
(SANITIZER_LINUX && !SANITIZER_ANDROID || SANITIZER_MAC)
#define CAN_SANITIZE_LEAKS 1
#elif defined(__arm__) && \
SANITIZER_LINUX && !SANITIZER_ANDROID
2014-09-21 17:33:12 +00:00
#define CAN_SANITIZE_LEAKS 1
#else
#define CAN_SANITIZE_LEAKS 0
#endif
2017-04-10 11:32:00 +00:00
namespace __sanitizer {
class FlagParser;
2019-06-02 15:48:37 +00:00
class ThreadRegistry;
struct DTLS;
2017-04-10 11:32:00 +00:00
}
2014-09-21 17:33:12 +00:00
namespace __lsan {
// Chunk tags.
enum ChunkTag {
kDirectlyLeaked = 0, // default
kIndirectlyLeaked = 1,
kReachable = 2,
kIgnored = 3
};
2018-12-28 15:30:48 +00:00
const u32 kInvalidTid = (u32) -1;
2014-09-21 17:33:12 +00:00
struct Flags {
2017-04-10 11:32:00 +00:00
#define LSAN_FLAG(Type, Name, DefaultValue, Description) Type Name;
#include "lsan_flags.inc"
#undef LSAN_FLAG
void SetDefaults();
2014-09-21 17:33:12 +00:00
uptr pointer_alignment() const {
return use_unaligned ? 1 : sizeof(uptr);
}
};
extern Flags lsan_flags;
inline Flags *flags() { return &lsan_flags; }
2017-04-10 11:32:00 +00:00
void RegisterLsanFlags(FlagParser *parser, Flags *f);
2014-09-21 17:33:12 +00:00
struct Leak {
2015-08-28 15:33:40 +00:00
u32 id;
2014-09-21 17:33:12 +00:00
uptr hit_count;
uptr total_size;
u32 stack_trace_id;
bool is_directly_leaked;
bool is_suppressed;
};
2015-08-28 15:33:40 +00:00
struct LeakedObject {
u32 leak_id;
uptr addr;
uptr size;
};
2014-09-21 17:33:12 +00:00
// Aggregates leaks by stack trace prefix.
class LeakReport {
public:
2019-06-02 15:48:37 +00:00
LeakReport() {}
2015-08-28 15:33:40 +00:00
void AddLeakedChunk(uptr chunk, u32 stack_trace_id, uptr leaked_size,
ChunkTag tag);
void ReportTopLeaks(uptr max_leaks);
2014-09-21 17:33:12 +00:00
void PrintSummary();
2015-08-28 15:33:40 +00:00
void ApplySuppressions();
uptr UnsuppressedLeakCount();
2014-09-21 17:33:12 +00:00
private:
2015-08-28 15:33:40 +00:00
void PrintReportForLeak(uptr index);
void PrintLeakedObjectsForLeak(uptr index);
2019-06-02 15:48:37 +00:00
u32 next_id_ = 0;
2014-09-21 17:33:12 +00:00
InternalMmapVector<Leak> leaks_;
2015-08-28 15:33:40 +00:00
InternalMmapVector<LeakedObject> leaked_objects_;
2014-09-21 17:33:12 +00:00
};
typedef InternalMmapVector<uptr> Frontier;
// Platform-specific functions.
void InitializePlatformSpecificModules();
void ProcessGlobalRegions(Frontier *frontier);
void ProcessPlatformSpecificAllocations(Frontier *frontier);
2018-12-28 15:30:48 +00:00
struct RootRegion {
uptr begin;
uptr size;
};
InternalMmapVector<RootRegion> const *GetRootRegions();
void ScanRootRegion(Frontier *frontier, RootRegion const &region,
uptr region_begin, uptr region_end, bool is_readable);
2017-04-10 11:32:00 +00:00
// Run stoptheworld while holding any platform-specific locks.
void DoStopTheWorld(StopTheWorldCallback callback, void* argument);
2014-09-21 17:33:12 +00:00
void ScanRangeForPointers(uptr begin, uptr end,
Frontier *frontier,
const char *region_type, ChunkTag tag);
2018-12-28 15:30:48 +00:00
void ScanGlobalRange(uptr begin, uptr end, Frontier *frontier);
2014-09-21 17:33:12 +00:00
enum IgnoreObjectResult {
kIgnoreObjectSuccess,
kIgnoreObjectAlreadyIgnored,
kIgnoreObjectInvalid
};
// Functions called from the parent tool.
2018-12-28 15:30:48 +00:00
const char *MaybeCallLsanDefaultOptions();
2017-04-10 11:32:00 +00:00
void InitCommonLsan();
2014-09-21 17:33:12 +00:00
void DoLeakCheck();
2018-12-28 15:30:48 +00:00
void DoRecoverableLeakCheckVoid();
void DisableCounterUnderflow();
2014-09-21 17:33:12 +00:00
bool DisabledInThisThread();
// Used to implement __lsan::ScopedDisabler.
void DisableInThisThread();
void EnableInThisThread();
// Can be used to ignore memory allocated by an intercepted
// function.
struct ScopedInterceptorDisabler {
ScopedInterceptorDisabler() { DisableInThisThread(); }
~ScopedInterceptorDisabler() { EnableInThisThread(); }
};
2018-12-28 15:30:48 +00:00
// According to Itanium C++ ABI array cookie is a one word containing
// size of allocated array.
static inline bool IsItaniumABIArrayCookie(uptr chunk_beg, uptr chunk_size,
uptr addr) {
return chunk_size == sizeof(uptr) && chunk_beg + chunk_size == addr &&
*reinterpret_cast<uptr *>(chunk_beg) == 0;
}
// According to ARM C++ ABI array cookie consists of two words:
// struct array_cookie {
// std::size_t element_size; // element_size != 0
// std::size_t element_count;
// };
static inline bool IsARMABIArrayCookie(uptr chunk_beg, uptr chunk_size,
uptr addr) {
return chunk_size == 2 * sizeof(uptr) && chunk_beg + chunk_size == addr &&
*reinterpret_cast<uptr *>(chunk_beg + sizeof(uptr)) == 0;
}
2015-08-28 15:33:40 +00:00
// Special case for "new T[0]" where T is a type with DTOR.
2018-12-28 15:30:48 +00:00
// new T[0] will allocate a cookie (one or two words) for the array size (0)
// and store a pointer to the end of allocated chunk. The actual cookie layout
// varies between platforms according to their C++ ABI implementation.
2015-08-28 15:33:40 +00:00
inline bool IsSpecialCaseOfOperatorNew0(uptr chunk_beg, uptr chunk_size,
uptr addr) {
2018-12-28 15:30:48 +00:00
#if defined(__arm__)
return IsARMABIArrayCookie(chunk_beg, chunk_size, addr);
#else
return IsItaniumABIArrayCookie(chunk_beg, chunk_size, addr);
#endif
2015-08-28 15:33:40 +00:00
}
2014-09-21 17:33:12 +00:00
// The following must be implemented in the parent tool.
void ForEachChunk(ForEachChunkCallback callback, void *arg);
// Returns the address range occupied by the global allocator object.
void GetAllocatorGlobalRange(uptr *begin, uptr *end);
// Wrappers for allocator's ForceLock()/ForceUnlock().
void LockAllocator();
void UnlockAllocator();
2015-08-28 15:33:40 +00:00
// Returns true if [addr, addr + sizeof(void *)) is poisoned.
bool WordIsPoisoned(uptr addr);
2014-09-21 17:33:12 +00:00
// Wrappers for ThreadRegistry access.
void LockThreadRegistry();
void UnlockThreadRegistry();
2019-06-02 15:48:37 +00:00
ThreadRegistry *GetThreadRegistryLocked();
2018-12-28 15:30:48 +00:00
bool GetThreadRangesLocked(tid_t os_id, uptr *stack_begin, uptr *stack_end,
uptr *tls_begin, uptr *tls_end, uptr *cache_begin,
uptr *cache_end, DTLS **dtls);
2018-12-28 15:30:48 +00:00
void ForEachExtraStackRange(tid_t os_id, RangeIteratorCallback callback,
2014-09-21 17:33:12 +00:00
void *arg);
// If called from the main thread, updates the main thread's TID in the thread
// registry. We need this to handle processes that fork() without a subsequent
// exec(), which invalidates the recorded TID. To update it, we must call
// gettid() from the main thread. Our solution is to call this function before
// leak checking and also before every call to pthread_create() (to handle cases
// where leak checking is initiated from a non-main thread).
void EnsureMainThreadIDIsCorrect();
// If p points into a chunk that has been allocated to the user, returns its
// user-visible address. Otherwise, returns 0.
uptr PointsIntoChunk(void *p);
// Returns address of user-visible chunk contained in this allocator chunk.
uptr GetUserBegin(uptr chunk);
// Helper for __lsan_ignore_object().
IgnoreObjectResult IgnoreObjectLocked(const void *p);
2018-12-28 15:30:48 +00:00
// Return the linker module, if valid for the platform.
LoadedModule *GetLinker();
// Return true if LSan has finished leak checking and reported leaks.
bool HasReportedLeaks();
// Run platform-specific leak handlers.
void HandleLeaks();
2014-09-21 17:33:12 +00:00
// Wrapper for chunk metadata operations.
class LsanMetadata {
public:
// Constructor accepts address of user-visible chunk.
explicit LsanMetadata(uptr chunk);
bool allocated() const;
ChunkTag tag() const;
void set_tag(ChunkTag value);
uptr requested_size() const;
u32 stack_trace_id() const;
private:
void *metadata_;
};
} // namespace __lsan
extern "C" {
2018-12-28 15:30:48 +00:00
SANITIZER_INTERFACE_ATTRIBUTE SANITIZER_WEAK_ATTRIBUTE
const char *__lsan_default_options();
2014-09-21 17:33:12 +00:00
SANITIZER_INTERFACE_ATTRIBUTE SANITIZER_WEAK_ATTRIBUTE
int __lsan_is_turned_off();
SANITIZER_INTERFACE_ATTRIBUTE SANITIZER_WEAK_ATTRIBUTE
const char *__lsan_default_suppressions();
} // extern "C"
#endif // LSAN_COMMON_H