
Instead of writing out in native endian, write out the raw profile bytes in little endian. Also update the MIB data in little endian. Also clean up some lint and unused includes in rawprofile.cpp.
298 lines
9.4 KiB
C++
298 lines
9.4 KiB
C++
#include <stdint.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#include "memprof_rawprofile.h"
|
|
#include "profile/MemProfData.inc"
|
|
#include "sanitizer_common/sanitizer_allocator_internal.h"
|
|
#include "sanitizer_common/sanitizer_array_ref.h"
|
|
#include "sanitizer_common/sanitizer_common.h"
|
|
#include "sanitizer_common/sanitizer_stackdepot.h"
|
|
#include "sanitizer_common/sanitizer_stacktrace.h"
|
|
#include "sanitizer_common/sanitizer_vector.h"
|
|
|
|
namespace __memprof {
|
|
using ::__sanitizer::Vector;
|
|
using ::llvm::memprof::MemInfoBlock;
|
|
using SegmentEntry = ::llvm::memprof::SegmentEntry;
|
|
using Header = ::llvm::memprof::Header;
|
|
using ::llvm::memprof::encodeHistogramCount;
|
|
|
|
namespace {
|
|
template <class T> char *WriteBytes(const T &Pod, char *Buffer) {
|
|
static_assert(is_trivially_copyable<T>::value, "T must be POD");
|
|
const uint8_t *Src = reinterpret_cast<const uint8_t *>(&Pod);
|
|
|
|
for (size_t I = 0; I < sizeof(T); ++I)
|
|
#if defined(__BYTE_ORDER__) && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
|
|
// Reverse byte order since reader is little-endian.
|
|
Buffer[I] = Src[sizeof(T) - 1 - I];
|
|
#else
|
|
Buffer[I] = Src[I];
|
|
#endif
|
|
return Buffer + sizeof(T);
|
|
}
|
|
|
|
void RecordStackId(const uptr Key, UNUSED LockedMemInfoBlock *const &MIB,
|
|
void *Arg) {
|
|
// No need to touch the MIB value here since we are only recording the key.
|
|
auto *StackIds = reinterpret_cast<Vector<u64> *>(Arg);
|
|
StackIds->PushBack(Key);
|
|
}
|
|
|
|
u64 SegmentSizeBytes(ArrayRef<LoadedModule> Modules) {
|
|
u64 NumSegmentsToRecord = 0;
|
|
for (const auto &Module : Modules) {
|
|
for (const auto &Segment : Module.ranges()) {
|
|
if (Segment.executable)
|
|
NumSegmentsToRecord++;
|
|
}
|
|
}
|
|
|
|
return sizeof(u64) // A header which stores the number of records.
|
|
+ sizeof(SegmentEntry) * NumSegmentsToRecord;
|
|
}
|
|
|
|
// The segment section uses the following format:
|
|
// ---------- Segment Info
|
|
// Num Entries
|
|
// ---------- Segment Entry
|
|
// Start
|
|
// End
|
|
// Offset
|
|
// UuidSize
|
|
// Uuid 32B
|
|
// ----------
|
|
// ...
|
|
void SerializeSegmentsToBuffer(ArrayRef<LoadedModule> Modules,
|
|
const u64 ExpectedNumBytes, char *&Buffer) {
|
|
char *Ptr = Buffer;
|
|
// Reserve space for the final count.
|
|
Ptr += sizeof(u64);
|
|
|
|
u64 NumSegmentsRecorded = 0;
|
|
|
|
for (const auto &Module : Modules) {
|
|
for (const auto &Segment : Module.ranges()) {
|
|
if (Segment.executable) {
|
|
SegmentEntry Entry(Segment.beg, Segment.end, Module.base_address());
|
|
CHECK(Module.uuid_size() <= MEMPROF_BUILDID_MAX_SIZE);
|
|
Entry.BuildIdSize = Module.uuid_size();
|
|
memcpy(Entry.BuildId, Module.uuid(), Module.uuid_size());
|
|
memcpy(Ptr, &Entry, sizeof(SegmentEntry));
|
|
Ptr += sizeof(SegmentEntry);
|
|
NumSegmentsRecorded++;
|
|
}
|
|
}
|
|
}
|
|
// Store the number of segments we recorded in the space we reserved.
|
|
*((u64 *)Buffer) = NumSegmentsRecorded;
|
|
CHECK(ExpectedNumBytes >= static_cast<u64>(Ptr - Buffer) &&
|
|
"Expected num bytes != actual bytes written");
|
|
}
|
|
|
|
u64 StackSizeBytes(const Vector<u64> &StackIds) {
|
|
u64 NumBytesToWrite = sizeof(u64);
|
|
|
|
const u64 NumIds = StackIds.Size();
|
|
for (unsigned k = 0; k < NumIds; ++k) {
|
|
const u64 Id = StackIds[k];
|
|
// One entry for the id and then one more for the number of stack pcs.
|
|
NumBytesToWrite += 2 * sizeof(u64);
|
|
const StackTrace St = StackDepotGet(Id);
|
|
|
|
CHECK(St.trace != nullptr && St.size > 0 && "Empty stack trace");
|
|
for (uptr i = 0; i < St.size && St.trace[i] != 0; i++) {
|
|
NumBytesToWrite += sizeof(u64);
|
|
}
|
|
}
|
|
return NumBytesToWrite;
|
|
}
|
|
|
|
// The stack info section uses the following format:
|
|
//
|
|
// ---------- Stack Info
|
|
// Num Entries
|
|
// ---------- Stack Entry
|
|
// Num Stacks
|
|
// PC1
|
|
// PC2
|
|
// ...
|
|
// ----------
|
|
void SerializeStackToBuffer(const Vector<u64> &StackIds,
|
|
const u64 ExpectedNumBytes, char *&Buffer) {
|
|
const u64 NumIds = StackIds.Size();
|
|
char *Ptr = Buffer;
|
|
Ptr = WriteBytes(static_cast<u64>(NumIds), Ptr);
|
|
|
|
for (unsigned k = 0; k < NumIds; ++k) {
|
|
const u64 Id = StackIds[k];
|
|
Ptr = WriteBytes(Id, Ptr);
|
|
Ptr += sizeof(u64); // Bump it by u64, we will fill this in later.
|
|
u64 Count = 0;
|
|
const StackTrace St = StackDepotGet(Id);
|
|
for (uptr i = 0; i < St.size && St.trace[i] != 0; i++) {
|
|
// PCs in stack traces are actually the return addresses, that is,
|
|
// addresses of the next instructions after the call.
|
|
uptr pc = StackTrace::GetPreviousInstructionPc(St.trace[i]);
|
|
Ptr = WriteBytes(static_cast<u64>(pc), Ptr);
|
|
++Count;
|
|
}
|
|
// Store the count in the space we reserved earlier.
|
|
*(u64 *)(Ptr - (Count + 1) * sizeof(u64)) = Count;
|
|
}
|
|
|
|
CHECK(ExpectedNumBytes >= static_cast<u64>(Ptr - Buffer) &&
|
|
"Expected num bytes != actual bytes written");
|
|
}
|
|
|
|
// The MIB section has the following format:
|
|
// ---------- MIB Info
|
|
// Num Entries
|
|
// ---------- MIB Entry 0
|
|
// Alloc Count
|
|
// ...
|
|
// ---- AccessHistogram Entry 0
|
|
// ...
|
|
// ---- AccessHistogram Entry AccessHistogramSize - 1
|
|
// ---------- MIB Entry 1
|
|
// Alloc Count
|
|
// ...
|
|
// ---- AccessHistogram Entry 0
|
|
// ...
|
|
// ---- AccessHistogram Entry AccessHistogramSize - 1
|
|
// ----------
|
|
void SerializeMIBInfoToBuffer(MIBMapTy &MIBMap, const Vector<u64> &StackIds,
|
|
const u64 ExpectedNumBytes, char *&Buffer) {
|
|
char *Ptr = Buffer;
|
|
const u64 NumEntries = StackIds.Size();
|
|
Ptr = WriteBytes(NumEntries, Ptr);
|
|
for (u64 i = 0; i < NumEntries; i++) {
|
|
const u64 Key = StackIds[i];
|
|
MIBMapTy::Handle h(&MIBMap, Key, /*remove=*/true, /*create=*/false);
|
|
CHECK(h.exists());
|
|
Ptr = WriteBytes(Key, Ptr);
|
|
// FIXME: We unnecessarily serialize the AccessHistogram pointer. Adding a
|
|
// serialization schema will fix this issue. See also FIXME in
|
|
// deserialization.
|
|
auto &MIB = (*h)->mib;
|
|
Ptr = WriteBytes(MIB, Ptr);
|
|
for (u64 j = 0; j < MIB.AccessHistogramSize; ++j) {
|
|
u16 HistogramEntry =
|
|
encodeHistogramCount(((u64 *)(MIB.AccessHistogram))[j]);
|
|
Ptr = WriteBytes(HistogramEntry, Ptr);
|
|
}
|
|
if (MIB.AccessHistogramSize > 0) {
|
|
InternalFree((void *)MIB.AccessHistogram);
|
|
}
|
|
}
|
|
CHECK(ExpectedNumBytes >= static_cast<u64>(Ptr - Buffer) &&
|
|
"Expected num bytes != actual bytes written");
|
|
}
|
|
} // namespace
|
|
|
|
// Format
|
|
// ---------- Header
|
|
// Magic
|
|
// Version
|
|
// Total Size
|
|
// Segment Offset
|
|
// MIB Info Offset
|
|
// Stack Offset
|
|
// ---------- Segment Info
|
|
// Num Entries
|
|
// ---------- Segment Entry
|
|
// Start
|
|
// End
|
|
// Offset
|
|
// BuildID 32B
|
|
// ----------
|
|
// ...
|
|
// ----------
|
|
// Optional Padding Bytes
|
|
// ---------- MIB Info
|
|
// Num Entries
|
|
// ---------- MIB Entry
|
|
// Alloc Count
|
|
// ...
|
|
// ---- AccessHistogram Entry 0
|
|
// ...
|
|
// ---- AccessHistogram Entry AccessHistogramSize - 1
|
|
// ---------- MIB Entry 1
|
|
// Alloc Count
|
|
// ...
|
|
// ---- AccessHistogram Entry 0
|
|
// ...
|
|
// ---- AccessHistogram Entry AccessHistogramSize - 1
|
|
// Optional Padding Bytes
|
|
// ---------- Stack Info
|
|
// Num Entries
|
|
// ---------- Stack Entry
|
|
// Num Stacks
|
|
// PC1
|
|
// PC2
|
|
// ...
|
|
// ----------
|
|
// Optional Padding Bytes
|
|
// ...
|
|
u64 SerializeToRawProfile(MIBMapTy &MIBMap, ArrayRef<LoadedModule> Modules,
|
|
char *&Buffer) {
|
|
// Each section size is rounded up to 8b since the first entry in each section
|
|
// is a u64 which holds the number of entries in the section by convention.
|
|
const u64 NumSegmentBytes = RoundUpTo(SegmentSizeBytes(Modules), 8);
|
|
|
|
Vector<u64> StackIds;
|
|
MIBMap.ForEach(RecordStackId, reinterpret_cast<void *>(&StackIds));
|
|
// The first 8b are for the total number of MIB records. Each MIB record is
|
|
// preceded by a 8b stack id which is associated with stack frames in the next
|
|
// section.
|
|
const u64 NumMIBInfoBytes = RoundUpTo(
|
|
sizeof(u64) + StackIds.Size() * (sizeof(u64) + sizeof(MemInfoBlock)), 8);
|
|
|
|
// Get Number of AccessHistogram entries in total
|
|
u64 TotalAccessHistogramEntries = 0;
|
|
MIBMap.ForEach(
|
|
[](const uptr Key, UNUSED LockedMemInfoBlock *const &MIB, void *Arg) {
|
|
u64 *TotalAccessHistogramEntries = (u64 *)Arg;
|
|
*TotalAccessHistogramEntries += MIB->mib.AccessHistogramSize;
|
|
},
|
|
reinterpret_cast<void *>(&TotalAccessHistogramEntries));
|
|
const u64 NumHistogramBytes =
|
|
RoundUpTo(TotalAccessHistogramEntries * sizeof(uint16_t), 8);
|
|
|
|
const u64 NumStackBytes = RoundUpTo(StackSizeBytes(StackIds), 8);
|
|
|
|
// Ensure that the profile is 8b aligned. We allow for some optional padding
|
|
// at the end so that any subsequent profile serialized to the same file does
|
|
// not incur unaligned accesses.
|
|
const u64 TotalSizeBytes =
|
|
RoundUpTo(sizeof(Header) + NumSegmentBytes + NumStackBytes +
|
|
NumMIBInfoBytes + NumHistogramBytes,
|
|
8);
|
|
|
|
// Allocate the memory for the entire buffer incl. info blocks.
|
|
Buffer = (char *)InternalAlloc(TotalSizeBytes);
|
|
char *Ptr = Buffer;
|
|
|
|
Header header{MEMPROF_RAW_MAGIC_64,
|
|
MEMPROF_RAW_VERSION,
|
|
static_cast<u64>(TotalSizeBytes),
|
|
sizeof(Header),
|
|
sizeof(Header) + NumSegmentBytes,
|
|
sizeof(Header) + NumSegmentBytes + NumMIBInfoBytes +
|
|
NumHistogramBytes};
|
|
Ptr = WriteBytes(header, Ptr);
|
|
|
|
SerializeSegmentsToBuffer(Modules, NumSegmentBytes, Ptr);
|
|
Ptr += NumSegmentBytes;
|
|
|
|
SerializeMIBInfoToBuffer(MIBMap, StackIds,
|
|
NumMIBInfoBytes + NumHistogramBytes, Ptr);
|
|
Ptr += NumMIBInfoBytes + NumHistogramBytes;
|
|
|
|
SerializeStackToBuffer(StackIds, NumStackBytes, Ptr);
|
|
|
|
return TotalSizeBytes;
|
|
}
|
|
} // namespace __memprof
|