Documentation: - Created docs/DEFENSIVE_LAYERS_MAPPING.md documenting all 5 defensive layers - Maps which symptoms each layer suppresses - Defines safe removal order after root cause fix - Includes test methods for each layer removal Diagnostic Logging Enhancements (ChatGPT work): - TLS_SLL_HEAD_SET log with count and backtrace for NORMALIZE_USERPTR - tiny_next_store_log with filtering capability - Environment variables for log filtering: - HAKMEM_TINY_SLL_NEXTCLS: class filter for next store (-1 disables) - HAKMEM_TINY_SLL_NEXTTAG: tag filter (substring match) - HAKMEM_TINY_SLL_HEADCLS: class filter for head trace Current Investigation Status: - sh8bench 60/120s: crash-free, zero NEXT_INVALID/HDR_RESET/SANITIZE - BUT: shot limit (256) exhausted by class3 tls_push before class1/drain - Need: Add tags to pop/clear paths, or increase shot limit for class1 Purpose of this commit: - Document defensive layers for safe removal later - Enable targeted diagnostic logging - Prepare for final root cause identification Next Steps: 1. Add tags to tls_sll_pop tiny_next_write (e.g., "tls_pop_clear") 2. Re-run with HAKMEM_TINY_SLL_NEXTTAG=tls_pop 3. Capture class1 writes that lead to corruption 🤖 Generated with Claude Code (https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
162 lines
6.7 KiB
C
162 lines
6.7 KiB
C
// ptr_trace.h - Pointer link read/write tracing (Debug-only, macro-replaceable)
|
||
//
|
||
// Purpose:
|
||
// - Centrally instrument single-linked list next pointer reads/writes
|
||
// - Zero overhead in release builds (macros devolve to raw ops)
|
||
// - Compact TLS ring for triage; optional dump at exit via env
|
||
//
|
||
// Usage:
|
||
// - Replace direct next ops with PTR_NEXT_WRITE / PTR_NEXT_READ
|
||
// - Tag with a short literal (e.g., "tls_push", "tls_pop", "splice")
|
||
// - Offsets should be header-aware (C0–C6: +1, C7: +0)
|
||
//
|
||
// Control:
|
||
// - Compile-time: HAKMEM_PTR_TRACE (default: 1 for debug, 0 for release)
|
||
// - Runtime dump: HAKMEM_PTR_TRACE_DUMP=1 (prints ring at exit)
|
||
//
|
||
// Phase E1-CORRECT: Uses Box API (tiny_next_ptr_box.h) for offset calculation
|
||
|
||
#pragma once
|
||
|
||
#include <stdint.h>
|
||
#include <stddef.h>
|
||
#include "box/tiny_next_ptr_box.h" // Box API: tiny_next_read/write
|
||
|
||
#ifndef HAKMEM_PTR_TRACE
|
||
# if !HAKMEM_BUILD_RELEASE
|
||
# define HAKMEM_PTR_TRACE 1
|
||
# else
|
||
# define HAKMEM_PTR_TRACE 0
|
||
# endif
|
||
#endif
|
||
|
||
#if HAKMEM_PTR_TRACE
|
||
#include <stdio.h>
|
||
|
||
typedef struct {
|
||
const char* tag; // literal tag
|
||
int class_idx; // tiny class (0..7)
|
||
void* node; // node address (base)
|
||
void* val; // written/read value
|
||
size_t off; // offset used for access
|
||
} ptr_trace_ev_t;
|
||
|
||
#ifndef PTR_TRACE_CAP
|
||
# define PTR_TRACE_CAP 256
|
||
#endif
|
||
|
||
static __thread ptr_trace_ev_t g_ptr_trace_ring[PTR_TRACE_CAP];
|
||
static __thread uint32_t g_ptr_trace_idx = 0;
|
||
static __thread int g_ptr_trace_dump_registered = 0;
|
||
|
||
static inline void ptr_trace_record(const char* tag, int cls, void* node, void* val, size_t off) {
|
||
uint32_t i = g_ptr_trace_idx++;
|
||
g_ptr_trace_ring[i & (PTR_TRACE_CAP - 1)] = (ptr_trace_ev_t){ tag, cls, node, val, off };
|
||
}
|
||
|
||
static inline void ptr_trace_dump(void) {
|
||
fprintf(stderr, "\n[PTR_TRACE_DUMP] last=%u (cap=%u)\n", g_ptr_trace_idx, (unsigned)PTR_TRACE_CAP);
|
||
uint32_t n = (g_ptr_trace_idx < PTR_TRACE_CAP) ? g_ptr_trace_idx : PTR_TRACE_CAP;
|
||
for (uint32_t k = 0; k < n; k++) {
|
||
const ptr_trace_ev_t* e = &g_ptr_trace_ring[(g_ptr_trace_idx - n + k) & (PTR_TRACE_CAP - 1)];
|
||
fprintf(stderr, "[%3u] tag=%s cls=%d node=%p val=%p off=%zu\n",
|
||
k, e->tag ? e->tag : "?", e->class_idx, e->node, e->val, e->off);
|
||
}
|
||
}
|
||
|
||
static inline void ptr_trace_try_register_dump(void) {
|
||
#if HAKMEM_BUILD_RELEASE
|
||
return;
|
||
#else
|
||
if (g_ptr_trace_dump_registered) return;
|
||
g_ptr_trace_dump_registered = 1;
|
||
|
||
// Gate ENV check behind !HAKMEM_BUILD_RELEASE
|
||
static int g_dump_enabled = -1;
|
||
if (__builtin_expect(g_dump_enabled == -1, 0)) {
|
||
const char* env = getenv("HAKMEM_PTR_TRACE_DUMP");
|
||
g_dump_enabled = (env && *env && *env != '0') ? 1 : 0;
|
||
}
|
||
if (!g_dump_enabled) return;
|
||
atexit(ptr_trace_dump);
|
||
#endif
|
||
}
|
||
|
||
// Immediate dump (Debug only) — static inline to avoid ODR/link conflicts under LTO
|
||
// Only dumps if HAKMEM_PTR_TRACE_VERBOSE=1 to avoid excessive output in debug builds
|
||
#if HAKMEM_PTR_TRACE
|
||
static inline void ptr_trace_dump_now(const char* reason) {
|
||
#if HAKMEM_BUILD_RELEASE
|
||
(void)reason;
|
||
return;
|
||
#else
|
||
// Gate ENV check behind !HAKMEM_BUILD_RELEASE
|
||
static int verbose_mode = -1;
|
||
if (__builtin_expect(verbose_mode == -1, 0)) {
|
||
const char* env = getenv("HAKMEM_PTR_TRACE_VERBOSE");
|
||
verbose_mode = (env && *env && *env != '0') ? 1 : 0;
|
||
}
|
||
if (!verbose_mode) return; // Skip verbose logging unless explicitly enabled
|
||
|
||
fprintf(stderr, "\n[PTR_TRACE_NOW] reason=%s last=%u (cap=%u)\n",
|
||
reason ? reason : "(null)", g_ptr_trace_idx, (unsigned)PTR_TRACE_CAP);
|
||
uint32_t n = (g_ptr_trace_idx < PTR_TRACE_CAP) ? g_ptr_trace_idx : PTR_TRACE_CAP;
|
||
for (uint32_t k = 0; k < n; k++) {
|
||
const ptr_trace_ev_t* e = &g_ptr_trace_ring[(g_ptr_trace_idx - n + k) & (PTR_TRACE_CAP - 1)];
|
||
fprintf(stderr, "[%3u] tag=%s cls=%d node=%p val=%p off=%zu\n",
|
||
k, e->tag ? e->tag : "?", e->class_idx, e->node, e->val, e->off);
|
||
}
|
||
#endif
|
||
}
|
||
#else
|
||
static inline void ptr_trace_dump_now(const char* reason) { (void)reason; }
|
||
#endif
|
||
|
||
// Phase E1-CORRECT: Use Box API for all next pointer operations
|
||
// Box API handles offset calculation internally based on class_idx.
|
||
// `off` は呼び出し元互換用に受け取るが、アドレス計算には使わない(ログ専用)。
|
||
#define PTR_NEXT_WRITE(tag, cls, node, off, value) do { \
|
||
g_tiny_next_tag = (tag); \
|
||
g_tiny_next_file = __FILE__; \
|
||
g_tiny_next_line = __LINE__; \
|
||
g_tiny_next_ra0 = __builtin_return_address(0); \
|
||
g_tiny_next_ra1 = __builtin_return_address(1); \
|
||
g_tiny_next_ra2 = __builtin_return_address(2); \
|
||
(void)(off); \
|
||
tiny_next_write((cls), (node), (value)); \
|
||
ptr_trace_record((tag), (cls), (node), (value), (size_t)(off)); \
|
||
ptr_trace_try_register_dump(); \
|
||
} while (0)
|
||
|
||
#define PTR_NEXT_READ(tag, cls, node, off, out_var) do { \
|
||
(void)(off); \
|
||
void* _tmp = tiny_next_read((cls), (node)); \
|
||
(out_var) = _tmp; \
|
||
ptr_trace_record((tag), (cls), (node), (out_var), (size_t)(off)); \
|
||
ptr_trace_try_register_dump(); \
|
||
} while (0)
|
||
|
||
#else // HAKMEM_PTR_TRACE == 0
|
||
|
||
// Phase E1-CORRECT: Use Box API for all next pointer operations (Release mode)
|
||
// `off` は互換用のダミーで、Box API が offset を決定する。
|
||
#define PTR_NEXT_WRITE(tag, cls, node, off, value) \
|
||
do { \
|
||
g_tiny_next_tag = (tag); \
|
||
g_tiny_next_file = __FILE__; \
|
||
g_tiny_next_line = __LINE__; \
|
||
g_tiny_next_ra0 = __builtin_return_address(0); \
|
||
g_tiny_next_ra1 = __builtin_return_address(1); \
|
||
g_tiny_next_ra2 = __builtin_return_address(2); \
|
||
(void)(tag); (void)(off); \
|
||
tiny_next_write((cls), (node), (value)); \
|
||
} while (0)
|
||
|
||
#define PTR_NEXT_READ(tag, cls, node, off, out_var) \
|
||
do { (void)(tag); (void)(off); (out_var) = tiny_next_read((cls), (node)); } while (0)
|
||
|
||
// Always provide a stub for release builds so callers can link
|
||
static inline void ptr_trace_dump_now(const char* reason) { (void)reason; }
|
||
|
||
#endif // HAKMEM_PTR_TRACE
|