Files
hakmem/core/tiny_region_id.h
Moe Charm (CI) 25d963a4aa Code Cleanup: Remove false positives, redundant validations, and reduce verbose logging
Following the C7 stride upgrade fix (commit 23c0d9541), this commit performs
comprehensive cleanup to improve code quality and reduce debug noise.

## Changes

### 1. Disable False Positive Checks (tiny_nextptr.h)
- **Disabled**: NXT_MISALIGN validation block with `#if 0`
- **Reason**: Produces false positives due to slab base offsets (2048, 65536)
  not being stride-aligned, causing all blocks to appear "misaligned"
- **TODO**: Reimplement to check stride DISTANCE between consecutive blocks
  instead of absolute alignment to stride boundaries

### 2. Remove Redundant Geometry Validations

**hakmem_tiny_refill_p0.inc.h (P0 batch refill)**
- Removed 25-line CARVE_GEOMETRY_FIX validation block
- Replaced with NOTE explaining redundancy
- **Reason**: Stride table is now correct in tiny_block_stride_for_class(),
  defense-in-depth validation adds overhead without benefit

**ss_legacy_backend_box.c (legacy backend)**
- Removed 18-line LEGACY_FIX_GEOMETRY validation block
- Replaced with NOTE explaining redundancy
- **Reason**: Shared_pool validates geometry at acquisition time

### 3. Reduce Verbose Logging

**hakmem_shared_pool.c (sp_fix_geometry_if_needed)**
- Made SP_FIX_GEOMETRY logging conditional on `!HAKMEM_BUILD_RELEASE`
- **Reason**: Geometry fixes are expected during stride upgrades,
  no need to log in release builds

### 4. Verification
- Build:  Successful (LTO warnings expected)
- Test:  10K iterations (1.87M ops/s, no crashes)
- NXT_MISALIGN false positives:  Eliminated

## Files Modified
- core/tiny_nextptr.h - Disabled false positive NXT_MISALIGN check
- core/hakmem_tiny_refill_p0.inc.h - Removed redundant CARVE validation
- core/box/ss_legacy_backend_box.c - Removed redundant LEGACY validation
- core/hakmem_shared_pool.c - Made SP_FIX_GEOMETRY logging debug-only

## Impact
- **Code clarity**: Removed 43 lines of redundant validation code
- **Debug noise**: Reduced false positive diagnostics
- **Performance**: Eliminated overhead from redundant geometry checks
- **Maintainability**: Single source of truth for geometry validation

🧹 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-11-21 23:00:24 +09:00

268 lines
9.3 KiB
C

// tiny_region_id.h - Region-ID Direct Lookup API (Phase 7)
// Purpose: O(1) class_idx lookup from pointer (eliminates SuperSlab lookup)
// Design: Smart Headers - 1-byte class_idx embedded before each block
// Performance: 2-3 cycles (vs 100+ cycles for SuperSlab lookup)
//
// Expected Impact: 1.2M → 40-60M ops/s (30-50x improvement)
#ifndef TINY_REGION_ID_H
#define TINY_REGION_ID_H
#include <stdint.h>
#include <stddef.h>
#include <execinfo.h>
#include <dlfcn.h>
#include "hakmem_build_flags.h"
#include "tiny_box_geometry.h"
#include "ptr_track.h"
#include "hakmem_super_registry.h"
#include "superslab/superslab_inline.h"
// Feature flag: Enable header-based class_idx lookup
#ifndef HAKMEM_TINY_HEADER_CLASSIDX
#define HAKMEM_TINY_HEADER_CLASSIDX 0
#endif
#if HAKMEM_TINY_HEADER_CLASSIDX
// ========== Header Layout ==========
//
// Memory layout:
// [Header: 1 byte] [User block: N bytes]
// ^ ^
// ptr-1 ptr (returned to user)
//
// Header format (1 byte):
// - Bits 0-3: class_idx (0-15, only 0-7 used for Tiny)
// - Bits 4-7: magic (0xA for validation in debug mode)
//
// Example:
// class_idx = 3 → header = 0xA3 (debug) or 0x03 (release)
#define HEADER_MAGIC 0xA0
#define HEADER_CLASS_MASK 0x0F
// ========== Write Header (Allocation) ==========
// Write class_idx to header (called after allocation)
// Input: base (block start from SuperSlab)
// Returns: user pointer (base + 1, skipping header)
static inline void* tiny_region_id_write_header(void* base, int class_idx) {
if (!base) return base;
// Phase E1-CORRECT: ALL classes (C0-C7) have 1-byte header (no exceptions)
// Rationale: Unified box structure enables:
// - O(1) class identification (no registry lookup)
// - All classes use same fast path
// - Zero special cases across all layers
// Cost: 0.1% memory overhead for C7 (1024B → 1023B usable)
// Benefit: 100% safety, architectural simplicity, maximum performance
// Write header at block start (ALL classes including C7)
uint8_t* header_ptr = (uint8_t*)base;
// Debug: detect header writes with class_idx that disagrees with slab metadata.
do {
static _Atomic uint32_t g_hdr_meta_mis = 0;
struct SuperSlab* ss = hak_super_lookup(base);
if (ss && ss->magic == SUPERSLAB_MAGIC) {
int slab_idx = slab_index_for(ss, base);
if (slab_idx >= 0 && slab_idx < ss_slabs_capacity(ss)) {
uint8_t meta_cls = ss->slabs[slab_idx].class_idx;
if (meta_cls < TINY_NUM_CLASSES && meta_cls != (uint8_t)class_idx) {
uint32_t n = atomic_fetch_add_explicit(&g_hdr_meta_mis, 1, memory_order_relaxed);
if (n < 8) {
void* ra = __builtin_return_address(0);
const char* sym = "(unknown)";
#ifdef __GLIBC__
Dl_info info;
if (dladdr(ra, &info) && info.dli_sname) {
sym = info.dli_sname;
}
#endif
fprintf(stderr,
"[HDR_META_MISMATCH] cls=%d meta_cls=%u base=%p slab_idx=%d ss=%p ra=%p fn=%s\n",
class_idx,
(unsigned)meta_cls,
base,
slab_idx,
(void*)ss,
ra,
sym);
if (n < 4) {
void* bt[8];
int frames = backtrace(bt, 8);
backtrace_symbols_fd(bt, frames, fileno(stderr));
}
fflush(stderr);
}
}
}
}
} while (0);
*header_ptr = HEADER_MAGIC | (class_idx & HEADER_CLASS_MASK);
PTR_TRACK_HEADER_WRITE(base, HEADER_MAGIC | (class_idx & HEADER_CLASS_MASK));
void* user = header_ptr + 1; // skip header for user pointer
PTR_TRACK_MALLOC(base, 0, class_idx); // Track at BASE (where header is)
// Optional guard: log stride/base/user for targeted class
extern int tiny_guard_is_enabled(void);
extern void tiny_guard_on_alloc(int cls, void* base, void* user, size_t stride);
if (tiny_guard_is_enabled()) {
size_t stride = tiny_stride_for_class(class_idx);
tiny_guard_on_alloc(class_idx, base, user, stride);
}
return user;
}
// ========== Read Header (Free) ==========
// Read class_idx from header (called during free)
// Returns: class_idx (0-7), or -1 if invalid
static inline int tiny_region_id_read_header(void* ptr) {
if (!ptr) return -1;
if ((uintptr_t)ptr < 4096) return -1; // reject invalid tiny values
uint8_t* header_ptr = (uint8_t*)ptr - 1;
uint8_t header = *header_ptr;
// CRITICAL FIX (Pool TLS Phase 1): ALWAYS validate magic when Pool TLS is enabled
// Reason: Pool TLS uses different magic (0xb0 vs 0xa0), MUST distinguish them!
// Without this, Pool TLS allocations are wrongly routed to Tiny freelist → corruption
#if !HAKMEM_BUILD_RELEASE || defined(HAKMEM_POOL_TLS_PHASE1)
// Debug/Development OR Pool TLS: Validate magic byte to catch non-header allocations
// Reason: Mid/Large allocations don't have headers, must detect and reject them
uint8_t magic = header & 0xF0;
#if HAKMEM_DEBUG_VERBOSE
static int debug_count = 0;
if (debug_count < 5) {
fprintf(stderr, "[TINY_READ_HEADER] ptr=%p header=0x%02x magic=0x%02x expected=0x%02x\n",
ptr, header, magic, HEADER_MAGIC);
debug_count++;
}
#endif
if (magic != HEADER_MAGIC) {
// Invalid header - likely non-header allocation (Mid/Large/Pool TLS)
#if HAKMEM_DEBUG_VERBOSE
if (debug_count < 6) { // One more after the 5 above
fprintf(stderr, "[TINY_READ_HEADER] REJECTING ptr=%p (magic mismatch)\n", ptr);
}
#endif
#if !HAKMEM_BUILD_RELEASE
static int invalid_count = 0;
if (invalid_count < 5) {
fprintf(stderr, "[HEADER_INVALID] ptr=%p, header=%02x, magic=%02x (expected %02x)\n",
ptr, header, magic, HEADER_MAGIC);
invalid_count++;
}
#endif
// Optional guard hook for invalid header
extern void tiny_guard_on_invalid(void* user_ptr, uint8_t hdr);
if (tiny_guard_is_enabled()) tiny_guard_on_invalid(ptr, header);
return -1;
}
#else
// Release (without Pool TLS): Skip magic validation (save 2-3 cycles)
// Safety: Bounds check below still prevents out-of-bounds array access
// Trade-off: Mid/Large frees may corrupt TLS freelist (rare, ~0.1% of frees)
// NOTE: This optimization is DISABLED when Pool TLS is enabled (different magic bytes!)
#endif
int class_idx = (int)(header & HEADER_CLASS_MASK);
// CRITICAL: Always validate class_idx range (even in release builds)
// Reason: Corrupted headers could cause out-of-bounds array access
#ifndef TINY_NUM_CLASSES
#define TINY_NUM_CLASSES 8
#endif
if (class_idx < 0 || class_idx >= TINY_NUM_CLASSES) {
// Corrupted header
return -1;
}
return class_idx;
}
// ========== Header Validation ==========
// Check if pointer has valid header (debug mode)
static inline int tiny_region_id_has_header(void* ptr) {
#if !HAKMEM_BUILD_RELEASE
if (!ptr) return 0;
if ((uintptr_t)ptr < 4096) return 0;
uint8_t* header_ptr = (uint8_t*)ptr - 1;
uint8_t header = *header_ptr;
uint8_t magic = header & 0xF0;
return (magic == HEADER_MAGIC);
#else
// Release: Assume all allocations have headers
(void)ptr;
return 1;
#endif
}
// ========== Allocation Size Adjustment ==========
// Calculate allocation size including header (1 byte)
static inline size_t tiny_region_id_alloc_size(size_t user_size) {
return user_size + 1; // Add 1 byte for header
}
// Calculate user size from allocation size
static inline size_t tiny_region_id_user_size(size_t alloc_size) {
return alloc_size - 1;
}
// ========== Performance Notes ==========
//
// Header Read Performance:
// - Best case: 2 cycles (L1 hit, no validation)
// - Average: 3 cycles (with class_idx extraction)
// - Worst case: 5 cycles (debug validation)
// - vs SuperSlab lookup: 100+ cycles (50x faster!)
//
// Memory Overhead:
// - Per block: 1 byte
// - 8-byte blocks: 12.5% overhead
// - 128-byte blocks: 0.8% overhead
// - Average (typical workload): ~1.5%
// - Slab[0]: 0% (reuses 960B wasted padding)
//
// Cache Impact:
// - Excellent: Header is inline with user data
// - Prefetch: Header loaded with first user data access
// - No additional cache lines required
#else // !HAKMEM_TINY_HEADER_CLASSIDX
// Disabled: No-op implementations
static inline void* tiny_region_id_write_header(void* ptr, int class_idx) {
(void)class_idx;
return ptr;
}
static inline int tiny_region_id_read_header(void* ptr) {
(void)ptr;
return -1; // Not supported
}
static inline int tiny_region_id_has_header(void* ptr) {
(void)ptr;
return 0; // No headers
}
static inline size_t tiny_region_id_alloc_size(size_t user_size) {
return user_size; // No header
}
static inline size_t tiny_region_id_user_size(size_t alloc_size) {
return alloc_size;
}
#endif // HAKMEM_TINY_HEADER_CLASSIDX
#endif // TINY_REGION_ID_H