Summary of Deletions:
- Remove core/box/unified_batch_box.c (26 lines)
* Legacy batch allocation logic superseded by Alloc Gatekeeper Box
* unified_cache now handles allocation aggregation
- Remove core/box/unified_batch_box.h (29 lines)
* Header declarations for deprecated unified_batch_box module
- Remove core/tiny_free_fast.inc.h (329 lines)
* Legacy fast-path free implementation
* Functionality consolidated into:
- tiny_free_gate_box.h (Fail-Fast layer + diagnostics)
- malloc_tiny_fast.h (Free path integration)
- unified_cache (return to freelist)
* Code path now routes through Gatekeeper Box for consistency
Build System Updates:
- Update Makefile
* Remove unified_batch_box.o from OBJS_BASE
* Remove unified_batch_box_shared.o from SHARED_OBJS
* Remove unified_batch_box.o from BENCH_HAKMEM_OBJS_BASE
- Update core/hakmem_tiny_phase6_wrappers_box.inc
* Remove unified_batch_box references
* Simplify allocation wrapper to use new Gatekeeper architecture
Impact:
- Removes ~385 lines of redundant/superseded code
- Consolidates allocation logic through unified Gatekeeper entry points
- All functionality preserved via new Box-based architecture
- Simplifies codebase and reduces maintenance burden
Testing:
- Build verification: make clean && make RELEASE=0/1
- Smoke tests: All pass (simple_alloc, loop 10M, pool_tls)
- No functional regressions
Rationale:
After implementing Alloc/Free Gatekeeper Boxes with Fail-Fast layers
and Unified Cache type safety, the legacy separate implementations
became redundant. This commit completes the architectural consolidation
and simplifies the allocator codebase.
🤖 Generated with [Claude Code](https://claude.com/claude-code)
Co-Authored-By: Claude <noreply@anthropic.com>
102 lines
4.2 KiB
C++
102 lines
4.2 KiB
C++
// Phase 6-1.7: Box Theory Refactoring - Mutual exclusion check
|
|
#if HAKMEM_TINY_PHASE6_BOX_REFACTOR
|
|
#if defined(HAKMEM_TINY_PHASE6_METADATA) || defined(HAKMEM_TINY_PHASE6_ULTRA_SIMPLE)
|
|
#error "Cannot enable PHASE6_BOX_REFACTOR with other Phase 6 options"
|
|
#endif
|
|
|
|
// Box 1: Atomic Operations (Layer 0 - Foundation)
|
|
#include "tiny_atomic.h"
|
|
|
|
// Box 5: Allocation Fast Path (Layer 1 - 3-4 instructions)
|
|
#include "tiny_alloc_fast.inc.h"
|
|
|
|
// ---------------- Refill count (Front) global config ----------------
|
|
// Parsed once at init; hot path reads plain ints (no getenv).
|
|
int g_refill_count_global = 0; // HAKMEM_TINY_REFILL_COUNT
|
|
int g_refill_count_hot = 0; // HAKMEM_TINY_REFILL_COUNT_HOT
|
|
int g_refill_count_mid = 0; // HAKMEM_TINY_REFILL_COUNT_MID
|
|
int g_refill_count_class[TINY_NUM_CLASSES] = {0}; // HAKMEM_TINY_REFILL_COUNT_C{0..7}
|
|
|
|
// Export wrapper functions for hakmem.c to call
|
|
// Phase 6-1.7 Optimization: Remove diagnostic overhead, rely on LTO for inlining
|
|
void* hak_tiny_alloc_fast_wrapper(size_t size) {
|
|
static _Atomic int g_alloc_fast_trace = 0;
|
|
if (atomic_fetch_add_explicit(&g_alloc_fast_trace, 1, memory_order_relaxed) < 128) {
|
|
HAK_TRACE("[tiny_alloc_fast_wrapper_enter]\n");
|
|
}
|
|
// Phase E5: Ultra fast path (8-instruction alloc, bypasses all layers)
|
|
// Enable with: HAKMEM_ULTRA_FAST_PATH=1 (compile-time)
|
|
#if HAKMEM_ULTRA_FAST_PATH
|
|
void* ret = tiny_alloc_fast_ultra(size);
|
|
if (ret) return ret;
|
|
// Miss → fallback to full fast path
|
|
#endif
|
|
|
|
// Bench-only ultra-short path: bypass diagnostics and pointer tracking
|
|
// Enable with: HAKMEM_BENCH_FAST_FRONT=1
|
|
static int g_bench_fast_front = -1;
|
|
if (__builtin_expect(g_bench_fast_front == -1, 0)) {
|
|
const char* e = getenv("HAKMEM_BENCH_FAST_FRONT");
|
|
g_bench_fast_front = (e && *e && *e != '0') ? 1 : 0;
|
|
}
|
|
if (__builtin_expect(g_bench_fast_front, 0)) {
|
|
return tiny_alloc_fast(size);
|
|
}
|
|
|
|
static _Atomic uint64_t wrapper_call_count = 0;
|
|
uint64_t call_num = atomic_fetch_add(&wrapper_call_count, 1);
|
|
|
|
// Pointer tracking init (first call only)
|
|
PTR_TRACK_INIT();
|
|
|
|
// PRIORITY 3: Periodic canary validation (every 1000 ops)
|
|
periodic_canary_check(call_num, "hak_tiny_alloc_fast_wrapper");
|
|
|
|
// Box I: Periodic full integrity check (every 5000 ops)
|
|
#if HAKMEM_INTEGRITY_LEVEL >= 3
|
|
if ((call_num % 5000) == 0) {
|
|
extern void integrity_periodic_full_check(const char*);
|
|
integrity_periodic_full_check("periodic check in alloc wrapper");
|
|
}
|
|
#endif
|
|
|
|
#if !HAKMEM_BUILD_RELEASE
|
|
if (call_num > 14250 && call_num < 14280 && size <= 1024) {
|
|
fprintf(stderr, "[HAK_TINY_ALLOC_FAST_WRAPPER] call=%lu size=%zu\n", call_num, size);
|
|
fflush(stderr);
|
|
}
|
|
#endif
|
|
|
|
void* result = tiny_alloc_fast(size);
|
|
|
|
#if !HAKMEM_BUILD_RELEASE
|
|
if (call_num > 14250 && call_num < 14280 && size <= 1024) {
|
|
fprintf(stderr, "[HAK_TINY_ALLOC_FAST_WRAPPER] call=%lu returned %p\n", call_num, result);
|
|
fflush(stderr);
|
|
}
|
|
#endif
|
|
return result;
|
|
}
|
|
|
|
void hak_tiny_free_fast_wrapper(void* ptr) {
|
|
static _Atomic uint64_t free_call_count = 0;
|
|
uint64_t call_num = atomic_fetch_add(&free_call_count, 1);
|
|
if (call_num > 14135 && call_num < 14145) {
|
|
fprintf(stderr, "[HAK_TINY_FREE_FAST_WRAPPER] call=%lu ptr=%p\n", call_num, ptr);
|
|
fflush(stderr);
|
|
}
|
|
// Box 6 v1 (tiny_free_fast) は v2 に置き換え済み。
|
|
// Wrapper からは Box 経路の slow/fast 判定に委ねる。
|
|
hak_tiny_free(ptr);
|
|
if (call_num > 14135 && call_num < 14145) {
|
|
fprintf(stderr, "[HAK_TINY_FREE_FAST_WRAPPER] call=%lu completed\n", call_num);
|
|
fflush(stderr);
|
|
}
|
|
}
|
|
|
|
// HAKMEM_TINY_PHASE6_ULTRA_SIMPLE - REMOVED (dead code cleanup 2025-11-27)
|
|
#elif defined(HAKMEM_TINY_PHASE6_METADATA)
|
|
// Phase 6-1.6: Metadata header (recommended)
|
|
#include "hakmem_tiny_metadata.inc"
|
|
#endif
|