Phase 9-1: O(1) SuperSlab lookup optimization - Created ss_addr_map_box: Hash table (8192 buckets) for O(1) SuperSlab lookup - Created ss_tls_hint_box: TLS caching layer for SuperSlab hints - Integrated hash table into registry (init, insert, remove, lookup) - Modified hak_super_lookup() to use new hash table - Expected: 50-80 cycles → 10-20 cycles (not verified - SuperSlab disabled by default) Phase 9-2: EMPTY slab recycling implementation - Created slab_recycling_box: SLAB_TRY_RECYCLE() macro following Box pattern - Integrated into remote drain (superslab_slab.c) - Integrated into TLS SLL drain (tls_sll_drain_box.h) with touched slab tracking - Observable: Debug tracing via HAKMEM_SLAB_RECYCLE_TRACE - Updated Makefile: Added new box objects to 3 build targets Known Issues: - SuperSlab registry exhaustion still occurs (unregistration not working) - shared_pool_release_slab() may not be removing from g_super_reg[] - Needs investigation before Phase 9-2 can be completed Expected Impact (when fixed): - Stage 1 hit rate: 0% → 80% - shared_fail events: 4 → 0 - Kernel overhead: 55% → 15% - Throughput: 16.5M → 25-30M ops/s (+50-80%) 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
321 lines
13 KiB
C
321 lines
13 KiB
C
// tls_sll_drain_box.h - Box: TLS SLL Periodic Drain
|
|
// Purpose: Restore slab accounting consistency by periodically draining TLS SLL to slab freelists
|
|
//
|
|
// Problem:
|
|
// - Fast free path (hak_tiny_free_fast_v2) pushes to TLS SLL without decrementing meta->used
|
|
// - Slabs never appear empty → SuperSlabs never freed → LRU cache never populated
|
|
// - Result: 6,455 mmap/munmap syscalls per 200K iterations (74.8% time)
|
|
//
|
|
// Solution:
|
|
// - Every N frees (default: 1024), drain TLS SLL → slab freelist
|
|
// - This path decrements meta->used properly via tiny_free_local_box()
|
|
// - Enables empty detection → SuperSlabs freed → LRU cache functional
|
|
//
|
|
// Expected Impact:
|
|
// - mmap/munmap: 6,455 → ~100 calls (-96-97%)
|
|
// - Throughput: 563K → 8-10M ops/s (+1,300-1,700%)
|
|
//
|
|
// References:
|
|
// - Root cause: PHASE9_LRU_ARCHITECTURE_ISSUE.md
|
|
// - Design: Option B (Periodic TLS SLL Drain)
|
|
|
|
#pragma once
|
|
|
|
#include <stdint.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <pthread.h>
|
|
#include "tls_sll_box.h" // TLS SLL operations (tls_sll_pop)
|
|
#include "tiny_header_box.h" // Header Box: Single Source of Truth for header operations
|
|
#include "slab_recycling_box.h" // Phase 9-2: EMPTY slab recycling (SLAB_TRY_RECYCLE)
|
|
#include "../hakmem_tiny_config.h" // TINY_NUM_CLASSES
|
|
#include "../hakmem_super_registry.h" // SuperSlab lookup
|
|
#include "../tiny_region_id.h" // HEADER_MAGIC, HEADER_CLASS_MASK
|
|
#include "free_local_box.h" // tiny_free_local_box (decrements meta->used)
|
|
|
|
// ========== ENV Configuration ==========
|
|
|
|
// Check if TLS SLL drain is enabled
|
|
// ENV: HAKMEM_TINY_SLL_DRAIN_ENABLE=1/0 (default: 1)
|
|
static inline int tls_sll_drain_is_enabled(void) {
|
|
static int g_drain_enable = -1;
|
|
if (__builtin_expect(g_drain_enable == -1, 0)) {
|
|
const char* env = getenv("HAKMEM_TINY_SLL_DRAIN_ENABLE");
|
|
if (env && *env == '0') {
|
|
g_drain_enable = 0;
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Drain DISABLED via ENV\n");
|
|
} else {
|
|
g_drain_enable = 1;
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Drain ENABLED (default)\n");
|
|
}
|
|
}
|
|
return g_drain_enable;
|
|
}
|
|
|
|
// Get drain interval (number of frees before triggering drain)
|
|
// ENV: HAKMEM_TINY_SLL_DRAIN_INTERVAL=N (default: 2048)
|
|
static inline uint32_t tls_sll_drain_get_interval(void) {
|
|
static uint32_t g_drain_interval = 0;
|
|
if (__builtin_expect(g_drain_interval == 0, 0)) {
|
|
const char* env = getenv("HAKMEM_TINY_SLL_DRAIN_INTERVAL");
|
|
if (env && *env) {
|
|
int val = atoi(env);
|
|
if (val > 0 && val <= 65536) {
|
|
g_drain_interval = (uint32_t)val;
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Interval=%u (from ENV)\n", g_drain_interval);
|
|
} else {
|
|
g_drain_interval = 2048;
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Invalid ENV value, using default=2048\n");
|
|
}
|
|
} else {
|
|
g_drain_interval = 2048;
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Interval=%u (default)\n", g_drain_interval);
|
|
}
|
|
}
|
|
return g_drain_interval;
|
|
}
|
|
|
|
// ========== Drain Counter (TLS) ==========
|
|
|
|
// Per-class drain counter (TLS, one per size class)
|
|
// Incremented on each free, triggers drain when reaching interval
|
|
static __thread uint32_t g_tls_sll_drain_counter[TINY_NUM_CLASSES] = {0};
|
|
|
|
// Debug: Total drain operations performed (all classes)
|
|
static __thread uint64_t g_tls_sll_drain_total_calls = 0;
|
|
static __thread uint64_t g_tls_sll_drain_total_blocks = 0;
|
|
|
|
// ========== Drain Implementation (Skeleton) ==========
|
|
|
|
// Box: TLS SLL Drain
|
|
// Purpose: Pop blocks from TLS SLL and push to slab freelist
|
|
//
|
|
// Flow:
|
|
// 1. Pop up to batch_size blocks from TLS SLL (g_tls_sll_head[class_idx])
|
|
// 2. For each block:
|
|
// a. Resolve SuperSlab/Slab (like slow path does)
|
|
// b. Call tiny_free_local_box() → decrements meta->used properly
|
|
// 3. Result: meta->used reflects true state, empty detection works
|
|
//
|
|
// Args:
|
|
// class_idx: Size class to drain
|
|
// batch_size: Max blocks to drain (0 = drain all)
|
|
//
|
|
// Returns: Number of blocks drained
|
|
static inline uint32_t tiny_tls_sll_drain(int class_idx, uint32_t batch_size) {
|
|
if (class_idx < 0 || class_idx >= TINY_NUM_CLASSES) {
|
|
return 0;
|
|
}
|
|
|
|
// Sanity check: TLS SLL count
|
|
extern __thread TinyTLSSLL g_tls_sll[TINY_NUM_CLASSES];
|
|
uint32_t avail = g_tls_sll[class_idx].count;
|
|
if (avail == 0) {
|
|
return 0; // Nothing to drain
|
|
}
|
|
|
|
// Drain up to batch_size blocks (0 = drain all)
|
|
uint32_t to_drain = (batch_size == 0) ? avail : (avail < batch_size ? avail : batch_size);
|
|
uint32_t drained = 0;
|
|
|
|
// Phase 9-2: Track touched slabs for EMPTY recycling after drain completes
|
|
// We can't recycle inside the loop (other blocks from same slab may be queued),
|
|
// but we CAN check after all blocks are drained
|
|
#define MAX_TOUCHED_SLABS 64
|
|
struct { SuperSlab* ss; int slab_idx; } touched[MAX_TOUCHED_SLABS];
|
|
int num_touched = 0;
|
|
|
|
// Debug logging
|
|
static int g_debug = -1;
|
|
if (__builtin_expect(g_debug == -1, 0)) {
|
|
const char* env = getenv("HAKMEM_TINY_SLL_DRAIN_DEBUG");
|
|
g_debug = (env && *env && *env != '0') ? 1 : 0;
|
|
}
|
|
|
|
if (g_debug) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] START: class=%d avail=%u to_drain=%u\n",
|
|
class_idx, avail, to_drain);
|
|
}
|
|
|
|
// External functions needed for drain
|
|
// Note: hak_super_lookup() is defined in hakmem_super_registry.h (included transitively)
|
|
extern const size_t g_tiny_class_sizes[TINY_NUM_CLASSES]; // Block sizes (const)
|
|
|
|
// Get thread ID once (used for all blocks)
|
|
// Note: Use pthread_self() directly since tiny_self_u32() is static inline
|
|
uint32_t my_tid = (uint32_t)(uintptr_t)pthread_self();
|
|
|
|
// Drain loop: Pop blocks from TLS SLL and push to slab freelist
|
|
for (uint32_t i = 0; i < to_drain; i++) {
|
|
void* base = NULL;
|
|
if (!tls_sll_pop(class_idx, &base)) {
|
|
// TLS SLL exhausted (concurrent drain or count mismatch)
|
|
if (g_debug) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Pop failed at i=%u/%u (TLS SLL exhausted)\n", i, to_drain);
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (g_debug && i < 5) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] Popped %u/%u: class=%d base=%p\n", i+1, to_drain, class_idx, base);
|
|
}
|
|
|
|
// Resolve SuperSlab/Slab (like slow path does)
|
|
SuperSlab* ss = hak_super_lookup(base);
|
|
if (!ss || ss->magic != SUPERSLAB_MAGIC) {
|
|
// CRITICAL FIX (2025-11-27): Don't push back - causes duplicates!
|
|
// Problem: Pushback bypasses duplicate checking and creates cycles
|
|
// Old buggy approach: Push back to TLS SLL → pointer at BOTH position 0 and position N
|
|
// New approach: Skip this pointer (accept rare leak) to avoid duplicates
|
|
// Leak is acceptable because SuperSlab lookup failure is transient/rare
|
|
if (g_debug) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] SKIP: class=%d base=%p (invalid SuperSlab, pointer leaked)\n",
|
|
class_idx, base);
|
|
}
|
|
// DO NOT push back - would create duplicate!
|
|
// Just continue to next pointer
|
|
continue;
|
|
}
|
|
|
|
// Get slab index
|
|
int slab_idx = slab_index_for(ss, base);
|
|
if (slab_idx < 0 || slab_idx >= ss_slabs_capacity(ss)) {
|
|
// CRITICAL FIX (2025-11-27): Don't push back - causes duplicates!
|
|
if (g_debug) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] SKIP: class=%d base=%p (invalid slab_idx=%d, pointer leaked)\n",
|
|
class_idx, base, slab_idx);
|
|
}
|
|
// DO NOT push back - would create duplicate!
|
|
continue;
|
|
}
|
|
|
|
// Get slab metadata
|
|
TinySlabMeta* meta = &ss->slabs[slab_idx];
|
|
|
|
// CRITICAL FIX: Restore header BEFORE calling tiny_free_local_box()
|
|
// This ensures tiny_free_local_box() can read class_idx from header
|
|
// Uses Header Box API (ALL classes - tiny_free_local_box needs header)
|
|
tiny_header_write_for_alloc(base, class_idx);
|
|
|
|
// Convert BASE → USER pointer (add 1 byte header offset)
|
|
// Phase E1: ALL classes (C0-C7) have 1-byte header
|
|
void* user_ptr = (char*)base + 1;
|
|
|
|
// Call tiny_free_local_box() to:
|
|
// 1. Push block to slab freelist
|
|
// 2. Decrement meta->used (THIS IS THE KEY!)
|
|
tiny_free_local_box(ss, slab_idx, meta, user_ptr, my_tid);
|
|
|
|
#if !HAKMEM_BUILD_RELEASE
|
|
// Trace drain operation (debug only)
|
|
extern void ptr_trace_record_impl(int event, void* ptr, int class_idx, uint64_t op_num,
|
|
void* aux_ptr, uint32_t aux_u32, int aux_int,
|
|
const char* file, int line);
|
|
extern _Atomic uint64_t g_ptr_trace_op_counter;
|
|
uint64_t _trace_op = atomic_fetch_add_explicit(&g_ptr_trace_op_counter, 1, memory_order_relaxed);
|
|
ptr_trace_record_impl(5 /*PTR_EVENT_DRAIN_TO_FREELIST*/, base, class_idx, _trace_op,
|
|
NULL, avail, 0, __FILE__, __LINE__);
|
|
#endif
|
|
|
|
drained++;
|
|
|
|
// Phase 9-2: Track touched slab for later EMPTY check
|
|
// We track (ss, slab_idx) pairs to check after loop completes
|
|
int already_tracked = 0;
|
|
for (int t = 0; t < num_touched; t++) {
|
|
if (touched[t].ss == ss && touched[t].slab_idx == slab_idx) {
|
|
already_tracked = 1;
|
|
break;
|
|
}
|
|
}
|
|
if (!already_tracked && num_touched < MAX_TOUCHED_SLABS) {
|
|
touched[num_touched].ss = ss;
|
|
touched[num_touched].slab_idx = slab_idx;
|
|
num_touched++;
|
|
}
|
|
|
|
// BUG FIX: DO NOT release slab here even if meta->used == 0
|
|
// Reason: Other blocks from the same slab may still be queued in TLS SLL
|
|
// waiting to be drained. Releasing the slab prematurely causes:
|
|
// 1. SuperSlab reused for different class
|
|
// 2. hak_super_lookup() returns NULL for remaining blocks
|
|
// 3. TLS_SLL_POP_INVALID errors and corruption
|
|
// Solution: Let LRU eviction and normal lifecycle handle empty slab release.
|
|
// Empty slabs will naturally be reclaimed when SuperSlab is idle.
|
|
}
|
|
|
|
// Phase 9-2: Check touched slabs and recycle if EMPTY
|
|
// Now that ALL blocks have been drained, it's safe to check for EMPTY slabs
|
|
// This fixes the bug where EMPTY slabs accumulate and never return to freelist
|
|
for (int t = 0; t < num_touched; t++) {
|
|
SuperSlab* ss = touched[t].ss;
|
|
int slab_idx = touched[t].slab_idx;
|
|
TinySlabMeta* meta = &ss->slabs[slab_idx];
|
|
SLAB_TRY_RECYCLE(ss, slab_idx, meta);
|
|
}
|
|
|
|
if (g_debug && drained > 0) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN] END: class=%d drained=%u remaining=%u\n",
|
|
class_idx, drained, g_tls_sll[class_idx].count);
|
|
}
|
|
|
|
// Update stats
|
|
g_tls_sll_drain_total_calls++;
|
|
g_tls_sll_drain_total_blocks += drained;
|
|
|
|
return drained;
|
|
}
|
|
|
|
// ========== Drain Trigger (Called from Fast Free Path) ==========
|
|
|
|
// Box: Try Drain (with counter trigger)
|
|
// Purpose: Check drain counter and trigger drain if interval reached
|
|
//
|
|
// Flow:
|
|
// 1. Increment drain counter for this class
|
|
// 2. If counter >= interval, trigger drain and reset counter
|
|
// 3. Otherwise, do nothing (fast path continues)
|
|
//
|
|
// Args:
|
|
// class_idx: Size class that was just freed
|
|
//
|
|
// Returns: Number of blocks drained (0 if no drain)
|
|
static inline uint32_t tiny_tls_sll_try_drain(int class_idx) {
|
|
// Check if drain is enabled
|
|
if (__builtin_expect(!tls_sll_drain_is_enabled(), 0)) {
|
|
return 0;
|
|
}
|
|
|
|
// Increment counter
|
|
g_tls_sll_drain_counter[class_idx]++;
|
|
|
|
// Check if interval reached
|
|
uint32_t interval = tls_sll_drain_get_interval();
|
|
if (__builtin_expect(g_tls_sll_drain_counter[class_idx] >= interval, 0)) {
|
|
// Trigger drain (drain ALL blocks to enable empty detection)
|
|
// batch_size=0 means drain all available blocks
|
|
uint32_t drained = tiny_tls_sll_drain(class_idx, 0);
|
|
|
|
// Reset counter
|
|
g_tls_sll_drain_counter[class_idx] = 0;
|
|
|
|
return drained;
|
|
}
|
|
|
|
return 0; // No drain triggered
|
|
}
|
|
|
|
// ========== Debug Stats (Destructor) ==========
|
|
|
|
#if !HAKMEM_BUILD_RELEASE
|
|
static void tls_sll_drain_print_stats(void) __attribute__((destructor));
|
|
static void tls_sll_drain_print_stats(void) {
|
|
if (g_tls_sll_drain_total_calls > 0) {
|
|
fprintf(stderr, "[TLS_SLL_DRAIN_STATS] Total drains: %lu, Total blocks: %lu, Avg: %.2f\n",
|
|
g_tls_sll_drain_total_calls,
|
|
g_tls_sll_drain_total_blocks,
|
|
(double)g_tls_sll_drain_total_blocks / g_tls_sll_drain_total_calls);
|
|
}
|
|
}
|
|
#endif
|