Handle-with-cache.c
pthread_mutex_lock(&cache_lock);
// Create new cache entry CacheEntry *new_entry = malloc(sizeof(CacheEntry)); new_entry->profile = profile; new_entry->last_access = time(NULL); new_entry->ref_count = 1;
// Store in cache (use user_id as key) int *key = malloc(sizeof(int)); *key = user_id; g_hash_table_insert(handle_cache, key, new_entry);
// Find the entry for this profile (simplified; real code needs reverse mapping) GHashTableIter iter; gpointer key, value; g_hash_table_iter_init(&iter, handle_cache); while (g_hash_table_iter_next(&iter, &key, &value)) { CacheEntry *entry = value; if (entry->profile == profile) { entry->ref_count--; if (entry->ref_count == 0) { // Last reference - we could evict immediately or mark as stale printf("No more references to user %d, marking for eviction\n", *(int*)key); } break; } } handle-with-cache.c
// Improved get_handle() with double-check UserProfile* get_user_profile_handle_safe(int user_id) { pthread_mutex_lock(&cache_lock); CacheEntry *entry = g_hash_table_lookup(handle_cache, &user_id); if (entry) { entry->ref_count++; pthread_mutex_unlock(&cache_lock); return entry->profile; } pthread_mutex_unlock(&cache_lock); // Load outside lock UserProfile *profile = load_user_profile_from_disk(user_id);
// Cache miss - load the resource pthread_mutex_unlock(&cache_lock); // Unlock during I/O UserProfile *profile = load_user_profile_from_disk(user_id); pthread_mutex_lock(&cache_lock);
// The cache itself (often a global or passed context) static GHashTable *handle_cache = NULL; static pthread_mutex_t cache_lock = PTHREAD_MUTEX_INITIALIZER; This function does the actual heavy lifting – creating a handle from scratch. This is where caching becomes indispensable
// Remove stale entries for (GList *l = to_remove; l; l = l->next) { int *key = l->data; CacheEntry *entry = g_hash_table_lookup(handle_cache, key); free(entry->profile->name); free(entry->profile->email); free(entry->profile); free(entry); g_hash_table_remove(handle_cache, key); free(key); } g_list_free(to_remove);
In systems programming, efficiency is paramount. Repeatedly opening, reading, or computing the same resource (a file, a network socket, a database row, or a complex calculation result) is wasteful. This is where caching becomes indispensable.
static UserProfile* load_user_profile_from_disk(int user_id) { // Simulate expensive I/O printf("Loading user %d from disk...\n", user_id); sleep(1); // Pretend this is slow UserProfile *profile = malloc(sizeof(UserProfile)); profile->user_id = user_id; profile->name = malloc(32); profile->email = malloc(64); sprintf(profile->name, "User_%d", user_id); sprintf(profile->email, "user%d@example.com", user_id); return profile; } This is the heart of the module. The cache is transparent to the caller. A common optimization is or using a per-key
A common optimization is or using a per-key mutex:
// Cache entry wrapper typedef struct { UserProfile *profile; time_t last_access; unsigned int ref_count; // Reference counting for safety } CacheEntry;
// Background thread or called periodically void evict_stale_handles(int max_age_seconds, int max_size) { pthread_mutex_lock(&cache_lock); time_t now = time(NULL); GList *to_remove = NULL;