Handle-with-cache.c
// Cache miss - load the resource pthread_mutex_unlock(&cache_lock); // Unlock during I/O UserProfile *profile = load_user_profile_from_disk(user_id); pthread_mutex_lock(&cache_lock);
A handle cache solves this by storing active handles in a key-value store after the first access. Subsequent requests bypass the expensive operation and return the cached handle directly. A well-written handle-with-cache.c typically contains four main sections: 1. The Handle and Cache Structures First, we define our handle type (opaque to the user) and the cache entry. handle-with-cache.c
pthread_mutex_lock(&cache_lock); // Double-check: another thread might have inserted it while we were loading entry = g_hash_table_lookup(handle_cache, &user_id); if (entry) { // Discard our loaded profile and use the cached one free_user_profile(profile); entry->ref_count++; pthread_mutex_unlock(&cache_lock); return entry->profile; } The Handle and Cache Structures First, we define
pthread_mutex_unlock(&cache_lock); } The cache_lock mutex protects the hash table, but note that get_handle() releases the lock during the actual load_user_profile_from_disk() call. This is crucial to avoid blocking all threads during I/O. However, it introduces a race condition where two threads might simultaneously miss the cache and both load the same resource. However, it introduces a race condition where two
// Store in cache (use user_id as key) int *key = malloc(sizeof(int)); *key = user_id; g_hash_table_insert(handle_cache, key, new_entry);