summaryrefslogtreecommitdiff
path: root/examples/redis-unstable/src/memory_prefetch.c
diff options
context:
space:
mode:
Diffstat (limited to 'examples/redis-unstable/src/memory_prefetch.c')
-rw-r--r--examples/redis-unstable/src/memory_prefetch.c413
1 files changed, 413 insertions, 0 deletions
diff --git a/examples/redis-unstable/src/memory_prefetch.c b/examples/redis-unstable/src/memory_prefetch.c
new file mode 100644
index 0000000..0965f62
--- /dev/null
+++ b/examples/redis-unstable/src/memory_prefetch.c
@@ -0,0 +1,413 @@
+/*
+ * This file utilizes prefetching keys and data for multiple commands in a batch,
+ * to improve performance by amortizing memory access costs across multiple operations.
+ *
+ * Copyright (c) 2025-Present, Redis Ltd. and contributors.
+ * All rights reserved.
+ *
+ * Copyright (c) 2024-present, Valkey contributors.
+ * All rights reserved.
+ *
+ * Licensed under your choice of (a) the Redis Source Available License 2.0
+ * (RSALv2); or (b) the Server Side Public License v1 (SSPLv1); or (c) the
+ * GNU Affero General Public License v3 (AGPLv3).
+ *
+ * Portions of this file are available under BSD3 terms; see REDISCONTRIBUTIONS for more information.
+ */
+
+#include "memory_prefetch.h"
+#include "server.h"
+#include "dict.h"
+
+typedef enum { HT_IDX_FIRST = 0, HT_IDX_SECOND = 1, HT_IDX_INVALID = -1 } HashTableIndex;
+
+typedef enum {
+ PREFETCH_BUCKET, /* Initial state, determines which hash table to use and prefetch the table's bucket */
+ PREFETCH_ENTRY, /* prefetch entries associated with the given key's hash */
+ PREFETCH_KVOBJ, /* prefetch the kv object of the entry found in the previous step */
+ PREFETCH_VALDATA, /* prefetch the value data of the kv object found in the previous step */
+ PREFETCH_DONE /* Indicates that prefetching for this key is complete */
+} PrefetchState;
+
+
+/************************************ State machine diagram for the prefetch operation. ********************************
+ │
+ start
+ │
+ ┌────────▼─────────┐
+ ┌─────────►│ PREFETCH_BUCKET ├────►────────┐
+ │ └────────┬─────────┘ no more tables -> done
+ | bucket|found |
+ │ | │
+ entry not found - goto next table ┌────────▼────────┐ │
+ └────◄─────┤ PREFETCH_ENTRY | ▼
+ ┌────────────►└────────┬────────┘ │
+ | Entry│found │
+ │ | │
+ | ┌───────▼────────┐ │
+ │ | PREFETCH_KVOBJ | ▼
+ │ └───────┬────────┘ │
+ kvobj not found - goto next entry | |
+ │ ┌───────────▼────────────┐ │
+ └──────◄───│ PREFETCH_VALDATA │ ▼
+ └───────────┬────────────┘ │
+ | │
+ ┌───────-─▼─────────────┐ │
+ │ PREFETCH_DONE │◄────────┘
+ └───────────────────────┘
+**********************************************************************************************************************/
+
+typedef void *(*GetValueDataFunc)(const void *val);
+
+typedef struct KeyPrefetchInfo {
+ PrefetchState state; /* Current state of the prefetch operation */
+ HashTableIndex ht_idx; /* Index of the current hash table (0 or 1 for rehashing) */
+ uint64_t bucket_idx; /* Index of the bucket in the current hash table */
+ uint64_t key_hash; /* Hash value of the key being prefetched */
+ dictEntry *current_entry; /* Pointer to the current entry being processed */
+ kvobj *current_kv; /* Pointer to the kv object being prefetched */
+} KeyPrefetchInfo;
+
+/* PrefetchCommandsBatch structure holds the state of the current batch of client commands being processed. */
+typedef struct PrefetchCommandsBatch {
+ size_t cur_idx; /* Index of the current key being processed */
+ size_t key_count; /* Number of keys in the current batch */
+ size_t client_count; /* Number of clients in the current batch */
+ size_t max_prefetch_size; /* Maximum number of keys to prefetch in a batch */
+ void **keys; /* Array of keys to prefetch in the current batch */
+ client **clients; /* Array of clients in the current batch */
+ dict **keys_dicts; /* Main dict for each key */
+ dict **current_dicts; /* Points to dict to prefetch from */
+ KeyPrefetchInfo *prefetch_info; /* Prefetch info for each key */
+ GetValueDataFunc get_value_data_func; /* Function to get the value data */
+} PrefetchCommandsBatch;
+
+static PrefetchCommandsBatch *batch = NULL;
+
+void freePrefetchCommandsBatch(void) {
+ if (batch == NULL) {
+ return;
+ }
+
+ zfree(batch->clients);
+ zfree(batch->keys);
+ zfree(batch->keys_dicts);
+ zfree(batch->prefetch_info);
+ zfree(batch);
+ batch = NULL;
+}
+
+void prefetchCommandsBatchInit(void) {
+ serverAssert(!batch);
+
+ /* To avoid prefetching small batches, we set the max size to twice
+ * the configured size, so if not exceeding twice the limit, we can
+ * prefetch all of it. See also `determinePrefetchCount` */
+ size_t max_prefetch_size = server.prefetch_batch_max_size * 2;
+
+ if (max_prefetch_size == 0) {
+ return;
+ }
+
+ batch = zcalloc(sizeof(PrefetchCommandsBatch));
+ batch->max_prefetch_size = max_prefetch_size;
+ batch->clients = zcalloc(max_prefetch_size * sizeof(client *));
+ batch->keys = zcalloc(max_prefetch_size * sizeof(void *));
+ batch->keys_dicts = zcalloc(max_prefetch_size * sizeof(dict *));
+ batch->prefetch_info = zcalloc(max_prefetch_size * sizeof(KeyPrefetchInfo));
+}
+
+void onMaxBatchSizeChange(void) {
+ if (batch && batch->client_count > 0) {
+ /* We need to process the current batch before updating the size */
+ return;
+ }
+
+ freePrefetchCommandsBatch();
+ prefetchCommandsBatchInit();
+}
+
+/* Prefetch the given pointer and move to the next key in the batch. */
+static inline void prefetchAndMoveToNextKey(void *addr) {
+ redis_prefetch_read(addr);
+ /* While the prefetch is in progress, we can continue to the next key */
+ batch->cur_idx = (batch->cur_idx + 1) % batch->key_count;
+}
+
+static inline void markKeyAsdone(KeyPrefetchInfo *info) {
+ info->state = PREFETCH_DONE;
+ server.stat_total_prefetch_entries++;
+}
+
+/* Returns the next KeyPrefetchInfo structure that needs to be processed. */
+static KeyPrefetchInfo *getNextPrefetchInfo(void) {
+ size_t start_idx = batch->cur_idx;
+ do {
+ KeyPrefetchInfo *info = &batch->prefetch_info[batch->cur_idx];
+ if (info->state != PREFETCH_DONE) return info;
+ batch->cur_idx = (batch->cur_idx + 1) % batch->key_count;
+ } while (batch->cur_idx != start_idx);
+ return NULL;
+}
+
+static void initBatchInfo(dict **dicts, GetValueDataFunc func) {
+ batch->current_dicts = dicts;
+ batch->get_value_data_func = func;
+
+ /* Initialize the prefetch info */
+ for (size_t i = 0; i < batch->key_count; i++) {
+ KeyPrefetchInfo *info = &batch->prefetch_info[i];
+ if (!batch->current_dicts[i] || dictSize(batch->current_dicts[i]) == 0) {
+ info->state = PREFETCH_DONE;
+ continue;
+ }
+ info->ht_idx = HT_IDX_INVALID;
+ info->current_entry = NULL;
+ info->current_kv = NULL;
+ info->state = PREFETCH_BUCKET;
+ info->key_hash = dictGetHash(batch->current_dicts[i], batch->keys[i]);
+ }
+}
+
+/* Prefetch the bucket of the next hash table index.
+ * If no tables are left, move to the PREFETCH_DONE state. */
+static void prefetchBucket(KeyPrefetchInfo *info) {
+ size_t i = batch->cur_idx;
+
+ /* Determine which hash table to use */
+ if (info->ht_idx == HT_IDX_INVALID) {
+ info->ht_idx = HT_IDX_FIRST;
+ } else if (info->ht_idx == HT_IDX_FIRST && dictIsRehashing(batch->current_dicts[i])) {
+ info->ht_idx = HT_IDX_SECOND;
+ } else {
+ /* No more tables left - mark as done. */
+ markKeyAsdone(info);
+ return;
+ }
+
+ /* Prefetch the bucket */
+ info->bucket_idx = info->key_hash & DICTHT_SIZE_MASK(batch->current_dicts[i]->ht_size_exp[info->ht_idx]);
+ prefetchAndMoveToNextKey(&batch->current_dicts[i]->ht_table[info->ht_idx][info->bucket_idx]);
+ info->current_entry = NULL;
+ info->state = PREFETCH_ENTRY;
+}
+
+/* Prefetch the entry in the bucket and move to the PREFETCH_KVOBJ state.
+ * If no more entries in the bucket, move to the PREFETCH_BUCKET state to look at the next table. */
+static void prefetchEntry(KeyPrefetchInfo *info) {
+ size_t i = batch->cur_idx;
+
+ if (info->current_entry) {
+ /* We already found an entry in the bucket - move to the next entry */
+ info->current_entry = dictGetNext(info->current_entry);
+ } else {
+ /* Go to the first entry in the bucket */
+ info->current_entry = batch->current_dicts[i]->ht_table[info->ht_idx][info->bucket_idx];
+ }
+
+ if (info->current_entry) {
+ prefetchAndMoveToNextKey(info->current_entry);
+ info->current_kv = NULL;
+ info->state = PREFETCH_KVOBJ;
+ } else {
+ /* No entry found in the bucket - try the bucket in the next table */
+ info->state = PREFETCH_BUCKET;
+ }
+}
+
+/* Prefetch the kv object in the dict entry, and to the PREFETCH_VALDATA state. */
+static inline void prefetchKVOject(KeyPrefetchInfo *info) {
+ kvobj *kv = dictGetKey(info->current_entry);
+ int is_kv = dictEntryIsKey(info->current_entry);
+
+ info->current_kv = kv;
+ info->state = PREFETCH_VALDATA;
+ /* If the entry is a pointer of kv object, we don't need to prefetch it */
+ if (!is_kv) prefetchAndMoveToNextKey(kv);
+}
+
+/* Prefetch the value data of the kv object found in dict entry. */
+static void prefetchValueData(KeyPrefetchInfo *info) {
+ size_t i = batch->cur_idx;
+ kvobj *kv = info->current_kv;
+ sds key = kvobjGetKey(kv);
+
+ /* 1. If this is the last element, we assume a hit and don't compare the keys
+ * 2. This kv object is the target of the lookup. */
+ if ((!dictGetNext(info->current_entry) && !dictIsRehashing(batch->current_dicts[i])) ||
+ dictCompareKeys(batch->current_dicts[i], batch->keys[i], key))
+ {
+ if (batch->get_value_data_func) {
+ void *value_data = batch->get_value_data_func(kv);
+ if (value_data) prefetchAndMoveToNextKey(value_data);
+ }
+ markKeyAsdone(info);
+ } else {
+ /* Not found in the current entry, move to the next entry */
+ info->state = PREFETCH_ENTRY;
+ }
+}
+
+/* Prefetch dictionary data for an array of keys.
+ *
+ * This function takes an array of dictionaries and keys, attempting to bring
+ * data closer to the L1 cache that might be needed for dictionary operations
+ * on those keys.
+ *
+ * The dictFind algorithm:
+ * 1. Evaluate the hash of the key
+ * 2. Access the index in the first table
+ * 3. Walk the entries linked list until the key is found
+ * If the key hasn't been found and the dictionary is in the middle of rehashing,
+ * access the index on the second table and repeat step 3
+ *
+ * dictPrefetch executes the same algorithm as dictFind, but one step at a time
+ * for each key. Instead of waiting for data to be read from memory, it prefetches
+ * the data and then moves on to execute the next prefetch for another key.
+ *
+ * dicts - An array of dictionaries to prefetch data from.
+ * get_val_data_func - A callback function that dictPrefetch can invoke
+ * to bring the key's value data closer to the L1 cache as well.
+ */
+static void dictPrefetch(dict **dicts, GetValueDataFunc get_val_data_func) {
+ initBatchInfo(dicts, get_val_data_func);
+ KeyPrefetchInfo *info;
+ while ((info = getNextPrefetchInfo())) {
+ switch (info->state) {
+ case PREFETCH_BUCKET: prefetchBucket(info); break;
+ case PREFETCH_ENTRY: prefetchEntry(info); break;
+ case PREFETCH_KVOBJ: prefetchKVOject(info); break;
+ case PREFETCH_VALDATA: prefetchValueData(info); break;
+ default: serverPanic("Unknown prefetch state %d", info->state);
+ }
+ }
+}
+
+/* Helper function to get the value pointer of a kv object. */
+static void *getObjectValuePtr(const void *value) {
+ kvobj *kv = (kvobj *)value;
+ return (kv->type == OBJ_STRING && kv->encoding == OBJ_ENCODING_RAW) ? kv->ptr : NULL;
+}
+
+void resetCommandsBatch(void) {
+ if (batch == NULL) {
+ /* Handle the case where prefetching becomes enabled from disabled. */
+ if (server.prefetch_batch_max_size) prefetchCommandsBatchInit();
+ return;
+ }
+
+ batch->cur_idx = 0;
+ batch->key_count = 0;
+ batch->client_count = 0;
+
+ /* Handle the case where the max prefetch size has been changed. */
+ if (batch->max_prefetch_size != (size_t)server.prefetch_batch_max_size * 2) {
+ onMaxBatchSizeChange();
+ }
+}
+
+/* Prefetching in very small batches tends to be ineffective because the technique
+ * relies on a small gap—typically a few CPU cycles—between issuing the prefetch
+ * and performing the actual memory access. If the batch is too small, this delay
+ * cannot be effectively inserted, and the prefetching yields little to no benefit.
+ *
+ * To avoid wasting effort, when the remaining data is small (less than twice the
+ * maximum batch size), we simply prefetch all of it at once. Otherwise, we only
+ * prefetch a limited portion, capped at the configured maximum. */
+int determinePrefetchCount(int len) {
+ if (!batch) return 0;
+
+ /* The batch max size is double of the configured size. */
+ int config_size = batch->max_prefetch_size / 2;
+ return len < (int)batch->max_prefetch_size ? len : config_size;
+}
+
+/* Prefetch command-related data:
+ * 1. Prefetch the command arguments allocated by the I/O thread to bring them
+ * closer to the L1 cache.
+ * 2. Prefetch the keys and values for all commands in the current batch from
+ * the main dictionaries. */
+void prefetchCommands(void) {
+ if (!batch) return;
+
+ /* Prefetch argv's for all clients */
+ for (size_t i = 0; i < batch->client_count; i++) {
+ client *c = batch->clients[i];
+ if (!c || c->argc <= 1) continue;
+ /* Skip prefetching first argv (cmd name) it was already looked up by
+ * the I/O thread, and the main thread will not touch argv[0]. */
+ for (int j = 1; j < c->argc; j++) {
+ redis_prefetch_read(c->argv[j]);
+ }
+ }
+
+ /* Prefetch the argv->ptr if required */
+ for (size_t i = 0; i < batch->client_count; i++) {
+ client *c = batch->clients[i];
+ if (!c || c->argc <= 1) continue;
+ for (int j = 1; j < c->argc; j++) {
+ if (c->argv[j]->encoding == OBJ_ENCODING_RAW) {
+ redis_prefetch_read(c->argv[j]->ptr);
+ }
+ }
+ }
+
+ /* Get the keys ptrs - we do it here after the key obj was prefetched. */
+ for (size_t i = 0; i < batch->key_count; i++) {
+ batch->keys[i] = ((robj *)batch->keys[i])->ptr;
+ }
+
+ /* Prefetch dict keys for all commands.
+ * Prefetching is beneficial only if there are more than one key. */
+ if (batch->key_count > 1) {
+ server.stat_total_prefetch_batches++;
+ /* Prefetch keys from the main dict */
+ dictPrefetch(batch->keys_dicts, getObjectValuePtr);
+ }
+}
+
+/* Adds the client's command to the current batch.
+ *
+ * Returns C_OK if the command was added successfully, C_ERR otherwise. */
+int addCommandToBatch(client *c) {
+ if (unlikely(!batch)) return C_ERR;
+
+ /* If the batch is full, process it.
+ * We also check the client count to handle cases where
+ * no keys exist for the clients' commands. */
+ if (batch->client_count == batch->max_prefetch_size ||
+ batch->key_count == batch->max_prefetch_size)
+ {
+ return C_ERR;
+ }
+
+ /* Avoid partial prefetching: if the batch already has keys and adding this
+ * client's ready commands would likely exceed the batch size limit, reject
+ * the entire client. This is a conservative estimate using command count as
+ * a proxy for key count to ensure all keys from a client are either fully
+ * prefetched together or not prefetched at all. */
+ if (batch->key_count > 0 &&
+ c->pending_cmds.ready_len + batch->key_count > batch->max_prefetch_size)
+ {
+ return C_ERR;
+ }
+
+ batch->clients[batch->client_count++] = c;
+
+ pendingCommand *pcmd = c->pending_cmds.head;
+ while (pcmd != NULL) {
+ /* Skip commands that have not been preprocessed, or have errors. */
+ if ((pcmd->flags & PENDING_CMD_FLAG_INCOMPLETE) || !pcmd->cmd || pcmd->read_error) break;
+
+ serverAssert(pcmd->flags & PENDING_CMD_KEYS_RESULT_VALID);
+ for (int i = 0; i < pcmd->keys_result.numkeys && batch->key_count < batch->max_prefetch_size; i++) {
+ batch->keys[batch->key_count] = pcmd->argv[pcmd->keys_result.keys[i].pos];
+ batch->keys_dicts[batch->key_count] =
+ kvstoreGetDict(c->db->keys, pcmd->slot > 0 ? pcmd->slot : 0);
+ batch->key_count++;
+ }
+ pcmd = pcmd->next;
+ }
+
+ return C_OK;
+}