From 5d8dfe892a2ea89f706ee140c3bdcfd89fe03fda Mon Sep 17 00:00:00 2001 From: Mitja Felicijan Date: Wed, 21 Jan 2026 22:40:55 +0100 Subject: Add Redis source code for testing --- .../redis-unstable/tests/unit/memefficiency.tcl | 1176 ++++++++++++++++++++ 1 file changed, 1176 insertions(+) create mode 100644 examples/redis-unstable/tests/unit/memefficiency.tcl (limited to 'examples/redis-unstable/tests/unit/memefficiency.tcl') diff --git a/examples/redis-unstable/tests/unit/memefficiency.tcl b/examples/redis-unstable/tests/unit/memefficiency.tcl new file mode 100644 index 0000000..123bf37 --- /dev/null +++ b/examples/redis-unstable/tests/unit/memefficiency.tcl @@ -0,0 +1,1176 @@ +# +# Copyright (c) 2009-Present, Redis Ltd. +# All rights reserved. +# +# Copyright (c) 2024-present, Valkey contributors. +# All rights reserved. +# +# Licensed under your choice of (a) the Redis Source Available License 2.0 +# (RSALv2); or (b) the Server Side Public License v1 (SSPLv1); or (c) the +# GNU Affero General Public License v3 (AGPLv3). +# +# Portions of this file are available under BSD3 terms; see REDISCONTRIBUTIONS for more information. +# + +proc test_memory_efficiency {range} { + r flushall + set rd [redis_deferring_client] + set base_mem [s used_memory] + set written 0 + for {set j 0} {$j < 10000} {incr j} { + set key key:$j + set val [string repeat A [expr {int(rand()*$range)}]] + $rd set $key $val + incr written [string length $key] + incr written [string length $val] + incr written 2 ;# A separator is the minimum to store key-value data. + } + for {set j 0} {$j < 10000} {incr j} { + $rd read ; # Discard replies + } + + set current_mem [s used_memory] + set used [expr {$current_mem-$base_mem}] + set efficiency [expr {double($written)/$used}] + return $efficiency +} + +start_server {tags {"memefficiency external:skip"}} { + foreach {size_range expected_min_efficiency} { + 32 0.15 + 64 0.25 + 128 0.35 + 1024 0.75 + 16384 0.82 + } { + test "Memory efficiency with values in range $size_range" { + set efficiency [test_memory_efficiency $size_range] + assert {$efficiency >= $expected_min_efficiency} + } + } +} + +run_solo {defrag} { + proc wait_for_defrag_stop {maxtries delay {expect_frag 0}} { + wait_for_condition $maxtries $delay { + [s active_defrag_running] eq 0 && ($expect_frag == 0 || [s allocator_frag_ratio] <= $expect_frag) + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + if {$expect_frag != 0} { + fail "defrag didn't stop or failed to achieve expected frag ratio ([s allocator_frag_ratio] > $expect_frag)" + } else { + fail "defrag didn't stop." + } + } + } + + proc discard_replies_every {rd count frequency discard_num} { + if {$count % $frequency == 0} { + for {set k 0} {$k < $discard_num} {incr k} { + $rd read ; # Discard replies + } + } + } + + proc test_active_defrag {type} { + + # note: Disabling lookahead because it changes the number and order of allocations which interferes with defrag and causes tests to fail + r config set lookahead 1 + + r debug reply-copy-avoidance 0 ;# Disable copy avoidance because it affects memory usage + + if {[string match {*jemalloc*} [s mem_allocator]] && [r debug mallctl arenas.page] <= 8192} { + test "Active defrag main dictionary: $type" { + r config set hz 100 + r config set activedefrag no + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 2mb + r config set maxmemory 100mb + r config set maxmemory-policy allkeys-lru + + populate 700000 asdf1 150 + populate 100 asdf1 150 0 false 1000 + populate 170000 asdf2 300 + populate 100 asdf2 300 0 false 1000 + + assert {[scan [regexp -inline {expires\=([\d]*)} [r info keyspace]] expires=%d] > 0} + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + if {$::verbose} { + puts "frag $frag" + } + assert {$frag >= 1.4} + + r config set latency-monitor-threshold 5 + r latency reset + r config set maxmemory 110mb ;# prevent further eviction (not to fail the digest test) + set digest [debug_digest] + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + # Wait for the active defrag to start working (decision once a + # second). + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # This test usually runs for a while, during this interval, we test the range. + assert_range [s active_defrag_running] 65 75 + r config set active-defrag-cycle-min 1 + r config set active-defrag-cycle-max 1 + after 120 ;# serverCron only updates the info once in 100ms + assert_range [s active_defrag_running] 1 1 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + + # Wait for the active defrag to stop working. + wait_for_defrag_stop 2000 100 1.1 + + # Test the fragmentation is lower. + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + set max_latency 0 + foreach event [r latency latest] { + lassign $event eventname time latency max + if {$eventname == "active-defrag-cycle"} { + set max_latency $max + } + } + if {$::verbose} { + puts "frag $frag" + set misses [s active_defrag_misses] + set hits [s active_defrag_hits] + puts "hits: $hits" + puts "misses: $misses" + puts "max latency $max_latency" + puts [r latency latest] + puts [r latency history active-defrag-cycle] + } + # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, + # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher + if {!$::no_latency} { + assert {$max_latency <= 30} + } + } + # verify the data isn't corrupted or changed + set newdigest [debug_digest] + assert {$digest eq $newdigest} + r save ;# saving an rdb iterates over all the data / pointers + + # if defrag is supported, test AOF loading too + if {[r config get activedefrag] eq "activedefrag yes" && $type eq "standalone"} { + test "Active defrag - AOF loading" { + # reset stats and load the AOF file + r config resetstat + r config set key-load-delay -25 ;# sleep on average 1/25 usec + # Note: This test is checking if defrag is working DURING AOF loading (while + # timers are not active). So we don't give any extra time, and we deactivate + # defrag immediately after the AOF loading is complete. During loading, + # defrag will get invoked less often, causing starvation prevention. We + # should expect longer latency measurements. + r debug loadaof + r config set activedefrag no + # measure hits and misses right after aof loading + set misses [s active_defrag_misses] + set hits [s active_defrag_hits] + + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + set max_latency 0 + foreach event [r latency latest] { + lassign $event eventname time latency max + if {$eventname == "while-blocked-cron"} { + set max_latency $max + } + } + if {$::verbose} { + puts "AOF loading:" + puts "frag $frag" + puts "hits: $hits" + puts "misses: $misses" + puts "max latency $max_latency" + puts [r latency latest] + puts [r latency history "while-blocked-cron"] + } + # make sure we had defrag hits during AOF loading + assert {$hits > 100000} + # make sure the defragger did enough work to keep the fragmentation low during loading. + # we cannot check that it went all the way down, since we don't wait for full defrag cycle to complete. + assert {$frag < 1.4} + # since the AOF contains simple (fast) SET commands (and the cron during loading runs every 1024 commands), + # it'll still not block the loading for long periods of time. + if {!$::no_latency} { + assert {$max_latency <= 40} + } + } + } ;# Active defrag - AOF loading + } + r config set appendonly no + r config set key-load-delay 0 + + test "Active defrag eval scripts: $type" { + r flushdb + r script flush sync + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1500kb + r config set maxmemory 0 + + set n 50000 + + # Populate memory with interleaving script-key pattern of same size + set dummy_script "--[string repeat x 400]\nreturn " + set rd [redis_deferring_client] + # Send commands in batches and read responses to avoid TCP deadlock. + # Without interleaving reads, TCP congestion control can throttle + # the connection when buffers fill, causing the test to hang. + set batch_size 1000 + for {set j 0} {$j < $n} {incr j} { + set val "$dummy_script[format "%06d" $j]" + $rd script load $val + $rd set k$j $val + if {($j + 1) % $batch_size == 0} { + for {set i 0} {$i < $batch_size} {incr i} { + $rd read ; # Discard script load replies + $rd read ; # Discard set replies + } + } + } + # Read remaining responses + set remaining [expr {$n % $batch_size}] + for {set j 0} {$j < $remaining} {incr j} { + $rd read ; # Discard script load replies + $rd read ; # Discard set replies + } + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_lessthan [s allocator_frag_ratio] 1.05 + + # Delete all the keys to create fragmentation + # Use same batching pattern to avoid TCP deadlock + for {set j 0} {$j < $n} {incr j} { + $rd del k$j + if {($j + 1) % $batch_size == 0} { + for {set i 0} {$i < $batch_size} {incr i} { + $rd read + } + } + } + set remaining [expr {$n % $batch_size}] + for {set j 0} {$j < $remaining} {incr j} { $rd read } + if {$type eq "cluster"} { + $rd config resetstat + $rd read ; # Discard config resetstat reply + } + $rd close + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_morethan [s allocator_frag_ratio] 1.4 + + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.05 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + } + # Flush all script to make sure we don't crash after defragging them + r script flush sync + } {OK} + + test "Active defrag big keys: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-max-scan-fields 1000 + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 2mb + r config set maxmemory 0 + r config set list-max-ziplist-size 5 ;# list of 10k items will have 2000 quicklist nodes + r config set stream-node-max-entries 5 + r config set hash-max-listpack-entries 10 + r hmset hash_lp h1 v1 h2 v2 h3 v3 + assert_encoding listpack hash_lp + r hmset hash_ht h1 v1 h2 v2 h3 v3 h4 v4 h5 v5 h6 v6 h7 v7 h8 v8 h9 v9 h10 v10 h11 v11 + assert_encoding hashtable hash_ht + r lpush list a b c d + r zadd zset 0 a 1 b 2 c 3 d + r sadd set a b c d + r xadd stream * item 1 value a + r xadd stream * item 2 value b + r xgroup create stream mygroup 0 + r xreadgroup GROUP mygroup Alice COUNT 1 STREAMS stream > + + # create big keys with 10k items + # Use batching to avoid TCP deadlock + set rd [redis_deferring_client] + set batch_size 1000 + for {set j 0} {$j < 10000} {incr j} { + $rd hset bighash $j [concat "asdfasdfasdf" $j] + $rd lpush biglist [concat "asdfasdfasdf" $j] + $rd zadd bigzset $j [concat "asdfasdfasdf" $j] + $rd sadd bigset [concat "asdfasdfasdf" $j] + $rd xadd bigstream * item 1 value a + if {($j + 1) % $batch_size == 0} { + for {set i 0} {$i < [expr {$batch_size * 5}]} {incr i} { + $rd read + } + } + } + # Read remaining replies + set remaining [expr {(10000 % $batch_size) * 5}] + for {set j 0} {$j < $remaining} {incr j} { + $rd read + } + + # create some small items (effective in cluster-enabled) + r set "{bighash}smallitem" val + r set "{biglist}smallitem" val + r set "{bigzset}smallitem" val + r set "{bigset}smallitem" val + r set "{bigstream}smallitem" val + + + set expected_frag 1.49 + if {$::accurate} { + # scale the hash to 1m fields in order to have a measurable the latency + set count 0 + for {set j 10000} {$j < 1000000} {incr j} { + $rd hset bighash $j [concat "asdfasdfasdf" $j] + + incr count + discard_replies_every $rd $count 10000 10000 + } + # creating that big hash, increased used_memory, so the relative frag goes down + set expected_frag 1.3 + } + + # add a mass of string keys + set count 0 + for {set j 0} {$j < 500000} {incr j} { + $rd setrange $j 150 a + + incr count + discard_replies_every $rd $count 10000 10000 + } + assert_equal [r dbsize] 500016 + + # create some fragmentation + set count 0 + for {set j 0} {$j < 500000} {incr j 2} { + $rd del $j + + incr count + discard_replies_every $rd $count 10000 10000 + } + assert_equal [r dbsize] 250016 + + # start defrag + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + if {$::verbose} { + puts "frag $frag" + } + assert {$frag >= $expected_frag} + r config set latency-monitor-threshold 5 + r latency reset + + set digest [debug_digest] + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.1 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + set max_latency 0 + foreach event [r latency latest] { + lassign $event eventname time latency max + if {$eventname == "active-defrag-cycle"} { + set max_latency $max + } + } + if {$::verbose} { + puts "frag $frag" + set misses [s active_defrag_misses] + set hits [s active_defrag_hits] + puts "hits: $hits" + puts "misses: $misses" + puts "max latency $max_latency" + puts [r latency latest] + puts [r latency history active-defrag-cycle] + } + # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, + # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher + if {!$::no_latency} { + assert {$max_latency <= 30} + } + } + # verify the data isn't corrupted or changed + set newdigest [debug_digest] + assert {$digest eq $newdigest} + r save ;# saving an rdb iterates over all the data / pointers + } {OK} + + test "Active defrag pubsub: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1500kb + r config set maxmemory 0 + + # Populate memory with interleaving pubsub-key pattern of same size + set n 50000 + set dummy_channel "[string repeat x 400]" + set rd [redis_deferring_client] + set rd_pubsub [redis_deferring_client] + for {set j 0} {$j < $n} {incr j} { + set channel_name "$dummy_channel[format "%06d" $j]" + $rd_pubsub subscribe $channel_name + $rd_pubsub read ; # Discard subscribe replies + $rd_pubsub ssubscribe $channel_name + $rd_pubsub read ; # Discard ssubscribe replies + # Pub/Sub clients are handled in the main thread, so their memory is + # allocated there. Using the SETBIT command avoids the main thread + # referencing argv from IO threads. + $rd setbit k$j [expr {[string length $channel_name] * 8}] 1 + $rd read ; # Discard set replies + } + + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_lessthan [s allocator_frag_ratio] 1.05 + + # Delete all the keys to create fragmentation + # Use batching to avoid TCP deadlock + set batch_size 1000 + for {set j 0} {$j < $n} {incr j} { + $rd del k$j + if {($j + 1) % $batch_size == 0} { + for {set i 0} {$i < $batch_size} {incr i} { + $rd read + } + } + } + set remaining [expr {$n % $batch_size}] + for {set j 0} {$j < $remaining} {incr j} { $rd read } + if {$type eq "cluster"} { + $rd config resetstat + $rd read ; # Discard config resetstat reply + } + $rd close + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_morethan [s allocator_frag_ratio] 1.35 + + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.05 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + } + + # Publishes some message to all the pubsub clients to make sure that + # we didn't break the data structure. + for {set j 0} {$j < $n} {incr j} { + set channel "$dummy_channel[format "%06d" $j]" + r publish $channel "hello" + assert_equal "message $channel hello" [$rd_pubsub read] + $rd_pubsub unsubscribe $channel + $rd_pubsub read + r spublish $channel "hello" + assert_equal "smessage $channel hello" [$rd_pubsub read] + $rd_pubsub sunsubscribe $channel + $rd_pubsub read + } + $rd_pubsub close + } + + test "Active defrag IDMP streams: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1500kb + r config set maxmemory 0 + + set n 50000 + + # Create the stream first and configure IDMP limits + r xadd idmpstream * dummy value + r xcfgset idmpstream idmp-maxsize 10000 ;# Allow 10000 entries per producer + + # Populate memory with interleaving IDMP stream-key pattern of same size + set dummy_iid "[string repeat x 400]" + set rd [redis_deferring_client] + for {set j 0} {$j < $n} {incr j} { + set producer_id "producer[expr {$j % 10}]" + set iid "$dummy_iid[format "%06d" $j]" + $rd xadd idmpstream IDMP $producer_id $iid * field value + $rd set k$j $iid + } + for {set j 0} {$j < [expr {$n * 2}]} {incr j} { + $rd read ; # Discard replies + } + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_lessthan [s allocator_frag_ratio] 1.05 + + # Verify IDMP structures were created + set idmp_info [r xinfo stream idmpstream full] + set num_producers [dict get $idmp_info pids-tracked] + set num_entries [dict get $idmp_info iids-tracked] + assert {$num_producers == 10} + assert {$num_entries == $n} + + # Delete all the keys to create fragmentation + for {set j 0} {$j < $n} {incr j} { $rd del k$j } + for {set j 0} {$j < $n} {incr j} { $rd read } ; # Discard del replies + if {$type eq "cluster"} { + $rd config resetstat + $rd read ; # Discard config resetstat reply + } + $rd close + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_morethan [s allocator_frag_ratio] 1.35 + + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.1 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + + # Verify IDMP structures are intact after defrag + set idmp_info_after [r xinfo stream idmpstream full] + set num_producers_after [dict get $idmp_info_after pids-tracked] + set num_entries_after [dict get $idmp_info_after iids-tracked] + assert {$num_producers_after == 10} + assert {$num_entries_after == $n} + + # Verify IDMP deduplication still works after defrag + set original_length [r xlen idmpstream] + r xadd idmpstream IDMP producer0 "${dummy_iid}000000" * field newvalue + set new_length [r xlen idmpstream] + assert {$new_length == $original_length} + } + } + + foreach {eb_container fields n} {eblist 16 3000 ebrax 30 1600 large_ebrax 500 100} { + test "Active Defrag HFE with $eb_container: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-threshold-lower 7 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1000kb + r config set maxmemory 0 + r config set hash-max-listpack-value 512 + r config set hash-max-listpack-entries 10 + + # Populate memory with interleaving hash field of same size + # Interleave reads to avoid TCP deadlock + set dummy_field "[string repeat x 400]" + set rd [redis_deferring_client] + for {set i 0} {$i < $n} {incr i} { + for {set j 0} {$j < $fields} {incr j} { + $rd hset h$i $dummy_field$j v + $rd hexpire h$i 9999999 FIELDS 1 $dummy_field$j + $rd hset k$i $dummy_field$j v + $rd hexpire k$i 9999999 FIELDS 1 $dummy_field$j + } + $rd expire h$i 9999999 ;# Ensure expire is updated after kvobj reallocation + # Read replies for this iteration to avoid TCP deadlock + for {set j 0} {$j < $fields} {incr j} { + $rd read ; # Discard hset replies + $rd read ; # Discard hexpire replies + $rd read ; # Discard hset replies + $rd read ; # Discard hexpire replies + } + $rd read ; # Discard expire replies + } + + # Coverage for listpackex. + r hset h_lpex $dummy_field v + r hexpire h_lpex 9999999 FIELDS 1 $dummy_field + assert_encoding listpackex h_lpex + + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_lessthan [s allocator_frag_ratio] 1.07 + + # Delete all the keys to create fragmentation + for {set i 0} {$i < $n} {incr i} { + r del k$i + } + $rd close + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_morethan [s allocator_frag_ratio] 1.35 + + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.07 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + } + } + } ;# end of foreach + + test "Active defrag for argv retained by the main thread from IO thread: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + set io_threads [lindex [r config get io-threads] 1] + if {$io_threads == 1} { + r config set active-defrag-threshold-lower 5 + } else { + r config set active-defrag-threshold-lower 10 + } + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1000kb + r config set maxmemory 0 + + # Create some clients so that they are distributed among different io threads. + set clients {} + for {set i 0} {$i < 8} {incr i} { + lappend clients [redis_client] + } + + # Populate memory with interleaving key pattern of same size + set dummy "[string repeat x 400]" + set n 10000 + for {set i 0} {$i < [llength $clients]} {incr i} { + set rr [lindex $clients $i] + for {set j 0} {$j < $n} {incr j} { + $rr set "k$i-$j" $dummy + } + } + + # If io-threads is enable, verify that memory allocation is not from the main thread. + if {$io_threads != 1} { + # At least make sure that bin 448 is created in the main thread's arena. + r set k dummy + r del k + + # We created 10000 string keys of 400 bytes each for each client, so when the memory + # allocation for the 448 bin in the main thread is significantly smaller than this, + # we can conclude that the memory allocation is not coming from it. + set malloc_stats [r memory malloc-stats] + if {[regexp {(?s)arenas\[0\]:.*?448[ ]+[\d]+[ ]+([\d]+)[ ]} $malloc_stats - allocated]} { + # Ensure the allocation for bin 448 in the main thread’s arena + # is far less than 4375k (10000 * 448 bytes). + assert_lessthan $allocated 200000 + } else { + fail "Failed to get the main thread's malloc stats." + } + } + + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_lessthan [s allocator_frag_ratio] 1.05 + + # Delete keys with even indices to create fragmentation. + for {set i 0} {$i < [llength $clients]} {incr i} { + set rd [lindex $clients $i] + for {set j 0} {$j < $n} {incr j 2} { + $rd del "k$i-$j" + } + } + for {set i 0} {$i < [llength $clients]} {incr i} { + [lindex $clients $i] close + } + if {$type eq "cluster"} { + r config resetstat + } + + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + assert_morethan [s allocator_frag_ratio] 1.35 + + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + if {$io_threads == 1} { + wait_for_defrag_stop 500 100 1.05 + } else { + # TODO: When multithreading is enabled, argv may be created in the io thread + # and kept in the main thread, which can cause fragmentation to become worse. + wait_for_defrag_stop 500 100 1.1 + } + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag [s allocator_frag_ratio]" + puts "frag_bytes [s allocator_frag_bytes]" + } + } + } + + if {$type eq "standalone"} { ;# skip in cluster mode + test "Active defrag big list: $type" { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-max-scan-fields 1000 + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 2mb + r config set maxmemory 0 + r config set list-max-ziplist-size 1 ;# list of 100k items will have 100k quicklist nodes + + # create big keys with 10k items + set rd [redis_deferring_client] + + set expected_frag 1.5 + # add a mass of list nodes to two lists (allocations are interlaced) + set val [string repeat A 500] ;# 1 item of 500 bytes puts us in the 640 bytes bin, which has 32 regs, so high potential for fragmentation + set elements 100000 + set count 0 + for {set j 0} {$j < $elements} {incr j} { + $rd lpush biglist1 $val + $rd lpush biglist2 $val + + incr count + discard_replies_every $rd $count 10000 20000 + } + + # create some fragmentation + r del biglist2 + + # start defrag + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + if {$::verbose} { + puts "frag $frag" + } + + assert {$frag >= $expected_frag} + r config set latency-monitor-threshold 5 + r latency reset + + set digest [debug_digest] + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.1 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + set misses [s active_defrag_misses] + set hits [s active_defrag_hits] + set frag [s allocator_frag_ratio] + set max_latency 0 + foreach event [r latency latest] { + lassign $event eventname time latency max + if {$eventname == "active-defrag-cycle"} { + set max_latency $max + } + } + if {$::verbose} { + puts "used [s allocator_allocated]" + puts "rss [s allocator_active]" + puts "frag_bytes [s allocator_frag_bytes]" + puts "frag $frag" + puts "misses: $misses" + puts "hits: $hits" + puts "max latency $max_latency" + puts [r latency latest] + puts [r latency history active-defrag-cycle] + puts [r memory malloc-stats] + } + # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, + # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher + if {!$::no_latency} { + assert {$max_latency <= 30} + } + + # in extreme cases of stagnation, we see over 5m misses before the tests aborts with "defrag didn't stop", + # in normal cases we only see 100k misses out of 100k elements + assert {$misses < $elements * 2} + } + # verify the data isn't corrupted or changed + set newdigest [debug_digest] + assert {$digest eq $newdigest} + r save ;# saving an rdb iterates over all the data / pointers + r del biglist1 ;# coverage for quicklistBookmarksClear + } {1} + + test "Active defrag edge case: $type" { + # there was an edge case in defrag where all the slabs of a certain bin are exact the same + # % utilization, with the exception of the current slab from which new allocations are made + # if the current slab is lower in utilization the defragger would have ended up in stagnation, + # kept running and not move any allocation. + # this test is more consistent on a fresh server with no history + start_server {tags {"defrag"} overrides {save ""}} { + r flushdb + r config set hz 100 + r config set activedefrag no + wait_for_defrag_stop 500 100 + r config resetstat + r config set active-defrag-max-scan-fields 1000 + r config set active-defrag-threshold-lower 5 + r config set active-defrag-cycle-min 65 + r config set active-defrag-cycle-max 75 + r config set active-defrag-ignore-bytes 1mb + r config set maxmemory 0 + set expected_frag 1.3 + + r debug mallctl-str thread.tcache.flush VOID + # fill the first slab containing 32 regs of 640 bytes. + for {set j 0} {$j < 32} {incr j} { + r setrange "_$j" 600 x + r debug mallctl-str thread.tcache.flush VOID + } + + # add a mass of keys with 600 bytes values, fill the bin of 640 bytes which has 32 regs per slab. + set rd [redis_deferring_client] + set keys 640000 + set count 0 + for {set j 0} {$j < $keys} {incr j} { + $rd setrange $j 600 x + + incr count + discard_replies_every $rd $count 10000 10000 + } + + # create some fragmentation of 50% + set sent 0 + for {set j 0} {$j < $keys} {incr j 1} { + $rd del $j + incr sent + incr j 1 + + discard_replies_every $rd $sent 10000 10000 + } + + # create higher fragmentation in the first slab + for {set j 10} {$j < 32} {incr j} { + r del "_$j" + } + + # start defrag + after 120 ;# serverCron only updates the info once in 100ms + set frag [s allocator_frag_ratio] + if {$::verbose} { + puts "frag $frag" + } + + assert {$frag >= $expected_frag} + + set digest [debug_digest] + catch {r config set activedefrag yes} e + if {[r config get activedefrag] eq "activedefrag yes"} { + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [r info memory] + puts [r info stats] + puts [r memory malloc-stats] + fail "defrag not started." + } + + # wait for the active defrag to stop working + wait_for_defrag_stop 500 100 1.1 + + # test the fragmentation is lower + after 120 ;# serverCron only updates the info once in 100ms + set misses [s active_defrag_misses] + set hits [s active_defrag_hits] + set frag [s allocator_frag_ratio] + if {$::verbose} { + puts "frag $frag" + puts "hits: $hits" + puts "misses: $misses" + } + assert {$misses < 10000000} ;# when defrag doesn't stop, we have some 30m misses, when it does, we have 2m misses + } + + # verify the data isn't corrupted or changed + set newdigest [debug_digest] + assert {$digest eq $newdigest} + r save ;# saving an rdb iterates over all the data / pointers + } + } ;# standalone + } + } + } + + test "Active defrag can't be triggered during replicaof database flush. See issue #14267" { + start_server {tags {"repl"} overrides {save ""}} { + set master_host [srv 0 host] + set master_port [srv 0 port] + + start_server {overrides {save ""}} { + set replica [srv 0 client] + set rd [redis_deferring_client 0] + + $replica config set hz 100 + $replica config set activedefrag no + $replica config set active-defrag-threshold-lower 5 + $replica config set active-defrag-cycle-min 65 + $replica config set active-defrag-cycle-max 75 + $replica config set active-defrag-ignore-bytes 2mb + + # add a mass of string keys + set count 0 + for {set j 0} {$j < 500000} {incr j} { + $rd setrange $j 150 a + + incr count + discard_replies_every $rd $count 10000 10000 + } + assert_equal [$replica dbsize] 500000 + + # create some fragmentation + set count 0 + for {set j 0} {$j < 500000} {incr j 2} { + $rd del $j + + incr count + discard_replies_every $rd $count 10000 10000 + } + $rd close + assert_equal [$replica dbsize] 250000 + + catch {$replica config set activedefrag yes} e + if {[$replica config get activedefrag] eq "activedefrag yes"} { + # Start replication sync which will flush the replica's database, + # then enable defrag to run concurrently with the database flush. + $replica replicaof $master_host $master_port + + # wait for the active defrag to start working (decision once a second) + wait_for_condition 50 100 { + [s total_active_defrag_time] ne 0 + } else { + after 120 ;# serverCron only updates the info once in 100ms + puts [$replica info memory] + puts [$replica info stats] + puts [$replica memory malloc-stats] + fail "defrag not started." + } + + wait_for_sync $replica + + # wait for the active defrag to stop working (db has been emptied during replication sync) + wait_for_defrag_stop 500 100 + assert_equal [$replica dbsize] 0 + } + } + } + } {} {defrag external:skip tsan:skip debug_defrag:skip cluster} + + start_cluster 1 0 {tags {"defrag external:skip tsan:skip debug_defrag:skip cluster needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} { + test_active_defrag "cluster" + } + + start_server {tags {"defrag external:skip tsan:skip debug_defrag:skip standalone needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} { + test_active_defrag "standalone" + } +} ;# run_solo -- cgit v1.2.3