diff options
Diffstat (limited to 'examples/redis-unstable/tests/unit/memefficiency.tcl')
| -rw-r--r-- | examples/redis-unstable/tests/unit/memefficiency.tcl | 1176 |
1 files changed, 0 insertions, 1176 deletions
diff --git a/examples/redis-unstable/tests/unit/memefficiency.tcl b/examples/redis-unstable/tests/unit/memefficiency.tcl deleted file mode 100644 index 123bf37..0000000 --- a/examples/redis-unstable/tests/unit/memefficiency.tcl +++ /dev/null @@ -1,1176 +0,0 @@ -# -# Copyright (c) 2009-Present, Redis Ltd. -# All rights reserved. -# -# Copyright (c) 2024-present, Valkey contributors. -# All rights reserved. -# -# Licensed under your choice of (a) the Redis Source Available License 2.0 -# (RSALv2); or (b) the Server Side Public License v1 (SSPLv1); or (c) the -# GNU Affero General Public License v3 (AGPLv3). -# -# Portions of this file are available under BSD3 terms; see REDISCONTRIBUTIONS for more information. -# - -proc test_memory_efficiency {range} { - r flushall - set rd [redis_deferring_client] - set base_mem [s used_memory] - set written 0 - for {set j 0} {$j < 10000} {incr j} { - set key key:$j - set val [string repeat A [expr {int(rand()*$range)}]] - $rd set $key $val - incr written [string length $key] - incr written [string length $val] - incr written 2 ;# A separator is the minimum to store key-value data. - } - for {set j 0} {$j < 10000} {incr j} { - $rd read ; # Discard replies - } - - set current_mem [s used_memory] - set used [expr {$current_mem-$base_mem}] - set efficiency [expr {double($written)/$used}] - return $efficiency -} - -start_server {tags {"memefficiency external:skip"}} { - foreach {size_range expected_min_efficiency} { - 32 0.15 - 64 0.25 - 128 0.35 - 1024 0.75 - 16384 0.82 - } { - test "Memory efficiency with values in range $size_range" { - set efficiency [test_memory_efficiency $size_range] - assert {$efficiency >= $expected_min_efficiency} - } - } -} - -run_solo {defrag} { - proc wait_for_defrag_stop {maxtries delay {expect_frag 0}} { - wait_for_condition $maxtries $delay { - [s active_defrag_running] eq 0 && ($expect_frag == 0 || [s allocator_frag_ratio] <= $expect_frag) - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - if {$expect_frag != 0} { - fail "defrag didn't stop or failed to achieve expected frag ratio ([s allocator_frag_ratio] > $expect_frag)" - } else { - fail "defrag didn't stop." - } - } - } - - proc discard_replies_every {rd count frequency discard_num} { - if {$count % $frequency == 0} { - for {set k 0} {$k < $discard_num} {incr k} { - $rd read ; # Discard replies - } - } - } - - proc test_active_defrag {type} { - - # note: Disabling lookahead because it changes the number and order of allocations which interferes with defrag and causes tests to fail - r config set lookahead 1 - - r debug reply-copy-avoidance 0 ;# Disable copy avoidance because it affects memory usage - - if {[string match {*jemalloc*} [s mem_allocator]] && [r debug mallctl arenas.page] <= 8192} { - test "Active defrag main dictionary: $type" { - r config set hz 100 - r config set activedefrag no - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 2mb - r config set maxmemory 100mb - r config set maxmemory-policy allkeys-lru - - populate 700000 asdf1 150 - populate 100 asdf1 150 0 false 1000 - populate 170000 asdf2 300 - populate 100 asdf2 300 0 false 1000 - - assert {[scan [regexp -inline {expires\=([\d]*)} [r info keyspace]] expires=%d] > 0} - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - if {$::verbose} { - puts "frag $frag" - } - assert {$frag >= 1.4} - - r config set latency-monitor-threshold 5 - r latency reset - r config set maxmemory 110mb ;# prevent further eviction (not to fail the digest test) - set digest [debug_digest] - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - # Wait for the active defrag to start working (decision once a - # second). - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # This test usually runs for a while, during this interval, we test the range. - assert_range [s active_defrag_running] 65 75 - r config set active-defrag-cycle-min 1 - r config set active-defrag-cycle-max 1 - after 120 ;# serverCron only updates the info once in 100ms - assert_range [s active_defrag_running] 1 1 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - - # Wait for the active defrag to stop working. - wait_for_defrag_stop 2000 100 1.1 - - # Test the fragmentation is lower. - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - set max_latency 0 - foreach event [r latency latest] { - lassign $event eventname time latency max - if {$eventname == "active-defrag-cycle"} { - set max_latency $max - } - } - if {$::verbose} { - puts "frag $frag" - set misses [s active_defrag_misses] - set hits [s active_defrag_hits] - puts "hits: $hits" - puts "misses: $misses" - puts "max latency $max_latency" - puts [r latency latest] - puts [r latency history active-defrag-cycle] - } - # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, - # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher - if {!$::no_latency} { - assert {$max_latency <= 30} - } - } - # verify the data isn't corrupted or changed - set newdigest [debug_digest] - assert {$digest eq $newdigest} - r save ;# saving an rdb iterates over all the data / pointers - - # if defrag is supported, test AOF loading too - if {[r config get activedefrag] eq "activedefrag yes" && $type eq "standalone"} { - test "Active defrag - AOF loading" { - # reset stats and load the AOF file - r config resetstat - r config set key-load-delay -25 ;# sleep on average 1/25 usec - # Note: This test is checking if defrag is working DURING AOF loading (while - # timers are not active). So we don't give any extra time, and we deactivate - # defrag immediately after the AOF loading is complete. During loading, - # defrag will get invoked less often, causing starvation prevention. We - # should expect longer latency measurements. - r debug loadaof - r config set activedefrag no - # measure hits and misses right after aof loading - set misses [s active_defrag_misses] - set hits [s active_defrag_hits] - - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - set max_latency 0 - foreach event [r latency latest] { - lassign $event eventname time latency max - if {$eventname == "while-blocked-cron"} { - set max_latency $max - } - } - if {$::verbose} { - puts "AOF loading:" - puts "frag $frag" - puts "hits: $hits" - puts "misses: $misses" - puts "max latency $max_latency" - puts [r latency latest] - puts [r latency history "while-blocked-cron"] - } - # make sure we had defrag hits during AOF loading - assert {$hits > 100000} - # make sure the defragger did enough work to keep the fragmentation low during loading. - # we cannot check that it went all the way down, since we don't wait for full defrag cycle to complete. - assert {$frag < 1.4} - # since the AOF contains simple (fast) SET commands (and the cron during loading runs every 1024 commands), - # it'll still not block the loading for long periods of time. - if {!$::no_latency} { - assert {$max_latency <= 40} - } - } - } ;# Active defrag - AOF loading - } - r config set appendonly no - r config set key-load-delay 0 - - test "Active defrag eval scripts: $type" { - r flushdb - r script flush sync - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1500kb - r config set maxmemory 0 - - set n 50000 - - # Populate memory with interleaving script-key pattern of same size - set dummy_script "--[string repeat x 400]\nreturn " - set rd [redis_deferring_client] - # Send commands in batches and read responses to avoid TCP deadlock. - # Without interleaving reads, TCP congestion control can throttle - # the connection when buffers fill, causing the test to hang. - set batch_size 1000 - for {set j 0} {$j < $n} {incr j} { - set val "$dummy_script[format "%06d" $j]" - $rd script load $val - $rd set k$j $val - if {($j + 1) % $batch_size == 0} { - for {set i 0} {$i < $batch_size} {incr i} { - $rd read ; # Discard script load replies - $rd read ; # Discard set replies - } - } - } - # Read remaining responses - set remaining [expr {$n % $batch_size}] - for {set j 0} {$j < $remaining} {incr j} { - $rd read ; # Discard script load replies - $rd read ; # Discard set replies - } - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_lessthan [s allocator_frag_ratio] 1.05 - - # Delete all the keys to create fragmentation - # Use same batching pattern to avoid TCP deadlock - for {set j 0} {$j < $n} {incr j} { - $rd del k$j - if {($j + 1) % $batch_size == 0} { - for {set i 0} {$i < $batch_size} {incr i} { - $rd read - } - } - } - set remaining [expr {$n % $batch_size}] - for {set j 0} {$j < $remaining} {incr j} { $rd read } - if {$type eq "cluster"} { - $rd config resetstat - $rd read ; # Discard config resetstat reply - } - $rd close - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_morethan [s allocator_frag_ratio] 1.4 - - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.05 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - } - # Flush all script to make sure we don't crash after defragging them - r script flush sync - } {OK} - - test "Active defrag big keys: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-max-scan-fields 1000 - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 2mb - r config set maxmemory 0 - r config set list-max-ziplist-size 5 ;# list of 10k items will have 2000 quicklist nodes - r config set stream-node-max-entries 5 - r config set hash-max-listpack-entries 10 - r hmset hash_lp h1 v1 h2 v2 h3 v3 - assert_encoding listpack hash_lp - r hmset hash_ht h1 v1 h2 v2 h3 v3 h4 v4 h5 v5 h6 v6 h7 v7 h8 v8 h9 v9 h10 v10 h11 v11 - assert_encoding hashtable hash_ht - r lpush list a b c d - r zadd zset 0 a 1 b 2 c 3 d - r sadd set a b c d - r xadd stream * item 1 value a - r xadd stream * item 2 value b - r xgroup create stream mygroup 0 - r xreadgroup GROUP mygroup Alice COUNT 1 STREAMS stream > - - # create big keys with 10k items - # Use batching to avoid TCP deadlock - set rd [redis_deferring_client] - set batch_size 1000 - for {set j 0} {$j < 10000} {incr j} { - $rd hset bighash $j [concat "asdfasdfasdf" $j] - $rd lpush biglist [concat "asdfasdfasdf" $j] - $rd zadd bigzset $j [concat "asdfasdfasdf" $j] - $rd sadd bigset [concat "asdfasdfasdf" $j] - $rd xadd bigstream * item 1 value a - if {($j + 1) % $batch_size == 0} { - for {set i 0} {$i < [expr {$batch_size * 5}]} {incr i} { - $rd read - } - } - } - # Read remaining replies - set remaining [expr {(10000 % $batch_size) * 5}] - for {set j 0} {$j < $remaining} {incr j} { - $rd read - } - - # create some small items (effective in cluster-enabled) - r set "{bighash}smallitem" val - r set "{biglist}smallitem" val - r set "{bigzset}smallitem" val - r set "{bigset}smallitem" val - r set "{bigstream}smallitem" val - - - set expected_frag 1.49 - if {$::accurate} { - # scale the hash to 1m fields in order to have a measurable the latency - set count 0 - for {set j 10000} {$j < 1000000} {incr j} { - $rd hset bighash $j [concat "asdfasdfasdf" $j] - - incr count - discard_replies_every $rd $count 10000 10000 - } - # creating that big hash, increased used_memory, so the relative frag goes down - set expected_frag 1.3 - } - - # add a mass of string keys - set count 0 - for {set j 0} {$j < 500000} {incr j} { - $rd setrange $j 150 a - - incr count - discard_replies_every $rd $count 10000 10000 - } - assert_equal [r dbsize] 500016 - - # create some fragmentation - set count 0 - for {set j 0} {$j < 500000} {incr j 2} { - $rd del $j - - incr count - discard_replies_every $rd $count 10000 10000 - } - assert_equal [r dbsize] 250016 - - # start defrag - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - if {$::verbose} { - puts "frag $frag" - } - assert {$frag >= $expected_frag} - r config set latency-monitor-threshold 5 - r latency reset - - set digest [debug_digest] - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.1 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - set max_latency 0 - foreach event [r latency latest] { - lassign $event eventname time latency max - if {$eventname == "active-defrag-cycle"} { - set max_latency $max - } - } - if {$::verbose} { - puts "frag $frag" - set misses [s active_defrag_misses] - set hits [s active_defrag_hits] - puts "hits: $hits" - puts "misses: $misses" - puts "max latency $max_latency" - puts [r latency latest] - puts [r latency history active-defrag-cycle] - } - # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, - # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher - if {!$::no_latency} { - assert {$max_latency <= 30} - } - } - # verify the data isn't corrupted or changed - set newdigest [debug_digest] - assert {$digest eq $newdigest} - r save ;# saving an rdb iterates over all the data / pointers - } {OK} - - test "Active defrag pubsub: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1500kb - r config set maxmemory 0 - - # Populate memory with interleaving pubsub-key pattern of same size - set n 50000 - set dummy_channel "[string repeat x 400]" - set rd [redis_deferring_client] - set rd_pubsub [redis_deferring_client] - for {set j 0} {$j < $n} {incr j} { - set channel_name "$dummy_channel[format "%06d" $j]" - $rd_pubsub subscribe $channel_name - $rd_pubsub read ; # Discard subscribe replies - $rd_pubsub ssubscribe $channel_name - $rd_pubsub read ; # Discard ssubscribe replies - # Pub/Sub clients are handled in the main thread, so their memory is - # allocated there. Using the SETBIT command avoids the main thread - # referencing argv from IO threads. - $rd setbit k$j [expr {[string length $channel_name] * 8}] 1 - $rd read ; # Discard set replies - } - - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_lessthan [s allocator_frag_ratio] 1.05 - - # Delete all the keys to create fragmentation - # Use batching to avoid TCP deadlock - set batch_size 1000 - for {set j 0} {$j < $n} {incr j} { - $rd del k$j - if {($j + 1) % $batch_size == 0} { - for {set i 0} {$i < $batch_size} {incr i} { - $rd read - } - } - } - set remaining [expr {$n % $batch_size}] - for {set j 0} {$j < $remaining} {incr j} { $rd read } - if {$type eq "cluster"} { - $rd config resetstat - $rd read ; # Discard config resetstat reply - } - $rd close - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_morethan [s allocator_frag_ratio] 1.35 - - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.05 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - } - - # Publishes some message to all the pubsub clients to make sure that - # we didn't break the data structure. - for {set j 0} {$j < $n} {incr j} { - set channel "$dummy_channel[format "%06d" $j]" - r publish $channel "hello" - assert_equal "message $channel hello" [$rd_pubsub read] - $rd_pubsub unsubscribe $channel - $rd_pubsub read - r spublish $channel "hello" - assert_equal "smessage $channel hello" [$rd_pubsub read] - $rd_pubsub sunsubscribe $channel - $rd_pubsub read - } - $rd_pubsub close - } - - test "Active defrag IDMP streams: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1500kb - r config set maxmemory 0 - - set n 50000 - - # Create the stream first and configure IDMP limits - r xadd idmpstream * dummy value - r xcfgset idmpstream idmp-maxsize 10000 ;# Allow 10000 entries per producer - - # Populate memory with interleaving IDMP stream-key pattern of same size - set dummy_iid "[string repeat x 400]" - set rd [redis_deferring_client] - for {set j 0} {$j < $n} {incr j} { - set producer_id "producer[expr {$j % 10}]" - set iid "$dummy_iid[format "%06d" $j]" - $rd xadd idmpstream IDMP $producer_id $iid * field value - $rd set k$j $iid - } - for {set j 0} {$j < [expr {$n * 2}]} {incr j} { - $rd read ; # Discard replies - } - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_lessthan [s allocator_frag_ratio] 1.05 - - # Verify IDMP structures were created - set idmp_info [r xinfo stream idmpstream full] - set num_producers [dict get $idmp_info pids-tracked] - set num_entries [dict get $idmp_info iids-tracked] - assert {$num_producers == 10} - assert {$num_entries == $n} - - # Delete all the keys to create fragmentation - for {set j 0} {$j < $n} {incr j} { $rd del k$j } - for {set j 0} {$j < $n} {incr j} { $rd read } ; # Discard del replies - if {$type eq "cluster"} { - $rd config resetstat - $rd read ; # Discard config resetstat reply - } - $rd close - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_morethan [s allocator_frag_ratio] 1.35 - - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.1 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - - # Verify IDMP structures are intact after defrag - set idmp_info_after [r xinfo stream idmpstream full] - set num_producers_after [dict get $idmp_info_after pids-tracked] - set num_entries_after [dict get $idmp_info_after iids-tracked] - assert {$num_producers_after == 10} - assert {$num_entries_after == $n} - - # Verify IDMP deduplication still works after defrag - set original_length [r xlen idmpstream] - r xadd idmpstream IDMP producer0 "${dummy_iid}000000" * field newvalue - set new_length [r xlen idmpstream] - assert {$new_length == $original_length} - } - } - - foreach {eb_container fields n} {eblist 16 3000 ebrax 30 1600 large_ebrax 500 100} { - test "Active Defrag HFE with $eb_container: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-threshold-lower 7 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1000kb - r config set maxmemory 0 - r config set hash-max-listpack-value 512 - r config set hash-max-listpack-entries 10 - - # Populate memory with interleaving hash field of same size - # Interleave reads to avoid TCP deadlock - set dummy_field "[string repeat x 400]" - set rd [redis_deferring_client] - for {set i 0} {$i < $n} {incr i} { - for {set j 0} {$j < $fields} {incr j} { - $rd hset h$i $dummy_field$j v - $rd hexpire h$i 9999999 FIELDS 1 $dummy_field$j - $rd hset k$i $dummy_field$j v - $rd hexpire k$i 9999999 FIELDS 1 $dummy_field$j - } - $rd expire h$i 9999999 ;# Ensure expire is updated after kvobj reallocation - # Read replies for this iteration to avoid TCP deadlock - for {set j 0} {$j < $fields} {incr j} { - $rd read ; # Discard hset replies - $rd read ; # Discard hexpire replies - $rd read ; # Discard hset replies - $rd read ; # Discard hexpire replies - } - $rd read ; # Discard expire replies - } - - # Coverage for listpackex. - r hset h_lpex $dummy_field v - r hexpire h_lpex 9999999 FIELDS 1 $dummy_field - assert_encoding listpackex h_lpex - - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_lessthan [s allocator_frag_ratio] 1.07 - - # Delete all the keys to create fragmentation - for {set i 0} {$i < $n} {incr i} { - r del k$i - } - $rd close - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_morethan [s allocator_frag_ratio] 1.35 - - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.07 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - } - } - } ;# end of foreach - - test "Active defrag for argv retained by the main thread from IO thread: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - set io_threads [lindex [r config get io-threads] 1] - if {$io_threads == 1} { - r config set active-defrag-threshold-lower 5 - } else { - r config set active-defrag-threshold-lower 10 - } - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1000kb - r config set maxmemory 0 - - # Create some clients so that they are distributed among different io threads. - set clients {} - for {set i 0} {$i < 8} {incr i} { - lappend clients [redis_client] - } - - # Populate memory with interleaving key pattern of same size - set dummy "[string repeat x 400]" - set n 10000 - for {set i 0} {$i < [llength $clients]} {incr i} { - set rr [lindex $clients $i] - for {set j 0} {$j < $n} {incr j} { - $rr set "k$i-$j" $dummy - } - } - - # If io-threads is enable, verify that memory allocation is not from the main thread. - if {$io_threads != 1} { - # At least make sure that bin 448 is created in the main thread's arena. - r set k dummy - r del k - - # We created 10000 string keys of 400 bytes each for each client, so when the memory - # allocation for the 448 bin in the main thread is significantly smaller than this, - # we can conclude that the memory allocation is not coming from it. - set malloc_stats [r memory malloc-stats] - if {[regexp {(?s)arenas\[0\]:.*?448[ ]+[\d]+[ ]+([\d]+)[ ]} $malloc_stats - allocated]} { - # Ensure the allocation for bin 448 in the main thread’s arena - # is far less than 4375k (10000 * 448 bytes). - assert_lessthan $allocated 200000 - } else { - fail "Failed to get the main thread's malloc stats." - } - } - - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_lessthan [s allocator_frag_ratio] 1.05 - - # Delete keys with even indices to create fragmentation. - for {set i 0} {$i < [llength $clients]} {incr i} { - set rd [lindex $clients $i] - for {set j 0} {$j < $n} {incr j 2} { - $rd del "k$i-$j" - } - } - for {set i 0} {$i < [llength $clients]} {incr i} { - [lindex $clients $i] close - } - if {$type eq "cluster"} { - r config resetstat - } - - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - assert_morethan [s allocator_frag_ratio] 1.35 - - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - if {$io_threads == 1} { - wait_for_defrag_stop 500 100 1.05 - } else { - # TODO: When multithreading is enabled, argv may be created in the io thread - # and kept in the main thread, which can cause fragmentation to become worse. - wait_for_defrag_stop 500 100 1.1 - } - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag [s allocator_frag_ratio]" - puts "frag_bytes [s allocator_frag_bytes]" - } - } - } - - if {$type eq "standalone"} { ;# skip in cluster mode - test "Active defrag big list: $type" { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-max-scan-fields 1000 - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 2mb - r config set maxmemory 0 - r config set list-max-ziplist-size 1 ;# list of 100k items will have 100k quicklist nodes - - # create big keys with 10k items - set rd [redis_deferring_client] - - set expected_frag 1.5 - # add a mass of list nodes to two lists (allocations are interlaced) - set val [string repeat A 500] ;# 1 item of 500 bytes puts us in the 640 bytes bin, which has 32 regs, so high potential for fragmentation - set elements 100000 - set count 0 - for {set j 0} {$j < $elements} {incr j} { - $rd lpush biglist1 $val - $rd lpush biglist2 $val - - incr count - discard_replies_every $rd $count 10000 20000 - } - - # create some fragmentation - r del biglist2 - - # start defrag - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - if {$::verbose} { - puts "frag $frag" - } - - assert {$frag >= $expected_frag} - r config set latency-monitor-threshold 5 - r latency reset - - set digest [debug_digest] - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.1 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - set misses [s active_defrag_misses] - set hits [s active_defrag_hits] - set frag [s allocator_frag_ratio] - set max_latency 0 - foreach event [r latency latest] { - lassign $event eventname time latency max - if {$eventname == "active-defrag-cycle"} { - set max_latency $max - } - } - if {$::verbose} { - puts "used [s allocator_allocated]" - puts "rss [s allocator_active]" - puts "frag_bytes [s allocator_frag_bytes]" - puts "frag $frag" - puts "misses: $misses" - puts "hits: $hits" - puts "max latency $max_latency" - puts [r latency latest] - puts [r latency history active-defrag-cycle] - puts [r memory malloc-stats] - } - # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75, - # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher - if {!$::no_latency} { - assert {$max_latency <= 30} - } - - # in extreme cases of stagnation, we see over 5m misses before the tests aborts with "defrag didn't stop", - # in normal cases we only see 100k misses out of 100k elements - assert {$misses < $elements * 2} - } - # verify the data isn't corrupted or changed - set newdigest [debug_digest] - assert {$digest eq $newdigest} - r save ;# saving an rdb iterates over all the data / pointers - r del biglist1 ;# coverage for quicklistBookmarksClear - } {1} - - test "Active defrag edge case: $type" { - # there was an edge case in defrag where all the slabs of a certain bin are exact the same - # % utilization, with the exception of the current slab from which new allocations are made - # if the current slab is lower in utilization the defragger would have ended up in stagnation, - # kept running and not move any allocation. - # this test is more consistent on a fresh server with no history - start_server {tags {"defrag"} overrides {save ""}} { - r flushdb - r config set hz 100 - r config set activedefrag no - wait_for_defrag_stop 500 100 - r config resetstat - r config set active-defrag-max-scan-fields 1000 - r config set active-defrag-threshold-lower 5 - r config set active-defrag-cycle-min 65 - r config set active-defrag-cycle-max 75 - r config set active-defrag-ignore-bytes 1mb - r config set maxmemory 0 - set expected_frag 1.3 - - r debug mallctl-str thread.tcache.flush VOID - # fill the first slab containing 32 regs of 640 bytes. - for {set j 0} {$j < 32} {incr j} { - r setrange "_$j" 600 x - r debug mallctl-str thread.tcache.flush VOID - } - - # add a mass of keys with 600 bytes values, fill the bin of 640 bytes which has 32 regs per slab. - set rd [redis_deferring_client] - set keys 640000 - set count 0 - for {set j 0} {$j < $keys} {incr j} { - $rd setrange $j 600 x - - incr count - discard_replies_every $rd $count 10000 10000 - } - - # create some fragmentation of 50% - set sent 0 - for {set j 0} {$j < $keys} {incr j 1} { - $rd del $j - incr sent - incr j 1 - - discard_replies_every $rd $sent 10000 10000 - } - - # create higher fragmentation in the first slab - for {set j 10} {$j < 32} {incr j} { - r del "_$j" - } - - # start defrag - after 120 ;# serverCron only updates the info once in 100ms - set frag [s allocator_frag_ratio] - if {$::verbose} { - puts "frag $frag" - } - - assert {$frag >= $expected_frag} - - set digest [debug_digest] - catch {r config set activedefrag yes} e - if {[r config get activedefrag] eq "activedefrag yes"} { - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [r info memory] - puts [r info stats] - puts [r memory malloc-stats] - fail "defrag not started." - } - - # wait for the active defrag to stop working - wait_for_defrag_stop 500 100 1.1 - - # test the fragmentation is lower - after 120 ;# serverCron only updates the info once in 100ms - set misses [s active_defrag_misses] - set hits [s active_defrag_hits] - set frag [s allocator_frag_ratio] - if {$::verbose} { - puts "frag $frag" - puts "hits: $hits" - puts "misses: $misses" - } - assert {$misses < 10000000} ;# when defrag doesn't stop, we have some 30m misses, when it does, we have 2m misses - } - - # verify the data isn't corrupted or changed - set newdigest [debug_digest] - assert {$digest eq $newdigest} - r save ;# saving an rdb iterates over all the data / pointers - } - } ;# standalone - } - } - } - - test "Active defrag can't be triggered during replicaof database flush. See issue #14267" { - start_server {tags {"repl"} overrides {save ""}} { - set master_host [srv 0 host] - set master_port [srv 0 port] - - start_server {overrides {save ""}} { - set replica [srv 0 client] - set rd [redis_deferring_client 0] - - $replica config set hz 100 - $replica config set activedefrag no - $replica config set active-defrag-threshold-lower 5 - $replica config set active-defrag-cycle-min 65 - $replica config set active-defrag-cycle-max 75 - $replica config set active-defrag-ignore-bytes 2mb - - # add a mass of string keys - set count 0 - for {set j 0} {$j < 500000} {incr j} { - $rd setrange $j 150 a - - incr count - discard_replies_every $rd $count 10000 10000 - } - assert_equal [$replica dbsize] 500000 - - # create some fragmentation - set count 0 - for {set j 0} {$j < 500000} {incr j 2} { - $rd del $j - - incr count - discard_replies_every $rd $count 10000 10000 - } - $rd close - assert_equal [$replica dbsize] 250000 - - catch {$replica config set activedefrag yes} e - if {[$replica config get activedefrag] eq "activedefrag yes"} { - # Start replication sync which will flush the replica's database, - # then enable defrag to run concurrently with the database flush. - $replica replicaof $master_host $master_port - - # wait for the active defrag to start working (decision once a second) - wait_for_condition 50 100 { - [s total_active_defrag_time] ne 0 - } else { - after 120 ;# serverCron only updates the info once in 100ms - puts [$replica info memory] - puts [$replica info stats] - puts [$replica memory malloc-stats] - fail "defrag not started." - } - - wait_for_sync $replica - - # wait for the active defrag to stop working (db has been emptied during replication sync) - wait_for_defrag_stop 500 100 - assert_equal [$replica dbsize] 0 - } - } - } - } {} {defrag external:skip tsan:skip debug_defrag:skip cluster} - - start_cluster 1 0 {tags {"defrag external:skip tsan:skip debug_defrag:skip cluster needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} { - test_active_defrag "cluster" - } - - start_server {tags {"defrag external:skip tsan:skip debug_defrag:skip standalone needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} { - test_active_defrag "standalone" - } -} ;# run_solo |
