summaryrefslogtreecommitdiff
path: root/examples/redis-unstable/tests/unit/memefficiency.tcl
diff options
context:
space:
mode:
Diffstat (limited to 'examples/redis-unstable/tests/unit/memefficiency.tcl')
-rw-r--r--examples/redis-unstable/tests/unit/memefficiency.tcl1176
1 files changed, 0 insertions, 1176 deletions
diff --git a/examples/redis-unstable/tests/unit/memefficiency.tcl b/examples/redis-unstable/tests/unit/memefficiency.tcl
deleted file mode 100644
index 123bf37..0000000
--- a/examples/redis-unstable/tests/unit/memefficiency.tcl
+++ /dev/null
@@ -1,1176 +0,0 @@
-#
-# Copyright (c) 2009-Present, Redis Ltd.
-# All rights reserved.
-#
-# Copyright (c) 2024-present, Valkey contributors.
-# All rights reserved.
-#
-# Licensed under your choice of (a) the Redis Source Available License 2.0
-# (RSALv2); or (b) the Server Side Public License v1 (SSPLv1); or (c) the
-# GNU Affero General Public License v3 (AGPLv3).
-#
-# Portions of this file are available under BSD3 terms; see REDISCONTRIBUTIONS for more information.
-#
-
-proc test_memory_efficiency {range} {
- r flushall
- set rd [redis_deferring_client]
- set base_mem [s used_memory]
- set written 0
- for {set j 0} {$j < 10000} {incr j} {
- set key key:$j
- set val [string repeat A [expr {int(rand()*$range)}]]
- $rd set $key $val
- incr written [string length $key]
- incr written [string length $val]
- incr written 2 ;# A separator is the minimum to store key-value data.
- }
- for {set j 0} {$j < 10000} {incr j} {
- $rd read ; # Discard replies
- }
-
- set current_mem [s used_memory]
- set used [expr {$current_mem-$base_mem}]
- set efficiency [expr {double($written)/$used}]
- return $efficiency
-}
-
-start_server {tags {"memefficiency external:skip"}} {
- foreach {size_range expected_min_efficiency} {
- 32 0.15
- 64 0.25
- 128 0.35
- 1024 0.75
- 16384 0.82
- } {
- test "Memory efficiency with values in range $size_range" {
- set efficiency [test_memory_efficiency $size_range]
- assert {$efficiency >= $expected_min_efficiency}
- }
- }
-}
-
-run_solo {defrag} {
- proc wait_for_defrag_stop {maxtries delay {expect_frag 0}} {
- wait_for_condition $maxtries $delay {
- [s active_defrag_running] eq 0 && ($expect_frag == 0 || [s allocator_frag_ratio] <= $expect_frag)
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- if {$expect_frag != 0} {
- fail "defrag didn't stop or failed to achieve expected frag ratio ([s allocator_frag_ratio] > $expect_frag)"
- } else {
- fail "defrag didn't stop."
- }
- }
- }
-
- proc discard_replies_every {rd count frequency discard_num} {
- if {$count % $frequency == 0} {
- for {set k 0} {$k < $discard_num} {incr k} {
- $rd read ; # Discard replies
- }
- }
- }
-
- proc test_active_defrag {type} {
-
- # note: Disabling lookahead because it changes the number and order of allocations which interferes with defrag and causes tests to fail
- r config set lookahead 1
-
- r debug reply-copy-avoidance 0 ;# Disable copy avoidance because it affects memory usage
-
- if {[string match {*jemalloc*} [s mem_allocator]] && [r debug mallctl arenas.page] <= 8192} {
- test "Active defrag main dictionary: $type" {
- r config set hz 100
- r config set activedefrag no
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 2mb
- r config set maxmemory 100mb
- r config set maxmemory-policy allkeys-lru
-
- populate 700000 asdf1 150
- populate 100 asdf1 150 0 false 1000
- populate 170000 asdf2 300
- populate 100 asdf2 300 0 false 1000
-
- assert {[scan [regexp -inline {expires\=([\d]*)} [r info keyspace]] expires=%d] > 0}
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- if {$::verbose} {
- puts "frag $frag"
- }
- assert {$frag >= 1.4}
-
- r config set latency-monitor-threshold 5
- r latency reset
- r config set maxmemory 110mb ;# prevent further eviction (not to fail the digest test)
- set digest [debug_digest]
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
- # Wait for the active defrag to start working (decision once a
- # second).
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # This test usually runs for a while, during this interval, we test the range.
- assert_range [s active_defrag_running] 65 75
- r config set active-defrag-cycle-min 1
- r config set active-defrag-cycle-max 1
- after 120 ;# serverCron only updates the info once in 100ms
- assert_range [s active_defrag_running] 1 1
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
-
- # Wait for the active defrag to stop working.
- wait_for_defrag_stop 2000 100 1.1
-
- # Test the fragmentation is lower.
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- set max_latency 0
- foreach event [r latency latest] {
- lassign $event eventname time latency max
- if {$eventname == "active-defrag-cycle"} {
- set max_latency $max
- }
- }
- if {$::verbose} {
- puts "frag $frag"
- set misses [s active_defrag_misses]
- set hits [s active_defrag_hits]
- puts "hits: $hits"
- puts "misses: $misses"
- puts "max latency $max_latency"
- puts [r latency latest]
- puts [r latency history active-defrag-cycle]
- }
- # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75,
- # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher
- if {!$::no_latency} {
- assert {$max_latency <= 30}
- }
- }
- # verify the data isn't corrupted or changed
- set newdigest [debug_digest]
- assert {$digest eq $newdigest}
- r save ;# saving an rdb iterates over all the data / pointers
-
- # if defrag is supported, test AOF loading too
- if {[r config get activedefrag] eq "activedefrag yes" && $type eq "standalone"} {
- test "Active defrag - AOF loading" {
- # reset stats and load the AOF file
- r config resetstat
- r config set key-load-delay -25 ;# sleep on average 1/25 usec
- # Note: This test is checking if defrag is working DURING AOF loading (while
- # timers are not active). So we don't give any extra time, and we deactivate
- # defrag immediately after the AOF loading is complete. During loading,
- # defrag will get invoked less often, causing starvation prevention. We
- # should expect longer latency measurements.
- r debug loadaof
- r config set activedefrag no
- # measure hits and misses right after aof loading
- set misses [s active_defrag_misses]
- set hits [s active_defrag_hits]
-
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- set max_latency 0
- foreach event [r latency latest] {
- lassign $event eventname time latency max
- if {$eventname == "while-blocked-cron"} {
- set max_latency $max
- }
- }
- if {$::verbose} {
- puts "AOF loading:"
- puts "frag $frag"
- puts "hits: $hits"
- puts "misses: $misses"
- puts "max latency $max_latency"
- puts [r latency latest]
- puts [r latency history "while-blocked-cron"]
- }
- # make sure we had defrag hits during AOF loading
- assert {$hits > 100000}
- # make sure the defragger did enough work to keep the fragmentation low during loading.
- # we cannot check that it went all the way down, since we don't wait for full defrag cycle to complete.
- assert {$frag < 1.4}
- # since the AOF contains simple (fast) SET commands (and the cron during loading runs every 1024 commands),
- # it'll still not block the loading for long periods of time.
- if {!$::no_latency} {
- assert {$max_latency <= 40}
- }
- }
- } ;# Active defrag - AOF loading
- }
- r config set appendonly no
- r config set key-load-delay 0
-
- test "Active defrag eval scripts: $type" {
- r flushdb
- r script flush sync
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1500kb
- r config set maxmemory 0
-
- set n 50000
-
- # Populate memory with interleaving script-key pattern of same size
- set dummy_script "--[string repeat x 400]\nreturn "
- set rd [redis_deferring_client]
- # Send commands in batches and read responses to avoid TCP deadlock.
- # Without interleaving reads, TCP congestion control can throttle
- # the connection when buffers fill, causing the test to hang.
- set batch_size 1000
- for {set j 0} {$j < $n} {incr j} {
- set val "$dummy_script[format "%06d" $j]"
- $rd script load $val
- $rd set k$j $val
- if {($j + 1) % $batch_size == 0} {
- for {set i 0} {$i < $batch_size} {incr i} {
- $rd read ; # Discard script load replies
- $rd read ; # Discard set replies
- }
- }
- }
- # Read remaining responses
- set remaining [expr {$n % $batch_size}]
- for {set j 0} {$j < $remaining} {incr j} {
- $rd read ; # Discard script load replies
- $rd read ; # Discard set replies
- }
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_lessthan [s allocator_frag_ratio] 1.05
-
- # Delete all the keys to create fragmentation
- # Use same batching pattern to avoid TCP deadlock
- for {set j 0} {$j < $n} {incr j} {
- $rd del k$j
- if {($j + 1) % $batch_size == 0} {
- for {set i 0} {$i < $batch_size} {incr i} {
- $rd read
- }
- }
- }
- set remaining [expr {$n % $batch_size}]
- for {set j 0} {$j < $remaining} {incr j} { $rd read }
- if {$type eq "cluster"} {
- $rd config resetstat
- $rd read ; # Discard config resetstat reply
- }
- $rd close
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_morethan [s allocator_frag_ratio] 1.4
-
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.05
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- }
- # Flush all script to make sure we don't crash after defragging them
- r script flush sync
- } {OK}
-
- test "Active defrag big keys: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-max-scan-fields 1000
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 2mb
- r config set maxmemory 0
- r config set list-max-ziplist-size 5 ;# list of 10k items will have 2000 quicklist nodes
- r config set stream-node-max-entries 5
- r config set hash-max-listpack-entries 10
- r hmset hash_lp h1 v1 h2 v2 h3 v3
- assert_encoding listpack hash_lp
- r hmset hash_ht h1 v1 h2 v2 h3 v3 h4 v4 h5 v5 h6 v6 h7 v7 h8 v8 h9 v9 h10 v10 h11 v11
- assert_encoding hashtable hash_ht
- r lpush list a b c d
- r zadd zset 0 a 1 b 2 c 3 d
- r sadd set a b c d
- r xadd stream * item 1 value a
- r xadd stream * item 2 value b
- r xgroup create stream mygroup 0
- r xreadgroup GROUP mygroup Alice COUNT 1 STREAMS stream >
-
- # create big keys with 10k items
- # Use batching to avoid TCP deadlock
- set rd [redis_deferring_client]
- set batch_size 1000
- for {set j 0} {$j < 10000} {incr j} {
- $rd hset bighash $j [concat "asdfasdfasdf" $j]
- $rd lpush biglist [concat "asdfasdfasdf" $j]
- $rd zadd bigzset $j [concat "asdfasdfasdf" $j]
- $rd sadd bigset [concat "asdfasdfasdf" $j]
- $rd xadd bigstream * item 1 value a
- if {($j + 1) % $batch_size == 0} {
- for {set i 0} {$i < [expr {$batch_size * 5}]} {incr i} {
- $rd read
- }
- }
- }
- # Read remaining replies
- set remaining [expr {(10000 % $batch_size) * 5}]
- for {set j 0} {$j < $remaining} {incr j} {
- $rd read
- }
-
- # create some small items (effective in cluster-enabled)
- r set "{bighash}smallitem" val
- r set "{biglist}smallitem" val
- r set "{bigzset}smallitem" val
- r set "{bigset}smallitem" val
- r set "{bigstream}smallitem" val
-
-
- set expected_frag 1.49
- if {$::accurate} {
- # scale the hash to 1m fields in order to have a measurable the latency
- set count 0
- for {set j 10000} {$j < 1000000} {incr j} {
- $rd hset bighash $j [concat "asdfasdfasdf" $j]
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
- # creating that big hash, increased used_memory, so the relative frag goes down
- set expected_frag 1.3
- }
-
- # add a mass of string keys
- set count 0
- for {set j 0} {$j < 500000} {incr j} {
- $rd setrange $j 150 a
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
- assert_equal [r dbsize] 500016
-
- # create some fragmentation
- set count 0
- for {set j 0} {$j < 500000} {incr j 2} {
- $rd del $j
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
- assert_equal [r dbsize] 250016
-
- # start defrag
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- if {$::verbose} {
- puts "frag $frag"
- }
- assert {$frag >= $expected_frag}
- r config set latency-monitor-threshold 5
- r latency reset
-
- set digest [debug_digest]
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.1
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- set max_latency 0
- foreach event [r latency latest] {
- lassign $event eventname time latency max
- if {$eventname == "active-defrag-cycle"} {
- set max_latency $max
- }
- }
- if {$::verbose} {
- puts "frag $frag"
- set misses [s active_defrag_misses]
- set hits [s active_defrag_hits]
- puts "hits: $hits"
- puts "misses: $misses"
- puts "max latency $max_latency"
- puts [r latency latest]
- puts [r latency history active-defrag-cycle]
- }
- # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75,
- # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher
- if {!$::no_latency} {
- assert {$max_latency <= 30}
- }
- }
- # verify the data isn't corrupted or changed
- set newdigest [debug_digest]
- assert {$digest eq $newdigest}
- r save ;# saving an rdb iterates over all the data / pointers
- } {OK}
-
- test "Active defrag pubsub: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1500kb
- r config set maxmemory 0
-
- # Populate memory with interleaving pubsub-key pattern of same size
- set n 50000
- set dummy_channel "[string repeat x 400]"
- set rd [redis_deferring_client]
- set rd_pubsub [redis_deferring_client]
- for {set j 0} {$j < $n} {incr j} {
- set channel_name "$dummy_channel[format "%06d" $j]"
- $rd_pubsub subscribe $channel_name
- $rd_pubsub read ; # Discard subscribe replies
- $rd_pubsub ssubscribe $channel_name
- $rd_pubsub read ; # Discard ssubscribe replies
- # Pub/Sub clients are handled in the main thread, so their memory is
- # allocated there. Using the SETBIT command avoids the main thread
- # referencing argv from IO threads.
- $rd setbit k$j [expr {[string length $channel_name] * 8}] 1
- $rd read ; # Discard set replies
- }
-
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_lessthan [s allocator_frag_ratio] 1.05
-
- # Delete all the keys to create fragmentation
- # Use batching to avoid TCP deadlock
- set batch_size 1000
- for {set j 0} {$j < $n} {incr j} {
- $rd del k$j
- if {($j + 1) % $batch_size == 0} {
- for {set i 0} {$i < $batch_size} {incr i} {
- $rd read
- }
- }
- }
- set remaining [expr {$n % $batch_size}]
- for {set j 0} {$j < $remaining} {incr j} { $rd read }
- if {$type eq "cluster"} {
- $rd config resetstat
- $rd read ; # Discard config resetstat reply
- }
- $rd close
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_morethan [s allocator_frag_ratio] 1.35
-
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.05
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- }
-
- # Publishes some message to all the pubsub clients to make sure that
- # we didn't break the data structure.
- for {set j 0} {$j < $n} {incr j} {
- set channel "$dummy_channel[format "%06d" $j]"
- r publish $channel "hello"
- assert_equal "message $channel hello" [$rd_pubsub read]
- $rd_pubsub unsubscribe $channel
- $rd_pubsub read
- r spublish $channel "hello"
- assert_equal "smessage $channel hello" [$rd_pubsub read]
- $rd_pubsub sunsubscribe $channel
- $rd_pubsub read
- }
- $rd_pubsub close
- }
-
- test "Active defrag IDMP streams: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1500kb
- r config set maxmemory 0
-
- set n 50000
-
- # Create the stream first and configure IDMP limits
- r xadd idmpstream * dummy value
- r xcfgset idmpstream idmp-maxsize 10000 ;# Allow 10000 entries per producer
-
- # Populate memory with interleaving IDMP stream-key pattern of same size
- set dummy_iid "[string repeat x 400]"
- set rd [redis_deferring_client]
- for {set j 0} {$j < $n} {incr j} {
- set producer_id "producer[expr {$j % 10}]"
- set iid "$dummy_iid[format "%06d" $j]"
- $rd xadd idmpstream IDMP $producer_id $iid * field value
- $rd set k$j $iid
- }
- for {set j 0} {$j < [expr {$n * 2}]} {incr j} {
- $rd read ; # Discard replies
- }
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_lessthan [s allocator_frag_ratio] 1.05
-
- # Verify IDMP structures were created
- set idmp_info [r xinfo stream idmpstream full]
- set num_producers [dict get $idmp_info pids-tracked]
- set num_entries [dict get $idmp_info iids-tracked]
- assert {$num_producers == 10}
- assert {$num_entries == $n}
-
- # Delete all the keys to create fragmentation
- for {set j 0} {$j < $n} {incr j} { $rd del k$j }
- for {set j 0} {$j < $n} {incr j} { $rd read } ; # Discard del replies
- if {$type eq "cluster"} {
- $rd config resetstat
- $rd read ; # Discard config resetstat reply
- }
- $rd close
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_morethan [s allocator_frag_ratio] 1.35
-
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.1
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
-
- # Verify IDMP structures are intact after defrag
- set idmp_info_after [r xinfo stream idmpstream full]
- set num_producers_after [dict get $idmp_info_after pids-tracked]
- set num_entries_after [dict get $idmp_info_after iids-tracked]
- assert {$num_producers_after == 10}
- assert {$num_entries_after == $n}
-
- # Verify IDMP deduplication still works after defrag
- set original_length [r xlen idmpstream]
- r xadd idmpstream IDMP producer0 "${dummy_iid}000000" * field newvalue
- set new_length [r xlen idmpstream]
- assert {$new_length == $original_length}
- }
- }
-
- foreach {eb_container fields n} {eblist 16 3000 ebrax 30 1600 large_ebrax 500 100} {
- test "Active Defrag HFE with $eb_container: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-threshold-lower 7
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1000kb
- r config set maxmemory 0
- r config set hash-max-listpack-value 512
- r config set hash-max-listpack-entries 10
-
- # Populate memory with interleaving hash field of same size
- # Interleave reads to avoid TCP deadlock
- set dummy_field "[string repeat x 400]"
- set rd [redis_deferring_client]
- for {set i 0} {$i < $n} {incr i} {
- for {set j 0} {$j < $fields} {incr j} {
- $rd hset h$i $dummy_field$j v
- $rd hexpire h$i 9999999 FIELDS 1 $dummy_field$j
- $rd hset k$i $dummy_field$j v
- $rd hexpire k$i 9999999 FIELDS 1 $dummy_field$j
- }
- $rd expire h$i 9999999 ;# Ensure expire is updated after kvobj reallocation
- # Read replies for this iteration to avoid TCP deadlock
- for {set j 0} {$j < $fields} {incr j} {
- $rd read ; # Discard hset replies
- $rd read ; # Discard hexpire replies
- $rd read ; # Discard hset replies
- $rd read ; # Discard hexpire replies
- }
- $rd read ; # Discard expire replies
- }
-
- # Coverage for listpackex.
- r hset h_lpex $dummy_field v
- r hexpire h_lpex 9999999 FIELDS 1 $dummy_field
- assert_encoding listpackex h_lpex
-
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_lessthan [s allocator_frag_ratio] 1.07
-
- # Delete all the keys to create fragmentation
- for {set i 0} {$i < $n} {incr i} {
- r del k$i
- }
- $rd close
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_morethan [s allocator_frag_ratio] 1.35
-
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.07
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- }
- }
- } ;# end of foreach
-
- test "Active defrag for argv retained by the main thread from IO thread: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- set io_threads [lindex [r config get io-threads] 1]
- if {$io_threads == 1} {
- r config set active-defrag-threshold-lower 5
- } else {
- r config set active-defrag-threshold-lower 10
- }
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1000kb
- r config set maxmemory 0
-
- # Create some clients so that they are distributed among different io threads.
- set clients {}
- for {set i 0} {$i < 8} {incr i} {
- lappend clients [redis_client]
- }
-
- # Populate memory with interleaving key pattern of same size
- set dummy "[string repeat x 400]"
- set n 10000
- for {set i 0} {$i < [llength $clients]} {incr i} {
- set rr [lindex $clients $i]
- for {set j 0} {$j < $n} {incr j} {
- $rr set "k$i-$j" $dummy
- }
- }
-
- # If io-threads is enable, verify that memory allocation is not from the main thread.
- if {$io_threads != 1} {
- # At least make sure that bin 448 is created in the main thread's arena.
- r set k dummy
- r del k
-
- # We created 10000 string keys of 400 bytes each for each client, so when the memory
- # allocation for the 448 bin in the main thread is significantly smaller than this,
- # we can conclude that the memory allocation is not coming from it.
- set malloc_stats [r memory malloc-stats]
- if {[regexp {(?s)arenas\[0\]:.*?448[ ]+[\d]+[ ]+([\d]+)[ ]} $malloc_stats - allocated]} {
- # Ensure the allocation for bin 448 in the main thread’s arena
- # is far less than 4375k (10000 * 448 bytes).
- assert_lessthan $allocated 200000
- } else {
- fail "Failed to get the main thread's malloc stats."
- }
- }
-
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_lessthan [s allocator_frag_ratio] 1.05
-
- # Delete keys with even indices to create fragmentation.
- for {set i 0} {$i < [llength $clients]} {incr i} {
- set rd [lindex $clients $i]
- for {set j 0} {$j < $n} {incr j 2} {
- $rd del "k$i-$j"
- }
- }
- for {set i 0} {$i < [llength $clients]} {incr i} {
- [lindex $clients $i] close
- }
- if {$type eq "cluster"} {
- r config resetstat
- }
-
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- assert_morethan [s allocator_frag_ratio] 1.35
-
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- if {$io_threads == 1} {
- wait_for_defrag_stop 500 100 1.05
- } else {
- # TODO: When multithreading is enabled, argv may be created in the io thread
- # and kept in the main thread, which can cause fragmentation to become worse.
- wait_for_defrag_stop 500 100 1.1
- }
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag [s allocator_frag_ratio]"
- puts "frag_bytes [s allocator_frag_bytes]"
- }
- }
- }
-
- if {$type eq "standalone"} { ;# skip in cluster mode
- test "Active defrag big list: $type" {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-max-scan-fields 1000
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 2mb
- r config set maxmemory 0
- r config set list-max-ziplist-size 1 ;# list of 100k items will have 100k quicklist nodes
-
- # create big keys with 10k items
- set rd [redis_deferring_client]
-
- set expected_frag 1.5
- # add a mass of list nodes to two lists (allocations are interlaced)
- set val [string repeat A 500] ;# 1 item of 500 bytes puts us in the 640 bytes bin, which has 32 regs, so high potential for fragmentation
- set elements 100000
- set count 0
- for {set j 0} {$j < $elements} {incr j} {
- $rd lpush biglist1 $val
- $rd lpush biglist2 $val
-
- incr count
- discard_replies_every $rd $count 10000 20000
- }
-
- # create some fragmentation
- r del biglist2
-
- # start defrag
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- if {$::verbose} {
- puts "frag $frag"
- }
-
- assert {$frag >= $expected_frag}
- r config set latency-monitor-threshold 5
- r latency reset
-
- set digest [debug_digest]
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.1
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- set misses [s active_defrag_misses]
- set hits [s active_defrag_hits]
- set frag [s allocator_frag_ratio]
- set max_latency 0
- foreach event [r latency latest] {
- lassign $event eventname time latency max
- if {$eventname == "active-defrag-cycle"} {
- set max_latency $max
- }
- }
- if {$::verbose} {
- puts "used [s allocator_allocated]"
- puts "rss [s allocator_active]"
- puts "frag_bytes [s allocator_frag_bytes]"
- puts "frag $frag"
- puts "misses: $misses"
- puts "hits: $hits"
- puts "max latency $max_latency"
- puts [r latency latest]
- puts [r latency history active-defrag-cycle]
- puts [r memory malloc-stats]
- }
- # due to high fragmentation, 100hz, and active-defrag-cycle-max set to 75,
- # we expect max latency to be not much higher than 7.5ms but due to rare slowness threshold is set higher
- if {!$::no_latency} {
- assert {$max_latency <= 30}
- }
-
- # in extreme cases of stagnation, we see over 5m misses before the tests aborts with "defrag didn't stop",
- # in normal cases we only see 100k misses out of 100k elements
- assert {$misses < $elements * 2}
- }
- # verify the data isn't corrupted or changed
- set newdigest [debug_digest]
- assert {$digest eq $newdigest}
- r save ;# saving an rdb iterates over all the data / pointers
- r del biglist1 ;# coverage for quicklistBookmarksClear
- } {1}
-
- test "Active defrag edge case: $type" {
- # there was an edge case in defrag where all the slabs of a certain bin are exact the same
- # % utilization, with the exception of the current slab from which new allocations are made
- # if the current slab is lower in utilization the defragger would have ended up in stagnation,
- # kept running and not move any allocation.
- # this test is more consistent on a fresh server with no history
- start_server {tags {"defrag"} overrides {save ""}} {
- r flushdb
- r config set hz 100
- r config set activedefrag no
- wait_for_defrag_stop 500 100
- r config resetstat
- r config set active-defrag-max-scan-fields 1000
- r config set active-defrag-threshold-lower 5
- r config set active-defrag-cycle-min 65
- r config set active-defrag-cycle-max 75
- r config set active-defrag-ignore-bytes 1mb
- r config set maxmemory 0
- set expected_frag 1.3
-
- r debug mallctl-str thread.tcache.flush VOID
- # fill the first slab containing 32 regs of 640 bytes.
- for {set j 0} {$j < 32} {incr j} {
- r setrange "_$j" 600 x
- r debug mallctl-str thread.tcache.flush VOID
- }
-
- # add a mass of keys with 600 bytes values, fill the bin of 640 bytes which has 32 regs per slab.
- set rd [redis_deferring_client]
- set keys 640000
- set count 0
- for {set j 0} {$j < $keys} {incr j} {
- $rd setrange $j 600 x
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
-
- # create some fragmentation of 50%
- set sent 0
- for {set j 0} {$j < $keys} {incr j 1} {
- $rd del $j
- incr sent
- incr j 1
-
- discard_replies_every $rd $sent 10000 10000
- }
-
- # create higher fragmentation in the first slab
- for {set j 10} {$j < 32} {incr j} {
- r del "_$j"
- }
-
- # start defrag
- after 120 ;# serverCron only updates the info once in 100ms
- set frag [s allocator_frag_ratio]
- if {$::verbose} {
- puts "frag $frag"
- }
-
- assert {$frag >= $expected_frag}
-
- set digest [debug_digest]
- catch {r config set activedefrag yes} e
- if {[r config get activedefrag] eq "activedefrag yes"} {
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [r info memory]
- puts [r info stats]
- puts [r memory malloc-stats]
- fail "defrag not started."
- }
-
- # wait for the active defrag to stop working
- wait_for_defrag_stop 500 100 1.1
-
- # test the fragmentation is lower
- after 120 ;# serverCron only updates the info once in 100ms
- set misses [s active_defrag_misses]
- set hits [s active_defrag_hits]
- set frag [s allocator_frag_ratio]
- if {$::verbose} {
- puts "frag $frag"
- puts "hits: $hits"
- puts "misses: $misses"
- }
- assert {$misses < 10000000} ;# when defrag doesn't stop, we have some 30m misses, when it does, we have 2m misses
- }
-
- # verify the data isn't corrupted or changed
- set newdigest [debug_digest]
- assert {$digest eq $newdigest}
- r save ;# saving an rdb iterates over all the data / pointers
- }
- } ;# standalone
- }
- }
- }
-
- test "Active defrag can't be triggered during replicaof database flush. See issue #14267" {
- start_server {tags {"repl"} overrides {save ""}} {
- set master_host [srv 0 host]
- set master_port [srv 0 port]
-
- start_server {overrides {save ""}} {
- set replica [srv 0 client]
- set rd [redis_deferring_client 0]
-
- $replica config set hz 100
- $replica config set activedefrag no
- $replica config set active-defrag-threshold-lower 5
- $replica config set active-defrag-cycle-min 65
- $replica config set active-defrag-cycle-max 75
- $replica config set active-defrag-ignore-bytes 2mb
-
- # add a mass of string keys
- set count 0
- for {set j 0} {$j < 500000} {incr j} {
- $rd setrange $j 150 a
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
- assert_equal [$replica dbsize] 500000
-
- # create some fragmentation
- set count 0
- for {set j 0} {$j < 500000} {incr j 2} {
- $rd del $j
-
- incr count
- discard_replies_every $rd $count 10000 10000
- }
- $rd close
- assert_equal [$replica dbsize] 250000
-
- catch {$replica config set activedefrag yes} e
- if {[$replica config get activedefrag] eq "activedefrag yes"} {
- # Start replication sync which will flush the replica's database,
- # then enable defrag to run concurrently with the database flush.
- $replica replicaof $master_host $master_port
-
- # wait for the active defrag to start working (decision once a second)
- wait_for_condition 50 100 {
- [s total_active_defrag_time] ne 0
- } else {
- after 120 ;# serverCron only updates the info once in 100ms
- puts [$replica info memory]
- puts [$replica info stats]
- puts [$replica memory malloc-stats]
- fail "defrag not started."
- }
-
- wait_for_sync $replica
-
- # wait for the active defrag to stop working (db has been emptied during replication sync)
- wait_for_defrag_stop 500 100
- assert_equal [$replica dbsize] 0
- }
- }
- }
- } {} {defrag external:skip tsan:skip debug_defrag:skip cluster}
-
- start_cluster 1 0 {tags {"defrag external:skip tsan:skip debug_defrag:skip cluster needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} {
- test_active_defrag "cluster"
- }
-
- start_server {tags {"defrag external:skip tsan:skip debug_defrag:skip standalone needs:debug"} overrides {appendonly yes auto-aof-rewrite-percentage 0 save "" loglevel notice}} {
- test_active_defrag "standalone"
- }
-} ;# run_solo