summaryrefslogtreecommitdiff
path: root/examples/redis-unstable/tests/unit/other.tcl
diff options
context:
space:
mode:
authorMitja Felicijan <mitja.felicijan@gmail.com>2026-01-21 22:52:54 +0100
committerMitja Felicijan <mitja.felicijan@gmail.com>2026-01-21 22:52:54 +0100
commitdcacc00e3750300617ba6e16eb346713f91a783a (patch)
tree38e2d4fb5ed9d119711d4295c6eda4b014af73fd /examples/redis-unstable/tests/unit/other.tcl
parent58dac10aeb8f5a041c46bddbeaf4c7966a99b998 (diff)
downloadcrep-dcacc00e3750300617ba6e16eb346713f91a783a.tar.gz
Remove testing data
Diffstat (limited to 'examples/redis-unstable/tests/unit/other.tcl')
-rw-r--r--examples/redis-unstable/tests/unit/other.tcl733
1 files changed, 0 insertions, 733 deletions
diff --git a/examples/redis-unstable/tests/unit/other.tcl b/examples/redis-unstable/tests/unit/other.tcl
deleted file mode 100644
index 2faa7e9..0000000
--- a/examples/redis-unstable/tests/unit/other.tcl
+++ /dev/null
@@ -1,733 +0,0 @@
-start_server {tags {"other"}} {
- if {$::force_failure} {
- # This is used just for test suite development purposes.
- test {Failing test} {
- format err
- } {ok}
- }
-
- test {Coverage: HELP commands} {
- assert_match "*OBJECT <subcommand> *" [r OBJECT HELP]
- assert_match "*MEMORY <subcommand> *" [r MEMORY HELP]
- assert_match "*PUBSUB <subcommand> *" [r PUBSUB HELP]
- assert_match "*SLOWLOG <subcommand> *" [r SLOWLOG HELP]
- assert_match "*CLIENT <subcommand> *" [r CLIENT HELP]
- assert_match "*COMMAND <subcommand> *" [r COMMAND HELP]
- assert_match "*CONFIG <subcommand> *" [r CONFIG HELP]
- assert_match "*FUNCTION <subcommand> *" [r FUNCTION HELP]
- assert_match "*MODULE <subcommand> *" [r MODULE HELP]
- }
-
- test {Coverage: MEMORY MALLOC-STATS} {
- if {[string match {*jemalloc*} [s mem_allocator]]} {
- assert_match "*jemalloc*" [r memory malloc-stats]
- }
- }
-
- test {Coverage: MEMORY PURGE} {
- if {[string match {*jemalloc*} [s mem_allocator]]} {
- assert_equal {OK} [r memory purge]
- }
- }
-
- test {SAVE - make sure there are all the types as values} {
- # Wait for a background saving in progress to terminate
- waitForBgsave r
- r lpush mysavelist hello
- r lpush mysavelist world
- r set myemptykey {}
- r set mynormalkey {blablablba}
- r zadd mytestzset 10 a
- r zadd mytestzset 20 b
- r zadd mytestzset 30 c
- r save
- } {OK} {needs:save}
-
- tags {slow} {
- if {$::accurate} {set iterations 10000} else {set iterations 1000}
- foreach fuzztype {binary alpha compr} {
- test "FUZZ stresser with data model $fuzztype" {
- set err 0
- for {set i 0} {$i < $iterations} {incr i} {
- set fuzz [randstring 0 512 $fuzztype]
- r set foo $fuzz
- set got [r get foo]
- if {$got ne $fuzz} {
- set err [list $fuzz $got]
- break
- }
- }
- set _ $err
- } {0}
- }
- }
-
- start_server {overrides {save ""} tags {external:skip}} {
- test {FLUSHALL should not reset the dirty counter if we disable save} {
- r set key value
- r flushall
- assert_morethan [s rdb_changes_since_last_save] 0
- }
-
- test {FLUSHALL should reset the dirty counter to 0 if we enable save} {
- r config set save "3600 1 300 100 60 10000"
- r set key value
- r flushall
- assert_equal [s rdb_changes_since_last_save] 0
- }
-
- test {FLUSHALL and bgsave} {
- r config set save "3600 1 300 100 60 10000"
- r set x y
- r bgsave
- r set x y
- r multi
- r debug sleep 1
- # by the time we'll get to run flushall, the child will finish,
- # but the parent will be unaware of it, and it could wrongly set the dirty counter.
- r flushall
- r exec
- assert_equal [s rdb_changes_since_last_save] 0
- }
- }
-
- test {BGSAVE} {
- # Use FLUSHALL instead of FLUSHDB, FLUSHALL do a foreground save
- # and reset the dirty counter to 0, so we won't trigger an unexpected bgsave.
- r flushall
- r save
- r set x 10
- r bgsave
- waitForBgsave r
- r debug reload
- r get x
- } {10} {needs:debug needs:save}
-
- test {SELECT an out of range DB} {
- catch {r select 1000000} err
- set _ $err
- } {*index is out of range*} {cluster:skip}
-
- tags {consistency} {
- proc check_consistency {dumpname code} {
- set dump [csvdump r]
- set sha1 [debug_digest]
-
- uplevel 1 $code
-
- set sha1_after [debug_digest]
- if {$sha1 eq $sha1_after} {
- return 1
- }
-
- # Failed
- set newdump [csvdump r]
- puts "Consistency test failed!"
- puts "You can inspect the two dumps in /tmp/${dumpname}*.txt"
-
- set fd [open /tmp/${dumpname}1.txt w]
- puts $fd $dump
- close $fd
- set fd [open /tmp/${dumpname}2.txt w]
- puts $fd $newdump
- close $fd
-
- return 0
- }
-
- if {$::accurate} {set numops 10000} else {set numops 1000}
- test {Check consistency of different data types after a reload} {
- r flushdb
- # TODO: integrate usehexpire following next commit that will support replication
- createComplexDataset r $numops {usetag usehexpire}
- if {$::ignoredigest} {
- set _ 1
- } else {
- check_consistency {repldump} {
- r debug reload
- }
- }
- } {1} {needs:debug}
-
- test {Same dataset digest if saving/reloading as AOF?} {
- if {$::ignoredigest} {
- set _ 1
- } else {
- check_consistency {aofdump} {
- r config set aof-use-rdb-preamble no
- r bgrewriteaof
- waitForBgrewriteaof r
- r debug loadaof
- }
- }
- } {1} {needs:debug}
- }
-
- test {EXPIRES after a reload (snapshot + append only file rewrite)} {
- r flushdb
- r set x 10
- r expire x 1000
- r save
- r debug reload
- set ttl [r ttl x]
- set e1 [expr {$ttl > 900 && $ttl <= 1000}]
- r bgrewriteaof
- waitForBgrewriteaof r
- r debug loadaof
- set ttl [r ttl x]
- set e2 [expr {$ttl > 900 && $ttl <= 1000}]
- list $e1 $e2
- } {1 1} {needs:debug needs:save}
-
- test {EXPIRES after AOF reload (without rewrite)} {
- r flushdb
- r config set appendonly yes
- r config set aof-use-rdb-preamble no
- r set x somevalue
- r expire x 1000
- r setex y 2000 somevalue
- r set z somevalue
- r expireat z [expr {[clock seconds]+3000}]
-
- # Milliseconds variants
- r set px somevalue
- r pexpire px 1000000
- r psetex py 2000000 somevalue
- r set pz somevalue
- r pexpireat pz [expr {([clock seconds]+3000)*1000}]
-
- # Reload and check
- waitForBgrewriteaof r
- # We need to wait two seconds to avoid false positives here, otherwise
- # the DEBUG LOADAOF command may read a partial file.
- # Another solution would be to set the fsync policy to no, since this
- # prevents write() to be delayed by the completion of fsync().
- after 2000
- r debug loadaof
- set ttl [r ttl x]
- assert {$ttl > 900 && $ttl <= 1000}
- set ttl [r ttl y]
- assert {$ttl > 1900 && $ttl <= 2000}
- set ttl [r ttl z]
- assert {$ttl > 2900 && $ttl <= 3000}
- set ttl [r ttl px]
- assert {$ttl > 900 && $ttl <= 1000}
- set ttl [r ttl py]
- assert {$ttl > 1900 && $ttl <= 2000}
- set ttl [r ttl pz]
- assert {$ttl > 2900 && $ttl <= 3000}
- r config set appendonly no
- } {OK} {needs:debug}
-
- tags {protocol} {
- test {PIPELINING stresser (also a regression for the old epoll bug)} {
- if {$::tls} {
- set fd2 [::tls::socket [srv host] [srv port]]
- } else {
- set fd2 [socket [srv host] [srv port]]
- }
- fconfigure $fd2 -encoding binary -translation binary
- if {!$::singledb} {
- puts -nonewline $fd2 "SELECT 9\r\n"
- flush $fd2
- gets $fd2
- }
-
- for {set i 0} {$i < 100000} {incr i} {
- set q {}
- set val "0000${i}0000"
- append q "SET key:$i $val\r\n"
- puts -nonewline $fd2 $q
- set q {}
- append q "GET key:$i\r\n"
- puts -nonewline $fd2 $q
- }
- flush $fd2
-
- for {set i 0} {$i < 100000} {incr i} {
- gets $fd2 line
- gets $fd2 count
- set count [string range $count 1 end]
- set val [read $fd2 $count]
- read $fd2 2
- }
- close $fd2
- set _ 1
- } {1}
- }
-
- test {APPEND basics} {
- r del foo
- list [r append foo bar] [r get foo] \
- [r append foo 100] [r get foo]
- } {3 bar 6 bar100}
-
- test {APPEND basics, integer encoded values} {
- set res {}
- r del foo
- r append foo 1
- r append foo 2
- lappend res [r get foo]
- r set foo 1
- r append foo 2
- lappend res [r get foo]
- } {12 12}
-
- test {APPEND fuzzing} {
- set err {}
- foreach type {binary alpha compr} {
- set buf {}
- r del x
- for {set i 0} {$i < 1000} {incr i} {
- set bin [randstring 0 10 $type]
- append buf $bin
- r append x $bin
- }
- if {$buf != [r get x]} {
- set err "Expected '$buf' found '[r get x]'"
- break
- }
- }
- set _ $err
- } {}
-
- # Leave the user with a clean DB before to exit
- test {FLUSHDB} {
- set aux {}
- if {$::singledb} {
- r flushdb
- lappend aux 0 [r dbsize]
- } else {
- r select 9
- r flushdb
- lappend aux [r dbsize]
- r select 10
- r flushdb
- lappend aux [r dbsize]
- }
- } {0 0}
-
- test {Perform a final SAVE to leave a clean DB on disk} {
- waitForBgsave r
- r save
- } {OK} {needs:save}
-
- test {RESET clears client state} {
- r client setname test-client
- r client tracking on
-
- assert_equal [r reset] "RESET"
- set client [r client list]
- assert_match {*name= *} $client
- assert_match {*flags=N *} $client
- } {} {needs:reset}
-
- test {RESET clears MONITOR state} {
- set rd [redis_deferring_client]
- $rd monitor
- assert_equal [$rd read] "OK"
-
- $rd reset
- assert_equal [$rd read] "RESET"
- $rd close
-
- assert_no_match {*flags=O*} [r client list]
- } {} {needs:reset}
-
- test {RESET clears and discards MULTI state} {
- r multi
- r set key-a a
-
- r reset
- catch {r exec} err
- assert_match {*EXEC without MULTI*} $err
- } {} {needs:reset}
-
- test {RESET clears Pub/Sub state} {
- r subscribe channel-1
- r reset
-
- # confirm we're not subscribed by executing another command
- r set key val
- } {OK} {needs:reset}
-
- test {RESET clears authenticated state} {
- r acl setuser user1 on >secret +@all
- r auth user1 secret
- assert_equal [r acl whoami] user1
-
- r reset
-
- assert_equal [r acl whoami] default
- } {} {needs:reset}
-
- test "Subcommand syntax error crash (issue #10070)" {
- assert_error {*unknown command*} {r GET|}
- assert_error {*unknown command*} {r GET|SET}
- assert_error {*unknown command*} {r GET|SET|OTHER}
- assert_error {*unknown command*} {r CONFIG|GET GET_XX}
- assert_error {*unknown subcommand*} {r CONFIG GET_XX}
- }
-}
-
-start_server {tags {"other external:skip"}} {
- test {Don't rehash if redis has child process} {
- r config set save ""
- r config set rdb-key-save-delay 1000000
-
- populate 4095 "" 1
- r bgsave
- wait_for_condition 10 100 {
- [s rdb_bgsave_in_progress] eq 1
- } else {
- fail "bgsave did not start in time"
- }
-
- r mset k1 v1 k2 v2
- # Hash table should not rehash
- assert_no_match "*table size: 8192*" [r debug HTSTATS 9]
- exec kill -9 [get_child_pid 0]
- waitForBgsave r
-
- # Hash table should rehash since there is no child process,
- # size is power of two and over 4096, so it is 8192
- wait_for_condition 50 100 {
- [string match "*table size: 8192*" [r debug HTSTATS 9]]
- } else {
- fail "hash table did not rehash after child process killed"
- }
- } {} {needs:debug needs:local-process}
-}
-
-proc read_proc_title {pid} {
- set fd [open "/proc/$pid/cmdline" "r"]
- set cmdline [read $fd 1024]
- close $fd
-
- return $cmdline
-}
-
-start_server {tags {"other external:skip"}} {
- test {Process title set as expected} {
- # Test only on Linux where it's easy to get cmdline without relying on tools.
- # Skip valgrind as it messes up the arguments.
- set os [exec uname]
- if {$os == "Linux" && !$::valgrind} {
- # Set a custom template
- r config set "proc-title-template" "TEST {title} {listen-addr} {port} {tls-port} {unixsocket} {config-file}"
- set cmdline [read_proc_title [srv 0 pid]]
-
- assert_equal "TEST" [lindex $cmdline 0]
- assert_match "*/redis-server" [lindex $cmdline 1]
-
- if {$::tls} {
- set expect_port [srv 0 pport]
- set expect_tls_port [srv 0 port]
- set port [srv 0 pport]
- } else {
- set expect_port [srv 0 port]
- set expect_tls_port 0
- set port [srv 0 port]
- }
-
- assert_equal "$::host:$port" [lindex $cmdline 2]
- assert_equal $expect_port [lindex $cmdline 3]
- assert_equal $expect_tls_port [lindex $cmdline 4]
- assert_match "*/tests/tmp/server.*/socket" [lindex $cmdline 5]
- assert_match "*/tests/tmp/redis.conf.*" [lindex $cmdline 6]
-
- # Try setting a bad template
- catch {r config set "proc-title-template" "{invalid-var}"} err
- assert_match {*template format is invalid*} $err
- }
- }
-}
-
-start_cluster 1 0 {tags {"other external:skip cluster slow"}} {
- r config set dynamic-hz no hz 500
- test "Redis can trigger resizing" {
- r flushall
- # hashslot(foo) is 12182
- for {set j 1} {$j <= 128} {incr j} {
- r set "{foo}$j" a
- }
- assert_match "*table size: 128*" [r debug HTSTATS 0]
-
- # disable resizing, the reason for not using slow bgsave is because
- # it will hit the dict_force_resize_ratio.
- r debug dict-resizing 0
-
- # delete data to have lot's (96%) of empty buckets
- for {set j 1} {$j <= 123} {incr j} {
- r del "{foo}$j"
- }
- assert_match "*table size: 128*" [r debug HTSTATS 0]
-
- # enable resizing
- r debug dict-resizing 1
-
- # waiting for serverCron to resize the tables
- wait_for_condition 1000 10 {
- [string match {*table size: 8*} [r debug HTSTATS 0]]
- } else {
- puts [r debug HTSTATS 0]
- fail "hash tables weren't resize."
- }
- } {} {needs:debug}
-
- test "Redis can rewind and trigger smaller slot resizing" {
- # hashslot(foo) is 12182
- # hashslot(alice) is 749, smaller than hashslot(foo),
- # attempt to trigger a resize on it, see details in #12802.
- for {set j 1} {$j <= 128} {incr j} {
- r set "{alice}$j" a
- }
-
- # disable resizing, the reason for not using slow bgsave is because
- # it will hit the dict_force_resize_ratio.
- r debug dict-resizing 0
-
- for {set j 1} {$j <= 123} {incr j} {
- r del "{alice}$j"
- }
-
- # enable resizing
- r debug dict-resizing 1
-
- # waiting for serverCron to resize the tables
- wait_for_condition 1000 10 {
- [string match {*table size: 16*} [r debug HTSTATS 0]]
- } else {
- puts [r debug HTSTATS 0]
- fail "hash tables weren't resize."
- }
- } {} {needs:debug}
-}
-
-start_server {tags {"other external:skip"}} {
- test "Redis can resize empty dict" {
- # Write and then delete 128 keys, creating an empty dict
- r flushall
-
- # Add one key to the db just to create the dict and get its initial size
- r set x 1
- set initial_size [dict get [r memory stats] db.9 overhead.hashtable.main]
-
- # Now add 128 keys and then delete them
- for {set j 1} {$j <= 128} {incr j} {
- r set $j{b} a
- }
-
- for {set j 1} {$j <= 128} {incr j} {
- r del $j{b}
- }
-
- # dict must have expanded. Verify it eventually shrinks back to its initial size.
- wait_for_condition 100 50 {
- [dict get [r memory stats] db.9 overhead.hashtable.main] == $initial_size
- } else {
- fail "dict did not resize in time to its initial size"
- }
- }
-}
-
-start_server {tags {"other external:skip"} overrides {cluster-compatibility-sample-ratio 100}} {
- test {Cross DB command is incompatible with cluster mode} {
- set incompatible_ops [s cluster_incompatible_ops]
-
- # SELECT with 0 is compatible command in cluster mode
- assert_equal {OK} [r select 0]
- assert_equal $incompatible_ops [s cluster_incompatible_ops]
-
- # SELECT with nonzero is incompatible command in cluster mode
- assert_equal {OK} [r select 1]
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # SWAPDB is incompatible command in cluster mode
- assert_equal {OK} [r swapdb 0 1]
- assert_equal [expr $incompatible_ops + 2] [s cluster_incompatible_ops]
-
-
- # If destination db in COPY command is equal to source db, it is compatible
- # with cluster mode, otherwise it is incompatible.
- r select 0
- r set key1 value1
- set incompatible_ops [s cluster_incompatible_ops]
- assert_equal {1} [r copy key1 key2{key1}] ;# destination db is equal to source db
- assert_equal $incompatible_ops [s cluster_incompatible_ops]
- assert_equal {1} [r copy key2{key1} key1 db 1] ;# destination db is not equal to source db
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # If destination db in MOVE command is not equal to source db, it is incompatible
- # with cluster mode.
- r set key3 value3
- assert_equal {1} [r move key3 1]
- assert_equal [expr $incompatible_ops + 2] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Function no-cluster flag is incompatible with cluster mode} {
- set incompatible_ops [s cluster_incompatible_ops]
-
- # no-cluster flag is incompatible with cluster mode
- r function load {#!lua name=test
- redis.register_function{function_name='f1', callback=function() return 'hello' end, flags={'no-cluster'}}
- }
- r fcall f1 0
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # It is compatible without no-cluster flag, should not increase the cluster_incompatible_ops
- r function load {#!lua name=test2
- redis.register_function{function_name='f2', callback=function() return 'hello' end}
- }
- r fcall f2 0
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Script no-cluster flag is incompatible with cluster mode} {
- set incompatible_ops [s cluster_incompatible_ops]
-
- # no-cluster flag is incompatible with cluster mode
- r eval {#!lua flags=no-cluster
- return 1
- } 0
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # It is compatible without no-cluster flag, should not increase the cluster_incompatible_ops
- r eval {#!lua
- return 1
- } 0
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {SORT command incompatible operations with cluster mode} {
- set incompatible_ops [s cluster_incompatible_ops]
-
- # If the BY pattern slot is not equal with the slot of keys, we consider
- # an incompatible behavior, otherwise it is compatible, should not increase
- # the cluster_incompatible_ops
- r lpush mylist 1 2 3
- for {set i 1} {$i < 4} {incr i} {
- r set weight_$i [expr 4 - $i]
- }
- assert_equal {3 2 1} [r sort mylist BY weight_*]
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
- # weight{mylist}_* and mylist have the same slot
- for {set i 1} {$i < 4} {incr i} {
- r set weight{mylist}_$i [expr 4 - $i]
- }
- assert_equal {3 2 1} [r sort mylist BY weight{mylist}_*]
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # If the GET pattern slot is not equal with the slot of keys, we consider
- # an incompatible behavior, otherwise it is compatible, should not increase
- # the cluster_incompatible_ops
- for {set i 1} {$i < 4} {incr i} {
- r set object_$i o_$i
- }
- assert_equal {o_3 o_2 o_1} [r sort mylist BY weight{mylist}_* GET object_*]
- assert_equal [expr $incompatible_ops + 2] [s cluster_incompatible_ops]
- # object{mylist}_*, weight{mylist}_* and mylist have the same slot
- for {set i 1} {$i < 4} {incr i} {
- r set object{mylist}_$i o_$i
- }
- assert_equal {o_3 o_2 o_1} [r sort mylist BY weight{mylist}_* GET object{mylist}_*]
- assert_equal [expr $incompatible_ops + 2] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Normal cross slot commands are incompatible with cluster mode} {
- # Normal cross slot command
- set incompatible_ops [s cluster_incompatible_ops]
- r mset foo bar bar foo
- r del foo bar
- assert_equal [expr $incompatible_ops + 2] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Transaction is incompatible with cluster mode} {
- set incompatible_ops [s cluster_incompatible_ops]
-
- # Incomplete transaction
- catch {r EXEC}
- r multi
- r exec
- assert_equal $incompatible_ops [s cluster_incompatible_ops]
-
- # Transaction, SET and DEL have keys with different slots
- r multi
- r set foo bar
- r del bar
- r exec
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Lua scripts are incompatible with cluster mode} {
- # Lua script, declared keys have different slots, it is not a compatible operation
- set incompatible_ops [s cluster_incompatible_ops]
- r eval {#!lua
- redis.call('mset', KEYS[1], 0, KEYS[2], 0)
- } 2 foo bar
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # Lua script, no declared keys, but accessing keys have different slots,
- # it is not a compatible operation
- set incompatible_ops [s cluster_incompatible_ops]
- r eval {#!lua
- redis.call('mset', 'foo', 0, 'bar', 0)
- } 0
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # Lua script, declared keys have the same slot, but accessing keys
- # have different slots in one command, even with flag 'allow-cross-slot-keys',
- # it still is not a compatible operation
- set incompatible_ops [s cluster_incompatible_ops]
- r eval {#!lua flags=allow-cross-slot-keys
- redis.call('mset', 'foo', 0, 'bar', 0)
- redis.call('mset', KEYS[1], 0, KEYS[2], 0)
- } 2 foo bar{foo}
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
-
- # Lua script, declared keys have the same slot, but accessing keys have different slots
- # in multiple commands, and with flag 'allow-cross-slot-keys', it is a compatible operation
- set incompatible_ops [s cluster_incompatible_ops]
- r eval {#!lua flags=allow-cross-slot-keys
- redis.call('set', 'foo', 0)
- redis.call('set', 'bar', 0)
- redis.call('mset', KEYS[1], 0, KEYS[2], 0)
- } 2 foo bar{foo}
- assert_equal $incompatible_ops [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {Shard subscribe commands are incompatible with cluster mode} {
- set rd1 [redis_deferring_client]
- set incompatible_ops [s cluster_incompatible_ops]
- assert_equal {1 2} [ssubscribe $rd1 {foo bar}]
- assert_equal [expr $incompatible_ops + 1] [s cluster_incompatible_ops]
- } {} {cluster:skip}
-
- test {cluster-compatibility-sample-ratio configuration can work} {
- # Disable cluster compatibility sampling, no increase in cluster_incompatible_ops
- set incompatible_ops [s cluster_incompatible_ops]
- r config set cluster-compatibility-sample-ratio 0
- for {set i 0} {$i < 100} {incr i} {
- r mset foo bar$i bar foo$i
- }
- # Enable cluster compatibility sampling again to show the metric
- r config set cluster-compatibility-sample-ratio 1
- assert_equal $incompatible_ops [s cluster_incompatible_ops]
-
- # 100% sample ratio, all operations should increase cluster_incompatible_ops
- set incompatible_ops [s cluster_incompatible_ops]
- r config set cluster-compatibility-sample-ratio 100
- for {set i 0} {$i < 100} {incr i} {
- r mset foo bar$i bar foo$i
- }
- assert_equal [expr $incompatible_ops + 100] [s cluster_incompatible_ops]
-
- # 30% sample ratio, cluster_incompatible_ops should increase between 20% and 40%
- set incompatible_ops [s cluster_incompatible_ops]
- r config set cluster-compatibility-sample-ratio 30
- for {set i 0} {$i < 1000} {incr i} {
- r mset foo bar$i bar foo$i
- }
- assert_range [s cluster_incompatible_ops] [expr $incompatible_ops + 200] [expr $incompatible_ops + 400]
- } {} {cluster:skip}
-}