summaryrefslogtreecommitdiff
path: root/examples/redis-unstable/tests/unit/violations.tcl
diff options
context:
space:
mode:
authorMitja Felicijan <mitja.felicijan@gmail.com>2026-01-21 22:52:54 +0100
committerMitja Felicijan <mitja.felicijan@gmail.com>2026-01-21 22:52:54 +0100
commitdcacc00e3750300617ba6e16eb346713f91a783a (patch)
tree38e2d4fb5ed9d119711d4295c6eda4b014af73fd /examples/redis-unstable/tests/unit/violations.tcl
parent58dac10aeb8f5a041c46bddbeaf4c7966a99b998 (diff)
downloadcrep-dcacc00e3750300617ba6e16eb346713f91a783a.tar.gz
Remove testing data
Diffstat (limited to 'examples/redis-unstable/tests/unit/violations.tcl')
-rw-r--r--examples/redis-unstable/tests/unit/violations.tcl103
1 files changed, 0 insertions, 103 deletions
diff --git a/examples/redis-unstable/tests/unit/violations.tcl b/examples/redis-unstable/tests/unit/violations.tcl
deleted file mode 100644
index 783f306..0000000
--- a/examples/redis-unstable/tests/unit/violations.tcl
+++ /dev/null
@@ -1,103 +0,0 @@
-# One XADD with one huge 5GB field
-# Expected to fail resulting in an empty stream
-run_solo {violations} {
-start_server [list overrides [list save ""] ] {
- test {XADD one huge field} {
- r config set proto-max-bulk-len 10000000000 ;#10gb
- r config set client-query-buffer-limit 10000000000 ;#10gb
- r write "*5\r\n\$4\r\nXADD\r\n\$2\r\nS1\r\n\$1\r\n*\r\n"
- r write "\$1\r\nA\r\n"
- catch {
- write_big_bulk 5000000000 ;#5gb
- } err
- assert_match {*too large*} $err
- r xlen S1
- } {0} {large-memory}
-}
-
-# One XADD with one huge (exactly nearly) 4GB field
-# This uncovers the overflow in lpEncodeGetType
-# Expected to fail resulting in an empty stream
-start_server [list overrides [list save ""] ] {
- test {XADD one huge field - 1} {
- r config set proto-max-bulk-len 10000000000 ;#10gb
- r config set client-query-buffer-limit 10000000000 ;#10gb
- r write "*5\r\n\$4\r\nXADD\r\n\$2\r\nS1\r\n\$1\r\n*\r\n"
- r write "\$1\r\nA\r\n"
- catch {
- write_big_bulk 4294967295 ;#4gb-1
- } err
- assert_match {*too large*} $err
- r xlen S1
- } {0} {large-memory}
-}
-
-# Gradually add big stream fields using repeated XADD calls
-start_server [list overrides [list save ""] ] {
- test {several XADD big fields} {
- r config set stream-node-max-bytes 0
- for {set j 0} {$j<10} {incr j} {
- r xadd stream * 1 $::str500 2 $::str500
- }
- r ping
- r xlen stream
- } {10} {large-memory}
-}
-
-# Add over 4GB to a single stream listpack (one XADD command)
-# Expected to fail resulting in an empty stream
-start_server [list overrides [list save ""] ] {
- test {single XADD big fields} {
- r write "*23\r\n\$4\r\nXADD\r\n\$1\r\nS\r\n\$1\r\n*\r\n"
- for {set j 0} {$j<10} {incr j} {
- r write "\$1\r\n$j\r\n"
- write_big_bulk 500000000 "" yes ;#500mb
- }
- r flush
- catch {r read} err
- assert_match {*too large*} $err
- r xlen S
- } {0} {large-memory}
-}
-
-# Gradually add big hash fields using repeated HSET calls
-# This reproduces the overflow in the call to ziplistResize
-# Object will be converted to hashtable encoding
-start_server [list overrides [list save ""] ] {
- r config set hash-max-ziplist-value 1000000000 ;#1gb
- test {hash with many big fields} {
- for {set j 0} {$j<10} {incr j} {
- r hset h $j $::str500
- }
- r object encoding h
- } {hashtable} {large-memory}
-}
-
-# Add over 4GB to a single hash field (one HSET command)
-# Object will be converted to hashtable encoding
-start_server [list overrides [list save ""] ] {
- test {hash with one huge field} {
- catch {r config set hash-max-ziplist-value 10000000000} ;#10gb
- r config set proto-max-bulk-len 10000000000 ;#10gb
- r config set client-query-buffer-limit 10000000000 ;#10gb
- r write "*4\r\n\$4\r\nHSET\r\n\$2\r\nH1\r\n"
- r write "\$1\r\nA\r\n"
- write_big_bulk 5000000000 ;#5gb
- r object encoding H1
- } {hashtable} {large-memory}
-}
-} ;# run_solo
-
-# SORT which stores an integer encoded element into a list.
-# Just for coverage, no news here.
-start_server [list overrides [list save ""] ] {
- test {SORT adds integer field to list} {
- r set S1 asdf
- r set S2 123 ;# integer encoded
- assert_encoding "int" S2
- r sadd myset 1 2
- r mset D1 1 D2 2
- r sort myset by D* get S* store mylist
- r llen mylist
- } {2} {cluster:skip}
-}