diff options
| author | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-01-21 22:52:54 +0100 |
|---|---|---|
| committer | Mitja Felicijan <mitja.felicijan@gmail.com> | 2026-01-21 22:52:54 +0100 |
| commit | dcacc00e3750300617ba6e16eb346713f91a783a (patch) | |
| tree | 38e2d4fb5ed9d119711d4295c6eda4b014af73fd /examples/redis-unstable/tests/unit/violations.tcl | |
| parent | 58dac10aeb8f5a041c46bddbeaf4c7966a99b998 (diff) | |
| download | crep-dcacc00e3750300617ba6e16eb346713f91a783a.tar.gz | |
Remove testing data
Diffstat (limited to 'examples/redis-unstable/tests/unit/violations.tcl')
| -rw-r--r-- | examples/redis-unstable/tests/unit/violations.tcl | 103 |
1 files changed, 0 insertions, 103 deletions
diff --git a/examples/redis-unstable/tests/unit/violations.tcl b/examples/redis-unstable/tests/unit/violations.tcl deleted file mode 100644 index 783f306..0000000 --- a/examples/redis-unstable/tests/unit/violations.tcl +++ /dev/null @@ -1,103 +0,0 @@ -# One XADD with one huge 5GB field -# Expected to fail resulting in an empty stream -run_solo {violations} { -start_server [list overrides [list save ""] ] { - test {XADD one huge field} { - r config set proto-max-bulk-len 10000000000 ;#10gb - r config set client-query-buffer-limit 10000000000 ;#10gb - r write "*5\r\n\$4\r\nXADD\r\n\$2\r\nS1\r\n\$1\r\n*\r\n" - r write "\$1\r\nA\r\n" - catch { - write_big_bulk 5000000000 ;#5gb - } err - assert_match {*too large*} $err - r xlen S1 - } {0} {large-memory} -} - -# One XADD with one huge (exactly nearly) 4GB field -# This uncovers the overflow in lpEncodeGetType -# Expected to fail resulting in an empty stream -start_server [list overrides [list save ""] ] { - test {XADD one huge field - 1} { - r config set proto-max-bulk-len 10000000000 ;#10gb - r config set client-query-buffer-limit 10000000000 ;#10gb - r write "*5\r\n\$4\r\nXADD\r\n\$2\r\nS1\r\n\$1\r\n*\r\n" - r write "\$1\r\nA\r\n" - catch { - write_big_bulk 4294967295 ;#4gb-1 - } err - assert_match {*too large*} $err - r xlen S1 - } {0} {large-memory} -} - -# Gradually add big stream fields using repeated XADD calls -start_server [list overrides [list save ""] ] { - test {several XADD big fields} { - r config set stream-node-max-bytes 0 - for {set j 0} {$j<10} {incr j} { - r xadd stream * 1 $::str500 2 $::str500 - } - r ping - r xlen stream - } {10} {large-memory} -} - -# Add over 4GB to a single stream listpack (one XADD command) -# Expected to fail resulting in an empty stream -start_server [list overrides [list save ""] ] { - test {single XADD big fields} { - r write "*23\r\n\$4\r\nXADD\r\n\$1\r\nS\r\n\$1\r\n*\r\n" - for {set j 0} {$j<10} {incr j} { - r write "\$1\r\n$j\r\n" - write_big_bulk 500000000 "" yes ;#500mb - } - r flush - catch {r read} err - assert_match {*too large*} $err - r xlen S - } {0} {large-memory} -} - -# Gradually add big hash fields using repeated HSET calls -# This reproduces the overflow in the call to ziplistResize -# Object will be converted to hashtable encoding -start_server [list overrides [list save ""] ] { - r config set hash-max-ziplist-value 1000000000 ;#1gb - test {hash with many big fields} { - for {set j 0} {$j<10} {incr j} { - r hset h $j $::str500 - } - r object encoding h - } {hashtable} {large-memory} -} - -# Add over 4GB to a single hash field (one HSET command) -# Object will be converted to hashtable encoding -start_server [list overrides [list save ""] ] { - test {hash with one huge field} { - catch {r config set hash-max-ziplist-value 10000000000} ;#10gb - r config set proto-max-bulk-len 10000000000 ;#10gb - r config set client-query-buffer-limit 10000000000 ;#10gb - r write "*4\r\n\$4\r\nHSET\r\n\$2\r\nH1\r\n" - r write "\$1\r\nA\r\n" - write_big_bulk 5000000000 ;#5gb - r object encoding H1 - } {hashtable} {large-memory} -} -} ;# run_solo - -# SORT which stores an integer encoded element into a list. -# Just for coverage, no news here. -start_server [list overrides [list save ""] ] { - test {SORT adds integer field to list} { - r set S1 asdf - r set S2 123 ;# integer encoded - assert_encoding "int" S2 - r sadd myset 1 2 - r mset D1 1 D2 2 - r sort myset by D* get S* store mylist - r llen mylist - } {2} {cluster:skip} -} |
