7a7c73905d
of elements in indices. Fix the handling of large integers. (CVS 317) FossilOrigin-Name: fc2aae04578c305304a45ec6b76d3ab216cc7526
216 lines
5.3 KiB
Plaintext
216 lines
5.3 KiB
Plaintext
# 2001 September 23
|
|
#
|
|
# The author disclaims copyright to this source code. In place of
|
|
# a legal notice, here is a blessing:
|
|
#
|
|
# May you do good and not evil.
|
|
# May you find forgiveness for yourself and forgive others.
|
|
# May you share freely, never taking more than you give.
|
|
#
|
|
#***********************************************************************
|
|
# This file implements regression tests for SQLite library. The
|
|
# focus of this file is stressing the library by putting large amounts
|
|
# of data in a single row of a table.
|
|
#
|
|
# $Id: bigrow.test,v 1.4 2001/11/24 00:31:47 drh Exp $
|
|
|
|
set testdir [file dirname $argv0]
|
|
source $testdir/tester.tcl
|
|
|
|
# Make a big string that we can use for test data
|
|
#
|
|
do_test bigrow-1.0 {
|
|
set ::bigstr {}
|
|
for {set i 1} {$i<=9999} {incr i} {
|
|
set sep [string index "abcdefghijklmnopqrstuvwxyz" [expr {$i%26}]]
|
|
append ::bigstr "$sep [format %04d $i] "
|
|
}
|
|
string length $::bigstr
|
|
} {69993}
|
|
|
|
# Make a table into which we can insert some but records.
|
|
#
|
|
do_test bigrow-1.1 {
|
|
execsql {
|
|
CREATE TABLE t1(a text, b text, c text);
|
|
SELECT name FROM sqlite_master
|
|
WHERE type='table' OR type='index'
|
|
ORDER BY name
|
|
}
|
|
} {t1}
|
|
|
|
do_test bigrow-1.2 {
|
|
set ::big1 [string range $::bigstr 0 65519]
|
|
set sql "INSERT INTO t1 VALUES('abc',"
|
|
append sql "'$::big1', 'xyz');"
|
|
execsql $sql
|
|
execsql {SELECT a, c FROM t1}
|
|
} {abc xyz}
|
|
do_test bigrow-1.3 {
|
|
execsql {SELECT b FROM t1}
|
|
} [list $::big1]
|
|
do_test bigrow-1.4 {
|
|
set ::big2 [string range $::bigstr 0 65520]
|
|
set sql "INSERT INTO t1 VALUES('abc2',"
|
|
append sql "'$::big2', 'xyz2');"
|
|
set r [catch {execsql $sql} msg]
|
|
lappend r $msg
|
|
} {0 {}}
|
|
do_test bigrow-1.4.1 {
|
|
execsql {SELECT b FROM t1 ORDER BY c}
|
|
} [list $::big1 $::big2]
|
|
do_test bigrow-1.4.2 {
|
|
execsql {SELECT c FROM t1 ORDER BY c}
|
|
} {xyz xyz2}
|
|
do_test bigrow-1.4.3 {
|
|
execsql {DELETE FROM t1 WHERE a='abc2'}
|
|
execsql {SELECT c FROM t1}
|
|
} {xyz}
|
|
|
|
do_test bigrow-1.5 {
|
|
execsql {
|
|
UPDATE t1 SET a=b, b=a;
|
|
SELECT b,c FROM t1
|
|
}
|
|
} {abc xyz}
|
|
do_test bigrow-1.6 {
|
|
execsql {
|
|
SELECT * FROM t1
|
|
}
|
|
} [list $::big1 abc xyz]
|
|
do_test bigrow-1.7 {
|
|
execsql {
|
|
INSERT INTO t1 VALUES('1','2','3');
|
|
INSERT INTO t1 VALUES('A','B','C');
|
|
SELECT b FROM t1 WHERE a=='1';
|
|
}
|
|
} {2}
|
|
do_test bigrow-1.8 {
|
|
execsql "SELECT b FROM t1 WHERE a=='$::big1'"
|
|
} {abc}
|
|
do_test bigrow-1.9 {
|
|
execsql "SELECT b FROM t1 WHERE a!='$::big1' ORDER BY a"
|
|
} {2 B}
|
|
|
|
# Try doing some indexing on big columns
|
|
#
|
|
do_test bigrow-2.1 {
|
|
execsql {
|
|
CREATE INDEX i1 ON t1(a)
|
|
}
|
|
execsql "SELECT b FROM t1 WHERE a=='$::big1'"
|
|
} {abc}
|
|
do_test bigrow-2.2 {
|
|
execsql {
|
|
UPDATE t1 SET a=b, b=a
|
|
}
|
|
execsql "SELECT b FROM t1 WHERE a=='abc'"
|
|
} [list $::big1]
|
|
do_test bigrow-2.3 {
|
|
execsql {
|
|
UPDATE t1 SET a=b, b=a
|
|
}
|
|
execsql "SELECT b FROM t1 WHERE a=='$::big1'"
|
|
} {abc}
|
|
catch {unset ::bigstr}
|
|
catch {unset ::big1}
|
|
catch {unset ::big2}
|
|
|
|
# Mosts of the tests above were created back when rows were limited in
|
|
# size to 64K. Now rows can be much bigger. Test that logic. Also
|
|
# make sure things work correctly at the transition boundries between
|
|
# row sizes of 256 to 257 bytes and from 65536 to 65537 bytes.
|
|
#
|
|
# We begin by testing the 256..257 transition.
|
|
#
|
|
do_test bigrow-3.1 {
|
|
execsql {
|
|
DELETE FROM t1;
|
|
INSERT INTO t1(a,b,c) VALUES('one','abcdefghijklmnopqrstuvwxyz0123','hi');
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 30 hi}
|
|
do_test bigrow-3.2 {
|
|
execsql {
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 240 hi}
|
|
for {set i 1} {$i<10} {incr i} {
|
|
do_test bigrow-3.3.$i {
|
|
execsql "UPDATE t1 SET b=b||'$i'"
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} "one [expr {240+$i}] hi"
|
|
}
|
|
|
|
# Now test the 65536..65537 row-size transition.
|
|
#
|
|
do_test bigrow-4.1 {
|
|
execsql {
|
|
DELETE FROM t1;
|
|
INSERT INTO t1(a,b,c) VALUES('one','abcdefghijklmnopqrstuvwxyz0123','hi');
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 30 hi}
|
|
do_test bigrow-4.2 {
|
|
execsql {
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
UPDATE t1 SET b=b||b;
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 122880 hi}
|
|
do_test bigrow-4.3 {
|
|
execsql {
|
|
UPDATE t1 SET b=substr(b,1,65515)
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 65515 hi}
|
|
for {set i 1} {$i<10} {incr i} {
|
|
do_test bigrow-4.4.$i {
|
|
execsql "UPDATE t1 SET b=b||'$i'"
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} "one [expr {65515+$i}] hi"
|
|
}
|
|
|
|
# Check to make sure the library recovers safely if a row contains
|
|
# too much data.
|
|
#
|
|
do_test bigrow-5.1 {
|
|
execsql {
|
|
DELETE FROM t1;
|
|
INSERT INTO t1(a,b,c) VALUES('one','abcdefghijklmnopqrstuvwxyz0123','hi');
|
|
}
|
|
execsql {SELECT a,length(b),c FROM t1}
|
|
} {one 30 hi}
|
|
set i 1
|
|
for {set sz 60} {$sz<1048560} {incr sz $sz} {
|
|
do_test bigrow-5.2.$i {
|
|
execsql {
|
|
UPDATE t1 SET b=b||b;
|
|
SELECT a,length(b),c FROM t1;
|
|
}
|
|
} "one $sz hi"
|
|
incr i
|
|
}
|
|
do_test bigrow-5.3 {
|
|
set r [catch {execsql {UPDATE t1 SET b=b||b}} msg]
|
|
lappend r $msg
|
|
} {1 {too much data for one table row}}
|
|
do_test bigrow-5.4 {
|
|
execsql {DROP TABLE t1}
|
|
} {}
|
|
|
|
finish_test
|