| 1 | #include "catch.hpp" |
| 2 | #include "test_helpers.hpp" |
| 3 | |
| 4 | using namespace duckdb; |
| 5 | using namespace std; |
| 6 | |
| 7 | TEST_CASE("Insert big varchar strings" , "[varchar]" ) { |
| 8 | unique_ptr<QueryResult> result; |
| 9 | DuckDB db(nullptr); |
| 10 | Connection con(db); |
| 11 | |
| 12 | REQUIRE_NO_FAIL(con.Query("CREATE TABLE test (a VARCHAR);" )); |
| 13 | // insert a big varchar |
| 14 | REQUIRE_NO_FAIL(con.Query("INSERT INTO test VALUES ('aaaaaaaaaa')" )); |
| 15 | // sizes: 10, 100, 1000, 10000 |
| 16 | for (idx_t i = 0; i < 3; i++) { |
| 17 | REQUIRE_NO_FAIL(con.Query("INSERT INTO test SELECT a||a||a||a||a||a||a||a||a||a FROM test WHERE " |
| 18 | "LENGTH(a)=(SELECT MAX(LENGTH(a)) FROM test)" )); |
| 19 | } |
| 20 | |
| 21 | result = con.Query("SELECT LENGTH(a) FROM test ORDER BY 1" ); |
| 22 | REQUIRE(CHECK_COLUMN(result, 0, {10, 100, 1000, 10000})); |
| 23 | } |
| 24 | |
| 25 | TEST_CASE("Test scanning many big varchar strings with limited memory" , "[varchar][.]" ) { |
| 26 | auto temp_dir = TestCreatePath("temp_buf" ); |
| 27 | |
| 28 | auto config = GetTestConfig(); |
| 29 | config->maximum_memory = 100000000; |
| 30 | config->temporary_directory = temp_dir; |
| 31 | |
| 32 | unique_ptr<QueryResult> result; |
| 33 | DuckDB db(nullptr, config.get()); |
| 34 | Connection con(db); |
| 35 | |
| 36 | REQUIRE_NO_FAIL(con.Query("CREATE TABLE test (a VARCHAR);" )); |
| 37 | // create a big varchar (10K characters) |
| 38 | REQUIRE_NO_FAIL(con.Query("INSERT INTO test VALUES ('aaaaaaaaaa')" )); |
| 39 | // sizes: 10, 100, 1000, 10000 |
| 40 | for (idx_t i = 0; i < 3; i++) { |
| 41 | REQUIRE_NO_FAIL(con.Query("INSERT INTO test SELECT a||a||a||a||a||a||a||a||a||a FROM test WHERE " |
| 42 | "LENGTH(a)=(SELECT MAX(LENGTH(a)) FROM test)" )); |
| 43 | } |
| 44 | // now create a second table, we only insert the big varchar string in there |
| 45 | REQUIRE_NO_FAIL(con.Query("CREATE TABLE bigtable (a VARCHAR);" )); |
| 46 | REQUIRE_NO_FAIL( |
| 47 | con.Query("INSERT INTO bigtable SELECT a FROM test WHERE LENGTH(a)=(SELECT MAX(LENGTH(a)) FROM test)" )); |
| 48 | |
| 49 | idx_t entries = 1; |
| 50 | |
| 51 | // verify that the append worked |
| 52 | result = con.Query("SELECT COUNT(*), COUNT(a), MAX(LENGTH(a)), SUM(LENGTH(a)) FROM bigtable" ); |
| 53 | REQUIRE(CHECK_COLUMN(result, 0, {Value::BIGINT(entries)})); |
| 54 | REQUIRE(CHECK_COLUMN(result, 1, {Value::BIGINT(entries)})); |
| 55 | REQUIRE(CHECK_COLUMN(result, 2, {10000})); |
| 56 | REQUIRE(CHECK_COLUMN(result, 3, {Value::BIGINT(entries * 10000)})); |
| 57 | // we create a total of 16K entries in the big table |
| 58 | // the total size of this table is 16K*10K = 160MB |
| 59 | // we then scan the table at every step, as our buffer pool is limited to 100MB not all strings fit in memory |
| 60 | while (entries < 10000) { |
| 61 | REQUIRE_NO_FAIL(con.Query("INSERT INTO bigtable SELECT * FROM bigtable" )); |
| 62 | entries *= 2; |
| 63 | |
| 64 | result = con.Query("SELECT COUNT(*), COUNT(a), MAX(LENGTH(a)), SUM(LENGTH(a)) FROM bigtable" ); |
| 65 | REQUIRE(CHECK_COLUMN(result, 0, {Value::BIGINT(entries)})); |
| 66 | REQUIRE(CHECK_COLUMN(result, 1, {Value::BIGINT(entries)})); |
| 67 | REQUIRE(CHECK_COLUMN(result, 2, {10000})); |
| 68 | REQUIRE(CHECK_COLUMN(result, 3, {Value::BIGINT(entries * 10000)})); |
| 69 | } |
| 70 | |
| 71 | TestDeleteDirectory(temp_dir); |
| 72 | } |
| 73 | |