mirror of
https://github.com/timescale/timescaledb.git
synced 2025-05-14 01:10:37 +08:00
On compressed hypertables 3 schema levels are in use simultaneously * main - hypertable level * chunk - inheritance level * compressed chunk In the build_scankeys method all of them appear - as slot have their fields represented as a for a row of the main hypertable. Accessing the slot by the attribut numbers of the chunks may lead to indexing mismatches if there are differences between the schemes. Fixes: #5577
471 lines
21 KiB
PL/PgSQL
471 lines
21 KiB
PL/PgSQL
-- This file and its contents are licensed under the Timescale License.
|
|
-- Please see the included NOTICE for copyright information and
|
|
-- LICENSE-TIMESCALE for a copy of the license.
|
|
|
|
\set ON_ERROR_STOP 0
|
|
\set VERBOSITY default
|
|
|
|
\set ECHO none
|
|
\o /dev/null
|
|
\ir ../../../test/sql/include/test_utils.sql
|
|
\o
|
|
\set ECHO all
|
|
|
|
--table with special column names --
|
|
create table foo2 (a integer, "bacB toD" integer, c integer, d integer);
|
|
select table_name from create_hypertable('foo2', 'a', chunk_time_interval=> 10);
|
|
|
|
create table foo3 (a integer, "bacB toD" integer, c integer, d integer);
|
|
select table_name from create_hypertable('foo3', 'a', chunk_time_interval=> 10);
|
|
|
|
create table non_compressed (a integer, "bacB toD" integer, c integer, d integer);
|
|
select table_name from create_hypertable('non_compressed', 'a', chunk_time_interval=> 10);
|
|
insert into non_compressed values( 3 , 16 , 20, 4);
|
|
|
|
ALTER TABLE foo2 set (timescaledb.compress_segmentby = '"bacB toD",c' , timescaledb.compress_orderby = 'c');
|
|
ALTER TABLE foo2 set (timescaledb.compress, timescaledb.compress_segmentby = '"bacB toD",c' , timescaledb.compress_orderby = 'c');
|
|
ALTER TABLE foo2 set (timescaledb.compress, timescaledb.compress_segmentby = '"bacB toD",c' , timescaledb.compress_orderby = 'd DESC');
|
|
ALTER TABLE foo2 set (timescaledb.compress, timescaledb.compress_segmentby = '"bacB toD",c' , timescaledb.compress_orderby = 'd');
|
|
|
|
-- this is acceptable: having previously set the default value for orderby
|
|
-- and skipping orderby on a subsequent alter command
|
|
create table default_skipped (a integer not null, b integer, c integer, d integer);
|
|
select create_hypertable('default_skipped', 'a', chunk_time_interval=> 10);
|
|
alter table default_skipped set (timescaledb.compress, timescaledb.compress_segmentby = 'c');
|
|
alter table default_skipped set (timescaledb.compress, timescaledb.compress_segmentby = 'c');
|
|
|
|
create table with_rls (a integer, b integer);
|
|
ALTER TABLE with_rls ENABLE ROW LEVEL SECURITY;
|
|
select table_name from create_hypertable('with_rls', 'a', chunk_time_interval=> 10);
|
|
ALTER TABLE with_rls set (timescaledb.compress, timescaledb.compress_orderby='a');
|
|
|
|
--note that the time column "a" should be added to the end of the orderby list
|
|
select * from _timescaledb_catalog.hypertable_compression order by attname;
|
|
|
|
ALTER TABLE foo3 set (timescaledb.compress, timescaledb.compress_orderby='d DeSc NullS lAsT');
|
|
--shold allow alter since segment by was empty
|
|
ALTER TABLE foo3 set (timescaledb.compress, timescaledb.compress_orderby='d Asc NullS lAsT');
|
|
--this is ok too
|
|
ALTER TABLE foo3 set (timescaledb.compress, timescaledb.compress_segmentby = '"bacB toD",c', timescaledb.compress_orderby = 'd DeSc NullS lAsT');
|
|
|
|
-- Negative test cases ---
|
|
|
|
ALTER TABLE foo2 set (timescaledb.compress, timescaledb.compress_segmentby = '"bacB toD",c');
|
|
|
|
alter table default_skipped set (timescaledb.compress, timescaledb.compress_orderby = 'a asc', timescaledb.compress_segmentby = 'c');
|
|
alter table default_skipped set (timescaledb.compress, timescaledb.compress_segmentby = 'c');
|
|
|
|
create table reserved_column_prefix (a integer, _ts_meta_foo integer, "bacB toD" integer, c integer, d integer);
|
|
select table_name from create_hypertable('reserved_column_prefix', 'a', chunk_time_interval=> 10);
|
|
ALTER TABLE reserved_column_prefix set (timescaledb.compress);
|
|
|
|
--basic test with count
|
|
create table foo (a integer, b integer, c integer, t text, p point);
|
|
ALTER TABLE foo ADD CONSTRAINT chk_existing CHECK(b > 0);
|
|
select table_name from create_hypertable('foo', 'a', chunk_time_interval=> 10);
|
|
|
|
insert into foo values( 3 , 16 , 20);
|
|
insert into foo values( 10 , 10 , 20);
|
|
insert into foo values( 20 , 11 , 20);
|
|
insert into foo values( 30 , 12 , 20);
|
|
|
|
-- should error out --
|
|
ALTER TABLE foo ALTER b SET NOT NULL, set (timescaledb.compress);
|
|
|
|
ALTER TABLE foo ALTER b SET NOT NULL;
|
|
select attname, attnotnull from pg_attribute where attrelid = (select oid from pg_class where relname like 'foo') and attname like 'b';
|
|
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'd');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'd');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c desc nulls');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c desc nulls thirsty');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c climb nulls first');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c nulls first asC');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c desc nulls first asc');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c desc hurry');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c descend');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c; SELECT 1');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = '1,2');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c + 1');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'random()');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c LIMIT 1');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'c USING <');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 't COLLATE "en_US"');
|
|
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'c asc' , timescaledb.compress_orderby = 'c');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'c nulls last');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'c + 1');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'random()');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'c LIMIT 1');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_segmentby = 'c + b');
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'a, p');
|
|
|
|
--should succeed
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'a, b');
|
|
|
|
--ddl on ht with compression
|
|
ALTER TABLE foo DROP COLUMN a;
|
|
ALTER TABLE foo DROP COLUMN b;
|
|
ALTER TABLE foo ALTER COLUMN t SET NOT NULL;
|
|
ALTER TABLE foo RESET (timescaledb.compress);
|
|
ALTER TABLE foo ADD CONSTRAINT chk CHECK(b > 0);
|
|
ALTER TABLE foo ADD CONSTRAINT chk UNIQUE(b);
|
|
ALTER TABLE foo DROP CONSTRAINT chk_existing;
|
|
|
|
--note that the time column "a" should not be added to the end of the order by list again (should appear first)
|
|
select hc.* from _timescaledb_catalog.hypertable_compression hc inner join _timescaledb_catalog.hypertable h on (h.id = hc.hypertable_id) where h.table_name = 'foo' order by attname;
|
|
|
|
select decompress_chunk(ch1.schema_name|| '.' || ch1.table_name)
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht where ch1.hypertable_id = ht.id and ht.table_name like 'foo' ORDER BY ch1.id limit 1;
|
|
|
|
--test changing the segment by columns
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'a', timescaledb.compress_segmentby = 'b');
|
|
|
|
select ch1.schema_name|| '.' || ch1.table_name AS "CHUNK_NAME"
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht where ch1.hypertable_id = ht.id and ht.table_name like 'foo' ORDER BY ch1.id limit 1 \gset
|
|
|
|
|
|
select decompress_chunk(:'CHUNK_NAME');
|
|
select decompress_chunk(:'CHUNK_NAME', if_compressed=>true);
|
|
|
|
--should succeed
|
|
select compress_chunk(:'CHUNK_NAME');
|
|
select compress_chunk(:'CHUNK_NAME');
|
|
select compress_chunk(:'CHUNK_NAME', if_not_compressed=>true);
|
|
|
|
select compress_chunk(ch1.schema_name|| '.' || ch1.table_name)
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht where ch1.hypertable_id = ht.id and ht.table_name like 'non_compressed' ORDER BY ch1.id limit 1;
|
|
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'a', timescaledb.compress_segmentby = 'c');
|
|
ALTER TABLE foo set (timescaledb.compress='f');
|
|
ALTER TABLE foo reset (timescaledb.compress);
|
|
|
|
select decompress_chunk(ch1.schema_name|| '.' || ch1.table_name)
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht where ch1.hypertable_id = ht.id and ht.table_name like 'non_compressed' ORDER BY ch1.id limit 1;
|
|
|
|
--should succeed
|
|
select decompress_chunk(ch1.schema_name|| '.' || ch1.table_name)
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht where ch1.hypertable_id = ht.id and ht.table_name like 'foo' and ch1.compressed_chunk_id IS NOT NULL;
|
|
|
|
--should succeed
|
|
ALTER TABLE foo set (timescaledb.compress, timescaledb.compress_orderby = 'a', timescaledb.compress_segmentby = 'b');
|
|
|
|
select hc.* from _timescaledb_catalog.hypertable_compression hc inner join _timescaledb_catalog.hypertable h on (h.id = hc.hypertable_id) where h.table_name = 'foo' order by attname;
|
|
|
|
SELECT comp_hyper.schema_name|| '.' || comp_hyper.table_name as "COMPRESSED_HYPER_NAME"
|
|
FROM _timescaledb_catalog.hypertable comp_hyper
|
|
INNER JOIN _timescaledb_catalog.hypertable uncomp_hyper ON (comp_hyper.id = uncomp_hyper.compressed_hypertable_id)
|
|
WHERE uncomp_hyper.table_name like 'foo' ORDER BY comp_hyper.id LIMIT 1 \gset
|
|
|
|
select add_retention_policy(:'COMPRESSED_HYPER_NAME', INTERVAL '4 months', true);
|
|
|
|
--Constraint checking for compression
|
|
create table fortable(col integer primary key);
|
|
create table table_constr( device_id integer,
|
|
timec integer ,
|
|
location integer ,
|
|
c integer constraint valid_cval check (c > 20) ,
|
|
d integer,
|
|
primary key ( device_id, timec)
|
|
|
|
);
|
|
select table_name from create_hypertable('table_constr', 'timec', chunk_time_interval=> 10);
|
|
BEGIN;
|
|
ALTER TABLE table_constr set (timescaledb.compress, timescaledb.compress_segmentby = 'd');
|
|
ROLLBACK;
|
|
alter table table_constr add constraint table_constr_uk unique (location, timec, device_id);
|
|
BEGIN;
|
|
ALTER TABLE table_constr set (timescaledb.compress, timescaledb.compress_orderby = 'timec', timescaledb.compress_segmentby = 'device_id');
|
|
ROLLBACK;
|
|
|
|
alter table table_constr add constraint table_constr_fk FOREIGN KEY(d) REFERENCES fortable(col) on delete cascade;
|
|
ALTER TABLE table_constr set (timescaledb.compress, timescaledb.compress_orderby = 'timec', timescaledb.compress_segmentby = 'device_id, location');
|
|
|
|
--exclusion constraints not allowed
|
|
alter table table_constr add constraint table_constr_exclu exclude using btree (timec with = );
|
|
ALTER TABLE table_constr set (timescaledb.compress, timescaledb.compress_orderby = 'timec', timescaledb.compress_segmentby = 'device_id, location, d');
|
|
alter table table_constr drop constraint table_constr_exclu ;
|
|
--now it works
|
|
ALTER TABLE table_constr set (timescaledb.compress, timescaledb.compress_orderby = 'timec', timescaledb.compress_segmentby = 'device_id, location, d');
|
|
--can't add fks after compression enabled
|
|
alter table table_constr add constraint table_constr_fk_add_after FOREIGN KEY(d) REFERENCES fortable(col) on delete cascade;
|
|
|
|
-- ddl ADD column variants that are not supported
|
|
ALTER TABLE table_constr ADD COLUMN newcol integer CHECK ( newcol < 10 );
|
|
ALTER TABLE table_constr ADD COLUMN newcol integer UNIQUE;
|
|
ALTER TABLE table_constr ADD COLUMN newcol integer PRIMARY KEY;
|
|
ALTER TABLE table_constr ADD COLUMN newcol integer NOT NULL;
|
|
ALTER TABLE table_constr ADD COLUMN newcol integer DEFAULT random() + random();
|
|
ALTER TABLE table_constr ADD COLUMN IF NOT EXISTS newcol integer REFERENCES fortable(col);
|
|
ALTER TABLE table_constr ADD COLUMN IF NOT EXISTS newcol integer GENERATED ALWAYS AS IDENTITY;
|
|
ALTER TABLE table_constr ADD COLUMN IF NOT EXISTS newcol integer GENERATED BY DEFAULT AS IDENTITY;
|
|
ALTER TABLE table_constr ADD COLUMN newcol nonexistent_type;
|
|
|
|
--FK check should not error even with dropped columns (previously had a bug related to this)
|
|
CREATE TABLE table_fk (
|
|
time timestamptz NOT NULL,
|
|
id1 int8 NOT NULL,
|
|
id2 int8 NOT NULL,
|
|
value float8 NULL,
|
|
CONSTRAINT fk1 FOREIGN KEY (id1) REFERENCES fortable(col),
|
|
CONSTRAINT fk2 FOREIGN KEY (id2) REFERENCES fortable(col)
|
|
);
|
|
|
|
SELECT create_hypertable('table_fk', 'time');
|
|
ALTER TABLE table_fk DROP COLUMN id1;
|
|
ALTER TABLE table_fk SET (timescaledb.compress,timescaledb.compress_segmentby = 'id2');
|
|
|
|
-- TEST fk cascade delete behavior on compressed chunk --
|
|
insert into fortable values(1);
|
|
insert into fortable values(10);
|
|
--we want 2 chunks here --
|
|
insert into table_constr values(1000, 1, 44, 44, 1);
|
|
insert into table_constr values(1000, 10, 44, 44, 10);
|
|
|
|
select ch1.schema_name|| '.' || ch1.table_name AS "CHUNK_NAME"
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht
|
|
where ch1.hypertable_id = ht.id and ht.table_name like 'table_constr'
|
|
ORDER BY ch1.id limit 1 \gset
|
|
|
|
-- we have 1 compressed and 1 uncompressed chunk after this.
|
|
select compress_chunk(:'CHUNK_NAME');
|
|
|
|
SELECT total_chunks , number_compressed_chunks
|
|
FROM hypertable_compression_stats('table_constr');
|
|
|
|
--github issue 1661
|
|
--disable compression after enabling it on a table that has fk constraints
|
|
CREATE TABLE table_constr2( device_id integer,
|
|
timec integer ,
|
|
location integer ,
|
|
d integer references fortable(col),
|
|
primary key ( device_id, timec)
|
|
);
|
|
SELECT table_name from create_hypertable('table_constr2', 'timec', chunk_time_interval=> 10);
|
|
INSERT INTO fortable VALUES( 99 );
|
|
INSERT INTO table_constr2 VALUES( 1000, 10, 5, 99);
|
|
|
|
ALTER TABLE table_constr2 SET (timescaledb.compress, timescaledb.compress_segmentby = 'device_id');
|
|
|
|
ALTER TABLE table_constr2 SET (timescaledb.compress, timescaledb.compress_segmentby = 'device_id, d');
|
|
|
|
--compress a chunk and try to disable compression, it should fail --
|
|
SELECT ch1.schema_name|| '.' || ch1.table_name AS "CHUNK_NAME"
|
|
FROM _timescaledb_catalog.chunk ch1, _timescaledb_catalog.hypertable ht
|
|
WHERE ch1.hypertable_id = ht.id and ht.table_name like 'table_constr2' \gset
|
|
SELECT compress_chunk(:'CHUNK_NAME');
|
|
ALTER TABLE table_constr2 set (timescaledb.compress=false);
|
|
|
|
--decompress all chunks and disable compression.
|
|
SELECT decompress_chunk(:'CHUNK_NAME');
|
|
ALTER TABLE table_constr2 SET (timescaledb.compress=false);
|
|
|
|
-- TEST compression policy
|
|
-- modify the config to trigger errors at runtime
|
|
CREATE TABLE test_table_int(time bigint, val int);
|
|
SELECT create_hypertable('test_table_int', 'time', chunk_time_interval => 1);
|
|
|
|
CREATE OR REPLACE function dummy_now() returns BIGINT LANGUAGE SQL IMMUTABLE as 'SELECT 5::BIGINT';
|
|
SELECT set_integer_now_func('test_table_int', 'dummy_now');
|
|
INSERT INTO test_table_int SELECT generate_series(1,5), 10;
|
|
ALTER TABLE test_table_int set (timescaledb.compress);
|
|
SELECT add_compression_policy('test_table_int', 2::int) AS compressjob_id
|
|
\gset
|
|
|
|
\c :TEST_DBNAME :ROLE_SUPERUSER
|
|
UPDATE _timescaledb_config.bgw_job
|
|
SET config = config - 'compress_after'
|
|
WHERE id = :compressjob_id;
|
|
SELECT config FROM _timescaledb_config.bgw_job WHERE id = :compressjob_id;
|
|
--should fail
|
|
CALL run_job(:compressjob_id);
|
|
|
|
SELECT remove_compression_policy('test_table_int');
|
|
|
|
--again add a new policy that we'll tamper with
|
|
SELECT add_compression_policy('test_table_int', 2::int) AS compressjob_id
|
|
\gset
|
|
UPDATE _timescaledb_config.bgw_job
|
|
SET config = config - 'hypertable_id'
|
|
WHERE id = :compressjob_id;
|
|
SELECT config FROM _timescaledb_config.bgw_job WHERE id = :compressjob_id;
|
|
|
|
--should fail
|
|
CALL run_job(:compressjob_id);
|
|
|
|
UPDATE _timescaledb_config.bgw_job
|
|
SET config = NULL
|
|
WHERE id = :compressjob_id;
|
|
SELECT config FROM _timescaledb_config.bgw_job WHERE id = :compressjob_id;
|
|
|
|
--should fail
|
|
CALL run_job(:compressjob_id);
|
|
|
|
-- test ADD COLUMN IF NOT EXISTS
|
|
CREATE TABLE metric (time TIMESTAMPTZ NOT NULL, val FLOAT8 NOT NULL, dev_id INT4 NOT NULL);
|
|
SELECT create_hypertable('metric', 'time', 'dev_id', 10);
|
|
ALTER TABLE metric SET (
|
|
timescaledb.compress,
|
|
timescaledb.compress_segmentby = 'dev_id',
|
|
timescaledb.compress_orderby = 'time DESC'
|
|
);
|
|
|
|
INSERT INTO metric(time, val, dev_id)
|
|
SELECT s.*, 3.14+1, 1
|
|
FROM generate_series('2021-08-17 00:00:00'::timestamp,
|
|
'2021-08-17 00:02:00'::timestamp, '1 s'::interval) s;
|
|
SELECT compress_chunk(show_chunks('metric'));
|
|
-- column does not exist the first time
|
|
ALTER TABLE metric ADD COLUMN IF NOT EXISTS "medium" VARCHAR ;
|
|
-- column already exists the second time
|
|
ALTER TABLE metric ADD COLUMN IF NOT EXISTS "medium" VARCHAR ;
|
|
-- also add one without IF NOT EXISTS
|
|
ALTER TABLE metric ADD COLUMN "medium_1" VARCHAR ;
|
|
ALTER TABLE metric ADD COLUMN "medium_1" VARCHAR ;
|
|
|
|
--github issue 3481
|
|
--GROUP BY error when setting compress_segmentby with an enum column
|
|
|
|
CREATE TYPE an_enum_type AS ENUM ('home', 'school');
|
|
|
|
CREATE TABLE test (
|
|
time timestamp NOT NULL,
|
|
enum_col an_enum_type NOT NULL
|
|
);
|
|
|
|
SELECT create_hypertable(
|
|
'test', 'time'
|
|
);
|
|
INSERT INTO test VALUES ('2001-01-01 00:00', 'home'),
|
|
('2001-01-01 01:00', 'school'),
|
|
('2001-01-01 02:00', 'home');
|
|
|
|
--enable compression on enum_col
|
|
ALTER TABLE test SET (
|
|
timescaledb.compress,
|
|
timescaledb.compress_segmentby = 'enum_col',
|
|
timescaledb.compress_orderby = 'time'
|
|
);
|
|
|
|
--below queries will pass before chunks are compressed
|
|
SELECT 1 FROM test GROUP BY enum_col;
|
|
EXPLAIN SELECT DISTINCT 1 FROM test;
|
|
|
|
--compress chunks
|
|
SELECT COMPRESS_CHUNK(X) FROM SHOW_CHUNKS('test') X;
|
|
|
|
--below query should pass after chunks are compressed
|
|
SELECT 1 FROM test GROUP BY enum_col;
|
|
EXPLAIN SELECT DISTINCT 1 FROM test;
|
|
|
|
--github issue 4398
|
|
SELECT format('CREATE TABLE data_table AS SELECT now() AS tm, %s', array_to_string(array_agg(format('125 AS c%s',a)), ', ')) FROM generate_series(1,550)a \gexec
|
|
CREATE TABLE ts_table (LIKE data_table);
|
|
SELECT * FROM create_hypertable('ts_table', 'tm');
|
|
--should report a warning
|
|
\set VERBOSITY terse
|
|
ALTER TABLE ts_table SET(timescaledb.compress, timescaledb.compress_segmentby = 'c1',
|
|
timescaledb.compress_orderby = 'tm');
|
|
INSERT INTO ts_table SELECT * FROM data_table;
|
|
--cleanup tables
|
|
DROP TABLE data_table cascade;
|
|
DROP TABLE ts_table cascade;
|
|
|
|
-- #5458 invalid reads for row expressions after column dropped on compressed tables
|
|
CREATE TABLE readings(
|
|
"time" TIMESTAMPTZ NOT NULL,
|
|
battery_status TEXT,
|
|
battery_temperature DOUBLE PRECISION
|
|
);
|
|
|
|
INSERT INTO readings ("time") VALUES ('2022-11-11 11:11:11-00');
|
|
|
|
SELECT create_hypertable('readings', 'time', chunk_time_interval => interval '12 hour', migrate_data=>true);
|
|
|
|
create unique index readings_uniq_idx on readings("time",battery_temperature);
|
|
ALTER TABLE readings SET (timescaledb.compress,timescaledb.compress_segmentby = 'battery_temperature');
|
|
SELECT compress_chunk(show_chunks('readings'));
|
|
|
|
ALTER TABLE readings DROP COLUMN battery_status;
|
|
INSERT INTO readings ("time", battery_temperature) VALUES ('2022-11-11 11:11:11', 0.2);
|
|
SELECT readings FROM readings;
|
|
|
|
-- #5577 On-insert decompression after schema changes may not work properly
|
|
|
|
SELECT decompress_chunk(show_chunks('readings'),true);
|
|
SELECT compress_chunk(show_chunks('readings'),true);
|
|
|
|
\set ON_ERROR_STOP 0
|
|
INSERT INTO readings ("time", battery_temperature) VALUES
|
|
('2022-11-11 11:11:11',0.2) -- same record as inserted
|
|
;
|
|
|
|
\set ON_ERROR_STOP 1
|
|
|
|
SELECT * from readings;
|
|
SELECT assert_equal(count(1), 2::bigint) FROM readings;
|
|
|
|
-- no unique check failure during decompression
|
|
SELECT decompress_chunk(show_chunks('readings'),true);
|
|
|
|
-- #5553 Unique constraints are not always respected on compressed tables
|
|
CREATE TABLE main_table AS
|
|
SELECT '2011-11-11 11:11:11'::timestamptz AS time, 'foo' AS device_id;
|
|
|
|
CREATE UNIQUE INDEX xm ON main_table(time, device_id);
|
|
|
|
SELECT create_hypertable('main_table', 'time', chunk_time_interval => interval '12 hour', migrate_data => TRUE);
|
|
|
|
ALTER TABLE main_table SET (
|
|
timescaledb.compress,
|
|
timescaledb.compress_segmentby = 'device_id',
|
|
timescaledb.compress_orderby = '');
|
|
|
|
SELECT compress_chunk(show_chunks('main_table'));
|
|
|
|
-- insert rejected
|
|
\set ON_ERROR_STOP 0
|
|
INSERT INTO main_table VALUES
|
|
('2011-11-11 11:11:11', 'foo');
|
|
|
|
-- insert rejected in case 1st row doesn't violate constraint with different segmentby
|
|
INSERT INTO main_table VALUES
|
|
('2011-11-11 11:12:11', 'bar'),
|
|
('2011-11-11 11:11:11', 'foo');
|
|
|
|
\set ON_ERROR_STOP 1
|
|
SELECT assert_equal(count(1), 1::bigint) FROM main_table;
|
|
|
|
-- no unique check failure during decompression
|
|
SELECT decompress_chunk(show_chunks('main_table'), TRUE);
|
|
|
|
|
|
DROP TABLE IF EXISTS readings;
|
|
|
|
CREATE TABLE readings(
|
|
"time" timestamptz NOT NULL,
|
|
battery_status text,
|
|
candy integer,
|
|
battery_status2 text,
|
|
battery_temperature text
|
|
);
|
|
|
|
SELECT create_hypertable('readings', 'time', chunk_time_interval => interval '12 hour');
|
|
|
|
CREATE UNIQUE INDEX readings_uniq_idx ON readings("time", battery_temperature);
|
|
|
|
ALTER TABLE readings SET (timescaledb.compress, timescaledb.compress_segmentby = 'battery_temperature');
|
|
|
|
ALTER TABLE readings DROP COLUMN battery_status;
|
|
ALTER TABLE readings DROP COLUMN battery_status2;
|
|
|
|
INSERT INTO readings("time", candy, battery_temperature)
|
|
VALUES ('2022-11-11 11:11:11', 88, '0.2');
|
|
|
|
SELECT compress_chunk(show_chunks('readings'), TRUE);
|
|
|
|
-- no error happens
|
|
INSERT INTO readings("time", candy, battery_temperature)
|
|
VALUES ('2022-11-11 11:11:11', 33, 0.3)
|
|
;
|