From 251948148dd0d1bc0d1cb5e78eb1fb09332ca35e Mon Sep 17 00:00:00 2001 From: github-actions <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 21 Sep 2023 21:18:19 +0000 Subject: [PATCH] Update query snapshots --- .../test/__snapshots__/test_schema.ambr | 94 +++++++++++++++++++ 1 file changed, 94 insertions(+) diff --git a/posthog/clickhouse/test/__snapshots__/test_schema.ambr b/posthog/clickhouse/test/__snapshots__/test_schema.ambr index ac21b1ac5989fd..c0a744a746ac0d 100644 --- a/posthog/clickhouse/test/__snapshots__/test_schema.ambr +++ b/posthog/clickhouse/test/__snapshots__/test_schema.ambr @@ -54,6 +54,22 @@ ' --- +# name: test_create_kafka_table_with_different_kafka_host[kafka_batch_exports_log_entries] + ' + + CREATE TABLE IF NOT EXISTS kafka_batch_exports_log_entries ON CLUSTER 'posthog' + ( + team_id UInt64, + batch_export_id VARCHAR, + run_id VARCHAR, + timestamp DateTime64(6, 'UTC'), + level VARCHAR, + message VARCHAR + + ) ENGINE = Kafka('test.kafka.broker:9092', 'batch_exports_log_entries_test', 'group1', 'JSONEachRow') + + ' +--- # name: test_create_kafka_table_with_different_kafka_host[kafka_events_dead_letter_queue] ' @@ -390,6 +406,46 @@ ' --- +# name: test_create_table_query[batch_exports_log_entries] + ' + + CREATE TABLE IF NOT EXISTS batch_exports_log_entries ON CLUSTER 'posthog' + ( + team_id UInt64, + batch_export_id VARCHAR, + run_id VARCHAR, + timestamp DateTime64(6, 'UTC'), + level VARCHAR, + message VARCHAR + + , _timestamp DateTime + , _offset UInt64 + + ) ENGINE = ReplicatedReplacingMergeTree('/clickhouse/tables/77f1df52-4b43-11e9-910f-b8ca3a9b9f3e_noshard/posthog.batch_exports_log_entries', '{replica}-{shard}', _timestamp) + ORDER BY (team_id, batch_export_id, run_id, timestamp) + + SETTINGS index_granularity=512 + + ' +--- +# name: test_create_table_query[batch_exports_log_entries_mv] + ' + + CREATE MATERIALIZED VIEW IF NOT EXISTS batch_exports_log_entries_mv ON CLUSTER 'posthog' + TO posthog_test.batch_exports_log_entries + AS SELECT + team_id, + batch_export_id, + run_id, + timestamp, + level, + message, + _timestamp, + _offset + FROM posthog_test.kafka_batch_exports_log_entries + + ' +--- # name: test_create_table_query[cohortpeople] ' @@ -642,6 +698,22 @@ ' --- +# name: test_create_table_query[kafka_batch_exports_log_entries] + ' + + CREATE TABLE IF NOT EXISTS kafka_batch_exports_log_entries ON CLUSTER 'posthog' + ( + team_id UInt64, + batch_export_id VARCHAR, + run_id VARCHAR, + timestamp DateTime64(6, 'UTC'), + level VARCHAR, + message VARCHAR + + ) ENGINE = Kafka('kafka:9092', 'batch_exports_log_entries_test', 'group1', 'JSONEachRow') + + ' +--- # name: test_create_table_query[kafka_events_dead_letter_queue] ' @@ -1770,6 +1842,28 @@ ' --- +# name: test_create_table_query_replicated_and_storage[batch_exports_log_entries] + ' + + CREATE TABLE IF NOT EXISTS batch_exports_log_entries ON CLUSTER 'posthog' + ( + team_id UInt64, + batch_export_id VARCHAR, + run_id VARCHAR, + timestamp DateTime64(6, 'UTC'), + level VARCHAR, + message VARCHAR + + , _timestamp DateTime + , _offset UInt64 + + ) ENGINE = ReplicatedReplacingMergeTree('/clickhouse/tables/77f1df52-4b43-11e9-910f-b8ca3a9b9f3e_noshard/posthog.batch_exports_log_entries', '{replica}-{shard}', _timestamp) + ORDER BY (team_id, batch_export_id, run_id, timestamp) + + SETTINGS index_granularity=512 + + ' +--- # name: test_create_table_query_replicated_and_storage[cohortpeople] '