mirror of
https://github.com/PostHog/posthog.git
synced 2024-12-01 04:12:23 +01:00
067d73cb4f
Problem see #15200 (comment) When we store session recording events we materialize a lot of information using the snapshot data column. We'll soon not be storing the snapshot data so won't be able to use that to materialize that information, so we need to capture it earlier in the pipeline. Since this is only used for searching for/summarizing recordings we don't need to store every event. Changes We'll push a summary event to a new kafka topic during ingestion. ClickHouse can ingest from that topic into an aggregating merge tree table. So that we store (in theory, although not in practice) only one row per session. add config to turn this on and off by team in plugin server add code behind that to write session recording summary events to a new topic in kafka add ClickHouse tables to ingest and aggregate those summary events |
||
---|---|---|
.. | ||
Celery.run.xml | ||
Frontend.run.xml | ||
Jest Tests.run.xml | ||
Plugin Server.run.xml | ||
PostHog.run.xml | ||
Storybook.run.xml |