0
0
mirror of https://github.com/PostHog/posthog.git synced 2024-11-27 16:26:50 +01:00
posthog/ee/management/commands/materialize_columns.py

103 lines
3.8 KiB
Python

import logging
from django.core.management.base import BaseCommand
from ee.clickhouse.materialized_columns.analyze import (
logger,
materialize_properties_task,
)
from ee.clickhouse.materialized_columns.columns import DEFAULT_TABLE_COLUMN
from posthog.settings import (
MATERIALIZE_COLUMNS_ANALYSIS_PERIOD_HOURS,
MATERIALIZE_COLUMNS_BACKFILL_PERIOD_DAYS,
MATERIALIZE_COLUMNS_MAX_AT_ONCE,
MATERIALIZE_COLUMNS_MINIMUM_QUERY_TIME,
)
class Command(BaseCommand):
help = "Materialize properties into columns in clickhouse"
def add_arguments(self, parser):
parser.add_argument("--dry-run", action="store_true", help="Print plan instead of executing it")
parser.add_argument(
"--property",
help="Properties to materialize. Skips analysis. Allows multiple arguments --property abc '$.abc.def'",
nargs="+",
)
parser.add_argument(
"--property-table",
type=str,
default="events",
choices=["events", "person"],
help="Table of --property",
)
parser.add_argument(
"--table-column",
help="The column to which --property should be materialised from.",
default=DEFAULT_TABLE_COLUMN,
)
parser.add_argument(
"--backfill-period",
type=int,
default=MATERIALIZE_COLUMNS_BACKFILL_PERIOD_DAYS,
help="How many days worth of data to backfill. 0 to disable. Same as MATERIALIZE_COLUMNS_BACKFILL_PERIOD_DAYS env variable.",
)
parser.add_argument(
"--min-query-time",
type=int,
default=MATERIALIZE_COLUMNS_MINIMUM_QUERY_TIME,
help="Minimum query time (ms) before a query if considered for optimization. Same as MATERIALIZE_COLUMNS_MINIMUM_QUERY_TIME env variable.",
)
parser.add_argument(
"--analyze-period",
type=int,
default=MATERIALIZE_COLUMNS_ANALYSIS_PERIOD_HOURS,
help="How long of a time period to analyze. Same as MATERIALIZE_COLUMNS_ANALYSIS_PERIOD_HOURS env variable.",
)
parser.add_argument(
"--analyze-team-id",
type=int,
default=None,
help="Analyze queries only for a specific team_id",
)
parser.add_argument(
"--max-columns",
type=int,
default=MATERIALIZE_COLUMNS_MAX_AT_ONCE,
help="Max number of columns to materialize via single invocation. Same as MATERIALIZE_COLUMNS_MAX_AT_ONCE env variable.",
)
def handle(self, *args, **options):
logger.setLevel(logging.INFO)
if options["dry_run"]:
logger.warn("Dry run: No changes to the tables will be made!")
if options.get("property"):
logger.info(f"Materializing column. table={options['property_table']}, property_name={options['property']}")
materialize_properties_task(
columns_to_materialize=[
(
options["property_table"],
options["table_column"],
prop,
)
for prop in options.get("property")
],
backfill_period_days=options["backfill_period"],
dry_run=options["dry_run"],
)
else:
materialize_properties_task(
time_to_analyze_hours=options["analyze_period"],
maximum=options["max_columns"],
min_query_time=options["min_query_time"],
backfill_period_days=options["backfill_period"],
dry_run=options["dry_run"],
team_id_to_analyze=options["analyze_team_id"],
)