mirror of
https://github.com/PostHog/posthog.git
synced 2024-11-28 18:26:15 +01:00
cd7ccc779e
* Use clickhouse connection pooler to avoid connections from stepping on eachother * build reqs
199 lines
5.9 KiB
Python
199 lines
5.9 KiB
Python
import asyncio
|
|
import hashlib
|
|
import json
|
|
from time import time
|
|
from typing import Any, Dict, List, Tuple
|
|
|
|
import sqlparse
|
|
import statsd
|
|
from aioch import Client
|
|
from asgiref.sync import async_to_sync
|
|
from clickhouse_driver import Client as SyncClient
|
|
from clickhouse_pool import ChPool
|
|
from django.conf import settings as app_settings
|
|
from django.core.cache import cache
|
|
from django.utils.timezone import now
|
|
from sentry_sdk.api import capture_exception
|
|
|
|
from posthog import redis
|
|
from posthog.constants import RDBMS
|
|
from posthog.settings import (
|
|
CLICKHOUSE_ASYNC,
|
|
CLICKHOUSE_CA,
|
|
CLICKHOUSE_DATABASE,
|
|
CLICKHOUSE_HOST,
|
|
CLICKHOUSE_PASSWORD,
|
|
CLICKHOUSE_SECURE,
|
|
CLICKHOUSE_USER,
|
|
CLICKHOUSE_VERIFY,
|
|
PRIMARY_DB,
|
|
STATSD_HOST,
|
|
STATSD_PORT,
|
|
STATSD_PREFIX,
|
|
TEST,
|
|
)
|
|
from posthog.utils import get_safe_cache
|
|
|
|
if STATSD_HOST is not None:
|
|
statsd.Connection.set_defaults(host=STATSD_HOST, port=STATSD_PORT)
|
|
|
|
CACHE_TTL = 60 # seconds
|
|
|
|
_save_query_user_id = False
|
|
|
|
if PRIMARY_DB != RDBMS.CLICKHOUSE:
|
|
ch_client = None # type: Client
|
|
|
|
def async_execute(query, args=None, settings=None):
|
|
return
|
|
|
|
def sync_execute(query, args=None, settings=None):
|
|
return
|
|
|
|
def cache_sync_execute(query, args=None, redis_client=None, ttl=None, settings=None):
|
|
return
|
|
|
|
|
|
else:
|
|
if not TEST and CLICKHOUSE_ASYNC:
|
|
ch_client = Client(
|
|
host=CLICKHOUSE_HOST,
|
|
database=CLICKHOUSE_DATABASE,
|
|
secure=CLICKHOUSE_SECURE,
|
|
user=CLICKHOUSE_USER,
|
|
password=CLICKHOUSE_PASSWORD,
|
|
ca_certs=CLICKHOUSE_CA,
|
|
verify=CLICKHOUSE_VERIFY,
|
|
)
|
|
|
|
ch_pool = ChPool(
|
|
host=CLICKHOUSE_HOST,
|
|
database=CLICKHOUSE_DATABASE,
|
|
secure=CLICKHOUSE_SECURE,
|
|
user=CLICKHOUSE_USER,
|
|
password=CLICKHOUSE_PASSWORD,
|
|
ca_certs=CLICKHOUSE_CA,
|
|
verify=CLICKHOUSE_VERIFY,
|
|
connections_min=20,
|
|
connections_max=1000,
|
|
)
|
|
|
|
@async_to_sync
|
|
async def async_execute(query, args=None, settings=None):
|
|
loop = asyncio.get_event_loop()
|
|
task = loop.create_task(ch_client.execute(query, args, settings=settings))
|
|
return task
|
|
|
|
else:
|
|
# if this is a test use the sync client
|
|
ch_client = SyncClient(
|
|
host=CLICKHOUSE_HOST,
|
|
database=CLICKHOUSE_DATABASE,
|
|
secure=CLICKHOUSE_SECURE,
|
|
user=CLICKHOUSE_USER,
|
|
password=CLICKHOUSE_PASSWORD,
|
|
ca_certs=CLICKHOUSE_CA,
|
|
verify=CLICKHOUSE_VERIFY,
|
|
)
|
|
|
|
ch_pool = ChPool(
|
|
host=CLICKHOUSE_HOST,
|
|
database=CLICKHOUSE_DATABASE,
|
|
secure=CLICKHOUSE_SECURE,
|
|
user=CLICKHOUSE_USER,
|
|
password=CLICKHOUSE_PASSWORD,
|
|
ca_certs=CLICKHOUSE_CA,
|
|
verify=CLICKHOUSE_VERIFY,
|
|
connections_min=20,
|
|
connections_max=1000,
|
|
)
|
|
|
|
def async_execute(query, args=None, settings=None):
|
|
return sync_execute(query, args, settings=settings)
|
|
|
|
def cache_sync_execute(query, args=None, redis_client=None, ttl=CACHE_TTL, settings=None):
|
|
if not redis_client:
|
|
redis_client = redis.get_client()
|
|
key = _key_hash(query, args)
|
|
if redis_client.exists(key):
|
|
result = _deserialize(redis_client.get(key))
|
|
return result
|
|
else:
|
|
result = sync_execute(query, args, settings=settings)
|
|
redis_client.set(key, _serialize(result), ex=ttl)
|
|
return result
|
|
|
|
def sync_execute(query, args=None, settings=None):
|
|
with ch_pool.get_client() as client:
|
|
start_time = time()
|
|
try:
|
|
result = client.execute(query, args, settings=settings)
|
|
finally:
|
|
execution_time = time() - start_time
|
|
g = statsd.Gauge("%s_clickhouse_sync_execution_time" % (STATSD_PREFIX,))
|
|
g.send("clickhouse_sync_query_time", execution_time)
|
|
if app_settings.SHELL_PLUS_PRINT_SQL:
|
|
print(format_sql(query, args))
|
|
print("Execution time: %.6fs" % (execution_time,))
|
|
if _save_query_user_id:
|
|
save_query(query, args, execution_time)
|
|
return result
|
|
|
|
|
|
def _deserialize(result_bytes: bytes) -> List[Tuple]:
|
|
results = []
|
|
for x in json.loads(result_bytes):
|
|
results.append(tuple(x))
|
|
return results
|
|
|
|
|
|
def _serialize(result: Any) -> bytes:
|
|
return json.dumps(result).encode("utf-8")
|
|
|
|
|
|
def _key_hash(query: str, args: Any) -> bytes:
|
|
key = hashlib.md5(query.encode("utf-8") + json.dumps(args).encode("utf-8")).digest()
|
|
return key
|
|
|
|
|
|
def format_sql(sql, params, colorize=True):
|
|
substitute_params = (
|
|
ch_client.substitute_params if isinstance(ch_client, SyncClient) else ch_client._client.substitute_params
|
|
)
|
|
sql = substitute_params(sql, params or {})
|
|
sql = sqlparse.format(sql, reindent_aligned=True)
|
|
if colorize:
|
|
try:
|
|
import pygments.formatters
|
|
import pygments.lexers
|
|
|
|
sql = pygments.highlight(
|
|
sql, pygments.lexers.get_lexer_by_name("sql"), pygments.formatters.TerminalFormatter()
|
|
)
|
|
except:
|
|
pass
|
|
|
|
return sql
|
|
|
|
|
|
def save_query(sql: str, params: Dict, execution_time: float) -> None:
|
|
"""
|
|
Save query for debugging purposes
|
|
"""
|
|
|
|
try:
|
|
key = "save_query_{}".format(_save_query_user_id)
|
|
queries = json.loads(get_safe_cache(key) or "[]")
|
|
|
|
queries.insert(
|
|
0,
|
|
{
|
|
"timestamp": now().isoformat(),
|
|
"query": format_sql(sql, params, colorize=False),
|
|
"execution_time": execution_time,
|
|
},
|
|
)
|
|
cache.set(key, json.dumps(queries), timeout=120)
|
|
except Exception as e:
|
|
capture_exception(e)
|