2020-04-06 21:39:52 +00:00
|
|
|
import asyncio
|
|
|
|
import contextlib
|
2020-04-07 17:42:02 +00:00
|
|
|
import dataclasses
|
2020-04-06 21:39:52 +00:00
|
|
|
import logging
|
|
|
|
import os
|
2020-04-07 17:42:02 +00:00
|
|
|
import time
|
2020-04-06 21:39:52 +00:00
|
|
|
import typing
|
|
|
|
from distutils import util
|
|
|
|
|
|
|
|
import aiopg
|
|
|
|
import psycopg2
|
|
|
|
from aiopg import pool
|
2020-04-07 17:42:02 +00:00
|
|
|
from psycopg2 import errors, extras
|
2020-04-06 21:39:52 +00:00
|
|
|
from tornado import ioloop, web
|
|
|
|
|
|
|
|
LOGGER = logging.getLogger('sprockets-postgres')
|
|
|
|
|
|
|
|
DEFAULT_POSTGRES_CONNECTION_TIMEOUT = 10
|
|
|
|
DEFAULT_POSTGRES_CONNECTION_TTL = 300
|
|
|
|
DEFAULT_POSTGRES_HSTORE = 'FALSE'
|
|
|
|
DEFAULT_POSTGRES_JSON = 'FALSE'
|
|
|
|
DEFAULT_POSTGRES_MAX_POOL_SIZE = 0
|
|
|
|
DEFAULT_POSTGRES_MIN_POOL_SIZE = 1
|
|
|
|
DEFAULT_POSTGRES_QUERY_TIMEOUT = 120
|
|
|
|
DEFAULT_POSTGRES_UUID = 'TRUE'
|
|
|
|
|
2020-04-07 20:59:06 +00:00
|
|
|
QueryParameters = typing.Union[dict, list, tuple, None]
|
2020-04-07 17:42:02 +00:00
|
|
|
Timeout = typing.Union[int, float, None]
|
|
|
|
|
|
|
|
|
|
|
|
@dataclasses.dataclass
|
|
|
|
class QueryResult:
|
|
|
|
row_count: int
|
|
|
|
row: typing.Optional[dict]
|
|
|
|
rows: typing.Optional[typing.List[dict]]
|
|
|
|
|
|
|
|
|
|
|
|
class PostgresConnector:
|
|
|
|
|
|
|
|
def __init__(self,
|
|
|
|
cursor: aiopg.Cursor,
|
|
|
|
on_error: typing.Callable,
|
|
|
|
record_duration: typing.Optional[typing.Callable] = None,
|
|
|
|
timeout: Timeout = None):
|
|
|
|
self.cursor = cursor
|
|
|
|
self._on_error = on_error
|
|
|
|
self._record_duration = record_duration
|
|
|
|
self._timeout = timeout or int(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_QUERY_TIMEOUT',
|
|
|
|
DEFAULT_POSTGRES_QUERY_TIMEOUT))
|
|
|
|
|
|
|
|
async def callproc(self,
|
|
|
|
name: str,
|
|
|
|
parameters: QueryParameters = None,
|
|
|
|
metric_name: str = '',
|
|
|
|
*,
|
|
|
|
timeout: Timeout = None) -> QueryResult:
|
|
|
|
return await self._query(
|
|
|
|
self.cursor.callproc,
|
|
|
|
metric_name,
|
|
|
|
procname=name,
|
|
|
|
parameters=parameters,
|
|
|
|
timeout=timeout)
|
|
|
|
|
|
|
|
async def execute(self,
|
|
|
|
sql: str,
|
|
|
|
parameters: QueryParameters = None,
|
|
|
|
metric_name: str = '',
|
|
|
|
*,
|
|
|
|
timeout: Timeout = None) -> QueryResult:
|
|
|
|
return await self._query(
|
|
|
|
self.cursor.execute,
|
|
|
|
metric_name,
|
|
|
|
operation=sql,
|
|
|
|
parameters=parameters,
|
|
|
|
timeout=timeout)
|
|
|
|
|
|
|
|
@contextlib.asynccontextmanager
|
|
|
|
async def transaction(self) \
|
|
|
|
-> typing.AsyncContextManager['PostgresConnector']:
|
|
|
|
async with self.cursor.begin():
|
|
|
|
yield self
|
|
|
|
|
|
|
|
async def _query(self,
|
|
|
|
method: typing.Callable,
|
|
|
|
metric_name: str,
|
|
|
|
**kwargs):
|
|
|
|
if kwargs['timeout'] is None:
|
|
|
|
kwargs['timeout'] = self._timeout
|
|
|
|
start_time = time.monotonic()
|
|
|
|
try:
|
|
|
|
await method(**kwargs)
|
|
|
|
except (asyncio.TimeoutError, psycopg2.Error) as err:
|
|
|
|
exc = self._on_error(metric_name, err)
|
|
|
|
if exc:
|
|
|
|
raise exc
|
2020-04-07 20:59:06 +00:00
|
|
|
else:
|
2020-04-07 17:42:02 +00:00
|
|
|
if self._record_duration:
|
|
|
|
self._record_duration(
|
|
|
|
metric_name, time.monotonic() - start_time)
|
|
|
|
return await self._query_results()
|
|
|
|
|
|
|
|
async def _query_results(self) -> QueryResult:
|
2020-04-07 20:59:06 +00:00
|
|
|
count, row, rows = self.cursor.rowcount, None, None
|
2020-04-07 17:42:02 +00:00
|
|
|
if self.cursor.rowcount == 1:
|
|
|
|
try:
|
|
|
|
row = dict(await self.cursor.fetchone())
|
|
|
|
except psycopg2.ProgrammingError:
|
|
|
|
pass
|
|
|
|
elif self.cursor.rowcount > 1:
|
|
|
|
try:
|
|
|
|
rows = [dict(row) for row in await self.cursor.fetchall()]
|
|
|
|
except psycopg2.ProgrammingError:
|
|
|
|
pass
|
2020-04-07 20:59:06 +00:00
|
|
|
return QueryResult(count, row, rows)
|
2020-04-07 17:42:02 +00:00
|
|
|
|
|
|
|
|
|
|
|
class ConnectionException(Exception):
|
|
|
|
"""Raised when the connection to Postgres can not be established"""
|
|
|
|
|
2020-04-06 21:39:52 +00:00
|
|
|
|
|
|
|
class ApplicationMixin:
|
2020-04-07 17:42:02 +00:00
|
|
|
"""
|
|
|
|
:class:`sprockets.http.app.Application` mixin for handling the connection
|
|
|
|
to Postgres and exporting functions for querying the database,
|
|
|
|
getting the status, and proving a cursor.
|
2020-04-06 21:39:52 +00:00
|
|
|
|
2020-04-07 17:42:02 +00:00
|
|
|
Automatically creates and shuts down :class:`aio.pool.Pool` on startup
|
|
|
|
and shutdown.
|
|
|
|
|
|
|
|
"""
|
2020-04-06 21:39:52 +00:00
|
|
|
POSTGRES_STATUS_TIMEOUT = 3
|
|
|
|
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
|
|
super().__init__(*args, **kwargs)
|
|
|
|
self._postgres_pool: typing.Optional[pool.Pool] = None
|
|
|
|
self.runner_callbacks['on_start'].append(self._postgres_setup)
|
|
|
|
self.runner_callbacks['shutdown'].append(self._postgres_shutdown)
|
|
|
|
|
|
|
|
@contextlib.asynccontextmanager
|
2020-04-07 17:42:02 +00:00
|
|
|
async def postgres_connector(self,
|
|
|
|
on_error: typing.Callable,
|
|
|
|
record_duration: typing.Optional[
|
|
|
|
typing.Callable] = None,
|
|
|
|
timeout: Timeout = None) \
|
|
|
|
-> typing.AsyncContextManager[PostgresConnector]:
|
2020-04-06 21:39:52 +00:00
|
|
|
try:
|
|
|
|
async with self._postgres_pool.acquire() as conn:
|
|
|
|
async with conn.cursor(
|
|
|
|
cursor_factory=extras.RealDictCursor,
|
2020-04-07 17:42:02 +00:00
|
|
|
timeout=timeout) as cursor:
|
|
|
|
yield PostgresConnector(
|
|
|
|
cursor, on_error, record_duration, timeout)
|
|
|
|
except (asyncio.TimeoutError, psycopg2.Error) as err:
|
|
|
|
on_error('postgres_connector', ConnectionException(str(err)))
|
2020-04-06 21:39:52 +00:00
|
|
|
|
|
|
|
async def postgres_status(self) -> dict:
|
|
|
|
"""Invoke from the ``/status`` RequestHandler to check that there is
|
|
|
|
a Postgres connection handler available and return info about the
|
|
|
|
pool.
|
|
|
|
|
2020-04-07 17:42:02 +00:00
|
|
|
"""
|
|
|
|
query_error = asyncio.Event()
|
|
|
|
|
|
|
|
def on_error(_metric_name, _exc) -> None:
|
|
|
|
query_error.set()
|
|
|
|
return None
|
|
|
|
|
|
|
|
async with self.postgres_connector(
|
|
|
|
on_error,
|
|
|
|
timeout=self.POSTGRES_STATUS_TIMEOUT) as connector:
|
|
|
|
await connector.execute('SELECT 1')
|
2020-04-06 21:39:52 +00:00
|
|
|
return {
|
2020-04-07 17:42:02 +00:00
|
|
|
'available': not query_error.is_set(),
|
2020-04-06 21:39:52 +00:00
|
|
|
'pool_size': self._postgres_pool.size,
|
|
|
|
'pool_free': self._postgres_pool.freesize
|
|
|
|
}
|
|
|
|
|
|
|
|
async def _postgres_setup(self,
|
|
|
|
_app: web.Application,
|
2020-04-07 17:42:02 +00:00
|
|
|
loop: ioloop.IOLoop) -> None:
|
2020-04-06 21:39:52 +00:00
|
|
|
"""Setup the Postgres pool of connections and log if there is an error.
|
|
|
|
|
|
|
|
This is invoked by the Application on start callback mechanism.
|
|
|
|
|
|
|
|
"""
|
2020-04-07 17:42:02 +00:00
|
|
|
if 'POSTGRES_URL' not in os.environ:
|
|
|
|
LOGGER.critical('Missing POSTGRES_URL environment variable')
|
|
|
|
return self.stop(loop)
|
2020-04-06 21:39:52 +00:00
|
|
|
self._postgres_pool = pool.Pool(
|
2020-04-07 17:42:02 +00:00
|
|
|
os.environ['POSTGRES_URL'],
|
2020-04-06 21:39:52 +00:00
|
|
|
minsize=int(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_MIN_POOL_SIZE',
|
|
|
|
DEFAULT_POSTGRES_MIN_POOL_SIZE)),
|
|
|
|
maxsize=int(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_MAX_POOL_SIZE',
|
|
|
|
DEFAULT_POSTGRES_MAX_POOL_SIZE)),
|
|
|
|
timeout=int(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_CONNECT_TIMEOUT',
|
|
|
|
DEFAULT_POSTGRES_CONNECTION_TIMEOUT)),
|
|
|
|
enable_hstore=util.strtobool(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_HSTORE', DEFAULT_POSTGRES_HSTORE)),
|
|
|
|
enable_json=util.strtobool(
|
|
|
|
os.environ.get('POSTGRES_JSON', DEFAULT_POSTGRES_JSON)),
|
|
|
|
enable_uuid=util.strtobool(
|
|
|
|
os.environ.get('POSTGRES_UUID', DEFAULT_POSTGRES_UUID)),
|
|
|
|
echo=False,
|
|
|
|
on_connect=None,
|
|
|
|
pool_recycle=int(
|
|
|
|
os.environ.get(
|
|
|
|
'POSTGRES_CONNECTION_TTL',
|
|
|
|
DEFAULT_POSTGRES_CONNECTION_TTL)))
|
|
|
|
try:
|
|
|
|
async with self._postgres_pool._cond:
|
|
|
|
await self._postgres_pool._fill_free_pool(False)
|
|
|
|
except (psycopg2.OperationalError,
|
|
|
|
psycopg2.Error) as error: # pragma: nocover
|
|
|
|
LOGGER.warning('Error connecting to PostgreSQL on startup: %s',
|
|
|
|
error)
|
|
|
|
|
|
|
|
async def _postgres_shutdown(self, _ioloop: ioloop.IOLoop) -> None:
|
|
|
|
"""Shutdown the Postgres connections and wait for them to close.
|
|
|
|
|
|
|
|
This is invoked by the Application shutdown callback mechanism.
|
|
|
|
|
|
|
|
"""
|
|
|
|
self._postgres_pool.close()
|
|
|
|
await self._postgres_pool.wait_closed()
|
2020-04-07 17:42:02 +00:00
|
|
|
|
|
|
|
|
|
|
|
class RequestHandlerMixin:
|
|
|
|
"""
|
|
|
|
RequestHandler mixin class exposing functions for querying the database,
|
|
|
|
recording the duration to either `sprockets-influxdb` or
|
|
|
|
`sprockets.mixins.metrics`, and handling exceptions.
|
|
|
|
|
|
|
|
"""
|
|
|
|
async def postgres_callproc(self,
|
|
|
|
name: str,
|
|
|
|
parameters: QueryParameters = None,
|
|
|
|
metric_name: str = '',
|
|
|
|
*,
|
|
|
|
timeout: Timeout = None) -> QueryResult:
|
2020-04-07 20:59:06 +00:00
|
|
|
async with self.application.postgres_connector(
|
|
|
|
self._on_postgres_error,
|
|
|
|
self._on_postgres_timing,
|
|
|
|
timeout) as connector:
|
2020-04-07 17:42:02 +00:00
|
|
|
return await connector.callproc(
|
|
|
|
name, parameters, metric_name, timeout=timeout)
|
|
|
|
|
|
|
|
async def postgres_execute(self,
|
|
|
|
sql: str,
|
|
|
|
parameters: QueryParameters = None,
|
|
|
|
metric_name: str = '',
|
|
|
|
*,
|
|
|
|
timeout: Timeout = None) -> QueryResult:
|
|
|
|
"""Execute a query, specifying a name for the query, the SQL statement,
|
|
|
|
and optional positional arguments to pass in with the query.
|
|
|
|
|
|
|
|
Parameters may be provided as sequence or mapping and will be
|
|
|
|
bound to variables in the operation. Variables are specified
|
|
|
|
either with positional ``%s`` or named ``%({name})s`` placeholders.
|
|
|
|
|
|
|
|
"""
|
2020-04-07 20:59:06 +00:00
|
|
|
async with self.application.postgres_connector(
|
|
|
|
self._on_postgres_error,
|
|
|
|
self._on_postgres_timing,
|
|
|
|
timeout) as connector:
|
2020-04-07 17:42:02 +00:00
|
|
|
return await connector.execute(
|
|
|
|
sql, parameters, metric_name, timeout=timeout)
|
|
|
|
|
|
|
|
@contextlib.asynccontextmanager
|
|
|
|
async def postgres_transaction(self, timeout: Timeout = None) \
|
|
|
|
-> typing.AsyncContextManager[PostgresConnector]:
|
|
|
|
"""Yields a :class:`PostgresConnector` instance in a transaction.
|
|
|
|
Will automatically commit or rollback based upon exception.
|
|
|
|
|
|
|
|
"""
|
|
|
|
async with self.application.postgres_connector(
|
2020-04-07 20:59:06 +00:00
|
|
|
self._on_postgres_error,
|
|
|
|
self._on_postgres_timing,
|
2020-04-07 17:42:02 +00:00
|
|
|
timeout) as connector:
|
2020-04-07 20:59:06 +00:00
|
|
|
async with connector.transaction():
|
|
|
|
yield connector
|
2020-04-07 17:42:02 +00:00
|
|
|
|
2020-04-07 20:59:06 +00:00
|
|
|
def _on_postgres_error(self,
|
|
|
|
metric_name: str,
|
|
|
|
exc: Exception) -> typing.Optional[Exception]:
|
2020-04-07 17:42:02 +00:00
|
|
|
"""Override for different error handling behaviors"""
|
|
|
|
LOGGER.error('%s in %s for %s (%s)',
|
2020-04-07 20:59:06 +00:00
|
|
|
exc.__class__.__name__, self.__class__.__name__,
|
|
|
|
metric_name, str(exc).split('\n')[0])
|
2020-04-07 17:42:02 +00:00
|
|
|
if isinstance(exc, ConnectionException):
|
|
|
|
raise web.HTTPError(503, reason='Database Connection Error')
|
|
|
|
elif isinstance(exc, asyncio.TimeoutError):
|
|
|
|
raise web.HTTPError(500, reason='Query Timeout')
|
|
|
|
elif isinstance(exc, errors.UniqueViolation):
|
|
|
|
raise web.HTTPError(409, reason='Unique Violation')
|
|
|
|
elif isinstance(exc, psycopg2.Error):
|
|
|
|
raise web.HTTPError(500, reason='Database Error')
|
|
|
|
return exc
|
|
|
|
|
2020-04-07 20:59:06 +00:00
|
|
|
def _on_postgres_timing(self,
|
|
|
|
metric_name: str,
|
|
|
|
duration: float) -> None:
|
2020-04-07 17:42:02 +00:00
|
|
|
"""Override for custom metric recording"""
|
|
|
|
if hasattr(self, 'influxdb'): # sprockets-influxdb
|
|
|
|
self.influxdb.set_field(metric_name, duration)
|
|
|
|
elif hasattr(self, 'record_timing'): # sprockets.mixins.metrics
|
|
|
|
self.record_timing(metric_name, duration)
|
|
|
|
else:
|
|
|
|
LOGGER.debug('Postgres query %s duration: %s',
|
|
|
|
metric_name, duration)
|