Easily monitor your Tornado-based application.
Find a file
2016-08-02 17:31:32 -04:00
docs Metadata bumps for 2.0.1. 2016-03-21 11:16:39 -04:00
examples Add install for statsd mixin 2016-07-18 15:40:44 -04:00
requires Require tornado 4.x to be present. 2016-02-01 14:51:48 -05:00
sprockets Fix typos 2016-08-02 17:31:32 -04:00
.gitignore SYN 2016-01-19 08:33:53 -05:00
.travis.yml travis: Enable PyPI uploads for tagged versions. 2016-01-27 10:54:38 -05:00
LICENSE SYN 2016-01-19 08:33:53 -05:00
MANIFEST.in SYN 2016-01-19 08:33:53 -05:00
README.rst Update README 2016-07-18 16:49:40 -04:00
setup.cfg setup.cfg: Enable universal wheels. 2016-01-27 10:56:57 -05:00
setup.py Metadata bumps for 1.0.0. 2016-02-01 10:18:04 -05:00
tests.py Add install for statsd mixin 2016-07-18 15:40:44 -04:00
tox.ini Clean up docs/environment etc. 2016-01-19 11:43:24 -05:00

sprockets.mixins.metrics
========================
Adjust counter and timer metrics in `InfluxDB`_ or `StatsD`_ using the same API.

The mix-in is configured through the ``tornado.web.Application`` settings
property using a key defined by the specific mix-in.

Statsd Mixin
------------

The following snippet configures the StatsD mix-in from common environment
variables. This simple handler will emit a timer metric that identifies each
call to the ``get`` method as well as a separate metric for the database query.

.. code-block:: python

   import os

   from sprockets.mixins import mediatype
   from sprockets.mixins.metrics import statsd
   from tornado import gen, web
   import queries

   def make_application():
       application = web.Application([
           web.url(r'/', MyHandler),
       ], **settings)

       statsd.install({'namespace': 'my-application',
                       'host': os.environ.get('STATSD_HOST', '127.0.0.1'),
                       'port': os.environ.get('STATSD_PORT', '8125')})
       return application

   class MyHandler(statsd.StatsdMixin,
                   mediatype.ContentMixin,
                   web.RequestHandler):

       def initialize(self):
           super(MyHandler, self).initialize()
           self.db = queries.TornadoSession(os.environ['MY_PGSQL_DSN'])

       @gen.coroutine
       def get(self, obj_id):
           with self.execution_timer('dbquery', 'get'):
              result = yield self.db.query('SELECT * FROM foo WHERE id=%s',
                                           obj_id)
           self.send_response(result)

Settings
^^^^^^^^

:namespace: The namespace for the measurements
:host: The Statsd host
:port: The Statsd port
:prepend_metric_type: Optional flag to prepend bucket path with the StatsD
    metric type
:prepend_hostname: Optional flag to prepend bucket path with the hostname

InfluxDB Mixin
--------------

The following snippet configures the InfluxDB mix-in from common environment
variables:

.. code-block:: python

   import os

   from sprockets.mixins.metrics import influxdb
   from sprockets.mixins import postgresql
   from tornado import gen, web

   def make_app(**settings):
       settings[influxdb.SETTINGS_KEY] = {
           'measurement': 'rollup',
       }

       application = web.Application(
           [
               web.url(r'/', MyHandler),
           ], **settings)

       influxdb.install({'url': 'http://localhost:8086',
                         'database': 'tornado-app'})
       return application


   class MyHandler(influxdb.InfluxDBMixin,
                   postgresql.HandlerMixin,
                   web.RequestHandler):

       @gen.coroutine
       def get(self, obj_id):
           with self.execution_timer('dbquery', 'get'):
              result = yield self.postgresql_session.query(
                  'SELECT * FROM foo WHERE id=%s', obj_id)
           self.send_response(result)

If your application handles signal handling for shutdowns, the
:meth:`~sprockets.mixins.influxdb.shutdown` method will try to cleanly ensure
that any buffered metrics in the InfluxDB collector are written prior to
shutting down. The method returns a :cls:`~tornado.concurrent.TracebackFuture`
that should be waited on prior to shutting down.

Settings
^^^^^^^^

:url: The InfluxDB API URL
:database: the database to write measurements into
:submission_interval: How often to submit metric batches in
   milliseconds. Default: ``5000``
:max_batch_size: The number of measurements to be submitted in a
   single HTTP request. Default: ``1000``
:tags: Default tags that are to be submitted with each metric. The tag
   ``hostname`` is added by default along with ``environment`` and ``service``
   if the corresponding ``ENVIRONMENT`` or ``SERVICE`` environment variables
   are set.

Development Quickstart
----------------------
.. code-block:: bash

   $ python3.4 -mvenv env
   $ . ./env/bin/activate
   (env)$ env/bin/pip install -r requires/development.txt
   (env)$ nosetests
   test_metrics_with_buffer_not_flush (tests.InfluxDbTests) ... ok
   test_that_cached_db_connection_is_used (tests.InfluxDbTests) ... ok
   test_that_counter_is_tracked (tests.InfluxDbTests) ... ok
   test_that_execution_timer_is_tracked (tests.InfluxDbTests) ... ok
   test_that_http_method_call_details_are_recorded (tests.InfluxDbTests) ... ok
   test_that_metric_tag_is_tracked (tests.InfluxDbTests) ... ok
   test_that_add_metric_tag_is_ignored (tests.StatsdMethodTimingTests) ... ok
   test_that_cached_socket_is_used (tests.StatsdMethodTimingTests) ... ok
   test_that_counter_accepts_increment_value (tests.StatsdMethodTimingTests) ... ok
   test_that_counter_increment_defaults_to_one (tests.StatsdMethodTimingTests) ... ok
   test_that_default_prefix_is_stored (tests.StatsdMethodTimingTests) ... ok
   test_that_execution_timer_records_time_spent (tests.StatsdMethodTimingTests) ... ok
   test_that_http_method_call_is_recorded (tests.StatsdMethodTimingTests) ... ok

   ----------------------------------------------------------------------
   Ran 13 tests in 3.572s

   OK
   (env)$ ./setup.py build_sphinx -q
   running build_sphinx
   (env)$ open build/sphinx/html/index.html

.. _StatsD: https://github.com/etsy/statsd
.. _InfluxDB: https://influxdata.com