Update longqueries tool: add client IP, waiting attributes.
# -*- coding: utf-8 -*-
#
# PgManager - manage database connections
#
# Requires: Python 3.2, psycopg2
#
# Part of pgtoolkit
# http://hg.devl.cz/pgtoolkit
#
# Copyright (c) 2010, 2011, 2012 Radek Brich <radek.brich@devl.cz>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
# THE SOFTWARE.
"""Postgres database connection manager
PgManager wraps psycopg2 connect function, adding following features:
* Manage database connection parameters - link connection parameters
to an unique identifier, retrieve connection object by this identifier
* Connection pooling - connections with same identifier are pooled and reused
* Easy query using the with statement - retrieve cursor directly by connection
identifier, don't worry about connections
* Dict rows - cursor has additional methods like fetchall_dict(), which
returns dict row instead of ordinary list-like row
Example:
import pgmanager
pgm = pgmanager.get_instance()
pgm.create_conn(hostaddr='127.0.0.1', dbname='postgres')
with pgm.cursor() as curs:
curs.execute('SELECT now() AS now')
row = curs.fetchone_dict()
print row.now
First, we have obtained PgManager instance. This is like calling
PgManager(), although in our example the instance is global. That means
getting the instance in another module brings us all the defined connections
etc.
On second line we have created connection named 'default' (this name can be left out).
The with statement obtains connection (actually connects to database when needed),
then returns cursor for this connection. At the end of with statement,
the connection is returned to the pool or closed (depending on number of connections
in pool and on setting of pool_size parameter).
The row returned by fetchone_dict() is special dict object, which can be accessed
using item or attribute access, that is row['now'] or row.now.
"""
from contextlib import contextmanager
from collections import OrderedDict
import logging
import threading
import multiprocessing
import select
import socket
import psycopg2
import psycopg2.extensions
from psycopg2 import DatabaseError, IntegrityError, OperationalError
log_sql = logging.getLogger("pgmanager_sql")
log_notices = logging.getLogger("pgmanager_notices")
class PgManagerError(Exception):
pass
class ConnectionInfo:
def __init__(self, name, dsn, isolation_level=None, keep_alive=True,
init_statement=None, pool_size=1):
self.name = name # connection name is logged with SQL queries
self.dsn = dsn # dsn or string with connection parameters
self.isolation_level = isolation_level
self.keep_alive = keep_alive
self.init_statement = init_statement
self.pool_size = pool_size
class RowDict(OrderedDict):
"""Special read-only dictionary used for rows returned from queries.
Initialization is same as for dict:
row = RowDict([('id', 123), ('name', 'hello')])
Allows key and attribute access to contained items:
row['id']
row.id
Items keep order in which columns where returned from database.
Tuple style access is also supported:
row[0]
id, name = row
"""
def __init__(self, data):
self._dict = OrderedDict(data)
def __getitem__(self, key):
if isinstance(key, int):
return tuple(self._dict.values())[key]
return self._dict[key]
def __getattr__(self, key):
try:
return self._dict[key]
except KeyError:
raise AttributeError(key)
def __contains__(self, key):
return key in self._dict
class Cursor(psycopg2.extensions.cursor):
def execute(self, query, args=None):
self._log_query(query, args)
try:
return super(Cursor, self).execute(query, args)
except DatabaseError:
self._log_exception()
def callproc(self, procname, args=None):
self._log_query('CALL %s(%s)' % (procname, ','.join(args)))
try:
return super(Cursor, self).callproc(procname, args)
except DatabaseError:
self._log_exception()
def row_dict(self, row, lstrip=None):
adjustname = lambda a: a
if lstrip:
adjustname = lambda a: a.lstrip(lstrip)
return RowDict(zip([adjustname(desc[0]) for desc in self.description], row))
def fetchone_dict(self, lstrip=None):
'''Return one row as OrderedDict'''
row = super(Cursor, self).fetchone()
if row is None:
return None
return self.row_dict(row, lstrip)
def fetchall_dict(self, lstrip=None):
'''Return all rows as OrderedDict'''
rows = super(Cursor, self).fetchall()
return [self.row_dict(row, lstrip) for row in rows]
def adapt(self, row):
if isinstance(row, RowDict):
# dict
adapted = RowDict()
for key in row.keys():
adapted[key] = self.mogrify('%s', [row[key]]).decode('utf8')
else:
# list
adapted = [self.mogrify('%s', [x]).decode('utf8') for x in row]
return adapted
def fetchone_adapted(self, lstrip=None):
'''Like fetchone_dict() but values are quoted for direct inclusion in SQL query.
This is useful when you need to generate SQL script from data returned
by the query. Use mogrify() for simple cases.
'''
row = super(Cursor, self).fetchone()
if row is None:
return None
return self.row_dict([self.mogrify('%s', [x]).decode('utf8') for x in row], lstrip)
def fetchall_adapted(self, lstrip=None):
'''Like fetchall_dict() but values are quoted for direct inclusion in SQL query.'''
rows = super(Cursor, self).fetchall()
return [self.row_dict([self.mogrify('%s', [x]).decode('utf8') for x in row], lstrip) for row in rows]
def _log_query(self, query, args=None):
name = self.connection.name if hasattr(self.connection, 'name') else '-'
log_sql.info('[%s] %s' % (name, self.mogrify(query, args).decode('utf8')))
def _log_exception(self):
name = self.connection.name if hasattr(self.connection, 'name') else '-'
log_sql.exception('[%s] exception:' % (name,))
class Connection(psycopg2.extensions.connection):
def cursor(self, name=None):
if name is None:
return super(Connection, self).cursor(cursor_factory=Cursor)
else:
return super(Connection, self).cursor(name, cursor_factory=Cursor)
def keep_alive(self):
'''Set socket to keepalive mode. Must be called before any query.'''
sock = socket.fromfd(self.fileno(), socket.AF_INET, socket.SOCK_STREAM)
sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1)
try:
# Maximum keep-alive probes before asuming the connection is lost
sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPCNT, 5)
# Interval (in seconds) between keep-alive probes
sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPINTVL, 2)
# Maximum idle time (in seconds) before start sending keep-alive probes
sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPIDLE, 10)
except socket.error:
pass
# close duplicated fd, options set for socket stays
sock.close()
class PgManager:
def __init__(self):
self.conn_known = {} # available connections
self.conn_pool = {} # active connetions
self.lock = threading.Lock() # mutual exclusion for threads
self.pid = multiprocessing.current_process().pid # forking check
def __del__(self):
for conn in tuple(self.conn_known.keys()):
self.destroy_conn(conn)
def create_conn(self, name='default', isolation_level=None, keep_alive=True, init_statement=None,
pool_size=1, dsn=None, **kwargs):
'''Create named connection.
name -- name for connection (default is "default")
pool_size -- how many connections will be kept open in pool
(more connections will still be created but they will be closed by put_conn)
None - disable pool, always return same connection
isolation_level -- "autocommit", "read_committed", "serializable" or None for driver default
keep_alive -- set socket to keepalive mode
dsn -- connection string (parameters or data source name)
Other keyword args are used as connection parameters.
'''
if name in self.conn_known:
raise PgManagerError('Connection name "%s" already registered.' % name)
if dsn is None:
dsn = ' '.join([x[0]+'='+str(x[1]) for x in kwargs.items() if x[1] is not None])
isolation_level = self._normalize_isolation_level(isolation_level)
ci = ConnectionInfo(name, dsn, isolation_level, keep_alive, init_statement, pool_size)
self.conn_known[name] = ci
self.conn_pool[name] = []
def create_conn_listen(self, name, channel, dsn=None, copy_dsn=None, **kwargs):
'''Create connection listening for notifies.
Disables pool. If you want to use pool, create other connection for that.
This connection can be used as usual: conn.cursor() etc.
Don't use PgManager's cursor() and put_conn().
name -- name for connection
channel -- listen on this channel
copy_dsn -- specify name of other connection and its dsn will be used
Other parameters forwarded to create_conn().
'''
if dsn is None and copy_dsn:
try:
dsn = self.conn_known[copy_dsn].dsn
except KeyError:
raise PgManagerError("Connection name '%s' not registered." % copy_dsn)
listen_query = "LISTEN " + channel
self.create_conn(name=name, pool_size=None, isolation_level='autocommit', init_statement=listen_query,
dsn=dsn, **kwargs)
def close_conn(self, name='default'):
'''Close all connections of given name.
Connection credentials are still saved.
'''
while len(self.conn_pool[name]):
conn = self.conn_pool[name].pop()
conn.close()
def destroy_conn(self, name='default'):
'''Destroy connection.
Counterpart of create_conn.
'''
if not name in self.conn_known:
raise PgManagerError('Connection name "%s" not registered.' % name)
self.close_conn(name)
del self.conn_known[name]
del self.conn_pool[name]
def get_conn(self, name='default'):
'''Get connection of name 'name' from pool.'''
self._check_fork()
self.lock.acquire()
try:
try:
ci = self.conn_known[name]
except KeyError:
raise PgManagerError("Connection name '%s' not registered." % name)
# no pool, just one static connection
if ci.pool_size is None:
# check for existing connection
try:
conn = self.conn_pool[name][0]
if conn.closed:
conn = None
except IndexError:
conn = None
self.conn_pool[name].append(conn)
# if no existing connection is valid, connect new one and save it
if conn is None:
conn = self._connect(ci)
self.conn_pool[name][0] = conn
# connection from pool
else:
conn = None
while len(self.conn_pool[name]) and conn is None:
conn = self.conn_pool[name].pop()
if conn.closed:
conn = None
if conn is None:
conn = self._connect(ci)
finally:
self.lock.release()
return conn
def put_conn(self, conn, name='default'):
'''Put connection back to pool.
Name must be same as used for get_conn,
otherwise things become broken.
'''
self.lock.acquire()
try:
if not name in self.conn_known:
raise PgManagerError("Connection name '%s' not registered." % name)
if len(self.conn_pool[name]) >= self.conn_known[name].pool_size:
conn.close()
return
if conn.get_transaction_status() == psycopg2.extensions.TRANSACTION_STATUS_UNKNOWN:
conn.close()
return
# connection returned to the pool must not be in transaction
if conn.get_transaction_status() != psycopg2.extensions.TRANSACTION_STATUS_IDLE:
try:
conn.rollback()
except OperationalError:
if not conn.closed:
conn.close()
return
self.conn_pool[name].append(conn)
finally:
self.lock.release()
@contextmanager
def cursor(self, name='default'):
'''Cursor context.
Uses any connection of name 'name' from pool
and returns cursor for that connection.
'''
conn = self.get_conn(name)
try:
curs = conn.cursor()
yield curs
finally:
curs.close()
self.log_notices(conn)
self.put_conn(conn, name)
def log_notices(self, conn):
for notice in conn.notices:
log_notices.info(notice.rstrip())
conn.notices[:] = []
def wait_for_notify(self, name='default', timeout=5.0):
'''Wait for asynchronous notifies, return the last one.
name -- name of connection, must be created using create_conn_listen()
timeout -- in seconds, floating point (None - wait forever)
Returns None on timeout.
'''
conn = self.get_conn(name)
# return any notifies on stack
if conn.notifies:
return conn.notifies.pop()
if select.select([conn], [], [], timeout) == ([], [], []):
# timeout
return None
else:
conn.poll()
# return just the last notify (we do not care for older ones)
if conn.notifies:
return conn.notifies.pop()
return None
def _connect(self, ci):
conn = psycopg2.connect(ci.dsn, connection_factory=Connection)
conn.name = ci.name
if ci.keep_alive:
conn.keep_alive()
if not ci.isolation_level is None:
conn.set_isolation_level(ci.isolation_level)
if ci.init_statement:
curs = conn.cursor()
curs.execute(ci.init_statement)
curs.close()
return conn
def _normalize_isolation_level(self, level):
if type(level) == str:
if level.lower() == 'autocommit':
return psycopg2.extensions.ISOLATION_LEVEL_AUTOCOMMIT
if level.lower() == 'read_committed':
return psycopg2.extensions.ISOLATION_LEVEL_READ_COMMITTED
if level.lower() == 'serializable':
return psycopg2.extensions.ISOLATION_LEVEL_SERIALIZABLE
raise PgManagerError('Unknown isolation level name: "%s"', level)
return level
def _check_fork(self):
'''Check if process was forked (PID has changed).
If it was, clean parent's connections.
New connections are created for children.
Known connection credentials are inherited, but not shared.
'''
if self.pid == multiprocessing.current_process().pid:
# PID has not changed
return
# update saved PID
self.pid = multiprocessing.current_process().pid
# reinitialize lock
self.lock = threading.Lock()
# clean parent's connections
for name in self.conn_pool:
self.conn_pool[name] = []
@classmethod
def get_instance(cls):
if not hasattr(cls, '_instance'):
cls._instance = cls()
return cls._instance
def get_instance():
return PgManager.get_instance()