Connection Pooling  {@name=pooling}
======================

This section describes the connection pool module of SQLAlchemy.  The `Pool` object it provides is normally embedded within an `Engine` instance.  For most cases, explicit access to the pool module is not required.  However, the `Pool` object can be used on its own, without the rest of SA, to manage DBAPI connections; this section describes that usage.  Also, this section will describe in more detail how to customize the pooling strategy used by an `Engine`.

At the base of any database helper library is a system of efficiently acquiring connections to the database.  Since the establishment of a database connection is typically a somewhat expensive operation, an application needs a way to get at database connections repeatedly without incurring the full overhead each time.  Particularly for server-side web applications, a connection pool is the standard way to maintain a "pool" of database connections which are used over and over again among many requests.  Connection pools typically are configured to maintain a certain "size", which represents how many connections can be used simultaneously without resorting to creating more newly-established connections.

### Establishing a Transparent Connection Pool {@name=establishing}

Any DBAPI module can be "proxied" through the connection pool using the following technique (note that the usage of 'psycopg2' is **just an example**; substitute whatever DBAPI module you'd like):
    
    {python}
    import sqlalchemy.pool as pool
    import psycopg2 as psycopg
    psycopg = pool.manage(psycopg)
    
    # then connect normally
    connection = psycopg.connect(database='test', username='scott', password='tiger')

This produces a `sqlalchemy.pool.DBProxy` object which supports the same `connect()` function as the original DBAPI module.  Upon connection, a connection proxy object is returned, which delegates its calls to a real DBAPI connection object.  This connection object is stored persistently within a connection pool (an instance of `sqlalchemy.pool.Pool`) that corresponds to the exact connection arguments sent to the `connect()` function.  

The connection proxy supports all of the methods on the original connection object, most of which are proxied via `__getattr__()`.  The `close()` method will return the connection to the pool, and the `cursor()` method will return a proxied cursor object.  Both the connection proxy and the cursor proxy will also return the underlying connection to the pool after they have both been garbage collected, which is detected via the `__del__()` method.

Additionally, when connections are returned to the pool, a `rollback()` is issued on the connection unconditionally.  This is to release any locks still held by the connection that may have resulted from normal activity.

By default, the `connect()` method will return the same connection that is already checked out in the current thread.  This allows a particular connection to be used in a given thread without needing to pass it around between functions.  To disable this behavior, specify `use_threadlocal=False` to the `manage()` function.

### Connection Pool Configuration {@name=configuration}

For all types of Pool construction, which includes the "transparent proxy" described in the previous section, using an `Engine` via `create_engine()`, or constructing a pool through direct class instantiation, the options are generally the same.  Additional options may be available based on the specific subclass of `Pool` being used.

For a description of all pool classes, see the [generated documentation](rel:docstrings_sqlalchemy.pool).

Common options include:

 * echo=False : if set to True, connections being pulled and retrieved from/to the pool will
   be logged to the standard output, as well as pool sizing information.  Echoing can also
   be achieved by enabling logging for the "sqlalchemy.pool" namespace.  When using create_engine(), 
   this option is specified as `echo_pool`.
 * use_threadlocal=False : if set to True, repeated calls to connect() within the same
   application thread will be guaranteed to return the same connection object, if one has
   already been retrieved from the pool and has not been returned yet. This allows code to
   retrieve a connection from the pool, and then while still holding on to that connection,
   to call other functions which also ask the pool for a connection of the same arguments;
   those functions will act upon the same connection that the calling method is using.
   This option is overridden during `create_engine()`, corresponding to the "plain" or 
   "threadlocal" connection strategy.
 * recycle=-1 : if set to non -1, a number of seconds between connection recycling, which
   means upon checkout, if this timeout is surpassed the connection will be closed and replaced
   with a newly opened connection.
 * auto_close_cursors = True : cursors, returned by connection.cursor(), are tracked and are 
   automatically closed when the connection is returned to the pool.  some DBAPIs like MySQLDB
   become unstable if cursors remain open.
 *  disallow_open_cursors = False : if auto_close_cursors is False, and disallow_open_cursors is True,
   will raise an exception if an open cursor is detected upon connection checkin.  If auto_close_cursors and disallow_open_cursors are both False, then no cursor processing occurs upon checkin.

QueuePool options include:

 * pool_size=5 : the size of the pool to be maintained. This is the
 largest number of connections that will be kept persistently in the pool. Note that the
 pool begins with no connections; once this number of connections is requested, that
 number of connections will remain.
 * max_overflow=10 : the maximum overflow size of the pool. When the number of checked-out
 connections reaches the size set in pool_size, additional connections will be returned up
 to this limit. When those additional connections are returned to the pool, they are
 disconnected and discarded. It follows then that the total number of simultaneous
 connections the pool will allow is pool_size + max_overflow, and the total number of
 "sleeping" connections the pool will allow is pool_size. max_overflow can be set to -1 to
 indicate no overflow limit; no limit will be placed on the total number of concurrent
 connections.
 * timeout=30 : the number of seconds to wait before giving up on returning a connection
 
### Custom Pool Construction {@name=custom}

Besides using the transparent proxy, instances of `sqlalchemy.pool.Pool` can be created directly.  Constructing your own pool involves passing a callable used to create a connection.  Through this method, custom connection schemes can be made, such as a connection that automatically executes some initialization commands to start.  

    {python title="Constructing a QueuePool"}
    import sqlalchemy.pool as pool
    import psycopg2
    
    def getconn():
        c = psycopg2.connect(username='ed', host='127.0.0.1', dbname='test')
        # execute an initialization function on the connection before returning
        c.cursor.execute("setup_encodings()")
        return c
        
    p = pool.QueuePool(getconn, max_overflow=10, pool_size=5, use_threadlocal=True)
    
Or with SingletonThreadPool:

    {python title="Constructing a SingletonThreadPool"}
    import sqlalchemy.pool as pool
    import sqlite
    
    def getconn():
        return sqlite.connect(filename='myfile.db')
    
    # SQLite connections require the SingletonThreadPool    
    p = pool.SingletonThreadPool(getconn)
    
