--- title: Connection latency and timeouts subtitle: Learn about strategies to manage connection latencies and timeouts enableTableOfContents: true isDraft: false updatedOn: '2025-08-26T15:57:17.416Z' --- Neon's _Scale to zero_ feature is designed to minimize costs by automatically scaling a compute resource down to zero after a period of inactivity. By default, Neon scales a compute to zero after 5 minutes of inactivity. A characteristic of this feature is the concept of a "cold start". During this process, a compute transitions from an idle state to an active state to process requests. Currently, activating a Neon compute from an idle state typically takes a few hundred milliseconds not counting other factors that can add to latencies such as the physical distance between your application and database or startup times of other services that participate in your connection process. Services you integrate with Neon may also have startup times, which can add to connection latencies. This topic does not address latencies of other vendors, but if your application connects to Neon via another service, remember to consider startup times for those services as well. ## Check the status of a compute You can check the current status of a compute on the **Branches** page in the Neon Console. A compute will report either an **Active** or **Idle** status. ![Compute status](/docs/connect/compute_endpoint_state.png) You can also view compute state transitions in the **Branches** widget on the Neon **Dashboard**. User actions that activate an idle compute include connecting from a client or application, running a query on your database from the [Neon SQL Editor](/docs/get-started/query-with-neon-sql-editor), or accessing the compute via the [Neon API](https://api-docs.neon.tech/reference/getting-started-with-neon-api). The Neon API includes [Start endpoint](https://api-docs.neon.tech/reference/startprojectendpoint) and [Suspend endpoint](https://api-docs.neon.tech/reference/startprojectendpoint) APIs for the specific purpose of activating and suspending a compute. You can try any of these methods and watch the status of your compute as it changes from an **Idle** to an **Active** state. By default, a compute is suspended after 300 seconds (5 minutes) of inactivity. Users on the Neon [Scale plan](/docs/introduction/plans) can configure this delay period, which is described later in this topic. ## Strategies for managing latency and timeouts Given the potential impact on application responsiveness, it's important to have strategies in place to manage connection latencies and timeouts. Here are some methods you can implement: - [Adjust your Scale to zero configuration](#adjust-your-scale-to-zero-configuration) - [Place your application and database in the same region](#place-your-application-and-database-in-the-same-region) - [Increase your connection timeout](#increase-your-connection-timeout) - [Build connection timeout handling into your application](#build-connection-timeout-handling-into-your-application) - [Use application-level caching](#use-application-level-caching) ### Adjust your scale to zero configuration Users on paid plans can configure the length of time that the system remains in an inactive state before Neon scales your compute down to zero. This lets you set the balance between performance (never scaling down) and cost (scaling to zero at reasonable intervals). The scale to zero setting is set to 5 minutes by default. You can set a custom period of up to a maximum of 7 days, or disable scale to zero entirely. To disable scale to zero, see [Edit a compute](/docs/manage/endpoints#edit-a-compute). If you disable scale to zero entirely or your compute is never idle long enough to be automatically suspended, you will have to manually restart your compute to pick up the latest updates to Neon's compute images. Neon typically releases compute-related updates weekly. Not all releases contain critical updates, but a weekly compute restart is recommended to ensure that you do not miss anything important. For how to restart a compute, see [Restart a compute](/docs/manage/endpoints#restart-a-compute). To configure a custom scale to zero setting, modify `suspend_timeout_seconds` using the [Update compute endpoint API](https://api-docs.neon.tech/reference/updateprojectendpoint) API, as shown below. To use this API, you need to specify your project ID and compute endpoint ID. You can find your project ID in your project's settings. You can find the compute endpoint ID on your branch page. ```bash curl --request PATCH \ --url https://console.neon.tech/api/v2/projects/{project_id}/endpoints/{endpoint_id} \ --header 'accept: application/json' \ --header 'authorization: Bearer $NEON_API_KEY' \ --header 'content-type: application/json' \ --data ' { "endpoint": { "suspend_timeout_seconds": 300 } } ' ``` Consider combining this strategy with Neon's _Autoscaling_ feature, which allows you to run a compute with minimal resources and scale up on demand. For example, with autoscaling, you can configure a minimum compute size to reduce costs during off-peak times. In the image shown below, the scale to zero setting is set to 1 hour so that your compute only suspends after an hour of inactivity, and autoscaling is configured with a minimum compute size that keep costs low during periods of light usage. ![Connection warmup scale to zero and autoscaling configuration](/docs/connect/cold_start_compute_config.png) For autoscaling configuration instructions, see [Compute size and autoscaling configuration](/docs/manage/computes#compute-size-and-autoscaling-configuration). ### Place your application and database in the same region A key strategy for reducing connection latency is ensuring that your application and database are hosted in the same region, or as close as possible, geographically. For the regions supported by Neon, see [Regions](/docs/introduction/regions). For information about moving your database to a different region, see [Import data from another Neon project](/docs/import/migrate-from-neon). ### Increase your connection timeout By configuring longer connection timeout durations, your application has more time to accommodate cold starts and other factors that contribute to latency. Connection timeout settings are typically configured in your application or the database client library you're using, and the specific way to do it depends on the language or framework you're using. Here are examples of how to increase connection timeout settings in a few common programming languages and frameworks: ```javascript const { Pool } = require('pg'); const pool = new Pool({ connectionString: process.env.DATABASE_URL, connectionTimeoutMillis: 10000, // connection timeout in milliseconds idleTimeoutMillis: 10000, // idle timeout in milliseconds }); ``` ```python import psycopg2 from psycopg2 import connect from psycopg2.extensions import ISOLATION_LEVEL_AUTOCOMMIT import os DATABASE_URL = os.environ['DATABASE_URL'] conn = psycopg2.connect(DATABASE_URL, connect_timeout=10) ``` ```java import java.sql.Connection; import java.sql.DriverManager; import java.util.Properties; String dbUrl = System.getenv("DATABASE_URL"); Properties properties = new Properties(); properties.setProperty("connectTimeout", "10"); Connection conn = DriverManager.getConnection(dbUrl, properties); ``` ```prisma DATABASE_URL=postgresql://[user]:[password]@[neon_hostname]/[dbname]?connect_timeout=15&pool_timeout=15` ``` If you are using Prisma Client, your timeout issue could be related to Prisma's connection pool configuration. The Prisma Client query engine instantiates its own connection pool when it opens a first connection to the database. If you encounter a `Timed out fetching a new connection from the connection pool` error, refer to [Prisma connection pool timeouts](/docs/guides/prisma#connection-pool-timeouts) for information about configuring your Prisma connection pool size and pool timeout settings. Remember that increasing connection timeout settings might impact the responsiveness of your application, and users could end up waiting longer for their requests to be processed. Always test and monitor your application's performance when making changes like these. ### Build connection timeout handling into your application You can prepare your application to handle connection timeouts when latency is unavoidable. This might involve using retries with exponential backoff. This Javascript example connects to the database using the `pg` library and uses the `node-retry` library to handle connection retries with an exponential backoff. The general logic can be easily translated into other languages. ```javascript require('dotenv').config(); var Client = require('pg').Client; var retry = require('retry'); // Connection string from .env file var connectionString = process.env.DATABASE_URL; function connectWithRetry() { var operation = retry.operation({ retries: 5, // number of retries before giving up minTimeout: 4000, // minimum time between retries in milliseconds randomize: true, // adds randomness to timeouts to prevent retries from overwhelming the server }); operation.attempt(function (currentAttempt) { var client = new Client({ connectionString }); client .connect() .then(function () { console.log('Connected to the database'); // Perform your operations with the client // For example, let's run a simple SELECT query return client.query('SELECT NOW()'); }) .then(function (res) { console.log(res.rows[0]); return client.end(); }) .catch(function (err) { if (operation.retry(err)) { console.warn(`Failed to connect on attempt ${currentAttempt}, retrying...`); } else { console.error('Failed to connect to the database after multiple attempts:', err); } }); }); } // Usage connectWithRetry(); ``` In the example above, the `operation.attempt` function initiates the connection logic. If the connection fails (i.e., `client.connect()` returns a rejected Promise), the error is passed to `operation.retry`(err). If there are retries left, the retry function schedules another attempt with a delay based on the parameters defined in the `retry.operation`. The delay between retries is controlled by the `minTimeout` and `randomize` options. The randomize option adds a degree of randomness to the delay to prevent a large number of retries from potentially overwhelming the server. The `minTimeout` option defines the minimum time between retries in milliseconds. However, this example is a simplification. In a production application, you might want to use a more sophisticated strategy. For example, you could initially attempt to reconnect quickly in the event of a transient network issue, then fall back to slower retries if the problem persists. #### Connection retry references - [SQL Alchemy: Dealing with disconnects](https://arc.net/l/quote/nojcaewr) - [Fast API blog post: Recycling connections for Neon's scale to zero](/blog/deploy-a-serverless-fastapi-app-with-neon-postgres-and-aws-app-runner-at-any-scale) ### Use application-level caching Implement a caching system like [Redis](https://redis.io/) to store frequently accessed data, which can be rapidly served to users. This approach can help reduce occurrences of latency, but only if the data requested is available in the cache. Challenges with this strategy include cache invalidation due to frequently changing data, and cache misses when queries request uncached data. This strategy will not avoid latency entirely, but you may be able to combine it with other strategies to improve application responsiveness overall. ### Optimizing connection latency with sslnegotiation Starting with PostgreSQL 17, you can use the `sslnegotiation` connection parameter to control how SSL negotiation is handled when establishing a connection. The `sslnegotiation=direct` option reduces connection latency by skipping unnecessary negotiation steps. Neon has implemented support for `sslnegotiation=direct` in our proxy layer, allowing you to benefit from faster connection times even if your database runs on an older PostgreSQL version. You just need a PostgreSQL 17 client to use this feature. Here's a comparison of connection times with and without the `sslnegotiation=direct` parameter: **Without sslnegotiation=direct:** ```bash $ time psql "postgresql://neondb_owner@your-neon-endpoint/neondb?sslmode=require&channel_binding=require" -c "SELECT version();" version --------------------------------------------------------------------------------------------------------- PostgreSQL 16.4 on x86_64-pc-linux-gnu, compiled by gcc (Debian 10.2.1-6) 10.2.1 20210110, 64-bit (1 row) real 0m0.872s user 0m0.019s sys 0m0.000s ``` **With sslnegotiation=direct:** ```bash $ time psql "postgresql://neondb_owner@your-neon-endpoint/neondb?sslmode=require&channel_binding=require&sslnegotiation=direct" -c "SELECT version();" version --------------------------------------------------------------------------------------------------------- PostgreSQL 17.0 on x86_64-pc-linux-gnu, compiled by gcc (Debian 10.2.1-6) 10.2.1 20210110, 64-bit (1 row) real 0m0.753s user 0m0.016s sys 0m0.005s ``` As shown in the example above, using `sslnegotiation=direct` reduces the connection time by skipping the initial SSL negotiation step. To use this optimization, simply append `sslnegotiation=direct` to your connection string: ```text shouldWrap postgresql://[user]:[password]@[neon_hostname]/[dbname]?sslmode=verify-full&sslnegotiation=direct ``` ## Latency benchmarking See [Benchmarking latency in Neon's serverless Postgres](/docs/guides/benchmarking-latency) to learn how to measure and optimize query latency in your Neon database. ## Conclusion With the right strategies, you can optimize your system to handle connection latencies and timeouts, ensuring your application delivers a consistently high level of performance. The best solution often involves a combination of strategies, so experiment and find the right configuration for your specific use case. ## Related resources - [Neon latency benchmarks dashboard](/demos/regional-latency) - Interactive dashboard showing real-world latency measurements across different regions and workloads ([source code](https://github.com/neondatabase-labs/latency-benchmarks)) - [Connection pooling guide](/docs/connect/connection-pooling) - Reduce latency with efficient connection management - [Regional deployment options](/docs/introduction/regions) - Choose the optimal region for lowest latency - [Ship faster with Postgres](https://neon.tech/faster) - Explore examples and case studies demonstrating rapid development workflows