Running in Production
Last updated
Last updated
When running in production, we highly recommend using Queue mode with the following settings:
At least 2 main servers with load balancing, each starting from 2 CPU 4GB RAM
At least 2 workers, each starting from 1 CPU 2GB RAM
You can configure auto scaling depending on the traffic and volume.
By default, Flowise will use SQLite as the database. However when running at scale, its recommended to use PostgresQL.
Currently Flowise only supports AWS S3 with plan to support more blob storage providers. This will allow files and logs to be stored on S3, instead of local file path. Refer
Flowise uses an encryption key to encrypt/decrypt credentials you use such as OpenAI API keys. AWS Secret Manager is recommended to be used in production for better security control and key rotation. Refer
Artillery can be used to load testing your deployed Flowise application. Example script can be found here.
Users can create multiple API keys within Flowise in order to authenticate with the APIs. By default, keys get stored as a JSON file to your local file path. However when you have multiple instances, each instance will create a new JSON file, causing confusion. You can change the behaviour to store into database instead. Refer
When deployed to cloud/on-prem, most likely the instances are behind a proxy/load balancer. The IP address of the request might be the IP of the load balancer/reverse proxy, making the rate limiter effectively a global one and blocking all requests once the limit is reached or undefined
. Setting the correct NUMBER_OF_PROXIES
can resolve the issue. Refer