PostgreSQL Requirements for Salesforce
To store a local copy of the Salesforce data and to perform certain functions, store your Salesforce data in a database. Each Salesforce organization needs a separate database. For example, if you back up two production organizations and one sandbox organization, configure three databases.
The following functions require a database:
- Object comparisons
- Record-level restores
- Restore to the Salesforce cloud
- Sandbox seeding
- PostgreSQL Database
- Amazon RDS for PostgreSQL
The Commvault Salesforce application connects to a database with default options such as connection timeout, host, instance, and so on. If your setup requires additional connection properties, configure the sf_dbConnectionProperties additional setting.
- The PostgreSQL version must be V9.3 or higher.
- The database host must have at least 8 CPUs/processors and 16 GB of memory. We recommend 32 GB of memory.
- The database host must have free space that is at a minimum equal to four times the size of the Salesforce used data storage. This amount of free space accounts for the following:
- Salesforce data change frequency because the database maintains all versions
- Data used space
- Additional table size, such as share and history tables
- Index size
- Log size
Additional space is needed for out-of-place restore operations. When you perform an out-of-place restore, the source data is copied, the copy is used for the restore, and then the copy is removed.
For instructions about checking the storage usage values, see Checking the Values for Salesforce Storage Usage.
- You must have a user who has super user permissions.
- You must be able to access the database from the internal network.
- For operations that restore data from the PostgreSQL database, such as record-level restores, the database must be accessible from the Web Server.
Best Practice: To avoid filling the disk with transaction logs during the Salesforce backup, disable Write-Ahead Logging (WAL).
Last modified: 3/17/2021 1:47:51 PM