7.8 KiB
stage | group | info | type |
---|---|---|---|
Enablement | Database | To determine the technical writer assigned to the Stage/Group associated with this page, see https://about.gitlab.com/handbook/engineering/ux/technical-writing/#assignments | reference |
PostgreSQL
This page contains information about PostgreSQL the GitLab Support team uses when troubleshooting. GitLab makes this information public, so that anyone can make use of the Support team's collected knowledge.
WARNING: Some procedures documented here may break your GitLab instance. Use at your own risk.
If you're on a paid tier and aren't sure how to use these commands, contact Support for assistance with any issues you're having.
Other GitLab PostgreSQL documentation
This section is for links to information elsewhere in the GitLab documentation.
Procedures
-
Omnibus database procedures including:
- SSL: enabling, disabling, and verifying.
- Enabling Write Ahead Log (WAL) archiving.
- Using an external (non-Omnibus) PostgreSQL installation; and backing it up.
- Listening on TCP/IP as well as or instead of sockets.
- Storing data in another location.
- Destructively reseeding the GitLab database.
- Guidance around updating packaged PostgreSQL, including how to stop it from happening automatically.
-
Consuming PostgreSQL from within CI runners.
-
Using Slony to update PostgreSQL.
- Uses replication to handle PostgreSQL upgrades if the schemas are the same.
- Reduces downtime to a short window for switching to the newer version.
-
Managing Omnibus PostgreSQL versions from the development docs.
-
- Including troubleshooting
gitlab-ctl repmgr-check-master
(orgitlab-ctl patroni check-leader
if you're using Patroni) and PgBouncer errors.
- Including troubleshooting
-
Developer database documentation, some of which is absolutely not for production use. Including:
- Understanding EXPLAIN plans.
Troubleshooting/Fixes
-
GitLab database requirements, including
- Support for MySQL was removed in GitLab 12.1; migrate to PostgreSQL.
- Required extension:
pg_trgm
- Required extension:
btree_gist
-
Errors like this in the
production/sidekiq
log; see: Set default_transaction_isolation into read committed:ActiveRecord::StatementInvalid PG::TRSerializationFailure: ERROR: could not serialize access due to concurrent update
-
PostgreSQL HA replication slot errors:
pg_basebackup: could not create temporary replication slot "pg_basebackup_12345": ERROR: all replication slots are in use HINT: Free one or increase max_replication_slots.
-
Geo replication errors including:
ERROR: replication slots can only be used if max_replication_slots > 0 FATAL: could not start WAL streaming: ERROR: replication slot "geo_secondary_my_domain_com" does not exist Command exceeded allowed execution time PANIC: could not write to file 'pg_xlog/xlogtemp.123': No space left on device
-
Checking Geo configuration, including:
- Reconfiguring hosts/ports.
- Checking and fixing user/password mappings.
Support topics
Database deadlocks
References:
- Issue #1 Deadlocks with GitLab 12.1, PostgreSQL 10.7.
- Customer ticket (internal) GitLab 12.1.6 and Google doc (internal).
- Issue #2 deadlocks can occur if an instance is flooded with pushes. Provided for context about how GitLab code can have this sort of unanticipated effect in unusual situations.
ERROR: deadlock detected
Three applicable timeouts are identified in the issue #1; our recommended settings are as follows:
deadlock_timeout = 5s
statement_timeout = 15s
idle_in_transaction_session_timeout = 60s
Quoting from issue #1:
"If a deadlock is hit, and we resolve it through aborting the transaction after a short period, then the retry mechanisms we already have will make the deadlocked piece of work try again, and it's unlikely we'll deadlock multiple times in a row."
NOTE: In Support, our general approach to reconfiguring timeouts (applies also to the HTTP stack) is that it's acceptable to do it temporarily as a workaround. If it makes GitLab usable for the customer, then it buys time to understand the problem more completely, implement a hot fix, or make some other change that addresses the root cause. Generally, the timeouts should be put back to reasonable defaults after the root cause is resolved.
In this case, the guidance we had from development was to drop deadlock_timeout or statement_timeout, but to leave the third setting at 60s. Setting idle_in_transaction protects the database from sessions potentially hanging for days. There's more discussion in the issue relating to introducing this timeout on GitLab.com.
PostgresSQL defaults:
statement_timeout = 0
(never)idle_in_transaction_session_timeout = 0
(never)
Comments in issue #1 indicate that these should both be set to at least a number of minutes for all Omnibus GitLab installations (so they don't hang indefinitely). However, 15s for statement_timeout is very short, and will only be effective if the underlying infrastructure is very performant.
See current settings with:
sudo gitlab-rails runner "c = ApplicationRecord.connection ; puts c.execute('SHOW statement_timeout').to_a ;
puts c.execute('SHOW deadlock_timeout').to_a ;
puts c.execute('SHOW idle_in_transaction_session_timeout').to_a ;"
It may take a little while to respond.
{"statement_timeout"=>"1min"}
{"deadlock_timeout"=>"0"}
{"idle_in_transaction_session_timeout"=>"1min"}
These settings can be updated in /etc/gitlab/gitlab.rb
with:
postgresql['deadlock_timeout'] = '5s'
postgresql['statement_timeout'] = '15s'
postgresql['idle_in_transaction_session_timeout'] = '60s'
Once saved, reconfigure GitLab for the changes to take effect.
NOTE: These are Omnibus GitLab settings. If an external database, such as a customer's PostgreSQL installation or Amazon RDS is being used, these values don't get set, and would have to be set externally.