debian-mirror-gitlab/lib/gitlab/database/migration_helpers.rb

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

1368 lines
54 KiB
Ruby
Raw Normal View History

2019-02-15 15:39:39 +05:30
# frozen_string_literal: true
2016-06-02 11:05:42 +05:30
module Gitlab
module Database
module MigrationHelpers
2022-07-16 23:28:13 +05:30
include Migrations::ReestablishedConnectionStack
2020-07-28 23:09:34 +05:30
include Migrations::BackgroundMigrationHelpers
2022-01-26 12:08:38 +05:30
include Migrations::BatchedBackgroundMigrationHelpers
2023-01-13 00:05:48 +05:30
include Migrations::LockRetriesHelpers
include Migrations::TimeoutHelpers
include Migrations::ConstraintsHelpers
include Migrations::ExtensionHelpers
2021-04-29 21:17:54 +05:30
include DynamicModelHelpers
2021-09-04 01:27:46 +05:30
include RenameTableHelpers
2021-10-27 15:23:28 +05:30
include AsyncIndexes::MigrationHelpers
2020-07-28 23:09:34 +05:30
2022-04-04 11:22:00 +05:30
def define_batchable_model(table_name, connection: self.connection)
super(table_name, connection: connection)
end
def each_batch(table_name, connection: self.connection, **kwargs)
super(table_name, connection: connection, **kwargs)
end
def each_batch_range(table_name, connection: self.connection, **kwargs)
super(table_name, connection: connection, **kwargs)
end
2019-10-12 21:52:04 +05:30
DEFAULT_TIMESTAMP_COLUMNS = %i[created_at updated_at].freeze
2017-09-10 17:25:29 +05:30
# Adds `created_at` and `updated_at` columns with timezone information.
#
# This method is an improved version of Rails' built-in method `add_timestamps`.
#
2019-10-12 21:52:04 +05:30
# By default, adds `created_at` and `updated_at` columns, but these can be specified as:
#
# add_timestamps_with_timezone(:my_table, columns: [:created_at, :deleted_at])
#
# This allows you to create just the timestamps you need, saving space.
#
2017-09-10 17:25:29 +05:30
# Available options are:
2019-10-12 21:52:04 +05:30
# :default - The default value for the column.
# :null - When set to `true` the column will allow NULL values.
2017-09-10 17:25:29 +05:30
# The default is to not allow NULL values.
2021-12-11 22:18:48 +05:30
# :columns - the column names to create. Must end with `_at`.
2019-10-12 21:52:04 +05:30
# Default value: `DEFAULT_TIMESTAMP_COLUMNS`
#
# All options are optional.
2017-09-10 17:25:29 +05:30
def add_timestamps_with_timezone(table_name, options = {})
2019-10-12 21:52:04 +05:30
columns = options.fetch(:columns, DEFAULT_TIMESTAMP_COLUMNS)
columns.each do |column_name|
validate_timestamp_column_name!(column_name)
2017-09-10 17:25:29 +05:30
2021-12-11 22:18:48 +05:30
add_column(
table_name,
column_name,
:datetime_with_timezone,
default: options[:default],
null: options[:null] || false
)
2017-09-10 17:25:29 +05:30
end
end
2019-10-12 21:52:04 +05:30
# To be used in the `#down` method of migrations that
# use `#add_timestamps_with_timezone`.
2016-06-02 11:05:42 +05:30
#
2019-10-12 21:52:04 +05:30
# Available options are:
# :columns - the column names to remove. Must be one
# Default value: `DEFAULT_TIMESTAMP_COLUMNS`
#
# All options are optional.
def remove_timestamps(table_name, options = {})
columns = options.fetch(:columns, DEFAULT_TIMESTAMP_COLUMNS)
columns.each do |column_name|
remove_column(table_name, column_name)
end
end
2021-11-11 11:23:49 +05:30
# @deprecated Use `create_table` in V2 instead
2021-03-08 18:12:59 +05:30
#
# Creates a new table, optionally allowing the caller to add check constraints to the table.
# Aside from that addition, this method should behave identically to Rails' `create_table` method.
#
# Example:
#
# create_table_with_constraints :some_table do |t|
# t.integer :thing, null: false
# t.text :other_thing
#
# t.check_constraint :thing_is_not_null, 'thing IS NOT NULL'
# t.text_limit :other_thing, 255
# end
#
# See Rails' `create_table` for more info on the available arguments.
def create_table_with_constraints(table_name, **options, &block)
helper_context = self
with_lock_retries do
2021-04-17 20:07:23 +05:30
check_constraints = []
2021-03-08 18:12:59 +05:30
create_table(table_name, **options) do |t|
t.define_singleton_method(:check_constraint) do |name, definition|
helper_context.send(:validate_check_constraint_name!, name) # rubocop:disable GitlabSecurity/PublicSend
check_constraints << { name: name, definition: definition }
end
t.define_singleton_method(:text_limit) do |column_name, limit, name: nil|
# rubocop:disable GitlabSecurity/PublicSend
name = helper_context.send(:text_limit_name, table_name, column_name, name: name)
helper_context.send(:validate_check_constraint_name!, name)
# rubocop:enable GitlabSecurity/PublicSend
column_name = helper_context.quote_column_name(column_name)
definition = "char_length(#{column_name}) <= #{limit}"
check_constraints << { name: name, definition: definition }
end
t.instance_eval(&block) unless block.nil?
end
next if check_constraints.empty?
constraint_clauses = check_constraints.map do |constraint|
"ADD CONSTRAINT #{quote_table_name(constraint[:name])} CHECK (#{constraint[:definition]})"
end
execute(<<~SQL)
ALTER TABLE #{quote_table_name(table_name)}
#{constraint_clauses.join(",\n")}
SQL
end
end
2019-10-12 21:52:04 +05:30
# Creates a new index, concurrently
2016-06-02 11:05:42 +05:30
#
# Example:
#
# add_concurrent_index :users, :some_column
#
# See Rails' `add_index` for more info on the available arguments.
def add_concurrent_index(table_name, column_name, options = {})
2016-06-02 11:05:42 +05:30
if transaction_open?
raise 'add_concurrent_index can not be run inside a transaction, ' \
'you can disable transactions by calling disable_ddl_transaction! ' \
'in the body of your migration class'
end
2023-01-13 00:05:48 +05:30
if !options.delete(:allow_partition) && partition?(table_name)
raise ArgumentError, 'add_concurrent_index can not be used on a partitioned ' \
'table. Please use add_concurrent_partitioned_index on the partitioned table ' \
'as we need to create indexes on each partition and an index on the parent table'
end
2019-10-12 21:52:04 +05:30
options = options.merge({ algorithm: :concurrently })
2016-06-02 11:05:42 +05:30
2021-06-08 01:23:25 +05:30
if index_exists?(table_name, column_name, **options)
2021-12-11 22:18:48 +05:30
name = options[:name] || index_name(table_name, column_name)
_, schema = table_name.to_s.split('.').reverse
if index_invalid?(name, schema: schema)
say "Index being recreated because the existing version was INVALID: table_name: #{table_name}, column_name: #{column_name}"
remove_concurrent_index_by_name(table_name, name)
else
say "Index not created because it already exists (this may be due to an aborted migration or similar): table_name: #{table_name}, column_name: #{column_name}"
return
end
2018-05-09 12:01:36 +05:30
end
2018-11-20 20:47:30 +05:30
disable_statement_timeout do
2021-06-08 01:23:25 +05:30
add_index(table_name, column_name, **options)
2018-11-20 20:47:30 +05:30
end
2021-10-27 15:23:28 +05:30
# We created this index. Now let's remove the queuing entry for async creation in case it's still there.
unprepare_async_index(table_name, column_name, **options)
2016-06-02 11:05:42 +05:30
end
2021-12-11 22:18:48 +05:30
def index_invalid?(index_name, schema: nil)
index_name = connection.quote(index_name)
schema = connection.quote(schema) if schema
schema ||= 'current_schema()'
connection.select_value(<<~SQL)
select not i.indisvalid
from pg_class c
inner join pg_index i
on c.oid = i.indexrelid
inner join pg_namespace n
on n.oid = c.relnamespace
where n.nspname = #{schema}
and c.relname = #{index_name}
SQL
end
2019-10-12 21:52:04 +05:30
# Removes an existed index, concurrently
2017-08-17 22:00:37 +05:30
#
# Example:
#
# remove_concurrent_index :users, :some_column
#
# See Rails' `remove_index` for more info on the available arguments.
def remove_concurrent_index(table_name, column_name, options = {})
if transaction_open?
raise 'remove_concurrent_index can not be run inside a transaction, ' \
'you can disable transactions by calling disable_ddl_transaction! ' \
'in the body of your migration class'
end
2023-01-13 00:05:48 +05:30
if partition?(table_name)
raise ArgumentError, 'remove_concurrent_index can not be used on a partitioned ' \
'table. Please use remove_concurrent_partitioned_index_by_name on the partitioned table ' \
'as we need to remove the index on the parent table'
end
2019-12-26 22:10:19 +05:30
options = options.merge({ algorithm: :concurrently })
2017-08-17 22:00:37 +05:30
2021-06-08 01:23:25 +05:30
unless index_exists?(table_name, column_name, **options)
2020-11-24 15:15:51 +05:30
Gitlab::AppLogger.warn "Index not removed because it does not exist (this may be due to an aborted migration or similar): table_name: #{table_name}, column_name: #{column_name}"
2018-05-09 12:01:36 +05:30
return
end
2018-11-20 20:47:30 +05:30
disable_statement_timeout do
2021-06-08 01:23:25 +05:30
remove_index(table_name, **options.merge({ column: column_name }))
2018-11-20 20:47:30 +05:30
end
2021-10-27 15:23:28 +05:30
# We removed this index. Now let's make sure it's not queued for async creation.
unprepare_async_index(table_name, column_name, **options)
2017-08-17 22:00:37 +05:30
end
2019-10-12 21:52:04 +05:30
# Removes an existing index, concurrently
2017-09-10 17:25:29 +05:30
#
# Example:
#
# remove_concurrent_index :users, "index_X_by_Y"
#
# See Rails' `remove_index` for more info on the available arguments.
def remove_concurrent_index_by_name(table_name, index_name, options = {})
if transaction_open?
raise 'remove_concurrent_index_by_name can not be run inside a transaction, ' \
'you can disable transactions by calling disable_ddl_transaction! ' \
'in the body of your migration class'
end
2023-01-13 00:05:48 +05:30
if partition?(table_name)
raise ArgumentError, 'remove_concurrent_index_by_name can not be used on a partitioned ' \
'table. Please use remove_concurrent_partitioned_index_by_name on the partitioned table ' \
'as we need to remove the index on the parent table'
end
2020-07-28 23:09:34 +05:30
index_name = index_name[:name] if index_name.is_a?(Hash)
raise 'remove_concurrent_index_by_name must get an index name as the second argument' if index_name.blank?
2019-12-26 22:10:19 +05:30
options = options.merge({ algorithm: :concurrently })
2017-09-10 17:25:29 +05:30
2018-05-09 12:01:36 +05:30
unless index_exists_by_name?(table_name, index_name)
2020-11-24 15:15:51 +05:30
Gitlab::AppLogger.warn "Index not removed because it does not exist (this may be due to an aborted migration or similar): table_name: #{table_name}, index_name: #{index_name}"
2018-05-09 12:01:36 +05:30
return
end
2018-11-20 20:47:30 +05:30
disable_statement_timeout do
2021-06-08 01:23:25 +05:30
remove_index(table_name, **options.merge({ name: index_name }))
2018-11-20 20:47:30 +05:30
end
2021-10-27 15:23:28 +05:30
# We removed this index. Now let's make sure it's not queued for async creation.
unprepare_async_index_by_name(table_name, index_name, **options)
2017-09-10 17:25:29 +05:30
end
2017-08-17 22:00:37 +05:30
# Adds a foreign key with only minimal locking on the tables involved.
#
2019-10-12 21:52:04 +05:30
# This method only requires minimal locking
2017-08-17 22:00:37 +05:30
#
# source - The source table containing the foreign key.
# target - The target table the key points to.
# column - The name of the column to create the foreign key on.
2021-09-30 23:02:18 +05:30
# target_column - The name of the referenced column, defaults to "id".
2017-08-17 22:00:37 +05:30
# on_delete - The action to perform when associated data is removed,
# defaults to "CASCADE".
2020-01-01 13:55:28 +05:30
# name - The name of the foreign key.
2021-10-27 15:23:28 +05:30
# validate - Flag that controls whether the new foreign key will be validated after creation.
# If the flag is not set, the constraint will only be enforced for new data.
# reverse_lock_order - Flag that controls whether we should attempt to acquire locks in the reverse
# order of the ALTER TABLE. This can be useful in situations where the foreign
# key creation could deadlock with another process.
2019-09-30 21:07:59 +05:30
#
2021-10-27 15:23:28 +05:30
def add_concurrent_foreign_key(source, target, column:, on_delete: :cascade, target_column: :id, name: nil, validate: true, reverse_lock_order: false)
2017-08-17 22:00:37 +05:30
# Transactions would result in ALTER TABLE locks being held for the
# duration of the transaction, defeating the purpose of this method.
if transaction_open?
raise 'add_concurrent_foreign_key can not be run inside a transaction'
end
2020-01-01 13:55:28 +05:30
options = {
column: column,
on_delete: on_delete,
2021-09-30 23:02:18 +05:30
name: name.presence || concurrent_foreign_key_name(source, column),
primary_key: target_column
2020-01-01 13:55:28 +05:30
}
2017-08-17 22:00:37 +05:30
2021-01-03 14:25:43 +05:30
if foreign_key_exists?(source, target, **options)
2020-01-01 13:55:28 +05:30
warning_message = "Foreign key not created because it exists already " \
2018-05-09 12:01:36 +05:30
"(this may be due to an aborted migration or similar): " \
2020-01-01 13:55:28 +05:30
"source: #{source}, target: #{target}, column: #{options[:column]}, "\
"name: #{options[:name]}, on_delete: #{options[:on_delete]}"
2018-05-09 12:01:36 +05:30
2020-11-24 15:15:51 +05:30
Gitlab::AppLogger.warn warning_message
2020-01-01 13:55:28 +05:30
else
2018-05-09 12:01:36 +05:30
# Using NOT VALID allows us to create a key without immediately
# validating it. This means we keep the ALTER TABLE lock only for a
# short period of time. The key _is_ enforced for any newly created
# data.
2020-01-01 13:55:28 +05:30
2020-04-22 19:07:51 +05:30
with_lock_retries do
2021-10-27 15:23:28 +05:30
execute("LOCK TABLE #{target}, #{source} IN SHARE ROW EXCLUSIVE MODE") if reverse_lock_order
2020-04-22 19:07:51 +05:30
execute <<-EOF.strip_heredoc
ALTER TABLE #{source}
ADD CONSTRAINT #{options[:name]}
2022-11-25 23:54:43 +05:30
FOREIGN KEY (#{multiple_columns(options[:column])})
REFERENCES #{target} (#{multiple_columns(target_column)})
2020-04-22 19:07:51 +05:30
#{on_delete_statement(options[:on_delete])}
NOT VALID;
EOF
end
2018-05-09 12:01:36 +05:30
end
2017-08-17 22:00:37 +05:30
# Validate the existing constraint. This can potentially take a very
# long time to complete, but fortunately does not lock the source table
# while running.
2020-03-13 15:44:24 +05:30
# Disable this check by passing `validate: false` to the method call
# The check will be enforced for new data (inserts) coming in,
# but validating existing data is delayed.
2018-05-09 12:01:36 +05:30
#
# Note this is a no-op in case the constraint is VALID already
2020-03-13 15:44:24 +05:30
if validate
disable_statement_timeout do
execute("ALTER TABLE #{source} VALIDATE CONSTRAINT #{options[:name]};")
end
2018-11-20 20:47:30 +05:30
end
2017-08-17 22:00:37 +05:30
end
2020-03-13 15:44:24 +05:30
def validate_foreign_key(source, column, name: nil)
fk_name = name || concurrent_foreign_key_name(source, column)
unless foreign_key_exists?(source, name: fk_name)
2020-04-08 14:13:33 +05:30
raise missing_schema_object_message(source, "foreign key", fk_name)
2020-03-13 15:44:24 +05:30
end
disable_statement_timeout do
execute("ALTER TABLE #{source} VALIDATE CONSTRAINT #{fk_name};")
end
end
2020-01-01 13:55:28 +05:30
def foreign_key_exists?(source, target = nil, **options)
foreign_keys(source).any? do |foreign_key|
tables_match?(target.to_s, foreign_key.to_table.to_s) &&
options_match?(foreign_key.options, options)
2018-05-09 12:01:36 +05:30
end
end
2017-08-17 22:00:37 +05:30
# Returns the name for a concurrent foreign key.
#
# PostgreSQL constraint names have a limit of 63 bytes. The logic used
# here is based on Rails' foreign_key_name() method, which unfortunately
# is private so we can't rely on it directly.
2020-04-08 14:13:33 +05:30
#
# prefix:
# - The default prefix is `fk_` for backward compatibility with the existing
# concurrent foreign key helpers.
# - For standard rails foreign keys the prefix is `fk_rails_`
#
def concurrent_foreign_key_name(table, column, prefix: 'fk_')
2022-11-25 23:54:43 +05:30
identifier = "#{table}_#{multiple_columns(column, separator: '_')}_fk"
2019-09-30 21:07:59 +05:30
hashed_identifier = Digest::SHA256.hexdigest(identifier).first(10)
2020-04-08 14:13:33 +05:30
"#{prefix}#{hashed_identifier}"
2017-08-17 22:00:37 +05:30
end
def true_value
Database.true_value
end
def false_value
Database.false_value
2016-08-24 12:49:21 +05:30
end
2016-06-02 11:05:42 +05:30
# Updates the value of a column in batches.
#
# This method updates the table in batches of 5% of the total row count.
2019-07-07 11:18:12 +05:30
# A `batch_size` option can also be passed to set this to a fixed number.
2016-06-22 15:30:34 +05:30
# This method will continue updating rows until no rows remain.
#
# When given a block this method will yield two values to the block:
#
# 1. An instance of `Arel::Table` for the table that is being updated.
# 2. The query to run as an Arel object.
#
# By supplying a block one can add extra conditions to the queries being
# executed. Note that the same block is used for _all_ queries.
#
# Example:
#
# update_column_in_batches(:projects, :foo, 10) do |table, query|
# query.where(table[:some_column].eq('hello'))
# end
#
# This would result in this method updating only rows where
# `projects.some_column` equals "hello".
2016-06-02 11:05:42 +05:30
#
# table - The name of the table.
# column - The name of the column to update.
# value - The value for the column.
2016-06-22 15:30:34 +05:30
#
2018-03-17 18:26:18 +05:30
# The `value` argument is typically a literal. To perform a computed
# update, an Arel literal can be used instead:
#
# update_value = Arel.sql('bar * baz')
#
# update_column_in_batches(:projects, :foo, update_value) do |table, query|
# query.where(table[:some_column].eq('hello'))
# end
#
2016-06-22 15:30:34 +05:30
# Rubocop's Metrics/AbcSize metric is disabled for this method as Rubocop
# determines this method to be too complex while there's no way to make it
# less "complex" without introducing extra methods (which actually will
# make things _more_ complex).
#
2020-04-22 19:07:51 +05:30
# `batch_column_name` option is for tables without primary key, in this
2020-05-24 23:13:21 +05:30
# case another unique integer column can be used. Example: :user_id
2020-04-22 19:07:51 +05:30
#
2016-06-22 15:30:34 +05:30
# rubocop: disable Metrics/AbcSize
2022-08-27 11:52:29 +05:30
def update_column_in_batches(table_name, column, value, batch_size: nil, batch_column_name: :id, disable_lock_writes: false)
2017-09-10 17:25:29 +05:30
if transaction_open?
raise 'update_column_in_batches can not be run inside a transaction, ' \
'you can disable transactions by calling disable_ddl_transaction! ' \
'in the body of your migration class'
end
2022-08-27 11:52:29 +05:30
table = Arel::Table.new(table_name)
2016-06-22 15:30:34 +05:30
count_arel = table.project(Arel.star.count.as('count'))
count_arel = yield table, count_arel if block_given?
2020-03-13 15:44:24 +05:30
total = exec_query(count_arel.to_sql).to_a.first['count'].to_i
2016-06-22 15:30:34 +05:30
return if total == 0
2016-06-02 11:05:42 +05:30
2019-07-07 11:18:12 +05:30
if batch_size.nil?
# Update in batches of 5% until we run out of any rows to update.
batch_size = ((total / 100.0) * 5.0).ceil
max_size = 1000
2017-09-10 17:25:29 +05:30
2019-07-07 11:18:12 +05:30
# The upper limit is 1000 to ensure we don't lock too many rows. For
# example, for "merge_requests" even 1% of the table is around 35 000
# rows for GitLab.com.
batch_size = max_size if batch_size > max_size
end
2016-06-02 11:05:42 +05:30
2020-04-22 19:07:51 +05:30
start_arel = table.project(table[batch_column_name]).order(table[batch_column_name].asc).take(1)
2016-06-22 15:30:34 +05:30
start_arel = yield table, start_arel if block_given?
2020-04-22 19:07:51 +05:30
start_id = exec_query(start_arel.to_sql).to_a.first[batch_column_name.to_s].to_i
2016-06-22 15:30:34 +05:30
loop do
2020-04-22 19:07:51 +05:30
stop_arel = table.project(table[batch_column_name])
.where(table[batch_column_name].gteq(start_id))
.order(table[batch_column_name].asc)
2017-09-10 17:25:29 +05:30
.take(1)
.skip(batch_size)
2016-06-22 15:30:34 +05:30
stop_arel = yield table, stop_arel if block_given?
2020-03-13 15:44:24 +05:30
stop_row = exec_query(stop_arel.to_sql).to_a.first
2016-06-22 15:30:34 +05:30
2019-02-15 15:39:39 +05:30
update_arel = Arel::UpdateManager.new
2017-09-10 17:25:29 +05:30
.table(table)
.set([[table[column], value]])
2020-04-22 19:07:51 +05:30
.where(table[batch_column_name].gteq(start_id))
2016-06-02 11:05:42 +05:30
if stop_row
2020-04-22 19:07:51 +05:30
stop_id = stop_row[batch_column_name.to_s].to_i
2016-06-22 15:30:34 +05:30
start_id = stop_id
2020-04-22 19:07:51 +05:30
update_arel = update_arel.where(table[batch_column_name].lt(stop_id))
2016-06-02 11:05:42 +05:30
end
2016-06-22 15:30:34 +05:30
update_arel = yield table, update_arel if block_given?
2022-08-27 11:52:29 +05:30
transaction do
execute("SELECT set_config('lock_writes.#{table_name}', 'false', true)") if disable_lock_writes
execute(update_arel.to_sql)
end
2016-06-02 11:05:42 +05:30
2016-06-22 15:30:34 +05:30
# There are no more rows left to update.
break unless stop_row
2016-06-02 11:05:42 +05:30
end
end
# Adds a column with a default value without locking an entire table.
#
2020-05-24 23:13:21 +05:30
# @deprecated With PostgreSQL 11, adding columns with a default does not lead to a table rewrite anymore.
# As such, this method is not needed anymore and the default `add_column` helper should be used.
# This helper is subject to be removed in a >13.0 release.
def add_column_with_default(table, column, type, default:, limit: nil, allow_null: false)
raise 'Deprecated: add_column_with_default does not support being passed blocks anymore' if block_given?
2020-04-22 19:07:51 +05:30
2020-05-24 23:13:21 +05:30
add_column(table, column, type, default: default, limit: limit, null: allow_null)
2016-06-02 11:05:42 +05:30
end
2017-08-17 22:00:37 +05:30
# Renames a column without requiring downtime.
#
# Concurrent renames work by using database triggers to ensure both the
# old and new column are in sync. However, this method will _not_ remove
# the triggers or the old column automatically; this needs to be done
# manually in a post-deployment migration. This can be done using the
# method `cleanup_concurrent_column_rename`.
#
# table - The name of the database table containing the column.
# old - The old column name.
# new - The new column name.
# type - The type of the new column. If no type is given the old column's
# type is used.
2020-05-24 23:13:21 +05:30
# batch_column_name - option is for tables without primary key, in this
# case another unique integer column can be used. Example: :user_id
2020-07-28 23:09:34 +05:30
def rename_column_concurrently(table, old, new, type: nil, type_cast_function: nil, batch_column_name: :id)
2020-05-24 23:13:21 +05:30
unless column_exists?(table, batch_column_name)
raise "Column #{batch_column_name} does not exist on #{table}"
end
2017-08-17 22:00:37 +05:30
if transaction_open?
raise 'rename_column_concurrently can not be run inside a transaction'
end
2018-03-17 18:26:18 +05:30
check_trigger_permissions!(table)
2020-07-28 23:09:34 +05:30
create_column_from(table, old, new, type: type, batch_column_name: batch_column_name, type_cast_function: type_cast_function)
2017-08-17 22:00:37 +05:30
2018-03-17 18:26:18 +05:30
install_rename_triggers(table, old, new)
end
2019-12-04 20:38:33 +05:30
# Reverses operations performed by rename_column_concurrently.
#
# This method takes care of removing previously installed triggers as well
# as removing the new column.
#
# table - The name of the database table.
# old - The name of the old column.
# new - The name of the new column.
2019-10-12 21:52:04 +05:30
def undo_rename_column_concurrently(table, old, new)
trigger_name = rename_trigger_name(table, old, new)
check_trigger_permissions!(table)
2021-06-08 01:23:25 +05:30
remove_rename_triggers(table, trigger_name)
2019-10-12 21:52:04 +05:30
remove_column(table, new)
end
2018-03-17 18:26:18 +05:30
# Installs triggers in a table that keep a new column in sync with an old
# one.
#
# table - The name of the table to install the trigger in.
# old_column - The name of the old column.
# new_column - The name of the new column.
2021-06-08 01:23:25 +05:30
# trigger_name - The name of the trigger to use (optional).
def install_rename_triggers(table, old, new, trigger_name: nil)
2021-11-11 11:23:49 +05:30
Gitlab::Database::UnidirectionalCopyTrigger.on_table(table, connection: connection).create(old, new, trigger_name: trigger_name)
2021-06-08 01:23:25 +05:30
end
# Removes the triggers used for renaming a column concurrently.
def remove_rename_triggers(table, trigger)
2021-11-11 11:23:49 +05:30
Gitlab::Database::UnidirectionalCopyTrigger.on_table(table, connection: connection).drop(trigger)
2021-06-08 01:23:25 +05:30
end
# Returns the (base) name to use for triggers when renaming columns.
def rename_trigger_name(table, old, new)
2021-11-11 11:23:49 +05:30
Gitlab::Database::UnidirectionalCopyTrigger.on_table(table, connection: connection).name(old, new)
2017-08-17 22:00:37 +05:30
end
# Changes the type of a column concurrently.
#
# table - The table containing the column.
# column - The name of the column to change.
# new_type - The new column type.
2020-11-24 15:15:51 +05:30
def change_column_type_concurrently(table, column, new_type, type_cast_function: nil, batch_column_name: :id)
2017-08-17 22:00:37 +05:30
temp_column = "#{column}_for_type_change"
2020-11-24 15:15:51 +05:30
rename_column_concurrently(table, column, temp_column, type: new_type, type_cast_function: type_cast_function, batch_column_name: batch_column_name)
2017-08-17 22:00:37 +05:30
end
2021-01-03 14:25:43 +05:30
# Reverses operations performed by change_column_type_concurrently.
#
# table - The table containing the column.
# column - The name of the column to change.
def undo_change_column_type_concurrently(table, column)
temp_column = "#{column}_for_type_change"
undo_rename_column_concurrently(table, column, temp_column)
end
2017-08-17 22:00:37 +05:30
# Performs cleanup of a concurrent type change.
#
# table - The table containing the column.
# column - The name of the column to change.
# new_type - The new column type.
def cleanup_concurrent_column_type_change(table, column)
temp_column = "#{column}_for_type_change"
transaction do
# This has to be performed in a transaction as otherwise we might have
# inconsistent data.
cleanup_concurrent_column_rename(table, column, temp_column)
rename_column(table, temp_column, column)
end
end
2021-01-03 14:25:43 +05:30
# Reverses operations performed by cleanup_concurrent_column_type_change.
#
# table - The table containing the column.
# column - The name of the column to change.
# old_type - The type of the original column used with change_column_type_concurrently.
# type_cast_function - Required if the conversion back to the original type is not automatic
# batch_column_name - option for tables without a primary key, in this case
# another unique integer column can be used. Example: :user_id
2021-02-22 17:27:13 +05:30
def undo_cleanup_concurrent_column_type_change(table, column, old_type, type_cast_function: nil, batch_column_name: :id, limit: nil)
2022-06-21 17:19:12 +05:30
Gitlab::Database::QueryAnalyzers::RestrictAllowedSchemas.require_ddl_mode!
2021-01-03 14:25:43 +05:30
temp_column = "#{column}_for_type_change"
# Using a descriptive name that includes orinal column's name risks
# taking us above the 63 character limit, so we use a hash
identifier = "#{table}_#{column}_for_type_change"
hashed_identifier = Digest::SHA256.hexdigest(identifier).first(10)
temp_undo_cleanup_column = "tmp_undo_cleanup_column_#{hashed_identifier}"
unless column_exists?(table, batch_column_name)
raise "Column #{batch_column_name} does not exist on #{table}"
end
if transaction_open?
raise 'undo_cleanup_concurrent_column_type_change can not be run inside a transaction'
end
check_trigger_permissions!(table)
begin
create_column_from(
table,
column,
temp_undo_cleanup_column,
type: old_type,
batch_column_name: batch_column_name,
2021-02-22 17:27:13 +05:30
type_cast_function: type_cast_function,
limit: limit
2021-01-03 14:25:43 +05:30
)
transaction do
# This has to be performed in a transaction as otherwise we might
# have inconsistent data.
rename_column(table, column, temp_column)
rename_column(table, temp_undo_cleanup_column, column)
install_rename_triggers(table, column, temp_column)
end
2021-06-08 01:23:25 +05:30
rescue StandardError
2021-01-03 14:25:43 +05:30
# create_column_from can not run inside a transaction, which means
# that there is a risk that if any of the operations that follow it
# fail, we'll be left with an inconsistent schema
# For those reasons, we make sure that we drop temp_undo_cleanup_column
# if an error is caught
if column_exists?(table, temp_undo_cleanup_column)
remove_column(table, temp_undo_cleanup_column)
end
raise
end
end
2017-08-17 22:00:37 +05:30
# Cleans up a concurrent column name.
#
# This method takes care of removing previously installed triggers as well
# as removing the old column.
#
# table - The name of the database table.
# old - The name of the old column.
# new - The name of the new column.
def cleanup_concurrent_column_rename(table, old, new)
trigger_name = rename_trigger_name(table, old, new)
2018-03-17 18:26:18 +05:30
check_trigger_permissions!(table)
2021-06-08 01:23:25 +05:30
remove_rename_triggers(table, trigger_name)
2017-08-17 22:00:37 +05:30
remove_column(table, old)
end
2019-12-04 20:38:33 +05:30
# Reverses the operations performed by cleanup_concurrent_column_rename.
#
# This method adds back the old_column removed
# by cleanup_concurrent_column_rename.
# It also adds back the (old_column > new_column) trigger that is removed
# by cleanup_concurrent_column_rename.
#
# table - The name of the database table containing the column.
# old - The old column name.
# new - The new column name.
# type - The type of the old column. If no type is given the new column's
# type is used.
2020-05-24 23:13:21 +05:30
# batch_column_name - option is for tables without primary key, in this
# case another unique integer column can be used. Example: :user_id
def undo_cleanup_concurrent_column_rename(table, old, new, type: nil, batch_column_name: :id)
unless column_exists?(table, batch_column_name)
raise "Column #{batch_column_name} does not exist on #{table}"
end
2019-10-12 21:52:04 +05:30
if transaction_open?
raise 'undo_cleanup_concurrent_column_rename can not be run inside a transaction'
end
check_trigger_permissions!(table)
2020-05-24 23:13:21 +05:30
create_column_from(table, new, old, type: type, batch_column_name: batch_column_name)
2019-10-12 21:52:04 +05:30
install_rename_triggers(table, old, new)
end
2023-01-13 00:05:48 +05:30
def convert_to_type_column(column, from_type, to_type)
"#{column}_convert_#{from_type}_to_#{to_type}"
end
2021-06-08 01:23:25 +05:30
def convert_to_bigint_column(column)
"#{column}_convert_to_bigint"
end
# Initializes the conversion of a set of integer columns to bigint
2021-03-08 18:12:59 +05:30
#
# It can be used for converting both a Primary Key and any Foreign Keys
# that may reference it or any other integer column that we may want to
# upgrade (e.g. columns that store IDs, but are not set as FKs).
#
# - For primary keys and Foreign Keys (or other columns) defined as NOT NULL,
# the new bigint column is added with a hardcoded NOT NULL DEFAULT 0
# which allows us to skip a very costly verification step once we
# are ready to switch it.
# This is crucial for Primary Key conversions, because setting a column
# as the PK converts even check constraints to NOT NULL constraints
# and forces an inline re-verification of the whole table.
2021-04-29 21:17:54 +05:30
# - It sets up a trigger to keep the two columns in sync.
#
# Note: this helper is intended to be used in a regular (pre-deployment) migration.
#
# This helper is part 1 of a multi-step migration process:
2021-06-08 01:23:25 +05:30
# 1. initialize_conversion_of_integer_to_bigint to create the new columns and database trigger
2021-04-29 21:17:54 +05:30
# 2. backfill_conversion_of_integer_to_bigint to copy historic data using background migrations
# 3. remaining steps TBD, see #288005
#
# table - The name of the database table containing the column
2021-06-08 01:23:25 +05:30
# columns - The name, or array of names, of the column(s) that we want to convert to bigint.
2021-04-29 21:17:54 +05:30
# primary_key - The name of the primary key column (most often :id)
2021-06-08 01:23:25 +05:30
def initialize_conversion_of_integer_to_bigint(table, columns, primary_key: :id)
2023-01-13 00:05:48 +05:30
mappings = Array(columns).map do |c|
{
c => {
from_type: :int,
to_type: :bigint,
default_value: 0
}
}
end.reduce(&:merge)
create_temporary_columns_and_triggers(
table,
mappings,
primary_key: primary_key,
old_bigint_column_naming: true
)
2021-04-29 21:17:54 +05:30
end
2021-06-08 01:23:25 +05:30
# Reverts `initialize_conversion_of_integer_to_bigint`
#
# table - The name of the database table containing the columns
# columns - The name, or array of names, of the column(s) that we're converting to bigint.
def revert_initialize_conversion_of_integer_to_bigint(table, columns)
columns = Array.wrap(columns)
temporary_columns = columns.map { |column| convert_to_bigint_column(column) }
trigger_name = rename_trigger_name(table, columns, temporary_columns)
remove_rename_triggers(table, trigger_name)
temporary_columns.each { |column| remove_column(table, column) }
end
2021-11-11 11:23:49 +05:30
alias_method :cleanup_conversion_of_integer_to_bigint, :revert_initialize_conversion_of_integer_to_bigint
# Reverts `cleanup_conversion_of_integer_to_bigint`
#
# table - The name of the database table containing the columns
# columns - The name, or array of names, of the column(s) that we have converted to bigint.
# primary_key - The name of the primary key column (most often :id)
def restore_conversion_of_integer_to_bigint(table, columns, primary_key: :id)
2023-01-13 00:05:48 +05:30
mappings = Array(columns).map do |c|
{
c => {
from_type: :bigint,
to_type: :int,
default_value: 0
}
}
end.reduce(&:merge)
create_temporary_columns_and_triggers(
table,
mappings,
primary_key: primary_key,
old_bigint_column_naming: true
)
2021-11-11 11:23:49 +05:30
end
2021-06-08 01:23:25 +05:30
# Backfills the new columns used in an integer-to-bigint conversion using background migrations.
2021-04-29 21:17:54 +05:30
#
# - This helper should be called from a post-deployment migration.
2021-06-08 01:23:25 +05:30
# - In order for this helper to work properly, the new columns must be first initialized with
2021-04-29 21:17:54 +05:30
# the `initialize_conversion_of_integer_to_bigint` helper.
# - It tracks the scheduled background jobs through Gitlab::Database::BackgroundMigration::BatchedMigration,
2021-03-08 18:12:59 +05:30
# which allows a more thorough check that all jobs succeeded in the
# cleanup migration and is way faster for very large tables.
#
2021-04-29 21:17:54 +05:30
# Note: this helper is intended to be used in a post-deployment migration, to ensure any new code is
# deployed (including background job changes) before we begin processing the background migration.
#
# This helper is part 2 of a multi-step migration process:
2021-06-08 01:23:25 +05:30
# 1. initialize_conversion_of_integer_to_bigint to create the new columns and database trigger
2021-04-29 21:17:54 +05:30
# 2. backfill_conversion_of_integer_to_bigint to copy historic data using background migrations
# 3. remaining steps TBD, see #288005
2021-03-08 18:12:59 +05:30
#
# table - The name of the database table containing the column
2021-06-08 01:23:25 +05:30
# columns - The name, or an array of names, of the column(s) we want to convert to bigint.
2021-03-08 18:12:59 +05:30
# primary_key - The name of the primary key column (most often :id)
# batch_size - The number of rows to schedule in a single background migration
# sub_batch_size - The smaller batches that will be used by each scheduled job
# to update the table. Useful to keep each update at ~100ms while executing
# more updates per interval (2.minutes)
# Note that each execution of a sub-batch adds a constant 100ms sleep
# time in between the updates, which must be taken into account
# while calculating the batch, sub_batch and interval values.
# interval - The time interval between every background migration
#
# example:
# Assume that we have figured out that updating 200 records of the events
# table takes ~100ms on average.
# We can set the sub_batch_size to 200, leave the interval to the default
# and set the batch_size to 50_000 which will require
# ~50s = (50000 / 200) * (0.1 + 0.1) to complete and leaves breathing space
# between the scheduled jobs
2021-04-29 21:17:54 +05:30
def backfill_conversion_of_integer_to_bigint(
2021-03-08 18:12:59 +05:30
table,
2021-06-08 01:23:25 +05:30
columns,
2021-03-08 18:12:59 +05:30
primary_key: :id,
batch_size: 20_000,
sub_batch_size: 1000,
interval: 2.minutes
)
unless table_exists?(table)
raise "Table #{table} does not exist"
end
unless column_exists?(table, primary_key)
raise "Column #{primary_key} does not exist on #{table}"
end
2021-06-08 01:23:25 +05:30
conversions = Array.wrap(columns).to_h do |column|
raise ArgumentError, "Column #{column} does not exist on #{table}" unless column_exists?(table, column)
2021-03-08 18:12:59 +05:30
2021-06-08 01:23:25 +05:30
temporary_name = convert_to_bigint_column(column)
raise ArgumentError, "Column #{temporary_name} does not exist on #{table}" unless column_exists?(table, temporary_name)
2021-03-08 18:12:59 +05:30
2021-06-08 01:23:25 +05:30
[column, temporary_name]
2021-03-08 18:12:59 +05:30
end
2021-06-08 01:23:25 +05:30
queue_batched_background_migration(
2021-03-08 18:12:59 +05:30
'CopyColumnUsingBackgroundMigrationJob',
2021-04-29 21:17:54 +05:30
table,
primary_key,
2021-06-08 01:23:25 +05:30
conversions.keys,
conversions.values,
2021-04-29 21:17:54 +05:30
job_interval: interval,
2021-03-08 18:12:59 +05:30
batch_size: batch_size,
2021-04-29 21:17:54 +05:30
sub_batch_size: sub_batch_size)
2017-08-17 22:00:37 +05:30
end
2021-06-08 01:23:25 +05:30
# Reverts `backfill_conversion_of_integer_to_bigint`
#
# table - The name of the database table containing the column
# columns - The name, or an array of names, of the column(s) we want to convert to bigint.
# primary_key - The name of the primary key column (most often :id)
def revert_backfill_conversion_of_integer_to_bigint(table, columns, primary_key: :id)
columns = Array.wrap(columns)
2022-10-11 01:57:18 +05:30
conditions = ActiveRecord::Base.sanitize_sql(
[
'job_class_name = :job_class_name AND table_name = :table_name AND column_name = :column_name AND job_arguments = :job_arguments',
job_class_name: 'CopyColumnUsingBackgroundMigrationJob',
table_name: table,
column_name: primary_key,
job_arguments: [columns, columns.map { |column| convert_to_bigint_column(column) }].to_json
])
2021-06-08 01:23:25 +05:30
execute("DELETE FROM batched_background_migrations WHERE #{conditions}")
2017-08-17 22:00:37 +05:30
end
# Returns an Array containing the indexes for the given column
def indexes_for(table, column)
column = column.to_s
indexes(table).select { |index| index.columns.include?(column) }
end
# Returns an Array containing the foreign keys for the given column.
def foreign_keys_for(table, column)
column = column.to_s
foreign_keys(table).select { |fk| fk.column == column }
end
# Copies all indexes for the old column to a new column.
#
# table - The table containing the columns and indexes.
# old - The old column.
# new - The new column.
def copy_indexes(table, old, new)
old = old.to_s
new = new.to_s
indexes_for(table, old).each do |index|
new_columns = index.columns.map do |column|
column == old ? new : column
end
# This is necessary as we can't properly rename indexes such as
# "ci_taggings_idx".
unless index.name.include?(old)
raise "The index #{index.name} can not be copied as it does not "\
"mention the old column. You have to rename this index manually first."
end
name = index.name.gsub(old, new)
options = {
unique: index.unique,
name: name,
length: index.lengths,
order: index.orders
}
options[:using] = index.using if index.using
options[:where] = index.where if index.where
unless index.opclasses.blank?
opclasses = index.opclasses.dup
# Copy the operator classes for the old column (if any) to the new
# column.
opclasses[new] = opclasses.delete(old) if opclasses[old]
2021-01-03 14:25:43 +05:30
options[:opclass] = opclasses
2017-08-17 22:00:37 +05:30
end
add_concurrent_index(table, new_columns, options)
end
end
# Copies all foreign keys for the old column to the new column.
#
# table - The table containing the columns and indexes.
# old - The old column.
# new - The new column.
def copy_foreign_keys(table, old, new)
foreign_keys_for(table, old).each do |fk|
add_concurrent_foreign_key(fk.from_table,
fk.to_table,
column: new,
on_delete: fk.on_delete)
end
end
# Returns the column for the given table and column name.
def column_for(table, name)
name = name.to_s
2020-04-08 14:13:33 +05:30
column = columns(table).find { |column| column.name == name }
raise(missing_schema_object_message(table, "column", name)) if column.nil?
column
2017-08-17 22:00:37 +05:30
end
2018-12-13 13:39:08 +05:30
# This will replace the first occurrence of a string in a column with
2019-10-12 21:52:04 +05:30
# the replacement using `regexp_replace`
2017-08-17 22:00:37 +05:30
def replace_sql(column, pattern, replacement)
quoted_pattern = Arel::Nodes::Quoted.new(pattern.to_s)
quoted_replacement = Arel::Nodes::Quoted.new(replacement.to_s)
2019-10-12 21:52:04 +05:30
replace = Arel::Nodes::NamedFunction.new(
"regexp_replace", [column, quoted_pattern, quoted_replacement]
)
2017-08-17 22:00:37 +05:30
2019-10-12 21:52:04 +05:30
Arel::Nodes::SqlLiteral.new(replace.to_sql)
2017-08-17 22:00:37 +05:30
end
2017-09-10 17:25:29 +05:30
2022-04-04 11:22:00 +05:30
def remove_foreign_key_if_exists(source, target = nil, **kwargs)
reverse_lock_order = kwargs.delete(:reverse_lock_order)
return unless foreign_key_exists?(source, target, **kwargs)
if target && reverse_lock_order && transaction_open?
execute("LOCK TABLE #{target}, #{source} IN ACCESS EXCLUSIVE MODE")
end
if target
remove_foreign_key(source, target, **kwargs)
else
remove_foreign_key(source, **kwargs)
2019-07-07 11:18:12 +05:30
end
end
2021-06-08 01:23:25 +05:30
def remove_foreign_key_without_error(*args, **kwargs)
remove_foreign_key(*args, **kwargs)
2017-09-10 17:25:29 +05:30
rescue ArgumentError
end
2018-03-17 18:26:18 +05:30
2023-01-13 00:05:48 +05:30
# Remove any instances of deprecated job classes lingering in queues.
#
# rubocop:disable Cop/SidekiqApiUsage
def sidekiq_remove_jobs(job_klass:)
Sidekiq::Queue.new(job_klass.queue).each do |job|
job.delete if job.klass == job_klass.to_s
end
Sidekiq::RetrySet.new.each do |retri|
retri.delete if retri.klass == job_klass.to_s
end
Sidekiq::ScheduledSet.new.each do |scheduled|
scheduled.delete if scheduled.klass == job_klass.to_s
end
end
# rubocop:enable Cop/SidekiqApiUsage
2018-03-17 18:26:18 +05:30
def sidekiq_queue_migrate(queue_from, to:)
while sidekiq_queue_length(queue_from) > 0
Sidekiq.redis do |conn|
conn.rpoplpush "queue:#{queue_from}", "queue:#{to}"
end
end
end
def sidekiq_queue_length(queue_name)
Sidekiq.redis do |conn|
conn.llen("queue:#{queue_name}")
end
end
def check_trigger_permissions!(table)
unless Grant.create_and_execute_trigger?(table)
2021-12-11 22:18:48 +05:30
dbname = ApplicationRecord.database.database_name
user = ApplicationRecord.database.username
2018-03-17 18:26:18 +05:30
raise <<-EOF
Your database user is not allowed to create, drop, or execute triggers on the
table #{table}.
If you are using PostgreSQL you can solve this by logging in to the GitLab
database (#{dbname}) using a super user and running:
ALTER #{user} WITH SUPERUSER
2019-10-12 21:52:04 +05:30
This query will grant the user super user permissions, ensuring you don't run
2018-03-17 18:26:18 +05:30
into similar problems in the future (e.g. when new tables are created).
EOF
end
end
2018-05-09 12:01:36 +05:30
# Fetches indexes on a column by name for postgres.
#
# This will include indexes using an expression on the column, for example:
# `CREATE INDEX CONCURRENTLY index_name ON table (LOWER(column));`
#
# We can remove this when upgrading to Rails 5 with an updated `index_exists?`:
# - https://github.com/rails/rails/commit/edc2b7718725016e988089b5fb6d6fb9d6e16882
#
# Or this can be removed when we no longer support postgres < 9.5, so we
# can use `CREATE INDEX IF NOT EXISTS`.
def index_exists_by_name?(table, index)
# We can't fall back to the normal `index_exists?` method because that
# does not find indexes without passing a column name.
if indexes(table).map(&:name).include?(index.to_s)
true
else
2019-10-12 21:52:04 +05:30
postgres_exists_by_name?(table, index)
2018-03-27 19:54:05 +05:30
end
end
2018-05-09 12:01:36 +05:30
def postgres_exists_by_name?(table, name)
index_sql = <<~SQL
SELECT COUNT(*)
2021-01-03 14:25:43 +05:30
FROM pg_catalog.pg_indexes
WHERE schemaname = #{connection.quote(current_schema)}
AND tablename = #{connection.quote(table)}
AND indexname = #{connection.quote(name)}
2018-05-09 12:01:36 +05:30
SQL
connection.select_value(index_sql).to_i > 0
2018-03-27 19:54:05 +05:30
end
2018-12-05 23:21:45 +05:30
2020-01-01 13:55:28 +05:30
def create_or_update_plan_limit(limit_name, plan_name, limit_value)
2020-04-22 19:07:51 +05:30
limit_name_quoted = quote_column_name(limit_name)
plan_name_quoted = quote(plan_name)
limit_value_quoted = quote(limit_value)
2020-01-01 13:55:28 +05:30
execute <<~SQL
2020-04-22 19:07:51 +05:30
INSERT INTO plan_limits (plan_id, #{limit_name_quoted})
SELECT id, #{limit_value_quoted} FROM plans WHERE name = #{plan_name_quoted} LIMIT 1
ON CONFLICT (plan_id) DO UPDATE SET #{limit_name_quoted} = EXCLUDED.#{limit_name_quoted};
2020-01-01 13:55:28 +05:30
SQL
end
2020-03-13 15:44:24 +05:30
# Note this should only be used with very small tables
def backfill_iids(table)
sql = <<-END
UPDATE #{table}
SET iid = #{table}_with_calculated_iid.iid_num
FROM (
SELECT id, ROW_NUMBER() OVER (PARTITION BY project_id ORDER BY id ASC) AS iid_num FROM #{table}
) AS #{table}_with_calculated_iid
WHERE #{table}.id = #{table}_with_calculated_iid.id
END
execute(sql)
end
2022-11-25 23:54:43 +05:30
def add_primary_key_using_index(table_name, pk_name, index_to_use)
execute <<~SQL
ALTER TABLE #{quote_table_name(table_name)} ADD CONSTRAINT #{quote_table_name(pk_name)} PRIMARY KEY USING INDEX #{quote_table_name(index_to_use)}
SQL
end
def swap_primary_key(table_name, primary_key_name, index_to_use)
with_lock_retries(raise_on_exhaustion: true) do
drop_constraint(table_name, primary_key_name, cascade: true)
add_primary_key_using_index(table_name, primary_key_name, index_to_use)
end
end
alias_method :unswap_primary_key, :swap_primary_key
2022-07-23 23:45:48 +05:30
def drop_sequence(table_name, column_name, sequence_name)
execute <<~SQL
ALTER TABLE #{quote_table_name(table_name)} ALTER COLUMN #{quote_column_name(column_name)} DROP DEFAULT;
DROP SEQUENCE IF EXISTS #{quote_table_name(sequence_name)}
SQL
end
def add_sequence(table_name, column_name, sequence_name, start_value)
execute <<~SQL
CREATE SEQUENCE #{quote_table_name(sequence_name)} START #{start_value};
ALTER TABLE #{quote_table_name(table_name)} ALTER COLUMN #{quote_column_name(column_name)} SET DEFAULT nextval(#{quote(sequence_name)})
SQL
end
2023-01-13 00:05:48 +05:30
# rubocop:disable Metrics/CyclomaticComplexity,Metrics/PerceivedComplexity
def create_temporary_columns_and_triggers(table, mappings, primary_key: :id, old_bigint_column_naming: false)
raise ArgumentError, "No mappings for column conversion provided" if mappings.blank?
2019-10-12 21:52:04 +05:30
2023-01-13 00:05:48 +05:30
unless mappings.values.all? { |values| mapping_has_required_columns?(values) }
raise ArgumentError, "Some mappings don't have required keys provided"
end
2022-11-25 23:54:43 +05:30
2023-01-13 00:05:48 +05:30
neutral_values_for_type = {
int: 0,
bigint: 0,
uuid: '00000000-0000-0000-0000-000000000000'
}
2022-11-25 23:54:43 +05:30
2021-11-11 11:23:49 +05:30
unless table_exists?(table)
raise "Table #{table} does not exist"
end
unless column_exists?(table, primary_key)
raise "Column #{primary_key} does not exist on #{table}"
end
2023-01-13 00:05:48 +05:30
columns = mappings.keys
2021-11-11 11:23:49 +05:30
columns.each do |column|
next if column_exists?(table, column)
raise ArgumentError, "Column #{column} does not exist on #{table}"
end
check_trigger_permissions!(table)
2023-01-13 00:05:48 +05:30
if old_bigint_column_naming
mappings.each do |column, params|
params.merge!(
temporary_column_name: convert_to_bigint_column(column)
)
end
else
mappings.each do |column, params|
params.merge!(
temporary_column_name: convert_to_type_column(column, params[:from_type], params[:to_type])
)
end
end
2021-11-11 11:23:49 +05:30
with_lock_retries do
2023-01-13 00:05:48 +05:30
mappings.each do |(column_name, params)|
column = column_for(table, column_name)
temporary_name = params[:temporary_column_name]
data_type = params[:to_type]
default_value = params[:default_value]
2021-11-11 11:23:49 +05:30
if (column.name.to_s == primary_key.to_s) || !column.null
# If the column to be converted is either a PK or is defined as NOT NULL,
# set it to `NOT NULL DEFAULT 0` and we'll copy paste the correct values bellow
# That way, we skip the expensive validation step required to add
# a NOT NULL constraint at the end of the process
2023-01-13 00:05:48 +05:30
add_column(
table,
temporary_name,
data_type,
default: column.default || default_value || neutral_values_for_type.fetch(data_type),
null: false
)
2021-11-11 11:23:49 +05:30
else
2023-01-13 00:05:48 +05:30
add_column(
table,
temporary_name,
data_type,
default: column.default
)
2021-11-11 11:23:49 +05:30
end
end
2023-01-13 00:05:48 +05:30
old_column_names = mappings.keys
temporary_column_names = mappings.values.map { |v| v[:temporary_column_name] }
install_rename_triggers(table, old_column_names, temporary_column_names)
2021-11-11 11:23:49 +05:30
end
end
2023-01-13 00:05:48 +05:30
# rubocop:enable Metrics/CyclomaticComplexity,Metrics/PerceivedComplexity
2021-11-11 11:23:49 +05:30
2023-01-13 00:05:48 +05:30
def partition?(table_name)
if view_exists?(:postgres_partitions)
Gitlab::Database::PostgresPartition.partition_exists?(table_name)
else
Gitlab::Database::PostgresPartition.legacy_partition_exists?(table_name)
2020-06-23 00:09:42 +05:30
end
end
2023-01-13 00:05:48 +05:30
private
def multiple_columns(columns, separator: ', ')
Array.wrap(columns).join(separator)
end
def cascade_statement(cascade)
cascade ? 'CASCADE' : ''
end
2021-01-03 14:25:43 +05:30
2023-01-13 00:05:48 +05:30
def validate_check_constraint_name!(constraint_name)
if constraint_name.to_s.length > MAX_IDENTIFIER_NAME_LENGTH
raise "The maximum allowed constraint name is #{MAX_IDENTIFIER_NAME_LENGTH} characters"
end
2021-01-03 14:25:43 +05:30
end
2023-01-13 00:05:48 +05:30
# mappings => {} where keys are column names and values are hashes with the following keys:
# from_type - from which type we're migrating
# to_type - to which type we're migrating
# default_value - custom default value, if not provided will be taken from neutral_values_for_type
def mapping_has_required_columns?(mapping)
%i[from_type to_type].map do |required_key|
mapping.has_key?(required_key)
end.all?
2020-05-24 23:13:21 +05:30
end
def column_is_nullable?(table, column)
# Check if table.column has not been defined with NOT NULL
check_sql = <<~SQL
SELECT c.is_nullable
FROM information_schema.columns c
2021-01-03 14:25:43 +05:30
WHERE c.table_schema = #{connection.quote(current_schema)}
AND c.table_name = #{connection.quote(table)}
AND c.column_name = #{connection.quote(column)}
2020-05-24 23:13:21 +05:30
SQL
connection.select_value(check_sql) == 'YES'
end
2020-04-08 14:13:33 +05:30
def missing_schema_object_message(table, type, name)
<<~MESSAGE
Could not find #{type} "#{name}" on table "#{table}" which was referenced during the migration.
This issue could be caused by the database schema straying from the expected state.
To resolve this issue, please verify:
1. all previous migrations have completed
2. the database objects used in this migration match the Rails definition in schema.rb or structure.sql
MESSAGE
end
2020-01-01 13:55:28 +05:30
def tables_match?(target_table, foreign_key_table)
target_table.blank? || foreign_key_table == target_table
end
def options_match?(foreign_key_options, options)
options.all? { |k, v| foreign_key_options[k].to_s == v.to_s }
end
def on_delete_statement(on_delete)
return '' if on_delete.blank?
return 'ON DELETE SET NULL' if on_delete == :nullify
"ON DELETE #{on_delete.upcase}"
end
2021-02-22 17:27:13 +05:30
def create_column_from(table, old, new, type: nil, batch_column_name: :id, type_cast_function: nil, limit: nil)
2019-12-04 20:38:33 +05:30
old_col = column_for(table, old)
new_type = type || old_col.type
2021-02-22 17:27:13 +05:30
new_limit = limit || old_col.limit
2019-12-04 20:38:33 +05:30
add_column(table, new, new_type,
2021-02-22 17:27:13 +05:30
limit: new_limit,
2019-12-04 20:38:33 +05:30
precision: old_col.precision,
scale: old_col.scale)
# We set the default value _after_ adding the column so we don't end up
# updating any existing data with the default value. This isn't
# necessary since we copy over old values further down.
change_column_default(table, new, old_col.default) unless old_col.default.nil?
2020-07-28 23:09:34 +05:30
old_value = Arel::Table.new(table)[old]
if type_cast_function.present?
old_value = Arel::Nodes::NamedFunction.new(type_cast_function, [old_value])
end
2022-06-21 17:19:12 +05:30
Gitlab::Database::QueryAnalyzers::RestrictAllowedSchemas.with_suppressed do
2022-08-27 11:52:29 +05:30
Gitlab::Database::QueryAnalyzers::GitlabSchemasValidateConnection.with_suppressed do
update_column_in_batches(table, new, old_value, batch_column_name: batch_column_name, disable_lock_writes: true)
end
2022-06-21 17:19:12 +05:30
end
2019-12-04 20:38:33 +05:30
2020-05-24 23:13:21 +05:30
add_not_null_constraint(table, new) unless old_col.null
2019-12-04 20:38:33 +05:30
copy_indexes(table, old, new)
copy_foreign_keys(table, old, new)
2021-01-03 14:25:43 +05:30
copy_check_constraints(table, old, new)
2019-12-04 20:38:33 +05:30
end
2019-10-12 21:52:04 +05:30
def validate_timestamp_column_name!(column_name)
2021-12-11 22:18:48 +05:30
return if column_name.to_s.end_with?('_at')
2019-10-12 21:52:04 +05:30
raise <<~MESSAGE
Illegal timestamp column name! Got #{column_name}.
2021-12-11 22:18:48 +05:30
Must end with `_at`}
2019-10-12 21:52:04 +05:30
MESSAGE
end
2016-06-02 11:05:42 +05:30
end
end
end