debian-mirror-gitlab/app/uploaders/object_storage.rb

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

470 lines
13 KiB
Ruby
Raw Normal View History

2018-11-08 19:23:39 +05:30
# frozen_string_literal: true
2018-05-09 12:01:36 +05:30
require 'fog/aws'
require 'carrierwave/storage/fog'
#
# This concern should add object storage support
# to the GitlabUploader class
#
module ObjectStorage
RemoteStoreError = Class.new(StandardError)
UnknownStoreError = Class.new(StandardError)
ObjectStorageUnavailable = Class.new(StandardError)
2018-11-08 19:23:39 +05:30
class ExclusiveLeaseTaken < StandardError
def initialize(lease_key)
@lease_key = lease_key
end
def message
*lease_key_group, _ = *@lease_key.split(":")
"Exclusive lease for #{lease_key_group.join(':')} is already taken."
end
end
2019-12-04 20:38:33 +05:30
TMP_UPLOAD_PATH = 'tmp/uploads'
2018-05-09 12:01:36 +05:30
module Store
LOCAL = 1
REMOTE = 2
end
2020-11-24 15:15:51 +05:30
SUPPORTED_STORES = [Store::LOCAL, Store::REMOTE].freeze
2018-05-09 12:01:36 +05:30
module Extension
# this extension is the glue between the ObjectStorage::Concern and RecordsUploads::Concern
module RecordsUploads
extend ActiveSupport::Concern
2018-11-08 19:23:39 +05:30
prepended do |base|
2018-05-09 12:01:36 +05:30
raise "#{base} must include ObjectStorage::Concern to use extensions." unless base < Concern
2018-11-08 19:23:39 +05:30
base.include(::RecordsUploads::Concern)
2018-05-09 12:01:36 +05:30
end
def retrieve_from_store!(identifier)
2018-11-08 19:23:39 +05:30
paths = upload_paths(identifier)
2018-05-09 12:01:36 +05:30
unless current_upload_satisfies?(paths, model)
# the upload we already have isn't right, find the correct one
2018-11-08 19:23:39 +05:30
self.upload = model&.retrieve_upload(identifier, paths)
2018-05-09 12:01:36 +05:30
end
super
end
def build_upload
super.tap do |upload|
upload.store = object_store
end
end
def upload=(upload)
2018-11-08 19:23:39 +05:30
return if upload.nil?
2018-05-09 12:01:36 +05:30
self.object_store = upload.store
super
end
2018-11-08 19:23:39 +05:30
def exclusive_lease_key
# For FileUploaders, model may have many uploaders. In that case
# we want to use exclusive key per upload, not per model to allow
# parallel migration
key_object = upload || model
"object_storage_migrate:#{key_object.class}:#{key_object.id}"
end
2018-05-09 12:01:36 +05:30
private
def current_upload_satisfies?(paths, model)
return false unless upload
return false unless model
paths.include?(upload.path) &&
upload.model_id == model.id &&
upload.model_type == model.class.base_class.sti_name
end
end
end
module Concern
extend ActiveSupport::Concern
included do |base|
base.include(ObjectStorage)
after :migrate, :delete_migrated_file
end
class_methods do
def object_store_options
options.object_store
end
def object_store_enabled?
object_store_options.enabled
end
def direct_upload_enabled?
object_store_options&.direct_upload
end
def proxy_download_enabled?
object_store_options.proxy_download
end
def direct_download_enabled?
!proxy_download_enabled?
end
def object_store_credentials
object_store_options.connection.to_hash.deep_symbolize_keys
end
def remote_store_path
object_store_options.remote_directory
end
def serialization_column(model_class, mount_point)
model_class.uploader_options.dig(mount_point, :mount_on) || mount_point
end
2018-11-08 19:23:39 +05:30
def workhorse_authorize(has_length:, maximum_size: nil)
2020-11-24 15:15:51 +05:30
{}.tap do |hash|
if self.object_store_enabled? && self.direct_upload_enabled?
hash[:RemoteObject] = workhorse_remote_upload_options(has_length: has_length, maximum_size: maximum_size)
else
hash[:TempPath] = workhorse_local_upload_path
end
2023-03-17 16:20:25 +05:30
hash[:UploadHashFunctions] = %w[sha1 sha256 sha512] if ::Gitlab::FIPS.enabled?
2020-11-24 15:15:51 +05:30
hash[:MaximumSize] = maximum_size if maximum_size.present?
2019-12-21 20:55:43 +05:30
end
2018-05-09 12:01:36 +05:30
end
def workhorse_local_upload_path
File.join(self.root, TMP_UPLOAD_PATH)
end
2020-10-24 23:57:45 +05:30
def object_store_config
ObjectStorage::Config.new(object_store_options)
end
2018-11-08 19:23:39 +05:30
def workhorse_remote_upload_options(has_length:, maximum_size: nil)
2018-05-09 12:01:36 +05:30
return unless self.object_store_enabled?
return unless self.direct_upload_enabled?
id = [CarrierWave.generate_cache_id, SecureRandom.hex].join('-')
upload_path = File.join(TMP_UPLOAD_PATH, id)
2020-10-24 23:57:45 +05:30
direct_upload = ObjectStorage::DirectUpload.new(self.object_store_config, upload_path,
has_length: has_length, maximum_size: maximum_size)
2018-05-09 12:01:36 +05:30
2018-11-08 19:23:39 +05:30
direct_upload.to_hash.merge(ID: id)
2018-05-09 12:01:36 +05:30
end
end
2020-11-24 15:15:51 +05:30
class OpenFile
extend Forwardable
# Explicitly exclude :path, because rubyzip uses that to detect "real" files.
def_delegators :@file, *(Zip::File::IO_METHODS - [:path])
# Even though :size is not in IO_METHODS, we do need it.
def_delegators :@file, :size
def initialize(file)
@file = file
end
2022-08-27 11:52:29 +05:30
def file_path
@file.path
end
2020-11-24 15:15:51 +05:30
end
2018-05-09 12:01:36 +05:30
# allow to configure and overwrite the filename
def filename
@filename || super || file&.filename # rubocop:disable Gitlab/ModuleWithInstanceVariables
end
def filename=(filename)
@filename = filename # rubocop:disable Gitlab/ModuleWithInstanceVariables
end
def file_storage?
storage.is_a?(CarrierWave::Storage::File)
end
def file_cache_storage?
cache_storage.is_a?(CarrierWave::Storage::File)
end
def object_store
# We use Store::LOCAL as null value indicates the local storage
@object_store ||= model.try(store_serialization_column) || Store::LOCAL
end
# rubocop:disable Gitlab/ModuleWithInstanceVariables
def object_store=(value)
@object_store = value || Store::LOCAL
@storage = storage_for(object_store)
end
# rubocop:enable Gitlab/ModuleWithInstanceVariables
# Return true if the current file is part or the model (i.e. is mounted in the model)
#
def persist_object_store?
model.respond_to?(:"#{store_serialization_column}=")
end
# Save the current @object_store to the model <mounted_as>_store column
def persist_object_store!
return unless persist_object_store?
updated = model.update_column(store_serialization_column, object_store)
raise 'Failed to update object store' unless updated
end
def use_file(&blk)
with_exclusive_lease do
unsafe_use_file(&blk)
end
end
2022-08-27 11:52:29 +05:30
def use_open_file(unlink_early: true)
2020-11-24 15:15:51 +05:30
Tempfile.open(path) do |file|
2022-08-27 11:52:29 +05:30
file.unlink if unlink_early
2020-11-24 15:15:51 +05:30
file.binmode
if file_storage?
IO.copy_stream(path, file)
else
2021-09-04 01:27:46 +05:30
Faraday.get(url) do |req|
req.options.on_data = proc { |chunk, _| file.write(chunk) }
end
2020-11-24 15:15:51 +05:30
end
file.seek(0, IO::SEEK_SET)
yield OpenFile.new(file)
2022-08-27 11:52:29 +05:30
ensure
file.unlink unless unlink_early
2020-11-24 15:15:51 +05:30
end
end
2018-05-09 12:01:36 +05:30
#
# Move the file to another store
#
# new_store: Enum (Store::LOCAL, Store::REMOTE)
#
def migrate!(new_store)
with_exclusive_lease do
unsafe_migrate!(new_store)
end
end
def fog_directory
self.class.remote_store_path
end
def fog_credentials
self.class.object_store_credentials
end
2020-10-24 23:57:45 +05:30
def fog_attributes
@fog_attributes ||= self.class.object_store_config.fog_attributes
end
2019-07-07 11:18:12 +05:30
# Set ACL of uploaded objects to not-public (fog-aws)[1] or no ACL at all
2023-05-27 22:25:52 +05:30
# (fog-google). Value is ignored by fog-aliyun
2019-07-07 11:18:12 +05:30
# [1]: https://github.com/fog/fog-aws/blob/daa50bb3717a462baf4d04d0e0cbfc18baacb541/lib/fog/aws/models/storage/file.rb#L152-L159
2018-05-09 12:01:36 +05:30
def fog_public
2019-07-07 11:18:12 +05:30
nil
2018-05-09 12:01:36 +05:30
end
def delete_migrated_file(migrated_file)
2018-11-08 19:23:39 +05:30
migrated_file.delete
2018-05-09 12:01:36 +05:30
end
def exists?
file.present?
end
def store_dir(store = nil)
store_dirs[store || object_store]
end
def store_dirs
{
Store::LOCAL => File.join(base_dir, dynamic_segment),
Store::REMOTE => File.join(dynamic_segment)
}
end
2022-07-23 23:45:48 +05:30
def store_path(*args)
if self.object_store == Store::REMOTE
# We allow administrators to create "sub buckets" by setting a prefix.
# This makes it possible to deploy GitLab with only one object storage
# bucket. Because the prefix is configuration data we do not want to
# store it in the uploads table via RecordsUploads. That means that the
# prefix cannot be part of store_dir. This is why we chose to implement
# the prefix support here in store_path.
File.join([self.class.object_store_options.bucket_prefix, super].compact)
else
super
end
end
2018-11-08 19:23:39 +05:30
# Returns all the possible paths for an upload.
# the `upload.path` is a lookup parameter, and it may change
# depending on the `store` param.
def upload_paths(identifier)
store_dirs.map { |store, path| File.join(path, identifier) }
end
2018-05-09 12:01:36 +05:30
def cache!(new_file = sanitized_file)
# We intercept ::UploadedFile which might be stored on remote storage
# We use that for "accelerated" uploads, where we store result on remote storage
2020-03-28 13:19:24 +05:30
if new_file.is_a?(::UploadedFile) && new_file.remote_id.present?
2018-05-09 12:01:36 +05:30
return cache_remote_file!(new_file.remote_id, new_file.original_filename)
end
super
end
def store!(new_file = nil)
# when direct upload is enabled, always store on remote storage
if self.class.object_store_enabled? && self.class.direct_upload_enabled?
self.object_store = Store::REMOTE
end
super
end
2018-11-08 19:23:39 +05:30
def exclusive_lease_key
"object_storage_migrate:#{model.class}:#{model.id}"
end
2018-05-09 12:01:36 +05:30
private
def cache_remote_file!(remote_object_id, original_filename)
file_path = File.join(TMP_UPLOAD_PATH, remote_object_id)
file_path = Pathname.new(file_path).cleanpath.to_s
raise RemoteStoreError, 'Bad file path' unless file_path.start_with?(TMP_UPLOAD_PATH + '/')
# TODO:
# This should be changed to make use of `tmp/cache` mechanism
# instead of using custom upload directory,
# using tmp/cache makes this implementation way easier than it is today
CarrierWave::Storage::Fog::File.new(self, storage_for(Store::REMOTE), file_path).tap do |file|
raise RemoteStoreError, 'Missing file' unless file.exists?
# Remote stored file, we force to store on remote storage
self.object_store = Store::REMOTE
# TODO:
# We store file internally and force it to be considered as `cached`
# This makes CarrierWave to store file in permament location (copy/delete)
# once this object is saved, but not sooner
@cache_id = "force-to-use-cache" # rubocop:disable Gitlab/ModuleWithInstanceVariables
@file = file # rubocop:disable Gitlab/ModuleWithInstanceVariables
@filename = original_filename # rubocop:disable Gitlab/ModuleWithInstanceVariables
end
end
# this is a hack around CarrierWave. The #migrate method needs to be
# able to force the current file to the migrated file upon success.
def file=(file)
@file = file # rubocop:disable Gitlab/ModuleWithInstanceVariables
end
def serialization_column
self.class.serialization_column(model.class, mounted_as)
end
# Returns the column where the 'store' is saved
# defaults to 'store'
def store_serialization_column
[serialization_column, 'store'].compact.join('_').to_sym
end
def storage
@storage ||= storage_for(object_store)
end
def storage_for(store)
case store
when Store::REMOTE
2020-04-08 14:13:33 +05:30
raise "Object Storage is not enabled for #{self.class}" unless self.class.object_store_enabled?
2018-05-09 12:01:36 +05:30
CarrierWave::Storage::Fog.new(self)
when Store::LOCAL
CarrierWave::Storage::File.new(self)
else
raise UnknownStoreError
end
end
def with_exclusive_lease
2018-11-08 19:23:39 +05:30
lease_key = exclusive_lease_key
uuid = Gitlab::ExclusiveLease.new(lease_key, timeout: 1.hour.to_i).try_obtain
2021-06-08 01:23:25 +05:30
raise ExclusiveLeaseTaken, lease_key unless uuid
2018-05-09 12:01:36 +05:30
yield uuid
ensure
2018-11-08 19:23:39 +05:30
Gitlab::ExclusiveLease.cancel(lease_key, uuid)
2018-05-09 12:01:36 +05:30
end
#
# Move the file to another store
#
# new_store: Enum (Store::LOCAL, Store::REMOTE)
#
def unsafe_migrate!(new_store)
return unless object_store != new_store
return unless file
new_file = nil
file_to_delete = file
from_object_store = object_store
self.object_store = new_store # changes the storage and file
cache_stored_file! if file_storage?
with_callbacks(:migrate, file_to_delete) do
with_callbacks(:store, file_to_delete) do # for #store_versions!
new_file = storage.store!(file)
persist_object_store!
self.file = new_file
end
end
file
2021-06-08 01:23:25 +05:30
rescue StandardError => e
2018-05-09 12:01:36 +05:30
# in case of failure delete new file
new_file.delete unless new_file.nil?
# revert back to the old file
self.object_store = from_object_store
self.file = file_to_delete
raise e
end
end
def unsafe_use_file
if file_storage?
return yield path
end
begin
cache_stored_file!
yield cache_path
ensure
FileUtils.rm_f(cache_path)
cache_storage.delete_dir!(cache_path(nil))
end
end
end
2020-04-22 19:07:51 +05:30
2021-06-08 01:23:25 +05:30
ObjectStorage::Concern.include_mod_with('ObjectStorage::Concern')