debian-mirror-gitlab/spec/uploaders/object_storage_spec.rb

1213 lines
39 KiB
Ruby
Raw Normal View History

2019-12-21 20:55:43 +05:30
# frozen_string_literal: true
2019-12-04 20:38:33 +05:30
require 'spec_helper'
2018-05-09 12:01:36 +05:30
require 'carrierwave/storage/fog'
class Implementation < GitlabUploader
include ObjectStorage::Concern
include ::RecordsUploads::Concern
prepend ::ObjectStorage::Extension::RecordsUploads
2023-05-27 22:25:52 +05:30
storage_location :uploads
2018-05-09 12:01:36 +05:30
private
# user/:id
def dynamic_segment
2019-09-04 21:01:54 +05:30
File.join(model.class.underscore, model.id.to_s)
2018-05-09 12:01:36 +05:30
end
end
2023-07-09 08:55:56 +05:30
# TODO: Update feature_category once object storage group ownership has been determined.
RSpec.describe ObjectStorage, :clean_gitlab_redis_shared_state, feature_category: :shared do
2018-05-09 12:01:36 +05:30
let(:uploader_class) { Implementation }
let(:object) { build_stubbed(:user) }
2023-07-09 08:55:56 +05:30
let(:file_column) { :file }
let(:uploader) { uploader_class.new(object, file_column) }
2018-05-09 12:01:36 +05:30
describe '#object_store=' do
before do
allow(uploader_class).to receive(:object_store_enabled?).and_return(true)
end
it "reload the local storage" do
uploader.object_store = described_class::Store::LOCAL
expect(uploader.file_storage?).to be_truthy
end
it "reload the REMOTE storage" do
uploader.object_store = described_class::Store::REMOTE
expect(uploader.file_storage?).to be_falsey
end
context 'object_store is Store::LOCAL' do
before do
uploader.object_store = described_class::Store::LOCAL
end
describe '#store_dir' do
it 'is the composition of (base_dir, dynamic_segment)' do
expect(uploader.store_dir).to start_with("uploads/-/system/user/")
end
end
2022-07-23 23:45:48 +05:30
describe '#store_path' do
subject { uploader.store_path('filename') }
it 'uses store_dir' do
expect(subject).to eq("uploads/-/system/user/#{object.id}/filename")
end
context 'when a bucket prefix is configured' do
before do
allow(uploader_class).to receive(:object_store_options) do
double(
bucket_prefix: 'my/prefix'
)
end
end
it 'uses store_dir and ignores prefix' do
expect(subject).to eq("uploads/-/system/user/#{object.id}/filename")
end
end
end
2018-05-09 12:01:36 +05:30
end
context 'object_store is Store::REMOTE' do
before do
uploader.object_store = described_class::Store::REMOTE
end
describe '#store_dir' do
it 'is the composition of (dynamic_segment)' do
expect(uploader.store_dir).to start_with("user/")
end
end
2022-07-23 23:45:48 +05:30
describe '#store_path' do
subject { uploader.store_path('filename') }
it 'uses store_dir' do
expect(subject).to eq("user/#{object.id}/filename")
end
context 'when a bucket prefix is configured' do
before do
allow(uploader_class).to receive(:object_store_options) do
double(
bucket_prefix: 'my/prefix'
)
end
end
it 'uses the prefix and store_dir' do
expect(subject).to eq("my/prefix/user/#{object.id}/filename")
end
end
2023-07-09 08:55:56 +05:30
context 'when model has final path defined for the file column' do
# Changing this to `foo` to make a point that not all uploaders are mounted
# as `file`. They can be mounted as different names, for example, `avatar`.
let(:file_column) { :foo }
before do
allow(object).to receive(:foo_final_path).and_return('123-final-path')
end
it 'uses the final path instead' do
expect(subject).to eq('123-final-path')
end
context 'and a bucket prefix is configured' do
before do
allow(uploader_class).to receive(:object_store_options) do
double(
bucket_prefix: 'my/prefix'
)
end
end
it 'uses the prefix with the final path' do
expect(subject).to eq("my/prefix/123-final-path")
end
end
end
2022-07-23 23:45:48 +05:30
end
2018-05-09 12:01:36 +05:30
end
end
describe '#object_store' do
subject { uploader.object_store }
it "delegates to <mount>_store on model" do
expect(object).to receive(:file_store)
subject
end
context 'when store is null' do
before do
expect(object).to receive(:file_store).and_return(nil)
end
it "uses Store::LOCAL" do
is_expected.to eq(described_class::Store::LOCAL)
end
end
context 'when value is set' do
before do
expect(object).to receive(:file_store).and_return(described_class::Store::REMOTE)
end
it "returns the given value" do
is_expected.to eq(described_class::Store::REMOTE)
end
end
end
describe '#file_cache_storage?' do
context 'when file storage is used' do
before do
expect(uploader_class).to receive(:cache_storage) { CarrierWave::Storage::File }
end
it { expect(uploader).to be_file_cache_storage }
end
context 'when is remote storage' do
before do
expect(uploader_class).to receive(:cache_storage) { CarrierWave::Storage::Fog }
end
it { expect(uploader).not_to be_file_cache_storage }
end
end
# this means the model shall include
# include RecordsUpload::Concern
# prepend ObjectStorage::Extension::RecordsUploads
# the object_store persistence is delegated to the `Upload` model.
#
context 'when persist_object_store? is false' do
let(:object) { create(:project, :with_avatar) }
let(:uploader) { object.avatar }
it { expect(object).to be_a(Avatarable) }
it { expect(uploader.persist_object_store?).to be_falsey }
describe 'delegates the object_store logic to the `Upload` model' do
it 'sets @upload to the found `upload`' do
expect(uploader.upload).to eq(uploader.upload)
end
it 'sets @object_store to the `Upload` value' do
expect(uploader.object_store).to eq(uploader.upload.store)
end
end
describe '#migrate!' do
let(:new_store) { ObjectStorage::Store::REMOTE }
before do
stub_uploads_object_storage(uploader: AvatarUploader)
end
subject { uploader.migrate!(new_store) }
it 'persist @object_store to the recorded upload' do
subject
expect(uploader.upload.store).to eq(new_store)
end
describe 'fails' do
it 'is handled gracefully' do
store = uploader.object_store
2020-01-01 13:55:28 +05:30
expect_next_instance_of(Upload) do |instance|
expect(instance).to receive(:save!).and_raise("An error")
end
2018-05-09 12:01:36 +05:30
expect { subject }.to raise_error("An error")
expect(uploader.exists?).to be_truthy
expect(uploader.upload.store).to eq(store)
end
end
end
end
# this means the model holds an <mounted_as>_store attribute directly
# and do not delegate the object_store persistence to the `Upload` model.
#
context 'persist_object_store? is true' do
context 'when using JobArtifactsUploader' do
let(:store) { described_class::Store::LOCAL }
let(:object) { create(:ci_job_artifact, :archive, file_store: store) }
let(:uploader) { object.file }
context 'checking described_class' do
it "uploader include described_class::Concern" do
expect(uploader).to be_a(described_class::Concern)
end
end
describe '#use_file' do
context 'when file is stored locally' do
it "calls a regular path" do
expect { |b| uploader.use_file(&b) }.not_to yield_with_args(%r[tmp/cache])
end
end
context 'when file is stored remotely' do
let(:store) { described_class::Store::REMOTE }
before do
stub_artifacts_object_storage
end
it "calls a cache path" do
expect { |b| uploader.use_file(&b) }.to yield_with_args(%r[tmp/cache])
end
2018-11-08 19:23:39 +05:30
it "cleans up the cached file" do
cached_path = ''
uploader.use_file do |path|
cached_path = path
expect(File.exist?(cached_path)).to be_truthy
end
expect(File.exist?(cached_path)).to be_falsey
end
2018-05-09 12:01:36 +05:30
end
end
2020-11-24 15:15:51 +05:30
describe '#use_open_file' do
context 'when file is stored locally' do
2022-08-27 11:52:29 +05:30
it "returns the file unlinked" do
expect { |b| uploader.use_open_file(&b) }.to yield_with_args(
satisfying do |f|
expect(f).to be_an_instance_of(ObjectStorage::Concern::OpenFile)
expect(f.file_path).to be_nil
end
)
end
it "returns the file not unlinked" do
expect { |b| uploader.use_open_file(unlink_early: false, &b) }.to yield_with_args(
satisfying do |f|
expect(f).to be_an_instance_of(ObjectStorage::Concern::OpenFile)
expect(File.exist?(f.file_path)).to be_truthy
end
)
2020-11-24 15:15:51 +05:30
end
end
context 'when file is stored remotely' do
let(:store) { described_class::Store::REMOTE }
before do
stub_artifacts_object_storage
2021-09-04 01:27:46 +05:30
# We need to check the Host header not including the port because AWS does not accept
stub_request(:get, %r{s3.amazonaws.com/#{uploader.path}})
.with { |request| !request.headers['Host'].to_s.include?(':443') }
.to_return(status: 200, body: '')
2020-11-24 15:15:51 +05:30
end
it "returns the file" do
expect { |b| uploader.use_open_file(&b) }.to yield_with_args(an_instance_of(ObjectStorage::Concern::OpenFile))
end
end
end
2018-05-09 12:01:36 +05:30
describe '#migrate!' do
subject { uploader.migrate!(new_store) }
shared_examples "updates the underlying <mounted>_store" do
it do
subject
expect(object.file_store).to eq(new_store)
end
end
context 'when using the same storage' do
let(:new_store) { store }
it "to not migrate the storage" do
subject
expect(uploader).not_to receive(:store!)
expect(uploader.object_store).to eq(store)
end
end
context 'when migrating to local storage' do
let(:store) { described_class::Store::REMOTE }
let(:new_store) { described_class::Store::LOCAL }
before do
stub_artifacts_object_storage
end
include_examples "updates the underlying <mounted>_store"
it "local file does not exist" do
expect(File.exist?(uploader.path)).to eq(false)
end
it "remote file exist" do
expect(uploader.file.exists?).to be_truthy
end
it "does migrate the file" do
subject
expect(uploader.object_store).to eq(new_store)
expect(File.exist?(uploader.path)).to eq(true)
end
end
context 'when migrating to remote storage' do
let(:new_store) { described_class::Store::REMOTE }
let!(:current_path) { uploader.path }
it "file does exist" do
expect(File.exist?(current_path)).to eq(true)
end
context 'when storage is disabled' do
before do
stub_artifacts_object_storage(enabled: false)
end
it "to raise an error" do
2020-04-08 14:13:33 +05:30
expect { subject }.to raise_error(/Object Storage is not enabled for JobArtifactUploader/)
2018-05-09 12:01:36 +05:30
end
end
context 'when credentials are set' do
before do
stub_artifacts_object_storage
end
include_examples "updates the underlying <mounted>_store"
it "does migrate the file" do
subject
expect(uploader.object_store).to eq(new_store)
end
it "does delete original file" do
subject
expect(File.exist?(current_path)).to eq(false)
end
context 'when subject save fails' do
before do
expect(uploader).to receive(:persist_object_store!).and_raise(RuntimeError, "exception")
end
it "original file is not removed" do
expect { subject }.to raise_error(/exception/)
expect(File.exist?(current_path)).to eq(true)
end
end
end
end
end
end
end
describe '#fog_directory' do
let(:remote_directory) { 'directory' }
before do
allow(uploader_class).to receive(:options) do
double(object_store: double(remote_directory: remote_directory))
end
end
subject { uploader.fog_directory }
it { is_expected.to eq(remote_directory) }
end
context 'when file is in use' do
def when_file_is_in_use
uploader.use_file do
yield
end
end
it 'cannot migrate' do
when_file_is_in_use do
expect(uploader).not_to receive(:unsafe_migrate!)
2018-11-08 19:23:39 +05:30
expect { uploader.migrate!(described_class::Store::REMOTE) }.to raise_error(ObjectStorage::ExclusiveLeaseTaken)
2018-05-09 12:01:36 +05:30
end
end
it 'cannot use_file' do
when_file_is_in_use do
expect(uploader).not_to receive(:unsafe_use_file)
2018-11-08 19:23:39 +05:30
expect { uploader.use_file }.to raise_error(ObjectStorage::ExclusiveLeaseTaken)
end
end
it 'can still migrate other files of the same model' do
uploader2 = uploader_class.new(object, :file)
uploader2.upload = create(:upload)
uploader.upload = create(:upload)
when_file_is_in_use do
expect(uploader2).to receive(:unsafe_migrate!)
uploader2.migrate!(described_class::Store::REMOTE)
2018-05-09 12:01:36 +05:30
end
end
end
describe '#fog_credentials' do
2023-07-09 08:55:56 +05:30
let(:connection) { GitlabSettings::Options.build("provider" => "AWS") }
2018-05-09 12:01:36 +05:30
before do
allow(uploader_class).to receive(:options) do
double(object_store: double(connection: connection))
end
end
subject { uploader.fog_credentials }
it { is_expected.to eq(provider: 'AWS') }
end
describe '#fog_public' do
subject { uploader.fog_public }
2019-07-07 11:18:12 +05:30
it { is_expected.to eq(nil) }
2018-05-09 12:01:36 +05:30
end
2020-10-24 23:57:45 +05:30
describe '#fog_attributes' do
subject { uploader.fog_attributes }
it { is_expected.to eq({}) }
context 'with encryption configured' do
let(:raw_options) do
{
"enabled" => true,
"connection" => { "provider" => 'AWS' },
"storage_options" => { "server_side_encryption" => "AES256" }
}
end
2023-07-09 08:55:56 +05:30
let(:options) { GitlabSettings::Options.build(raw_options) }
2020-10-24 23:57:45 +05:30
before do
allow(uploader_class).to receive(:options) do
double(object_store: options)
end
end
it { is_expected.to eq({ "x-amz-server-side-encryption" => "AES256" }) }
end
end
2018-05-09 12:01:36 +05:30
describe '.workhorse_authorize' do
2018-11-08 19:23:39 +05:30
let(:has_length) { true }
let(:maximum_size) { nil }
2023-07-09 08:55:56 +05:30
let(:use_final_store_path) { false }
2018-05-09 12:01:36 +05:30
2023-07-09 08:55:56 +05:30
subject do
uploader_class.workhorse_authorize(
has_length: has_length,
maximum_size: maximum_size,
use_final_store_path: use_final_store_path
)
end
2018-05-09 12:01:36 +05:30
2023-03-17 16:20:25 +05:30
context 'when FIPS is enabled', :fips_mode do
it 'response enables FIPS' do
expect(subject[:UploadHashFunctions]).to match_array(%w[sha1 sha256 sha512])
end
end
context 'when FIPS is disabled' do
it 'response disables FIPS' do
expect(subject[:UploadHashFunctions]).to be nil
end
end
2020-11-24 15:15:51 +05:30
shared_examples 'returns the maximum size given' do
2018-05-09 12:01:36 +05:30
it "returns temporary path" do
2020-11-24 15:15:51 +05:30
expect(subject[:MaximumSize]).to eq(maximum_size)
end
end
shared_examples 'uses local storage' do
it_behaves_like 'returns the maximum size given' do
it "returns temporary path" do
is_expected.to have_key(:TempPath)
2018-05-09 12:01:36 +05:30
2020-11-24 15:15:51 +05:30
expect(subject[:TempPath]).to start_with(uploader_class.root)
expect(subject[:TempPath]).to include(described_class::TMP_UPLOAD_PATH)
end
2018-05-09 12:01:36 +05:30
end
end
shared_examples 'uses remote storage' do
2020-11-24 15:15:51 +05:30
it_behaves_like 'returns the maximum size given' do
2023-07-09 08:55:56 +05:30
it "returns remote object properties for a temporary upload" do
2020-11-24 15:15:51 +05:30
is_expected.to have_key(:RemoteObject)
2018-05-09 12:01:36 +05:30
2020-11-24 15:15:51 +05:30
expect(subject[:RemoteObject]).to have_key(:ID)
expect(subject[:RemoteObject]).to include(Timeout: a_kind_of(Integer))
expect(subject[:RemoteObject][:Timeout]).to be(ObjectStorage::DirectUpload::TIMEOUT)
2023-07-09 08:55:56 +05:30
upload_path = File.join(described_class::TMP_UPLOAD_PATH, subject[:RemoteObject][:ID])
expect(subject[:RemoteObject][:GetURL]).to include(upload_path)
expect(subject[:RemoteObject][:DeleteURL]).to include(upload_path)
expect(subject[:RemoteObject][:StoreURL]).to include(upload_path)
expect(subject[:RemoteObject][:SkipDelete]).to eq(false)
::Gitlab::Redis::SharedState.with do |redis|
expect(redis.hlen(ObjectStorage::PendingDirectUpload::KEY)).to be_zero
end
2020-11-24 15:15:51 +05:30
end
2018-05-09 12:01:36 +05:30
end
2018-11-08 19:23:39 +05:30
end
shared_examples 'uses remote storage with multipart uploads' do
it_behaves_like 'uses remote storage' do
it "returns multipart upload" do
is_expected.to have_key(:RemoteObject)
expect(subject[:RemoteObject]).to have_key(:MultipartUpload)
expect(subject[:RemoteObject][:MultipartUpload]).to have_key(:PartSize)
2023-07-09 08:55:56 +05:30
upload_path = File.join(described_class::TMP_UPLOAD_PATH, subject[:RemoteObject][:ID])
expect(subject[:RemoteObject][:MultipartUpload][:PartURLs]).to all(include(upload_path))
expect(subject[:RemoteObject][:MultipartUpload][:CompleteURL]).to include(upload_path)
expect(subject[:RemoteObject][:MultipartUpload][:AbortURL]).to include(upload_path)
2018-11-08 19:23:39 +05:30
end
end
end
2018-05-09 12:01:36 +05:30
2018-11-08 19:23:39 +05:30
shared_examples 'uses remote storage without multipart uploads' do
it_behaves_like 'uses remote storage' do
it "does not return multipart upload" do
is_expected.to have_key(:RemoteObject)
expect(subject[:RemoteObject]).not_to have_key(:MultipartUpload)
end
2018-05-09 12:01:36 +05:30
end
end
2023-07-09 08:55:56 +05:30
shared_examples 'handling object storage final upload path' do |multipart|
context 'when use_final_store_path is true' do
let(:use_final_store_path) { true }
let(:final_store_path) { File.join('@final', 'abc', '123', 'somefilename') }
let(:escaped_path) { escape_path(final_store_path) }
before do
stub_object_storage_multipart_init_with_final_store_path("#{storage_url}#{final_store_path}") if multipart
allow(uploader_class).to receive(:generate_final_store_path).and_return(final_store_path)
end
it 'uses the full path instead of the temporary one' do
expect(subject[:RemoteObject][:ID]).to eq(final_store_path)
expect(subject[:RemoteObject][:GetURL]).to include(escaped_path)
expect(subject[:RemoteObject][:StoreURL]).to include(escaped_path)
if multipart
expect(subject[:RemoteObject][:MultipartUpload][:PartURLs]).to all(include(escaped_path))
expect(subject[:RemoteObject][:MultipartUpload][:CompleteURL]).to include(escaped_path)
expect(subject[:RemoteObject][:MultipartUpload][:AbortURL]).to include(escaped_path)
end
expect(subject[:RemoteObject][:SkipDelete]).to eq(true)
expect(
ObjectStorage::PendingDirectUpload.exists?(uploader_class.storage_location_identifier, final_store_path)
).to eq(true)
end
context 'and bucket prefix is configured' do
let(:prefixed_final_store_path) { "my/prefix/#{final_store_path}" }
let(:escaped_path) { escape_path(prefixed_final_store_path) }
before do
allow(uploader_class.object_store_options).to receive(:bucket_prefix).and_return('my/prefix')
if multipart
stub_object_storage_multipart_init_with_final_store_path("#{storage_url}#{prefixed_final_store_path}")
end
end
it 'sets the remote object ID to the final path without prefix' do
expect(subject[:RemoteObject][:ID]).to eq(final_store_path)
end
it 'returns the final path with prefix' do
expect(subject[:RemoteObject][:GetURL]).to include(escaped_path)
expect(subject[:RemoteObject][:StoreURL]).to include(escaped_path)
if multipart
expect(subject[:RemoteObject][:MultipartUpload][:PartURLs]).to all(include(escaped_path))
expect(subject[:RemoteObject][:MultipartUpload][:CompleteURL]).to include(escaped_path)
expect(subject[:RemoteObject][:MultipartUpload][:AbortURL]).to include(escaped_path)
end
end
it 'creates the pending upload entry without the prefix' do
is_expected.to have_key(:RemoteObject)
expect(
ObjectStorage::PendingDirectUpload.exists?(uploader_class.storage_location_identifier, final_store_path)
).to eq(true)
end
end
end
def escape_path(path)
# This is what the private method Fog::AWS::Storage#object_to_path will do to the object name
Fog::AWS.escape(path).gsub('%2F', '/')
end
end
2018-05-09 12:01:36 +05:30
context 'when object storage is disabled' do
before do
allow(Gitlab.config.uploads.object_store).to receive(:enabled) { false }
end
it_behaves_like 'uses local storage'
end
context 'when object storage is enabled' do
before do
allow(Gitlab.config.uploads.object_store).to receive(:enabled) { true }
end
context 'when direct upload is enabled' do
before do
allow(Gitlab.config.uploads.object_store).to receive(:direct_upload) { true }
end
context 'uses AWS' do
2021-02-22 17:27:13 +05:30
let(:storage_url) { "https://uploads.s3.eu-central-1.amazonaws.com/" }
2020-10-24 23:57:45 +05:30
let(:credentials) do
{
provider: "AWS",
aws_access_key_id: "AWS_ACCESS_KEY_ID",
aws_secret_access_key: "AWS_SECRET_ACCESS_KEY",
region: "eu-central-1"
}
end
2018-11-08 19:23:39 +05:30
2018-05-09 12:01:36 +05:30
before do
2020-10-24 23:57:45 +05:30
expect_next_instance_of(ObjectStorage::Config) do |instance|
allow(instance).to receive(:credentials).and_return(credentials)
2018-05-09 12:01:36 +05:30
end
end
2018-11-08 19:23:39 +05:30
context 'for known length' do
it_behaves_like 'uses remote storage without multipart uploads' do
it 'returns links for S3' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
end
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path'
2018-11-08 19:23:39 +05:30
end
context 'for unknown length' do
let(:has_length) { false }
let(:maximum_size) { 1.gigabyte }
2018-05-09 12:01:36 +05:30
2018-11-08 19:23:39 +05:30
before do
stub_object_storage_multipart_init(storage_url)
end
it_behaves_like 'uses remote storage with multipart uploads' do
it 'returns links for S3' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:MultipartUpload][:PartURLs]).to all(start_with(storage_url))
expect(subject[:RemoteObject][:MultipartUpload][:CompleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:MultipartUpload][:AbortURL]).to start_with(storage_url)
end
2018-05-09 12:01:36 +05:30
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path', :multipart
2018-05-09 12:01:36 +05:30
end
end
context 'uses Google' do
2018-11-08 19:23:39 +05:30
let(:storage_url) { "https://storage.googleapis.com/uploads/" }
2020-10-24 23:57:45 +05:30
let(:credentials) do
{
provider: "Google",
google_storage_access_key_id: 'ACCESS_KEY_ID',
google_storage_secret_access_key: 'SECRET_ACCESS_KEY'
}
end
2018-11-08 19:23:39 +05:30
2018-05-09 12:01:36 +05:30
before do
2020-10-24 23:57:45 +05:30
expect_next_instance_of(ObjectStorage::Config) do |instance|
allow(instance).to receive(:credentials).and_return(credentials)
2018-05-09 12:01:36 +05:30
end
end
2018-11-08 19:23:39 +05:30
context 'for known length' do
it_behaves_like 'uses remote storage without multipart uploads' do
it 'returns links for Google Cloud' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
end
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path'
2018-11-08 19:23:39 +05:30
end
context 'for unknown length' do
let(:has_length) { false }
let(:maximum_size) { 1.gigabyte }
2018-05-09 12:01:36 +05:30
2018-11-08 19:23:39 +05:30
it_behaves_like 'uses remote storage without multipart uploads' do
it 'returns links for Google Cloud' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
end
2018-05-09 12:01:36 +05:30
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path'
2018-05-09 12:01:36 +05:30
end
end
context 'uses GDK/minio' do
2018-11-08 19:23:39 +05:30
let(:storage_url) { "http://minio:9000/uploads/" }
2020-10-24 23:57:45 +05:30
let(:credentials) do
{ provider: "AWS",
aws_access_key_id: "AWS_ACCESS_KEY_ID",
aws_secret_access_key: "AWS_SECRET_ACCESS_KEY",
endpoint: 'http://minio:9000',
path_style: true,
region: "gdk" }
end
2018-11-08 19:23:39 +05:30
2018-05-09 12:01:36 +05:30
before do
2020-10-24 23:57:45 +05:30
expect_next_instance_of(ObjectStorage::Config) do |instance|
allow(instance).to receive(:credentials).and_return(credentials)
2018-05-09 12:01:36 +05:30
end
end
2018-11-08 19:23:39 +05:30
context 'for known length' do
it_behaves_like 'uses remote storage without multipart uploads' do
it 'returns links for S3' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
end
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path'
2018-11-08 19:23:39 +05:30
end
context 'for unknown length' do
let(:has_length) { false }
let(:maximum_size) { 1.gigabyte }
2018-05-09 12:01:36 +05:30
2018-11-08 19:23:39 +05:30
before do
stub_object_storage_multipart_init(storage_url)
end
it_behaves_like 'uses remote storage with multipart uploads' do
it 'returns links for S3' do
expect(subject[:RemoteObject][:GetURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:DeleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:StoreURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:MultipartUpload][:PartURLs]).to all(start_with(storage_url))
expect(subject[:RemoteObject][:MultipartUpload][:CompleteURL]).to start_with(storage_url)
expect(subject[:RemoteObject][:MultipartUpload][:AbortURL]).to start_with(storage_url)
end
2018-05-09 12:01:36 +05:30
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'handling object storage final upload path', :multipart
2018-05-09 12:01:36 +05:30
end
end
end
context 'when direct upload is disabled' do
before do
allow(Gitlab.config.uploads.object_store).to receive(:direct_upload) { false }
end
it_behaves_like 'uses local storage'
end
end
end
describe '#cache!' do
subject do
uploader.cache!(uploaded_file)
end
context 'when local file is used' do
let(:temp_file) { Tempfile.new("test") }
before do
FileUtils.touch(temp_file)
end
after do
FileUtils.rm_f(temp_file)
end
context 'when valid file is used' do
context 'when valid file is specified' do
let(:uploaded_file) { temp_file }
2020-04-22 19:07:51 +05:30
it 'properly caches the file' do
subject
expect(uploader).to be_exists
expect(uploader.path).to start_with(uploader_class.root)
expect(uploader.filename).to eq(File.basename(uploaded_file.path))
end
2018-05-09 12:01:36 +05:30
context 'when object storage and direct upload is specified' do
before do
stub_uploads_object_storage(uploader_class, enabled: true, direct_upload: true)
end
context 'when file is stored' do
subject do
uploader.store!(uploaded_file)
end
it 'file to be remotely stored in permament location' do
subject
expect(uploader).to be_exists
expect(uploader).not_to be_cached
expect(uploader).not_to be_file_storage
expect(uploader.path).not_to be_nil
expect(uploader.path).not_to include('tmp/upload')
expect(uploader.path).not_to include('tmp/cache')
expect(uploader.object_store).to eq(described_class::Store::REMOTE)
end
end
end
context 'when object storage and direct upload is not used' do
before do
stub_uploads_object_storage(uploader_class, enabled: true, direct_upload: false)
end
context 'when file is stored' do
subject do
uploader.store!(uploaded_file)
end
it 'file to be remotely stored in permament location' do
subject
expect(uploader).to be_exists
expect(uploader).not_to be_cached
expect(uploader).to be_file_storage
expect(uploader.path).not_to be_nil
expect(uploader.path).not_to include('tmp/upload')
expect(uploader.path).not_to include('tmp/cache')
expect(uploader.object_store).to eq(described_class::Store::LOCAL)
end
end
end
end
end
end
context 'when remote file is used' do
let(:temp_file) { Tempfile.new("test") }
let!(:fog_connection) do
stub_uploads_object_storage(uploader_class)
end
before do
FileUtils.touch(temp_file)
end
after do
FileUtils.rm_f(temp_file)
end
context 'when valid file is used' do
context 'when invalid file is specified' do
let(:uploaded_file) do
UploadedFile.new(temp_file.path, remote_id: "../test/123123")
end
it 'raises an error' do
expect { subject }.to raise_error(uploader_class::RemoteStoreError, /Bad file path/)
end
end
context 'when non existing file is specified' do
let(:uploaded_file) do
UploadedFile.new(temp_file.path, remote_id: "test/123123")
end
it 'raises an error' do
expect { subject }.to raise_error(uploader_class::RemoteStoreError, /Missing file/)
end
end
2020-03-28 13:19:24 +05:30
context 'when empty remote_id is specified' do
let(:uploaded_file) do
UploadedFile.new(temp_file.path, remote_id: '')
end
it 'uses local storage' do
subject
expect(uploader).to be_file_storage
expect(uploader.object_store).to eq(described_class::Store::LOCAL)
end
end
2018-05-09 12:01:36 +05:30
context 'when valid file is specified' do
let(:uploaded_file) do
UploadedFile.new(temp_file.path, filename: "my_file.txt", remote_id: "test/123123")
end
let!(:fog_file) do
2021-01-03 14:25:43 +05:30
fog_connection.directories.new(key: 'uploads').files.create( # rubocop:disable Rails/SaveBang
2018-11-08 19:23:39 +05:30
key: 'tmp/uploads/test/123123',
2018-05-09 12:01:36 +05:30
body: 'content'
)
end
it 'file to be cached and remote stored' do
expect { subject }.not_to raise_error
expect(uploader).to be_exists
expect(uploader).to be_cached
2023-07-09 08:55:56 +05:30
expect(uploader.cache_only).to be_falsey
2018-05-09 12:01:36 +05:30
expect(uploader).not_to be_file_storage
expect(uploader.path).not_to be_nil
2023-07-09 08:55:56 +05:30
expect(uploader.path).to include('tmp/uploads')
2018-05-09 12:01:36 +05:30
expect(uploader.path).not_to include('tmp/cache')
expect(uploader.object_store).to eq(described_class::Store::REMOTE)
end
context 'when file is stored' do
subject do
uploader.store!(uploaded_file)
end
it 'file to be remotely stored in permament location' do
subject
expect(uploader).to be_exists
expect(uploader).not_to be_cached
expect(uploader).not_to be_file_storage
expect(uploader.path).not_to be_nil
expect(uploader.path).not_to include('tmp/upload')
expect(uploader.path).not_to include('tmp/cache')
expect(uploader.url).to include('/my_file.txt')
expect(uploader.object_store).to eq(described_class::Store::REMOTE)
end
end
2023-07-09 08:55:56 +05:30
context 'when uploaded file remote_id matches a pending direct upload entry' do
let(:object) { build_stubbed(:ci_job_artifact) }
let(:final_path) { '@final/test/123123' }
let(:fog_config) { Gitlab.config.uploads.object_store }
let(:bucket) { 'uploads' }
let(:uploaded_file) { UploadedFile.new(temp_file.path, filename: "my_file.txt", remote_id: final_path) }
let(:fog_file_path) { final_path }
let(:fog_connection_2) do
stub_object_storage_uploader(
config: fog_config,
uploader: uploader_class,
direct_upload: true
)
end
let!(:fog_file_2) do
fog_connection_2.directories.new(key: bucket).files.create( # rubocop:disable Rails/SaveBang
key: fog_file_path,
body: 'content'
)
end
before do
ObjectStorage::PendingDirectUpload.prepare(
uploader_class.storage_location_identifier,
final_path
)
end
it 'file to be cached and remote stored with final path set' do
expect { subject }.not_to raise_error
expect(uploader).to be_exists
expect(uploader).to be_cached
expect(uploader.cache_only).to be_falsey
expect(uploader).not_to be_file_storage
expect(uploader.path).to eq(uploaded_file.remote_id)
expect(uploader.object_store).to eq(described_class::Store::REMOTE)
expect(object.file_final_path).to eq(uploaded_file.remote_id)
end
context 'when bucket prefix is configured' do
let(:fog_config) do
Gitlab.config.uploads.object_store.tap do |config|
config[:remote_directory] = 'main-bucket'
config[:bucket_prefix] = 'uploads'
end
end
let(:bucket) { 'main-bucket' }
let(:fog_file_path) { "uploads/#{final_path}" }
it 'stores the file final path in the db without the prefix' do
expect { subject }.not_to raise_error
expect(uploader.store_path).to eq("uploads/#{final_path}")
expect(object.file_final_path).to eq(final_path)
end
end
context 'when file is stored' do
subject do
uploader.store!(uploaded_file)
end
it 'file to be remotely stored in permament location' do
subject
expect(uploader).to be_exists
expect(uploader).not_to be_cached
expect(uploader.path).to eq(uploaded_file.remote_id)
end
it 'does not trigger Carrierwave copy and delete because it is already in the final location' do
expect_next_instance_of(CarrierWave::Storage::Fog::File) do |instance|
expect(instance).not_to receive(:copy_to)
expect(instance).not_to receive(:delete)
end
subject
end
end
end
2018-05-09 12:01:36 +05:30
end
end
end
end
2018-11-08 19:23:39 +05:30
describe '#retrieve_from_store!' do
2023-07-09 08:55:56 +05:30
context 'uploaders that includes the RecordsUploads extension' do
[:group, :project, :user].each do |model|
context "for #{model}s" do
let(:models) { create_list(model, 3, :with_avatar).map(&:reload) }
let(:avatars) { models.map(&:avatar) }
it 'batches fetching uploads from the database' do
# Ensure that these are all created and fully loaded before we start
# running queries for avatars
models
expect { avatars }.not_to exceed_query_limit(1)
end
it 'does not attempt to replace methods' do
models.each do |model|
expect(model.avatar.upload).to receive(:method_missing).and_call_original
2018-11-08 19:23:39 +05:30
2023-07-09 08:55:56 +05:30
model.avatar.upload.path
end
end
2018-11-08 19:23:39 +05:30
2023-07-09 08:55:56 +05:30
it 'fetches a unique upload for each model' do
expect(avatars.map(&:url).uniq).to eq(avatars.map(&:url))
expect(avatars.map(&:upload).uniq).to eq(avatars.map(&:upload))
end
2018-11-08 19:23:39 +05:30
end
2023-07-09 08:55:56 +05:30
end
end
2018-11-08 19:23:39 +05:30
2023-07-09 08:55:56 +05:30
describe 'filename' do
let(:model) { create(:ci_job_artifact, :remote_store, :archive) }
2019-07-31 22:56:46 +05:30
2023-07-09 08:55:56 +05:30
before do
stub_artifacts_object_storage
end
shared_examples 'ensuring correct filename' do
it 'uses the original filename' do
expect(model.reload.file.filename).to eq('ci_build_artifacts.zip')
2019-07-31 22:56:46 +05:30
end
2023-07-09 08:55:56 +05:30
end
2019-07-31 22:56:46 +05:30
2023-07-09 08:55:56 +05:30
context 'when model has final path defined for the file column' do
before do
model.update_column(:file_final_path, 'some/final/path/abc-123')
2018-11-08 19:23:39 +05:30
end
2023-07-09 08:55:56 +05:30
it_behaves_like 'ensuring correct filename'
end
context 'when model has no final path defined for the file column' do
it_behaves_like 'ensuring correct filename'
2018-11-08 19:23:39 +05:30
end
end
end
2020-11-24 15:15:51 +05:30
2023-07-09 08:55:56 +05:30
describe '.generate_final_store_path' do
subject(:final_path) { uploader_class.generate_final_store_path }
before do
allow(Digest::SHA2).to receive(:hexdigest).and_return('somehash1234')
end
it 'returns the generated hashed path' do
expect(final_path).to eq('@final/so/me/hash1234')
end
end
2020-11-24 15:15:51 +05:30
describe 'OpenFile' do
subject { ObjectStorage::Concern::OpenFile.new(file) }
let(:file) { double(read: true, size: true, path: true) }
it 'delegates read and size methods' do
expect(subject.read).to eq(true)
expect(subject.size).to eq(true)
end
it 'does not delegate path method' do
expect { subject.path }.to raise_error(NoMethodError)
end
end
2018-05-09 12:01:36 +05:30
end