2019-07-31 22:56:46 +05:30
|
|
|
# frozen_string_literal: true
|
|
|
|
|
|
|
|
require 'spec_helper'
|
|
|
|
|
|
|
|
describe Git::BranchHooksService do
|
|
|
|
include RepoHelpers
|
2019-10-12 21:52:04 +05:30
|
|
|
include ProjectForksHelper
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
let(:project) { create(:project, :repository) }
|
|
|
|
let(:user) { project.creator }
|
|
|
|
|
|
|
|
let(:branch) { project.default_branch }
|
|
|
|
let(:ref) { "refs/heads/#{branch}" }
|
|
|
|
let(:commit) { project.commit(sample_commit.id) }
|
|
|
|
let(:oldrev) { commit.parent_id }
|
|
|
|
let(:newrev) { commit.id }
|
|
|
|
|
|
|
|
let(:service) do
|
|
|
|
described_class.new(project, user, oldrev: oldrev, newrev: newrev, ref: ref)
|
|
|
|
end
|
|
|
|
|
2019-09-04 21:01:54 +05:30
|
|
|
it 'update remote mirrors' do
|
|
|
|
expect(service).to receive(:update_remote_mirrors).and_call_original
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
|
2019-07-31 22:56:46 +05:30
|
|
|
describe "Git Push Data" do
|
2019-10-12 21:52:04 +05:30
|
|
|
subject(:push_data) { service.send(:push_data) }
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
it 'has expected push data attributes' do
|
|
|
|
is_expected.to match a_hash_including(
|
|
|
|
object_kind: 'push',
|
|
|
|
before: oldrev,
|
|
|
|
after: newrev,
|
|
|
|
ref: ref,
|
|
|
|
user_id: user.id,
|
|
|
|
user_name: user.name,
|
|
|
|
project_id: project.id
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
|
|
|
context "with repository data" do
|
|
|
|
subject { push_data[:repository] }
|
|
|
|
|
|
|
|
it 'has expected attributes' do
|
|
|
|
is_expected.to match a_hash_including(
|
|
|
|
name: project.name,
|
|
|
|
url: project.url_to_repo,
|
|
|
|
description: project.description,
|
|
|
|
homepage: project.web_url
|
|
|
|
)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context "with commits" do
|
|
|
|
subject { push_data[:commits] }
|
|
|
|
|
|
|
|
it { is_expected.to be_an(Array) }
|
|
|
|
|
|
|
|
it 'has 1 element' do
|
|
|
|
expect(subject.size).to eq(1)
|
|
|
|
end
|
|
|
|
|
|
|
|
context "the commit" do
|
|
|
|
subject { push_data[:commits].first }
|
|
|
|
|
|
|
|
it { expect(subject[:timestamp].in_time_zone).to eq(commit.date.in_time_zone) }
|
|
|
|
|
|
|
|
it 'includes expected commit data' do
|
|
|
|
is_expected.to match a_hash_including(
|
|
|
|
id: commit.id,
|
|
|
|
message: commit.safe_message,
|
|
|
|
url: [
|
|
|
|
Gitlab.config.gitlab.url,
|
|
|
|
project.namespace.to_param,
|
|
|
|
project.to_param,
|
|
|
|
'commit',
|
|
|
|
commit.id
|
|
|
|
].join('/')
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
|
|
|
context "with a author" do
|
|
|
|
subject { push_data[:commits].first[:author] }
|
|
|
|
|
|
|
|
it 'includes expected author data' do
|
|
|
|
is_expected.to match a_hash_including(
|
|
|
|
name: commit.author_name,
|
|
|
|
email: commit.author_email
|
|
|
|
)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
describe 'Push Event' do
|
|
|
|
let(:event) { Event.find_by_action(Event::PUSHED) }
|
|
|
|
|
|
|
|
before do
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
|
|
|
|
context "with an existing branch" do
|
|
|
|
it 'generates a push event with one commit' do
|
|
|
|
expect(event).to be_an_instance_of(PushEvent)
|
|
|
|
expect(event.project).to eq(project)
|
|
|
|
expect(event.action).to eq(Event::PUSHED)
|
|
|
|
expect(event.push_event_payload).to be_an_instance_of(PushEventPayload)
|
|
|
|
expect(event.push_event_payload.commit_from).to eq(oldrev)
|
|
|
|
expect(event.push_event_payload.commit_to).to eq(newrev)
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(event.push_event_payload.commit_title).to eq('Change some files')
|
2019-07-31 22:56:46 +05:30
|
|
|
expect(event.push_event_payload.ref).to eq('master')
|
|
|
|
expect(event.push_event_payload.commit_count).to eq(1)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context "with a new branch" do
|
|
|
|
let(:oldrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'generates a push event with more than one commit' do
|
|
|
|
expect(event).to be_an_instance_of(PushEvent)
|
|
|
|
expect(event.project).to eq(project)
|
|
|
|
expect(event.action).to eq(Event::PUSHED)
|
|
|
|
expect(event.push_event_payload).to be_an_instance_of(PushEventPayload)
|
|
|
|
expect(event.push_event_payload.commit_from).to be_nil
|
|
|
|
expect(event.push_event_payload.commit_to).to eq(newrev)
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(event.push_event_payload.commit_title).to eq('Initial commit')
|
2019-07-31 22:56:46 +05:30
|
|
|
expect(event.push_event_payload.ref).to eq('master')
|
|
|
|
expect(event.push_event_payload.commit_count).to be > 1
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'removing a branch' do
|
|
|
|
let(:newrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'generates a push event with no commits' do
|
|
|
|
expect(event).to be_an_instance_of(PushEvent)
|
|
|
|
expect(event.project).to eq(project)
|
|
|
|
expect(event.action).to eq(Event::PUSHED)
|
|
|
|
expect(event.push_event_payload).to be_an_instance_of(PushEventPayload)
|
|
|
|
expect(event.push_event_payload.commit_from).to eq(oldrev)
|
|
|
|
expect(event.push_event_payload.commit_to).to be_nil
|
|
|
|
expect(event.push_event_payload.ref).to eq('master')
|
|
|
|
expect(event.push_event_payload.commit_count).to eq(0)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
describe 'Invalidating project cache' do
|
|
|
|
let(:commit_id) do
|
|
|
|
project.repository.update_file(
|
|
|
|
user, 'README.md', '', message: 'Update', branch_name: branch
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
|
|
|
let(:commit) { project.repository.commit(commit_id) }
|
|
|
|
let(:blank_sha) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
def clears_cache(extended: [])
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(service).to receive(:invalidated_file_types).and_return(extended)
|
|
|
|
|
|
|
|
if extended.present?
|
|
|
|
expect(ProjectCacheWorker)
|
|
|
|
.to receive(:perform_async)
|
|
|
|
.with(project.id, extended, [], false)
|
|
|
|
end
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
|
|
|
|
def clears_extended_cache
|
|
|
|
clears_cache(extended: %i[readme])
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'on default branch' do
|
|
|
|
context 'create' do
|
|
|
|
# FIXME: When creating the default branch,the cache worker runs twice
|
|
|
|
before do
|
|
|
|
allow(ProjectCacheWorker).to receive(:perform_async)
|
|
|
|
end
|
|
|
|
|
|
|
|
let(:oldrev) { blank_sha }
|
|
|
|
|
|
|
|
it { clears_cache }
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'update' do
|
|
|
|
it { clears_extended_cache }
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'remove' do
|
|
|
|
let(:newrev) { blank_sha }
|
|
|
|
|
|
|
|
# TODO: this case should pass, but we only take account of added files
|
|
|
|
it { clears_cache }
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'on ordinary branch' do
|
|
|
|
let(:branch) { 'fix' }
|
|
|
|
|
|
|
|
context 'create' do
|
|
|
|
let(:oldrev) { blank_sha }
|
|
|
|
|
|
|
|
it { clears_cache }
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'update' do
|
|
|
|
it { clears_cache }
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'remove' do
|
|
|
|
let(:newrev) { blank_sha }
|
|
|
|
|
|
|
|
it { clears_cache }
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
describe 'GPG signatures' do
|
|
|
|
context 'when the commit has a signature' do
|
|
|
|
context 'when the signature is already cached' do
|
|
|
|
before do
|
|
|
|
create(:gpg_signature, commit_sha: commit.id)
|
|
|
|
end
|
|
|
|
|
|
|
|
it 'does not queue a CreateGpgSignatureWorker' do
|
|
|
|
expect(CreateGpgSignatureWorker).not_to receive(:perform_async)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'when the signature is not yet cached' do
|
|
|
|
it 'queues a CreateGpgSignatureWorker' do
|
|
|
|
expect(CreateGpgSignatureWorker).to receive(:perform_async).with([commit.id], project.id)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
|
|
|
|
it 'can queue several commits to create the gpg signature' do
|
|
|
|
allow(Gitlab::Git::Commit)
|
|
|
|
.to receive(:shas_with_signatures)
|
|
|
|
.and_return([sample_commit.id, another_sample_commit.id])
|
|
|
|
|
|
|
|
expect(CreateGpgSignatureWorker)
|
|
|
|
.to receive(:perform_async)
|
|
|
|
.with([sample_commit.id, another_sample_commit.id], project.id)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'when the commit does not have a signature' do
|
|
|
|
before do
|
|
|
|
allow(Gitlab::Git::Commit)
|
|
|
|
.to receive(:shas_with_signatures)
|
|
|
|
.with(project.repository, [sample_commit.id])
|
|
|
|
.and_return([])
|
|
|
|
end
|
|
|
|
|
|
|
|
it 'does not queue a CreateGpgSignatureWorker' do
|
|
|
|
expect(CreateGpgSignatureWorker)
|
|
|
|
.not_to receive(:perform_async)
|
|
|
|
.with(sample_commit.id, project.id)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
describe 'Processing commit messages' do
|
2019-10-12 21:52:04 +05:30
|
|
|
# Create 6 commits, 3 of which have references. Limiting to 4 commits, we
|
|
|
|
# expect to see two commit message processors enqueued.
|
|
|
|
let!(:commit_ids) do
|
|
|
|
Array.new(6) do |i|
|
2019-07-31 22:56:46 +05:30
|
|
|
message = "Issue #{'#' if i.even?}#{i}"
|
|
|
|
project.repository.update_file(
|
|
|
|
user, 'README.md', '', message: message, branch_name: branch
|
|
|
|
)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2019-10-12 21:52:04 +05:30
|
|
|
let(:oldrev) { project.commit(commit_ids.first).parent_id }
|
2019-07-31 22:56:46 +05:30
|
|
|
let(:newrev) { commit_ids.last }
|
|
|
|
|
|
|
|
before do
|
2019-10-12 21:52:04 +05:30
|
|
|
stub_const("::Git::BaseHooksService::PROCESS_COMMIT_LIMIT", 4)
|
2019-07-31 22:56:46 +05:30
|
|
|
end
|
|
|
|
|
|
|
|
context 'creating the default branch' do
|
|
|
|
let(:oldrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
2019-09-04 21:01:54 +05:30
|
|
|
it 'processes a limited number of commit messages' do
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).twice
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'updating the default branch' do
|
|
|
|
it 'processes a limited number of commit messages' do
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).twice
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'removing the default branch' do
|
|
|
|
let(:newrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'does not process commit messages' do
|
|
|
|
expect(ProcessCommitWorker).not_to receive(:perform_async)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'creating a normal branch' do
|
|
|
|
let(:branch) { 'fix' }
|
|
|
|
let(:oldrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'processes a limited number of commit messages' do
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).twice
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'updating a normal branch' do
|
|
|
|
let(:branch) { 'fix' }
|
|
|
|
|
|
|
|
it 'processes a limited number of commit messages' do
|
2019-10-12 21:52:04 +05:30
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).twice
|
2019-07-31 22:56:46 +05:30
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'removing a normal branch' do
|
|
|
|
let(:branch) { 'fix' }
|
|
|
|
let(:newrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'does not process commit messages' do
|
|
|
|
expect(ProcessCommitWorker).not_to receive(:perform_async)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
2019-10-12 21:52:04 +05:30
|
|
|
|
|
|
|
context 'when the project is forked' do
|
|
|
|
let(:upstream_project) { project }
|
|
|
|
let(:forked_project) { fork_project(upstream_project, user, repository: true) }
|
|
|
|
|
|
|
|
let!(:forked_service) do
|
|
|
|
described_class.new(forked_project, user, oldrev: oldrev, newrev: newrev, ref: ref)
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'when commits already exists in the upstream project' do
|
|
|
|
it 'does not process commit messages' do
|
|
|
|
expect(ProcessCommitWorker).not_to receive(:perform_async)
|
|
|
|
|
|
|
|
forked_service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'when a commit does not exist in the upstream repo' do
|
|
|
|
# On top of the existing 6 commits, 3 of which have references,
|
|
|
|
# create 2 more, 1 of which has a reference. Limiting to 4 commits, we
|
|
|
|
# expect to see one commit message processor enqueued.
|
|
|
|
let!(:forked_commit_ids) do
|
|
|
|
Array.new(2) do |i|
|
|
|
|
message = "Issue #{'#' if i.even?}#{i}"
|
|
|
|
forked_project.repository.update_file(
|
|
|
|
user, 'README.md', '', message: message, branch_name: branch
|
|
|
|
)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
let(:newrev) { forked_commit_ids.last }
|
|
|
|
|
|
|
|
it 'processes the commit message' do
|
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).once
|
|
|
|
|
|
|
|
forked_service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'when the upstream project no longer exists' do
|
|
|
|
it 'processes the commit messages' do
|
|
|
|
upstream_project.destroy!
|
|
|
|
|
|
|
|
expect(ProcessCommitWorker).to receive(:perform_async).twice
|
|
|
|
|
|
|
|
forked_service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
2019-07-31 22:56:46 +05:30
|
|
|
end
|
2019-09-30 21:07:59 +05:30
|
|
|
|
|
|
|
describe 'New branch detection' do
|
|
|
|
let(:branch) { 'fix' }
|
|
|
|
|
|
|
|
context 'oldrev is the blank SHA' do
|
|
|
|
let(:oldrev) { Gitlab::Git::BLANK_SHA }
|
|
|
|
|
|
|
|
it 'is treated as a new branch' do
|
|
|
|
expect(service).to receive(:branch_create_hooks)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'oldrev is set' do
|
|
|
|
context 'Gitaly does not know about the branch' do
|
|
|
|
it 'is treated as a new branch' do
|
|
|
|
allow(project.repository).to receive(:branch_names) { [] }
|
|
|
|
|
|
|
|
expect(service).to receive(:branch_create_hooks)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
context 'Gitaly knows about the branch' do
|
|
|
|
it 'is not treated as a new branch' do
|
|
|
|
expect(service).not_to receive(:branch_create_hooks)
|
|
|
|
|
|
|
|
service.execute
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
2019-07-31 22:56:46 +05:30
|
|
|
end
|