app/models/ci/pipeline.rb (1,150 lines of code) (raw):

# frozen_string_literal: true module Ci class Pipeline < Ci::ApplicationRecord include Ci::Partitionable include Ci::HasStatus include Ci::HasCompletionReason include Importable include AfterCommitQueue include Presentable include Gitlab::Allowable include Gitlab::OptimisticLocking include Gitlab::Utils::StrongMemoize include AtomicInternalId include Ci::HasRef include ShaAttribute include FromUnion include UpdatedAtFilterable include EachBatch include FastDestroyAll::Helpers include Gitlab::InternalEventsTracking self.table_name = :p_ci_pipelines self.primary_key = :id self.sequence_name = :ci_pipelines_id_seq MAX_OPEN_MERGE_REQUESTS_REFS = 4 YAML_ERRORS_MAX_LENGTH = 100.kilobytes MAX_SHA_LENGTH = 64 MAX_REF_LENGTH = 300 PROJECT_ROUTE_AND_NAMESPACE_ROUTE = { project: [:project_feature, :route, { namespace: :route }] }.freeze DEFAULT_CONFIG_PATH = '.gitlab-ci.yml' CANCELABLE_STATUSES = (Ci::HasStatus::CANCELABLE_STATUSES + ['manual']).freeze UNLOCKABLE_STATUSES = (Ci::Pipeline.completed_statuses + [:manual]).freeze # UI only shows 100+. TODO: pass constant to UI for SSoT COUNT_FAILED_JOBS_LIMIT = 101 INPUTS_LIMIT = 20 paginates_per 15 sha_attribute :source_sha sha_attribute :target_sha query_constraints :id, :partition_id partitionable scope: ->(_) { Ci::Pipeline.current_partition_value }, partitioned: true # Ci::CreatePipelineService returns Ci::Pipeline so this is the only place # where we can pass additional information from the service. This accessor # is used for storing the processed metadata for linting purposes. # There is an open issue to address this: # https://gitlab.com/gitlab-org/gitlab/-/issues/259010 attr_accessor :config_metadata # This is used to retain access to the method defined by `Ci::HasRef` # before being overridden in this class. alias_method :jobs_git_ref, :git_ref belongs_to :project, inverse_of: :all_pipelines belongs_to :project_mirror, primary_key: :project_id, foreign_key: :project_id, inverse_of: :pipelines belongs_to :user belongs_to :auto_canceled_by, class_name: 'Ci::Pipeline', inverse_of: :auto_canceled_pipelines belongs_to :pipeline_schedule, class_name: 'Ci::PipelineSchedule' belongs_to :merge_request, class_name: 'MergeRequest' belongs_to :external_pull_request, class_name: 'Ci::ExternalPullRequest' belongs_to :ci_ref, class_name: 'Ci::Ref', foreign_key: :ci_ref_id, inverse_of: :pipelines belongs_to :trigger, class_name: 'Ci::Trigger', inverse_of: :pipelines has_internal_id :iid, scope: :project, presence: false, track_if: -> { !importing? }, ensure_if: -> { !importing? }, init: ->(pipeline, scope) do if pipeline pipeline.project&.all_pipelines&.maximum(:iid) || pipeline.project&.all_pipelines&.count elsif scope ::Ci::Pipeline.where(**scope).maximum(:iid) end end has_many :stages, ->(pipeline) { in_partition(pipeline).order(position: :asc) }, partition_foreign_key: :partition_id, inverse_of: :pipeline has_one :workload, ->(pipeline) { in_partition(pipeline) }, class_name: "Ci::Workloads::Workload", partition_foreign_key: :partition_id, inverse_of: :pipeline # # In https://gitlab.com/groups/gitlab-org/-/epics/9991, we aim to convert all CommitStatus related models to # Ci::Job models. With that epic, we aim to replace `statuses` with `jobs`. # # DEPRECATED: has_many :statuses, ->(pipeline) { in_partition(pipeline) }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :processables, ->(pipeline) { in_partition(pipeline) }, class_name: 'Ci::Processable', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :latest_statuses_ordered_by_stage, ->(pipeline) { latest.in_partition(pipeline).order(:stage_idx, :stage) }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :latest_statuses, ->(pipeline) { latest.in_partition(pipeline) }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :statuses_order_id_desc, ->(pipeline) { in_partition(pipeline).order_id_desc }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :bridges, ->(pipeline) { in_partition(pipeline) }, class_name: 'Ci::Bridge', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :builds, ->(pipeline) { in_partition(pipeline) }, foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :build_execution_configs, ->(pipeline) { in_partition(pipeline) }, class_name: 'Ci::BuildExecutionConfig', inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :generic_commit_statuses, ->(pipeline) { in_partition(pipeline) }, foreign_key: :commit_id, inverse_of: :pipeline, class_name: 'GenericCommitStatus', partition_foreign_key: :partition_id # # NEW: has_many :all_jobs, ->(pipeline) { in_partition(pipeline) }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :current_jobs, ->(pipeline) { latest.in_partition(pipeline) }, class_name: 'CommitStatus', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :all_processable_jobs, ->(pipeline) { in_partition(pipeline) }, class_name: 'Ci::Processable', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :current_processable_jobs, ->(pipeline) { latest.in_partition(pipeline) }, class_name: 'Ci::Processable', foreign_key: :commit_id, inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :job_artifacts, through: :builds has_many :build_trace_chunks, class_name: 'Ci::BuildTraceChunk', through: :builds, source: :trace_chunks has_many :variables, ->(pipeline) { in_partition(pipeline) }, class_name: 'Ci::PipelineVariable', inverse_of: :pipeline, partition_foreign_key: :partition_id has_many :latest_builds, ->(pipeline) { in_partition(pipeline).latest.with_project_and_metadata }, foreign_key: :commit_id, inverse_of: :pipeline, class_name: 'Ci::Build' has_many :downloadable_artifacts, -> do not_expired.or(where_exists(Ci::Pipeline.artifacts_locked.where("#{Ci::Pipeline.quoted_table_name}.id = #{Ci::Build.quoted_table_name}.commit_id"))).downloadable.with_job end, through: :latest_builds, source: :job_artifacts has_many :latest_successful_jobs, ->(pipeline) { in_partition(pipeline).latest.success.with_project_and_metadata }, foreign_key: :commit_id, inverse_of: :pipeline, class_name: 'Ci::Processable' has_many :latest_finished_jobs, ->(pipeline) { in_partition(pipeline).latest.finished.with_project_and_metadata }, foreign_key: :commit_id, inverse_of: :pipeline, class_name: 'Ci::Processable' has_many :messages, class_name: 'Ci::PipelineMessage', inverse_of: :pipeline # Merge requests for which the current pipeline is running against # the merge request's latest commit. has_many :merge_requests_as_head_pipeline, foreign_key: :head_pipeline_id, class_name: 'MergeRequest', inverse_of: :head_pipeline has_many :pending_builds, ->(pipeline) { in_partition(pipeline).pending }, foreign_key: :commit_id, class_name: 'Ci::Build', inverse_of: :pipeline has_many :failed_builds, ->(pipeline) { in_partition(pipeline).latest.failed }, foreign_key: :commit_id, class_name: 'Ci::Build', inverse_of: :pipeline has_many :limited_failed_builds, ->(pipeline) { in_partition(pipeline).latest.failed.limit(COUNT_FAILED_JOBS_LIMIT) }, foreign_key: :commit_id, class_name: 'Ci::Build', inverse_of: :pipeline has_many :retryable_builds, ->(pipeline) { in_partition(pipeline).latest.failed_or_canceled.includes(:project) }, foreign_key: :commit_id, class_name: 'Ci::Build', inverse_of: :pipeline has_many :cancelable_statuses, ->(pipeline) { in_partition(pipeline).cancelable }, foreign_key: :commit_id, class_name: 'CommitStatus', inverse_of: :pipeline has_many :manual_actions, ->(pipeline) { in_partition(pipeline).latest.manual_actions.includes(:project) }, foreign_key: :commit_id, class_name: 'Ci::Processable', inverse_of: :pipeline has_many :scheduled_actions, ->(pipeline) { in_partition(pipeline).latest.scheduled_actions.includes(:project) }, foreign_key: :commit_id, class_name: 'Ci::Build', inverse_of: :pipeline has_many :auto_canceled_pipelines, class_name: 'Ci::Pipeline', foreign_key: :auto_canceled_by_id, inverse_of: :auto_canceled_by has_many :auto_canceled_jobs, class_name: 'CommitStatus', foreign_key: :auto_canceled_by_id, inverse_of: :auto_canceled_by has_many :sourced_pipelines, class_name: 'Ci::Sources::Pipeline', foreign_key: :source_pipeline_id, inverse_of: :source_pipeline has_one :source_pipeline, class_name: 'Ci::Sources::Pipeline', inverse_of: :pipeline has_one :chat_data, class_name: 'Ci::PipelineChatData' has_many :triggered_pipelines, through: :sourced_pipelines, source: :pipeline # Only includes direct and not nested children has_many :child_pipelines, -> { merge(Ci::Sources::Pipeline.same_project) }, through: :sourced_pipelines, source: :pipeline has_one :triggered_by_pipeline, through: :source_pipeline, source: :source_pipeline has_one :parent_pipeline, -> { merge(Ci::Sources::Pipeline.same_project) }, through: :source_pipeline, source: :source_pipeline has_one :source_job, through: :source_pipeline, source: :source_job has_one :source_bridge, through: :source_pipeline, source: :source_bridge has_one :pipeline_config, class_name: 'Ci::PipelineConfig', inverse_of: :pipeline has_one :pipeline_metadata, class_name: 'Ci::PipelineMetadata', inverse_of: :pipeline has_many :daily_build_group_report_results, class_name: 'Ci::DailyBuildGroupReportResult', foreign_key: :last_pipeline_id, inverse_of: :last_pipeline has_many :latest_builds_report_results, through: :latest_builds, source: :report_results has_many :pipeline_artifacts, class_name: 'Ci::PipelineArtifact', inverse_of: :pipeline, dependent: :destroy # rubocop:disable Cop/ActiveRecordDependent accepts_nested_attributes_for :variables, reject_if: :persisted? delegate :full_path, to: :project, prefix: true delegate :name, to: :pipeline_metadata, allow_nil: true validates :sha, presence: { unless: :importing? } validates :ref, presence: { unless: :importing? } validates :tag, inclusion: { in: [false], if: :merge_request? } validates :external_pull_request, presence: { if: :external_pull_request_event? } validates :external_pull_request, absence: { unless: :external_pull_request_event? } validates :tag, inclusion: { in: [false], if: :external_pull_request_event? } validates :status, presence: { unless: :importing? } validate :valid_commit_sha, unless: :importing? validates :source, exclusion: { in: %w[unknown], unless: :importing? }, on: :create validates :project, presence: true validates :ref, length: { maximum: MAX_REF_LENGTH }, if: :ref_changed? validates :sha, length: { maximum: MAX_SHA_LENGTH }, if: :sha_changed? validates :before_sha, length: { maximum: MAX_SHA_LENGTH }, if: :before_sha_changed? validates :source_sha, length: { maximum: MAX_SHA_LENGTH }, if: :source_sha_changed? validates :target_sha, length: { maximum: MAX_SHA_LENGTH }, if: :target_sha_changed? validates :yaml_errors, bytesize: { maximum: -> { YAML_ERRORS_MAX_LENGTH } }, if: :yaml_errors_changed? after_create :keep_around_commits, unless: :importing? after_commit :trigger_pipeline_status_change_subscription, if: :saved_change_to_status? after_commit :track_ci_pipeline_created_event, on: :create, if: :internal_pipeline? after_find :observe_age_in_minutes, unless: :importing? use_fast_destroy :job_artifacts use_fast_destroy :build_trace_chunks # We use `Enums::Ci::Pipeline.sources` here so that EE can more easily extend # this `Hash` with new values. enum :source, Enums::Ci::Pipeline.sources enum :config_source, Enums::Ci::Pipeline.config_sources # We use `Enums::Ci::Pipeline.failure_reasons` here so that EE can more easily # extend this `Hash` with new values. enum :failure_reason, Enums::Ci::Pipeline.failure_reasons enum :locked, { unlocked: 0, artifacts_locked: 1 } state_machine :status, initial: :created do event :enqueue do transition [:created, :manual, :waiting_for_resource, :preparing, :skipped, :scheduled] => :pending transition [:success, :failed, :canceling, :canceled] => :running # this is needed to ensure tests to be covered transition [:running] => :running transition [:waiting_for_callback] => :waiting_for_callback end event :request_resource do transition any - [:waiting_for_resource] => :waiting_for_resource end event :prepare do transition any - [:preparing] => :preparing end event :run do transition any - [:running] => :running end event :wait_for_callback do transition any - [:waiting_for_callback] => :waiting_for_callback end event :skip do transition any - [:skipped] => :skipped end event :drop do transition any - [:failed] => :failed end event :succeed do # A success pipeline can also be retried, for example; a pipeline with a failed manual job. # When retrying the pipeline, the status of the pipeline is not changed because the failed # manual job transitions to the `manual` status. # More info: https://gitlab.com/gitlab-org/gitlab/-/merge_requests/98967#note_1144718316 transition any => :success end event :start_cancel do transition any - [:canceling, :canceled] => :canceling end event :cancel do transition any - [:canceled] => :canceled end event :block do transition any - [:manual] => :manual end event :delay do transition any - [:scheduled] => :scheduled end # IMPORTANT # Do not add any operations to this state_machine # Create a separate worker for each new operation before_transition [:created, :waiting_for_resource, :preparing, :pending] => :running do |pipeline| pipeline.started_at ||= Time.current end before_transition any => [:success, :failed, :canceled] do |pipeline| pipeline.finished_at = Time.current pipeline.update_duration end before_transition any => [:manual] do |pipeline| pipeline.update_duration end before_transition canceled: any - [:canceled] do |pipeline| pipeline.auto_canceled_by = nil end before_transition any => :failed do |pipeline, transition| transition.args.first.try do |reason| pipeline.failure_reason = reason end end after_transition [:created, :waiting_for_resource, :preparing, :pending] => :running do |pipeline| pipeline.run_after_commit { PipelineMetricsWorker.perform_async(pipeline.id) } end after_transition any => [:success] do |pipeline| pipeline.run_after_commit { PipelineMetricsWorker.perform_async(pipeline.id) } end after_transition any => UNLOCKABLE_STATUSES do |pipeline| pipeline.run_after_commit do Ci::PipelineFinishedWorker.perform_async(pipeline.id) Ci::Refs::UnlockPreviousPipelinesWorker.perform_async(pipeline.ci_ref_id) end end after_transition [:created, :waiting_for_resource, :preparing, :pending, :running] => :success do |pipeline| # We wait a little bit to ensure that all Ci::BuildFinishedWorkers finish first # because this is where some metrics like code coverage is parsed and stored # in CI build records which the daily build metrics worker relies on. pipeline.run_after_commit { Ci::DailyBuildGroupReportResultsWorker.perform_in(10.minutes, pipeline.id) } end after_transition do |pipeline, transition| next if transition.loopback? pipeline.run_after_commit do unless pipeline.user&.blocked? Gitlab::AppLogger.info( message: "Enqueuing hooks for Pipeline #{pipeline.id}: #{pipeline.status}", class: self.class.name, pipeline_id: pipeline.id, project_id: pipeline.project_id, pipeline_status: pipeline.status) PipelineHooksWorker.perform_async(pipeline.id) end if pipeline.project.jira_subscription_exists? # Passing the seq-id ensures this is idempotent seq_id = ::Atlassian::JiraConnect::Client.generate_update_sequence_id ::JiraConnect::SyncBuildsWorker.perform_async(pipeline.id, seq_id) end Ci::ExpirePipelineCacheService.new.execute(pipeline) # rubocop: disable CodeReuse/ServiceClass end end after_transition any => ::Ci::Pipeline.completed_statuses do |pipeline| pipeline.run_after_commit do AutoMergeProcessWorker.perform_async({ 'pipeline_id' => self.id }) if pipeline.auto_devops_source? self.class.auto_devops_pipelines_completed_total.increment(status: pipeline.status) end end end after_transition any => ::Ci::Pipeline.completed_statuses do |pipeline| pipeline.run_after_commit do ::Ci::PipelineArtifacts::CoverageReportWorker.perform_async(pipeline.id) ::Ci::PipelineArtifacts::CreateQualityReportWorker.perform_async(pipeline.id) end end after_transition any => ::Ci::Pipeline.completed_statuses do |pipeline| next unless pipeline.bridge_waiting? pipeline.run_after_commit do ::Ci::PipelineBridgeStatusWorker.perform_async(pipeline.id) end end after_transition any => ::Ci::Pipeline.completed_statuses do |pipeline| pipeline.run_after_commit do ::Ci::TestFailureHistoryService.new(pipeline).async.perform_if_needed # rubocop: disable CodeReuse/ServiceClass end end after_transition any => ::Ci::Pipeline.completed_statuses do |pipeline| pipeline.run_after_commit do ::Ci::JobArtifacts::TrackArtifactReportWorker.perform_async(pipeline.id) track_internal_event( 'completed_pipeline_execution', project: pipeline.project, user: pipeline.user, additional_properties: { label: pipeline.status } ) end end # This needs to be kept in sync with `Ci::PipelineRef#should_delete?` after_transition any => ::Ci::Pipeline.stopped_statuses do |pipeline| pipeline.run_after_commit do Ci::Pipelines::ClearPersistentRefService.new(pipeline).execute # rubocop: disable CodeReuse/ServiceClass end end after_transition any => [:success, :failed] do |pipeline| ref_status = pipeline.ci_ref&.update_status_by!(pipeline) pipeline.run_after_commit do # We don't send notifications for a pipeline dropped due to the # user been blocked. unless pipeline.user&.blocked? PipelineNotificationWorker .perform_async(pipeline.id, 'ref_status' => ref_status&.to_s) end end end after_transition any => [:failed] do |pipeline| pipeline.run_after_commit do ::Gitlab::Ci::Pipeline::Metrics.pipeline_failure_reason_counter.increment(reason: pipeline.failure_reason) AutoDevops::DisableWorker.perform_async(pipeline.id) if pipeline.auto_devops_source? end end after_transition any => [:running, *::Ci::Pipeline.completed_statuses] do |pipeline| project = pipeline&.project next unless project pipeline.run_after_commit do next if pipeline.child? pipeline.all_merge_requests.opened.each do |merge_request| GraphqlTriggers.merge_request_merge_status_updated(merge_request) end end end end scope :with_unlockable_status, -> { with_status(*UNLOCKABLE_STATUSES) } scope :internal, -> { where(source: internal_sources) } scope :no_tag, -> { where(tag: false) } scope :no_child, -> { where.not(source: :parent_pipeline) } scope :ci_sources, -> { where(source: Enums::Ci::Pipeline.ci_sources.values) } scope :ci_branch_sources, -> { where(source: Enums::Ci::Pipeline.ci_branch_sources.values) } scope :ci_and_parent_sources, -> { where(source: Enums::Ci::Pipeline.ci_and_parent_sources.values) } scope :ci_and_security_orchestration_sources, -> do where(source: Enums::Ci::Pipeline.ci_and_security_orchestration_sources.values) end scope :for_user, ->(user) { where(user: user) } scope :for_sha, ->(sha) { where(sha: sha) } scope :where_not_sha, ->(sha) { where.not(sha: sha) } scope :for_source_sha, ->(source_sha) { where(source_sha: source_sha) } scope :for_sha_or_source_sha, ->(sha) { for_sha(sha).or(for_source_sha(sha)) } scope :for_ref, ->(ref) { where(ref: ref) } scope :for_branch, ->(branch) { for_ref(branch).where(tag: false) } scope :for_iid, ->(iid) { where(iid: iid) } scope :for_project, ->(project_id) { where(project_id: project_id) } scope :for_name, ->(name) do name_column = Ci::PipelineMetadata.arel_table[:name] joins(:pipeline_metadata).where(name_column.eq(name)) end scope :for_status, ->(status) { where(status: status) } scope :created_after, ->(time) { where(arel_table[:created_at].gt(time)) } scope :created_before, ->(time) { where(arel_table[:created_at].lt(time)) } scope :created_before_id, ->(id) { where(arel_table[:id].lt(id)) } scope :before_pipeline, ->(pipeline) { created_before_id(pipeline.id).outside_pipeline_family(pipeline) } scope :with_pipeline_source, ->(source) { where(source: source) } scope :preload_pipeline_metadata, -> { preload(:pipeline_metadata) } scope :not_ref_protected, -> { where("#{quoted_table_name}.protected IS NOT true") } scope :outside_pipeline_family, ->(pipeline) do where.not(id: pipeline.same_family_pipeline_ids) end scope :with_reports, ->(reports_scope) do where_exists(Ci::Build.latest.scoped_pipeline.with_artifacts(reports_scope)) end scope :conservative_interruptible, -> do where_not_exists( Ci::Build.scoped_pipeline.with_status(STARTED_STATUSES).not_interruptible ) end # Returns the pipelines that associated with the given merge request. # In general, please use `Ci::PipelinesForMergeRequestFinder` instead, # for checking permission of the actor. scope :triggered_by_merge_request, ->(merge_request) do where( source: :merge_request_event, merge_request: merge_request, project: [merge_request.source_project, merge_request.target_project] ) end scope :order_id_asc, -> { order(id: :asc) } scope :order_id_desc, -> { order(id: :desc) } # Returns the pipelines in descending order (= newest first), optionally # limited to a number of references. # # ref - The name (or names) of the branch(es)/tag(s) to limit the list of # pipelines to. # sha - The commit SHA (or multiple SHAs) to limit the list of pipelines to. # limit - Number of pipelines to return. Chaining with sampling methods (#pick, #take) # will cause unnecessary subqueries. def self.newest_first(ref: nil, sha: nil, limit: nil, source: nil) relation = order(id: :desc) relation = relation.where(ref: ref) if ref relation = relation.where(sha: sha) if sha if source && Feature.enabled?(:source_filter_pipelines, :current_request) relation = relation.where(source: source) end if limit ids = relation.limit(limit).select(:id) relation = relation.where(id: ids) end relation end def self.latest_status(ref = nil) newest_first(ref: ref).pick(:status) end def self.latest_successful_for_ref(ref) newest_first(ref: ref).success.take end def self.latest_successful_for_sha(sha) newest_first(sha: sha).success.take end def self.latest_successful_for_refs(refs) return Ci::Pipeline.none if refs.empty? refs_values = refs.map { |ref| "(#{connection.quote(ref)})" }.join(",") query = Arel.sql(sanitize_sql_array(["refs_values.ref = #{quoted_table_name}.ref"])) join_query = success.where(query).order(id: :desc).limit(1) Ci::Pipeline .from("(VALUES #{refs_values}) refs_values (ref)") .joins("INNER JOIN LATERAL (#{join_query.to_sql}) #{quoted_table_name} ON TRUE") .index_by(&:ref) end def self.latest_running_for_ref(ref) newest_first(ref: ref).running.take end def self.latest_failed_for_ref(ref) newest_first(ref: ref).failed.take end def self.jobs_count_in_alive_pipelines created_after(24.hours.ago).alive.joins(:statuses).count end def self.builds_count_in_alive_pipelines created_after(24.hours.ago).alive.joins(:builds).count end # Returns a Hash containing the latest pipeline for every given # commit. # # The keys of this Hash are the commit SHAs, the values the pipelines. # # commits - The list of commit SHAs to get the pipelines for. # ref - The ref to scope the data to (e.g. "master"). If the ref is not # given we simply get the latest pipelines for the commits, regardless # of what refs the pipelines belong to. def self.latest_pipeline_per_commit(commits, ref = nil) sql = select('DISTINCT ON (sha) *') .where(sha: commits) .order(:sha, id: :desc) sql = sql.where(ref: ref) if ref sql.index_by(&:sha) end def self.latest_successful_ids_per_project success.group(:project_id).select('max(id) as id') end def self.last_finished_for_ref_id(ci_ref_id) where(ci_ref_id: ci_ref_id).ci_sources.finished.order(id: :desc).select(:id).take end def self.truncate_sha(sha) sha[0...8] end def self.total_duration where.not(duration: nil).sum(:duration) end def self.internal_sources sources.reject { |source| source == "external" }.values end def self.bridgeable_statuses ::Ci::Pipeline::AVAILABLE_STATUSES - %w[created waiting_for_resource waiting_for_callback preparing pending] end def self.auto_devops_pipelines_completed_total @auto_devops_pipelines_completed_total ||= Gitlab::Metrics.counter(:auto_devops_pipelines_completed_total, 'Number of completed auto devops pipelines') end def self.current_partition_value Gitlab::SafeRequestStore.fetch(:ci_current_partition_value) do Ci::Partition.current&.id || Ci::Partition::INITIAL_PARTITION_VALUE end end def self.object_hierarchy(relation, options = {}) ::Gitlab::Ci::PipelineObjectHierarchy.new(relation, options: options) end def self.internal_id_scope_usage :ci_pipelines end def trigger_pipeline_status_change_subscription GraphqlTriggers.ci_pipeline_status_updated(self) end def uses_needs? processables.where(scheduling_type: :dag).any? end def stages_count stages.count end def total_size statuses.count(:id) end def tags_count Ci::BuildTag.in_partition(self).where(build: builds).count end def distinct_tags_count Ci::BuildTag.in_partition(self).where(build: builds).count('distinct(tag_id)') end def stages_names stages.order(:position).pluck(:name) end def ref_exists? project.repository.ref_exists?(git_ref) rescue Gitlab::Git::Repository::NoRepository false end def triggered_pipelines_with_preloads triggered_pipelines.preload(:source_job) end def valid_commit_sha self.errors.add(:sha, "can't be 00000000 (branch removal)") if Gitlab::Git.blank_ref?(self.sha) end def git_author_name strong_memoize(:git_author_name) do commit.try(:author_name) end end def git_author_email strong_memoize(:git_author_email) do commit.try(:author_email) end end def git_author_full_text strong_memoize(:git_author_full_text) do commit.try(:author_full_text) end end def git_commit_message strong_memoize(:git_commit_message) do commit.try(:message) end end def git_commit_title strong_memoize(:git_commit_title) do commit.try(:title) end end def git_commit_full_title strong_memoize(:git_commit_full_title) do commit.try(:full_title) end end def git_commit_description strong_memoize(:git_commit_description) do commit.try(:description) end end def git_commit_timestamp strong_memoize(:git_commit_timestamp) do commit.try(:timestamp) end end def before_sha super || project.repository.blank_ref end def short_sha Ci::Pipeline.truncate_sha(sha) end # NOTE: This is loaded lazily and will never be nil, even if the commit # cannot be found. # # Use constructs like: `pipeline.commit.present?` def commit @commit ||= Commit.lazy(project, sha) end def stuck? pending_builds.any?(&:stuck?) end def retryable? return false if archived? retryable_builds.any? end def archived? archive_builds_older_than = Gitlab::CurrentSettings.current_application_settings.archive_builds_older_than archive_builds_older_than.present? && created_at < archive_builds_older_than end def cancelable? cancelable_statuses.any? && internal_pipeline? end def auto_canceled? canceled? && auto_canceled_by_id? end # rubocop: disable CodeReuse/ServiceClass def retry_failed(current_user) Ci::RetryPipelineService.new(project, current_user) .execute(self) end # rubocop: enable CodeReuse/ServiceClass def lazy_ref_commit BatchLoader.for(ref).batch(key: project.id) do |refs, loader| next unless project.repository_exists? project.repository.list_commits_by_ref_name(refs).then do |commits| commits.each { |key, commit| loader.call(key, commits[key]) } end end end def latest? return false unless git_ref && commit.present? return false if lazy_ref_commit.nil? lazy_ref_commit.id == commit.id end def retried @retried ||= (statuses.order(id: :desc) - latest_statuses) end def coverage coverage_array = latest_statuses.filter_map(&:coverage) coverage_array.sum / coverage_array.size if coverage_array.size >= 1 end def update_builds_coverage builds.with_coverage_regex.without_coverage.each(&:update_coverage) end def batch_lookup_report_artifact_for_file_type(file_type) batch_lookup_report_artifact_for_file_types([file_type]) end def batch_lookup_report_artifact_for_file_types(file_types) file_types_to_search = [] file_types.each { |file_type| file_types_to_search.append(*::Ci::JobArtifact.associated_file_types_for(file_type.to_s)) } latest_report_artifacts .values_at(*file_types_to_search.uniq) .flatten .compact .last end # This batch loads the latest reports for each CI job artifact # type (e.g. sast, dast, etc.) in a single SQL query to eliminate # the need to do N different `job_artifacts.where(file_type: # X).last` calls. # # Return a hash of file type => array of 1 job artifact def latest_report_artifacts ::Gitlab::SafeRequestStore.fetch("pipeline:#{self.id}:latest_report_artifacts") do ::Ci::JobArtifact.where( id: job_artifacts.all_reports .select("max(#{Ci::JobArtifact.quoted_table_name}.id) as id") .group(:file_type) ) .preload(:job) .group_by(&:file_type) end end def has_kubernetes_active? strong_memoize(:has_kubernetes_active) do project.deployment_platform&.active? end end def freeze_period? strong_memoize(:freeze_period) do project.freeze_periods.any?(&:active?) end end def has_warnings? number_of_warnings > 0 end def number_of_warnings BatchLoader.for(id).batch(default_value: 0) do |pipeline_ids, loader| ::CommitStatus.where(commit_id: pipeline_ids) .latest .failed_but_allowed .group(:commit_id) .count .each { |id, amount| loader.call(id, amount) } end end def needs_processing? statuses .where(processed: [false, nil]) .latest .exists? end def has_yaml_errors? yaml_errors.present? end def add_error_message(content) add_message(:error, content) end def add_warning_message(content) add_message(:warning, content) end # Like #drop!, but does not persist the pipeline nor trigger any state # machine callbacks. def set_failed(failure_reason) self.failure_reason = failure_reason.to_s self.status = 'failed' end # We can't use `messages.error` scope here because messages should also be # read when the pipeline is not persisted. Using the scope will return no # results as it would query persisted data. def error_messages messages.select(&:error?) end def warning_messages(limit: nil) messages.select(&:warning?).tap do |warnings| break warnings.take(limit) if limit end end # Manually set the notes for a Ci::Pipeline # There is no ActiveRecord relation between Ci::Pipeline and notes # as they are related to a commit sha. This method helps importing # them using the +Gitlab::ImportExport::Project::RelationFactory+ class. def notes=(notes_to_save) notes_to_save.reject! do |note_to_save| notes.any? do |note| [note_to_save.note, note_to_save.created_at.to_i] == [note.note, note.created_at.to_i] end end notes_to_save.each do |note| note[:id] = nil note[:commit_id] = sha note[:noteable_id] = self['id'] note.save! end end def notes project.notes.for_commit_id(sha) end # rubocop: disable Metrics/CyclomaticComplexity -- breaking apart hurts readability def set_status(new_status) retry_optimistic_lock(self, name: 'ci_pipeline_set_status') do case new_status when 'created' then nil when 'waiting_for_resource' then request_resource when 'preparing' then prepare when 'waiting_for_callback' then wait_for_callback when 'pending' then enqueue when 'running' then run when 'success' then succeed when 'failed' then drop when 'canceling' then start_cancel when 'canceled' then cancel when 'skipped' then skip when 'manual' then block when 'scheduled' then delay else raise Ci::HasStatus::UnknownStatusError, "Unknown status `#{new_status}`" end end end # rubocop: enable Metrics/CyclomaticComplexity def protected_ref? strong_memoize(:protected_ref) { project.protected_for?(git_ref) } end def variables_builder @variables_builder ||= ::Gitlab::Ci::Variables::Builder.new(self) end def persisted_variables Gitlab::Ci::Variables::Collection.new.tap do |variables| break variables unless persisted? variables.append(key: 'CI_PIPELINE_ID', value: id.to_s) variables.append(key: 'CI_PIPELINE_URL', value: Gitlab::Routing.url_helpers.project_pipeline_url(project, self)) end end def queued_duration return unless started_at seconds = (started_at - created_at).to_i seconds unless seconds == 0 end def update_duration return unless started_at self.duration = Gitlab::Ci::Pipeline::Duration.from_pipeline(self) end # All the merge requests for which the current pipeline runs/ran against def all_merge_requests @all_merge_requests ||= if merge_request? MergeRequest.where(id: merge_request_id) else MergeRequest.where(source_project_id: project_id, source_branch: ref) .by_commit_sha(sha) end end def all_merge_requests_by_recency all_merge_requests.order(id: :desc) end # This returns a list of MRs that point # to the same source project/branch def related_merge_requests if merge_request? # We look for all other MRs that this branch might be pointing to MergeRequest.where( source_project_id: merge_request.source_project_id, source_branch: merge_request.source_branch) else MergeRequest.where( source_project_id: project_id, source_branch: ref) end end # We cannot use `all_merge_requests`, due to race condition # This returns a list of at most 4 open MRs def open_merge_requests_refs strong_memoize(:open_merge_requests_refs) do # We ensure that triggering user can actually read the pipeline related_merge_requests .opened .limit(MAX_OPEN_MERGE_REQUESTS_REFS) .order(id: :desc) .preload(:target_project) .select { |mr| can?(user, :read_merge_request, mr) } .map { |mr| mr.to_reference(project, full: true) } end end def same_family_pipeline_ids ::Gitlab::Ci::PipelineObjectHierarchy.new( self.class.default_scoped.where(id: root_ancestor), options: { project_condition: :same } ).base_and_descendants.select(:id) end def build_with_artifacts_in_self_and_project_descendants(name) builds_in_self_and_project_descendants .ordered_by_pipeline # find job in hierarchical order .with_downloadable_artifacts .find_by_name(name) end def builds_in_self_and_project_descendants Ci::Build.in_partition(self).latest.where(pipeline: self_and_project_descendants) end def bridges_in_self_and_project_descendants Ci::Bridge.in_partition(self).latest.where(pipeline: self_and_project_descendants) end def jobs_in_self_and_project_descendants Ci::Processable.in_partition(self).latest.where(pipeline: self_and_project_descendants) end def environments_in_self_and_project_descendants(deployment_status: nil) # We limit to 100 unique environments for application safety. # See: https://gitlab.com/gitlab-org/gitlab/-/issues/340781#note_699114700 expanded_environment_names = jobs_in_self_and_project_descendants.joins(:metadata) .where.not(Ci::BuildMetadata.table_name => { expanded_environment_name: nil }) .distinct("#{Ci::BuildMetadata.quoted_table_name}.expanded_environment_name") .limit(100) .pluck(:expanded_environment_name) Environment.where(project: project, name: expanded_environment_names).with_deployment(sha, status: deployment_status) end # With multi-project and parent-child pipelines def self_and_upstreams object_hierarchy.base_and_ancestors end # With multi-project and parent-child pipelines def self_and_downstreams object_hierarchy.base_and_descendants end # With multi-project and parent-child pipelines def upstream_and_all_downstreams object_hierarchy.all_objects end # With only parent-child pipelines def self_and_project_ancestors object_hierarchy(project_condition: :same).base_and_ancestors end # With only parent-child pipelines def self_and_project_descendants object_hierarchy(project_condition: :same).base_and_descendants end # With only parent-child pipelines def all_child_pipelines object_hierarchy(project_condition: :same).descendants end def self_and_project_descendants_complete? self_and_project_descendants.all?(&:complete?) end # Follow the parent-child relationships and return the top-level parent def root_ancestor return self unless child? object_hierarchy(project_condition: :same) .base_and_ancestors(hierarchy_order: :desc) .first end # Follow the upstream pipeline relationships, regardless of multi-project or # parent-child, and return the top-level ancestor. def upstream_root @upstream_root ||= object_hierarchy.base_and_ancestors(hierarchy_order: :desc).first end # Applies to all parent-child and multi-project pipelines def complete_hierarchy_count upstream_root.self_and_downstreams.count end def bridge_triggered? source_bridge.present? end def bridge_waiting? source_bridge&.dependent? end def child? parent_pipeline? && # child pipelines have `parent_pipeline` source parent_pipeline.present? end def parent? child_pipelines.exists? end def created_successfully? persisted? && failure_reason.blank? end def filtered_as_empty? filtered_by_rules? || filtered_by_workflow_rules? end def detailed_status(current_user) Gitlab::Ci::Status::Pipeline::Factory .new(self.present, current_user) .fabricate! end def find_job_with_archive_artifacts(name) builds.latest.with_downloadable_artifacts.find_by_name(name) end def latest_builds_with_artifacts # We purposely cast the builds to an Array here. Because we always use the # rows if there are more than 0 this prevents us from having to run two # queries: one to get the count and one to get the rows. @latest_builds_with_artifacts ||= builds.latest.with_artifacts_not_expired.to_a end def latest_report_builds(reports_scope = ::Ci::JobArtifact.all_reports) builds.latest.with_artifacts(reports_scope) end def latest_test_report_builds latest_report_builds(Ci::JobArtifact.of_report_type(:test)).preload(:project, :metadata, job_artifacts: :artifact_report) end def latest_report_builds_in_self_and_project_descendants(reports_scope = ::Ci::JobArtifact.all_reports) builds_in_self_and_project_descendants.with_artifacts(reports_scope) end def builds_with_coverage builds.latest.with_coverage end def builds_with_failed_tests(limit: nil) latest_test_report_builds.failed.limit(limit) end def has_reports?(reports_scope) latest_report_builds(reports_scope).exists? end def complete_and_has_reports?(reports_scope) complete? && has_reports?(reports_scope) end def complete_and_has_self_or_descendant_reports?(reports_scope) complete? && latest_report_builds_in_self_and_project_descendants(reports_scope).exists? end def complete_or_manual_and_has_reports?(reports_scope) complete_or_manual? && has_reports?(reports_scope) end def has_coverage_reports? pipeline_artifacts&.report_exists?(:code_coverage) end def has_codequality_mr_diff_report? pipeline_artifacts&.report_exists?(:code_quality_mr_diff) end def can_generate_codequality_reports? complete_and_has_reports?(Ci::JobArtifact.of_report_type(:codequality)) end def test_report_summary strong_memoize(:test_report_summary) do Gitlab::Ci::Reports::TestReportSummary.new(latest_builds_report_results) end end def test_reports Gitlab::Ci::Reports::TestReport.new.tap do |test_reports| latest_test_report_builds.find_each do |build| build.collect_test_reports!(test_reports) end end end def accessibility_reports Gitlab::Ci::Reports::AccessibilityReports.new.tap do |accessibility_reports| latest_report_builds(Ci::JobArtifact.of_report_type(:accessibility)).each do |build| build.collect_accessibility_reports!(accessibility_reports) end end end def codequality_reports Gitlab::Ci::Reports::CodequalityReports.new.tap do |codequality_reports| latest_report_builds(Ci::JobArtifact.of_report_type(:codequality)).each do |build| build.collect_codequality_reports!(codequality_reports) end end end def terraform_reports if Feature.enabled?(:show_child_reports_in_mr_page, project) ::Gitlab::Ci::Reports::TerraformReports.new.tap do |terraform_reports| latest_report_builds_in_self_and_project_descendants(::Ci::JobArtifact.of_report_type(:terraform)).each do |build| build.collect_terraform_reports!(terraform_reports) end end else ::Gitlab::Ci::Reports::TerraformReports.new.tap do |terraform_reports| latest_report_builds(::Ci::JobArtifact.of_report_type(:terraform)).each do |build| build.collect_terraform_reports!(terraform_reports) end end end end def has_archive_artifacts? complete? && builds.latest.with_existing_job_artifacts(Ci::JobArtifact.archive.or(Ci::JobArtifact.metadata)).exists? end def has_exposed_artifacts? complete? && builds.latest.with_exposed_artifacts.exists? end def has_erasable_artifacts? complete? && builds.latest.with_erasable_artifacts.exists? end def branch_updated? strong_memoize(:branch_updated) do push_details.branch_updated? end end # Returns the modified paths. # # The returned value is # * Array: List of modified paths that should be evaluated # * nil: Modified path can not be evaluated def modified_paths strong_memoize(:modified_paths) do if merge_request? merge_request.modified_paths elsif branch_updated? push_details.modified_paths elsif external_pull_request? external_pull_request.modified_paths end end end def modified_paths_since(compare_to_sha) strong_memoize_with(:modified_paths_since, compare_to_sha) do project.repository.diff_stats(project.repository.merge_base(compare_to_sha, sha), sha).paths end end def all_worktree_paths strong_memoize(:all_worktree_paths) do project.repository.ls_files(sha) end end def top_level_worktree_paths strong_memoize(:top_level_worktree_paths) do project.repository.tree(sha).blobs.map(&:path) end end def default_branch? ref == project.default_branch end def merge_request? merge_request_id.present? && merge_request.present? end def external_pull_request? external_pull_request_id.present? end def detached_merge_request_pipeline? merge_request? && target_sha.nil? end def legacy_detached_merge_request_pipeline? detached_merge_request_pipeline? && !merge_request_ref? end def merged_result_pipeline? merge_request? && target_sha.present? end def tag_pipeline? tag? end def type if merge_train_pipeline? 'merge_train' elsif merged_result_pipeline? 'merged_result' elsif merge_request? 'merge_request' elsif tag_pipeline? 'tag' else 'branch' end end def merge_request_ref? MergeRequest.merge_request_ref?(ref) end def matches_sha_or_source_sha?(sha) self.sha == sha || self.source_sha == sha end def triggered_by?(current_user) user == current_user end def source_ref if merge_request? merge_request.source_branch else ref end end def source_ref_slug Gitlab::Utils.slugify(source_ref.to_s) end def stage(name) stages.find_by(name: name) end def full_error_messages errors ? errors.full_messages.to_sentence : "" end def merge_request_event_type return unless merge_request? strong_memoize(:merge_request_event_type) do if merged_result_pipeline? :merged_result elsif detached_merge_request_pipeline? :detached end end end def persistent_ref @persistent_ref ||= PersistentRef.new(pipeline: self) end def dangling? Enums::Ci::Pipeline.dangling_sources.key?(source.to_sym) end def only_workload_variables? Enums::Ci::Pipeline.workload_sources.key?(source.to_sym) end def source_ref_path if branch? || merge_request? Gitlab::Git::BRANCH_REF_PREFIX + source_ref.to_s elsif tag? Gitlab::Git::TAG_REF_PREFIX + source_ref.to_s end end # Set scheduling type of processables if they were created before scheduling_type # data was deployed (https://gitlab.com/gitlab-org/gitlab/-/merge_requests/22246). def ensure_scheduling_type! processables.populate_scheduling_type! end def ensure_ci_ref! self.ci_ref = Ci::Ref.ensure_for(self) end def ensure_persistent_ref Pipelines::CreatePersistentRefService.new(self).execute # rubocop: disable CodeReuse/ServiceClass end # For dependent bridge jobs we reset the upstream bridge recursively # to reflect that a downstream pipeline is running again def reset_source_bridge!(current_user) # break recursion when no source_pipeline bridge (first upstream pipeline) return unless bridge_waiting? return unless current_user.can?(:update_pipeline, source_bridge.pipeline) # Before enqueuing the trigger job again, its status must be one of :created, :skipped, :manual, and :scheduled. # Also, we use `skip_pipeline_processing` to prevent processing the pipeline to avoid redundant process. source_bridge.created!(current_user, skip_pipeline_processing: true) Ci::EnqueueJobService.new(source_bridge, current_user: current_user).execute # rubocop:disable CodeReuse/ServiceClass end # EE-only def merge_train_pipeline? false end def build_matchers self.builds.latest.build_matchers(project) end def cluster_agent_authorizations strong_memoize(:cluster_agent_authorizations) do ::Clusters::Agents::Authorizations::CiAccess::Finder.new(project).execute end end def has_test_reports? strong_memoize(:has_test_reports) do has_reports?(::Ci::JobArtifact.of_report_type(:test)) end end def age_in_minutes return 0 unless persisted? raise ArgumentError, 'pipeline not fully loaded' unless has_attribute?(:created_at) return 0 unless created_at (Time.current - created_at).ceil / 60 end def merge_request_diff return unless merge_request? merge_request.merge_request_diff_for(merge_request_diff_sha) end def auto_cancel_on_job_failure pipeline_metadata&.auto_cancel_on_job_failure || 'none' end def auto_cancel_on_new_commit pipeline_metadata&.auto_cancel_on_new_commit || 'conservative' end def cancel_async_on_job_failure case auto_cancel_on_job_failure when 'none' # no-op when 'all' ::Ci::UserCancelPipelineWorker.perform_async(id, id, user.id) else raise ArgumentError, "Unknown auto_cancel_on_job_failure value: #{auto_cancel_on_job_failure}" end end private def add_message(severity, content) messages.build(severity: severity, content: content, project_id: project_id) end def merge_request_diff_sha return unless merge_request? if merged_result_pipeline? source_sha else sha end end def push_details strong_memoize(:push_details) do Gitlab::Git::Push.new(project, before_sha, sha, git_ref) end end def git_ref strong_memoize(:git_ref) do if merge_request? ## # In the future, we're going to change this ref to # merge request's merged reference, such as "refs/merge-requests/:iid/merge". # In order to do that, we have to update GitLab-Runner's source pulling # logic. # See https://gitlab.com/gitlab-org/gitlab-runner/merge_requests/1092 Gitlab::Git::BRANCH_REF_PREFIX + ref.to_s else super end end end def keep_around_commits return unless project project.repository.keep_around(self.sha, self.before_sha, source: self.class.name) end def observe_age_in_minutes return unless age_metric_enabled? return unless persisted? && has_attribute?(:created_at) ::Gitlab::Ci::Pipeline::Metrics .pipeline_age_histogram .observe({}, age_in_minutes) end def age_metric_enabled? ::Gitlab::SafeRequestStore.fetch(:age_metric_enabled) do ::Feature.enabled?(:ci_pipeline_age_histogram, type: :ops) end end # Without using `unscoped`, caller scope is also included into the query. # Using `unscoped` here will be redundant after Rails 6.1 def object_hierarchy(options = {}) ::Gitlab::Ci::PipelineObjectHierarchy .new(self.class.unscoped.where(id: id), options: options) end def internal_pipeline? source != "external" end def track_ci_pipeline_created_event Gitlab::InternalEvents.track_event( 'create_ci_internal_pipeline', project: project, user: user, additional_properties: { label: source, property: config_source } ) end end end Ci::Pipeline.prepend_mod_with('Ci::Pipeline')