Check against a list of known-broken specs in ci generate (#22690)

* Strip leading slash from S3 key in url_exists()

* Check against a list of known-broken specs in `ci generate`
This commit is contained in:
Zack Galbreath 2021-04-02 19:40:47 -04:00 committed by GitHub
parent 69d123a1a0
commit 7cc2db1b4e
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
7 changed files with 124 additions and 16 deletions

View file

@ -240,6 +240,13 @@ takes a boolean and determines whether the pipeline uses artifacts to store and
pass along the buildcaches from one stage to the next (the default if you don't pass along the buildcaches from one stage to the next (the default if you don't
provide this option is ``False``). provide this option is ``False``).
The optional ``broken-specs-url`` key tells Spack to check against a list of
specs that are known to be currently broken in ``develop``. If any such specs
are found, the ``spack ci generate`` command will fail with an error message
informing the user what broken specs were encountered. This allows the pipeline
to fail early and avoid wasting compute resources attempting to build packages
that will not succeed.
The optional ``cdash`` section provides information that will be used by the The optional ``cdash`` section provides information that will be used by the
``spack ci generate`` command (invoked by ``spack ci start``) for reporting ``spack ci generate`` command (invoked by ``spack ci start``) for reporting
to CDash. All the jobs generated from this environment will belong to a to CDash. All the jobs generated from this environment will belong to a
@ -554,7 +561,7 @@ provision of a custom ``script`` section. The reason for this is to run
Now imagine you have long pipelines with many specs to be built, and you Now imagine you have long pipelines with many specs to be built, and you
are pointing to a spack repository and branch that has a tendency to change are pointing to a spack repository and branch that has a tendency to change
frequently, such as the main repo and it's ``develop`` branch. If each child frequently, such as the main repo and its ``develop`` branch. If each child
job checks out the ``develop`` branch, that could result in some jobs running job checks out the ``develop`` branch, that could result in some jobs running
with one SHA of spack, while later jobs run with another. To help avoid this with one SHA of spack, while later jobs run with another. To help avoid this
issue, the pipeline generation process saves global variables called issue, the pipeline generation process saves global variables called

View file

@ -575,6 +575,13 @@ def generate_gitlab_ci_yaml(env, print_summary, output_file, prune_dag=False,
ci_mirrors = yaml_root['mirrors'] ci_mirrors = yaml_root['mirrors']
mirror_urls = [url for url in ci_mirrors.values()] mirror_urls = [url for url in ci_mirrors.values()]
# Check for a list of "known broken" specs that we should not bother
# trying to build.
broken_specs_url = ''
known_broken_specs_encountered = []
if 'broken-specs-url' in gitlab_ci:
broken_specs_url = gitlab_ci['broken-specs-url']
enable_artifacts_buildcache = False enable_artifacts_buildcache = False
if 'enable-artifacts-buildcache' in gitlab_ci: if 'enable-artifacts-buildcache' in gitlab_ci:
enable_artifacts_buildcache = gitlab_ci['enable-artifacts-buildcache'] enable_artifacts_buildcache = gitlab_ci['enable-artifacts-buildcache']
@ -665,6 +672,14 @@ def generate_gitlab_ci_yaml(env, print_summary, output_file, prune_dag=False,
pkg_name = pkg_name_from_spec_label(spec_label) pkg_name = pkg_name_from_spec_label(spec_label)
release_spec = root_spec[pkg_name] release_spec = root_spec[pkg_name]
# Check if this spec is in our list of known failures.
if broken_specs_url:
full_hash = release_spec.full_hash()
broken_spec_path = url_util.join(broken_specs_url, full_hash)
if web_util.url_exists(broken_spec_path):
known_broken_specs_encountered.append('{0} ({1})'.format(
release_spec, full_hash))
runner_attribs = find_matching_config( runner_attribs = find_matching_config(
release_spec, gitlab_ci) release_spec, gitlab_ci)
@ -1029,6 +1044,14 @@ def generate_gitlab_ci_yaml(env, print_summary, output_file, prune_dag=False,
sorted_output = {'no-specs-to-rebuild': noop_job} sorted_output = {'no-specs-to-rebuild': noop_job}
if known_broken_specs_encountered:
error_msg = (
'Pipeline generation failed due to the presence of the '
'following specs that are known to be broken in develop:\n')
for broken_spec in known_broken_specs_encountered:
error_msg += '* {0}\n'.format(broken_spec)
tty.die(error_msg)
with open(output_file, 'w') as outf: with open(output_file, 'w') as outf:
outf.write(syaml.dump_config(sorted_output, default_flow_style=True)) outf.write(syaml.dump_config(sorted_output, default_flow_style=True))

View file

@ -111,6 +111,7 @@
}, },
'service-job-attributes': runner_selector_schema, 'service-job-attributes': runner_selector_schema,
'rebuild-index': {'type': 'boolean'}, 'rebuild-index': {'type': 'boolean'},
'broken-specs-url': {'type': 'string'},
}, },
) )

View file

@ -1404,3 +1404,55 @@ def test_ci_generate_temp_storage_url(tmpdir, mutable_mock_env_path,
# Cleanup job should be 2nd to last, just before rebuild-index # Cleanup job should be 2nd to last, just before rebuild-index
assert('stage' in cleanup_job) assert('stage' in cleanup_job)
assert(cleanup_job['stage'] == stages[-2]) assert(cleanup_job['stage'] == stages[-2])
def test_ci_generate_read_broken_specs_url(tmpdir, mutable_mock_env_path,
env_deactivate, install_mockery,
mock_packages, monkeypatch):
"""Verify that `broken-specs-url` works as intended"""
spec_a = Spec('a')
spec_a.concretize()
a_full_hash = spec_a.full_hash()
spec_flattendeps = Spec('flatten-deps')
spec_flattendeps.concretize()
flattendeps_full_hash = spec_flattendeps.full_hash()
# Mark 'a' as broken (but not 'flatten-deps')
broken_spec_a_path = str(tmpdir.join(a_full_hash))
with open(broken_spec_a_path, 'w') as bsf:
bsf.write('')
# Test that `spack ci generate` notices this broken spec and fails.
filename = str(tmpdir.join('spack.yaml'))
with open(filename, 'w') as f:
f.write("""\
spack:
specs:
- flatten-deps
- a
mirrors:
some-mirror: https://my.fake.mirror
gitlab-ci:
broken-specs-url: "{0}"
mappings:
- match:
- archive-files
runner-attributes:
tags:
- donotcare
image: donotcare
""".format(tmpdir.strpath))
with tmpdir.as_cwd():
env_cmd('create', 'test', './spack.yaml')
with ev.read('test'):
# Check output of the 'generate' subcommand
output = ci_cmd('generate', output=str, fail_on_error=False)
assert('known to be broken' in output)
ex = '({0})'.format(a_full_hash)
assert(ex in output)
ex = '({0})'.format(flattendeps_full_hash)
assert(ex not in output)

View file

@ -215,6 +215,11 @@ def paginate(self, *args, **kwargs):
return MockPages() return MockPages()
class MockClientError(Exception):
def __init__(self):
self.response = {'Error': {'Code': 'NoSuchKey'}}
class MockS3Client(object): class MockS3Client(object):
def get_paginator(self, *args, **kwargs): def get_paginator(self, *args, **kwargs):
return MockPaginator() return MockPaginator()
@ -233,6 +238,12 @@ def delete_objects(self, *args, **kwargs):
def delete_object(self, *args, **kwargs): def delete_object(self, *args, **kwargs):
pass pass
def get_object(self, Bucket=None, Key=None):
self.ClientError = MockClientError
if Bucket == 'my-bucket' and Key == 'subdirectory/my-file':
return True
raise self.ClientError
def test_remove_s3_url(monkeypatch, capfd): def test_remove_s3_url(monkeypatch, capfd):
fake_s3_url = 's3://my-bucket/subdirectory/mirror' fake_s3_url = 's3://my-bucket/subdirectory/mirror'
@ -254,3 +265,16 @@ def mock_create_s3_session(url):
assert('Failed to delete keyone (Access Denied)' in err) assert('Failed to delete keyone (Access Denied)' in err)
assert('Deleted keythree' in err) assert('Deleted keythree' in err)
assert('Deleted keytwo' in err) assert('Deleted keytwo' in err)
def test_s3_url_exists(monkeypatch, capfd):
def mock_create_s3_session(url):
return MockS3Client()
monkeypatch.setattr(
spack.util.s3, 'create_s3_session', mock_create_s3_session)
fake_s3_url_exists = 's3://my-bucket/subdirectory/my-file'
assert(spack.util.web.url_exists(fake_s3_url_exists))
fake_s3_url_does_not_exist = 's3://my-bucket/subdirectory/my-notfound-file'
assert(not spack.util.web.url_exists(fake_s3_url_does_not_exist))

View file

@ -22,6 +22,7 @@ def create_s3_session(url):
# want to require boto as a dependency unless the user actually wants to # want to require boto as a dependency unless the user actually wants to
# access S3 mirrors. # access S3 mirrors.
from boto3 import Session from boto3 import Session
from botocore.exceptions import ClientError
session = Session() session = Session()
@ -41,4 +42,6 @@ def create_s3_session(url):
s3_client_args["config"] = Config(signature_version=UNSIGNED) s3_client_args["config"] = Config(signature_version=UNSIGNED)
return session.client('s3', **s3_client_args) client = session.client('s3', **s3_client_args)
client.ClientError = ClientError
return client

View file

@ -20,17 +20,6 @@
from six.moves.urllib.error import URLError from six.moves.urllib.error import URLError
from six.moves.urllib.request import urlopen, Request from six.moves.urllib.request import urlopen, Request
if sys.version_info < (3, 0):
# Python 2 had these in the HTMLParser package.
from HTMLParser import HTMLParser, HTMLParseError # novm
else:
# In Python 3, things moved to html.parser
from html.parser import HTMLParser
# Also, HTMLParseError is deprecated and never raised.
class HTMLParseError(Exception):
pass
from llnl.util.filesystem import mkdirp from llnl.util.filesystem import mkdirp
import llnl.util.tty as tty import llnl.util.tty as tty
@ -45,6 +34,16 @@ class HTMLParseError(Exception):
from spack.util.compression import ALLOWED_ARCHIVE_TYPES from spack.util.compression import ALLOWED_ARCHIVE_TYPES
if sys.version_info < (3, 0):
# Python 2 had these in the HTMLParser package.
from HTMLParser import HTMLParser, HTMLParseError # novm
else:
# In Python 3, things moved to html.parser
from html.parser import HTMLParser
# Also, HTMLParseError is deprecated and never raised.
class HTMLParseError(Exception):
pass
# Timeout in seconds for web requests # Timeout in seconds for web requests
_timeout = 10 _timeout = 10
@ -211,11 +210,10 @@ def url_exists(url):
if url.scheme == 's3': if url.scheme == 's3':
s3 = s3_util.create_s3_session(url) s3 = s3_util.create_s3_session(url)
from botocore.exceptions import ClientError
try: try:
s3.get_object(Bucket=url.netloc, Key=url.path) s3.get_object(Bucket=url.netloc, Key=url.path.lstrip('/'))
return True return True
except ClientError as err: except s3.ClientError as err:
if err.response['Error']['Code'] == 'NoSuchKey': if err.response['Error']['Code'] == 'NoSuchKey':
return False return False
raise err raise err