Tom Pollard pushed to branch tpollard/494 at BuildStream / buildstream
Commits:
-
516e990e
by ctolentino8 at 2018-10-31T11:36:46Z
-
b8a37a63
by Tristan Van Berkom at 2018-11-01T10:16:25Z
-
b27b592a
by Benjamin Schubert at 2018-11-01T10:49:57Z
-
89ace5d7
by Benjamin Schubert at 2018-11-01T11:16:36Z
-
4cfabce8
by Angelos Evripiotis at 2018-11-01T11:35:02Z
-
48860aac
by Tristan Van Berkom at 2018-11-01T12:01:04Z
-
d868b409
by Daniel Silverstone at 2018-11-01T13:40:24Z
-
7f79b9ce
by Tristan Van Berkom at 2018-11-01T14:25:57Z
-
de59ebdb
by ctolentino8 at 2018-11-02T16:41:54Z
-
8d7cf806
by ctolentino8 at 2018-11-02T16:41:54Z
-
9c2f9bf7
by Chandan Singh at 2018-11-02T17:09:46Z
-
3788e701
by Jürg Billeter at 2018-11-03T11:52:00Z
-
82e971ef
by Jürg Billeter at 2018-11-05T11:33:20Z
-
62942bfd
by Valentin David at 2018-11-05T12:14:20Z
-
442da2f9
by Javier Jardón at 2018-11-05T12:41:54Z
-
73ca4453
by Tom Pollard at 2018-11-05T13:16:53Z
-
02980640
by Tom Pollard at 2018-11-05T13:32:29Z
28 changed files:
- .gitlab-ci.yml
- NEWS
- buildstream/_artifactcache/artifactcache.py
- buildstream/_artifactcache/cascache.py
- buildstream/_context.py
- buildstream/_frontend/app.py
- buildstream/_frontend/cli.py
- buildstream/_scheduler/queues/pullqueue.py
- buildstream/_versions.py
- buildstream/_yaml.py
- buildstream/data/userconfig.yaml
- buildstream/element.py
- buildstream/plugins/elements/manual.yaml
- buildstream/plugins/sources/pip.py
- setup.py
- tests/completions/completions.py
- tests/frontend/init.py
- tests/integration/build-tree.py
- tests/integration/pip_source.py
- − tests/integration/project/files/pypi-repo/app2/App2-0.1.tar.gz
- − tests/integration/project/files/pypi-repo/app2/index.html
- − tests/integration/project/files/pypi-repo/hellolib/HelloLib-0.1.tar.gz
- − tests/integration/project/files/pypi-repo/hellolib/index.html
- + tests/integration/pullbuildtrees.py
- tests/sources/pip.py
- tests/testutils/__init__.py
- tests/testutils/artifactshare.py
- + tests/testutils/python_repo.py
Changes:
... | ... | @@ -166,6 +166,12 @@ docs: |
166 | 166 |
BST_EXT_REF: 1d6ab71151b93c8cbc0a91a36ffe9270f3b835f1 # 0.5.1
|
167 | 167 |
FD_SDK_REF: 88d7c22c2281b987faa02edd57df80d430eecf1f # 18.08.11-35-g88d7c22c
|
168 | 168 |
before_script:
|
169 |
+ - |
|
|
170 |
+ mkdir -p "${HOME}/.config"
|
|
171 |
+ cat <<EOF >"${HOME}/.config/buildstream.conf"
|
|
172 |
+ scheduler:
|
|
173 |
+ fetchers: 2
|
|
174 |
+ EOF
|
|
169 | 175 |
- (cd dist && ./unpack.sh && cd buildstream && pip3 install .)
|
170 | 176 |
- pip3 install --user -e ${BST_EXT_URL}@${BST_EXT_REF}#egg=bst_ext
|
171 | 177 |
- git clone https://gitlab.com/freedesktop-sdk/freedesktop-sdk.git
|
... | ... | @@ -2,6 +2,12 @@ |
2 | 2 |
buildstream 1.3.1
|
3 | 3 |
=================
|
4 | 4 |
|
5 |
+ o BREAKING CHANGE: The 'manual' element lost its default 'MAKEFLAGS' and 'V'
|
|
6 |
+ environment variables. There is already a 'make' element with the same
|
|
7 |
+ variables. Note that this is a breaking change, it will require users to
|
|
8 |
+ make changes to their .bst files if they are expecting these environment
|
|
9 |
+ variables to be set.
|
|
10 |
+ |
|
5 | 11 |
o Failed builds are included in the cache as well.
|
6 | 12 |
`bst checkout` will provide anything in `%{install-root}`.
|
7 | 13 |
A build including cached fails will cause any dependant elements
|
... | ... | @@ -32,13 +38,23 @@ buildstream 1.3.1 |
32 | 38 |
a bug fix to workspaces so they can be build in workspaces too.
|
33 | 39 |
|
34 | 40 |
o Creating a build shell through the interactive mode or `bst shell --build`
|
35 |
- will now use the cached build tree. It is now easier to debug local build
|
|
36 |
- failures.
|
|
41 |
+ will now use the cached buildtree if available locally. It is now easier to
|
|
42 |
+ debug local build failures.
|
|
37 | 43 |
|
38 | 44 |
o `bst shell --sysroot` now takes any directory that contains a sysroot,
|
39 | 45 |
instead of just a specially-formatted build-root with a `root` and `scratch`
|
40 | 46 |
subdirectory.
|
41 | 47 |
|
48 |
+ o Due to the element `build tree` being cached in the respective artifact their
|
|
49 |
+ size in some cases has significantly increased. In *most* cases the build tree
|
|
50 |
+ is not utilised when building targets, as such by default bst 'pull' & 'build'
|
|
51 |
+ will not fetch buildtrees from remotes. This behaviour can be overriden with
|
|
52 |
+ the cli main option '--pull-build-trees', or the user configuration option
|
|
53 |
+ 'pullbuildtrees = True'. The override will also add the build tree to already
|
|
54 |
+ cached artifacts. When attempting to populate an artifactcache server with
|
|
55 |
+ cached artifacts, only 'complete' elements can be pushed. If the element is
|
|
56 |
+ expected to have a populated build tree then it must be cached before pushing.
|
|
57 |
+ |
|
42 | 58 |
|
43 | 59 |
=================
|
44 | 60 |
buildstream 1.1.5
|
... | ... | @@ -428,6 +428,22 @@ class ArtifactCache(): |
428 | 428 |
raise ImplError("Cache '{kind}' does not implement contains()"
|
429 | 429 |
.format(kind=type(self).__name__))
|
430 | 430 |
|
431 |
+ # contains_subdir_artifact():
|
|
432 |
+ #
|
|
433 |
+ # Check whether an artifact element contains a digest for a subdir
|
|
434 |
+ # which is populated in the cache, i.e non dangling.
|
|
435 |
+ #
|
|
436 |
+ # Args:
|
|
437 |
+ # element (Element): The Element to check
|
|
438 |
+ # key (str): The cache key to use
|
|
439 |
+ # subdir (str): The subdir to check
|
|
440 |
+ #
|
|
441 |
+ # Returns: True if the subdir exists & is populated in the cache, False otherwise
|
|
442 |
+ #
|
|
443 |
+ def contains_subdir_artifact(self, element, key, subdir):
|
|
444 |
+ raise ImplError("Cache '{kind}' does not implement contains_subdir_artifact()"
|
|
445 |
+ .format(kind=type(self).__name__))
|
|
446 |
+ |
|
431 | 447 |
# list_artifacts():
|
432 | 448 |
#
|
433 | 449 |
# List artifacts in this cache in LRU order.
|
... | ... | @@ -464,6 +480,7 @@ class ArtifactCache(): |
464 | 480 |
# Args:
|
465 | 481 |
# element (Element): The Element to extract
|
466 | 482 |
# key (str): The cache key to use
|
483 |
+ # subdir (str): Optional specific dir to extract
|
|
467 | 484 |
#
|
468 | 485 |
# Raises:
|
469 | 486 |
# ArtifactError: In cases there was an OSError, or if the artifact
|
... | ... | @@ -471,7 +488,7 @@ class ArtifactCache(): |
471 | 488 |
#
|
472 | 489 |
# Returns: path to extracted artifact
|
473 | 490 |
#
|
474 |
- def extract(self, element, key):
|
|
491 |
+ def extract(self, element, key, subdir=None):
|
|
475 | 492 |
raise ImplError("Cache '{kind}' does not implement extract()"
|
476 | 493 |
.format(kind=type(self).__name__))
|
477 | 494 |
|
... | ... | @@ -553,11 +570,13 @@ class ArtifactCache(): |
553 | 570 |
# element (Element): The Element whose artifact is to be fetched
|
554 | 571 |
# key (str): The cache key to use
|
555 | 572 |
# progress (callable): The progress callback, if any
|
573 |
+ # subdir (str): The optional specific subdir to pull
|
|
574 |
+ # excluded_subdirs (list): The optional list of subdirs to not pull
|
|
556 | 575 |
#
|
557 | 576 |
# Returns:
|
558 | 577 |
# (bool): True if pull was successful, False if artifact was not available
|
559 | 578 |
#
|
560 |
- def pull(self, element, key, *, progress=None):
|
|
579 |
+ def pull(self, element, key, *, progress=None, subdir=None, excluded_subdirs=None):
|
|
561 | 580 |
raise ImplError("Cache '{kind}' does not implement pull()"
|
562 | 581 |
.format(kind=type(self).__name__))
|
563 | 582 |
|
... | ... | @@ -93,16 +93,36 @@ class CASCache(ArtifactCache): |
93 | 93 |
# This assumes that the repository doesn't have any dangling pointers
|
94 | 94 |
return os.path.exists(refpath)
|
95 | 95 |
|
96 |
- def extract(self, element, key):
|
|
96 |
+ def contains_subdir_artifact(self, element, key, subdir):
|
|
97 |
+ tree = self.resolve_ref(self.get_artifact_fullname(element, key))
|
|
98 |
+ |
|
99 |
+ # This assumes that the subdir digest is present in the element tree
|
|
100 |
+ subdirdigest = self._get_subdir(tree, subdir)
|
|
101 |
+ objpath = self.objpath(subdirdigest)
|
|
102 |
+ |
|
103 |
+ # True if subdir content is cached or if empty as expected
|
|
104 |
+ return os.path.exists(objpath)
|
|
105 |
+ |
|
106 |
+ def extract(self, element, key, subdir=None):
|
|
97 | 107 |
ref = self.get_artifact_fullname(element, key)
|
98 | 108 |
|
99 | 109 |
tree = self.resolve_ref(ref, update_mtime=True)
|
100 | 110 |
|
101 |
- dest = os.path.join(self.extractdir, element._get_project().name,
|
|
102 |
- element.normal_name, tree.hash)
|
|
111 |
+ elementdest = dest = os.path.join(self.extractdir, element._get_project().name,
|
|
112 |
+ element.normal_name, tree.hash)
|
|
113 |
+ |
|
114 |
+ # If artifact is already extracted, check if the optional subdir
|
|
115 |
+ # has also been extracted. If the artifact has not been extracted
|
|
116 |
+ # a full extraction would include the optional subdir
|
|
103 | 117 |
if os.path.isdir(dest):
|
104 |
- # artifact has already been extracted
|
|
105 |
- return dest
|
|
118 |
+ if subdir:
|
|
119 |
+ if not os.path.isdir(os.path.join(dest, subdir)):
|
|
120 |
+ dest = os.path.join(dest, subdir)
|
|
121 |
+ tree = self._get_subdir(tree, subdir)
|
|
122 |
+ else:
|
|
123 |
+ return dest
|
|
124 |
+ else:
|
|
125 |
+ return dest
|
|
106 | 126 |
|
107 | 127 |
with tempfile.TemporaryDirectory(prefix='tmp', dir=self.extractdir) as tmpdir:
|
108 | 128 |
checkoutdir = os.path.join(tmpdir, ref)
|
... | ... | @@ -121,7 +141,7 @@ class CASCache(ArtifactCache): |
121 | 141 |
raise ArtifactError("Failed to extract artifact for ref '{}': {}"
|
122 | 142 |
.format(ref, e)) from e
|
123 | 143 |
|
124 |
- return dest
|
|
144 |
+ return elementdest
|
|
125 | 145 |
|
126 | 146 |
def commit(self, element, content, keys):
|
127 | 147 |
refs = [self.get_artifact_fullname(element, key) for key in keys]
|
... | ... | @@ -229,7 +249,7 @@ class CASCache(ArtifactCache): |
229 | 249 |
remotes_for_project = self._remotes[element._get_project()]
|
230 | 250 |
return any(remote.spec.push for remote in remotes_for_project)
|
231 | 251 |
|
232 |
- def pull(self, element, key, *, progress=None):
|
|
252 |
+ def pull(self, element, key, *, progress=None, subdir=None, excluded_subdirs=None):
|
|
233 | 253 |
ref = self.get_artifact_fullname(element, key)
|
234 | 254 |
|
235 | 255 |
project = element._get_project()
|
... | ... | @@ -248,8 +268,18 @@ class CASCache(ArtifactCache): |
248 | 268 |
tree.hash = response.digest.hash
|
249 | 269 |
tree.size_bytes = response.digest.size_bytes
|
250 | 270 |
|
251 |
- self._fetch_directory(remote, tree)
|
|
271 |
+ # Check if the element artifact is present, if so just fetch the subdir
|
|
272 |
+ # and extract into artifact extract dir if it already exists. If the
|
|
273 |
+ # respective artifact extract dir does not exist a complete extraction
|
|
274 |
+ # will complete
|
|
275 |
+ if subdir and os.path.exists(self.objpath(tree)):
|
|
276 |
+ self._fetch_subdir(remote, tree, subdir)
|
|
277 |
+ self.extract(element, key, subdir)
|
|
278 |
+ else:
|
|
279 |
+ # Fetch artifact, excluded_subdirs determined in pullqueue
|
|
280 |
+ self._fetch_directory(remote, tree, excluded_subdirs=excluded_subdirs)
|
|
252 | 281 |
|
282 |
+ # tree is the remote value, so is the same without or without dangling ref locally
|
|
253 | 283 |
self.set_ref(ref, tree)
|
254 | 284 |
|
255 | 285 |
element.info("Pulled artifact {} <- {}".format(display_key, remote.spec.url))
|
... | ... | @@ -673,8 +703,10 @@ class CASCache(ArtifactCache): |
673 | 703 |
stat.S_IRGRP | stat.S_IXGRP | stat.S_IROTH | stat.S_IXOTH)
|
674 | 704 |
|
675 | 705 |
for dirnode in directory.directories:
|
676 |
- fullpath = os.path.join(dest, dirnode.name)
|
|
677 |
- self._checkout(fullpath, dirnode.digest)
|
|
706 |
+ # Don't try to checkout a dangling ref
|
|
707 |
+ if os.path.exists(self.objpath(dirnode.digest)):
|
|
708 |
+ fullpath = os.path.join(dest, dirnode.name)
|
|
709 |
+ self._checkout(fullpath, dirnode.digest)
|
|
678 | 710 |
|
679 | 711 |
for symlinknode in directory.symlinks:
|
680 | 712 |
# symlink
|
... | ... | @@ -952,11 +984,14 @@ class CASCache(ArtifactCache): |
952 | 984 |
# Args:
|
953 | 985 |
# remote (Remote): The remote to use.
|
954 | 986 |
# dir_digest (Digest): Digest object for the directory to fetch.
|
987 |
+ # excluded_subdirs (list): The optional list of subdirs to not fetch
|
|
955 | 988 |
#
|
956 |
- def _fetch_directory(self, remote, dir_digest):
|
|
989 |
+ def _fetch_directory(self, remote, dir_digest, *, excluded_subdirs=None):
|
|
957 | 990 |
fetch_queue = [dir_digest]
|
958 | 991 |
fetch_next_queue = []
|
959 | 992 |
batch = _CASBatchRead(remote)
|
993 |
+ if not excluded_subdirs:
|
|
994 |
+ excluded_subdirs = []
|
|
960 | 995 |
|
961 | 996 |
while len(fetch_queue) + len(fetch_next_queue) > 0:
|
962 | 997 |
if not fetch_queue:
|
... | ... | @@ -971,8 +1006,9 @@ class CASCache(ArtifactCache): |
971 | 1006 |
directory.ParseFromString(f.read())
|
972 | 1007 |
|
973 | 1008 |
for dirnode in directory.directories:
|
974 |
- batch = self._fetch_directory_node(remote, dirnode.digest, batch,
|
|
975 |
- fetch_queue, fetch_next_queue, recursive=True)
|
|
1009 |
+ if dirnode.name not in excluded_subdirs:
|
|
1010 |
+ batch = self._fetch_directory_node(remote, dirnode.digest, batch,
|
|
1011 |
+ fetch_queue, fetch_next_queue, recursive=True)
|
|
976 | 1012 |
|
977 | 1013 |
for filenode in directory.files:
|
978 | 1014 |
batch = self._fetch_directory_node(remote, filenode.digest, batch,
|
... | ... | @@ -981,6 +1017,10 @@ class CASCache(ArtifactCache): |
981 | 1017 |
# Fetch final batch
|
982 | 1018 |
self._fetch_directory_batch(remote, batch, fetch_queue, fetch_next_queue)
|
983 | 1019 |
|
1020 |
+ def _fetch_subdir(self, remote, tree, subdir):
|
|
1021 |
+ subdirdigest = self._get_subdir(tree, subdir)
|
|
1022 |
+ self._fetch_directory(remote, subdirdigest)
|
|
1023 |
+ |
|
984 | 1024 |
def _fetch_tree(self, remote, digest):
|
985 | 1025 |
# download but do not store the Tree object
|
986 | 1026 |
with tempfile.NamedTemporaryFile(dir=self.tmpdir) as out:
|
... | ... | @@ -105,6 +105,9 @@ class Context(): |
105 | 105 |
# What to do when a build fails in non interactive mode
|
106 | 106 |
self.sched_error_action = 'continue'
|
107 | 107 |
|
108 |
+ # Whether or not to attempt to pull build trees globally
|
|
109 |
+ self.pull_build_trees = False
|
|
110 |
+ |
|
108 | 111 |
# Whether elements must be rebuilt when their dependencies have changed
|
109 | 112 |
self._strict_build_plan = None
|
110 | 113 |
|
... | ... | @@ -161,7 +164,7 @@ class Context(): |
161 | 164 |
_yaml.node_validate(defaults, [
|
162 | 165 |
'sourcedir', 'builddir', 'artifactdir', 'logdir',
|
163 | 166 |
'scheduler', 'artifacts', 'logging', 'projects',
|
164 |
- 'cache'
|
|
167 |
+ 'cache', 'pullbuildtrees'
|
|
165 | 168 |
])
|
166 | 169 |
|
167 | 170 |
for directory in ['sourcedir', 'builddir', 'artifactdir', 'logdir']:
|
... | ... | @@ -186,6 +189,9 @@ class Context(): |
186 | 189 |
# Load artifact share configuration
|
187 | 190 |
self.artifact_cache_specs = ArtifactCache.specs_from_config_node(defaults)
|
188 | 191 |
|
192 |
+ # Load pull build trees configuration
|
|
193 |
+ self.pull_build_trees = _yaml.node_get(defaults, bool, 'pullbuildtrees', default_value='False')
|
|
194 |
+ |
|
189 | 195 |
# Load logging config
|
190 | 196 |
logging = _yaml.node_get(defaults, Mapping, 'logging')
|
191 | 197 |
_yaml.node_validate(logging, [
|
... | ... | @@ -182,7 +182,8 @@ class App(): |
182 | 182 |
'fetchers': 'sched_fetchers',
|
183 | 183 |
'builders': 'sched_builders',
|
184 | 184 |
'pushers': 'sched_pushers',
|
185 |
- 'network_retries': 'sched_network_retries'
|
|
185 |
+ 'network_retries': 'sched_network_retries',
|
|
186 |
+ 'pull_build_trees': 'pull_build_trees'
|
|
186 | 187 |
}
|
187 | 188 |
for cli_option, context_attr in override_map.items():
|
188 | 189 |
option_value = self._main_options.get(cli_option)
|
... | ... | @@ -305,7 +306,6 @@ class App(): |
305 | 306 |
directory = self._main_options['directory']
|
306 | 307 |
directory = os.path.abspath(directory)
|
307 | 308 |
project_path = os.path.join(directory, 'project.conf')
|
308 |
- elements_path = os.path.join(directory, element_path)
|
|
309 | 309 |
|
310 | 310 |
try:
|
311 | 311 |
# Abort if the project.conf already exists, unless `--force` was specified in `bst init`
|
... | ... | @@ -335,6 +335,7 @@ class App(): |
335 | 335 |
raise AppError("Error creating project directory {}: {}".format(directory, e)) from e
|
336 | 336 |
|
337 | 337 |
# Create the elements sub-directory if it doesnt exist
|
338 |
+ elements_path = os.path.join(directory, element_path)
|
|
338 | 339 |
try:
|
339 | 340 |
os.makedirs(elements_path, exist_ok=True)
|
340 | 341 |
except IOError as e:
|
... | ... | @@ -219,6 +219,8 @@ def print_version(ctx, param, value): |
219 | 219 |
help="Specify a project option")
|
220 | 220 |
@click.option('--default-mirror', default=None,
|
221 | 221 |
help="The mirror to fetch from first, before attempting other mirrors")
|
222 |
+@click.option('--pull-build-trees', is_flag=True, default=None,
|
|
223 |
+ help="Include an element's build trees when pulling remote element artifacts")
|
|
222 | 224 |
@click.pass_context
|
223 | 225 |
def cli(context, **kwargs):
|
224 | 226 |
"""Build and manipulate BuildStream projects
|
... | ... | @@ -32,9 +32,20 @@ class PullQueue(Queue): |
32 | 32 |
complete_name = "Pulled"
|
33 | 33 |
resources = [ResourceType.DOWNLOAD, ResourceType.CACHE]
|
34 | 34 |
|
35 |
+ def __init__(self, scheduler):
|
|
36 |
+ super().__init__(scheduler)
|
|
37 |
+ |
|
38 |
+ # Current default exclusions on pull
|
|
39 |
+ self._excluded_subdirs = ["buildtree"]
|
|
40 |
+ self._subdir = None
|
|
41 |
+ # If build trees are to be pulled, remove the value from exclusion list
|
|
42 |
+ if self._scheduler.context.pull_build_trees:
|
|
43 |
+ self._subdir = "buildtree"
|
|
44 |
+ self._excluded_subdirs.remove(self._subdir)
|
|
45 |
+ |
|
35 | 46 |
def process(self, element):
|
36 | 47 |
# returns whether an artifact was downloaded or not
|
37 |
- if not element._pull():
|
|
48 |
+ if not element._pull(subdir=self._subdir, excluded_subdirs=self._excluded_subdirs):
|
|
38 | 49 |
raise SkipJob(self.action_name)
|
39 | 50 |
|
40 | 51 |
def status(self, element):
|
... | ... | @@ -49,7 +60,7 @@ class PullQueue(Queue): |
49 | 60 |
if not element._can_query_cache():
|
50 | 61 |
return QueueStatus.WAIT
|
51 | 62 |
|
52 |
- if element._pull_pending():
|
|
63 |
+ if element._pull_pending(subdir=self._subdir):
|
|
53 | 64 |
return QueueStatus.READY
|
54 | 65 |
else:
|
55 | 66 |
return QueueStatus.SKIP
|
... | ... | @@ -23,7 +23,7 @@ |
23 | 23 |
# This version is bumped whenever enhancements are made
|
24 | 24 |
# to the `project.conf` format or the core element format.
|
25 | 25 |
#
|
26 |
-BST_FORMAT_VERSION = 17
|
|
26 |
+BST_FORMAT_VERSION = 18
|
|
27 | 27 |
|
28 | 28 |
|
29 | 29 |
# The base BuildStream artifact version
|
... | ... | @@ -1049,6 +1049,12 @@ class ChainMap(collections.ChainMap): |
1049 | 1049 |
for key in clearable:
|
1050 | 1050 |
del self[key]
|
1051 | 1051 |
|
1052 |
+ def get(self, key, default=None):
|
|
1053 |
+ try:
|
|
1054 |
+ return self[key]
|
|
1055 |
+ except KeyError:
|
|
1056 |
+ return default
|
|
1057 |
+ |
|
1052 | 1058 |
|
1053 | 1059 |
def node_chain_copy(source):
|
1054 | 1060 |
copy = ChainMap({}, source)
|
... | ... | @@ -97,3 +97,5 @@ logging: |
97 | 97 |
|
98 | 98 |
[%{elapsed}][%{key}][%{element}] %{action} %{message}
|
99 | 99 |
|
100 |
+# Whether to pull buildtrees when downloading element artifacts
|
|
101 |
+pullbuildtrees: False
|
... | ... | @@ -1399,9 +1399,18 @@ class Element(Plugin): |
1399 | 1399 |
.format(workspace.get_absolute_path())):
|
1400 | 1400 |
workspace.stage(temp_staging_directory)
|
1401 | 1401 |
elif self._cached():
|
1402 |
- # We have a cached buildtree to use, instead
|
|
1403 |
- artifact_base, _ = self.__extract()
|
|
1404 |
- import_dir = os.path.join(artifact_base, 'buildtree')
|
|
1402 |
+ # Check if we have a cached buildtree to use
|
|
1403 |
+ context = self._get_context()
|
|
1404 |
+ if context.get_strict():
|
|
1405 |
+ if self.__artifacts.contains_subdir_artifact(self, self.__strict_cache_key, 'buildtree'):
|
|
1406 |
+ artifact_base, _ = self.__extract()
|
|
1407 |
+ import_dir = os.path.join(artifact_base, 'buildtree')
|
|
1408 |
+ elif self.__artifacts.contains_subdir_artifact(self, self.__weak_cache_key, 'buildtree'):
|
|
1409 |
+ artifact_base, _ = self.__extract()
|
|
1410 |
+ import_dir = os.path.join(artifact_base, 'buildtree')
|
|
1411 |
+ else:
|
|
1412 |
+ self.warn("{} is cached without a buildtree, the source will be staged instead"
|
|
1413 |
+ .format(self.name))
|
|
1405 | 1414 |
else:
|
1406 | 1415 |
# No workspace, stage directly
|
1407 | 1416 |
for source in self.sources():
|
... | ... | @@ -1699,18 +1708,26 @@ class Element(Plugin): |
1699 | 1708 |
|
1700 | 1709 |
# _pull_pending()
|
1701 | 1710 |
#
|
1702 |
- # Check whether the artifact will be pulled.
|
|
1711 |
+ # Check whether the artifact will be pulled. If the pull operation is to
|
|
1712 |
+ # include a specific subdir of the element artifact (from cli or user conf)
|
|
1713 |
+ # then the local cache is queried for the subdirs existence.
|
|
1714 |
+ #
|
|
1715 |
+ # Args:
|
|
1716 |
+ # subdir (str): Whether the pull has been invoked with a specific subdir set
|
|
1703 | 1717 |
#
|
1704 | 1718 |
# Returns:
|
1705 | 1719 |
# (bool): Whether a pull operation is pending
|
1706 | 1720 |
#
|
1707 |
- def _pull_pending(self):
|
|
1721 |
+ def _pull_pending(self, subdir=None):
|
|
1708 | 1722 |
if self._get_workspace():
|
1709 | 1723 |
# Workspace builds are never pushed to artifact servers
|
1710 | 1724 |
return False
|
1711 | 1725 |
|
1712 |
- if self.__strong_cached:
|
|
1713 |
- # Artifact already in local cache
|
|
1726 |
+ if self.__strong_cached and subdir:
|
|
1727 |
+ # If we've specified a subdir, check if the subdir is cached locally
|
|
1728 |
+ if self.__artifacts.contains_subdir_artifact(self, self.__strict_cache_key, subdir):
|
|
1729 |
+ return False
|
|
1730 |
+ elif self.__strong_cached:
|
|
1714 | 1731 |
return False
|
1715 | 1732 |
|
1716 | 1733 |
# Pull is pending if artifact remote server available
|
... | ... | @@ -1732,50 +1749,27 @@ class Element(Plugin): |
1732 | 1749 |
|
1733 | 1750 |
self._update_state()
|
1734 | 1751 |
|
1735 |
- def _pull_strong(self, *, progress=None):
|
|
1736 |
- weak_key = self._get_cache_key(strength=_KeyStrength.WEAK)
|
|
1737 |
- |
|
1738 |
- key = self.__strict_cache_key
|
|
1739 |
- if not self.__artifacts.pull(self, key, progress=progress):
|
|
1740 |
- return False
|
|
1741 |
- |
|
1742 |
- # update weak ref by pointing it to this newly fetched artifact
|
|
1743 |
- self.__artifacts.link_key(self, key, weak_key)
|
|
1744 |
- |
|
1745 |
- return True
|
|
1746 |
- |
|
1747 |
- def _pull_weak(self, *, progress=None):
|
|
1748 |
- weak_key = self._get_cache_key(strength=_KeyStrength.WEAK)
|
|
1749 |
- |
|
1750 |
- if not self.__artifacts.pull(self, weak_key, progress=progress):
|
|
1751 |
- return False
|
|
1752 |
- |
|
1753 |
- # extract strong cache key from this newly fetched artifact
|
|
1754 |
- self._pull_done()
|
|
1755 |
- |
|
1756 |
- # create tag for strong cache key
|
|
1757 |
- key = self._get_cache_key(strength=_KeyStrength.STRONG)
|
|
1758 |
- self.__artifacts.link_key(self, weak_key, key)
|
|
1759 |
- |
|
1760 |
- return True
|
|
1761 |
- |
|
1762 | 1752 |
# _pull():
|
1763 | 1753 |
#
|
1764 | 1754 |
# Pull artifact from remote artifact repository into local artifact cache.
|
1765 | 1755 |
#
|
1756 |
+ # Args:
|
|
1757 |
+ # subdir (str): The optional specific subdir to pull
|
|
1758 |
+ # excluded_subdirs (list): The optional list of subdirs to not pull
|
|
1759 |
+ #
|
|
1766 | 1760 |
# Returns: True if the artifact has been downloaded, False otherwise
|
1767 | 1761 |
#
|
1768 |
- def _pull(self):
|
|
1762 |
+ def _pull(self, subdir=None, excluded_subdirs=None):
|
|
1769 | 1763 |
context = self._get_context()
|
1770 | 1764 |
|
1771 | 1765 |
def progress(percent, message):
|
1772 | 1766 |
self.status(message)
|
1773 | 1767 |
|
1774 | 1768 |
# Attempt to pull artifact without knowing whether it's available
|
1775 |
- pulled = self._pull_strong(progress=progress)
|
|
1769 |
+ pulled = self.__pull_strong(progress=progress, subdir=subdir, excluded_subdirs=excluded_subdirs)
|
|
1776 | 1770 |
|
1777 | 1771 |
if not pulled and not self._cached() and not context.get_strict():
|
1778 |
- pulled = self._pull_weak(progress=progress)
|
|
1772 |
+ pulled = self.__pull_weak(progress=progress, subdir=subdir, excluded_subdirs=excluded_subdirs)
|
|
1779 | 1773 |
|
1780 | 1774 |
if not pulled:
|
1781 | 1775 |
return False
|
... | ... | @@ -1798,10 +1792,21 @@ class Element(Plugin): |
1798 | 1792 |
if not self._cached():
|
1799 | 1793 |
return True
|
1800 | 1794 |
|
1801 |
- # Do not push tained artifact
|
|
1795 |
+ # Do not push tainted artifact
|
|
1802 | 1796 |
if self.__get_tainted():
|
1803 | 1797 |
return True
|
1804 | 1798 |
|
1799 |
+ # strict_cache_key can't be relied on to be available when running in non strict mode
|
|
1800 |
+ context = self._get_context()
|
|
1801 |
+ |
|
1802 |
+ # Do not push elements that have a dangling buildtree artifact unless element type is
|
|
1803 |
+ # expected to have an empty buildtree directory
|
|
1804 |
+ if context.get_strict():
|
|
1805 |
+ if not self.__artifacts.contains_subdir_artifact(self, self.__strict_cache_key, 'buildtree'):
|
|
1806 |
+ return True
|
|
1807 |
+ elif not self.__artifacts.contains_subdir_artifact(self, self.__weak_cache_key, 'buildtree'):
|
|
1808 |
+ return True
|
|
1809 |
+ |
|
1805 | 1810 |
return False
|
1806 | 1811 |
|
1807 | 1812 |
# _push():
|
... | ... | @@ -2180,6 +2185,7 @@ class Element(Plugin): |
2180 | 2185 |
stderr=stderr,
|
2181 | 2186 |
config=config,
|
2182 | 2187 |
server_url=self.__remote_execution_url,
|
2188 |
+ bare_directory=bare_directory,
|
|
2183 | 2189 |
allow_real_directory=False)
|
2184 | 2190 |
yield sandbox
|
2185 | 2191 |
|
... | ... | @@ -2681,6 +2687,59 @@ class Element(Plugin): |
2681 | 2687 |
|
2682 | 2688 |
return utils._deduplicate(keys)
|
2683 | 2689 |
|
2690 |
+ # __pull_strong():
|
|
2691 |
+ #
|
|
2692 |
+ # Attempt pulling given element from configured artifact caches with
|
|
2693 |
+ # the strict cache key
|
|
2694 |
+ #
|
|
2695 |
+ # Args:
|
|
2696 |
+ # progress (callable): The progress callback, if any
|
|
2697 |
+ # subdir (str): The optional specific subdir to pull
|
|
2698 |
+ # excluded_subdirs (list): The optional list of subdirs to not pull
|
|
2699 |
+ #
|
|
2700 |
+ # Returns:
|
|
2701 |
+ # (bool): Whether or not the pull was successful
|
|
2702 |
+ #
|
|
2703 |
+ def __pull_strong(self, *, progress=None, subdir=None, excluded_subdirs=None):
|
|
2704 |
+ weak_key = self._get_cache_key(strength=_KeyStrength.WEAK)
|
|
2705 |
+ key = self.__strict_cache_key
|
|
2706 |
+ if not self.__artifacts.pull(self, key, progress=progress, subdir=subdir,
|
|
2707 |
+ excluded_subdirs=excluded_subdirs):
|
|
2708 |
+ return False
|
|
2709 |
+ |
|
2710 |
+ # update weak ref by pointing it to this newly fetched artifact
|
|
2711 |
+ self.__artifacts.link_key(self, key, weak_key)
|
|
2712 |
+ |
|
2713 |
+ return True
|
|
2714 |
+ |
|
2715 |
+ # __pull_weak():
|
|
2716 |
+ #
|
|
2717 |
+ # Attempt pulling given element from configured artifact caches with
|
|
2718 |
+ # the weak cache key
|
|
2719 |
+ #
|
|
2720 |
+ # Args:
|
|
2721 |
+ # progress (callable): The progress callback, if any
|
|
2722 |
+ # subdir (str): The optional specific subdir to pull
|
|
2723 |
+ # excluded_subdirs (list): The optional list of subdirs to not pull
|
|
2724 |
+ #
|
|
2725 |
+ # Returns:
|
|
2726 |
+ # (bool): Whether or not the pull was successful
|
|
2727 |
+ #
|
|
2728 |
+ def __pull_weak(self, *, progress=None, subdir=None, excluded_subdirs=None):
|
|
2729 |
+ weak_key = self._get_cache_key(strength=_KeyStrength.WEAK)
|
|
2730 |
+ if not self.__artifacts.pull(self, weak_key, progress=progress, subdir=subdir,
|
|
2731 |
+ excluded_subdirs=excluded_subdirs):
|
|
2732 |
+ return False
|
|
2733 |
+ |
|
2734 |
+ # extract strong cache key from this newly fetched artifact
|
|
2735 |
+ self._pull_done()
|
|
2736 |
+ |
|
2737 |
+ # create tag for strong cache key
|
|
2738 |
+ key = self._get_cache_key(strength=_KeyStrength.STRONG)
|
|
2739 |
+ self.__artifacts.link_key(self, weak_key, key)
|
|
2740 |
+ |
|
2741 |
+ return True
|
|
2742 |
+ |
|
2684 | 2743 |
|
2685 | 2744 |
def _overlap_error_detail(f, forbidden_overlap_elements, elements):
|
2686 | 2745 |
if forbidden_overlap_elements:
|
1 |
-# No variables added for the manual element by default, set
|
|
2 |
-# this if you plan to use make, and the sources cannot handle
|
|
3 |
-# parallelization.
|
|
4 |
-#
|
|
5 |
-# variables:
|
|
6 |
-#
|
|
7 |
-# notparallel: True
|
|
8 |
- |
|
9 | 1 |
# Manual build element does not provide any default
|
10 | 2 |
# build commands
|
11 | 3 |
config:
|
... | ... | @@ -28,14 +20,3 @@ config: |
28 | 20 |
strip-commands:
|
29 | 21 |
- |
|
30 | 22 |
%{strip-binaries}
|
31 |
- |
|
32 |
-# Use max-jobs CPUs for building and enable verbosity
|
|
33 |
-environment:
|
|
34 |
- MAKEFLAGS: -j%{max-jobs}
|
|
35 |
- V: 1
|
|
36 |
- |
|
37 |
-# And dont consider MAKEFLAGS or V as something which may
|
|
38 |
-# affect build output.
|
|
39 |
-environment-nocache:
|
|
40 |
-- MAKEFLAGS
|
|
41 |
-- V
|
... | ... | @@ -96,7 +96,7 @@ _PYTHON_VERSIONS = [ |
96 | 96 |
# Names of source distribution archives must be of the form
|
97 | 97 |
# '%{package-name}-%{version}.%{extension}'.
|
98 | 98 |
_SDIST_RE = re.compile(
|
99 |
- r'^([a-zA-Z0-9]+?)-(.+).(?:tar|tar.bz2|tar.gz|tar.xz|tar.Z|zip)$',
|
|
99 |
+ r'^([\w.-]+?)-((?:[\d.]+){2,})\.(?:tar|tar.bz2|tar.gz|tar.xz|tar.Z|zip)$',
|
|
100 | 100 |
re.IGNORECASE)
|
101 | 101 |
|
102 | 102 |
|
... | ... | @@ -225,12 +225,27 @@ class PipSource(Source): |
225 | 225 |
def _parse_sdist_names(self, basedir):
|
226 | 226 |
reqs = []
|
227 | 227 |
for f in os.listdir(basedir):
|
228 |
- pkg_match = _SDIST_RE.match(f)
|
|
229 |
- if pkg_match:
|
|
230 |
- reqs.append(pkg_match.groups())
|
|
228 |
+ pkg = _match_package_name(f)
|
|
229 |
+ if pkg is not None:
|
|
230 |
+ reqs.append(pkg)
|
|
231 | 231 |
|
232 | 232 |
return sorted(reqs)
|
233 | 233 |
|
234 | 234 |
|
235 |
+# Extract the package name and version of a source distribution
|
|
236 |
+#
|
|
237 |
+# Args:
|
|
238 |
+# filename (str): Filename of the source distribution
|
|
239 |
+#
|
|
240 |
+# Returns:
|
|
241 |
+# (tuple): A tuple of (package_name, version)
|
|
242 |
+#
|
|
243 |
+def _match_package_name(filename):
|
|
244 |
+ pkg_match = _SDIST_RE.match(filename)
|
|
245 |
+ if pkg_match is None:
|
|
246 |
+ return None
|
|
247 |
+ return pkg_match.groups()
|
|
248 |
+ |
|
249 |
+ |
|
235 | 250 |
def setup():
|
236 | 251 |
return PipSource
|
... | ... | @@ -39,6 +39,7 @@ if sys.version_info[0] != REQUIRED_PYTHON_MAJOR or sys.version_info[1] < REQUIRE |
39 | 39 |
try:
|
40 | 40 |
from setuptools import setup, find_packages, Command
|
41 | 41 |
from setuptools.command.easy_install import ScriptWriter
|
42 |
+ from setuptools.command.test import test as TestCommand
|
|
42 | 43 |
except ImportError:
|
43 | 44 |
print("BuildStream requires setuptools in order to build. Install it using"
|
44 | 45 |
" your package manager (usually python3-setuptools) or via pip (pip3"
|
... | ... | @@ -219,9 +220,48 @@ class BuildGRPC(Command): |
219 | 220 |
f.write(code)
|
220 | 221 |
|
221 | 222 |
|
223 |
+#####################################################
|
|
224 |
+# Pytest command #
|
|
225 |
+#####################################################
|
|
226 |
+class PyTest(TestCommand):
|
|
227 |
+ """Defines a pytest command class to run tests from setup.py"""
|
|
228 |
+ |
|
229 |
+ user_options = TestCommand.user_options + [
|
|
230 |
+ ("addopts=", None, "Arguments to pass to pytest"),
|
|
231 |
+ ('index-url=''build_grpc': BuildGRPC,
|
|
264 |
+ 'pytest': PyTest,
|
|
225 | 265 |
}
|
226 | 266 |
cmdclass.update(versioneer.get_cmdclass())
|
227 | 267 |
return cmdclass
|
... | ... | @@ -305,6 +345,5 @@ setup(name='BuildStream', |
305 | 345 |
'grpcio >= 1.10',
|
306 | 346 |
],
|
307 | 347 |
entry_points=bst_install_entry_points,
|
308 |
- setup_requires=['pytest-runner'],
|
|
309 | 348 |
tests_require=dev_requires,
|
310 | 349 |
zip_safe=False)
|
... | ... | @@ -42,6 +42,7 @@ MAIN_OPTIONS = [ |
42 | 42 |
"-o ",
|
43 | 43 |
"--option ",
|
44 | 44 |
"--on-error ",
|
45 |
+ "--pull-build-trees ",
|
|
45 | 46 |
"--pushers ",
|
46 | 47 |
"--strict ",
|
47 | 48 |
"--verbose ",
|
... | ... | @@ -3,6 +3,7 @@ import pytest |
3 | 3 |
from tests.testutils import cli
|
4 | 4 |
|
5 | 5 |
from buildstream import _yaml
|
6 |
+from buildstream._frontend.app import App
|
|
6 | 7 |
from buildstream._exceptions import ErrorDomain, LoadErrorReason
|
7 | 8 |
from buildstream._versions import BST_FORMAT_VERSION
|
8 | 9 |
|
... | ... | @@ -98,3 +99,34 @@ def test_bad_element_path(cli, tmpdir, element_path): |
98 | 99 |
'init', '--project-name', 'foo', '--element-path', element_path
|
99 | 100 |
])
|
100 | 101 |
result.assert_main_error(ErrorDomain.APP, 'invalid-element-path')
|
102 |
+ |
|
103 |
+ |
|
104 |
+@pytest.mark.parametrize("element_path", [('foo'), ('foo/bar')])
|
|
105 |
+def test_element_path_interactive(cli, tmp_path, monkeypatch, element_path):
|
|
106 |
+ project = tmp_path
|
|
107 |
+ project_conf_path = project.joinpath('project.conf')
|
|
108 |
+ |
|
109 |
+ class DummyInteractiveApp(App):
|
|
110 |
+ def __init__(self, *args, **kwargs):
|
|
111 |
+ super().__init__(*args, **kwargs)
|
|
112 |
+ self.interactive = True
|
|
113 |
+ |
|
114 |
+ @classmethod
|
|
115 |
+ def create(cls, *args, **kwargs):
|
|
116 |
+ return DummyInteractiveApp(*args, **kwargs)
|
|
117 |
+ |
|
118 |
+ def _init_project_interactive(self, *args, **kwargs):
|
|
119 |
+ return ('project_name', '0', element_path)
|
|
120 |
+ |
|
121 |
+ monkeypatch.setattr(App, 'create', DummyInteractiveApp.create)
|
|
122 |
+ |
|
123 |
+ result = cli.run(project=str(project), args=['init'])
|
|
124 |
+ result.assert_success()
|
|
125 |
+ |
|
126 |
+ full_element_path = project.joinpath(element_path)
|
|
127 |
+ assert full_element_path.exists()
|
|
128 |
+ |
|
129 |
+ project_conf = _yaml.load(str(project_conf_path))
|
|
130 |
+ assert project_conf['name'] == 'project_name'
|
|
131 |
+ assert project_conf['format-version'] == '0'
|
|
132 |
+ assert project_conf['element-path'] == element_path
|
... | ... | @@ -70,8 +70,8 @@ def test_buildtree_pulled(cli, tmpdir, datafiles): |
70 | 70 |
})
|
71 | 71 |
assert cli.get_element_state(project, element_name) != 'cached'
|
72 | 72 |
|
73 |
- # Pull from cache
|
|
74 |
- result = cli.run(project=project, args=['pull', '--deps', 'all', element_name])
|
|
73 |
+ # Pull from cache, ensuring cli options is set to pull the buildtree
|
|
74 |
+ result = cli.run(project=project, args=['--pull-build-trees', 'pull', '--deps', 'all', element_name])
|
|
75 | 75 |
result.assert_success()
|
76 | 76 |
|
77 | 77 |
# Check it's using the cached build tree
|
... | ... | @@ -4,6 +4,7 @@ import pytest |
4 | 4 |
from buildstream import _yaml
|
5 | 5 |
|
6 | 6 |
from tests.testutils import cli_integration as cli
|
7 |
+from tests.testutils.python_repo import setup_pypi_repo
|
|
7 | 8 |
from tests.testutils.integration import assert_contains
|
8 | 9 |
|
9 | 10 |
|
... | ... | @@ -17,12 +18,21 @@ DATA_DIR = os.path.join( |
17 | 18 |
|
18 | 19 |
|
19 | 20 |
@pytest.mark.datafiles(DATA_DIR)
|
20 |
-def test_pip_source_import(cli, tmpdir, datafiles):
|
|
21 |
+def test_pip_source_import(cli, tmpdir, datafiles, setup_pypi_repo):
|
|
21 | 22 |
project = os.path.join(datafiles.dirname, datafiles.basename)
|
22 | 23 |
checkout = os.path.join(cli.directory, 'checkout')
|
23 | 24 |
element_path = os.path.join(project, 'elements')
|
24 | 25 |
element_name = 'pip/hello.bst'
|
25 | 26 |
|
27 |
+ # check that exotically named packages are imported correctly
|
|
28 |
+ myreqs_packages = ['hellolib']
|
|
29 |
+ packages = ['app2', 'app.3', 'app-4', 'app_5', 'app.no.6', 'app-no-7', 'app_no_8']
|
|
30 |
+ |
|
31 |
+ # create mock pypi repository
|
|
32 |
+ pypi_repo = os.path.join(project, 'files', 'pypi-repo')
|
|
33 |
+ os.makedirs(pypi_repo, exist_ok=True)
|
|
34 |
+ setup_pypi_repo(myreqs_packages + packages, pypi_repo)
|
|
35 |
+ |
|
26 | 36 |
element = {
|
27 | 37 |
'kind': 'import',
|
28 | 38 |
'sources': [
|
... | ... | @@ -32,9 +42,9 @@ def test_pip_source_import(cli, tmpdir, datafiles): |
32 | 42 |
},
|
33 | 43 |
{
|
34 | 44 |
'kind': 'pip',
|
35 |
- 'url': 'file://{}'.format(os.path.realpath(os.path.join(project, 'files', 'pypi-repo'))),
|
|
45 |
+ 'url': 'file://{}'.format(os.path.realpath(pypi_repo)),
|
|
36 | 46 |
'requirements-files': ['myreqs.txt'],
|
37 |
- 'packages': ['app2']
|
|
47 |
+ 'packages': packages
|
|
38 | 48 |
}
|
39 | 49 |
]
|
40 | 50 |
}
|
... | ... | @@ -51,16 +61,31 @@ def test_pip_source_import(cli, tmpdir, datafiles): |
51 | 61 |
assert result.exit_code == 0
|
52 | 62 |
|
53 | 63 |
assert_contains(checkout, ['/.bst_pip_downloads',
|
54 |
- '/.bst_pip_downloads/HelloLib-0.1.tar.gz',
|
|
55 |
- '/.bst_pip_downloads/App2-0.1.tar.gz'])
|
|
64 |
+ '/.bst_pip_downloads/hellolib-0.1.tar.gz',
|
|
65 |
+ '/.bst_pip_downloads/app2-0.1.tar.gz',
|
|
66 |
+ '/.bst_pip_downloads/app.3-0.1.tar.gz',
|
|
67 |
+ '/.bst_pip_downloads/app-4-0.1.tar.gz',
|
|
68 |
+ '/.bst_pip_downloads/app_5-0.1.tar.gz',
|
|
69 |
+ '/.bst_pip_downloads/app.no.6-0.1.tar.gz',
|
|
70 |
+ '/.bst_pip_downloads/app-no-7-0.1.tar.gz',
|
|
71 |
+ '/.bst_pip_downloads/app_no_8-0.1.tar.gz'])
|
|
56 | 72 |
|
57 | 73 |
|
58 | 74 |
@pytest.mark.datafiles(DATA_DIR)
|
59 |
-def test_pip_source_build(cli, tmpdir, datafiles):
|
|
75 |
+def test_pip_source_build(cli, tmpdir, datafiles, setup_pypi_repo):
|
|
60 | 76 |
project = os.path.join(datafiles.dirname, datafiles.basename)
|
61 | 77 |
element_path = os.path.join(project, 'elements')
|
62 | 78 |
element_name = 'pip/hello.bst'
|
63 | 79 |
|
80 |
+ # check that exotically named packages are imported correctly
|
|
81 |
+ myreqs_packages = ['hellolib']
|
|
82 |
+ packages = ['app2', 'app.3', 'app-4', 'app_5', 'app.no.6', 'app-no-7', 'app_no_8']
|
|
83 |
+ |
|
84 |
+ # create mock pypi repository
|
|
85 |
+ pypi_repo = os.path.join(project, 'files', 'pypi-repo')
|
|
86 |
+ os.makedirs(pypi_repo, exist_ok=True)
|
|
87 |
+ setup_pypi_repo(myreqs_packages + packages, pypi_repo)
|
|
88 |
+ |
|
64 | 89 |
element = {
|
65 | 90 |
'kind': 'manual',
|
66 | 91 |
'depends': ['base.bst'],
|
... | ... | @@ -71,16 +96,15 @@ def test_pip_source_build(cli, tmpdir, datafiles): |
71 | 96 |
},
|
72 | 97 |
{
|
73 | 98 |
'kind': 'pip',
|
74 |
- 'url': 'file://{}'.format(os.path.realpath(os.path.join(project, 'files', 'pypi-repo'))),
|
|
99 |
+ 'url': 'file://{}'.format(os.path.realpath(pypi_repo)),
|
|
75 | 100 |
'requirements-files': ['myreqs.txt'],
|
76 |
- 'packages': ['app2']
|
|
101 |
+ 'packages': packages
|
|
77 | 102 |
}
|
78 | 103 |
],
|
79 | 104 |
'config': {
|
80 | 105 |
'install-commands': [
|
81 | 106 |
'pip3 install --no-index --prefix %{install-root}/usr .bst_pip_downloads/*.tar.gz',
|
82 |
- 'chmod +x app1.py',
|
|
83 |
- 'install app1.py %{install-root}/usr/bin/'
|
|
107 |
+ 'install app1.py %{install-root}/usr/bin/'
|
|
84 | 108 |
]
|
85 | 109 |
}
|
86 | 110 |
}
|
... | ... | @@ -95,5 +119,4 @@ def test_pip_source_build(cli, tmpdir, datafiles): |
95 | 119 |
|
96 | 120 |
result = cli.run(project=project, args=['shell', element_name, '/usr/bin/app1.py'])
|
97 | 121 |
assert result.exit_code == 0
|
98 |
- assert result.output == """Hello App1!
|
|
99 |
-"""
|
|
122 |
+ assert result.output == "Hello App1! This is hellolib\n"
|
No preview for this file type
1 |
-<html>
|
|
2 |
- <head>
|
|
3 |
- <title>Links for app1</title>
|
|
4 |
- </head>
|
|
5 |
- <body>
|
|
6 |
- <a href="">'App2-0.1.tar.gz'>App2-0.1.tar.gz</a><br />
|
|
7 |
- </body>
|
|
8 |
-</html>
|
No preview for this file type
1 |
-<html>
|
|
2 |
- <head>
|
|
3 |
- <title>Links for app1</title>
|
|
4 |
- </head>
|
|
5 |
- <body>
|
|
6 |
- <a href="">'HelloLib-0.1.tar.gz'>HelloLib-0.1.tar.gz</a><br />
|
|
7 |
- </body>
|
|
8 |
-</html>
|
1 |
+import os
|
|
2 |
+import shutil
|
|
3 |
+import pytest
|
|
4 |
+ |
|
5 |
+from tests.testutils import cli_integration as cli, create_artifact_share
|
|
6 |
+from tests.testutils.integration import assert_contains
|
|
7 |
+ |
|
8 |
+ |
|
9 |
+DATA_DIR = os.path.join(
|
|
10 |
+ os.path.dirname(os.path.realpath(__file__)),
|
|
11 |
+ "project"
|
|
12 |
+)
|
|
13 |
+ |
|
14 |
+ |
|
15 |
+# Remove artifact cache & set cli.config value of pullbuildtrees
|
|
16 |
+# to false, which is the default user context. The cache has to be
|
|
17 |
+# cleared as just forcefully removing the refpath leaves dangling objects.
|
|
18 |
+def default_state(cli, tmpdir, share):
|
|
19 |
+ shutil.rmtree(os.path.join(str(tmpdir), 'artifacts'))
|
|
20 |
+ cli.configure({
|
|
21 |
+ 'pullbuildtrees': False,
|
|
22 |
+ 'artifacts': {'url': share.repo, 'push': False},
|
|
23 |
+ 'artifactdir': os.path.join(str(tmpdir), 'artifacts')
|
|
24 |
+ })
|
|
25 |
+ |
|
26 |
+ |
|
27 |
+# A test to capture the integration of the pullbuildtrees
|
|
28 |
+# behaviour, which by default is to not include the buildtree
|
|
29 |
+# directory of an element.
|
|
30 |
+@pytest.mark.integration
|
|
31 |
+@pytest.mark.datafiles(DATA_DIR)
|
|
32 |
+def test_pullbuildtrees(cli, tmpdir, datafiles, integration_cache):
|
|
33 |
+ |
|
34 |
+ project = os.path.join(datafiles.dirname, datafiles.basename)
|
|
35 |
+ element_name = 'autotools/amhello.bst'
|
|
36 |
+ |
|
37 |
+ # Create artifact shares for pull & push testing
|
|
38 |
+ with create_artifact_share(os.path.join(str(tmpdir), 'share1')) as share1,\
|
|
39 |
+ create_artifact_share(os.path.join(str(tmpdir), 'share2')) as share2:
|
|
40 |
+ cli.configure({
|
|
41 |
+ 'artifacts': {'url': share1.repo, 'push': True},
|
|
42 |
+ 'artifactdir': os.path.join(str(tmpdir), 'artifacts')
|
|
43 |
+ })
|
|
44 |
+ |
|
45 |
+ # Build autotools element, checked pushed, delete local
|
|
46 |
+ result = cli.run(project=project, args=['build', element_name])
|
|
47 |
+ assert result.exit_code == 0
|
|
48 |
+ assert cli.get_element_state(project, element_name) == 'cached'
|
|
49 |
+ assert share1.has_artifact('test', element_name, cli.get_element_key(project, element_name))
|
|
50 |
+ default_state(cli, tmpdir, share1)
|
|
51 |
+ |
|
52 |
+ # Pull artifact with default config, assert that pulling again
|
|
53 |
+ # doesn't create a pull job, then assert with buildtrees user
|
|
54 |
+ # config set creates a pull job.
|
|
55 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
56 |
+ assert element_name in result.get_pulled_elements()
|
|
57 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
58 |
+ assert element_name not in result.get_pulled_elements()
|
|
59 |
+ cli.configure({'pullbuildtrees': True})
|
|
60 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
61 |
+ assert element_name in result.get_pulled_elements()
|
|
62 |
+ default_state(cli, tmpdir, share1)
|
|
63 |
+ |
|
64 |
+ # Pull artifact with default config, then assert that pulling
|
|
65 |
+ # with buildtrees cli flag set creates a pull job.
|
|
66 |
+ # Also assert that the buildtree is added to the artifact's
|
|
67 |
+ # extract dir
|
|
68 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
69 |
+ assert element_name in result.get_pulled_elements()
|
|
70 |
+ elementdigest = share1.has_artifact('test', element_name, cli.get_element_key(project, element_name))
|
|
71 |
+ buildtreedir = os.path.join(str(tmpdir), 'artifacts', 'extract', 'test', 'autotools-amhello',
|
|
72 |
+ elementdigest.hash, 'buildtree')
|
|
73 |
+ assert not os.path.isdir(buildtreedir)
|
|
74 |
+ result = cli.run(project=project, args=['--pull-build-trees', 'pull', element_name])
|
|
75 |
+ assert element_name in result.get_pulled_elements()
|
|
76 |
+ assert os.path.isdir(buildtreedir)
|
|
77 |
+ default_state(cli, tmpdir, share1)
|
|
78 |
+ |
|
79 |
+ # Pull artifact with pullbuildtrees set in user config, then assert
|
|
80 |
+ # that pulling with the same user config doesn't creates a pull job,
|
|
81 |
+ # or when buildtrees cli flag is set.
|
|
82 |
+ cli.configure({'pullbuildtrees': True})
|
|
83 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
84 |
+ assert element_name in result.get_pulled_elements()
|
|
85 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
86 |
+ assert element_name not in result.get_pulled_elements()
|
|
87 |
+ result = cli.run(project=project, args=['--pull-build-trees', 'pull', element_name])
|
|
88 |
+ assert element_name not in result.get_pulled_elements()
|
|
89 |
+ default_state(cli, tmpdir, share1)
|
|
90 |
+ |
|
91 |
+ # Pull artifact with default config and buildtrees cli flag set, then assert
|
|
92 |
+ # that pulling with pullbuildtrees set in user config doesn't create a pull
|
|
93 |
+ # job.
|
|
94 |
+ result = cli.run(project=project, args=['--pull-build-trees', 'pull', element_name])
|
|
95 |
+ assert element_name in result.get_pulled_elements()
|
|
96 |
+ cli.configure({'pullbuildtrees': True})
|
|
97 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
98 |
+ assert element_name not in result.get_pulled_elements()
|
|
99 |
+ default_state(cli, tmpdir, share1)
|
|
100 |
+ |
|
101 |
+ # Assert that a partial build element (not containing a populated buildtree dir)
|
|
102 |
+ # can't be pushed to an artifact share, then assert that a complete build element
|
|
103 |
+ # can be. This will attempt a partial pull from share1 and then a partial push
|
|
104 |
+ # to share2
|
|
105 |
+ result = cli.run(project=project, args=['pull', element_name])
|
|
106 |
+ assert element_name in result.get_pulled_elements()
|
|
107 |
+ cli.configure({'artifacts': {'url': share2.repo, 'push': True}})
|
|
108 |
+ result = cli.run(project=project, args=['push', element_name])
|
|
109 |
+ assert element_name not in result.get_pushed_elements()
|
|
110 |
+ assert not share2.has_artifact('test', element_name, cli.get_element_key(project, element_name))
|
|
111 |
+ |
|
112 |
+ # Assert that after pulling the missing buildtree the element artifact can be
|
|
113 |
+ # successfully pushed to the remote. This will attempt to pull the buildtree
|
|
114 |
+ # from share1 and then a 'complete' push to share2
|
|
115 |
+ cli.configure({'artifacts': {'url': share1.repo, 'push': False}})
|
|
116 |
+ result = cli.run(project=project, args=['--pull-build-trees', 'pull', element_name])
|
|
117 |
+ assert element_name in result.get_pulled_elements()
|
|
118 |
+ cli.configure({'artifacts': {'url': share2.repo, 'push': True}})
|
|
119 |
+ result = cli.run(project=project, args=['push', element_name])
|
|
120 |
+ assert element_name in result.get_pushed_elements()
|
|
121 |
+ assert share2.has_artifact('test', element_name, cli.get_element_key(project, element_name))
|
|
122 |
+ default_state(cli, tmpdir, share1)
|
... | ... | @@ -3,6 +3,7 @@ import pytest |
3 | 3 |
|
4 | 4 |
from buildstream._exceptions import ErrorDomain
|
5 | 5 |
from buildstream import _yaml
|
6 |
+from buildstream.plugins.sources.pip import _match_package_name
|
|
6 | 7 |
from tests.testutils import cli
|
7 | 8 |
|
8 | 9 |
DATA_DIR = os.path.join(
|
... | ... | @@ -45,3 +46,22 @@ def test_no_packages(cli, tmpdir, datafiles): |
45 | 46 |
'show', 'target.bst'
|
46 | 47 |
])
|
47 | 48 |
result.assert_main_error(ErrorDomain.SOURCE, None)
|
49 |
+ |
|
50 |
+ |
|
51 |
+# Test that pip source parses tar ball names correctly for the ref
|
|
52 |
+@pytest.mark.parametrize(
|
|
53 |
+ 'tarball, expected_name, expected_version',
|
|
54 |
+ [
|
|
55 |
+ ('dotted.package-0.9.8.tar.gz', 'dotted.package', '0.9.8'),
|
|
56 |
+ ('hyphenated-package-2.6.0.tar.gz', 'hyphenated-package', '2.6.0'),
|
|
57 |
+ ('underscore_pkg-3.1.0.tar.gz', 'underscore_pkg', '3.1.0'),
|
|
58 |
+ ('numbers2and5-1.0.1.tar.gz', 'numbers2and5', '1.0.1'),
|
|
59 |
+ ('multiple.dots.package-5.6.7.tar.gz', 'multiple.dots.package', '5.6.7'),
|
|
60 |
+ ('multiple-hyphens-package-1.2.3.tar.gz', 'multiple-hyphens-package', '1.2.3'),
|
|
61 |
+ ('multiple_underscore_pkg-3.4.5.tar.gz', 'multiple_underscore_pkg', '3.4.5'),
|
|
62 |
+ ('shortversion-1.0.tar.gz', 'shortversion', '1.0'),
|
|
63 |
+ ('longversion-1.2.3.4.tar.gz', 'longversion', '1.2.3.4')
|
|
64 |
+ ])
|
|
65 |
+def test_match_package_name(tarball, expected_name, expected_version):
|
|
66 |
+ name, version = _match_package_name(tarball)
|
|
67 |
+ assert (expected_name, expected_version) == (name, version)
|
... | ... | @@ -29,3 +29,4 @@ from .artifactshare import create_artifact_share |
29 | 29 |
from .element_generators import create_element_size, update_element_size
|
30 | 30 |
from .junction import generate_junction
|
31 | 31 |
from .runner_integration import wait_for_cache_granularity
|
32 |
+from .python_repo import setup_pypi_repo
|
... | ... | @@ -114,7 +114,7 @@ class ArtifactShare(): |
114 | 114 |
# cache_key (str): The cache key
|
115 | 115 |
#
|
116 | 116 |
# Returns:
|
117 |
- # (bool): True if the artifact exists in the share, otherwise false.
|
|
117 |
+ # (str): artifact digest if the artifact exists in the share, otherwise None.
|
|
118 | 118 |
def has_artifact(self, project_name, element_name, cache_key):
|
119 | 119 |
|
120 | 120 |
# NOTE: This should be kept in line with our
|
... | ... | @@ -134,9 +134,9 @@ class ArtifactShare(): |
134 | 134 |
|
135 | 135 |
try:
|
136 | 136 |
tree = self.cas.resolve_ref(artifact_key)
|
137 |
- return True
|
|
137 |
+ return tree
|
|
138 | 138 |
except ArtifactError:
|
139 |
- return False
|
|
139 |
+ return None
|
|
140 | 140 |
|
141 | 141 |
# close():
|
142 | 142 |
#
|
1 |
+from setuptools.sandbox import run_setup
|
|
2 |
+import os
|
|
3 |
+import pytest
|
|
4 |
+import re
|
|
5 |
+import shutil
|
|
6 |
+ |
|
7 |
+ |
|
8 |
+SETUP_TEMPLATE = '''\
|
|
9 |
+from setuptools import setup
|
|
10 |
+ |
|
11 |
+setup(
|
|
12 |
+ name='{name}',
|
|
13 |
+ version='{version}',
|
|
14 |
+ description='{name}',
|
|
15 |
+ packages=['{pkgdirname}'],
|
|
16 |
+ entry_points={{
|
|
17 |
+ 'console_scripts': [
|
|
18 |
+ '{pkgdirname}={pkgdirname}:main'
|
|
19 |
+ ]
|
|
20 |
+ }}
|
|
21 |
+)
|
|
22 |
+'''
|
|
23 |
+ |
|
24 |
+# All packages generated via generate_pip_package will have the functions below
|
|
25 |
+INIT_TEMPLATE = '''\
|
|
26 |
+def main():
|
|
27 |
+ print('This is {name}')
|
|
28 |
+ |
|
29 |
+def hello(actor='world'):
|
|
30 |
+ print('Hello {{}}! This is {name}'.format(actor))
|
|
31 |
+'''
|
|
32 |
+ |
|
33 |
+HTML_TEMPLATE = '''\
|
|
34 |
+<html>
|
|
35 |
+ <head>
|
|
36 |
+ <title>Links for {name}</title>
|
|
37 |
+ </head>
|
|
38 |
+ <body>
|
|
39 |
+ <a href=''>{name}-{version}.tar.gz</a><br />
|
|
40 |
+ </body>
|
|
41 |
+</html>
|
|
42 |
+'''
|
|
43 |
+ |
|
44 |
+ |
|
45 |
+# Creates a simple python source distribution and copies this into a specified
|
|
46 |
+# directory which is to serve as a mock python repository
|
|
47 |
+#
|
|
48 |
+# Args:
|
|
49 |
+# tmpdir (str): Directory in which the source files will be created
|
|
50 |
+# pypi (str): Directory serving as a mock python repository
|
|
51 |
+# name (str): The name of the package to be created
|
|
52 |
+# version (str): The version of the package to be created
|
|
53 |
+#
|
|
54 |
+# Returns:
|
|
55 |
+# None
|
|
56 |
+#
|
|
57 |
+def generate_pip_package(tmpdir, pypi, name, version='0.1'):
|
|
58 |
+ # check if package already exists in pypi
|
|
59 |
+ pypi_package = os.path.join(pypi, re.sub('[^0-9a-zA-Z]+', '-', name))
|
|
60 |
+ if os.path.exists(pypi_package):
|
|
61 |
+ return
|
|
62 |
+ |
|
63 |
+ # create the package source files in tmpdir resulting in a directory
|
|
64 |
+ # tree resembling the following structure:
|
|
65 |
+ #
|
|
66 |
+ # tmpdir
|
|
67 |
+ # |-- setup.py
|
|
68 |
+ # `-- package
|
|
69 |
+ # `-- __init__.py
|
|
70 |
+ #
|
|
71 |
+ setup_file = os.path.join(tmpdir, 'setup.py')
|
|
72 |
+ pkgdirname = re.sub('[^0-9a-zA-Z]+', '', name)
|
|
73 |
+ with open(setup_file, 'w') as f:
|
|
74 |
+ f.write(
|
|
75 |
+ SETUP_TEMPLATE.format(
|
|
76 |
+ name=name,
|
|
77 |
+ version=version,
|
|
78 |
+ pkgdirname=pkgdirname
|
|
79 |
+ )
|
|
80 |
+ )
|
|
81 |
+ os.chmod(setup_file, 0o755)
|
|
82 |
+ |
|
83 |
+ package = os.path.join(tmpdir, pkgdirname)
|
|
84 |
+ os.makedirs(package)
|
|
85 |
+ |
|
86 |
+ main_file = os.path.join(package, '__init__.py')
|
|
87 |
+ with open(main_file, 'w') as f:
|
|
88 |
+ f.write(INIT_TEMPLATE.format(name=name))
|
|
89 |
+ os.chmod(main_file, 0o644)
|
|
90 |
+ |
|
91 |
+ run_setup(setup_file, ['sdist'])
|
|
92 |
+ |
|
93 |
+ # create directory for this package in pypi resulting in a directory
|
|
94 |
+ # tree resembling the following structure:
|
|
95 |
+ #
|
|
96 |
+ # pypi
|
|
97 |
+ # `-- pypi_package
|
|
98 |
+ # |-- index.html
|
|
99 |
+ # `-- foo-0.1.tar.gz
|
|
100 |
+ #
|
|
101 |
+ os.makedirs(pypi_package)
|
|
102 |
+ |
|
103 |
+ # add an index html page
|
|
104 |
+ index_html = os.path.join(pypi_package, 'index.html')
|
|
105 |
+ with open(index_html, 'w') as f:
|
|
106 |
+ f.write(HTML_TEMPLATE.format(name=name, version=version))
|
|
107 |
+ |
|
108 |
+ # copy generated tarfile to pypi package
|
|
109 |
+ dist_dir = os.path.join(tmpdir, 'dist')
|
|
110 |
+ for tar in os.listdir(dist_dir):
|
|
111 |
+ tarpath = os.path.join(dist_dir, tar)
|
|
112 |
+ shutil.copy(tarpath, pypi_package)
|
|
113 |
+ |
|
114 |
+ |
|
115 |
+@pytest.fixture
|
|
116 |
+def setup_pypi_repo(tmpdir):
|
|
117 |
+ def create_pkgdir(package):
|
|
118 |
+ pkgdirname = re.sub('[^0-9a-zA-Z]+', '', package)
|
|
119 |
+ pkgdir = os.path.join(str(tmpdir), pkgdirname)
|
|
120 |
+ os.makedirs(pkgdir)
|
|
121 |
+ return pkgdir
|
|
122 |
+ |
|
123 |
+ def add_packages(packages, pypi_repo):
|
|
124 |
+ for package in packages:
|
|
125 |
+ pkgdir = create_pkgdir(package)
|
|
126 |
+ generate_pip_package(pkgdir, pypi_repo, package)
|
|
127 |
+ |
|
128 |
+ return add_packages
|