Benjamin Schubert pushed to branch bschubert/pipeline at BuildStream / buildstream
Commits:
-
f122cfe8
by Jürg Billeter at 2019-01-06T17:57:33Z
-
7dfb85b3
by Tristan Van Berkom at 2019-01-06T18:31:25Z
-
891fcb0e
by Tristan Van Berkom at 2019-01-07T16:47:01Z
-
5de42d43
by Tristan Van Berkom at 2019-01-07T18:00:37Z
-
059035b9
by Tristan Van Berkom at 2019-01-07T18:02:00Z
-
b83d1b1f
by Tristan Van Berkom at 2019-01-07T18:02:00Z
-
16a8816f
by Tristan Van Berkom at 2019-01-07T18:02:00Z
-
c2fc2a5e
by Tristan Van Berkom at 2019-01-07T18:02:00Z
-
3e3984ad
by Tristan Van Berkom at 2019-01-07T18:50:23Z
-
512c726e
by Tristan Van Berkom at 2019-01-08T03:38:11Z
-
01171988
by Tristan Van Berkom at 2019-01-08T04:20:14Z
-
6c1d06d6
by Phil Dawson at 2019-01-08T10:24:32Z
-
914ecb72
by Jürg Billeter at 2019-01-08T10:54:02Z
-
81f7a964
by Tristan Van Berkom at 2019-01-08T16:55:58Z
-
edcc43ed
by Tristan Van Berkom at 2019-01-08T19:19:10Z
-
17c5ca2d
by Valentin David at 2019-01-09T10:45:23Z
-
605f8d11
by Valentin David at 2019-01-09T10:45:23Z
-
65ea03ab
by Valentin David at 2019-01-09T10:45:23Z
-
ba05c4fa
by Valentin David at 2019-01-09T12:58:59Z
-
eb0dbcfc
by Valentin David at 2019-01-09T13:56:58Z
-
f7681925
by Valentin David at 2019-01-09T14:44:22Z
-
f87d1c93
by Jim MacArthur at 2019-01-09T15:01:18Z
-
ed8bc5bc
by Jim MacArthur at 2019-01-09T15:30:28Z
-
16e7c447
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
af57299f
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
3ae2429a
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
34a415ae
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
55d58733
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
240845dc
by Benjamin Schubert at 2019-01-09T20:19:44Z
-
315042d8
by Benjamin Schubert at 2019-01-09T20:19:44Z
23 changed files:
- buildstream/_artifactcache/cascache.py
- buildstream/_context.py
- buildstream/_frontend/app.py
- buildstream/_project.py
- buildstream/_scheduler/__init__.py
- buildstream/_scheduler/jobs/__init__.py
- buildstream/_scheduler/jobs/cachesizejob.py
- buildstream/_scheduler/jobs/cleanupjob.py
- buildstream/_scheduler/jobs/elementjob.py
- buildstream/_scheduler/jobs/job.py
- buildstream/_scheduler/queues/buildqueue.py
- buildstream/_scheduler/queues/fetchqueue.py
- buildstream/_scheduler/queues/pullqueue.py
- buildstream/_scheduler/queues/queue.py
- buildstream/_scheduler/queues/trackqueue.py
- buildstream/_scheduler/scheduler.py
- buildstream/element.py
- buildstream/plugins/elements/script.py
- buildstream/sandbox/_sandboxremote.py
- buildstream/sandbox/sandbox.py
- buildstream/utils.py
- doc/source/format_project.rst
- doc/source/using_config.rst
Changes:
| ... | ... | @@ -53,7 +53,7 @@ class CASRemoteSpec(namedtuple('CASRemoteSpec', 'url push server_cert client_key |
| 53 | 53 |
#
|
| 54 | 54 |
@staticmethod
|
| 55 | 55 |
def _new_from_config_node(spec_node, basedir=None):
|
| 56 |
- _yaml.node_validate(spec_node, ['url', 'push', 'server-cert', 'client-key', 'client-cert', 'instance_name'])
|
|
| 56 |
+ _yaml.node_validate(spec_node, ['url', 'push', 'server-cert', 'client-key', 'client-cert', 'instance-name'])
|
|
| 57 | 57 |
url = _yaml.node_get(spec_node, str, 'url')
|
| 58 | 58 |
push = _yaml.node_get(spec_node, bool, 'push', default_value=False)
|
| 59 | 59 |
if not url:
|
| ... | ... | @@ -61,7 +61,7 @@ class CASRemoteSpec(namedtuple('CASRemoteSpec', 'url push server_cert client_key |
| 61 | 61 |
raise LoadError(LoadErrorReason.INVALID_DATA,
|
| 62 | 62 |
"{}: empty artifact cache URL".format(provenance))
|
| 63 | 63 |
|
| 64 |
- instance_name = _yaml.node_get(spec_node, str, 'instance_name', default_value=None)
|
|
| 64 |
+ instance_name = _yaml.node_get(spec_node, str, 'instance-name', default_value=None)
|
|
| 65 | 65 |
|
| 66 | 66 |
server_cert = _yaml.node_get(spec_node, str, 'server-cert', default_value=None)
|
| 67 | 67 |
if server_cert and basedir:
|
| ... | ... | @@ -34,6 +34,7 @@ from ._artifactcache import ArtifactCache |
| 34 | 34 |
from ._artifactcache.cascache import CASCache
|
| 35 | 35 |
from ._workspaces import Workspaces, WorkspaceProjectCache, WORKSPACE_PROJECT_FILE
|
| 36 | 36 |
from .plugin import _plugin_lookup
|
| 37 |
+from .sandbox import SandboxRemote
|
|
| 37 | 38 |
|
| 38 | 39 |
|
| 39 | 40 |
# Context()
|
| ... | ... | @@ -72,6 +73,9 @@ class Context(): |
| 72 | 73 |
# The locations from which to push and pull prebuilt artifacts
|
| 73 | 74 |
self.artifact_cache_specs = None
|
| 74 | 75 |
|
| 76 |
+ # The global remote execution configuration
|
|
| 77 |
+ self.remote_execution_specs = None
|
|
| 78 |
+ |
|
| 75 | 79 |
# The directory to store build logs
|
| 76 | 80 |
self.logdir = None
|
| 77 | 81 |
|
| ... | ... | @@ -187,7 +191,7 @@ class Context(): |
| 187 | 191 |
_yaml.node_validate(defaults, [
|
| 188 | 192 |
'sourcedir', 'builddir', 'artifactdir', 'logdir',
|
| 189 | 193 |
'scheduler', 'artifacts', 'logging', 'projects',
|
| 190 |
- 'cache', 'prompt', 'workspacedir',
|
|
| 194 |
+ 'cache', 'prompt', 'workspacedir', 'remote-execution'
|
|
| 191 | 195 |
])
|
| 192 | 196 |
|
| 193 | 197 |
for directory in ['sourcedir', 'builddir', 'artifactdir', 'logdir', 'workspacedir']:
|
| ... | ... | @@ -212,6 +216,8 @@ class Context(): |
| 212 | 216 |
# Load artifact share configuration
|
| 213 | 217 |
self.artifact_cache_specs = ArtifactCache.specs_from_config_node(defaults)
|
| 214 | 218 |
|
| 219 |
+ self.remote_execution_specs = SandboxRemote.specs_from_config_node(defaults)
|
|
| 220 |
+ |
|
| 215 | 221 |
# Load pull build trees configuration
|
| 216 | 222 |
self.pull_buildtrees = _yaml.node_get(cache, bool, 'pull-buildtrees')
|
| 217 | 223 |
|
| ... | ... | @@ -271,7 +277,8 @@ class Context(): |
| 271 | 277 |
# Shallow validation of overrides, parts of buildstream which rely
|
| 272 | 278 |
# on the overrides are expected to validate elsewhere.
|
| 273 | 279 |
for _, overrides in _yaml.node_items(self._project_overrides):
|
| 274 |
- _yaml.node_validate(overrides, ['artifacts', 'options', 'strict', 'default-mirror'])
|
|
| 280 |
+ _yaml.node_validate(overrides, ['artifacts', 'options', 'strict', 'default-mirror',
|
|
| 281 |
+ 'remote-execution'])
|
|
| 275 | 282 |
|
| 276 | 283 |
profile_end(Topics.LOAD_CONTEXT, 'load')
|
| 277 | 284 |
|
| ... | ... | @@ -38,7 +38,7 @@ from .._message import Message, MessageType, unconditional_messages |
| 38 | 38 |
from .._stream import Stream
|
| 39 | 39 |
from .._versions import BST_FORMAT_VERSION
|
| 40 | 40 |
from .. import _yaml
|
| 41 |
-from .._scheduler import ElementJob
|
|
| 41 |
+from .._scheduler import ElementJob, JobStatus
|
|
| 42 | 42 |
|
| 43 | 43 |
# Import frontend assets
|
| 44 | 44 |
from . import Profile, LogLine, Status
|
| ... | ... | @@ -515,13 +515,13 @@ class App(): |
| 515 | 515 |
self._status.add_job(job)
|
| 516 | 516 |
self._maybe_render_status()
|
| 517 | 517 |
|
| 518 |
- def _job_completed(self, job, success):
|
|
| 518 |
+ def _job_completed(self, job, status):
|
|
| 519 | 519 |
self._status.remove_job(job)
|
| 520 | 520 |
self._maybe_render_status()
|
| 521 | 521 |
|
| 522 | 522 |
# Dont attempt to handle a failure if the user has already opted to
|
| 523 | 523 |
# terminate
|
| 524 |
- if not success and not self.stream.terminated:
|
|
| 524 |
+ if status == JobStatus.FAIL and not self.stream.terminated:
|
|
| 525 | 525 |
|
| 526 | 526 |
if isinstance(job, ElementJob):
|
| 527 | 527 |
element = job.element
|
| ... | ... | @@ -599,7 +599,7 @@ class App(): |
| 599 | 599 |
click.echo("\nDropping into an interactive shell in the failed build sandbox\n", err=True)
|
| 600 | 600 |
try:
|
| 601 | 601 |
prompt = self.shell_prompt(element)
|
| 602 |
- self.stream.shell(element, Scope.BUILD, prompt, isolate=True)
|
|
| 602 |
+ self.stream.shell(element, Scope.BUILD, prompt, isolate=True, usebuildtree=True)
|
|
| 603 | 603 |
except BstError as e:
|
| 604 | 604 |
click.echo("Error while attempting to create interactive shell: {}".format(e), err=True)
|
| 605 | 605 |
elif choice == 'log':
|
| ... | ... | @@ -507,7 +507,16 @@ class Project(): |
| 507 | 507 |
self.artifact_cache_specs = ArtifactCache.specs_from_config_node(config, self.directory)
|
| 508 | 508 |
|
| 509 | 509 |
# Load remote-execution configuration for this project
|
| 510 |
- self.remote_execution_specs = SandboxRemote.specs_from_config_node(config, self.directory)
|
|
| 510 |
+ project_specs = SandboxRemote.specs_from_config_node(config, self.directory)
|
|
| 511 |
+ override_specs = SandboxRemote.specs_from_config_node(
|
|
| 512 |
+ self._context.get_overrides(self.name), self.directory)
|
|
| 513 |
+ |
|
| 514 |
+ if override_specs is not None:
|
|
| 515 |
+ self.remote_execution_specs = override_specs
|
|
| 516 |
+ elif project_specs is not None:
|
|
| 517 |
+ self.remote_execution_specs = project_specs
|
|
| 518 |
+ else:
|
|
| 519 |
+ self.remote_execution_specs = self._context.remote_execution_specs
|
|
| 511 | 520 |
|
| 512 | 521 |
# Load sandbox environment variables
|
| 513 | 522 |
self.base_environment = _yaml.node_get(config, Mapping, 'environment')
|
| ... | ... | @@ -26,4 +26,4 @@ from .queues.pushqueue import PushQueue |
| 26 | 26 |
from .queues.pullqueue import PullQueue
|
| 27 | 27 |
|
| 28 | 28 |
from .scheduler import Scheduler, SchedStatus
|
| 29 |
-from .jobs import ElementJob
|
|
| 29 |
+from .jobs import ElementJob, JobStatus
|
| ... | ... | @@ -20,3 +20,4 @@ |
| 20 | 20 |
from .elementjob import ElementJob
|
| 21 | 21 |
from .cachesizejob import CacheSizeJob
|
| 22 | 22 |
from .cleanupjob import CleanupJob
|
| 23 |
+from .job import JobStatus
|
| ... | ... | @@ -16,7 +16,7 @@ |
| 16 | 16 |
# Author:
|
| 17 | 17 |
# Tristan Daniël Maat <tristan maat codethink co uk>
|
| 18 | 18 |
#
|
| 19 |
-from .job import Job
|
|
| 19 |
+from .job import Job, JobStatus
|
|
| 20 | 20 |
|
| 21 | 21 |
|
| 22 | 22 |
class CacheSizeJob(Job):
|
| ... | ... | @@ -30,8 +30,8 @@ class CacheSizeJob(Job): |
| 30 | 30 |
def child_process(self):
|
| 31 | 31 |
return self._artifacts.compute_cache_size()
|
| 32 | 32 |
|
| 33 |
- def parent_complete(self, success, result):
|
|
| 34 |
- if success:
|
|
| 33 |
+ def parent_complete(self, status, result):
|
|
| 34 |
+ if status == JobStatus.OK:
|
|
| 35 | 35 |
self._artifacts.set_cache_size(result)
|
| 36 | 36 |
|
| 37 | 37 |
if self._complete_cb:
|
| ... | ... | @@ -16,7 +16,7 @@ |
| 16 | 16 |
# Author:
|
| 17 | 17 |
# Tristan Daniël Maat <tristan maat codethink co uk>
|
| 18 | 18 |
#
|
| 19 |
-from .job import Job
|
|
| 19 |
+from .job import Job, JobStatus
|
|
| 20 | 20 |
|
| 21 | 21 |
|
| 22 | 22 |
class CleanupJob(Job):
|
| ... | ... | @@ -29,6 +29,6 @@ class CleanupJob(Job): |
| 29 | 29 |
def child_process(self):
|
| 30 | 30 |
return self._artifacts.clean()
|
| 31 | 31 |
|
| 32 |
- def parent_complete(self, success, result):
|
|
| 33 |
- if success:
|
|
| 32 |
+ def parent_complete(self, status, result):
|
|
| 33 |
+ if status == JobStatus.OK:
|
|
| 34 | 34 |
self._artifacts.set_cache_size(result)
|
| ... | ... | @@ -60,7 +60,7 @@ from .job import Job |
| 60 | 60 |
# Args:
|
| 61 | 61 |
# job (Job): The job object which completed
|
| 62 | 62 |
# element (Element): The element passed to the Job() constructor
|
| 63 |
-# success (bool): True if the action_cb did not raise an exception
|
|
| 63 |
+# status (JobStatus): The status of whether the workload raised an exception
|
|
| 64 | 64 |
# result (object): The deserialized object returned by the `action_cb`, or None
|
| 65 | 65 |
# if `success` is False
|
| 66 | 66 |
#
|
| ... | ... | @@ -93,8 +93,8 @@ class ElementJob(Job): |
| 93 | 93 |
# Run the action
|
| 94 | 94 |
return self._action_cb(self._element)
|
| 95 | 95 |
|
| 96 |
- def parent_complete(self, success, result):
|
|
| 97 |
- self._complete_cb(self, self._element, success, self._result)
|
|
| 96 |
+ def parent_complete(self, status, result):
|
|
| 97 |
+ self._complete_cb(self, self._element, status, self._result)
|
|
| 98 | 98 |
|
| 99 | 99 |
def message(self, message_type, message, **kwargs):
|
| 100 | 100 |
args = dict(kwargs)
|
| ... | ... | @@ -28,8 +28,6 @@ import traceback |
| 28 | 28 |
import asyncio
|
| 29 | 29 |
import multiprocessing
|
| 30 | 30 |
|
| 31 |
-import psutil
|
|
| 32 |
- |
|
| 33 | 31 |
# BuildStream toplevel imports
|
| 34 | 32 |
from ..._exceptions import ImplError, BstError, set_last_task_error, SkipJob
|
| 35 | 33 |
from ..._message import Message, MessageType, unconditional_messages
|
| ... | ... | @@ -43,6 +41,22 @@ RC_PERM_FAIL = 2 |
| 43 | 41 |
RC_SKIPPED = 3
|
| 44 | 42 |
|
| 45 | 43 |
|
| 44 |
+# JobStatus:
|
|
| 45 |
+#
|
|
| 46 |
+# The job completion status, passed back through the
|
|
| 47 |
+# complete callbacks.
|
|
| 48 |
+#
|
|
| 49 |
+class JobStatus():
|
|
| 50 |
+ # Job succeeded
|
|
| 51 |
+ OK = 0
|
|
| 52 |
+ |
|
| 53 |
+ # A temporary BstError was raised
|
|
| 54 |
+ FAIL = 1
|
|
| 55 |
+ |
|
| 56 |
+ # A SkipJob was raised
|
|
| 57 |
+ SKIPPED = 3
|
|
| 58 |
+ |
|
| 59 |
+ |
|
| 46 | 60 |
# Used to distinguish between status messages and return values
|
| 47 | 61 |
class Envelope():
|
| 48 | 62 |
def __init__(self, message_type, message):
|
| ... | ... | @@ -118,7 +132,6 @@ class Job(): |
| 118 | 132 |
self._max_retries = max_retries # Maximum number of automatic retries
|
| 119 | 133 |
self._result = None # Return value of child action in the parent
|
| 120 | 134 |
self._tries = 0 # Try count, for retryable jobs
|
| 121 |
- self._skipped_flag = False # Indicate whether the job was skipped.
|
|
| 122 | 135 |
self._terminated = False # Whether this job has been explicitly terminated
|
| 123 | 136 |
|
| 124 | 137 |
# If False, a retry will not be attempted regardless of whether _tries is less than _max_retries.
|
| ... | ... | @@ -215,17 +228,10 @@ class Job(): |
| 215 | 228 |
# Forcefully kill the process, and any children it might have.
|
| 216 | 229 |
#
|
| 217 | 230 |
def kill(self):
|
| 218 |
- |
|
| 219 | 231 |
# Force kill
|
| 220 | 232 |
self.message(MessageType.WARN,
|
| 221 | 233 |
"{} did not terminate gracefully, killing".format(self.action_name))
|
| 222 |
- |
|
| 223 |
- try:
|
|
| 224 |
- utils._kill_process_tree(self._process.pid)
|
|
| 225 |
- # This can happen if the process died of its own accord before
|
|
| 226 |
- # we try to kill it
|
|
| 227 |
- except psutil.NoSuchProcess:
|
|
| 228 |
- return
|
|
| 234 |
+ utils._kill_process_tree(self._process.pid)
|
|
| 229 | 235 |
|
| 230 | 236 |
# suspend()
|
| 231 | 237 |
#
|
| ... | ... | @@ -282,18 +288,6 @@ class Job(): |
| 282 | 288 |
def set_task_id(self, task_id):
|
| 283 | 289 |
self._task_id = task_id
|
| 284 | 290 |
|
| 285 |
- # skipped
|
|
| 286 |
- #
|
|
| 287 |
- # This will evaluate to True if the job was skipped
|
|
| 288 |
- # during processing, or if it was forcefully terminated.
|
|
| 289 |
- #
|
|
| 290 |
- # Returns:
|
|
| 291 |
- # (bool): Whether the job should appear as skipped
|
|
| 292 |
- #
|
|
| 293 |
- @property
|
|
| 294 |
- def skipped(self):
|
|
| 295 |
- return self._skipped_flag or self._terminated
|
|
| 296 |
- |
|
| 297 | 291 |
#######################################################
|
| 298 | 292 |
# Abstract Methods #
|
| 299 | 293 |
#######################################################
|
| ... | ... | @@ -304,10 +298,10 @@ class Job(): |
| 304 | 298 |
# pass the result to the main thread.
|
| 305 | 299 |
#
|
| 306 | 300 |
# Args:
|
| 307 |
- # success (bool): Whether the job was successful.
|
|
| 301 |
+ # status (JobStatus): The job exit status
|
|
| 308 | 302 |
# result (any): The result returned by child_process().
|
| 309 | 303 |
#
|
| 310 |
- def parent_complete(self, success, result):
|
|
| 304 |
+ def parent_complete(self, status, result):
|
|
| 311 | 305 |
raise ImplError("Job '{kind}' does not implement parent_complete()"
|
| 312 | 306 |
.format(kind=type(self).__name__))
|
| 313 | 307 |
|
| ... | ... | @@ -571,16 +565,23 @@ class Job(): |
| 571 | 565 |
#
|
| 572 | 566 |
self._retry_flag = returncode == RC_FAIL
|
| 573 | 567 |
|
| 574 |
- # Set the flag to alert Queue that this job skipped.
|
|
| 575 |
- self._skipped_flag = returncode == RC_SKIPPED
|
|
| 576 |
- |
|
| 577 | 568 |
if self._retry_flag and (self._tries <= self._max_retries) and not self._scheduler.terminated:
|
| 578 | 569 |
self.spawn()
|
| 579 | 570 |
return
|
| 580 | 571 |
|
| 581 |
- success = returncode in (RC_OK, RC_SKIPPED)
|
|
| 582 |
- self.parent_complete(success, self._result)
|
|
| 583 |
- self._scheduler.job_completed(self, success)
|
|
| 572 |
+ # Resolve the outward facing overall job completion status
|
|
| 573 |
+ #
|
|
| 574 |
+ if returncode == RC_OK:
|
|
| 575 |
+ status = JobStatus.OK
|
|
| 576 |
+ elif returncode == RC_SKIPPED:
|
|
| 577 |
+ status = JobStatus.SKIPPED
|
|
| 578 |
+ elif returncode in (RC_FAIL, RC_PERM_FAIL):
|
|
| 579 |
+ status = JobStatus.FAIL
|
|
| 580 |
+ else:
|
|
| 581 |
+ status = JobStatus.FAIL
|
|
| 582 |
+ |
|
| 583 |
+ self.parent_complete(status, self._result)
|
|
| 584 |
+ self._scheduler.job_completed(self, status)
|
|
| 584 | 585 |
|
| 585 | 586 |
# Force the deletion of the queue and process objects to try and clean up FDs
|
| 586 | 587 |
self._queue = self._process = None
|
| ... | ... | @@ -21,7 +21,7 @@ |
| 21 | 21 |
from datetime import timedelta
|
| 22 | 22 |
|
| 23 | 23 |
from . import Queue, QueueStatus
|
| 24 |
-from ..jobs import ElementJob
|
|
| 24 |
+from ..jobs import ElementJob, JobStatus
|
|
| 25 | 25 |
from ..resources import ResourceType
|
| 26 | 26 |
from ..._message import MessageType
|
| 27 | 27 |
|
| ... | ... | @@ -71,9 +71,6 @@ class BuildQueue(Queue): |
| 71 | 71 |
return element._assemble()
|
| 72 | 72 |
|
| 73 | 73 |
def status(self, element):
|
| 74 |
- # state of dependencies may have changed, recalculate element state
|
|
| 75 |
- element._update_state()
|
|
| 76 |
- |
|
| 77 | 74 |
if not element._is_required():
|
| 78 | 75 |
# Artifact is not currently required but it may be requested later.
|
| 79 | 76 |
# Keep it in the queue.
|
| ... | ... | @@ -104,11 +101,16 @@ class BuildQueue(Queue): |
| 104 | 101 |
if artifacts.has_quota_exceeded():
|
| 105 | 102 |
self._scheduler.check_cache_size()
|
| 106 | 103 |
|
| 107 |
- def done(self, job, element, result, success):
|
|
| 104 |
+ def done(self, job, element, result, status):
|
|
| 108 | 105 |
|
| 109 | 106 |
# Inform element in main process that assembly is done
|
| 110 | 107 |
element._assemble_done()
|
| 111 | 108 |
|
| 109 |
+ if element.is_workspaced:
|
|
| 110 |
+ # Update the state of all reverse dependencies
|
|
| 111 |
+ for reverse_dependency in element.reverse_build_dependencies(recurse=True):
|
|
| 112 |
+ reverse_dependency._update_state()
|
|
| 113 |
+ |
|
| 112 | 114 |
# This has to be done after _assemble_done, such that the
|
| 113 | 115 |
# element may register its cache key as required
|
| 114 | 116 |
#
|
| ... | ... | @@ -117,5 +119,5 @@ class BuildQueue(Queue): |
| 117 | 119 |
# artifact cache size for a successful build even though we know a
|
| 118 | 120 |
# failed build also grows the artifact cache size.
|
| 119 | 121 |
#
|
| 120 |
- if success:
|
|
| 122 |
+ if status == JobStatus.OK:
|
|
| 121 | 123 |
self._check_cache_size(job, element, result)
|
| ... | ... | @@ -24,6 +24,7 @@ from ... import Consistency |
| 24 | 24 |
# Local imports
|
| 25 | 25 |
from . import Queue, QueueStatus
|
| 26 | 26 |
from ..resources import ResourceType
|
| 27 |
+from ..jobs import JobStatus
|
|
| 27 | 28 |
|
| 28 | 29 |
|
| 29 | 30 |
# A queue which fetches element sources
|
| ... | ... | @@ -43,9 +44,6 @@ class FetchQueue(Queue): |
| 43 | 44 |
element._fetch()
|
| 44 | 45 |
|
| 45 | 46 |
def status(self, element):
|
| 46 |
- # state of dependencies may have changed, recalculate element state
|
|
| 47 |
- element._update_state()
|
|
| 48 |
- |
|
| 49 | 47 |
if not element._is_required():
|
| 50 | 48 |
# Artifact is not currently required but it may be requested later.
|
| 51 | 49 |
# Keep it in the queue.
|
| ... | ... | @@ -66,9 +64,9 @@ class FetchQueue(Queue): |
| 66 | 64 |
|
| 67 | 65 |
return QueueStatus.READY
|
| 68 | 66 |
|
| 69 |
- def done(self, _, element, result, success):
|
|
| 67 |
+ def done(self, _, element, result, status):
|
|
| 70 | 68 |
|
| 71 |
- if not success:
|
|
| 69 |
+ if status == JobStatus.FAIL:
|
|
| 72 | 70 |
return
|
| 73 | 71 |
|
| 74 | 72 |
element._update_state()
|
| ... | ... | @@ -21,6 +21,7 @@ |
| 21 | 21 |
# Local imports
|
| 22 | 22 |
from . import Queue, QueueStatus
|
| 23 | 23 |
from ..resources import ResourceType
|
| 24 |
+from ..jobs import JobStatus
|
|
| 24 | 25 |
from ..._exceptions import SkipJob
|
| 25 | 26 |
|
| 26 | 27 |
|
| ... | ... | @@ -38,9 +39,6 @@ class PullQueue(Queue): |
| 38 | 39 |
raise SkipJob(self.action_name)
|
| 39 | 40 |
|
| 40 | 41 |
def status(self, element):
|
| 41 |
- # state of dependencies may have changed, recalculate element state
|
|
| 42 |
- element._update_state()
|
|
| 43 |
- |
|
| 44 | 42 |
if not element._is_required():
|
| 45 | 43 |
# Artifact is not currently required but it may be requested later.
|
| 46 | 44 |
# Keep it in the queue.
|
| ... | ... | @@ -54,9 +52,9 @@ class PullQueue(Queue): |
| 54 | 52 |
else:
|
| 55 | 53 |
return QueueStatus.SKIP
|
| 56 | 54 |
|
| 57 |
- def done(self, _, element, result, success):
|
|
| 55 |
+ def done(self, _, element, result, status):
|
|
| 58 | 56 |
|
| 59 |
- if not success:
|
|
| 57 |
+ if status == JobStatus.FAIL:
|
|
| 60 | 58 |
return
|
| 61 | 59 |
|
| 62 | 60 |
element._pull_done()
|
| ... | ... | @@ -64,4 +62,5 @@ class PullQueue(Queue): |
| 64 | 62 |
# Build jobs will check the "approximate" size first. Since we
|
| 65 | 63 |
# do not get an artifact size from pull jobs, we have to
|
| 66 | 64 |
# actually check the cache size.
|
| 67 |
- self._scheduler.check_cache_size()
|
|
| 65 |
+ if status == JobStatus.OK:
|
|
| 66 |
+ self._scheduler.check_cache_size()
|
| ... | ... | @@ -25,7 +25,7 @@ from enum import Enum |
| 25 | 25 |
import traceback
|
| 26 | 26 |
|
| 27 | 27 |
# Local imports
|
| 28 |
-from ..jobs import ElementJob
|
|
| 28 |
+from ..jobs import ElementJob, JobStatus
|
|
| 29 | 29 |
from ..resources import ResourceType
|
| 30 | 30 |
|
| 31 | 31 |
# BuildStream toplevel imports
|
| ... | ... | @@ -133,10 +133,9 @@ class Queue(): |
| 133 | 133 |
# job (Job): The job which completed processing
|
| 134 | 134 |
# element (Element): The element which completed processing
|
| 135 | 135 |
# result (any): The return value of the process() implementation
|
| 136 |
- # success (bool): True if the process() implementation did not
|
|
| 137 |
- # raise any exception
|
|
| 136 |
+ # status (JobStatus): The return status of the Job
|
|
| 138 | 137 |
#
|
| 139 |
- def done(self, job, element, result, success):
|
|
| 138 |
+ def done(self, job, element, result, status):
|
|
| 140 | 139 |
pass
|
| 141 | 140 |
|
| 142 | 141 |
#####################################################
|
| ... | ... | @@ -291,7 +290,7 @@ class Queue(): |
| 291 | 290 |
#
|
| 292 | 291 |
# See the Job object for an explanation of the call signature
|
| 293 | 292 |
#
|
| 294 |
- def _job_done(self, job, element, success, result):
|
|
| 293 |
+ def _job_done(self, job, element, status, result):
|
|
| 295 | 294 |
|
| 296 | 295 |
# Update values that need to be synchronized in the main task
|
| 297 | 296 |
# before calling any queue implementation
|
| ... | ... | @@ -301,7 +300,7 @@ class Queue(): |
| 301 | 300 |
# and determine if it should be considered as processed
|
| 302 | 301 |
# or skipped.
|
| 303 | 302 |
try:
|
| 304 |
- self.done(job, element, result, success)
|
|
| 303 |
+ self.done(job, element, result, status)
|
|
| 305 | 304 |
except BstError as e:
|
| 306 | 305 |
|
| 307 | 306 |
# Report error and mark as failed
|
| ... | ... | @@ -332,12 +331,10 @@ class Queue(): |
| 332 | 331 |
# All jobs get placed on the done queue for later processing.
|
| 333 | 332 |
self._done_queue.append(job)
|
| 334 | 333 |
|
| 335 |
- # A Job can be skipped whether or not it has failed,
|
|
| 336 |
- # we want to only bookkeep them as processed or failed
|
|
| 337 |
- # if they are not skipped.
|
|
| 338 |
- if job.skipped:
|
|
| 334 |
+ # These lists are for bookkeeping purposes for the UI and logging.
|
|
| 335 |
+ if status == JobStatus.SKIPPED:
|
|
| 339 | 336 |
self.skipped_elements.append(element)
|
| 340 |
- elif success:
|
|
| 337 |
+ elif status == JobStatus.OK:
|
|
| 341 | 338 |
self.processed_elements.append(element)
|
| 342 | 339 |
else:
|
| 343 | 340 |
self.failed_elements.append(element)
|
| ... | ... | @@ -24,6 +24,7 @@ from ...plugin import _plugin_lookup |
| 24 | 24 |
# Local imports
|
| 25 | 25 |
from . import Queue, QueueStatus
|
| 26 | 26 |
from ..resources import ResourceType
|
| 27 |
+from ..jobs import JobStatus
|
|
| 27 | 28 |
|
| 28 | 29 |
|
| 29 | 30 |
# A queue which tracks sources
|
| ... | ... | @@ -47,9 +48,9 @@ class TrackQueue(Queue): |
| 47 | 48 |
|
| 48 | 49 |
return QueueStatus.READY
|
| 49 | 50 |
|
| 50 |
- def done(self, _, element, result, success):
|
|
| 51 |
+ def done(self, _, element, result, status):
|
|
| 51 | 52 |
|
| 52 |
- if not success:
|
|
| 53 |
+ if status == JobStatus.FAIL:
|
|
| 53 | 54 |
return
|
| 54 | 55 |
|
| 55 | 56 |
# Set the new refs in the main process one by one as they complete
|
| ... | ... | @@ -58,3 +59,6 @@ class TrackQueue(Queue): |
| 58 | 59 |
source._save_ref(new_ref)
|
| 59 | 60 |
|
| 60 | 61 |
element._tracking_done()
|
| 62 |
+ |
|
| 63 |
+ for rdep in element.reverse_build_dependencies(recurse=True):
|
|
| 64 |
+ rdep._update_state()
|
| ... | ... | @@ -38,6 +38,16 @@ class SchedStatus(): |
| 38 | 38 |
TERMINATED = 1
|
| 39 | 39 |
|
| 40 | 40 |
|
| 41 |
+# Our _REDUNDANT_EXCLUSIVE_ACTIONS jobs are special ones
|
|
| 42 |
+# which we launch dynamically, they have the property of being
|
|
| 43 |
+# meaningless to queue if one is already queued, and it also
|
|
| 44 |
+# doesnt make sense to run them in parallel
|
|
| 45 |
+#
|
|
| 46 |
+_ACTION_NAME_CLEANUP = 'cleanup'
|
|
| 47 |
+_ACTION_NAME_CACHE_SIZE = 'cache_size'
|
|
| 48 |
+_REDUNDANT_EXCLUSIVE_ACTIONS = [_ACTION_NAME_CLEANUP, _ACTION_NAME_CACHE_SIZE]
|
|
| 49 |
+ |
|
| 50 |
+ |
|
| 41 | 51 |
# Scheduler()
|
| 42 | 52 |
#
|
| 43 | 53 |
# The scheduler operates on a list queues, each of which is meant to accomplish
|
| ... | ... | @@ -94,6 +104,15 @@ class Scheduler(): |
| 94 | 104 |
self._suspendtime = None
|
| 95 | 105 |
self._queue_jobs = True # Whether we should continue to queue jobs
|
| 96 | 106 |
|
| 107 |
+ # Whether our exclusive jobs, like 'cleanup' are currently already
|
|
| 108 |
+ # waiting or active.
|
|
| 109 |
+ #
|
|
| 110 |
+ # This is just a bit quicker than scanning the wait queue and active
|
|
| 111 |
+ # queue and comparing job action names.
|
|
| 112 |
+ #
|
|
| 113 |
+ self._exclusive_waiting = set()
|
|
| 114 |
+ self._exclusive_active = set()
|
|
| 115 |
+ |
|
| 97 | 116 |
self._resources = Resources(context.sched_builders,
|
| 98 | 117 |
context.sched_fetchers,
|
| 99 | 118 |
context.sched_pushers)
|
| ... | ... | @@ -211,19 +230,6 @@ class Scheduler(): |
| 211 | 230 |
starttime = timenow
|
| 212 | 231 |
return timenow - starttime
|
| 213 | 232 |
|
| 214 |
- # schedule_jobs()
|
|
| 215 |
- #
|
|
| 216 |
- # Args:
|
|
| 217 |
- # jobs ([Job]): A list of jobs to schedule
|
|
| 218 |
- #
|
|
| 219 |
- # Schedule 'Job's for the scheduler to run. Jobs scheduled will be
|
|
| 220 |
- # run as soon any other queueing jobs finish, provided sufficient
|
|
| 221 |
- # resources are available for them to run
|
|
| 222 |
- #
|
|
| 223 |
- def schedule_jobs(self, jobs):
|
|
| 224 |
- for job in jobs:
|
|
| 225 |
- self.waiting_jobs.append(job)
|
|
| 226 |
- |
|
| 227 | 233 |
# job_completed():
|
| 228 | 234 |
#
|
| 229 | 235 |
# Called when a Job completes
|
| ... | ... | @@ -231,12 +237,14 @@ class Scheduler(): |
| 231 | 237 |
# Args:
|
| 232 | 238 |
# queue (Queue): The Queue holding a complete job
|
| 233 | 239 |
# job (Job): The completed Job
|
| 234 |
- # success (bool): Whether the Job completed with a success status
|
|
| 240 |
+ # status (JobStatus): The status of the completed job
|
|
| 235 | 241 |
#
|
| 236 |
- def job_completed(self, job, success):
|
|
| 242 |
+ def job_completed(self, job, status):
|
|
| 237 | 243 |
self._resources.clear_job_resources(job)
|
| 238 | 244 |
self.active_jobs.remove(job)
|
| 239 |
- self._job_complete_callback(job, success)
|
|
| 245 |
+ if job.action_name in _REDUNDANT_EXCLUSIVE_ACTIONS:
|
|
| 246 |
+ self._exclusive_active.remove(job.action_name)
|
|
| 247 |
+ self._job_complete_callback(job, status)
|
|
| 240 | 248 |
self._schedule_queue_jobs()
|
| 241 | 249 |
self._sched()
|
| 242 | 250 |
|
| ... | ... | @@ -246,18 +254,13 @@ class Scheduler(): |
| 246 | 254 |
# size is calculated, a cleanup job will be run automatically
|
| 247 | 255 |
# if needed.
|
| 248 | 256 |
#
|
| 249 |
- # FIXME: This should ensure that only one cache size job
|
|
| 250 |
- # is ever pending at a given time. If a cache size
|
|
| 251 |
- # job is already running, it is correct to queue
|
|
| 252 |
- # a new one, it is incorrect to have more than one
|
|
| 253 |
- # of these jobs pending at a given time, though.
|
|
| 254 |
- #
|
|
| 255 | 257 |
def check_cache_size(self):
|
| 256 |
- job = CacheSizeJob(self, 'cache_size', 'cache_size/cache_size',
|
|
| 258 |
+ job = CacheSizeJob(self, _ACTION_NAME_CACHE_SIZE,
|
|
| 259 |
+ 'cache_size/cache_size',
|
|
| 257 | 260 |
resources=[ResourceType.CACHE,
|
| 258 | 261 |
ResourceType.PROCESS],
|
| 259 | 262 |
complete_cb=self._run_cleanup)
|
| 260 |
- self.schedule_jobs([job])
|
|
| 263 |
+ self._schedule_jobs([job])
|
|
| 261 | 264 |
|
| 262 | 265 |
#######################################################
|
| 263 | 266 |
# Local Private Methods #
|
| ... | ... | @@ -276,10 +279,19 @@ class Scheduler(): |
| 276 | 279 |
if not self._resources.reserve_job_resources(job):
|
| 277 | 280 |
continue
|
| 278 | 281 |
|
| 282 |
+ # Postpone these jobs if one is already running
|
|
| 283 |
+ if job.action_name in _REDUNDANT_EXCLUSIVE_ACTIONS and \
|
|
| 284 |
+ job.action_name in self._exclusive_active:
|
|
| 285 |
+ continue
|
|
| 286 |
+ |
|
| 279 | 287 |
job.spawn()
|
| 280 | 288 |
self.waiting_jobs.remove(job)
|
| 281 | 289 |
self.active_jobs.append(job)
|
| 282 | 290 |
|
| 291 |
+ if job.action_name in _REDUNDANT_EXCLUSIVE_ACTIONS:
|
|
| 292 |
+ self._exclusive_waiting.remove(job.action_name)
|
|
| 293 |
+ self._exclusive_active.add(job.action_name)
|
|
| 294 |
+ |
|
| 283 | 295 |
if self._job_start_callback:
|
| 284 | 296 |
self._job_start_callback(job)
|
| 285 | 297 |
|
| ... | ... | @@ -287,6 +299,33 @@ class Scheduler(): |
| 287 | 299 |
if not self.active_jobs and not self.waiting_jobs:
|
| 288 | 300 |
self.loop.stop()
|
| 289 | 301 |
|
| 302 |
+ # _schedule_jobs()
|
|
| 303 |
+ #
|
|
| 304 |
+ # The main entry point for jobs to be scheduled.
|
|
| 305 |
+ #
|
|
| 306 |
+ # This is called either as a result of scanning the queues
|
|
| 307 |
+ # in _schedule_queue_jobs(), or directly by the Scheduler
|
|
| 308 |
+ # to insert special jobs like cleanups.
|
|
| 309 |
+ #
|
|
| 310 |
+ # Args:
|
|
| 311 |
+ # jobs ([Job]): A list of jobs to schedule
|
|
| 312 |
+ #
|
|
| 313 |
+ def _schedule_jobs(self, jobs):
|
|
| 314 |
+ for job in jobs:
|
|
| 315 |
+ |
|
| 316 |
+ # Special treatment of our redundant exclusive jobs
|
|
| 317 |
+ #
|
|
| 318 |
+ if job.action_name in _REDUNDANT_EXCLUSIVE_ACTIONS:
|
|
| 319 |
+ |
|
| 320 |
+ # Drop the job if one is already queued
|
|
| 321 |
+ if job.action_name in self._exclusive_waiting:
|
|
| 322 |
+ continue
|
|
| 323 |
+ |
|
| 324 |
+ # Mark this action type as queued
|
|
| 325 |
+ self._exclusive_waiting.add(job.action_name)
|
|
| 326 |
+ |
|
| 327 |
+ self.waiting_jobs.append(job)
|
|
| 328 |
+ |
|
| 290 | 329 |
# _schedule_queue_jobs()
|
| 291 | 330 |
#
|
| 292 | 331 |
# Ask the queues what jobs they want to schedule and schedule
|
| ... | ... | @@ -331,7 +370,7 @@ class Scheduler(): |
| 331 | 370 |
# the next queue and process them.
|
| 332 | 371 |
process_queues = any(q.dequeue_ready() for q in self.queues)
|
| 333 | 372 |
|
| 334 |
- self.schedule_jobs(ready)
|
|
| 373 |
+ self._schedule_jobs(ready)
|
|
| 335 | 374 |
self._sched()
|
| 336 | 375 |
|
| 337 | 376 |
# _run_cleanup()
|
| ... | ... | @@ -353,11 +392,11 @@ class Scheduler(): |
| 353 | 392 |
if not artifacts.has_quota_exceeded():
|
| 354 | 393 |
return
|
| 355 | 394 |
|
| 356 |
- job = CleanupJob(self, 'cleanup', 'cleanup/cleanup',
|
|
| 395 |
+ job = CleanupJob(self, _ACTION_NAME_CLEANUP, 'cleanup/cleanup',
|
|
| 357 | 396 |
resources=[ResourceType.CACHE,
|
| 358 | 397 |
ResourceType.PROCESS],
|
| 359 | 398 |
exclusive_resources=[ResourceType.CACHE])
|
| 360 |
- self.schedule_jobs([job])
|
|
| 399 |
+ self._schedule_jobs([job])
|
|
| 361 | 400 |
|
| 362 | 401 |
# _suspend_jobs()
|
| 363 | 402 |
#
|
| ... | ... | @@ -65,7 +65,7 @@ Miscellaneous abstract methods also exist: |
| 65 | 65 |
|
| 66 | 66 |
* :func:`Element.generate_script() <buildstream.element.Element.generate_script>`
|
| 67 | 67 |
|
| 68 |
- For the purpose of ``bst source bundle``, an Element may optionally implement this.
|
|
| 68 |
+ For the purpose of ``bst source checkout --include-build-scripts``, an Element may optionally implement this.
|
|
| 69 | 69 |
|
| 70 | 70 |
|
| 71 | 71 |
Class Reference
|
| ... | ... | @@ -197,6 +197,7 @@ class Element(Plugin): |
| 197 | 197 |
|
| 198 | 198 |
self.__runtime_dependencies = [] # Direct runtime dependency Elements
|
| 199 | 199 |
self.__build_dependencies = [] # Direct build dependency Elements
|
| 200 |
+ self.__reverse_build_dependencies = [] # Direct reverse build dependency Elements
|
|
| 200 | 201 |
self.__sources = [] # List of Sources
|
| 201 | 202 |
self.__weak_cache_key = None # Our cached weak cache key
|
| 202 | 203 |
self.__strict_cache_key = None # Our cached cache key for strict builds
|
| ... | ... | @@ -227,6 +228,8 @@ class Element(Plugin): |
| 227 | 228 |
self.__metadata_workspaced = {} # Boolean of whether it's workspaced
|
| 228 | 229 |
self.__metadata_workspaced_dependencies = {} # List of which dependencies are workspaced
|
| 229 | 230 |
|
| 231 |
+ self.__is_workspaced = None
|
|
| 232 |
+ |
|
| 230 | 233 |
# Ensure we have loaded this class's defaults
|
| 231 | 234 |
self.__init_defaults(plugin_conf)
|
| 232 | 235 |
|
| ... | ... | @@ -370,6 +373,13 @@ class Element(Plugin): |
| 370 | 373 |
#############################################################
|
| 371 | 374 |
# Public Methods #
|
| 372 | 375 |
#############################################################
|
| 376 |
+ @property
|
|
| 377 |
+ def is_workspaced(self):
|
|
| 378 |
+ if self.__is_workspaced is None:
|
|
| 379 |
+ self.__is_workspaced = self._get_workspace() is not None
|
|
| 380 |
+ |
|
| 381 |
+ return self.__is_workspaced
|
|
| 382 |
+ |
|
| 373 | 383 |
def sources(self):
|
| 374 | 384 |
"""A generator function to enumerate the element sources
|
| 375 | 385 |
|
| ... | ... | @@ -439,6 +449,27 @@ class Element(Plugin): |
| 439 | 449 |
if should_yield and (recurse or recursed) and scope != Scope.BUILD:
|
| 440 | 450 |
yield self
|
| 441 | 451 |
|
| 452 |
+ def reverse_build_dependencies(self, recurse=False):
|
|
| 453 |
+ if not recurse:
|
|
| 454 |
+ yield from self.__reverse_build_dependencies
|
|
| 455 |
+ return
|
|
| 456 |
+ |
|
| 457 |
+ visited = set()
|
|
| 458 |
+ |
|
| 459 |
+ def recurse_rdeps(element):
|
|
| 460 |
+ if element in visited:
|
|
| 461 |
+ return
|
|
| 462 |
+ |
|
| 463 |
+ visited.add(element)
|
|
| 464 |
+ |
|
| 465 |
+ yield element
|
|
| 466 |
+ |
|
| 467 |
+ for rdep in element.__reverse_build_dependencies:
|
|
| 468 |
+ yield from recurse_rdeps(rdep)
|
|
| 469 |
+ |
|
| 470 |
+ for rdep in self.__reverse_build_dependencies:
|
|
| 471 |
+ yield from recurse_rdeps(rdep)
|
|
| 472 |
+ |
|
| 442 | 473 |
def search(self, scope, name):
|
| 443 | 474 |
"""Search for a dependency by name
|
| 444 | 475 |
|
| ... | ... | @@ -930,6 +961,7 @@ class Element(Plugin): |
| 930 | 961 |
for meta_dep in meta.build_dependencies:
|
| 931 | 962 |
dependency = Element._new_from_meta(meta_dep)
|
| 932 | 963 |
element.__build_dependencies.append(dependency)
|
| 964 |
+ dependency.__reverse_build_dependencies.append(element)
|
|
| 933 | 965 |
|
| 934 | 966 |
return element
|
| 935 | 967 |
|
| ... | ... | @@ -1446,6 +1478,7 @@ class Element(Plugin): |
| 1446 | 1478 |
# This unblocks pull/fetch/build.
|
| 1447 | 1479 |
#
|
| 1448 | 1480 |
def _set_required(self):
|
| 1481 |
+ # FIXME: this should enqueue stuff in the queue, it should not be here by default
|
|
| 1449 | 1482 |
if self.__required:
|
| 1450 | 1483 |
# Already done
|
| 1451 | 1484 |
return
|
| ... | ... | @@ -1456,6 +1489,7 @@ class Element(Plugin): |
| 1456 | 1489 |
for dep in self.dependencies(Scope.RUN, recurse=False):
|
| 1457 | 1490 |
dep._set_required()
|
| 1458 | 1491 |
|
| 1492 |
+ # FIXME: this should not be done at all here
|
|
| 1459 | 1493 |
self._update_state()
|
| 1460 | 1494 |
|
| 1461 | 1495 |
# _is_required():
|
| ... | ... | @@ -42,6 +42,9 @@ import buildstream |
| 42 | 42 |
class ScriptElement(buildstream.ScriptElement):
|
| 43 | 43 |
# pylint: disable=attribute-defined-outside-init
|
| 44 | 44 |
|
| 45 |
+ # This plugin has been modified to avoid the use of Sandbox.get_directory
|
|
| 46 |
+ BST_VIRTUAL_DIRECTORY = True
|
|
| 47 |
+ |
|
| 45 | 48 |
def configure(self, node):
|
| 46 | 49 |
for n in self.node_get_member(node, list, 'layout', []):
|
| 47 | 50 |
dst = self.node_subst_member(n, 'destination')
|
| ... | ... | @@ -62,10 +62,32 @@ class SandboxRemote(Sandbox): |
| 62 | 62 |
self.storage_url = config.storage_service['url']
|
| 63 | 63 |
self.exec_url = config.exec_service['url']
|
| 64 | 64 |
|
| 65 |
+ exec_certs = {}
|
|
| 66 |
+ for key in ['client-cert', 'client-key', 'server-cert']:
|
|
| 67 |
+ if key in config.exec_service:
|
|
| 68 |
+ with open(config.exec_service[key], 'rb') as f:
|
|
| 69 |
+ exec_certs[key] = f.read()
|
|
| 70 |
+ |
|
| 71 |
+ self.exec_credentials = grpc.ssl_channel_credentials(
|
|
| 72 |
+ root_certificates=exec_certs.get('server-cert'),
|
|
| 73 |
+ private_key=exec_certs.get('client-key'),
|
|
| 74 |
+ certificate_chain=exec_certs.get('client-cert'))
|
|
| 75 |
+ |
|
| 76 |
+ action_certs = {}
|
|
| 77 |
+ for key in ['client-cert', 'client-key', 'server-cert']:
|
|
| 78 |
+ if key in config.action_service:
|
|
| 79 |
+ with open(config.action_service[key], 'rb') as f:
|
|
| 80 |
+ action_certs[key] = f.read()
|
|
| 81 |
+ |
|
| 65 | 82 |
if config.action_service:
|
| 66 | 83 |
self.action_url = config.action_service['url']
|
| 84 |
+ self.action_credentials = grpc.ssl_channel_credentials(
|
|
| 85 |
+ root_certificates=action_certs.get('server-cert'),
|
|
| 86 |
+ private_key=action_certs.get('client-key'),
|
|
| 87 |
+ certificate_chain=action_certs.get('client-cert'))
|
|
| 67 | 88 |
else:
|
| 68 | 89 |
self.action_url = None
|
| 90 |
+ self.action_credentials = None
|
|
| 69 | 91 |
|
| 70 | 92 |
self.server_instance = config.exec_service.get('instance', None)
|
| 71 | 93 |
self.storage_instance = config.storage_service.get('instance', None)
|
| ... | ... | @@ -81,7 +103,7 @@ class SandboxRemote(Sandbox): |
| 81 | 103 |
self._get_context().message(Message(None, MessageType.INFO, msg))
|
| 82 | 104 |
|
| 83 | 105 |
@staticmethod
|
| 84 |
- def specs_from_config_node(config_node, basedir):
|
|
| 106 |
+ def specs_from_config_node(config_node, basedir=None):
|
|
| 85 | 107 |
|
| 86 | 108 |
def require_node(config, keyname):
|
| 87 | 109 |
val = config.get(keyname)
|
| ... | ... | @@ -109,10 +131,10 @@ class SandboxRemote(Sandbox): |
| 109 | 131 |
remote_exec_storage_config = require_node(remote_config, 'storage-service')
|
| 110 | 132 |
remote_exec_action_config = remote_config.get('action-cache-service', {})
|
| 111 | 133 |
|
| 112 |
- _yaml.node_validate(remote_exec_service_config, ['url', 'instance'])
|
|
| 134 |
+ _yaml.node_validate(remote_exec_service_config, ['url', 'instance'] + tls_keys)
|
|
| 113 | 135 |
_yaml.node_validate(remote_exec_storage_config, ['url', 'instance'] + tls_keys)
|
| 114 | 136 |
if remote_exec_action_config:
|
| 115 |
- _yaml.node_validate(remote_exec_action_config, ['url'])
|
|
| 137 |
+ _yaml.node_validate(remote_exec_action_config, ['url'] + tls_keys)
|
|
| 116 | 138 |
else:
|
| 117 | 139 |
remote_config['action-service'] = None
|
| 118 | 140 |
|
| ... | ... | @@ -135,6 +157,19 @@ class SandboxRemote(Sandbox): |
| 135 | 157 |
"remote-execution configuration. Your config is missing '{}'."
|
| 136 | 158 |
.format(str(provenance), tls_keys, key))
|
| 137 | 159 |
|
| 160 |
+ def resolve_path(path):
|
|
| 161 |
+ if basedir and path:
|
|
| 162 |
+ return os.path.join(basedir, path)
|
|
| 163 |
+ else:
|
|
| 164 |
+ return path
|
|
| 165 |
+ |
|
| 166 |
+ for key in tls_keys:
|
|
| 167 |
+ for d in (remote_config['execution-service'],
|
|
| 168 |
+ remote_config['storage-service'],
|
|
| 169 |
+ remote_exec_action_config):
|
|
| 170 |
+ if key in d:
|
|
| 171 |
+ d[key] = resolve_path(d[key])
|
|
| 172 |
+ |
|
| 138 | 173 |
spec = RemoteExecutionSpec(remote_config['execution-service'],
|
| 139 | 174 |
remote_config['storage-service'],
|
| 140 | 175 |
remote_exec_action_config)
|
| ... | ... | @@ -295,6 +330,8 @@ class SandboxRemote(Sandbox): |
| 295 | 330 |
"for example: http://buildservice:50051.")
|
| 296 | 331 |
if url.scheme == 'http':
|
| 297 | 332 |
channel = grpc.insecure_channel('{}:{}'.format(url.hostname, url.port))
|
| 333 |
+ elif url.scheme == 'https':
|
|
| 334 |
+ channel = grpc.secure_channel('{}:{}'.format(url.hostname, url.port), self.exec_credentials)
|
|
| 298 | 335 |
else:
|
| 299 | 336 |
raise SandboxError("Remote execution currently only supports the 'http' protocol "
|
| 300 | 337 |
"and '{}' was supplied.".format(url.scheme))
|
| ... | ... | @@ -352,11 +389,11 @@ class SandboxRemote(Sandbox): |
| 352 | 389 |
if not url.port:
|
| 353 | 390 |
raise SandboxError("You must supply a protocol and port number in the action-cache-service url, "
|
| 354 | 391 |
"for example: http://buildservice:50051.")
|
| 355 |
- if not url.scheme == "http":
|
|
| 356 |
- raise SandboxError("Currently only support http for the action cache"
|
|
| 357 |
- "and {} was supplied".format(url.scheme))
|
|
| 392 |
+ if url.scheme == 'http':
|
|
| 393 |
+ channel = grpc.insecure_channel('{}:{}'.format(url.hostname, url.port))
|
|
| 394 |
+ elif url.scheme == 'https':
|
|
| 395 |
+ channel = grpc.secure_channel('{}:{}'.format(url.hostname, url.port), self.action_credentials)
|
|
| 358 | 396 |
|
| 359 |
- channel = grpc.insecure_channel('{}:{}'.format(url.hostname, url.port))
|
|
| 360 | 397 |
request = remote_execution_pb2.GetActionResultRequest(action_digest=action_digest)
|
| 361 | 398 |
stub = remote_execution_pb2_grpc.ActionCacheStub(channel)
|
| 362 | 399 |
try:
|
| ... | ... | @@ -288,8 +288,8 @@ class Sandbox(): |
| 288 | 288 |
command = [command]
|
| 289 | 289 |
|
| 290 | 290 |
if self.__batch:
|
| 291 |
- if flags != self.__batch.flags:
|
|
| 292 |
- raise SandboxError("Inconsistent sandbox flags in single command batch")
|
|
| 291 |
+ assert flags == self.__batch.flags, \
|
|
| 292 |
+ "Inconsistent sandbox flags in single command batch"
|
|
| 293 | 293 |
|
| 294 | 294 |
batch_command = _SandboxBatchCommand(command, cwd=cwd, env=env, label=label)
|
| 295 | 295 |
|
| ... | ... | @@ -326,8 +326,8 @@ class Sandbox(): |
| 326 | 326 |
|
| 327 | 327 |
if self.__batch:
|
| 328 | 328 |
# Nested batch
|
| 329 |
- if flags != self.__batch.flags:
|
|
| 330 |
- raise SandboxError("Inconsistent sandbox flags in single command batch")
|
|
| 329 |
+ assert flags == self.__batch.flags, \
|
|
| 330 |
+ "Inconsistent sandbox flags in single command batch"
|
|
| 331 | 331 |
|
| 332 | 332 |
parent_group = self.__batch.current_group
|
| 333 | 333 |
parent_group.append(group)
|
| ... | ... | @@ -592,7 +592,7 @@ class _SandboxBatch(): |
| 592 | 592 |
if command.label:
|
| 593 | 593 |
context = self.sandbox._get_context()
|
| 594 | 594 |
message = Message(self.sandbox._get_plugin_id(), MessageType.STATUS,
|
| 595 |
- 'Running {}'.format(command.label))
|
|
| 595 |
+ 'Running command', detail=command.label)
|
|
| 596 | 596 |
context.message(message)
|
| 597 | 597 |
|
| 598 | 598 |
exitcode = self.sandbox._run(command.command, self.flags, cwd=command.cwd, env=command.env)
|
| ... | ... | @@ -1050,6 +1050,11 @@ def _kill_process_tree(pid): |
| 1050 | 1050 |
# Ignore this error, it can happen with
|
| 1051 | 1051 |
# some setuid bwrap processes.
|
| 1052 | 1052 |
pass
|
| 1053 |
+ except psutil.NoSuchProcess:
|
|
| 1054 |
+ # It is certain that this has already been sent
|
|
| 1055 |
+ # SIGTERM, so there is a window where the process
|
|
| 1056 |
+ # could have exited already.
|
|
| 1057 |
+ pass
|
|
| 1053 | 1058 |
|
| 1054 | 1059 |
# Bloody Murder
|
| 1055 | 1060 |
for child in children:
|
| ... | ... | @@ -218,6 +218,7 @@ The use of ports are required to distinguish between pull only access and |
| 218 | 218 |
push/pull access. For information regarding the server/client certificates
|
| 219 | 219 |
and keys, please see: :ref:`Key pair for the server <server_authentication>`.
|
| 220 | 220 |
|
| 221 |
+.. _project_remote_execution:
|
|
| 221 | 222 |
|
| 222 | 223 |
Remote execution
|
| 223 | 224 |
~~~~~~~~~~~~~~~~
|
| ... | ... | @@ -243,9 +244,6 @@ using the `remote-execution` option: |
| 243 | 244 |
action-cache-service:
|
| 244 | 245 |
url: http://bar.action.com:50052
|
| 245 | 246 |
|
| 246 |
-The execution-service part of remote execution does not support encrypted
|
|
| 247 |
-connections yet, so the protocol must always be http.
|
|
| 248 |
- |
|
| 249 | 247 |
storage-service specifies a remote CAS store and the parameters are the
|
| 250 | 248 |
same as those used to specify an :ref:`artifact server <artifacts>`.
|
| 251 | 249 |
|
| ... | ... | @@ -268,6 +266,9 @@ instance names. |
| 268 | 266 |
|
| 269 | 267 |
The Remote Execution API can be found via https://github.com/bazelbuild/remote-apis.
|
| 270 | 268 |
|
| 269 |
+Remote execution configuration can be also provided in the `user
|
|
| 270 |
+configuration <user_config_remote_execution>`.
|
|
| 271 |
+ |
|
| 271 | 272 |
.. _project_essentials_mirrors:
|
| 272 | 273 |
|
| 273 | 274 |
Mirrors
|
| ... | ... | @@ -100,6 +100,54 @@ pull only access and push/pull access. For information regarding this and the |
| 100 | 100 |
server/client certificates and keys, please see:
|
| 101 | 101 |
:ref:`Key pair for the server <server_authentication>`.
|
| 102 | 102 |
|
| 103 |
+.. _user_config_remote_execution:
|
|
| 104 |
+ |
|
| 105 |
+Remote execution
|
|
| 106 |
+~~~~~~~~~~~~~~~~
|
|
| 107 |
+ |
|
| 108 |
+The same configuration for :ref:`remote execution <project_remote_execution>`
|
|
| 109 |
+in ``project.conf`` can be provided in the user configuation.
|
|
| 110 |
+ |
|
| 111 |
+There is only one remote execution configuration used per project.
|
|
| 112 |
+ |
|
| 113 |
+The project overrides will be taken in priority. The global
|
|
| 114 |
+configuration will be used as fallback.
|
|
| 115 |
+ |
|
| 116 |
+1. Global remote execution fallback:
|
|
| 117 |
+ |
|
| 118 |
+.. code:: yaml
|
|
| 119 |
+ |
|
| 120 |
+ remote-execution:
|
|
| 121 |
+ execution-service:
|
|
| 122 |
+ url: http://execution.fallback.example.com:50051
|
|
| 123 |
+ instance-name: main
|
|
| 124 |
+ storage-service:
|
|
| 125 |
+ url: https://storage.fallback.example.com:11002/
|
|
| 126 |
+ server-cert: /keys/server.crt
|
|
| 127 |
+ client-cert: /keys/client.crt
|
|
| 128 |
+ client-key: /keys/client.key
|
|
| 129 |
+ instance-name: main
|
|
| 130 |
+ action-cache-service:
|
|
| 131 |
+ url: http://action.flalback.example.com:50052
|
|
| 132 |
+ |
|
| 133 |
+2. Project override:
|
|
| 134 |
+ |
|
| 135 |
+.. code:: yaml
|
|
| 136 |
+ |
|
| 137 |
+ projects:
|
|
| 138 |
+ some_project:
|
|
| 139 |
+ remote-execution:
|
|
| 140 |
+ execution-service:
|
|
| 141 |
+ url: http://execution.some_project.example.com:50051
|
|
| 142 |
+ instance-name: main
|
|
| 143 |
+ storage-service:
|
|
| 144 |
+ url: https://storage.some_project.example.com:11002/
|
|
| 145 |
+ server-cert: /some_project_keys/server.crt
|
|
| 146 |
+ client-cert: /some_project_keys/client.crt
|
|
| 147 |
+ client-key: /some_project_keys/client.key
|
|
| 148 |
+ instance-name: main
|
|
| 149 |
+ action-cache-service:
|
|
| 150 |
+ url: http://action.some_project.example.com:50052
|
|
| 103 | 151 |
|
| 104 | 152 |
|
| 105 | 153 |
Strict build plan
|
