[Notes] [Git][BuildStream/buildstream][jonathan/wsl-tests-allow-fail] 17 commits: README.rst: Add table with distros with packaged buildstream



Title: GitLab

Jonathan Maw pushed to branch jonathan/wsl-tests-allow-fail at BuildStream / buildstream

Commits:

9 changed files:

Changes:

  • .gitlab-ci.yml
    ... ... @@ -160,7 +160,7 @@ tests-wsl:
    160 160
     
    
    161 161
       script:
    
    162 162
       - "${TEST_COMMAND}"
    
    163
    -  when: manual
    
    163
    +  allow_failure: true
    
    164 164
     
    
    165 165
     # Automatically build documentation for every commit, we want to know
    
    166 166
     # if building documentation fails even if we're not deploying it.
    
    ... ... @@ -281,10 +281,12 @@ coverage:
    281 281
         - tox -e coverage
    
    282 282
         - cp -a .coverage-reports/ ./coverage-report
    
    283 283
       dependencies:
    
    284
    +  - tests-centos-7.6
    
    284 285
       - tests-debian-9
    
    285 286
       - tests-fedora-28
    
    286 287
       - tests-fedora-29
    
    287 288
       - tests-fedora-missing-deps
    
    289
    +  - tests-fedora-update-deps
    
    288 290
       - tests-ubuntu-18.04
    
    289 291
       - tests-unix
    
    290 292
       except:
    

  • CONTRIBUTING.rst
    ... ... @@ -1234,6 +1234,8 @@ This will give you a ``doc/build/html`` directory with the html docs which
    1234 1234
     you can view in your browser locally to test.
    
    1235 1235
     
    
    1236 1236
     
    
    1237
    +.. _contributing_session_html:
    
    1238
    +
    
    1237 1239
     Regenerating session html
    
    1238 1240
     '''''''''''''''''''''''''
    
    1239 1241
     The documentation build will build the session files if they are missing,
    
    ... ... @@ -1252,6 +1254,8 @@ To force rebuild session html while building the doc, simply run `tox` with the
    1252 1254
       env BST_FORCE_SESSION_REBUILD=1 tox -e docs
    
    1253 1255
     
    
    1254 1256
     
    
    1257
    +.. _contributing_man_pages:
    
    1258
    +
    
    1255 1259
     Man pages
    
    1256 1260
     ~~~~~~~~~
    
    1257 1261
     Unfortunately it is quite difficult to integrate the man pages build
    
    ... ... @@ -1779,3 +1783,257 @@ changing the ``.in`` file, run the following to update the matching ``.txt``
    1779 1783
     file::
    
    1780 1784
     
    
    1781 1785
        make -C requirements
    
    1786
    +
    
    1787
    +
    
    1788
    +Making releases
    
    1789
    +---------------
    
    1790
    +This is a checklist of activities which must be observed when creating
    
    1791
    +BuildStream releases, it is important to keep this section up to date
    
    1792
    +whenever the release process changes.
    
    1793
    +
    
    1794
    +
    
    1795
    +Requirements
    
    1796
    +~~~~~~~~~~~~
    
    1797
    +There are a couple of requirements and accounts required in order
    
    1798
    +to publish a release.
    
    1799
    +
    
    1800
    +* Ability to send email to ``buildstream-list gnome org`` and
    
    1801
    +  to ``gnome-announce-list gnome org``.
    
    1802
    +
    
    1803
    +* Shell account at ``master.gnome.org``.
    
    1804
    +
    
    1805
    +* Access to the `BuildStream project on PyPI <https://pypi.org/project/BuildStream/>`_
    
    1806
    +
    
    1807
    +* An email client which still knows how to send emails in plain text.
    
    1808
    +
    
    1809
    +
    
    1810
    +Pre-release changes
    
    1811
    +~~~~~~~~~~~~~~~~~~~
    
    1812
    +Before actually rolling the release, here is a list of changes which
    
    1813
    +might need to be done in preparation of the release.
    
    1814
    +
    
    1815
    +* Ensure that the man pages are up to date
    
    1816
    +
    
    1817
    +  The man pages are committed to the repository because we are
    
    1818
    +  currently unable to integrate this generation into the setuptools
    
    1819
    +  build phase, as outlined in issue #8.
    
    1820
    +
    
    1821
    +  If any of the user facing CLI has changed, or if any of the
    
    1822
    +  related docstrings have changed, then you should
    
    1823
    +  :ref:`regenerate the man pages <contributing_man_pages>` and
    
    1824
    +  add/commit the results before wrapping a release.
    
    1825
    +
    
    1826
    +* Ensure the documentation session HTML is up to date
    
    1827
    +
    
    1828
    +  The session HTML files are committed to the repository for multiple
    
    1829
    +  reasons, one of them being that the documentation must be buildable
    
    1830
    +  from within a release build environment so that downstream distribution
    
    1831
    +  packagers can easily create the docs package.
    
    1832
    +
    
    1833
    +  This is currently only needed for the first stable release
    
    1834
    +  in a stable line of releases, after this point the API is frozen
    
    1835
    +  and will not change for the remainder of the stable release lifetime,
    
    1836
    +  so nothing interesting will have changed in these session files.
    
    1837
    +
    
    1838
    +  If regeneration is needed, follow :ref:`the instructions above <contributing_session_html>`.
    
    1839
    +
    
    1840
    +* Ensure the NEWS entry is up to date and ready
    
    1841
    +
    
    1842
    +  For a stable release where features have not been added, we
    
    1843
    +  should at least add some entries about the issues which have
    
    1844
    +  been fixed since the last stable release.
    
    1845
    +
    
    1846
    +  For development releases, it is worthwhile going over the
    
    1847
    +  existing entries and ensuring all the major feature additions
    
    1848
    +  are mentioned and there are no redundancies.
    
    1849
    +
    
    1850
    +* Push pre-release changes
    
    1851
    +
    
    1852
    +  Now that any final pre-release changes to generated files or NEWS have
    
    1853
    +  been made, push these directly to the upstream repository.
    
    1854
    +
    
    1855
    +  Do not sit around waiting for CI or approval, these superficial changes
    
    1856
    +  do not affect CI and you are intended to push these changes directly
    
    1857
    +  to the upstream repository.
    
    1858
    +
    
    1859
    +
    
    1860
    +Release process
    
    1861
    +~~~~~~~~~~~~~~~
    
    1862
    +
    
    1863
    +* Ensure that the latest commit is passing in CI
    
    1864
    +
    
    1865
    +  Of course, we do not release software which does not pass it's own
    
    1866
    +  tests.
    
    1867
    +
    
    1868
    +* Get the list of contributors
    
    1869
    +
    
    1870
    +  The list of contributors for a given list is a list of
    
    1871
    +  any contributors who have landed any patches since the
    
    1872
    +  last release.
    
    1873
    +
    
    1874
    +  An easy way to get this list is to ask git to summarize
    
    1875
    +  the authors of commits since the *last release tag*. For
    
    1876
    +  example, if we are about to create the ``1.1.1`` release, then
    
    1877
    +  we need to observe all of the commits since the ``1.1.0``
    
    1878
    +  release:
    
    1879
    +
    
    1880
    +  .. code:: shell
    
    1881
    +
    
    1882
    +     git shortlog -s 1.1.0...@
    
    1883
    +
    
    1884
    +  At times, the same contributor might make contributions from different
    
    1885
    +  machines which they have setup their author names differently, you
    
    1886
    +  can see that some of the authors are actually duplicates, then
    
    1887
    +  remove the duplicates.
    
    1888
    +
    
    1889
    +* Start composing the release announcement email
    
    1890
    +
    
    1891
    +  The first thing to do when composing the release email is to
    
    1892
    +  ensure your mail client has disabled any HTML formatting and will
    
    1893
    +  safely use plain text only.
    
    1894
    +
    
    1895
    +  Try to make the release announcement consistent with other release
    
    1896
    +  announcements as much as possible, an example of the email
    
    1897
    +  can be `found here <https://mail.gnome.org/archives/buildstream-list/2019-February/msg00039.html>`_.
    
    1898
    +
    
    1899
    +  The recipients of the email are ``buildstream-list gnome org`` and
    
    1900
    +  ``gnome-announce-list gnome org`` and the title of the email should
    
    1901
    +  be of the form: ``BuildStream 1.1.1 released``, without any exclamation point.
    
    1902
    +
    
    1903
    +  The format of the email is essentially::
    
    1904
    +
    
    1905
    +    Hi all,
    
    1906
    +
    
    1907
    +    This is the personalized message written to you about this
    
    1908
    +    release.
    
    1909
    +
    
    1910
    +    If this is an unstable release, this should include a warning
    
    1911
    +    to this effect and an invitation to users to please help us
    
    1912
    +    test this release.
    
    1913
    +
    
    1914
    +    This is also a good place to highlight specific bug fixes which
    
    1915
    +    users may have been waiting for, or highlight a new feature we
    
    1916
    +    want users to try out.
    
    1917
    +
    
    1918
    +
    
    1919
    +    What is BuildStream ?
    
    1920
    +    =====================
    
    1921
    +    This is a concise blurb which describes BuildStream in a couple of
    
    1922
    +    sentences, and is taken from the the README.rst.
    
    1923
    +
    
    1924
    +    The easiest thing is to just copy this over from the last release email.
    
    1925
    +
    
    1926
    +
    
    1927
    +    =================
    
    1928
    +    buildstream 1.1.1
    
    1929
    +    =================
    
    1930
    +    This section is directly copy pasted from the top of the NEWS file
    
    1931
    +
    
    1932
    +
    
    1933
    +    Contributors
    
    1934
    +    ============
    
    1935
    +     - This is Where
    
    1936
    +     - You Put
    
    1937
    +     - The Contributor
    
    1938
    +     - Names Which
    
    1939
    +     - You Extracted
    
    1940
    +     - Using git shortlog -s
    
    1941
    +
    
    1942
    +
    
    1943
    +    Where can I get it ?
    
    1944
    +    ====================
    
    1945
    +    https://download.gnome.org/sources/BuildStream/1.1/
    
    1946
    +
    
    1947
    +    For more information on the BuildStream project, visit our home page
    
    1948
    +    at https://buildstream.build/
    
    1949
    +
    
    1950
    +* Publish the release tag
    
    1951
    +
    
    1952
    +  Now that any pre-release changes are upstream, create and push the
    
    1953
    +  signed release tag like so:
    
    1954
    +
    
    1955
    +  .. code:: shell
    
    1956
    +
    
    1957
    +     git tag -s 1.1.1
    
    1958
    +     git push origin 1.1.1
    
    1959
    +
    
    1960
    +* Upload the release tarball
    
    1961
    +
    
    1962
    +  First get yourself into a clean repository state, ensure that you
    
    1963
    +  don't have any unfinished work or precious, uncommitted files lying
    
    1964
    +  around in your checkout and then run:
    
    1965
    +
    
    1966
    +  .. code:: shell
    
    1967
    +
    
    1968
    +     git clean -xdff
    
    1969
    +
    
    1970
    +  Create the tarball with the following command:
    
    1971
    +
    
    1972
    +  .. code:: shell
    
    1973
    +
    
    1974
    +     python3 setup.py sdist
    
    1975
    +
    
    1976
    +  And upload the resulting tarball to the master GNOME server:
    
    1977
    +
    
    1978
    +  .. code:: shell
    
    1979
    +
    
    1980
    +     scp dist/BuildStream-1.1.1.tar.gz <user>@master.gnome.org:
    
    1981
    +
    
    1982
    +  And finally login to your account at ``master.gnome.org`` and run
    
    1983
    +  the install scripts to publish the tarball and update the mirrors:
    
    1984
    +
    
    1985
    +  .. code:: shell
    
    1986
    +
    
    1987
    +     ftpadmin install BuildStream-1.1.1.tar.gz
    
    1988
    +
    
    1989
    +* Send the release email
    
    1990
    +
    
    1991
    +  Now that the release tag is up and the tarball is published,
    
    1992
    +  you can send the release email.
    
    1993
    +
    
    1994
    +
    
    1995
    +Post-release activities
    
    1996
    +~~~~~~~~~~~~~~~~~~~~~~~
    
    1997
    +Once the release has been published, there are some activities
    
    1998
    +which need to be done to ensure everything is up to date.
    
    1999
    +
    
    2000
    +* If this is a stable release, then the tarball should also be
    
    2001
    +  uploaded to PyPI.
    
    2002
    +
    
    2003
    +  Make sure you have ``twine`` installed and upload the tarball
    
    2004
    +  like so:
    
    2005
    +
    
    2006
    +  .. code:: shell
    
    2007
    +
    
    2008
    +     pip3 install --user twine
    
    2009
    +     twine upload -r pypi dist/BuildStream-1.0.1.tar.gz
    
    2010
    +
    
    2011
    +* Update the topic line in the #buildstream IRC channel if needed
    
    2012
    +
    
    2013
    +  The IRC channel usually advertizes the latest stable release
    
    2014
    +  in the topic line, now is the right time to update it.
    
    2015
    +
    
    2016
    +* Update the website repository
    
    2017
    +
    
    2018
    +  The website wants to link to release announcements, but this
    
    2019
    +  cannot be automated because we cannot guess what the link to
    
    2020
    +  the release email will be in the mailing list archive.
    
    2021
    +
    
    2022
    +  Find the URL to the announcement you just published
    
    2023
    +  `in the mailing list archives <https://mail.gnome.org/archives/buildstream-list/>`_,
    
    2024
    +  and use that URL to update the ``anouncements.json`` file in the website
    
    2025
    +  repository.
    
    2026
    +
    
    2027
    +  Commit and push this change to the the ``anouncements.json`` file to
    
    2028
    +  the upstream website repository, and gitlab will take care of automatically
    
    2029
    +  updating the website accordingly.
    
    2030
    +
    
    2031
    +* Regenerate BuildStream documentation
    
    2032
    +
    
    2033
    +  In order to update the badges we use in various documentation
    
    2034
    +  which reflects what is the latest stable releases and the latest
    
    2035
    +  development snapshots, we simply need to ensure a pipeline runs
    
    2036
    +  for the master branch in the BuildStream repository.
    
    2037
    +
    
    2038
    +  You can do this by using the "Run Pipeline" feature on the
    
    2039
    +  `pipelines page in the gitlab UI <https://gitlab.com/BuildStream/buildstream/pipelines>`_.

  • README.rst
    ... ... @@ -100,3 +100,9 @@ We also recommend exploring some existing BuildStream projects:
    100 100
     * https://gitlab.com/baserock/definitions
    
    101 101
     
    
    102 102
     If you have any questions please ask on our `#buildstream <irc://irc.gnome.org/buildstream>`_ channel in `irc.gnome.org <irc://irc.gnome.org>`_
    
    103
    +
    
    104
    +
    
    105
    +Availability in distros
    
    106
    +=======================
    
    107
    +.. image:: https://repology.org/badge/vertical-allrepos/buildstream.svg
    
    108
    +   :target: https://repology.org/metapackage/buildstream/versions

  • buildstream/_artifactcache.py
    ... ... @@ -588,13 +588,16 @@ class ArtifactCache():
    588 588
         #
    
    589 589
         # Args:
    
    590 590
         #     element (Element): The Element commit an artifact for
    
    591
    -    #     content (str): The element's content directory
    
    591
    +    #     content (Directory): The element's content directory
    
    592 592
         #     keys (list): The cache keys to use
    
    593 593
         #
    
    594 594
         def commit(self, element, content, keys):
    
    595 595
             refs = [element.get_artifact_name(key) for key in keys]
    
    596 596
     
    
    597
    -        self.cas.commit(refs, content)
    
    597
    +        tree = content._get_digest()
    
    598
    +
    
    599
    +        for ref in refs:
    
    600
    +            self.cas.set_ref(ref, tree)
    
    598 601
     
    
    599 602
         # diff():
    
    600 603
         #
    

  • buildstream/element.py
    ... ... @@ -103,6 +103,7 @@ from .types import _KeyStrength, CoreWarnings
    103 103
     
    
    104 104
     from .storage.directory import Directory
    
    105 105
     from .storage._filebaseddirectory import FileBasedDirectory
    
    106
    +from .storage._casbaseddirectory import CasBasedDirectory
    
    106 107
     from .storage.directory import VirtualDirectoryError
    
    107 108
     
    
    108 109
     
    
    ... ... @@ -1670,106 +1671,109 @@ class Element(Plugin):
    1670 1671
                         cleanup_rootdir()
    
    1671 1672
     
    
    1672 1673
         def _cache_artifact(self, rootdir, sandbox, collect):
    
    1673
    -        if collect is not None:
    
    1674
    -            try:
    
    1675
    -                sandbox_vroot = sandbox.get_virtual_directory()
    
    1676
    -                collectvdir = sandbox_vroot.descend(collect.lstrip(os.sep).split(os.sep))
    
    1677
    -            except VirtualDirectoryError:
    
    1678
    -                # No collect directory existed
    
    1679
    -                collectvdir = None
    
    1674
    +        with self.timed_activity("Caching artifact"):
    
    1675
    +            if collect is not None:
    
    1676
    +                try:
    
    1677
    +                    sandbox_vroot = sandbox.get_virtual_directory()
    
    1678
    +                    collectvdir = sandbox_vroot.descend(collect.lstrip(os.sep).split(os.sep))
    
    1679
    +                except VirtualDirectoryError:
    
    1680
    +                    # No collect directory existed
    
    1681
    +                    collectvdir = None
    
    1680 1682
     
    
    1681
    -        context = self._get_context()
    
    1683
    +            context = self._get_context()
    
    1682 1684
     
    
    1683
    -        # Create artifact directory structure
    
    1684
    -        assembledir = os.path.join(rootdir, 'artifact')
    
    1685
    -        filesdir = os.path.join(assembledir, 'files')
    
    1686
    -        logsdir = os.path.join(assembledir, 'logs')
    
    1687
    -        metadir = os.path.join(assembledir, 'meta')
    
    1688
    -        buildtreedir = os.path.join(assembledir, 'buildtree')
    
    1689
    -        os.mkdir(assembledir)
    
    1690
    -        if collect is not None and collectvdir is not None:
    
    1691
    -            os.mkdir(filesdir)
    
    1692
    -        os.mkdir(logsdir)
    
    1693
    -        os.mkdir(metadir)
    
    1694
    -        os.mkdir(buildtreedir)
    
    1695
    -
    
    1696
    -        # Hard link files from collect dir to files directory
    
    1697
    -        if collect is not None and collectvdir is not None:
    
    1698
    -            collectvdir.export_files(filesdir, can_link=True)
    
    1699
    -
    
    1700
    -        cache_buildtrees = context.cache_buildtrees
    
    1701
    -        build_success = self.__build_result[0]
    
    1702
    -
    
    1703
    -        # cache_buildtrees defaults to 'always', as such the
    
    1704
    -        # default behaviour is to attempt to cache them. If only
    
    1705
    -        # caching failed artifact buildtrees, then query the build
    
    1706
    -        # result. Element types without a build-root dir will be cached
    
    1707
    -        # with an empty buildtreedir regardless of this configuration.
    
    1708
    -
    
    1709
    -        if cache_buildtrees == 'always' or (cache_buildtrees == 'failure' and not build_success):
    
    1710
    -            try:
    
    1685
    +            assemblevdir = CasBasedDirectory(cas_cache=context.artifactcache.cas, ref=None)
    
    1686
    +            logsvdir = assemblevdir.descend("logs", create=True)
    
    1687
    +            metavdir = assemblevdir.descend("meta", create=True)
    
    1688
    +            buildtreevdir = assemblevdir.descend("buildtree", create=True)
    
    1689
    +
    
    1690
    +            # Create artifact directory structure
    
    1691
    +            assembledir = os.path.join(rootdir, 'artifact')
    
    1692
    +            logsdir = os.path.join(assembledir, 'logs')
    
    1693
    +            metadir = os.path.join(assembledir, 'meta')
    
    1694
    +            os.mkdir(assembledir)
    
    1695
    +            os.mkdir(logsdir)
    
    1696
    +            os.mkdir(metadir)
    
    1697
    +
    
    1698
    +            if collect is not None and collectvdir is not None:
    
    1699
    +                filesvdir = assemblevdir.descend("files", create=True)
    
    1700
    +                filesvdir.import_files(collectvdir)
    
    1701
    +
    
    1702
    +            cache_buildtrees = context.cache_buildtrees
    
    1703
    +            build_success = self.__build_result[0]
    
    1704
    +
    
    1705
    +            # cache_buildtrees defaults to 'always', as such the
    
    1706
    +            # default behaviour is to attempt to cache them. If only
    
    1707
    +            # caching failed artifact buildtrees, then query the build
    
    1708
    +            # result. Element types without a build-root dir will be cached
    
    1709
    +            # with an empty buildtreedir regardless of this configuration.
    
    1710
    +
    
    1711
    +            if cache_buildtrees == 'always' or (cache_buildtrees == 'failure' and not build_success):
    
    1711 1712
                     sandbox_vroot = sandbox.get_virtual_directory()
    
    1712
    -                sandbox_build_dir = sandbox_vroot.descend(
    
    1713
    -                    self.get_variable('build-root').lstrip(os.sep).split(os.sep))
    
    1714
    -                # Hard link files from build-root dir to buildtreedir directory
    
    1715
    -                sandbox_build_dir.export_files(buildtreedir)
    
    1716
    -            except VirtualDirectoryError:
    
    1717
    -                # Directory could not be found. Pre-virtual
    
    1718
    -                # directory behaviour was to continue silently
    
    1719
    -                # if the directory could not be found.
    
    1720
    -                pass
    
    1713
    +                try:
    
    1714
    +                    sandbox_build_dir = sandbox_vroot.descend(
    
    1715
    +                        self.get_variable('build-root').lstrip(os.sep).split(os.sep))
    
    1716
    +                    buildtreevdir.import_files(sandbox_build_dir)
    
    1717
    +                except VirtualDirectoryError:
    
    1718
    +                    # Directory could not be found. Pre-virtual
    
    1719
    +                    # directory behaviour was to continue silently
    
    1720
    +                    # if the directory could not be found.
    
    1721
    +                    pass
    
    1722
    +
    
    1723
    +            # Write some logs out to normal directories: logsdir and metadir
    
    1724
    +            # Copy build log
    
    1725
    +            log_filename = context.get_log_filename()
    
    1726
    +            self._build_log_path = os.path.join(logsdir, 'build.log')
    
    1727
    +            if log_filename:
    
    1728
    +                shutil.copyfile(log_filename, self._build_log_path)
    
    1729
    +
    
    1730
    +            # Store public data
    
    1731
    +            _yaml.dump(_yaml.node_sanitize(self.__dynamic_public), os.path.join(metadir, 'public.yaml'))
    
    1732
    +
    
    1733
    +            # Store result
    
    1734
    +            build_result_dict = {"success": self.__build_result[0], "description": self.__build_result[1]}
    
    1735
    +            if self.__build_result[2] is not None:
    
    1736
    +                build_result_dict["detail"] = self.__build_result[2]
    
    1737
    +            _yaml.dump(build_result_dict, os.path.join(metadir, 'build-result.yaml'))
    
    1738
    +
    
    1739
    +            # ensure we have cache keys
    
    1740
    +            self._assemble_done()
    
    1741
    +
    
    1742
    +            # Store keys.yaml
    
    1743
    +            _yaml.dump(_yaml.node_sanitize({
    
    1744
    +                'strong': self._get_cache_key(),
    
    1745
    +                'weak': self._get_cache_key(_KeyStrength.WEAK),
    
    1746
    +            }), os.path.join(metadir, 'keys.yaml'))
    
    1747
    +
    
    1748
    +            # Store dependencies.yaml
    
    1749
    +            _yaml.dump(_yaml.node_sanitize({
    
    1750
    +                e.name: e._get_cache_key() for e in self.dependencies(Scope.BUILD)
    
    1751
    +            }), os.path.join(metadir, 'dependencies.yaml'))
    
    1752
    +
    
    1753
    +            # Store workspaced.yaml
    
    1754
    +            _yaml.dump(_yaml.node_sanitize({
    
    1755
    +                'workspaced': bool(self._get_workspace())
    
    1756
    +            }), os.path.join(metadir, 'workspaced.yaml'))
    
    1757
    +
    
    1758
    +            # Store workspaced-dependencies.yaml
    
    1759
    +            _yaml.dump(_yaml.node_sanitize({
    
    1760
    +                'workspaced-dependencies': [
    
    1761
    +                    e.name for e in self.dependencies(Scope.BUILD)
    
    1762
    +                    if e._get_workspace()
    
    1763
    +                ]
    
    1764
    +            }), os.path.join(metadir, 'workspaced-dependencies.yaml'))
    
    1721 1765
     
    
    1722
    -        # Copy build log
    
    1723
    -        log_filename = context.get_log_filename()
    
    1724
    -        self._build_log_path = os.path.join(logsdir, 'build.log')
    
    1725
    -        if log_filename:
    
    1726
    -            shutil.copyfile(log_filename, self._build_log_path)
    
    1727
    -
    
    1728
    -        # Store public data
    
    1729
    -        _yaml.dump(_yaml.node_sanitize(self.__dynamic_public), os.path.join(metadir, 'public.yaml'))
    
    1730
    -
    
    1731
    -        # Store result
    
    1732
    -        build_result_dict = {"success": self.__build_result[0], "description": self.__build_result[1]}
    
    1733
    -        if self.__build_result[2] is not None:
    
    1734
    -            build_result_dict["detail"] = self.__build_result[2]
    
    1735
    -        _yaml.dump(build_result_dict, os.path.join(metadir, 'build-result.yaml'))
    
    1736
    -
    
    1737
    -        # ensure we have cache keys
    
    1738
    -        self._assemble_done()
    
    1739
    -
    
    1740
    -        # Store keys.yaml
    
    1741
    -        _yaml.dump(_yaml.node_sanitize({
    
    1742
    -            'strong': self._get_cache_key(),
    
    1743
    -            'weak': self._get_cache_key(_KeyStrength.WEAK),
    
    1744
    -        }), os.path.join(metadir, 'keys.yaml'))
    
    1745
    -
    
    1746
    -        # Store dependencies.yaml
    
    1747
    -        _yaml.dump(_yaml.node_sanitize({
    
    1748
    -            e.name: e._get_cache_key() for e in self.dependencies(Scope.BUILD)
    
    1749
    -        }), os.path.join(metadir, 'dependencies.yaml'))
    
    1750
    -
    
    1751
    -        # Store workspaced.yaml
    
    1752
    -        _yaml.dump(_yaml.node_sanitize({
    
    1753
    -            'workspaced': bool(self._get_workspace())
    
    1754
    -        }), os.path.join(metadir, 'workspaced.yaml'))
    
    1755
    -
    
    1756
    -        # Store workspaced-dependencies.yaml
    
    1757
    -        _yaml.dump(_yaml.node_sanitize({
    
    1758
    -            'workspaced-dependencies': [
    
    1759
    -                e.name for e in self.dependencies(Scope.BUILD)
    
    1760
    -                if e._get_workspace()
    
    1761
    -            ]
    
    1762
    -        }), os.path.join(metadir, 'workspaced-dependencies.yaml'))
    
    1766
    +            metavdir.import_files(metadir)
    
    1767
    +            logsvdir.import_files(logsdir)
    
    1763 1768
     
    
    1764
    -        with self.timed_activity("Caching artifact"):
    
    1765
    -            artifact_size = utils._get_dir_size(assembledir)
    
    1766
    -            self.__artifacts.commit(self, assembledir, self.__get_cache_keys_for_commit())
    
    1767
    -
    
    1768
    -        if collect is not None and collectvdir is None:
    
    1769
    -            raise ElementError(
    
    1770
    -                "Directory '{}' was not found inside the sandbox, "
    
    1771
    -                "unable to collect artifact contents"
    
    1772
    -                .format(collect))
    
    1769
    +            artifact_size = assemblevdir.get_size()
    
    1770
    +            self.__artifacts.commit(self, assemblevdir, self.__get_cache_keys_for_commit())
    
    1771
    +
    
    1772
    +            if collect is not None and collectvdir is None:
    
    1773
    +                raise ElementError(
    
    1774
    +                    "Directory '{}' was not found inside the sandbox, "
    
    1775
    +                    "unable to collect artifact contents"
    
    1776
    +                    .format(collect))
    
    1773 1777
     
    
    1774 1778
             return artifact_size
    
    1775 1779
     
    

  • buildstream/storage/_casbaseddirectory.py
    ... ... @@ -136,10 +136,10 @@ class CasBasedDirectory(Directory):
    136 136
             the parent).
    
    137 137
     
    
    138 138
             """
    
    139
    -        self.ref = self.cas_cache.add_object(buffer=self.pb2_directory.SerializeToString())
    
    140 139
             if caller:
    
    141 140
                 old_dir = self._find_pb2_entry(caller.filename)
    
    142 141
                 self.cas_cache.add_object(digest=old_dir.digest, buffer=caller.pb2_directory.SerializeToString())
    
    142
    +        self.ref = self.cas_cache.add_object(buffer=self.pb2_directory.SerializeToString())
    
    143 143
             if self.parent:
    
    144 144
                 self.parent._recalculate_recursing_up(self)
    
    145 145
     
    
    ... ... @@ -277,14 +277,6 @@ class CasBasedDirectory(Directory):
    277 277
                                                              directory_list))
    
    278 278
             return None
    
    279 279
     
    
    280
    -    def find_root(self):
    
    281
    -        """ Finds the root of this directory tree by following 'parent' until there is
    
    282
    -        no parent. """
    
    283
    -        if self.parent:
    
    284
    -            return self.parent.find_root()
    
    285
    -        else:
    
    286
    -            return self
    
    287
    -
    
    288 280
         def _check_replacement(self, name, path_prefix, fileListResult):
    
    289 281
             """ Checks whether 'name' exists, and if so, whether we can overwrite it.
    
    290 282
             If we can, add the name to 'overwritten_files' and delete the existing entry.
    
    ... ... @@ -451,7 +443,7 @@ class CasBasedDirectory(Directory):
    451 443
                     files = external_pathspec.list_relative_paths()
    
    452 444
     
    
    453 445
             if isinstance(external_pathspec, FileBasedDirectory):
    
    454
    -            source_directory = external_pathspec.get_underlying_directory()
    
    446
    +            source_directory = external_pathspec._get_underlying_directory()
    
    455 447
                 result = self._import_files_from_directory(source_directory, files=files)
    
    456 448
             elif isinstance(external_pathspec, str):
    
    457 449
                 source_directory = external_pathspec
    
    ... ... @@ -635,6 +627,18 @@ class CasBasedDirectory(Directory):
    635 627
             self._recalculate_recursing_up()
    
    636 628
             self._recalculate_recursing_down()
    
    637 629
     
    
    630
    +    def get_size(self):
    
    631
    +        total = len(self.pb2_directory.SerializeToString())
    
    632
    +        for i in self.index.values():
    
    633
    +            if isinstance(i.buildstream_object, CasBasedDirectory):
    
    634
    +                total += i.buildstream_object.get_size()
    
    635
    +            elif isinstance(i.pb_object, remote_execution_pb2.FileNode):
    
    636
    +                src_name = self.cas_cache.objpath(i.pb_object.digest)
    
    637
    +                filesize = os.stat(src_name).st_size
    
    638
    +                total += filesize
    
    639
    +            # Symlink nodes are encoded as part of the directory serialization.
    
    640
    +        return total
    
    641
    +
    
    638 642
         def _get_identifier(self):
    
    639 643
             path = ""
    
    640 644
             if self.parent:
    
    ... ... @@ -653,3 +657,15 @@ class CasBasedDirectory(Directory):
    653 657
             throw an exception. """
    
    654 658
             raise VirtualDirectoryError("_get_underlying_directory was called on a CAS-backed directory," +
    
    655 659
                                         " which has no underlying directory.")
    
    660
    +
    
    661
    +    # _get_digest():
    
    662
    +    #
    
    663
    +    # Return the Digest for this directory.
    
    664
    +    #
    
    665
    +    # Returns:
    
    666
    +    #   (Digest): The Digest protobuf object for the Directory protobuf
    
    667
    +    #
    
    668
    +    def _get_digest(self):
    
    669
    +        if not self.ref:
    
    670
    +            self.ref = self.cas_cache.add_object(buffer=self.pb2_directory.SerializeToString())
    
    671
    +        return self.ref

  • buildstream/storage/_filebaseddirectory.py
    ... ... @@ -30,6 +30,7 @@ See also: :ref:`sandboxing`.
    30 30
     import os
    
    31 31
     import time
    
    32 32
     from .directory import Directory, VirtualDirectoryError
    
    33
    +from .. import utils
    
    33 34
     from ..utils import link_files, copy_files, list_relative_paths, _get_link_mtime, _magic_timestamp
    
    34 35
     from ..utils import _set_deterministic_user, _set_deterministic_mtime
    
    35 36
     
    
    ... ... @@ -201,6 +202,9 @@ class FileBasedDirectory(Directory):
    201 202
     
    
    202 203
             return list_relative_paths(self.external_directory)
    
    203 204
     
    
    205
    +    def get_size(self):
    
    206
    +        return utils._get_dir_size(self.external_directory)
    
    207
    +
    
    204 208
         def __str__(self):
    
    205 209
             # This returns the whole path (since we don't know where the directory started)
    
    206 210
             # which exposes the sandbox directory; we will have to assume for the time being
    

  • buildstream/storage/directory.py
    ... ... @@ -177,3 +177,9 @@ class Directory():
    177 177
     
    
    178 178
             """
    
    179 179
             raise NotImplementedError()
    
    180
    +
    
    181
    +    def get_size(self):
    
    182
    +        """ Get an approximation of the storage space in bytes used by this directory
    
    183
    +        and all files and subdirectories in it. Storage space varies by implementation
    
    184
    +        and effective space used may be lower than this number due to deduplication. """
    
    185
    +        raise NotImplementedError()

  • requirements/Makefile
    ... ... @@ -17,5 +17,5 @@ all: $(REQUIREMENTS_TXT)
    17 17
     	$(eval VENVDIR := $(shell mktemp -d $(CURDIR)/.bst-venv.XXXXXX))
    
    18 18
     	$(VENV) $(VENVDIR)
    
    19 19
     	$(VENV_PIP) install -r $^
    
    20
    -	$(VENV_PIP) freeze -r $^ > $@
    
    20
    +	$(VENV_PIP) freeze -r $^ | grep -v pkg-resources > $@
    
    21 21
     	rm -rf $(VENVDIR)



  • [Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]