mirror of
https://github.com/pyTooling/Actions.git
synced 2026-02-15 04:26:55 +08:00
Compare commits
59 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
c81d139080 | ||
|
|
c64e054bcd | ||
|
|
78fdb584aa | ||
|
|
a456635686 | ||
|
|
befc59f22d | ||
|
|
d6fc0efd47 | ||
|
|
c018acc3c1 | ||
|
|
d74c610bb4 | ||
|
|
edc4ab3e86 | ||
|
|
0a338ae8b7 | ||
|
|
4069da0a74 | ||
|
|
679ec24c80 | ||
|
|
3a13486ea6 | ||
|
|
34fb9c9869 | ||
|
|
7523c4adca | ||
|
|
530ad7a4a1 | ||
|
|
bd3f2afaf3 | ||
|
|
b1e4cb961f | ||
|
|
1e6b71e87b | ||
|
|
5d8a608893 | ||
|
|
c3b7b3ca64 | ||
|
|
f61b77ee72 | ||
|
|
5f18024dd4 | ||
|
|
d0f07e1af4 | ||
|
|
2b5a9bdeff | ||
|
|
d324bdacee | ||
|
|
9296bd6e7d | ||
|
|
c37f727e2d | ||
|
|
a4559e8e63 | ||
|
|
bc94fba95e | ||
|
|
d6ec94fea7 | ||
|
|
4bc4ec5cf4 | ||
|
|
e21f89670a | ||
|
|
c13dd2521c | ||
|
|
cf2e89a622 | ||
|
|
a9a3c400fd | ||
|
|
970b4ae021 | ||
|
|
1807741b0a | ||
|
|
cf095afe77 | ||
|
|
12991ee38a | ||
|
|
70134b31c3 | ||
|
|
38514a9005 | ||
|
|
9ceefdbf5d | ||
|
|
b1bc6e50a3 | ||
|
|
fdee9e011f | ||
|
|
3e50c2ed5b | ||
|
|
9e0b1c69f1 | ||
|
|
f084e02f01 | ||
|
|
c34d4e240e | ||
|
|
6d04009bd6 | ||
|
|
50d32d1950 | ||
|
|
7733e8998f | ||
|
|
4c28b9d003 | ||
|
|
bafea7d082 | ||
|
|
9ca7b04f37 | ||
|
|
7a0ee75fd5 | ||
|
|
bc876f7171 | ||
|
|
edca070047 | ||
|
|
21c2f48dad |
2
.github/dependabot.yml
vendored
2
.github/dependabot.yml
vendored
@@ -10,9 +10,7 @@ updates:
|
|||||||
- Dependencies
|
- Dependencies
|
||||||
assignees:
|
assignees:
|
||||||
- Paebbels
|
- Paebbels
|
||||||
- umarcor
|
|
||||||
reviewers:
|
reviewers:
|
||||||
- Paebbels
|
- Paebbels
|
||||||
- umarcor
|
|
||||||
schedule:
|
schedule:
|
||||||
interval: "daily" # Checks on Monday trough Friday.
|
interval: "daily" # Checks on Monday trough Friday.
|
||||||
|
|||||||
15
.github/workflows/ApplicationTesting.yml
vendored
15
.github/workflows/ApplicationTesting.yml
vendored
@@ -89,7 +89,7 @@ jobs:
|
|||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.wheel }}' from 'Package' job
|
- name: 📥 Download artifacts '${{ inputs.wheel }}' from 'Package' job
|
||||||
uses: actions/download-artifact@v4
|
uses: pyTooling/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.wheel }}
|
name: ${{ inputs.wheel }}
|
||||||
path: install
|
path: install
|
||||||
@@ -188,8 +188,8 @@ jobs:
|
|||||||
f.write(f"pacboy_packages={' '.join(pacboyPackages)}\n")
|
f.write(f"pacboy_packages={' '.join(pacboyPackages)}\n")
|
||||||
|
|
||||||
- name: '🟦 Setup MSYS2 for ${{ matrix.runtime }}'
|
- name: '🟦 Setup MSYS2 for ${{ matrix.runtime }}'
|
||||||
if: matrix.system == 'msys2'
|
|
||||||
uses: msys2/setup-msys2@v2
|
uses: msys2/setup-msys2@v2
|
||||||
|
if: matrix.system == 'msys2'
|
||||||
with:
|
with:
|
||||||
msystem: ${{ matrix.runtime }}
|
msystem: ${{ matrix.runtime }}
|
||||||
update: true
|
update: true
|
||||||
@@ -198,8 +198,8 @@ jobs:
|
|||||||
${{ inputs.pacboy }}
|
${{ inputs.pacboy }}
|
||||||
|
|
||||||
- name: 🐍 Setup Python ${{ matrix.python }}
|
- name: 🐍 Setup Python ${{ matrix.python }}
|
||||||
if: matrix.system != 'msys2'
|
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
|
if: matrix.system != 'msys2'
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python }}
|
python-version: ${{ matrix.python }}
|
||||||
|
|
||||||
@@ -231,10 +231,10 @@ jobs:
|
|||||||
cd "${{ inputs.root_directory || '.' }}"
|
cd "${{ inputs.root_directory || '.' }}"
|
||||||
[ -n '${{ inputs.apptest_xml_artifact }}' ] && PYTEST_ARGS='--junitxml=report/unit/TestReportSummary.xml' || unset PYTEST_ARGS
|
[ -n '${{ inputs.apptest_xml_artifact }}' ] && PYTEST_ARGS='--junitxml=report/unit/TestReportSummary.xml' || unset PYTEST_ARGS
|
||||||
if [ -n '${{ inputs.coverage_config }}' ]; then
|
if [ -n '${{ inputs.coverage_config }}' ]; then
|
||||||
echo "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}"
|
printf "%s\n" "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}"
|
||||||
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}
|
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}
|
||||||
else
|
else
|
||||||
echo "python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}"
|
printf "%s\n" "python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}"
|
||||||
python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}
|
python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.apptest_directory }}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -255,9 +255,10 @@ jobs:
|
|||||||
|
|
||||||
- name: 📤 Upload 'TestReportSummary.xml' artifact
|
- name: 📤 Upload 'TestReportSummary.xml' artifact
|
||||||
if: inputs.apptest_xml_artifact != ''
|
if: inputs.apptest_xml_artifact != ''
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.apptest_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.apptest_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: report/unit/TestReportSummary.xml
|
working-directory: report/unit
|
||||||
|
path: TestReportSummary.xml
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|||||||
4
.github/workflows/ArtifactCleanUp.yml
vendored
4
.github/workflows/ArtifactCleanUp.yml
vendored
@@ -47,13 +47,13 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: 🗑️ Delete package Artifacts
|
- name: 🗑️ Delete package Artifacts
|
||||||
if: ${{ ! startsWith(github.ref, 'refs/tags') }}
|
|
||||||
uses: geekyeggo/delete-artifact@v5
|
uses: geekyeggo/delete-artifact@v5
|
||||||
|
if: ${{ ! startsWith(github.ref, 'refs/tags') }}
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.package }}
|
name: ${{ inputs.package }}
|
||||||
|
|
||||||
- name: 🗑️ Delete remaining Artifacts
|
- name: 🗑️ Delete remaining Artifacts
|
||||||
if: ${{ inputs.remaining != '' }}
|
|
||||||
uses: geekyeggo/delete-artifact@v5
|
uses: geekyeggo/delete-artifact@v5
|
||||||
|
if: ${{ inputs.remaining != '' }}
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.remaining }}
|
name: ${{ inputs.remaining }}
|
||||||
|
|||||||
7
.github/workflows/BuildTheDocs.yml
vendored
7
.github/workflows/BuildTheDocs.yml
vendored
@@ -38,7 +38,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: '❗ Deprecation message'
|
- name: '❗ Deprecation message'
|
||||||
run: echo "::warning title=Deprecated::'BuildTheDocs.yml' is not maintained anymore. Please switch to 'SphinxDocumentation.yml', 'LaTeXDocumentation.yml' and 'ExtractConfiguration.yml'."
|
run: printf "%s\n" "::warning title=Deprecated::'BuildTheDocs.yml' is not maintained anymore. Please switch to 'SphinxDocumentation.yml', 'LaTeXDocumentation.yml' and 'ExtractConfiguration.yml'."
|
||||||
|
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
@@ -49,11 +49,12 @@ jobs:
|
|||||||
skip-deploy: true
|
skip-deploy: true
|
||||||
|
|
||||||
- name: 📤 Upload 'documentation' artifacts
|
- name: 📤 Upload 'documentation' artifacts
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.artifact != ''
|
if: inputs.artifact != ''
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.artifact }}
|
name: ${{ inputs.artifact }}
|
||||||
path: doc/_build/html
|
working-directory: doc/_build/html
|
||||||
|
path: '*'
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: '📓 Publish site to GitHub Pages'
|
- name: '📓 Publish site to GitHub Pages'
|
||||||
|
|||||||
6
.github/workflows/CheckDocumentation.yml
vendored
6
.github/workflows/CheckDocumentation.yml
vendored
@@ -59,14 +59,14 @@ jobs:
|
|||||||
|
|
||||||
- name: 🔧 Install wheel,tomli and pip dependencies (native)
|
- name: 🔧 Install wheel,tomli and pip dependencies (native)
|
||||||
run: |
|
run: |
|
||||||
python -m pip install --disable-pip-version-check -U docstr_coverage interrogate
|
python -m pip install --disable-pip-version-check -U docstr_coverage interrogate[png]
|
||||||
|
|
||||||
- name: Run 'interrogate' Documentation Coverage Check
|
- name: Run 'interrogate' Documentation Coverage Check
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
interrogate -c pyproject.toml --fail-under=${{ inputs.fail_under }} && echo "::error title=interrogate::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
interrogate -c pyproject.toml --fail-under=${{ inputs.fail_under }} && printf "%s\n" "::error title=interrogate::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
||||||
|
|
||||||
- name: Run 'docstr_coverage' Documentation Coverage Check
|
- name: Run 'docstr_coverage' Documentation Coverage Check
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
docstr-coverage -v 2 --fail-under=${{ inputs.fail_under }} ${{ inputs.directory }} && echo "::error title=docstr-coverage::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
docstr-coverage -v 2 --fail-under=${{ inputs.fail_under }} ${{ inputs.directory }} && printf "%s\n" "::error title=docstr-coverage::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
||||||
|
|||||||
33
.github/workflows/CompletePipeline.yml
vendored
33
.github/workflows/CompletePipeline.yml
vendored
@@ -1,3 +1,24 @@
|
|||||||
|
# ==================================================================================================================== #
|
||||||
|
# Authors: #
|
||||||
|
# Patrick Lehmann #
|
||||||
|
# #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
# Copyright 2020-2024 The pyTooling Authors #
|
||||||
|
# #
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
|
# you may not use this file except in compliance with the License. #
|
||||||
|
# You may obtain a copy of the License at #
|
||||||
|
# #
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0 #
|
||||||
|
# #
|
||||||
|
# Unless required by applicable law or agreed to in writing, software #
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, #
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #
|
||||||
|
# See the License for the specific language governing permissions and #
|
||||||
|
# limitations under the License. #
|
||||||
|
# #
|
||||||
|
# SPDX-License-Identifier: Apache-2.0 #
|
||||||
|
# ==================================================================================================================== #
|
||||||
name: Namespace Package
|
name: Namespace Package
|
||||||
|
|
||||||
on:
|
on:
|
||||||
@@ -114,13 +135,17 @@ jobs:
|
|||||||
UnitTesting:
|
UnitTesting:
|
||||||
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- ConfigParams
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
jobs: ${{ needs.UnitTestingParams.outputs.python_jobs }}
|
jobs: ${{ needs.UnitTestingParams.outputs.python_jobs }}
|
||||||
requirements: "-r tests/unit/requirements.txt"
|
requirements: "-r tests/unit/requirements.txt"
|
||||||
# pacboy: "msys/git python-lxml:p"
|
# pacboy: "msys/git python-lxml:p"
|
||||||
unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
unittest_report_xml_directory: ${{ needs.ConfigParams.outputs.unittest_report_xml_directory }}
|
||||||
coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
unittest_report_xml_filename: ${{ needs.ConfigParams.outputs.unittest_report_xml_filename }}
|
||||||
|
coverage_report_html_directory: ${{ needs.ConfigParams.outputs.coverage_report_html_directory }}
|
||||||
|
unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
|
coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
||||||
|
|
||||||
StaticTypeCheck:
|
StaticTypeCheck:
|
||||||
uses: pyTooling/Actions/.github/workflows/StaticTypeCheck.yml@main
|
uses: pyTooling/Actions/.github/workflows/StaticTypeCheck.yml@main
|
||||||
@@ -138,6 +163,7 @@ jobs:
|
|||||||
DocCoverage:
|
DocCoverage:
|
||||||
uses: pyTooling/Actions/.github/workflows/CheckDocumentation.yml@main
|
uses: pyTooling/Actions/.github/workflows/CheckDocumentation.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- ConfigParams
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
@@ -183,6 +209,7 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
with:
|
with:
|
||||||
|
additional_merge_args: '-d "--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit"'
|
||||||
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
|
|
||||||
# VerifyDocs:
|
# VerifyDocs:
|
||||||
@@ -195,8 +222,8 @@ jobs:
|
|||||||
Documentation:
|
Documentation:
|
||||||
uses: pyTooling/Actions/.github/workflows/SphinxDocumentation.yml@main
|
uses: pyTooling/Actions/.github/workflows/SphinxDocumentation.yml@main
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
|
||||||
- ConfigParams
|
- ConfigParams
|
||||||
|
- UnitTestingParams
|
||||||
- PublishTestResults
|
- PublishTestResults
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
# - VerifyDocs
|
# - VerifyDocs
|
||||||
|
|||||||
14
.github/workflows/CoverageCollection.yml
vendored
14
.github/workflows/CoverageCollection.yml
vendored
@@ -72,10 +72,13 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: '❗ Deprecation message'
|
- name: '❗ Deprecation message'
|
||||||
run: echo "::warning title=Deprecated::'CoverageCollection.yml' is not maintained anymore. Please switch to 'UnitTesting.yml', 'PublishCoverageResults.yml' and 'PublishTestResults.yml'."
|
run: printf "%s\n" "::warning title=Deprecated::'CoverageCollection.yml' is not maintained anymore. Please switch to 'UnitTesting.yml', 'PublishCoverageResults.yml' and 'PublishTestResults.yml'."
|
||||||
|
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
@@ -147,7 +150,7 @@ jobs:
|
|||||||
ABSDIR=$(pwd)
|
ABSDIR=$(pwd)
|
||||||
cd "${{ inputs.tests_directory || '.' }}"
|
cd "${{ inputs.tests_directory || '.' }}"
|
||||||
[ -n '${{ inputs.coverage_config }}' ] && PYCOV_ARGS="--cov-config=${ABSDIR}/${{ inputs.coverage_config }}" || unset PYCOV_ARGS
|
[ -n '${{ inputs.coverage_config }}' ] && PYCOV_ARGS="--cov-config=${ABSDIR}/${{ inputs.coverage_config }}" || unset PYCOV_ARGS
|
||||||
echo "python -m pytest -rA --cov=${ABSDIR} ${PYCOV_ARGS} ${{ inputs.unittest_directory }} --color=yes"
|
printf "%s\n" "python -m pytest -rA --cov=${ABSDIR} ${PYCOV_ARGS} ${{ inputs.unittest_directory }} --color=yes"
|
||||||
python -m pytest -rA --cov=${ABSDIR} $PYCOV_ARGS ${{ inputs.unittest_directory }} --color=yes
|
python -m pytest -rA --cov=${ABSDIR} $PYCOV_ARGS ${{ inputs.unittest_directory }} --color=yes
|
||||||
|
|
||||||
- name: Convert to cobertura format
|
- name: Convert to cobertura format
|
||||||
@@ -160,16 +163,17 @@ jobs:
|
|||||||
|
|
||||||
- name: 📤 Upload 'Coverage Report' artifact
|
- name: 📤 Upload 'Coverage Report' artifact
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.artifact }}
|
name: ${{ inputs.artifact }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
working-directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📊 Publish coverage at CodeCov
|
- name: 📊 Publish coverage at CodeCov
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: codecov/codecov-action@v3
|
uses: codecov/codecov-action@v5
|
||||||
with:
|
with:
|
||||||
files: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
files: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
flags: unittests
|
flags: unittests
|
||||||
|
|||||||
59
.github/workflows/ExtractConfiguration.yml
vendored
59
.github/workflows/ExtractConfiguration.yml
vendored
@@ -59,18 +59,33 @@ on:
|
|||||||
mypy_prepare_command:
|
mypy_prepare_command:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.mypy_prepare_command }}
|
value: ${{ jobs.Extract.outputs.mypy_prepare_command }}
|
||||||
|
unittest_report_xml_directory:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Extract.outputs.unittest_report_xml_directory }}
|
||||||
|
unittest_report_xml_filename:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Extract.outputs.unittest_report_xml_filename }}
|
||||||
|
unittest_report_xml:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Extract.outputs.unittest_report_xml }}
|
||||||
coverage_report_html_directory:
|
coverage_report_html_directory:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.coverage_report_html_directory }}
|
value: ${{ jobs.Extract.outputs.coverage_report_html_directory }}
|
||||||
coverage_report_xml_directory:
|
coverage_report_xml_directory:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.coverage_report_xml_directory }}
|
value: ${{ jobs.Extract.outputs.coverage_report_xml_directory }}
|
||||||
|
coverage_report_xml_filename:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Extract.outputs.coverage_report_xml_filename }}
|
||||||
coverage_report_xml:
|
coverage_report_xml:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.coverage_report_xml }}
|
value: ${{ jobs.Extract.outputs.coverage_report_xml }}
|
||||||
coverage_report_json_directory:
|
coverage_report_json_directory:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.coverage_report_json_directory }}
|
value: ${{ jobs.Extract.outputs.coverage_report_json_directory }}
|
||||||
|
coverage_report_json_filename:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Extract.outputs.coverage_report_json_filename }}
|
||||||
coverage_report_json:
|
coverage_report_json:
|
||||||
description: ""
|
description: ""
|
||||||
value: ${{ jobs.Extract.outputs.coverage_report_json }}
|
value: ${{ jobs.Extract.outputs.coverage_report_json }}
|
||||||
@@ -83,10 +98,15 @@ jobs:
|
|||||||
package_fullname: ${{ steps.getPackageName.outputs.package_fullname }}
|
package_fullname: ${{ steps.getPackageName.outputs.package_fullname }}
|
||||||
package_directory: ${{ steps.getPackageName.outputs.package_directory }}
|
package_directory: ${{ steps.getPackageName.outputs.package_directory }}
|
||||||
mypy_prepare_command: ${{ steps.getPackageName.outputs.mypy_prepare_command }}
|
mypy_prepare_command: ${{ steps.getPackageName.outputs.mypy_prepare_command }}
|
||||||
|
unittest_report_xml_directory: ${{ steps.getVariables.outputs.unittest_report_xml_directory }}
|
||||||
|
unittest_report_xml_filename: ${{ steps.getVariables.outputs.unittest_report_xml_filename }}
|
||||||
|
unittest_report_xml: ${{ steps.getVariables.outputs.unittest_report_xml }}
|
||||||
coverage_report_html_directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
coverage_report_html_directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
||||||
coverage_report_xml_directory: ${{ steps.getVariables.outputs.coverage_report_xml_directory }}
|
coverage_report_xml_directory: ${{ steps.getVariables.outputs.coverage_report_xml_directory }}
|
||||||
|
coverage_report_xml_filename: ${{ steps.getVariables.outputs.coverage_report_xml_filename }}
|
||||||
coverage_report_xml: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
coverage_report_xml: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
coverage_report_json_directory: ${{ steps.getVariables.outputs.coverage_report_json_directory }}
|
coverage_report_json_directory: ${{ steps.getVariables.outputs.coverage_report_json_directory }}
|
||||||
|
coverage_report_json_filename: ${{ steps.getVariables.outputs.coverage_report_json_filename }}
|
||||||
coverage_report_json: ${{ steps.getVariables.outputs.coverage_report_json }}
|
coverage_report_json: ${{ steps.getVariables.outputs.coverage_report_json }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
@@ -144,10 +164,11 @@ jobs:
|
|||||||
|
|
||||||
from tomli import load as tomli_load
|
from tomli import load as tomli_load
|
||||||
|
|
||||||
htmlDirectory = Path("htmlcov")
|
unittestXMLFile = Path("./unittest.xml")
|
||||||
xmlFile = Path("./coverage.xml")
|
coverageHTMLDirectory = Path("htmlcov")
|
||||||
jsonFile = Path("./coverage.json")
|
coverageXMLFile = Path("./coverage.xml")
|
||||||
coverageRC = "${{ inputs.coverage_config }}".strip()
|
coverageJSONFile = Path("./coverage.json")
|
||||||
|
coverageRC = "${{ inputs.coverage_config }}".strip()
|
||||||
|
|
||||||
# Read output paths from 'pyproject.toml' file
|
# Read output paths from 'pyproject.toml' file
|
||||||
if coverageRC == "pyproject.toml":
|
if coverageRC == "pyproject.toml":
|
||||||
@@ -156,9 +177,10 @@ jobs:
|
|||||||
with pyProjectFile.open("rb") as file:
|
with pyProjectFile.open("rb") as file:
|
||||||
pyProjectSettings = tomli_load(file)
|
pyProjectSettings = tomli_load(file)
|
||||||
|
|
||||||
htmlDirectory = Path(pyProjectSettings["tool"]["coverage"]["html"]["directory"])
|
unittestXMLFile = Path(pyProjectSettings["tool"]["pytest"]["junit_xml"])
|
||||||
xmlFile = Path(pyProjectSettings["tool"]["coverage"]["xml"]["output"])
|
coverageHTMLDirectory = Path(pyProjectSettings["tool"]["coverage"]["html"]["directory"])
|
||||||
jsonFile = Path(pyProjectSettings["tool"]["coverage"]["json"]["output"])
|
coverageXMLFile = Path(pyProjectSettings["tool"]["coverage"]["xml"]["output"])
|
||||||
|
coverageJSONFile= Path(pyProjectSettings["tool"]["coverage"]["json"]["output"])
|
||||||
else:
|
else:
|
||||||
print(f"File '{pyProjectFile}' not found.")
|
print(f"File '{pyProjectFile}' not found.")
|
||||||
print(f"::error title=FileNotFoundError::File '{pyProjectFile}' not found.")
|
print(f"::error title=FileNotFoundError::File '{pyProjectFile}' not found.")
|
||||||
@@ -171,9 +193,9 @@ jobs:
|
|||||||
with coverageRCFile.open("rb") as file:
|
with coverageRCFile.open("rb") as file:
|
||||||
coverageRCSettings = tomli_load(file)
|
coverageRCSettings = tomli_load(file)
|
||||||
|
|
||||||
htmlDirectory = Path(coverageRCSettings["html"]["directory"])
|
coverageHTMLDirectory = Path(coverageRCSettings["html"]["directory"])
|
||||||
xmlFile = Path(coverageRCSettings["xml"]["output"])
|
coverageXMLFile = Path(coverageRCSettings["xml"]["output"])
|
||||||
jsonFile = Path(coverageRCSettings["json"]["output"])
|
coverageJSONFile = Path(coverageRCSettings["json"]["output"])
|
||||||
else:
|
else:
|
||||||
print(f"File '{coverageRCFile}' not found.")
|
print(f"File '{coverageRCFile}' not found.")
|
||||||
print(f"::error title=FileNotFoundError::File '{coverageRCFile}' not found.")
|
print(f"::error title=FileNotFoundError::File '{coverageRCFile}' not found.")
|
||||||
@@ -184,11 +206,16 @@ jobs:
|
|||||||
print(f"GITHUB_OUTPUT: {github_output}")
|
print(f"GITHUB_OUTPUT: {github_output}")
|
||||||
with github_output.open("a+", encoding="utf-8") as f:
|
with github_output.open("a+", encoding="utf-8") as f:
|
||||||
f.write(dedent(f"""\
|
f.write(dedent(f"""\
|
||||||
coverage_report_html_directory={htmlDirectory.as_posix()}
|
unittest_report_xml_directory={unittestXMLFile.parent.as_posix()}
|
||||||
coverage_report_xml_directory={xmlFile.parent.as_posix()}
|
unittest_report_xml_filename={unittestXMLFile.name}
|
||||||
coverage_report_xml={xmlFile.as_posix()}
|
unittest_report_xml={unittestXMLFile.as_posix()}
|
||||||
coverage_report_json_directory={jsonFile.parent.as_posix()}
|
coverage_report_html_directory={coverageHTMLDirectory.as_posix()}
|
||||||
coverage_report_json={jsonFile.as_posix()}
|
coverage_report_xml_directory={coverageXMLFile.parent.as_posix()}
|
||||||
|
coverage_report_xml_filename={coverageXMLFile.name}
|
||||||
|
coverage_report_xml={coverageXMLFile.as_posix()}
|
||||||
|
coverage_report_json_directory={coverageJSONFile.parent.as_posix()}
|
||||||
|
coverage_report_json_filename={coverageJSONFile.name}
|
||||||
|
coverage_report_json={coverageJSONFile.as_posix()}
|
||||||
"""))
|
"""))
|
||||||
|
|
||||||
print(f"DEBUG:\n html={htmlDirectory}\n xml={xmlFile}\n json={jsonFile}")
|
print(f"DEBUG:\n unittest xml: {unittestXMLFile}\n coverage html: {coverageHTMLDirectory}\n coverage xml: {coverageXMLFile}\n coverage json: {coverageJSONFile}")
|
||||||
|
|||||||
23
.github/workflows/LaTeXDocumentation.yml
vendored
23
.github/workflows/LaTeXDocumentation.yml
vendored
@@ -50,22 +50,33 @@ jobs:
|
|||||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
||||||
steps:
|
steps:
|
||||||
- name: 📥 Download artifacts '${{ inputs.latex_artifact }}' from 'SphinxDocumentation' job
|
- name: 📥 Download artifacts '${{ inputs.latex_artifact }}' from 'SphinxDocumentation' job
|
||||||
uses: actions/download-artifact@v4
|
uses: pyTooling/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.latex_artifact }}
|
name: ${{ inputs.latex_artifact }}
|
||||||
path: latex
|
path: latex
|
||||||
|
|
||||||
- name: Compile LaTeX document
|
- name: Debug
|
||||||
uses: xu-cheng/latex-action@master
|
run: |
|
||||||
|
tree -pash .
|
||||||
|
|
||||||
|
- name: Build LaTeX document using 'pytooling/miktex:sphinx'
|
||||||
|
uses: addnab/docker-run-action@v3
|
||||||
with:
|
with:
|
||||||
working_directory: latex
|
image: pytooling/miktex:sphinx
|
||||||
root_file: ${{ inputs.document }}.tex
|
options: -v ${{ github.workspace }}/latex:/latex --workdir /latex
|
||||||
|
run: |
|
||||||
|
which pdflatex
|
||||||
|
pwd
|
||||||
|
ls -lAh
|
||||||
|
|
||||||
|
latexmk -xelatex ${{ inputs.document }}.tex
|
||||||
|
|
||||||
- name: 📤 Upload 'PDF Documentation' artifact
|
- name: 📤 Upload 'PDF Documentation' artifact
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.pdf_artifact != ''
|
if: inputs.pdf_artifact != ''
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.pdf_artifact }}
|
name: ${{ inputs.pdf_artifact }}
|
||||||
|
working-directory: latex
|
||||||
path: ${{ inputs.document }}.pdf
|
path: ${{ inputs.document }}.pdf
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|||||||
530
.github/workflows/NightlyRelease.yml
vendored
Normal file
530
.github/workflows/NightlyRelease.yml
vendored
Normal file
@@ -0,0 +1,530 @@
|
|||||||
|
# ==================================================================================================================== #
|
||||||
|
# Authors: #
|
||||||
|
# Patrick Lehmann #
|
||||||
|
# #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
# Copyright 2020-2024 The pyTooling Authors #
|
||||||
|
# #
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
|
# you may not use this file except in compliance with the License. #
|
||||||
|
# You may obtain a copy of the License at #
|
||||||
|
# #
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0 #
|
||||||
|
# #
|
||||||
|
# Unless required by applicable law or agreed to in writing, software #
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, #
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #
|
||||||
|
# See the License for the specific language governing permissions and #
|
||||||
|
# limitations under the License. #
|
||||||
|
# #
|
||||||
|
# SPDX-License-Identifier: Apache-2.0 #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
name: Nightly
|
||||||
|
|
||||||
|
on:
|
||||||
|
workflow_call:
|
||||||
|
inputs:
|
||||||
|
ubuntu_image:
|
||||||
|
description: 'Name of the Ubuntu image.'
|
||||||
|
required: false
|
||||||
|
default: 'ubuntu-24.04'
|
||||||
|
type: string
|
||||||
|
nightly_name:
|
||||||
|
description: 'Name of the nightly release.'
|
||||||
|
required: false
|
||||||
|
default: 'nightly'
|
||||||
|
type: string
|
||||||
|
nightly_title:
|
||||||
|
description: 'Title of the nightly release.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
nightly_description:
|
||||||
|
description: 'Description of the nightly release.'
|
||||||
|
required: false
|
||||||
|
default: 'Release of artifacts from latest CI pipeline.'
|
||||||
|
type: string
|
||||||
|
draft:
|
||||||
|
description: 'Specify if this is a draft.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
prerelease:
|
||||||
|
description: 'Specify if this is a pre-release.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
latest:
|
||||||
|
description: 'Specify if this is the latest release.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
replacements:
|
||||||
|
description: 'Multi-line string containing search=replace patterns.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
assets:
|
||||||
|
description: 'Multi-line string containing artifact:file:title asset descriptions.'
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
inventory-json:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
inventory-version:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
inventory-categories:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
tarball-name:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: '__pyTooling_upload_artifact__.tar'
|
||||||
|
can-fail:
|
||||||
|
type: boolean
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
Release:
|
||||||
|
name: 📝 Update 'Nightly Page' on GitHub
|
||||||
|
runs-on: ${{ inputs.ubuntu_image }}
|
||||||
|
continue-on-error: ${{ inputs.can-fail }}
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
# attestations: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: ⏬ Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
# The command 'git describe' (used for version) needs the history.
|
||||||
|
fetch-depth: 0
|
||||||
|
|
||||||
|
- name: 🔧 Install zstd
|
||||||
|
run: sudo apt-get install -y --no-install-recommends zstd
|
||||||
|
|
||||||
|
- name: 📑 Delete (old) Release Page
|
||||||
|
id: deleteReleasePage
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
printf "%s" "Deleting release '${{ inputs.nightly_name }}' ... "
|
||||||
|
message="$(gh release delete ${{ inputs.nightly_name }} --yes 2>&1)"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
elif [[ "${message}" == "release not found" ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_YELLOW}[NOT FOUND]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=InternalError::Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📑 (Re)create (new) Release Page
|
||||||
|
id: createReleasePage
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
addDraft="--draft"
|
||||||
|
|
||||||
|
if [[ "${{ inputs.prerelease }}" == "true" ]]; then
|
||||||
|
addPreRelease="--prerelease"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.latest }}" == "false" ]]; then
|
||||||
|
addLatest="--latest=false"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.nightly_title }}" != "" ]]; then
|
||||||
|
addTitle=("--title" "${{ inputs.nightly_title }}")
|
||||||
|
fi
|
||||||
|
|
||||||
|
cat <<'EOF' > __NoTeS__.md
|
||||||
|
${{ inputs.nightly_description }}
|
||||||
|
EOF
|
||||||
|
if [[ -s __NoTeS__.md ]]; then
|
||||||
|
addNotes=("--notes-file" "__NoTeS__.md")
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Apply replacements
|
||||||
|
while IFS=$'\r\n' read -r patternLine; do
|
||||||
|
# skip empty lines
|
||||||
|
[[ "$patternLine" == "" ]] && continue
|
||||||
|
|
||||||
|
pattern="${patternLine%%=*}"
|
||||||
|
replacement="${patternLine#*=}"
|
||||||
|
sed -i -e "s/%$pattern%/$replacement/g" "__NoTeS__.md"
|
||||||
|
done <<<'${{ inputs.replacements }}'
|
||||||
|
|
||||||
|
# Add footer line
|
||||||
|
cat <<EOF >> __NoTeS__.md
|
||||||
|
|
||||||
|
--------
|
||||||
|
Published from [${{ github.workflow }}](https://github.com/Paebbels/ghdl/actions/runs/${{ github.run_id }}) workflow triggered by @${{ github.actor }} on $(date '+%Y-%m-%d %H:%M:%S %Z').
|
||||||
|
EOF
|
||||||
|
|
||||||
|
printf "%s\n" "Creating release '${{ inputs.nightly_name }}' ... "
|
||||||
|
message="$(gh release create "${{ inputs.nightly_name }}" --verify-tag $addDraft $addPreRelease $addLatest "${addTitle[@]}" "${addNotes[@]}" 2>&1)"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=InternalError::Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📥 Download artifacts and upload as assets
|
||||||
|
id: uploadAssets
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
Replace() {
|
||||||
|
line="$1"
|
||||||
|
while IFS=$'\r\n' read -r patternLine; do
|
||||||
|
# skip empty lines
|
||||||
|
[[ "$patternLine" == "" ]] && continue
|
||||||
|
|
||||||
|
pattern="${patternLine%%=*}"
|
||||||
|
replacement="${patternLine#*=}"
|
||||||
|
line="${line//"%$pattern%"/"$replacement"}"
|
||||||
|
done <<<'${{ inputs.replacements }}'
|
||||||
|
printf "%s\n" "$line"
|
||||||
|
}
|
||||||
|
|
||||||
|
# Create JSON inventory
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
VERSION="1.0"
|
||||||
|
|
||||||
|
# Split categories by ',' into a Bash array.
|
||||||
|
# See https://stackoverflow.com/a/45201229/3719459
|
||||||
|
if [[ "${{ inputs.inventory-categories }}" != "" ]]; then
|
||||||
|
readarray -td, inventoryCategories <<<"${{ inputs.inventory-categories }},"
|
||||||
|
unset 'inventoryCategories[-1]'
|
||||||
|
declare -p inventoryCategories
|
||||||
|
else
|
||||||
|
inventoryCategories=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
jsonInventory=$(jq -c -n \
|
||||||
|
--arg version "${VERSION}" \
|
||||||
|
--arg date "$(date +"%Y-%m-%dT%H-%M-%S%:z")" \
|
||||||
|
--argjson jsonMeta "$(jq -c -n \
|
||||||
|
--arg tag "${{ inputs.nightly_name }}" \
|
||||||
|
--arg version "${{ inputs.inventory-version }}" \
|
||||||
|
--arg hash "${{ github.sha }}" \
|
||||||
|
--arg repo "${{ github.server_url }}/${{ github.repository }}" \
|
||||||
|
--arg release "${{ github.server_url }}/${{ github.repository }}/releases/download/${{ inputs.nightly_name }}" \
|
||||||
|
--argjson categories "$(jq -c -n \
|
||||||
|
'$ARGS.positional' \
|
||||||
|
--args "${inventoryCategories[@]}" \
|
||||||
|
)" \
|
||||||
|
'{"tag": $tag, "version": $version, "git-hash": $hash, "repository-url": $repo, "release-url": $release, "categories": $categories}' \
|
||||||
|
)" \
|
||||||
|
'{"version": 1.0, "timestamp": $date, "meta": $jsonMeta, "files": {}}'
|
||||||
|
)
|
||||||
|
fi
|
||||||
|
|
||||||
|
ERRORS=0
|
||||||
|
# A dictionary of 0/1 to avoid duplicate downloads
|
||||||
|
declare -A downloadedArtifacts
|
||||||
|
# A dictionary to check for duplicate asset files in release
|
||||||
|
declare -A assetFilenames
|
||||||
|
while IFS=$'\r\n' read -r assetLine; do
|
||||||
|
if [[ "${assetLine}" == "" || "${assetLine:0:1}" == "#" ]]; then
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
# split assetLine colon separated triple: artifact:asset:title
|
||||||
|
artifact="${assetLine%%:*}"
|
||||||
|
assetLine="${assetLine#*:}"
|
||||||
|
asset="${assetLine%%:*}"
|
||||||
|
assetLine="${assetLine#*:}"
|
||||||
|
if [[ "${{ inputs.inventory-json }}" == "" ]]; then
|
||||||
|
categories=""
|
||||||
|
title="${assetLine##*:}"
|
||||||
|
else
|
||||||
|
categories="${assetLine%%:*}"
|
||||||
|
title="${assetLine##*:}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# remove leading whitespace
|
||||||
|
asset="${asset#"${asset%%[![:space:]]*}"}"
|
||||||
|
categories="${categories#"${categories%%[![:space:]]*}"}"
|
||||||
|
title="${title#"${title%%[![:space:]]*}"}"
|
||||||
|
|
||||||
|
# apply replacements
|
||||||
|
asset="$(Replace "${asset}")"
|
||||||
|
title="$(Replace "${title}")"
|
||||||
|
|
||||||
|
printf "%s\n" "Publish asset '${asset}' from artifact '${artifact}' with title '${title}'"
|
||||||
|
printf " %s" "Checked asset for duplicates ... "
|
||||||
|
if [[ -n "${assetFilenames[$asset]}" ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=DuplicateAsset::Asset '${asset}' from artifact '${artifact}' was already uploaded to release '${{ inputs.nightly_name }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
assetFilenames[$asset]=1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Download artifact by artifact name
|
||||||
|
if [[ -n "${downloadedArtifacts[$artifact]}" ]]; then
|
||||||
|
printf " %s\n" "downloading '${artifact}' ... ${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
echo " downloading '${artifact}' ... "
|
||||||
|
printf " %s" "gh run download $GITHUB_RUN_ID --dir \"${artifact}\" --name \"${artifact}\" "
|
||||||
|
gh run download $GITHUB_RUN_ID --dir "${artifact}" --name "${artifact}"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't download artifact '${artifact}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=ArtifactNotFound::Couldn't download artifact '${artifact}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
downloadedArtifacts[$artifact]=1
|
||||||
|
|
||||||
|
printf " %s" "Checking for embedded tarball ... "
|
||||||
|
if [[ -f "${artifact}/${{ inputs.tarball-name }}" ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[FOUND]${ANSI_NOCOLOR}"
|
||||||
|
|
||||||
|
pushd "${artifact}" > /dev/null
|
||||||
|
|
||||||
|
printf " %s" "Extracting embedded tarball ... "
|
||||||
|
tar -xf "${{ inputs.tarball-name }}"
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf " %s" "Removing temporary tarball ... "
|
||||||
|
rm -f "${{ inputs.tarball-name }}"
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
popd > /dev/null
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check if artifact should be compressed (zip, tgz) or if asset was part of the downloaded artifact.
|
||||||
|
printf " %s" "checking asset '${artifact}/${asset}' ... "
|
||||||
|
if [[ "${asset}" == !*.zip ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[ZIP]${ANSI_NOCOLOR}"
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf "::group:: %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
zip -r "../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zip file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to zip file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ "${asset}" == !*.tgz || "${asset}" == !*.tar.gz || "${asset}" == \$*.tgz || "${asset}" == \$*.tar.gz ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[TAR/GZ]${ANSI_NOCOLOR}"
|
||||||
|
|
||||||
|
if [[ "${asset:0:1}" == "\$" ]]; then
|
||||||
|
asset="${asset##*$}"
|
||||||
|
dirName="${asset%.*}"
|
||||||
|
printf " %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
tar -c --gzip --owner=0 --group=0 --file="${asset}" --directory="${artifact}" --transform "s|^\.|${dirName%.tar}|" .
|
||||||
|
retCode=$?
|
||||||
|
else
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf " %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
tar -c --gzip --owner=0 --group=0 --file="../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to tgz file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to tgz file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ "${asset}" == !*.tzst || "${asset}" == !*.tar.zst || "${asset}" == \$*.tzst || "${asset}" == \$*.tar.zst ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[ZST]${ANSI_NOCOLOR}"
|
||||||
|
|
||||||
|
if [[ "${asset:0:1}" == "\$" ]]; then
|
||||||
|
asset="${asset##*$}"
|
||||||
|
dirName="${asset%.*}"
|
||||||
|
printf " %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
tar -c --zstd --owner=0 --group=0 --file="${asset}" --directory="${artifact}" --transform "s|^\.|${dirName%.tar}|" .
|
||||||
|
retCode=$?
|
||||||
|
else
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf " %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
tar -c --zstd --owner=0 --group=0 --file="../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zst file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to zst file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ -e "${artifact}/${asset}" ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
uploadFile="${artifact}/${asset}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't find asset '${asset}' in artifact '${artifact}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=FileNotFound::Couldn't find asset '${asset}' in artifact '${artifact}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Add asset to JSON inventory
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
if [[ "${categories}" != "${title}" ]]; then
|
||||||
|
printf " %s\n" "adding file '${uploadFile#*/}' with '${categories//;/ → }' to JSON inventory ..."
|
||||||
|
category=""
|
||||||
|
jsonEntry=$(jq -c -n \
|
||||||
|
--arg title "${title}" \
|
||||||
|
--arg file "${uploadFile#*/}" \
|
||||||
|
'{"file": $file, "title": $title}' \
|
||||||
|
)
|
||||||
|
|
||||||
|
while [[ "${categories}" != "${category}" ]]; do
|
||||||
|
category="${categories##*,}"
|
||||||
|
categories="${categories%,*}"
|
||||||
|
jsonEntry=$(jq -c -n --arg cat "${category}" --argjson value "${jsonEntry}" '{$cat: $value}')
|
||||||
|
done
|
||||||
|
|
||||||
|
jsonInventory=$(jq -c -n \
|
||||||
|
--argjson inventory "${jsonInventory}" \
|
||||||
|
--argjson file "${jsonEntry}" \
|
||||||
|
'$inventory * {"files": $file}' \
|
||||||
|
)
|
||||||
|
else
|
||||||
|
printf " %s\n" "adding file '${uploadFile#*/}' to JSON inventory ... ${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Upload asset to existing release page
|
||||||
|
printf " %s" "uploading asset '${asset}' from '${uploadFile}' with title '${title}' ... "
|
||||||
|
gh release upload ${{ inputs.nightly_name }} "${uploadFile}#${title}" --clobber
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=UploadError::Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
done <<<'${{ inputs.assets }}'
|
||||||
|
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
inventoryTitle="Release Inventory (JSON)"
|
||||||
|
|
||||||
|
printf "%s\n" "Publish asset '${{ inputs.inventory-json }}' with title '${inventoryTitle}'"
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Writing JSON inventory to '${{ inputs.inventory-json }}' ...."
|
||||||
|
printf "%s\n" "$(jq -n --argjson inventory "${jsonInventory}" '$inventory')" > "${{ inputs.inventory-json }}"
|
||||||
|
cat "${{ inputs.inventory-json }}"
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
|
||||||
|
# Upload inventory asset to existing release page
|
||||||
|
printf " %s" "uploading asset '${{ inputs.inventory-json }}' title '${inventoryTitle}' ... "
|
||||||
|
gh release upload ${{ inputs.nightly_name }} "${{ inputs.inventory-json }}#${inventoryTitle}" --clobber
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=UploadError::Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Inspecting downloaded artifacts ..."
|
||||||
|
tree -pash -L 3 .
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
|
||||||
|
if [[ $ERRORS -ne 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}${ERRORS} errors detected in previous steps.${ANSI_NOCOLOR}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📑 Remove draft state from Release Page
|
||||||
|
if: ${{ ! inputs.draft }}
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
# Remove draft-state from release page
|
||||||
|
printf "%s" "Remove draft-state from release '${title}' ... "
|
||||||
|
gh release edit --draft=false "${{ inputs.nightly_name }}"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}"
|
||||||
|
else
|
||||||
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't remove draft-state from release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
|
printf "%s\n" "::error title=ReleasePage::Couldn't remove draft-state from release '${{ inputs.nightly_name }}'."
|
||||||
|
fi
|
||||||
8
.github/workflows/Package.yml
vendored
8
.github/workflows/Package.yml
vendored
@@ -54,6 +54,9 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
@@ -103,9 +106,10 @@ jobs:
|
|||||||
run: python setup.py bdist_wheel
|
run: python setup.py bdist_wheel
|
||||||
|
|
||||||
- name: 📤 Upload wheel artifact
|
- name: 📤 Upload wheel artifact
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.artifact }}
|
name: ${{ inputs.artifact }}
|
||||||
path: dist/
|
working-directory: dist
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|||||||
21
.github/workflows/Parameters.yml
vendored
21
.github/workflows/Parameters.yml
vendored
@@ -83,7 +83,7 @@ on:
|
|||||||
windows_image:
|
windows_image:
|
||||||
description: 'The used GitHub Action image for Windows based jobs.'
|
description: 'The used GitHub Action image for Windows based jobs.'
|
||||||
required: false
|
required: false
|
||||||
default: 'windows-latest'
|
default: 'windows-2022'
|
||||||
type: string
|
type: string
|
||||||
macos_intel_image:
|
macos_intel_image:
|
||||||
description: 'The used GitHub Action image for macOS (Intel x86-64) based jobs.'
|
description: 'The used GitHub Action image for macOS (Intel x86-64) based jobs.'
|
||||||
@@ -93,7 +93,7 @@ on:
|
|||||||
macos_arm_image:
|
macos_arm_image:
|
||||||
description: 'The used GitHub Action image for macOS (ARM aarch64) based jobs.'
|
description: 'The used GitHub Action image for macOS (ARM aarch64) based jobs.'
|
||||||
required: false
|
required: false
|
||||||
default: 'macos-latest'
|
default: 'macos-14'
|
||||||
type: string
|
type: string
|
||||||
|
|
||||||
outputs:
|
outputs:
|
||||||
@@ -147,7 +147,7 @@ jobs:
|
|||||||
else:
|
else:
|
||||||
name = f"{package_namespace}.{package_name}"
|
name = f"{package_namespace}.{package_name}"
|
||||||
|
|
||||||
currentMSYS2Version = "3.11"
|
currentMSYS2Version = "3.12"
|
||||||
currentAlphaVersion = "3.14"
|
currentAlphaVersion = "3.14"
|
||||||
currentAlphaRelease = "3.14.0-alpha.1"
|
currentAlphaRelease = "3.14.0-alpha.1"
|
||||||
|
|
||||||
@@ -197,9 +197,10 @@ jobs:
|
|||||||
"3.13": { "icon": "🟢", "until": "2029.10" },
|
"3.13": { "icon": "🟢", "until": "2029.10" },
|
||||||
"3.14": { "icon": "🟣", "until": "2030.10" },
|
"3.14": { "icon": "🟣", "until": "2030.10" },
|
||||||
"pypy-3.7": { "icon": "⟲⚫", "until": "????.??" },
|
"pypy-3.7": { "icon": "⟲⚫", "until": "????.??" },
|
||||||
"pypy-3.8": { "icon": "⟲🔴", "until": "????.??" },
|
"pypy-3.8": { "icon": "⟲⚫", "until": "????.??" },
|
||||||
"pypy-3.9": { "icon": "⟲🟠", "until": "????.??" },
|
"pypy-3.9": { "icon": "⟲🔴", "until": "????.??" },
|
||||||
"pypy-3.10": { "icon": "⟲🟡", "until": "????.??" },
|
"pypy-3.10": { "icon": "⟲🟠", "until": "????.??" },
|
||||||
|
"pypy-3.11": { "icon": "⟲🟡", "until": "????.??" },
|
||||||
},
|
},
|
||||||
# Runner systems (runner images) supported by GitHub Actions
|
# Runner systems (runner images) supported by GitHub Actions
|
||||||
"sys": {
|
"sys": {
|
||||||
@@ -337,7 +338,7 @@ jobs:
|
|||||||
- name: Verify out parameters
|
- name: Verify out parameters
|
||||||
id: verify
|
id: verify
|
||||||
run: |
|
run: |
|
||||||
echo 'python_version: ${{ steps.params.outputs.python_version }}'
|
printf "python_version: %s\n" '${{ steps.params.outputs.python_version }}'
|
||||||
echo 'python_jobs: ${{ steps.params.outputs.python_jobs }}'
|
printf "python_jobs: %s\n" '${{ steps.params.outputs.python_jobs }}'
|
||||||
echo 'artifact_names: ${{ steps.params.outputs.artifact_names }}'
|
printf "artifact_names: %s\n" '${{ steps.params.outputs.artifact_names }}'
|
||||||
echo 'params: ${{ steps.params.outputs.params }}'
|
printf "params: %s\n" '${{ steps.params.outputs.params }}'
|
||||||
|
|||||||
41
.github/workflows/PublishCoverageResults.yml
vendored
41
.github/workflows/PublishCoverageResults.yml
vendored
@@ -29,6 +29,10 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: '24.04'
|
default: '24.04'
|
||||||
type: string
|
type: string
|
||||||
|
coverage_artifacts_pattern:
|
||||||
|
required: false
|
||||||
|
default: '*-CodeCoverage-SQLite-*'
|
||||||
|
type: string
|
||||||
coverage_config:
|
coverage_config:
|
||||||
description: 'Path to the .coveragerc file. Use pyproject.toml by default.'
|
description: 'Path to the .coveragerc file. Use pyproject.toml by default.'
|
||||||
required: false
|
required: false
|
||||||
@@ -68,12 +72,20 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Download Artifacts
|
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
|
- name: 📥 Download Artifacts
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
|
with:
|
||||||
|
pattern: ${{ inputs.coverage_artifacts_pattern }}
|
||||||
path: artifacts
|
path: artifacts
|
||||||
|
|
||||||
|
- name: 🔎 Inspect extracted artifact (tarball)
|
||||||
|
run: |
|
||||||
|
tree -pash artifacts
|
||||||
|
|
||||||
- name: 🔧 Install coverage and tomli
|
- name: 🔧 Install coverage and tomli
|
||||||
run: |
|
run: |
|
||||||
python -m pip install -U --disable-pip-version-check --break-system-packages coverage[toml] tomli
|
python -m pip install -U --disable-pip-version-check --break-system-packages coverage[toml] tomli
|
||||||
@@ -138,13 +150,11 @@ jobs:
|
|||||||
|
|
||||||
print(f"DEBUG:\n html={htmlDirectory}\n xml={xmlFile}\n json={jsonFile}")
|
print(f"DEBUG:\n html={htmlDirectory}\n xml={xmlFile}\n json={jsonFile}")
|
||||||
|
|
||||||
- name: Rename .coverage files and collect them all to coverage/
|
- name: Rename .coverage files and move them all into 'coverage/'
|
||||||
run: |
|
run: |
|
||||||
ls -lAh artifacts/
|
|
||||||
ls -lAh artifacts/*/.coverage
|
|
||||||
mkdir -p coverage
|
mkdir -p coverage
|
||||||
find artifacts/ -type f -path "*SQLite*.coverage" -exec sh -c 'cp -v $0 "coverage/$(basename $0).$(basename $(dirname $0))"' {} ';'
|
find artifacts/ -type f -path "*SQLite*.coverage" -exec sh -c 'cp -v $0 "coverage/$(basename $0).$(basename $(dirname $0))"' {} ';'
|
||||||
tree -a coverage
|
tree -pash coverage
|
||||||
|
|
||||||
- name: Combine SQLite files (using Coverage.py)
|
- name: Combine SQLite files (using Coverage.py)
|
||||||
run: coverage combine --data-file=.coverage coverage/
|
run: coverage combine --data-file=.coverage coverage/
|
||||||
@@ -165,12 +175,12 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
coverage html --data-file=.coverage -d report/coverage/html
|
coverage html --data-file=.coverage -d report/coverage/html
|
||||||
rm report/coverage/html/.gitignore
|
rm report/coverage/html/.gitignore
|
||||||
tree -a report/coverage/html
|
tree -pash report/coverage/html
|
||||||
|
|
||||||
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.coverage_sqlite_artifact != ''
|
if: inputs.coverage_sqlite_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_sqlite_artifact }}
|
name: ${{ inputs.coverage_sqlite_artifact }}
|
||||||
path: .coverage
|
path: .coverage
|
||||||
@@ -178,9 +188,9 @@ jobs:
|
|||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📤 Upload 'Coverage XML Report' artifact
|
- name: 📤 Upload 'Coverage XML Report' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.coverage_xml_artifact != ''
|
if: inputs.coverage_xml_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_xml_artifact }}
|
name: ${{ inputs.coverage_xml_artifact }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
path: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
@@ -188,9 +198,9 @@ jobs:
|
|||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📤 Upload 'Coverage JSON Report' artifact
|
- name: 📤 Upload 'Coverage JSON Report' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.coverage_json_artifact != ''
|
if: inputs.coverage_json_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_json_artifact }}
|
name: ${{ inputs.coverage_json_artifact }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_json }}
|
path: ${{ steps.getVariables.outputs.coverage_report_json }}
|
||||||
@@ -198,28 +208,29 @@ jobs:
|
|||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📤 Upload 'Coverage HTML Report' artifact
|
- name: 📤 Upload 'Coverage HTML Report' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.coverage_html_artifact != ''
|
if: inputs.coverage_html_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_html_artifact }}
|
name: ${{ inputs.coverage_html_artifact }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
working-directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📊 Publish code coverage at CodeCov
|
- name: 📊 Publish code coverage at CodeCov
|
||||||
|
uses: codecov/codecov-action@v5
|
||||||
if: inputs.CodeCov == true
|
if: inputs.CodeCov == true
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: codecov/codecov-action@v4
|
|
||||||
with:
|
with:
|
||||||
files: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
files: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
flags: unittests
|
flags: unittests
|
||||||
env_vars: PYTHON
|
env_vars: PYTHON
|
||||||
|
|
||||||
- name: 📉 Publish code coverage at Codacy
|
- name: 📉 Publish code coverage at Codacy
|
||||||
|
uses: codacy/codacy-coverage-reporter-action@v1
|
||||||
if: inputs.Codacy == true
|
if: inputs.Codacy == true
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: codacy/codacy-coverage-reporter-action@v1
|
|
||||||
with:
|
with:
|
||||||
project-token: ${{ secrets.codacy_token }}
|
project-token: ${{ secrets.codacy_token }}
|
||||||
coverage-reports: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
coverage-reports: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
|
|||||||
4
.github/workflows/PublishOnPyPI.yml
vendored
4
.github/workflows/PublishOnPyPI.yml
vendored
@@ -57,10 +57,10 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: 📥 Download artifacts '${{ inputs.artifact }}' from 'Package' job
|
- name: 📥 Download artifacts '${{ inputs.artifact }}' from 'Package' job
|
||||||
uses: actions/download-artifact@v4
|
uses: pyTooling/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.artifact }}
|
name: ${{ inputs.artifact }}
|
||||||
path: dist/
|
path: dist
|
||||||
|
|
||||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
|
|||||||
30
.github/workflows/PublishTestResults.yml
vendored
30
.github/workflows/PublishTestResults.yml
vendored
@@ -30,6 +30,10 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: '24.04'
|
default: '24.04'
|
||||||
type: string
|
type: string
|
||||||
|
unittest_artifacts_pattern:
|
||||||
|
required: false
|
||||||
|
default: '*-UnitTestReportSummary-XML-*'
|
||||||
|
type: string
|
||||||
merged_junit_artifact:
|
merged_junit_artifact:
|
||||||
description: 'Name of the merged JUnit Test Summary artifact.'
|
description: 'Name of the merged JUnit Test Summary artifact.'
|
||||||
required: false
|
required: false
|
||||||
@@ -40,6 +44,11 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: '"--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit"'
|
default: '"--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit"'
|
||||||
type: string
|
type: string
|
||||||
|
publish:
|
||||||
|
description: 'Publish test report summary via Dorny Test-Reporter'
|
||||||
|
required: false
|
||||||
|
default: true
|
||||||
|
type: boolean
|
||||||
report_title:
|
report_title:
|
||||||
description: 'Title of the summary report in the pipeline''s sidebar'
|
description: 'Title of the summary report in the pipeline''s sidebar'
|
||||||
required: false
|
required: false
|
||||||
@@ -56,38 +65,43 @@ jobs:
|
|||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Download Artifacts
|
- name: 📥 Download Artifacts
|
||||||
uses: actions/download-artifact@v4
|
uses: pyTooling/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
|
pattern: ${{ inputs.unittest_artifacts_pattern }}
|
||||||
path: artifacts
|
path: artifacts
|
||||||
|
|
||||||
|
- name: 🔎 Inspect extracted artifact (tarball)
|
||||||
|
run: |
|
||||||
|
tree -pash artifacts
|
||||||
|
|
||||||
- name: 🔧 Install pyEDAA.Reports (JUunit Parser and Merger)
|
- name: 🔧 Install pyEDAA.Reports (JUunit Parser and Merger)
|
||||||
run: |
|
run: |
|
||||||
python -m pip install --disable-pip-version-check --break-system-packages -U pyEDAA.Reports
|
python -m pip install --disable-pip-version-check --break-system-packages -U pyEDAA.Reports
|
||||||
|
|
||||||
- name: Move JUnit files and collect them all to junit/
|
- name: Rename JUnit files and move them all into 'junit/'
|
||||||
run: |
|
run: |
|
||||||
mkdir -p junit
|
mkdir -p junit
|
||||||
ls -lAh artifacts/*/*.xml
|
find artifacts/ -type f -path "*.xml" -exec sh -c 'cp -v $0 "junit/$(basename $(dirname $0)).$(basename $0)"' {} ';'
|
||||||
find artifacts/ -type f -path "*TestReportSummary*.xml" -exec sh -c 'cp -v $0 "junit/$(basename $(dirname $0)).$(basename $0)"' {} ';'
|
tree -pash junit
|
||||||
tree -a junit
|
|
||||||
|
|
||||||
- name: 🔁 Merge JUnit Unit Test Summaries
|
- name: 🔁 Merge JUnit Unit Test Summaries
|
||||||
run: |
|
run: |
|
||||||
pyedaa-reports -v unittest "--merge=pyTest-JUnit:junit/*.xml" ${{ inputs.additional_merge_args }} "--output=pyTest-JUnit:Unittesting.xml"
|
pyedaa-reports -v unittest "--merge=pyTest-JUnit:junit/*.xml" ${{ inputs.additional_merge_args }} "--output=pyTest-JUnit:Unittesting.xml"
|
||||||
echo "cat Unittesting.xml"
|
printf "%s\n" "cat Unittesting.xml"
|
||||||
cat Unittesting.xml
|
cat Unittesting.xml
|
||||||
|
|
||||||
- name: 📊 Publish Unit Test Results
|
- name: 📊 Publish Unit Test Results
|
||||||
uses: dorny/test-reporter@v1
|
uses: dorny/test-reporter@v1
|
||||||
|
if: inputs.publish && inputs.report_title != ''
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.report_title }}
|
name: ${{ inputs.report_title }}
|
||||||
path: Unittesting.xml
|
path: Unittesting.xml
|
||||||
reporter: java-junit
|
reporter: java-junit
|
||||||
|
|
||||||
- name: 📤 Upload merged 'JUnit Test Summary' artifact
|
- name: 📤 Upload merged 'JUnit Test Summary' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.merged_junit_artifact != ''
|
if: inputs.merged_junit_artifact != ''
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.merged_junit_artifact }}
|
name: ${{ inputs.merged_junit_artifact }}
|
||||||
path: Unittesting.xml
|
path: Unittesting.xml
|
||||||
|
|||||||
8
.github/workflows/PublishToGitHubPages.yml
vendored
8
.github/workflows/PublishToGitHubPages.yml
vendored
@@ -55,22 +55,22 @@ jobs:
|
|||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.doc }}' from 'BuildTheDocs' job
|
- name: 📥 Download artifacts '${{ inputs.doc }}' from 'SphinxDocumentation' job
|
||||||
uses: actions/download-artifact@v4
|
uses: pyTooling/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.doc }}
|
name: ${{ inputs.doc }}
|
||||||
path: public
|
path: public
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.coverage }}' from 'Coverage' job
|
- name: 📥 Download artifacts '${{ inputs.coverage }}' from 'Coverage' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: ${{ inputs.coverage != '' }}
|
if: ${{ inputs.coverage != '' }}
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage }}
|
name: ${{ inputs.coverage }}
|
||||||
path: public/coverage
|
path: public/coverage
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.typing }}' from 'StaticTypeCheck' job
|
- name: 📥 Download artifacts '${{ inputs.typing }}' from 'StaticTypeCheck' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: ${{ inputs.typing != '' }}
|
if: ${{ inputs.typing != '' }}
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.typing }}
|
name: ${{ inputs.typing }}
|
||||||
path: public/typing
|
path: public/typing
|
||||||
|
|||||||
2
.github/workflows/Release.yml
vendored
2
.github/workflows/Release.yml
vendored
@@ -49,8 +49,8 @@ jobs:
|
|||||||
echo "datetime=${RELEASE_DATETIME}" >> $GITHUB_OUTPUT
|
echo "datetime=${RELEASE_DATETIME}" >> $GITHUB_OUTPUT
|
||||||
|
|
||||||
- name: 📑 Create Release Page
|
- name: 📑 Create Release Page
|
||||||
id: createReleasePage
|
|
||||||
uses: actions/create-release@v1
|
uses: actions/create-release@v1
|
||||||
|
id: createReleasePage
|
||||||
env:
|
env:
|
||||||
GITHUB_TOKEN: ${{ github.token }}
|
GITHUB_TOKEN: ${{ github.token }}
|
||||||
with:
|
with:
|
||||||
|
|||||||
106
.github/workflows/SphinxDocumentation.yml
vendored
106
.github/workflows/SphinxDocumentation.yml
vendored
@@ -82,6 +82,9 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
- name: 🔧 Install graphviz
|
- name: 🔧 Install graphviz
|
||||||
run: sudo apt-get install -y --no-install-recommends graphviz
|
run: sudo apt-get install -y --no-install-recommends graphviz
|
||||||
@@ -97,15 +100,15 @@ jobs:
|
|||||||
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: inputs.unittest_xml_artifact != ''
|
if: inputs.unittest_xml_artifact != ''
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.unittest_xml_artifact }}
|
name: ${{ inputs.unittest_xml_artifact }}
|
||||||
path: ${{ inputs.unittest_xml_directory }}
|
path: ${{ inputs.unittest_xml_directory }}
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: inputs.coverage_json_artifact != ''
|
if: inputs.coverage_json_artifact != ''
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_json_artifact }}
|
name: ${{ inputs.coverage_json_artifact }}
|
||||||
path: ${{ inputs.coverage_report_json_directory }}
|
path: ${{ inputs.coverage_report_json_directory }}
|
||||||
@@ -119,12 +122,13 @@ jobs:
|
|||||||
sphinx-build -v -n -b html -d _build/doctrees -j $(nproc) -w _build/html.log . _build/html
|
sphinx-build -v -n -b html -d _build/doctrees -j $(nproc) -w _build/html.log . _build/html
|
||||||
|
|
||||||
- name: 📤 Upload 'HTML Documentation' artifact
|
- name: 📤 Upload 'HTML Documentation' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.html_artifact != ''
|
if: inputs.html_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.html_artifact }}
|
name: ${{ inputs.html_artifact }}
|
||||||
path: ${{ inputs.doc_directory }}/_build/html
|
working-directory: ${{ inputs.doc_directory }}/_build/html
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
@@ -135,6 +139,9 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
- name: 🔧 Install graphviz
|
- name: 🔧 Install graphviz
|
||||||
run: sudo apt-get install -y --no-install-recommends graphviz
|
run: sudo apt-get install -y --no-install-recommends graphviz
|
||||||
@@ -150,15 +157,15 @@ jobs:
|
|||||||
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: inputs.unittest_xml_artifact != ''
|
if: inputs.unittest_xml_artifact != ''
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.unittest_xml_artifact }}
|
name: ${{ inputs.unittest_xml_artifact }}
|
||||||
path: ${{ inputs.unittest_xml_directory }}
|
path: ${{ inputs.unittest_xml_directory }}
|
||||||
|
|
||||||
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
if: inputs.coverage_json_artifact != ''
|
if: inputs.coverage_json_artifact != ''
|
||||||
uses: actions/download-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_json_artifact }}
|
name: ${{ inputs.coverage_json_artifact }}
|
||||||
path: ${{ inputs.coverage_report_json_directory }}
|
path: ${{ inputs.coverage_report_json_directory }}
|
||||||
@@ -173,12 +180,95 @@ jobs:
|
|||||||
sphinx-build -v -n -b latex -d _build/doctrees -j $(nproc) -w _build/latex.log . _build/latex
|
sphinx-build -v -n -b latex -d _build/doctrees -j $(nproc) -w _build/latex.log . _build/latex
|
||||||
# --builder html --doctree-dir _build/doctrees --verbose --fresh-env --write-all --nitpicky --warning-file _build/html.log . _build/html
|
# --builder html --doctree-dir _build/doctrees --verbose --fresh-env --write-all --nitpicky --warning-file _build/html.log . _build/html
|
||||||
|
|
||||||
|
- name: Workaround I - https://github.com/sphinx-doc/sphinx/issues/13190
|
||||||
|
if: inputs.latex_artifact != ''
|
||||||
|
run: |
|
||||||
|
printf "Changing directory to 'doc/_build/latex' ...\n"
|
||||||
|
cd doc/_build/latex
|
||||||
|
|
||||||
|
MIMETYPE_EXTENSIONS=(
|
||||||
|
"image/png:png"
|
||||||
|
"image/jpeg:jpg"
|
||||||
|
"image/svg+xml:svg"
|
||||||
|
)
|
||||||
|
|
||||||
|
printf "Changing file extension according to MIME type ...\n"
|
||||||
|
while IFS=$'\n' read -r file; do
|
||||||
|
printf " Checking '%s' ... " "${file}"
|
||||||
|
mime="$(file --mime-type -b "${file}")"
|
||||||
|
printf "[%s]\n" "${mime}"
|
||||||
|
|
||||||
|
found=0
|
||||||
|
for MIME in "${MIMETYPE_EXTENSIONS[@]}"; do
|
||||||
|
mimetype="${MIME%%:*}"
|
||||||
|
extension="${MIME#*:}"
|
||||||
|
|
||||||
|
if [[ "${mime}" == "${mimetype}" && "${file##*.}" != "${extension}" ]]; then
|
||||||
|
printf " Rename file to '%s' " "${file}.${extension}"
|
||||||
|
mv "${file}" "${file}.${extension}"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "[OK]\n"
|
||||||
|
else
|
||||||
|
printf "[FAILED]\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf " Patching LaTeX file for '%s' " "${file}"
|
||||||
|
sed -i "s:{{${file%.*}}\.${file##*.}}:{{${file}}.${extension}}:g" *.tex
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "[OK]\n"
|
||||||
|
else
|
||||||
|
printf "[FAILED]\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
found=1
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
if [[ $found -eq 0 ]]; then
|
||||||
|
printf "[SKIPPED]\n"
|
||||||
|
fi
|
||||||
|
done < <(find . -type f -not -iname "*.cls" -not -iname "*.sty" -not -iname "*.xdy" -not -iname "*.svg" -not -iname "*.png" -not -iname "*.jpg" | sed 's:./::')
|
||||||
|
|
||||||
|
- name: Workaround II - https://github.com/sphinx-doc/sphinx/issues/13189
|
||||||
|
if: inputs.latex_artifact != ''
|
||||||
|
run: |
|
||||||
|
printf "Changing directory to 'doc/_build/latex' ...\n"
|
||||||
|
cd doc/_build/latex
|
||||||
|
|
||||||
|
printf "Searching for downloaded images, that need normalization ...\n"
|
||||||
|
for imageExt in png svg jpg jpeg; do
|
||||||
|
printf " Processing '%s' ...\n" "${imageExt}"
|
||||||
|
while IFS=$'\n' read -r imageFile; do
|
||||||
|
newFile="${imageFile//%/_}";
|
||||||
|
|
||||||
|
printf " %s\n" "$imageFile";
|
||||||
|
if [[ "${imageFile}" != "${newFile}" ]]; then
|
||||||
|
printf " Rename file to '%s' " "${newFile}"
|
||||||
|
mv "${imageFile}" "${newFile}"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "[OK]\n"
|
||||||
|
else
|
||||||
|
printf "[FAILED]\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf " Patching LaTeX file for '%s' " "${newFile}"
|
||||||
|
sed -i "s:{{${imageFile%.*}}\.${imageFile##*.}}:{{${newFile%.*}}.${newFile##*.}}:g" *.tex
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "[OK]\n"
|
||||||
|
else
|
||||||
|
printf "[FAILED]\n"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done < <(find . -type f -iname "*.$imageExt" | sed 's:./::')
|
||||||
|
done
|
||||||
|
|
||||||
- name: 📤 Upload 'LaTeX Documentation' artifact
|
- name: 📤 Upload 'LaTeX Documentation' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.latex_artifact != ''
|
if: inputs.latex_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.latex_artifact }}
|
name: ${{ inputs.latex_artifact }}
|
||||||
path: ${{ inputs.doc_directory }}/_build/latex
|
working-directory: ${{ inputs.doc_directory }}/_build/latex
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|||||||
7
.github/workflows/StaticTypeCheck.yml
vendored
7
.github/workflows/StaticTypeCheck.yml
vendored
@@ -87,19 +87,20 @@ jobs:
|
|||||||
run: ${{ inputs.commands }}
|
run: ${{ inputs.commands }}
|
||||||
|
|
||||||
- name: 📤 Upload 'Static Typing Report' HTML artifact
|
- name: 📤 Upload 'Static Typing Report' HTML artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: ${{ inputs.html_artifact != '' }}
|
if: ${{ inputs.html_artifact != '' }}
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.html_artifact }}
|
name: ${{ inputs.html_artifact }}
|
||||||
path: ${{ inputs.html_report }}
|
working-directory: ${{ inputs.html_report }}
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
- name: 📤 Upload 'Static Typing Report' JUnit artifact
|
- name: 📤 Upload 'Static Typing Report' JUnit artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: ${{ inputs.junit_artifact != '' }}
|
if: ${{ inputs.junit_artifact != '' }}
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.junit_artifact }}
|
name: ${{ inputs.junit_artifact }}
|
||||||
path: ${{ inputs.junit_report }}
|
path: ${{ inputs.junit_report }}
|
||||||
|
|||||||
20
.github/workflows/TestReleaser.yml
vendored
20
.github/workflows/TestReleaser.yml
vendored
@@ -64,7 +64,7 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
- run: echo "Build some tool and generate some (versioned) artifacts" > artifact-$(date -u +"%Y-%m-%dT%H-%M-%SZ").txt
|
- run: printf "%s\n" "Build some tool and generate some (versioned) artifacts" > artifact-$(date -u +"%Y-%m-%dT%H-%M-%SZ").txt
|
||||||
|
|
||||||
- name: Single
|
- name: Single
|
||||||
uses: ./releaser/composite
|
uses: ./releaser/composite
|
||||||
@@ -84,7 +84,7 @@ jobs:
|
|||||||
- name: Add artifacts/*.txt
|
- name: Add artifacts/*.txt
|
||||||
run: |
|
run: |
|
||||||
mkdir artifacts
|
mkdir artifacts
|
||||||
echo "Build some tool and generate some artifacts" > artifacts/artifact.txt
|
printf "%s\n" "Build some tool and generate some artifacts" > artifacts/artifact.txt
|
||||||
touch artifacts/empty_file.txt
|
touch artifacts/empty_file.txt
|
||||||
|
|
||||||
- name: Single in subdir
|
- name: Single in subdir
|
||||||
@@ -95,8 +95,8 @@ jobs:
|
|||||||
|
|
||||||
- name: Add artifacts/*.md
|
- name: Add artifacts/*.md
|
||||||
run: |
|
run: |
|
||||||
echo "releaser hello" > artifacts/hello.md
|
printf "%s\n" "releaser hello" > artifacts/hello.md
|
||||||
echo "releaser world" > artifacts/world.md
|
printf "%s\n" "releaser world" > artifacts/world.md
|
||||||
|
|
||||||
- name: Directory wildcard
|
- name: Directory wildcard
|
||||||
uses: ./releaser/composite
|
uses: ./releaser/composite
|
||||||
@@ -107,7 +107,7 @@ jobs:
|
|||||||
- name: Add artifacts/subdir
|
- name: Add artifacts/subdir
|
||||||
run: |
|
run: |
|
||||||
mkdir artifacts/subdir
|
mkdir artifacts/subdir
|
||||||
echo "Test recursive glob" > artifacts/subdir/deep_file.txt
|
printf "%s\n" "Test recursive glob" > artifacts/subdir/deep_file.txt
|
||||||
|
|
||||||
- name: Directory wildcard (recursive)
|
- name: Directory wildcard (recursive)
|
||||||
uses: ./releaser/composite
|
uses: ./releaser/composite
|
||||||
@@ -124,7 +124,7 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
- run: echo "Build some tool and generate some (versioned) artifacts" > artifact-$(date -u +"%Y-%m-%dT%H-%M-%SZ").txt
|
- run: printf "%s\n" "Build some tool and generate some (versioned) artifacts" > artifact-$(date -u +"%Y-%m-%dT%H-%M-%SZ").txt
|
||||||
|
|
||||||
- name: Single
|
- name: Single
|
||||||
uses: ./releaser
|
uses: ./releaser
|
||||||
@@ -144,7 +144,7 @@ jobs:
|
|||||||
- name: Add artifacts/*.txt
|
- name: Add artifacts/*.txt
|
||||||
run: |
|
run: |
|
||||||
mkdir artifacts
|
mkdir artifacts
|
||||||
echo "Build some tool and generate some artifacts" > artifacts/artifact.txt
|
printf "%s\n" "Build some tool and generate some artifacts" > artifacts/artifact.txt
|
||||||
touch artifacts/empty_file.txt
|
touch artifacts/empty_file.txt
|
||||||
|
|
||||||
- name: Single in subdir
|
- name: Single in subdir
|
||||||
@@ -155,8 +155,8 @@ jobs:
|
|||||||
|
|
||||||
- name: Add artifacts/*.md
|
- name: Add artifacts/*.md
|
||||||
run: |
|
run: |
|
||||||
echo "releaser hello" > artifacts/hello.md
|
printf "%s\n" "releaser hello" > artifacts/hello.md
|
||||||
echo "releaser world" > artifacts/world.md
|
printf "%s\n" "releaser world" > artifacts/world.md
|
||||||
|
|
||||||
- name: Directory wildcard
|
- name: Directory wildcard
|
||||||
uses: ./releaser
|
uses: ./releaser
|
||||||
@@ -167,7 +167,7 @@ jobs:
|
|||||||
- name: Add artifacts/subdir
|
- name: Add artifacts/subdir
|
||||||
run: |
|
run: |
|
||||||
mkdir artifacts/subdir
|
mkdir artifacts/subdir
|
||||||
echo "Test recursive glob" > artifacts/subdir/deep_file.txt
|
printf "%s\n" "Test recursive glob" > artifacts/subdir/deep_file.txt
|
||||||
|
|
||||||
- name: Directory wildcard (recursive)
|
- name: Directory wildcard (recursive)
|
||||||
uses: ./releaser
|
uses: ./releaser
|
||||||
|
|||||||
117
.github/workflows/UnitTesting.yml
vendored
117
.github/workflows/UnitTesting.yml
vendored
@@ -94,11 +94,26 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: 'unit'
|
default: 'unit'
|
||||||
type: string
|
type: string
|
||||||
|
unittest_report_xml_directory:
|
||||||
|
description: 'Path where to save the unittest summary report XML.'
|
||||||
|
required: false
|
||||||
|
default: 'report/unit'
|
||||||
|
type: string
|
||||||
|
unittest_report_xml_filename:
|
||||||
|
description: 'Filename of the unittest summary report XML.'
|
||||||
|
required: false
|
||||||
|
default: 'TestReportSummary.xml'
|
||||||
|
type: string
|
||||||
coverage_config:
|
coverage_config:
|
||||||
description: 'Path to the .coveragerc file. Use pyproject.toml by default.'
|
description: 'Path to the .coveragerc file. Use pyproject.toml by default.'
|
||||||
required: false
|
required: false
|
||||||
default: 'pyproject.toml'
|
default: 'pyproject.toml'
|
||||||
type: string
|
type: string
|
||||||
|
coverage_report_html_directory:
|
||||||
|
description: ''
|
||||||
|
required: false
|
||||||
|
default: 'report/coverage/html'
|
||||||
|
type: string
|
||||||
unittest_xml_artifact:
|
unittest_xml_artifact:
|
||||||
description: "Generate unit test report with junitxml and upload results as an artifact."
|
description: "Generate unit test report with junitxml and upload results as an artifact."
|
||||||
required: false
|
required: false
|
||||||
@@ -147,6 +162,9 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
lfs: true
|
||||||
|
submodules: true
|
||||||
|
|
||||||
# Package Manager steps
|
# Package Manager steps
|
||||||
- name: 🔧 Install homebrew dependencies on macOS
|
- name: 🔧 Install homebrew dependencies on macOS
|
||||||
@@ -264,8 +282,8 @@ jobs:
|
|||||||
# Python setup
|
# Python setup
|
||||||
|
|
||||||
- name: '🟦 Setup MSYS2 for ${{ matrix.runtime }}'
|
- name: '🟦 Setup MSYS2 for ${{ matrix.runtime }}'
|
||||||
if: matrix.system == 'msys2'
|
|
||||||
uses: msys2/setup-msys2@v2
|
uses: msys2/setup-msys2@v2
|
||||||
|
if: matrix.system == 'msys2'
|
||||||
with:
|
with:
|
||||||
msystem: ${{ matrix.runtime }}
|
msystem: ${{ matrix.runtime }}
|
||||||
update: true
|
update: true
|
||||||
@@ -274,8 +292,8 @@ jobs:
|
|||||||
${{ inputs.pacboy }}
|
${{ inputs.pacboy }}
|
||||||
|
|
||||||
- name: 🐍 Setup Python ${{ matrix.python }}
|
- name: 🐍 Setup Python ${{ matrix.python }}
|
||||||
if: matrix.system != 'msys2'
|
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v5
|
||||||
|
if: matrix.system != 'msys2'
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python }}
|
python-version: ${{ matrix.python }}
|
||||||
|
|
||||||
@@ -320,69 +338,6 @@ jobs:
|
|||||||
if: matrix.system == 'msys2' && matrix.runtime == 'UCRT64' && inputs.ucrt64_before_script != ''
|
if: matrix.system == 'msys2' && matrix.runtime == 'UCRT64' && inputs.ucrt64_before_script != ''
|
||||||
run: ${{ inputs.ucrt64_before_script }}
|
run: ${{ inputs.ucrt64_before_script }}
|
||||||
|
|
||||||
# Read pyproject.toml
|
|
||||||
|
|
||||||
- name: 🔁 Extract configurations from pyproject.toml
|
|
||||||
id: getVariables
|
|
||||||
shell: python
|
|
||||||
run: |
|
|
||||||
from os import getenv
|
|
||||||
from pathlib import Path
|
|
||||||
from sys import version
|
|
||||||
from textwrap import dedent
|
|
||||||
|
|
||||||
print(f"Python: {version}")
|
|
||||||
|
|
||||||
from tomli import load as tomli_load
|
|
||||||
|
|
||||||
htmlDirectory = Path("htmlcov")
|
|
||||||
xmlFile = Path("./coverage.xml")
|
|
||||||
jsonFile = Path("./coverage.json")
|
|
||||||
coverageRC = "${{ inputs.coverage_config }}".strip()
|
|
||||||
|
|
||||||
# Read output paths from 'pyproject.toml' file
|
|
||||||
if coverageRC == "pyproject.toml":
|
|
||||||
pyProjectFile = Path("pyproject.toml")
|
|
||||||
if pyProjectFile.exists():
|
|
||||||
with pyProjectFile.open("rb") as file:
|
|
||||||
pyProjectSettings = tomli_load(file)
|
|
||||||
|
|
||||||
htmlDirectory = Path(pyProjectSettings["tool"]["coverage"]["html"]["directory"])
|
|
||||||
xmlFile = Path(pyProjectSettings["tool"]["coverage"]["xml"]["output"])
|
|
||||||
jsonFile = Path(pyProjectSettings["tool"]["coverage"]["json"]["output"])
|
|
||||||
else:
|
|
||||||
print(f"File '{pyProjectFile}' not found.")
|
|
||||||
print(f"::error title=FileNotFoundError::File '{pyProjectFile}' not found.")
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
# Read output paths from '.coveragerc' file
|
|
||||||
elif len(coverageRC) > 0:
|
|
||||||
coverageRCFile = Path(coverageRC)
|
|
||||||
if coverageRCFile.exists():
|
|
||||||
with coverageRCFile.open("rb") as file:
|
|
||||||
coverageRCSettings = tomli_load(file)
|
|
||||||
|
|
||||||
htmlDirectory = Path(coverageRCSettings["html"]["directory"])
|
|
||||||
xmlFile = Path(coverageRCSettings["xml"]["output"])
|
|
||||||
jsonFile = Path(coverageRCSettings["json"]["output"])
|
|
||||||
else:
|
|
||||||
print(f"File '{coverageRCFile}' not found.")
|
|
||||||
print(f"::error title=FileNotFoundError::File '{coverageRCFile}' not found.")
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
# Write jobs to special file
|
|
||||||
github_output = Path(getenv("GITHUB_OUTPUT"))
|
|
||||||
print(f"GITHUB_OUTPUT: {github_output}")
|
|
||||||
with github_output.open("a+", encoding="utf-8") as f:
|
|
||||||
f.write(dedent(f"""\
|
|
||||||
unittest_report_html_directory={htmlDirectory}
|
|
||||||
coverage_report_html_directory={htmlDirectory.as_posix()}
|
|
||||||
coverage_report_xml={xmlFile}
|
|
||||||
coverage_report_json={jsonFile}
|
|
||||||
"""))
|
|
||||||
|
|
||||||
print(f"DEBUG:\n html={htmlDirectory}\n xml={xmlFile}\n json={jsonFile}")
|
|
||||||
|
|
||||||
# Run pytests
|
# Run pytests
|
||||||
|
|
||||||
- name: ✅ Run unit tests (Ubuntu/macOS)
|
- name: ✅ Run unit tests (Ubuntu/macOS)
|
||||||
@@ -392,12 +347,12 @@ jobs:
|
|||||||
export PYTHONPATH=$(pwd)
|
export PYTHONPATH=$(pwd)
|
||||||
|
|
||||||
cd "${{ inputs.root_directory || '.' }}"
|
cd "${{ inputs.root_directory || '.' }}"
|
||||||
[ -n '${{ inputs.unittest_xml_artifact }}' ] && PYTEST_ARGS='--junitxml=report/unit/TestReportSummary.xml' || unset PYTEST_ARGS
|
[ -n '${{ inputs.unittest_xml_artifact }}' ] && PYTEST_ARGS='--junitxml=${{ inputs.unittest_report_xml_directory }}/${{ inputs.unittest_report_xml_filename }}' || unset PYTEST_ARGS
|
||||||
if [ -n '${{ inputs.coverage_config }}' ]; then
|
if [ -n '${{ inputs.coverage_config }}' ]; then
|
||||||
echo "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
printf "%s\n" "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
||||||
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
||||||
else
|
else
|
||||||
echo "python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
printf "%s\n" "python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
||||||
python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
python -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -408,7 +363,7 @@ jobs:
|
|||||||
$env:PYTHONPATH = (Get-Location).ToString()
|
$env:PYTHONPATH = (Get-Location).ToString()
|
||||||
|
|
||||||
cd "${{ inputs.root_directory || '.' }}"
|
cd "${{ inputs.root_directory || '.' }}"
|
||||||
$PYTEST_ARGS = if ("${{ inputs.unittest_xml_artifact }}") { "--junitxml=report/unit/TestReportSummary.xml" } else { "" }
|
$PYTEST_ARGS = if ("${{ inputs.unittest_xml_artifact }}") { "--junitxml=${{ inputs.unittest_report_xml_directory }}/${{ inputs.unittest_report_xml_filename }}" } else { "" }
|
||||||
if ("${{ inputs.coverage_config }}") {
|
if ("${{ inputs.coverage_config }}") {
|
||||||
Write-Host "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
Write-Host "coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}"
|
||||||
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
coverage run --data-file=.coverage --rcfile=pyproject.toml -m pytest -raP $PYTEST_ARGS --color=yes ${{ inputs.tests_directory || '.' }}/${{ inputs.unittest_directory }}
|
||||||
@@ -431,25 +386,26 @@ jobs:
|
|||||||
if: inputs.coverage_html_artifact != ''
|
if: inputs.coverage_html_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
coverage html --data-file=.coverage -d ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
coverage html --data-file=.coverage -d ${{ inputs.coverage_report_html_directory }}
|
||||||
rm ${{ steps.getVariables.outputs.coverage_report_html_directory }}/.gitignore
|
rm ${{ inputs.coverage_report_html_directory }}/.gitignore
|
||||||
|
|
||||||
# Upload artifacts
|
# Upload artifacts
|
||||||
|
|
||||||
- name: 📤 Upload 'TestReportSummary.xml' artifact
|
- name: 📤 Upload '${{ inputs.unittest_report_xml_filename }}' artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.unittest_xml_artifact != ''
|
if: inputs.unittest_xml_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.unittest_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.unittest_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: report/unit/TestReportSummary.xml
|
working-directory: ${{ inputs.unittest_report_xml_directory }}
|
||||||
|
path: ${{ inputs.unittest_report_xml_filename }}
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|
||||||
# - name: 📤 Upload 'Unit Tests HTML Report' artifact
|
# - name: 📤 Upload 'Unit Tests HTML Report' artifact
|
||||||
# if: inputs.unittest_html_artifact != ''
|
# if: inputs.unittest_html_artifact != ''
|
||||||
# continue-on-error: true
|
# continue-on-error: true
|
||||||
# uses: actions/upload-artifact@v4
|
# uses: pyTooling/upload-artifact@v4
|
||||||
# with:
|
# with:
|
||||||
# name: ${{ inputs.unittest_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
# name: ${{ inputs.unittest_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
# path: ${{ steps.getVariables.outputs.unittest_report_html_directory }}
|
# path: ${{ steps.getVariables.outputs.unittest_report_html_directory }}
|
||||||
@@ -459,7 +415,7 @@ jobs:
|
|||||||
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
||||||
if: inputs.coverage_sqlite_artifact != ''
|
if: inputs.coverage_sqlite_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_sqlite_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.coverage_sqlite_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: .coverage
|
path: .coverage
|
||||||
@@ -470,7 +426,7 @@ jobs:
|
|||||||
- name: 📤 Upload 'Coverage XML Report' artifact
|
- name: 📤 Upload 'Coverage XML Report' artifact
|
||||||
if: inputs.coverage_xml_artifact != ''
|
if: inputs.coverage_xml_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.coverage_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
path: ${{ steps.getVariables.outputs.coverage_report_xml }}
|
||||||
@@ -480,7 +436,7 @@ jobs:
|
|||||||
- name: 📤 Upload 'Coverage JSON Report' artifact
|
- name: 📤 Upload 'Coverage JSON Report' artifact
|
||||||
if: inputs.coverage_json_artifact != ''
|
if: inputs.coverage_json_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_json_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.coverage_json_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_json }}
|
path: ${{ steps.getVariables.outputs.coverage_report_json }}
|
||||||
@@ -490,9 +446,10 @@ jobs:
|
|||||||
- name: 📤 Upload 'Coverage HTML Report' artifact
|
- name: 📤 Upload 'Coverage HTML Report' artifact
|
||||||
if: inputs.coverage_html_artifact != ''
|
if: inputs.coverage_html_artifact != ''
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.coverage_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
name: ${{ inputs.coverage_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||||
path: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
working-directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
||||||
|
path: '*'
|
||||||
if-no-files-found: error
|
if-no-files-found: error
|
||||||
retention-days: 1
|
retention-days: 1
|
||||||
|
|||||||
@@ -22,10 +22,10 @@ jobs:
|
|||||||
include: ${{ fromJson(needs.Params.outputs.python_jobs) }}
|
include: ${{ fromJson(needs.Params.outputs.python_jobs) }}
|
||||||
steps:
|
steps:
|
||||||
- name: Content creation for ${{ matrix.system }}-${{ matrix.python }}
|
- name: Content creation for ${{ matrix.system }}-${{ matrix.python }}
|
||||||
run: echo "${{ matrix.runs-on }}-${{ matrix.python }}" >> artifact.txt
|
run: printf "%s\n" "${{ matrix.runs-on }}-${{ matrix.python }}" >> artifact.txt
|
||||||
|
|
||||||
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ fromJson(needs.Params.outputs.artifact_names).unittesting_xml }}-${{ matrix.system }}-${{ matrix.python }}
|
name: ${{ fromJson(needs.Params.outputs.artifact_names).unittesting_xml }}-${{ matrix.system }}-${{ matrix.python }}
|
||||||
path: artifact.txt
|
path: artifact.txt
|
||||||
@@ -39,10 +39,10 @@ jobs:
|
|||||||
runs-on: ubuntu-24.04
|
runs-on: ubuntu-24.04
|
||||||
steps:
|
steps:
|
||||||
- name: Package creation
|
- name: Package creation
|
||||||
run: echo "Package" >> package.txt
|
run: printf "%s\n" "Package" >> package.txt
|
||||||
|
|
||||||
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
||||||
uses: actions/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
name: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
||||||
path: package.txt
|
path: package.txt
|
||||||
|
|||||||
79
.github/workflows/_Checking_JobTemplates.yml
vendored
79
.github/workflows/_Checking_JobTemplates.yml
vendored
@@ -7,8 +7,6 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
ConfigParams:
|
ConfigParams:
|
||||||
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
||||||
needs:
|
|
||||||
- DocCoverage
|
|
||||||
with:
|
with:
|
||||||
package_name: pyDummy
|
package_name: pyDummy
|
||||||
|
|
||||||
@@ -24,17 +22,21 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
name: Platform
|
name: Platform
|
||||||
python_version_list: ""
|
python_version_list: ""
|
||||||
system_list: "ubuntu windows macos mingw32 mingw64 clang64 ucrt64"
|
system_list: "ubuntu windows macos mingw64 clang64 ucrt64"
|
||||||
|
|
||||||
UnitTesting:
|
UnitTesting:
|
||||||
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- ConfigParams
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
jobs: ${{ needs.UnitTestingParams.outputs.python_jobs }}
|
jobs: ${{ needs.UnitTestingParams.outputs.python_jobs }}
|
||||||
unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
unittest_report_xml_directory: ${{ needs.ConfigParams.outputs.unittest_report_xml_directory }}
|
||||||
unittest_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_html }}
|
unittest_report_xml_filename: ${{ needs.ConfigParams.outputs.unittest_report_xml_filename }}
|
||||||
# coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
coverage_report_html_directory: ${{ needs.ConfigParams.outputs.coverage_report_html_directory }}
|
||||||
|
unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
|
unittest_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_html }}
|
||||||
|
coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
||||||
# coverage_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}
|
# coverage_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}
|
||||||
# coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
# coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
||||||
# coverage_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
# coverage_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
@@ -42,27 +44,21 @@ jobs:
|
|||||||
PlatformTesting:
|
PlatformTesting:
|
||||||
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- ConfigParams
|
||||||
- PlatformTestingParams
|
- PlatformTestingParams
|
||||||
with:
|
with:
|
||||||
jobs: ${{ needs.PlatformTestingParams.outputs.python_jobs }}
|
jobs: ${{ needs.PlatformTestingParams.outputs.python_jobs }}
|
||||||
# tests_directory: ""
|
# tests_directory: ""
|
||||||
unittest_directory: platform
|
unittest_directory: platform
|
||||||
unittest_xml_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_xml }}
|
unittest_report_xml_directory: ${{ needs.ConfigParams.outputs.unittest_report_xml_directory }}
|
||||||
unittest_html_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_html }}
|
unittest_report_xml_filename: ${{ needs.ConfigParams.outputs.unittest_report_xml_filename }}
|
||||||
coverage_sqlite_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
coverage_report_html_directory: ${{ needs.ConfigParams.outputs.coverage_report_html_directory }}
|
||||||
coverage_xml_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_xml }}
|
unittest_xml_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
coverage_json_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_json }}
|
unittest_html_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_html }}
|
||||||
coverage_html_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_html }}
|
coverage_sqlite_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
||||||
|
coverage_xml_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_xml }}
|
||||||
# Coverage:
|
coverage_json_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_json }}
|
||||||
# uses: pyTooling/Actions/.github/workflows/CoverageCollection.yml@main
|
coverage_html_artifact: ${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
# needs:
|
|
||||||
# - UnitTestingParams
|
|
||||||
# with:
|
|
||||||
# python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
|
||||||
# artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
|
||||||
# secrets:
|
|
||||||
# codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
|
||||||
|
|
||||||
StaticTypeCheck:
|
StaticTypeCheck:
|
||||||
uses: pyTooling/Actions/.github/workflows/StaticTypeCheck.yml@main
|
uses: pyTooling/Actions/.github/workflows/StaticTypeCheck.yml@main
|
||||||
@@ -78,12 +74,13 @@ jobs:
|
|||||||
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
||||||
|
|
||||||
DocCoverage:
|
DocCoverage:
|
||||||
uses: pyTooling/Actions/.github/workflows/CheckDocumentation.yml@r1
|
uses: pyTooling/Actions/.github/workflows/CheckDocumentation.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- ConfigParams
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
directory: sphinx_reports
|
directory: ${{ needs.ConfigParams.outputs.package_directors }}
|
||||||
# fail_below: 70
|
# fail_below: 70
|
||||||
|
|
||||||
Package:
|
Package:
|
||||||
@@ -91,7 +88,6 @@ jobs:
|
|||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
# - Coverage
|
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
@@ -103,22 +99,23 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
# - Coverage
|
|
||||||
with:
|
with:
|
||||||
coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
coverage_sqlite_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}
|
||||||
coverage_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}
|
coverage_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}
|
||||||
coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
||||||
coverage_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
coverage_html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
secrets:
|
secrets:
|
||||||
codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
||||||
|
|
||||||
PublishTestResults:
|
PublishTestResults:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
||||||
needs:
|
needs:
|
||||||
|
- UnitTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
with:
|
with:
|
||||||
additional_merge_args: '-d "--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit;reduce-depth:pytest.tests.platform"'
|
additional_merge_args: '-d "--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit;reduce-depth:pytest.tests.platform"'
|
||||||
|
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
|
|
||||||
# VerifyDocs:
|
# VerifyDocs:
|
||||||
# uses: pyTooling/Actions/.github/workflows/VerifyDocs.yml@main
|
# uses: pyTooling/Actions/.github/workflows/VerifyDocs.yml@main
|
||||||
@@ -130,21 +127,21 @@ jobs:
|
|||||||
Documentation:
|
Documentation:
|
||||||
uses: pyTooling/Actions/.github/workflows/SphinxDocumentation.yml@main
|
uses: pyTooling/Actions/.github/workflows/SphinxDocumentation.yml@main
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
|
||||||
- ConfigParams
|
- ConfigParams
|
||||||
|
- UnitTestingParams
|
||||||
- PublishTestResults
|
- PublishTestResults
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
# - VerifyDocs
|
# - VerifyDocs
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
coverage_report_json_directory: ${{ needs.ConfigParams.outputs.coverage_report_json_directory }}
|
coverage_report_json_directory: ${{ needs.ConfigParams.outputs.coverage_report_json_directory }}
|
||||||
# unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
unittest_xml_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
# coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
coverage_json_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}
|
||||||
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_html }}
|
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_html }}
|
||||||
latex_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_latex }}
|
latex_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_latex }}
|
||||||
|
|
||||||
IntermediateCleanUp:
|
IntermediateCleanUp:
|
||||||
uses: pyTooling/Actions/.github/workflows/IntermediateCleanUp.yml@r1
|
uses: pyTooling/Actions/.github/workflows/IntermediateCleanUp.yml@main
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
@@ -152,7 +149,7 @@ jobs:
|
|||||||
- Documentation
|
- Documentation
|
||||||
with:
|
with:
|
||||||
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
||||||
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
||||||
|
|
||||||
PDFDocumentation:
|
PDFDocumentation:
|
||||||
uses: pyTooling/Actions/.github/workflows/LaTeXDocumentation.yml@main
|
uses: pyTooling/Actions/.github/workflows/LaTeXDocumentation.yml@main
|
||||||
@@ -160,7 +157,7 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- Documentation
|
- Documentation
|
||||||
with:
|
with:
|
||||||
document: actions
|
document: Actions
|
||||||
latex_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_latex }}
|
latex_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_latex }}
|
||||||
pdf_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_pdf }}
|
pdf_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_pdf }}
|
||||||
|
|
||||||
@@ -169,8 +166,7 @@ jobs:
|
|||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- Documentation
|
- Documentation
|
||||||
# - PDFDocumentation
|
- PDFDocumentation
|
||||||
# - Coverage
|
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
- StaticTypeCheck
|
- StaticTypeCheck
|
||||||
with:
|
with:
|
||||||
@@ -184,7 +180,6 @@ jobs:
|
|||||||
needs:
|
needs:
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
# - Coverage
|
|
||||||
# - StaticTypeCheck
|
# - StaticTypeCheck
|
||||||
- Package
|
- Package
|
||||||
- PublishToGitHubPages
|
- PublishToGitHubPages
|
||||||
@@ -209,11 +204,10 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- PlatformTestingParams
|
- PlatformTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
# - Coverage
|
|
||||||
- StaticTypeCheck
|
- StaticTypeCheck
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
- Documentation
|
- Documentation
|
||||||
# - PDFDocumentation
|
- PDFDocumentation
|
||||||
- PublishTestResults
|
- PublishTestResults
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
- PublishToGitHubPages
|
- PublishToGitHubPages
|
||||||
@@ -223,7 +217,6 @@ jobs:
|
|||||||
remaining: |
|
remaining: |
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-*
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-*
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_html }}-*
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_html }}-*
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-*
|
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}-*
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_xml }}-*
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}-*
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_json }}-*
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}-*
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}-*
|
||||||
@@ -235,9 +228,9 @@ jobs:
|
|||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
||||||
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_html }}
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_html }}
|
||||||
|
${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).documentation_latex }}
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_xml }}-*
|
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_xml }}-*
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_html }}-*
|
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).unittesting_html }}-*
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_sqlite }}-*
|
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_xml }}-*
|
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_xml }}-*
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_json }}-*
|
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_json }}-*
|
||||||
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_html }}-*
|
${{ fromJson(needs.PlatformTestingParams.outputs.artifact_names).codecoverage_html }}-*
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
name: Verification of Pipeline Templates
|
name: Verification of Pipeline Templates (Namespace Package)
|
||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
|
|||||||
123
.github/workflows/_Checking_Nightly.yml
vendored
Normal file
123
.github/workflows/_Checking_Nightly.yml
vendored
Normal file
@@ -0,0 +1,123 @@
|
|||||||
|
name: Verification of Nightly Releases
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
Build:
|
||||||
|
name: Build something
|
||||||
|
runs-on: ubuntu-24.04
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: 🖉 Build 1
|
||||||
|
run: |
|
||||||
|
printf "%s\n" "Document 1 $(date --utc '+%d.%m.%Y - %H:%M:%S')" > document1.txt
|
||||||
|
printf "%s\n" "Analysis log $(date --utc '+%d.%m.%Y - %H:%M:%S')" > analysis.log
|
||||||
|
printf "%s\n" "Build log $(date --utc '+%d.%m.%Y - %H:%M:%S')" > build.log
|
||||||
|
|
||||||
|
- name: 📤 Upload artifact
|
||||||
|
uses: pyTooling/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: document
|
||||||
|
path: |
|
||||||
|
document1.txt
|
||||||
|
*.log
|
||||||
|
if-no-files-found: error
|
||||||
|
retention-days: 1
|
||||||
|
|
||||||
|
- name: 🖉 Program
|
||||||
|
run: |
|
||||||
|
printf "%s\n" "Document other $(date --utc '+%d.%m.%Y - %H:%M:%S')" > document1.txt
|
||||||
|
printf "%s\n" "Program $(date --utc '+%d.%m.%Y - %H:%M:%S')" > program.py
|
||||||
|
|
||||||
|
- name: 📤 Upload artifact
|
||||||
|
uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: other
|
||||||
|
path: |
|
||||||
|
*.txt
|
||||||
|
*.py
|
||||||
|
if-no-files-found: error
|
||||||
|
retention-days: 1
|
||||||
|
|
||||||
|
NightlyPage:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/NightlyRelease.yml@main
|
||||||
|
needs:
|
||||||
|
- Build
|
||||||
|
secrets: inherit
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
# attestations: write
|
||||||
|
with:
|
||||||
|
can-fail: true
|
||||||
|
prerelease: true
|
||||||
|
replacements: |
|
||||||
|
version=4.2.0
|
||||||
|
tool=myTool
|
||||||
|
prog=program
|
||||||
|
nightly_title: "Nightly Test Release"
|
||||||
|
nightly_description: |
|
||||||
|
This *nightly* release contains all latest and important artifacts created by %tool%'s CI pipeline.
|
||||||
|
|
||||||
|
# %tool% %version%
|
||||||
|
|
||||||
|
* %prog%
|
||||||
|
assets: |
|
||||||
|
document: document1.txt: Documentation
|
||||||
|
document: build.log: Logfile - %tool% - %tool%
|
||||||
|
other: document1.txt: SBOM - %version%
|
||||||
|
other: %prog%.py: Application - %tool% - %version%
|
||||||
|
document:!archive1.zip: Archive 1 - zip
|
||||||
|
document:!archive2.tgz: Archive 2 - tgz
|
||||||
|
document:!archive3.tar.gz: Archive 3 - tar.gz
|
||||||
|
document:!archive4.tzst: Archive 4 - tzst
|
||||||
|
document:!archive5.tar.zst:Archive 5 - tar.zst
|
||||||
|
document:$archive6.tgz: Archive 6 - tgz + dir
|
||||||
|
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
||||||
|
document:$archive8.tzst: Archive 8 - tzst + dir
|
||||||
|
document:$archive9.tar.zst:Archive 9 - tar.zst + dir
|
||||||
|
|
||||||
|
NightlyPageWithInventory:
|
||||||
|
uses: ./.github/workflows/NightlyRelease.yml
|
||||||
|
needs:
|
||||||
|
- Build
|
||||||
|
secrets: inherit
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
# attestations: write
|
||||||
|
with:
|
||||||
|
can-fail: true
|
||||||
|
replacements: |
|
||||||
|
version=4.2.0
|
||||||
|
tool=myTool
|
||||||
|
prog=program
|
||||||
|
nightly_name: inventory
|
||||||
|
nightly_title: "Nightly Test Release with Inventory"
|
||||||
|
nightly_description: |
|
||||||
|
This *nightly* release contains all latest and important artifacts created by %tool%'s CI pipeline.
|
||||||
|
|
||||||
|
# %tool% %version%
|
||||||
|
|
||||||
|
* %prog%
|
||||||
|
* iventory.json
|
||||||
|
inventory-json: "inventory.json"
|
||||||
|
inventory-version: 4.2.5
|
||||||
|
inventory-categories: "kind1,kind2"
|
||||||
|
assets: |
|
||||||
|
# artifact: file: labels: asset title
|
||||||
|
document: document1.txt: doc,html: Documentation
|
||||||
|
document: build.log: build,log: Logfile - %tool% - %tool%
|
||||||
|
other: document1.txt: build,SBOM:SBOM - %version%
|
||||||
|
other: %prog%.py: app,binary:Application - %tool% - %version%
|
||||||
|
document:!archive1.zip: Archive 1 - zip
|
||||||
|
document:!archive2.tgz: Archive 2 - tgz
|
||||||
|
document:!archive3.tar.gz: Archive 3 - tar.gz
|
||||||
|
document:!archive4.tzst: Archive 4 - tzst
|
||||||
|
document:!archive5.tar.zst: Archive 5 - tar.zst
|
||||||
|
document:$archive6.tgz: Archive 6 - tgz + dir
|
||||||
|
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
||||||
|
document:$archive8.tzst: Archive 8 - tzst + dir
|
||||||
|
document:$archive9.tar.zst: Archive 9 - tar.zst + dir
|
||||||
6
.github/workflows/_Checking_Parameters.yml
vendored
6
.github/workflows/_Checking_Parameters.yml
vendored
@@ -83,7 +83,7 @@ jobs:
|
|||||||
expectedPythonVersion = "3.13"
|
expectedPythonVersion = "3.13"
|
||||||
expectedPythons = ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
expectedPythons = ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
||||||
expectedSystems = ["ubuntu", "windows", "macos", "macos-arm"]
|
expectedSystems = ["ubuntu", "windows", "macos", "macos-arm"]
|
||||||
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw64:3.11", "ucrt64:3.11"]
|
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw64:3.12", "ucrt64:3.11"]
|
||||||
expectedName = "Example"
|
expectedName = "Example"
|
||||||
expectedArtifacts = {
|
expectedArtifacts = {
|
||||||
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
||||||
@@ -146,7 +146,7 @@ jobs:
|
|||||||
expectedPythonVersion = "3.13"
|
expectedPythonVersion = "3.13"
|
||||||
expectedPythons = ["3.11", "3.12", "pypy-3.9", "pypy-3.10"]
|
expectedPythons = ["3.11", "3.12", "pypy-3.9", "pypy-3.10"]
|
||||||
expectedSystems = ["ubuntu", "windows", "macos", "macos-arm"]
|
expectedSystems = ["ubuntu", "windows", "macos", "macos-arm"]
|
||||||
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw64:3.11", "ucrt64:3.11"]
|
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw64:3.12", "ucrt64:3.11"]
|
||||||
expectedName = "Example"
|
expectedName = "Example"
|
||||||
expectedArtifacts = {
|
expectedArtifacts = {
|
||||||
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
||||||
@@ -209,7 +209,7 @@ jobs:
|
|||||||
expectedPythonVersion = "3.13"
|
expectedPythonVersion = "3.13"
|
||||||
expectedPythons = ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
expectedPythons = ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
||||||
expectedSystems = ["windows"]
|
expectedSystems = ["windows"]
|
||||||
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw32:3.11", "mingw64:3.11"]
|
expectedJobs = [f"{system}:{python}" for system in expectedSystems for python in expectedPythons] + ["mingw32:3.12", "mingw64:3.11"]
|
||||||
expectedName = "Example"
|
expectedName = "Example"
|
||||||
expectedArtifacts = {
|
expectedArtifacts = {
|
||||||
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
"unittesting_xml": f"{expectedName}-UnitTestReportSummary-XML",
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
name: Verification of Pipeline Templates
|
name: Verification of Pipeline Templates (Simple Package)
|
||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
|
|||||||
2
.gitignore
vendored
2
.gitignore
vendored
@@ -31,7 +31,7 @@ doc/pyDummy/**/*.*
|
|||||||
# BuildTheDocs
|
# BuildTheDocs
|
||||||
doc/_theme/**/*.*
|
doc/_theme/**/*.*
|
||||||
|
|
||||||
# IntelliJ project files
|
# PyCharm project files
|
||||||
/.idea/workspace.xml
|
/.idea/workspace.xml
|
||||||
|
|
||||||
# Git files
|
# Git files
|
||||||
|
|||||||
@@ -1,178 +0,0 @@
|
|||||||
# ==================================================================================================================== #
|
|
||||||
# Authors: #
|
|
||||||
# Patrick Lehmann #
|
|
||||||
# Unai Martinez-Corral #
|
|
||||||
# #
|
|
||||||
# ==================================================================================================================== #
|
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
|
||||||
# #
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
|
||||||
# you may not use this file except in compliance with the License. #
|
|
||||||
# You may obtain a copy of the License at #
|
|
||||||
# #
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0 #
|
|
||||||
# #
|
|
||||||
# Unless required by applicable law or agreed to in writing, software #
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS, #
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #
|
|
||||||
# See the License for the specific language governing permissions and #
|
|
||||||
# limitations under the License. #
|
|
||||||
# #
|
|
||||||
# SPDX-License-Identifier: Apache-2.0 #
|
|
||||||
# ==================================================================================================================== #
|
|
||||||
name: Pipeline
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_dispatch:
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
|
|
||||||
# This job is a workaround for global variables
|
|
||||||
# See https://github.com/actions/runner/issues/480
|
|
||||||
Params:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
|
||||||
with:
|
|
||||||
name: ToolName
|
|
||||||
# Optional
|
|
||||||
system_list: 'ubuntu windows msys2 macos'
|
|
||||||
python_version: '3.10'
|
|
||||||
python_version_list: '3.8 3.9 3.10'
|
|
||||||
|
|
||||||
UnitTesting:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
with:
|
|
||||||
jobs: ${{ needs.Params.outputs.python_jobs }}
|
|
||||||
# Optional
|
|
||||||
requirements: '-r tests/requirements.txt'
|
|
||||||
pacboy: >-
|
|
||||||
python-pip:p
|
|
||||||
python-wheel:p
|
|
||||||
python-coverage:p
|
|
||||||
python-lxml:p
|
|
||||||
mingw_requirements: '-r tests/requirements.mingw.txt'
|
|
||||||
tests_directory: 'tests'
|
|
||||||
unittest_directory: 'unit'
|
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).unittesting_xml }}
|
|
||||||
|
|
||||||
Coverage:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/CoverageCollection.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
with:
|
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).codecoverage_html }}
|
|
||||||
# Optional
|
|
||||||
python_version: ${{ needs..Params.outputs.python_version }}
|
|
||||||
requirements: '-r tests/requirements.txt'
|
|
||||||
tests_directory: 'tests'
|
|
||||||
unittest_directory: 'unit'
|
|
||||||
secrets:
|
|
||||||
codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
|
||||||
|
|
||||||
StaticTypeCheck:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/StaticTypeCheck.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
with:
|
|
||||||
commands: |
|
|
||||||
mypy --junit-xml StaticTypingSummary.xml --html-report htmlmypy -p ToolName
|
|
||||||
html_artifact: ${{ fromJson(needs.Params.outputs.artifact_names).statictyping_html }}
|
|
||||||
junit_artifact: ${{ fromJson(needs.Params.outputs.artifact_names).statictyping_junit }}
|
|
||||||
# Optional
|
|
||||||
python_version: ${{ needs..Params.outputs.python_version }}
|
|
||||||
requirements: '-r tests/requirements.txt'
|
|
||||||
html_report: 'htmlmypy'
|
|
||||||
junit_report: 'StaticTypingSummary.xml'
|
|
||||||
allow_failure: true
|
|
||||||
|
|
||||||
PublishTestResults:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
|
||||||
needs:
|
|
||||||
- UnitTesting
|
|
||||||
- StaticTypeCheck
|
|
||||||
with:
|
|
||||||
# Optional
|
|
||||||
report_files: artifacts/**/*.xml
|
|
||||||
|
|
||||||
Package:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/Package.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
- Coverage
|
|
||||||
with:
|
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
|
||||||
# Optional
|
|
||||||
python_version: ${{ needs..Params.outputs.python_version }}
|
|
||||||
requirements: 'wheel'
|
|
||||||
|
|
||||||
Release:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/Release.yml@main
|
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
|
||||||
- UnitTesting
|
|
||||||
- Coverage
|
|
||||||
- StaticTypeCheck
|
|
||||||
- Package
|
|
||||||
|
|
||||||
PublishOnPyPI:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishOnPyPI.yml@main
|
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
- Release
|
|
||||||
- Package
|
|
||||||
with:
|
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
|
||||||
# Optional
|
|
||||||
python_version: ${{ needs..Params.outputs.python_version }}
|
|
||||||
requirements: 'wheel twine'
|
|
||||||
secrets:
|
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
|
||||||
|
|
||||||
VerifyDocs:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/VerifyDocs.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
with:
|
|
||||||
# Optional
|
|
||||||
python_version: ${{ needs..Params.outputs.python_version }}
|
|
||||||
|
|
||||||
BuildTheDocs:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/BuildTheDocs.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
- VerifyDocs
|
|
||||||
with:
|
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).documentation_html }}
|
|
||||||
|
|
||||||
PublishToGitHubPages:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishToGitHubPages.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
- BuildTheDocs
|
|
||||||
- Coverage
|
|
||||||
- StaticTypeCheck
|
|
||||||
with:
|
|
||||||
doc: ${{ fromJson(needs.Params.outputs.artifact_names).documentation_html }}
|
|
||||||
# Optional
|
|
||||||
coverage: ${{ fromJson(needs.Params.outputs.artifact_names).codecoverage_html }}
|
|
||||||
typing: ${{ fromJson(needs.Params.outputs.artifact_names).statictyping_html }}
|
|
||||||
|
|
||||||
ArtifactCleanUp:
|
|
||||||
uses: pyTooling/Actions/.github/workflows/ArtifactCleanUp.yml@main
|
|
||||||
needs:
|
|
||||||
- Params
|
|
||||||
- PublishTestResults
|
|
||||||
- Coverage
|
|
||||||
- StaticTypeCheck
|
|
||||||
- BuildTheDocs
|
|
||||||
- PublishToGitHubPages
|
|
||||||
with:
|
|
||||||
package: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
|
||||||
remaining: |
|
|
||||||
${{ fromJson(needs.Params.outputs.artifact_names).unittesting_xml }}-*
|
|
||||||
${{ fromJson(needs.Params.outputs.artifact_names).codecoverage_html }}
|
|
||||||
${{ fromJson(needs.Params.outputs.artifact_names).statictyping_html }}
|
|
||||||
${{ fromJson(needs.Params.outputs.artifact_names).statictyping_junit }}
|
|
||||||
${{ fromJson(needs.Params.outputs.artifact_names).documentation_html }}
|
|
||||||
Binary file not shown.
|
Before Width: | Height: | Size: 74 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 135 KiB |
111
README.md
111
README.md
@@ -10,52 +10,97 @@ This repository gathers reusable CI tooling for testing, packaging and distribut
|
|||||||
See [GitHub Actions and GitHub Reusable Workflows](https://pytooling.github.io/Actions/Background.html) for more
|
See [GitHub Actions and GitHub Reusable Workflows](https://pytooling.github.io/Actions/Background.html) for more
|
||||||
background information.
|
background information.
|
||||||
|
|
||||||
## Reusable workflows
|
## Reusable Actions
|
||||||
|
|
||||||
|
- **Artifacts:**
|
||||||
|
[**pyTooling/upload-artifact**](https://github.com/pyTooling/upload-artifact): The upload-artifact action will
|
||||||
|
preserve file attributes like permissions.
|
||||||
|
|
||||||
|
[**pyTooling/download-artifact**](https://github.com/pyTooling/download-artifact): The download-artifact action will
|
||||||
|
preserve file attributes like permissions.
|
||||||
|
|
||||||
|
## Predefined Docker Images
|
||||||
|
|
||||||
|
- **Documentation:**
|
||||||
|
[**MikTeX**](https://github.com/pyTooling/MikTeX): A predefined MikTeX image based on Debian Bookworm + Python 3.13
|
||||||
|
with specific tools for documentation generation using e.g. Sphinx and related extensions.
|
||||||
|
|
||||||
|
## Reusable Workflows
|
||||||
|
|
||||||
This repository provides 10+ *Reusable Workflows* based on the CI pipelines of the repos in this GitHub organisation,
|
This repository provides 10+ *Reusable Workflows* based on the CI pipelines of the repos in this GitHub organisation,
|
||||||
[EDA²](https://github.com/edaa-org), [VHDL](https://github.com/vhdl), and others. By combining them, Python packages can
|
[EDA²](https://github.com/edaa-org), [VHDL](https://github.com/vhdl), and others. By combining them, Python packages can
|
||||||
be continuously tested and released along with Sphinx documentation sites, to GitHub Releases, GitHub Pages and PyPI.
|
be continuously tested and released along with Sphinx documentation sites, to GitHub Releases, GitHub Pages and PyPI.
|
||||||
Optionally, coverage and static type check reports can be gathered and integrated into the online documentation.
|
Optionally, coverage and static type check reports can be gathered and integrated into the online documentation.
|
||||||
|
|
||||||
[](ExamplePipeline_dark.png)
|
[](doc/_static/pyTooling-Actions-SimplePackage.png)
|
||||||
|
|
||||||
[](ExamplePipeline_light.png)
|
|
||||||
|
|
||||||
As shown in the screenshots above, the expected order is:
|
As shown in the screenshots above, the expected order is:
|
||||||
|
|
||||||
- Global:
|
- **Global:**
|
||||||
- [Parameters](.github/workflows/Parameters.yml): a workaround for the limitations to handle global variables in
|
[**Parameters**](.github/workflows/Parameters.yml): It generates output parameters with artifact names and job matrices
|
||||||
GitHub Actions workflows (see [actions/runner#480](https://github.com/actions/runner/issues/480)).
|
to be used in later running jobs.
|
||||||
It generates outputs with artifact names and job matrices to be used in later running jobs.
|
It's a workaround for the limitations to handle global variables in GitHub Actions workflows (see
|
||||||
- Code testing/analysis:
|
[actions/runner#480](https://github.com/actions/runner/issues/480)).
|
||||||
- [UnitTesting](.github/workflows/UnitTesting.yml): run unit test with `pytest` using multiple versions of Python, and
|
|
||||||
optionally upload results as XML reports. Configuration options to `pytest` should be given via section
|
[**ExtractConfiguration**](.github/workflows/ExtractConfiguration.yml): extracts configuration values from
|
||||||
`[tool.pytest.ini_options]` in a `pyproject.toml` file.
|
`pyproject.toml` and exposes configured paths and filenames as job output parameters.
|
||||||
- [CoverageCollection](.github/workflows/CoverageCollection.yml): collect code coverage data (incl. branch coverage)
|
- **Predefined pipelines:**
|
||||||
with `pytest`/`pytest-cov`/`coverage.py` using a single version of Python (latest). It generates HTML and Cobertura
|
[**CompletePipeline**](.github/workflows/CompletePipeline.yml): is a predefined pipeline for typical Python projects
|
||||||
(XML)reports, upload the HTML report as an artifact, and upload the test results to Codecov and Codacy. Configuration
|
using all predefined job templates of pyTooling at once: (unit testing, code coverage, static typing, documentation
|
||||||
options to `pytest` and `coverage.py` should be given via section `[tool.pytest.ini_options]` and `[tool.coverage.*]`
|
report generation and publishing, packaging, releasing, ...)
|
||||||
in a `pyproject.toml` file.
|
- **Code testing/analysis:**
|
||||||
- [StaticTypeCheck](.github/workflows/StaticTypeCheck.yml): collect static type check result with `mypy`, and
|
[**ApplicationTesting**](.github/workflows/ApplicationTesting.yml): like UnitTesting, but running tests using an
|
||||||
optionally upload results as an HTML report.
|
installed Python package.
|
||||||
Example `commands`:
|
|
||||||
|
[**UnitTesting**](.github/workflows/UnitTesting.yml): run unit test with `pytest` using multiple versions of Python, and
|
||||||
|
optionally upload results as XML reports. Configuration options to `pytest` should be given via section
|
||||||
|
`[tool.pytest.ini_options]` in a `pyproject.toml` file.
|
||||||
|
Besides test results, also code coverage data (incl. branch coverage) can be collected using
|
||||||
|
`pytest`/`pytest-cov`/`coverage.py`. Configuration options to `coverage.py` should be given via section
|
||||||
|
`[tool.coverage.*]` in a `pyproject.toml` file.
|
||||||
|
While multiple report formats can be created in the job, it's recommended to use `PublishTestResults` and/or
|
||||||
|
`PublishCoverageResults` to merge results from matrix runs and then generate final reports as XML, JSON or HTML.
|
||||||
|
Finally, reports can be published to GitHub Pages or cloud services like Codecov and Codacy.
|
||||||
|
|
||||||
|
[**StaticTypeCheck**](.github/workflows/StaticTypeCheck.yml): collect static type check result with `mypy`, and
|
||||||
|
optionally upload results as an HTML report.
|
||||||
|
|
||||||
|
[**VerifyDocs**](.github/workflows/VerifyDocs.yml): extract code examples from the README and test these code snippets.
|
||||||
|
- **Packaging and releasing:**
|
||||||
|
[**Package**](.github/workflows/Package.yml): generate source and wheel packages, and upload them as an artifact.
|
||||||
|
|
||||||
|
[**PublishOnPyPI**](.github/workflows/PublishOnPyPI.yml): publish source and wheel packages to PyPI.
|
||||||
|
|
||||||
|
[**PublishTestResults**](.github/workflows/PublishTestResults.yml): publish unit test results through GH action `dorny/test-reporter`.
|
||||||
|
|
||||||
|
[**PublishCoverageResults**](.github/workflows/PublishCoverageResults.yml): publish ucode coverage results.
|
||||||
|
|
||||||
|
[**NightlyRelease**](.github/workflows/NightlyRelease.yml): publish GitHub Release.
|
||||||
|
|
||||||
|
[**Release**](.github/workflows/Release.yml): publish GitHub Release.
|
||||||
|
- **Documentation:**
|
||||||
|
[**SphinxDocumentation**](.github/workflows/PublishCoverageResults.yml): create HTML and LaTeX documentation using
|
||||||
|
Sphinx.
|
||||||
|
|
||||||
|
[**LaTeXDocumentation**](.github/workflows/LaTeXDocumentation.yml): compile LaTeX documentation to a PDF file using
|
||||||
|
MikTeX.
|
||||||
|
|
||||||
|
[**PublishToGitHubPages**](.github/workflows/PublishToGitHubPages.yml): publish HTML documentation to GitHub Pages.
|
||||||
|
- **Cleanup:**
|
||||||
|
[**IntermediateCleanUp**](.github/workflows/IntermediateCleanUp.yml): delete intermediate artifacts.
|
||||||
|
|
||||||
|
[**ArtifactCleanUp**](.github/workflows/ArtifactCleanUp.yml): delete artifacts.
|
||||||
|
- **⚠ Deprecated ⚠:**
|
||||||
|
[**CoverageCollection**](.github/workflows/CoverageCollection.yml): Use `UnitTesting`, because is can collect code
|
||||||
|
coverage too. This avoids code duplication in job templates.
|
||||||
|
|
||||||
|
[**BuildTheDocs**](.github/workflows/BuildTheDocs.yml): Use `SphinxDocumentation`, `LaTeXDocumentation` and
|
||||||
|
`PublishToGitHubPages`. BuildTheDocs isn't maintained anymore.
|
||||||
|
|
||||||
- [VerifyDocs](.github/workflows/VerifyDocs.yml): extract code examples from the README and test these code snippets.
|
|
||||||
- Packaging and releasing:
|
|
||||||
- [Release](.github/workflows/Release.yml): publish GitHub Release.
|
|
||||||
- [Package](.github/workflows/Package.yml): generate source and wheel packages, and upload them as an artifact.
|
|
||||||
- [PublishOnPyPI](.github/workflows/PublishOnPyPI.yml): publish source and wheel packages to PyPI.
|
|
||||||
- [PublishTestResults](.github/workflows/PublishTestResults.yml): publish unit test results through GH action `dorny/test-reporter`.
|
|
||||||
- Documentation:
|
|
||||||
- [BuildTheDocs](.github/workflows/BuildTheDocs.yml): build Sphinx documentation with BuildTheDocs, and upload HTML as
|
|
||||||
an artifact.
|
|
||||||
- [PublishToGitHubPages](.github/workflows/PublishToGitHubPages.yml): publish HTML documentation to GitHub Pages.
|
|
||||||
- Cleanup:
|
|
||||||
- [ArtifactCleanUp](.github/workflows/ArtifactCleanUp.yml): delete artifacts.
|
|
||||||
|
|
||||||
### Example pipeline
|
### Example pipeline
|
||||||
|
|
||||||
[ExamplePipeline.yml](ExamplePipeline.yml) is an example Workflow which uses all of the Reusable Workflows.
|
ExamplePipeline.yml is an example Workflow which uses all of the Reusable Workflows.
|
||||||
Python package/tool developers can copy it into their repos, in order to use al the reusable workflows straightaway.
|
Python package/tool developers can copy it into their repos, in order to use al the reusable workflows straightaway.
|
||||||
Minimal required modifications are the following:
|
Minimal required modifications are the following:
|
||||||
|
|
||||||
|
|||||||
2
dist/requirements.txt
vendored
2
dist/requirements.txt
vendored
@@ -1,2 +1,2 @@
|
|||||||
wheel ~= 0.45
|
wheel ~= 0.45
|
||||||
twine ~= 5.1
|
twine ~= 6.1
|
||||||
|
|||||||
BIN
doc/_static/pyTooling-Actions-SimplePackage.png
vendored
Normal file
BIN
doc/_static/pyTooling-Actions-SimplePackage.png
vendored
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 555 KiB |
@@ -1,9 +1,9 @@
|
|||||||
-r ../requirements.txt
|
-r ../requirements.txt
|
||||||
|
|
||||||
pyTooling ~= 8.0
|
pyTooling ~= 8.2
|
||||||
|
|
||||||
# Enforce latest version on ReadTheDocs
|
# Enforce latest version on ReadTheDocs
|
||||||
sphinx ~= 8.1
|
sphinx ~= 8.2
|
||||||
docutils ~= 0.21
|
docutils ~= 0.21
|
||||||
docutils_stubs ~= 0.0.22
|
docutils_stubs ~= 0.0.22
|
||||||
|
|
||||||
@@ -11,17 +11,9 @@ docutils_stubs ~= 0.0.22
|
|||||||
sphinx_rtd_theme ~= 3.0
|
sphinx_rtd_theme ~= 3.0
|
||||||
|
|
||||||
# Sphinx Extenstions
|
# Sphinx Extenstions
|
||||||
#sphinx.ext.coverage
|
sphinxcontrib-mermaid ~= 1.0
|
||||||
#sphinxcontrib-actdiag>=0.8.5
|
|
||||||
sphinxcontrib-mermaid>=0.9.2
|
|
||||||
#sphinxcontrib-seqdiag>=0.8.5
|
|
||||||
#sphinxcontrib-textstyle>=0.2.1
|
|
||||||
#sphinxcontrib-spelling>=2.2.0
|
|
||||||
autoapi >= 2.0.1
|
autoapi >= 2.0.1
|
||||||
sphinx_design ~= 0.6.1
|
sphinx_design ~= 0.6.1
|
||||||
sphinx-copybutton >= 0.5.2
|
sphinx-copybutton >= 0.5.2
|
||||||
sphinx_autodoc_typehints ~= 2.5
|
sphinx_autodoc_typehints ~= 3.1
|
||||||
# changelog>=0.3.5
|
|
||||||
sphinx_reports ~= 0.7
|
sphinx_reports ~= 0.7
|
||||||
|
|
||||||
# BuildTheDocs Extensions (mostly patched Sphinx extensions)
|
|
||||||
|
|||||||
@@ -4,11 +4,11 @@
|
|||||||
# percent encoding so that the URL is properly parsed.
|
# percent encoding so that the URL is properly parsed.
|
||||||
|
|
||||||
.. # Sourcecode link to GitHub
|
.. # Sourcecode link to GitHub
|
||||||
.. |SHIELD:svg:pyTooling-github| image:: https://img.shields.io/badge/pyTooling-Actions-63bf7f.svg?longCache=true&style=flat-square&longCache=true&logo=GitHub
|
.. |SHIELD:svg:pyTooling-github| image:: https://img.shields.io/badge/pyTooling-Actions-63bf7f?longCache=true&style=flat-square&longCache=true&logo=GitHub
|
||||||
:alt: Sourcecode on GitHub
|
:alt: Sourcecode on GitHub
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: https://GitHub.com/pyTooling/Actions
|
:target: https://GitHub.com/pyTooling/Actions
|
||||||
.. |SHIELD:png:pyTooling-github| image:: https://raster.shields.io/badge/pyTooling-Actions-63bf7f.svg?longCache=true&style=flat-square&longCache=true&logo=GitHub
|
.. |SHIELD:png:pyTooling-github| image:: https://raster.shields.io/badge/pyTooling-Actions-63bf7f?longCache=true&style=flat-square&longCache=true&logo=GitHub
|
||||||
:alt: Sourcecode on GitHub
|
:alt: Sourcecode on GitHub
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: https://GitHub.com/pyTooling/Actions
|
:target: https://GitHub.com/pyTooling/Actions
|
||||||
@@ -18,7 +18,7 @@
|
|||||||
:alt: Code license
|
:alt: Code license
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: Code-License.html
|
:target: Code-License.html
|
||||||
.. |SHIELD:png:pyTooling-src-license| image:: https://img.shields.io/pypi/l/pyTooling?longCache=true&style=flat-square&logo=Apache&label=code
|
.. |SHIELD:png:pyTooling-src-license| image:: https://raster.shields.io/pypi/l/pyTooling?longCache=true&style=flat-square&logo=Apache&label=code
|
||||||
:alt: Code license
|
:alt: Code license
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: https://GitHub.com/pyTooling/Actions/blob/main/LICENSE.md
|
:target: https://GitHub.com/pyTooling/Actions/blob/main/LICENSE.md
|
||||||
@@ -64,11 +64,11 @@
|
|||||||
:target: https://pyTooling.github.io/pyTooling/
|
:target: https://pyTooling.github.io/pyTooling/
|
||||||
|
|
||||||
.. # Gitter
|
.. # Gitter
|
||||||
.. |SHIELD:svg:pyTooling-gitter| image:: https://img.shields.io/badge/chat-on%20gitter-4db797.svg?longCache=true&style=flat-square&logo=gitter&logoColor=e8ecef
|
.. |SHIELD:svg:pyTooling-gitter| image:: https://img.shields.io/badge/chat-on%20gitter-4db797?longCache=true&style=flat-square&logo=gitter&logoColor=e8ecef
|
||||||
:alt: Documentation License
|
:alt: Documentation License
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: https://gitter.im/hdl/community
|
:target: https://gitter.im/hdl/community
|
||||||
.. |SHIELD:png:pyTooling-gitter| image:: https://raster.shields.io/badge/chat-on%20gitter-4db797.svg?longCache=true&style=flat-square&logo=gitter&logoColor=e8ecef
|
.. |SHIELD:png:pyTooling-gitter| image:: https://raster.shields.io/badge/chat-on%20gitter-4db797?longCache=true&style=flat-square&logo=gitter&logoColor=e8ecef
|
||||||
:alt: Documentation License
|
:alt: Documentation License
|
||||||
:height: 22
|
:height: 22
|
||||||
:target: https://gitter.im/hdl/community
|
:target: https://gitter.im/hdl/community
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -34,7 +34,7 @@ A module for a set of dummy classes.
|
|||||||
|
|
||||||
__author__ = "Patrick Lehmann"
|
__author__ = "Patrick Lehmann"
|
||||||
__email__ = "Paebbels@gmail.com"
|
__email__ = "Paebbels@gmail.com"
|
||||||
__copyright__ = "2017-2024, Patrick Lehmann"
|
__copyright__ = "2017-2025, Patrick Lehmann"
|
||||||
__license__ = "Apache License, Version 2.0"
|
__license__ = "Apache License, Version 2.0"
|
||||||
__version__ = "0.4.4"
|
__version__ = "0.4.4"
|
||||||
__keywords__ = ["GitHub Actions"]
|
__keywords__ = ["GitHub Actions"]
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -34,7 +34,7 @@ A module for a set of dummy classes.
|
|||||||
|
|
||||||
__author__ = "Patrick Lehmann"
|
__author__ = "Patrick Lehmann"
|
||||||
__email__ = "Paebbels@gmail.com"
|
__email__ = "Paebbels@gmail.com"
|
||||||
__copyright__ = "2017-2024, Patrick Lehmann"
|
__copyright__ = "2017-2025, Patrick Lehmann"
|
||||||
__license__ = "Apache License, Version 2.0"
|
__license__ = "Apache License, Version 2.0"
|
||||||
__version__ = "0.14.8"
|
__version__ = "0.14.8"
|
||||||
__keywords__ = ["GitHub Actions"]
|
__keywords__ = ["GitHub Actions"]
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
[build-system]
|
[build-system]
|
||||||
requires = [
|
requires = [
|
||||||
"setuptools ~= 75.3",
|
"setuptools ~= 75.8",
|
||||||
"wheel ~= 0.45",
|
"wheel ~= 0.45",
|
||||||
"pyTooling ~= 8.0"
|
"pyTooling ~= 8.2"
|
||||||
]
|
]
|
||||||
build-backend = "setuptools.build_meta"
|
build-backend = "setuptools.build_meta"
|
||||||
|
|
||||||
@@ -20,6 +20,9 @@ show_error_codes = true
|
|||||||
namespace_packages = true
|
namespace_packages = true
|
||||||
html_report = "report/typing"
|
html_report = "report/typing"
|
||||||
|
|
||||||
|
[tool.pytest]
|
||||||
|
junit_xml = "report/unit/TestReportSummary.xml"
|
||||||
|
|
||||||
[tool.pytest.ini_options]
|
[tool.pytest.ini_options]
|
||||||
addopts = "--tb=native"
|
addopts = "--tb=native"
|
||||||
# Don't set 'python_classes = *' otherwise, pytest doesn't search for classes
|
# Don't set 'python_classes = *' otherwise, pytest doesn't search for classes
|
||||||
@@ -30,6 +33,7 @@ filterwarnings = [
|
|||||||
"error::DeprecationWarning",
|
"error::DeprecationWarning",
|
||||||
"error::PendingDeprecationWarning"
|
"error::PendingDeprecationWarning"
|
||||||
]
|
]
|
||||||
|
junit_logging = "all"
|
||||||
|
|
||||||
[tool.interrogate]
|
[tool.interrogate]
|
||||||
color = true
|
color = true
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
pyTooling ~= 8.0
|
pyTooling ~= 8.2
|
||||||
|
|||||||
6
run.ps1
6
run.ps1
@@ -88,7 +88,7 @@ if ($build)
|
|||||||
rm -Force .\build\bdist.win-amd64
|
rm -Force .\build\bdist.win-amd64
|
||||||
rm -Force .\build\lib
|
rm -Force .\build\lib
|
||||||
Write-Host -ForegroundColor Yellow "[live][BUILD] Building $PackageName package as wheel ..."
|
Write-Host -ForegroundColor Yellow "[live][BUILD] Building $PackageName package as wheel ..."
|
||||||
py -3.12 -m build --wheel
|
py -3.13 -m build --wheel
|
||||||
|
|
||||||
Write-Host -ForegroundColor Yellow "[live][BUILD] Building wheel finished"
|
Write-Host -ForegroundColor Yellow "[live][BUILD] Building wheel finished"
|
||||||
}
|
}
|
||||||
@@ -102,9 +102,9 @@ if ($install)
|
|||||||
}
|
}
|
||||||
else
|
else
|
||||||
{ Write-Host -ForegroundColor Cyan "[ADMIN][UNINSTALL] Uninstalling $PackageName ..."
|
{ Write-Host -ForegroundColor Cyan "[ADMIN][UNINSTALL] Uninstalling $PackageName ..."
|
||||||
py -3.12 -m pip uninstall -y $PackageName
|
py -3.13 -m pip uninstall -y $PackageName
|
||||||
Write-Host -ForegroundColor Cyan "[ADMIN][INSTALL] Installing $PackageName from wheel ..."
|
Write-Host -ForegroundColor Cyan "[ADMIN][INSTALL] Installing $PackageName from wheel ..."
|
||||||
py -3.12 -m pip install .\dist\$PackageName-6.7.0-py3-none-any.whl
|
py -3.13 -m pip install .\dist\$PackageName-8.1.0-py3-none-any.whl
|
||||||
|
|
||||||
Write-Host -ForegroundColor Cyan "[ADMIN][INSTALL] Closing window in 5 seconds ..."
|
Write-Host -ForegroundColor Cyan "[ADMIN][INSTALL] Closing window in 5 seconds ..."
|
||||||
Start-Sleep -Seconds 5
|
Start-Sleep -Seconds 5
|
||||||
|
|||||||
2
setup.py
2
setup.py
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ print(f"Python: {version}")
|
|||||||
|
|
||||||
def loadRequirementsFile(requirementsFile: Path):
|
def loadRequirementsFile(requirementsFile: Path):
|
||||||
requirements = []
|
requirements = []
|
||||||
with requirementsFile.open("r") as file:
|
with requirementsFile.open("r", encoding="utf-8") as file:
|
||||||
for line in file.readlines():
|
for line in file.readlines():
|
||||||
line = line.strip()
|
line = line.strip()
|
||||||
if line.startswith("#") or line.startswith("https") or line == "":
|
if line.startswith("#") or line.startswith("https") or line == "":
|
||||||
@@ -84,7 +84,7 @@ for dependency in dependencies:
|
|||||||
# Write jobs to special file
|
# Write jobs to special file
|
||||||
github_output = Path(getenv("GITHUB_OUTPUT"))
|
github_output = Path(getenv("GITHUB_OUTPUT"))
|
||||||
print(f"GITHUB_OUTPUT: {github_output}")
|
print(f"GITHUB_OUTPUT: {github_output}")
|
||||||
with github_output.open("a+") as f:
|
with github_output.open("a+", encoding="utf-8") as f:
|
||||||
f.write(f"pacboy_packages={' '.join(pacboyPackages)}\n")
|
f.write(f"pacboy_packages={' '.join(pacboyPackages)}\n")
|
||||||
|
|
||||||
print(f"GITHUB_OUTPUT:")
|
print(f"GITHUB_OUTPUT:")
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -8,6 +8,6 @@ pytest ~= 8.3
|
|||||||
pytest-cov ~= 6.0
|
pytest-cov ~= 6.0
|
||||||
|
|
||||||
# Static Type Checking
|
# Static Type Checking
|
||||||
mypy ~= 1.13
|
mypy ~= 1.15
|
||||||
typing_extensions ~= 4.12
|
typing_extensions ~= 4.12
|
||||||
lxml ~= 5.3
|
lxml ~= 5.3
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
# #
|
# #
|
||||||
# License: #
|
# License: #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2017-2024 Patrick Lehmann - Bötzingen, Germany #
|
# Copyright 2017-2025 Patrick Lehmann - Bötzingen, Germany #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
Reference in New Issue
Block a user