mirror of
https://github.com/pyTooling/Actions.git
synced 2026-02-12 02:56:56 +08:00
v5.0.0
This commit is contained in:
3
.github/CODEOWNERS
vendored
Normal file
3
.github/CODEOWNERS
vendored
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
* @Paebbels
|
||||||
|
|
||||||
|
/.github/ @Paebbels
|
||||||
2
.github/dependabot.yml
vendored
2
.github/dependabot.yml
vendored
@@ -10,7 +10,5 @@ updates:
|
|||||||
- Dependencies
|
- Dependencies
|
||||||
assignees:
|
assignees:
|
||||||
- Paebbels
|
- Paebbels
|
||||||
reviewers:
|
|
||||||
- Paebbels
|
|
||||||
schedule:
|
schedule:
|
||||||
interval: "daily" # Checks on Monday trough Friday.
|
interval: "daily" # Checks on Monday trough Friday.
|
||||||
|
|||||||
7
.github/workflows/ApplicationTesting.yml
vendored
7
.github/workflows/ApplicationTesting.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -134,20 +134,21 @@ jobs:
|
|||||||
|
|
||||||
packages = {
|
packages = {
|
||||||
"coverage": "python-coverage:p",
|
"coverage": "python-coverage:p",
|
||||||
"docstr_coverage": "python-pyyaml:p",
|
"docstr_coverage": "python-pyyaml:p python-types-pyyaml:p",
|
||||||
"igraph": "igraph:p",
|
"igraph": "igraph:p",
|
||||||
"jinja2": "python-markupsafe:p",
|
"jinja2": "python-markupsafe:p",
|
||||||
"lxml": "python-lxml:p",
|
"lxml": "python-lxml:p",
|
||||||
"numpy": "python-numpy:p",
|
"numpy": "python-numpy:p",
|
||||||
"markupsafe": "python-markupsafe:p",
|
"markupsafe": "python-markupsafe:p",
|
||||||
"pip": "python-pip:p",
|
"pip": "python-pip:p",
|
||||||
"pyyaml": "python-pyyaml:p",
|
"pyyaml": "python-pyyaml:p python-types-pyyaml:p",
|
||||||
"ruamel.yaml": "python-ruamel-yaml:p python-ruamel.yaml.clib:p",
|
"ruamel.yaml": "python-ruamel-yaml:p python-ruamel.yaml.clib:p",
|
||||||
"sphinx": "python-markupsafe:p",
|
"sphinx": "python-markupsafe:p",
|
||||||
"tomli": "python-tomli:p",
|
"tomli": "python-tomli:p",
|
||||||
"wheel": "python-wheel:p",
|
"wheel": "python-wheel:p",
|
||||||
"pyEDAA.ProjectModel": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
"pyEDAA.ProjectModel": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
||||||
"pyEDAA.Reports": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
"pyEDAA.Reports": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
||||||
|
"sphinx-reports": "python-markupsafe:p python-pyaml:p python-types-pyyaml:p",
|
||||||
}
|
}
|
||||||
subPackages = {
|
subPackages = {
|
||||||
"pytooling": {
|
"pytooling": {
|
||||||
|
|||||||
2
.github/workflows/ArtifactCleanUp.yml
vendored
2
.github/workflows/ArtifactCleanUp.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
4
.github/workflows/BuildTheDocs.yml
vendored
4
.github/workflows/BuildTheDocs.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -38,7 +38,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: '❗ Deprecation message'
|
- name: '❗ Deprecation message'
|
||||||
run: printf "%s\n" "::warning title=Deprecated::'BuildTheDocs.yml' is not maintained anymore. Please switch to 'SphinxDocumentation.yml', 'LaTeXDocumentation.yml' and 'ExtractConfiguration.yml'."
|
run: printf "::warning title=%s::%s\n" "Deprecated" "'BuildTheDocs.yml' is not maintained anymore. Please switch to 'SphinxDocumentation.yml', 'LaTeXDocumentation.yml' and 'ExtractConfiguration.yml'."
|
||||||
|
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|||||||
17
.github/workflows/CheckDocumentation.yml
vendored
17
.github/workflows/CheckDocumentation.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -64,9 +64,20 @@ jobs:
|
|||||||
- name: Run 'interrogate' Documentation Coverage Check
|
- name: Run 'interrogate' Documentation Coverage Check
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
interrogate -c pyproject.toml --fail-under=${{ inputs.fail_under }} && printf "%s\n" "::error title=interrogate::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
set +e
|
||||||
|
|
||||||
|
interrogate -c pyproject.toml --fail-under=${{ inputs.fail_under }} ${{ inputs.directory }}
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "::error title=%s::%s\n" "interrogate" "Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
||||||
|
fi
|
||||||
|
|
||||||
- name: Run 'docstr_coverage' Documentation Coverage Check
|
- name: Run 'docstr_coverage' Documentation Coverage Check
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
docstr-coverage -v 2 --fail-under=${{ inputs.fail_under }} ${{ inputs.directory }} && printf "%s\n" "::error title=docstr-coverage::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
set +e
|
||||||
|
|
||||||
|
docstr-coverage --fail-under=${{ inputs.fail_under }} ${{ inputs.directory }}
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "%s\n" "::error title=docstr-coverage::Insufficient documentation quality (goal: ${{ inputs.fail_under }})"
|
||||||
|
fi
|
||||||
|
|
||||||
|
|||||||
118
.github/workflows/CompletePipeline.yml
vendored
118
.github/workflows/CompletePipeline.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -96,23 +96,23 @@ on:
|
|||||||
codecov:
|
codecov:
|
||||||
description: 'Publish merged coverage and unittest reports to Codecov.'
|
description: 'Publish merged coverage and unittest reports to Codecov.'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
codacy:
|
codacy:
|
||||||
description: 'Publish merged coverage report to Codacy.'
|
description: 'Publish merged coverage report to Codacy.'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
dorny:
|
dorny:
|
||||||
description: 'Publish merged unittest report via Dorny Test-Reporter.'
|
description: 'Publish merged unittest report via Dorny Test-Reporter.'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
cleanup:
|
cleanup:
|
||||||
description: 'Cleanup artifacts afterwards.'
|
description: 'Cleanup artifacts afterwards.'
|
||||||
required: false
|
required: false
|
||||||
default: true
|
default: 'true'
|
||||||
type: boolean
|
type: string
|
||||||
secrets:
|
secrets:
|
||||||
PYPI_TOKEN:
|
PYPI_TOKEN:
|
||||||
description: "Token for pushing releases to PyPI."
|
description: "Token for pushing releases to PyPI."
|
||||||
@@ -120,11 +120,14 @@ on:
|
|||||||
CODECOV_TOKEN:
|
CODECOV_TOKEN:
|
||||||
description: "Token for pushing coverage and unittest results to Codecov."
|
description: "Token for pushing coverage and unittest results to Codecov."
|
||||||
required: false
|
required: false
|
||||||
CODACY_PROJECT_TOKEN:
|
CODACY_TOKEN:
|
||||||
description: "Token for pushing coverage results to Codacy."
|
description: "Token for pushing coverage results to Codacy."
|
||||||
required: false
|
required: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
Prepare:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/PrepareJob.yml@main
|
||||||
|
|
||||||
ConfigParams:
|
ConfigParams:
|
||||||
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
||||||
with:
|
with:
|
||||||
@@ -143,17 +146,29 @@ jobs:
|
|||||||
exclude_list: ${{ inputs.unittest_exclude_list }}
|
exclude_list: ${{ inputs.unittest_exclude_list }}
|
||||||
disable_list: ${{ inputs.unittest_disable_list }}
|
disable_list: ${{ inputs.unittest_disable_list }}
|
||||||
|
|
||||||
AppTestingParams:
|
# AppTestingParams:
|
||||||
|
# uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
||||||
|
# with:
|
||||||
|
# package_namespace: ${{ inputs.package_namespace }}
|
||||||
|
# package_name: ${{ inputs.package_name }}
|
||||||
|
# python_version: ${{ inputs.apptest_python_version }}
|
||||||
|
# python_version_list: ${{ inputs.apptest_python_version_list }}
|
||||||
|
# system_list: ${{ inputs.apptest_system_list }}
|
||||||
|
# include_list: ${{ inputs.apptest_include_list }}
|
||||||
|
# exclude_list: ${{ inputs.apptest_exclude_list }}
|
||||||
|
# disable_list: ${{ inputs.apptest_disable_list }}
|
||||||
|
|
||||||
|
InstallParams:
|
||||||
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
||||||
with:
|
with:
|
||||||
package_namespace: ${{ inputs.package_namespace }}
|
package_namespace: ${{ inputs.package_namespace }}
|
||||||
package_name: ${{ inputs.package_name }}
|
package_name: ${{ inputs.package_name }}
|
||||||
python_version: ${{ inputs.apptest_python_version }}
|
python_version: ${{ inputs.unittest_python_version }}
|
||||||
python_version_list: ${{ inputs.apptest_python_version_list }}
|
python_version_list: ''
|
||||||
system_list: ${{ inputs.apptest_system_list }}
|
system_list: ${{ inputs.unittest_system_list }}
|
||||||
include_list: ${{ inputs.apptest_include_list }}
|
include_list: ${{ inputs.unittest_include_list }}
|
||||||
exclude_list: ${{ inputs.apptest_exclude_list }}
|
exclude_list: ${{ inputs.unittest_exclude_list }}
|
||||||
disable_list: ${{ inputs.apptest_disable_list }}
|
disable_list: ${{ inputs.unittest_disable_list }}
|
||||||
|
|
||||||
UnitTesting:
|
UnitTesting:
|
||||||
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
uses: pyTooling/Actions/.github/workflows/UnitTesting.yml@main
|
||||||
@@ -190,18 +205,29 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
directory: ${{ inputs.package_namespace }}/${{ inputs.package_name }}
|
directory: ${{ needs.ConfigParams.outputs.package_directory }}
|
||||||
# fail_below: 70
|
|
||||||
|
|
||||||
Package:
|
Package:
|
||||||
uses: pyTooling/Actions/.github/workflows/Package.yml@main
|
uses: pyTooling/Actions/.github/workflows/Package.yml@main
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
# - UnitTesting
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
|
|
||||||
|
Install:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/InstallPackage.yml@main
|
||||||
|
needs:
|
||||||
|
- ConfigParams
|
||||||
|
- UnitTestingParams
|
||||||
|
- InstallParams
|
||||||
|
- Package
|
||||||
|
with:
|
||||||
|
jobs: ${{ needs.InstallParams.outputs.python_jobs }}
|
||||||
|
wheel: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
|
package_name: ${{ needs.ConfigParams.outputs.package_fullname }}
|
||||||
|
|
||||||
# AppTesting:
|
# AppTesting:
|
||||||
# uses: pyTooling/Actions/.github/workflows/ApplicationTesting.yml@main
|
# uses: pyTooling/Actions/.github/workflows/ApplicationTesting.yml@main
|
||||||
# needs:
|
# needs:
|
||||||
@@ -233,7 +259,7 @@ jobs:
|
|||||||
codacy: ${{ inputs.codacy }}
|
codacy: ${{ inputs.codacy }}
|
||||||
secrets:
|
secrets:
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
||||||
CODACY_TOKEN: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
CODACY_TOKEN: ${{ secrets.CODACY_TOKEN }}
|
||||||
|
|
||||||
PublishTestResults:
|
PublishTestResults:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
||||||
@@ -242,13 +268,11 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
with:
|
with:
|
||||||
additional_merge_args: '-d "--pytest=rewrite-dunder-init;reduce-depth:pytest.tests.unit"'
|
testsuite-summary-name: ${{ needs.ConfigParams.outputs.package_fullname }}
|
||||||
testsuite-summary-name: ${{ inputs.package_name }}
|
|
||||||
merged_junit_filename: ${{ needs.ConfigParams.outputs.unittest_merged_report_xml_filename }}
|
merged_junit_filename: ${{ needs.ConfigParams.outputs.unittest_merged_report_xml_filename }}
|
||||||
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
dorny: ${{ inputs.dorny }}
|
dorny: ${{ inputs.dorny }}
|
||||||
codecov: ${{ inputs.codecov }}
|
codecov: ${{ inputs.codecov }}
|
||||||
|
|
||||||
secrets:
|
secrets:
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
||||||
|
|
||||||
@@ -281,8 +305,7 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
- PublishTestResults
|
- PublishTestResults
|
||||||
- Documentation
|
if: inputs.cleanup == 'true'
|
||||||
if: ${{ inputs.cleanup }}
|
|
||||||
with:
|
with:
|
||||||
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
||||||
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
||||||
@@ -310,20 +333,51 @@ jobs:
|
|||||||
coverage: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
coverage: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
typing: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
typing: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
||||||
|
|
||||||
ReleasePage:
|
TriggerTaggedRelease:
|
||||||
uses: pyTooling/Actions/.github/workflows/Release.yml@main
|
uses: pyTooling/Actions/.github/workflows/TagReleaseCommit.yml@main
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
needs:
|
||||||
- Package
|
- Prepare
|
||||||
|
- UnitTesting
|
||||||
|
- Install
|
||||||
# - AppTesting
|
# - AppTesting
|
||||||
|
# - StaticTypeCheck
|
||||||
|
- Package
|
||||||
- PublishToGitHubPages
|
- PublishToGitHubPages
|
||||||
|
if: needs.Prepare.outputs.is_release_commit
|
||||||
|
permissions:
|
||||||
|
contents: write # required for create tag
|
||||||
|
actions: write # required for trigger workflow
|
||||||
|
with:
|
||||||
|
version: ${{ needs.Prepare.outputs.version }}
|
||||||
|
auto_tag: ${{ needs.Prepare.outputs.is_release_commit }}
|
||||||
|
secrets: inherit
|
||||||
|
|
||||||
|
ReleasePage:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/PublishReleaseNotes.yml@main
|
||||||
|
needs:
|
||||||
|
- Prepare
|
||||||
|
- UnitTesting
|
||||||
|
- Install
|
||||||
|
# - AppTesting
|
||||||
|
# - StaticTypeCheck
|
||||||
|
- Package
|
||||||
|
- PublishToGitHubPages
|
||||||
|
if: needs.Prepare.outputs.is_release_tag == 'true'
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
with:
|
||||||
|
tag: ${{ needs.Prepare.outputs.version }}
|
||||||
|
secrets: inherit
|
||||||
|
|
||||||
PublishOnPyPI:
|
PublishOnPyPI:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishOnPyPI.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishOnPyPI.yml@main
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
needs:
|
||||||
|
- Prepare
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
|
- Package
|
||||||
- ReleasePage
|
- ReleasePage
|
||||||
|
if: needs.Prepare.outputs.is_release_tag == 'true'
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
requirements: -r dist/requirements.txt
|
requirements: -r dist/requirements.txt
|
||||||
@@ -344,7 +398,7 @@ jobs:
|
|||||||
- PublishToGitHubPages
|
- PublishToGitHubPages
|
||||||
# - PublishOnPyPI
|
# - PublishOnPyPI
|
||||||
- IntermediateCleanUp
|
- IntermediateCleanUp
|
||||||
if: ${{ inputs.cleanup }}
|
if: inputs.cleanup == 'true'
|
||||||
with:
|
with:
|
||||||
package: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
package: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
remaining: |
|
remaining: |
|
||||||
|
|||||||
4
.github/workflows/CoverageCollection.yml
vendored
4
.github/workflows/CoverageCollection.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -72,7 +72,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: '❗ Deprecation message'
|
- name: '❗ Deprecation message'
|
||||||
run: printf "%s\n" "::warning title=Deprecated::'CoverageCollection.yml' is not maintained anymore. Please switch to 'UnitTesting.yml', 'PublishCoverageResults.yml' and 'PublishTestResults.yml'."
|
run: printf "::warning title=%s::%s\n" "Deprecated" "'CoverageCollection.yml' is not maintained anymore. Please switch to 'UnitTesting.yml', 'PublishCoverageResults.yml' and 'PublishTestResults.yml'."
|
||||||
|
|
||||||
- name: ⏬ Checkout repository
|
- name: ⏬ Checkout repository
|
||||||
uses: actions/checkout@v4
|
uses: actions/checkout@v4
|
||||||
|
|||||||
24
.github/workflows/ExtractConfiguration.yml
vendored
24
.github/workflows/ExtractConfiguration.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -145,6 +145,12 @@ jobs:
|
|||||||
namespace = "${{ inputs.package_namespace }}".strip()
|
namespace = "${{ inputs.package_namespace }}".strip()
|
||||||
name = "${{ inputs.package_name }}".strip()
|
name = "${{ inputs.package_name }}".strip()
|
||||||
|
|
||||||
|
print(dedent(f"""\
|
||||||
|
INPUTS:
|
||||||
|
package_namespace: {namespace}
|
||||||
|
package_name: {name}
|
||||||
|
"""))
|
||||||
|
|
||||||
if namespace == "" or namespace == ".":
|
if namespace == "" or namespace == ".":
|
||||||
fullname = f"{name}"
|
fullname = f"{name}"
|
||||||
directory = f"{name}"
|
directory = f"{name}"
|
||||||
@@ -154,6 +160,13 @@ jobs:
|
|||||||
directory = f"{namespace}/{name}"
|
directory = f"{namespace}/{name}"
|
||||||
mypy_prepare_command = f"touch {namespace}/__init__.py"
|
mypy_prepare_command = f"touch {namespace}/__init__.py"
|
||||||
|
|
||||||
|
print(dedent(f"""\
|
||||||
|
OUTPUTS:
|
||||||
|
package_fullname: {fullname}
|
||||||
|
package_directory: {directory}
|
||||||
|
mypy_prepare_command: {mypy_prepare_command}
|
||||||
|
"""))
|
||||||
|
|
||||||
github_output = Path(getenv("GITHUB_OUTPUT"))
|
github_output = Path(getenv("GITHUB_OUTPUT"))
|
||||||
print(f"GITHUB_OUTPUT: {github_output}")
|
print(f"GITHUB_OUTPUT: {github_output}")
|
||||||
with github_output.open("a+", encoding="utf-8") as f:
|
with github_output.open("a+", encoding="utf-8") as f:
|
||||||
@@ -234,4 +247,11 @@ jobs:
|
|||||||
coverage_report_json={coverageJSONFile.as_posix()}
|
coverage_report_json={coverageJSONFile.as_posix()}
|
||||||
"""))
|
"""))
|
||||||
|
|
||||||
print(f"DEBUG:\n unittest xml: {unittestXMLFile}\n merged unittest xml: {mergedUnittestXMLFile}\n coverage html: {coverageHTMLDirectory}\n coverage xml: {coverageXMLFile}\n coverage json: {coverageJSONFile}")
|
print(dedent(f"""\
|
||||||
|
DEBUG:
|
||||||
|
unittest xml: {unittestXMLFile}
|
||||||
|
merged unittest xml: {mergedUnittestXMLFile}
|
||||||
|
coverage html: {coverageHTMLDirectory}
|
||||||
|
coverage xml: {coverageXMLFile}
|
||||||
|
coverage json: {coverageJSONFile}
|
||||||
|
"""))
|
||||||
|
|||||||
130
.github/workflows/InstallPackage.yml
vendored
Normal file
130
.github/workflows/InstallPackage.yml
vendored
Normal file
@@ -0,0 +1,130 @@
|
|||||||
|
# ==================================================================================================================== #
|
||||||
|
# Authors: #
|
||||||
|
# Patrick Lehmann #
|
||||||
|
# #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
# Copyright 2025-2025 The pyTooling Authors #
|
||||||
|
# #
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
|
# you may not use this file except in compliance with the License. #
|
||||||
|
# You may obtain a copy of the License at #
|
||||||
|
# #
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0 #
|
||||||
|
# #
|
||||||
|
# Unless required by applicable law or agreed to in writing, software #
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, #
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #
|
||||||
|
# See the License for the specific language governing permissions and #
|
||||||
|
# limitations under the License. #
|
||||||
|
# #
|
||||||
|
# SPDX-License-Identifier: Apache-2.0 #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
name: Install Package
|
||||||
|
|
||||||
|
on:
|
||||||
|
workflow_call:
|
||||||
|
inputs:
|
||||||
|
jobs:
|
||||||
|
description: 'JSON list with environment fields, telling the system and Python versions to run tests with.'
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
wheel:
|
||||||
|
description: "Wheel package as input artifact."
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
package_name:
|
||||||
|
description: "Name of the Python package."
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
PackageInstallation:
|
||||||
|
name: ${{ matrix.sysicon }} ${{ matrix.pyicon }} Package installation using Python ${{ matrix.python }}
|
||||||
|
runs-on: ${{ matrix.runs-on }}
|
||||||
|
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
include: ${{ fromJson(inputs.jobs) }}
|
||||||
|
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
shell: ${{ matrix.shell }}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: 📥 Download artifacts '${{ inputs.wheel }}' from 'Package' job
|
||||||
|
uses: pyTooling/download-artifact@v4
|
||||||
|
with:
|
||||||
|
name: ${{ inputs.wheel }}
|
||||||
|
path: install
|
||||||
|
|
||||||
|
- name: '🟦 Setup MSYS2 for ${{ matrix.runtime }}'
|
||||||
|
uses: msys2/setup-msys2@v2
|
||||||
|
if: matrix.system == 'msys2'
|
||||||
|
with:
|
||||||
|
msystem: ${{ matrix.runtime }}
|
||||||
|
update: true
|
||||||
|
pacboy: >-
|
||||||
|
python-pip:p python-wheel:p
|
||||||
|
python-lxml:p
|
||||||
|
python-markupsafe:p
|
||||||
|
python-pyaml:p python-types-pyyaml:p
|
||||||
|
python-ruamel-yaml:p python-ruamel.yaml.clib:p
|
||||||
|
python-tomli:p
|
||||||
|
|
||||||
|
- name: 🐍 Setup Python ${{ matrix.python }}
|
||||||
|
uses: actions/setup-python@v5
|
||||||
|
if: matrix.system != 'msys2'
|
||||||
|
with:
|
||||||
|
python-version: ${{ matrix.python }}
|
||||||
|
|
||||||
|
- name: 🔧 Install wheel and pip dependencies (native)
|
||||||
|
if: matrix.system != 'msys2'
|
||||||
|
run: |
|
||||||
|
python -m pip install --disable-pip-version-check -U wheel
|
||||||
|
|
||||||
|
- name: 🔧 Install wheel from artifact (Ubuntu/macOS)
|
||||||
|
if: matrix.system != 'windows'
|
||||||
|
run: |
|
||||||
|
python -m pip install --disable-pip-version-check -U install/*.whl
|
||||||
|
|
||||||
|
- name: 🔧 Install wheel from artifact (Windows)
|
||||||
|
if: matrix.system == 'windows'
|
||||||
|
run: |
|
||||||
|
python -m pip install -v --disable-pip-version-check (Get-Item .\install\*.whl).FullName
|
||||||
|
|
||||||
|
- name: 📦 Run application tests (Ubuntu/macOS)
|
||||||
|
if: matrix.system != 'windows'
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
printf "Import package and checking package version ...\n "
|
||||||
|
python3 - << EOF | tee ImportTest.log | grep -E "^Package version:\s+[0-9]+\.[0-9]+\.[0-9]+"
|
||||||
|
from ${{ inputs.package_name }} import __version__
|
||||||
|
|
||||||
|
print(f"Package version: {__version__}")
|
||||||
|
EOF
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf " ${ANSI_LIGHT_GREEN}[PASSED]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf " ${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "::error title=%s::%s\n" "InstallPackage" "Couldn't check package version of '${{ inputs.package_name }}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📦 Run application tests (Windows)
|
||||||
|
if: matrix.system == 'windows'
|
||||||
|
run: |
|
||||||
|
$result=$(python -c "from ${{ inputs.package_name }} import __version__; print(f""Package version: {__version__}"")")
|
||||||
|
Write-Host $result
|
||||||
|
if ($result -match "Package version:\s+\d+\.\d+\.\d+") {
|
||||||
|
Write-Host -ForegroundColor Green "[PASSED]"
|
||||||
|
} else {
|
||||||
|
Write-Host -ForegroundColor Red "[FAILED]"
|
||||||
|
Write-Host ("::error title={0}::{1}" -f "InstallPackage", "Couldn't check package version of '${{ inputs.package_name }}'.")
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
2
.github/workflows/IntermediateCleanUp.yml
vendored
2
.github/workflows/IntermediateCleanUp.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
2
.github/workflows/LaTeXDocumentation.yml
vendored
2
.github/workflows/LaTeXDocumentation.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
24
.github/workflows/NightlyRelease.yml
vendored
24
.github/workflows/NightlyRelease.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -130,7 +130,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=InternalError::Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
printf "::error title=%s::%s\n" "InternalError" "Couldn't delete release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -190,7 +190,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=InternalError::Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
printf "::error title=%s::%s\n" "InternalError" "Couldn't create release '${{ inputs.nightly_name }}' -> Error: '${message}'."
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -289,7 +289,7 @@ jobs:
|
|||||||
printf " %s" "Checked asset for duplicates ... "
|
printf " %s" "Checked asset for duplicates ... "
|
||||||
if [[ -n "${assetFilenames[$asset]}" ]]; then
|
if [[ -n "${assetFilenames[$asset]}" ]]; then
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=DuplicateAsset::Asset '${asset}' from artifact '${artifact}' was already uploaded to release '${{ inputs.nightly_name }}'."
|
printf "::error title=%s::%s\n" "DuplicateAsset" "Asset '${asset}' from artifact '${artifact}' was already uploaded to release '${{ inputs.nightly_name }}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
else
|
else
|
||||||
@@ -309,7 +309,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't download artifact '${artifact}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't download artifact '${artifact}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=ArtifactNotFound::Couldn't download artifact '${artifact}'."
|
printf "::error title=%s::%s\n" "ArtifactNotFound" "Couldn't download artifact '${artifact}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -361,7 +361,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zip file '${asset}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zip file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to zip file '${asset}'."
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to zip file '${asset}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -390,7 +390,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to tgz file '${asset}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to tgz file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to tgz file '${asset}'."
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to tgz file '${asset}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -419,7 +419,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf " %s\n" "Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zst file '${asset}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't compress '${artifact}' to zst file '${asset}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=CompressionError::Couldn't compress '${artifact}' to zst file '${asset}'."
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to zst file '${asset}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -429,7 +429,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't find asset '${asset}' in artifact '${artifact}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't find asset '${asset}' in artifact '${artifact}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=FileNotFound::Couldn't find asset '${asset}' in artifact '${artifact}'."
|
printf "::error title=%s::%s\n" "FileNotFound" "Couldn't find asset '${asset}' in artifact '${artifact}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -469,7 +469,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=UploadError::Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'."
|
printf "::error title=%s::%s\n" "UploadError" "Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.nightly_name }}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -492,7 +492,7 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=UploadError::Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'."
|
printf "::error title=%s::%s\n" "UploadError" "Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.nightly_name }}'."
|
||||||
ERRORS=$((ERRORS + 1))
|
ERRORS=$((ERRORS + 1))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
@@ -526,5 +526,5 @@ jobs:
|
|||||||
else
|
else
|
||||||
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
printf "%s\n" "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}"
|
||||||
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't remove draft-state from release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
printf " %s\n" "${ANSI_LIGHT_RED}Couldn't remove draft-state from release '${{ inputs.nightly_name }}'.${ANSI_NOCOLOR}"
|
||||||
printf "%s\n" "::error title=ReleasePage::Couldn't remove draft-state from release '${{ inputs.nightly_name }}'."
|
printf "::error title=%s::%s\n" "ReleasePage" "Couldn't remove draft-state from release '${{ inputs.nightly_name }}'."
|
||||||
fi
|
fi
|
||||||
|
|||||||
2
.github/workflows/Package.yml
vendored
2
.github/workflows/Package.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
13
.github/workflows/Parameters.yml
vendored
13
.github/workflows/Parameters.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -95,6 +95,11 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: 'macos-14'
|
default: 'macos-14'
|
||||||
type: string
|
type: string
|
||||||
|
pipeline-delay:
|
||||||
|
description: 'Slow down this job, to delay the startup of the GitHub Action pipline.'
|
||||||
|
required: false
|
||||||
|
default: 0
|
||||||
|
type: number
|
||||||
|
|
||||||
outputs:
|
outputs:
|
||||||
python_version:
|
python_version:
|
||||||
@@ -121,6 +126,12 @@ jobs:
|
|||||||
params: ${{ steps.params.outputs.params }}
|
params: ${{ steps.params.outputs.params }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
|
- name: Generate a startup delay of ${{ inputs.pipeline-delay }} seconds
|
||||||
|
id: delay
|
||||||
|
if: inputs.pipeline-delay >= 0
|
||||||
|
run: |
|
||||||
|
sleep ${{ inputs.pipeline-delay }}
|
||||||
|
|
||||||
- name: Generate 'params' and 'python_jobs'
|
- name: Generate 'params' and 'python_jobs'
|
||||||
id: params
|
id: params
|
||||||
shell: python
|
shell: python
|
||||||
|
|||||||
334
.github/workflows/PrepareJob.yml
vendored
Normal file
334
.github/workflows/PrepareJob.yml
vendored
Normal file
@@ -0,0 +1,334 @@
|
|||||||
|
name: Prepare Variables
|
||||||
|
|
||||||
|
on:
|
||||||
|
workflow_call:
|
||||||
|
inputs:
|
||||||
|
ubuntu_image:
|
||||||
|
description: 'Name of the Ubuntu image.'
|
||||||
|
required: false
|
||||||
|
default: 'ubuntu-24.04'
|
||||||
|
type: string
|
||||||
|
main_branch:
|
||||||
|
description: 'Name of the branch containing releases.'
|
||||||
|
required: false
|
||||||
|
default: 'main'
|
||||||
|
type: string
|
||||||
|
development_branch:
|
||||||
|
description: 'Name of the development branch containing features.'
|
||||||
|
required: false
|
||||||
|
default: 'dev'
|
||||||
|
type: string
|
||||||
|
release_branch:
|
||||||
|
description: 'Name of the branch containing releases and nightly builds.'
|
||||||
|
required: false
|
||||||
|
default: 'main'
|
||||||
|
type: string
|
||||||
|
nightly_tag_pattern:
|
||||||
|
description: 'Pattern for nightly tags on the release branch.'
|
||||||
|
required: false
|
||||||
|
default: 'nightly'
|
||||||
|
type: string
|
||||||
|
release_tag_pattern:
|
||||||
|
description: 'Pattern for release tags on the release branch. Usually: vXX.YY.ZZ'
|
||||||
|
required: false
|
||||||
|
default: '(v|r)?[0-9]+(\.[0-9]+){0,2}(-(dev|alpha|beta|rc)([0-9]*))?'
|
||||||
|
type: string
|
||||||
|
|
||||||
|
outputs:
|
||||||
|
on_main_branch:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.on_main_branch }}
|
||||||
|
on_dev_branch:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.on_dev_branch }}
|
||||||
|
on_release_branch:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.on_release_branch }}
|
||||||
|
is_regular_commit:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.is_regular_commit }}
|
||||||
|
is_merge_commit:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.is_merge_commit }}
|
||||||
|
is_release_commit:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.is_release_commit }}
|
||||||
|
is_nightly_tag:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.is_nightly_tag }}
|
||||||
|
is_release_tag:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.is_release_tag }}
|
||||||
|
ref_kind:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.ref_kind }}
|
||||||
|
branch:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.branch }}
|
||||||
|
tag:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.tag }}
|
||||||
|
version:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.version }}
|
||||||
|
pr_title:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.pr_title }}
|
||||||
|
pr_number:
|
||||||
|
description: ""
|
||||||
|
value: ${{ jobs.Prepare.outputs.pr_number }}
|
||||||
|
# pr_mergedby:
|
||||||
|
# description: ""
|
||||||
|
# value: ${{ jobs.Prepare.outputs.pr_mergedby }}
|
||||||
|
# pr_mergedat:
|
||||||
|
# description: ""
|
||||||
|
# value: ${{ jobs.Prepare.outputs.pr_mergedat }}
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
Prepare:
|
||||||
|
name: Extract Information
|
||||||
|
runs-on: ubuntu-24.04
|
||||||
|
outputs:
|
||||||
|
on_main_branch: ${{ steps.Classify.outputs.on_main_branch }}
|
||||||
|
on_dev_branch: ${{ steps.Classify.outputs.on_dev_branch }}
|
||||||
|
on_release_branch: ${{ steps.Classify.outputs.on_release_branch }}
|
||||||
|
is_regular_commit: ${{ steps.Classify.outputs.is_regular_commit }}
|
||||||
|
is_merge_commit: ${{ steps.Classify.outputs.is_merge_commit }}
|
||||||
|
is_release_commit: ${{ steps.Classify.outputs.is_release_commit }}
|
||||||
|
is_nightly_tag: ${{ steps.Classify.outputs.is_nightly_tag }}
|
||||||
|
is_release_tag: ${{ steps.Classify.outputs.is_release_tag }}
|
||||||
|
ref_kind: ${{ steps.Classify.outputs.ref_kind }}
|
||||||
|
branch: ${{ steps.Classify.outputs.branch }}
|
||||||
|
tag: ${{ steps.Classify.outputs.tag }}
|
||||||
|
version: ${{ steps.Classify.outputs.version || steps.FindPullRequest.outputs.pr_version }}
|
||||||
|
# release_version: ${{ steps.FindPullRequest.outputs.release_version }}
|
||||||
|
pr_title: ${{ steps.FindPullRequest.outputs.pr_title }}
|
||||||
|
pr_number: ${{ steps.FindPullRequest.outputs.pr_number }}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: ⏬ Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
# The command 'git describe' (used for version) needs the history.
|
||||||
|
fetch-depth: 0
|
||||||
|
|
||||||
|
- name: 🖉 GitHub context information
|
||||||
|
run: |
|
||||||
|
printf "%s\n" "github.event_name: ${{ github.event_name }}"
|
||||||
|
printf "%s\n" "github.actor: ${{ github.actor }}"
|
||||||
|
printf "%s\n" "github.ref: ${{ github.ref }}"
|
||||||
|
printf "%s\n" "github.base_ref: ${{ github.base_ref }}"
|
||||||
|
printf "%s\n" "github.head_ref: ${{ github.head_ref }}"
|
||||||
|
printf "%s\n" "github.sha: ${{ github.sha }}"
|
||||||
|
|
||||||
|
- name: 🖉 Classify commit
|
||||||
|
id: Classify
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
ref="${{ github.ref }}"
|
||||||
|
on_main_branch="false"
|
||||||
|
on_dev_branch="false"
|
||||||
|
on_release_branch="false"
|
||||||
|
is_regular_commit="false"
|
||||||
|
is_merge_commit="false"
|
||||||
|
is_release_commit="false"
|
||||||
|
is_nightly_tag="false"
|
||||||
|
is_release_tag="false"
|
||||||
|
ref_kind="unknown"
|
||||||
|
branch=""
|
||||||
|
tag=""
|
||||||
|
version=""
|
||||||
|
|
||||||
|
if [[ "${ref:0:11}" == "refs/heads/" ]]; then
|
||||||
|
ref_kind="branch"
|
||||||
|
branch="${ref:11}"
|
||||||
|
|
||||||
|
printf "Commit check:\n"
|
||||||
|
|
||||||
|
if [[ "${branch}" == "${{ inputs.main_branch }}" ]]; then
|
||||||
|
on_main_branch="true"
|
||||||
|
|
||||||
|
if [[ -z "$(git rev-list -1 --merges ${{ github.sha }}~1..${{ github.sha }})" ]]; then
|
||||||
|
is_regular_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_YELLOW}regular "
|
||||||
|
else
|
||||||
|
is_merge_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_GREEN}merge "
|
||||||
|
fi
|
||||||
|
printf "commit${ANSI_NOCOLOR} on main branch ${ANSI_LIGHT_BLUE}'%s'${ANSI_NOCOLOR}\n" "${{ inputs.main_branch }}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${branch}" == "${{ inputs.development_branch }}" ]]; then
|
||||||
|
on_dev_branch="true"
|
||||||
|
|
||||||
|
if [[ -z "$(git rev-list -1 --merges ${{ github.sha }}~1..${{ github.sha }})" ]]; then
|
||||||
|
is_regular_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_YELLOW}regular "
|
||||||
|
else
|
||||||
|
is_merge_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_GREEN}merge "
|
||||||
|
fi
|
||||||
|
printf "commit${ANSI_NOCOLOR} on development branch ${ANSI_LIGHT_BLUE}'%s'${ANSI_NOCOLOR}\n" "${{ inputs.development_branch }}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${branch}" == "${{ inputs.release_branch }}" ]]; then
|
||||||
|
on_release_branch="true"
|
||||||
|
|
||||||
|
if [[ -z "$(git rev-list -1 --merges ${{ github.sha }}~1..${{ github.sha }})" ]]; then
|
||||||
|
is_regular_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_YELLOW}regular "
|
||||||
|
else
|
||||||
|
is_release_commit="true"
|
||||||
|
printf " ${ANSI_LIGHT_GREEN}release "
|
||||||
|
fi
|
||||||
|
printf "commit${ANSI_NOCOLOR} on release branch ${ANSI_LIGHT_BLUE}'%s'${ANSI_NOCOLOR}\n" "${{ inputs.release_branch }}"
|
||||||
|
fi
|
||||||
|
elif [[ "${ref:0:10}" == "refs/tags/" ]]; then
|
||||||
|
ref_kind="tag"
|
||||||
|
tag="${ref:10}"
|
||||||
|
|
||||||
|
printf "Tag check:\n"
|
||||||
|
|
||||||
|
printf " Check if tag is on release branch '%s' ... " "${{ inputs.release_branch }}"
|
||||||
|
git branch --remotes --contains $(git rev-parse --verify "tags/${tag}~0") | grep "origin/${{ inputs.release_branch }}" > /dev/null
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Tag '%s' isn't on branch '%s'.${ANSI_NOCOLOR}\n" "${tag}" "${{ inputs.release_branch }}"
|
||||||
|
printf "::error title=TagCheck::Tag '%s' isn't on branch '%s'.\n" "${tag}" "${{ inputs.release_branch }}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
NIGHTLY_TAG_PATTERN='^${{ inputs.nightly_tag_pattern }}$'
|
||||||
|
RELEASE_TAG_PATTERN='^${{ inputs.release_tag_pattern }}$'
|
||||||
|
printf " Check tag name against regexp '%s' ... " "${RELEASE_TAG_PATTERN}"
|
||||||
|
if [[ "${tag}" =~ NIGHTLY_TAG_PATTERN ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[NIGHTLY]${ANSI_NOCOLOR}\n"
|
||||||
|
is_nightly_tag="true"
|
||||||
|
elif [[ "${tag}" =~ $RELEASE_TAG_PATTERN ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[RELEASE]${ANSI_NOCOLOR}\n"
|
||||||
|
version="${tag}"
|
||||||
|
is_release_tag="true"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Tag name '%s' doesn't conform to regexp${ANSI_NOCOLOR}\n" "${tag}"
|
||||||
|
printf " ${ANSI_LIGHT_RED}nightly tag: %s${ANSI_NOCOLOR}\n" "${NIGHTLY_TAG_PATTERN}"
|
||||||
|
printf " ${ANSI_LIGHT_RED}release tag: %s${ANSI_NOCOLOR}\n" "${RELEASE_TAG_PATTERN}"
|
||||||
|
printf "::error title=RexExpCheck::Tag name '%s' doesn't conform to regexp '%s' nor '%s'.\n" "${tag}" "${NIGHTLY_TAG_PATTERN}" "${RELEASE_TAG_PATTERN}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}Unknown Git reference '%s'.${ANSI_NOCOLOR}\n" "${{ github.ref }}"
|
||||||
|
printf "::error title=Classify Commit::Unknown Git reference '%s'.\n" "${{ github.ref }}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
tee --append "${GITHUB_OUTPUT}" <<EOF
|
||||||
|
on_main_branch=${on_main_branch}
|
||||||
|
on_dev_branch=${on_dev_branch}
|
||||||
|
on_release_branch=${on_release_branch}
|
||||||
|
is_regular_commit=${is_regular_commit}
|
||||||
|
is_merge_commit=${is_merge_commit}
|
||||||
|
is_release_commit=${is_release_commit}
|
||||||
|
is_nightly_tag=${is_nightly_tag}
|
||||||
|
is_release_tag=${is_release_tag}
|
||||||
|
ref_kind=${ref_kind}
|
||||||
|
branch=${branch}
|
||||||
|
tag=${tag}
|
||||||
|
version=${version}
|
||||||
|
EOF
|
||||||
|
|
||||||
|
- name: 🔁 Find merged PullRequest from second parent of current SHA (${{ github.sha }})
|
||||||
|
id: FindPullRequest
|
||||||
|
if: steps.Classify.outputs.is_merge_commit == 'true'
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
printf "Read second parent of current SHA (%s) ... " "${{ github.ref }}"
|
||||||
|
FATHER_SHA=$(git rev-parse ${{ github.ref }}^2)
|
||||||
|
if [[ $? -ne 0 || "{FATHER_SHA}" == "" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Couldn't read second parent (father) of '%s'.${ANSI_NOCOLOR}\n" "${{ github.ref }}^2"
|
||||||
|
printf "::error title=GitCommitHistoryError::Couldn't read second parent (father) of '%s'. -> %s\n" "${{ github.ref }}^2" "${FATHER_SHA}"
|
||||||
|
exit 1
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "Search Pull Request to '%s' and branch containing SHA %s ... " "${{ inputs.release_branch }}" "${FATHER_SHA}"
|
||||||
|
PULL_REQUESTS=$(gh pr list --base "${{ inputs.release_branch }}" --search "${FATHER_SHA}" --state "merged" --json "title,number,mergedBy,mergedAt")
|
||||||
|
if [[ $? -ne 0 || "${PULL_REQUESTS}" == "" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Couldn't find a merged Pull Request to '%s'. -> %s${ANSI_NOCOLOR}\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||||
|
printf "::error title=PullRequest::Couldn't find a merged Pull Request to '%s'. -> %s\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||||
|
exit 1
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
PR_TITLE="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].title")"
|
||||||
|
PR_NUMBER="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].number")"
|
||||||
|
PR_MERGED_BY="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedBy.login")"
|
||||||
|
PR_MERGED_AT="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedAt")"
|
||||||
|
|
||||||
|
printf "${ANSI_LIGHT_BLUE}Found Pull Request:${ANSI_NOCOLOR}\n"
|
||||||
|
printf " %s\n" "Title: ${PR_TITLE}"
|
||||||
|
printf " %s\n" "Number: ${PR_NUMBER}"
|
||||||
|
printf " %s\n" "MergedBy: ${PR_MERGED_BY}"
|
||||||
|
printf " %s\n" "MergedAt: ${PR_MERGED_AT} ($(date -d"${PR_MERGED_AT}" '+%d.%m.%Y - %H:%M:%S'))"
|
||||||
|
fi
|
||||||
|
|
||||||
|
RELEASE_TAG_PATTERN='^${{ inputs.release_tag_pattern }}$'
|
||||||
|
printf "Check Pull Request title against regexp '%s' ... " "${RELEASE_TAG_PATTERN}"
|
||||||
|
if [[ "${PR_TITLE}" =~ $RELEASE_TAG_PATTERN ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
RELEASE_VERSION="${PR_TITLE}"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Pull Request title '%s' doesn't conform to regexp '%s'.${ANSI_NOCOLOR}\n" "${PR_TITLE}" "${RELEASE_TAG_PATTERN}"
|
||||||
|
printf "::error title=RexExpCheck::Pull Request title '%s' doesn't conform to regexp '%s'.\n" "${PR_TITLE}" "${RELEASE_TAG_PATTERN}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "Release tag: ${ANSI_LIGHT_GREEN}%s${ANSI_NOCOLOR}\n" "${RELEASE_VERSION}"
|
||||||
|
tee --append "${GITHUB_OUTPUT}" <<EOF
|
||||||
|
pr_version=${RELEASE_VERSION}
|
||||||
|
pr_title=${PR_TITLE}
|
||||||
|
pr_number=${PR_NUMBER}
|
||||||
|
pr_mergedby=${PR_MERGED_BY}
|
||||||
|
pr_mergedat=${PR_MERGED_AT}
|
||||||
|
EOF
|
||||||
|
|
||||||
|
- name: Debug
|
||||||
|
run: |
|
||||||
|
printf "on_main_branch: %s\n" "${{ steps.Classify.outputs.on_main_branch }}"
|
||||||
|
printf "on_dev_branch: %s\n" "${{ steps.Classify.outputs.on_dev_branch }}"
|
||||||
|
printf "on_release_branch: %s\n" "${{ steps.Classify.outputs.on_release_branch }}"
|
||||||
|
printf "is_regular_commit: %s\n" "${{ steps.Classify.outputs.is_regular_commit }}"
|
||||||
|
printf "is_merge_commit: %s\n" "${{ steps.Classify.outputs.is_merge_commit }}"
|
||||||
|
printf "is_release_commit: %s\n" "${{ steps.Classify.outputs.is_release_commit }}"
|
||||||
|
printf "is_nightly_tag: %s\n" "${{ steps.Classify.outputs.is_nightly_tag }}"
|
||||||
|
printf "is_release_tag: %s\n" "${{ steps.Classify.outputs.is_release_tag }}"
|
||||||
|
printf "ref_kind: %s\n" "${{ steps.Classify.outputs.ref_kind }}"
|
||||||
|
printf "branch: %s\n" "${{ steps.Classify.outputs.branch }}"
|
||||||
|
printf "tag: %s\n" "${{ steps.Classify.outputs.tag }}"
|
||||||
|
printf "version from tag: %s\n" "${{ steps.Classify.outputs.version }}"
|
||||||
|
printf "version from pr: %s\n" "${{ steps.FindPullRequest.outputs.pr_version }}"
|
||||||
|
printf "version: %s\n" "${{ steps.Classify.outputs.version || steps.FindPullRequest.outputs.pr_version }}"
|
||||||
|
printf "pr title: %s\n" "${{ steps.FindPullRequest.outputs.pr_title }}"
|
||||||
|
printf "pr number: %s\n" "${{ steps.FindPullRequest.outputs.pr_number }}"
|
||||||
33
.github/workflows/PublishCoverageResults.yml
vendored
33
.github/workflows/PublishCoverageResults.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -86,20 +86,20 @@ on:
|
|||||||
codecov:
|
codecov:
|
||||||
description: 'Publish merged coverage report to Codecov.'
|
description: 'Publish merged coverage report to Codecov.'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
codacy:
|
codacy:
|
||||||
description: 'Publish merged coverage report to Codacy.'
|
description: 'Publish merged coverage report to Codacy.'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
secrets:
|
secrets:
|
||||||
CODECOV_TOKEN:
|
CODECOV_TOKEN:
|
||||||
description: 'Token to push result to Codecov.'
|
description: 'Token to push result to Codecov.'
|
||||||
required: true
|
required: false
|
||||||
CODACY_TOKEN:
|
CODACY_TOKEN:
|
||||||
description: 'Token to push result to Codacy.'
|
description: 'Token to push result to Codacy.'
|
||||||
required: true
|
required: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
PublishCoverageResults:
|
PublishCoverageResults:
|
||||||
@@ -200,7 +200,8 @@ jobs:
|
|||||||
|
|
||||||
- name: 📊 Publish code coverage at CodeCov
|
- name: 📊 Publish code coverage at CodeCov
|
||||||
uses: codecov/codecov-action@v5
|
uses: codecov/codecov-action@v5
|
||||||
if: inputs.codecov
|
id: codecov
|
||||||
|
if: inputs.codecov == 'true'
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
with:
|
with:
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
@@ -212,8 +213,22 @@ jobs:
|
|||||||
|
|
||||||
- name: 📉 Publish code coverage at Codacy
|
- name: 📉 Publish code coverage at Codacy
|
||||||
uses: codacy/codacy-coverage-reporter-action@v1
|
uses: codacy/codacy-coverage-reporter-action@v1
|
||||||
if: inputs.codacy
|
id: codacy
|
||||||
|
if: inputs.codacy == 'true'
|
||||||
continue-on-error: true
|
continue-on-error: true
|
||||||
with:
|
with:
|
||||||
project-token: ${{ secrets.CODACY_TOKEN }}
|
project-token: ${{ secrets.CODACY_TOKEN }}
|
||||||
coverage-reports: ${{ inputs.coverage_report_xml_directory }}/${{ inputs.coverage_report_xml_filename }}
|
coverage-reports: ${{ inputs.coverage_report_xml_directory }}/${{ inputs.coverage_report_xml_filename }}
|
||||||
|
|
||||||
|
- name: Generate error messages
|
||||||
|
run: |
|
||||||
|
if [[ "${{ steps.codecov.outcome }}" == "failure" ]]; then
|
||||||
|
printf "::error title=%s::%s\n" "Publish Code Coverage Results / Codecov" "Failed to publish code coverage results."
|
||||||
|
else
|
||||||
|
printf "Codecov: No errors to report.\n"
|
||||||
|
fi
|
||||||
|
if [[ "${{ steps.codacy.outcome }}" == "failure" ]]; then
|
||||||
|
printf "::error title=%s::%s\n" "Publish Code Coverage Results / Codacy" "Failed to publish code coverage results."
|
||||||
|
else
|
||||||
|
printf "Codacy: No errors to report.\n"
|
||||||
|
fi
|
||||||
|
|||||||
3
.github/workflows/PublishOnPyPI.yml
vendored
3
.github/workflows/PublishOnPyPI.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -50,7 +50,6 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
|
||||||
PublishOnPyPI:
|
PublishOnPyPI:
|
||||||
name: 🚀 Publish to PyPI
|
name: 🚀 Publish to PyPI
|
||||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
||||||
|
|||||||
801
.github/workflows/PublishReleaseNotes.yml
vendored
Normal file
801
.github/workflows/PublishReleaseNotes.yml
vendored
Normal file
@@ -0,0 +1,801 @@
|
|||||||
|
# ==================================================================================================================== #
|
||||||
|
# Authors: #
|
||||||
|
# Patrick Lehmann #
|
||||||
|
# #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
|
# #
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
|
# you may not use this file except in compliance with the License. #
|
||||||
|
# You may obtain a copy of the License at #
|
||||||
|
# #
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0 #
|
||||||
|
# #
|
||||||
|
# Unless required by applicable law or agreed to in writing, software #
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, #
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. #
|
||||||
|
# See the License for the specific language governing permissions and #
|
||||||
|
# limitations under the License. #
|
||||||
|
# #
|
||||||
|
# SPDX-License-Identifier: Apache-2.0 #
|
||||||
|
# ==================================================================================================================== #
|
||||||
|
name: Release
|
||||||
|
|
||||||
|
on:
|
||||||
|
workflow_call:
|
||||||
|
inputs:
|
||||||
|
ubuntu_image:
|
||||||
|
description: 'Name of the Ubuntu image.'
|
||||||
|
required: false
|
||||||
|
default: 'ubuntu-24.04'
|
||||||
|
type: string
|
||||||
|
release_branch:
|
||||||
|
description: 'Name of the branch containing releases.'
|
||||||
|
required: false
|
||||||
|
default: 'main'
|
||||||
|
type: string
|
||||||
|
mode:
|
||||||
|
description: 'Release mode: nightly or release.'
|
||||||
|
required: false
|
||||||
|
default: 'release'
|
||||||
|
type: string
|
||||||
|
tag:
|
||||||
|
description: 'Name of the release (tag).'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
title:
|
||||||
|
description: 'Title of the release.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
description:
|
||||||
|
description: 'Multi-line description of the release.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
description_file:
|
||||||
|
description: 'Description of the release from a Markdown file.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
description_footer:
|
||||||
|
description: 'Footer line(s) in every release.'
|
||||||
|
required: false
|
||||||
|
default: |
|
||||||
|
|
||||||
|
--------
|
||||||
|
Published from [%%gh_workflow_name%%](%%gh_server%%/%%gh_owner_repo%%/actions/runs/%%gh_runid%%) workflow triggered by %%gh_actor%% on %%datetime%%.
|
||||||
|
|
||||||
|
This automatic release was created by [pyTooling/Actions](http://github.com/pyTooling/Actions)::Release.yml
|
||||||
|
type: string
|
||||||
|
draft:
|
||||||
|
description: 'Specify if this is a draft.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
prerelease:
|
||||||
|
description: 'Specify if this is a pre-release.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
latest:
|
||||||
|
description: 'Specify if this is the latest release.'
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
type: boolean
|
||||||
|
replacements:
|
||||||
|
description: 'Multi-line string containing search=replace patterns.'
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
assets:
|
||||||
|
description: 'Multi-line string containing artifact:file:title asset descriptions.'
|
||||||
|
required: false
|
||||||
|
type: string
|
||||||
|
default: ''
|
||||||
|
inventory-json:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
inventory-version:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
inventory-categories:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
tarball-name:
|
||||||
|
type: string
|
||||||
|
required: false
|
||||||
|
default: '__pyTooling_upload_artifact__.tar'
|
||||||
|
can-fail:
|
||||||
|
type: boolean
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
|
outputs:
|
||||||
|
release-page:
|
||||||
|
description: "URL to the release page."
|
||||||
|
value: ${{ jobs.Release.outputs.release-page }}
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
Release:
|
||||||
|
name: 📝 Create or Update Release Page on GitHub
|
||||||
|
runs-on: ${{ inputs.ubuntu_image }}
|
||||||
|
continue-on-error: ${{ inputs.can-fail }}
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
# attestations: write
|
||||||
|
outputs:
|
||||||
|
release-page: ${{ steps.removeDraft.outputs.release_page }}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: ⏬ Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
with:
|
||||||
|
# The command 'git describe' (used for version) needs the history.
|
||||||
|
fetch-depth: 0
|
||||||
|
|
||||||
|
- name: 🔧 Install zstd
|
||||||
|
run: sudo apt-get install -y --no-install-recommends zstd
|
||||||
|
|
||||||
|
- name: 📑 Prepare
|
||||||
|
id: prepare
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
printf "Release mode: ${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "${{ inputs.mode }}"
|
||||||
|
case "${{ inputs.mode }}" in
|
||||||
|
"release")
|
||||||
|
;;
|
||||||
|
"nightly")
|
||||||
|
printf "→ Allow deletion and recreation of existing release pages for rolling releases (nightly releases)\n"
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
printf "Unknown mode '%s'\n" "${{ inputs.mode }}"
|
||||||
|
printf "::error title=%s::%s\n" "InternalError" "Unknown mode '${{ inputs.mode }}'."
|
||||||
|
exit 1
|
||||||
|
esac
|
||||||
|
|
||||||
|
- name: 📑 Delete (old) Release Page
|
||||||
|
id: deleteReleasePage
|
||||||
|
if: inputs.mode == 'nightly'
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
printf "Deleting release '%s' ... " "${{ inputs.tag }}"
|
||||||
|
message="$(gh release delete ${{ inputs.tag }} --yes 2>&1)"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
elif [[ "${message}" == "release not found" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_YELLOW}[NOT FOUND]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't delete release '%s' -> Error: '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.tag }}" "${message}"
|
||||||
|
printf "::error title=%s::%s\n" "InternalError" "Couldn't delete release '${{ inputs.tag }}' -> Error: '${message}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📑 Assemble Release Notes
|
||||||
|
id: createReleaseNotes
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
# Save release description (from parameter in a file)
|
||||||
|
head -c -1 <<'EOF' > __DESCRIPTION__.md
|
||||||
|
${{ inputs.description }}
|
||||||
|
EOF
|
||||||
|
|
||||||
|
# Save release footer (from parameter in a file)
|
||||||
|
head -c -1 <<'EOF' > __FOOTER__.md
|
||||||
|
${{ inputs.description_footer }}
|
||||||
|
EOF
|
||||||
|
|
||||||
|
# Download Markdown from PullRequest
|
||||||
|
# Readout second parent's SHA
|
||||||
|
# Search PR with that SHA
|
||||||
|
# Load description of that PR
|
||||||
|
printf "Read second parent of current SHA (%s) ... " "${{ github.ref }}"
|
||||||
|
FATHER_SHA=$(git rev-parse ${{ github.ref }}^2 -- 2> /dev/null)
|
||||||
|
if [[ $? -ne 0 || "{FATHER_SHA}" == "" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "→ ${ANSI_LIGHT_YELLOW}Skipped readout of pull request description. This is not a merge commit.${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
printf "Search Pull Request to '%s' and branch containing SHA %s ... " "${{ inputs.release_branch }}" "${FATHER_SHA}"
|
||||||
|
PULL_REQUESTS=$(gh pr list --base "${{ inputs.release_branch }}" --search "${FATHER_SHA}" --state "merged" --json "title,number,mergedBy,mergedAt,body")
|
||||||
|
if [[ $? -ne 0 || "${PULL_REQUESTS}" == "" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "${ANSI_LIGHT_RED}Couldn't find a merged Pull Request to '%s'. -> %s${ANSI_NOCOLOR}\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||||
|
printf "::error title=PullRequest::Couldn't find a merged Pull Request to '%s'. -> %s\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||||
|
exit 1
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
PR_TITLE="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].title")"
|
||||||
|
PR_NUMBER="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].number")"
|
||||||
|
PR_BODY="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].body")"
|
||||||
|
PR_MERGED_BY="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedBy.login")"
|
||||||
|
PR_MERGED_AT="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedAt")"
|
||||||
|
|
||||||
|
printf "Found Pull Request:\n"
|
||||||
|
printf " %s\n" "Title: ${PR_TITLE}"
|
||||||
|
printf " %s\n" "Number: ${PR_NUMBER}"
|
||||||
|
printf " %s\n" "MergedBy: ${PR_MERGED_BY}"
|
||||||
|
printf " %s\n" "MergedAt: ${PR_MERGED_AT} ($(date -d"${PR_MERGED_AT}" '+%d.%m.%Y - %H:%M:%S'))"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "${PR_BODY}" > __PULLREQUEST__.md
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check if a release description file should be used and exists.
|
||||||
|
if [[ "${{ inputs.description_file }}" != "" ]]; then
|
||||||
|
if [[ ! -f "${{ inputs.description_file }}" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}Release description file '%s' not found.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||||
|
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' not found."
|
||||||
|
exit 1
|
||||||
|
elif [[ -s "${{ inputs.description_file }}" ]]; then
|
||||||
|
printf "Use '%s' as main release description.\n" "${{ inputs.description_file }}"
|
||||||
|
cp -v "${{ inputs.description_file }}" __NOTES__.md
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}Release description file '%s' is empty.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||||
|
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' is empty."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
# Check if the main release description is provided by a template parameter
|
||||||
|
elif [[ -s __DESCRIPTION__.md ]]; then
|
||||||
|
printf "Use '__DESCRIPTION__.md' as main release description.\n"
|
||||||
|
mv -v __DESCRIPTION__.md __NOTES__.md
|
||||||
|
# Check if the pull request serves as the main release description text.
|
||||||
|
elif [[ -s __PULLREQUEST__.md ]]; then
|
||||||
|
printf "Use '__PULLREQUEST__.md' as main release description.\n"
|
||||||
|
mv -v __PULLREQUEST__.md __NOTES__.md
|
||||||
|
|
||||||
|
printf "Append '%%%%FOOTER%%%%' to '__NOTES__.md'.\n"
|
||||||
|
printf "\n%%%%FOOTER%%%%\n" >> __NOTES__.md
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}No release description specified (file, parameter, PR text).${ANSI_NOCOLOR}\n"
|
||||||
|
printf "::error title=%s::%s\n" "MissingDescription" "No release description specified (file, parameter, PR text)."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Read release notes main file for placeholder substitution
|
||||||
|
NOTES=$(<__NOTES__.md)
|
||||||
|
|
||||||
|
# Inline description
|
||||||
|
if [[ -s __DESCRIPTION__.md ]]; then
|
||||||
|
NOTES="${NOTES//%%DESCRIPTION%%/$(<__DESCRIPTION__.md)}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Inline PullRequest and increase headline levels
|
||||||
|
if [[ -s __PULLREQUEST__.md ]]; then
|
||||||
|
while [[ "${NOTES}" =~ %%(PULLREQUEST(\+[0-3])?)%% ]]; do
|
||||||
|
case "${BASH_REMATCH[1]}" in
|
||||||
|
"PULLREQUEST+0" | "PULLREQUEST")
|
||||||
|
NOTES="${NOTES//${BASH_REMATCH[0]}/$(<__PULLREQUEST__.md)}"
|
||||||
|
;;
|
||||||
|
"PULLREQUEST+1")
|
||||||
|
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1# /gm;t')}"
|
||||||
|
;;
|
||||||
|
"PULLREQUEST+2")
|
||||||
|
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||||
|
;;
|
||||||
|
"PULLREQUEST+3")
|
||||||
|
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
# inline Footer
|
||||||
|
if [[ -s __FOOTER__.md ]]; then
|
||||||
|
NOTES="${NOTES//%%FOOTER%%/$(<__FOOTER__.md)}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Apply replacements
|
||||||
|
while IFS=$'\r\n' read -r patternLine; do
|
||||||
|
# skip empty lines
|
||||||
|
[[ "$patternLine" == "" ]] && continue
|
||||||
|
|
||||||
|
pattern="%${patternLine%%=*}%"
|
||||||
|
replacement="${patternLine#*=}"
|
||||||
|
NOTES="${NOTES//$pattern/$replacement}"
|
||||||
|
done <<<'${{ inputs.replacements }}'
|
||||||
|
|
||||||
|
# Workarounds for stupid GitHub variables
|
||||||
|
owner_repo="${{ github.repository }}"
|
||||||
|
repo=${owner_repo##*/}
|
||||||
|
|
||||||
|
# Replace special identifiers
|
||||||
|
NOTES="${NOTES//%%gh_server%%/${{ github.server_url }}}"
|
||||||
|
NOTES="${NOTES//%%gh_workflow_name%%/${{ github.workflow }}}"
|
||||||
|
NOTES="${NOTES//%%gh_owner%%/${{ github.repository_owner }}}"
|
||||||
|
NOTES="${NOTES//%%gh_repo%%/${repo}}"
|
||||||
|
NOTES="${NOTES//%%gh_owner_repo%%/${{ github.repository_owner }}}"
|
||||||
|
#NOTES="${NOTES//%%gh_pages%%/https://${{ github.repository_owner }}.github.io/${repo}/}"
|
||||||
|
NOTES="${NOTES//%%gh_runid%%/${{ github.run_id }}}"
|
||||||
|
NOTES="${NOTES//%%gh_actor%%/${{ github.actor }}}"
|
||||||
|
NOTES="${NOTES//%%gh_sha%%/${{ github.sha }}}"
|
||||||
|
NOTES="${NOTES//%%date%%/$(date '+%Y-%m-%d')}"
|
||||||
|
NOTES="${NOTES//%%time%%/$(date '+%H:%M:%S %Z')}"
|
||||||
|
NOTES="${NOTES//%%datetime%%/$(date '+%Y-%m-%d %H:%M:%S %Z')}"
|
||||||
|
|
||||||
|
# Write final release notes to file
|
||||||
|
echo "${NOTES}" > __NOTES__.md
|
||||||
|
|
||||||
|
# Display partial contents for debugging
|
||||||
|
if [[ -s __DESCRIPTION__.md ]]; then
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__DESCRIPTION__.md' ($(stat --printf="%s" "__DESCRIPTION__.md") B) ...."
|
||||||
|
cat __DESCRIPTION__.md
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_YELLOW}No '__DESCRIPTION__.md' found.${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
if [[ -s __PULLREQUEST__.md ]]; then
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__PULLREQUEST__.md' ($(stat --printf="%s" "__PULLREQUEST__.md") B) ...."
|
||||||
|
cat __PULLREQUEST__.md
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_YELLOW}No '__PULLREQUEST__.md' found.${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
if [[ -s __FOOTER__.md ]]; then
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__FOOTER__.md' ($(stat --printf="%s" "__FOOTER__.md") B) ...."
|
||||||
|
cat __FOOTER__.md
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_YELLOW}No '__FOOTER__.md' found.${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Print final release notes
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__NOTES__.md' ($(stat --printf="%s" "__NOTES__.md") B) ...."
|
||||||
|
cat __NOTES__.md
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
|
||||||
|
- name: 📑 Create new Release Page
|
||||||
|
id: createReleasePage
|
||||||
|
if: inputs.mode == 'release'
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
if [[ "${{ inputs.prerelease }}" == "true" ]]; then
|
||||||
|
addPreRelease="--prerelease"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.latest }}" == "false" ]]; then
|
||||||
|
addLatest="--latest=false"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.title }}" != "" ]]; then
|
||||||
|
addTitle=("--title" "${{ inputs.title }}")
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ -s __NOTES__.md ]]; then
|
||||||
|
addNotes=("--notes-file" "__NOTES__.md")
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "Creating release '%s' ... " "${{ inputs.tag }}"
|
||||||
|
message="$(gh release create "${{ inputs.tag }}" --verify-tag --draft $addPreRelease $addLatest "${addTitle[@]}" "${addNotes[@]}" 2>&1)"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " Release page: %s\n" "${message}"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't create release '%s' -> Error: '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.tag }}" "${message}"
|
||||||
|
printf "::error title=%s::%s\n" "InternalError" "Couldn't create release '${{ inputs.tag }}' -> Error: '${message}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📑 Recreate Release Page
|
||||||
|
id: recreateReleasePage
|
||||||
|
if: inputs.mode == 'nightly'
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
addDraft="--draft"
|
||||||
|
if [[ "${{ inputs.prerelease }}" == "true" ]]; then
|
||||||
|
addPreRelease="--prerelease"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.latest }}" == "false" ]]; then
|
||||||
|
addLatest="--latest=false"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${{ inputs.title }}" != "" ]]; then
|
||||||
|
addTitle=("--title" "${{ inputs.title }}")
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ -s __NOTES__.md ]]; then
|
||||||
|
addNotes=("--notes-file" "__NOTES__.md")
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "Creating release '%s' ... " "${{ inputs.tag }}"
|
||||||
|
message="$(gh release create "${{ inputs.tag }}" --verify-tag --draft $addPreRelease $addLatest "${addTitle[@]}" "${addNotes[@]}" 2>&1)"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " Release page: %s\n" "${message}"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't recreate release '%s' -> Error: '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.tag }}" "${message}"
|
||||||
|
printf "::error title=%s::%s\n" "InternalError" "Couldn't recreate release '${{ inputs.tag }}' -> Error: '${message}'."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📥 Download artifacts and upload as assets
|
||||||
|
id: uploadAssets
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||||
|
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
Replace() {
|
||||||
|
line="$1"
|
||||||
|
while IFS=$'\r\n' read -r patternLine; do
|
||||||
|
# skip empty lines
|
||||||
|
[[ "$patternLine" == "" ]] && continue
|
||||||
|
|
||||||
|
pattern="${patternLine%%=*}"
|
||||||
|
replacement="${patternLine#*=}"
|
||||||
|
line="${line//"%$pattern%"/"$replacement"}"
|
||||||
|
done <<<'${{ inputs.replacements }}'
|
||||||
|
printf "%s\n" "$line"
|
||||||
|
}
|
||||||
|
|
||||||
|
# Create JSON inventory
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
VERSION="1.0"
|
||||||
|
|
||||||
|
# Split categories by ',' into a Bash array.
|
||||||
|
# See https://stackoverflow.com/a/45201229/3719459
|
||||||
|
if [[ "${{ inputs.inventory-categories }}" != "" ]]; then
|
||||||
|
readarray -td, inventoryCategories <<<"${{ inputs.inventory-categories }},"
|
||||||
|
unset 'inventoryCategories[-1]'
|
||||||
|
declare -p inventoryCategories
|
||||||
|
else
|
||||||
|
inventoryCategories=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
jsonInventory=$(jq -c -n \
|
||||||
|
--arg version "${VERSION}" \
|
||||||
|
--arg date "$(date +"%Y-%m-%dT%H-%M-%S%:z")" \
|
||||||
|
--argjson jsonMeta "$(jq -c -n \
|
||||||
|
--arg tag "${{ inputs.tag }}" \
|
||||||
|
--arg version "${{ inputs.inventory-version }}" \
|
||||||
|
--arg hash "${{ github.sha }}" \
|
||||||
|
--arg repo "${{ github.server_url }}/${{ github.repository }}" \
|
||||||
|
--arg release "${{ github.server_url }}/${{ github.repository }}/releases/download/${{ inputs.tag }}" \
|
||||||
|
--argjson categories "$(jq -c -n \
|
||||||
|
'$ARGS.positional' \
|
||||||
|
--args "${inventoryCategories[@]}" \
|
||||||
|
)" \
|
||||||
|
'{"tag": $tag, "version": $version, "git-hash": $hash, "repository-url": $repo, "release-url": $release, "categories": $categories}' \
|
||||||
|
)" \
|
||||||
|
'{"version": 1.0, "timestamp": $date, "meta": $jsonMeta, "files": {}}'
|
||||||
|
)
|
||||||
|
fi
|
||||||
|
|
||||||
|
ERRORS=0
|
||||||
|
# A dictionary of 0/1 to avoid duplicate downloads
|
||||||
|
declare -A downloadedArtifacts
|
||||||
|
# A dictionary to check for duplicate asset files in release
|
||||||
|
declare -A assetFilenames
|
||||||
|
while IFS=$'\r\n' read -r assetLine; do
|
||||||
|
if [[ "${assetLine}" == "" || "${assetLine:0:1}" == "#" ]]; then
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
# split assetLine colon separated triple: artifact:asset:title
|
||||||
|
artifact="${assetLine%%:*}"
|
||||||
|
assetLine="${assetLine#*:}"
|
||||||
|
asset="${assetLine%%:*}"
|
||||||
|
assetLine="${assetLine#*:}"
|
||||||
|
if [[ "${{ inputs.inventory-json }}" == "" ]]; then
|
||||||
|
categories=""
|
||||||
|
title="${assetLine##*:}"
|
||||||
|
else
|
||||||
|
categories="${assetLine%%:*}"
|
||||||
|
title="${assetLine##*:}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# remove leading whitespace
|
||||||
|
asset="${asset#"${asset%%[![:space:]]*}"}"
|
||||||
|
categories="${categories#"${categories%%[![:space:]]*}"}"
|
||||||
|
title="${title#"${title%%[![:space:]]*}"}"
|
||||||
|
|
||||||
|
# apply replacements
|
||||||
|
asset="$(Replace "${asset}")"
|
||||||
|
title="$(Replace "${title}")"
|
||||||
|
|
||||||
|
printf "Publish asset '%s' from artifact '%s' with title '%s'\n" "${asset}" "${artifact}" "${title}"
|
||||||
|
printf " Checked asset for duplicates ... "
|
||||||
|
if [[ -n "${assetFilenames[$asset]}" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf "::error title=%s::%s\n" "DuplicateAsset" "Asset '${asset}' from artifact '${artifact}' was already uploaded to release '${{ inputs.tag }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
assetFilenames[$asset]=1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Download artifact by artifact name
|
||||||
|
if [[ -n "${downloadedArtifacts[$artifact]}" ]]; then
|
||||||
|
printf " downloading '%s' ... ${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}\n" "${artifact}"
|
||||||
|
else
|
||||||
|
echo " downloading '${artifact}' ... "
|
||||||
|
printf " gh run download $GITHUB_RUN_ID --dir \"%s\" --name \"%s\" " "${artifact}" "${artifact}"
|
||||||
|
gh run download $GITHUB_RUN_ID --dir "${artifact}" --name "${artifact}"
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't download artifact '%s'.${ANSI_NOCOLOR}\n" "${artifact}"
|
||||||
|
printf "::error title=%s::%s\n" "ArtifactNotFound" "Couldn't download artifact '${artifact}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
downloadedArtifacts[$artifact]=1
|
||||||
|
|
||||||
|
printf " Checking for embedded tarball ... "
|
||||||
|
if [[ -f "${artifact}/${{ inputs.tarball-name }}" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[FOUND]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
pushd "${artifact}" > /dev/null
|
||||||
|
|
||||||
|
printf " Extracting embedded tarball ... "
|
||||||
|
tar -xf "${{ inputs.tarball-name }}"
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf " Removing temporary tarball ... "
|
||||||
|
rm -f "${{ inputs.tarball-name }}"
|
||||||
|
if [[ $? -ne 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
|
||||||
|
popd > /dev/null
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}\n"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check if artifact should be compressed (zip, tgz) or if asset was part of the downloaded artifact.
|
||||||
|
printf " checking asset '%s' ... " "${artifact}/${asset}"
|
||||||
|
if [[ "${asset}" == !*.zip ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[ZIP]${ANSI_NOCOLOR}\n"
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf "::group:: %s\n" "Compressing artifact '${artifact}' to '${asset}' ..."
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
zip -r "../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't compress '%s' to zip file '%s'.${ANSI_NOCOLOR}\n" "${artifact}" "${asset}"
|
||||||
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to zip file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ "${asset}" == !*.tgz || "${asset}" == !*.tar.gz || "${asset}" == \$*.tgz || "${asset}" == \$*.tar.gz ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[TAR/GZ]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
if [[ "${asset:0:1}" == "\$" ]]; then
|
||||||
|
asset="${asset##*$}"
|
||||||
|
dirName="${asset%.*}"
|
||||||
|
printf " Compressing artifact '%s' to '%s' ...\n" "${artifact}" "${asset}"
|
||||||
|
tar -c --gzip --owner=0 --group=0 --file="${asset}" --directory="${artifact}" --transform "s|^\.|${dirName%.tar}|" .
|
||||||
|
retCode=$?
|
||||||
|
else
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf " Compressing artifact '%s' to '%s' ...\n" "${artifact}" "${asset}"
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
tar -c --gzip --owner=0 --group=0 --file="../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't compress '%s' to tgz file '%s'.${ANSI_NOCOLOR}\n" "${artifact}" "${asset}"
|
||||||
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to tgz file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ "${asset}" == !*.tzst || "${asset}" == !*.tar.zst || "${asset}" == \$*.tzst || "${asset}" == \$*.tar.zst ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[ZST]${ANSI_NOCOLOR}\n"
|
||||||
|
|
||||||
|
if [[ "${asset:0:1}" == "\$" ]]; then
|
||||||
|
asset="${asset##*$}"
|
||||||
|
dirName="${asset%.*}"
|
||||||
|
printf " Compressing artifact '%s' to '%s' ...\n" "${artifact}" "${asset}"
|
||||||
|
tar -c --zstd --owner=0 --group=0 --file="${asset}" --directory="${artifact}" --transform "s|^\.|${dirName%.tar}|" .
|
||||||
|
retCode=$?
|
||||||
|
else
|
||||||
|
asset="${asset##*!}"
|
||||||
|
printf " Compressing artifact '%s' to '%s' ...\n" "${artifact}" "${asset}"
|
||||||
|
(
|
||||||
|
cd "${artifact}" && \
|
||||||
|
tar -c --zstd --owner=0 --group=0 --file="../${asset}" *
|
||||||
|
)
|
||||||
|
retCode=$?
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ $retCode -eq 0 ]]; then
|
||||||
|
printf " Compression ${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
uploadFile="${asset}"
|
||||||
|
else
|
||||||
|
printf " Compression ${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't compress '%s' to zst file '%s'.${ANSI_NOCOLOR}\n" "${artifact}" "${asset}"
|
||||||
|
printf "::error title=%s::%s\n" "CompressionError" "Couldn't compress '${artifact}' to zst file '${asset}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
elif [[ -e "${artifact}/${asset}" ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
uploadFile="${artifact}/${asset}"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't find asset '%s' in artifact '%s'.${ANSI_NOCOLOR}\n" "${asset}" "${artifact}"
|
||||||
|
printf "::error title=%s::%s\n" "FileNotFound" "Couldn't find asset '${asset}' in artifact '${artifact}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Add asset to JSON inventory
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
if [[ "${categories}" != "${title}" ]]; then
|
||||||
|
printf " adding file '%s' with '%s' to JSON inventory ...\n" "${uploadFile#*/}" "${categories//;/ → }"
|
||||||
|
category=""
|
||||||
|
jsonEntry=$(jq -c -n \
|
||||||
|
--arg title "${title}" \
|
||||||
|
--arg file "${uploadFile#*/}" \
|
||||||
|
'{"file": $file, "title": $title}' \
|
||||||
|
)
|
||||||
|
|
||||||
|
while [[ "${categories}" != "${category}" ]]; do
|
||||||
|
category="${categories##*,}"
|
||||||
|
categories="${categories%,*}"
|
||||||
|
jsonEntry=$(jq -c -n --arg cat "${category}" --argjson value "${jsonEntry}" '{$cat: $value}')
|
||||||
|
done
|
||||||
|
|
||||||
|
jsonInventory=$(jq -c -n \
|
||||||
|
--argjson inventory "${jsonInventory}" \
|
||||||
|
--argjson file "${jsonEntry}" \
|
||||||
|
'$inventory * {"files": $file}' \
|
||||||
|
)
|
||||||
|
else
|
||||||
|
printf " adding file '%s' to JSON inventory ... ${ANSI_LIGHT_YELLOW}[SKIPPED]${ANSI_NOCOLOR}\n" "${uploadFile#*/}"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Upload asset to existing release page
|
||||||
|
printf " uploading asset '%s' from '%s' with title '%s' ... " "${asset}" "${uploadFile}" "${title}"
|
||||||
|
gh release upload ${{ inputs.tag }} "${uploadFile}#${title}" --clobber
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't upload asset '%s' from '%s' to release '%s'.${ANSI_NOCOLOR}\n" "${asset}" "${uploadFile}" "${{ inputs.tag }}"
|
||||||
|
printf "::error title=%s::%s\n" "UploadError" "Couldn't upload asset '${asset}' from '${uploadFile}' to release '${{ inputs.tag }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
done <<<'${{ inputs.assets }}'
|
||||||
|
|
||||||
|
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||||
|
inventoryTitle="Release Inventory (JSON)"
|
||||||
|
|
||||||
|
printf "Publish asset '%s' with title '%s'\n" "${{ inputs.inventory-json }}" "${inventoryTitle}"
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Writing JSON inventory to '${{ inputs.inventory-json }}' ...."
|
||||||
|
printf "%s\n" "$(jq -n --argjson inventory "${jsonInventory}" '$inventory')" > "${{ inputs.inventory-json }}"
|
||||||
|
cat "${{ inputs.inventory-json }}"
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
|
||||||
|
# Upload inventory asset to existing release page
|
||||||
|
printf " uploading asset '%s' title '%s' ... " "${{ inputs.inventory-json }}" "${inventoryTitle}"
|
||||||
|
gh release upload ${{ inputs.tag }} "${{ inputs.inventory-json }}#${inventoryTitle}" --clobber
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't upload asset '%s' to release '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.inventory-json }}" "${{ inputs.tag }}"
|
||||||
|
printf "::error title=%s::%s\n" "UploadError" "Couldn't upload asset '${{ inputs.inventory-json }}' to release '${{ inputs.tag }}'."
|
||||||
|
ERRORS=$((ERRORS + 1))
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Inspecting downloaded artifacts ..."
|
||||||
|
tree -pash -L 3 .
|
||||||
|
printf "::endgroup::\n"
|
||||||
|
|
||||||
|
if [[ $ERRORS -ne 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_RED}%s errors detected in previous steps.${ANSI_NOCOLOR}\n" "${ERRORS}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: 📑 Remove draft state from Release Page
|
||||||
|
id: removeDraft
|
||||||
|
if: ${{ ! inputs.draft }}
|
||||||
|
run: |
|
||||||
|
set +e
|
||||||
|
|
||||||
|
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||||
|
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||||
|
ANSI_NOCOLOR=$'\x1b[0m'
|
||||||
|
|
||||||
|
export GH_TOKEN=${{ github.token }}
|
||||||
|
|
||||||
|
# Remove draft-state from release page
|
||||||
|
printf "Remove draft-state from release '%s' ... " "${title}"
|
||||||
|
releasePage=$(gh release edit --draft=false "${{ inputs.tag }}")
|
||||||
|
if [[ $? -eq 0 ]]; then
|
||||||
|
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " Release page: %s\n" "${releasePage}"
|
||||||
|
|
||||||
|
printf "release_page=%s\n" "${releasePage}" >> "${GITHUB_OUTPUT}"
|
||||||
|
else
|
||||||
|
printf "${ANSI_LIGHT_RED}[ERROR]${ANSI_NOCOLOR}\n"
|
||||||
|
printf " ${ANSI_LIGHT_RED}Couldn't remove draft-state from release '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.tag }}"
|
||||||
|
printf "::error title=%s::%s\n" "ReleasePage" "Couldn't remove draft-state from release '${{ inputs.tag }}'."
|
||||||
|
fi
|
||||||
36
.github/workflows/PublishTestResults.yml
vendored
36
.github/workflows/PublishTestResults.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -57,8 +57,8 @@ on:
|
|||||||
publish:
|
publish:
|
||||||
description: 'Publish test report summary via Dorny Test-Reporter'
|
description: 'Publish test report summary via Dorny Test-Reporter'
|
||||||
required: false
|
required: false
|
||||||
default: true
|
default: 'true'
|
||||||
type: boolean
|
type: string
|
||||||
report_title:
|
report_title:
|
||||||
description: 'Title of the summary report in the pipeline''s sidebar'
|
description: 'Title of the summary report in the pipeline''s sidebar'
|
||||||
required: false
|
required: false
|
||||||
@@ -67,17 +67,22 @@ on:
|
|||||||
dorny:
|
dorny:
|
||||||
description: 'Publish merged unittest results via Dorny Test-Reporter.'
|
description: 'Publish merged unittest results via Dorny Test-Reporter.'
|
||||||
required: false
|
required: false
|
||||||
default: true
|
default: 'true'
|
||||||
type: boolean
|
type: string
|
||||||
codecov:
|
codecov:
|
||||||
description: 'Publish merged unittest results to Codecov.'
|
description: 'Publish merged unittest results to Codecov.'
|
||||||
required: false
|
required: false
|
||||||
default: true
|
default: 'false'
|
||||||
type: boolean
|
type: string
|
||||||
|
codecov_flags:
|
||||||
|
description: 'Flags applied to the upload to Codecov'
|
||||||
|
required: false
|
||||||
|
default: 'unittest'
|
||||||
|
type: string
|
||||||
secrets:
|
secrets:
|
||||||
CODECOV_TOKEN:
|
CODECOV_TOKEN:
|
||||||
description: 'Token to push result to Codecov.'
|
description: 'Token to push result to Codecov.'
|
||||||
required: true
|
required: false
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
PublishTestResults:
|
PublishTestResults:
|
||||||
@@ -117,7 +122,7 @@ jobs:
|
|||||||
|
|
||||||
- name: 📊 Publish Unit Test Results
|
- name: 📊 Publish Unit Test Results
|
||||||
uses: dorny/test-reporter@v2
|
uses: dorny/test-reporter@v2
|
||||||
if: (inputs.dorny || inputs.publish) && inputs.report_title != ''
|
if: ( inputs.dorny == 'true' || inputs.publish == 'true' ) && inputs.report_title != ''
|
||||||
with:
|
with:
|
||||||
name: ${{ inputs.report_title }}
|
name: ${{ inputs.report_title }}
|
||||||
path: ${{ inputs.merged_junit_filename }}
|
path: ${{ inputs.merged_junit_filename }}
|
||||||
@@ -125,13 +130,24 @@ jobs:
|
|||||||
|
|
||||||
- name: 📊 Publish unittest results at CodeCov
|
- name: 📊 Publish unittest results at CodeCov
|
||||||
uses: codecov/test-results-action@v1
|
uses: codecov/test-results-action@v1
|
||||||
if: inputs.codecov
|
id: codecov
|
||||||
|
if: inputs.codecov == 'true'
|
||||||
|
continue-on-error: true
|
||||||
with:
|
with:
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
disable_search: true
|
disable_search: true
|
||||||
files: ${{ inputs.merged_junit_filename }}
|
files: ${{ inputs.merged_junit_filename }}
|
||||||
|
flags: ${{ inputs.codecov_flags }}
|
||||||
fail_ci_if_error: true
|
fail_ci_if_error: true
|
||||||
|
|
||||||
|
- name: Generate error messages
|
||||||
|
run: |
|
||||||
|
if [[ "${{ steps.codecov.outcome }}" == "failure" ]]; then
|
||||||
|
printf "::error title=%s::%s\n" "Publish Unit Test Results / Codecov" "Failed to publish unittest results."
|
||||||
|
else
|
||||||
|
printf "Codecov: No errors to report.\n"
|
||||||
|
fi
|
||||||
|
|
||||||
- name: 📤 Upload merged 'JUnit Test Summary' artifact
|
- name: 📤 Upload merged 'JUnit Test Summary' artifact
|
||||||
uses: pyTooling/upload-artifact@v4
|
uses: pyTooling/upload-artifact@v4
|
||||||
if: inputs.merged_junit_artifact != ''
|
if: inputs.merged_junit_artifact != ''
|
||||||
|
|||||||
2
.github/workflows/PublishToGitHubPages.yml
vendored
2
.github/workflows/PublishToGitHubPages.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
2
.github/workflows/SphinxDocumentation.yml
vendored
2
.github/workflows/SphinxDocumentation.yml
vendored
@@ -3,7 +3,7 @@
|
|||||||
# Patrick Lehmann #
|
# Patrick Lehmann #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
2
.github/workflows/StaticTypeCheck.yml
vendored
2
.github/workflows/StaticTypeCheck.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -20,74 +20,65 @@
|
|||||||
# #
|
# #
|
||||||
# SPDX-License-Identifier: Apache-2.0 #
|
# SPDX-License-Identifier: Apache-2.0 #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
name: Release
|
name: Auto Tag
|
||||||
|
|
||||||
on:
|
on:
|
||||||
workflow_call:
|
workflow_call:
|
||||||
inputs:
|
inputs:
|
||||||
ubuntu_image_version:
|
ubuntu_image:
|
||||||
description: 'Ubuntu image version.'
|
description: 'Name of the Ubuntu image.'
|
||||||
required: false
|
required: false
|
||||||
default: '24.04'
|
default: 'ubuntu-24.04'
|
||||||
|
type: string
|
||||||
|
version:
|
||||||
|
description: 'Version used as tag name.'
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
auto_tag:
|
||||||
|
description: 'Automatically add and push a tag.'
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
workflow:
|
||||||
|
description: 'Workflow to start after adding a tag.'
|
||||||
|
required: false
|
||||||
|
default: 'Pipeline.yml'
|
||||||
type: string
|
type: string
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
Release:
|
AutoTag:
|
||||||
name: 📝 Create 'Release Page' on GitHub
|
name: "🏷 Create tag '${{ inputs.version}}' on GitHub"
|
||||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
runs-on: ${{ inputs.ubuntu_image }}
|
||||||
|
if: inputs.auto_tag == 'true'
|
||||||
|
|
||||||
|
# if: github.ref == 'refs/heads/${{ inputs.release_branch }}'
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: write # required for tag creation
|
||||||
|
actions: write # required to start a new pipeline
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: 🔁 Extract Git tag from GITHUB_REF
|
- name: 🏷 Create release tag '${{ steps.FindPullRequest.outputs.version }}'
|
||||||
id: getVariables
|
uses: actions/github-script@v7
|
||||||
run: |
|
id: createReleaseTag
|
||||||
GIT_TAG=${GITHUB_REF#refs/*/}
|
# if: inputs.auto_tag == 'true'
|
||||||
RELEASE_VERSION=${GIT_TAG#v}
|
|
||||||
RELEASE_DATETIME="$(date --utc '+%d.%m.%Y - %H:%M:%S')"
|
|
||||||
# write to step outputs
|
|
||||||
printf "%s\n" "gitTag=${GIT_TAG}" >> $GITHUB_OUTPUT
|
|
||||||
printf "%s\n" "version=${RELEASE_VERSION}" >> $GITHUB_OUTPUT
|
|
||||||
printf "%s\n" "datetime=${RELEASE_DATETIME}" >> $GITHUB_OUTPUT
|
|
||||||
|
|
||||||
- name: 📑 Create Release Page
|
|
||||||
uses: actions/create-release@v1
|
|
||||||
id: createReleasePage
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ github.token }}
|
|
||||||
with:
|
with:
|
||||||
tag_name: ${{ steps.getVariables.outputs.gitTag }}
|
script: |
|
||||||
# release_name: ${{ steps.getVariables.outputs.gitTag }}
|
github.rest.git.createRef({
|
||||||
body: |
|
owner: context.repo.owner,
|
||||||
**Automated Release created on: ${{ steps.getVariables.outputs.datetime }}**
|
repo: context.repo.repo,
|
||||||
|
ref: 'refs/tags/${{ inputs.version }}',
|
||||||
|
sha: context.sha
|
||||||
|
})
|
||||||
|
|
||||||
# New Features
|
- name: Trigger Workflow
|
||||||
|
uses: actions/github-script@v7
|
||||||
* tbd
|
id: runReleaseTag
|
||||||
* tbd
|
# if: inputs.auto_tag == 'true'
|
||||||
|
with:
|
||||||
# Changes
|
script: |
|
||||||
|
github.rest.actions.createWorkflowDispatch({
|
||||||
* tbd
|
owner: context.repo.owner,
|
||||||
* tbd
|
repo: context.repo.repo,
|
||||||
|
workflow_id: '${{ inputs.workflow }}',
|
||||||
# Bug Fixes
|
ref: '${{ inputs.version }}'
|
||||||
|
})
|
||||||
* tbd
|
|
||||||
* tbd
|
|
||||||
|
|
||||||
# Documentation
|
|
||||||
|
|
||||||
* tbd
|
|
||||||
* tbd
|
|
||||||
|
|
||||||
# Unit Tests
|
|
||||||
|
|
||||||
* tbd
|
|
||||||
* tbd
|
|
||||||
|
|
||||||
----------
|
|
||||||
# Related Issues and Pull-Requests
|
|
||||||
|
|
||||||
* tbd
|
|
||||||
* tbd
|
|
||||||
draft: true
|
|
||||||
prerelease: false
|
|
||||||
8
.github/workflows/UnitTesting.yml
vendored
8
.github/workflows/UnitTesting.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
@@ -225,14 +225,14 @@ jobs:
|
|||||||
|
|
||||||
packages = {
|
packages = {
|
||||||
"coverage": "python-coverage:p",
|
"coverage": "python-coverage:p",
|
||||||
"docstr_coverage": "python-pyaml:p",
|
"docstr_coverage": "python-pyaml:p python-types-pyyaml:p",
|
||||||
"igraph": "igraph:p",
|
"igraph": "igraph:p",
|
||||||
"jinja2": "python-markupsafe:p",
|
"jinja2": "python-markupsafe:p",
|
||||||
"lxml": "python-lxml:p",
|
"lxml": "python-lxml:p",
|
||||||
"numpy": "python-numpy:p",
|
"numpy": "python-numpy:p",
|
||||||
"markupsafe": "python-markupsafe:p",
|
"markupsafe": "python-markupsafe:p",
|
||||||
"pip": "python-pip:p",
|
"pip": "python-pip:p",
|
||||||
"pyyaml": "python-pyyaml:p",
|
"pyyaml": "python-pyyaml:p python-types-pyyaml:p",
|
||||||
"ruamel.yaml": "python-ruamel-yaml:p",
|
"ruamel.yaml": "python-ruamel-yaml:p",
|
||||||
# "ruamel.yaml": "python-ruamel-yaml:p python-ruamel.yaml.clib:p",
|
# "ruamel.yaml": "python-ruamel-yaml:p python-ruamel.yaml.clib:p",
|
||||||
"sphinx": "python-markupsafe:p",
|
"sphinx": "python-markupsafe:p",
|
||||||
@@ -240,6 +240,7 @@ jobs:
|
|||||||
"wheel": "python-wheel:p",
|
"wheel": "python-wheel:p",
|
||||||
"pyedaa.projectmodel": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
"pyedaa.projectmodel": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
||||||
"pyedaa.reports": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
"pyedaa.reports": "python-ruamel-yaml:p python-ruamel.yaml.clib:p python-lxml:p",
|
||||||
|
"sphinx-reports": "python-markupsafe:p python-pyaml:p python-types-pyyaml:p",
|
||||||
}
|
}
|
||||||
subPackages = {
|
subPackages = {
|
||||||
"pytooling": {
|
"pytooling": {
|
||||||
@@ -342,6 +343,7 @@ jobs:
|
|||||||
|
|
||||||
- name: ✅ Run unit tests (Ubuntu/macOS)
|
- name: ✅ Run unit tests (Ubuntu/macOS)
|
||||||
if: matrix.system != 'windows'
|
if: matrix.system != 'windows'
|
||||||
|
continue-on-error: true
|
||||||
run: |
|
run: |
|
||||||
export ENVIRONMENT_NAME="${{ matrix.envname }}"
|
export ENVIRONMENT_NAME="${{ matrix.envname }}"
|
||||||
export PYTHONPATH=$(pwd)
|
export PYTHONPATH=$(pwd)
|
||||||
|
|||||||
2
.github/workflows/VerifyDocs.yml
vendored
2
.github/workflows/VerifyDocs.yml
vendored
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
76
.github/workflows/_Checking_JobTemplates.yml
vendored
76
.github/workflows/_Checking_JobTemplates.yml
vendored
@@ -5,11 +5,20 @@ on:
|
|||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
Prepare:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/PrepareJob.yml@main
|
||||||
|
|
||||||
ConfigParams:
|
ConfigParams:
|
||||||
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
uses: pyTooling/Actions/.github/workflows/ExtractConfiguration.yml@main
|
||||||
with:
|
with:
|
||||||
package_name: pyDummy
|
package_name: pyDummy
|
||||||
|
|
||||||
|
InstallParams:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
||||||
|
with:
|
||||||
|
package_name: pyDummy
|
||||||
|
python_version_list: ''
|
||||||
|
|
||||||
UnitTestingParams:
|
UnitTestingParams:
|
||||||
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
uses: pyTooling/Actions/.github/workflows/Parameters.yml@main
|
||||||
with:
|
with:
|
||||||
@@ -69,8 +78,8 @@ jobs:
|
|||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
commands: |
|
commands: |
|
||||||
${{ needs.ConfigParams.outputs.mypy_prepare_command }}
|
${{ needs.ConfigParams.outputs.mypy_prepare_command }}
|
||||||
mypy --html-report htmlmypy -p ${{ needs.ConfigParams.outputs.package_fullname }}
|
mypy --html-report report/typing -p ${{ needs.ConfigParams.outputs.package_fullname }}
|
||||||
html_report: 'htmlmypy'
|
html_report: 'report/typing'
|
||||||
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
html_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
||||||
|
|
||||||
DocCoverage:
|
DocCoverage:
|
||||||
@@ -80,19 +89,31 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
directory : ${{ needs.ConfigParams.outputs.package_directors }}
|
directory : ${{ needs.ConfigParams.outputs.package_directory }}
|
||||||
# fail_below: 70
|
# fail_below: 70
|
||||||
|
|
||||||
Package:
|
Package:
|
||||||
uses: pyTooling/Actions/.github/workflows/Package.yml@main
|
uses: pyTooling/Actions/.github/workflows/Package.yml@main
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- UnitTesting
|
# - UnitTesting
|
||||||
- PlatformTesting
|
# - PlatformTesting
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
|
|
||||||
|
Install:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/InstallPackage.yml@main
|
||||||
|
needs:
|
||||||
|
- ConfigParams
|
||||||
|
- UnitTestingParams
|
||||||
|
- InstallParams
|
||||||
|
- Package
|
||||||
|
with:
|
||||||
|
jobs: ${{ needs.InstallParams.outputs.python_jobs }}
|
||||||
|
wheel: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
|
package_name: ${{ needs.ConfigParams.outputs.package_fullname }}
|
||||||
|
|
||||||
PublishCoverageResults:
|
PublishCoverageResults:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishCoverageResults.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishCoverageResults.yml@main
|
||||||
needs:
|
needs:
|
||||||
@@ -112,9 +133,7 @@ jobs:
|
|||||||
coverage_report_html_directory: ${{ needs.ConfigParams.outputs.coverage_report_html_directory }}
|
coverage_report_html_directory: ${{ needs.ConfigParams.outputs.coverage_report_html_directory }}
|
||||||
codecov: true
|
codecov: true
|
||||||
codacy: true
|
codacy: true
|
||||||
secrets:
|
secrets: inherit
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
|
||||||
CODACY_TOKEN: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
|
||||||
|
|
||||||
PublishTestResults:
|
PublishTestResults:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishTestResults.yml@main
|
||||||
@@ -129,8 +148,7 @@ jobs:
|
|||||||
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
merged_junit_artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}
|
||||||
codecov: true
|
codecov: true
|
||||||
dorny: true
|
dorny: true
|
||||||
secrets:
|
secrets: inherit
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
|
||||||
|
|
||||||
# VerifyDocs:
|
# VerifyDocs:
|
||||||
# uses: pyTooling/Actions/.github/workflows/VerifyDocs.yml@main
|
# uses: pyTooling/Actions/.github/workflows/VerifyDocs.yml@main
|
||||||
@@ -161,7 +179,6 @@ jobs:
|
|||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- PublishCoverageResults
|
- PublishCoverageResults
|
||||||
- PublishTestResults
|
- PublishTestResults
|
||||||
- Documentation
|
|
||||||
with:
|
with:
|
||||||
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
sqlite_coverage_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_sqlite }}-
|
||||||
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
xml_unittest_artifacts_prefix: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).unittesting_xml }}-
|
||||||
@@ -189,29 +206,54 @@ jobs:
|
|||||||
coverage: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
coverage: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).codecoverage_html }}
|
||||||
typing: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
typing: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).statictyping_html }}
|
||||||
|
|
||||||
ReleasePage:
|
TriggerTaggedRelease:
|
||||||
uses: pyTooling/Actions/.github/workflows/Release.yml@main
|
uses: pyTooling/Actions/.github/workflows/TagReleaseCommit.yml@main
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
needs:
|
||||||
|
- Prepare
|
||||||
- UnitTesting
|
- UnitTesting
|
||||||
- PlatformTesting
|
- PlatformTesting
|
||||||
|
- Install
|
||||||
# - StaticTypeCheck
|
# - StaticTypeCheck
|
||||||
- Package
|
- Package
|
||||||
- PublishToGitHubPages
|
- PublishToGitHubPages
|
||||||
|
permissions:
|
||||||
|
contents: write # required for create tag
|
||||||
|
actions: write # required for trigger workflow
|
||||||
|
with:
|
||||||
|
version: ${{ needs.Prepare.outputs.version }}
|
||||||
|
auto_tag: ${{ needs.Prepare.outputs.is_release_commit }}
|
||||||
|
secrets: inherit
|
||||||
|
|
||||||
|
ReleasePage:
|
||||||
|
uses: pyTooling/Actions/.github/workflows/PublishReleaseNotes.yml@main
|
||||||
|
needs:
|
||||||
|
- Prepare
|
||||||
|
- UnitTesting
|
||||||
|
- PlatformTesting
|
||||||
|
- Install
|
||||||
|
# - StaticTypeCheck
|
||||||
|
- Package
|
||||||
|
- PublishToGitHubPages
|
||||||
|
if: needs.Prepare.outputs.is_release_tag == 'true'
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
actions: write
|
||||||
|
with:
|
||||||
|
tag: ${{ needs.Prepare.outputs.version }}
|
||||||
|
secrets: inherit
|
||||||
|
|
||||||
PublishOnPyPI:
|
PublishOnPyPI:
|
||||||
uses: pyTooling/Actions/.github/workflows/PublishOnPyPI.yml@main
|
uses: pyTooling/Actions/.github/workflows/PublishOnPyPI.yml@main
|
||||||
if: startsWith(github.ref, 'refs/tags')
|
|
||||||
needs:
|
needs:
|
||||||
- UnitTestingParams
|
- UnitTestingParams
|
||||||
- ReleasePage
|
- ReleasePage
|
||||||
# - Package
|
# - Package
|
||||||
|
if: needs.Prepare.outputs.is_release_tag == 'true'
|
||||||
with:
|
with:
|
||||||
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
python_version: ${{ needs.UnitTestingParams.outputs.python_version }}
|
||||||
requirements: -r dist/requirements.txt
|
requirements: -r dist/requirements.txt
|
||||||
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
artifact: ${{ fromJson(needs.UnitTestingParams.outputs.artifact_names).package_all }}
|
||||||
secrets:
|
secrets: inherit
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
|
||||||
|
|
||||||
ArtifactCleanUp:
|
ArtifactCleanUp:
|
||||||
uses: pyTooling/Actions/.github/workflows/ArtifactCleanUp.yml@main
|
uses: pyTooling/Actions/.github/workflows/ArtifactCleanUp.yml@main
|
||||||
|
|||||||
@@ -16,4 +16,4 @@ jobs:
|
|||||||
secrets:
|
secrets:
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
||||||
CODACY_PROJECT_TOKEN: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
CODACY_TOKEN: ${{ secrets.CODACY_TOKEN }}
|
||||||
|
|||||||
6
.github/workflows/_Checking_Nightly.yml
vendored
6
.github/workflows/_Checking_Nightly.yml
vendored
@@ -45,11 +45,9 @@ jobs:
|
|||||||
uses: pyTooling/Actions/.github/workflows/NightlyRelease.yml@main
|
uses: pyTooling/Actions/.github/workflows/NightlyRelease.yml@main
|
||||||
needs:
|
needs:
|
||||||
- Build
|
- Build
|
||||||
secrets: inherit
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: write
|
contents: write
|
||||||
actions: write
|
actions: write
|
||||||
# attestations: write
|
|
||||||
with:
|
with:
|
||||||
can-fail: true
|
can-fail: true
|
||||||
prerelease: true
|
prerelease: true
|
||||||
@@ -78,16 +76,15 @@ jobs:
|
|||||||
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
||||||
document:$archive8.tzst: Archive 8 - tzst + dir
|
document:$archive8.tzst: Archive 8 - tzst + dir
|
||||||
document:$archive9.tar.zst:Archive 9 - tar.zst + dir
|
document:$archive9.tar.zst:Archive 9 - tar.zst + dir
|
||||||
|
secrets: inherit
|
||||||
|
|
||||||
NightlyPageWithInventory:
|
NightlyPageWithInventory:
|
||||||
uses: ./.github/workflows/NightlyRelease.yml
|
uses: ./.github/workflows/NightlyRelease.yml
|
||||||
needs:
|
needs:
|
||||||
- Build
|
- Build
|
||||||
secrets: inherit
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: write
|
contents: write
|
||||||
actions: write
|
actions: write
|
||||||
# attestations: write
|
|
||||||
with:
|
with:
|
||||||
can-fail: true
|
can-fail: true
|
||||||
replacements: |
|
replacements: |
|
||||||
@@ -121,3 +118,4 @@ jobs:
|
|||||||
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
document:$archive7.tar.gz: Archive 7 - tar.gz + dir
|
||||||
document:$archive8.tzst: Archive 8 - tzst + dir
|
document:$archive8.tzst: Archive 8 - tzst + dir
|
||||||
document:$archive9.tar.zst: Archive 9 - tar.zst + dir
|
document:$archive9.tar.zst: Archive 9 - tar.zst + dir
|
||||||
|
secrets: inherit
|
||||||
|
|||||||
@@ -16,4 +16,4 @@ jobs:
|
|||||||
secrets:
|
secrets:
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
||||||
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}
|
||||||
CODACY_PROJECT_TOKEN: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
CODACY_TOKEN: ${{ secrets.CODACY_TOKEN }}
|
||||||
|
|||||||
@@ -77,7 +77,7 @@ As shown in the screenshots above, the expected order is:
|
|||||||
|
|
||||||
[**NightlyRelease**](.github/workflows/NightlyRelease.yml): publish GitHub Release.
|
[**NightlyRelease**](.github/workflows/NightlyRelease.yml): publish GitHub Release.
|
||||||
|
|
||||||
[**Release**](.github/workflows/Release.yml): publish GitHub Release.
|
[**PublishReleaseNotes**](.github/workflows/PublishReleaseNotes.yml): publish GitHub Release.
|
||||||
- **Documentation:**
|
- **Documentation:**
|
||||||
[**SphinxDocumentation**](.github/workflows/PublishCoverageResults.yml): create HTML and LaTeX documentation using
|
[**SphinxDocumentation**](.github/workflows/PublishCoverageResults.yml): create HTML and LaTeX documentation using
|
||||||
Sphinx.
|
Sphinx.
|
||||||
|
|||||||
@@ -55,8 +55,7 @@ Simple Example
|
|||||||
uses: pyTooling/Actions/.github/workflows/CoverageCollection.yml@r0
|
uses: pyTooling/Actions/.github/workflows/CoverageCollection.yml@r0
|
||||||
with:
|
with:
|
||||||
artifact: Coverage
|
artifact: Coverage
|
||||||
secrets:
|
secrets: inherit
|
||||||
codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
|
||||||
|
|
||||||
Complex Example
|
Complex Example
|
||||||
===============
|
===============
|
||||||
@@ -71,8 +70,7 @@ Complex Example
|
|||||||
with:
|
with:
|
||||||
python_version: ${{ needs.Params.outputs.python_version }}
|
python_version: ${{ needs.Params.outputs.python_version }}
|
||||||
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).codecoverage_html }}
|
artifact: ${{ fromJson(needs.Params.outputs.artifact_names).codecoverage_html }}
|
||||||
secrets:
|
secrets: inherit
|
||||||
codacy_token: ${{ secrets.CODACY_PROJECT_TOKEN }}
|
|
||||||
|
|
||||||
Parameters
|
Parameters
|
||||||
**********
|
**********
|
||||||
|
|||||||
45
doc/conf.py
45
doc/conf.py
@@ -1,20 +1,29 @@
|
|||||||
# If extensions (or modules to document with autodoc) are in another directory,
|
# If extensions (or modules to document with autodoc) are in another directory,
|
||||||
# add these directories to sys.path here. If the directory is relative to the
|
# add these directories to sys.path here. If the directory is relative to the
|
||||||
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||||
from importlib.util import find_spec
|
|
||||||
from sys import path as sys_path
|
from sys import path as sys_path
|
||||||
from os.path import abspath
|
from os.path import abspath
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from json import loads
|
|
||||||
|
|
||||||
from pyTooling.Packaging import extractVersionInformation
|
from pyTooling.Packaging import extractVersionInformation
|
||||||
|
|
||||||
|
# ==============================================================================
|
||||||
|
# Project configuration
|
||||||
|
# ==============================================================================
|
||||||
|
githubNamespace = "pyTooling"
|
||||||
|
githubProject = "Actions"
|
||||||
|
pythonProject = "pyDummy"
|
||||||
|
directoryName = pythonProject.replace('.', '/')
|
||||||
|
|
||||||
|
|
||||||
|
# ==============================================================================
|
||||||
|
# Project paths
|
||||||
|
# ==============================================================================
|
||||||
ROOT = Path(__file__).resolve().parent
|
ROOT = Path(__file__).resolve().parent
|
||||||
|
|
||||||
sys_path.insert(0, abspath("."))
|
sys_path.insert(0, abspath("."))
|
||||||
sys_path.insert(0, abspath(".."))
|
sys_path.insert(0, abspath(".."))
|
||||||
sys_path.insert(0, abspath("../pyDummy"))
|
sys_path.insert(0, abspath(f"../{directoryName}"))
|
||||||
# sys_path.insert(0, abspath("_extensions"))
|
|
||||||
|
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
@@ -23,11 +32,7 @@ sys_path.insert(0, abspath("../pyDummy"))
|
|||||||
# The version info for the project you're documenting, acts as replacement for
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
# |version| and |release|, also used in various other places throughout the
|
# |version| and |release|, also used in various other places throughout the
|
||||||
# built documents.
|
# built documents.
|
||||||
githubNamespace = "pyTooling"
|
packageInformationFile = Path(f"../{directoryName}/__init__.py")
|
||||||
githubProject = "Actions"
|
|
||||||
project = "pyDummy"
|
|
||||||
|
|
||||||
packageInformationFile = Path(f"../{project}/__init__.py")
|
|
||||||
versionInformation = extractVersionInformation(packageInformationFile)
|
versionInformation = extractVersionInformation(packageInformationFile)
|
||||||
|
|
||||||
author = versionInformation.Author
|
author = versionInformation.Author
|
||||||
@@ -105,7 +110,7 @@ html_last_updated_fmt = "%d.%m.%Y"
|
|||||||
# Python settings
|
# Python settings
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
modindex_common_prefix = [
|
modindex_common_prefix = [
|
||||||
f"{project}."
|
f"{pythonProject}."
|
||||||
]
|
]
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
@@ -257,13 +262,13 @@ todo_link_only = True
|
|||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
# report_unittest_testsuites = {
|
# report_unittest_testsuites = {
|
||||||
# "src": {
|
# "src": {
|
||||||
# "name": f"{project}",
|
# "name": f"{pythonProject}",
|
||||||
# "xml_report": "../report/unit/unittest.xml",
|
# "xml_report": "../report/unit/unittest.xml",
|
||||||
# }
|
# }
|
||||||
# }
|
# }
|
||||||
# report_codecov_packages = {
|
# report_codecov_packages = {
|
||||||
# "src": {
|
# "src": {
|
||||||
# "name": f"{project}",
|
# "name": f"{pythonProject}",
|
||||||
# "json_report": "../report/coverage/coverage.json",
|
# "json_report": "../report/coverage/coverage.json",
|
||||||
# "fail_below": 80,
|
# "fail_below": 80,
|
||||||
# "levels": "default"
|
# "levels": "default"
|
||||||
@@ -271,8 +276,8 @@ todo_link_only = True
|
|||||||
# }
|
# }
|
||||||
# report_doccov_packages = {
|
# report_doccov_packages = {
|
||||||
# "src": {
|
# "src": {
|
||||||
# "name": f"{project}",
|
# "name": f"{pythonProject}",
|
||||||
# "directory": f"../{project}",
|
# "directory": f"../{directoryName}",
|
||||||
# "fail_below": 80,
|
# "fail_below": 80,
|
||||||
# "levels": "default"
|
# "levels": "default"
|
||||||
# }
|
# }
|
||||||
@@ -289,17 +294,17 @@ todo_link_only = True
|
|||||||
# AutoAPI.Sphinx
|
# AutoAPI.Sphinx
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
autoapi_modules = {
|
autoapi_modules = {
|
||||||
f"{project}": {
|
f"{pythonProject}": {
|
||||||
"template": "package",
|
"template": "package",
|
||||||
"output": project,
|
"output": pythonProject,
|
||||||
"override": True
|
"override": True
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for directory in [mod for mod in Path(f"../{project}").iterdir() if mod.is_dir() and mod.name != "__pycache__"]:
|
for directory in [mod for mod in Path(f"../{directoryName}").iterdir() if mod.is_dir() and mod.name != "__pycache__"]:
|
||||||
print(f"Adding module rule for '{project}.{directory.name}'")
|
print(f"Adding module rule for '{pythonProject}.{directory.name}'")
|
||||||
autoapi_modules[f"{project}.{directory.name}"] = {
|
autoapi_modules[f"{pythonProject}.{directory.name}"] = {
|
||||||
"template": "module",
|
"template": "module",
|
||||||
"output": project,
|
"output": pythonProject,
|
||||||
"override": True
|
"override": True
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,4 +4,4 @@ Code Coverage Report
|
|||||||
Code coverage report generated with `pytest <https://github.com/pytest-dev/pytest>`__ and `Coverage.py <https://github.com/nedbat/coveragepy/tree/master>`__.
|
Code coverage report generated with `pytest <https://github.com/pytest-dev/pytest>`__ and `Coverage.py <https://github.com/nedbat/coveragepy/tree/master>`__.
|
||||||
|
|
||||||
.. #report:code-coverage::
|
.. #report:code-coverage::
|
||||||
:packageid: src
|
:reportid: src
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ pushd %~dp0
|
|||||||
REM Command file for Sphinx documentation
|
REM Command file for Sphinx documentation
|
||||||
|
|
||||||
if "%SPHINXBUILD%" == "" (
|
if "%SPHINXBUILD%" == "" (
|
||||||
set SPHINXBUILD=sphinx-build
|
set SPHINXBUILD=py -3.13 -m sphinx.cmd.build
|
||||||
)
|
)
|
||||||
set SOURCEDIR=.
|
set SOURCEDIR=.
|
||||||
set BUILDDIR=_build
|
set BUILDDIR=_build
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
-r ../requirements.txt
|
-r ../requirements.txt
|
||||||
|
|
||||||
pyTooling ~= 8.4
|
pyTooling ~= 8.5
|
||||||
|
|
||||||
# Enforce latest version on ReadTheDocs
|
# Enforce latest version on ReadTheDocs
|
||||||
sphinx ~= 8.2
|
sphinx ~= 8.2
|
||||||
@@ -15,5 +15,5 @@ sphinxcontrib-mermaid ~= 1.0
|
|||||||
autoapi >= 2.0.1
|
autoapi >= 2.0.1
|
||||||
sphinx_design ~= 0.6.1
|
sphinx_design ~= 0.6.1
|
||||||
sphinx-copybutton >= 0.5.2
|
sphinx-copybutton >= 0.5.2
|
||||||
sphinx_autodoc_typehints ~= 3.1
|
sphinx_autodoc_typehints ~= 3.2
|
||||||
sphinx_reports ~= 0.7
|
sphinx_reports ~= 0.9
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ __author__ = "Patrick Lehmann"
|
|||||||
__email__ = "Paebbels@gmail.com"
|
__email__ = "Paebbels@gmail.com"
|
||||||
__copyright__ = "2017-2025, Patrick Lehmann"
|
__copyright__ = "2017-2025, Patrick Lehmann"
|
||||||
__license__ = "Apache License, Version 2.0"
|
__license__ = "Apache License, Version 2.0"
|
||||||
__version__ = "0.4.4"
|
__version__ = "0.4.5"
|
||||||
__keywords__ = ["GitHub Actions"]
|
__keywords__ = ["GitHub Actions"]
|
||||||
__issue_tracker__ = "https://GitHub.com/pyTooling/Actions/issues"
|
__issue_tracker__ = "https://GitHub.com/pyTooling/Actions/issues"
|
||||||
|
|
||||||
@@ -53,9 +53,9 @@ class Base:
|
|||||||
_value: int #: An internal value.
|
_value: int #: An internal value.
|
||||||
|
|
||||||
def __init__(self) -> None:
|
def __init__(self) -> None:
|
||||||
"""
|
# """
|
||||||
Initializes the base-class.
|
# Initializes the base-class.
|
||||||
"""
|
# """
|
||||||
self._value = 0
|
self._value = 0
|
||||||
|
|
||||||
@readonly
|
@readonly
|
||||||
@@ -75,9 +75,9 @@ class Application(Base):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self) -> None:
|
def __init__(self) -> None:
|
||||||
"""
|
# """
|
||||||
Initializes the dummy application.
|
# Initializes the dummy application.
|
||||||
"""
|
# """
|
||||||
super().__init__()
|
super().__init__()
|
||||||
|
|
||||||
platform = Platform()
|
platform = Platform()
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
[build-system]
|
[build-system]
|
||||||
requires = [
|
requires = [
|
||||||
"setuptools ~= 78.1",
|
"setuptools >= 80.0",
|
||||||
"wheel ~= 0.45",
|
"wheel ~= 0.45",
|
||||||
"pyTooling ~= 8.4"
|
"pyTooling ~= 8.5"
|
||||||
]
|
]
|
||||||
build-backend = "setuptools.build_meta"
|
build-backend = "setuptools.build_meta"
|
||||||
|
|
||||||
|
|||||||
@@ -3,7 +3,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -3,7 +3,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
pyTooling ~= 8.4
|
pyTooling ~= 8.5
|
||||||
|
|||||||
2
run.ps1
2
run.ps1
@@ -88,7 +88,7 @@ if ($build)
|
|||||||
rm -Force .\build\bdist.win-amd64
|
rm -Force .\build\bdist.win-amd64
|
||||||
rm -Force .\build\lib
|
rm -Force .\build\lib
|
||||||
Write-Host -ForegroundColor Yellow "[live][BUILD] Building $PackageName package as wheel ..."
|
Write-Host -ForegroundColor Yellow "[live][BUILD] Building $PackageName package as wheel ..."
|
||||||
py -3.13 -m build --wheel
|
py -3.13 -m build --wheel --no-isolation
|
||||||
|
|
||||||
Write-Host -ForegroundColor Yellow "[live][BUILD] Building wheel finished"
|
Write-Host -ForegroundColor Yellow "[live][BUILD] Building wheel finished"
|
||||||
}
|
}
|
||||||
|
|||||||
9
setup.py
9
setup.py
@@ -39,7 +39,8 @@ packageName = "pyDummy"
|
|||||||
packageDirectory = packageName
|
packageDirectory = packageName
|
||||||
packageInformationFile = Path(f"{packageDirectory}/__init__.py")
|
packageInformationFile = Path(f"{packageDirectory}/__init__.py")
|
||||||
|
|
||||||
setup(**DescribePythonPackageHostedOnGitHub(
|
setup(
|
||||||
|
**DescribePythonPackageHostedOnGitHub(
|
||||||
packageName=packageName,
|
packageName=packageName,
|
||||||
description="pyDummy is a test package to verify GitHub actions for Python projects.",
|
description="pyDummy is a test package to verify GitHub actions for Python projects.",
|
||||||
gitHubNamespace=gitHubNamespace,
|
gitHubNamespace=gitHubNamespace,
|
||||||
@@ -47,5 +48,7 @@ setup(**DescribePythonPackageHostedOnGitHub(
|
|||||||
sourceFileWithVersion=packageInformationFile,
|
sourceFileWithVersion=packageInformationFile,
|
||||||
dataFiles={
|
dataFiles={
|
||||||
packageName: ["py.typed"]
|
packageName: ["py.typed"]
|
||||||
}
|
},
|
||||||
))
|
debug=True
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|||||||
@@ -1,13 +1,13 @@
|
|||||||
-r ../requirements.txt
|
-r ../requirements.txt
|
||||||
|
|
||||||
# Coverage collection
|
# Coverage collection
|
||||||
Coverage ~= 7.8
|
Coverage ~= 7.9
|
||||||
|
|
||||||
# Test Runner
|
# Test Runner
|
||||||
pytest ~= 8.3
|
pytest ~= 8.4
|
||||||
pytest-cov ~= 6.1
|
pytest-cov ~= 6.2
|
||||||
|
|
||||||
# Static Type Checking
|
# Static Type Checking
|
||||||
mypy ~= 1.15
|
mypy ~= 1.16
|
||||||
typing_extensions ~= 4.13
|
typing_extensions ~= 4.13
|
||||||
lxml ~= 5.3
|
lxml ~= 5.4
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# Unai Martinez-Corral #
|
# Unai Martinez-Corral #
|
||||||
# #
|
# #
|
||||||
# ==================================================================================================================== #
|
# ==================================================================================================================== #
|
||||||
# Copyright 2020-2024 The pyTooling Authors #
|
# Copyright 2020-2025 The pyTooling Authors #
|
||||||
# #
|
# #
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
# Licensed under the Apache License, Version 2.0 (the "License"); #
|
||||||
# you may not use this file except in compliance with the License. #
|
# you may not use this file except in compliance with the License. #
|
||||||
|
|||||||
Reference in New Issue
Block a user