mirror of
https://github.com/pyTooling/Actions.git
synced 2026-02-12 02:56:56 +08:00
v7.0.0
This commit is contained in:
6
.github/workflows/ApplicationTesting.yml
vendored
6
.github/workflows/ApplicationTesting.yml
vendored
@@ -86,10 +86,10 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.wheel }}' from 'Package' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
name: ${{ inputs.wheel }}
|
||||
path: install
|
||||
@@ -262,7 +262,7 @@ jobs:
|
||||
|
||||
- name: 📤 Upload 'TestReportSummary.xml' artifact
|
||||
if: inputs.apptest_xml_artifact != ''
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.apptest_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
working-directory: report/unit
|
||||
|
||||
4
.github/workflows/BuildTheDocs.yml
vendored
4
.github/workflows/BuildTheDocs.yml
vendored
@@ -41,7 +41,7 @@ jobs:
|
||||
run: printf "::warning title=%s::%s\n" "Deprecated" "'BuildTheDocs.yml' template is deprecated. Please switch to 'SphinxDocumentation.yml'. See https://pytooling.github.io/Actions/JobTemplate/Documentation/SphinxDocumentation.html"
|
||||
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 🛳️ Build documentation
|
||||
uses: buildthedocs/btd@v0
|
||||
@@ -49,7 +49,7 @@ jobs:
|
||||
skip-deploy: true
|
||||
|
||||
- name: 📤 Upload 'documentation' artifacts
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.artifact }}
|
||||
|
||||
6
.github/workflows/CheckCodeQuality.yml
vendored
6
.github/workflows/CheckCodeQuality.yml
vendored
@@ -71,7 +71,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -140,7 +140,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -180,7 +180,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
|
||||
2
.github/workflows/CheckDocumentation.yml
vendored
2
.github/workflows/CheckDocumentation.yml
vendored
@@ -50,7 +50,7 @@ jobs:
|
||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||
uses: actions/setup-python@v6
|
||||
|
||||
4
.github/workflows/CompletePipeline.yml
vendored
4
.github/workflows/CompletePipeline.yml
vendored
@@ -188,7 +188,7 @@ jobs:
|
||||
code_version: ${{ steps.extract.outputs.code_version }}
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
# The command 'git describe' (used for version) needs the history.
|
||||
fetch-depth: 0
|
||||
@@ -237,8 +237,6 @@ jobs:
|
||||
with:
|
||||
jobs: ${{ needs.UnitTestingParams.outputs.python_jobs }}
|
||||
# TODO: shouldn't this be configured by a parameter? Same as directories
|
||||
requirements: "-r tests/unit/requirements.txt"
|
||||
# pacboy: "msys/git python-lxml:p"
|
||||
unittest_report_xml: ${{ needs.ConfigParams.outputs.unittest_report_xml }}
|
||||
coverage_report_xml: ${{ needs.ConfigParams.outputs.coverage_report_xml }}
|
||||
coverage_report_json: ${{ needs.ConfigParams.outputs.coverage_report_json }}
|
||||
|
||||
4
.github/workflows/CoverageCollection.yml
vendored
4
.github/workflows/CoverageCollection.yml
vendored
@@ -75,7 +75,7 @@ jobs:
|
||||
run: printf "::warning title=%s::%s\n" "Deprecated" "'CoverageCollection.yml' template is deprecated. Please switch to 'PublishReleaseNotes.yml'. See https://pytooling.github.io/Actions/JobTemplate/Testing/UnitTesting.html"
|
||||
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -163,7 +163,7 @@ jobs:
|
||||
|
||||
- name: 📤 Upload 'Coverage Report' artifact
|
||||
continue-on-error: true
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.artifact }}
|
||||
working-directory: ${{ steps.getVariables.outputs.coverage_report_html_directory }}
|
||||
|
||||
4
.github/workflows/ExtractConfiguration.yml
vendored
4
.github/workflows/ExtractConfiguration.yml
vendored
@@ -68,7 +68,7 @@ on:
|
||||
|
||||
jobs:
|
||||
Extract:
|
||||
name: 📓 Extract configurations from pyproject.toml
|
||||
name: 🔬 Extract configurations from pyproject.toml
|
||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
||||
outputs:
|
||||
unittest_report_xml: ${{ steps.getVariables.outputs.unittest_report_xml }}
|
||||
@@ -82,7 +82,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||
uses: actions/setup-python@v6
|
||||
|
||||
2
.github/workflows/InstallPackage.yml
vendored
2
.github/workflows/InstallPackage.yml
vendored
@@ -53,7 +53,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: 📥 Download artifacts '${{ inputs.wheel }}' from 'Package' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
name: ${{ inputs.wheel }}
|
||||
path: install
|
||||
|
||||
4
.github/workflows/LaTeXDocumentation.yml
vendored
4
.github/workflows/LaTeXDocumentation.yml
vendored
@@ -60,7 +60,7 @@ jobs:
|
||||
continue-on-error: ${{ inputs.can-fail == 'true' }}
|
||||
steps:
|
||||
- name: 📥 Download artifacts '${{ inputs.latex_artifact }}' from 'SphinxDocumentation' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
name: ${{ inputs.latex_artifact }}
|
||||
path: latex
|
||||
@@ -83,7 +83,7 @@ jobs:
|
||||
latexmk -${{ inputs.processor }} "${{ inputs.document }}.tex"
|
||||
|
||||
- name: 📤 Upload 'PDF Documentation' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.pdf_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.pdf_artifact }}
|
||||
|
||||
2
.github/workflows/NightlyRelease.yml
vendored
2
.github/workflows/NightlyRelease.yml
vendored
@@ -104,7 +104,7 @@ jobs:
|
||||
run: printf "::warning title=%s::%s\n" "NightlyRelease" "'NightlyRelease.yml' template is deprecated. Please switch to 'PublishReleaseNotes.yml'. See https://pytooling.github.io/Actions/JobTemplate/Release/PublishReleaseNotes.html"
|
||||
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
# The command 'git describe' (used for version) needs the history.
|
||||
fetch-depth: 0
|
||||
|
||||
4
.github/workflows/Package.yml
vendored
4
.github/workflows/Package.yml
vendored
@@ -53,7 +53,7 @@ jobs:
|
||||
artifact: ${{ inputs.artifact }}
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -106,7 +106,7 @@ jobs:
|
||||
run: python setup.py bdist_wheel
|
||||
|
||||
- name: 📤 Upload wheel artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.artifact }}
|
||||
working-directory: dist
|
||||
|
||||
4
.github/workflows/Parameters.yml
vendored
4
.github/workflows/Parameters.yml
vendored
@@ -103,7 +103,7 @@ on:
|
||||
macos_intel_image:
|
||||
description: 'The used GitHub Action image for macOS (Intel x86-64) based jobs.'
|
||||
required: false
|
||||
default: 'macos-13'
|
||||
default: 'macos-15-intel'
|
||||
type: string
|
||||
macos_arm_image:
|
||||
description: 'The used GitHub Action image for macOS (ARM aarch64) based jobs.'
|
||||
@@ -154,7 +154,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
# The command 'git describe' (used for version) needs the history.
|
||||
fetch-depth: 0
|
||||
|
||||
2
.github/workflows/PrepareJob.yml
vendored
2
.github/workflows/PrepareJob.yml
vendored
@@ -131,7 +131,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
# The command 'git describe' (used for version) needs the history.
|
||||
fetch-depth: 0
|
||||
|
||||
12
.github/workflows/PublishCoverageResults.yml
vendored
12
.github/workflows/PublishCoverageResults.yml
vendored
@@ -109,13 +109,13 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
|
||||
- name: 📥 Download Artifacts
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
pattern: ${{ inputs.coverage_artifacts_pattern }}
|
||||
path: artifacts
|
||||
@@ -156,7 +156,7 @@ jobs:
|
||||
tree -pash ${{ fromJson(inputs.coverage_report_html).directory }}
|
||||
|
||||
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.coverage_sqlite_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -166,7 +166,7 @@ jobs:
|
||||
retention-days: 1
|
||||
|
||||
- name: 📤 Upload 'Coverage XML Report' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.coverage_xml_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -177,7 +177,7 @@ jobs:
|
||||
retention-days: 1
|
||||
|
||||
- name: 📤 Upload 'Coverage JSON Report' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.coverage_json_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -188,7 +188,7 @@ jobs:
|
||||
retention-days: 1
|
||||
|
||||
- name: 📤 Upload 'Coverage HTML Report' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.coverage_html_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
|
||||
2
.github/workflows/PublishOnPyPI.yml
vendored
2
.github/workflows/PublishOnPyPI.yml
vendored
@@ -56,7 +56,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: 📥 Download artifacts '${{ inputs.artifact }}' from 'Package' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
name: ${{ inputs.artifact }}
|
||||
path: dist
|
||||
|
||||
471
.github/workflows/PublishReleaseNotes.yml
vendored
471
.github/workflows/PublishReleaseNotes.yml
vendored
@@ -132,7 +132,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
# The command 'git describe' (used for version) needs the history.
|
||||
fetch-depth: 0
|
||||
@@ -191,198 +191,6 @@ jobs:
|
||||
exit 1
|
||||
fi
|
||||
|
||||
- name: 📑 Assemble Release Notes
|
||||
id: createReleaseNotes
|
||||
run: |
|
||||
set +e
|
||||
|
||||
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||
ANSI_NOCOLOR=$'\x1b[0m'
|
||||
|
||||
export GH_TOKEN=${{ github.token }}
|
||||
|
||||
# Save release description (from parameter in a file)
|
||||
head -c -1 <<'EOF' > __DESCRIPTION__.md
|
||||
${{ inputs.description }}
|
||||
EOF
|
||||
|
||||
# Save release footer (from parameter in a file)
|
||||
head -c -1 <<'EOF' > __FOOTER__.md
|
||||
${{ inputs.description_footer }}
|
||||
EOF
|
||||
|
||||
# Download Markdown from PullRequest
|
||||
# Readout second parent's SHA
|
||||
# Search PR with that SHA
|
||||
# Load description of that PR
|
||||
printf "Read second parent of current SHA (%s) ... " "${{ github.ref }}"
|
||||
FATHER_SHA=$(git rev-parse ${{ github.ref }}^2 -- 2> /dev/null)
|
||||
if [[ $? -ne 0 || "{FATHER_SHA}" == "" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||
printf "→ ${ANSI_LIGHT_YELLOW}Skipped readout of pull request description. This is not a merge commit.${ANSI_NOCOLOR}\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
|
||||
printf "Search Pull Request to '%s' and branch containing SHA %s ... " "${{ inputs.release_branch }}" "${FATHER_SHA}"
|
||||
PULL_REQUESTS=$(gh pr list --base "${{ inputs.release_branch }}" --search "${FATHER_SHA}" --state "merged" --json "title,number,mergedBy,mergedAt,body")
|
||||
if [[ $? -ne 0 || "${PULL_REQUESTS}" == "" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||
printf "${ANSI_LIGHT_RED}Couldn't find a merged Pull Request to '%s'. -> %s${ANSI_NOCOLOR}\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||
printf "::error title=PullRequest::Couldn't find a merged Pull Request to '%s'. -> %s\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||
exit 1
|
||||
else
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
|
||||
PR_TITLE="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].title")"
|
||||
PR_NUMBER="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].number")"
|
||||
PR_BODY="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].body")"
|
||||
PR_MERGED_BY="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedBy.login")"
|
||||
PR_MERGED_AT="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedAt")"
|
||||
|
||||
printf "Found Pull Request:\n"
|
||||
printf " %s\n" "Title: ${PR_TITLE}"
|
||||
printf " %s\n" "Number: ${PR_NUMBER}"
|
||||
printf " %s\n" "MergedBy: ${PR_MERGED_BY}"
|
||||
printf " %s\n" "MergedAt: ${PR_MERGED_AT} ($(date -d"${PR_MERGED_AT}" '+%d.%m.%Y - %H:%M:%S'))"
|
||||
fi
|
||||
|
||||
printf "%s\n" "${PR_BODY}" > __PULLREQUEST__.md
|
||||
fi
|
||||
|
||||
# Check if a release description file should be used and exists.
|
||||
if [[ "${{ inputs.description_file }}" != "" ]]; then
|
||||
if [[ ! -f "${{ inputs.description_file }}" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}Release description file '%s' not found.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' not found."
|
||||
exit 1
|
||||
elif [[ -s "${{ inputs.description_file }}" ]]; then
|
||||
printf "Use '%s' as main release description.\n" "${{ inputs.description_file }}"
|
||||
cp -v "${{ inputs.description_file }}" __NOTES__.md
|
||||
else
|
||||
printf "${ANSI_LIGHT_RED}Release description file '%s' is empty.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' is empty."
|
||||
exit 1
|
||||
fi
|
||||
# Check if the main release description is provided by a template parameter
|
||||
elif [[ -s __DESCRIPTION__.md ]]; then
|
||||
printf "Use '__DESCRIPTION__.md' as main release description.\n"
|
||||
mv -v __DESCRIPTION__.md __NOTES__.md
|
||||
# Check if the pull request serves as the main release description text.
|
||||
elif [[ -s __PULLREQUEST__.md ]]; then
|
||||
printf "Use '__PULLREQUEST__.md' as main release description.\n"
|
||||
mv -v __PULLREQUEST__.md __NOTES__.md
|
||||
|
||||
printf "Append '%%%%FOOTER%%%%' to '__NOTES__.md'.\n"
|
||||
printf "\n%%%%FOOTER%%%%\n" >> __NOTES__.md
|
||||
else
|
||||
printf "${ANSI_LIGHT_RED}No release description specified (file, parameter, PR text).${ANSI_NOCOLOR}\n"
|
||||
printf "::error title=%s::%s\n" "MissingDescription" "No release description specified (file, parameter, PR text)."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Read release notes main file for placeholder substitution
|
||||
NOTES=$(<__NOTES__.md)
|
||||
|
||||
# Inline description
|
||||
if [[ -s __DESCRIPTION__.md ]]; then
|
||||
NOTES="${NOTES//%%DESCRIPTION%%/$(<__DESCRIPTION__.md)}"
|
||||
else
|
||||
NOTES="${NOTES//%%DESCRIPTION%%/}"
|
||||
fi
|
||||
|
||||
# Inline PullRequest and increase headline levels
|
||||
if [[ -s __PULLREQUEST__.md ]]; then
|
||||
while [[ "${NOTES}" =~ %%(PULLREQUEST(\+[0-3])?)%% ]]; do
|
||||
case "${BASH_REMATCH[1]}" in
|
||||
"PULLREQUEST+0" | "PULLREQUEST")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(<__PULLREQUEST__.md)}"
|
||||
;;
|
||||
"PULLREQUEST+1")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1# /gm;t')}"
|
||||
;;
|
||||
"PULLREQUEST+2")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||
;;
|
||||
"PULLREQUEST+3")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||
;;
|
||||
esac
|
||||
done
|
||||
else
|
||||
while [[ "${NOTES}" =~ %%(PULLREQUEST(\+[0-3])?)%% ]]; do
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/}"
|
||||
done
|
||||
fi
|
||||
|
||||
# inline Footer
|
||||
if [[ -s __FOOTER__.md ]]; then
|
||||
NOTES="${NOTES//%%FOOTER%%/$(<__FOOTER__.md)}"
|
||||
else
|
||||
NOTES="${NOTES//%%FOOTER%%/}"
|
||||
fi
|
||||
|
||||
# Apply replacements
|
||||
while IFS=$'\r\n' read -r patternLine; do
|
||||
# skip empty lines
|
||||
[[ "$patternLine" == "" ]] && continue
|
||||
|
||||
pattern="%${patternLine%%=*}%"
|
||||
replacement="${patternLine#*=}"
|
||||
NOTES="${NOTES//$pattern/$replacement}"
|
||||
done <<<'${{ inputs.replacements }}'
|
||||
|
||||
# Workarounds for stupid GitHub variables
|
||||
owner_repo="${{ github.repository }}"
|
||||
repo=${owner_repo##*/}
|
||||
|
||||
# Replace special identifiers
|
||||
NOTES="${NOTES//%%gh_server%%/${{ github.server_url }}}"
|
||||
NOTES="${NOTES//%%gh_workflow_name%%/${{ github.workflow }}}"
|
||||
NOTES="${NOTES//%%gh_owner%%/${{ github.repository_owner }}}"
|
||||
NOTES="${NOTES//%%gh_repo%%/${repo}}"
|
||||
NOTES="${NOTES//%%gh_owner_repo%%/${{ github.repository }}}"
|
||||
#NOTES="${NOTES//%%gh_pages%%/https://${{ github.repository_owner }}.github.io/${repo}/}"
|
||||
NOTES="${NOTES//%%gh_runid%%/${{ github.run_id }}}"
|
||||
NOTES="${NOTES//%%gh_actor%%/${{ github.actor }}}"
|
||||
NOTES="${NOTES//%%gh_sha%%/${{ github.sha }}}"
|
||||
NOTES="${NOTES//%%date%%/$(date '+%Y-%m-%d')}"
|
||||
NOTES="${NOTES//%%time%%/$(date '+%H:%M:%S %Z')}"
|
||||
NOTES="${NOTES//%%datetime%%/$(date '+%Y-%m-%d %H:%M:%S %Z')}"
|
||||
|
||||
# Write final release notes to file
|
||||
printf "%s\n" "${NOTES}" > __NOTES__.md
|
||||
|
||||
# Display partial contents for debugging
|
||||
if [[ -s __DESCRIPTION__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__DESCRIPTION__.md' ($(stat --printf="%s" "__DESCRIPTION__.md") B) ...."
|
||||
cat __DESCRIPTION__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__DESCRIPTION__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
if [[ -s __PULLREQUEST__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__PULLREQUEST__.md' ($(stat --printf="%s" "__PULLREQUEST__.md") B) ...."
|
||||
cat __PULLREQUEST__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__PULLREQUEST__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
if [[ -s __FOOTER__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__FOOTER__.md' ($(stat --printf="%s" "__FOOTER__.md") B) ...."
|
||||
cat __FOOTER__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__FOOTER__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
|
||||
# Print final release notes
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__NOTES__.md' ($(stat --printf="%s" "__NOTES__.md") B) ...."
|
||||
cat __NOTES__.md
|
||||
printf "::endgroup::\n"
|
||||
|
||||
- name: 📑 Create new Release Page
|
||||
id: createReleasePage
|
||||
if: inputs.mode == 'release'
|
||||
@@ -397,6 +205,15 @@ jobs:
|
||||
|
||||
export GH_TOKEN=${{ github.token }}
|
||||
|
||||
tee "__PRELIMINARY_NOTES__.md" <<EOF
|
||||
Release notes for ${{ inputs.tag }} are created right now ...
|
||||
|
||||
1. download artifacts → (compression?) → upload as assets
|
||||
2. optional: create inventory.json
|
||||
3. assemble release notes → update this text
|
||||
4. optional: remove draft state
|
||||
EOF
|
||||
|
||||
if [[ "${{ inputs.prerelease }}" == "true" ]]; then
|
||||
addPreRelease="--prerelease"
|
||||
fi
|
||||
@@ -409,9 +226,7 @@ jobs:
|
||||
addTitle=("--title" "${{ inputs.title }}")
|
||||
fi
|
||||
|
||||
if [[ -s __NOTES__.md ]]; then
|
||||
addNotes=("--notes-file" "__NOTES__.md")
|
||||
fi
|
||||
addNotes=("--notes-file" "__PRELIMINARY_NOTES__.md")
|
||||
|
||||
printf "Creating release '%s' ... " "${{ inputs.tag }}"
|
||||
message="$(gh release create "${{ inputs.tag }}" --verify-tag --draft $addPreRelease $addLatest "${addTitle[@]}" "${addNotes[@]}" 2>&1)"
|
||||
@@ -439,6 +254,14 @@ jobs:
|
||||
|
||||
export GH_TOKEN=${{ github.token }}
|
||||
|
||||
tee "__PRELIMINARY_NOTES__.md" <<EOF
|
||||
Release notes for ${{ inputs.tag }} are updated right now ...
|
||||
|
||||
1. download artifacts → (compression?) → upload as assets
|
||||
2. optional: create inventory.json
|
||||
3. assemble release notes → update this text
|
||||
EOF
|
||||
|
||||
addDraft="--draft"
|
||||
if [[ "${{ inputs.prerelease }}" == "true" ]]; then
|
||||
addPreRelease="--prerelease"
|
||||
@@ -452,9 +275,7 @@ jobs:
|
||||
addTitle=("--title" "${{ inputs.title }}")
|
||||
fi
|
||||
|
||||
if [[ -s __NOTES__.md ]]; then
|
||||
addNotes=("--notes-file" "__NOTES__.md")
|
||||
fi
|
||||
addNotes=("--notes-file" "__PRELIMINARY_NOTES__.md")
|
||||
|
||||
printf "Creating release '%s' ... " "${{ inputs.tag }}"
|
||||
message="$(gh release create "${{ inputs.tag }}" --verify-tag --draft $addPreRelease $addLatest "${addTitle[@]}" "${addNotes[@]}" 2>&1)"
|
||||
@@ -553,6 +374,10 @@ jobs:
|
||||
)
|
||||
fi
|
||||
|
||||
# Write Markdown table header
|
||||
printf "| Asset Name | File Size | SHA256 |\n" > __ASSETS__.md
|
||||
printf "|------------|-----------|--------|\n" >> __ASSETS__.md
|
||||
|
||||
ERRORS=0
|
||||
# A dictionary of 0/1 to avoid duplicate downloads
|
||||
declare -A downloadedArtifacts
|
||||
@@ -741,6 +566,13 @@ jobs:
|
||||
sha256Checksums[$asset]="sha256:${sha256}"
|
||||
printf "${ANSI_LIGHT_BLUE}${sha256}${ANSI_NOCOLOR}\n"
|
||||
|
||||
# Add asset to Markdown table
|
||||
printf "| %s | %s | %s |\n" \
|
||||
"[${title}](${{ github.server_url }}/${{ github.repository }}/releases/download/${{ inputs.tag }}/${uploadFile#*/})" \
|
||||
"$(stat --printf="%s" "${uploadFile}" | numfmt --format "%.1f" --suffix=B --to=iec-i)" \
|
||||
"\`${sha256}\`" \
|
||||
>> __ASSETS__.md
|
||||
|
||||
# Add asset to JSON inventory
|
||||
if [[ "${{ inputs.inventory-json }}" != "" ]]; then
|
||||
if [[ "${categories}" != "${title}" ]]; then
|
||||
@@ -775,7 +607,7 @@ jobs:
|
||||
if [[ $? -eq 0 ]]; then
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
|
||||
printf " checking assets SHA256 checksum ... \n"
|
||||
printf " checking assets SHA256 checksum ... "
|
||||
ghSHA256=$(gh release view --json assets --jq ".assets[] | select(.name == \"${asset}\") | .digest" ${{ inputs.tag }})
|
||||
if [[ "${ghSHA256}" == "${sha256Checksums[$asset]}" ]]; then
|
||||
printf "${ANSI_LIGHT_GREEN}[PASSED]${ANSI_NOCOLOR}\n"
|
||||
@@ -829,6 +661,245 @@ jobs:
|
||||
exit 1
|
||||
fi
|
||||
|
||||
- name: 📑 Assemble Release Notes
|
||||
id: createReleaseNotes
|
||||
run: |
|
||||
set +e
|
||||
|
||||
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||
ANSI_NOCOLOR=$'\x1b[0m'
|
||||
|
||||
export GH_TOKEN=${{ github.token }}
|
||||
|
||||
# Save release description (from parameter in a file)
|
||||
head -c -1 <<'EOF' > __DESCRIPTION__.md
|
||||
${{ inputs.description }}
|
||||
EOF
|
||||
|
||||
# Save release footer (from parameter in a file)
|
||||
head -c -1 <<'EOF' > __FOOTER__.md
|
||||
${{ inputs.description_footer }}
|
||||
EOF
|
||||
|
||||
# Download Markdown from PullRequest
|
||||
# Readout second parent's SHA
|
||||
# Search PR with that SHA
|
||||
# Load description of that PR
|
||||
printf "Read second parent of current SHA (%s) ... " "${{ github.ref }}"
|
||||
FATHER_SHA=$(git rev-parse ${{ github.ref }}^2 -- 2> /dev/null)
|
||||
if [[ $? -ne 0 || "{FATHER_SHA}" == "" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||
printf "→ ${ANSI_LIGHT_YELLOW}Skipped readout of pull request description. This is not a merge commit.${ANSI_NOCOLOR}\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
|
||||
printf "Search Pull Request to '%s' and branch containing SHA %s ... " "${{ inputs.release_branch }}" "${FATHER_SHA}"
|
||||
PULL_REQUESTS=$(gh pr list --base "${{ inputs.release_branch }}" --search "${FATHER_SHA}" --state "merged" --json "title,number,mergedBy,mergedAt,body")
|
||||
if [[ $? -ne 0 || "${PULL_REQUESTS}" == "" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||
printf "${ANSI_LIGHT_RED}Couldn't find a merged Pull Request to '%s'. -> %s${ANSI_NOCOLOR}\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||
printf "::error title=PullRequest::Couldn't find a merged Pull Request to '%s'. -> %s\n" "${{ inputs.release_branch }}" "${PULL_REQUESTS}"
|
||||
exit 1
|
||||
else
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
|
||||
PR_TITLE="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].title")"
|
||||
PR_NUMBER="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].number")"
|
||||
PR_BODY="$( printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].body")"
|
||||
PR_MERGED_BY="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedBy.login")"
|
||||
PR_MERGED_AT="$(printf "%s\n" "${PULL_REQUESTS}" | jq --raw-output ".[0].mergedAt")"
|
||||
|
||||
printf "Found Pull Request:\n"
|
||||
printf " %s\n" "Title: ${PR_TITLE}"
|
||||
printf " %s\n" "Number: ${PR_NUMBER}"
|
||||
printf " %s\n" "MergedBy: ${PR_MERGED_BY}"
|
||||
printf " %s\n" "MergedAt: ${PR_MERGED_AT} ($(date -d"${PR_MERGED_AT}" '+%d.%m.%Y - %H:%M:%S'))"
|
||||
fi
|
||||
|
||||
printf "%s\n" "${PR_BODY}" > __PULLREQUEST__.md
|
||||
fi
|
||||
|
||||
# Check if a release description file should be used and exists.
|
||||
if [[ "${{ inputs.description_file }}" != "" ]]; then
|
||||
if [[ ! -f "${{ inputs.description_file }}" ]]; then
|
||||
printf "${ANSI_LIGHT_RED}Release description file '%s' not found.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' not found."
|
||||
exit 1
|
||||
elif [[ -s "${{ inputs.description_file }}" ]]; then
|
||||
printf "Use '%s' as main release description.\n" "${{ inputs.description_file }}"
|
||||
cp -v "${{ inputs.description_file }}" __NOTES__.md
|
||||
else
|
||||
printf "${ANSI_LIGHT_RED}Release description file '%s' is empty.${ANSI_NOCOLOR}\n" "${{ inputs.description_file }}"
|
||||
printf "::error title=%s::%s\n" "FileNotFound" "Release description file '${{ inputs.description_file }}' is empty."
|
||||
exit 1
|
||||
fi
|
||||
# Check if the main release description is provided by a template parameter
|
||||
elif [[ -s __DESCRIPTION__.md ]]; then
|
||||
printf "Use '__DESCRIPTION__.md' as main release description.\n"
|
||||
mv -v __DESCRIPTION__.md __NOTES__.md
|
||||
# Check if the pull request serves as the main release description text.
|
||||
elif [[ -s __PULLREQUEST__.md ]]; then
|
||||
printf "Use '__PULLREQUEST__.md' as main release description.\n"
|
||||
mv -v __PULLREQUEST__.md __NOTES__.md
|
||||
|
||||
printf "Append '%%%%FOOTER%%%%' to '__NOTES__.md'.\n"
|
||||
printf "\n%%%%FOOTER%%%%\n" >> __NOTES__.md
|
||||
else
|
||||
printf "${ANSI_LIGHT_RED}No release description specified (file, parameter, PR text).${ANSI_NOCOLOR}\n"
|
||||
printf "::error title=%s::%s\n" "MissingDescription" "No release description specified (file, parameter, PR text)."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Read release notes main file for placeholder substitution
|
||||
NOTES=$(<__NOTES__.md)
|
||||
|
||||
# Inline description
|
||||
if [[ -s __DESCRIPTION__.md ]]; then
|
||||
NOTES="${NOTES//%%DESCRIPTION%%/$(<__DESCRIPTION__.md)}"
|
||||
else
|
||||
NOTES="${NOTES//%%DESCRIPTION%%/}"
|
||||
fi
|
||||
|
||||
# Inline PullRequest and increase headline levels
|
||||
if [[ -s __PULLREQUEST__.md ]]; then
|
||||
while [[ "${NOTES}" =~ %%(PULLREQUEST(\+[0-3])?)%% ]]; do
|
||||
case "${BASH_REMATCH[1]}" in
|
||||
"PULLREQUEST+0" | "PULLREQUEST")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(<__PULLREQUEST__.md)}"
|
||||
;;
|
||||
"PULLREQUEST+1")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1# /gm;t')}"
|
||||
;;
|
||||
"PULLREQUEST+2")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||
;;
|
||||
"PULLREQUEST+3")
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/$(cat __PULLREQUEST__.md | sed -E 's/^(#+) /\1### /gm;t')}"
|
||||
;;
|
||||
esac
|
||||
done
|
||||
else
|
||||
while [[ "${NOTES}" =~ %%(PULLREQUEST(\+[0-3])?)%% ]]; do
|
||||
NOTES="${NOTES//${BASH_REMATCH[0]}/}"
|
||||
done
|
||||
fi
|
||||
|
||||
# Inline Files table
|
||||
if [[ -s __ASSETS__.md ]]; then
|
||||
NOTES="${NOTES//%%ASSETS%%/$(<__ASSETS__.md)}"
|
||||
else
|
||||
NOTES="${NOTES//%%ASSETS%%/}"
|
||||
fi
|
||||
|
||||
# Inline Footer
|
||||
if [[ -s __FOOTER__.md ]]; then
|
||||
NOTES="${NOTES//%%FOOTER%%/$(<__FOOTER__.md)}"
|
||||
else
|
||||
NOTES="${NOTES//%%FOOTER%%/}"
|
||||
fi
|
||||
|
||||
# Apply replacements
|
||||
while IFS=$'\r\n' read -r patternLine; do
|
||||
# skip empty lines
|
||||
[[ "$patternLine" == "" ]] && continue
|
||||
|
||||
pattern="%${patternLine%%=*}%"
|
||||
replacement="${patternLine#*=}"
|
||||
NOTES="${NOTES//$pattern/$replacement}"
|
||||
done <<<'${{ inputs.replacements }}'
|
||||
|
||||
# Workarounds for stupid GitHub variables
|
||||
owner_repo="${{ github.repository }}"
|
||||
repo=${owner_repo##*/}
|
||||
|
||||
# Replace special identifiers
|
||||
NOTES="${NOTES//%%gh_server%%/${{ github.server_url }}}"
|
||||
NOTES="${NOTES//%%gh_workflow_name%%/${{ github.workflow }}}"
|
||||
NOTES="${NOTES//%%gh_owner%%/${{ github.repository_owner }}}"
|
||||
NOTES="${NOTES//%%gh_repo%%/${repo}}"
|
||||
NOTES="${NOTES//%%gh_owner_repo%%/${{ github.repository }}}"
|
||||
#NOTES="${NOTES//%%gh_pages%%/https://${{ github.repository_owner }}.github.io/${repo}/}"
|
||||
NOTES="${NOTES//%%gh_runid%%/${{ github.run_id }}}"
|
||||
NOTES="${NOTES//%%gh_actor%%/${{ github.actor }}}"
|
||||
NOTES="${NOTES//%%gh_sha%%/${{ github.sha }}}"
|
||||
NOTES="${NOTES//%%date%%/$(date '+%Y-%m-%d')}"
|
||||
NOTES="${NOTES//%%time%%/$(date '+%H:%M:%S %Z')}"
|
||||
NOTES="${NOTES//%%datetime%%/$(date '+%Y-%m-%d %H:%M:%S %Z')}"
|
||||
|
||||
# Write final release notes to file
|
||||
printf "%s\n" "${NOTES}" > __NOTES__.md
|
||||
|
||||
# Display partial contents for debugging
|
||||
if [[ -s __DESCRIPTION__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__DESCRIPTION__.md' ($(stat --printf="%s" "__DESCRIPTION__.md") B) ...."
|
||||
cat __DESCRIPTION__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__DESCRIPTION__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
if [[ -s __PULLREQUEST__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__PULLREQUEST__.md' ($(stat --printf="%s" "__PULLREQUEST__.md") B) ...."
|
||||
cat __PULLREQUEST__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__PULLREQUEST__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
if [[ -s __ASSETS__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__ASSETS__.md' ($(stat --printf="%s" "__ASSETS__.md") B) ...."
|
||||
cat __ASSETS__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__ASSETS__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
if [[ -s __FOOTER__.md ]]; then
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__FOOTER__.md' ($(stat --printf="%s" "__FOOTER__.md") B) ...."
|
||||
cat __FOOTER__.md
|
||||
printf "::endgroup::\n"
|
||||
else
|
||||
printf "${ANSI_LIGHT_YELLOW}No '__FOOTER__.md' found.${ANSI_NOCOLOR}\n"
|
||||
fi
|
||||
|
||||
# Print final release notes
|
||||
printf "::group::${ANSI_LIGHT_BLUE}%s${ANSI_NOCOLOR}\n" "Content of '__NOTES__.md' ($(stat --printf="%s" "__NOTES__.md") B) ...."
|
||||
cat __NOTES__.md
|
||||
printf "::endgroup::\n"
|
||||
|
||||
- name: 📑 Update release notes
|
||||
id: updateReleaseNotes
|
||||
run: |
|
||||
set +e
|
||||
|
||||
ANSI_LIGHT_RED=$'\x1b[91m'
|
||||
ANSI_LIGHT_GREEN=$'\x1b[92m'
|
||||
ANSI_LIGHT_YELLOW=$'\x1b[93m'
|
||||
ANSI_LIGHT_BLUE=$'\x1b[94m'
|
||||
ANSI_NOCOLOR=$'\x1b[0m'
|
||||
|
||||
export GH_TOKEN=${{ github.token }}
|
||||
|
||||
if [[ -s __ASSETS__.md ]]; then
|
||||
addNotes=("--notes-file" "__ASSETS__.md")
|
||||
else
|
||||
printf " ${ANSI_LIGHT_RED}File '%s' not found.${ANSI_NOCOLOR}\n" "__ASSETS__.md"
|
||||
printf "::error title=%s::%s\n" "InternalError" "File '__ASSETS__.md' not found."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
printf "Updating release '%s' ... " "${{ inputs.tag }}"
|
||||
message="$(gh release edit "${addNotes[@]}" "${{ inputs.tag }}" 2>&1)"
|
||||
if [[ $? -eq 0 ]]; then
|
||||
printf "${ANSI_LIGHT_GREEN}[OK]${ANSI_NOCOLOR}\n"
|
||||
printf " Release page: %s\n" "${message}"
|
||||
else
|
||||
printf "${ANSI_LIGHT_RED}[FAILED]${ANSI_NOCOLOR}\n"
|
||||
printf " ${ANSI_LIGHT_RED}Couldn't update release '%s' -> Error: '%s'.${ANSI_NOCOLOR}\n" "${{ inputs.tag }}" "${message}"
|
||||
printf "::error title=%s::%s\n" "InternalError" "Couldn't update release '${{ inputs.tag }}' -> Error: '${message}'."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
- name: 📑 Remove draft state from Release Page
|
||||
id: removeDraft
|
||||
if: ${{ ! inputs.draft }}
|
||||
|
||||
6
.github/workflows/PublishTestResults.yml
vendored
6
.github/workflows/PublishTestResults.yml
vendored
@@ -102,10 +102,10 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 📥 Download Artifacts
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
pattern: ${{ inputs.unittest_artifacts_pattern }}
|
||||
path: artifacts
|
||||
@@ -156,7 +156,7 @@ jobs:
|
||||
fail_ci_if_error: true
|
||||
|
||||
- name: 📤 Upload merged 'JUnit Test Summary' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.merged_junit_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.merged_junit_artifact }}
|
||||
|
||||
48
.github/workflows/PublishToGitHubPages.yml
vendored
48
.github/workflows/PublishToGitHubPages.yml
vendored
@@ -45,45 +45,51 @@ on:
|
||||
default: ''
|
||||
type: string
|
||||
|
||||
outputs:
|
||||
github_pages_url:
|
||||
description: "URL to GitHub Pages."
|
||||
value: ${{ jobs.PrepareGitHubPages.outputs.github_pages_url }}
|
||||
|
||||
jobs:
|
||||
|
||||
PublishToGitHubPages:
|
||||
name: 📚 Publish to GH-Pages
|
||||
PrepareGitHubPages:
|
||||
name: 📖 Merge multiple contents for publishing
|
||||
runs-on: "ubuntu-${{ inputs.ubuntu_image_version }}"
|
||||
|
||||
permissions:
|
||||
pages: write # to deploy to Pages
|
||||
id-token: write # to verify the deployment originates from an appropriate source
|
||||
environment:
|
||||
name: github-pages
|
||||
url: ${{ steps.deployment.outputs.page_url }}
|
||||
outputs:
|
||||
github_pages_url: ${{ steps.deployment.outputs.page_url }}
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.doc }}' from 'SphinxDocumentation' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
with:
|
||||
name: ${{ inputs.doc }}
|
||||
path: public
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.coverage }}' from 'Coverage' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: ${{ inputs.coverage != '' }}
|
||||
with:
|
||||
name: ${{ inputs.coverage }}
|
||||
path: public/coverage
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.typing }}' from 'StaticTypeCheck' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: ${{ inputs.typing != '' }}
|
||||
with:
|
||||
name: ${{ inputs.typing }}
|
||||
path: public/typing
|
||||
|
||||
- name: '📓 Publish site to GitHub Pages'
|
||||
- name: 📑 Upload static files as artifact
|
||||
if: github.event_name != 'pull_request'
|
||||
run: |
|
||||
cd public
|
||||
touch .nojekyll
|
||||
git init
|
||||
cp ../.git/config ./.git/config
|
||||
git add .
|
||||
git config --local user.email "BuildTheDocs@GitHubActions"
|
||||
git config --local user.name "GitHub Actions"
|
||||
git commit -a -m "update ${{ github.sha }}"
|
||||
git push -u origin +HEAD:gh-pages
|
||||
uses: actions/upload-pages-artifact@v4
|
||||
with:
|
||||
path: public/
|
||||
|
||||
- name: 📖 Deploy to GitHub Pages
|
||||
id: deployment
|
||||
if: github.event_name != 'pull_request'
|
||||
uses: actions/deploy-pages@v4
|
||||
|
||||
16
.github/workflows/SphinxDocumentation.yml
vendored
16
.github/workflows/SphinxDocumentation.yml
vendored
@@ -86,7 +86,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -105,7 +105,7 @@ jobs:
|
||||
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: inputs.unittest_xml_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.unittest_xml_artifact }}
|
||||
@@ -113,7 +113,7 @@ jobs:
|
||||
investigate: true
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: inputs.coverage_json_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.coverage_json_artifact }}
|
||||
@@ -129,7 +129,7 @@ jobs:
|
||||
sphinx-build -v -n -b html -d _build/doctrees -j $(nproc) -w _build/html.log . _build/html
|
||||
|
||||
- name: 📤 Upload 'HTML Documentation' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.html_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -145,7 +145,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -164,7 +164,7 @@ jobs:
|
||||
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.unittest_xml_artifact }}' from 'Unittesting' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: inputs.unittest_xml_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.unittest_xml_artifact }}
|
||||
@@ -172,7 +172,7 @@ jobs:
|
||||
investigate: true
|
||||
|
||||
- name: 📥 Download artifacts '${{ inputs.coverage_json_artifact }}' from 'PublishCoverageResults' job
|
||||
uses: pyTooling/download-artifact@v6
|
||||
uses: pyTooling/download-artifact@v7
|
||||
if: inputs.coverage_json_artifact != ''
|
||||
with:
|
||||
name: ${{ inputs.coverage_json_artifact }}
|
||||
@@ -272,7 +272,7 @@ jobs:
|
||||
done
|
||||
|
||||
- name: 📤 Upload 'LaTeX Documentation' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.latex_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
|
||||
18
.github/workflows/StaticTypeCheck.yml
vendored
18
.github/workflows/StaticTypeCheck.yml
vendored
@@ -38,7 +38,7 @@ on:
|
||||
requirements:
|
||||
description: 'Python dependencies to be installed through pip.'
|
||||
required: false
|
||||
default: '-r tests/requirements.txt'
|
||||
default: '-r tests/typing/requirements.txt'
|
||||
type: string
|
||||
mypy_options:
|
||||
description: 'Additional mypy options.'
|
||||
@@ -49,18 +49,18 @@ on:
|
||||
description: 'Cobertura file to upload as an artifact.'
|
||||
required: false
|
||||
default: >-
|
||||
{ "fullpath": "report/typing/cobertura.xml",
|
||||
{ "fullpath": "report/typing/cobertura.xml",
|
||||
"directory": "report/typing",
|
||||
"filename": "cobertura.xml"
|
||||
"filename": "cobertura.xml"
|
||||
}
|
||||
type: string
|
||||
junit_report:
|
||||
description: 'JUnit file to upload as an artifact.'
|
||||
required: false
|
||||
default: >-
|
||||
{ "fullpath": "report/typing/StaticTypingSummary.xml",
|
||||
{ "fullpath": "report/typing/StaticTypingSummary.xml",
|
||||
"directory": "report/typing",
|
||||
"filename": "StaticTypingSummary.xml"
|
||||
"filename": "StaticTypingSummary.xml"
|
||||
}
|
||||
type: string
|
||||
html_report:
|
||||
@@ -94,7 +94,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 🐍 Setup Python ${{ inputs.python_version }}
|
||||
uses: actions/setup-python@v6
|
||||
@@ -142,7 +142,7 @@ jobs:
|
||||
fi
|
||||
|
||||
- name: 📤 Upload '${{ inputs.html_artifact }}' HTML artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: ${{ inputs.html_artifact != '' }}
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -153,7 +153,7 @@ jobs:
|
||||
retention-days: 1
|
||||
|
||||
- name: 📤 Upload '${{ inputs.junit_artifact }}' JUnit artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: ${{ inputs.junit_artifact != '' }}
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -164,7 +164,7 @@ jobs:
|
||||
retention-days: 1
|
||||
|
||||
- name: 📤 Upload '${{ inputs.cobertura_artifact }}' Cobertura artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: ${{ inputs.cobertura_artifact != '' }}
|
||||
continue-on-error: true
|
||||
with:
|
||||
|
||||
61
.github/workflows/UnitTesting.yml
vendored
61
.github/workflows/UnitTesting.yml
vendored
@@ -47,7 +47,7 @@ on:
|
||||
requirements:
|
||||
description: 'Python dependencies to be installed through pip.'
|
||||
required: false
|
||||
default: '-r tests/requirements.txt'
|
||||
default: '-r ./requirements.txt'
|
||||
type: string
|
||||
mingw_requirements:
|
||||
description: 'Override Python dependencies to be installed through pip on MSYS2 (MINGW64) only.'
|
||||
@@ -82,7 +82,7 @@ on:
|
||||
root_directory:
|
||||
description: 'Working directory for running tests.'
|
||||
required: false
|
||||
default: ''
|
||||
default: '.'
|
||||
type: string
|
||||
tests_directory:
|
||||
description: 'Path to the directory containing tests (relative from root_directory).'
|
||||
@@ -181,7 +181,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
with:
|
||||
lfs: true
|
||||
submodules: true
|
||||
@@ -205,6 +205,38 @@ jobs:
|
||||
run: |
|
||||
py -3.9 -m pip install --disable-pip-version-check --break-system-packages -U tomli
|
||||
|
||||
- name: Compute path to requirements file
|
||||
id: requirements
|
||||
shell: python
|
||||
run: |
|
||||
from os import getenv
|
||||
from pathlib import Path
|
||||
from sys import version
|
||||
|
||||
print(f"Python: {version}")
|
||||
|
||||
requirements = "${{ inputs.requirements }}"
|
||||
if requirements.startswith("-r"):
|
||||
requirements = requirements[2:].lstrip()
|
||||
if requirements.startswith("./"):
|
||||
requirementsFile = Path("${{ inputs.root_directory || '.' }}") / Path("${{ inputs.tests_directory || '.' }}") / Path("${{ inputs.unittest_directory || '.' }}") / Path(requirements[2:])
|
||||
else:
|
||||
requirementsFile = Path(requirements)
|
||||
|
||||
if not requirementsFile.exists():
|
||||
print(f"::error title=FileNotFoundError::{ex}")
|
||||
exit(1)
|
||||
|
||||
print(f"requirements file: {requirementsFile.as_posix()}")
|
||||
|
||||
# Write requirements path to special file
|
||||
github_output = Path(getenv("GITHUB_OUTPUT"))
|
||||
print(f"GITHUB_OUTPUT: {github_output}")
|
||||
with github_output.open("a+") as f:
|
||||
f.write(f"requirements=-r {requirementsFile.as_posix()}\n")
|
||||
else:
|
||||
print(f"requirements list: {requirements}")
|
||||
|
||||
- name: Compute pacman/pacboy packages
|
||||
id: pacboy
|
||||
if: matrix.system == 'msys2'
|
||||
@@ -215,8 +247,6 @@ jobs:
|
||||
from re import compile
|
||||
from sys import version
|
||||
|
||||
print(f"Python: {version}")
|
||||
|
||||
def loadRequirementsFile(requirementsFile: Path):
|
||||
requirements = []
|
||||
with requirementsFile.open("r") as file:
|
||||
@@ -232,11 +262,10 @@ jobs:
|
||||
|
||||
return requirements
|
||||
|
||||
requirements = "${{ inputs.requirements }}"
|
||||
requirements = "${{ steps.requirements.outputs.requirements }}"
|
||||
if requirements.startswith("-r"):
|
||||
requirementsFile = Path(requirements[2:].lstrip())
|
||||
try:
|
||||
dependencies = loadRequirementsFile(requirementsFile)
|
||||
dependencies = loadRequirementsFile(Path(requirements[2:].lstrip()))
|
||||
except FileNotFoundError as ex:
|
||||
print(f"::error title=FileNotFoundError::{ex}")
|
||||
exit(1)
|
||||
@@ -324,7 +353,7 @@ jobs:
|
||||
if: matrix.system != 'msys2'
|
||||
run: |
|
||||
python -m pip install --disable-pip-version-check -U wheel tomli
|
||||
python -m pip install --disable-pip-version-check ${{ inputs.requirements }}
|
||||
python -m pip install --disable-pip-version-check ${{ steps.requirements.outputs.requirements }}
|
||||
|
||||
- name: 🔧 Install pip dependencies (MSYS2)
|
||||
if: matrix.system == 'msys2'
|
||||
@@ -332,7 +361,7 @@ jobs:
|
||||
if [ -n '${{ inputs.mingw_requirements }}' ]; then
|
||||
python -m pip install --disable-pip-version-check --break-system-packages ${{ inputs.mingw_requirements }}
|
||||
else
|
||||
python -m pip install --disable-pip-version-check --break-system-packages ${{ inputs.requirements }}
|
||||
python -m pip install --disable-pip-version-check --break-system-packages ${{ steps.requirements.outputs.requirements }}
|
||||
fi
|
||||
|
||||
# Before scripts
|
||||
@@ -421,7 +450,7 @@ jobs:
|
||||
# Upload artifacts
|
||||
|
||||
- name: 📤 Upload '${{ fromJson(inputs.unittest_report_xml).filename }}' artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
if: inputs.unittest_xml_artifact != ''
|
||||
continue-on-error: true
|
||||
with:
|
||||
@@ -434,7 +463,7 @@ jobs:
|
||||
# - name: 📤 Upload 'Unit Tests HTML Report' artifact
|
||||
# if: inputs.unittest_html_artifact != ''
|
||||
# continue-on-error: true
|
||||
# uses: pyTooling/upload-artifact@v5
|
||||
# uses: pyTooling/upload-artifact@v6
|
||||
# with:
|
||||
# name: ${{ inputs.unittest_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
# path: ${{ inputs.unittest_report_html_directory }}
|
||||
@@ -444,7 +473,7 @@ jobs:
|
||||
- name: 📤 Upload 'Coverage SQLite Database' artifact
|
||||
if: inputs.coverage_sqlite_artifact != ''
|
||||
continue-on-error: true
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.coverage_sqlite_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
path: .coverage
|
||||
@@ -455,7 +484,7 @@ jobs:
|
||||
- name: 📤 Upload 'Coverage XML Report' artifact
|
||||
if: inputs.coverage_xml_artifact != '' && steps.convert_xml.outcome == 'success'
|
||||
continue-on-error: true
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.coverage_xml_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
working-directory: ${{ fromJson(inputs.coverage_report_xml).directory }}
|
||||
@@ -466,7 +495,7 @@ jobs:
|
||||
- name: 📤 Upload 'Coverage JSON Report' artifact
|
||||
if: inputs.coverage_json_artifact != '' && steps.convert_json.outcome == 'success'
|
||||
continue-on-error: true
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.coverage_json_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
working-directory: ${{ fromJson(inputs.coverage_report_json).directory }}
|
||||
@@ -477,7 +506,7 @@ jobs:
|
||||
- name: 📤 Upload 'Coverage HTML Report' artifact
|
||||
if: inputs.coverage_html_artifact != '' && steps.convert_html.outcome == 'success'
|
||||
continue-on-error: true
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ inputs.coverage_html_artifact }}-${{ matrix.system }}-${{ matrix.runtime }}-${{ matrix.python }}
|
||||
working-directory: ${{ fromJson(inputs.coverage_report_html).directory }}
|
||||
|
||||
2
.github/workflows/VerifyDocs.yml
vendored
2
.github/workflows/VerifyDocs.yml
vendored
@@ -44,7 +44,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: ⏬ Checkout repository
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: 🐍 Setup Python
|
||||
uses: actions/setup-python@v6
|
||||
|
||||
@@ -25,7 +25,7 @@ jobs:
|
||||
run: printf "%s\n" "${{ matrix.runs-on }}-${{ matrix.python }}" >> artifact.txt
|
||||
|
||||
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ fromJson(needs.Params.outputs.artifact_names).unittesting_xml }}-${{ matrix.system }}-${{ matrix.python }}
|
||||
path: artifact.txt
|
||||
@@ -42,7 +42,7 @@ jobs:
|
||||
run: printf "%s\n" "Package" >> package.txt
|
||||
|
||||
- name: 📤 Upload artifact for ${{ matrix.system }}-${{ matrix.python }}
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: ${{ fromJson(needs.Params.outputs.artifact_names).package_all }}
|
||||
path: package.txt
|
||||
|
||||
@@ -16,10 +16,9 @@ jobs:
|
||||
include:
|
||||
- {icon: '🐧', name: 'Ubuntu 22.04 (x86-64)', image: 'ubuntu-22.04', shell: 'bash', can-fail: false}
|
||||
- {icon: '🐧', name: 'Ubuntu 24.04 (x86-64)', image: 'ubuntu-24.04', shell: 'bash', can-fail: false} # latest
|
||||
- {icon: '🍎', name: 'macOS-13 (x86-64)', image: 'macos-13', shell: 'bash', can-fail: false}
|
||||
- {icon: '🍎', name: 'macOS-14 (x86-64)', image: 'macos-14-large', shell: 'bash', can-fail: true } # not in free plan
|
||||
- {icon: '🍎', name: 'macOS-15 (x86-64)', image: 'macos-15-large', shell: 'bash', can-fail: true } # not in free plan
|
||||
- {icon: '🍏', name: 'macOS-13 (aarch64)', image: 'macos-13-xlarge', shell: 'bash', can-fail: true } # not in free plan
|
||||
### - {icon: '🍎', name: 'macOS-15 (x86-64)', image: 'macos-15-large', shell: 'bash', can-fail: true } # same as -intel; not in free plan
|
||||
- {icon: '🍎', name: 'macOS-15 (x86-64)', image: 'macos-15-intel', shell: 'bash', can-fail: false}
|
||||
- {icon: '🍏', name: 'macOS-14 (aarch64)', image: 'macos-14', shell: 'bash', can-fail: false} # latest
|
||||
- {icon: '🍏', name: 'macOS-15 (aarch64)', image: 'macos-15', shell: 'bash', can-fail: false}
|
||||
- {icon: '🪟', name: 'Windows Server 2022', image: 'windows-2022', shell: 'bash', can-fail: false}
|
||||
|
||||
15
.github/workflows/_Checking_Nightly.yml
vendored
15
.github/workflows/_Checking_Nightly.yml
vendored
@@ -17,7 +17,7 @@ jobs:
|
||||
printf "%s\n" "Build log $(date --utc '+%d.%m.%Y - %H:%M:%S')" > build.log
|
||||
|
||||
- name: 📤 Upload artifact
|
||||
uses: pyTooling/upload-artifact@v5
|
||||
uses: pyTooling/upload-artifact@v6
|
||||
with:
|
||||
name: document
|
||||
path: |
|
||||
@@ -29,10 +29,11 @@ jobs:
|
||||
- name: 🖉 Program
|
||||
run: |
|
||||
printf "%s\n" "Document other $(date --utc '+%d.%m.%Y - %H:%M:%S')" > document1.txt
|
||||
printf "%s\n" "Document other $(date --utc '+%d.%m.%Y - %H:%M:%S')" > document2.txt
|
||||
printf "%s\n" "Program $(date --utc '+%d.%m.%Y - %H:%M:%S')" > program.py
|
||||
|
||||
- name: 📤 Upload artifact
|
||||
uses: actions/upload-artifact@v5
|
||||
uses: actions/upload-artifact@v6
|
||||
with:
|
||||
name: other
|
||||
path: |
|
||||
@@ -55,7 +56,7 @@ jobs:
|
||||
version=4.2.0
|
||||
tool=myTool
|
||||
prog=program
|
||||
tag: 4.2.0
|
||||
tag: v4.2.0
|
||||
title: "Nightly Test Release"
|
||||
description: |
|
||||
This *nightly* release contains all latest and important artifacts created by %tool%'s CI pipeline.
|
||||
@@ -63,10 +64,14 @@ jobs:
|
||||
# %tool% %version%
|
||||
|
||||
* %prog%
|
||||
|
||||
# Attached files:
|
||||
|
||||
%%ASSETS%%
|
||||
assets: |
|
||||
document: document1.txt: Documentation
|
||||
document: build.log: Logfile - %tool% - %tool%
|
||||
other: document1.txt: SBOM - %version%
|
||||
other: document2.txt: SBOM - %version%
|
||||
other: %prog%.py: Application - %tool% - %version%
|
||||
document:!archive1.zip: Archive 1 - zip
|
||||
document:!archive2.tgz: Archive 2 - tgz
|
||||
@@ -108,7 +113,7 @@ jobs:
|
||||
# artifact: file: labels: asset title
|
||||
document: document1.txt: doc,html: Documentation
|
||||
document: build.log: build,log: Logfile - %tool% - %tool%
|
||||
other: document1.txt: build,SBOM:SBOM - %version%
|
||||
other: document2.txt: build,SBOM:SBOM - %version%
|
||||
other: %prog%.py: app,binary:Application - %tool% - %version%
|
||||
document:!archive1.zip: Archive 1 - zip
|
||||
document:!archive2.tgz: Archive 2 - tgz
|
||||
|
||||
14
.github/workflows/_Checking_Parameters.yml
vendored
14
.github/workflows/_Checking_Parameters.yml
vendored
@@ -64,7 +64,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_Default'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -92,7 +92,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_PythonVersions'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -114,7 +114,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_Systems'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -136,7 +136,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_Include'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -158,7 +158,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_Exclude'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -180,7 +180,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_Disable'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
@@ -202,7 +202,7 @@ jobs:
|
||||
shell: python
|
||||
steps:
|
||||
- name: Checkout repository to access local Action
|
||||
uses: actions/checkout@v5
|
||||
uses: actions/checkout@v6
|
||||
|
||||
- name: Checking job matrix from 'Params_All'
|
||||
uses: ./.github/actions/CheckJobMatrix
|
||||
|
||||
Reference in New Issue
Block a user