Skip to content

1. On Pull Request

1. On Pull Request #6

name: 1. On Pull Request
# This is the primary workflow to process pathway edits coming in as GPML files from
# PathVisio. A new pull request should be create per pathway. Subsequent edits to the
# same GPML can accumulate in the same pull request while in review.
#
# The outputs of this workflow will be temporarily stored as PR artifacts to allow access
# by subsequent jobs and to allow parallelization. At the end of the worflow, the outputs
# will be pushed to the main website (wikipathways.github.io repo) for transparent review.
#
# The acceptance/merge of a processed pull request will trigger a secondary workflow to
# migrate output files to published folders in main website repo, as well as the assets
# and homology repos.
#
# Inputs: Each pull request should have one and only one GPML file attached.
#
# Outputs:
# - info.json
# - datanodes.tsv
# - bibliography.tsv (via refs.tsv)
# - .json .svg and .png files
# - .md file
# - TODO: test results?
#
on:
workflow_dispatch:
inputs:
manual-pr-number:
description: 'PR Number'
required: true
default: '10'
#pull_request:
#paths:
# - '**/*.gpml'
concurrency:
group: ${{ github.workflow }}-${{ github.head_ref || github.ref }}
cancel-in-progress: false # to allow multiple runs to queue up rather than clobber
jobs:
get-pr:
# This job determines the PR number and stores it as a GITHUB_OUTPUT environment variable
# that can accessed by subsequent jobs.
runs-on: ubuntu-latest
outputs:
pr-number: ${{ steps.get-pr.outputs.pr-number }}
steps:
- name: Get PR
id: get-pr
# Get PR from triggering event, unless provided via manual workflow run
run: |
if [ -z "${{ inputs.manual-pr-number }}" ]; then
echo "pr-number=${{ github.event.pull_request.number }}" >> $GITHUB_OUTPUT
else
echo "pr-number=${{ inputs.manual-pr-number }}" >> $GITHUB_OUTPUT
fi
get-gpml:
# This job gets the GPML file, renames it (if new), and uploads it as an artifact
# that can accessed by subsequent jobs. The GPML filename is also stored as a
# GITHUB_OUTPUT environment variable. The job also extracts key information from
# the GPML file and starts a new PR description to let reviewers know the processing status.
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
#
# The branch name from the submitter's fork is also extracted and stored (if needed).
needs: [get-pr]
#if: ${{ needs.get-pr.outputs.pr-number }}
runs-on: ubuntu-latest
outputs:
gpml-filepath: ${{ steps.get-gpml.outputs.gpml-filepath }}
gpml-file: ${{ steps.get-gpml.outputs.gpml-file }}
pr-desc: ${{ steps.get-gpml.outputs.pr-desc }}
branch-name: ${{ steps.get-branch.outputs.branch-name }}
env:
PR_NUMBER: ${{ needs.get-pr.outputs.pr-number }}
steps:
- name: Checkout repository
uses: actions/checkout@v4
with:
ref: refs/pull/${{ needs.get-pr.outputs.pr-number }}/head
- name: Install XML processing tools
# Used to parse XML values from GPML
run: sudo apt-get install -y libxml2-utils
- name: Get GPML
id: get-gpml
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
run: |
GPML_FILEPATH=$(gh pr view $PR_NUMBER --json files --jq '.files.[].path' | grep '.gpml$')
echo "Found GPML file at: $GPML_FILEPATH"
GPML_FILE="$(basename ""$GPML_FILEPATH"")"
# Rename if new pathway
first_digit=${GPML_FILE:2:1}
if [ "$first_digit" == "0" ]; then
echo "Temporarily renaming new pathway per PR number."
GPML_FILE="WP0${PR_NUMBER}.gpml"
fi
cp "$GPML_FILEPATH" ./"$GPML_FILE"
GPML_FILEPATH="./${GPML_FILE}"
echo "gpml-file=$GPML_FILE" >> $GITHUB_OUTPUT
echo "gpml-filepath=$GPML_FILEPATH" >> $GITHUB_OUTPUT #for subsequent jobs
# Extract information from GPML
wpid="$(echo $GPML_FILE | sed 's/.gpml//')"
org="$(xmllint --xpath 'string(//*[local-name()="Pathway"]/@Organism)' "$GPML_FILEPATH")"
name="$(xmllint --xpath 'string(//*[local-name()="Pathway"]/@Name)' "$GPML_FILEPATH")"
desc="$(xmllint --xpath 'string(//*[local-name()="Comment" and @Source="WikiPathways-description"])' "$GPML_FILEPATH")"
# Start PR description (newlines are important in NEW_DESCRIPTION string)
NEW_DESCRIPTION="
## Pathway Information
**WPID**: $wpid
**TITLE**: $name
**ORGANISM**: $org
**DESCRIPTION**: $desc
---
"
# Store PR description for compilation in final job
# EOF base64 needed for storing multi-line strings as output variables
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
# Temporarily append PR description
NEW_DESCRIPTION="$NEW_DESCRIPTION
Processing...
"
# Edit PR description
gh pr edit $PR_NUMBER --body "$NEW_DESCRIPTION"
- name: Upload GPML file as artifact
uses: actions/upload-artifact@v3
with:
name: gpml-file
path: ${{ steps.get-gpml.outputs.gpml-filepath }}
retention-days: 1
if-no-files-found: error # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Get branch name
id: get-branch
# Determine PR branch from submitter's fork, if needed #TODO: prune if not needed
run: |
BRANCH_NAME=$(curl -s -H "Authorization: token ${{ secrets.GITHUB_TOKEN }}" \
https://api.github.com/repos/${{ github.repository }}/pulls/${{needs.get-pr.outputs.pr-number}} \
| jq -r .head.ref)
echo "branch-name=$BRANCH_NAME"
echo "branch-name=$BRANCH_NAME" >> $GITHUB_OUTPUT
if ! git ls-remote --heads origin $BRANCH_NAME; then
git push origin HEAD:refs/heads/$BRANCH_NAME
fi
metadata:
# This job generates files dervied from the GPML, including info.json, datanodes.tsv and refs.tsv.
# The meta-data-action Java application performs the bulk of the work:
# https://github.com/wikipathways/meta-data-action
#
# The generated files are uploaded as artifacts to be downloaded and pushed together with other files
# to another repo in a final sync job. This strategy allows for parallel job processing!
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
needs: [get-pr,get-gpml]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
# TODO: impl more efficient way to get scripts
uses: actions/checkout@v4
- name: Download GPML artifact
# Downloads gpml-file artifact to working dir
# Note: Automatically unzips as folders containing files with original filenames
uses: actions/download-artifact@v3
with:
name: gpml-file
- name: Setup Java
uses: actions/setup-java@v3
with:
distribution: 'temurin'
java-version: '11'
- name: Cache meta-data-action with dependencies
uses: actions/cache@v3
id: cacheMetaJar
with:
path: ./meta-data-action-1.1.2-jar-with-dependencies.jar
key: cached-meta-data-action-${{ hashFiles('meta-data-action-1.1.2-jar-with-dependencies.jar') }}
restore-keys: |
cached-meta-data-action-${{ hashFiles('meta-data-action-1.1.2-jar-with-dependencies.jar') }}
cached-meta-data-action-
- name: Install deps
run: |
echo "Refreshing cached-meta-data-action"
if [ ! -e ./meta-data-action-1.1.2-jar-with-dependencies.jar ]; then
wget -O meta-data-action-1.1.2-jar-with-dependencies.jar https://github.com/wikipathways/meta-data-action/releases/download/v1.1.2/meta-data-action-1.1.2-jar-with-dependencies.jar
fi
- name: Cache and install dependencies
uses: actions/cache@v3
id: cache
with:
path: ${{ github.workspace }}/Hs_Derby_Ensembl_108.bridge
key: ${{ runner.os }}-java-Hs_Derby_Ensembl_108
restore-keys: |
${{ runner.os }}-java-Hs_Derby_Ensembl_108
${{ runner.os }}-java-Hs_Derby_Ensembl_
- if: steps.cache.outputs.cache-hit != 'true'
name: Install deps
run: |
cd "${{ github.workspace }}"
if [ ! -e ./Hs_Derby_Ensembl_108.bridge ]; then
wget -O Hs_Derby_Ensembl_108.bridge https://zenodo.org/record/7781913/files/Hs_Derby_Ensembl_108.bridge?download=1
fi
- name: Generate gdb.config, fileNames.config, and fileDownloads.config
run: scripts/meta-data-action/configGenerator.sh ./$GPML_FILE
- name: Cache all bridge files
uses: actions/cache@v3
id: cacheAllBridge
with:
path: |
./metabolites*.bridge
./Ag*.bridge
./An*.bridge
./At*.bridge
./Bs*.bridge
./Bt*.bridge
./Ce*.bridge
./Cf*.bridge
./Ci*.bridge
./Dr*.bridge
./Da*.bridge
./Dp*.bridge
./Dm*.bridge
./Ec*.bridge
./Gg*.bridge
./Fg*.bridge
./Gm*.bridge
./Hs*.bridge
./Hv*.bridge
./Ml*.bridge
./Mm*.bridge
./Mx*.bridge
./Oa*.bridge
./Ova*.bridge
./Oi*.bridge
./Oj*.bridge
./Pi*.bridge
./Pt*.bridge
./Qc*.bridge
./Rn*.bridge
./Sc*.bridge
./Sl*.bridge
./Ss*.bridge
./Vv*.bridge
./Xt*.bridge
./Zm*.bridge
key: cached-bridge-files
restore-keys: |
cached-bridge-files
- if: steps.cacheAllBridge.outputs.cache-hit != 'true'
name: Install all bridge files
run: |
echo "Cache not found: cached-bridge-files"
declare -a OrganismNames=("Metabolites" "Anopheles gambiae" "Aspergillus niger" "Arabidopsis thaliana" "Bacillus subtilis" "Bos taurus" "Caenorhabditis elegans" "Canis familiaris" "Ciona intestinalis" "Danio rerio" "Daphnia magna" "Daphnia pulex" "Drosophila melanogaster" "Escherichia coli" "Gallus gallus" "Fusarium graminearum" "Glycine max" "Homo sapiens" "Hordeum vulgare" "Macaca mulatta" "Mus musculus" "Mycobacterium tuberculosis" "Ornithorhynchus anatinus" "Ovis aries" "Oryza indica" "Oryza japonica" "Populus trichocarpa" "Pan troglodytes" "Equus caballus" "Rattus norvegicus" "Saccharomyces cerevisiae" "Solanum lycopersicum" "Sus scrofa" "Vitis vinifera" "Xenopus tropicalis" "Zea mays")
for org in "${OrganismNames[@]}"; do
echo "generating configuration files for "$org""
scripts/meta-data-action/configGenerator.sh "$org"
echo "installing bridgedb files for "$org""
scripts/meta-data-action/installDependencies.sh "$org"
done
- name: Generate configs, install bridgeDb, generate info and tsv files
run: |
chmod 777 meta-data-action-1.1.2-jar-with-dependencies.jar
f=./$GPML_FILE
org="$(sed -n '/<Pathway /s/.*Organism=\(.*\)[^\n]*/\1/p' $f | tr -d '"' | tr -d '>' | tr -d '\r')"
echo "generating configuration files for "$org""
scripts/meta-data-action/configGenerator.sh "$org"
wpid="$(echo $GPML_FILE | sed 's/.gpml//')"
cat gdb.config
echo "generating info and datanode files for $wpid, organism "$org""
java -jar meta-data-action-1.1.2-jar-with-dependencies.jar local "$f" $(date --utc +%F) gdb.config "$org"
# generated files are placed in "pathways" dir
ls -R
mkdir ./metadata
mv pathways/"$wpid"-info.json metadata/.
mv pathways/"$wpid"*.tsv metadata/.
- name: Upload metadata files as artifacts
uses: actions/upload-artifact@v3
with:
name: metadata
path: metadata/
retention-days: 1
if-no-files-found: error # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Report on meta-data-action
id: report
run: |
# Verify generated files
infojson=$(find . -path "./metadata/WP*info.json" -print -quit)
dntsv=$(find . -path "./metadata/WP*datanodes.tsv" -print -quit)
refstsv=$(find . -path "./metadata/WP*refs.tsv" -print -quit)
# Update PR description
NEW_DESCRIPTION="
## Generate Metadata Files
"
if [[ -n $infojson ]]; then
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [x] info.json generated"
else
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [] info.json generated"
fi
if [[ -n $dntsv ]]; then
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [x] datanodes.tsv generated"
else
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [] datanodes.tsv generated"
fi
if [[ -n $refstsv ]]; then
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [x] refs.tsv generated"
else
NEW_DESCRIPTION="$NEW_DESCRIPTION
- [] refs.tsv generated"
fi
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
authors:
# This job generates new author profiles if needed. It checks the list of authors from the GPML file
# against the list of prior authors in scripts/authors-list.csv and updates the list accordingly.
#
# An artifact is created and uploaded only if new author profiles are generated. A later job will
# check to see if an "authors" artifact is present and commit files at that time.
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
if: true
needs: [get-pr,get-gpml,metadata]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
uses: actions/checkout@v4
- name: Download GPML artifact
# Downloads gpml-file artifact to working dir
# Note: Automatically unzips file with original filename
uses: actions/download-artifact@v3
with:
name: gpml-file
- name: Update author list and create md profiles
run: |
# Read list of previous authors
authorList=()
{
read #skip header line
while IFS=, read -r username realname orcid wikidata github; do
authorList+=("$username")
done
}< scripts/author_list.csv
echo ${#authorList[@]}
# Generate unique list of authors from this GPML
uniqueAuthors=()
auth="$(sed -n '/<Pathway /s/.*Author=\"\[\(.*\)\]\".*/\1/p' $GPML_FILE )"
IFS=',' read -r -a curAuthors <<< "$auth"
echo $curAuthors
echo ${#curAuthors[@]}
for i in "${curAuthors[@]}"; do
skip=
for j in "${uniqueAuthors[@]}"; do
[[ $i == $j ]] && { skip=1; break; }
done
[[ -n $skip ]] || uniqueAuthors+=("$i")
done
echo ${uniqueAuthors[@]}
echo ${#uniqueAuthors[@]}
# For any new authors, generate a profile and add them to the list
mkdir ./authors
k=0
for a in "${uniqueAuthors[@]}"; do
a=$(echo "$a" | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//')
a1=${a//AAR&Co/AARandCo}
a1=${a//Önder/Onder}
a2=$(echo "$a" | tr ' ' '_')
if [[ ! " ${authorList[*]} " =~ " ${a} " ]]; then
echo "Adding $a"
echo $a","$a",,," >> scripts/author_list.csv
echo "---" > "authors/$a1.md"
echo "username: $a1" >> "authors/$a1.md"
echo "realname: $a" >> "authors/$a1.md"
echo "website: " >> "authors/$a1.md"
echo "affiliation: " >> "authors/$a1.md"
echo "bio: " >> "authors/$a1.md"
echo "github: " >> "authors/$a1.md"
echo "orcid: " >> "authors/$a1.md"
echo "linkedin: " >> "authors/$a1.md"
echo "googlescholar: " >> "authors/$a1.md"
echo "wikidata: " >> "authors/$a1.md"
echo "twitter: " >> "authors/$a1.md"
echo "mastodon-url: " >> "authors/$a1.md"
echo "meta:" >> "authors/$a1.md"
echo "instagram:" >> "authors/$a1.md"
echo "email:" >> "authors/$a1.md"
echo "redirect_from:" >> "authors/$a1.md"
echo "- /index.php/User:$a2" >> "authors/$a1.md"
echo "- /index.php/Special:Contributions/$a2" >> "authors/$a1.md"
echo "---" >> "authors/$a1.md"
$k=$k + 1
fi
done
if [ $k -gt 0 ]; then
mv scripts/author_list.csv authors/.
fi
echo "NEW_AUTHORS=$k" >> $GITHUB_ENV
- name: Upload author files as artifacts
uses: actions/upload-artifact@v3
with:
name: authors
path: authors/
retention-days: 1
if-no-files-found: ignore # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Report on authors
id: report
run: |
# Update PR description
if [ ${{ env.NEW_AUTHORS }} -eq 0 ]; then
NEW_DESCRIPTION="- [x] No new author profiles needed"
elif [ ${{ env.NEW_AUTHORS }} -eq 1 ]; then
NEW_DESCRIPTION="- [x] 1 author profile generated"
else
NEW_DESCRIPTION="- [x] ${{ env.NEW_AUTHORS }} author profiles generated"
fi
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
pubmed:
# This job generates a bibliography.tsv file based on the refs.tsv generated by the metadata job.
#
# A "pubmed" artifact is created and uploaded containing this new tsv file. A later job will
# commit the file to appropriate draft folders in the website repo.
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
if: true
needs: [get-pr,get-gpml,metadata]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
uses: actions/checkout@v4
- name: Download Metadata artifact
# Downloads metadata artifact to working dir
# Note: Automatically unzips as a folder containing files with original filenames
uses: actions/download-artifact@v3
with:
name: metadata
- name: Setup Node
uses: actions/setup-node@v3
with:
node-version: 16
- name: Cache dependencies
uses: actions/cache@v3
with:
path: ~/.npm
key: ${{ runner.os }}-node-${{ hashFiles('**/package-lock.json') }}
restore-keys: |
${{ runner.os }}-node-
- name: Install dependencies
working-directory: scripts/generate-references
run: npm install
- name: Generate references
working-directory: scripts/generate-references
run: |
mkdir -p ../../pubmed
cp ../../*-refs.tsv ../../pubmed/.
node index.js
# where are generated files placed
rm ../../*-refs.tsv
- name: Upload pubmed file as artifact
uses: actions/upload-artifact@v3
with:
name: pubmed
path: pubmed/
retention-days: 1
if-no-files-found: error # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Report on pubmed
id: report
run: |
# Update PR description
NEW_DESCRIPTION="- [x] bibliography.tsv generated"
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
frontmatter:
# This job generates a wpid.md file based on the wpid-info.json file generated by the metadata job.
#
# A "frontmatter" artifact is created and uploaded containing this .md file. A later job will
# commit the file to the appropriate draft folder in the website repo.
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
if: true
needs: [get-pr,get-gpml,metadata]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
uses: actions/checkout@v4
- name: Download Metadata artifact
# Downloads metadata artifacts to working dir
# Note: Automatically unzips as folders containing files with original filenames
uses: actions/download-artifact@v3
with:
name: metadata
- name: Setup Python
uses: actions/setup-python@v4
with:
python-version: '3.x'
- name: Install deps
run: |
pip install python-frontmatter
- name: Run create_pathway_frontmatter.py
run: |
wpid="$(echo $GPML_FILE | sed 's/.gpml//')"
echo "generating frontmatter file for $wpid"
json_info_f=./"$wpid"-info.json
echo "*****input for python script : $json_info_f"
if [ -e "$json_info_f" ]; then
python scripts/create_pathway_frontmatter.py "$json_info_f"
sed -i 's/&amp;/and/g' ./"$wpid".md
sed -i 's/AAR&amp;Co/AARandCo/g' ./"$wpid".md
sed -i 's/&Ouml;nder/Onder/g' ./"$wpid".md
else
echo "info.json file missing for $wpid" >2
fi
#move the generated file to the frontmatter folder
mkdir ./frontmatter
mv ./"$wpid".md frontmatter/.
- name: Upload frontmatter file as artifact
uses: actions/upload-artifact@v3
with:
name: frontmatter
path: frontmatter/
retention-days: 1
if-no-files-found: error # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Report on frontmatter
id: report
run: |
# Update PR description
NEW_DESCRIPTION="- [x] .md generated"
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
json-svg:
# This job generates json, svg and png files based on the GPML file.
#
# A "json-svg" artifact is created and uploaded containing the JSON, SVG and PNG files. A later job will
# commit the file to the appropriate draft folder in the website repo.
#
# A markdown text report is stored as pr-desc to be compiled and written to the PR description.
if: true
needs: [get-pr,get-gpml,metadata]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
uses: actions/checkout@v4
- name: Download GPML and Metadata artifacts
# Downloads gpml-file and metadata artifacts to working dir
# Note: Automatically unzips as folders containing files with original filenames
uses: actions/download-artifact@v3
- name: Setup Node
uses: actions/setup-node@v3
with:
node-version: 16
- name: Cache NPM dependencies
uses: actions/cache@v3
with:
path: ~/.npm
key: ${{ runner.os }}-node-${{ hashFiles('**/package-lock.json') }}
restore-keys: |
${{ runner.os }}-node-
- name: Install deps
working-directory: scripts/generate-svgs
run: |
npm install
sudo apt-get update
sudo apt-get install -y xmlstarlet
- name: Run generate-svgs to convert GPML to JSON and SVG
run: |
echo "SANDBOX: Adapted to produce JSON and SVG files in sandbox-wp-db repo only"
DB_PATH="${{ github.workspace }}"
wpid="$(echo $GPML_FILE | sed 's/.gpml//')"
echo "generating JSON and SVG files for $wpid"
cd "$DB_PATH"
mkdir -p pathways/"$wpid"
for old_f in pathways/"$wpid"/"$wpid".{json,svg}; do
if [ -e "$old_f" ]; then
rm "$old_f"
fi
done
cd "scripts/generate-svgs"
./gpmlconverter --id "$wpid" -i $DB_PATH/gpml-file/"$wpid".gpml -o $DB_PATH/pathways/"$wpid"/"$wpid".svg
# delete intermediate JSON files
rm $DB_PATH/pathways/"$wpid"/"$wpid".json.b4bridgedb.json || true
rm $DB_PATH/pathways/"$wpid"/"$wpid".b4wd.json || true
rm $DB_PATH/pathways/"$wpid"/"$wpid".b4hgnc.json || true
# pretty print the JSON
for json_f in $DB_PATH/pathways/"$wpid"/"$wpid".json; do
mv "$json_f" "$json_f".tmp.json
jq -S . "$json_f".tmp.json >"$json_f"
rm "$json_f".tmp.json
done
#move the generated files to the json-svg folder
mkdir ../../json-svg
mv $DB_PATH/pathways/"$wpid"/"$wpid".json ../../json-svg/.
mv $DB_PATH/pathways/"$wpid"/"$wpid".png ../../json-svg/.
mv $DB_PATH/pathways/"$wpid"/"$wpid".svg ../../json-svg/.
mv $DB_PATH/pathways/"$wpid"/"$wpid"-thumb.png ../../json-svg/.
- name: Upload json-svg file as artifact
uses: actions/upload-artifact@v3
with:
name: json-svg
path: json-svg/
retention-days: 1
if-no-files-found: error # 'warn' or 'ignore' are also available, defaults to `warn`
- name: Report on json-svg
id: report
run: |
# Update PR description
NEW_DESCRIPTION="- [x] .json generated
- [x] .svg generated
- [x] .png generated
*TODO: insert PNG here*
"
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
testing:
if: true
needs: [get-pr,get-gpml,metadata]
runs-on: ubuntu-latest
outputs:
pr-desc: ${{ steps.report.outputs.pr-desc }}
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
steps:
- name: Checkout repository for scripts
uses: actions/checkout@v4
- name: Download GPML and Metadata artifacts
# Downloads gpml-file and metadata artifacts to working dir
# Note: Automatically unzips as folders containing files with original filenames
uses: actions/download-artifact@v3
- name: Perform automatic test
run: |
# TODO: perform testing
- name: Report on pubmed
id: report
run: |
# Update PR description
NEW_DESCRIPTION='
---
## Automated Testing
*As table*
| # | Test name | Link | Notes | Result |
|---|---|---|---|---|
| 1 | Test one | | | \$\${\color{green}PASS}\$\$ |
| 2 | Test two | | | \$\${\color{red}FAIL}\$\$ |
| 3 | Test three | | | \$\${\color{green}PASS}\$\$ |
*As checklist*
- [ ] Interactions are connected
- [ ] Datanodes are annotated with database references
- [ ] Decription, consisting of 2-3 sentence overview of processes described in the pathway
- [ ] At least one literature reference
- [ ] At least one pathway ontology term
- [ ] Pathway title conforms to the [guidelines](https://github.com/wikipathways/wikipathways-faq/discussions/24)
---
'
# Store PR description for compilation in final job
EOF=$(dd if=/dev/urandom bs=15 count=1 status=none | base64)
echo "pr-desc<<$EOF" >> $GITHUB_OUTPUT
echo "$NEW_DESCRIPTION" >> $GITHUB_OUTPUT
echo "$EOF" >> $GITHUB_OUTPUT
commit-outputs:
needs: [get-pr,get-gpml,metadata,authors,pubmed,frontmatter,json-svg,testing]
runs-on: ubuntu-latest
steps:
- name: Download all output files from artifacts
uses: actions/download-artifact@v3
- name: Checkout jekyll repo
# TESTING: change repo and ssh-key from sandbox to real
uses: actions/checkout@v3
with:
repository: wikipathways/sandbox-wp.gh.io
path: wikipathways.github.io
fetch-depth: 1
ref: main
ssh-key: ${{ secrets.ACTIONS_SANDBOX_DEPLOY_KEY }}
- name: Commit report
env:
GPML_FILE: ${{ needs.get-gpml.outputs.gpml-file }}
PR_NUMBER: ${{ needs.get-pr.outputs.pr-number }}
run: |
wpid="$(echo $GPML_FILE | sed 's/.gpml//')"
## Copy to draft_assets
mkdir -p wikipathways.github.io/draft_assets/"$wpid"
cp gpml-file/"$wpid".gpml wikipathways.github.io/draft_assets/"$wpid"/.
cp metadata/"$wpid"*.* wikipathways.github.io/draft_assets/"$wpid"/.
cp pubmed/"$wpid"-bibliography.tsv wikipathways.github.io/draft_assets/"$wpid"/.
cp json-svg/"$wpid".json wikipathways.github.io/draft_assets/"$wpid"/.
cp json-svg/"$wpid".png wikipathways.github.io/draft_assets/"$wpid"/.
cp json-svg/"$wpid".svg wikipathways.github.io/draft_assets/"$wpid"/.
cp json-svg/"$wpid"-thumb.png wikipathways.github.io/draft_assets/"$wpid"/.
## Copy to _data/drafts
cp metadata/"$wpid"-datanodes.tsv wikipathways.github.io/_data/drafts/.
cp pubmed/"$wpid"-bibliography.tsv wikipathways.github.io/_data/drafts/.
## Copy to _drafts
cp frontmatter/"$wpid".md wikipathways.github.io/_drafts/.
## Copy to _authors
if [ -d "authors" ]; then
cp authors/*.md wikipathways.github.io/_authors/.
cp author_list.csv wikipathways.github.io/scripts/.
fi
## Commit, pull and push
cd "${{ github.workspace }}/wikipathways.github.io"
git config --global user.name 'GitHub Action'
git config --global user.email '[email protected]'
git add .
if git diff --exit-code --staged; then
echo "No changes"
else
git commit -m 'Adds outputs from PR processing'
git pull --rebase
git push
fi
update-pr-desc:
needs: [get-pr,get-gpml,metadata,authors,pubmed,frontmatter,json-svg,testing]
runs-on: ubuntu-latest
steps:
- name: Check out code
uses: actions/checkout@v4
with:
ref: refs/pull/${{ needs.get-pr.outputs.pr-number }}/head
- name: Compile and update PR description
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
PR_NUMBER: ${{ needs.get-pr.outputs.pr-number }}
run: |
NEW_DESCRIPTION="${{ needs.get-gpml.outputs.pr-desc }}
${{ needs.metadata.outputs.pr-desc }}
${{ needs.authors.outputs.pr-desc }}
${{ needs.pubmed.outputs.pr-desc }}
${{ needs.frontmatter.outputs.pr-desc }}
${{ needs.json-svg.outputs.pr-desc }}
${{ needs.testing.outputs.pr-desc }}
"
# Edit PR description
gh pr edit $PR_NUMBER --body "$NEW_DESCRIPTION"