steps
262 rows
This data as json, CSV (advanced)
with 19 ✖
- {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} 12
- {"python-version": "${{ matrix.python-version }}"} 7
- {"path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n"} 5
- {"python-version": "3.9"} 4
- {"python-version": 3.9} 4
- {"python-version": "3.11"} 3
- {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "pyproject.toml"} 2
- {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "setup.py"} 2
- {"token": "${{ secrets.CODECOV_TOKEN }}", "file": "coverage.xml"} 2
- {"version": "318.0.0", "service_account_email": "${{ secrets.GCP_SA_EMAIL }}", "service_account_key": "${{ secrets.GCP_SA_KEY }}"} 2
- {"fetch-depth": 2} 1
- {"languages": "${{ matrix.language }}"} 1
- {"path": "~/.cache/ms-playwright/", "key": "${{ runner.os }}-browsers"} 1
- {"path": "~/.npm", "key": "${{ runner.OS }}-npm-${{ hashFiles('**/package-lock.json') }}", "restore-keys": "${{ runner.OS }}-npm-\n"} 1
- {"project-slug": "datasette"} 1
- {"project-slug": "sqlite-utils"} 1
- {"python-version": "3.10", "cache": "pip", "cache-dependency-path": "**/setup.py"} 1
- {"python-version": "3.11", "cache": "pip", "cache-dependency-path": "pyproject.toml"} 1
- {"python-version": "3.11", "cache": "pip", "cache-dependency-path": "setup.py"} 1
id ▼ | seq | job | repo | uses | name | with | run | env | if |
---|---|---|---|---|---|---|---|---|---|
136912 | 2 | deploy 27475 | healthkit-to-sqlite 197882382 | actions/setup-python@v2 | Set up Python | {"python-version": "3.9"} | |||
136913 | 3 | deploy 27475 | healthkit-to-sqlite 197882382 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n"} | |||
136914 | 4 | deploy 27475 | healthkit-to-sqlite 197882382 | Install dependencies | pip install setuptools wheel twine | ||||
136915 | 5 | deploy 27475 | healthkit-to-sqlite 197882382 | Publish | python setup.py sdist bdist_wheel twine upload dist/* | {"TWINE_USERNAME": "__token__", "TWINE_PASSWORD": "${{ secrets.PYPI_TOKEN }}"} | |||
136916 | 1 | test 27476 | healthkit-to-sqlite 197882382 | actions/checkout@v2 | |||||
136917 | 2 | test 27476 | healthkit-to-sqlite 197882382 | actions/setup-python@v2 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}"} | |||
136918 | 3 | test 27476 | healthkit-to-sqlite 197882382 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
136919 | 4 | test 27476 | healthkit-to-sqlite 197882382 | Install dependencies | pip install -e '.[test]' | ||||
136920 | 5 | test 27476 | healthkit-to-sqlite 197882382 | Run tests | pytest | ||||
136921 | 1 | test 27477 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
136922 | 2 | test 27477 | pocket-to-sqlite 213286752 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "pyproject.toml"} | |||
136923 | 3 | test 27477 | pocket-to-sqlite 213286752 | Install dependencies | pip install '.[test]' | ||||
136924 | 4 | test 27477 | pocket-to-sqlite 213286752 | Run tests | pytest | ||||
136925 | 1 | deploy 27478 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
136926 | 2 | deploy 27478 | pocket-to-sqlite 213286752 | actions/setup-python@v4 | Set up Python | {"python-version": "3.11", "cache": "pip", "cache-dependency-path": "pyproject.toml"} | |||
136927 | 3 | deploy 27478 | pocket-to-sqlite 213286752 | Install dependencies | pip install setuptools wheel twine build | ||||
136928 | 4 | deploy 27478 | pocket-to-sqlite 213286752 | Publish | python -m build twine upload dist/* | {"TWINE_USERNAME": "__token__", "TWINE_PASSWORD": "${{ secrets.PYPI_TOKEN }}"} | |||
136929 | 1 | test 27479 | pocket-to-sqlite 213286752 | actions/checkout@v3 | |||||
136930 | 2 | test 27479 | pocket-to-sqlite 213286752 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "pyproject.toml"} | |||
136931 | 3 | test 27479 | pocket-to-sqlite 213286752 | Install dependencies | pip install '.[test]' | ||||
136932 | 4 | test 27479 | pocket-to-sqlite 213286752 | Run tests | pytest | ||||
136933 | 1 | test 27480 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
136934 | 2 | test 27480 | swarm-to-sqlite 205429375 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "setup.py"} | |||
136935 | 3 | test 27480 | swarm-to-sqlite 205429375 | Install dependencies | pip install -e '.[test]' | ||||
136936 | 4 | test 27480 | swarm-to-sqlite 205429375 | Run tests | pytest | ||||
136937 | 1 | deploy 27481 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
136938 | 2 | deploy 27481 | swarm-to-sqlite 205429375 | actions/setup-python@v4 | Set up Python | {"python-version": "3.11", "cache": "pip", "cache-dependency-path": "setup.py"} | |||
136939 | 3 | deploy 27481 | swarm-to-sqlite 205429375 | Install dependencies | pip install setuptools wheel twine | ||||
136940 | 4 | deploy 27481 | swarm-to-sqlite 205429375 | Publish | python setup.py sdist bdist_wheel twine upload dist/* | {"TWINE_USERNAME": "__token__", "TWINE_PASSWORD": "${{ secrets.PYPI_TOKEN }}"} | |||
136941 | 1 | test 27482 | swarm-to-sqlite 205429375 | actions/checkout@v3 | |||||
136942 | 2 | test 27482 | swarm-to-sqlite 205429375 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}", "cache": "pip", "cache-dependency-path": "setup.py"} | |||
136943 | 3 | test 27482 | swarm-to-sqlite 205429375 | Install dependencies | pip install -e '.[test]' | ||||
136944 | 4 | test 27482 | swarm-to-sqlite 205429375 | Run tests | pytest | ||||
136945 | 1 | test 27483 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
136946 | 2 | test 27483 | twitter-to-sqlite 206156866 | actions/setup-python@v2 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}"} | |||
136947 | 3 | test 27483 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
136948 | 4 | test 27483 | twitter-to-sqlite 206156866 | Install dependencies | pip install -e '.[test]' | ||||
136949 | 5 | test 27483 | twitter-to-sqlite 206156866 | Run tests | pytest | ||||
136950 | 1 | deploy 27484 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
136951 | 2 | deploy 27484 | twitter-to-sqlite 206156866 | actions/setup-python@v2 | Set up Python | {"python-version": "3.9"} | |||
136952 | 3 | deploy 27484 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n"} | |||
136953 | 4 | deploy 27484 | twitter-to-sqlite 206156866 | Install dependencies | pip install setuptools wheel twine | ||||
136954 | 5 | deploy 27484 | twitter-to-sqlite 206156866 | Publish | python setup.py sdist bdist_wheel twine upload dist/* | {"TWINE_USERNAME": "__token__", "TWINE_PASSWORD": "${{ secrets.PYPI_TOKEN }}"} | |||
136955 | 1 | build 27485 | twitter-to-sqlite 206156866 | actions/checkout@v2 | Check out repo | ||||
136956 | 2 | build 27485 | twitter-to-sqlite 206156866 | Update TOC | npx markdown-toc README.md -i | ||||
136957 | 3 | build 27485 | twitter-to-sqlite 206156866 | Commit and push if README changed | git diff git config --global user.email "readme-bot@example.com" git config --global user.name "README-bot" git diff --quiet || (git add README.md && git commit -m "Updated README") git push | ||||
136958 | 1 | test 27486 | twitter-to-sqlite 206156866 | actions/checkout@v2 | |||||
136959 | 2 | test 27486 | twitter-to-sqlite 206156866 | actions/setup-python@v2 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}"} | |||
136960 | 3 | test 27486 | twitter-to-sqlite 206156866 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
136961 | 4 | test 27486 | twitter-to-sqlite 206156866 | Install dependencies | pip install -e '.[test]' | ||||
136962 | 5 | test 27486 | twitter-to-sqlite 206156866 | Run tests | pytest | ||||
136963 | 1 | deploy-branch-preview 27487 | datasette 107914493 | actions/checkout@v3 | |||||
136964 | 2 | deploy-branch-preview 27487 | datasette 107914493 | actions/setup-python@v4 | Set up Python 3.11 | {"python-version": "3.11"} | |||
136965 | 3 | deploy-branch-preview 27487 | datasette 107914493 | Install dependencies | pip install datasette-publish-vercel | ||||
136966 | 4 | deploy-branch-preview 27487 | datasette 107914493 | Deploy the preview | export BRANCH="${{ github.event.inputs.branch }}" wget https://latest.datasette.io/fixtures.db datasette publish vercel fixtures.db \ --branch $BRANCH \ --project "datasette-preview-$BRANCH" \ --token $VERCEL_TOKEN \ --scope datasette \ --about "Preview of $BRANCH" \ --about_url "https://github.com/simonw/datasette/tree/$BRANCH" | {"VERCEL_TOKEN": "${{ secrets.BRANCH_PREVIEW_VERCEL_TOKEN }}"} | |||
136967 | 1 | deploy 27488 | datasette 107914493 | actions/checkout@v3 | Check out datasette | ||||
136968 | 2 | deploy 27488 | datasette 107914493 | actions/setup-python@v4 | Set up Python | {"python-version": "3.9"} | |||
136969 | 3 | deploy 27488 | datasette 107914493 | actions/cache@v3 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
136970 | 4 | deploy 27488 | datasette 107914493 | Install Python dependencies | python -m pip install --upgrade pip python -m pip install -e .[test] python -m pip install -e .[docs] python -m pip install sphinx-to-sqlite==0.1a1 | ||||
136971 | 5 | deploy 27488 | datasette 107914493 | Run tests | pytest -n auto -m "not serial" pytest -m "serial" | ${{ github.ref == 'refs/heads/main' }} | |||
136972 | 6 | deploy 27488 | datasette 107914493 | Build fixtures.db | python tests/fixtures.py fixtures.db fixtures.json plugins --extra-db-filename extra_database.db | ||||
136973 | 7 | deploy 27488 | datasette 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ${{ github.ref == 'refs/heads/main' }} | |||
136974 | 8 | deploy 27488 | datasette 107914493 | Set up the alternate-route demo | echo ' from datasette import hookimpl @hookimpl def startup(datasette): db = datasette.get_database("fixtures2") db.route = "alternative-route" ' > plugins/alternative_route.py cp fixtures.db fixtures2.db | ||||
136975 | 9 | deploy 27488 | datasette 107914493 | Make some modifications to metadata.json | cat fixtures.json | \ jq '.databases |= . + {"ephemeral": {"allow": {"id": "*"}}}' | \ jq '.plugins |= . + {"datasette-ephemeral-tables": {"table_ttl": 900}}' \ > metadata.json cat metadata.json | ||||
136976 | 10 | deploy 27488 | datasette 107914493 | google-github-actions/setup-gcloud@v0 | Set up Cloud Run | {"version": "318.0.0", "service_account_email": "${{ secrets.GCP_SA_EMAIL }}", "service_account_key": "${{ secrets.GCP_SA_KEY }}"} | |||
136977 | 11 | deploy 27488 | datasette 107914493 | Deploy to Cloud Run | gcloud config set run/region us-central1 gcloud config set project datasette-222320 export SUFFIX="-${GITHUB_REF#refs/heads/}" export SUFFIX=${SUFFIX#-main} # Replace 1.0 with one-dot-zero in SUFFIX export SUFFIX=${SUFFIX//1.0/one-dot-zero} datasette publish cloudrun fixtures.db fixtures2.db extra_database.db \ -m metadata.json \ --plugins-dir=plugins \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options="--setting template_debug 1 --setting trace_debug 1 --crossdb" \ --install 'datasette-ephemeral-tables>=0.2.2' \ --service "datasette-latest$SUFFIX" \ --secret $LATEST_DATASETTE_SECRET | {"LATEST_DATASETTE_SECRET": "${{ secrets.LATEST_DATASETTE_SECRET }}"} | |||
136978 | 12 | deploy 27488 | datasette 107914493 | Deploy to docs as well (only for main) | # Deploy docs.db to a different service datasette publish cloudrun docs.db \ --branch=$GITHUB_SHA \ --version-note=$GITHUB_SHA \ --extra-options="--setting template_debug 1" \ --service=datasette-docs-latest | ${{ github.ref == 'refs/heads/main' }} | |||
136979 | 1 | documentation-links 27489 | datasette 107914493 | readthedocs/actions/preview@v1 | {"project-slug": "datasette"} | ||||
136980 | 1 | prettier 27490 | datasette 107914493 | actions/checkout@v2 | Check out repo | ||||
136981 | 2 | prettier 27490 | datasette 107914493 | actions/cache@v2 | Configure npm caching | {"path": "~/.npm", "key": "${{ runner.OS }}-npm-${{ hashFiles('**/package-lock.json') }}", "restore-keys": "${{ runner.OS }}-npm-\n"} | |||
136982 | 3 | prettier 27490 | datasette 107914493 | Install dependencies | npm ci | ||||
136983 | 4 | prettier 27490 | datasette 107914493 | Run prettier | npm run prettier -- --check | ||||
136984 | 1 | test 27491 | datasette 107914493 | actions/checkout@v3 | |||||
136985 | 2 | test 27491 | datasette 107914493 | actions/setup-python@v4 | Set up Python ${{ matrix.python-version }} | {"python-version": "${{ matrix.python-version }}"} | |||
136986 | 3 | test 27491 | datasette 107914493 | actions/cache@v3 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
136987 | 4 | test 27491 | datasette 107914493 | Install dependencies | pip install -e '.[test]' | ||||
136988 | 5 | test 27491 | datasette 107914493 | Run tests | pytest | ||||
136989 | 1 | deploy 27492 | datasette 107914493 | actions/checkout@v3 | |||||
136990 | 2 | deploy 27492 | datasette 107914493 | actions/setup-python@v4 | Set up Python | {"python-version": "3.11"} | |||
136991 | 3 | deploy 27492 | datasette 107914493 | actions/cache@v3 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n"} | |||
136992 | 4 | deploy 27492 | datasette 107914493 | Install dependencies | pip install setuptools wheel twine | ||||
136993 | 5 | deploy 27492 | datasette 107914493 | Publish | python setup.py sdist bdist_wheel twine upload dist/* | {"TWINE_USERNAME": "__token__", "TWINE_PASSWORD": "${{ secrets.PYPI_TOKEN }}"} | |||
136994 | 1 | deploy_static_docs 27493 | datasette 107914493 | actions/checkout@v2 | |||||
136995 | 2 | deploy_static_docs 27493 | datasette 107914493 | actions/setup-python@v2 | Set up Python | {"python-version": "3.9"} | |||
136996 | 3 | deploy_static_docs 27493 | datasette 107914493 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-publish-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-publish-pip-\n"} | |||
136997 | 4 | deploy_static_docs 27493 | datasette 107914493 | Install dependencies | python -m pip install -e .[docs] python -m pip install sphinx-to-sqlite==0.1a1 | ||||
136998 | 5 | deploy_static_docs 27493 | datasette 107914493 | Build docs.db | cd docs DISABLE_SPHINX_INLINE_TABS=1 sphinx-build -b xml . _build sphinx-to-sqlite ../docs.db _build cd .. | ||||
136999 | 6 | deploy_static_docs 27493 | datasette 107914493 | google-github-actions/setup-gcloud@v0 | Set up Cloud Run | {"version": "318.0.0", "service_account_email": "${{ secrets.GCP_SA_EMAIL }}", "service_account_key": "${{ secrets.GCP_SA_KEY }}"} | |||
137000 | 7 | deploy_static_docs 27493 | datasette 107914493 | Deploy stable-docs.datasette.io to Cloud Run | gcloud config set run/region us-central1 gcloud config set project datasette-222320 datasette publish cloudrun docs.db \ --service=datasette-docs-stable | ||||
137001 | 1 | deploy_docker 27494 | datasette 107914493 | actions/checkout@v2 | |||||
137002 | 2 | deploy_docker 27494 | datasette 107914493 | Build and push to Docker Hub | sleep 60 # Give PyPI time to make the new release available docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${GITHUB_REF#refs/tags/} \ --build-arg VERSION=${GITHUB_REF#refs/tags/} . docker tag $REPO:${GITHUB_REF#refs/tags/} $REPO:latest docker push $REPO:${GITHUB_REF#refs/tags/} docker push $REPO:latest | {"DOCKER_USER": "${{ secrets.DOCKER_USER }}", "DOCKER_PASS": "${{ secrets.DOCKER_PASS }}"} | |||
137003 | 1 | deploy_docker 27495 | datasette 107914493 | actions/checkout@v2 | |||||
137004 | 2 | deploy_docker 27495 | datasette 107914493 | Build and push to Docker Hub | docker login -u $DOCKER_USER -p $DOCKER_PASS export REPO=datasetteproject/datasette docker build -f Dockerfile \ -t $REPO:${VERSION_TAG} \ --build-arg VERSION=${VERSION_TAG} . docker push $REPO:${VERSION_TAG} | {"DOCKER_USER": "${{ secrets.DOCKER_USER }}", "DOCKER_PASS": "${{ secrets.DOCKER_PASS }}", "VERSION_TAG": "${{ github.event.inputs.version_tag }}"} | |||
137005 | 1 | spellcheck 27496 | datasette 107914493 | actions/checkout@v2 | |||||
137006 | 2 | spellcheck 27496 | datasette 107914493 | actions/setup-python@v2 | Set up Python ${{ matrix.python-version }} | {"python-version": 3.9} | |||
137007 | 3 | spellcheck 27496 | datasette 107914493 | actions/cache@v2 | Configure pip caching | {"path": "~/.cache/pip", "key": "${{ runner.os }}-pip-${{ hashFiles('**/setup.py') }}", "restore-keys": "${{ runner.os }}-pip-\n"} | |||
137008 | 4 | spellcheck 27496 | datasette 107914493 | Install dependencies | pip install -e '.[docs]' | ||||
137009 | 5 | spellcheck 27496 | datasette 107914493 | Check spelling | codespell README.md --ignore-words docs/codespell-ignore-words.txt codespell docs/*.rst --ignore-words docs/codespell-ignore-words.txt codespell datasette -S datasette/static --ignore-words docs/codespell-ignore-words.txt | ||||
137010 | 1 | test 27497 | datasette 107914493 | actions/checkout@v2 | Check out datasette | ||||
137011 | 2 | test 27497 | datasette 107914493 | actions/setup-python@v2 | Set up Python | {"python-version": 3.9} |
Advanced export
JSON shape: default, array, newline-delimited, object
CREATE TABLE [steps] ( [id] INTEGER PRIMARY KEY, [seq] INTEGER, [job] INTEGER REFERENCES [jobs]([id]), [repo] INTEGER REFERENCES [repos]([id]), [uses] TEXT, [name] TEXT, [with] TEXT, [run] TEXT , [env] TEXT, [if] TEXT); CREATE INDEX [idx_steps_repo] ON [steps] ([repo]); CREATE INDEX [idx_steps_job] ON [steps] ([job]);