mirror of
https://github.com/Swatinem/rust-cache.git
synced 2025-08-14 20:55:13 +00:00
Compare commits
62 Commits
Author | SHA1 | Date | |
---|---|---|---|
|
267a8a94c9 | ||
|
46cb408fba | ||
|
203c9eb82a | ||
|
a8b36e3f74 | ||
|
7e1e2d0a10 | ||
|
98c8021b55 | ||
|
14d3bc39c4 | ||
|
52ea1434f8 | ||
|
eaa85be6b1 | ||
|
901019c0f8 | ||
|
9d47c6ad4b | ||
|
27b8ea9368 | ||
|
f0deed1e0e | ||
|
008623fb83 | ||
|
720f7e45cc | ||
|
4b1f006ad2 | ||
|
e8e63cdbf2 | ||
|
9a2e0d3212 | ||
|
c00f3025ca | ||
|
68b3cb7503 | ||
|
82a92a6e8f | ||
|
598fe25fa1 | ||
|
8f842c2d45 | ||
|
96a8d65dba | ||
|
9bdad043e8 | ||
|
f7a52f6914 | ||
|
2bceda3912 | ||
|
640a22190e | ||
|
1582741630 | ||
|
23bce251a8 | ||
|
378c8285a4 | ||
|
a226033982 | ||
|
d30f1144e8 | ||
|
b1db5f9d5f | ||
|
3cf7f8cc28 | ||
|
e03705e031 | ||
|
b86d1c6caa | ||
|
f27990c89a | ||
|
a95ba19544 | ||
|
82c8487d00 | ||
|
67c46e7159 | ||
|
44b6087283 | ||
|
e207df5d26 | ||
|
decb69d790 | ||
|
ab6b2769d1 | ||
|
578b235f6e | ||
|
5113490c3f | ||
|
c0e052c18c | ||
|
4e0f4b19dd | ||
|
b919e1427f | ||
|
b8a6852b4f | ||
|
80c47cc945 | ||
|
5ec9842c14 | ||
|
3312b3ab47 | ||
|
f6987ea139 | ||
|
e97a782690 | ||
|
b00faf5858 | ||
|
9de8f90afb | ||
|
fd201ad913 | ||
|
cf3f88254c | ||
|
4fedae9bcf | ||
|
1e604afb09 |
50
.github/dependabot.yaml
vendored
Normal file
50
.github/dependabot.yaml
vendored
Normal file
@ -0,0 +1,50 @@
|
|||||||
|
# https://docs.github.com/en/code-security/dependabot/dependabot-version-updates/optimizing-pr-creation-version-updates#setting-up-a-cooldown-period-for-dependency-updates
|
||||||
|
|
||||||
|
version: 2
|
||||||
|
updates:
|
||||||
|
- package-ecosystem: cargo
|
||||||
|
directories:
|
||||||
|
- tests
|
||||||
|
- tests/wasm-workspace
|
||||||
|
schedule:
|
||||||
|
interval: weekly
|
||||||
|
# https://docs.github.com/en/code-security/dependabot/working-with-dependabot/dependabot-options-reference#groups--
|
||||||
|
# 1 PR per week and group
|
||||||
|
groups:
|
||||||
|
cargo-major:
|
||||||
|
update-types: ["major"]
|
||||||
|
cargo-minor:
|
||||||
|
update-types: ["minor"]
|
||||||
|
cargo-patch:
|
||||||
|
update-types: ["patch"]
|
||||||
|
- package-ecosystem: github-actions
|
||||||
|
directory: /
|
||||||
|
schedule:
|
||||||
|
interval: weekly
|
||||||
|
groups:
|
||||||
|
actions:
|
||||||
|
# Combine all images of the last week
|
||||||
|
patterns: ["*"]
|
||||||
|
- package-ecosystem: npm
|
||||||
|
directory: /
|
||||||
|
schedule:
|
||||||
|
interval: weekly
|
||||||
|
groups:
|
||||||
|
prd-major:
|
||||||
|
dependency-type: "production"
|
||||||
|
update-types: ["major"]
|
||||||
|
prd-minor:
|
||||||
|
dependency-type: "production"
|
||||||
|
update-types: ["minor"]
|
||||||
|
prd-patch:
|
||||||
|
dependency-type: "production"
|
||||||
|
update-types: ["patch"]
|
||||||
|
dev-major:
|
||||||
|
dependency-type: "development"
|
||||||
|
update-types: ["major"]
|
||||||
|
dev-minor:
|
||||||
|
dependency-type: "development"
|
||||||
|
update-types: ["minor"]
|
||||||
|
dev-patch:
|
||||||
|
dependency-type: "development"
|
||||||
|
update-types: ["patch"]
|
33
.github/workflows/buildjet.yml
vendored
Normal file
33
.github/workflows/buildjet.yml
vendored
Normal file
@ -0,0 +1,33 @@
|
|||||||
|
name: buildjet
|
||||||
|
|
||||||
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
buildjet:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest, macos-latest, windows-latest]
|
||||||
|
|
||||||
|
name: Test buildjet provider on ${{ matrix.os }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
|
||||||
|
env:
|
||||||
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
|
- run: rustup toolchain install stable --profile minimal --no-self-update
|
||||||
|
|
||||||
|
- uses: ./
|
||||||
|
with:
|
||||||
|
workspaces: tests
|
||||||
|
cache-provider: buildjet
|
||||||
|
|
||||||
|
- run: |
|
||||||
|
cargo check
|
||||||
|
cargo test
|
||||||
|
cargo build --release
|
||||||
|
working-directory: tests
|
15
.github/workflows/check-dist.yml
vendored
15
.github/workflows/check-dist.yml
vendored
@ -5,22 +5,23 @@ on:
|
|||||||
branches:
|
branches:
|
||||||
- master
|
- master
|
||||||
paths-ignore:
|
paths-ignore:
|
||||||
- '**.md'
|
- "**.md"
|
||||||
pull_request:
|
pull_request:
|
||||||
paths-ignore:
|
paths-ignore:
|
||||||
- '**.md'
|
- "**.md"
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
check-dist:
|
check-dist:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
- name: Setup Node.js 16.x
|
- name: Setup Node.js 20.x
|
||||||
uses: actions/setup-node@v3
|
uses: actions/setup-node@v4
|
||||||
with:
|
with:
|
||||||
node-version: 16.x
|
node-version: 20.x
|
||||||
cache: npm
|
cache: npm
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
@ -38,7 +39,7 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
id: diff
|
id: diff
|
||||||
|
|
||||||
- uses: actions/upload-artifact@v3
|
- uses: actions/upload-artifact@v4
|
||||||
if: ${{ failure() && steps.diff.conclusion == 'failure' }}
|
if: ${{ failure() && steps.diff.conclusion == 'failure' }}
|
||||||
with:
|
with:
|
||||||
name: dist
|
name: dist
|
||||||
|
9
.github/workflows/coverage.yml
vendored
9
.github/workflows/coverage.yml
vendored
@ -4,6 +4,7 @@ on: [push, pull_request]
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
coverage:
|
coverage:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@ -16,13 +17,9 @@ jobs:
|
|||||||
CARGO_TERM_COLOR: always
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
- run: rustup toolchain install stable --profile minimal --component llvm-tools-preview --no-self-update
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install stable --profile minimal --component llvm-tools-preview
|
|
||||||
|
|
||||||
- uses: taiki-e/install-action@cargo-llvm-cov
|
- uses: taiki-e/install-action@cargo-llvm-cov
|
||||||
|
|
||||||
|
25
.github/workflows/dependabot.yml
vendored
Normal file
25
.github/workflows/dependabot.yml
vendored
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
# https://docs.github.com/en/code-security/dependabot/working-with-dependabot/automating-dependabot-with-github-actions#enabling-automerge-on-a-pull-request
|
||||||
|
|
||||||
|
name: Dependabot Automation
|
||||||
|
on: pull_request
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
pull-requests: write
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
automerge:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
if: github.event.pull_request.user.login == 'dependabot[bot]' && github.repository == 'Swatinem/rust-cache'
|
||||||
|
steps:
|
||||||
|
- name: Fetch metadata
|
||||||
|
id: metadata
|
||||||
|
uses: dependabot/fetch-metadata@v2
|
||||||
|
with:
|
||||||
|
github-token: "${{ secrets.GITHUB_TOKEN }}"
|
||||||
|
- name: Auto-merge Patch PRs
|
||||||
|
if: steps.metadata.outputs.update-type == 'version-update:semver-patch'
|
||||||
|
run: gh pr merge --auto --merge "$PR_URL"
|
||||||
|
env:
|
||||||
|
PR_URL: ${{github.event.pull_request.html_url}}
|
||||||
|
GH_TOKEN: ${{secrets.GITHUB_TOKEN}}
|
32
.github/workflows/git-registry.yml
vendored
Normal file
32
.github/workflows/git-registry.yml
vendored
Normal file
@ -0,0 +1,32 @@
|
|||||||
|
name: git-registry
|
||||||
|
|
||||||
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
git-registry:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest, macos-latest, windows-latest]
|
||||||
|
|
||||||
|
name: Test cargo "git" registry on ${{ matrix.os }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
|
||||||
|
env:
|
||||||
|
CARGO_TERM_COLOR: always
|
||||||
|
CARGO_REGISTRIES_CRATES_IO_PROTOCOL: git
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
|
- run: rustup toolchain install stable --profile minimal --no-self-update
|
||||||
|
|
||||||
|
- uses: ./
|
||||||
|
with:
|
||||||
|
workspaces: tests
|
||||||
|
|
||||||
|
- run: |
|
||||||
|
cargo check
|
||||||
|
cargo test
|
||||||
|
working-directory: tests
|
23
.github/workflows/globbed-directories.yml
vendored
23
.github/workflows/globbed-directories.yml
vendored
@ -1,23 +0,0 @@
|
|||||||
name: globbed-directories
|
|
||||||
|
|
||||||
on: [push, pull_request]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
simple:
|
|
||||||
strategy:
|
|
||||||
fail-fast: false
|
|
||||||
matrix:
|
|
||||||
os: [ubuntu-latest]
|
|
||||||
|
|
||||||
name: Test `cargo check/test` on ${{ matrix.os }}
|
|
||||||
runs-on: ${{ matrix.os }}
|
|
||||||
|
|
||||||
env:
|
|
||||||
CARGO_TERM_COLOR: always
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
- name: Create folder that will be globbed
|
|
||||||
run: |
|
|
||||||
mkdir -p folder/rust-toolchain
|
|
||||||
- uses: ./
|
|
9
.github/workflows/install.yml
vendored
9
.github/workflows/install.yml
vendored
@ -4,6 +4,7 @@ on: [push, pull_request]
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
install:
|
install:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@ -16,13 +17,9 @@ jobs:
|
|||||||
CARGO_TERM_COLOR: always
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
- run: rustup toolchain install stable --profile minimal --no-self-update
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install stable --profile minimal
|
|
||||||
|
|
||||||
- uses: ./
|
- uses: ./
|
||||||
|
|
||||||
|
12
.github/workflows/simple.yml
vendored
12
.github/workflows/simple.yml
vendored
@ -4,25 +4,22 @@ on: [push, pull_request]
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
simple:
|
simple:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
os: [ubuntu-latest, macos-latest, windows-latest]
|
os: [ubuntu-latest, macos-latest, windows-latest]
|
||||||
|
|
||||||
name: Test `cargo check/test` on ${{ matrix.os }}
|
name: Test `cargo check/test/build` on ${{ matrix.os }}
|
||||||
runs-on: ${{ matrix.os }}
|
runs-on: ${{ matrix.os }}
|
||||||
|
|
||||||
env:
|
env:
|
||||||
CARGO_TERM_COLOR: always
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
- run: rustup toolchain install stable --profile minimal --no-self-update
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install stable --profile minimal
|
|
||||||
|
|
||||||
- uses: ./
|
- uses: ./
|
||||||
with:
|
with:
|
||||||
@ -31,4 +28,5 @@ jobs:
|
|||||||
- run: |
|
- run: |
|
||||||
cargo check
|
cargo check
|
||||||
cargo test
|
cargo test
|
||||||
|
cargo build --release
|
||||||
working-directory: tests
|
working-directory: tests
|
||||||
|
36
.github/workflows/sparse-registry.yml
vendored
36
.github/workflows/sparse-registry.yml
vendored
@ -1,36 +0,0 @@
|
|||||||
name: sparse-registry
|
|
||||||
|
|
||||||
on: [push, pull_request]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
sparse-registry:
|
|
||||||
strategy:
|
|
||||||
fail-fast: false
|
|
||||||
matrix:
|
|
||||||
os: [ubuntu-latest, macos-latest, windows-latest]
|
|
||||||
|
|
||||||
name: Test `cargo check/test` with sparse registry on ${{ matrix.os }}
|
|
||||||
runs-on: ${{ matrix.os }}
|
|
||||||
|
|
||||||
env:
|
|
||||||
CARGO_TERM_COLOR: always
|
|
||||||
CARGO_UNSTABLE_SPARSE_REGISTRY: true
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install nightly --profile minimal
|
|
||||||
rustup default nightly
|
|
||||||
|
|
||||||
- uses: ./
|
|
||||||
with:
|
|
||||||
workspaces: tests
|
|
||||||
|
|
||||||
- run: |
|
|
||||||
cargo check
|
|
||||||
cargo test
|
|
||||||
working-directory: tests
|
|
9
.github/workflows/target-dir.yml
vendored
9
.github/workflows/target-dir.yml
vendored
@ -4,6 +4,7 @@ on: [push, pull_request]
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
target-dir:
|
target-dir:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@ -16,13 +17,9 @@ jobs:
|
|||||||
CARGO_TERM_COLOR: always
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
- run: rustup toolchain install stable --profile minimal --no-self-update
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install stable --profile minimal
|
|
||||||
|
|
||||||
# the `workspaces` option has the format `$workspace -> $target-dir`
|
# the `workspaces` option has the format `$workspace -> $target-dir`
|
||||||
# and the `$target-dir` is relative to the `$workspace`.
|
# and the `$target-dir` is relative to the `$workspace`.
|
||||||
|
15
.github/workflows/workspaces.yml
vendored
15
.github/workflows/workspaces.yml
vendored
@ -4,6 +4,7 @@ on: [push, pull_request]
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
workspaces:
|
workspaces:
|
||||||
|
if: github.repository == 'Swatinem/rust-cache'
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@ -16,13 +17,9 @@ jobs:
|
|||||||
CARGO_TERM_COLOR: always
|
CARGO_TERM_COLOR: always
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# When rustup is updated, it tries to replace its binary, which on Windows is somehow locked.
|
- run: rustup toolchain install stable --profile minimal --target wasm32-unknown-unknown --no-self-update
|
||||||
# This can result in the CI failure, see: https://github.com/rust-lang/rustup/issues/3029
|
|
||||||
- run: |
|
|
||||||
rustup set auto-self-update disable
|
|
||||||
rustup toolchain install stable --profile minimal --target wasm32-unknown-unknown
|
|
||||||
|
|
||||||
- uses: ./
|
- uses: ./
|
||||||
with:
|
with:
|
||||||
@ -30,8 +27,10 @@ jobs:
|
|||||||
tests
|
tests
|
||||||
tests/wasm-workspace
|
tests/wasm-workspace
|
||||||
|
|
||||||
- run: cargo check
|
- name: cargo check (tests)
|
||||||
working-directory: tests
|
working-directory: tests
|
||||||
|
run: cargo check
|
||||||
|
|
||||||
- run: cargo check
|
- name: cargo check (tests/wasm-workspace)
|
||||||
working-directory: tests/wasm-workspace
|
working-directory: tests/wasm-workspace
|
||||||
|
run: cargo check
|
||||||
|
55
CHANGELOG.md
55
CHANGELOG.md
@ -1,5 +1,60 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 2.8.0
|
||||||
|
|
||||||
|
- Add support for `warpbuild` cache provider
|
||||||
|
- Add new `cache-workspace-crates` feature
|
||||||
|
|
||||||
|
## 2.7.8
|
||||||
|
|
||||||
|
- Include CPU arch in the cache key
|
||||||
|
|
||||||
|
## 2.7.7
|
||||||
|
|
||||||
|
- Also cache `cargo install` metadata
|
||||||
|
|
||||||
|
## 2.7.6
|
||||||
|
|
||||||
|
- Allow opting out of caching $CARGO_HOME/bin
|
||||||
|
- Add runner OS in cache key
|
||||||
|
- Adds an option to do lookup-only of the cache
|
||||||
|
|
||||||
|
## 2.7.5
|
||||||
|
|
||||||
|
- Support Cargo.lock format cargo-lock v4
|
||||||
|
- Only run macOsWorkaround() on macOS
|
||||||
|
|
||||||
|
## 2.7.3
|
||||||
|
|
||||||
|
- Work around upstream problem that causes cache saving to hang for minutes.
|
||||||
|
|
||||||
|
## 2.7.2
|
||||||
|
|
||||||
|
- Only key by `Cargo.toml` and `Cargo.lock` files of workspace members.
|
||||||
|
|
||||||
|
## 2.7.1
|
||||||
|
|
||||||
|
- Update toml parser to fix parsing errors.
|
||||||
|
|
||||||
|
## 2.7.0
|
||||||
|
|
||||||
|
- Properly cache `trybuild` tests.
|
||||||
|
|
||||||
|
## 2.6.2
|
||||||
|
|
||||||
|
- Fix `toml` parsing.
|
||||||
|
|
||||||
|
## 2.6.1
|
||||||
|
|
||||||
|
- Fix hash contributions of `Cargo.lock`/`Cargo.toml` files.
|
||||||
|
|
||||||
|
## 2.6.0
|
||||||
|
|
||||||
|
- Add "buildjet" as a second `cache-provider` backend.
|
||||||
|
- Clean up sparse registry index.
|
||||||
|
- Do not clean up src of `-sys` crates.
|
||||||
|
- Remove `.cargo/credentials.toml` before saving.
|
||||||
|
|
||||||
## 2.5.1
|
## 2.5.1
|
||||||
|
|
||||||
- Fix hash contribution of `Cargo.lock`.
|
- Fix hash contribution of `Cargo.lock`.
|
||||||
|
39
README.md
39
README.md
@ -6,7 +6,7 @@ sensible defaults.
|
|||||||
## Example usage
|
## Example usage
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v5
|
||||||
|
|
||||||
# selecting a toolchain either by action or manual `rustup` calls should happen
|
# selecting a toolchain either by action or manual `rustup` calls should happen
|
||||||
# before the plugin, as the cache uses the current rustc version as its cache key
|
# before the plugin, as the cache uses the current rustc version as its cache key
|
||||||
@ -60,11 +60,35 @@ sensible defaults.
|
|||||||
# default: "false"
|
# default: "false"
|
||||||
cache-all-crates: ""
|
cache-all-crates: ""
|
||||||
|
|
||||||
# Determiners whether the cache should be saved.
|
# Similar to cache-all-crates.
|
||||||
|
# If `true` the workspace crates will be cached.
|
||||||
|
# Useful if the workspace contains libraries that are only updated sporadically.
|
||||||
|
# default: "false"
|
||||||
|
cache-workspace-crates: ""
|
||||||
|
|
||||||
|
# Determines whether the cache should be saved.
|
||||||
# If `false`, the cache is only restored.
|
# If `false`, the cache is only restored.
|
||||||
# Useful for jobs where the matrix is additive e.g. additional Cargo features.
|
# Useful for jobs where the matrix is additive e.g. additional Cargo features,
|
||||||
|
# or when only runs from `master` should be saved to the cache.
|
||||||
# default: "true"
|
# default: "true"
|
||||||
save-if: ""
|
save-if: ""
|
||||||
|
# To only cache runs from `master`:
|
||||||
|
save-if: ${{ github.ref == 'refs/heads/master' }}
|
||||||
|
|
||||||
|
# Determines whether the cache should be restored.
|
||||||
|
# If `true` the cache key will be checked and the `cache-hit` output will be set
|
||||||
|
# but the cache itself won't be restored
|
||||||
|
# default: "false"
|
||||||
|
lookup-only: ""
|
||||||
|
|
||||||
|
# Specifies what to use as the backend providing cache
|
||||||
|
# Can be set to "github", "buildjet", or "warpbuild"
|
||||||
|
# default: "github"
|
||||||
|
cache-provider: ""
|
||||||
|
|
||||||
|
# Determines whether to cache the ~/.cargo/bin directory.
|
||||||
|
# default: "true"
|
||||||
|
cache-bin: ""
|
||||||
```
|
```
|
||||||
|
|
||||||
Further examples are available in the [.github/workflows](./.github/workflows/) directory.
|
Further examples are available in the [.github/workflows](./.github/workflows/) directory.
|
||||||
@ -85,7 +109,8 @@ repositories with only a `Cargo.toml` file have limited benefits, as cargo will
|
|||||||
_always_ use the most up-to-date dependency versions, which may not be cached.
|
_always_ use the most up-to-date dependency versions, which may not be cached.
|
||||||
|
|
||||||
Usage with Stable Rust is most effective, as a cache is tied to the Rust version.
|
Usage with Stable Rust is most effective, as a cache is tied to the Rust version.
|
||||||
Using it with Nightly Rust is less effective as it will throw away the cache every day.
|
Using it with Nightly Rust is less effective as it will throw away the cache every day,
|
||||||
|
unless a specific nightly build is being pinned.
|
||||||
|
|
||||||
## Cache Details
|
## Cache Details
|
||||||
|
|
||||||
@ -101,6 +126,7 @@ This cache is automatically keyed by:
|
|||||||
- the value of some compiler-specific environment variables (eg. RUSTFLAGS, etc), and
|
- the value of some compiler-specific environment variables (eg. RUSTFLAGS, etc), and
|
||||||
- a hash of all `Cargo.lock` / `Cargo.toml` files found anywhere in the repository (if present).
|
- a hash of all `Cargo.lock` / `Cargo.toml` files found anywhere in the repository (if present).
|
||||||
- a hash of all `rust-toolchain` / `rust-toolchain.toml` files in the root of the repository (if present).
|
- a hash of all `rust-toolchain` / `rust-toolchain.toml` files in the root of the repository (if present).
|
||||||
|
- a hash of all `.cargo/config.toml` files in the root of the repository (if present).
|
||||||
|
|
||||||
An additional input `key` can be provided if the builtin keys are not sufficient.
|
An additional input `key` can be provided if the builtin keys are not sufficient.
|
||||||
|
|
||||||
@ -159,4 +185,7 @@ to see those details as well as further details related to caching operations.
|
|||||||
## Known issues
|
## Known issues
|
||||||
|
|
||||||
- The cache cleaning process currently removes all the files from `~/.cargo/bin`
|
- The cache cleaning process currently removes all the files from `~/.cargo/bin`
|
||||||
that were present before the action ran (for example `rustc`).
|
that were present before the action ran (for example `rustc`), by default.
|
||||||
|
This can be an issue on long-running self-hosted runners, where such state
|
||||||
|
is expected to be preserved across runs. You can work around this by setting
|
||||||
|
`cache-bin: "false"`.
|
||||||
|
1
TODO.md
1
TODO.md
@ -1,4 +1,3 @@
|
|||||||
- better .cargo/bin handling:
|
- better .cargo/bin handling:
|
||||||
- get a list of all the files on "pre"/"restore"
|
- get a list of all the files on "pre"/"restore"
|
||||||
- move the files out of the way on "post"/"save" and move them back afterwards
|
- move the files out of the way on "post"/"save" and move them back afterwards
|
||||||
- properly clean sparse registry
|
|
||||||
|
18
action.yml
18
action.yml
@ -32,15 +32,31 @@ inputs:
|
|||||||
description: "Determines which crates are cached. If `true` all crates will be cached, otherwise only dependent crates will be cached."
|
description: "Determines which crates are cached. If `true` all crates will be cached, otherwise only dependent crates will be cached."
|
||||||
required: false
|
required: false
|
||||||
default: "false"
|
default: "false"
|
||||||
|
cache-workspace-crates:
|
||||||
|
description: "Similar to cache-all-crates. If `true` the workspace crates will be cached."
|
||||||
|
required: false
|
||||||
|
default: "false"
|
||||||
save-if:
|
save-if:
|
||||||
description: "Determiners whether the cache should be saved. If `false`, the cache is only restored."
|
description: "Determiners whether the cache should be saved. If `false`, the cache is only restored."
|
||||||
required: false
|
required: false
|
||||||
default: "true"
|
default: "true"
|
||||||
|
cache-provider:
|
||||||
|
description: "Determines which provider to use for caching. Options are github, buildjet, or warpbuild. Defaults to github."
|
||||||
|
required: false
|
||||||
|
default: "github"
|
||||||
|
cache-bin:
|
||||||
|
description: "Determines whether to cache ${CARGO_HOME}/bin."
|
||||||
|
required: false
|
||||||
|
default: "true"
|
||||||
|
lookup-only:
|
||||||
|
description: "Check if a cache entry exists without downloading the cache"
|
||||||
|
required: false
|
||||||
|
default: "false"
|
||||||
outputs:
|
outputs:
|
||||||
cache-hit:
|
cache-hit:
|
||||||
description: "A boolean value that indicates an exact match was found."
|
description: "A boolean value that indicates an exact match was found."
|
||||||
runs:
|
runs:
|
||||||
using: "node16"
|
using: "node20"
|
||||||
main: "dist/restore/index.js"
|
main: "dist/restore/index.js"
|
||||||
post: "dist/save/index.js"
|
post: "dist/save/index.js"
|
||||||
post-if: "success() || env.CACHE_ON_FAILURE == 'true'"
|
post-if: "success() || env.CACHE_ON_FAILURE == 'true'"
|
||||||
|
155098
dist/restore/index.js
vendored
155098
dist/restore/index.js
vendored
File diff suppressed because one or more lines are too long
155108
dist/save/index.js
vendored
155108
dist/save/index.js
vendored
File diff suppressed because one or more lines are too long
1962
package-lock.json
generated
1962
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
17
package.json
17
package.json
@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"private": true,
|
"private": true,
|
||||||
"name": "rust-cache",
|
"name": "rust-cache",
|
||||||
"version": "2.5.1",
|
"version": "2.8.0",
|
||||||
"description": "A GitHub Action that implements smart caching for rust/cargo projects with sensible defaults.",
|
"description": "A GitHub Action that implements smart caching for rust/cargo projects with sensible defaults.",
|
||||||
"keywords": [
|
"keywords": [
|
||||||
"actions",
|
"actions",
|
||||||
@ -22,17 +22,20 @@
|
|||||||
},
|
},
|
||||||
"homepage": "https://github.com/Swatinem/rust-cache#readme",
|
"homepage": "https://github.com/Swatinem/rust-cache#readme",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/cache": "^3.2.1",
|
"@actions/buildjet-cache": "npm:github-actions.cache-buildjet@0.2.0",
|
||||||
"@actions/core": "^1.10.0",
|
"@actions/warpbuild-cache": "npm:github-actions.warp-cache@1.4.7",
|
||||||
|
"@actions/cache": "^4.0.5",
|
||||||
|
"@actions/core": "^1.11.1",
|
||||||
"@actions/exec": "^1.1.1",
|
"@actions/exec": "^1.1.1",
|
||||||
"@actions/glob": "^0.4.0",
|
"@actions/glob": "^0.5.0",
|
||||||
"@actions/io": "^1.1.3",
|
"@actions/io": "^1.1.3",
|
||||||
"toml": "^3.0.0"
|
"smol-toml": "^1.4.2"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@vercel/ncc": "^0.36.1",
|
"@types/node": "^22.16.0",
|
||||||
|
"@vercel/ncc": "^0.38.3",
|
||||||
"linefix": "^0.1.1",
|
"linefix": "^0.1.1",
|
||||||
"typescript": "5.0.4"
|
"typescript": "5.8.3"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"prepare": "ncc build --target es2020 -o dist/restore src/restore.ts && ncc build --target es2020 -o dist/save src/save.ts && linefix dist"
|
"prepare": "ncc build --target es2020 -o dist/restore src/restore.ts && ncc build --target es2020 -o dist/save src/save.ts && linefix dist"
|
||||||
|
@ -4,6 +4,7 @@ import fs from "fs";
|
|||||||
import path from "path";
|
import path from "path";
|
||||||
|
|
||||||
import { CARGO_HOME } from "./config";
|
import { CARGO_HOME } from "./config";
|
||||||
|
import { exists } from "./utils";
|
||||||
import { Packages } from "./workspace";
|
import { Packages } from "./workspace";
|
||||||
|
|
||||||
export async function cleanTargetDir(targetDir: string, packages: Packages, checkTimestamp = false) {
|
export async function cleanTargetDir(targetDir: string, packages: Packages, checkTimestamp = false) {
|
||||||
@ -34,6 +35,26 @@ export async function cleanTargetDir(targetDir: string, packages: Packages, chec
|
|||||||
async function cleanProfileTarget(profileDir: string, packages: Packages, checkTimestamp = false) {
|
async function cleanProfileTarget(profileDir: string, packages: Packages, checkTimestamp = false) {
|
||||||
core.debug(`cleaning profile directory "${profileDir}"`);
|
core.debug(`cleaning profile directory "${profileDir}"`);
|
||||||
|
|
||||||
|
// Quite a few testing utility crates store compilation artifacts as nested
|
||||||
|
// workspaces under `target/tests`. Notably, `target/tests/target` and
|
||||||
|
// `target/tests/trybuild`.
|
||||||
|
if (path.basename(profileDir) === "tests") {
|
||||||
|
try {
|
||||||
|
// https://github.com/vertexclique/kaos/blob/9876f6c890339741cc5be4b7cb9df72baa5a6d79/src/cargo.rs#L25
|
||||||
|
// https://github.com/eupn/macrotest/blob/c4151a5f9f545942f4971980b5d264ebcd0b1d11/src/cargo.rs#L27
|
||||||
|
cleanTargetDir(path.join(profileDir, "target"), packages, checkTimestamp);
|
||||||
|
} catch {}
|
||||||
|
try {
|
||||||
|
// https://github.com/dtolnay/trybuild/blob/eec8ca6cb9b8f53d0caf1aa499d99df52cae8b40/src/cargo.rs#L50
|
||||||
|
cleanTargetDir(path.join(profileDir, "trybuild"), packages, checkTimestamp);
|
||||||
|
} catch {}
|
||||||
|
|
||||||
|
// Delete everything else.
|
||||||
|
await rmExcept(profileDir, new Set(["target", "trybuild"]), checkTimestamp);
|
||||||
|
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
let keepProfile = new Set(["build", ".fingerprint", "deps"]);
|
let keepProfile = new Set(["build", ".fingerprint", "deps"]);
|
||||||
await rmExcept(profileDir, keepProfile);
|
await rmExcept(profileDir, keepProfile);
|
||||||
|
|
||||||
@ -91,11 +112,15 @@ export async function cleanBin(oldBins: Array<string>) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export async function cleanRegistry(packages: Packages, crates = true) {
|
export async function cleanRegistry(packages: Packages, crates = true) {
|
||||||
// `.cargo/registry/src`
|
// remove `.cargo/credentials.toml`
|
||||||
// we can remove this completely, as cargo will recreate this from `cache`
|
try {
|
||||||
await rmRF(path.join(CARGO_HOME, "registry", "src"));
|
const credentials = path.join(CARGO_HOME, ".cargo", "credentials.toml");
|
||||||
|
core.debug(`deleting "${credentials}"`);
|
||||||
|
await fs.promises.unlink(credentials);
|
||||||
|
} catch {}
|
||||||
|
|
||||||
// `.cargo/registry/index`
|
// `.cargo/registry/index`
|
||||||
|
let pkgSet = new Set(packages.map((p) => p.name));
|
||||||
const indexDir = await fs.promises.opendir(path.join(CARGO_HOME, "registry", "index"));
|
const indexDir = await fs.promises.opendir(path.join(CARGO_HOME, "registry", "index"));
|
||||||
for await (const dirent of indexDir) {
|
for await (const dirent of indexDir) {
|
||||||
if (dirent.isDirectory()) {
|
if (dirent.isDirectory()) {
|
||||||
@ -106,19 +131,38 @@ export async function cleanRegistry(packages: Packages, crates = true) {
|
|||||||
// for a git registry, we can remove `.cache`, as cargo will recreate it from git
|
// for a git registry, we can remove `.cache`, as cargo will recreate it from git
|
||||||
if (await exists(path.join(dirPath, ".git"))) {
|
if (await exists(path.join(dirPath, ".git"))) {
|
||||||
await rmRF(path.join(dirPath, ".cache"));
|
await rmRF(path.join(dirPath, ".cache"));
|
||||||
|
} else {
|
||||||
|
await cleanRegistryIndexCache(dirPath, pkgSet);
|
||||||
}
|
}
|
||||||
// TODO: else, clean `.cache` based on the `packages`
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!crates) {
|
if (!crates) {
|
||||||
core.debug(`skipping crate cleanup`);
|
core.debug("skipping registry cache and src cleanup");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const pkgSet = new Set(packages.map((p) => `${p.name}-${p.version}.crate`));
|
// `.cargo/registry/src`
|
||||||
|
// Cargo usually re-creates these from the `.crate` cache below,
|
||||||
|
// but for some reason that does not work for `-sys` crates that check timestamps
|
||||||
|
// to decide if rebuilds are necessary.
|
||||||
|
pkgSet = new Set(packages.filter((p) => p.name.endsWith("-sys")).map((p) => `${p.name}-${p.version}`));
|
||||||
|
const srcDir = await fs.promises.opendir(path.join(CARGO_HOME, "registry", "src"));
|
||||||
|
for await (const dirent of srcDir) {
|
||||||
|
if (dirent.isDirectory()) {
|
||||||
|
// eg `.cargo/registry/src/github.com-1ecc6299db9ec823`
|
||||||
|
// or `.cargo/registry/src/index.crates.io-e139d0d48fed7772`
|
||||||
|
const dir = await fs.promises.opendir(path.join(srcDir.path, dirent.name));
|
||||||
|
for await (const dirent of dir) {
|
||||||
|
if (dirent.isDirectory() && !pkgSet.has(dirent.name)) {
|
||||||
|
await rmRF(path.join(dir.path, dirent.name));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// `.cargo/registry/cache`
|
// `.cargo/registry/cache`
|
||||||
|
pkgSet = new Set(packages.map((p) => `${p.name}-${p.version}.crate`));
|
||||||
const cacheDir = await fs.promises.opendir(path.join(CARGO_HOME, "registry", "cache"));
|
const cacheDir = await fs.promises.opendir(path.join(CARGO_HOME, "registry", "cache"));
|
||||||
for await (const dirent of cacheDir) {
|
for await (const dirent of cacheDir) {
|
||||||
if (dirent.isDirectory()) {
|
if (dirent.isDirectory()) {
|
||||||
@ -135,6 +179,28 @@ export async function cleanRegistry(packages: Packages, crates = true) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Recursively walks and cleans the index `.cache`
|
||||||
|
async function cleanRegistryIndexCache(dirName: string, keepPkg: Set<string>) {
|
||||||
|
let dirIsEmpty = true;
|
||||||
|
const cacheDir = await fs.promises.opendir(dirName);
|
||||||
|
for await (const dirent of cacheDir) {
|
||||||
|
if (dirent.isDirectory()) {
|
||||||
|
if (await cleanRegistryIndexCache(path.join(dirName, dirent.name), keepPkg)) {
|
||||||
|
await rm(dirName, dirent);
|
||||||
|
} else {
|
||||||
|
dirIsEmpty &&= false;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if (keepPkg.has(dirent.name)) {
|
||||||
|
dirIsEmpty &&= false;
|
||||||
|
} else {
|
||||||
|
await rm(dirName, dirent);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return dirIsEmpty;
|
||||||
|
}
|
||||||
|
|
||||||
export async function cleanGit(packages: Packages) {
|
export async function cleanGit(packages: Packages) {
|
||||||
const coPath = path.join(CARGO_HOME, "git", "checkouts");
|
const coPath = path.join(CARGO_HOME, "git", "checkouts");
|
||||||
const dbPath = path.join(CARGO_HOME, "git", "db");
|
const dbPath = path.join(CARGO_HOME, "git", "db");
|
||||||
@ -243,12 +309,3 @@ async function rmRF(dirName: string) {
|
|||||||
core.debug(`deleting "${dirName}"`);
|
core.debug(`deleting "${dirName}"`);
|
||||||
await io.rmRF(dirName);
|
await io.rmRF(dirName);
|
||||||
}
|
}
|
||||||
|
|
||||||
async function exists(path: string) {
|
|
||||||
try {
|
|
||||||
await fs.promises.access(path);
|
|
||||||
return true;
|
|
||||||
} catch {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
@ -1,15 +1,15 @@
|
|||||||
import * as core from "@actions/core";
|
import * as core from "@actions/core";
|
||||||
import * as glob from "@actions/glob";
|
import * as glob from "@actions/glob";
|
||||||
import * as toml from "toml";
|
|
||||||
import crypto from "crypto";
|
import crypto from "crypto";
|
||||||
import fs from "fs";
|
import fs from "fs";
|
||||||
import fs_promises from "fs/promises";
|
import fs_promises from "fs/promises";
|
||||||
import os from "os";
|
import os from "os";
|
||||||
import path from "path";
|
import path from "path";
|
||||||
|
import * as toml from "smol-toml";
|
||||||
|
|
||||||
import { getCmdOutput } from "./utils";
|
|
||||||
import { Workspace } from "./workspace";
|
|
||||||
import { getCargoBins } from "./cleanup";
|
import { getCargoBins } from "./cleanup";
|
||||||
|
import { CacheProvider, exists, getCmdOutput } from "./utils";
|
||||||
|
import { Workspace } from "./workspace";
|
||||||
|
|
||||||
const HOME = os.homedir();
|
const HOME = os.homedir();
|
||||||
export const CARGO_HOME = process.env.CARGO_HOME || path.join(HOME, ".cargo");
|
export const CARGO_HOME = process.env.CARGO_HOME || path.join(HOME, ".cargo");
|
||||||
@ -25,6 +25,9 @@ export class CacheConfig {
|
|||||||
/** The secondary (restore) key that only contains the prefix and environment */
|
/** The secondary (restore) key that only contains the prefix and environment */
|
||||||
public restoreKey = "";
|
public restoreKey = "";
|
||||||
|
|
||||||
|
/** Whether to cache CARGO_HOME/.bin */
|
||||||
|
public cacheBin: boolean = true;
|
||||||
|
|
||||||
/** The workspace configurations */
|
/** The workspace configurations */
|
||||||
public workspaces: Array<Workspace> = [];
|
public workspaces: Array<Workspace> = [];
|
||||||
|
|
||||||
@ -71,6 +74,11 @@ export class CacheConfig {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Add runner OS and CPU architecture to the key to avoid cross-contamination of cache
|
||||||
|
const runnerOS = os.type();
|
||||||
|
const runnerArch = os.arch();
|
||||||
|
key += `-${runnerOS}-${runnerArch}`;
|
||||||
|
|
||||||
self.keyPrefix = key;
|
self.keyPrefix = key;
|
||||||
|
|
||||||
// Construct environment portion of the key:
|
// Construct environment portion of the key:
|
||||||
@ -116,6 +124,8 @@ export class CacheConfig {
|
|||||||
// This considers all the files found via globbing for various manifests
|
// This considers all the files found via globbing for various manifests
|
||||||
// and lockfiles.
|
// and lockfiles.
|
||||||
|
|
||||||
|
self.cacheBin = core.getInput("cache-bin").toLowerCase() == "true";
|
||||||
|
|
||||||
// Constructs the workspace config and paths to restore:
|
// Constructs the workspace config and paths to restore:
|
||||||
// The workspaces are given using a `$workspace -> $target` syntax.
|
// The workspaces are given using a `$workspace -> $target` syntax.
|
||||||
|
|
||||||
@ -142,17 +152,20 @@ export class CacheConfig {
|
|||||||
)),
|
)),
|
||||||
);
|
);
|
||||||
|
|
||||||
const cargo_manifests = sort_and_uniq(await globFiles(`${root}/**/Cargo.toml`));
|
const workspaceMembers = await workspace.getWorkspaceMembers();
|
||||||
|
|
||||||
|
const cargo_manifests = sort_and_uniq(workspaceMembers.map((member) => path.join(member.path, "Cargo.toml")));
|
||||||
|
|
||||||
for (const cargo_manifest of cargo_manifests) {
|
for (const cargo_manifest of cargo_manifests) {
|
||||||
try {
|
try {
|
||||||
const content = await fs_promises.readFile(cargo_manifest, { encoding: 'utf8' });
|
const content = await fs_promises.readFile(cargo_manifest, { encoding: "utf8" });
|
||||||
const parsed = toml.parse(content);
|
// Use any since TomlPrimitive is not exposed
|
||||||
|
const parsed = toml.parse(content) as { [key: string]: any };
|
||||||
|
|
||||||
if ("package" in parsed) {
|
if ("package" in parsed) {
|
||||||
const pack = parsed.package;
|
const pack = parsed.package;
|
||||||
if ("version" in pack) {
|
if ("version" in pack) {
|
||||||
pack.version = "0.0.0";
|
pack["version"] = "0.0.0";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -166,8 +179,15 @@ export class CacheConfig {
|
|||||||
for (const key of Object.keys(deps)) {
|
for (const key of Object.keys(deps)) {
|
||||||
const dep = deps[key];
|
const dep = deps[key];
|
||||||
|
|
||||||
|
try {
|
||||||
if ("path" in dep) {
|
if ("path" in dep) {
|
||||||
dep.version = '0.0.0'
|
dep.version = "0.0.0";
|
||||||
|
dep.path = "";
|
||||||
|
}
|
||||||
|
} catch (_e) {
|
||||||
|
// Not an object, probably a string (version),
|
||||||
|
// continue.
|
||||||
|
continue;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -175,35 +195,37 @@ export class CacheConfig {
|
|||||||
hasher.update(JSON.stringify(parsed));
|
hasher.update(JSON.stringify(parsed));
|
||||||
|
|
||||||
parsedKeyFiles.push(cargo_manifest);
|
parsedKeyFiles.push(cargo_manifest);
|
||||||
} catch (_e) { // Fallback to caching them as regular file
|
} catch (e) {
|
||||||
|
// Fallback to caching them as regular file
|
||||||
|
core.warning(`Error parsing Cargo.toml manifest, fallback to caching entire file: ${e}`);
|
||||||
keyFiles.push(cargo_manifest);
|
keyFiles.push(cargo_manifest);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const cargo_locks = sort_and_uniq(await globFiles(`${root}/**/Cargo.lock`));
|
const cargo_lock = path.join(workspace.root, "Cargo.lock");
|
||||||
|
if (await exists(cargo_lock)) {
|
||||||
for (const cargo_lock of cargo_locks) {
|
|
||||||
try {
|
try {
|
||||||
const content = await fs_promises.readFile(cargo_lock, { encoding: 'utf8' });
|
const content = await fs_promises.readFile(cargo_lock, { encoding: "utf8" });
|
||||||
const parsed = toml.parse(content);
|
const parsed = toml.parse(content);
|
||||||
|
|
||||||
if (parsed.version !== 3 || !("package" in parsed)) {
|
if ((parsed.version !== 3 && parsed.version !== 4) || !("package" in parsed)) {
|
||||||
// Fallback to caching them as regular file since this action
|
// Fallback to caching them as regular file since this action
|
||||||
// can only handle Cargo.lock format version 3
|
// can only handle Cargo.lock format version 3
|
||||||
|
core.warning("Unsupported Cargo.lock format, fallback to caching entire file");
|
||||||
keyFiles.push(cargo_lock);
|
keyFiles.push(cargo_lock);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Package without `[[package]].source` and `[[package]].checksum`
|
// Package without `[[package]].source` and `[[package]].checksum`
|
||||||
// are the one with `path = "..."` to crates within the workspace.
|
// are the one with `path = "..."` to crates within the workspace.
|
||||||
const packages = parsed.package.filter((p: any) => {
|
const packages = (parsed.package as any[]).filter((p: any) => "source" in p || "checksum" in p);
|
||||||
"source" in p || "checksum" in p
|
|
||||||
});
|
|
||||||
|
|
||||||
hasher.update(JSON.stringify(packages));
|
hasher.update(JSON.stringify(packages));
|
||||||
|
|
||||||
parsedKeyFiles.push(cargo_lock);
|
parsedKeyFiles.push(cargo_lock);
|
||||||
} catch (_e) { // Fallback to caching them as regular file
|
} catch (e) {
|
||||||
|
// Fallback to caching them as regular file
|
||||||
|
core.warning(`Error parsing Cargo.lock manifest, fallback to caching entire file: ${e}`);
|
||||||
keyFiles.push(cargo_lock);
|
keyFiles.push(cargo_lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -224,7 +246,15 @@ export class CacheConfig {
|
|||||||
key += `-${lockHash}`;
|
key += `-${lockHash}`;
|
||||||
self.cacheKey = key;
|
self.cacheKey = key;
|
||||||
|
|
||||||
self.cachePaths = [CARGO_HOME];
|
self.cachePaths = [path.join(CARGO_HOME, "registry"), path.join(CARGO_HOME, "git")];
|
||||||
|
if (self.cacheBin) {
|
||||||
|
self.cachePaths = [
|
||||||
|
path.join(CARGO_HOME, "bin"),
|
||||||
|
path.join(CARGO_HOME, ".crates.toml"),
|
||||||
|
path.join(CARGO_HOME, ".crates2.json"),
|
||||||
|
...self.cachePaths,
|
||||||
|
];
|
||||||
|
}
|
||||||
const cacheTargets = core.getInput("cache-targets").toLowerCase() || "true";
|
const cacheTargets = core.getInput("cache-targets").toLowerCase() || "true";
|
||||||
if (cacheTargets === "true") {
|
if (cacheTargets === "true") {
|
||||||
self.cachePaths.push(...workspaces.map((ws) => ws.target));
|
self.cachePaths.push(...workspaces.map((ws) => ws.target));
|
||||||
@ -257,8 +287,7 @@ export class CacheConfig {
|
|||||||
|
|
||||||
const self = new CacheConfig();
|
const self = new CacheConfig();
|
||||||
Object.assign(self, JSON.parse(source));
|
Object.assign(self, JSON.parse(source));
|
||||||
self.workspaces = self.workspaces
|
self.workspaces = self.workspaces.map((w: any) => new Workspace(w.root, w.target));
|
||||||
.map((w: any) => new Workspace(w.root, w.target));
|
|
||||||
|
|
||||||
return self;
|
return self;
|
||||||
}
|
}
|
||||||
@ -266,8 +295,10 @@ export class CacheConfig {
|
|||||||
/**
|
/**
|
||||||
* Prints the configuration to the action log.
|
* Prints the configuration to the action log.
|
||||||
*/
|
*/
|
||||||
printInfo() {
|
printInfo(cacheProvider: CacheProvider) {
|
||||||
core.startGroup("Cache Configuration");
|
core.startGroup("Cache Configuration");
|
||||||
|
core.info(`Cache Provider:`);
|
||||||
|
core.info(` ${cacheProvider.name}`);
|
||||||
core.info(`Workspaces:`);
|
core.info(`Workspaces:`);
|
||||||
for (const workspace of this.workspaces) {
|
for (const workspace of this.workspaces) {
|
||||||
core.info(` ${workspace.root}`);
|
core.info(` ${workspace.root}`);
|
||||||
@ -345,14 +376,13 @@ async function globFiles(pattern: string): Promise<string[]> {
|
|||||||
// fs.statSync resolve the symbolic link and returns stat for the
|
// fs.statSync resolve the symbolic link and returns stat for the
|
||||||
// file it pointed to, so isFile would make sure the resolved
|
// file it pointed to, so isFile would make sure the resolved
|
||||||
// file is actually a regular file.
|
// file is actually a regular file.
|
||||||
return (await globber.glob()).filter(file => fs.statSync(file).isFile());
|
return (await globber.glob()).filter((file) => fs.statSync(file).isFile());
|
||||||
}
|
}
|
||||||
|
|
||||||
function sort_and_uniq(a: string[]) {
|
function sort_and_uniq(a: string[]) {
|
||||||
return a
|
return a
|
||||||
.sort((a, b) => a.localeCompare(b))
|
.sort((a, b) => a.localeCompare(b))
|
||||||
.reduce(
|
.reduce((accumulator: string[], currentValue: string) => {
|
||||||
(accumulator: string[], currentValue: string) => {
|
|
||||||
const len = accumulator.length;
|
const len = accumulator.length;
|
||||||
// If accumulator is empty or its last element != currentValue
|
// If accumulator is empty or its last element != currentValue
|
||||||
// Since array is already sorted, elements with the same value
|
// Since array is already sorted, elements with the same value
|
||||||
@ -363,7 +393,5 @@ function sort_and_uniq(a: string[]) {
|
|||||||
accumulator.push(currentValue);
|
accumulator.push(currentValue);
|
||||||
}
|
}
|
||||||
return accumulator;
|
return accumulator;
|
||||||
},
|
}, []);
|
||||||
[]
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
@ -1,8 +1,8 @@
|
|||||||
import * as cache from "@actions/cache";
|
|
||||||
import * as core from "@actions/core";
|
import * as core from "@actions/core";
|
||||||
|
|
||||||
import { cleanTargetDir } from "./cleanup";
|
import { cleanTargetDir } from "./cleanup";
|
||||||
import { CacheConfig } from "./config";
|
import { CacheConfig } from "./config";
|
||||||
|
import { getCacheProvider, reportError } from "./utils";
|
||||||
|
|
||||||
process.on("uncaughtException", (e) => {
|
process.on("uncaughtException", (e) => {
|
||||||
core.error(e.message);
|
core.error(e.message);
|
||||||
@ -12,7 +12,9 @@ process.on("uncaughtException", (e) => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
async function run() {
|
async function run() {
|
||||||
if (!cache.isFeatureAvailable()) {
|
const cacheProvider = getCacheProvider();
|
||||||
|
|
||||||
|
if (!cacheProvider.cache.isFeatureAvailable()) {
|
||||||
setCacheHitOutput(false);
|
setCacheHitOutput(false);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@ -22,22 +24,26 @@ async function run() {
|
|||||||
if (cacheOnFailure !== "true") {
|
if (cacheOnFailure !== "true") {
|
||||||
cacheOnFailure = "false";
|
cacheOnFailure = "false";
|
||||||
}
|
}
|
||||||
|
var lookupOnly = core.getInput("lookup-only").toLowerCase() === "true";
|
||||||
|
|
||||||
core.exportVariable("CACHE_ON_FAILURE", cacheOnFailure);
|
core.exportVariable("CACHE_ON_FAILURE", cacheOnFailure);
|
||||||
core.exportVariable("CARGO_INCREMENTAL", 0);
|
core.exportVariable("CARGO_INCREMENTAL", 0);
|
||||||
|
|
||||||
const config = await CacheConfig.new();
|
const config = await CacheConfig.new();
|
||||||
config.printInfo();
|
config.printInfo(cacheProvider);
|
||||||
core.info("");
|
core.info("");
|
||||||
|
|
||||||
core.info(`... Restoring cache ...`);
|
core.info(`... ${lookupOnly ? "Checking" : "Restoring"} cache ...`);
|
||||||
const key = config.cacheKey;
|
const key = config.cacheKey;
|
||||||
// Pass a copy of cachePaths to avoid mutating the original array as reported by:
|
// Pass a copy of cachePaths to avoid mutating the original array as reported by:
|
||||||
// https://github.com/actions/toolkit/pull/1378
|
// https://github.com/actions/toolkit/pull/1378
|
||||||
// TODO: remove this once the underlying bug is fixed.
|
// TODO: remove this once the underlying bug is fixed.
|
||||||
const restoreKey = await cache.restoreCache(config.cachePaths.slice(), key, [config.restoreKey]);
|
const restoreKey = await cacheProvider.cache.restoreCache(config.cachePaths.slice(), key, [config.restoreKey], {
|
||||||
|
lookupOnly,
|
||||||
|
});
|
||||||
if (restoreKey) {
|
if (restoreKey) {
|
||||||
const match = restoreKey === key;
|
const match = restoreKey === key;
|
||||||
core.info(`Restored from cache key "${restoreKey}" full match: ${match}.`);
|
core.info(`${lookupOnly ? "Found" : "Restored from"} cache key "${restoreKey}" full match: ${match}.`);
|
||||||
if (!match) {
|
if (!match) {
|
||||||
// pre-clean the target directory on cache mismatch
|
// pre-clean the target directory on cache mismatch
|
||||||
for (const workspace of config.workspaces) {
|
for (const workspace of config.workspaces) {
|
||||||
@ -60,8 +66,9 @@ async function run() {
|
|||||||
} catch (e) {
|
} catch (e) {
|
||||||
setCacheHitOutput(false);
|
setCacheHitOutput(false);
|
||||||
|
|
||||||
core.error(`${(e as any).stack}`);
|
reportError(e);
|
||||||
}
|
}
|
||||||
|
process.exit();
|
||||||
}
|
}
|
||||||
|
|
||||||
function setCacheHitOutput(cacheHit: boolean): void {
|
function setCacheHitOutput(cacheHit: boolean): void {
|
||||||
|
36
src/save.ts
36
src/save.ts
@ -1,9 +1,9 @@
|
|||||||
import * as cache from "@actions/cache";
|
|
||||||
import * as core from "@actions/core";
|
import * as core from "@actions/core";
|
||||||
import * as exec from "@actions/exec";
|
import * as exec from "@actions/exec";
|
||||||
|
|
||||||
import { cleanBin, cleanGit, cleanRegistry, cleanTargetDir } from "./cleanup";
|
import { cleanBin, cleanGit, cleanRegistry, cleanTargetDir } from "./cleanup";
|
||||||
import { CacheConfig, isCacheUpToDate } from "./config";
|
import { CacheConfig, isCacheUpToDate } from "./config";
|
||||||
|
import { getCacheProvider, reportError } from "./utils";
|
||||||
|
|
||||||
process.on("uncaughtException", (e) => {
|
process.on("uncaughtException", (e) => {
|
||||||
core.error(e.message);
|
core.error(e.message);
|
||||||
@ -13,9 +13,11 @@ process.on("uncaughtException", (e) => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
async function run() {
|
async function run() {
|
||||||
|
const cacheProvider = getCacheProvider();
|
||||||
|
|
||||||
const save = core.getInput("save-if").toLowerCase() || "true";
|
const save = core.getInput("save-if").toLowerCase() || "true";
|
||||||
|
|
||||||
if (!(cache.isFeatureAvailable() && save === "true")) {
|
if (!(cacheProvider.cache.isFeatureAvailable() && save === "true")) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -26,54 +28,64 @@ async function run() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const config = CacheConfig.fromState();
|
const config = CacheConfig.fromState();
|
||||||
config.printInfo();
|
config.printInfo(cacheProvider);
|
||||||
core.info("");
|
core.info("");
|
||||||
|
|
||||||
// TODO: remove this once https://github.com/actions/toolkit/pull/553 lands
|
// TODO: remove this once https://github.com/actions/toolkit/pull/553 lands
|
||||||
|
if (process.env["RUNNER_OS"] == "macOS") {
|
||||||
await macOsWorkaround();
|
await macOsWorkaround();
|
||||||
|
}
|
||||||
|
|
||||||
|
const workspaceCrates = core.getInput("cache-workspace-crates").toLowerCase() || "false";
|
||||||
const allPackages = [];
|
const allPackages = [];
|
||||||
for (const workspace of config.workspaces) {
|
for (const workspace of config.workspaces) {
|
||||||
const packages = await workspace.getPackages();
|
const packages = await workspace.getPackagesOutsideWorkspaceRoot();
|
||||||
|
if (workspaceCrates === "true") {
|
||||||
|
const wsMembers = await workspace.getWorkspaceMembers();
|
||||||
|
packages.push(...wsMembers);
|
||||||
|
}
|
||||||
allPackages.push(...packages);
|
allPackages.push(...packages);
|
||||||
try {
|
try {
|
||||||
core.info(`... Cleaning ${workspace.target} ...`);
|
core.info(`... Cleaning ${workspace.target} ...`);
|
||||||
await cleanTargetDir(workspace.target, packages);
|
await cleanTargetDir(workspace.target, packages);
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`${(e as any).stack}`);
|
core.debug(`${(e as any).stack}`);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const crates = core.getInput("cache-all-crates").toLowerCase() || "false"
|
const crates = core.getInput("cache-all-crates").toLowerCase() || "false";
|
||||||
core.info(`... Cleaning cargo registry cache-all-crates: ${crates} ...`);
|
core.info(`... Cleaning cargo registry (cache-all-crates: ${crates}) ...`);
|
||||||
await cleanRegistry(allPackages, crates !== "true");
|
await cleanRegistry(allPackages, crates !== "true");
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`${(e as any).stack}`);
|
core.debug(`${(e as any).stack}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (config.cacheBin) {
|
||||||
try {
|
try {
|
||||||
core.info(`... Cleaning cargo/bin ...`);
|
core.info(`... Cleaning cargo/bin ...`);
|
||||||
await cleanBin(config.cargoBins);
|
await cleanBin(config.cargoBins);
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`${(e as any).stack}`);
|
core.debug(`${(e as any).stack}`);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
core.info(`... Cleaning cargo git cache ...`);
|
core.info(`... Cleaning cargo git cache ...`);
|
||||||
await cleanGit(allPackages);
|
await cleanGit(allPackages);
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`${(e as any).stack}`);
|
core.debug(`${(e as any).stack}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
core.info(`... Saving cache ...`);
|
core.info(`... Saving cache ...`);
|
||||||
// Pass a copy of cachePaths to avoid mutating the original array as reported by:
|
// Pass a copy of cachePaths to avoid mutating the original array as reported by:
|
||||||
// https://github.com/actions/toolkit/pull/1378
|
// https://github.com/actions/toolkit/pull/1378
|
||||||
// TODO: remove this once the underlying bug is fixed.
|
// TODO: remove this once the underlying bug is fixed.
|
||||||
await cache.saveCache(config.cachePaths.slice(), config.cacheKey);
|
await cacheProvider.cache.saveCache(config.cachePaths.slice(), config.cacheKey);
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`${(e as any).stack}`);
|
reportError(e);
|
||||||
}
|
}
|
||||||
|
process.exit();
|
||||||
}
|
}
|
||||||
|
|
||||||
run();
|
run();
|
||||||
|
63
src/utils.ts
63
src/utils.ts
@ -1,5 +1,19 @@
|
|||||||
import * as core from "@actions/core";
|
import * as core from "@actions/core";
|
||||||
import * as exec from "@actions/exec";
|
import * as exec from "@actions/exec";
|
||||||
|
import * as buildjetCache from "@actions/buildjet-cache";
|
||||||
|
import * as warpbuildCache from "@actions/warpbuild-cache";
|
||||||
|
import * as ghCache from "@actions/cache";
|
||||||
|
import fs from "fs";
|
||||||
|
|
||||||
|
export function reportError(e: any) {
|
||||||
|
const { commandFailed } = e;
|
||||||
|
if (commandFailed) {
|
||||||
|
core.error(`Command failed: ${commandFailed.command}`);
|
||||||
|
core.error(commandFailed.stderr);
|
||||||
|
} else {
|
||||||
|
core.error(`${e.stack}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
export async function getCmdOutput(
|
export async function getCmdOutput(
|
||||||
cmd: string,
|
cmd: string,
|
||||||
@ -22,9 +36,54 @@ export async function getCmdOutput(
|
|||||||
...options,
|
...options,
|
||||||
});
|
});
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
core.error(`Command failed: ${cmd} ${args.join(" ")}`);
|
(e as any).commandFailed = {
|
||||||
core.error(stderr);
|
command: `${cmd} ${args.join(" ")}`,
|
||||||
|
stderr,
|
||||||
|
};
|
||||||
throw e;
|
throw e;
|
||||||
}
|
}
|
||||||
return stdout;
|
return stdout;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export interface GhCache {
|
||||||
|
isFeatureAvailable: typeof ghCache.isFeatureAvailable;
|
||||||
|
restoreCache: typeof ghCache.restoreCache;
|
||||||
|
saveCache: (paths: string[], key: string) => Promise<string | number>;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface CacheProvider {
|
||||||
|
name: string;
|
||||||
|
cache: GhCache;
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getCacheProvider(): CacheProvider {
|
||||||
|
const cacheProvider = core.getInput("cache-provider");
|
||||||
|
let cache: GhCache;
|
||||||
|
switch (cacheProvider) {
|
||||||
|
case "github":
|
||||||
|
cache = ghCache;
|
||||||
|
break;
|
||||||
|
case "buildjet":
|
||||||
|
cache = buildjetCache;
|
||||||
|
break;
|
||||||
|
case "warpbuild":
|
||||||
|
cache = warpbuildCache;
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
throw new Error(`The \`cache-provider\` \`${cacheProvider}\` is not valid.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
name: cacheProvider,
|
||||||
|
cache: cache,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function exists(path: string) {
|
||||||
|
try {
|
||||||
|
await fs.promises.access(path);
|
||||||
|
return true;
|
||||||
|
} catch {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@ -8,26 +8,34 @@ const SAVE_TARGETS = new Set(["lib", "proc-macro"]);
|
|||||||
export class Workspace {
|
export class Workspace {
|
||||||
constructor(public root: string, public target: string) {}
|
constructor(public root: string, public target: string) {}
|
||||||
|
|
||||||
public async getPackages(): Promise<Packages> {
|
async getPackages(filter: (p: Meta["packages"][0]) => boolean, ...extraArgs: string[]): Promise<Packages> {
|
||||||
let packages: Packages = [];
|
let packages: Packages = [];
|
||||||
try {
|
try {
|
||||||
core.debug(`collecting metadata for "${this.root}"`);
|
core.debug(`collecting metadata for "${this.root}"`);
|
||||||
const meta: Meta = JSON.parse(
|
const meta: Meta = JSON.parse(
|
||||||
await getCmdOutput("cargo", ["metadata", "--all-features", "--format-version", "1"], {
|
await getCmdOutput("cargo", ["metadata", "--all-features", "--format-version", "1", ...extraArgs], {
|
||||||
cwd: this.root,
|
cwd: this.root,
|
||||||
|
env: { "CARGO_ENCODED_RUSTFLAGS": "" },
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
core.debug(`workspace "${this.root}" has ${meta.packages.length} packages`);
|
core.debug(`workspace "${this.root}" has ${meta.packages.length} packages`);
|
||||||
for (const pkg of meta.packages) {
|
for (const pkg of meta.packages.filter(filter)) {
|
||||||
if (pkg.manifest_path.startsWith(this.root)) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
const targets = pkg.targets.filter((t) => t.kind.some((kind) => SAVE_TARGETS.has(kind))).map((t) => t.name);
|
const targets = pkg.targets.filter((t) => t.kind.some((kind) => SAVE_TARGETS.has(kind))).map((t) => t.name);
|
||||||
packages.push({ name: pkg.name, version: pkg.version, targets, path: path.dirname(pkg.manifest_path) });
|
packages.push({ name: pkg.name, version: pkg.version, targets, path: path.dirname(pkg.manifest_path) });
|
||||||
}
|
}
|
||||||
} catch {}
|
} catch (err) {
|
||||||
|
console.error(err);
|
||||||
|
}
|
||||||
return packages;
|
return packages;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public async getPackagesOutsideWorkspaceRoot(): Promise<Packages> {
|
||||||
|
return await this.getPackages((pkg) => !pkg.manifest_path.startsWith(this.root));
|
||||||
|
}
|
||||||
|
|
||||||
|
public async getWorkspaceMembers(): Promise<Packages> {
|
||||||
|
return await this.getPackages((_) => true, "--no-deps");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export interface PackageDefinition {
|
export interface PackageDefinition {
|
||||||
|
1419
tests/Cargo.lock
generated
1419
tests/Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
@ -6,5 +6,12 @@ authors = ["Arpad Borsos <arpad.borsos@googlemail.com>"]
|
|||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
reqwest = "0.11.11"
|
reqwest = "0.12.1"
|
||||||
watto = { git = "https://github.com/getsentry/watto", rev = "d71c8218506bddba102a124a460d64da25e303dc", features = ["strings"] }
|
jsonpath_lib_polars_vendor = "0.0.1"
|
||||||
|
watto = { git = "https://github.com/getsentry/watto", rev = "39ccb9add289c1f23c89f40506f4a80b2f4011b9", features = ["strings"] }
|
||||||
|
|
||||||
|
[dev-dependencies]
|
||||||
|
trybuild = "1"
|
||||||
|
|
||||||
|
[target.'cfg(not(target_env = "msvc"))'.dependencies]
|
||||||
|
tikv-jemallocator = "0.6.0"
|
||||||
|
2
tests/rust-toolchain/.keep
Normal file
2
tests/rust-toolchain/.keep
Normal file
@ -0,0 +1,2 @@
|
|||||||
|
the `rust-toolchain` directory will be globbed,
|
||||||
|
and should not lead to any errors down the road
|
@ -1,3 +1,10 @@
|
|||||||
|
#[cfg(not(target_env = "msvc"))]
|
||||||
|
use tikv_jemallocator::Jemalloc;
|
||||||
|
|
||||||
|
#[cfg(not(target_env = "msvc"))]
|
||||||
|
#[global_allocator]
|
||||||
|
static GLOBAL: Jemalloc = Jemalloc;
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
println!("Hello, world!");
|
println!("Hello, world!");
|
||||||
}
|
}
|
||||||
|
6
tests/tests/trybuild.rs
Normal file
6
tests/tests/trybuild.rs
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
#[test]
|
||||||
|
fn test_trybuild() {
|
||||||
|
let t = trybuild::TestCases::new();
|
||||||
|
t.pass("tests/trybuild/empty_main.rs");
|
||||||
|
t.compile_fail("tests/trybuild/fail_to_compile.rs");
|
||||||
|
}
|
1
tests/tests/trybuild/empty_main.rs
Normal file
1
tests/tests/trybuild/empty_main.rs
Normal file
@ -0,0 +1 @@
|
|||||||
|
fn main() {}
|
3
tests/tests/trybuild/fail_to_compile.rs
Normal file
3
tests/tests/trybuild/fail_to_compile.rs
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
fn main() {
|
||||||
|
"foobar".foobar();
|
||||||
|
}
|
5
tests/tests/trybuild/fail_to_compile.stderr
Normal file
5
tests/tests/trybuild/fail_to_compile.stderr
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
error[E0599]: no method named `foobar` found for reference `&'static str` in the current scope
|
||||||
|
--> tests/trybuild/fail_to_compile.rs:2:14
|
||||||
|
|
|
||||||
|
2 | "foobar".foobar();
|
||||||
|
| ^^^^^^ method not found in `&'static str`
|
1578
tests/wasm-workspace/Cargo.lock
generated
1578
tests/wasm-workspace/Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
@ -1,4 +1,5 @@
|
|||||||
[workspace]
|
[workspace]
|
||||||
|
resolver = "2"
|
||||||
members = [
|
members = [
|
||||||
"crates/one",
|
"crates/one",
|
||||||
"crates/two",
|
"crates/two",
|
||||||
|
@ -5,7 +5,7 @@ version = "0.1.0"
|
|||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
reqwest = "0.11.0"
|
reqwest = "0.12"
|
||||||
async-std = "1"
|
async-std = "1"
|
||||||
tracing = "0.1"
|
tracing = "0.1"
|
||||||
tracing-futures = "0.2"
|
tracing-futures = "0.2"
|
||||||
|
@ -5,4 +5,4 @@ version = "0.1.0"
|
|||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
clap = "3"
|
clap = "4"
|
||||||
|
Loading…
x
Reference in New Issue
Block a user