mirror of
https://github.com/Swatinem/rust-cache.git
synced 2025-08-14 20:55:13 +00:00
Fixed #146 - Set all `package.version` in `Cargo.toml` to `0.0.0` - Set `{build-, dev-, }dependencies` of workspace crates to `0.0.0` - Remove workspace crates from `Cargo.lock` before caching - Sort all toml objects before hashing them as json Signed-off-by: Jiahao XU <Jiahao_XU@outlook.com>
370 lines
11 KiB
TypeScript
370 lines
11 KiB
TypeScript
import * as core from "@actions/core";
|
|
import * as glob from "@actions/glob";
|
|
import * as toml from "toml";
|
|
import crypto from "crypto";
|
|
import fs from "fs";
|
|
import fs_promises from "fs/promises";
|
|
import os from "os";
|
|
import path from "path";
|
|
|
|
import { getCmdOutput } from "./utils";
|
|
import { Workspace } from "./workspace";
|
|
import { getCargoBins } from "./cleanup";
|
|
|
|
const HOME = os.homedir();
|
|
export const CARGO_HOME = process.env.CARGO_HOME || path.join(HOME, ".cargo");
|
|
|
|
const STATE_CONFIG = "RUST_CACHE_CONFIG";
|
|
const HASH_LENGTH = 8;
|
|
|
|
export class CacheConfig {
|
|
/** All the paths we want to cache */
|
|
public cachePaths: Array<string> = [];
|
|
/** The primary cache key */
|
|
public cacheKey = "";
|
|
/** The secondary (restore) key that only contains the prefix and environment */
|
|
public restoreKey = "";
|
|
|
|
/** The workspace configurations */
|
|
public workspaces: Array<Workspace> = [];
|
|
|
|
/** The cargo binaries present during main step */
|
|
public cargoBins: Array<string> = [];
|
|
|
|
/** The prefix portion of the cache key */
|
|
private keyPrefix = "";
|
|
/** The rust version considered for the cache key */
|
|
private keyRust = "";
|
|
/** The environment variables considered for the cache key */
|
|
private keyEnvs: Array<string> = [];
|
|
/** The files considered for the cache key */
|
|
private keyFiles: Array<string> = [];
|
|
|
|
private constructor() {}
|
|
|
|
/**
|
|
* Constructs a [`CacheConfig`] with all the paths and keys.
|
|
*
|
|
* This will read the action `input`s, and read and persist `state` as necessary.
|
|
*/
|
|
static async new(): Promise<CacheConfig> {
|
|
const self = new CacheConfig();
|
|
|
|
// Construct key prefix:
|
|
// This uses either the `shared-key` input,
|
|
// or the `key` input combined with the `job` key.
|
|
|
|
let key = core.getInput("prefix-key") || "v0-rust";
|
|
|
|
const sharedKey = core.getInput("shared-key");
|
|
if (sharedKey) {
|
|
key += `-${sharedKey}`;
|
|
} else {
|
|
const inputKey = core.getInput("key");
|
|
if (inputKey) {
|
|
key += `-${inputKey}`;
|
|
}
|
|
|
|
const job = process.env.GITHUB_JOB;
|
|
if (job) {
|
|
key += `-${job}`;
|
|
}
|
|
}
|
|
|
|
self.keyPrefix = key;
|
|
|
|
// Construct environment portion of the key:
|
|
// This consists of a hash that considers the rust version
|
|
// as well as all the environment variables as given by a default list
|
|
// and the `env-vars` input.
|
|
// The env vars are sorted, matched by prefix and hashed into the
|
|
// resulting environment hash.
|
|
|
|
let hasher = crypto.createHash("sha1");
|
|
const rustVersion = await getRustVersion();
|
|
|
|
let keyRust = `${rustVersion.release} ${rustVersion.host}`;
|
|
hasher.update(keyRust);
|
|
hasher.update(rustVersion["commit-hash"]);
|
|
|
|
keyRust += ` (${rustVersion["commit-hash"]})`;
|
|
self.keyRust = keyRust;
|
|
|
|
// these prefixes should cover most of the compiler / rust / cargo keys
|
|
const envPrefixes = ["CARGO", "CC", "CFLAGS", "CXX", "CMAKE", "RUST"];
|
|
envPrefixes.push(...core.getInput("env-vars").split(/\s+/).filter(Boolean));
|
|
|
|
// sort the available env vars so we have a more stable hash
|
|
const keyEnvs = [];
|
|
const envKeys = Object.keys(process.env);
|
|
envKeys.sort((a, b) => a.localeCompare(b));
|
|
for (const key of envKeys) {
|
|
const value = process.env[key];
|
|
if (envPrefixes.some((prefix) => key.startsWith(prefix)) && value) {
|
|
hasher.update(`${key}=${value}`);
|
|
keyEnvs.push(key);
|
|
}
|
|
}
|
|
|
|
self.keyEnvs = keyEnvs;
|
|
|
|
key += `-${digest(hasher)}`;
|
|
|
|
self.restoreKey = key;
|
|
|
|
// Construct the lockfiles portion of the key:
|
|
// This considers all the files found via globbing for various manifests
|
|
// and lockfiles.
|
|
|
|
// Constructs the workspace config and paths to restore:
|
|
// The workspaces are given using a `$workspace -> $target` syntax.
|
|
|
|
const workspaces: Array<Workspace> = [];
|
|
const workspacesInput = core.getInput("workspaces") || ".";
|
|
for (const workspace of workspacesInput.trim().split("\n")) {
|
|
let [root, target = "target"] = workspace.split("->").map((s) => s.trim());
|
|
root = path.resolve(root);
|
|
target = path.join(root, target);
|
|
workspaces.push(new Workspace(root, target));
|
|
}
|
|
self.workspaces = workspaces;
|
|
|
|
let keyFiles = await globFiles("rust-toolchain\nrust-toolchain.toml");
|
|
const parsedKeyFiles = []; // keyFiles that are parsed, pre-processed and hashed
|
|
|
|
hasher = crypto.createHash("sha1");
|
|
|
|
for (const workspace of workspaces) {
|
|
const root = workspace.root;
|
|
keyFiles.push(
|
|
...(await globFiles(
|
|
`${root}/**/rust-toolchain\n${root}/**/rust-toolchain.toml`,
|
|
)),
|
|
);
|
|
|
|
const cargo_manifests = (await globFiles(`${root}/**/Cargo.toml`))
|
|
.filter(file => !fs.statSync(file).isDirectory());
|
|
cargo_manifests.sort((a, b) => a.localeCompare(b));
|
|
|
|
for (const cargo_manifest of cargo_manifests) {
|
|
try {
|
|
const content = await fs_promises.readFile(cargo_manifest, { encoding: 'utf8' });
|
|
const parsed = toml.parse(content);
|
|
|
|
if ("package" in parsed) {
|
|
const pack = parsed.package;
|
|
if ("version" in pack) {
|
|
pack.version = "0.0.0";
|
|
}
|
|
}
|
|
|
|
for (const prefix of ["", "build-", "dev-"]) {
|
|
const section_name = `${prefix}dependencies`;
|
|
if (!(section_name in parsed)) {
|
|
continue;
|
|
}
|
|
const deps = parsed[section_name];
|
|
|
|
for (const key of Object.keys(deps)) {
|
|
const dep = deps[key];
|
|
|
|
if ("path" in dep) {
|
|
dep.version = '0.0.0'
|
|
}
|
|
}
|
|
}
|
|
|
|
hasher.update(JSON.stringify(sort_object(parsed)));
|
|
|
|
parsedKeyFiles.push(cargo_manifest);
|
|
} catch (_e) { // Fallback to caching them as regular file
|
|
keyFiles.push(cargo_manifest);
|
|
}
|
|
}
|
|
|
|
const cargo_locks = (await globFiles(`${root}/**/Cargo.lock`))
|
|
.filter(file => !fs.statSync(file).isDirectory());
|
|
cargo_locks.sort((a, b) => a.localeCompare(b));
|
|
|
|
for (const cargo_lock of cargo_locks) {
|
|
try {
|
|
const content = await fs_promises.readFile(cargo_lock, { encoding: 'utf8' });
|
|
const parsed = toml.parse(content);
|
|
|
|
if (parsed.version !== 3 || !("package" in parsed)) {
|
|
// Fallback to caching them as regular file since this action
|
|
// can only handle Cargo.lock format version 3
|
|
keyFiles.push(cargo_lock);
|
|
continue;
|
|
}
|
|
|
|
// Package without `[[package]].source` and `[[package]].checksum`
|
|
// are the one with `path = "..."` to crates within the workspace.
|
|
const packages = parsed.package.filter((p: any) => {
|
|
"source" in p || "checksum" in p
|
|
});
|
|
|
|
hasher.update(JSON.stringify(sort_object(packages)));
|
|
|
|
parsedKeyFiles.push(cargo_lock);
|
|
} catch (_e) { // Fallback to caching them as regular file
|
|
keyFiles.push(cargo_lock);
|
|
}
|
|
}
|
|
}
|
|
keyFiles = keyFiles.filter(file => !fs.statSync(file).isDirectory());
|
|
keyFiles.sort((a, b) => a.localeCompare(b));
|
|
|
|
for (const file of keyFiles) {
|
|
for await (const chunk of fs.createReadStream(file)) {
|
|
hasher.update(chunk);
|
|
}
|
|
}
|
|
|
|
let lockHash = digest(hasher);
|
|
|
|
keyFiles.push(...parsedKeyFiles);
|
|
keyFiles.sort((a, b) => a.localeCompare(b));
|
|
self.keyFiles = keyFiles;
|
|
|
|
key += `-${lockHash}`;
|
|
self.cacheKey = key;
|
|
|
|
self.cachePaths = [CARGO_HOME];
|
|
const cacheTargets = core.getInput("cache-targets").toLowerCase() || "true";
|
|
if (cacheTargets === "true") {
|
|
self.cachePaths.push(...workspaces.map((ws) => ws.target));
|
|
}
|
|
|
|
const cacheDirectories = core.getInput("cache-directories");
|
|
for (const dir of cacheDirectories.trim().split(/\s+/).filter(Boolean)) {
|
|
self.cachePaths.push(dir);
|
|
}
|
|
|
|
const bins = await getCargoBins();
|
|
self.cargoBins = Array.from(bins.values());
|
|
|
|
return self;
|
|
}
|
|
|
|
/**
|
|
* Reads and returns the cache config from the action `state`.
|
|
*
|
|
* @throws {Error} if the state is not present.
|
|
* @returns {CacheConfig} the configuration.
|
|
* @see {@link CacheConfig#saveState}
|
|
* @see {@link CacheConfig#new}
|
|
*/
|
|
static fromState(): CacheConfig {
|
|
const source = core.getState(STATE_CONFIG);
|
|
if (!source) {
|
|
throw new Error("Cache configuration not found in state");
|
|
}
|
|
|
|
const self = new CacheConfig();
|
|
Object.assign(self, JSON.parse(source));
|
|
self.workspaces = self.workspaces
|
|
.map((w: any) => new Workspace(w.root, w.target));
|
|
|
|
return self;
|
|
}
|
|
|
|
/**
|
|
* Prints the configuration to the action log.
|
|
*/
|
|
printInfo() {
|
|
core.startGroup("Cache Configuration");
|
|
core.info(`Workspaces:`);
|
|
for (const workspace of this.workspaces) {
|
|
core.info(` ${workspace.root}`);
|
|
}
|
|
core.info(`Cache Paths:`);
|
|
for (const path of this.cachePaths) {
|
|
core.info(` ${path}`);
|
|
}
|
|
core.info(`Restore Key:`);
|
|
core.info(` ${this.restoreKey}`);
|
|
core.info(`Cache Key:`);
|
|
core.info(` ${this.cacheKey}`);
|
|
core.info(`.. Prefix:`);
|
|
core.info(` - ${this.keyPrefix}`);
|
|
core.info(`.. Environment considered:`);
|
|
core.info(` - Rust Version: ${this.keyRust}`);
|
|
for (const env of this.keyEnvs) {
|
|
core.info(` - ${env}`);
|
|
}
|
|
core.info(`.. Lockfiles considered:`);
|
|
for (const file of this.keyFiles) {
|
|
core.info(` - ${file}`);
|
|
}
|
|
core.endGroup();
|
|
}
|
|
|
|
/**
|
|
* Saves the configuration to the state store.
|
|
* This is used to restore the configuration in the post action.
|
|
*/
|
|
saveState() {
|
|
core.saveState(STATE_CONFIG, this);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Checks if the cache is up to date.
|
|
*
|
|
* @returns `true` if the cache is up to date, `false` otherwise.
|
|
*/
|
|
export function isCacheUpToDate(): boolean {
|
|
return core.getState(STATE_CONFIG) === "";
|
|
}
|
|
|
|
/**
|
|
* Returns a hex digest of the given hasher truncated to `HASH_LENGTH`.
|
|
*
|
|
* @param hasher The hasher to digest.
|
|
* @returns The hex digest.
|
|
*/
|
|
function digest(hasher: crypto.Hash): string {
|
|
return hasher.digest("hex").substring(0, HASH_LENGTH);
|
|
}
|
|
|
|
interface RustVersion {
|
|
host: string;
|
|
release: string;
|
|
"commit-hash": string;
|
|
}
|
|
|
|
async function getRustVersion(): Promise<RustVersion> {
|
|
const stdout = await getCmdOutput("rustc", ["-vV"]);
|
|
let splits = stdout
|
|
.split(/[\n\r]+/)
|
|
.filter(Boolean)
|
|
.map((s) => s.split(":").map((s) => s.trim()))
|
|
.filter((s) => s.length === 2);
|
|
return Object.fromEntries(splits);
|
|
}
|
|
|
|
async function globFiles(pattern: string): Promise<string[]> {
|
|
const globber = await glob.create(pattern, {
|
|
followSymbolicLinks: false,
|
|
});
|
|
return await globber.glob();
|
|
}
|
|
|
|
function sort_object(o: any): any {
|
|
if (Array.isArray(o)) {
|
|
return o.sort().map(sort_object);
|
|
} else if (typeof o === 'object' && o != null) {
|
|
return Object
|
|
.keys(o)
|
|
.sort()
|
|
.reduce(function(a: any, k) {
|
|
a[k] = sort_object(o[k]);
|
|
|
|
return a;
|
|
}, {});
|
|
} else {
|
|
return o;
|
|
}
|
|
}
|