2019-10-31 01:48:49 +07:00
|
|
|
import * as core from "@actions/core";
|
|
|
|
import * as path from "path";
|
|
|
|
import * as cacheHttpClient from "./cacheHttpClient";
|
2019-11-22 02:37:54 +07:00
|
|
|
import { Events, Inputs, State } from "./constants";
|
2019-12-14 05:24:37 +07:00
|
|
|
import { createTar } from "./tar";
|
2019-10-31 01:48:49 +07:00
|
|
|
import * as utils from "./utils/actionUtils";
|
|
|
|
|
2019-11-13 04:48:02 +07:00
|
|
|
async function run(): Promise<void> {
|
2019-10-31 01:48:49 +07:00
|
|
|
try {
|
2019-11-22 02:37:54 +07:00
|
|
|
if (!utils.isValidEvent()) {
|
|
|
|
utils.logWarning(
|
|
|
|
`Event Validation Error: The event type ${
|
|
|
|
process.env[Events.Key]
|
|
|
|
} is not supported. Only ${utils
|
|
|
|
.getSupportedEvents()
|
|
|
|
.join(", ")} events are supported at this time.`
|
|
|
|
);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-10-31 01:48:49 +07:00
|
|
|
const state = utils.getCacheState();
|
|
|
|
|
|
|
|
// Inputs are re-evaluted before the post action, so we want the original key used for restore
|
|
|
|
const primaryKey = core.getState(State.CacheKey);
|
|
|
|
if (!primaryKey) {
|
2019-11-22 02:37:54 +07:00
|
|
|
utils.logWarning(`Error retrieving key from state.`);
|
2019-10-31 01:48:49 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (utils.isExactKeyMatch(primaryKey, state)) {
|
|
|
|
core.info(
|
|
|
|
`Cache hit occurred on the primary key ${primaryKey}, not saving cache.`
|
|
|
|
);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2020-01-07 01:05:50 +07:00
|
|
|
core.debug("Reserving Cache");
|
|
|
|
const cacheId = await cacheHttpClient.reserveCache(primaryKey);
|
|
|
|
if (cacheId == -1) {
|
|
|
|
core.info(
|
|
|
|
`Unable to reserve cache with key ${primaryKey}, another job may be creating this cache.`
|
|
|
|
);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
core.debug(`Cache ID: ${cacheId}`);
|
2019-11-15 05:14:16 +07:00
|
|
|
const cachePath = utils.resolvePath(
|
2019-10-31 01:48:49 +07:00
|
|
|
core.getInput(Inputs.Path, { required: true })
|
|
|
|
);
|
|
|
|
core.debug(`Cache Path: ${cachePath}`);
|
|
|
|
|
2019-11-15 05:14:16 +07:00
|
|
|
const archivePath = path.join(
|
2019-10-31 01:48:49 +07:00
|
|
|
await utils.createTempDirectory(),
|
|
|
|
"cache.tgz"
|
|
|
|
);
|
|
|
|
core.debug(`Archive Path: ${archivePath}`);
|
|
|
|
|
2019-12-14 05:24:37 +07:00
|
|
|
await createTar(archivePath, cachePath);
|
2019-10-31 01:48:49 +07:00
|
|
|
|
2020-01-07 01:05:50 +07:00
|
|
|
const fileSizeLimit = 2 * 1024 * 1024 * 1024; // 2GB per repo limit
|
2019-11-15 05:14:16 +07:00
|
|
|
const archiveFileSize = utils.getArchiveFileSize(archivePath);
|
2019-10-31 01:48:49 +07:00
|
|
|
core.debug(`File Size: ${archiveFileSize}`);
|
|
|
|
if (archiveFileSize > fileSizeLimit) {
|
2019-11-22 02:37:54 +07:00
|
|
|
utils.logWarning(
|
2019-11-13 23:00:46 +07:00
|
|
|
`Cache size of ~${Math.round(
|
|
|
|
archiveFileSize / (1024 * 1024)
|
2020-01-07 01:05:50 +07:00
|
|
|
)} MB (${archiveFileSize} B) is over the 2GB limit, not saving cache.`
|
2019-10-31 01:48:49 +07:00
|
|
|
);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2020-01-07 01:05:50 +07:00
|
|
|
core.debug(`Saving Cache (ID: ${cacheId})`);
|
|
|
|
await cacheHttpClient.saveCache(cacheId, archivePath);
|
2019-10-31 01:48:49 +07:00
|
|
|
} catch (error) {
|
2019-11-22 02:37:54 +07:00
|
|
|
utils.logWarning(error.message);
|
2019-10-31 01:48:49 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
run();
|
|
|
|
|
|
|
|
export default run;
|