mirror of
				https://github.com/actions/cache.git
				synced 2025-10-31 07:16:24 +07:00 
			
		
		
		
	Compare commits
	
		
			28 Commits
		
	
	
		
			anuragc617
			...
			joshmgross
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| cdec5dec0d | |||
| 2ce22df8c4 | |||
| 8c77f01f0b | |||
| 4fcbc07edb | |||
| 16019b42a9 | |||
| 73a15dc5a9 | |||
| 574cd74b58 | |||
| 14055801c2 | |||
| 289c5d2518 | |||
| ba6476e454 | |||
| b425e87f79 | |||
| 83f86c103f | |||
| 64668e22dd | |||
| 1c77f64ab3 | |||
| a70833fb48 | |||
| b25804d19e | |||
| 577b274c51 | |||
| 0816faf84c | |||
| 131e247bd2 | |||
| 2cbd952179 | |||
| 994e3b75fc | |||
| 21dc9a47e6 | |||
| 436418ea07 | |||
| 7f6523f535 | |||
| 4d3086b6b8 | |||
| d788427754 | |||
| a8adbe4b05 | |||
| bad827c28e | 
| @ -255,7 +255,10 @@ test("restore with cache found", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key]); |     expect(getCacheMock).toHaveBeenCalledWith([key]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
|     expect(mkdirMock).toHaveBeenCalledWith(cachePath); |     expect(mkdirMock).toHaveBeenCalledWith(cachePath); | ||||||
|  |  | ||||||
| @ -333,7 +336,10 @@ test("restore with a pull request event and cache found", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key]); |     expect(getCacheMock).toHaveBeenCalledWith([key]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
|     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~60 MB (62915000 B)`); |     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~60 MB (62915000 B)`); | ||||||
|     expect(mkdirMock).toHaveBeenCalledWith(cachePath); |     expect(mkdirMock).toHaveBeenCalledWith(cachePath); | ||||||
| @ -412,7 +418,10 @@ test("restore with cache found for restore key", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key, restoreKey]); |     expect(getCacheMock).toHaveBeenCalledWith([key, restoreKey]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
|     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~0 MB (142 B)`); |     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~0 MB (142 B)`); | ||||||
|     expect(mkdirMock).toHaveBeenCalledWith(cachePath); |     expect(mkdirMock).toHaveBeenCalledWith(cachePath); | ||||||
|  | |||||||
| @ -200,7 +200,7 @@ test("save with large cache outputs warning", async () => { | |||||||
|  |  | ||||||
|     const execMock = jest.spyOn(exec, "exec"); |     const execMock = jest.spyOn(exec, "exec"); | ||||||
|  |  | ||||||
|     const cacheSize = 1024 * 1024 * 1024; //~1GB, over the 400MB limit |     const cacheSize = 4 * 1024 * 1024 * 1024; //~4GB, over the 2GB limit | ||||||
|     jest.spyOn(actionUtils, "getArchiveFileSize").mockImplementationOnce(() => { |     jest.spyOn(actionUtils, "getArchiveFileSize").mockImplementationOnce(() => { | ||||||
|         return cacheSize; |         return cacheSize; | ||||||
|     }); |     }); | ||||||
| @ -212,14 +212,14 @@ test("save with large cache outputs warning", async () => { | |||||||
|     const IS_WINDOWS = process.platform === "win32"; |     const IS_WINDOWS = process.platform === "win32"; | ||||||
|     const args = IS_WINDOWS |     const args = IS_WINDOWS | ||||||
|         ? [ |         ? [ | ||||||
|               "-cz", |             "-cz", | ||||||
|               "--force-local", |             "--force-local", | ||||||
|               "-f", |             "-f", | ||||||
|               archivePath.replace(/\\/g, "/"), |             archivePath.replace(/\\/g, "/"), | ||||||
|               "-C", |             "-C", | ||||||
|               cachePath.replace(/\\/g, "/"), |             cachePath.replace(/\\/g, "/"), | ||||||
|               "." |             "." | ||||||
|           ] |         ] | ||||||
|         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; |         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; | ||||||
|  |  | ||||||
|     expect(execMock).toHaveBeenCalledTimes(1); |     expect(execMock).toHaveBeenCalledTimes(1); | ||||||
| @ -227,7 +227,7 @@ test("save with large cache outputs warning", async () => { | |||||||
|  |  | ||||||
|     expect(logWarningMock).toHaveBeenCalledTimes(1); |     expect(logWarningMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(logWarningMock).toHaveBeenCalledWith( |     expect(logWarningMock).toHaveBeenCalledWith( | ||||||
|         "Cache size of ~1024 MB (1073741824 B) is over the 400MB limit, not saving cache." |         "Cache size of ~4 GB (4294967296 B) is over the 2GB limit, not saving cache." | ||||||
|     ); |     ); | ||||||
|  |  | ||||||
|     expect(failedMock).toHaveBeenCalledTimes(0); |     expect(failedMock).toHaveBeenCalledTimes(0); | ||||||
| @ -259,6 +259,11 @@ test("save with server error outputs warning", async () => { | |||||||
|     const cachePath = path.resolve(inputPath); |     const cachePath = path.resolve(inputPath); | ||||||
|     testUtils.setInput(Inputs.Path, inputPath); |     testUtils.setInput(Inputs.Path, inputPath); | ||||||
|  |  | ||||||
|  |     const cacheId = 4; | ||||||
|  |     const reserveCacheMock = jest.spyOn(cacheHttpClient, "reserveCache").mockImplementationOnce(() => { | ||||||
|  |         return Promise.resolve(cacheId); | ||||||
|  |     }); | ||||||
|  |  | ||||||
|     const execMock = jest.spyOn(exec, "exec"); |     const execMock = jest.spyOn(exec, "exec"); | ||||||
|  |  | ||||||
|     const saveCacheMock = jest |     const saveCacheMock = jest | ||||||
| @ -269,26 +274,29 @@ test("save with server error outputs warning", async () => { | |||||||
|  |  | ||||||
|     await run(); |     await run(); | ||||||
|  |  | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledWith(primaryKey); | ||||||
|  |  | ||||||
|     const archivePath = path.join("/foo/bar", "cache.tgz"); |     const archivePath = path.join("/foo/bar", "cache.tgz"); | ||||||
|  |  | ||||||
|     const IS_WINDOWS = process.platform === "win32"; |     const IS_WINDOWS = process.platform === "win32"; | ||||||
|     const args = IS_WINDOWS |     const args = IS_WINDOWS | ||||||
|         ? [ |         ? [ | ||||||
|               "-cz", |             "-cz", | ||||||
|               "--force-local", |             "--force-local", | ||||||
|               "-f", |             "-f", | ||||||
|               archivePath.replace(/\\/g, "/"), |             archivePath.replace(/\\/g, "/"), | ||||||
|               "-C", |             "-C", | ||||||
|               cachePath.replace(/\\/g, "/"), |             cachePath.replace(/\\/g, "/"), | ||||||
|               "." |             "." | ||||||
|           ] |         ] | ||||||
|         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; |         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; | ||||||
|  |  | ||||||
|     expect(execMock).toHaveBeenCalledTimes(1); |     expect(execMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(execMock).toHaveBeenCalledWith(`"tar"`, args); |     expect(execMock).toHaveBeenCalledWith(`"tar"`, args); | ||||||
|  |  | ||||||
|     expect(saveCacheMock).toHaveBeenCalledTimes(1); |     expect(saveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(saveCacheMock).toHaveBeenCalledWith(primaryKey, archivePath); |     expect(saveCacheMock).toHaveBeenCalledWith(cacheId, archivePath); | ||||||
|  |  | ||||||
|     expect(logWarningMock).toHaveBeenCalledTimes(1); |     expect(logWarningMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(logWarningMock).toHaveBeenCalledWith("HTTP Error Occurred"); |     expect(logWarningMock).toHaveBeenCalledWith("HTTP Error Occurred"); | ||||||
| @ -321,32 +329,40 @@ test("save with valid inputs uploads a cache", async () => { | |||||||
|     const cachePath = path.resolve(inputPath); |     const cachePath = path.resolve(inputPath); | ||||||
|     testUtils.setInput(Inputs.Path, inputPath); |     testUtils.setInput(Inputs.Path, inputPath); | ||||||
|  |  | ||||||
|  |     const cacheId = 4; | ||||||
|  |     const reserveCacheMock = jest.spyOn(cacheHttpClient, "reserveCache").mockImplementationOnce(() => { | ||||||
|  |         return Promise.resolve(cacheId); | ||||||
|  |     }); | ||||||
|  |  | ||||||
|     const execMock = jest.spyOn(exec, "exec"); |     const execMock = jest.spyOn(exec, "exec"); | ||||||
|  |  | ||||||
|     const saveCacheMock = jest.spyOn(cacheHttpClient, "saveCache"); |     const saveCacheMock = jest.spyOn(cacheHttpClient, "saveCache"); | ||||||
|  |  | ||||||
|     await run(); |     await run(); | ||||||
|  |  | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledWith(primaryKey); | ||||||
|  |  | ||||||
|     const archivePath = path.join("/foo/bar", "cache.tgz"); |     const archivePath = path.join("/foo/bar", "cache.tgz"); | ||||||
|  |  | ||||||
|     const IS_WINDOWS = process.platform === "win32"; |     const IS_WINDOWS = process.platform === "win32"; | ||||||
|     const args = IS_WINDOWS |     const args = IS_WINDOWS | ||||||
|         ? [ |         ? [ | ||||||
|               "-cz", |             "-cz", | ||||||
|               "--force-local", |             "--force-local", | ||||||
|               "-f", |             "-f", | ||||||
|               archivePath.replace(/\\/g, "/"), |             archivePath.replace(/\\/g, "/"), | ||||||
|               "-C", |             "-C", | ||||||
|               cachePath.replace(/\\/g, "/"), |             cachePath.replace(/\\/g, "/"), | ||||||
|               "." |             "." | ||||||
|           ] |         ] | ||||||
|         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; |         : ["-cz", "-f", archivePath, "-C", cachePath, "."]; | ||||||
|  |  | ||||||
|     expect(execMock).toHaveBeenCalledTimes(1); |     expect(execMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(execMock).toHaveBeenCalledWith(`"tar"`, args); |     expect(execMock).toHaveBeenCalledWith(`"tar"`, args); | ||||||
|  |  | ||||||
|     expect(saveCacheMock).toHaveBeenCalledTimes(1); |     expect(saveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(saveCacheMock).toHaveBeenCalledWith(primaryKey, archivePath); |     expect(saveCacheMock).toHaveBeenCalledWith(cacheId, archivePath); | ||||||
|  |  | ||||||
|     expect(failedMock).toHaveBeenCalledTimes(0); |     expect(failedMock).toHaveBeenCalledTimes(0); | ||||||
| }); | }); | ||||||
|  | |||||||
							
								
								
									
										5287
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										5287
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							
							
								
								
									
										5267
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										5267
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							
							
								
								
									
										12
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										12
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							| @ -4859,9 +4859,9 @@ | |||||||
|       "dev": true |       "dev": true | ||||||
|     }, |     }, | ||||||
|     "prettier": { |     "prettier": { | ||||||
|       "version": "1.18.2", |       "version": "1.19.1", | ||||||
|       "resolved": "https://registry.npmjs.org/prettier/-/prettier-1.18.2.tgz", |       "resolved": "https://registry.npmjs.org/prettier/-/prettier-1.19.1.tgz", | ||||||
|       "integrity": "sha512-OeHeMc0JhFE9idD4ZdtNibzY0+TPHSpSSb9h8FqtP+YnoZZ1sl8Vc9b1sasjfymH3SonAF4QcA2+mzHPhMvIiw==", |       "integrity": "sha512-s7PoyDv/II1ObgQunCbB9PdLmUcBZcnWOcxDh7O0N/UwDEsHyqkW+Qh28jW+mVuCdx7gLB0BotYI1Y6uI9iyew==", | ||||||
|       "dev": true |       "dev": true | ||||||
|     }, |     }, | ||||||
|     "prettier-linter-helpers": { |     "prettier-linter-helpers": { | ||||||
| @ -5983,9 +5983,9 @@ | |||||||
|       } |       } | ||||||
|     }, |     }, | ||||||
|     "typescript": { |     "typescript": { | ||||||
|       "version": "3.6.4", |       "version": "3.7.3", | ||||||
|       "resolved": "https://registry.npmjs.org/typescript/-/typescript-3.6.4.tgz", |       "resolved": "https://registry.npmjs.org/typescript/-/typescript-3.7.3.tgz", | ||||||
|       "integrity": "sha512-unoCll1+l+YK4i4F8f22TaNVPRHcD9PA3yCuZ8g5e0qGqlVlJ/8FSateOLLSagn+Yg5+ZwuPkL8LFUc0Jcvksg==", |       "integrity": "sha512-Mcr/Qk7hXqFBXMN7p7Lusj1ktCBydylfQM/FZCk5glCNQJrCUKPkMHdo9R0MTFWsC/4kPFvDS0fDPvukfCkFsw==", | ||||||
|       "dev": true |       "dev": true | ||||||
|     }, |     }, | ||||||
|     "uglify-js": { |     "uglify-js": { | ||||||
|  | |||||||
| @ -1,6 +1,6 @@ | |||||||
| { | { | ||||||
|   "name": "cache", |   "name": "cache", | ||||||
|   "version": "1.0.3", |   "version": "1.1.0", | ||||||
|   "private": true, |   "private": true, | ||||||
|   "description": "Cache dependencies and build outputs", |   "description": "Cache dependencies and build outputs", | ||||||
|   "main": "dist/restore/index.js", |   "main": "dist/restore/index.js", | ||||||
| @ -46,8 +46,8 @@ | |||||||
|     "jest": "^24.8.0", |     "jest": "^24.8.0", | ||||||
|     "jest-circus": "^24.7.1", |     "jest-circus": "^24.7.1", | ||||||
|     "nock": "^11.7.0", |     "nock": "^11.7.0", | ||||||
|     "prettier": "1.18.2", |     "prettier": "^1.19.1", | ||||||
|     "ts-jest": "^24.0.2", |     "ts-jest": "^24.0.2", | ||||||
|     "typescript": "^3.6.4" |     "typescript": "^3.7.3" | ||||||
|   } |   } | ||||||
| } | } | ||||||
|  | |||||||
| @ -3,24 +3,37 @@ import * as fs from "fs"; | |||||||
| import { BearerCredentialHandler } from "typed-rest-client/Handlers"; | import { BearerCredentialHandler } from "typed-rest-client/Handlers"; | ||||||
| import { HttpClient } from "typed-rest-client/HttpClient"; | import { HttpClient } from "typed-rest-client/HttpClient"; | ||||||
| import { IHttpClientResponse } from "typed-rest-client/Interfaces"; | import { IHttpClientResponse } from "typed-rest-client/Interfaces"; | ||||||
| import { IRequestOptions, RestClient } from "typed-rest-client/RestClient"; | import { | ||||||
| import { ArtifactCacheEntry } from "./contracts"; |     IRequestOptions, | ||||||
|  |     RestClient, | ||||||
|  |     IRestResponse | ||||||
|  | } from "typed-rest-client/RestClient"; | ||||||
|  | import { | ||||||
|  |     ArtifactCacheEntry, | ||||||
|  |     CommitCacheRequest, | ||||||
|  |     ReserveCacheRequest, | ||||||
|  |     ReserverCacheResponse | ||||||
|  | } from "./contracts"; | ||||||
|  | import * as utils from "./utils/actionUtils"; | ||||||
|  |  | ||||||
| function getCacheUrl(): string { | function isSuccessStatusCode(statusCode: number): boolean { | ||||||
|  |     return statusCode >= 200 && statusCode < 300; | ||||||
|  | } | ||||||
|  | function getCacheApiUrl(): string { | ||||||
|     // Ideally we just use ACTIONS_CACHE_URL |     // Ideally we just use ACTIONS_CACHE_URL | ||||||
|     const cacheUrl: string = ( |     const baseUrl: string = ( | ||||||
|         process.env["ACTIONS_CACHE_URL"] || |         process.env["ACTIONS_CACHE_URL"] || | ||||||
|         process.env["ACTIONS_RUNTIME_URL"] || |         process.env["ACTIONS_RUNTIME_URL"] || | ||||||
|         "" |         "" | ||||||
|     ).replace("pipelines", "artifactcache"); |     ).replace("pipelines", "artifactcache"); | ||||||
|     if (!cacheUrl) { |     if (!baseUrl) { | ||||||
|         throw new Error( |         throw new Error( | ||||||
|             "Cache Service Url not found, unable to restore cache." |             "Cache Service Url not found, unable to restore cache." | ||||||
|         ); |         ); | ||||||
|     } |     } | ||||||
|  |  | ||||||
|     core.debug(`Cache Url: ${cacheUrl}`); |     core.debug(`Cache Url: ${baseUrl}`); | ||||||
|     return cacheUrl; |     return `${baseUrl}_apis/artifactcache/`; | ||||||
| } | } | ||||||
|  |  | ||||||
| function createAcceptHeader(type: string, apiVersion: string): string { | function createAcceptHeader(type: string, apiVersion: string): string { | ||||||
| @ -29,26 +42,26 @@ function createAcceptHeader(type: string, apiVersion: string): string { | |||||||
|  |  | ||||||
| function getRequestOptions(): IRequestOptions { | function getRequestOptions(): IRequestOptions { | ||||||
|     const requestOptions: IRequestOptions = { |     const requestOptions: IRequestOptions = { | ||||||
|         acceptHeader: createAcceptHeader("application/json", "5.2-preview.1") |         acceptHeader: createAcceptHeader("application/json", "6.0-preview.1") | ||||||
|     }; |     }; | ||||||
|  |  | ||||||
|     return requestOptions; |     return requestOptions; | ||||||
| } | } | ||||||
|  |  | ||||||
| export async function getCacheEntry( | function createRestClient(): RestClient { | ||||||
|     keys: string[] |  | ||||||
| ): Promise<ArtifactCacheEntry | null> { |  | ||||||
|     const cacheUrl = getCacheUrl(); |  | ||||||
|     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; |     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; | ||||||
|     const bearerCredentialHandler = new BearerCredentialHandler(token); |     const bearerCredentialHandler = new BearerCredentialHandler(token); | ||||||
|  |  | ||||||
|     const resource = `_apis/artifactcache/cache?keys=${encodeURIComponent( |     return new RestClient("actions/cache", getCacheApiUrl(), [ | ||||||
|         keys.join(",") |  | ||||||
|     )}`; |  | ||||||
|  |  | ||||||
|     const restClient = new RestClient("actions/cache", cacheUrl, [ |  | ||||||
|         bearerCredentialHandler |         bearerCredentialHandler | ||||||
|     ]); |     ]); | ||||||
|  | } | ||||||
|  |  | ||||||
|  | export async function getCacheEntry( | ||||||
|  |     keys: string[] | ||||||
|  | ): Promise<ArtifactCacheEntry | null> { | ||||||
|  |     const restClient = createRestClient(); | ||||||
|  |     const resource = `cache?keys=${encodeURIComponent(keys.join(","))}`; | ||||||
|  |  | ||||||
|     const response = await restClient.get<ArtifactCacheEntry>( |     const response = await restClient.get<ArtifactCacheEntry>( | ||||||
|         resource, |         resource, | ||||||
| @ -57,14 +70,15 @@ export async function getCacheEntry( | |||||||
|     if (response.statusCode === 204) { |     if (response.statusCode === 204) { | ||||||
|         return null; |         return null; | ||||||
|     } |     } | ||||||
|     if (response.statusCode !== 200) { |     if (!isSuccessStatusCode(response.statusCode)) { | ||||||
|         throw new Error(`Cache service responded with ${response.statusCode}`); |         throw new Error(`Cache service responded with ${response.statusCode}`); | ||||||
|     } |     } | ||||||
|     const cacheResult = response.result; |     const cacheResult = response.result; | ||||||
|     if (!cacheResult || !cacheResult.archiveLocation) { |     const cacheDownloadUrl = cacheResult?.archiveLocation; | ||||||
|  |     if (!cacheDownloadUrl) { | ||||||
|         throw new Error("Cache not found."); |         throw new Error("Cache not found."); | ||||||
|     } |     } | ||||||
|     core.setSecret(cacheResult.archiveLocation); |     core.setSecret(cacheDownloadUrl); | ||||||
|     core.debug(`Cache Result:`); |     core.debug(`Cache Result:`); | ||||||
|     core.debug(JSON.stringify(cacheResult)); |     core.debug(JSON.stringify(cacheResult)); | ||||||
|  |  | ||||||
| @ -83,46 +97,165 @@ async function pipeResponseToStream( | |||||||
| } | } | ||||||
|  |  | ||||||
| export async function downloadCache( | export async function downloadCache( | ||||||
|     cacheEntry: ArtifactCacheEntry, |     archiveLocation: string, | ||||||
|     archivePath: string |     archivePath: string | ||||||
| ): Promise<void> { | ): Promise<void> { | ||||||
|     const stream = fs.createWriteStream(archivePath); |     const stream = fs.createWriteStream(archivePath); | ||||||
|     const httpClient = new HttpClient("actions/cache"); |     const httpClient = new HttpClient("actions/cache"); | ||||||
|     // eslint-disable-next-line @typescript-eslint/no-non-null-assertion |     const downloadResponse = await httpClient.get(archiveLocation); | ||||||
|     const downloadResponse = await httpClient.get(cacheEntry.archiveLocation!); |  | ||||||
|     await pipeResponseToStream(downloadResponse, stream); |     await pipeResponseToStream(downloadResponse, stream); | ||||||
| } | } | ||||||
|  |  | ||||||
| export async function saveCache( | // Reserve Cache | ||||||
|     key: string, | export async function reserveCache(key: string): Promise<number> { | ||||||
|     archivePath: string |     const restClient = createRestClient(); | ||||||
| ): Promise<void> { |  | ||||||
|     const stream = fs.createReadStream(archivePath); |  | ||||||
|  |  | ||||||
|     const cacheUrl = getCacheUrl(); |     const reserveCacheRequest: ReserveCacheRequest = { | ||||||
|     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; |         key | ||||||
|     const bearerCredentialHandler = new BearerCredentialHandler(token); |     }; | ||||||
|  |     const response = await restClient.create<ReserverCacheResponse>( | ||||||
|  |         "caches", | ||||||
|  |         reserveCacheRequest, | ||||||
|  |         getRequestOptions() | ||||||
|  |     ); | ||||||
|  |  | ||||||
|     const resource = `_apis/artifactcache/cache/${encodeURIComponent(key)}`; |     return response?.result?.cacheId ?? -1; | ||||||
|     const postUrl = cacheUrl + resource; | } | ||||||
|  |  | ||||||
|     const restClient = new RestClient("actions/cache", undefined, [ | function getContentRange(start: number, end: number): string { | ||||||
|         bearerCredentialHandler |     // Format: `bytes start-end/filesize | ||||||
|     ]); |     // start and end are inclusive | ||||||
|  |     // filesize can be * | ||||||
|  |     // For a 200 byte chunk starting at byte 0: | ||||||
|  |     // Content-Range: bytes 0-199/* | ||||||
|  |     return `bytes ${start}-${end}/*`; | ||||||
|  | } | ||||||
|  |  | ||||||
|  | async function uploadChunk( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     resourceUrl: string, | ||||||
|  |     data: NodeJS.ReadableStream, | ||||||
|  |     start: number, | ||||||
|  |     end: number | ||||||
|  | ): Promise<IRestResponse<void>> { | ||||||
|  |     core.debug( | ||||||
|  |         `Uploading chunk of size ${end - | ||||||
|  |             start + | ||||||
|  |             1} bytes at offset ${start} with content range: ${getContentRange( | ||||||
|  |             start, | ||||||
|  |             end | ||||||
|  |         )}` | ||||||
|  |     ); | ||||||
|     const requestOptions = getRequestOptions(); |     const requestOptions = getRequestOptions(); | ||||||
|     requestOptions.additionalHeaders = { |     requestOptions.additionalHeaders = { | ||||||
|         "Content-Type": "application/octet-stream" |         "Content-Type": "application/octet-stream", | ||||||
|  |         "Content-Range": getContentRange(start, end) | ||||||
|     }; |     }; | ||||||
|  |  | ||||||
|     const response = await restClient.uploadStream<void>( |     return await restClient.uploadStream<void>( | ||||||
|         "POST", |         "PATCH", | ||||||
|         postUrl, |         resourceUrl, | ||||||
|         stream, |         data, | ||||||
|         requestOptions |         requestOptions | ||||||
|     ); |     ); | ||||||
|     if (response.statusCode !== 200) { | } | ||||||
|         throw new Error(`Cache service responded with ${response.statusCode}`); |  | ||||||
|  | async function uploadFile( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     cacheId: number, | ||||||
|  |     archivePath: string | ||||||
|  | ): Promise<void> { | ||||||
|  |     // Upload Chunks | ||||||
|  |     const fileSize = fs.statSync(archivePath).size; | ||||||
|  |     const resourceUrl = getCacheApiUrl() + "caches/" + cacheId.toString(); | ||||||
|  |     const responses: IRestResponse<void>[] = []; | ||||||
|  |     const fd = fs.openSync(archivePath, "r"); | ||||||
|  |  | ||||||
|  |     const concurrency = 4; // # of HTTP requests in parallel | ||||||
|  |     const MAX_CHUNK_SIZE = 32000000; // 32 MB Chunks | ||||||
|  |     core.debug(`Concurrency: ${concurrency} and Chunk Size: ${MAX_CHUNK_SIZE}`); | ||||||
|  |  | ||||||
|  |     const parallelUploads = [...new Array(concurrency).keys()]; | ||||||
|  |     core.debug("Awaiting all uploads"); | ||||||
|  |     let offset = 0; | ||||||
|  |     await Promise.all( | ||||||
|  |         parallelUploads.map(async () => { | ||||||
|  |             while (offset < fileSize) { | ||||||
|  |                 const chunkSize = | ||||||
|  |                     offset + MAX_CHUNK_SIZE > fileSize | ||||||
|  |                         ? fileSize - offset | ||||||
|  |                         : MAX_CHUNK_SIZE; | ||||||
|  |                 const start = offset; | ||||||
|  |                 const end = offset + chunkSize - 1; | ||||||
|  |                 offset += MAX_CHUNK_SIZE; | ||||||
|  |                 const chunk = fs.createReadStream(archivePath, { | ||||||
|  |                     fd, | ||||||
|  |                     start, | ||||||
|  |                     end, | ||||||
|  |                     autoClose: false | ||||||
|  |                 }); | ||||||
|  |                 responses.push( | ||||||
|  |                     await uploadChunk( | ||||||
|  |                         restClient, | ||||||
|  |                         resourceUrl, | ||||||
|  |                         chunk, | ||||||
|  |                         start, | ||||||
|  |                         end | ||||||
|  |                     ) | ||||||
|  |                 ); | ||||||
|  |             } | ||||||
|  |         }) | ||||||
|  |     ); | ||||||
|  |  | ||||||
|  |     fs.closeSync(fd); | ||||||
|  |  | ||||||
|  |     const failedResponse = responses.find( | ||||||
|  |         x => !isSuccessStatusCode(x.statusCode) | ||||||
|  |     ); | ||||||
|  |     if (failedResponse) { | ||||||
|  |         throw new Error( | ||||||
|  |             `Cache service responded with ${failedResponse.statusCode} during chunk upload.` | ||||||
|  |         ); | ||||||
|  |     } | ||||||
|  |  | ||||||
|  |     return; | ||||||
|  | } | ||||||
|  |  | ||||||
|  | async function commitCache( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     cacheId: number, | ||||||
|  |     filesize: number | ||||||
|  | ): Promise<IRestResponse<void>> { | ||||||
|  |     const requestOptions = getRequestOptions(); | ||||||
|  |     const commitCacheRequest: CommitCacheRequest = { size: filesize }; | ||||||
|  |     return await restClient.create( | ||||||
|  |         `caches/${cacheId.toString()}`, | ||||||
|  |         commitCacheRequest, | ||||||
|  |         requestOptions | ||||||
|  |     ); | ||||||
|  | } | ||||||
|  |  | ||||||
|  | export async function saveCache( | ||||||
|  |     cacheId: number, | ||||||
|  |     archivePath: string | ||||||
|  | ): Promise<void> { | ||||||
|  |     const restClient = createRestClient(); | ||||||
|  |  | ||||||
|  |     core.debug("Upload cache"); | ||||||
|  |     await uploadFile(restClient, cacheId, archivePath); | ||||||
|  |  | ||||||
|  |     // Commit Cache | ||||||
|  |     core.debug("Commiting cache"); | ||||||
|  |     const cacheSize = utils.getArchiveFileSize(archivePath); | ||||||
|  |     const commitCacheResponse = await commitCache( | ||||||
|  |         restClient, | ||||||
|  |         cacheId, | ||||||
|  |         cacheSize | ||||||
|  |     ); | ||||||
|  |     if (!isSuccessStatusCode(commitCacheResponse.statusCode)) { | ||||||
|  |         throw new Error( | ||||||
|  |             `Cache service responded with ${commitCacheResponse.statusCode} during commit cache.` | ||||||
|  |         ); | ||||||
|     } |     } | ||||||
|  |  | ||||||
|     core.info("Cache saved successfully"); |     core.info("Cache saved successfully"); | ||||||
|  | |||||||
							
								
								
									
										13
									
								
								src/contracts.d.ts
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										13
									
								
								src/contracts.d.ts
									
									
									
									
										vendored
									
									
								
							| @ -4,3 +4,16 @@ export interface ArtifactCacheEntry { | |||||||
|     creationTime?: string; |     creationTime?: string; | ||||||
|     archiveLocation?: string; |     archiveLocation?: string; | ||||||
| } | } | ||||||
|  |  | ||||||
|  | export interface CommitCacheRequest { | ||||||
|  |     size: number; | ||||||
|  | } | ||||||
|  |  | ||||||
|  | export interface ReserveCacheRequest { | ||||||
|  |     key: string; | ||||||
|  |     version?: string; | ||||||
|  | } | ||||||
|  |  | ||||||
|  | export interface ReserverCacheResponse { | ||||||
|  |     cacheId: number; | ||||||
|  | } | ||||||
|  | |||||||
| @ -61,7 +61,7 @@ async function run(): Promise<void> { | |||||||
|  |  | ||||||
|         try { |         try { | ||||||
|             const cacheEntry = await cacheHttpClient.getCacheEntry(keys); |             const cacheEntry = await cacheHttpClient.getCacheEntry(keys); | ||||||
|             if (!cacheEntry) { |             if (!cacheEntry || !cacheEntry?.archiveLocation) { | ||||||
|                 core.info( |                 core.info( | ||||||
|                     `Cache not found for input keys: ${keys.join(", ")}.` |                     `Cache not found for input keys: ${keys.join(", ")}.` | ||||||
|                 ); |                 ); | ||||||
| @ -78,7 +78,10 @@ async function run(): Promise<void> { | |||||||
|             utils.setCacheState(cacheEntry); |             utils.setCacheState(cacheEntry); | ||||||
|  |  | ||||||
|             // Download the cache from the cache entry |             // Download the cache from the cache entry | ||||||
|             await cacheHttpClient.downloadCache(cacheEntry, archivePath); |             await cacheHttpClient.downloadCache( | ||||||
|  |                 cacheEntry?.archiveLocation, | ||||||
|  |                 archivePath | ||||||
|  |             ); | ||||||
|  |  | ||||||
|             const archiveFileSize = utils.getArchiveFileSize(archivePath); |             const archiveFileSize = utils.getArchiveFileSize(archivePath); | ||||||
|             core.info( |             core.info( | ||||||
|  | |||||||
							
								
								
									
										18
									
								
								src/save.ts
									
									
									
									
									
								
							
							
						
						
									
										18
									
								
								src/save.ts
									
									
									
									
									
								
							| @ -35,6 +35,15 @@ async function run(): Promise<void> { | |||||||
|             return; |             return; | ||||||
|         } |         } | ||||||
|  |  | ||||||
|  |         core.debug("Reserving Cache"); | ||||||
|  |         const cacheId = await cacheHttpClient.reserveCache(primaryKey); | ||||||
|  |         if (cacheId < 0) { | ||||||
|  |             core.info( | ||||||
|  |                 `Unable to reserve cache with key ${primaryKey}, another job may be creating this cache.` | ||||||
|  |             ); | ||||||
|  |             return; | ||||||
|  |         } | ||||||
|  |         core.debug(`Cache ID: ${cacheId}`); | ||||||
|         const cachePath = utils.resolvePath( |         const cachePath = utils.resolvePath( | ||||||
|             core.getInput(Inputs.Path, { required: true }) |             core.getInput(Inputs.Path, { required: true }) | ||||||
|         ); |         ); | ||||||
| @ -65,19 +74,20 @@ async function run(): Promise<void> { | |||||||
|         core.debug(`Tar Path: ${tarPath}`); |         core.debug(`Tar Path: ${tarPath}`); | ||||||
|         await exec(`"${tarPath}"`, args); |         await exec(`"${tarPath}"`, args); | ||||||
|  |  | ||||||
|         const fileSizeLimit = 400 * 1024 * 1024; // 400MB |         const fileSizeLimit = 2 * 1024 * 1024 * 1024; // 2GB per repo limit | ||||||
|         const archiveFileSize = utils.getArchiveFileSize(archivePath); |         const archiveFileSize = utils.getArchiveFileSize(archivePath); | ||||||
|         core.debug(`File Size: ${archiveFileSize}`); |         core.debug(`File Size: ${archiveFileSize}`); | ||||||
|         if (archiveFileSize > fileSizeLimit) { |         if (archiveFileSize > fileSizeLimit) { | ||||||
|             utils.logWarning( |             utils.logWarning( | ||||||
|                 `Cache size of ~${Math.round( |                 `Cache size of ~${Math.round( | ||||||
|                     archiveFileSize / (1024 * 1024) |                     archiveFileSize / (1024 * 1024 * 1024) | ||||||
|                 )} MB (${archiveFileSize} B) is over the 400MB limit, not saving cache.` |                 )} GB (${archiveFileSize} B) is over the 2GB limit, not saving cache.` | ||||||
|             ); |             ); | ||||||
|             return; |             return; | ||||||
|         } |         } | ||||||
|  |  | ||||||
|         await cacheHttpClient.saveCache(primaryKey, archivePath); |         core.debug("Saving Cache"); | ||||||
|  |         await cacheHttpClient.saveCache(cacheId, archivePath); | ||||||
|     } catch (error) { |     } catch (error) { | ||||||
|         utils.logWarning(error.message); |         utils.logWarning(error.message); | ||||||
|     } |     } | ||||||
|  | |||||||
		Reference in New Issue
	
	Block a user
	