mirror of
				https://gitea.com/actions/cache.git
				synced 2025-11-04 07:47:08 +00:00 
			
		
		
		
	Compare commits
	
		
			27 Commits
		
	
	
		
			v3.0.3
			...
			tiwarishub
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| 
						 | 
					402b512df8 | ||
| 
						 | 
					71e3ee5cce | ||
| 
						 | 
					c316eb7911 | ||
| 
						 | 
					0865c47f36 | ||
| 
						 | 
					354a2ae15e | ||
| 
						 | 
					baed3516c3 | ||
| 
						 | 
					8829e97be1 | ||
| 
						 | 
					eec8cd3f5f | ||
| 
						 | 
					5cc84c0123 | ||
| 
						 | 
					afc669e7fc | ||
| 
						 | 
					a0efc56c52 | ||
| 
						 | 
					d25c51bbfd | ||
| 
						 | 
					a080a3bda4 | ||
| 
						 | 
					02be3a9c73 | ||
| 
						 | 
					c3f1317a9e | ||
| 
						 | 
					d0a54b996c | ||
| 
						 | 
					8c5bd0c82d | ||
| 
						 | 
					c9c0f73558 | ||
| 
						 | 
					2b6caae3c9 | ||
| 
						 | 
					dd58d1315f | ||
| 
						 | 
					acace7fa1b | ||
| 
						 | 
					438628ac22 | ||
| 
						 | 
					c296e6a08c | ||
| 
						 | 
					7ed7f22ed8 | ||
| 
						 | 
					3767bf0386 | ||
| 
						 | 
					2086306d9c | ||
| 
						 | 
					009fe4e01c | 
							
								
								
									
										1
									
								
								.github/auto_assign.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										1
									
								
								.github/auto_assign.yml
									
									
									
									
										vendored
									
									
								
							@@ -7,6 +7,7 @@ addAssignees: false
 | 
			
		||||
# A list of reviewers to be added to pull requests (GitHub user name)
 | 
			
		||||
reviewers:
 | 
			
		||||
  - phantsure
 | 
			
		||||
  - kotewar
 | 
			
		||||
  - aparna-ravindra
 | 
			
		||||
  - tiwarishub
 | 
			
		||||
  - vsvipul
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										2
									
								
								.github/workflows/auto-assign-issues.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.github/workflows/auto-assign-issues.yml
									
									
									
									
										vendored
									
									
								
							@@ -11,5 +11,5 @@ jobs:
 | 
			
		||||
            - name: 'Auto-assign issue'
 | 
			
		||||
              uses: pozil/auto-assign-issue@v1.4.0
 | 
			
		||||
              with:
 | 
			
		||||
                  assignees: phantsure,tiwarishub,aparna-ravindra,vsvipul,bishal-pdmsft
 | 
			
		||||
                  assignees: phantsure,kotewar,tiwarishub,aparna-ravindra,vsvipul,bishal-pdmsft
 | 
			
		||||
                  numOfAssignee: 1
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										2
									
								
								.github/workflows/auto-assign.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.github/workflows/auto-assign.yml
									
									
									
									
										vendored
									
									
								
							@@ -1,6 +1,6 @@
 | 
			
		||||
name: 'Auto Assign'
 | 
			
		||||
on:
 | 
			
		||||
  pull_request:
 | 
			
		||||
  pull_request_target:
 | 
			
		||||
    types: [opened, ready_for_review]
 | 
			
		||||
 | 
			
		||||
jobs:
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										2
									
								
								.licenses/npm/@actions/cache.dep.yml
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										2
									
								
								.licenses/npm/@actions/cache.dep.yml
									
									
									
										generated
									
									
									
								
							@@ -1,6 +1,6 @@
 | 
			
		||||
---
 | 
			
		||||
name: "@actions/cache"
 | 
			
		||||
version: 2.0.5
 | 
			
		||||
version: 3.0.0
 | 
			
		||||
type: npm
 | 
			
		||||
summary: 
 | 
			
		||||
homepage: 
 | 
			
		||||
 
 | 
			
		||||
@@ -14,6 +14,7 @@ See ["Caching dependencies to speed up workflows"](https://help.github.com/githu
 | 
			
		||||
* Fixed download issue for files > 2GB during restore.
 | 
			
		||||
* Updated the minimum runner version support from node 12 -> node 16.
 | 
			
		||||
* Fixed avoiding empty cache save when no files are available for caching.
 | 
			
		||||
* Fixed tar creation error while trying to create tar with path as `~/` home folder on `ubuntu-latest`.
 | 
			
		||||
 | 
			
		||||
Refer [here](https://github.com/actions/cache/blob/v2/README.md) for previous versions
 | 
			
		||||
 | 
			
		||||
@@ -28,7 +29,8 @@ If you are using this inside a container, a POSIX-compliant `tar` needs to be in
 | 
			
		||||
 | 
			
		||||
* `path` - A list of files, directories, and wildcard patterns to cache and restore. See [`@actions/glob`](https://github.com/actions/toolkit/tree/main/packages/glob) for supported patterns.
 | 
			
		||||
* `key` - An explicit key for restoring and saving the cache
 | 
			
		||||
* `restore-keys` - An ordered list of keys to use for restoring the cache if no cache hit occurred for key
 | 
			
		||||
* `restore-keys` - An ordered list of keys to use for restoring stale cache if no cache hit occurred for key. Note
 | 
			
		||||
`cache-hit` returns false in this case.
 | 
			
		||||
 | 
			
		||||
### Outputs
 | 
			
		||||
 | 
			
		||||
@@ -70,6 +72,8 @@ jobs:
 | 
			
		||||
      run: /primes.sh -d prime-numbers
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
> Note: You must use the `cache` action in your workflow before you need to use the files that might be restored from the cache. If the provided `key` doesn't match an existing cache, a new cache is automatically created if the job completes successfully.
 | 
			
		||||
 | 
			
		||||
## Implementation Examples
 | 
			
		||||
 | 
			
		||||
Every programming language and framework has its own way of caching.
 | 
			
		||||
 
 | 
			
		||||
@@ -12,3 +12,9 @@
 | 
			
		||||
 | 
			
		||||
### 3.0.3
 | 
			
		||||
- Fixed avoiding empty cache save when no files are available for caching. ([issue](https://github.com/actions/cache/issues/624))
 | 
			
		||||
 | 
			
		||||
### 3.0.4
 | 
			
		||||
- Fixed tar creation error while trying to create tar with path as `~/` home folder on `ubuntu-latest`. ([issue](https://github.com/actions/cache/issues/689))
 | 
			
		||||
 | 
			
		||||
### 3.0.5
 | 
			
		||||
- Removed error handling by consuming actions/cache 3.0 toolkit, Now cache server error handling will be done by toolkit. ([PR](https://github.com/actions/cache/pull/834))
 | 
			
		||||
@@ -227,40 +227,6 @@ test("restore with no cache found", async () => {
 | 
			
		||||
    );
 | 
			
		||||
});
 | 
			
		||||
 | 
			
		||||
test("restore with server error should fail", async () => {
 | 
			
		||||
    const path = "node_modules";
 | 
			
		||||
    const key = "node-test";
 | 
			
		||||
    testUtils.setInputs({
 | 
			
		||||
        path: path,
 | 
			
		||||
        key
 | 
			
		||||
    });
 | 
			
		||||
 | 
			
		||||
    const logWarningMock = jest.spyOn(actionUtils, "logWarning");
 | 
			
		||||
    const failedMock = jest.spyOn(core, "setFailed");
 | 
			
		||||
    const stateMock = jest.spyOn(core, "saveState");
 | 
			
		||||
    const restoreCacheMock = jest
 | 
			
		||||
        .spyOn(cache, "restoreCache")
 | 
			
		||||
        .mockImplementationOnce(() => {
 | 
			
		||||
            throw new Error("HTTP Error Occurred");
 | 
			
		||||
        });
 | 
			
		||||
    const setCacheHitOutputMock = jest.spyOn(actionUtils, "setCacheHitOutput");
 | 
			
		||||
 | 
			
		||||
    await run();
 | 
			
		||||
 | 
			
		||||
    expect(restoreCacheMock).toHaveBeenCalledTimes(1);
 | 
			
		||||
    expect(restoreCacheMock).toHaveBeenCalledWith([path], key, []);
 | 
			
		||||
 | 
			
		||||
    expect(stateMock).toHaveBeenCalledWith("CACHE_KEY", key);
 | 
			
		||||
 | 
			
		||||
    expect(logWarningMock).toHaveBeenCalledTimes(1);
 | 
			
		||||
    expect(logWarningMock).toHaveBeenCalledWith("HTTP Error Occurred");
 | 
			
		||||
 | 
			
		||||
    expect(setCacheHitOutputMock).toHaveBeenCalledTimes(1);
 | 
			
		||||
    expect(setCacheHitOutputMock).toHaveBeenCalledWith(false);
 | 
			
		||||
 | 
			
		||||
    expect(failedMock).toHaveBeenCalledTimes(0);
 | 
			
		||||
});
 | 
			
		||||
 | 
			
		||||
test("restore with restore keys and no cache found", async () => {
 | 
			
		||||
    const path = "node_modules";
 | 
			
		||||
    const key = "node-test";
 | 
			
		||||
 
 | 
			
		||||
@@ -267,7 +267,6 @@ test("save with large cache outputs warning", async () => {
 | 
			
		||||
});
 | 
			
		||||
 | 
			
		||||
test("save with reserve cache failure outputs warning", async () => {
 | 
			
		||||
    const infoMock = jest.spyOn(core, "info");
 | 
			
		||||
    const logWarningMock = jest.spyOn(actionUtils, "logWarning");
 | 
			
		||||
    const failedMock = jest.spyOn(core, "setFailed");
 | 
			
		||||
 | 
			
		||||
@@ -306,10 +305,10 @@ test("save with reserve cache failure outputs warning", async () => {
 | 
			
		||||
        expect.anything()
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    expect(infoMock).toHaveBeenCalledWith(
 | 
			
		||||
    expect(logWarningMock).toHaveBeenCalledWith(
 | 
			
		||||
        `Unable to reserve cache with key ${primaryKey}, another job may be creating this cache.`
 | 
			
		||||
    );
 | 
			
		||||
    expect(logWarningMock).toHaveBeenCalledTimes(0);
 | 
			
		||||
    expect(logWarningMock).toHaveBeenCalledTimes(1);
 | 
			
		||||
    expect(failedMock).toHaveBeenCalledTimes(0);
 | 
			
		||||
});
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
@@ -9,7 +9,7 @@ inputs:
 | 
			
		||||
    description: 'An explicit key for restoring and saving the cache'
 | 
			
		||||
    required: true
 | 
			
		||||
  restore-keys:
 | 
			
		||||
    description: 'An ordered list of keys to use for restoring the cache if no cache hit occurred for key'
 | 
			
		||||
    description: 'An ordered list of keys to use for restoring stale cache if no cache hit occurred for key. Note `cache-hit` returns false in this case.'
 | 
			
		||||
    required: false
 | 
			
		||||
  upload-chunk-size:
 | 
			
		||||
    description: 'The chunk size used to split up large files during upload, in bytes'
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										
											BIN
										
									
								
								actions-cache-3.0.1.tgz
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										
											BIN
										
									
								
								actions-cache-3.0.1.tgz
									
									
									
									
									
										Normal file
									
								
							
										
											Binary file not shown.
										
									
								
							
							
								
								
									
										80
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										80
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							@@ -1113,9 +1113,15 @@ function resolvePaths(patterns) {
 | 
			
		||||
                    .replace(new RegExp(`\\${path.sep}`, 'g'), '/');
 | 
			
		||||
                core.debug(`Matched: ${relativeFile}`);
 | 
			
		||||
                // Paths are made relative so the tar entries are all relative to the root of the workspace.
 | 
			
		||||
                if (relativeFile === '') {
 | 
			
		||||
                    // path.relative returns empty string if workspace and file are equal
 | 
			
		||||
                    paths.push('.');
 | 
			
		||||
                }
 | 
			
		||||
                else {
 | 
			
		||||
                    paths.push(`${relativeFile}`);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        catch (e_1_1) { e_1 = { error: e_1_1 }; }
 | 
			
		||||
        finally {
 | 
			
		||||
            try {
 | 
			
		||||
@@ -5464,6 +5470,7 @@ const buffer = __importStar(__webpack_require__(293));
 | 
			
		||||
const fs = __importStar(__webpack_require__(747));
 | 
			
		||||
const stream = __importStar(__webpack_require__(794));
 | 
			
		||||
const util = __importStar(__webpack_require__(669));
 | 
			
		||||
const timer = __importStar(__webpack_require__(581));
 | 
			
		||||
const utils = __importStar(__webpack_require__(15));
 | 
			
		||||
const constants_1 = __webpack_require__(931);
 | 
			
		||||
const requestUtils_1 = __webpack_require__(899);
 | 
			
		||||
@@ -5654,10 +5661,14 @@ function downloadCacheStorageSDK(archiveLocation, archivePath, options) {
 | 
			
		||||
                    const segmentStart = downloadProgress.segmentOffset + downloadProgress.segmentSize;
 | 
			
		||||
                    const segmentSize = Math.min(maxSegmentSize, contentLength - segmentStart);
 | 
			
		||||
                    downloadProgress.nextSegment(segmentSize);
 | 
			
		||||
                    const result = yield client.downloadToBuffer(segmentStart, segmentSize, {
 | 
			
		||||
                    const result = yield Promise.race([client.downloadToBuffer(segmentStart, segmentSize, {
 | 
			
		||||
                            concurrency: options.downloadConcurrency,
 | 
			
		||||
                            onProgress: downloadProgress.onProgress()
 | 
			
		||||
                    });
 | 
			
		||||
                        }),
 | 
			
		||||
                        timer.setTimeout(60 * 60 * 1000, 'timeout')]);
 | 
			
		||||
                    if (result === 'timeout') {
 | 
			
		||||
                        throw new Error("Segment download timed out");
 | 
			
		||||
                    }
 | 
			
		||||
                    fs.writeFileSync(fd, result);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
@@ -37272,9 +37283,9 @@ function extractTar(archivePath, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -37305,9 +37316,9 @@ function createTar(archiveFolder, sourceDirectories, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -T0 --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'zstdmt --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -T0'];
 | 
			
		||||
                    return ['--use-compress-program', 'zstdmt'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -37317,6 +37328,8 @@ function createTar(archiveFolder, sourceDirectories, compressionMethod) {
 | 
			
		||||
            ...getCompressionProgram(),
 | 
			
		||||
            '-cf',
 | 
			
		||||
            cacheFileName.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
            '--exclude',
 | 
			
		||||
            cacheFileName.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
            '-P',
 | 
			
		||||
            '-C',
 | 
			
		||||
            workingDirectory.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
@@ -37336,9 +37349,9 @@ function listTar(archivePath, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -42341,7 +42354,12 @@ function clean(key)
 | 
			
		||||
/* 578 */,
 | 
			
		||||
/* 579 */,
 | 
			
		||||
/* 580 */,
 | 
			
		||||
/* 581 */,
 | 
			
		||||
/* 581 */
 | 
			
		||||
/***/ (function(module) {
 | 
			
		||||
 | 
			
		||||
module.exports = require("timers/promises");
 | 
			
		||||
 | 
			
		||||
/***/ }),
 | 
			
		||||
/* 582 */
 | 
			
		||||
/***/ (function(module) {
 | 
			
		||||
 | 
			
		||||
@@ -46848,6 +46866,8 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
            checkKey(key);
 | 
			
		||||
        }
 | 
			
		||||
        const compressionMethod = yield utils.getCompressionMethod();
 | 
			
		||||
        let archivePath = '';
 | 
			
		||||
        try {
 | 
			
		||||
            // path are needed to compute version
 | 
			
		||||
            const cacheEntry = yield cacheHttpClient.getCacheEntry(keys, paths, {
 | 
			
		||||
                compressionMethod
 | 
			
		||||
@@ -46856,9 +46876,8 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
                // Cache not found
 | 
			
		||||
                return undefined;
 | 
			
		||||
            }
 | 
			
		||||
        const archivePath = path.join(yield utils.createTempDirectory(), utils.getCacheFileName(compressionMethod));
 | 
			
		||||
            archivePath = path.join(yield utils.createTempDirectory(), utils.getCacheFileName(compressionMethod));
 | 
			
		||||
            core.debug(`Archive Path: ${archivePath}`);
 | 
			
		||||
        try {
 | 
			
		||||
            // Download the cache from the cache entry
 | 
			
		||||
            yield cacheHttpClient.downloadCache(cacheEntry.archiveLocation, archivePath, options);
 | 
			
		||||
            if (core.isDebug()) {
 | 
			
		||||
@@ -46868,6 +46887,17 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
            core.info(`Cache Size: ~${Math.round(archiveFileSize / (1024 * 1024))} MB (${archiveFileSize} B)`);
 | 
			
		||||
            yield tar_1.extractTar(archivePath, compressionMethod);
 | 
			
		||||
            core.info('Cache restored successfully');
 | 
			
		||||
            return cacheEntry.cacheKey;
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            const typedError = error;
 | 
			
		||||
            if (typedError.name === ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            }
 | 
			
		||||
            else {
 | 
			
		||||
                // Supress all non-validation cache related errors because caching should be optional
 | 
			
		||||
                core.warning(`Failed to restore: ${error.message}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        finally {
 | 
			
		||||
            // Try to delete the archive to save space
 | 
			
		||||
@@ -46878,7 +46908,7 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
                core.debug(`Failed to delete archive: ${error}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        return cacheEntry.cacheKey;
 | 
			
		||||
        return undefined;
 | 
			
		||||
    });
 | 
			
		||||
}
 | 
			
		||||
exports.restoreCache = restoreCache;
 | 
			
		||||
@@ -46896,7 +46926,7 @@ function saveCache(paths, key, options) {
 | 
			
		||||
        checkPaths(paths);
 | 
			
		||||
        checkKey(key);
 | 
			
		||||
        const compressionMethod = yield utils.getCompressionMethod();
 | 
			
		||||
        let cacheId = null;
 | 
			
		||||
        let cacheId = -1;
 | 
			
		||||
        const cachePaths = yield utils.resolvePaths(paths);
 | 
			
		||||
        core.debug('Cache Paths:');
 | 
			
		||||
        core.debug(`${JSON.stringify(cachePaths)}`);
 | 
			
		||||
@@ -46935,6 +46965,18 @@ function saveCache(paths, key, options) {
 | 
			
		||||
            core.debug(`Saving Cache (ID: ${cacheId})`);
 | 
			
		||||
            yield cacheHttpClient.saveCache(cacheId, archivePath, options);
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            const typedError = error;
 | 
			
		||||
            if (typedError.name === ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            }
 | 
			
		||||
            else if (typedError.name === ReserveCacheError.name) {
 | 
			
		||||
                core.info(`Failed to save: ${typedError.message}`);
 | 
			
		||||
            }
 | 
			
		||||
            else {
 | 
			
		||||
                core.warning(`Failed to save: ${typedError.message}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        finally {
 | 
			
		||||
            // Try to delete the archive to save space
 | 
			
		||||
            try {
 | 
			
		||||
@@ -48994,7 +49036,6 @@ function run() {
 | 
			
		||||
            const cachePaths = utils.getInputAsArray(constants_1.Inputs.Path, {
 | 
			
		||||
                required: true
 | 
			
		||||
            });
 | 
			
		||||
            try {
 | 
			
		||||
            const cacheKey = yield cache.restoreCache(cachePaths, primaryKey, restoreKeys);
 | 
			
		||||
            if (!cacheKey) {
 | 
			
		||||
                core.info(`Cache not found for input keys: ${[
 | 
			
		||||
@@ -49009,17 +49050,6 @@ function run() {
 | 
			
		||||
            utils.setCacheHitOutput(isExactKeyMatch);
 | 
			
		||||
            core.info(`Cache restored from key: ${cacheKey}`);
 | 
			
		||||
        }
 | 
			
		||||
            catch (error) {
 | 
			
		||||
                const typedError = error;
 | 
			
		||||
                if (typedError.name === cache.ValidationError.name) {
 | 
			
		||||
                    throw error;
 | 
			
		||||
                }
 | 
			
		||||
                else {
 | 
			
		||||
                    utils.logWarning(typedError.message);
 | 
			
		||||
                    utils.setCacheHitOutput(false);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            core.setFailed(error.message);
 | 
			
		||||
        }
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										84
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										84
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							@@ -1113,9 +1113,15 @@ function resolvePaths(patterns) {
 | 
			
		||||
                    .replace(new RegExp(`\\${path.sep}`, 'g'), '/');
 | 
			
		||||
                core.debug(`Matched: ${relativeFile}`);
 | 
			
		||||
                // Paths are made relative so the tar entries are all relative to the root of the workspace.
 | 
			
		||||
                if (relativeFile === '') {
 | 
			
		||||
                    // path.relative returns empty string if workspace and file are equal
 | 
			
		||||
                    paths.push('.');
 | 
			
		||||
                }
 | 
			
		||||
                else {
 | 
			
		||||
                    paths.push(`${relativeFile}`);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        catch (e_1_1) { e_1 = { error: e_1_1 }; }
 | 
			
		||||
        finally {
 | 
			
		||||
            try {
 | 
			
		||||
@@ -5464,6 +5470,7 @@ const buffer = __importStar(__webpack_require__(293));
 | 
			
		||||
const fs = __importStar(__webpack_require__(747));
 | 
			
		||||
const stream = __importStar(__webpack_require__(794));
 | 
			
		||||
const util = __importStar(__webpack_require__(669));
 | 
			
		||||
const timer = __importStar(__webpack_require__(581));
 | 
			
		||||
const utils = __importStar(__webpack_require__(15));
 | 
			
		||||
const constants_1 = __webpack_require__(931);
 | 
			
		||||
const requestUtils_1 = __webpack_require__(899);
 | 
			
		||||
@@ -5654,10 +5661,14 @@ function downloadCacheStorageSDK(archiveLocation, archivePath, options) {
 | 
			
		||||
                    const segmentStart = downloadProgress.segmentOffset + downloadProgress.segmentSize;
 | 
			
		||||
                    const segmentSize = Math.min(maxSegmentSize, contentLength - segmentStart);
 | 
			
		||||
                    downloadProgress.nextSegment(segmentSize);
 | 
			
		||||
                    const result = yield client.downloadToBuffer(segmentStart, segmentSize, {
 | 
			
		||||
                    const result = yield Promise.race([client.downloadToBuffer(segmentStart, segmentSize, {
 | 
			
		||||
                            concurrency: options.downloadConcurrency,
 | 
			
		||||
                            onProgress: downloadProgress.onProgress()
 | 
			
		||||
                    });
 | 
			
		||||
                        }),
 | 
			
		||||
                        timer.setTimeout(60 * 60 * 1000, 'timeout')]);
 | 
			
		||||
                    if (result === 'timeout') {
 | 
			
		||||
                        throw new Error("Segment download timed out");
 | 
			
		||||
                    }
 | 
			
		||||
                    fs.writeFileSync(fd, result);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
@@ -37272,9 +37283,9 @@ function extractTar(archivePath, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -37305,9 +37316,9 @@ function createTar(archiveFolder, sourceDirectories, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -T0 --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'zstdmt --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -T0'];
 | 
			
		||||
                    return ['--use-compress-program', 'zstdmt'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -37317,6 +37328,8 @@ function createTar(archiveFolder, sourceDirectories, compressionMethod) {
 | 
			
		||||
            ...getCompressionProgram(),
 | 
			
		||||
            '-cf',
 | 
			
		||||
            cacheFileName.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
            '--exclude',
 | 
			
		||||
            cacheFileName.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
            '-P',
 | 
			
		||||
            '-C',
 | 
			
		||||
            workingDirectory.replace(new RegExp(`\\${path.sep}`, 'g'), '/'),
 | 
			
		||||
@@ -37336,9 +37349,9 @@ function listTar(archivePath, compressionMethod) {
 | 
			
		||||
        function getCompressionProgram() {
 | 
			
		||||
            switch (compressionMethod) {
 | 
			
		||||
                case constants_1.CompressionMethod.Zstd:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d --long=30'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd --long=30'];
 | 
			
		||||
                case constants_1.CompressionMethod.ZstdWithoutLong:
 | 
			
		||||
                    return ['--use-compress-program', 'zstd -d'];
 | 
			
		||||
                    return ['--use-compress-program', 'unzstd'];
 | 
			
		||||
                default:
 | 
			
		||||
                    return ['-z'];
 | 
			
		||||
            }
 | 
			
		||||
@@ -42341,7 +42354,12 @@ function clean(key)
 | 
			
		||||
/* 578 */,
 | 
			
		||||
/* 579 */,
 | 
			
		||||
/* 580 */,
 | 
			
		||||
/* 581 */,
 | 
			
		||||
/* 581 */
 | 
			
		||||
/***/ (function(module) {
 | 
			
		||||
 | 
			
		||||
module.exports = require("timers/promises");
 | 
			
		||||
 | 
			
		||||
/***/ }),
 | 
			
		||||
/* 582 */
 | 
			
		||||
/***/ (function(module) {
 | 
			
		||||
 | 
			
		||||
@@ -46790,24 +46808,12 @@ function run() {
 | 
			
		||||
            const cachePaths = utils.getInputAsArray(constants_1.Inputs.Path, {
 | 
			
		||||
                required: true
 | 
			
		||||
            });
 | 
			
		||||
            try {
 | 
			
		||||
                yield cache.saveCache(cachePaths, primaryKey, {
 | 
			
		||||
            const cacheId = yield cache.saveCache(cachePaths, primaryKey, {
 | 
			
		||||
                uploadChunkSize: utils.getInputAsInt(constants_1.Inputs.UploadChunkSize)
 | 
			
		||||
            });
 | 
			
		||||
            if (cacheId != -1) {
 | 
			
		||||
                core.info(`Cache saved with key: ${primaryKey}`);
 | 
			
		||||
            }
 | 
			
		||||
            catch (error) {
 | 
			
		||||
                const typedError = error;
 | 
			
		||||
                if (typedError.name === cache.ValidationError.name) {
 | 
			
		||||
                    throw error;
 | 
			
		||||
                }
 | 
			
		||||
                else if (typedError.name === cache.ReserveCacheError.name) {
 | 
			
		||||
                    core.info(typedError.message);
 | 
			
		||||
                }
 | 
			
		||||
                else {
 | 
			
		||||
                    utils.logWarning(typedError.message);
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            utils.logWarning(error.message);
 | 
			
		||||
@@ -46946,6 +46952,8 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
            checkKey(key);
 | 
			
		||||
        }
 | 
			
		||||
        const compressionMethod = yield utils.getCompressionMethod();
 | 
			
		||||
        let archivePath = '';
 | 
			
		||||
        try {
 | 
			
		||||
            // path are needed to compute version
 | 
			
		||||
            const cacheEntry = yield cacheHttpClient.getCacheEntry(keys, paths, {
 | 
			
		||||
                compressionMethod
 | 
			
		||||
@@ -46954,9 +46962,8 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
                // Cache not found
 | 
			
		||||
                return undefined;
 | 
			
		||||
            }
 | 
			
		||||
        const archivePath = path.join(yield utils.createTempDirectory(), utils.getCacheFileName(compressionMethod));
 | 
			
		||||
            archivePath = path.join(yield utils.createTempDirectory(), utils.getCacheFileName(compressionMethod));
 | 
			
		||||
            core.debug(`Archive Path: ${archivePath}`);
 | 
			
		||||
        try {
 | 
			
		||||
            // Download the cache from the cache entry
 | 
			
		||||
            yield cacheHttpClient.downloadCache(cacheEntry.archiveLocation, archivePath, options);
 | 
			
		||||
            if (core.isDebug()) {
 | 
			
		||||
@@ -46966,6 +46973,17 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
            core.info(`Cache Size: ~${Math.round(archiveFileSize / (1024 * 1024))} MB (${archiveFileSize} B)`);
 | 
			
		||||
            yield tar_1.extractTar(archivePath, compressionMethod);
 | 
			
		||||
            core.info('Cache restored successfully');
 | 
			
		||||
            return cacheEntry.cacheKey;
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            const typedError = error;
 | 
			
		||||
            if (typedError.name === ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            }
 | 
			
		||||
            else {
 | 
			
		||||
                // Supress all non-validation cache related errors because caching should be optional
 | 
			
		||||
                core.warning(`Failed to restore: ${error.message}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        finally {
 | 
			
		||||
            // Try to delete the archive to save space
 | 
			
		||||
@@ -46976,7 +46994,7 @@ function restoreCache(paths, primaryKey, restoreKeys, options) {
 | 
			
		||||
                core.debug(`Failed to delete archive: ${error}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        return cacheEntry.cacheKey;
 | 
			
		||||
        return undefined;
 | 
			
		||||
    });
 | 
			
		||||
}
 | 
			
		||||
exports.restoreCache = restoreCache;
 | 
			
		||||
@@ -46994,7 +47012,7 @@ function saveCache(paths, key, options) {
 | 
			
		||||
        checkPaths(paths);
 | 
			
		||||
        checkKey(key);
 | 
			
		||||
        const compressionMethod = yield utils.getCompressionMethod();
 | 
			
		||||
        let cacheId = null;
 | 
			
		||||
        let cacheId = -1;
 | 
			
		||||
        const cachePaths = yield utils.resolvePaths(paths);
 | 
			
		||||
        core.debug('Cache Paths:');
 | 
			
		||||
        core.debug(`${JSON.stringify(cachePaths)}`);
 | 
			
		||||
@@ -47033,6 +47051,18 @@ function saveCache(paths, key, options) {
 | 
			
		||||
            core.debug(`Saving Cache (ID: ${cacheId})`);
 | 
			
		||||
            yield cacheHttpClient.saveCache(cacheId, archivePath, options);
 | 
			
		||||
        }
 | 
			
		||||
        catch (error) {
 | 
			
		||||
            const typedError = error;
 | 
			
		||||
            if (typedError.name === ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            }
 | 
			
		||||
            else if (typedError.name === ReserveCacheError.name) {
 | 
			
		||||
                core.info(`Failed to save: ${typedError.message}`);
 | 
			
		||||
            }
 | 
			
		||||
            else {
 | 
			
		||||
                core.warning(`Failed to save: ${typedError.message}`);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
        finally {
 | 
			
		||||
            // Try to delete the archive to save space
 | 
			
		||||
            try {
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										36
									
								
								examples.md
									
									
									
									
									
								
							
							
						
						
									
										36
									
								
								examples.md
									
									
									
									
									
								
							@@ -9,6 +9,7 @@
 | 
			
		||||
  - [macOS](#macos)
 | 
			
		||||
  - [Windows](#windows-1)
 | 
			
		||||
- [Elixir - Mix](#elixir---mix)
 | 
			
		||||
- [Erlang - Rebar3](#erlang--rebar3)
 | 
			
		||||
- [Go - Modules](#go---modules)
 | 
			
		||||
  - [Linux](#linux-1)
 | 
			
		||||
  - [macOS](#macos-1)
 | 
			
		||||
@@ -151,6 +152,18 @@ steps:
 | 
			
		||||
      ${{ runner.os }}-mix-
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
## Erlang - Rebar3
 | 
			
		||||
```yaml
 | 
			
		||||
- uses: actions/cache@v2
 | 
			
		||||
  with:
 | 
			
		||||
    path: |
 | 
			
		||||
      ~/.cache/rebar3
 | 
			
		||||
      _build
 | 
			
		||||
    key: ${{ runner.os }}-erlang-${{ env.OTP_VERSION }}-${{ hashFiles('**/*rebar.lock') }}
 | 
			
		||||
    restore-keys: |
 | 
			
		||||
      ${{ runner.os }}-erlang-${{ env.OTP_VERSION }}-
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
## Go - Modules
 | 
			
		||||
 | 
			
		||||
### Linux
 | 
			
		||||
@@ -210,6 +223,8 @@ We cache the elements of the Cabal store separately, as the entirety of `~/.caba
 | 
			
		||||
 | 
			
		||||
## Haskell - Stack
 | 
			
		||||
 | 
			
		||||
### Linux or macOS
 | 
			
		||||
 | 
			
		||||
```yaml
 | 
			
		||||
- uses: actions/cache@v3
 | 
			
		||||
  name: Cache ~/.stack
 | 
			
		||||
@@ -227,6 +242,27 @@ We cache the elements of the Cabal store separately, as the entirety of `~/.caba
 | 
			
		||||
      ${{ runner.os }}-stack-work-
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
### Windows
 | 
			
		||||
 | 
			
		||||
```yaml
 | 
			
		||||
- uses: actions/cache@v3
 | 
			
		||||
  name: Cache %APPDATA%\stack %LOCALAPPDATA%\Programs\stack
 | 
			
		||||
  with:
 | 
			
		||||
    path: |
 | 
			
		||||
      ~\AppData\Roaming\stack
 | 
			
		||||
      ~\AppData\Local\Programs\stack    
 | 
			
		||||
    key: ${{ runner.os }}-stack-global-${{ hashFiles('stack.yaml') }}-${{ hashFiles('package.yaml') }}
 | 
			
		||||
    restore-keys: |
 | 
			
		||||
      ${{ runner.os }}-stack-global-
 | 
			
		||||
- uses: actions/cache@v3
 | 
			
		||||
  name: Cache .stack-work
 | 
			
		||||
  with:
 | 
			
		||||
    path: .stack-work
 | 
			
		||||
    key: ${{ runner.os }}-stack-work-${{ hashFiles('stack.yaml') }}-${{ hashFiles('package.yaml') }}-${{ hashFiles('**/*.hs') }}
 | 
			
		||||
    restore-keys: |
 | 
			
		||||
      ${{ runner.os }}-stack-work-
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
## Java - Gradle
 | 
			
		||||
 | 
			
		||||
>Note: Ensure no Gradle daemons are running anymore when your workflow completes. Creating the cache package might fail due to locks being held by Gradle. Refer to the [Gradle Daemon documentation](https://docs.gradle.org/current/userguide/gradle_daemon.html) on how to disable or stop the Gradle Daemons.
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										18
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										18
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							@@ -1,15 +1,15 @@
 | 
			
		||||
{
 | 
			
		||||
  "name": "cache",
 | 
			
		||||
  "version": "3.0.3",
 | 
			
		||||
  "version": "3.0.5",
 | 
			
		||||
  "lockfileVersion": 2,
 | 
			
		||||
  "requires": true,
 | 
			
		||||
  "packages": {
 | 
			
		||||
    "": {
 | 
			
		||||
      "name": "cache",
 | 
			
		||||
      "version": "3.0.3",
 | 
			
		||||
      "version": "3.0.5",
 | 
			
		||||
      "license": "MIT",
 | 
			
		||||
      "dependencies": {
 | 
			
		||||
        "@actions/cache": "^2.0.5",
 | 
			
		||||
        "@actions/cache": "file:actions-cache-3.0.1.tgz",
 | 
			
		||||
        "@actions/core": "^1.7.0",
 | 
			
		||||
        "@actions/exec": "^1.1.1",
 | 
			
		||||
        "@actions/io": "^1.1.2"
 | 
			
		||||
@@ -36,9 +36,10 @@
 | 
			
		||||
      }
 | 
			
		||||
    },
 | 
			
		||||
    "node_modules/@actions/cache": {
 | 
			
		||||
      "version": "2.0.5",
 | 
			
		||||
      "resolved": "https://registry.npmjs.org/@actions/cache/-/cache-2.0.5.tgz",
 | 
			
		||||
      "integrity": "sha512-aG06dsgcVtiuHLJsIfwrDtvzNNJQ+Iqk8DQt1IeI6gG7ezmLaSdZkHEwA/DNrm5TtOahLcgGEo2SXqbFElVMQg==",
 | 
			
		||||
      "version": "3.0.1",
 | 
			
		||||
      "resolved": "file:actions-cache-3.0.1.tgz",
 | 
			
		||||
      "integrity": "sha512-ucvw0xvFpe0/vfNQ/rc11ste0nidCdBAJ5j5F01BxBqjxmGH2doVzfPlqSIGhcN7wKI074x2ATb9+7HSrTqGHg==",
 | 
			
		||||
      "license": "MIT",
 | 
			
		||||
      "dependencies": {
 | 
			
		||||
        "@actions/core": "^1.2.6",
 | 
			
		||||
        "@actions/exec": "^1.0.1",
 | 
			
		||||
@@ -9533,9 +9534,8 @@
 | 
			
		||||
  },
 | 
			
		||||
  "dependencies": {
 | 
			
		||||
    "@actions/cache": {
 | 
			
		||||
      "version": "2.0.5",
 | 
			
		||||
      "resolved": "https://registry.npmjs.org/@actions/cache/-/cache-2.0.5.tgz",
 | 
			
		||||
      "integrity": "sha512-aG06dsgcVtiuHLJsIfwrDtvzNNJQ+Iqk8DQt1IeI6gG7ezmLaSdZkHEwA/DNrm5TtOahLcgGEo2SXqbFElVMQg==",
 | 
			
		||||
      "version": "file:actions-cache-3.0.1.tgz",
 | 
			
		||||
      "integrity": "sha512-ucvw0xvFpe0/vfNQ/rc11ste0nidCdBAJ5j5F01BxBqjxmGH2doVzfPlqSIGhcN7wKI074x2ATb9+7HSrTqGHg==",
 | 
			
		||||
      "requires": {
 | 
			
		||||
        "@actions/core": "^1.2.6",
 | 
			
		||||
        "@actions/exec": "^1.0.1",
 | 
			
		||||
 
 | 
			
		||||
@@ -1,6 +1,6 @@
 | 
			
		||||
{
 | 
			
		||||
  "name": "cache",
 | 
			
		||||
  "version": "3.0.3",
 | 
			
		||||
  "version": "3.0.5",
 | 
			
		||||
  "private": true,
 | 
			
		||||
  "description": "Cache dependencies and build outputs",
 | 
			
		||||
  "main": "dist/restore/index.js",
 | 
			
		||||
@@ -23,7 +23,7 @@
 | 
			
		||||
  "author": "GitHub",
 | 
			
		||||
  "license": "MIT",
 | 
			
		||||
  "dependencies": {
 | 
			
		||||
    "@actions/cache": "^2.0.5",
 | 
			
		||||
    "@actions/cache": "file:actions-cache-3.0.1.tgz",
 | 
			
		||||
    "@actions/core": "^1.7.0",
 | 
			
		||||
    "@actions/exec": "^1.1.1",
 | 
			
		||||
    "@actions/io": "^1.1.2"
 | 
			
		||||
 
 | 
			
		||||
@@ -29,12 +29,12 @@ async function run(): Promise<void> {
 | 
			
		||||
            required: true
 | 
			
		||||
        });
 | 
			
		||||
 | 
			
		||||
        try {
 | 
			
		||||
        const cacheKey = await cache.restoreCache(
 | 
			
		||||
            cachePaths,
 | 
			
		||||
            primaryKey,
 | 
			
		||||
            restoreKeys
 | 
			
		||||
        );
 | 
			
		||||
 | 
			
		||||
        if (!cacheKey) {
 | 
			
		||||
            core.info(
 | 
			
		||||
                `Cache not found for input keys: ${[
 | 
			
		||||
@@ -42,6 +42,7 @@ async function run(): Promise<void> {
 | 
			
		||||
                    ...restoreKeys
 | 
			
		||||
                ].join(", ")}`
 | 
			
		||||
            );
 | 
			
		||||
 | 
			
		||||
            return;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
@@ -50,17 +51,7 @@ async function run(): Promise<void> {
 | 
			
		||||
 | 
			
		||||
        const isExactKeyMatch = utils.isExactKeyMatch(primaryKey, cacheKey);
 | 
			
		||||
        utils.setCacheHitOutput(isExactKeyMatch);
 | 
			
		||||
 | 
			
		||||
        core.info(`Cache restored from key: ${cacheKey}`);
 | 
			
		||||
        } catch (error: unknown) {
 | 
			
		||||
            const typedError = error as Error;
 | 
			
		||||
            if (typedError.name === cache.ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            } else {
 | 
			
		||||
                utils.logWarning(typedError.message);
 | 
			
		||||
                utils.setCacheHitOutput(false);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
    } catch (error: unknown) {
 | 
			
		||||
        core.setFailed((error as Error).message);
 | 
			
		||||
    }
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										14
									
								
								src/save.ts
									
									
									
									
									
								
							
							
						
						
									
										14
									
								
								src/save.ts
									
									
									
									
									
								
							@@ -44,20 +44,12 @@ async function run(): Promise<void> {
 | 
			
		||||
            required: true
 | 
			
		||||
        });
 | 
			
		||||
 | 
			
		||||
        try {
 | 
			
		||||
            await cache.saveCache(cachePaths, primaryKey, {
 | 
			
		||||
        const cacheId = await cache.saveCache(cachePaths, primaryKey, {
 | 
			
		||||
            uploadChunkSize: utils.getInputAsInt(Inputs.UploadChunkSize)
 | 
			
		||||
        });
 | 
			
		||||
 | 
			
		||||
        if (cacheId != -1) {
 | 
			
		||||
            core.info(`Cache saved with key: ${primaryKey}`);
 | 
			
		||||
        } catch (error: unknown) {
 | 
			
		||||
            const typedError = error as Error;
 | 
			
		||||
            if (typedError.name === cache.ValidationError.name) {
 | 
			
		||||
                throw error;
 | 
			
		||||
            } else if (typedError.name === cache.ReserveCacheError.name) {
 | 
			
		||||
                core.info(typedError.message);
 | 
			
		||||
            } else {
 | 
			
		||||
                utils.logWarning(typedError.message);
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
    } catch (error: unknown) {
 | 
			
		||||
        utils.logWarning((error as Error).message);
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user