mirror of
				https://github.com/actions/upload-artifact.git
				synced 2025-10-31 15:33:43 +00:00 
			
		
		
		
	Merge pull request #594 from actions/robherley/4.3.6
Revert to @actions/artifact 2.1.8
This commit is contained in:
		
						commit
						834a144ee9
					
				
							
								
								
									
										2
									
								
								.licenses/npm/@actions/artifact.dep.yml
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										2
									
								
								.licenses/npm/@actions/artifact.dep.yml
									
									
									
										generated
									
									
									
								
							| @ -1,6 +1,6 @@ | |||||||
| --- | --- | ||||||
| name: "@actions/artifact" | name: "@actions/artifact" | ||||||
| version: 2.1.9 | version: 2.1.8 | ||||||
| type: npm | type: npm | ||||||
| summary:  | summary:  | ||||||
| homepage:  | homepage:  | ||||||
|  | |||||||
							
								
								
									
										58
									
								
								dist/merge/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										58
									
								
								dist/merge/index.js
									
									
									
									
										vendored
									
									
								
							| @ -2997,7 +2997,7 @@ var __importDefault = (this && this.__importDefault) || function (mod) { | |||||||
|     return (mod && mod.__esModule) ? mod : { "default": mod }; |     return (mod && mod.__esModule) ? mod : { "default": mod }; | ||||||
| }; | }; | ||||||
| Object.defineProperty(exports, "__esModule", ({ value: true })); | Object.defineProperty(exports, "__esModule", ({ value: true })); | ||||||
| exports.getUploadChunkTimeout = exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0; | exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0; | ||||||
| const os_1 = __importDefault(__nccwpck_require__(22037)); | const os_1 = __importDefault(__nccwpck_require__(22037)); | ||||||
| // Used for controlling the highWaterMark value of the zip that is being streamed
 | // Used for controlling the highWaterMark value of the zip that is being streamed
 | ||||||
| // The same value is used as the chunk size that is use during upload to blob storage
 | // The same value is used as the chunk size that is use during upload to blob storage
 | ||||||
| @ -3050,10 +3050,6 @@ function getConcurrency() { | |||||||
|     return concurrency > 300 ? 300 : concurrency; |     return concurrency > 300 ? 300 : concurrency; | ||||||
| } | } | ||||||
| exports.getConcurrency = getConcurrency; | exports.getConcurrency = getConcurrency; | ||||||
| function getUploadChunkTimeout() { |  | ||||||
|     return 30000; // 30 seconds
 |  | ||||||
| } |  | ||||||
| exports.getUploadChunkTimeout = getUploadChunkTimeout; |  | ||||||
| //# sourceMappingURL=config.js.map
 | //# sourceMappingURL=config.js.map
 | ||||||
| 
 | 
 | ||||||
| /***/ }), | /***/ }), | ||||||
| @ -3302,34 +3298,37 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|     return __awaiter(this, void 0, void 0, function* () { |     return __awaiter(this, void 0, void 0, function* () { | ||||||
|         let uploadByteCount = 0; |         let uploadByteCount = 0; | ||||||
|         let lastProgressTime = Date.now(); |         let lastProgressTime = Date.now(); | ||||||
|         const abortController = new AbortController(); |         let timeoutId; | ||||||
|         const chunkTimer = (interval) => __awaiter(this, void 0, void 0, function* () { |         const chunkTimer = (timeout) => { | ||||||
|             return new Promise((resolve, reject) => { |             // clear the previous timeout
 | ||||||
|                 const timer = setInterval(() => { |             if (timeoutId) { | ||||||
|                     if (Date.now() - lastProgressTime > interval) { |                 clearTimeout(timeoutId); | ||||||
|                         reject(new Error('Upload progress stalled.')); |             } | ||||||
|                     } |             timeoutId = setTimeout(() => { | ||||||
|                 }, interval); |                 const now = Date.now(); | ||||||
|                 abortController.signal.addEventListener('abort', () => { |                 // if there's been more than 30 seconds since the
 | ||||||
|                     clearInterval(timer); |                 // last progress event, then we'll consider the upload stalled
 | ||||||
|                     resolve(); |                 if (now - lastProgressTime > timeout) { | ||||||
|                 }); |                     throw new Error('Upload progress stalled.'); | ||||||
|             }); |                 } | ||||||
|         }); |             }, timeout); | ||||||
|  |             return timeoutId; | ||||||
|  |         }; | ||||||
|         const maxConcurrency = (0, config_1.getConcurrency)(); |         const maxConcurrency = (0, config_1.getConcurrency)(); | ||||||
|         const bufferSize = (0, config_1.getUploadChunkSize)(); |         const bufferSize = (0, config_1.getUploadChunkSize)(); | ||||||
|         const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL); |         const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL); | ||||||
|         const blockBlobClient = blobClient.getBlockBlobClient(); |         const blockBlobClient = blobClient.getBlockBlobClient(); | ||||||
|  |         const timeoutDuration = 300000; // 30 seconds
 | ||||||
|         core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`); |         core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`); | ||||||
|         const uploadCallback = (progress) => { |         const uploadCallback = (progress) => { | ||||||
|             core.info(`Uploaded bytes ${progress.loadedBytes}`); |             core.info(`Uploaded bytes ${progress.loadedBytes}`); | ||||||
|             uploadByteCount = progress.loadedBytes; |             uploadByteCount = progress.loadedBytes; | ||||||
|  |             chunkTimer(timeoutDuration); | ||||||
|             lastProgressTime = Date.now(); |             lastProgressTime = Date.now(); | ||||||
|         }; |         }; | ||||||
|         const options = { |         const options = { | ||||||
|             blobHTTPHeaders: { blobContentType: 'zip' }, |             blobHTTPHeaders: { blobContentType: 'zip' }, | ||||||
|             onProgress: uploadCallback, |             onProgress: uploadCallback | ||||||
|             abortSignal: abortController.signal |  | ||||||
|         }; |         }; | ||||||
|         let sha256Hash = undefined; |         let sha256Hash = undefined; | ||||||
|         const uploadStream = new stream.PassThrough(); |         const uploadStream = new stream.PassThrough(); | ||||||
| @ -3338,10 +3337,9 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|         zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 |         zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | ||||||
|         core.info('Beginning upload of artifact content to blob storage'); |         core.info('Beginning upload of artifact content to blob storage'); | ||||||
|         try { |         try { | ||||||
|             yield Promise.race([ |             // Start the chunk timer
 | ||||||
|                 blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), |             timeoutId = chunkTimer(timeoutDuration); | ||||||
|                 chunkTimer((0, config_1.getUploadChunkTimeout)()) |             yield blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options); | ||||||
|             ]); |  | ||||||
|         } |         } | ||||||
|         catch (error) { |         catch (error) { | ||||||
|             if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) { |             if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) { | ||||||
| @ -3350,7 +3348,10 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|             throw error; |             throw error; | ||||||
|         } |         } | ||||||
|         finally { |         finally { | ||||||
|             abortController.abort(); |             // clear the timeout whether or not the upload completes
 | ||||||
|  |             if (timeoutId) { | ||||||
|  |                 clearTimeout(timeoutId); | ||||||
|  |             } | ||||||
|         } |         } | ||||||
|         core.info('Finished uploading artifact content to blob storage!'); |         core.info('Finished uploading artifact content to blob storage!'); | ||||||
|         hashStream.end(); |         hashStream.end(); | ||||||
| @ -3777,6 +3778,7 @@ exports.createZipUploadStream = exports.ZipUploadStream = exports.DEFAULT_COMPRE | |||||||
| const stream = __importStar(__nccwpck_require__(12781)); | const stream = __importStar(__nccwpck_require__(12781)); | ||||||
| const archiver = __importStar(__nccwpck_require__(43084)); | const archiver = __importStar(__nccwpck_require__(43084)); | ||||||
| const core = __importStar(__nccwpck_require__(42186)); | const core = __importStar(__nccwpck_require__(42186)); | ||||||
|  | const fs_1 = __nccwpck_require__(57147); | ||||||
| const config_1 = __nccwpck_require__(74610); | const config_1 = __nccwpck_require__(74610); | ||||||
| exports.DEFAULT_COMPRESSION_LEVEL = 6; | exports.DEFAULT_COMPRESSION_LEVEL = 6; | ||||||
| // Custom stream transformer so we can set the highWaterMark property
 | // Custom stream transformer so we can set the highWaterMark property
 | ||||||
| @ -3808,7 +3810,7 @@ function createZipUploadStream(uploadSpecification, compressionLevel = exports.D | |||||||
|         for (const file of uploadSpecification) { |         for (const file of uploadSpecification) { | ||||||
|             if (file.sourcePath !== null) { |             if (file.sourcePath !== null) { | ||||||
|                 // Add a normal file to the zip
 |                 // Add a normal file to the zip
 | ||||||
|                 zip.file(file.sourcePath, { |                 zip.append((0, fs_1.createReadStream)(file.sourcePath), { | ||||||
|                     name: file.destinationPath |                     name: file.destinationPath | ||||||
|                 }); |                 }); | ||||||
|             } |             } | ||||||
| @ -136150,7 +136152,7 @@ module.exports = index; | |||||||
| /***/ ((module) => { | /***/ ((module) => { | ||||||
| 
 | 
 | ||||||
| "use strict"; | "use strict"; | ||||||
| module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.9","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}'); | module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.8","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}'); | ||||||
| 
 | 
 | ||||||
| /***/ }), | /***/ }), | ||||||
| 
 | 
 | ||||||
|  | |||||||
							
								
								
									
										58
									
								
								dist/upload/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										58
									
								
								dist/upload/index.js
									
									
									
									
										vendored
									
									
								
							| @ -2997,7 +2997,7 @@ var __importDefault = (this && this.__importDefault) || function (mod) { | |||||||
|     return (mod && mod.__esModule) ? mod : { "default": mod }; |     return (mod && mod.__esModule) ? mod : { "default": mod }; | ||||||
| }; | }; | ||||||
| Object.defineProperty(exports, "__esModule", ({ value: true })); | Object.defineProperty(exports, "__esModule", ({ value: true })); | ||||||
| exports.getUploadChunkTimeout = exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0; | exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0; | ||||||
| const os_1 = __importDefault(__nccwpck_require__(22037)); | const os_1 = __importDefault(__nccwpck_require__(22037)); | ||||||
| // Used for controlling the highWaterMark value of the zip that is being streamed
 | // Used for controlling the highWaterMark value of the zip that is being streamed
 | ||||||
| // The same value is used as the chunk size that is use during upload to blob storage
 | // The same value is used as the chunk size that is use during upload to blob storage
 | ||||||
| @ -3050,10 +3050,6 @@ function getConcurrency() { | |||||||
|     return concurrency > 300 ? 300 : concurrency; |     return concurrency > 300 ? 300 : concurrency; | ||||||
| } | } | ||||||
| exports.getConcurrency = getConcurrency; | exports.getConcurrency = getConcurrency; | ||||||
| function getUploadChunkTimeout() { |  | ||||||
|     return 30000; // 30 seconds
 |  | ||||||
| } |  | ||||||
| exports.getUploadChunkTimeout = getUploadChunkTimeout; |  | ||||||
| //# sourceMappingURL=config.js.map
 | //# sourceMappingURL=config.js.map
 | ||||||
| 
 | 
 | ||||||
| /***/ }), | /***/ }), | ||||||
| @ -3302,34 +3298,37 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|     return __awaiter(this, void 0, void 0, function* () { |     return __awaiter(this, void 0, void 0, function* () { | ||||||
|         let uploadByteCount = 0; |         let uploadByteCount = 0; | ||||||
|         let lastProgressTime = Date.now(); |         let lastProgressTime = Date.now(); | ||||||
|         const abortController = new AbortController(); |         let timeoutId; | ||||||
|         const chunkTimer = (interval) => __awaiter(this, void 0, void 0, function* () { |         const chunkTimer = (timeout) => { | ||||||
|             return new Promise((resolve, reject) => { |             // clear the previous timeout
 | ||||||
|                 const timer = setInterval(() => { |             if (timeoutId) { | ||||||
|                     if (Date.now() - lastProgressTime > interval) { |                 clearTimeout(timeoutId); | ||||||
|                         reject(new Error('Upload progress stalled.')); |             } | ||||||
|                     } |             timeoutId = setTimeout(() => { | ||||||
|                 }, interval); |                 const now = Date.now(); | ||||||
|                 abortController.signal.addEventListener('abort', () => { |                 // if there's been more than 30 seconds since the
 | ||||||
|                     clearInterval(timer); |                 // last progress event, then we'll consider the upload stalled
 | ||||||
|                     resolve(); |                 if (now - lastProgressTime > timeout) { | ||||||
|                 }); |                     throw new Error('Upload progress stalled.'); | ||||||
|             }); |                 } | ||||||
|         }); |             }, timeout); | ||||||
|  |             return timeoutId; | ||||||
|  |         }; | ||||||
|         const maxConcurrency = (0, config_1.getConcurrency)(); |         const maxConcurrency = (0, config_1.getConcurrency)(); | ||||||
|         const bufferSize = (0, config_1.getUploadChunkSize)(); |         const bufferSize = (0, config_1.getUploadChunkSize)(); | ||||||
|         const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL); |         const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL); | ||||||
|         const blockBlobClient = blobClient.getBlockBlobClient(); |         const blockBlobClient = blobClient.getBlockBlobClient(); | ||||||
|  |         const timeoutDuration = 300000; // 30 seconds
 | ||||||
|         core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`); |         core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`); | ||||||
|         const uploadCallback = (progress) => { |         const uploadCallback = (progress) => { | ||||||
|             core.info(`Uploaded bytes ${progress.loadedBytes}`); |             core.info(`Uploaded bytes ${progress.loadedBytes}`); | ||||||
|             uploadByteCount = progress.loadedBytes; |             uploadByteCount = progress.loadedBytes; | ||||||
|  |             chunkTimer(timeoutDuration); | ||||||
|             lastProgressTime = Date.now(); |             lastProgressTime = Date.now(); | ||||||
|         }; |         }; | ||||||
|         const options = { |         const options = { | ||||||
|             blobHTTPHeaders: { blobContentType: 'zip' }, |             blobHTTPHeaders: { blobContentType: 'zip' }, | ||||||
|             onProgress: uploadCallback, |             onProgress: uploadCallback | ||||||
|             abortSignal: abortController.signal |  | ||||||
|         }; |         }; | ||||||
|         let sha256Hash = undefined; |         let sha256Hash = undefined; | ||||||
|         const uploadStream = new stream.PassThrough(); |         const uploadStream = new stream.PassThrough(); | ||||||
| @ -3338,10 +3337,9 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|         zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 |         zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | ||||||
|         core.info('Beginning upload of artifact content to blob storage'); |         core.info('Beginning upload of artifact content to blob storage'); | ||||||
|         try { |         try { | ||||||
|             yield Promise.race([ |             // Start the chunk timer
 | ||||||
|                 blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), |             timeoutId = chunkTimer(timeoutDuration); | ||||||
|                 chunkTimer((0, config_1.getUploadChunkTimeout)()) |             yield blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options); | ||||||
|             ]); |  | ||||||
|         } |         } | ||||||
|         catch (error) { |         catch (error) { | ||||||
|             if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) { |             if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) { | ||||||
| @ -3350,7 +3348,10 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) { | |||||||
|             throw error; |             throw error; | ||||||
|         } |         } | ||||||
|         finally { |         finally { | ||||||
|             abortController.abort(); |             // clear the timeout whether or not the upload completes
 | ||||||
|  |             if (timeoutId) { | ||||||
|  |                 clearTimeout(timeoutId); | ||||||
|  |             } | ||||||
|         } |         } | ||||||
|         core.info('Finished uploading artifact content to blob storage!'); |         core.info('Finished uploading artifact content to blob storage!'); | ||||||
|         hashStream.end(); |         hashStream.end(); | ||||||
| @ -3777,6 +3778,7 @@ exports.createZipUploadStream = exports.ZipUploadStream = exports.DEFAULT_COMPRE | |||||||
| const stream = __importStar(__nccwpck_require__(12781)); | const stream = __importStar(__nccwpck_require__(12781)); | ||||||
| const archiver = __importStar(__nccwpck_require__(43084)); | const archiver = __importStar(__nccwpck_require__(43084)); | ||||||
| const core = __importStar(__nccwpck_require__(42186)); | const core = __importStar(__nccwpck_require__(42186)); | ||||||
|  | const fs_1 = __nccwpck_require__(57147); | ||||||
| const config_1 = __nccwpck_require__(74610); | const config_1 = __nccwpck_require__(74610); | ||||||
| exports.DEFAULT_COMPRESSION_LEVEL = 6; | exports.DEFAULT_COMPRESSION_LEVEL = 6; | ||||||
| // Custom stream transformer so we can set the highWaterMark property
 | // Custom stream transformer so we can set the highWaterMark property
 | ||||||
| @ -3808,7 +3810,7 @@ function createZipUploadStream(uploadSpecification, compressionLevel = exports.D | |||||||
|         for (const file of uploadSpecification) { |         for (const file of uploadSpecification) { | ||||||
|             if (file.sourcePath !== null) { |             if (file.sourcePath !== null) { | ||||||
|                 // Add a normal file to the zip
 |                 // Add a normal file to the zip
 | ||||||
|                 zip.file(file.sourcePath, { |                 zip.append((0, fs_1.createReadStream)(file.sourcePath), { | ||||||
|                     name: file.destinationPath |                     name: file.destinationPath | ||||||
|                 }); |                 }); | ||||||
|             } |             } | ||||||
| @ -136160,7 +136162,7 @@ module.exports = index; | |||||||
| /***/ ((module) => { | /***/ ((module) => { | ||||||
| 
 | 
 | ||||||
| "use strict"; | "use strict"; | ||||||
| module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.9","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}'); | module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.8","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}'); | ||||||
| 
 | 
 | ||||||
| /***/ }), | /***/ }), | ||||||
| 
 | 
 | ||||||
|  | |||||||
							
								
								
									
										18
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										18
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							| @ -1,15 +1,15 @@ | |||||||
| { | { | ||||||
|   "name": "upload-artifact", |   "name": "upload-artifact", | ||||||
|   "version": "4.3.5", |   "version": "4.3.6", | ||||||
|   "lockfileVersion": 2, |   "lockfileVersion": 2, | ||||||
|   "requires": true, |   "requires": true, | ||||||
|   "packages": { |   "packages": { | ||||||
|     "": { |     "": { | ||||||
|       "name": "upload-artifact", |       "name": "upload-artifact", | ||||||
|       "version": "4.3.5", |       "version": "4.3.6", | ||||||
|       "license": "MIT", |       "license": "MIT", | ||||||
|       "dependencies": { |       "dependencies": { | ||||||
|         "@actions/artifact": "^2.1.9", |         "@actions/artifact": "2.1.8", | ||||||
|         "@actions/core": "^1.10.1", |         "@actions/core": "^1.10.1", | ||||||
|         "@actions/github": "^6.0.0", |         "@actions/github": "^6.0.0", | ||||||
|         "@actions/glob": "^0.3.0", |         "@actions/glob": "^0.3.0", | ||||||
| @ -34,9 +34,9 @@ | |||||||
|       } |       } | ||||||
|     }, |     }, | ||||||
|     "node_modules/@actions/artifact": { |     "node_modules/@actions/artifact": { | ||||||
|       "version": "2.1.9", |       "version": "2.1.8", | ||||||
|       "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.9.tgz", |       "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.8.tgz", | ||||||
|       "integrity": "sha512-f9JXC9JrwramDRJHZiIOKJo3PGw/V3riYegLj5kHi8YEJ2k72TNUd1zDW1BG50ILnzJ0cp1faDVJ2pSdolRQfg==", |       "integrity": "sha512-kxgbllgF5f6mEdMeSW6WXlUbV1U77V9ECpA7LOYaY+Tm6RfXOm36EdXbpm+T9VPeaVqXK4QHLAgqay9GSyClgw==", | ||||||
|       "dependencies": { |       "dependencies": { | ||||||
|         "@actions/core": "^1.10.0", |         "@actions/core": "^1.10.0", | ||||||
|         "@actions/github": "^5.1.1", |         "@actions/github": "^5.1.1", | ||||||
| @ -7902,9 +7902,9 @@ | |||||||
|   }, |   }, | ||||||
|   "dependencies": { |   "dependencies": { | ||||||
|     "@actions/artifact": { |     "@actions/artifact": { | ||||||
|       "version": "2.1.9", |       "version": "2.1.8", | ||||||
|       "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.9.tgz", |       "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.8.tgz", | ||||||
|       "integrity": "sha512-f9JXC9JrwramDRJHZiIOKJo3PGw/V3riYegLj5kHi8YEJ2k72TNUd1zDW1BG50ILnzJ0cp1faDVJ2pSdolRQfg==", |       "integrity": "sha512-kxgbllgF5f6mEdMeSW6WXlUbV1U77V9ECpA7LOYaY+Tm6RfXOm36EdXbpm+T9VPeaVqXK4QHLAgqay9GSyClgw==", | ||||||
|       "requires": { |       "requires": { | ||||||
|         "@actions/core": "^1.10.0", |         "@actions/core": "^1.10.0", | ||||||
|         "@actions/github": "^5.1.1", |         "@actions/github": "^5.1.1", | ||||||
|  | |||||||
| @ -1,6 +1,6 @@ | |||||||
| { | { | ||||||
|   "name": "upload-artifact", |   "name": "upload-artifact", | ||||||
|   "version": "4.3.5", |   "version": "4.3.6", | ||||||
|   "description": "Upload an Actions Artifact in a workflow run", |   "description": "Upload an Actions Artifact in a workflow run", | ||||||
|   "main": "dist/upload/index.js", |   "main": "dist/upload/index.js", | ||||||
|   "scripts": { |   "scripts": { | ||||||
| @ -29,7 +29,7 @@ | |||||||
|   }, |   }, | ||||||
|   "homepage": "https://github.com/actions/upload-artifact#readme", |   "homepage": "https://github.com/actions/upload-artifact#readme", | ||||||
|   "dependencies": { |   "dependencies": { | ||||||
|     "@actions/artifact": "^2.1.9", |     "@actions/artifact": "2.1.8", | ||||||
|     "@actions/core": "^1.10.1", |     "@actions/core": "^1.10.1", | ||||||
|     "@actions/github": "^6.0.0", |     "@actions/github": "^6.0.0", | ||||||
|     "@actions/glob": "^0.3.0", |     "@actions/glob": "^0.3.0", | ||||||
|  | |||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user