mirror of
				https://github.com/actions/cache.git
				synced 2025-10-31 03:38:36 +08:00 
			
		
		
		
	Test disabling concurrency
This commit is contained in:
		
							parent
							
								
									aced43a650
								
							
						
					
					
						commit
						6efe05572d
					
				
							
								
								
									
										30
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										30
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							|  | @ -2403,20 +2403,22 @@ function uploadFile(httpClient, cacheId, archivePath) { | ||||||
|         core.debug("Awaiting all uploads"); |         core.debug("Awaiting all uploads"); | ||||||
|         let offset = 0; |         let offset = 0; | ||||||
|         try { |         try { | ||||||
|             yield Promise.all(parallelUploads.map(() => __awaiter(this, void 0, void 0, function* () { |             //        await Promise.all(
 | ||||||
|                 while (offset < fileSize) { |             //            parallelUploads.map(async () => {
 | ||||||
|                     const chunkSize = Math.min(fileSize - offset, MAX_CHUNK_SIZE); |             while (offset < fileSize) { | ||||||
|                     const start = offset; |                 const chunkSize = Math.min(fileSize - offset, MAX_CHUNK_SIZE); | ||||||
|                     const end = offset + chunkSize - 1; |                 const start = offset; | ||||||
|                     offset += MAX_CHUNK_SIZE; |                 const end = offset + chunkSize - 1; | ||||||
|                     yield uploadChunk(httpClient, resourceUrl, () => fs.createReadStream(archivePath, { |                 offset += MAX_CHUNK_SIZE; | ||||||
|                         fd, |                 yield uploadChunk(httpClient, resourceUrl, () => fs.createReadStream(archivePath, { | ||||||
|                         start, |                     fd, | ||||||
|                         end, |                     start, | ||||||
|                         autoClose: false |                     end, | ||||||
|                     }), start, end); |                     autoClose: false | ||||||
|                 } |                 }), start, end); | ||||||
|             }))); |             } | ||||||
|  |             //            })
 | ||||||
|  |             //        );
 | ||||||
|         } |         } | ||||||
|         finally { |         finally { | ||||||
|             fs.closeSync(fd); |             fs.closeSync(fd); | ||||||
|  |  | ||||||
							
								
								
									
										30
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										30
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							|  | @ -2403,20 +2403,22 @@ function uploadFile(httpClient, cacheId, archivePath) { | ||||||
|         core.debug("Awaiting all uploads"); |         core.debug("Awaiting all uploads"); | ||||||
|         let offset = 0; |         let offset = 0; | ||||||
|         try { |         try { | ||||||
|             yield Promise.all(parallelUploads.map(() => __awaiter(this, void 0, void 0, function* () { |             //        await Promise.all(
 | ||||||
|                 while (offset < fileSize) { |             //            parallelUploads.map(async () => {
 | ||||||
|                     const chunkSize = Math.min(fileSize - offset, MAX_CHUNK_SIZE); |             while (offset < fileSize) { | ||||||
|                     const start = offset; |                 const chunkSize = Math.min(fileSize - offset, MAX_CHUNK_SIZE); | ||||||
|                     const end = offset + chunkSize - 1; |                 const start = offset; | ||||||
|                     offset += MAX_CHUNK_SIZE; |                 const end = offset + chunkSize - 1; | ||||||
|                     yield uploadChunk(httpClient, resourceUrl, () => fs.createReadStream(archivePath, { |                 offset += MAX_CHUNK_SIZE; | ||||||
|                         fd, |                 yield uploadChunk(httpClient, resourceUrl, () => fs.createReadStream(archivePath, { | ||||||
|                         start, |                     fd, | ||||||
|                         end, |                     start, | ||||||
|                         autoClose: false |                     end, | ||||||
|                     }), start, end); |                     autoClose: false | ||||||
|                 } |                 }), start, end); | ||||||
|             }))); |             } | ||||||
|  |             //            })
 | ||||||
|  |             //        );
 | ||||||
|         } |         } | ||||||
|         finally { |         finally { | ||||||
|             fs.closeSync(fd); |             fs.closeSync(fd); | ||||||
|  |  | ||||||
|  | @ -342,8 +342,8 @@ async function uploadFile( | ||||||
|     let offset = 0; |     let offset = 0; | ||||||
| 
 | 
 | ||||||
|     try { |     try { | ||||||
|         await Promise.all( | //        await Promise.all(
 | ||||||
|             parallelUploads.map(async () => { | //            parallelUploads.map(async () => {
 | ||||||
|                 while (offset < fileSize) { |                 while (offset < fileSize) { | ||||||
|                     const chunkSize = Math.min( |                     const chunkSize = Math.min( | ||||||
|                         fileSize - offset, |                         fileSize - offset, | ||||||
|  | @ -367,8 +367,8 @@ async function uploadFile( | ||||||
|                         end |                         end | ||||||
|                     ); |                     ); | ||||||
|                 } |                 } | ||||||
|             }) | //            })
 | ||||||
|         ); | //        );
 | ||||||
|     } finally { |     } finally { | ||||||
|         fs.closeSync(fd); |         fs.closeSync(fd); | ||||||
|     } |     } | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue
	
	Block a user
	 Dave Hadka
						Dave Hadka