mirror of
				https://github.com/actions/cache.git
				synced 2025-11-04 07:38:37 +08:00 
			
		
		
		
	Don't autoclose file
This commit is contained in:
		
							parent
							
								
									83f86c103f
								
							
						
					
					
						commit
						b425e87f79
					
				
							
								
								
									
										2
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								dist/restore/index.js
									
									
									
									
										vendored
									
									
								
							| 
						 | 
					@ -1628,7 +1628,7 @@ function saveCache(cacheId, archivePath) {
 | 
				
			||||||
        while (offset < fileSize) {
 | 
					        while (offset < fileSize) {
 | 
				
			||||||
            const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
					            const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
				
			||||||
            const end = offset + chunkSize - 1;
 | 
					            const end = offset + chunkSize - 1;
 | 
				
			||||||
            const chunk = fs.createReadStream(archivePath, { fd, start: offset, end });
 | 
					            const chunk = fs.createReadStream(archivePath, { fd, start: offset, end, autoClose: false });
 | 
				
			||||||
            uploads.push(yield uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
					            uploads.push(yield uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
				
			||||||
            offset += MAX_CHUNK_SIZE;
 | 
					            offset += MAX_CHUNK_SIZE;
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										2
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								dist/save/index.js
									
									
									
									
										vendored
									
									
								
							| 
						 | 
					@ -1628,7 +1628,7 @@ function saveCache(cacheId, archivePath) {
 | 
				
			||||||
        while (offset < fileSize) {
 | 
					        while (offset < fileSize) {
 | 
				
			||||||
            const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
					            const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
				
			||||||
            const end = offset + chunkSize - 1;
 | 
					            const end = offset + chunkSize - 1;
 | 
				
			||||||
            const chunk = fs.createReadStream(archivePath, { fd, start: offset, end });
 | 
					            const chunk = fs.createReadStream(archivePath, { fd, start: offset, end, autoClose: false });
 | 
				
			||||||
            uploads.push(yield uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
					            uploads.push(yield uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
				
			||||||
            offset += MAX_CHUNK_SIZE;
 | 
					            offset += MAX_CHUNK_SIZE;
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -191,7 +191,7 @@ export async function saveCache(
 | 
				
			||||||
    while (offset < fileSize) {
 | 
					    while (offset < fileSize) {
 | 
				
			||||||
        const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
					        const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
 | 
				
			||||||
        const end = offset + chunkSize - 1;
 | 
					        const end = offset + chunkSize - 1;
 | 
				
			||||||
        const chunk = fs.createReadStream(archivePath, { fd, start: offset, end });
 | 
					        const chunk = fs.createReadStream(archivePath, { fd, start: offset, end, autoClose: false });
 | 
				
			||||||
        uploads.push(await uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
					        uploads.push(await uploadChunk(restClient, resourceUrl, chunk, offset, end)); // Making this serial
 | 
				
			||||||
        offset += MAX_CHUNK_SIZE;
 | 
					        offset += MAX_CHUNK_SIZE;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in New Issue
	
	Block a user