1
0
Fork 0
mirror of https://code.forgejo.org/actions/cache.git synced 2024-12-23 20:50:05 +01:00
This commit is contained in:
Josh Gross 2019-12-17 17:08:48 -05:00
parent 14055801c2
commit 574cd74b58
3 changed files with 7 additions and 0 deletions

View file

@ -1631,10 +1631,12 @@ function uploadFile(restClient, cacheId, archivePath) {
const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE; const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
const start = offset; const start = offset;
const end = offset + chunkSize - 1; const end = offset + chunkSize - 1;
core.debug(`Start: ${start} End: ${end}`);
offset += MAX_CHUNK_SIZE; // Do this before losing thread during await? offset += MAX_CHUNK_SIZE; // Do this before losing thread during await?
const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false }); const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false });
responses.push(yield uploadChunk(restClient, resourceUrl, chunk, start, end)); responses.push(yield uploadChunk(restClient, resourceUrl, chunk, start, end));
} }
return Promise.resolve();
}))); })));
fs.closeSync(fd); fs.closeSync(fd);
const failedResponse = responses.find(x => !isSuccessStatusCode(x.statusCode)); const failedResponse = responses.find(x => !isSuccessStatusCode(x.statusCode));

2
dist/save/index.js vendored
View file

@ -1631,10 +1631,12 @@ function uploadFile(restClient, cacheId, archivePath) {
const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE; const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
const start = offset; const start = offset;
const end = offset + chunkSize - 1; const end = offset + chunkSize - 1;
core.debug(`Start: ${start} End: ${end}`);
offset += MAX_CHUNK_SIZE; // Do this before losing thread during await? offset += MAX_CHUNK_SIZE; // Do this before losing thread during await?
const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false }); const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false });
responses.push(yield uploadChunk(restClient, resourceUrl, chunk, start, end)); responses.push(yield uploadChunk(restClient, resourceUrl, chunk, start, end));
} }
return Promise.resolve();
}))); })));
fs.closeSync(fd); fs.closeSync(fd);
const failedResponse = responses.find(x => !isSuccessStatusCode(x.statusCode)); const failedResponse = responses.find(x => !isSuccessStatusCode(x.statusCode));

View file

@ -190,10 +190,13 @@ async function uploadFile(restClient: RestClient, cacheId: number, archivePath:
const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE; const chunkSize = offset + MAX_CHUNK_SIZE > fileSize ? fileSize - offset : MAX_CHUNK_SIZE;
const start = offset; const start = offset;
const end = offset + chunkSize - 1; const end = offset + chunkSize - 1;
core.debug(`Start: ${start} End: ${end}`);
offset += MAX_CHUNK_SIZE; // Do this before losing thread during await? offset += MAX_CHUNK_SIZE; // Do this before losing thread during await?
const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false }); const chunk = fs.createReadStream(archivePath, { fd, start, end, autoClose: false });
responses.push(await uploadChunk(restClient, resourceUrl, chunk, start, end)); responses.push(await uploadChunk(restClient, resourceUrl, chunk, start, end));
} }
return Promise.resolve();
})); }));
fs.closeSync(fd); fs.closeSync(fd);