| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231 |
- "use strict";
- var __importDefault = (this && this.__importDefault) || function (mod) {
- return (mod && mod.__esModule) ? mod : { "default": mod };
- };
- Object.defineProperty(exports, "__esModule", { value: true });
- exports.replace = void 0;
- // tar -r
- const fs_minipass_1 = require("@isaacs/fs-minipass");
- const node_fs_1 = __importDefault(require("node:fs"));
- const node_path_1 = __importDefault(require("node:path"));
- const header_js_1 = require("./header.js");
- const list_js_1 = require("./list.js");
- const make_command_js_1 = require("./make-command.js");
- const options_js_1 = require("./options.js");
- const pack_js_1 = require("./pack.js");
- // starting at the head of the file, read a Header
- // If the checksum is invalid, that's our position to start writing
- // If it is, jump forward by the specified size (round up to 512)
- // and try again.
- // Write the new Pack stream starting there.
- const replaceSync = (opt, files) => {
- const p = new pack_js_1.PackSync(opt);
- let threw = true;
- let fd;
- let position;
- try {
- try {
- fd = node_fs_1.default.openSync(opt.file, 'r+');
- }
- catch (er) {
- if (er?.code === 'ENOENT') {
- fd = node_fs_1.default.openSync(opt.file, 'w+');
- }
- else {
- throw er;
- }
- }
- const st = node_fs_1.default.fstatSync(fd);
- const headBuf = Buffer.alloc(512);
- POSITION: for (position = 0; position < st.size; position += 512) {
- for (let bufPos = 0, bytes = 0; bufPos < 512; bufPos += bytes) {
- bytes = node_fs_1.default.readSync(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos);
- if (position === 0 &&
- headBuf[0] === 0x1f &&
- headBuf[1] === 0x8b) {
- throw new Error('cannot append to compressed archives');
- }
- if (!bytes) {
- break POSITION;
- }
- }
- const h = new header_js_1.Header(headBuf);
- if (!h.cksumValid) {
- break;
- }
- const entryBlockSize = 512 * Math.ceil((h.size || 0) / 512);
- if (position + entryBlockSize + 512 > st.size) {
- break;
- }
- // the 512 for the header we just parsed will be added as well
- // also jump ahead all the blocks for the body
- position += entryBlockSize;
- if (opt.mtimeCache && h.mtime) {
- opt.mtimeCache.set(String(h.path), h.mtime);
- }
- }
- threw = false;
- streamSync(opt, p, position, fd, files);
- }
- finally {
- if (threw) {
- try {
- node_fs_1.default.closeSync(fd);
- }
- catch (er) { }
- }
- }
- };
- const streamSync = (opt, p, position, fd, files) => {
- const stream = new fs_minipass_1.WriteStreamSync(opt.file, {
- fd: fd,
- start: position,
- });
- p.pipe(stream);
- addFilesSync(p, files);
- };
- const replaceAsync = (opt, files) => {
- files = Array.from(files);
- const p = new pack_js_1.Pack(opt);
- const getPos = (fd, size, cb_) => {
- const cb = (er, pos) => {
- if (er) {
- node_fs_1.default.close(fd, _ => cb_(er));
- }
- else {
- cb_(null, pos);
- }
- };
- let position = 0;
- if (size === 0) {
- return cb(null, 0);
- }
- let bufPos = 0;
- const headBuf = Buffer.alloc(512);
- const onread = (er, bytes) => {
- if (er || typeof bytes === 'undefined') {
- return cb(er);
- }
- bufPos += bytes;
- if (bufPos < 512 && bytes) {
- return node_fs_1.default.read(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos, onread);
- }
- if (position === 0 &&
- headBuf[0] === 0x1f &&
- headBuf[1] === 0x8b) {
- return cb(new Error('cannot append to compressed archives'));
- }
- // truncated header
- if (bufPos < 512) {
- return cb(null, position);
- }
- const h = new header_js_1.Header(headBuf);
- if (!h.cksumValid) {
- return cb(null, position);
- }
- /* c8 ignore next */
- const entryBlockSize = 512 * Math.ceil((h.size ?? 0) / 512);
- if (position + entryBlockSize + 512 > size) {
- return cb(null, position);
- }
- position += entryBlockSize + 512;
- if (position >= size) {
- return cb(null, position);
- }
- if (opt.mtimeCache && h.mtime) {
- opt.mtimeCache.set(String(h.path), h.mtime);
- }
- bufPos = 0;
- node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
- };
- node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
- };
- const promise = new Promise((resolve, reject) => {
- p.on('error', reject);
- let flag = 'r+';
- const onopen = (er, fd) => {
- if (er && er.code === 'ENOENT' && flag === 'r+') {
- flag = 'w+';
- return node_fs_1.default.open(opt.file, flag, onopen);
- }
- if (er || !fd) {
- return reject(er);
- }
- node_fs_1.default.fstat(fd, (er, st) => {
- if (er) {
- return node_fs_1.default.close(fd, () => reject(er));
- }
- getPos(fd, st.size, (er, position) => {
- if (er) {
- return reject(er);
- }
- const stream = new fs_minipass_1.WriteStream(opt.file, {
- fd: fd,
- start: position,
- });
- p.pipe(stream);
- stream.on('error', reject);
- stream.on('close', resolve);
- addFilesAsync(p, files);
- });
- });
- };
- node_fs_1.default.open(opt.file, flag, onopen);
- });
- return promise;
- };
- const addFilesSync = (p, files) => {
- files.forEach(file => {
- if (file.charAt(0) === '@') {
- (0, list_js_1.list)({
- file: node_path_1.default.resolve(p.cwd, file.slice(1)),
- sync: true,
- noResume: true,
- onReadEntry: entry => p.add(entry),
- });
- }
- else {
- p.add(file);
- }
- });
- p.end();
- };
- const addFilesAsync = async (p, files) => {
- for (let i = 0; i < files.length; i++) {
- const file = String(files[i]);
- if (file.charAt(0) === '@') {
- await (0, list_js_1.list)({
- file: node_path_1.default.resolve(String(p.cwd), file.slice(1)),
- noResume: true,
- onReadEntry: entry => p.add(entry),
- });
- }
- else {
- p.add(file);
- }
- }
- p.end();
- };
- exports.replace = (0, make_command_js_1.makeCommand)(replaceSync, replaceAsync,
- /* c8 ignore start */
- () => {
- throw new TypeError('file is required');
- }, () => {
- throw new TypeError('file is required');
- },
- /* c8 ignore stop */
- (opt, entries) => {
- if (!(0, options_js_1.isFile)(opt)) {
- throw new TypeError('file is required');
- }
- if (opt.gzip ||
- opt.brotli ||
- opt.file.endsWith('.br') ||
- opt.file.endsWith('.tbr')) {
- throw new TypeError('cannot append to compressed archives');
- }
- if (!entries?.length) {
- throw new TypeError('no paths specified to add/replace');
- }
- });
- //# sourceMappingURL=replace.js.map
|