This takes part of the work being done in #96 that was reverted but still useful. Note Tailwind and Fresh weren't upgraded because there's no security vulnerability in either, and I have found the new versions to be worse in performance. Thos will likely stay at those fixed versions going forward.
10892 lines
331 KiB
JavaScript
10892 lines
331 KiB
JavaScript
// NOTE: This is mostly copied from https://github.com/sunsama/tsdav/tree/cc1c5a09b64c87bbee7e5f171cfcb6748e99469e/dist/tsdav.js and the goal would be to re-implement it as a much simpler class with only the necessary functionality, here. (https://github.com/bewcloud/bewcloud/issues/97)
|
||
var commonjsGlobal = typeof globalThis !== 'undefined' ? globalThis : typeof window !== 'undefined' ? window : typeof global !== 'undefined' ? global : typeof self !== 'undefined' ? self : {};
|
||
|
||
function getDefaultExportFromCjs (x) {
|
||
return x && x.__esModule && Object.prototype.hasOwnProperty.call(x, 'default') ? x['default'] : x;
|
||
}
|
||
|
||
function getAugmentedNamespace(n) {
|
||
if (n.__esModule) return n;
|
||
var f = n.default;
|
||
if (typeof f == "function") {
|
||
var a = function a () {
|
||
if (this instanceof a) {
|
||
return Reflect.construct(f, arguments, this.constructor);
|
||
}
|
||
return f.apply(this, arguments);
|
||
};
|
||
a.prototype = f.prototype;
|
||
} else a = {};
|
||
Object.defineProperty(a, '__esModule', {value: true});
|
||
Object.keys(n).forEach(function (k) {
|
||
var d = Object.getOwnPropertyDescriptor(n, k);
|
||
Object.defineProperty(a, k, d.get ? d : {
|
||
enumerable: true,
|
||
get: function () {
|
||
return n[k];
|
||
}
|
||
});
|
||
});
|
||
return a;
|
||
}
|
||
|
||
var browserPonyfill = {exports: {}};
|
||
|
||
var hasRequiredBrowserPonyfill;
|
||
|
||
function requireBrowserPonyfill () {
|
||
if (hasRequiredBrowserPonyfill) return browserPonyfill.exports;
|
||
hasRequiredBrowserPonyfill = 1;
|
||
(function (module, exports) {
|
||
// Save global object in a variable
|
||
var __global__ =
|
||
(typeof globalThis !== 'undefined' && globalThis) ||
|
||
(typeof self !== 'undefined' && self) ||
|
||
(typeof commonjsGlobal !== 'undefined' && commonjsGlobal);
|
||
// Create an object that extends from __global__ without the fetch function
|
||
var __globalThis__ = (function () {
|
||
function F() {
|
||
this.fetch = false;
|
||
this.DOMException = __global__.DOMException;
|
||
}
|
||
F.prototype = __global__; // Needed for feature detection on whatwg-fetch's code
|
||
return new F();
|
||
})();
|
||
// Wraps whatwg-fetch with a function scope to hijack the global object
|
||
// "globalThis" that's going to be patched
|
||
(function(globalThis) {
|
||
|
||
((function (exports) {
|
||
|
||
var global =
|
||
(typeof globalThis !== 'undefined' && globalThis) ||
|
||
(typeof self !== 'undefined' && self) ||
|
||
(typeof global !== 'undefined' && global);
|
||
|
||
var support = {
|
||
searchParams: 'URLSearchParams' in global,
|
||
iterable: 'Symbol' in global && 'iterator' in Symbol,
|
||
blob:
|
||
'FileReader' in global &&
|
||
'Blob' in global &&
|
||
(function() {
|
||
try {
|
||
new Blob();
|
||
return true
|
||
} catch (e) {
|
||
return false
|
||
}
|
||
})(),
|
||
formData: 'FormData' in global,
|
||
arrayBuffer: 'ArrayBuffer' in global
|
||
};
|
||
|
||
function isDataView(obj) {
|
||
return obj && DataView.prototype.isPrototypeOf(obj)
|
||
}
|
||
|
||
if (support.arrayBuffer) {
|
||
var viewClasses = [
|
||
'[object Int8Array]',
|
||
'[object Uint8Array]',
|
||
'[object Uint8ClampedArray]',
|
||
'[object Int16Array]',
|
||
'[object Uint16Array]',
|
||
'[object Int32Array]',
|
||
'[object Uint32Array]',
|
||
'[object Float32Array]',
|
||
'[object Float64Array]'
|
||
];
|
||
|
||
var isArrayBufferView =
|
||
ArrayBuffer.isView ||
|
||
function(obj) {
|
||
return obj && viewClasses.indexOf(Object.prototype.toString.call(obj)) > -1
|
||
};
|
||
}
|
||
|
||
function normalizeName(name) {
|
||
if (typeof name !== 'string') {
|
||
name = String(name);
|
||
}
|
||
if (/[^a-z0-9\-#$%&'*+.^_`|~!]/i.test(name) || name === '') {
|
||
throw new TypeError('Invalid character in header field name: "' + name + '"')
|
||
}
|
||
return name.toLowerCase()
|
||
}
|
||
|
||
function normalizeValue(value) {
|
||
if (typeof value !== 'string') {
|
||
value = String(value);
|
||
}
|
||
return value
|
||
}
|
||
|
||
// Build a destructive iterator for the value list
|
||
function iteratorFor(items) {
|
||
var iterator = {
|
||
next: function() {
|
||
var value = items.shift();
|
||
return {done: value === undefined, value: value}
|
||
}
|
||
};
|
||
|
||
if (support.iterable) {
|
||
iterator[Symbol.iterator] = function() {
|
||
return iterator
|
||
};
|
||
}
|
||
|
||
return iterator
|
||
}
|
||
|
||
function Headers(headers) {
|
||
this.map = {};
|
||
|
||
if (headers instanceof Headers) {
|
||
headers.forEach(function(value, name) {
|
||
this.append(name, value);
|
||
}, this);
|
||
} else if (Array.isArray(headers)) {
|
||
headers.forEach(function(header) {
|
||
this.append(header[0], header[1]);
|
||
}, this);
|
||
} else if (headers) {
|
||
Object.getOwnPropertyNames(headers).forEach(function(name) {
|
||
this.append(name, headers[name]);
|
||
}, this);
|
||
}
|
||
}
|
||
|
||
Headers.prototype.append = function(name, value) {
|
||
name = normalizeName(name);
|
||
value = normalizeValue(value);
|
||
var oldValue = this.map[name];
|
||
this.map[name] = oldValue ? oldValue + ', ' + value : value;
|
||
};
|
||
|
||
Headers.prototype['delete'] = function(name) {
|
||
delete this.map[normalizeName(name)];
|
||
};
|
||
|
||
Headers.prototype.get = function(name) {
|
||
name = normalizeName(name);
|
||
return this.has(name) ? this.map[name] : null
|
||
};
|
||
|
||
Headers.prototype.has = function(name) {
|
||
return this.map.hasOwnProperty(normalizeName(name))
|
||
};
|
||
|
||
Headers.prototype.set = function(name, value) {
|
||
this.map[normalizeName(name)] = normalizeValue(value);
|
||
};
|
||
|
||
Headers.prototype.forEach = function(callback, thisArg) {
|
||
for (var name in this.map) {
|
||
if (this.map.hasOwnProperty(name)) {
|
||
callback.call(thisArg, this.map[name], name, this);
|
||
}
|
||
}
|
||
};
|
||
|
||
Headers.prototype.keys = function() {
|
||
var items = [];
|
||
this.forEach(function(value, name) {
|
||
items.push(name);
|
||
});
|
||
return iteratorFor(items)
|
||
};
|
||
|
||
Headers.prototype.values = function() {
|
||
var items = [];
|
||
this.forEach(function(value) {
|
||
items.push(value);
|
||
});
|
||
return iteratorFor(items)
|
||
};
|
||
|
||
Headers.prototype.entries = function() {
|
||
var items = [];
|
||
this.forEach(function(value, name) {
|
||
items.push([name, value]);
|
||
});
|
||
return iteratorFor(items)
|
||
};
|
||
|
||
if (support.iterable) {
|
||
Headers.prototype[Symbol.iterator] = Headers.prototype.entries;
|
||
}
|
||
|
||
function consumed(body) {
|
||
if (body.bodyUsed) {
|
||
return Promise.reject(new TypeError('Already read'))
|
||
}
|
||
body.bodyUsed = true;
|
||
}
|
||
|
||
function fileReaderReady(reader) {
|
||
return new Promise(function(resolve, reject) {
|
||
reader.onload = function() {
|
||
resolve(reader.result);
|
||
};
|
||
reader.onerror = function() {
|
||
reject(reader.error);
|
||
};
|
||
})
|
||
}
|
||
|
||
function readBlobAsArrayBuffer(blob) {
|
||
var reader = new FileReader();
|
||
var promise = fileReaderReady(reader);
|
||
reader.readAsArrayBuffer(blob);
|
||
return promise
|
||
}
|
||
|
||
function readBlobAsText(blob) {
|
||
var reader = new FileReader();
|
||
var promise = fileReaderReady(reader);
|
||
reader.readAsText(blob);
|
||
return promise
|
||
}
|
||
|
||
function readArrayBufferAsText(buf) {
|
||
var view = new Uint8Array(buf);
|
||
var chars = new Array(view.length);
|
||
|
||
for (var i = 0; i < view.length; i++) {
|
||
chars[i] = String.fromCharCode(view[i]);
|
||
}
|
||
return chars.join('')
|
||
}
|
||
|
||
function bufferClone(buf) {
|
||
if (buf.slice) {
|
||
return buf.slice(0)
|
||
} else {
|
||
var view = new Uint8Array(buf.byteLength);
|
||
view.set(new Uint8Array(buf));
|
||
return view.buffer
|
||
}
|
||
}
|
||
|
||
function Body() {
|
||
this.bodyUsed = false;
|
||
|
||
this._initBody = function(body) {
|
||
/*
|
||
fetch-mock wraps the Response object in an ES6 Proxy to
|
||
provide useful test harness features such as flush. However, on
|
||
ES5 browsers without fetch or Proxy support pollyfills must be used;
|
||
the proxy-pollyfill is unable to proxy an attribute unless it exists
|
||
on the object before the Proxy is created. This change ensures
|
||
Response.bodyUsed exists on the instance, while maintaining the
|
||
semantic of setting Request.bodyUsed in the constructor before
|
||
_initBody is called.
|
||
*/
|
||
this.bodyUsed = this.bodyUsed;
|
||
this._bodyInit = body;
|
||
if (!body) {
|
||
this._bodyText = '';
|
||
} else if (typeof body === 'string') {
|
||
this._bodyText = body;
|
||
} else if (support.blob && Blob.prototype.isPrototypeOf(body)) {
|
||
this._bodyBlob = body;
|
||
} else if (support.formData && FormData.prototype.isPrototypeOf(body)) {
|
||
this._bodyFormData = body;
|
||
} else if (support.searchParams && URLSearchParams.prototype.isPrototypeOf(body)) {
|
||
this._bodyText = body.toString();
|
||
} else if (support.arrayBuffer && support.blob && isDataView(body)) {
|
||
this._bodyArrayBuffer = bufferClone(body.buffer);
|
||
// IE 10-11 can't handle a DataView body.
|
||
this._bodyInit = new Blob([this._bodyArrayBuffer]);
|
||
} else if (support.arrayBuffer && (ArrayBuffer.prototype.isPrototypeOf(body) || isArrayBufferView(body))) {
|
||
this._bodyArrayBuffer = bufferClone(body);
|
||
} else {
|
||
this._bodyText = body = Object.prototype.toString.call(body);
|
||
}
|
||
|
||
if (!this.headers.get('content-type')) {
|
||
if (typeof body === 'string') {
|
||
this.headers.set('content-type', 'text/plain;charset=UTF-8');
|
||
} else if (this._bodyBlob && this._bodyBlob.type) {
|
||
this.headers.set('content-type', this._bodyBlob.type);
|
||
} else if (support.searchParams && URLSearchParams.prototype.isPrototypeOf(body)) {
|
||
this.headers.set('content-type', 'application/x-www-form-urlencoded;charset=UTF-8');
|
||
}
|
||
}
|
||
};
|
||
|
||
if (support.blob) {
|
||
this.blob = function() {
|
||
var rejected = consumed(this);
|
||
if (rejected) {
|
||
return rejected
|
||
}
|
||
|
||
if (this._bodyBlob) {
|
||
return Promise.resolve(this._bodyBlob)
|
||
} else if (this._bodyArrayBuffer) {
|
||
return Promise.resolve(new Blob([this._bodyArrayBuffer]))
|
||
} else if (this._bodyFormData) {
|
||
throw new Error('could not read FormData body as blob')
|
||
} else {
|
||
return Promise.resolve(new Blob([this._bodyText]))
|
||
}
|
||
};
|
||
|
||
this.arrayBuffer = function() {
|
||
if (this._bodyArrayBuffer) {
|
||
var isConsumed = consumed(this);
|
||
if (isConsumed) {
|
||
return isConsumed
|
||
}
|
||
if (ArrayBuffer.isView(this._bodyArrayBuffer)) {
|
||
return Promise.resolve(
|
||
this._bodyArrayBuffer.buffer.slice(
|
||
this._bodyArrayBuffer.byteOffset,
|
||
this._bodyArrayBuffer.byteOffset + this._bodyArrayBuffer.byteLength
|
||
)
|
||
)
|
||
} else {
|
||
return Promise.resolve(this._bodyArrayBuffer)
|
||
}
|
||
} else {
|
||
return this.blob().then(readBlobAsArrayBuffer)
|
||
}
|
||
};
|
||
}
|
||
|
||
this.text = function() {
|
||
var rejected = consumed(this);
|
||
if (rejected) {
|
||
return rejected
|
||
}
|
||
|
||
if (this._bodyBlob) {
|
||
return readBlobAsText(this._bodyBlob)
|
||
} else if (this._bodyArrayBuffer) {
|
||
return Promise.resolve(readArrayBufferAsText(this._bodyArrayBuffer))
|
||
} else if (this._bodyFormData) {
|
||
throw new Error('could not read FormData body as text')
|
||
} else {
|
||
return Promise.resolve(this._bodyText)
|
||
}
|
||
};
|
||
|
||
if (support.formData) {
|
||
this.formData = function() {
|
||
return this.text().then(decode)
|
||
};
|
||
}
|
||
|
||
this.json = function() {
|
||
return this.text().then(JSON.parse)
|
||
};
|
||
|
||
return this
|
||
}
|
||
|
||
// HTTP methods whose capitalization should be normalized
|
||
var methods = ['DELETE', 'GET', 'HEAD', 'OPTIONS', 'POST', 'PUT'];
|
||
|
||
function normalizeMethod(method) {
|
||
var upcased = method.toUpperCase();
|
||
return methods.indexOf(upcased) > -1 ? upcased : method
|
||
}
|
||
|
||
function Request(input, options) {
|
||
if (!(this instanceof Request)) {
|
||
throw new TypeError('Please use the "new" operator, this DOM object constructor cannot be called as a function.')
|
||
}
|
||
|
||
options = options || {};
|
||
var body = options.body;
|
||
|
||
if (input instanceof Request) {
|
||
if (input.bodyUsed) {
|
||
throw new TypeError('Already read')
|
||
}
|
||
this.url = input.url;
|
||
this.credentials = input.credentials;
|
||
if (!options.headers) {
|
||
this.headers = new Headers(input.headers);
|
||
}
|
||
this.method = input.method;
|
||
this.mode = input.mode;
|
||
this.signal = input.signal;
|
||
if (!body && input._bodyInit != null) {
|
||
body = input._bodyInit;
|
||
input.bodyUsed = true;
|
||
}
|
||
} else {
|
||
this.url = String(input);
|
||
}
|
||
|
||
this.credentials = options.credentials || this.credentials || 'same-origin';
|
||
if (options.headers || !this.headers) {
|
||
this.headers = new Headers(options.headers);
|
||
}
|
||
this.method = normalizeMethod(options.method || this.method || 'GET');
|
||
this.mode = options.mode || this.mode || null;
|
||
this.signal = options.signal || this.signal;
|
||
this.referrer = null;
|
||
|
||
if ((this.method === 'GET' || this.method === 'HEAD') && body) {
|
||
throw new TypeError('Body not allowed for GET or HEAD requests')
|
||
}
|
||
this._initBody(body);
|
||
|
||
if (this.method === 'GET' || this.method === 'HEAD') {
|
||
if (options.cache === 'no-store' || options.cache === 'no-cache') {
|
||
// Search for a '_' parameter in the query string
|
||
var reParamSearch = /([?&])_=[^&]*/;
|
||
if (reParamSearch.test(this.url)) {
|
||
// If it already exists then set the value with the current time
|
||
this.url = this.url.replace(reParamSearch, '$1_=' + new Date().getTime());
|
||
} else {
|
||
// Otherwise add a new '_' parameter to the end with the current time
|
||
var reQueryString = /\?/;
|
||
this.url += (reQueryString.test(this.url) ? '&' : '?') + '_=' + new Date().getTime();
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
Request.prototype.clone = function() {
|
||
return new Request(this, {body: this._bodyInit})
|
||
};
|
||
|
||
function decode(body) {
|
||
var form = new FormData();
|
||
body
|
||
.trim()
|
||
.split('&')
|
||
.forEach(function(bytes) {
|
||
if (bytes) {
|
||
var split = bytes.split('=');
|
||
var name = split.shift().replace(/\+/g, ' ');
|
||
var value = split.join('=').replace(/\+/g, ' ');
|
||
form.append(decodeURIComponent(name), decodeURIComponent(value));
|
||
}
|
||
});
|
||
return form
|
||
}
|
||
|
||
function parseHeaders(rawHeaders) {
|
||
var headers = new Headers();
|
||
// Replace instances of \r\n and \n followed by at least one space or horizontal tab with a space
|
||
// https://tools.ietf.org/html/rfc7230#section-3.2
|
||
var preProcessedHeaders = rawHeaders.replace(/\r?\n[\t ]+/g, ' ');
|
||
// Avoiding split via regex to work around a common IE11 bug with the core-js 3.6.0 regex polyfill
|
||
// https://github.com/github/fetch/issues/748
|
||
// https://github.com/zloirock/core-js/issues/751
|
||
preProcessedHeaders
|
||
.split('\r')
|
||
.map(function(header) {
|
||
return header.indexOf('\n') === 0 ? header.substr(1, header.length) : header
|
||
})
|
||
.forEach(function(line) {
|
||
var parts = line.split(':');
|
||
var key = parts.shift().trim();
|
||
if (key) {
|
||
var value = parts.join(':').trim();
|
||
headers.append(key, value);
|
||
}
|
||
});
|
||
return headers
|
||
}
|
||
|
||
Body.call(Request.prototype);
|
||
|
||
function Response(bodyInit, options) {
|
||
if (!(this instanceof Response)) {
|
||
throw new TypeError('Please use the "new" operator, this DOM object constructor cannot be called as a function.')
|
||
}
|
||
if (!options) {
|
||
options = {};
|
||
}
|
||
|
||
this.type = 'default';
|
||
this.status = options.status === undefined ? 200 : options.status;
|
||
this.ok = this.status >= 200 && this.status < 300;
|
||
this.statusText = options.statusText === undefined ? '' : '' + options.statusText;
|
||
this.headers = new Headers(options.headers);
|
||
this.url = options.url || '';
|
||
this._initBody(bodyInit);
|
||
}
|
||
|
||
Body.call(Response.prototype);
|
||
|
||
Response.prototype.clone = function() {
|
||
return new Response(this._bodyInit, {
|
||
status: this.status,
|
||
statusText: this.statusText,
|
||
headers: new Headers(this.headers),
|
||
url: this.url
|
||
})
|
||
};
|
||
|
||
Response.error = function() {
|
||
var response = new Response(null, {status: 0, statusText: ''});
|
||
response.type = 'error';
|
||
return response
|
||
};
|
||
|
||
var redirectStatuses = [301, 302, 303, 307, 308];
|
||
|
||
Response.redirect = function(url, status) {
|
||
if (redirectStatuses.indexOf(status) === -1) {
|
||
throw new RangeError('Invalid status code')
|
||
}
|
||
|
||
return new Response(null, {status: status, headers: {location: url}})
|
||
};
|
||
|
||
exports.DOMException = global.DOMException;
|
||
try {
|
||
new exports.DOMException();
|
||
} catch (err) {
|
||
exports.DOMException = function(message, name) {
|
||
this.message = message;
|
||
this.name = name;
|
||
var error = Error(message);
|
||
this.stack = error.stack;
|
||
};
|
||
exports.DOMException.prototype = Object.create(Error.prototype);
|
||
exports.DOMException.prototype.constructor = exports.DOMException;
|
||
}
|
||
|
||
function fetch(input, init) {
|
||
return new Promise(function(resolve, reject) {
|
||
var request = new Request(input, init);
|
||
|
||
if (request.signal && request.signal.aborted) {
|
||
return reject(new exports.DOMException('Aborted', 'AbortError'))
|
||
}
|
||
|
||
var xhr = new XMLHttpRequest();
|
||
|
||
function abortXhr() {
|
||
xhr.abort();
|
||
}
|
||
|
||
xhr.onload = function() {
|
||
var options = {
|
||
status: xhr.status,
|
||
statusText: xhr.statusText,
|
||
headers: parseHeaders(xhr.getAllResponseHeaders() || '')
|
||
};
|
||
options.url = 'responseURL' in xhr ? xhr.responseURL : options.headers.get('X-Request-URL');
|
||
var body = 'response' in xhr ? xhr.response : xhr.responseText;
|
||
setTimeout(function() {
|
||
resolve(new Response(body, options));
|
||
}, 0);
|
||
};
|
||
|
||
xhr.onerror = function() {
|
||
setTimeout(function() {
|
||
reject(new TypeError('Network request failed'));
|
||
}, 0);
|
||
};
|
||
|
||
xhr.ontimeout = function() {
|
||
setTimeout(function() {
|
||
reject(new TypeError('Network request failed'));
|
||
}, 0);
|
||
};
|
||
|
||
xhr.onabort = function() {
|
||
setTimeout(function() {
|
||
reject(new exports.DOMException('Aborted', 'AbortError'));
|
||
}, 0);
|
||
};
|
||
|
||
function fixUrl(url) {
|
||
try {
|
||
return url === '' && global.location.href ? global.location.href : url
|
||
} catch (e) {
|
||
return url
|
||
}
|
||
}
|
||
|
||
xhr.open(request.method, fixUrl(request.url), true);
|
||
|
||
if (request.credentials === 'include') {
|
||
xhr.withCredentials = true;
|
||
} else if (request.credentials === 'omit') {
|
||
xhr.withCredentials = false;
|
||
}
|
||
|
||
if ('responseType' in xhr) {
|
||
if (support.blob) {
|
||
xhr.responseType = 'blob';
|
||
} else if (
|
||
support.arrayBuffer &&
|
||
request.headers.get('Content-Type') &&
|
||
request.headers.get('Content-Type').indexOf('application/octet-stream') !== -1
|
||
) {
|
||
xhr.responseType = 'arraybuffer';
|
||
}
|
||
}
|
||
|
||
if (init && typeof init.headers === 'object' && !(init.headers instanceof Headers)) {
|
||
Object.getOwnPropertyNames(init.headers).forEach(function(name) {
|
||
xhr.setRequestHeader(name, normalizeValue(init.headers[name]));
|
||
});
|
||
} else {
|
||
request.headers.forEach(function(value, name) {
|
||
xhr.setRequestHeader(name, value);
|
||
});
|
||
}
|
||
|
||
if (request.signal) {
|
||
request.signal.addEventListener('abort', abortXhr);
|
||
|
||
xhr.onreadystatechange = function() {
|
||
// DONE (success or failure)
|
||
if (xhr.readyState === 4) {
|
||
request.signal.removeEventListener('abort', abortXhr);
|
||
}
|
||
};
|
||
}
|
||
|
||
xhr.send(typeof request._bodyInit === 'undefined' ? null : request._bodyInit);
|
||
})
|
||
}
|
||
|
||
fetch.polyfill = true;
|
||
|
||
if (!global.fetch) {
|
||
global.fetch = fetch;
|
||
global.Headers = Headers;
|
||
global.Request = Request;
|
||
global.Response = Response;
|
||
}
|
||
|
||
exports.Headers = Headers;
|
||
exports.Request = Request;
|
||
exports.Response = Response;
|
||
exports.fetch = fetch;
|
||
|
||
return exports;
|
||
|
||
}))({});
|
||
})(__globalThis__);
|
||
// This is a ponyfill, so...
|
||
__globalThis__.fetch.ponyfill = true;
|
||
delete __globalThis__.fetch.polyfill;
|
||
// Choose between native implementation (__global__) or custom implementation (__globalThis__)
|
||
var ctx = __global__.fetch ? __global__ : __globalThis__;
|
||
exports = ctx.fetch; // To enable: import fetch from 'cross-fetch'
|
||
exports.default = ctx.fetch; // For TypeScript consumers without esModuleInterop.
|
||
exports.fetch = ctx.fetch; // To enable: import {fetch} from 'cross-fetch'
|
||
exports.Headers = ctx.Headers;
|
||
exports.Request = ctx.Request;
|
||
exports.Response = ctx.Response;
|
||
module.exports = exports;
|
||
} (browserPonyfill, browserPonyfill.exports));
|
||
return browserPonyfill.exports;
|
||
}
|
||
|
||
var browserPonyfillExports = requireBrowserPonyfill();
|
||
|
||
var global$1 = (typeof global !== "undefined" ? global :
|
||
typeof self !== "undefined" ? self :
|
||
typeof window !== "undefined" ? window : {});
|
||
|
||
// shim for using process in browser
|
||
// based off https://github.com/defunctzombie/node-process/blob/master/browser.js
|
||
|
||
function defaultSetTimout() {
|
||
throw new Error('setTimeout has not been defined');
|
||
}
|
||
function defaultClearTimeout () {
|
||
throw new Error('clearTimeout has not been defined');
|
||
}
|
||
var cachedSetTimeout = defaultSetTimout;
|
||
var cachedClearTimeout = defaultClearTimeout;
|
||
if (typeof global$1.setTimeout === 'function') {
|
||
cachedSetTimeout = setTimeout;
|
||
}
|
||
if (typeof global$1.clearTimeout === 'function') {
|
||
cachedClearTimeout = clearTimeout;
|
||
}
|
||
|
||
function runTimeout(fun) {
|
||
if (cachedSetTimeout === setTimeout) {
|
||
//normal enviroments in sane situations
|
||
return setTimeout(fun, 0);
|
||
}
|
||
// if setTimeout wasn't available but was latter defined
|
||
if ((cachedSetTimeout === defaultSetTimout || !cachedSetTimeout) && setTimeout) {
|
||
cachedSetTimeout = setTimeout;
|
||
return setTimeout(fun, 0);
|
||
}
|
||
try {
|
||
// when when somebody has screwed with setTimeout but no I.E. maddness
|
||
return cachedSetTimeout(fun, 0);
|
||
} catch(e){
|
||
try {
|
||
// When we are in I.E. but the script has been evaled so I.E. doesn't trust the global object when called normally
|
||
return cachedSetTimeout.call(null, fun, 0);
|
||
} catch(e){
|
||
// same as above but when it's a version of I.E. that must have the global object for 'this', hopfully our context correct otherwise it will throw a global error
|
||
return cachedSetTimeout.call(this, fun, 0);
|
||
}
|
||
}
|
||
|
||
|
||
}
|
||
function runClearTimeout(marker) {
|
||
if (cachedClearTimeout === clearTimeout) {
|
||
//normal enviroments in sane situations
|
||
return clearTimeout(marker);
|
||
}
|
||
// if clearTimeout wasn't available but was latter defined
|
||
if ((cachedClearTimeout === defaultClearTimeout || !cachedClearTimeout) && clearTimeout) {
|
||
cachedClearTimeout = clearTimeout;
|
||
return clearTimeout(marker);
|
||
}
|
||
try {
|
||
// when when somebody has screwed with setTimeout but no I.E. maddness
|
||
return cachedClearTimeout(marker);
|
||
} catch (e){
|
||
try {
|
||
// When we are in I.E. but the script has been evaled so I.E. doesn't trust the global object when called normally
|
||
return cachedClearTimeout.call(null, marker);
|
||
} catch (e){
|
||
// same as above but when it's a version of I.E. that must have the global object for 'this', hopfully our context correct otherwise it will throw a global error.
|
||
// Some versions of I.E. have different rules for clearTimeout vs setTimeout
|
||
return cachedClearTimeout.call(this, marker);
|
||
}
|
||
}
|
||
|
||
|
||
|
||
}
|
||
var queue = [];
|
||
var draining = false;
|
||
var currentQueue;
|
||
var queueIndex = -1;
|
||
|
||
function cleanUpNextTick() {
|
||
if (!draining || !currentQueue) {
|
||
return;
|
||
}
|
||
draining = false;
|
||
if (currentQueue.length) {
|
||
queue = currentQueue.concat(queue);
|
||
} else {
|
||
queueIndex = -1;
|
||
}
|
||
if (queue.length) {
|
||
drainQueue();
|
||
}
|
||
}
|
||
|
||
function drainQueue() {
|
||
if (draining) {
|
||
return;
|
||
}
|
||
var timeout = runTimeout(cleanUpNextTick);
|
||
draining = true;
|
||
|
||
var len = queue.length;
|
||
while(len) {
|
||
currentQueue = queue;
|
||
queue = [];
|
||
while (++queueIndex < len) {
|
||
if (currentQueue) {
|
||
currentQueue[queueIndex].run();
|
||
}
|
||
}
|
||
queueIndex = -1;
|
||
len = queue.length;
|
||
}
|
||
currentQueue = null;
|
||
draining = false;
|
||
runClearTimeout(timeout);
|
||
}
|
||
function nextTick(fun) {
|
||
var args = new Array(arguments.length - 1);
|
||
if (arguments.length > 1) {
|
||
for (var i = 1; i < arguments.length; i++) {
|
||
args[i - 1] = arguments[i];
|
||
}
|
||
}
|
||
queue.push(new Item(fun, args));
|
||
if (queue.length === 1 && !draining) {
|
||
runTimeout(drainQueue);
|
||
}
|
||
}
|
||
// v8 likes predictible objects
|
||
function Item(fun, array) {
|
||
this.fun = fun;
|
||
this.array = array;
|
||
}
|
||
Item.prototype.run = function () {
|
||
this.fun.apply(null, this.array);
|
||
};
|
||
var title = 'browser';
|
||
var platform = 'browser';
|
||
var browser$1 = true;
|
||
var env = {};
|
||
var argv = [];
|
||
var version = ''; // empty string to avoid regexp issues
|
||
var versions = {};
|
||
var release = {};
|
||
var config = {};
|
||
|
||
function noop() {}
|
||
|
||
var on = noop;
|
||
var addListener = noop;
|
||
var once = noop;
|
||
var off = noop;
|
||
var removeListener = noop;
|
||
var removeAllListeners = noop;
|
||
var emit = noop;
|
||
|
||
function binding(name) {
|
||
throw new Error('process.binding is not supported');
|
||
}
|
||
|
||
function cwd () { return '/' }
|
||
function chdir (dir) {
|
||
throw new Error('process.chdir is not supported');
|
||
}function umask() { return 0; }
|
||
|
||
// from https://github.com/kumavis/browser-process-hrtime/blob/master/index.js
|
||
var performance = global$1.performance || {};
|
||
var performanceNow =
|
||
performance.now ||
|
||
performance.mozNow ||
|
||
performance.msNow ||
|
||
performance.oNow ||
|
||
performance.webkitNow ||
|
||
function(){ return (new Date()).getTime() };
|
||
|
||
// generate timestamp or delta
|
||
// see http://nodejs.org/api/process.html#process_process_hrtime
|
||
function hrtime(previousTimestamp){
|
||
var clocktime = performanceNow.call(performance)*1e-3;
|
||
var seconds = Math.floor(clocktime);
|
||
var nanoseconds = Math.floor((clocktime%1)*1e9);
|
||
if (previousTimestamp) {
|
||
seconds = seconds - previousTimestamp[0];
|
||
nanoseconds = nanoseconds - previousTimestamp[1];
|
||
if (nanoseconds<0) {
|
||
seconds--;
|
||
nanoseconds += 1e9;
|
||
}
|
||
}
|
||
return [seconds,nanoseconds]
|
||
}
|
||
|
||
var startTime = new Date();
|
||
function uptime() {
|
||
var currentTime = new Date();
|
||
var dif = currentTime - startTime;
|
||
return dif / 1000;
|
||
}
|
||
|
||
var browser$1$1 = {
|
||
nextTick: nextTick,
|
||
title: title,
|
||
browser: browser$1,
|
||
env: env,
|
||
argv: argv,
|
||
version: version,
|
||
versions: versions,
|
||
on: on,
|
||
addListener: addListener,
|
||
once: once,
|
||
off: off,
|
||
removeListener: removeListener,
|
||
removeAllListeners: removeAllListeners,
|
||
emit: emit,
|
||
binding: binding,
|
||
cwd: cwd,
|
||
chdir: chdir,
|
||
umask: umask,
|
||
hrtime: hrtime,
|
||
platform: platform,
|
||
release: release,
|
||
config: config,
|
||
uptime: uptime
|
||
};
|
||
|
||
var browser = {exports: {}};
|
||
|
||
/**
|
||
* Helpers.
|
||
*/
|
||
|
||
var ms;
|
||
var hasRequiredMs;
|
||
|
||
function requireMs () {
|
||
if (hasRequiredMs) return ms;
|
||
hasRequiredMs = 1;
|
||
var s = 1000;
|
||
var m = s * 60;
|
||
var h = m * 60;
|
||
var d = h * 24;
|
||
var w = d * 7;
|
||
var y = d * 365.25;
|
||
|
||
/**
|
||
* Parse or format the given `val`.
|
||
*
|
||
* Options:
|
||
*
|
||
* - `long` verbose formatting [false]
|
||
*
|
||
* @param {String|Number} val
|
||
* @param {Object} [options]
|
||
* @throws {Error} throw an error if val is not a non-empty string or a number
|
||
* @return {String|Number}
|
||
* @api public
|
||
*/
|
||
|
||
ms = function(val, options) {
|
||
options = options || {};
|
||
var type = typeof val;
|
||
if (type === 'string' && val.length > 0) {
|
||
return parse(val);
|
||
} else if (type === 'number' && isFinite(val)) {
|
||
return options.long ? fmtLong(val) : fmtShort(val);
|
||
}
|
||
throw new Error(
|
||
'val is not a non-empty string or a valid number. val=' +
|
||
JSON.stringify(val)
|
||
);
|
||
};
|
||
|
||
/**
|
||
* Parse the given `str` and return milliseconds.
|
||
*
|
||
* @param {String} str
|
||
* @return {Number}
|
||
* @api private
|
||
*/
|
||
|
||
function parse(str) {
|
||
str = String(str);
|
||
if (str.length > 100) {
|
||
return;
|
||
}
|
||
var match = /^(-?(?:\d+)?\.?\d+) *(milliseconds?|msecs?|ms|seconds?|secs?|s|minutes?|mins?|m|hours?|hrs?|h|days?|d|weeks?|w|years?|yrs?|y)?$/i.exec(
|
||
str
|
||
);
|
||
if (!match) {
|
||
return;
|
||
}
|
||
var n = parseFloat(match[1]);
|
||
var type = (match[2] || 'ms').toLowerCase();
|
||
switch (type) {
|
||
case 'years':
|
||
case 'year':
|
||
case 'yrs':
|
||
case 'yr':
|
||
case 'y':
|
||
return n * y;
|
||
case 'weeks':
|
||
case 'week':
|
||
case 'w':
|
||
return n * w;
|
||
case 'days':
|
||
case 'day':
|
||
case 'd':
|
||
return n * d;
|
||
case 'hours':
|
||
case 'hour':
|
||
case 'hrs':
|
||
case 'hr':
|
||
case 'h':
|
||
return n * h;
|
||
case 'minutes':
|
||
case 'minute':
|
||
case 'mins':
|
||
case 'min':
|
||
case 'm':
|
||
return n * m;
|
||
case 'seconds':
|
||
case 'second':
|
||
case 'secs':
|
||
case 'sec':
|
||
case 's':
|
||
return n * s;
|
||
case 'milliseconds':
|
||
case 'millisecond':
|
||
case 'msecs':
|
||
case 'msec':
|
||
case 'ms':
|
||
return n;
|
||
default:
|
||
return undefined;
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Short format for `ms`.
|
||
*
|
||
* @param {Number} ms
|
||
* @return {String}
|
||
* @api private
|
||
*/
|
||
|
||
function fmtShort(ms) {
|
||
var msAbs = Math.abs(ms);
|
||
if (msAbs >= d) {
|
||
return Math.round(ms / d) + 'd';
|
||
}
|
||
if (msAbs >= h) {
|
||
return Math.round(ms / h) + 'h';
|
||
}
|
||
if (msAbs >= m) {
|
||
return Math.round(ms / m) + 'm';
|
||
}
|
||
if (msAbs >= s) {
|
||
return Math.round(ms / s) + 's';
|
||
}
|
||
return ms + 'ms';
|
||
}
|
||
|
||
/**
|
||
* Long format for `ms`.
|
||
*
|
||
* @param {Number} ms
|
||
* @return {String}
|
||
* @api private
|
||
*/
|
||
|
||
function fmtLong(ms) {
|
||
var msAbs = Math.abs(ms);
|
||
if (msAbs >= d) {
|
||
return plural(ms, msAbs, d, 'day');
|
||
}
|
||
if (msAbs >= h) {
|
||
return plural(ms, msAbs, h, 'hour');
|
||
}
|
||
if (msAbs >= m) {
|
||
return plural(ms, msAbs, m, 'minute');
|
||
}
|
||
if (msAbs >= s) {
|
||
return plural(ms, msAbs, s, 'second');
|
||
}
|
||
return ms + ' ms';
|
||
}
|
||
|
||
/**
|
||
* Pluralization helper.
|
||
*/
|
||
|
||
function plural(ms, msAbs, n, name) {
|
||
var isPlural = msAbs >= n * 1.5;
|
||
return Math.round(ms / n) + ' ' + name + (isPlural ? 's' : '');
|
||
}
|
||
return ms;
|
||
}
|
||
|
||
var common;
|
||
var hasRequiredCommon;
|
||
|
||
function requireCommon () {
|
||
if (hasRequiredCommon) return common;
|
||
hasRequiredCommon = 1;
|
||
/**
|
||
* This is the common logic for both the Node.js and web browser
|
||
* implementations of `debug()`.
|
||
*/
|
||
|
||
function setup(env) {
|
||
createDebug.debug = createDebug;
|
||
createDebug.default = createDebug;
|
||
createDebug.coerce = coerce;
|
||
createDebug.disable = disable;
|
||
createDebug.enable = enable;
|
||
createDebug.enabled = enabled;
|
||
createDebug.humanize = requireMs();
|
||
createDebug.destroy = destroy;
|
||
|
||
Object.keys(env).forEach(key => {
|
||
createDebug[key] = env[key];
|
||
});
|
||
|
||
/**
|
||
* The currently active debug mode names, and names to skip.
|
||
*/
|
||
|
||
createDebug.names = [];
|
||
createDebug.skips = [];
|
||
|
||
/**
|
||
* Map of special "%n" handling functions, for the debug "format" argument.
|
||
*
|
||
* Valid key names are a single, lower or upper-case letter, i.e. "n" and "N".
|
||
*/
|
||
createDebug.formatters = {};
|
||
|
||
/**
|
||
* Selects a color for a debug namespace
|
||
* @param {String} namespace The namespace string for the debug instance to be colored
|
||
* @return {Number|String} An ANSI color code for the given namespace
|
||
* @api private
|
||
*/
|
||
function selectColor(namespace) {
|
||
let hash = 0;
|
||
|
||
for (let i = 0; i < namespace.length; i++) {
|
||
hash = ((hash << 5) - hash) + namespace.charCodeAt(i);
|
||
hash |= 0; // Convert to 32bit integer
|
||
}
|
||
|
||
return createDebug.colors[Math.abs(hash) % createDebug.colors.length];
|
||
}
|
||
createDebug.selectColor = selectColor;
|
||
|
||
/**
|
||
* Create a debugger with the given `namespace`.
|
||
*
|
||
* @param {String} namespace
|
||
* @return {Function}
|
||
* @api public
|
||
*/
|
||
function createDebug(namespace) {
|
||
let prevTime;
|
||
let enableOverride = null;
|
||
let namespacesCache;
|
||
let enabledCache;
|
||
|
||
function debug(...args) {
|
||
// Disabled?
|
||
if (!debug.enabled) {
|
||
return;
|
||
}
|
||
|
||
const self = debug;
|
||
|
||
// Set `diff` timestamp
|
||
const curr = Number(new Date());
|
||
const ms = curr - (prevTime || curr);
|
||
self.diff = ms;
|
||
self.prev = prevTime;
|
||
self.curr = curr;
|
||
prevTime = curr;
|
||
|
||
args[0] = createDebug.coerce(args[0]);
|
||
|
||
if (typeof args[0] !== 'string') {
|
||
// Anything else let's inspect with %O
|
||
args.unshift('%O');
|
||
}
|
||
|
||
// Apply any `formatters` transformations
|
||
let index = 0;
|
||
args[0] = args[0].replace(/%([a-zA-Z%])/g, (match, format) => {
|
||
// If we encounter an escaped % then don't increase the array index
|
||
if (match === '%%') {
|
||
return '%';
|
||
}
|
||
index++;
|
||
const formatter = createDebug.formatters[format];
|
||
if (typeof formatter === 'function') {
|
||
const val = args[index];
|
||
match = formatter.call(self, val);
|
||
|
||
// Now we need to remove `args[index]` since it's inlined in the `format`
|
||
args.splice(index, 1);
|
||
index--;
|
||
}
|
||
return match;
|
||
});
|
||
|
||
// Apply env-specific formatting (colors, etc.)
|
||
createDebug.formatArgs.call(self, args);
|
||
|
||
const logFn = self.log || createDebug.log;
|
||
logFn.apply(self, args);
|
||
}
|
||
|
||
debug.namespace = namespace;
|
||
debug.useColors = createDebug.useColors();
|
||
debug.color = createDebug.selectColor(namespace);
|
||
debug.extend = extend;
|
||
debug.destroy = createDebug.destroy; // XXX Temporary. Will be removed in the next major release.
|
||
|
||
Object.defineProperty(debug, 'enabled', {
|
||
enumerable: true,
|
||
configurable: false,
|
||
get: () => {
|
||
if (enableOverride !== null) {
|
||
return enableOverride;
|
||
}
|
||
if (namespacesCache !== createDebug.namespaces) {
|
||
namespacesCache = createDebug.namespaces;
|
||
enabledCache = createDebug.enabled(namespace);
|
||
}
|
||
|
||
return enabledCache;
|
||
},
|
||
set: v => {
|
||
enableOverride = v;
|
||
}
|
||
});
|
||
|
||
// Env-specific initialization logic for debug instances
|
||
if (typeof createDebug.init === 'function') {
|
||
createDebug.init(debug);
|
||
}
|
||
|
||
return debug;
|
||
}
|
||
|
||
function extend(namespace, delimiter) {
|
||
const newDebug = createDebug(this.namespace + (typeof delimiter === 'undefined' ? ':' : delimiter) + namespace);
|
||
newDebug.log = this.log;
|
||
return newDebug;
|
||
}
|
||
|
||
/**
|
||
* Enables a debug mode by namespaces. This can include modes
|
||
* separated by a colon and wildcards.
|
||
*
|
||
* @param {String} namespaces
|
||
* @api public
|
||
*/
|
||
function enable(namespaces) {
|
||
createDebug.save(namespaces);
|
||
createDebug.namespaces = namespaces;
|
||
|
||
createDebug.names = [];
|
||
createDebug.skips = [];
|
||
|
||
let i;
|
||
const split = (typeof namespaces === 'string' ? namespaces : '').split(/[\s,]+/);
|
||
const len = split.length;
|
||
|
||
for (i = 0; i < len; i++) {
|
||
if (!split[i]) {
|
||
// ignore empty strings
|
||
continue;
|
||
}
|
||
|
||
namespaces = split[i].replace(/\*/g, '.*?');
|
||
|
||
if (namespaces[0] === '-') {
|
||
createDebug.skips.push(new RegExp('^' + namespaces.slice(1) + '$'));
|
||
} else {
|
||
createDebug.names.push(new RegExp('^' + namespaces + '$'));
|
||
}
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Disable debug output.
|
||
*
|
||
* @return {String} namespaces
|
||
* @api public
|
||
*/
|
||
function disable() {
|
||
const namespaces = [
|
||
...createDebug.names.map(toNamespace),
|
||
...createDebug.skips.map(toNamespace).map(namespace => '-' + namespace)
|
||
].join(',');
|
||
createDebug.enable('');
|
||
return namespaces;
|
||
}
|
||
|
||
/**
|
||
* Returns true if the given mode name is enabled, false otherwise.
|
||
*
|
||
* @param {String} name
|
||
* @return {Boolean}
|
||
* @api public
|
||
*/
|
||
function enabled(name) {
|
||
if (name[name.length - 1] === '*') {
|
||
return true;
|
||
}
|
||
|
||
let i;
|
||
let len;
|
||
|
||
for (i = 0, len = createDebug.skips.length; i < len; i++) {
|
||
if (createDebug.skips[i].test(name)) {
|
||
return false;
|
||
}
|
||
}
|
||
|
||
for (i = 0, len = createDebug.names.length; i < len; i++) {
|
||
if (createDebug.names[i].test(name)) {
|
||
return true;
|
||
}
|
||
}
|
||
|
||
return false;
|
||
}
|
||
|
||
/**
|
||
* Convert regexp to namespace
|
||
*
|
||
* @param {RegExp} regxep
|
||
* @return {String} namespace
|
||
* @api private
|
||
*/
|
||
function toNamespace(regexp) {
|
||
return regexp.toString()
|
||
.substring(2, regexp.toString().length - 2)
|
||
.replace(/\.\*\?$/, '*');
|
||
}
|
||
|
||
/**
|
||
* Coerce `val`.
|
||
*
|
||
* @param {Mixed} val
|
||
* @return {Mixed}
|
||
* @api private
|
||
*/
|
||
function coerce(val) {
|
||
if (val instanceof Error) {
|
||
return val.stack || val.message;
|
||
}
|
||
return val;
|
||
}
|
||
|
||
/**
|
||
* XXX DO NOT USE. This is a temporary stub function.
|
||
* XXX It WILL be removed in the next major release.
|
||
*/
|
||
function destroy() {
|
||
console.warn('Instance method `debug.destroy()` is deprecated and no longer does anything. It will be removed in the next major version of `debug`.');
|
||
}
|
||
|
||
createDebug.enable(createDebug.load());
|
||
|
||
return createDebug;
|
||
}
|
||
|
||
common = setup;
|
||
return common;
|
||
}
|
||
|
||
var hasRequiredBrowser;
|
||
|
||
function requireBrowser () {
|
||
if (hasRequiredBrowser) return browser.exports;
|
||
hasRequiredBrowser = 1;
|
||
(function (module, exports) {
|
||
/**
|
||
* This is the web browser implementation of `debug()`.
|
||
*/
|
||
|
||
exports.formatArgs = formatArgs;
|
||
exports.save = save;
|
||
exports.load = load;
|
||
exports.useColors = useColors;
|
||
exports.storage = localstorage();
|
||
exports.destroy = (() => {
|
||
let warned = false;
|
||
|
||
return () => {
|
||
if (!warned) {
|
||
warned = true;
|
||
console.warn('Instance method `debug.destroy()` is deprecated and no longer does anything. It will be removed in the next major version of `debug`.');
|
||
}
|
||
};
|
||
})();
|
||
|
||
/**
|
||
* Colors.
|
||
*/
|
||
|
||
exports.colors = [
|
||
'#0000CC',
|
||
'#0000FF',
|
||
'#0033CC',
|
||
'#0033FF',
|
||
'#0066CC',
|
||
'#0066FF',
|
||
'#0099CC',
|
||
'#0099FF',
|
||
'#00CC00',
|
||
'#00CC33',
|
||
'#00CC66',
|
||
'#00CC99',
|
||
'#00CCCC',
|
||
'#00CCFF',
|
||
'#3300CC',
|
||
'#3300FF',
|
||
'#3333CC',
|
||
'#3333FF',
|
||
'#3366CC',
|
||
'#3366FF',
|
||
'#3399CC',
|
||
'#3399FF',
|
||
'#33CC00',
|
||
'#33CC33',
|
||
'#33CC66',
|
||
'#33CC99',
|
||
'#33CCCC',
|
||
'#33CCFF',
|
||
'#6600CC',
|
||
'#6600FF',
|
||
'#6633CC',
|
||
'#6633FF',
|
||
'#66CC00',
|
||
'#66CC33',
|
||
'#9900CC',
|
||
'#9900FF',
|
||
'#9933CC',
|
||
'#9933FF',
|
||
'#99CC00',
|
||
'#99CC33',
|
||
'#CC0000',
|
||
'#CC0033',
|
||
'#CC0066',
|
||
'#CC0099',
|
||
'#CC00CC',
|
||
'#CC00FF',
|
||
'#CC3300',
|
||
'#CC3333',
|
||
'#CC3366',
|
||
'#CC3399',
|
||
'#CC33CC',
|
||
'#CC33FF',
|
||
'#CC6600',
|
||
'#CC6633',
|
||
'#CC9900',
|
||
'#CC9933',
|
||
'#CCCC00',
|
||
'#CCCC33',
|
||
'#FF0000',
|
||
'#FF0033',
|
||
'#FF0066',
|
||
'#FF0099',
|
||
'#FF00CC',
|
||
'#FF00FF',
|
||
'#FF3300',
|
||
'#FF3333',
|
||
'#FF3366',
|
||
'#FF3399',
|
||
'#FF33CC',
|
||
'#FF33FF',
|
||
'#FF6600',
|
||
'#FF6633',
|
||
'#FF9900',
|
||
'#FF9933',
|
||
'#FFCC00',
|
||
'#FFCC33'
|
||
];
|
||
|
||
/**
|
||
* Currently only WebKit-based Web Inspectors, Firefox >= v31,
|
||
* and the Firebug extension (any Firefox version) are known
|
||
* to support "%c" CSS customizations.
|
||
*
|
||
* TODO: add a `localStorage` variable to explicitly enable/disable colors
|
||
*/
|
||
|
||
// eslint-disable-next-line complexity
|
||
function useColors() {
|
||
// NB: In an Electron preload script, document will be defined but not fully
|
||
// initialized. Since we know we're in Chrome, we'll just detect this case
|
||
// explicitly
|
||
if (typeof window !== 'undefined' && window.process && (window.process.type === 'renderer' || window.process.__nwjs)) {
|
||
return true;
|
||
}
|
||
|
||
// Internet Explorer and Edge do not support colors.
|
||
if (typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/(edge|trident)\/(\d+)/)) {
|
||
return false;
|
||
}
|
||
|
||
// Is webkit? http://stackoverflow.com/a/16459606/376773
|
||
// document is undefined in react-native: https://github.com/facebook/react-native/pull/1632
|
||
return (typeof document !== 'undefined' && document.documentElement && document.documentElement.style && document.documentElement.style.WebkitAppearance) ||
|
||
// Is firebug? http://stackoverflow.com/a/398120/376773
|
||
(typeof window !== 'undefined' && window.console && (window.console.firebug || (window.console.exception && window.console.table))) ||
|
||
// Is firefox >= v31?
|
||
// https://developer.mozilla.org/en-US/docs/Tools/Web_Console#Styling_messages
|
||
(typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/firefox\/(\d+)/) && parseInt(RegExp.$1, 10) >= 31) ||
|
||
// Double check webkit in userAgent just in case we are in a worker
|
||
(typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/applewebkit\/(\d+)/));
|
||
}
|
||
|
||
/**
|
||
* Colorize log arguments if enabled.
|
||
*
|
||
* @api public
|
||
*/
|
||
|
||
function formatArgs(args) {
|
||
args[0] = (this.useColors ? '%c' : '') +
|
||
this.namespace +
|
||
(this.useColors ? ' %c' : ' ') +
|
||
args[0] +
|
||
(this.useColors ? '%c ' : ' ') +
|
||
'+' + module.exports.humanize(this.diff);
|
||
|
||
if (!this.useColors) {
|
||
return;
|
||
}
|
||
|
||
const c = 'color: ' + this.color;
|
||
args.splice(1, 0, c, 'color: inherit');
|
||
|
||
// The final "%c" is somewhat tricky, because there could be other
|
||
// arguments passed either before or after the %c, so we need to
|
||
// figure out the correct index to insert the CSS into
|
||
let index = 0;
|
||
let lastC = 0;
|
||
args[0].replace(/%[a-zA-Z%]/g, match => {
|
||
if (match === '%%') {
|
||
return;
|
||
}
|
||
index++;
|
||
if (match === '%c') {
|
||
// We only are interested in the *last* %c
|
||
// (the user may have provided their own)
|
||
lastC = index;
|
||
}
|
||
});
|
||
|
||
args.splice(lastC, 0, c);
|
||
}
|
||
|
||
/**
|
||
* Invokes `console.debug()` when available.
|
||
* No-op when `console.debug` is not a "function".
|
||
* If `console.debug` is not available, falls back
|
||
* to `console.log`.
|
||
*
|
||
* @api public
|
||
*/
|
||
exports.log = console.debug || console.log || (() => {});
|
||
|
||
/**
|
||
* Save `namespaces`.
|
||
*
|
||
* @param {String} namespaces
|
||
* @api private
|
||
*/
|
||
function save(namespaces) {
|
||
try {
|
||
if (namespaces) {
|
||
exports.storage.setItem('debug', namespaces);
|
||
} else {
|
||
exports.storage.removeItem('debug');
|
||
}
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Load `namespaces`.
|
||
*
|
||
* @return {String} returns the previously persisted debug modes
|
||
* @api private
|
||
*/
|
||
function load() {
|
||
let r;
|
||
try {
|
||
r = exports.storage.getItem('debug');
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
|
||
// If debug isn't set in LS, and we're in Electron, try to load $DEBUG
|
||
if (!r && typeof browser$1$1 !== 'undefined' && 'env' in browser$1$1) {
|
||
r = browser$1$1.env.DEBUG;
|
||
}
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* Localstorage attempts to return the localstorage.
|
||
*
|
||
* This is necessary because safari throws
|
||
* when a user disables cookies/localstorage
|
||
* and you attempt to access it.
|
||
*
|
||
* @return {LocalStorage}
|
||
* @api private
|
||
*/
|
||
|
||
function localstorage() {
|
||
try {
|
||
// TVMLKit (Apple TV JS Runtime) does not have a window object, just localStorage in the global context
|
||
// The Browser also has localStorage in the global context.
|
||
return localStorage;
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
}
|
||
|
||
module.exports = requireCommon()(exports);
|
||
|
||
const {formatters} = module.exports;
|
||
|
||
/**
|
||
* Map %j to `JSON.stringify()`, since no Web Inspectors do that by default.
|
||
*/
|
||
|
||
formatters.j = function (v) {
|
||
try {
|
||
return JSON.stringify(v);
|
||
} catch (error) {
|
||
return '[UnexpectedJSONParseError]: ' + error.message;
|
||
}
|
||
};
|
||
} (browser, browser.exports));
|
||
return browser.exports;
|
||
}
|
||
|
||
var browserExports = requireBrowser();
|
||
var getLogger = /*@__PURE__*/getDefaultExportFromCjs(browserExports);
|
||
|
||
var DAVNamespace;
|
||
(function (DAVNamespace) {
|
||
DAVNamespace["CALENDAR_SERVER"] = "http://calendarserver.org/ns/";
|
||
DAVNamespace["CALDAV_APPLE"] = "http://apple.com/ns/ical/";
|
||
DAVNamespace["CALDAV"] = "urn:ietf:params:xml:ns:caldav";
|
||
DAVNamespace["CARDDAV"] = "urn:ietf:params:xml:ns:carddav";
|
||
DAVNamespace["DAV"] = "DAV:";
|
||
})(DAVNamespace || (DAVNamespace = {}));
|
||
const DAVAttributeMap = {
|
||
[DAVNamespace.CALDAV]: 'xmlns:c',
|
||
[DAVNamespace.CARDDAV]: 'xmlns:card',
|
||
[DAVNamespace.CALENDAR_SERVER]: 'xmlns:cs',
|
||
[DAVNamespace.CALDAV_APPLE]: 'xmlns:ca',
|
||
[DAVNamespace.DAV]: 'xmlns:d',
|
||
};
|
||
var DAVNamespaceShort;
|
||
(function (DAVNamespaceShort) {
|
||
DAVNamespaceShort["CALDAV"] = "c";
|
||
DAVNamespaceShort["CARDDAV"] = "card";
|
||
DAVNamespaceShort["CALENDAR_SERVER"] = "cs";
|
||
DAVNamespaceShort["CALDAV_APPLE"] = "ca";
|
||
DAVNamespaceShort["DAV"] = "d";
|
||
})(DAVNamespaceShort || (DAVNamespaceShort = {}));
|
||
var ICALObjects;
|
||
(function (ICALObjects) {
|
||
ICALObjects["VEVENT"] = "VEVENT";
|
||
ICALObjects["VTODO"] = "VTODO";
|
||
ICALObjects["VJOURNAL"] = "VJOURNAL";
|
||
ICALObjects["VFREEBUSY"] = "VFREEBUSY";
|
||
ICALObjects["VTIMEZONE"] = "VTIMEZONE";
|
||
ICALObjects["VALARM"] = "VALARM";
|
||
})(ICALObjects || (ICALObjects = {}));
|
||
|
||
var lookup = [];
|
||
var revLookup = [];
|
||
var Arr = typeof Uint8Array !== 'undefined' ? Uint8Array : Array;
|
||
var inited = false;
|
||
function init () {
|
||
inited = true;
|
||
var code = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/';
|
||
for (var i = 0, len = code.length; i < len; ++i) {
|
||
lookup[i] = code[i];
|
||
revLookup[code.charCodeAt(i)] = i;
|
||
}
|
||
|
||
revLookup['-'.charCodeAt(0)] = 62;
|
||
revLookup['_'.charCodeAt(0)] = 63;
|
||
}
|
||
|
||
function toByteArray (b64) {
|
||
if (!inited) {
|
||
init();
|
||
}
|
||
var i, j, l, tmp, placeHolders, arr;
|
||
var len = b64.length;
|
||
|
||
if (len % 4 > 0) {
|
||
throw new Error('Invalid string. Length must be a multiple of 4')
|
||
}
|
||
|
||
// the number of equal signs (place holders)
|
||
// if there are two placeholders, than the two characters before it
|
||
// represent one byte
|
||
// if there is only one, then the three characters before it represent 2 bytes
|
||
// this is just a cheap hack to not do indexOf twice
|
||
placeHolders = b64[len - 2] === '=' ? 2 : b64[len - 1] === '=' ? 1 : 0;
|
||
|
||
// base64 is 4/3 + up to two characters of the original data
|
||
arr = new Arr(len * 3 / 4 - placeHolders);
|
||
|
||
// if there are placeholders, only get up to the last complete 4 chars
|
||
l = placeHolders > 0 ? len - 4 : len;
|
||
|
||
var L = 0;
|
||
|
||
for (i = 0, j = 0; i < l; i += 4, j += 3) {
|
||
tmp = (revLookup[b64.charCodeAt(i)] << 18) | (revLookup[b64.charCodeAt(i + 1)] << 12) | (revLookup[b64.charCodeAt(i + 2)] << 6) | revLookup[b64.charCodeAt(i + 3)];
|
||
arr[L++] = (tmp >> 16) & 0xFF;
|
||
arr[L++] = (tmp >> 8) & 0xFF;
|
||
arr[L++] = tmp & 0xFF;
|
||
}
|
||
|
||
if (placeHolders === 2) {
|
||
tmp = (revLookup[b64.charCodeAt(i)] << 2) | (revLookup[b64.charCodeAt(i + 1)] >> 4);
|
||
arr[L++] = tmp & 0xFF;
|
||
} else if (placeHolders === 1) {
|
||
tmp = (revLookup[b64.charCodeAt(i)] << 10) | (revLookup[b64.charCodeAt(i + 1)] << 4) | (revLookup[b64.charCodeAt(i + 2)] >> 2);
|
||
arr[L++] = (tmp >> 8) & 0xFF;
|
||
arr[L++] = tmp & 0xFF;
|
||
}
|
||
|
||
return arr
|
||
}
|
||
|
||
function tripletToBase64 (num) {
|
||
return lookup[num >> 18 & 0x3F] + lookup[num >> 12 & 0x3F] + lookup[num >> 6 & 0x3F] + lookup[num & 0x3F]
|
||
}
|
||
|
||
function encodeChunk (uint8, start, end) {
|
||
var tmp;
|
||
var output = [];
|
||
for (var i = start; i < end; i += 3) {
|
||
tmp = (uint8[i] << 16) + (uint8[i + 1] << 8) + (uint8[i + 2]);
|
||
output.push(tripletToBase64(tmp));
|
||
}
|
||
return output.join('')
|
||
}
|
||
|
||
function fromByteArray (uint8) {
|
||
if (!inited) {
|
||
init();
|
||
}
|
||
var tmp;
|
||
var len = uint8.length;
|
||
var extraBytes = len % 3; // if we have 1 byte left, pad 2 bytes
|
||
var output = '';
|
||
var parts = [];
|
||
var maxChunkLength = 16383; // must be multiple of 3
|
||
|
||
// go through the array every three bytes, we'll deal with trailing stuff later
|
||
for (var i = 0, len2 = len - extraBytes; i < len2; i += maxChunkLength) {
|
||
parts.push(encodeChunk(uint8, i, (i + maxChunkLength) > len2 ? len2 : (i + maxChunkLength)));
|
||
}
|
||
|
||
// pad the end with zeros, but make sure to not forget the extra bytes
|
||
if (extraBytes === 1) {
|
||
tmp = uint8[len - 1];
|
||
output += lookup[tmp >> 2];
|
||
output += lookup[(tmp << 4) & 0x3F];
|
||
output += '==';
|
||
} else if (extraBytes === 2) {
|
||
tmp = (uint8[len - 2] << 8) + (uint8[len - 1]);
|
||
output += lookup[tmp >> 10];
|
||
output += lookup[(tmp >> 4) & 0x3F];
|
||
output += lookup[(tmp << 2) & 0x3F];
|
||
output += '=';
|
||
}
|
||
|
||
parts.push(output);
|
||
|
||
return parts.join('')
|
||
}
|
||
|
||
function read (buffer, offset, isLE, mLen, nBytes) {
|
||
var e, m;
|
||
var eLen = nBytes * 8 - mLen - 1;
|
||
var eMax = (1 << eLen) - 1;
|
||
var eBias = eMax >> 1;
|
||
var nBits = -7;
|
||
var i = isLE ? (nBytes - 1) : 0;
|
||
var d = isLE ? -1 : 1;
|
||
var s = buffer[offset + i];
|
||
|
||
i += d;
|
||
|
||
e = s & ((1 << (-nBits)) - 1);
|
||
s >>= (-nBits);
|
||
nBits += eLen;
|
||
for (; nBits > 0; e = e * 256 + buffer[offset + i], i += d, nBits -= 8) {}
|
||
|
||
m = e & ((1 << (-nBits)) - 1);
|
||
e >>= (-nBits);
|
||
nBits += mLen;
|
||
for (; nBits > 0; m = m * 256 + buffer[offset + i], i += d, nBits -= 8) {}
|
||
|
||
if (e === 0) {
|
||
e = 1 - eBias;
|
||
} else if (e === eMax) {
|
||
return m ? NaN : ((s ? -1 : 1) * Infinity)
|
||
} else {
|
||
m = m + Math.pow(2, mLen);
|
||
e = e - eBias;
|
||
}
|
||
return (s ? -1 : 1) * m * Math.pow(2, e - mLen)
|
||
}
|
||
|
||
function write (buffer, value, offset, isLE, mLen, nBytes) {
|
||
var e, m, c;
|
||
var eLen = nBytes * 8 - mLen - 1;
|
||
var eMax = (1 << eLen) - 1;
|
||
var eBias = eMax >> 1;
|
||
var rt = (mLen === 23 ? Math.pow(2, -24) - Math.pow(2, -77) : 0);
|
||
var i = isLE ? 0 : (nBytes - 1);
|
||
var d = isLE ? 1 : -1;
|
||
var s = value < 0 || (value === 0 && 1 / value < 0) ? 1 : 0;
|
||
|
||
value = Math.abs(value);
|
||
|
||
if (isNaN(value) || value === Infinity) {
|
||
m = isNaN(value) ? 1 : 0;
|
||
e = eMax;
|
||
} else {
|
||
e = Math.floor(Math.log(value) / Math.LN2);
|
||
if (value * (c = Math.pow(2, -e)) < 1) {
|
||
e--;
|
||
c *= 2;
|
||
}
|
||
if (e + eBias >= 1) {
|
||
value += rt / c;
|
||
} else {
|
||
value += rt * Math.pow(2, 1 - eBias);
|
||
}
|
||
if (value * c >= 2) {
|
||
e++;
|
||
c /= 2;
|
||
}
|
||
|
||
if (e + eBias >= eMax) {
|
||
m = 0;
|
||
e = eMax;
|
||
} else if (e + eBias >= 1) {
|
||
m = (value * c - 1) * Math.pow(2, mLen);
|
||
e = e + eBias;
|
||
} else {
|
||
m = value * Math.pow(2, eBias - 1) * Math.pow(2, mLen);
|
||
e = 0;
|
||
}
|
||
}
|
||
|
||
for (; mLen >= 8; buffer[offset + i] = m & 0xff, i += d, m /= 256, mLen -= 8) {}
|
||
|
||
e = (e << mLen) | m;
|
||
eLen += mLen;
|
||
for (; eLen > 0; buffer[offset + i] = e & 0xff, i += d, e /= 256, eLen -= 8) {}
|
||
|
||
buffer[offset + i - d] |= s * 128;
|
||
}
|
||
|
||
var toString = {}.toString;
|
||
|
||
var isArray$1 = Array.isArray || function (arr) {
|
||
return toString.call(arr) == '[object Array]';
|
||
};
|
||
|
||
/*!
|
||
* The buffer module from node.js, for the browser.
|
||
*
|
||
* @author Feross Aboukhadijeh <feross@feross.org> <http://feross.org>
|
||
* @license MIT
|
||
*/
|
||
/* eslint-disable no-proto */
|
||
|
||
|
||
var INSPECT_MAX_BYTES = 50;
|
||
|
||
/**
|
||
* If `Buffer.TYPED_ARRAY_SUPPORT`:
|
||
* === true Use Uint8Array implementation (fastest)
|
||
* === false Use Object implementation (most compatible, even IE6)
|
||
*
|
||
* Browsers that support typed arrays are IE 10+, Firefox 4+, Chrome 7+, Safari 5.1+,
|
||
* Opera 11.6+, iOS 4.2+.
|
||
*
|
||
* Due to various browser bugs, sometimes the Object implementation will be used even
|
||
* when the browser supports typed arrays.
|
||
*
|
||
* Note:
|
||
*
|
||
* - Firefox 4-29 lacks support for adding new properties to `Uint8Array` instances,
|
||
* See: https://bugzilla.mozilla.org/show_bug.cgi?id=695438.
|
||
*
|
||
* - Chrome 9-10 is missing the `TypedArray.prototype.subarray` function.
|
||
*
|
||
* - IE10 has a broken `TypedArray.prototype.subarray` function which returns arrays of
|
||
* incorrect length in some situations.
|
||
|
||
* We detect these buggy browsers and set `Buffer.TYPED_ARRAY_SUPPORT` to `false` so they
|
||
* get the Object implementation, which is slower but behaves correctly.
|
||
*/
|
||
Buffer.TYPED_ARRAY_SUPPORT = global$1.TYPED_ARRAY_SUPPORT !== undefined
|
||
? global$1.TYPED_ARRAY_SUPPORT
|
||
: true;
|
||
|
||
/*
|
||
* Export kMaxLength after typed array support is determined.
|
||
*/
|
||
kMaxLength();
|
||
|
||
function kMaxLength () {
|
||
return Buffer.TYPED_ARRAY_SUPPORT
|
||
? 0x7fffffff
|
||
: 0x3fffffff
|
||
}
|
||
|
||
function createBuffer (that, length) {
|
||
if (kMaxLength() < length) {
|
||
throw new RangeError('Invalid typed array length')
|
||
}
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
// Return an augmented `Uint8Array` instance, for best performance
|
||
that = new Uint8Array(length);
|
||
that.__proto__ = Buffer.prototype;
|
||
} else {
|
||
// Fallback: Return an object instance of the Buffer class
|
||
if (that === null) {
|
||
that = new Buffer(length);
|
||
}
|
||
that.length = length;
|
||
}
|
||
|
||
return that
|
||
}
|
||
|
||
/**
|
||
* The Buffer constructor returns instances of `Uint8Array` that have their
|
||
* prototype changed to `Buffer.prototype`. Furthermore, `Buffer` is a subclass of
|
||
* `Uint8Array`, so the returned instances will have all the node `Buffer` methods
|
||
* and the `Uint8Array` methods. Square bracket notation works as expected -- it
|
||
* returns a single octet.
|
||
*
|
||
* The `Uint8Array` prototype remains unmodified.
|
||
*/
|
||
|
||
function Buffer (arg, encodingOrOffset, length) {
|
||
if (!Buffer.TYPED_ARRAY_SUPPORT && !(this instanceof Buffer)) {
|
||
return new Buffer(arg, encodingOrOffset, length)
|
||
}
|
||
|
||
// Common case.
|
||
if (typeof arg === 'number') {
|
||
if (typeof encodingOrOffset === 'string') {
|
||
throw new Error(
|
||
'If encoding is specified then the first argument must be a string'
|
||
)
|
||
}
|
||
return allocUnsafe(this, arg)
|
||
}
|
||
return from(this, arg, encodingOrOffset, length)
|
||
}
|
||
|
||
Buffer.poolSize = 8192; // not used by this implementation
|
||
|
||
// TODO: Legacy, not needed anymore. Remove in next major version.
|
||
Buffer._augment = function (arr) {
|
||
arr.__proto__ = Buffer.prototype;
|
||
return arr
|
||
};
|
||
|
||
function from (that, value, encodingOrOffset, length) {
|
||
if (typeof value === 'number') {
|
||
throw new TypeError('"value" argument must not be a number')
|
||
}
|
||
|
||
if (typeof ArrayBuffer !== 'undefined' && value instanceof ArrayBuffer) {
|
||
return fromArrayBuffer(that, value, encodingOrOffset, length)
|
||
}
|
||
|
||
if (typeof value === 'string') {
|
||
return fromString(that, value, encodingOrOffset)
|
||
}
|
||
|
||
return fromObject(that, value)
|
||
}
|
||
|
||
/**
|
||
* Functionally equivalent to Buffer(arg, encoding) but throws a TypeError
|
||
* if value is a number.
|
||
* Buffer.from(str[, encoding])
|
||
* Buffer.from(array)
|
||
* Buffer.from(buffer)
|
||
* Buffer.from(arrayBuffer[, byteOffset[, length]])
|
||
**/
|
||
Buffer.from = function (value, encodingOrOffset, length) {
|
||
return from(null, value, encodingOrOffset, length)
|
||
};
|
||
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
Buffer.prototype.__proto__ = Uint8Array.prototype;
|
||
Buffer.__proto__ = Uint8Array;
|
||
if (typeof Symbol !== 'undefined' && Symbol.species &&
|
||
Buffer[Symbol.species] === Buffer) ;
|
||
}
|
||
|
||
function assertSize (size) {
|
||
if (typeof size !== 'number') {
|
||
throw new TypeError('"size" argument must be a number')
|
||
} else if (size < 0) {
|
||
throw new RangeError('"size" argument must not be negative')
|
||
}
|
||
}
|
||
|
||
function alloc (that, size, fill, encoding) {
|
||
assertSize(size);
|
||
if (size <= 0) {
|
||
return createBuffer(that, size)
|
||
}
|
||
if (fill !== undefined) {
|
||
// Only pay attention to encoding if it's a string. This
|
||
// prevents accidentally sending in a number that would
|
||
// be interpretted as a start offset.
|
||
return typeof encoding === 'string'
|
||
? createBuffer(that, size).fill(fill, encoding)
|
||
: createBuffer(that, size).fill(fill)
|
||
}
|
||
return createBuffer(that, size)
|
||
}
|
||
|
||
/**
|
||
* Creates a new filled Buffer instance.
|
||
* alloc(size[, fill[, encoding]])
|
||
**/
|
||
Buffer.alloc = function (size, fill, encoding) {
|
||
return alloc(null, size, fill, encoding)
|
||
};
|
||
|
||
function allocUnsafe (that, size) {
|
||
assertSize(size);
|
||
that = createBuffer(that, size < 0 ? 0 : checked(size) | 0);
|
||
if (!Buffer.TYPED_ARRAY_SUPPORT) {
|
||
for (var i = 0; i < size; ++i) {
|
||
that[i] = 0;
|
||
}
|
||
}
|
||
return that
|
||
}
|
||
|
||
/**
|
||
* Equivalent to Buffer(num), by default creates a non-zero-filled Buffer instance.
|
||
* */
|
||
Buffer.allocUnsafe = function (size) {
|
||
return allocUnsafe(null, size)
|
||
};
|
||
/**
|
||
* Equivalent to SlowBuffer(num), by default creates a non-zero-filled Buffer instance.
|
||
*/
|
||
Buffer.allocUnsafeSlow = function (size) {
|
||
return allocUnsafe(null, size)
|
||
};
|
||
|
||
function fromString (that, string, encoding) {
|
||
if (typeof encoding !== 'string' || encoding === '') {
|
||
encoding = 'utf8';
|
||
}
|
||
|
||
if (!Buffer.isEncoding(encoding)) {
|
||
throw new TypeError('"encoding" must be a valid string encoding')
|
||
}
|
||
|
||
var length = byteLength(string, encoding) | 0;
|
||
that = createBuffer(that, length);
|
||
|
||
var actual = that.write(string, encoding);
|
||
|
||
if (actual !== length) {
|
||
// Writing a hex string, for example, that contains invalid characters will
|
||
// cause everything after the first invalid character to be ignored. (e.g.
|
||
// 'abxxcd' will be treated as 'ab')
|
||
that = that.slice(0, actual);
|
||
}
|
||
|
||
return that
|
||
}
|
||
|
||
function fromArrayLike (that, array) {
|
||
var length = array.length < 0 ? 0 : checked(array.length) | 0;
|
||
that = createBuffer(that, length);
|
||
for (var i = 0; i < length; i += 1) {
|
||
that[i] = array[i] & 255;
|
||
}
|
||
return that
|
||
}
|
||
|
||
function fromArrayBuffer (that, array, byteOffset, length) {
|
||
array.byteLength; // this throws if `array` is not a valid ArrayBuffer
|
||
|
||
if (byteOffset < 0 || array.byteLength < byteOffset) {
|
||
throw new RangeError('\'offset\' is out of bounds')
|
||
}
|
||
|
||
if (array.byteLength < byteOffset + (length || 0)) {
|
||
throw new RangeError('\'length\' is out of bounds')
|
||
}
|
||
|
||
if (byteOffset === undefined && length === undefined) {
|
||
array = new Uint8Array(array);
|
||
} else if (length === undefined) {
|
||
array = new Uint8Array(array, byteOffset);
|
||
} else {
|
||
array = new Uint8Array(array, byteOffset, length);
|
||
}
|
||
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
// Return an augmented `Uint8Array` instance, for best performance
|
||
that = array;
|
||
that.__proto__ = Buffer.prototype;
|
||
} else {
|
||
// Fallback: Return an object instance of the Buffer class
|
||
that = fromArrayLike(that, array);
|
||
}
|
||
return that
|
||
}
|
||
|
||
function fromObject (that, obj) {
|
||
if (internalIsBuffer(obj)) {
|
||
var len = checked(obj.length) | 0;
|
||
that = createBuffer(that, len);
|
||
|
||
if (that.length === 0) {
|
||
return that
|
||
}
|
||
|
||
obj.copy(that, 0, 0, len);
|
||
return that
|
||
}
|
||
|
||
if (obj) {
|
||
if ((typeof ArrayBuffer !== 'undefined' &&
|
||
obj.buffer instanceof ArrayBuffer) || 'length' in obj) {
|
||
if (typeof obj.length !== 'number' || isnan(obj.length)) {
|
||
return createBuffer(that, 0)
|
||
}
|
||
return fromArrayLike(that, obj)
|
||
}
|
||
|
||
if (obj.type === 'Buffer' && isArray$1(obj.data)) {
|
||
return fromArrayLike(that, obj.data)
|
||
}
|
||
}
|
||
|
||
throw new TypeError('First argument must be a string, Buffer, ArrayBuffer, Array, or array-like object.')
|
||
}
|
||
|
||
function checked (length) {
|
||
// Note: cannot use `length < kMaxLength()` here because that fails when
|
||
// length is NaN (which is otherwise coerced to zero.)
|
||
if (length >= kMaxLength()) {
|
||
throw new RangeError('Attempt to allocate Buffer larger than maximum ' +
|
||
'size: 0x' + kMaxLength().toString(16) + ' bytes')
|
||
}
|
||
return length | 0
|
||
}
|
||
Buffer.isBuffer = isBuffer;
|
||
function internalIsBuffer (b) {
|
||
return !!(b != null && b._isBuffer)
|
||
}
|
||
|
||
Buffer.compare = function compare (a, b) {
|
||
if (!internalIsBuffer(a) || !internalIsBuffer(b)) {
|
||
throw new TypeError('Arguments must be Buffers')
|
||
}
|
||
|
||
if (a === b) return 0
|
||
|
||
var x = a.length;
|
||
var y = b.length;
|
||
|
||
for (var i = 0, len = Math.min(x, y); i < len; ++i) {
|
||
if (a[i] !== b[i]) {
|
||
x = a[i];
|
||
y = b[i];
|
||
break
|
||
}
|
||
}
|
||
|
||
if (x < y) return -1
|
||
if (y < x) return 1
|
||
return 0
|
||
};
|
||
|
||
Buffer.isEncoding = function isEncoding (encoding) {
|
||
switch (String(encoding).toLowerCase()) {
|
||
case 'hex':
|
||
case 'utf8':
|
||
case 'utf-8':
|
||
case 'ascii':
|
||
case 'latin1':
|
||
case 'binary':
|
||
case 'base64':
|
||
case 'ucs2':
|
||
case 'ucs-2':
|
||
case 'utf16le':
|
||
case 'utf-16le':
|
||
return true
|
||
default:
|
||
return false
|
||
}
|
||
};
|
||
|
||
Buffer.concat = function concat (list, length) {
|
||
if (!isArray$1(list)) {
|
||
throw new TypeError('"list" argument must be an Array of Buffers')
|
||
}
|
||
|
||
if (list.length === 0) {
|
||
return Buffer.alloc(0)
|
||
}
|
||
|
||
var i;
|
||
if (length === undefined) {
|
||
length = 0;
|
||
for (i = 0; i < list.length; ++i) {
|
||
length += list[i].length;
|
||
}
|
||
}
|
||
|
||
var buffer = Buffer.allocUnsafe(length);
|
||
var pos = 0;
|
||
for (i = 0; i < list.length; ++i) {
|
||
var buf = list[i];
|
||
if (!internalIsBuffer(buf)) {
|
||
throw new TypeError('"list" argument must be an Array of Buffers')
|
||
}
|
||
buf.copy(buffer, pos);
|
||
pos += buf.length;
|
||
}
|
||
return buffer
|
||
};
|
||
|
||
function byteLength (string, encoding) {
|
||
if (internalIsBuffer(string)) {
|
||
return string.length
|
||
}
|
||
if (typeof ArrayBuffer !== 'undefined' && typeof ArrayBuffer.isView === 'function' &&
|
||
(ArrayBuffer.isView(string) || string instanceof ArrayBuffer)) {
|
||
return string.byteLength
|
||
}
|
||
if (typeof string !== 'string') {
|
||
string = '' + string;
|
||
}
|
||
|
||
var len = string.length;
|
||
if (len === 0) return 0
|
||
|
||
// Use a for loop to avoid recursion
|
||
var loweredCase = false;
|
||
for (;;) {
|
||
switch (encoding) {
|
||
case 'ascii':
|
||
case 'latin1':
|
||
case 'binary':
|
||
return len
|
||
case 'utf8':
|
||
case 'utf-8':
|
||
case undefined:
|
||
return utf8ToBytes(string).length
|
||
case 'ucs2':
|
||
case 'ucs-2':
|
||
case 'utf16le':
|
||
case 'utf-16le':
|
||
return len * 2
|
||
case 'hex':
|
||
return len >>> 1
|
||
case 'base64':
|
||
return base64ToBytes(string).length
|
||
default:
|
||
if (loweredCase) return utf8ToBytes(string).length // assume utf8
|
||
encoding = ('' + encoding).toLowerCase();
|
||
loweredCase = true;
|
||
}
|
||
}
|
||
}
|
||
Buffer.byteLength = byteLength;
|
||
|
||
function slowToString (encoding, start, end) {
|
||
var loweredCase = false;
|
||
|
||
// No need to verify that "this.length <= MAX_UINT32" since it's a read-only
|
||
// property of a typed array.
|
||
|
||
// This behaves neither like String nor Uint8Array in that we set start/end
|
||
// to their upper/lower bounds if the value passed is out of range.
|
||
// undefined is handled specially as per ECMA-262 6th Edition,
|
||
// Section 13.3.3.7 Runtime Semantics: KeyedBindingInitialization.
|
||
if (start === undefined || start < 0) {
|
||
start = 0;
|
||
}
|
||
// Return early if start > this.length. Done here to prevent potential uint32
|
||
// coercion fail below.
|
||
if (start > this.length) {
|
||
return ''
|
||
}
|
||
|
||
if (end === undefined || end > this.length) {
|
||
end = this.length;
|
||
}
|
||
|
||
if (end <= 0) {
|
||
return ''
|
||
}
|
||
|
||
// Force coersion to uint32. This will also coerce falsey/NaN values to 0.
|
||
end >>>= 0;
|
||
start >>>= 0;
|
||
|
||
if (end <= start) {
|
||
return ''
|
||
}
|
||
|
||
if (!encoding) encoding = 'utf8';
|
||
|
||
while (true) {
|
||
switch (encoding) {
|
||
case 'hex':
|
||
return hexSlice(this, start, end)
|
||
|
||
case 'utf8':
|
||
case 'utf-8':
|
||
return utf8Slice(this, start, end)
|
||
|
||
case 'ascii':
|
||
return asciiSlice(this, start, end)
|
||
|
||
case 'latin1':
|
||
case 'binary':
|
||
return latin1Slice(this, start, end)
|
||
|
||
case 'base64':
|
||
return base64Slice(this, start, end)
|
||
|
||
case 'ucs2':
|
||
case 'ucs-2':
|
||
case 'utf16le':
|
||
case 'utf-16le':
|
||
return utf16leSlice(this, start, end)
|
||
|
||
default:
|
||
if (loweredCase) throw new TypeError('Unknown encoding: ' + encoding)
|
||
encoding = (encoding + '').toLowerCase();
|
||
loweredCase = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
// The property is used by `Buffer.isBuffer` and `is-buffer` (in Safari 5-7) to detect
|
||
// Buffer instances.
|
||
Buffer.prototype._isBuffer = true;
|
||
|
||
function swap (b, n, m) {
|
||
var i = b[n];
|
||
b[n] = b[m];
|
||
b[m] = i;
|
||
}
|
||
|
||
Buffer.prototype.swap16 = function swap16 () {
|
||
var len = this.length;
|
||
if (len % 2 !== 0) {
|
||
throw new RangeError('Buffer size must be a multiple of 16-bits')
|
||
}
|
||
for (var i = 0; i < len; i += 2) {
|
||
swap(this, i, i + 1);
|
||
}
|
||
return this
|
||
};
|
||
|
||
Buffer.prototype.swap32 = function swap32 () {
|
||
var len = this.length;
|
||
if (len % 4 !== 0) {
|
||
throw new RangeError('Buffer size must be a multiple of 32-bits')
|
||
}
|
||
for (var i = 0; i < len; i += 4) {
|
||
swap(this, i, i + 3);
|
||
swap(this, i + 1, i + 2);
|
||
}
|
||
return this
|
||
};
|
||
|
||
Buffer.prototype.swap64 = function swap64 () {
|
||
var len = this.length;
|
||
if (len % 8 !== 0) {
|
||
throw new RangeError('Buffer size must be a multiple of 64-bits')
|
||
}
|
||
for (var i = 0; i < len; i += 8) {
|
||
swap(this, i, i + 7);
|
||
swap(this, i + 1, i + 6);
|
||
swap(this, i + 2, i + 5);
|
||
swap(this, i + 3, i + 4);
|
||
}
|
||
return this
|
||
};
|
||
|
||
Buffer.prototype.toString = function toString () {
|
||
var length = this.length | 0;
|
||
if (length === 0) return ''
|
||
if (arguments.length === 0) return utf8Slice(this, 0, length)
|
||
return slowToString.apply(this, arguments)
|
||
};
|
||
|
||
Buffer.prototype.equals = function equals (b) {
|
||
if (!internalIsBuffer(b)) throw new TypeError('Argument must be a Buffer')
|
||
if (this === b) return true
|
||
return Buffer.compare(this, b) === 0
|
||
};
|
||
|
||
Buffer.prototype.inspect = function inspect () {
|
||
var str = '';
|
||
var max = INSPECT_MAX_BYTES;
|
||
if (this.length > 0) {
|
||
str = this.toString('hex', 0, max).match(/.{2}/g).join(' ');
|
||
if (this.length > max) str += ' ... ';
|
||
}
|
||
return '<Buffer ' + str + '>'
|
||
};
|
||
|
||
Buffer.prototype.compare = function compare (target, start, end, thisStart, thisEnd) {
|
||
if (!internalIsBuffer(target)) {
|
||
throw new TypeError('Argument must be a Buffer')
|
||
}
|
||
|
||
if (start === undefined) {
|
||
start = 0;
|
||
}
|
||
if (end === undefined) {
|
||
end = target ? target.length : 0;
|
||
}
|
||
if (thisStart === undefined) {
|
||
thisStart = 0;
|
||
}
|
||
if (thisEnd === undefined) {
|
||
thisEnd = this.length;
|
||
}
|
||
|
||
if (start < 0 || end > target.length || thisStart < 0 || thisEnd > this.length) {
|
||
throw new RangeError('out of range index')
|
||
}
|
||
|
||
if (thisStart >= thisEnd && start >= end) {
|
||
return 0
|
||
}
|
||
if (thisStart >= thisEnd) {
|
||
return -1
|
||
}
|
||
if (start >= end) {
|
||
return 1
|
||
}
|
||
|
||
start >>>= 0;
|
||
end >>>= 0;
|
||
thisStart >>>= 0;
|
||
thisEnd >>>= 0;
|
||
|
||
if (this === target) return 0
|
||
|
||
var x = thisEnd - thisStart;
|
||
var y = end - start;
|
||
var len = Math.min(x, y);
|
||
|
||
var thisCopy = this.slice(thisStart, thisEnd);
|
||
var targetCopy = target.slice(start, end);
|
||
|
||
for (var i = 0; i < len; ++i) {
|
||
if (thisCopy[i] !== targetCopy[i]) {
|
||
x = thisCopy[i];
|
||
y = targetCopy[i];
|
||
break
|
||
}
|
||
}
|
||
|
||
if (x < y) return -1
|
||
if (y < x) return 1
|
||
return 0
|
||
};
|
||
|
||
// Finds either the first index of `val` in `buffer` at offset >= `byteOffset`,
|
||
// OR the last index of `val` in `buffer` at offset <= `byteOffset`.
|
||
//
|
||
// Arguments:
|
||
// - buffer - a Buffer to search
|
||
// - val - a string, Buffer, or number
|
||
// - byteOffset - an index into `buffer`; will be clamped to an int32
|
||
// - encoding - an optional encoding, relevant is val is a string
|
||
// - dir - true for indexOf, false for lastIndexOf
|
||
function bidirectionalIndexOf (buffer, val, byteOffset, encoding, dir) {
|
||
// Empty buffer means no match
|
||
if (buffer.length === 0) return -1
|
||
|
||
// Normalize byteOffset
|
||
if (typeof byteOffset === 'string') {
|
||
encoding = byteOffset;
|
||
byteOffset = 0;
|
||
} else if (byteOffset > 0x7fffffff) {
|
||
byteOffset = 0x7fffffff;
|
||
} else if (byteOffset < -0x80000000) {
|
||
byteOffset = -0x80000000;
|
||
}
|
||
byteOffset = +byteOffset; // Coerce to Number.
|
||
if (isNaN(byteOffset)) {
|
||
// byteOffset: it it's undefined, null, NaN, "foo", etc, search whole buffer
|
||
byteOffset = dir ? 0 : (buffer.length - 1);
|
||
}
|
||
|
||
// Normalize byteOffset: negative offsets start from the end of the buffer
|
||
if (byteOffset < 0) byteOffset = buffer.length + byteOffset;
|
||
if (byteOffset >= buffer.length) {
|
||
if (dir) return -1
|
||
else byteOffset = buffer.length - 1;
|
||
} else if (byteOffset < 0) {
|
||
if (dir) byteOffset = 0;
|
||
else return -1
|
||
}
|
||
|
||
// Normalize val
|
||
if (typeof val === 'string') {
|
||
val = Buffer.from(val, encoding);
|
||
}
|
||
|
||
// Finally, search either indexOf (if dir is true) or lastIndexOf
|
||
if (internalIsBuffer(val)) {
|
||
// Special case: looking for empty string/buffer always fails
|
||
if (val.length === 0) {
|
||
return -1
|
||
}
|
||
return arrayIndexOf(buffer, val, byteOffset, encoding, dir)
|
||
} else if (typeof val === 'number') {
|
||
val = val & 0xFF; // Search for a byte value [0-255]
|
||
if (Buffer.TYPED_ARRAY_SUPPORT &&
|
||
typeof Uint8Array.prototype.indexOf === 'function') {
|
||
if (dir) {
|
||
return Uint8Array.prototype.indexOf.call(buffer, val, byteOffset)
|
||
} else {
|
||
return Uint8Array.prototype.lastIndexOf.call(buffer, val, byteOffset)
|
||
}
|
||
}
|
||
return arrayIndexOf(buffer, [ val ], byteOffset, encoding, dir)
|
||
}
|
||
|
||
throw new TypeError('val must be string, number or Buffer')
|
||
}
|
||
|
||
function arrayIndexOf (arr, val, byteOffset, encoding, dir) {
|
||
var indexSize = 1;
|
||
var arrLength = arr.length;
|
||
var valLength = val.length;
|
||
|
||
if (encoding !== undefined) {
|
||
encoding = String(encoding).toLowerCase();
|
||
if (encoding === 'ucs2' || encoding === 'ucs-2' ||
|
||
encoding === 'utf16le' || encoding === 'utf-16le') {
|
||
if (arr.length < 2 || val.length < 2) {
|
||
return -1
|
||
}
|
||
indexSize = 2;
|
||
arrLength /= 2;
|
||
valLength /= 2;
|
||
byteOffset /= 2;
|
||
}
|
||
}
|
||
|
||
function read (buf, i) {
|
||
if (indexSize === 1) {
|
||
return buf[i]
|
||
} else {
|
||
return buf.readUInt16BE(i * indexSize)
|
||
}
|
||
}
|
||
|
||
var i;
|
||
if (dir) {
|
||
var foundIndex = -1;
|
||
for (i = byteOffset; i < arrLength; i++) {
|
||
if (read(arr, i) === read(val, foundIndex === -1 ? 0 : i - foundIndex)) {
|
||
if (foundIndex === -1) foundIndex = i;
|
||
if (i - foundIndex + 1 === valLength) return foundIndex * indexSize
|
||
} else {
|
||
if (foundIndex !== -1) i -= i - foundIndex;
|
||
foundIndex = -1;
|
||
}
|
||
}
|
||
} else {
|
||
if (byteOffset + valLength > arrLength) byteOffset = arrLength - valLength;
|
||
for (i = byteOffset; i >= 0; i--) {
|
||
var found = true;
|
||
for (var j = 0; j < valLength; j++) {
|
||
if (read(arr, i + j) !== read(val, j)) {
|
||
found = false;
|
||
break
|
||
}
|
||
}
|
||
if (found) return i
|
||
}
|
||
}
|
||
|
||
return -1
|
||
}
|
||
|
||
Buffer.prototype.includes = function includes (val, byteOffset, encoding) {
|
||
return this.indexOf(val, byteOffset, encoding) !== -1
|
||
};
|
||
|
||
Buffer.prototype.indexOf = function indexOf (val, byteOffset, encoding) {
|
||
return bidirectionalIndexOf(this, val, byteOffset, encoding, true)
|
||
};
|
||
|
||
Buffer.prototype.lastIndexOf = function lastIndexOf (val, byteOffset, encoding) {
|
||
return bidirectionalIndexOf(this, val, byteOffset, encoding, false)
|
||
};
|
||
|
||
function hexWrite (buf, string, offset, length) {
|
||
offset = Number(offset) || 0;
|
||
var remaining = buf.length - offset;
|
||
if (!length) {
|
||
length = remaining;
|
||
} else {
|
||
length = Number(length);
|
||
if (length > remaining) {
|
||
length = remaining;
|
||
}
|
||
}
|
||
|
||
// must be an even number of digits
|
||
var strLen = string.length;
|
||
if (strLen % 2 !== 0) throw new TypeError('Invalid hex string')
|
||
|
||
if (length > strLen / 2) {
|
||
length = strLen / 2;
|
||
}
|
||
for (var i = 0; i < length; ++i) {
|
||
var parsed = parseInt(string.substr(i * 2, 2), 16);
|
||
if (isNaN(parsed)) return i
|
||
buf[offset + i] = parsed;
|
||
}
|
||
return i
|
||
}
|
||
|
||
function utf8Write (buf, string, offset, length) {
|
||
return blitBuffer(utf8ToBytes(string, buf.length - offset), buf, offset, length)
|
||
}
|
||
|
||
function asciiWrite (buf, string, offset, length) {
|
||
return blitBuffer(asciiToBytes(string), buf, offset, length)
|
||
}
|
||
|
||
function latin1Write (buf, string, offset, length) {
|
||
return asciiWrite(buf, string, offset, length)
|
||
}
|
||
|
||
function base64Write (buf, string, offset, length) {
|
||
return blitBuffer(base64ToBytes(string), buf, offset, length)
|
||
}
|
||
|
||
function ucs2Write (buf, string, offset, length) {
|
||
return blitBuffer(utf16leToBytes(string, buf.length - offset), buf, offset, length)
|
||
}
|
||
|
||
Buffer.prototype.write = function write (string, offset, length, encoding) {
|
||
// Buffer#write(string)
|
||
if (offset === undefined) {
|
||
encoding = 'utf8';
|
||
length = this.length;
|
||
offset = 0;
|
||
// Buffer#write(string, encoding)
|
||
} else if (length === undefined && typeof offset === 'string') {
|
||
encoding = offset;
|
||
length = this.length;
|
||
offset = 0;
|
||
// Buffer#write(string, offset[, length][, encoding])
|
||
} else if (isFinite(offset)) {
|
||
offset = offset | 0;
|
||
if (isFinite(length)) {
|
||
length = length | 0;
|
||
if (encoding === undefined) encoding = 'utf8';
|
||
} else {
|
||
encoding = length;
|
||
length = undefined;
|
||
}
|
||
// legacy write(string, encoding, offset, length) - remove in v0.13
|
||
} else {
|
||
throw new Error(
|
||
'Buffer.write(string, encoding, offset[, length]) is no longer supported'
|
||
)
|
||
}
|
||
|
||
var remaining = this.length - offset;
|
||
if (length === undefined || length > remaining) length = remaining;
|
||
|
||
if ((string.length > 0 && (length < 0 || offset < 0)) || offset > this.length) {
|
||
throw new RangeError('Attempt to write outside buffer bounds')
|
||
}
|
||
|
||
if (!encoding) encoding = 'utf8';
|
||
|
||
var loweredCase = false;
|
||
for (;;) {
|
||
switch (encoding) {
|
||
case 'hex':
|
||
return hexWrite(this, string, offset, length)
|
||
|
||
case 'utf8':
|
||
case 'utf-8':
|
||
return utf8Write(this, string, offset, length)
|
||
|
||
case 'ascii':
|
||
return asciiWrite(this, string, offset, length)
|
||
|
||
case 'latin1':
|
||
case 'binary':
|
||
return latin1Write(this, string, offset, length)
|
||
|
||
case 'base64':
|
||
// Warning: maxLength not taken into account in base64Write
|
||
return base64Write(this, string, offset, length)
|
||
|
||
case 'ucs2':
|
||
case 'ucs-2':
|
||
case 'utf16le':
|
||
case 'utf-16le':
|
||
return ucs2Write(this, string, offset, length)
|
||
|
||
default:
|
||
if (loweredCase) throw new TypeError('Unknown encoding: ' + encoding)
|
||
encoding = ('' + encoding).toLowerCase();
|
||
loweredCase = true;
|
||
}
|
||
}
|
||
};
|
||
|
||
Buffer.prototype.toJSON = function toJSON () {
|
||
return {
|
||
type: 'Buffer',
|
||
data: Array.prototype.slice.call(this._arr || this, 0)
|
||
}
|
||
};
|
||
|
||
function base64Slice (buf, start, end) {
|
||
if (start === 0 && end === buf.length) {
|
||
return fromByteArray(buf)
|
||
} else {
|
||
return fromByteArray(buf.slice(start, end))
|
||
}
|
||
}
|
||
|
||
function utf8Slice (buf, start, end) {
|
||
end = Math.min(buf.length, end);
|
||
var res = [];
|
||
|
||
var i = start;
|
||
while (i < end) {
|
||
var firstByte = buf[i];
|
||
var codePoint = null;
|
||
var bytesPerSequence = (firstByte > 0xEF) ? 4
|
||
: (firstByte > 0xDF) ? 3
|
||
: (firstByte > 0xBF) ? 2
|
||
: 1;
|
||
|
||
if (i + bytesPerSequence <= end) {
|
||
var secondByte, thirdByte, fourthByte, tempCodePoint;
|
||
|
||
switch (bytesPerSequence) {
|
||
case 1:
|
||
if (firstByte < 0x80) {
|
||
codePoint = firstByte;
|
||
}
|
||
break
|
||
case 2:
|
||
secondByte = buf[i + 1];
|
||
if ((secondByte & 0xC0) === 0x80) {
|
||
tempCodePoint = (firstByte & 0x1F) << 0x6 | (secondByte & 0x3F);
|
||
if (tempCodePoint > 0x7F) {
|
||
codePoint = tempCodePoint;
|
||
}
|
||
}
|
||
break
|
||
case 3:
|
||
secondByte = buf[i + 1];
|
||
thirdByte = buf[i + 2];
|
||
if ((secondByte & 0xC0) === 0x80 && (thirdByte & 0xC0) === 0x80) {
|
||
tempCodePoint = (firstByte & 0xF) << 0xC | (secondByte & 0x3F) << 0x6 | (thirdByte & 0x3F);
|
||
if (tempCodePoint > 0x7FF && (tempCodePoint < 0xD800 || tempCodePoint > 0xDFFF)) {
|
||
codePoint = tempCodePoint;
|
||
}
|
||
}
|
||
break
|
||
case 4:
|
||
secondByte = buf[i + 1];
|
||
thirdByte = buf[i + 2];
|
||
fourthByte = buf[i + 3];
|
||
if ((secondByte & 0xC0) === 0x80 && (thirdByte & 0xC0) === 0x80 && (fourthByte & 0xC0) === 0x80) {
|
||
tempCodePoint = (firstByte & 0xF) << 0x12 | (secondByte & 0x3F) << 0xC | (thirdByte & 0x3F) << 0x6 | (fourthByte & 0x3F);
|
||
if (tempCodePoint > 0xFFFF && tempCodePoint < 0x110000) {
|
||
codePoint = tempCodePoint;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
if (codePoint === null) {
|
||
// we did not generate a valid codePoint so insert a
|
||
// replacement char (U+FFFD) and advance only 1 byte
|
||
codePoint = 0xFFFD;
|
||
bytesPerSequence = 1;
|
||
} else if (codePoint > 0xFFFF) {
|
||
// encode to utf16 (surrogate pair dance)
|
||
codePoint -= 0x10000;
|
||
res.push(codePoint >>> 10 & 0x3FF | 0xD800);
|
||
codePoint = 0xDC00 | codePoint & 0x3FF;
|
||
}
|
||
|
||
res.push(codePoint);
|
||
i += bytesPerSequence;
|
||
}
|
||
|
||
return decodeCodePointsArray(res)
|
||
}
|
||
|
||
// Based on http://stackoverflow.com/a/22747272/680742, the browser with
|
||
// the lowest limit is Chrome, with 0x10000 args.
|
||
// We go 1 magnitude less, for safety
|
||
var MAX_ARGUMENTS_LENGTH = 0x1000;
|
||
|
||
function decodeCodePointsArray (codePoints) {
|
||
var len = codePoints.length;
|
||
if (len <= MAX_ARGUMENTS_LENGTH) {
|
||
return String.fromCharCode.apply(String, codePoints) // avoid extra slice()
|
||
}
|
||
|
||
// Decode in chunks to avoid "call stack size exceeded".
|
||
var res = '';
|
||
var i = 0;
|
||
while (i < len) {
|
||
res += String.fromCharCode.apply(
|
||
String,
|
||
codePoints.slice(i, i += MAX_ARGUMENTS_LENGTH)
|
||
);
|
||
}
|
||
return res
|
||
}
|
||
|
||
function asciiSlice (buf, start, end) {
|
||
var ret = '';
|
||
end = Math.min(buf.length, end);
|
||
|
||
for (var i = start; i < end; ++i) {
|
||
ret += String.fromCharCode(buf[i] & 0x7F);
|
||
}
|
||
return ret
|
||
}
|
||
|
||
function latin1Slice (buf, start, end) {
|
||
var ret = '';
|
||
end = Math.min(buf.length, end);
|
||
|
||
for (var i = start; i < end; ++i) {
|
||
ret += String.fromCharCode(buf[i]);
|
||
}
|
||
return ret
|
||
}
|
||
|
||
function hexSlice (buf, start, end) {
|
||
var len = buf.length;
|
||
|
||
if (!start || start < 0) start = 0;
|
||
if (!end || end < 0 || end > len) end = len;
|
||
|
||
var out = '';
|
||
for (var i = start; i < end; ++i) {
|
||
out += toHex(buf[i]);
|
||
}
|
||
return out
|
||
}
|
||
|
||
function utf16leSlice (buf, start, end) {
|
||
var bytes = buf.slice(start, end);
|
||
var res = '';
|
||
for (var i = 0; i < bytes.length; i += 2) {
|
||
res += String.fromCharCode(bytes[i] + bytes[i + 1] * 256);
|
||
}
|
||
return res
|
||
}
|
||
|
||
Buffer.prototype.slice = function slice (start, end) {
|
||
var len = this.length;
|
||
start = ~~start;
|
||
end = end === undefined ? len : ~~end;
|
||
|
||
if (start < 0) {
|
||
start += len;
|
||
if (start < 0) start = 0;
|
||
} else if (start > len) {
|
||
start = len;
|
||
}
|
||
|
||
if (end < 0) {
|
||
end += len;
|
||
if (end < 0) end = 0;
|
||
} else if (end > len) {
|
||
end = len;
|
||
}
|
||
|
||
if (end < start) end = start;
|
||
|
||
var newBuf;
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
newBuf = this.subarray(start, end);
|
||
newBuf.__proto__ = Buffer.prototype;
|
||
} else {
|
||
var sliceLen = end - start;
|
||
newBuf = new Buffer(sliceLen, undefined);
|
||
for (var i = 0; i < sliceLen; ++i) {
|
||
newBuf[i] = this[i + start];
|
||
}
|
||
}
|
||
|
||
return newBuf
|
||
};
|
||
|
||
/*
|
||
* Need to make sure that buffer isn't trying to write out of bounds.
|
||
*/
|
||
function checkOffset (offset, ext, length) {
|
||
if ((offset % 1) !== 0 || offset < 0) throw new RangeError('offset is not uint')
|
||
if (offset + ext > length) throw new RangeError('Trying to access beyond buffer length')
|
||
}
|
||
|
||
Buffer.prototype.readUIntLE = function readUIntLE (offset, byteLength, noAssert) {
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) checkOffset(offset, byteLength, this.length);
|
||
|
||
var val = this[offset];
|
||
var mul = 1;
|
||
var i = 0;
|
||
while (++i < byteLength && (mul *= 0x100)) {
|
||
val += this[offset + i] * mul;
|
||
}
|
||
|
||
return val
|
||
};
|
||
|
||
Buffer.prototype.readUIntBE = function readUIntBE (offset, byteLength, noAssert) {
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) {
|
||
checkOffset(offset, byteLength, this.length);
|
||
}
|
||
|
||
var val = this[offset + --byteLength];
|
||
var mul = 1;
|
||
while (byteLength > 0 && (mul *= 0x100)) {
|
||
val += this[offset + --byteLength] * mul;
|
||
}
|
||
|
||
return val
|
||
};
|
||
|
||
Buffer.prototype.readUInt8 = function readUInt8 (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 1, this.length);
|
||
return this[offset]
|
||
};
|
||
|
||
Buffer.prototype.readUInt16LE = function readUInt16LE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 2, this.length);
|
||
return this[offset] | (this[offset + 1] << 8)
|
||
};
|
||
|
||
Buffer.prototype.readUInt16BE = function readUInt16BE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 2, this.length);
|
||
return (this[offset] << 8) | this[offset + 1]
|
||
};
|
||
|
||
Buffer.prototype.readUInt32LE = function readUInt32LE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
|
||
return ((this[offset]) |
|
||
(this[offset + 1] << 8) |
|
||
(this[offset + 2] << 16)) +
|
||
(this[offset + 3] * 0x1000000)
|
||
};
|
||
|
||
Buffer.prototype.readUInt32BE = function readUInt32BE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
|
||
return (this[offset] * 0x1000000) +
|
||
((this[offset + 1] << 16) |
|
||
(this[offset + 2] << 8) |
|
||
this[offset + 3])
|
||
};
|
||
|
||
Buffer.prototype.readIntLE = function readIntLE (offset, byteLength, noAssert) {
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) checkOffset(offset, byteLength, this.length);
|
||
|
||
var val = this[offset];
|
||
var mul = 1;
|
||
var i = 0;
|
||
while (++i < byteLength && (mul *= 0x100)) {
|
||
val += this[offset + i] * mul;
|
||
}
|
||
mul *= 0x80;
|
||
|
||
if (val >= mul) val -= Math.pow(2, 8 * byteLength);
|
||
|
||
return val
|
||
};
|
||
|
||
Buffer.prototype.readIntBE = function readIntBE (offset, byteLength, noAssert) {
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) checkOffset(offset, byteLength, this.length);
|
||
|
||
var i = byteLength;
|
||
var mul = 1;
|
||
var val = this[offset + --i];
|
||
while (i > 0 && (mul *= 0x100)) {
|
||
val += this[offset + --i] * mul;
|
||
}
|
||
mul *= 0x80;
|
||
|
||
if (val >= mul) val -= Math.pow(2, 8 * byteLength);
|
||
|
||
return val
|
||
};
|
||
|
||
Buffer.prototype.readInt8 = function readInt8 (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 1, this.length);
|
||
if (!(this[offset] & 0x80)) return (this[offset])
|
||
return ((0xff - this[offset] + 1) * -1)
|
||
};
|
||
|
||
Buffer.prototype.readInt16LE = function readInt16LE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 2, this.length);
|
||
var val = this[offset] | (this[offset + 1] << 8);
|
||
return (val & 0x8000) ? val | 0xFFFF0000 : val
|
||
};
|
||
|
||
Buffer.prototype.readInt16BE = function readInt16BE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 2, this.length);
|
||
var val = this[offset + 1] | (this[offset] << 8);
|
||
return (val & 0x8000) ? val | 0xFFFF0000 : val
|
||
};
|
||
|
||
Buffer.prototype.readInt32LE = function readInt32LE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
|
||
return (this[offset]) |
|
||
(this[offset + 1] << 8) |
|
||
(this[offset + 2] << 16) |
|
||
(this[offset + 3] << 24)
|
||
};
|
||
|
||
Buffer.prototype.readInt32BE = function readInt32BE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
|
||
return (this[offset] << 24) |
|
||
(this[offset + 1] << 16) |
|
||
(this[offset + 2] << 8) |
|
||
(this[offset + 3])
|
||
};
|
||
|
||
Buffer.prototype.readFloatLE = function readFloatLE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
return read(this, offset, true, 23, 4)
|
||
};
|
||
|
||
Buffer.prototype.readFloatBE = function readFloatBE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 4, this.length);
|
||
return read(this, offset, false, 23, 4)
|
||
};
|
||
|
||
Buffer.prototype.readDoubleLE = function readDoubleLE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 8, this.length);
|
||
return read(this, offset, true, 52, 8)
|
||
};
|
||
|
||
Buffer.prototype.readDoubleBE = function readDoubleBE (offset, noAssert) {
|
||
if (!noAssert) checkOffset(offset, 8, this.length);
|
||
return read(this, offset, false, 52, 8)
|
||
};
|
||
|
||
function checkInt (buf, value, offset, ext, max, min) {
|
||
if (!internalIsBuffer(buf)) throw new TypeError('"buffer" argument must be a Buffer instance')
|
||
if (value > max || value < min) throw new RangeError('"value" argument is out of bounds')
|
||
if (offset + ext > buf.length) throw new RangeError('Index out of range')
|
||
}
|
||
|
||
Buffer.prototype.writeUIntLE = function writeUIntLE (value, offset, byteLength, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) {
|
||
var maxBytes = Math.pow(2, 8 * byteLength) - 1;
|
||
checkInt(this, value, offset, byteLength, maxBytes, 0);
|
||
}
|
||
|
||
var mul = 1;
|
||
var i = 0;
|
||
this[offset] = value & 0xFF;
|
||
while (++i < byteLength && (mul *= 0x100)) {
|
||
this[offset + i] = (value / mul) & 0xFF;
|
||
}
|
||
|
||
return offset + byteLength
|
||
};
|
||
|
||
Buffer.prototype.writeUIntBE = function writeUIntBE (value, offset, byteLength, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
byteLength = byteLength | 0;
|
||
if (!noAssert) {
|
||
var maxBytes = Math.pow(2, 8 * byteLength) - 1;
|
||
checkInt(this, value, offset, byteLength, maxBytes, 0);
|
||
}
|
||
|
||
var i = byteLength - 1;
|
||
var mul = 1;
|
||
this[offset + i] = value & 0xFF;
|
||
while (--i >= 0 && (mul *= 0x100)) {
|
||
this[offset + i] = (value / mul) & 0xFF;
|
||
}
|
||
|
||
return offset + byteLength
|
||
};
|
||
|
||
Buffer.prototype.writeUInt8 = function writeUInt8 (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 1, 0xff, 0);
|
||
if (!Buffer.TYPED_ARRAY_SUPPORT) value = Math.floor(value);
|
||
this[offset] = (value & 0xff);
|
||
return offset + 1
|
||
};
|
||
|
||
function objectWriteUInt16 (buf, value, offset, littleEndian) {
|
||
if (value < 0) value = 0xffff + value + 1;
|
||
for (var i = 0, j = Math.min(buf.length - offset, 2); i < j; ++i) {
|
||
buf[offset + i] = (value & (0xff << (8 * (littleEndian ? i : 1 - i)))) >>>
|
||
(littleEndian ? i : 1 - i) * 8;
|
||
}
|
||
}
|
||
|
||
Buffer.prototype.writeUInt16LE = function writeUInt16LE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 2, 0xffff, 0);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value & 0xff);
|
||
this[offset + 1] = (value >>> 8);
|
||
} else {
|
||
objectWriteUInt16(this, value, offset, true);
|
||
}
|
||
return offset + 2
|
||
};
|
||
|
||
Buffer.prototype.writeUInt16BE = function writeUInt16BE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 2, 0xffff, 0);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value >>> 8);
|
||
this[offset + 1] = (value & 0xff);
|
||
} else {
|
||
objectWriteUInt16(this, value, offset, false);
|
||
}
|
||
return offset + 2
|
||
};
|
||
|
||
function objectWriteUInt32 (buf, value, offset, littleEndian) {
|
||
if (value < 0) value = 0xffffffff + value + 1;
|
||
for (var i = 0, j = Math.min(buf.length - offset, 4); i < j; ++i) {
|
||
buf[offset + i] = (value >>> (littleEndian ? i : 3 - i) * 8) & 0xff;
|
||
}
|
||
}
|
||
|
||
Buffer.prototype.writeUInt32LE = function writeUInt32LE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 4, 0xffffffff, 0);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset + 3] = (value >>> 24);
|
||
this[offset + 2] = (value >>> 16);
|
||
this[offset + 1] = (value >>> 8);
|
||
this[offset] = (value & 0xff);
|
||
} else {
|
||
objectWriteUInt32(this, value, offset, true);
|
||
}
|
||
return offset + 4
|
||
};
|
||
|
||
Buffer.prototype.writeUInt32BE = function writeUInt32BE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 4, 0xffffffff, 0);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value >>> 24);
|
||
this[offset + 1] = (value >>> 16);
|
||
this[offset + 2] = (value >>> 8);
|
||
this[offset + 3] = (value & 0xff);
|
||
} else {
|
||
objectWriteUInt32(this, value, offset, false);
|
||
}
|
||
return offset + 4
|
||
};
|
||
|
||
Buffer.prototype.writeIntLE = function writeIntLE (value, offset, byteLength, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) {
|
||
var limit = Math.pow(2, 8 * byteLength - 1);
|
||
|
||
checkInt(this, value, offset, byteLength, limit - 1, -limit);
|
||
}
|
||
|
||
var i = 0;
|
||
var mul = 1;
|
||
var sub = 0;
|
||
this[offset] = value & 0xFF;
|
||
while (++i < byteLength && (mul *= 0x100)) {
|
||
if (value < 0 && sub === 0 && this[offset + i - 1] !== 0) {
|
||
sub = 1;
|
||
}
|
||
this[offset + i] = ((value / mul) >> 0) - sub & 0xFF;
|
||
}
|
||
|
||
return offset + byteLength
|
||
};
|
||
|
||
Buffer.prototype.writeIntBE = function writeIntBE (value, offset, byteLength, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) {
|
||
var limit = Math.pow(2, 8 * byteLength - 1);
|
||
|
||
checkInt(this, value, offset, byteLength, limit - 1, -limit);
|
||
}
|
||
|
||
var i = byteLength - 1;
|
||
var mul = 1;
|
||
var sub = 0;
|
||
this[offset + i] = value & 0xFF;
|
||
while (--i >= 0 && (mul *= 0x100)) {
|
||
if (value < 0 && sub === 0 && this[offset + i + 1] !== 0) {
|
||
sub = 1;
|
||
}
|
||
this[offset + i] = ((value / mul) >> 0) - sub & 0xFF;
|
||
}
|
||
|
||
return offset + byteLength
|
||
};
|
||
|
||
Buffer.prototype.writeInt8 = function writeInt8 (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 1, 0x7f, -0x80);
|
||
if (!Buffer.TYPED_ARRAY_SUPPORT) value = Math.floor(value);
|
||
if (value < 0) value = 0xff + value + 1;
|
||
this[offset] = (value & 0xff);
|
||
return offset + 1
|
||
};
|
||
|
||
Buffer.prototype.writeInt16LE = function writeInt16LE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 2, 0x7fff, -0x8000);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value & 0xff);
|
||
this[offset + 1] = (value >>> 8);
|
||
} else {
|
||
objectWriteUInt16(this, value, offset, true);
|
||
}
|
||
return offset + 2
|
||
};
|
||
|
||
Buffer.prototype.writeInt16BE = function writeInt16BE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 2, 0x7fff, -0x8000);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value >>> 8);
|
||
this[offset + 1] = (value & 0xff);
|
||
} else {
|
||
objectWriteUInt16(this, value, offset, false);
|
||
}
|
||
return offset + 2
|
||
};
|
||
|
||
Buffer.prototype.writeInt32LE = function writeInt32LE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 4, 0x7fffffff, -0x80000000);
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value & 0xff);
|
||
this[offset + 1] = (value >>> 8);
|
||
this[offset + 2] = (value >>> 16);
|
||
this[offset + 3] = (value >>> 24);
|
||
} else {
|
||
objectWriteUInt32(this, value, offset, true);
|
||
}
|
||
return offset + 4
|
||
};
|
||
|
||
Buffer.prototype.writeInt32BE = function writeInt32BE (value, offset, noAssert) {
|
||
value = +value;
|
||
offset = offset | 0;
|
||
if (!noAssert) checkInt(this, value, offset, 4, 0x7fffffff, -0x80000000);
|
||
if (value < 0) value = 0xffffffff + value + 1;
|
||
if (Buffer.TYPED_ARRAY_SUPPORT) {
|
||
this[offset] = (value >>> 24);
|
||
this[offset + 1] = (value >>> 16);
|
||
this[offset + 2] = (value >>> 8);
|
||
this[offset + 3] = (value & 0xff);
|
||
} else {
|
||
objectWriteUInt32(this, value, offset, false);
|
||
}
|
||
return offset + 4
|
||
};
|
||
|
||
function checkIEEE754 (buf, value, offset, ext, max, min) {
|
||
if (offset + ext > buf.length) throw new RangeError('Index out of range')
|
||
if (offset < 0) throw new RangeError('Index out of range')
|
||
}
|
||
|
||
function writeFloat (buf, value, offset, littleEndian, noAssert) {
|
||
if (!noAssert) {
|
||
checkIEEE754(buf, value, offset, 4);
|
||
}
|
||
write(buf, value, offset, littleEndian, 23, 4);
|
||
return offset + 4
|
||
}
|
||
|
||
Buffer.prototype.writeFloatLE = function writeFloatLE (value, offset, noAssert) {
|
||
return writeFloat(this, value, offset, true, noAssert)
|
||
};
|
||
|
||
Buffer.prototype.writeFloatBE = function writeFloatBE (value, offset, noAssert) {
|
||
return writeFloat(this, value, offset, false, noAssert)
|
||
};
|
||
|
||
function writeDouble (buf, value, offset, littleEndian, noAssert) {
|
||
if (!noAssert) {
|
||
checkIEEE754(buf, value, offset, 8);
|
||
}
|
||
write(buf, value, offset, littleEndian, 52, 8);
|
||
return offset + 8
|
||
}
|
||
|
||
Buffer.prototype.writeDoubleLE = function writeDoubleLE (value, offset, noAssert) {
|
||
return writeDouble(this, value, offset, true, noAssert)
|
||
};
|
||
|
||
Buffer.prototype.writeDoubleBE = function writeDoubleBE (value, offset, noAssert) {
|
||
return writeDouble(this, value, offset, false, noAssert)
|
||
};
|
||
|
||
// copy(targetBuffer, targetStart=0, sourceStart=0, sourceEnd=buffer.length)
|
||
Buffer.prototype.copy = function copy (target, targetStart, start, end) {
|
||
if (!start) start = 0;
|
||
if (!end && end !== 0) end = this.length;
|
||
if (targetStart >= target.length) targetStart = target.length;
|
||
if (!targetStart) targetStart = 0;
|
||
if (end > 0 && end < start) end = start;
|
||
|
||
// Copy 0 bytes; we're done
|
||
if (end === start) return 0
|
||
if (target.length === 0 || this.length === 0) return 0
|
||
|
||
// Fatal error conditions
|
||
if (targetStart < 0) {
|
||
throw new RangeError('targetStart out of bounds')
|
||
}
|
||
if (start < 0 || start >= this.length) throw new RangeError('sourceStart out of bounds')
|
||
if (end < 0) throw new RangeError('sourceEnd out of bounds')
|
||
|
||
// Are we oob?
|
||
if (end > this.length) end = this.length;
|
||
if (target.length - targetStart < end - start) {
|
||
end = target.length - targetStart + start;
|
||
}
|
||
|
||
var len = end - start;
|
||
var i;
|
||
|
||
if (this === target && start < targetStart && targetStart < end) {
|
||
// descending copy from end
|
||
for (i = len - 1; i >= 0; --i) {
|
||
target[i + targetStart] = this[i + start];
|
||
}
|
||
} else if (len < 1000 || !Buffer.TYPED_ARRAY_SUPPORT) {
|
||
// ascending copy from start
|
||
for (i = 0; i < len; ++i) {
|
||
target[i + targetStart] = this[i + start];
|
||
}
|
||
} else {
|
||
Uint8Array.prototype.set.call(
|
||
target,
|
||
this.subarray(start, start + len),
|
||
targetStart
|
||
);
|
||
}
|
||
|
||
return len
|
||
};
|
||
|
||
// Usage:
|
||
// buffer.fill(number[, offset[, end]])
|
||
// buffer.fill(buffer[, offset[, end]])
|
||
// buffer.fill(string[, offset[, end]][, encoding])
|
||
Buffer.prototype.fill = function fill (val, start, end, encoding) {
|
||
// Handle string cases:
|
||
if (typeof val === 'string') {
|
||
if (typeof start === 'string') {
|
||
encoding = start;
|
||
start = 0;
|
||
end = this.length;
|
||
} else if (typeof end === 'string') {
|
||
encoding = end;
|
||
end = this.length;
|
||
}
|
||
if (val.length === 1) {
|
||
var code = val.charCodeAt(0);
|
||
if (code < 256) {
|
||
val = code;
|
||
}
|
||
}
|
||
if (encoding !== undefined && typeof encoding !== 'string') {
|
||
throw new TypeError('encoding must be a string')
|
||
}
|
||
if (typeof encoding === 'string' && !Buffer.isEncoding(encoding)) {
|
||
throw new TypeError('Unknown encoding: ' + encoding)
|
||
}
|
||
} else if (typeof val === 'number') {
|
||
val = val & 255;
|
||
}
|
||
|
||
// Invalid ranges are not set to a default, so can range check early.
|
||
if (start < 0 || this.length < start || this.length < end) {
|
||
throw new RangeError('Out of range index')
|
||
}
|
||
|
||
if (end <= start) {
|
||
return this
|
||
}
|
||
|
||
start = start >>> 0;
|
||
end = end === undefined ? this.length : end >>> 0;
|
||
|
||
if (!val) val = 0;
|
||
|
||
var i;
|
||
if (typeof val === 'number') {
|
||
for (i = start; i < end; ++i) {
|
||
this[i] = val;
|
||
}
|
||
} else {
|
||
var bytes = internalIsBuffer(val)
|
||
? val
|
||
: utf8ToBytes(new Buffer(val, encoding).toString());
|
||
var len = bytes.length;
|
||
for (i = 0; i < end - start; ++i) {
|
||
this[i + start] = bytes[i % len];
|
||
}
|
||
}
|
||
|
||
return this
|
||
};
|
||
|
||
// HELPER FUNCTIONS
|
||
// ================
|
||
|
||
var INVALID_BASE64_RE = /[^+\/0-9A-Za-z-_]/g;
|
||
|
||
function base64clean (str) {
|
||
// Node strips out invalid characters like \n and \t from the string, base64-js does not
|
||
str = stringtrim(str).replace(INVALID_BASE64_RE, '');
|
||
// Node converts strings with length < 2 to ''
|
||
if (str.length < 2) return ''
|
||
// Node allows for non-padded base64 strings (missing trailing ===), base64-js does not
|
||
while (str.length % 4 !== 0) {
|
||
str = str + '=';
|
||
}
|
||
return str
|
||
}
|
||
|
||
function stringtrim (str) {
|
||
if (str.trim) return str.trim()
|
||
return str.replace(/^\s+|\s+$/g, '')
|
||
}
|
||
|
||
function toHex (n) {
|
||
if (n < 16) return '0' + n.toString(16)
|
||
return n.toString(16)
|
||
}
|
||
|
||
function utf8ToBytes (string, units) {
|
||
units = units || Infinity;
|
||
var codePoint;
|
||
var length = string.length;
|
||
var leadSurrogate = null;
|
||
var bytes = [];
|
||
|
||
for (var i = 0; i < length; ++i) {
|
||
codePoint = string.charCodeAt(i);
|
||
|
||
// is surrogate component
|
||
if (codePoint > 0xD7FF && codePoint < 0xE000) {
|
||
// last char was a lead
|
||
if (!leadSurrogate) {
|
||
// no lead yet
|
||
if (codePoint > 0xDBFF) {
|
||
// unexpected trail
|
||
if ((units -= 3) > -1) bytes.push(0xEF, 0xBF, 0xBD);
|
||
continue
|
||
} else if (i + 1 === length) {
|
||
// unpaired lead
|
||
if ((units -= 3) > -1) bytes.push(0xEF, 0xBF, 0xBD);
|
||
continue
|
||
}
|
||
|
||
// valid lead
|
||
leadSurrogate = codePoint;
|
||
|
||
continue
|
||
}
|
||
|
||
// 2 leads in a row
|
||
if (codePoint < 0xDC00) {
|
||
if ((units -= 3) > -1) bytes.push(0xEF, 0xBF, 0xBD);
|
||
leadSurrogate = codePoint;
|
||
continue
|
||
}
|
||
|
||
// valid surrogate pair
|
||
codePoint = (leadSurrogate - 0xD800 << 10 | codePoint - 0xDC00) + 0x10000;
|
||
} else if (leadSurrogate) {
|
||
// valid bmp char, but last char was a lead
|
||
if ((units -= 3) > -1) bytes.push(0xEF, 0xBF, 0xBD);
|
||
}
|
||
|
||
leadSurrogate = null;
|
||
|
||
// encode utf8
|
||
if (codePoint < 0x80) {
|
||
if ((units -= 1) < 0) break
|
||
bytes.push(codePoint);
|
||
} else if (codePoint < 0x800) {
|
||
if ((units -= 2) < 0) break
|
||
bytes.push(
|
||
codePoint >> 0x6 | 0xC0,
|
||
codePoint & 0x3F | 0x80
|
||
);
|
||
} else if (codePoint < 0x10000) {
|
||
if ((units -= 3) < 0) break
|
||
bytes.push(
|
||
codePoint >> 0xC | 0xE0,
|
||
codePoint >> 0x6 & 0x3F | 0x80,
|
||
codePoint & 0x3F | 0x80
|
||
);
|
||
} else if (codePoint < 0x110000) {
|
||
if ((units -= 4) < 0) break
|
||
bytes.push(
|
||
codePoint >> 0x12 | 0xF0,
|
||
codePoint >> 0xC & 0x3F | 0x80,
|
||
codePoint >> 0x6 & 0x3F | 0x80,
|
||
codePoint & 0x3F | 0x80
|
||
);
|
||
} else {
|
||
throw new Error('Invalid code point')
|
||
}
|
||
}
|
||
|
||
return bytes
|
||
}
|
||
|
||
function asciiToBytes (str) {
|
||
var byteArray = [];
|
||
for (var i = 0; i < str.length; ++i) {
|
||
// Node's code seems to be doing this and not & 0x7F..
|
||
byteArray.push(str.charCodeAt(i) & 0xFF);
|
||
}
|
||
return byteArray
|
||
}
|
||
|
||
function utf16leToBytes (str, units) {
|
||
var c, hi, lo;
|
||
var byteArray = [];
|
||
for (var i = 0; i < str.length; ++i) {
|
||
if ((units -= 2) < 0) break
|
||
|
||
c = str.charCodeAt(i);
|
||
hi = c >> 8;
|
||
lo = c % 256;
|
||
byteArray.push(lo);
|
||
byteArray.push(hi);
|
||
}
|
||
|
||
return byteArray
|
||
}
|
||
|
||
|
||
function base64ToBytes (str) {
|
||
return toByteArray(base64clean(str))
|
||
}
|
||
|
||
function blitBuffer (src, dst, offset, length) {
|
||
for (var i = 0; i < length; ++i) {
|
||
if ((i + offset >= dst.length) || (i >= src.length)) break
|
||
dst[i + offset] = src[i];
|
||
}
|
||
return i
|
||
}
|
||
|
||
function isnan (val) {
|
||
return val !== val // eslint-disable-line no-self-compare
|
||
}
|
||
|
||
|
||
// the following is from is-buffer, also by Feross Aboukhadijeh and with same lisence
|
||
// The _isBuffer check is for Safari 5-7 support, because it's missing
|
||
// Object.prototype.constructor. Remove this eventually
|
||
function isBuffer(obj) {
|
||
return obj != null && (!!obj._isBuffer || isFastBuffer(obj) || isSlowBuffer(obj))
|
||
}
|
||
|
||
function isFastBuffer (obj) {
|
||
return !!obj.constructor && typeof obj.constructor.isBuffer === 'function' && obj.constructor.isBuffer(obj)
|
||
}
|
||
|
||
// For Node v0.10 support. Remove this eventually.
|
||
function isSlowBuffer (obj) {
|
||
return typeof obj.readFloatLE === 'function' && typeof obj.slice === 'function' && isFastBuffer(obj.slice(0, 0))
|
||
}
|
||
|
||
var sax = {};
|
||
|
||
var domain;
|
||
|
||
// This constructor is used to store event handlers. Instantiating this is
|
||
// faster than explicitly calling `Object.create(null)` to get a "clean" empty
|
||
// object (tested with v8 v4.9).
|
||
function EventHandlers() {}
|
||
EventHandlers.prototype = Object.create(null);
|
||
|
||
function EventEmitter() {
|
||
EventEmitter.init.call(this);
|
||
}
|
||
|
||
// nodejs oddity
|
||
// require('events') === require('events').EventEmitter
|
||
EventEmitter.EventEmitter = EventEmitter;
|
||
|
||
EventEmitter.usingDomains = false;
|
||
|
||
EventEmitter.prototype.domain = undefined;
|
||
EventEmitter.prototype._events = undefined;
|
||
EventEmitter.prototype._maxListeners = undefined;
|
||
|
||
// By default EventEmitters will print a warning if more than 10 listeners are
|
||
// added to it. This is a useful default which helps finding memory leaks.
|
||
EventEmitter.defaultMaxListeners = 10;
|
||
|
||
EventEmitter.init = function() {
|
||
this.domain = null;
|
||
if (EventEmitter.usingDomains) {
|
||
// if there is an active domain, then attach to it.
|
||
if (domain.active) ;
|
||
}
|
||
|
||
if (!this._events || this._events === Object.getPrototypeOf(this)._events) {
|
||
this._events = new EventHandlers();
|
||
this._eventsCount = 0;
|
||
}
|
||
|
||
this._maxListeners = this._maxListeners || undefined;
|
||
};
|
||
|
||
// Obviously not all Emitters should be limited to 10. This function allows
|
||
// that to be increased. Set to zero for unlimited.
|
||
EventEmitter.prototype.setMaxListeners = function setMaxListeners(n) {
|
||
if (typeof n !== 'number' || n < 0 || isNaN(n))
|
||
throw new TypeError('"n" argument must be a positive number');
|
||
this._maxListeners = n;
|
||
return this;
|
||
};
|
||
|
||
function $getMaxListeners(that) {
|
||
if (that._maxListeners === undefined)
|
||
return EventEmitter.defaultMaxListeners;
|
||
return that._maxListeners;
|
||
}
|
||
|
||
EventEmitter.prototype.getMaxListeners = function getMaxListeners() {
|
||
return $getMaxListeners(this);
|
||
};
|
||
|
||
// These standalone emit* functions are used to optimize calling of event
|
||
// handlers for fast cases because emit() itself often has a variable number of
|
||
// arguments and can be deoptimized because of that. These functions always have
|
||
// the same number of arguments and thus do not get deoptimized, so the code
|
||
// inside them can execute faster.
|
||
function emitNone(handler, isFn, self) {
|
||
if (isFn)
|
||
handler.call(self);
|
||
else {
|
||
var len = handler.length;
|
||
var listeners = arrayClone(handler, len);
|
||
for (var i = 0; i < len; ++i)
|
||
listeners[i].call(self);
|
||
}
|
||
}
|
||
function emitOne(handler, isFn, self, arg1) {
|
||
if (isFn)
|
||
handler.call(self, arg1);
|
||
else {
|
||
var len = handler.length;
|
||
var listeners = arrayClone(handler, len);
|
||
for (var i = 0; i < len; ++i)
|
||
listeners[i].call(self, arg1);
|
||
}
|
||
}
|
||
function emitTwo(handler, isFn, self, arg1, arg2) {
|
||
if (isFn)
|
||
handler.call(self, arg1, arg2);
|
||
else {
|
||
var len = handler.length;
|
||
var listeners = arrayClone(handler, len);
|
||
for (var i = 0; i < len; ++i)
|
||
listeners[i].call(self, arg1, arg2);
|
||
}
|
||
}
|
||
function emitThree(handler, isFn, self, arg1, arg2, arg3) {
|
||
if (isFn)
|
||
handler.call(self, arg1, arg2, arg3);
|
||
else {
|
||
var len = handler.length;
|
||
var listeners = arrayClone(handler, len);
|
||
for (var i = 0; i < len; ++i)
|
||
listeners[i].call(self, arg1, arg2, arg3);
|
||
}
|
||
}
|
||
|
||
function emitMany(handler, isFn, self, args) {
|
||
if (isFn)
|
||
handler.apply(self, args);
|
||
else {
|
||
var len = handler.length;
|
||
var listeners = arrayClone(handler, len);
|
||
for (var i = 0; i < len; ++i)
|
||
listeners[i].apply(self, args);
|
||
}
|
||
}
|
||
|
||
EventEmitter.prototype.emit = function emit(type) {
|
||
var er, handler, len, args, i, events, domain;
|
||
var doError = (type === 'error');
|
||
|
||
events = this._events;
|
||
if (events)
|
||
doError = (doError && events.error == null);
|
||
else if (!doError)
|
||
return false;
|
||
|
||
domain = this.domain;
|
||
|
||
// If there is no 'error' event listener then throw.
|
||
if (doError) {
|
||
er = arguments[1];
|
||
if (domain) {
|
||
if (!er)
|
||
er = new Error('Uncaught, unspecified "error" event');
|
||
er.domainEmitter = this;
|
||
er.domain = domain;
|
||
er.domainThrown = false;
|
||
domain.emit('error', er);
|
||
} else if (er instanceof Error) {
|
||
throw er; // Unhandled 'error' event
|
||
} else {
|
||
// At least give some kind of context to the user
|
||
var err = new Error('Uncaught, unspecified "error" event. (' + er + ')');
|
||
err.context = er;
|
||
throw err;
|
||
}
|
||
return false;
|
||
}
|
||
|
||
handler = events[type];
|
||
|
||
if (!handler)
|
||
return false;
|
||
|
||
var isFn = typeof handler === 'function';
|
||
len = arguments.length;
|
||
switch (len) {
|
||
// fast cases
|
||
case 1:
|
||
emitNone(handler, isFn, this);
|
||
break;
|
||
case 2:
|
||
emitOne(handler, isFn, this, arguments[1]);
|
||
break;
|
||
case 3:
|
||
emitTwo(handler, isFn, this, arguments[1], arguments[2]);
|
||
break;
|
||
case 4:
|
||
emitThree(handler, isFn, this, arguments[1], arguments[2], arguments[3]);
|
||
break;
|
||
// slower
|
||
default:
|
||
args = new Array(len - 1);
|
||
for (i = 1; i < len; i++)
|
||
args[i - 1] = arguments[i];
|
||
emitMany(handler, isFn, this, args);
|
||
}
|
||
|
||
return true;
|
||
};
|
||
|
||
function _addListener(target, type, listener, prepend) {
|
||
var m;
|
||
var events;
|
||
var existing;
|
||
|
||
if (typeof listener !== 'function')
|
||
throw new TypeError('"listener" argument must be a function');
|
||
|
||
events = target._events;
|
||
if (!events) {
|
||
events = target._events = new EventHandlers();
|
||
target._eventsCount = 0;
|
||
} else {
|
||
// To avoid recursion in the case that type === "newListener"! Before
|
||
// adding it to the listeners, first emit "newListener".
|
||
if (events.newListener) {
|
||
target.emit('newListener', type,
|
||
listener.listener ? listener.listener : listener);
|
||
|
||
// Re-assign `events` because a newListener handler could have caused the
|
||
// this._events to be assigned to a new object
|
||
events = target._events;
|
||
}
|
||
existing = events[type];
|
||
}
|
||
|
||
if (!existing) {
|
||
// Optimize the case of one listener. Don't need the extra array object.
|
||
existing = events[type] = listener;
|
||
++target._eventsCount;
|
||
} else {
|
||
if (typeof existing === 'function') {
|
||
// Adding the second element, need to change to array.
|
||
existing = events[type] = prepend ? [listener, existing] :
|
||
[existing, listener];
|
||
} else {
|
||
// If we've already got an array, just append.
|
||
if (prepend) {
|
||
existing.unshift(listener);
|
||
} else {
|
||
existing.push(listener);
|
||
}
|
||
}
|
||
|
||
// Check for listener leak
|
||
if (!existing.warned) {
|
||
m = $getMaxListeners(target);
|
||
if (m && m > 0 && existing.length > m) {
|
||
existing.warned = true;
|
||
var w = new Error('Possible EventEmitter memory leak detected. ' +
|
||
existing.length + ' ' + type + ' listeners added. ' +
|
||
'Use emitter.setMaxListeners() to increase limit');
|
||
w.name = 'MaxListenersExceededWarning';
|
||
w.emitter = target;
|
||
w.type = type;
|
||
w.count = existing.length;
|
||
emitWarning(w);
|
||
}
|
||
}
|
||
}
|
||
|
||
return target;
|
||
}
|
||
function emitWarning(e) {
|
||
typeof console.warn === 'function' ? console.warn(e) : console.log(e);
|
||
}
|
||
EventEmitter.prototype.addListener = function addListener(type, listener) {
|
||
return _addListener(this, type, listener, false);
|
||
};
|
||
|
||
EventEmitter.prototype.on = EventEmitter.prototype.addListener;
|
||
|
||
EventEmitter.prototype.prependListener =
|
||
function prependListener(type, listener) {
|
||
return _addListener(this, type, listener, true);
|
||
};
|
||
|
||
function _onceWrap(target, type, listener) {
|
||
var fired = false;
|
||
function g() {
|
||
target.removeListener(type, g);
|
||
if (!fired) {
|
||
fired = true;
|
||
listener.apply(target, arguments);
|
||
}
|
||
}
|
||
g.listener = listener;
|
||
return g;
|
||
}
|
||
|
||
EventEmitter.prototype.once = function once(type, listener) {
|
||
if (typeof listener !== 'function')
|
||
throw new TypeError('"listener" argument must be a function');
|
||
this.on(type, _onceWrap(this, type, listener));
|
||
return this;
|
||
};
|
||
|
||
EventEmitter.prototype.prependOnceListener =
|
||
function prependOnceListener(type, listener) {
|
||
if (typeof listener !== 'function')
|
||
throw new TypeError('"listener" argument must be a function');
|
||
this.prependListener(type, _onceWrap(this, type, listener));
|
||
return this;
|
||
};
|
||
|
||
// emits a 'removeListener' event iff the listener was removed
|
||
EventEmitter.prototype.removeListener =
|
||
function removeListener(type, listener) {
|
||
var list, events, position, i, originalListener;
|
||
|
||
if (typeof listener !== 'function')
|
||
throw new TypeError('"listener" argument must be a function');
|
||
|
||
events = this._events;
|
||
if (!events)
|
||
return this;
|
||
|
||
list = events[type];
|
||
if (!list)
|
||
return this;
|
||
|
||
if (list === listener || (list.listener && list.listener === listener)) {
|
||
if (--this._eventsCount === 0)
|
||
this._events = new EventHandlers();
|
||
else {
|
||
delete events[type];
|
||
if (events.removeListener)
|
||
this.emit('removeListener', type, list.listener || listener);
|
||
}
|
||
} else if (typeof list !== 'function') {
|
||
position = -1;
|
||
|
||
for (i = list.length; i-- > 0;) {
|
||
if (list[i] === listener ||
|
||
(list[i].listener && list[i].listener === listener)) {
|
||
originalListener = list[i].listener;
|
||
position = i;
|
||
break;
|
||
}
|
||
}
|
||
|
||
if (position < 0)
|
||
return this;
|
||
|
||
if (list.length === 1) {
|
||
list[0] = undefined;
|
||
if (--this._eventsCount === 0) {
|
||
this._events = new EventHandlers();
|
||
return this;
|
||
} else {
|
||
delete events[type];
|
||
}
|
||
} else {
|
||
spliceOne(list, position);
|
||
}
|
||
|
||
if (events.removeListener)
|
||
this.emit('removeListener', type, originalListener || listener);
|
||
}
|
||
|
||
return this;
|
||
};
|
||
|
||
// Alias for removeListener added in NodeJS 10.0
|
||
// https://nodejs.org/api/events.html#events_emitter_off_eventname_listener
|
||
EventEmitter.prototype.off = function(type, listener){
|
||
return this.removeListener(type, listener);
|
||
};
|
||
|
||
EventEmitter.prototype.removeAllListeners =
|
||
function removeAllListeners(type) {
|
||
var listeners, events;
|
||
|
||
events = this._events;
|
||
if (!events)
|
||
return this;
|
||
|
||
// not listening for removeListener, no need to emit
|
||
if (!events.removeListener) {
|
||
if (arguments.length === 0) {
|
||
this._events = new EventHandlers();
|
||
this._eventsCount = 0;
|
||
} else if (events[type]) {
|
||
if (--this._eventsCount === 0)
|
||
this._events = new EventHandlers();
|
||
else
|
||
delete events[type];
|
||
}
|
||
return this;
|
||
}
|
||
|
||
// emit removeListener for all listeners on all events
|
||
if (arguments.length === 0) {
|
||
var keys = Object.keys(events);
|
||
for (var i = 0, key; i < keys.length; ++i) {
|
||
key = keys[i];
|
||
if (key === 'removeListener') continue;
|
||
this.removeAllListeners(key);
|
||
}
|
||
this.removeAllListeners('removeListener');
|
||
this._events = new EventHandlers();
|
||
this._eventsCount = 0;
|
||
return this;
|
||
}
|
||
|
||
listeners = events[type];
|
||
|
||
if (typeof listeners === 'function') {
|
||
this.removeListener(type, listeners);
|
||
} else if (listeners) {
|
||
// LIFO order
|
||
do {
|
||
this.removeListener(type, listeners[listeners.length - 1]);
|
||
} while (listeners[0]);
|
||
}
|
||
|
||
return this;
|
||
};
|
||
|
||
EventEmitter.prototype.listeners = function listeners(type) {
|
||
var evlistener;
|
||
var ret;
|
||
var events = this._events;
|
||
|
||
if (!events)
|
||
ret = [];
|
||
else {
|
||
evlistener = events[type];
|
||
if (!evlistener)
|
||
ret = [];
|
||
else if (typeof evlistener === 'function')
|
||
ret = [evlistener.listener || evlistener];
|
||
else
|
||
ret = unwrapListeners(evlistener);
|
||
}
|
||
|
||
return ret;
|
||
};
|
||
|
||
EventEmitter.listenerCount = function(emitter, type) {
|
||
if (typeof emitter.listenerCount === 'function') {
|
||
return emitter.listenerCount(type);
|
||
} else {
|
||
return listenerCount$1.call(emitter, type);
|
||
}
|
||
};
|
||
|
||
EventEmitter.prototype.listenerCount = listenerCount$1;
|
||
function listenerCount$1(type) {
|
||
var events = this._events;
|
||
|
||
if (events) {
|
||
var evlistener = events[type];
|
||
|
||
if (typeof evlistener === 'function') {
|
||
return 1;
|
||
} else if (evlistener) {
|
||
return evlistener.length;
|
||
}
|
||
}
|
||
|
||
return 0;
|
||
}
|
||
|
||
EventEmitter.prototype.eventNames = function eventNames() {
|
||
return this._eventsCount > 0 ? Reflect.ownKeys(this._events) : [];
|
||
};
|
||
|
||
// About 1.5x faster than the two-arg version of Array#splice().
|
||
function spliceOne(list, index) {
|
||
for (var i = index, k = i + 1, n = list.length; k < n; i += 1, k += 1)
|
||
list[i] = list[k];
|
||
list.pop();
|
||
}
|
||
|
||
function arrayClone(arr, i) {
|
||
var copy = new Array(i);
|
||
while (i--)
|
||
copy[i] = arr[i];
|
||
return copy;
|
||
}
|
||
|
||
function unwrapListeners(arr) {
|
||
var ret = new Array(arr.length);
|
||
for (var i = 0; i < ret.length; ++i) {
|
||
ret[i] = arr[i].listener || arr[i];
|
||
}
|
||
return ret;
|
||
}
|
||
|
||
var inherits;
|
||
if (typeof Object.create === 'function'){
|
||
inherits = function inherits(ctor, superCtor) {
|
||
// implementation from standard node.js 'util' module
|
||
ctor.super_ = superCtor;
|
||
ctor.prototype = Object.create(superCtor.prototype, {
|
||
constructor: {
|
||
value: ctor,
|
||
enumerable: false,
|
||
writable: true,
|
||
configurable: true
|
||
}
|
||
});
|
||
};
|
||
} else {
|
||
inherits = function inherits(ctor, superCtor) {
|
||
ctor.super_ = superCtor;
|
||
var TempCtor = function () {};
|
||
TempCtor.prototype = superCtor.prototype;
|
||
ctor.prototype = new TempCtor();
|
||
ctor.prototype.constructor = ctor;
|
||
};
|
||
}
|
||
|
||
var formatRegExp = /%[sdj%]/g;
|
||
function format(f) {
|
||
if (!isString(f)) {
|
||
var objects = [];
|
||
for (var i = 0; i < arguments.length; i++) {
|
||
objects.push(inspect(arguments[i]));
|
||
}
|
||
return objects.join(' ');
|
||
}
|
||
|
||
var i = 1;
|
||
var args = arguments;
|
||
var len = args.length;
|
||
var str = String(f).replace(formatRegExp, function(x) {
|
||
if (x === '%%') return '%';
|
||
if (i >= len) return x;
|
||
switch (x) {
|
||
case '%s': return String(args[i++]);
|
||
case '%d': return Number(args[i++]);
|
||
case '%j':
|
||
try {
|
||
return JSON.stringify(args[i++]);
|
||
} catch (_) {
|
||
return '[Circular]';
|
||
}
|
||
default:
|
||
return x;
|
||
}
|
||
});
|
||
for (var x = args[i]; i < len; x = args[++i]) {
|
||
if (isNull(x) || !isObject(x)) {
|
||
str += ' ' + x;
|
||
} else {
|
||
str += ' ' + inspect(x);
|
||
}
|
||
}
|
||
return str;
|
||
}
|
||
|
||
// Mark that a method should not be used.
|
||
// Returns a modified function which warns once by default.
|
||
// If --no-deprecation is set, then it is a no-op.
|
||
function deprecate(fn, msg) {
|
||
// Allow for deprecating things in the process of starting up.
|
||
if (isUndefined(global$1.process)) {
|
||
return function() {
|
||
return deprecate(fn, msg).apply(this, arguments);
|
||
};
|
||
}
|
||
|
||
if (browser$1$1.noDeprecation === true) {
|
||
return fn;
|
||
}
|
||
|
||
var warned = false;
|
||
function deprecated() {
|
||
if (!warned) {
|
||
if (browser$1$1.throwDeprecation) {
|
||
throw new Error(msg);
|
||
} else if (browser$1$1.traceDeprecation) {
|
||
console.trace(msg);
|
||
} else {
|
||
console.error(msg);
|
||
}
|
||
warned = true;
|
||
}
|
||
return fn.apply(this, arguments);
|
||
}
|
||
|
||
return deprecated;
|
||
}
|
||
|
||
var debugs = {};
|
||
var debugEnviron;
|
||
function debuglog(set) {
|
||
if (isUndefined(debugEnviron))
|
||
debugEnviron = browser$1$1.env.NODE_DEBUG || '';
|
||
set = set.toUpperCase();
|
||
if (!debugs[set]) {
|
||
if (new RegExp('\\b' + set + '\\b', 'i').test(debugEnviron)) {
|
||
var pid = 0;
|
||
debugs[set] = function() {
|
||
var msg = format.apply(null, arguments);
|
||
console.error('%s %d: %s', set, pid, msg);
|
||
};
|
||
} else {
|
||
debugs[set] = function() {};
|
||
}
|
||
}
|
||
return debugs[set];
|
||
}
|
||
|
||
/**
|
||
* Echos the value of a value. Trys to print the value out
|
||
* in the best way possible given the different types.
|
||
*
|
||
* @param {Object} obj The object to print out.
|
||
* @param {Object} opts Optional options object that alters the output.
|
||
*/
|
||
/* legacy: obj, showHidden, depth, colors*/
|
||
function inspect(obj, opts) {
|
||
// default options
|
||
var ctx = {
|
||
seen: [],
|
||
stylize: stylizeNoColor
|
||
};
|
||
// legacy...
|
||
if (arguments.length >= 3) ctx.depth = arguments[2];
|
||
if (arguments.length >= 4) ctx.colors = arguments[3];
|
||
if (isBoolean(opts)) {
|
||
// legacy...
|
||
ctx.showHidden = opts;
|
||
} else if (opts) {
|
||
// got an "options" object
|
||
_extend(ctx, opts);
|
||
}
|
||
// set default options
|
||
if (isUndefined(ctx.showHidden)) ctx.showHidden = false;
|
||
if (isUndefined(ctx.depth)) ctx.depth = 2;
|
||
if (isUndefined(ctx.colors)) ctx.colors = false;
|
||
if (isUndefined(ctx.customInspect)) ctx.customInspect = true;
|
||
if (ctx.colors) ctx.stylize = stylizeWithColor;
|
||
return formatValue(ctx, obj, ctx.depth);
|
||
}
|
||
|
||
// http://en.wikipedia.org/wiki/ANSI_escape_code#graphics
|
||
inspect.colors = {
|
||
'bold' : [1, 22],
|
||
'italic' : [3, 23],
|
||
'underline' : [4, 24],
|
||
'inverse' : [7, 27],
|
||
'white' : [37, 39],
|
||
'grey' : [90, 39],
|
||
'black' : [30, 39],
|
||
'blue' : [34, 39],
|
||
'cyan' : [36, 39],
|
||
'green' : [32, 39],
|
||
'magenta' : [35, 39],
|
||
'red' : [31, 39],
|
||
'yellow' : [33, 39]
|
||
};
|
||
|
||
// Don't use 'blue' not visible on cmd.exe
|
||
inspect.styles = {
|
||
'special': 'cyan',
|
||
'number': 'yellow',
|
||
'boolean': 'yellow',
|
||
'undefined': 'grey',
|
||
'null': 'bold',
|
||
'string': 'green',
|
||
'date': 'magenta',
|
||
// "name": intentionally not styling
|
||
'regexp': 'red'
|
||
};
|
||
|
||
|
||
function stylizeWithColor(str, styleType) {
|
||
var style = inspect.styles[styleType];
|
||
|
||
if (style) {
|
||
return '\u001b[' + inspect.colors[style][0] + 'm' + str +
|
||
'\u001b[' + inspect.colors[style][1] + 'm';
|
||
} else {
|
||
return str;
|
||
}
|
||
}
|
||
|
||
|
||
function stylizeNoColor(str, styleType) {
|
||
return str;
|
||
}
|
||
|
||
|
||
function arrayToHash(array) {
|
||
var hash = {};
|
||
|
||
array.forEach(function(val, idx) {
|
||
hash[val] = true;
|
||
});
|
||
|
||
return hash;
|
||
}
|
||
|
||
|
||
function formatValue(ctx, value, recurseTimes) {
|
||
// Provide a hook for user-specified inspect functions.
|
||
// Check that value is an object with an inspect function on it
|
||
if (ctx.customInspect &&
|
||
value &&
|
||
isFunction(value.inspect) &&
|
||
// Filter out the util module, it's inspect function is special
|
||
value.inspect !== inspect &&
|
||
// Also filter out any prototype objects using the circular check.
|
||
!(value.constructor && value.constructor.prototype === value)) {
|
||
var ret = value.inspect(recurseTimes, ctx);
|
||
if (!isString(ret)) {
|
||
ret = formatValue(ctx, ret, recurseTimes);
|
||
}
|
||
return ret;
|
||
}
|
||
|
||
// Primitive types cannot have properties
|
||
var primitive = formatPrimitive(ctx, value);
|
||
if (primitive) {
|
||
return primitive;
|
||
}
|
||
|
||
// Look up the keys of the object.
|
||
var keys = Object.keys(value);
|
||
var visibleKeys = arrayToHash(keys);
|
||
|
||
if (ctx.showHidden) {
|
||
keys = Object.getOwnPropertyNames(value);
|
||
}
|
||
|
||
// IE doesn't make error fields non-enumerable
|
||
// http://msdn.microsoft.com/en-us/library/ie/dww52sbt(v=vs.94).aspx
|
||
if (isError(value)
|
||
&& (keys.indexOf('message') >= 0 || keys.indexOf('description') >= 0)) {
|
||
return formatError(value);
|
||
}
|
||
|
||
// Some type of object without properties can be shortcutted.
|
||
if (keys.length === 0) {
|
||
if (isFunction(value)) {
|
||
var name = value.name ? ': ' + value.name : '';
|
||
return ctx.stylize('[Function' + name + ']', 'special');
|
||
}
|
||
if (isRegExp(value)) {
|
||
return ctx.stylize(RegExp.prototype.toString.call(value), 'regexp');
|
||
}
|
||
if (isDate(value)) {
|
||
return ctx.stylize(Date.prototype.toString.call(value), 'date');
|
||
}
|
||
if (isError(value)) {
|
||
return formatError(value);
|
||
}
|
||
}
|
||
|
||
var base = '', array = false, braces = ['{', '}'];
|
||
|
||
// Make Array say that they are Array
|
||
if (isArray(value)) {
|
||
array = true;
|
||
braces = ['[', ']'];
|
||
}
|
||
|
||
// Make functions say that they are functions
|
||
if (isFunction(value)) {
|
||
var n = value.name ? ': ' + value.name : '';
|
||
base = ' [Function' + n + ']';
|
||
}
|
||
|
||
// Make RegExps say that they are RegExps
|
||
if (isRegExp(value)) {
|
||
base = ' ' + RegExp.prototype.toString.call(value);
|
||
}
|
||
|
||
// Make dates with properties first say the date
|
||
if (isDate(value)) {
|
||
base = ' ' + Date.prototype.toUTCString.call(value);
|
||
}
|
||
|
||
// Make error with message first say the error
|
||
if (isError(value)) {
|
||
base = ' ' + formatError(value);
|
||
}
|
||
|
||
if (keys.length === 0 && (!array || value.length == 0)) {
|
||
return braces[0] + base + braces[1];
|
||
}
|
||
|
||
if (recurseTimes < 0) {
|
||
if (isRegExp(value)) {
|
||
return ctx.stylize(RegExp.prototype.toString.call(value), 'regexp');
|
||
} else {
|
||
return ctx.stylize('[Object]', 'special');
|
||
}
|
||
}
|
||
|
||
ctx.seen.push(value);
|
||
|
||
var output;
|
||
if (array) {
|
||
output = formatArray(ctx, value, recurseTimes, visibleKeys, keys);
|
||
} else {
|
||
output = keys.map(function(key) {
|
||
return formatProperty(ctx, value, recurseTimes, visibleKeys, key, array);
|
||
});
|
||
}
|
||
|
||
ctx.seen.pop();
|
||
|
||
return reduceToSingleString(output, base, braces);
|
||
}
|
||
|
||
|
||
function formatPrimitive(ctx, value) {
|
||
if (isUndefined(value))
|
||
return ctx.stylize('undefined', 'undefined');
|
||
if (isString(value)) {
|
||
var simple = '\'' + JSON.stringify(value).replace(/^"|"$/g, '')
|
||
.replace(/'/g, "\\'")
|
||
.replace(/\\"/g, '"') + '\'';
|
||
return ctx.stylize(simple, 'string');
|
||
}
|
||
if (isNumber(value))
|
||
return ctx.stylize('' + value, 'number');
|
||
if (isBoolean(value))
|
||
return ctx.stylize('' + value, 'boolean');
|
||
// For some reason typeof null is "object", so special case here.
|
||
if (isNull(value))
|
||
return ctx.stylize('null', 'null');
|
||
}
|
||
|
||
|
||
function formatError(value) {
|
||
return '[' + Error.prototype.toString.call(value) + ']';
|
||
}
|
||
|
||
|
||
function formatArray(ctx, value, recurseTimes, visibleKeys, keys) {
|
||
var output = [];
|
||
for (var i = 0, l = value.length; i < l; ++i) {
|
||
if (hasOwnProperty(value, String(i))) {
|
||
output.push(formatProperty(ctx, value, recurseTimes, visibleKeys,
|
||
String(i), true));
|
||
} else {
|
||
output.push('');
|
||
}
|
||
}
|
||
keys.forEach(function(key) {
|
||
if (!key.match(/^\d+$/)) {
|
||
output.push(formatProperty(ctx, value, recurseTimes, visibleKeys,
|
||
key, true));
|
||
}
|
||
});
|
||
return output;
|
||
}
|
||
|
||
|
||
function formatProperty(ctx, value, recurseTimes, visibleKeys, key, array) {
|
||
var name, str, desc;
|
||
desc = Object.getOwnPropertyDescriptor(value, key) || { value: value[key] };
|
||
if (desc.get) {
|
||
if (desc.set) {
|
||
str = ctx.stylize('[Getter/Setter]', 'special');
|
||
} else {
|
||
str = ctx.stylize('[Getter]', 'special');
|
||
}
|
||
} else {
|
||
if (desc.set) {
|
||
str = ctx.stylize('[Setter]', 'special');
|
||
}
|
||
}
|
||
if (!hasOwnProperty(visibleKeys, key)) {
|
||
name = '[' + key + ']';
|
||
}
|
||
if (!str) {
|
||
if (ctx.seen.indexOf(desc.value) < 0) {
|
||
if (isNull(recurseTimes)) {
|
||
str = formatValue(ctx, desc.value, null);
|
||
} else {
|
||
str = formatValue(ctx, desc.value, recurseTimes - 1);
|
||
}
|
||
if (str.indexOf('\n') > -1) {
|
||
if (array) {
|
||
str = str.split('\n').map(function(line) {
|
||
return ' ' + line;
|
||
}).join('\n').substr(2);
|
||
} else {
|
||
str = '\n' + str.split('\n').map(function(line) {
|
||
return ' ' + line;
|
||
}).join('\n');
|
||
}
|
||
}
|
||
} else {
|
||
str = ctx.stylize('[Circular]', 'special');
|
||
}
|
||
}
|
||
if (isUndefined(name)) {
|
||
if (array && key.match(/^\d+$/)) {
|
||
return str;
|
||
}
|
||
name = JSON.stringify('' + key);
|
||
if (name.match(/^"([a-zA-Z_][a-zA-Z_0-9]*)"$/)) {
|
||
name = name.substr(1, name.length - 2);
|
||
name = ctx.stylize(name, 'name');
|
||
} else {
|
||
name = name.replace(/'/g, "\\'")
|
||
.replace(/\\"/g, '"')
|
||
.replace(/(^"|"$)/g, "'");
|
||
name = ctx.stylize(name, 'string');
|
||
}
|
||
}
|
||
|
||
return name + ': ' + str;
|
||
}
|
||
|
||
|
||
function reduceToSingleString(output, base, braces) {
|
||
var length = output.reduce(function(prev, cur) {
|
||
if (cur.indexOf('\n') >= 0) ;
|
||
return prev + cur.replace(/\u001b\[\d\d?m/g, '').length + 1;
|
||
}, 0);
|
||
|
||
if (length > 60) {
|
||
return braces[0] +
|
||
(base === '' ? '' : base + '\n ') +
|
||
' ' +
|
||
output.join(',\n ') +
|
||
' ' +
|
||
braces[1];
|
||
}
|
||
|
||
return braces[0] + base + ' ' + output.join(', ') + ' ' + braces[1];
|
||
}
|
||
|
||
|
||
// NOTE: These type checking functions intentionally don't use `instanceof`
|
||
// because it is fragile and can be easily faked with `Object.create()`.
|
||
function isArray(ar) {
|
||
return Array.isArray(ar);
|
||
}
|
||
|
||
function isBoolean(arg) {
|
||
return typeof arg === 'boolean';
|
||
}
|
||
|
||
function isNull(arg) {
|
||
return arg === null;
|
||
}
|
||
|
||
function isNumber(arg) {
|
||
return typeof arg === 'number';
|
||
}
|
||
|
||
function isString(arg) {
|
||
return typeof arg === 'string';
|
||
}
|
||
|
||
function isUndefined(arg) {
|
||
return arg === void 0;
|
||
}
|
||
|
||
function isRegExp(re) {
|
||
return isObject(re) && objectToString(re) === '[object RegExp]';
|
||
}
|
||
|
||
function isObject(arg) {
|
||
return typeof arg === 'object' && arg !== null;
|
||
}
|
||
|
||
function isDate(d) {
|
||
return isObject(d) && objectToString(d) === '[object Date]';
|
||
}
|
||
|
||
function isError(e) {
|
||
return isObject(e) &&
|
||
(objectToString(e) === '[object Error]' || e instanceof Error);
|
||
}
|
||
|
||
function isFunction(arg) {
|
||
return typeof arg === 'function';
|
||
}
|
||
|
||
function objectToString(o) {
|
||
return Object.prototype.toString.call(o);
|
||
}
|
||
|
||
function _extend(origin, add) {
|
||
// Don't do anything if add isn't an object
|
||
if (!add || !isObject(add)) return origin;
|
||
|
||
var keys = Object.keys(add);
|
||
var i = keys.length;
|
||
while (i--) {
|
||
origin[keys[i]] = add[keys[i]];
|
||
}
|
||
return origin;
|
||
}
|
||
function hasOwnProperty(obj, prop) {
|
||
return Object.prototype.hasOwnProperty.call(obj, prop);
|
||
}
|
||
|
||
function BufferList() {
|
||
this.head = null;
|
||
this.tail = null;
|
||
this.length = 0;
|
||
}
|
||
|
||
BufferList.prototype.push = function (v) {
|
||
var entry = { data: v, next: null };
|
||
if (this.length > 0) this.tail.next = entry;else this.head = entry;
|
||
this.tail = entry;
|
||
++this.length;
|
||
};
|
||
|
||
BufferList.prototype.unshift = function (v) {
|
||
var entry = { data: v, next: this.head };
|
||
if (this.length === 0) this.tail = entry;
|
||
this.head = entry;
|
||
++this.length;
|
||
};
|
||
|
||
BufferList.prototype.shift = function () {
|
||
if (this.length === 0) return;
|
||
var ret = this.head.data;
|
||
if (this.length === 1) this.head = this.tail = null;else this.head = this.head.next;
|
||
--this.length;
|
||
return ret;
|
||
};
|
||
|
||
BufferList.prototype.clear = function () {
|
||
this.head = this.tail = null;
|
||
this.length = 0;
|
||
};
|
||
|
||
BufferList.prototype.join = function (s) {
|
||
if (this.length === 0) return '';
|
||
var p = this.head;
|
||
var ret = '' + p.data;
|
||
while (p = p.next) {
|
||
ret += s + p.data;
|
||
}return ret;
|
||
};
|
||
|
||
BufferList.prototype.concat = function (n) {
|
||
if (this.length === 0) return Buffer.alloc(0);
|
||
if (this.length === 1) return this.head.data;
|
||
var ret = Buffer.allocUnsafe(n >>> 0);
|
||
var p = this.head;
|
||
var i = 0;
|
||
while (p) {
|
||
p.data.copy(ret, i);
|
||
i += p.data.length;
|
||
p = p.next;
|
||
}
|
||
return ret;
|
||
};
|
||
|
||
// Copyright Joyent, Inc. and other Node contributors.
|
||
//
|
||
// Permission is hereby granted, free of charge, to any person obtaining a
|
||
// copy of this software and associated documentation files (the
|
||
// "Software"), to deal in the Software without restriction, including
|
||
// without limitation the rights to use, copy, modify, merge, publish,
|
||
// distribute, sublicense, and/or sell copies of the Software, and to permit
|
||
// persons to whom the Software is furnished to do so, subject to the
|
||
// following conditions:
|
||
//
|
||
// The above copyright notice and this permission notice shall be included
|
||
// in all copies or substantial portions of the Software.
|
||
//
|
||
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
||
// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN
|
||
// NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
|
||
// DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
|
||
// OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
|
||
// USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
||
var isBufferEncoding = Buffer.isEncoding
|
||
|| function(encoding) {
|
||
switch (encoding && encoding.toLowerCase()) {
|
||
case 'hex': case 'utf8': case 'utf-8': case 'ascii': case 'binary': case 'base64': case 'ucs2': case 'ucs-2': case 'utf16le': case 'utf-16le': case 'raw': return true;
|
||
default: return false;
|
||
}
|
||
};
|
||
|
||
|
||
function assertEncoding(encoding) {
|
||
if (encoding && !isBufferEncoding(encoding)) {
|
||
throw new Error('Unknown encoding: ' + encoding);
|
||
}
|
||
}
|
||
|
||
// StringDecoder provides an interface for efficiently splitting a series of
|
||
// buffers into a series of JS strings without breaking apart multi-byte
|
||
// characters. CESU-8 is handled as part of the UTF-8 encoding.
|
||
//
|
||
// @TODO Handling all encodings inside a single object makes it very difficult
|
||
// to reason about this code, so it should be split up in the future.
|
||
// @TODO There should be a utf8-strict encoding that rejects invalid UTF-8 code
|
||
// points as used by CESU-8.
|
||
function StringDecoder(encoding) {
|
||
this.encoding = (encoding || 'utf8').toLowerCase().replace(/[-_]/, '');
|
||
assertEncoding(encoding);
|
||
switch (this.encoding) {
|
||
case 'utf8':
|
||
// CESU-8 represents each of Surrogate Pair by 3-bytes
|
||
this.surrogateSize = 3;
|
||
break;
|
||
case 'ucs2':
|
||
case 'utf16le':
|
||
// UTF-16 represents each of Surrogate Pair by 2-bytes
|
||
this.surrogateSize = 2;
|
||
this.detectIncompleteChar = utf16DetectIncompleteChar;
|
||
break;
|
||
case 'base64':
|
||
// Base-64 stores 3 bytes in 4 chars, and pads the remainder.
|
||
this.surrogateSize = 3;
|
||
this.detectIncompleteChar = base64DetectIncompleteChar;
|
||
break;
|
||
default:
|
||
this.write = passThroughWrite;
|
||
return;
|
||
}
|
||
|
||
// Enough space to store all bytes of a single character. UTF-8 needs 4
|
||
// bytes, but CESU-8 may require up to 6 (3 bytes per surrogate).
|
||
this.charBuffer = new Buffer(6);
|
||
// Number of bytes received for the current incomplete multi-byte character.
|
||
this.charReceived = 0;
|
||
// Number of bytes expected for the current incomplete multi-byte character.
|
||
this.charLength = 0;
|
||
}
|
||
|
||
// write decodes the given buffer and returns it as JS string that is
|
||
// guaranteed to not contain any partial multi-byte characters. Any partial
|
||
// character found at the end of the buffer is buffered up, and will be
|
||
// returned when calling write again with the remaining bytes.
|
||
//
|
||
// Note: Converting a Buffer containing an orphan surrogate to a String
|
||
// currently works, but converting a String to a Buffer (via `new Buffer`, or
|
||
// Buffer#write) will replace incomplete surrogates with the unicode
|
||
// replacement character. See https://codereview.chromium.org/121173009/ .
|
||
StringDecoder.prototype.write = function(buffer) {
|
||
var charStr = '';
|
||
// if our last write ended with an incomplete multibyte character
|
||
while (this.charLength) {
|
||
// determine how many remaining bytes this buffer has to offer for this char
|
||
var available = (buffer.length >= this.charLength - this.charReceived) ?
|
||
this.charLength - this.charReceived :
|
||
buffer.length;
|
||
|
||
// add the new bytes to the char buffer
|
||
buffer.copy(this.charBuffer, this.charReceived, 0, available);
|
||
this.charReceived += available;
|
||
|
||
if (this.charReceived < this.charLength) {
|
||
// still not enough chars in this buffer? wait for more ...
|
||
return '';
|
||
}
|
||
|
||
// remove bytes belonging to the current character from the buffer
|
||
buffer = buffer.slice(available, buffer.length);
|
||
|
||
// get the character that was split
|
||
charStr = this.charBuffer.slice(0, this.charLength).toString(this.encoding);
|
||
|
||
// CESU-8: lead surrogate (D800-DBFF) is also the incomplete character
|
||
var charCode = charStr.charCodeAt(charStr.length - 1);
|
||
if (charCode >= 0xD800 && charCode <= 0xDBFF) {
|
||
this.charLength += this.surrogateSize;
|
||
charStr = '';
|
||
continue;
|
||
}
|
||
this.charReceived = this.charLength = 0;
|
||
|
||
// if there are no more bytes in this buffer, just emit our char
|
||
if (buffer.length === 0) {
|
||
return charStr;
|
||
}
|
||
break;
|
||
}
|
||
|
||
// determine and set charLength / charReceived
|
||
this.detectIncompleteChar(buffer);
|
||
|
||
var end = buffer.length;
|
||
if (this.charLength) {
|
||
// buffer the incomplete character bytes we got
|
||
buffer.copy(this.charBuffer, 0, buffer.length - this.charReceived, end);
|
||
end -= this.charReceived;
|
||
}
|
||
|
||
charStr += buffer.toString(this.encoding, 0, end);
|
||
|
||
var end = charStr.length - 1;
|
||
var charCode = charStr.charCodeAt(end);
|
||
// CESU-8: lead surrogate (D800-DBFF) is also the incomplete character
|
||
if (charCode >= 0xD800 && charCode <= 0xDBFF) {
|
||
var size = this.surrogateSize;
|
||
this.charLength += size;
|
||
this.charReceived += size;
|
||
this.charBuffer.copy(this.charBuffer, size, 0, size);
|
||
buffer.copy(this.charBuffer, 0, 0, size);
|
||
return charStr.substring(0, end);
|
||
}
|
||
|
||
// or just emit the charStr
|
||
return charStr;
|
||
};
|
||
|
||
// detectIncompleteChar determines if there is an incomplete UTF-8 character at
|
||
// the end of the given buffer. If so, it sets this.charLength to the byte
|
||
// length that character, and sets this.charReceived to the number of bytes
|
||
// that are available for this character.
|
||
StringDecoder.prototype.detectIncompleteChar = function(buffer) {
|
||
// determine how many bytes we have to check at the end of this buffer
|
||
var i = (buffer.length >= 3) ? 3 : buffer.length;
|
||
|
||
// Figure out if one of the last i bytes of our buffer announces an
|
||
// incomplete char.
|
||
for (; i > 0; i--) {
|
||
var c = buffer[buffer.length - i];
|
||
|
||
// See http://en.wikipedia.org/wiki/UTF-8#Description
|
||
|
||
// 110XXXXX
|
||
if (i == 1 && c >> 5 == 0x06) {
|
||
this.charLength = 2;
|
||
break;
|
||
}
|
||
|
||
// 1110XXXX
|
||
if (i <= 2 && c >> 4 == 0x0E) {
|
||
this.charLength = 3;
|
||
break;
|
||
}
|
||
|
||
// 11110XXX
|
||
if (i <= 3 && c >> 3 == 0x1E) {
|
||
this.charLength = 4;
|
||
break;
|
||
}
|
||
}
|
||
this.charReceived = i;
|
||
};
|
||
|
||
StringDecoder.prototype.end = function(buffer) {
|
||
var res = '';
|
||
if (buffer && buffer.length)
|
||
res = this.write(buffer);
|
||
|
||
if (this.charReceived) {
|
||
var cr = this.charReceived;
|
||
var buf = this.charBuffer;
|
||
var enc = this.encoding;
|
||
res += buf.slice(0, cr).toString(enc);
|
||
}
|
||
|
||
return res;
|
||
};
|
||
|
||
function passThroughWrite(buffer) {
|
||
return buffer.toString(this.encoding);
|
||
}
|
||
|
||
function utf16DetectIncompleteChar(buffer) {
|
||
this.charReceived = buffer.length % 2;
|
||
this.charLength = this.charReceived ? 2 : 0;
|
||
}
|
||
|
||
function base64DetectIncompleteChar(buffer) {
|
||
this.charReceived = buffer.length % 3;
|
||
this.charLength = this.charReceived ? 3 : 0;
|
||
}
|
||
|
||
var _polyfillNode_string_decoder = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
StringDecoder: StringDecoder
|
||
});
|
||
|
||
Readable.ReadableState = ReadableState;
|
||
|
||
var debug$6 = debuglog('stream');
|
||
inherits(Readable, EventEmitter);
|
||
|
||
function prependListener(emitter, event, fn) {
|
||
// Sadly this is not cacheable as some libraries bundle their own
|
||
// event emitter implementation with them.
|
||
if (typeof emitter.prependListener === 'function') {
|
||
return emitter.prependListener(event, fn);
|
||
} else {
|
||
// This is a hack to make sure that our error handler is attached before any
|
||
// userland ones. NEVER DO THIS. This is here only because this code needs
|
||
// to continue to work with older versions of Node.js that do not include
|
||
// the prependListener() method. The goal is to eventually remove this hack.
|
||
if (!emitter._events || !emitter._events[event])
|
||
emitter.on(event, fn);
|
||
else if (Array.isArray(emitter._events[event]))
|
||
emitter._events[event].unshift(fn);
|
||
else
|
||
emitter._events[event] = [fn, emitter._events[event]];
|
||
}
|
||
}
|
||
function listenerCount (emitter, type) {
|
||
return emitter.listeners(type).length;
|
||
}
|
||
function ReadableState(options, stream) {
|
||
|
||
options = options || {};
|
||
|
||
// object stream flag. Used to make read(n) ignore n and to
|
||
// make all the buffer merging and length checks go away
|
||
this.objectMode = !!options.objectMode;
|
||
|
||
if (stream instanceof Duplex) this.objectMode = this.objectMode || !!options.readableObjectMode;
|
||
|
||
// the point at which it stops calling _read() to fill the buffer
|
||
// Note: 0 is a valid value, means "don't call _read preemptively ever"
|
||
var hwm = options.highWaterMark;
|
||
var defaultHwm = this.objectMode ? 16 : 16 * 1024;
|
||
this.highWaterMark = hwm || hwm === 0 ? hwm : defaultHwm;
|
||
|
||
// cast to ints.
|
||
this.highWaterMark = ~ ~this.highWaterMark;
|
||
|
||
// A linked list is used to store data chunks instead of an array because the
|
||
// linked list can remove elements from the beginning faster than
|
||
// array.shift()
|
||
this.buffer = new BufferList();
|
||
this.length = 0;
|
||
this.pipes = null;
|
||
this.pipesCount = 0;
|
||
this.flowing = null;
|
||
this.ended = false;
|
||
this.endEmitted = false;
|
||
this.reading = false;
|
||
|
||
// a flag to be able to tell if the onwrite cb is called immediately,
|
||
// or on a later tick. We set this to true at first, because any
|
||
// actions that shouldn't happen until "later" should generally also
|
||
// not happen before the first write call.
|
||
this.sync = true;
|
||
|
||
// whenever we return null, then we set a flag to say
|
||
// that we're awaiting a 'readable' event emission.
|
||
this.needReadable = false;
|
||
this.emittedReadable = false;
|
||
this.readableListening = false;
|
||
this.resumeScheduled = false;
|
||
|
||
// Crypto is kind of old and crusty. Historically, its default string
|
||
// encoding is 'binary' so we have to make this configurable.
|
||
// Everything else in the universe uses 'utf8', though.
|
||
this.defaultEncoding = options.defaultEncoding || 'utf8';
|
||
|
||
// when piping, we only care about 'readable' events that happen
|
||
// after read()ing all the bytes and not getting any pushback.
|
||
this.ranOut = false;
|
||
|
||
// the number of writers that are awaiting a drain event in .pipe()s
|
||
this.awaitDrain = 0;
|
||
|
||
// if true, a maybeReadMore has been scheduled
|
||
this.readingMore = false;
|
||
|
||
this.decoder = null;
|
||
this.encoding = null;
|
||
if (options.encoding) {
|
||
this.decoder = new StringDecoder(options.encoding);
|
||
this.encoding = options.encoding;
|
||
}
|
||
}
|
||
function Readable(options) {
|
||
|
||
if (!(this instanceof Readable)) return new Readable(options);
|
||
|
||
this._readableState = new ReadableState(options, this);
|
||
|
||
// legacy
|
||
this.readable = true;
|
||
|
||
if (options && typeof options.read === 'function') this._read = options.read;
|
||
|
||
EventEmitter.call(this);
|
||
}
|
||
|
||
// Manually shove something into the read() buffer.
|
||
// This returns true if the highWaterMark has not been hit yet,
|
||
// similar to how Writable.write() returns true if you should
|
||
// write() some more.
|
||
Readable.prototype.push = function (chunk, encoding) {
|
||
var state = this._readableState;
|
||
|
||
if (!state.objectMode && typeof chunk === 'string') {
|
||
encoding = encoding || state.defaultEncoding;
|
||
if (encoding !== state.encoding) {
|
||
chunk = Buffer.from(chunk, encoding);
|
||
encoding = '';
|
||
}
|
||
}
|
||
|
||
return readableAddChunk(this, state, chunk, encoding, false);
|
||
};
|
||
|
||
// Unshift should *always* be something directly out of read()
|
||
Readable.prototype.unshift = function (chunk) {
|
||
var state = this._readableState;
|
||
return readableAddChunk(this, state, chunk, '', true);
|
||
};
|
||
|
||
Readable.prototype.isPaused = function () {
|
||
return this._readableState.flowing === false;
|
||
};
|
||
|
||
function readableAddChunk(stream, state, chunk, encoding, addToFront) {
|
||
var er = chunkInvalid(state, chunk);
|
||
if (er) {
|
||
stream.emit('error', er);
|
||
} else if (chunk === null) {
|
||
state.reading = false;
|
||
onEofChunk(stream, state);
|
||
} else if (state.objectMode || chunk && chunk.length > 0) {
|
||
if (state.ended && !addToFront) {
|
||
var e = new Error('stream.push() after EOF');
|
||
stream.emit('error', e);
|
||
} else if (state.endEmitted && addToFront) {
|
||
var _e = new Error('stream.unshift() after end event');
|
||
stream.emit('error', _e);
|
||
} else {
|
||
var skipAdd;
|
||
if (state.decoder && !addToFront && !encoding) {
|
||
chunk = state.decoder.write(chunk);
|
||
skipAdd = !state.objectMode && chunk.length === 0;
|
||
}
|
||
|
||
if (!addToFront) state.reading = false;
|
||
|
||
// Don't add to the buffer if we've decoded to an empty string chunk and
|
||
// we're not in object mode
|
||
if (!skipAdd) {
|
||
// if we want the data now, just emit it.
|
||
if (state.flowing && state.length === 0 && !state.sync) {
|
||
stream.emit('data', chunk);
|
||
stream.read(0);
|
||
} else {
|
||
// update the buffer info.
|
||
state.length += state.objectMode ? 1 : chunk.length;
|
||
if (addToFront) state.buffer.unshift(chunk);else state.buffer.push(chunk);
|
||
|
||
if (state.needReadable) emitReadable(stream);
|
||
}
|
||
}
|
||
|
||
maybeReadMore(stream, state);
|
||
}
|
||
} else if (!addToFront) {
|
||
state.reading = false;
|
||
}
|
||
|
||
return needMoreData(state);
|
||
}
|
||
|
||
// if it's past the high water mark, we can push in some more.
|
||
// Also, if we have no data yet, we can stand some
|
||
// more bytes. This is to work around cases where hwm=0,
|
||
// such as the repl. Also, if the push() triggered a
|
||
// readable event, and the user called read(largeNumber) such that
|
||
// needReadable was set, then we ought to push more, so that another
|
||
// 'readable' event will be triggered.
|
||
function needMoreData(state) {
|
||
return !state.ended && (state.needReadable || state.length < state.highWaterMark || state.length === 0);
|
||
}
|
||
|
||
// backwards compatibility.
|
||
Readable.prototype.setEncoding = function (enc) {
|
||
this._readableState.decoder = new StringDecoder(enc);
|
||
this._readableState.encoding = enc;
|
||
return this;
|
||
};
|
||
|
||
// Don't raise the hwm > 8MB
|
||
var MAX_HWM = 0x800000;
|
||
function computeNewHighWaterMark(n) {
|
||
if (n >= MAX_HWM) {
|
||
n = MAX_HWM;
|
||
} else {
|
||
// Get the next highest power of 2 to prevent increasing hwm excessively in
|
||
// tiny amounts
|
||
n--;
|
||
n |= n >>> 1;
|
||
n |= n >>> 2;
|
||
n |= n >>> 4;
|
||
n |= n >>> 8;
|
||
n |= n >>> 16;
|
||
n++;
|
||
}
|
||
return n;
|
||
}
|
||
|
||
// This function is designed to be inlinable, so please take care when making
|
||
// changes to the function body.
|
||
function howMuchToRead(n, state) {
|
||
if (n <= 0 || state.length === 0 && state.ended) return 0;
|
||
if (state.objectMode) return 1;
|
||
if (n !== n) {
|
||
// Only flow one buffer at a time
|
||
if (state.flowing && state.length) return state.buffer.head.data.length;else return state.length;
|
||
}
|
||
// If we're asking for more than the current hwm, then raise the hwm.
|
||
if (n > state.highWaterMark) state.highWaterMark = computeNewHighWaterMark(n);
|
||
if (n <= state.length) return n;
|
||
// Don't have enough
|
||
if (!state.ended) {
|
||
state.needReadable = true;
|
||
return 0;
|
||
}
|
||
return state.length;
|
||
}
|
||
|
||
// you can override either this method, or the async _read(n) below.
|
||
Readable.prototype.read = function (n) {
|
||
debug$6('read', n);
|
||
n = parseInt(n, 10);
|
||
var state = this._readableState;
|
||
var nOrig = n;
|
||
|
||
if (n !== 0) state.emittedReadable = false;
|
||
|
||
// if we're doing read(0) to trigger a readable event, but we
|
||
// already have a bunch of data in the buffer, then just trigger
|
||
// the 'readable' event and move on.
|
||
if (n === 0 && state.needReadable && (state.length >= state.highWaterMark || state.ended)) {
|
||
debug$6('read: emitReadable', state.length, state.ended);
|
||
if (state.length === 0 && state.ended) endReadable(this);else emitReadable(this);
|
||
return null;
|
||
}
|
||
|
||
n = howMuchToRead(n, state);
|
||
|
||
// if we've ended, and we're now clear, then finish it up.
|
||
if (n === 0 && state.ended) {
|
||
if (state.length === 0) endReadable(this);
|
||
return null;
|
||
}
|
||
|
||
// All the actual chunk generation logic needs to be
|
||
// *below* the call to _read. The reason is that in certain
|
||
// synthetic stream cases, such as passthrough streams, _read
|
||
// may be a completely synchronous operation which may change
|
||
// the state of the read buffer, providing enough data when
|
||
// before there was *not* enough.
|
||
//
|
||
// So, the steps are:
|
||
// 1. Figure out what the state of things will be after we do
|
||
// a read from the buffer.
|
||
//
|
||
// 2. If that resulting state will trigger a _read, then call _read.
|
||
// Note that this may be asynchronous, or synchronous. Yes, it is
|
||
// deeply ugly to write APIs this way, but that still doesn't mean
|
||
// that the Readable class should behave improperly, as streams are
|
||
// designed to be sync/async agnostic.
|
||
// Take note if the _read call is sync or async (ie, if the read call
|
||
// has returned yet), so that we know whether or not it's safe to emit
|
||
// 'readable' etc.
|
||
//
|
||
// 3. Actually pull the requested chunks out of the buffer and return.
|
||
|
||
// if we need a readable event, then we need to do some reading.
|
||
var doRead = state.needReadable;
|
||
debug$6('need readable', doRead);
|
||
|
||
// if we currently have less than the highWaterMark, then also read some
|
||
if (state.length === 0 || state.length - n < state.highWaterMark) {
|
||
doRead = true;
|
||
debug$6('length less than watermark', doRead);
|
||
}
|
||
|
||
// however, if we've ended, then there's no point, and if we're already
|
||
// reading, then it's unnecessary.
|
||
if (state.ended || state.reading) {
|
||
doRead = false;
|
||
debug$6('reading or ended', doRead);
|
||
} else if (doRead) {
|
||
debug$6('do read');
|
||
state.reading = true;
|
||
state.sync = true;
|
||
// if the length is currently zero, then we *need* a readable event.
|
||
if (state.length === 0) state.needReadable = true;
|
||
// call internal read method
|
||
this._read(state.highWaterMark);
|
||
state.sync = false;
|
||
// If _read pushed data synchronously, then `reading` will be false,
|
||
// and we need to re-evaluate how much data we can return to the user.
|
||
if (!state.reading) n = howMuchToRead(nOrig, state);
|
||
}
|
||
|
||
var ret;
|
||
if (n > 0) ret = fromList(n, state);else ret = null;
|
||
|
||
if (ret === null) {
|
||
state.needReadable = true;
|
||
n = 0;
|
||
} else {
|
||
state.length -= n;
|
||
}
|
||
|
||
if (state.length === 0) {
|
||
// If we have nothing in the buffer, then we want to know
|
||
// as soon as we *do* get something into the buffer.
|
||
if (!state.ended) state.needReadable = true;
|
||
|
||
// If we tried to read() past the EOF, then emit end on the next tick.
|
||
if (nOrig !== n && state.ended) endReadable(this);
|
||
}
|
||
|
||
if (ret !== null) this.emit('data', ret);
|
||
|
||
return ret;
|
||
};
|
||
|
||
function chunkInvalid(state, chunk) {
|
||
var er = null;
|
||
if (!Buffer.isBuffer(chunk) && typeof chunk !== 'string' && chunk !== null && chunk !== undefined && !state.objectMode) {
|
||
er = new TypeError('Invalid non-string/buffer chunk');
|
||
}
|
||
return er;
|
||
}
|
||
|
||
function onEofChunk(stream, state) {
|
||
if (state.ended) return;
|
||
if (state.decoder) {
|
||
var chunk = state.decoder.end();
|
||
if (chunk && chunk.length) {
|
||
state.buffer.push(chunk);
|
||
state.length += state.objectMode ? 1 : chunk.length;
|
||
}
|
||
}
|
||
state.ended = true;
|
||
|
||
// emit 'readable' now to make sure it gets picked up.
|
||
emitReadable(stream);
|
||
}
|
||
|
||
// Don't emit readable right away in sync mode, because this can trigger
|
||
// another read() call => stack overflow. This way, it might trigger
|
||
// a nextTick recursion warning, but that's not so bad.
|
||
function emitReadable(stream) {
|
||
var state = stream._readableState;
|
||
state.needReadable = false;
|
||
if (!state.emittedReadable) {
|
||
debug$6('emitReadable', state.flowing);
|
||
state.emittedReadable = true;
|
||
if (state.sync) nextTick(emitReadable_, stream);else emitReadable_(stream);
|
||
}
|
||
}
|
||
|
||
function emitReadable_(stream) {
|
||
debug$6('emit readable');
|
||
stream.emit('readable');
|
||
flow(stream);
|
||
}
|
||
|
||
// at this point, the user has presumably seen the 'readable' event,
|
||
// and called read() to consume some data. that may have triggered
|
||
// in turn another _read(n) call, in which case reading = true if
|
||
// it's in progress.
|
||
// However, if we're not ended, or reading, and the length < hwm,
|
||
// then go ahead and try to read some more preemptively.
|
||
function maybeReadMore(stream, state) {
|
||
if (!state.readingMore) {
|
||
state.readingMore = true;
|
||
nextTick(maybeReadMore_, stream, state);
|
||
}
|
||
}
|
||
|
||
function maybeReadMore_(stream, state) {
|
||
var len = state.length;
|
||
while (!state.reading && !state.flowing && !state.ended && state.length < state.highWaterMark) {
|
||
debug$6('maybeReadMore read 0');
|
||
stream.read(0);
|
||
if (len === state.length)
|
||
// didn't get any data, stop spinning.
|
||
break;else len = state.length;
|
||
}
|
||
state.readingMore = false;
|
||
}
|
||
|
||
// abstract method. to be overridden in specific implementation classes.
|
||
// call cb(er, data) where data is <= n in length.
|
||
// for virtual (non-string, non-buffer) streams, "length" is somewhat
|
||
// arbitrary, and perhaps not very meaningful.
|
||
Readable.prototype._read = function (n) {
|
||
this.emit('error', new Error('not implemented'));
|
||
};
|
||
|
||
Readable.prototype.pipe = function (dest, pipeOpts) {
|
||
var src = this;
|
||
var state = this._readableState;
|
||
|
||
switch (state.pipesCount) {
|
||
case 0:
|
||
state.pipes = dest;
|
||
break;
|
||
case 1:
|
||
state.pipes = [state.pipes, dest];
|
||
break;
|
||
default:
|
||
state.pipes.push(dest);
|
||
break;
|
||
}
|
||
state.pipesCount += 1;
|
||
debug$6('pipe count=%d opts=%j', state.pipesCount, pipeOpts);
|
||
|
||
var doEnd = (!pipeOpts || pipeOpts.end !== false);
|
||
|
||
var endFn = doEnd ? onend : cleanup;
|
||
if (state.endEmitted) nextTick(endFn);else src.once('end', endFn);
|
||
|
||
dest.on('unpipe', onunpipe);
|
||
function onunpipe(readable) {
|
||
debug$6('onunpipe');
|
||
if (readable === src) {
|
||
cleanup();
|
||
}
|
||
}
|
||
|
||
function onend() {
|
||
debug$6('onend');
|
||
dest.end();
|
||
}
|
||
|
||
// when the dest drains, it reduces the awaitDrain counter
|
||
// on the source. This would be more elegant with a .once()
|
||
// handler in flow(), but adding and removing repeatedly is
|
||
// too slow.
|
||
var ondrain = pipeOnDrain(src);
|
||
dest.on('drain', ondrain);
|
||
|
||
var cleanedUp = false;
|
||
function cleanup() {
|
||
debug$6('cleanup');
|
||
// cleanup event handlers once the pipe is broken
|
||
dest.removeListener('close', onclose);
|
||
dest.removeListener('finish', onfinish);
|
||
dest.removeListener('drain', ondrain);
|
||
dest.removeListener('error', onerror);
|
||
dest.removeListener('unpipe', onunpipe);
|
||
src.removeListener('end', onend);
|
||
src.removeListener('end', cleanup);
|
||
src.removeListener('data', ondata);
|
||
|
||
cleanedUp = true;
|
||
|
||
// if the reader is waiting for a drain event from this
|
||
// specific writer, then it would cause it to never start
|
||
// flowing again.
|
||
// So, if this is awaiting a drain, then we just call it now.
|
||
// If we don't know, then assume that we are waiting for one.
|
||
if (state.awaitDrain && (!dest._writableState || dest._writableState.needDrain)) ondrain();
|
||
}
|
||
|
||
// If the user pushes more data while we're writing to dest then we'll end up
|
||
// in ondata again. However, we only want to increase awaitDrain once because
|
||
// dest will only emit one 'drain' event for the multiple writes.
|
||
// => Introduce a guard on increasing awaitDrain.
|
||
var increasedAwaitDrain = false;
|
||
src.on('data', ondata);
|
||
function ondata(chunk) {
|
||
debug$6('ondata');
|
||
increasedAwaitDrain = false;
|
||
var ret = dest.write(chunk);
|
||
if (false === ret && !increasedAwaitDrain) {
|
||
// If the user unpiped during `dest.write()`, it is possible
|
||
// to get stuck in a permanently paused state if that write
|
||
// also returned false.
|
||
// => Check whether `dest` is still a piping destination.
|
||
if ((state.pipesCount === 1 && state.pipes === dest || state.pipesCount > 1 && indexOf(state.pipes, dest) !== -1) && !cleanedUp) {
|
||
debug$6('false write response, pause', src._readableState.awaitDrain);
|
||
src._readableState.awaitDrain++;
|
||
increasedAwaitDrain = true;
|
||
}
|
||
src.pause();
|
||
}
|
||
}
|
||
|
||
// if the dest has an error, then stop piping into it.
|
||
// however, don't suppress the throwing behavior for this.
|
||
function onerror(er) {
|
||
debug$6('onerror', er);
|
||
unpipe();
|
||
dest.removeListener('error', onerror);
|
||
if (listenerCount(dest, 'error') === 0) dest.emit('error', er);
|
||
}
|
||
|
||
// Make sure our error handler is attached before userland ones.
|
||
prependListener(dest, 'error', onerror);
|
||
|
||
// Both close and finish should trigger unpipe, but only once.
|
||
function onclose() {
|
||
dest.removeListener('finish', onfinish);
|
||
unpipe();
|
||
}
|
||
dest.once('close', onclose);
|
||
function onfinish() {
|
||
debug$6('onfinish');
|
||
dest.removeListener('close', onclose);
|
||
unpipe();
|
||
}
|
||
dest.once('finish', onfinish);
|
||
|
||
function unpipe() {
|
||
debug$6('unpipe');
|
||
src.unpipe(dest);
|
||
}
|
||
|
||
// tell the dest that it's being piped to
|
||
dest.emit('pipe', src);
|
||
|
||
// start the flow if it hasn't been started already.
|
||
if (!state.flowing) {
|
||
debug$6('pipe resume');
|
||
src.resume();
|
||
}
|
||
|
||
return dest;
|
||
};
|
||
|
||
function pipeOnDrain(src) {
|
||
return function () {
|
||
var state = src._readableState;
|
||
debug$6('pipeOnDrain', state.awaitDrain);
|
||
if (state.awaitDrain) state.awaitDrain--;
|
||
if (state.awaitDrain === 0 && src.listeners('data').length) {
|
||
state.flowing = true;
|
||
flow(src);
|
||
}
|
||
};
|
||
}
|
||
|
||
Readable.prototype.unpipe = function (dest) {
|
||
var state = this._readableState;
|
||
|
||
// if we're not piping anywhere, then do nothing.
|
||
if (state.pipesCount === 0) return this;
|
||
|
||
// just one destination. most common case.
|
||
if (state.pipesCount === 1) {
|
||
// passed in one, but it's not the right one.
|
||
if (dest && dest !== state.pipes) return this;
|
||
|
||
if (!dest) dest = state.pipes;
|
||
|
||
// got a match.
|
||
state.pipes = null;
|
||
state.pipesCount = 0;
|
||
state.flowing = false;
|
||
if (dest) dest.emit('unpipe', this);
|
||
return this;
|
||
}
|
||
|
||
// slow case. multiple pipe destinations.
|
||
|
||
if (!dest) {
|
||
// remove all.
|
||
var dests = state.pipes;
|
||
var len = state.pipesCount;
|
||
state.pipes = null;
|
||
state.pipesCount = 0;
|
||
state.flowing = false;
|
||
|
||
for (var _i = 0; _i < len; _i++) {
|
||
dests[_i].emit('unpipe', this);
|
||
}return this;
|
||
}
|
||
|
||
// try to find the right one.
|
||
var i = indexOf(state.pipes, dest);
|
||
if (i === -1) return this;
|
||
|
||
state.pipes.splice(i, 1);
|
||
state.pipesCount -= 1;
|
||
if (state.pipesCount === 1) state.pipes = state.pipes[0];
|
||
|
||
dest.emit('unpipe', this);
|
||
|
||
return this;
|
||
};
|
||
|
||
// set up data events if they are asked for
|
||
// Ensure readable listeners eventually get something
|
||
Readable.prototype.on = function (ev, fn) {
|
||
var res = EventEmitter.prototype.on.call(this, ev, fn);
|
||
|
||
if (ev === 'data') {
|
||
// Start flowing on next tick if stream isn't explicitly paused
|
||
if (this._readableState.flowing !== false) this.resume();
|
||
} else if (ev === 'readable') {
|
||
var state = this._readableState;
|
||
if (!state.endEmitted && !state.readableListening) {
|
||
state.readableListening = state.needReadable = true;
|
||
state.emittedReadable = false;
|
||
if (!state.reading) {
|
||
nextTick(nReadingNextTick, this);
|
||
} else if (state.length) {
|
||
emitReadable(this);
|
||
}
|
||
}
|
||
}
|
||
|
||
return res;
|
||
};
|
||
Readable.prototype.addListener = Readable.prototype.on;
|
||
|
||
function nReadingNextTick(self) {
|
||
debug$6('readable nexttick read 0');
|
||
self.read(0);
|
||
}
|
||
|
||
// pause() and resume() are remnants of the legacy readable stream API
|
||
// If the user uses them, then switch into old mode.
|
||
Readable.prototype.resume = function () {
|
||
var state = this._readableState;
|
||
if (!state.flowing) {
|
||
debug$6('resume');
|
||
state.flowing = true;
|
||
resume(this, state);
|
||
}
|
||
return this;
|
||
};
|
||
|
||
function resume(stream, state) {
|
||
if (!state.resumeScheduled) {
|
||
state.resumeScheduled = true;
|
||
nextTick(resume_, stream, state);
|
||
}
|
||
}
|
||
|
||
function resume_(stream, state) {
|
||
if (!state.reading) {
|
||
debug$6('resume read 0');
|
||
stream.read(0);
|
||
}
|
||
|
||
state.resumeScheduled = false;
|
||
state.awaitDrain = 0;
|
||
stream.emit('resume');
|
||
flow(stream);
|
||
if (state.flowing && !state.reading) stream.read(0);
|
||
}
|
||
|
||
Readable.prototype.pause = function () {
|
||
debug$6('call pause flowing=%j', this._readableState.flowing);
|
||
if (false !== this._readableState.flowing) {
|
||
debug$6('pause');
|
||
this._readableState.flowing = false;
|
||
this.emit('pause');
|
||
}
|
||
return this;
|
||
};
|
||
|
||
function flow(stream) {
|
||
var state = stream._readableState;
|
||
debug$6('flow', state.flowing);
|
||
while (state.flowing && stream.read() !== null) {}
|
||
}
|
||
|
||
// wrap an old-style stream as the async data source.
|
||
// This is *not* part of the readable stream interface.
|
||
// It is an ugly unfortunate mess of history.
|
||
Readable.prototype.wrap = function (stream) {
|
||
var state = this._readableState;
|
||
var paused = false;
|
||
|
||
var self = this;
|
||
stream.on('end', function () {
|
||
debug$6('wrapped end');
|
||
if (state.decoder && !state.ended) {
|
||
var chunk = state.decoder.end();
|
||
if (chunk && chunk.length) self.push(chunk);
|
||
}
|
||
|
||
self.push(null);
|
||
});
|
||
|
||
stream.on('data', function (chunk) {
|
||
debug$6('wrapped data');
|
||
if (state.decoder) chunk = state.decoder.write(chunk);
|
||
|
||
// don't skip over falsy values in objectMode
|
||
if (state.objectMode && (chunk === null || chunk === undefined)) return;else if (!state.objectMode && (!chunk || !chunk.length)) return;
|
||
|
||
var ret = self.push(chunk);
|
||
if (!ret) {
|
||
paused = true;
|
||
stream.pause();
|
||
}
|
||
});
|
||
|
||
// proxy all the other methods.
|
||
// important when wrapping filters and duplexes.
|
||
for (var i in stream) {
|
||
if (this[i] === undefined && typeof stream[i] === 'function') {
|
||
this[i] = function (method) {
|
||
return function () {
|
||
return stream[method].apply(stream, arguments);
|
||
};
|
||
}(i);
|
||
}
|
||
}
|
||
|
||
// proxy certain important events.
|
||
var events = ['error', 'close', 'destroy', 'pause', 'resume'];
|
||
forEach(events, function (ev) {
|
||
stream.on(ev, self.emit.bind(self, ev));
|
||
});
|
||
|
||
// when we try to consume some more bytes, simply unpause the
|
||
// underlying stream.
|
||
self._read = function (n) {
|
||
debug$6('wrapped _read', n);
|
||
if (paused) {
|
||
paused = false;
|
||
stream.resume();
|
||
}
|
||
};
|
||
|
||
return self;
|
||
};
|
||
|
||
// exposed for testing purposes only.
|
||
Readable._fromList = fromList;
|
||
|
||
// Pluck off n bytes from an array of buffers.
|
||
// Length is the combined lengths of all the buffers in the list.
|
||
// This function is designed to be inlinable, so please take care when making
|
||
// changes to the function body.
|
||
function fromList(n, state) {
|
||
// nothing buffered
|
||
if (state.length === 0) return null;
|
||
|
||
var ret;
|
||
if (state.objectMode) ret = state.buffer.shift();else if (!n || n >= state.length) {
|
||
// read it all, truncate the list
|
||
if (state.decoder) ret = state.buffer.join('');else if (state.buffer.length === 1) ret = state.buffer.head.data;else ret = state.buffer.concat(state.length);
|
||
state.buffer.clear();
|
||
} else {
|
||
// read part of list
|
||
ret = fromListPartial(n, state.buffer, state.decoder);
|
||
}
|
||
|
||
return ret;
|
||
}
|
||
|
||
// Extracts only enough buffered data to satisfy the amount requested.
|
||
// This function is designed to be inlinable, so please take care when making
|
||
// changes to the function body.
|
||
function fromListPartial(n, list, hasStrings) {
|
||
var ret;
|
||
if (n < list.head.data.length) {
|
||
// slice is the same for buffers and strings
|
||
ret = list.head.data.slice(0, n);
|
||
list.head.data = list.head.data.slice(n);
|
||
} else if (n === list.head.data.length) {
|
||
// first chunk is a perfect match
|
||
ret = list.shift();
|
||
} else {
|
||
// result spans more than one buffer
|
||
ret = hasStrings ? copyFromBufferString(n, list) : copyFromBuffer(n, list);
|
||
}
|
||
return ret;
|
||
}
|
||
|
||
// Copies a specified amount of characters from the list of buffered data
|
||
// chunks.
|
||
// This function is designed to be inlinable, so please take care when making
|
||
// changes to the function body.
|
||
function copyFromBufferString(n, list) {
|
||
var p = list.head;
|
||
var c = 1;
|
||
var ret = p.data;
|
||
n -= ret.length;
|
||
while (p = p.next) {
|
||
var str = p.data;
|
||
var nb = n > str.length ? str.length : n;
|
||
if (nb === str.length) ret += str;else ret += str.slice(0, n);
|
||
n -= nb;
|
||
if (n === 0) {
|
||
if (nb === str.length) {
|
||
++c;
|
||
if (p.next) list.head = p.next;else list.head = list.tail = null;
|
||
} else {
|
||
list.head = p;
|
||
p.data = str.slice(nb);
|
||
}
|
||
break;
|
||
}
|
||
++c;
|
||
}
|
||
list.length -= c;
|
||
return ret;
|
||
}
|
||
|
||
// Copies a specified amount of bytes from the list of buffered data chunks.
|
||
// This function is designed to be inlinable, so please take care when making
|
||
// changes to the function body.
|
||
function copyFromBuffer(n, list) {
|
||
var ret = Buffer.allocUnsafe(n);
|
||
var p = list.head;
|
||
var c = 1;
|
||
p.data.copy(ret);
|
||
n -= p.data.length;
|
||
while (p = p.next) {
|
||
var buf = p.data;
|
||
var nb = n > buf.length ? buf.length : n;
|
||
buf.copy(ret, ret.length - n, 0, nb);
|
||
n -= nb;
|
||
if (n === 0) {
|
||
if (nb === buf.length) {
|
||
++c;
|
||
if (p.next) list.head = p.next;else list.head = list.tail = null;
|
||
} else {
|
||
list.head = p;
|
||
p.data = buf.slice(nb);
|
||
}
|
||
break;
|
||
}
|
||
++c;
|
||
}
|
||
list.length -= c;
|
||
return ret;
|
||
}
|
||
|
||
function endReadable(stream) {
|
||
var state = stream._readableState;
|
||
|
||
// If we get here before consuming all the bytes, then that is a
|
||
// bug in node. Should never happen.
|
||
if (state.length > 0) throw new Error('"endReadable()" called on non-empty stream');
|
||
|
||
if (!state.endEmitted) {
|
||
state.ended = true;
|
||
nextTick(endReadableNT, state, stream);
|
||
}
|
||
}
|
||
|
||
function endReadableNT(state, stream) {
|
||
// Check that we didn't get one last unshift.
|
||
if (!state.endEmitted && state.length === 0) {
|
||
state.endEmitted = true;
|
||
stream.readable = false;
|
||
stream.emit('end');
|
||
}
|
||
}
|
||
|
||
function forEach(xs, f) {
|
||
for (var i = 0, l = xs.length; i < l; i++) {
|
||
f(xs[i], i);
|
||
}
|
||
}
|
||
|
||
function indexOf(xs, x) {
|
||
for (var i = 0, l = xs.length; i < l; i++) {
|
||
if (xs[i] === x) return i;
|
||
}
|
||
return -1;
|
||
}
|
||
|
||
// A bit simpler than readable streams.
|
||
// Implement an async ._write(chunk, encoding, cb), and it'll handle all
|
||
// the drain event emission and buffering.
|
||
|
||
Writable.WritableState = WritableState;
|
||
inherits(Writable, EventEmitter);
|
||
|
||
function nop() {}
|
||
|
||
function WriteReq(chunk, encoding, cb) {
|
||
this.chunk = chunk;
|
||
this.encoding = encoding;
|
||
this.callback = cb;
|
||
this.next = null;
|
||
}
|
||
|
||
function WritableState(options, stream) {
|
||
Object.defineProperty(this, 'buffer', {
|
||
get: deprecate(function () {
|
||
return this.getBuffer();
|
||
}, '_writableState.buffer is deprecated. Use _writableState.getBuffer ' + 'instead.')
|
||
});
|
||
options = options || {};
|
||
|
||
// object stream flag to indicate whether or not this stream
|
||
// contains buffers or objects.
|
||
this.objectMode = !!options.objectMode;
|
||
|
||
if (stream instanceof Duplex) this.objectMode = this.objectMode || !!options.writableObjectMode;
|
||
|
||
// the point at which write() starts returning false
|
||
// Note: 0 is a valid value, means that we always return false if
|
||
// the entire buffer is not flushed immediately on write()
|
||
var hwm = options.highWaterMark;
|
||
var defaultHwm = this.objectMode ? 16 : 16 * 1024;
|
||
this.highWaterMark = hwm || hwm === 0 ? hwm : defaultHwm;
|
||
|
||
// cast to ints.
|
||
this.highWaterMark = ~ ~this.highWaterMark;
|
||
|
||
this.needDrain = false;
|
||
// at the start of calling end()
|
||
this.ending = false;
|
||
// when end() has been called, and returned
|
||
this.ended = false;
|
||
// when 'finish' is emitted
|
||
this.finished = false;
|
||
|
||
// should we decode strings into buffers before passing to _write?
|
||
// this is here so that some node-core streams can optimize string
|
||
// handling at a lower level.
|
||
var noDecode = options.decodeStrings === false;
|
||
this.decodeStrings = !noDecode;
|
||
|
||
// Crypto is kind of old and crusty. Historically, its default string
|
||
// encoding is 'binary' so we have to make this configurable.
|
||
// Everything else in the universe uses 'utf8', though.
|
||
this.defaultEncoding = options.defaultEncoding || 'utf8';
|
||
|
||
// not an actual buffer we keep track of, but a measurement
|
||
// of how much we're waiting to get pushed to some underlying
|
||
// socket or file.
|
||
this.length = 0;
|
||
|
||
// a flag to see when we're in the middle of a write.
|
||
this.writing = false;
|
||
|
||
// when true all writes will be buffered until .uncork() call
|
||
this.corked = 0;
|
||
|
||
// a flag to be able to tell if the onwrite cb is called immediately,
|
||
// or on a later tick. We set this to true at first, because any
|
||
// actions that shouldn't happen until "later" should generally also
|
||
// not happen before the first write call.
|
||
this.sync = true;
|
||
|
||
// a flag to know if we're processing previously buffered items, which
|
||
// may call the _write() callback in the same tick, so that we don't
|
||
// end up in an overlapped onwrite situation.
|
||
this.bufferProcessing = false;
|
||
|
||
// the callback that's passed to _write(chunk,cb)
|
||
this.onwrite = function (er) {
|
||
onwrite(stream, er);
|
||
};
|
||
|
||
// the callback that the user supplies to write(chunk,encoding,cb)
|
||
this.writecb = null;
|
||
|
||
// the amount that is being written when _write is called.
|
||
this.writelen = 0;
|
||
|
||
this.bufferedRequest = null;
|
||
this.lastBufferedRequest = null;
|
||
|
||
// number of pending user-supplied write callbacks
|
||
// this must be 0 before 'finish' can be emitted
|
||
this.pendingcb = 0;
|
||
|
||
// emit prefinish if the only thing we're waiting for is _write cbs
|
||
// This is relevant for synchronous Transform streams
|
||
this.prefinished = false;
|
||
|
||
// True if the error was already emitted and should not be thrown again
|
||
this.errorEmitted = false;
|
||
|
||
// count buffered requests
|
||
this.bufferedRequestCount = 0;
|
||
|
||
// allocate the first CorkedRequest, there is always
|
||
// one allocated and free to use, and we maintain at most two
|
||
this.corkedRequestsFree = new CorkedRequest(this);
|
||
}
|
||
|
||
WritableState.prototype.getBuffer = function writableStateGetBuffer() {
|
||
var current = this.bufferedRequest;
|
||
var out = [];
|
||
while (current) {
|
||
out.push(current);
|
||
current = current.next;
|
||
}
|
||
return out;
|
||
};
|
||
function Writable(options) {
|
||
|
||
// Writable ctor is applied to Duplexes, though they're not
|
||
// instanceof Writable, they're instanceof Readable.
|
||
if (!(this instanceof Writable) && !(this instanceof Duplex)) return new Writable(options);
|
||
|
||
this._writableState = new WritableState(options, this);
|
||
|
||
// legacy.
|
||
this.writable = true;
|
||
|
||
if (options) {
|
||
if (typeof options.write === 'function') this._write = options.write;
|
||
|
||
if (typeof options.writev === 'function') this._writev = options.writev;
|
||
}
|
||
|
||
EventEmitter.call(this);
|
||
}
|
||
|
||
// Otherwise people can pipe Writable streams, which is just wrong.
|
||
Writable.prototype.pipe = function () {
|
||
this.emit('error', new Error('Cannot pipe, not readable'));
|
||
};
|
||
|
||
function writeAfterEnd(stream, cb) {
|
||
var er = new Error('write after end');
|
||
// TODO: defer error events consistently everywhere, not just the cb
|
||
stream.emit('error', er);
|
||
nextTick(cb, er);
|
||
}
|
||
|
||
// If we get something that is not a buffer, string, null, or undefined,
|
||
// and we're not in objectMode, then that's an error.
|
||
// Otherwise stream chunks are all considered to be of length=1, and the
|
||
// watermarks determine how many objects to keep in the buffer, rather than
|
||
// how many bytes or characters.
|
||
function validChunk(stream, state, chunk, cb) {
|
||
var valid = true;
|
||
var er = false;
|
||
// Always throw error if a null is written
|
||
// if we are not in object mode then throw
|
||
// if it is not a buffer, string, or undefined.
|
||
if (chunk === null) {
|
||
er = new TypeError('May not write null values to stream');
|
||
} else if (!Buffer.isBuffer(chunk) && typeof chunk !== 'string' && chunk !== undefined && !state.objectMode) {
|
||
er = new TypeError('Invalid non-string/buffer chunk');
|
||
}
|
||
if (er) {
|
||
stream.emit('error', er);
|
||
nextTick(cb, er);
|
||
valid = false;
|
||
}
|
||
return valid;
|
||
}
|
||
|
||
Writable.prototype.write = function (chunk, encoding, cb) {
|
||
var state = this._writableState;
|
||
var ret = false;
|
||
|
||
if (typeof encoding === 'function') {
|
||
cb = encoding;
|
||
encoding = null;
|
||
}
|
||
|
||
if (Buffer.isBuffer(chunk)) encoding = 'buffer';else if (!encoding) encoding = state.defaultEncoding;
|
||
|
||
if (typeof cb !== 'function') cb = nop;
|
||
|
||
if (state.ended) writeAfterEnd(this, cb);else if (validChunk(this, state, chunk, cb)) {
|
||
state.pendingcb++;
|
||
ret = writeOrBuffer(this, state, chunk, encoding, cb);
|
||
}
|
||
|
||
return ret;
|
||
};
|
||
|
||
Writable.prototype.cork = function () {
|
||
var state = this._writableState;
|
||
|
||
state.corked++;
|
||
};
|
||
|
||
Writable.prototype.uncork = function () {
|
||
var state = this._writableState;
|
||
|
||
if (state.corked) {
|
||
state.corked--;
|
||
|
||
if (!state.writing && !state.corked && !state.finished && !state.bufferProcessing && state.bufferedRequest) clearBuffer(this, state);
|
||
}
|
||
};
|
||
|
||
Writable.prototype.setDefaultEncoding = function setDefaultEncoding(encoding) {
|
||
// node::ParseEncoding() requires lower case.
|
||
if (typeof encoding === 'string') encoding = encoding.toLowerCase();
|
||
if (!(['hex', 'utf8', 'utf-8', 'ascii', 'binary', 'base64', 'ucs2', 'ucs-2', 'utf16le', 'utf-16le', 'raw'].indexOf((encoding + '').toLowerCase()) > -1)) throw new TypeError('Unknown encoding: ' + encoding);
|
||
this._writableState.defaultEncoding = encoding;
|
||
return this;
|
||
};
|
||
|
||
function decodeChunk(state, chunk, encoding) {
|
||
if (!state.objectMode && state.decodeStrings !== false && typeof chunk === 'string') {
|
||
chunk = Buffer.from(chunk, encoding);
|
||
}
|
||
return chunk;
|
||
}
|
||
|
||
// if we're already writing something, then just put this
|
||
// in the queue, and wait our turn. Otherwise, call _write
|
||
// If we return false, then we need a drain event, so set that flag.
|
||
function writeOrBuffer(stream, state, chunk, encoding, cb) {
|
||
chunk = decodeChunk(state, chunk, encoding);
|
||
|
||
if (Buffer.isBuffer(chunk)) encoding = 'buffer';
|
||
var len = state.objectMode ? 1 : chunk.length;
|
||
|
||
state.length += len;
|
||
|
||
var ret = state.length < state.highWaterMark;
|
||
// we must ensure that previous needDrain will not be reset to false.
|
||
if (!ret) state.needDrain = true;
|
||
|
||
if (state.writing || state.corked) {
|
||
var last = state.lastBufferedRequest;
|
||
state.lastBufferedRequest = new WriteReq(chunk, encoding, cb);
|
||
if (last) {
|
||
last.next = state.lastBufferedRequest;
|
||
} else {
|
||
state.bufferedRequest = state.lastBufferedRequest;
|
||
}
|
||
state.bufferedRequestCount += 1;
|
||
} else {
|
||
doWrite(stream, state, false, len, chunk, encoding, cb);
|
||
}
|
||
|
||
return ret;
|
||
}
|
||
|
||
function doWrite(stream, state, writev, len, chunk, encoding, cb) {
|
||
state.writelen = len;
|
||
state.writecb = cb;
|
||
state.writing = true;
|
||
state.sync = true;
|
||
if (writev) stream._writev(chunk, state.onwrite);else stream._write(chunk, encoding, state.onwrite);
|
||
state.sync = false;
|
||
}
|
||
|
||
function onwriteError(stream, state, sync, er, cb) {
|
||
--state.pendingcb;
|
||
if (sync) nextTick(cb, er);else cb(er);
|
||
|
||
stream._writableState.errorEmitted = true;
|
||
stream.emit('error', er);
|
||
}
|
||
|
||
function onwriteStateUpdate(state) {
|
||
state.writing = false;
|
||
state.writecb = null;
|
||
state.length -= state.writelen;
|
||
state.writelen = 0;
|
||
}
|
||
|
||
function onwrite(stream, er) {
|
||
var state = stream._writableState;
|
||
var sync = state.sync;
|
||
var cb = state.writecb;
|
||
|
||
onwriteStateUpdate(state);
|
||
|
||
if (er) onwriteError(stream, state, sync, er, cb);else {
|
||
// Check if we're actually ready to finish, but don't emit yet
|
||
var finished = needFinish(state);
|
||
|
||
if (!finished && !state.corked && !state.bufferProcessing && state.bufferedRequest) {
|
||
clearBuffer(stream, state);
|
||
}
|
||
|
||
if (sync) {
|
||
/*<replacement>*/
|
||
nextTick(afterWrite, stream, state, finished, cb);
|
||
/*</replacement>*/
|
||
} else {
|
||
afterWrite(stream, state, finished, cb);
|
||
}
|
||
}
|
||
}
|
||
|
||
function afterWrite(stream, state, finished, cb) {
|
||
if (!finished) onwriteDrain(stream, state);
|
||
state.pendingcb--;
|
||
cb();
|
||
finishMaybe(stream, state);
|
||
}
|
||
|
||
// Must force callback to be called on nextTick, so that we don't
|
||
// emit 'drain' before the write() consumer gets the 'false' return
|
||
// value, and has a chance to attach a 'drain' listener.
|
||
function onwriteDrain(stream, state) {
|
||
if (state.length === 0 && state.needDrain) {
|
||
state.needDrain = false;
|
||
stream.emit('drain');
|
||
}
|
||
}
|
||
|
||
// if there's something in the buffer waiting, then process it
|
||
function clearBuffer(stream, state) {
|
||
state.bufferProcessing = true;
|
||
var entry = state.bufferedRequest;
|
||
|
||
if (stream._writev && entry && entry.next) {
|
||
// Fast case, write everything using _writev()
|
||
var l = state.bufferedRequestCount;
|
||
var buffer = new Array(l);
|
||
var holder = state.corkedRequestsFree;
|
||
holder.entry = entry;
|
||
|
||
var count = 0;
|
||
while (entry) {
|
||
buffer[count] = entry;
|
||
entry = entry.next;
|
||
count += 1;
|
||
}
|
||
|
||
doWrite(stream, state, true, state.length, buffer, '', holder.finish);
|
||
|
||
// doWrite is almost always async, defer these to save a bit of time
|
||
// as the hot path ends with doWrite
|
||
state.pendingcb++;
|
||
state.lastBufferedRequest = null;
|
||
if (holder.next) {
|
||
state.corkedRequestsFree = holder.next;
|
||
holder.next = null;
|
||
} else {
|
||
state.corkedRequestsFree = new CorkedRequest(state);
|
||
}
|
||
} else {
|
||
// Slow case, write chunks one-by-one
|
||
while (entry) {
|
||
var chunk = entry.chunk;
|
||
var encoding = entry.encoding;
|
||
var cb = entry.callback;
|
||
var len = state.objectMode ? 1 : chunk.length;
|
||
|
||
doWrite(stream, state, false, len, chunk, encoding, cb);
|
||
entry = entry.next;
|
||
// if we didn't call the onwrite immediately, then
|
||
// it means that we need to wait until it does.
|
||
// also, that means that the chunk and cb are currently
|
||
// being processed, so move the buffer counter past them.
|
||
if (state.writing) {
|
||
break;
|
||
}
|
||
}
|
||
|
||
if (entry === null) state.lastBufferedRequest = null;
|
||
}
|
||
|
||
state.bufferedRequestCount = 0;
|
||
state.bufferedRequest = entry;
|
||
state.bufferProcessing = false;
|
||
}
|
||
|
||
Writable.prototype._write = function (chunk, encoding, cb) {
|
||
cb(new Error('not implemented'));
|
||
};
|
||
|
||
Writable.prototype._writev = null;
|
||
|
||
Writable.prototype.end = function (chunk, encoding, cb) {
|
||
var state = this._writableState;
|
||
|
||
if (typeof chunk === 'function') {
|
||
cb = chunk;
|
||
chunk = null;
|
||
encoding = null;
|
||
} else if (typeof encoding === 'function') {
|
||
cb = encoding;
|
||
encoding = null;
|
||
}
|
||
|
||
if (chunk !== null && chunk !== undefined) this.write(chunk, encoding);
|
||
|
||
// .end() fully uncorks
|
||
if (state.corked) {
|
||
state.corked = 1;
|
||
this.uncork();
|
||
}
|
||
|
||
// ignore unnecessary end() calls.
|
||
if (!state.ending && !state.finished) endWritable(this, state, cb);
|
||
};
|
||
|
||
function needFinish(state) {
|
||
return state.ending && state.length === 0 && state.bufferedRequest === null && !state.finished && !state.writing;
|
||
}
|
||
|
||
function prefinish(stream, state) {
|
||
if (!state.prefinished) {
|
||
state.prefinished = true;
|
||
stream.emit('prefinish');
|
||
}
|
||
}
|
||
|
||
function finishMaybe(stream, state) {
|
||
var need = needFinish(state);
|
||
if (need) {
|
||
if (state.pendingcb === 0) {
|
||
prefinish(stream, state);
|
||
state.finished = true;
|
||
stream.emit('finish');
|
||
} else {
|
||
prefinish(stream, state);
|
||
}
|
||
}
|
||
return need;
|
||
}
|
||
|
||
function endWritable(stream, state, cb) {
|
||
state.ending = true;
|
||
finishMaybe(stream, state);
|
||
if (cb) {
|
||
if (state.finished) nextTick(cb);else stream.once('finish', cb);
|
||
}
|
||
state.ended = true;
|
||
stream.writable = false;
|
||
}
|
||
|
||
// It seems a linked list but it is not
|
||
// there will be only 2 of these for each stream
|
||
function CorkedRequest(state) {
|
||
var _this = this;
|
||
|
||
this.next = null;
|
||
this.entry = null;
|
||
|
||
this.finish = function (err) {
|
||
var entry = _this.entry;
|
||
_this.entry = null;
|
||
while (entry) {
|
||
var cb = entry.callback;
|
||
state.pendingcb--;
|
||
cb(err);
|
||
entry = entry.next;
|
||
}
|
||
if (state.corkedRequestsFree) {
|
||
state.corkedRequestsFree.next = _this;
|
||
} else {
|
||
state.corkedRequestsFree = _this;
|
||
}
|
||
};
|
||
}
|
||
|
||
inherits(Duplex, Readable);
|
||
|
||
var keys = Object.keys(Writable.prototype);
|
||
for (var v = 0; v < keys.length; v++) {
|
||
var method = keys[v];
|
||
if (!Duplex.prototype[method]) Duplex.prototype[method] = Writable.prototype[method];
|
||
}
|
||
function Duplex(options) {
|
||
if (!(this instanceof Duplex)) return new Duplex(options);
|
||
|
||
Readable.call(this, options);
|
||
Writable.call(this, options);
|
||
|
||
if (options && options.readable === false) this.readable = false;
|
||
|
||
if (options && options.writable === false) this.writable = false;
|
||
|
||
this.allowHalfOpen = true;
|
||
if (options && options.allowHalfOpen === false) this.allowHalfOpen = false;
|
||
|
||
this.once('end', onend);
|
||
}
|
||
|
||
// the no-half-open enforcer
|
||
function onend() {
|
||
// if we allow half-open state, or if the writable side ended,
|
||
// then we're ok.
|
||
if (this.allowHalfOpen || this._writableState.ended) return;
|
||
|
||
// no more data can be written.
|
||
// But allow more writes to happen in this tick.
|
||
nextTick(onEndNT, this);
|
||
}
|
||
|
||
function onEndNT(self) {
|
||
self.end();
|
||
}
|
||
|
||
// a transform stream is a readable/writable stream where you do
|
||
// something with the data. Sometimes it's called a "filter",
|
||
// but that's not a great name for it, since that implies a thing where
|
||
// some bits pass through, and others are simply ignored. (That would
|
||
// be a valid example of a transform, of course.)
|
||
//
|
||
// While the output is causally related to the input, it's not a
|
||
// necessarily symmetric or synchronous transformation. For example,
|
||
// a zlib stream might take multiple plain-text writes(), and then
|
||
// emit a single compressed chunk some time in the future.
|
||
//
|
||
// Here's how this works:
|
||
//
|
||
// The Transform stream has all the aspects of the readable and writable
|
||
// stream classes. When you write(chunk), that calls _write(chunk,cb)
|
||
// internally, and returns false if there's a lot of pending writes
|
||
// buffered up. When you call read(), that calls _read(n) until
|
||
// there's enough pending readable data buffered up.
|
||
//
|
||
// In a transform stream, the written data is placed in a buffer. When
|
||
// _read(n) is called, it transforms the queued up data, calling the
|
||
// buffered _write cb's as it consumes chunks. If consuming a single
|
||
// written chunk would result in multiple output chunks, then the first
|
||
// outputted bit calls the readcb, and subsequent chunks just go into
|
||
// the read buffer, and will cause it to emit 'readable' if necessary.
|
||
//
|
||
// This way, back-pressure is actually determined by the reading side,
|
||
// since _read has to be called to start processing a new chunk. However,
|
||
// a pathological inflate type of transform can cause excessive buffering
|
||
// here. For example, imagine a stream where every byte of input is
|
||
// interpreted as an integer from 0-255, and then results in that many
|
||
// bytes of output. Writing the 4 bytes {ff,ff,ff,ff} would result in
|
||
// 1kb of data being output. In this case, you could write a very small
|
||
// amount of input, and end up with a very large amount of output. In
|
||
// such a pathological inflating mechanism, there'd be no way to tell
|
||
// the system to stop doing the transform. A single 4MB write could
|
||
// cause the system to run out of memory.
|
||
//
|
||
// However, even in such a pathological case, only a single written chunk
|
||
// would be consumed, and then the rest would wait (un-transformed) until
|
||
// the results of the previous transformed chunk were consumed.
|
||
|
||
inherits(Transform, Duplex);
|
||
|
||
function TransformState(stream) {
|
||
this.afterTransform = function (er, data) {
|
||
return afterTransform(stream, er, data);
|
||
};
|
||
|
||
this.needTransform = false;
|
||
this.transforming = false;
|
||
this.writecb = null;
|
||
this.writechunk = null;
|
||
this.writeencoding = null;
|
||
}
|
||
|
||
function afterTransform(stream, er, data) {
|
||
var ts = stream._transformState;
|
||
ts.transforming = false;
|
||
|
||
var cb = ts.writecb;
|
||
|
||
if (!cb) return stream.emit('error', new Error('no writecb in Transform class'));
|
||
|
||
ts.writechunk = null;
|
||
ts.writecb = null;
|
||
|
||
if (data !== null && data !== undefined) stream.push(data);
|
||
|
||
cb(er);
|
||
|
||
var rs = stream._readableState;
|
||
rs.reading = false;
|
||
if (rs.needReadable || rs.length < rs.highWaterMark) {
|
||
stream._read(rs.highWaterMark);
|
||
}
|
||
}
|
||
function Transform(options) {
|
||
if (!(this instanceof Transform)) return new Transform(options);
|
||
|
||
Duplex.call(this, options);
|
||
|
||
this._transformState = new TransformState(this);
|
||
|
||
// when the writable side finishes, then flush out anything remaining.
|
||
var stream = this;
|
||
|
||
// start out asking for a readable event once data is transformed.
|
||
this._readableState.needReadable = true;
|
||
|
||
// we have implemented the _read method, and done the other things
|
||
// that Readable wants before the first _read call, so unset the
|
||
// sync guard flag.
|
||
this._readableState.sync = false;
|
||
|
||
if (options) {
|
||
if (typeof options.transform === 'function') this._transform = options.transform;
|
||
|
||
if (typeof options.flush === 'function') this._flush = options.flush;
|
||
}
|
||
|
||
this.once('prefinish', function () {
|
||
if (typeof this._flush === 'function') this._flush(function (er) {
|
||
done(stream, er);
|
||
});else done(stream);
|
||
});
|
||
}
|
||
|
||
Transform.prototype.push = function (chunk, encoding) {
|
||
this._transformState.needTransform = false;
|
||
return Duplex.prototype.push.call(this, chunk, encoding);
|
||
};
|
||
|
||
// This is the part where you do stuff!
|
||
// override this function in implementation classes.
|
||
// 'chunk' is an input chunk.
|
||
//
|
||
// Call `push(newChunk)` to pass along transformed output
|
||
// to the readable side. You may call 'push' zero or more times.
|
||
//
|
||
// Call `cb(err)` when you are done with this chunk. If you pass
|
||
// an error, then that'll put the hurt on the whole operation. If you
|
||
// never call cb(), then you'll never get another chunk.
|
||
Transform.prototype._transform = function (chunk, encoding, cb) {
|
||
throw new Error('Not implemented');
|
||
};
|
||
|
||
Transform.prototype._write = function (chunk, encoding, cb) {
|
||
var ts = this._transformState;
|
||
ts.writecb = cb;
|
||
ts.writechunk = chunk;
|
||
ts.writeencoding = encoding;
|
||
if (!ts.transforming) {
|
||
var rs = this._readableState;
|
||
if (ts.needTransform || rs.needReadable || rs.length < rs.highWaterMark) this._read(rs.highWaterMark);
|
||
}
|
||
};
|
||
|
||
// Doesn't matter what the args are here.
|
||
// _transform does all the work.
|
||
// That we got here means that the readable side wants more data.
|
||
Transform.prototype._read = function (n) {
|
||
var ts = this._transformState;
|
||
|
||
if (ts.writechunk !== null && ts.writecb && !ts.transforming) {
|
||
ts.transforming = true;
|
||
this._transform(ts.writechunk, ts.writeencoding, ts.afterTransform);
|
||
} else {
|
||
// mark that we need a transform, so that any data that comes in
|
||
// will get processed, now that we've asked for it.
|
||
ts.needTransform = true;
|
||
}
|
||
};
|
||
|
||
function done(stream, er) {
|
||
if (er) return stream.emit('error', er);
|
||
|
||
// if there's nothing in the write buffer, then that means
|
||
// that nothing more will ever be provided
|
||
var ws = stream._writableState;
|
||
var ts = stream._transformState;
|
||
|
||
if (ws.length) throw new Error('Calling transform done when ws.length != 0');
|
||
|
||
if (ts.transforming) throw new Error('Calling transform done when still transforming');
|
||
|
||
return stream.push(null);
|
||
}
|
||
|
||
inherits(PassThrough, Transform);
|
||
function PassThrough(options) {
|
||
if (!(this instanceof PassThrough)) return new PassThrough(options);
|
||
|
||
Transform.call(this, options);
|
||
}
|
||
|
||
PassThrough.prototype._transform = function (chunk, encoding, cb) {
|
||
cb(null, chunk);
|
||
};
|
||
|
||
inherits(Stream, EventEmitter);
|
||
Stream.Readable = Readable;
|
||
Stream.Writable = Writable;
|
||
Stream.Duplex = Duplex;
|
||
Stream.Transform = Transform;
|
||
Stream.PassThrough = PassThrough;
|
||
|
||
// Backwards-compat with node 0.4.x
|
||
Stream.Stream = Stream;
|
||
|
||
// old-style streams. Note that the pipe method (the only relevant
|
||
// part of this class) is overridden in the Readable class.
|
||
|
||
function Stream() {
|
||
EventEmitter.call(this);
|
||
}
|
||
|
||
Stream.prototype.pipe = function(dest, options) {
|
||
var source = this;
|
||
|
||
function ondata(chunk) {
|
||
if (dest.writable) {
|
||
if (false === dest.write(chunk) && source.pause) {
|
||
source.pause();
|
||
}
|
||
}
|
||
}
|
||
|
||
source.on('data', ondata);
|
||
|
||
function ondrain() {
|
||
if (source.readable && source.resume) {
|
||
source.resume();
|
||
}
|
||
}
|
||
|
||
dest.on('drain', ondrain);
|
||
|
||
// If the 'end' option is not supplied, dest.end() will be called when
|
||
// source gets the 'end' or 'close' events. Only dest.end() once.
|
||
if (!dest._isStdio && (!options || options.end !== false)) {
|
||
source.on('end', onend);
|
||
source.on('close', onclose);
|
||
}
|
||
|
||
var didOnEnd = false;
|
||
function onend() {
|
||
if (didOnEnd) return;
|
||
didOnEnd = true;
|
||
|
||
dest.end();
|
||
}
|
||
|
||
|
||
function onclose() {
|
||
if (didOnEnd) return;
|
||
didOnEnd = true;
|
||
|
||
if (typeof dest.destroy === 'function') dest.destroy();
|
||
}
|
||
|
||
// don't leave dangling pipes when there are errors.
|
||
function onerror(er) {
|
||
cleanup();
|
||
if (EventEmitter.listenerCount(this, 'error') === 0) {
|
||
throw er; // Unhandled stream error in pipe.
|
||
}
|
||
}
|
||
|
||
source.on('error', onerror);
|
||
dest.on('error', onerror);
|
||
|
||
// remove all the event listeners that were added.
|
||
function cleanup() {
|
||
source.removeListener('data', ondata);
|
||
dest.removeListener('drain', ondrain);
|
||
|
||
source.removeListener('end', onend);
|
||
source.removeListener('close', onclose);
|
||
|
||
source.removeListener('error', onerror);
|
||
dest.removeListener('error', onerror);
|
||
|
||
source.removeListener('end', cleanup);
|
||
source.removeListener('close', cleanup);
|
||
|
||
dest.removeListener('close', cleanup);
|
||
}
|
||
|
||
source.on('end', cleanup);
|
||
source.on('close', cleanup);
|
||
|
||
dest.on('close', cleanup);
|
||
|
||
dest.emit('pipe', source);
|
||
|
||
// Allow for unix-like usage: A.pipe(B).pipe(C)
|
||
return dest;
|
||
};
|
||
|
||
var _polyfillNode_stream = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
Duplex: Duplex,
|
||
PassThrough: PassThrough,
|
||
Readable: Readable,
|
||
Stream: Stream,
|
||
Transform: Transform,
|
||
Writable: Writable,
|
||
default: Stream
|
||
});
|
||
|
||
var require$$0 = /*@__PURE__*/getAugmentedNamespace(_polyfillNode_stream);
|
||
|
||
var require$$1 = /*@__PURE__*/getAugmentedNamespace(_polyfillNode_string_decoder);
|
||
|
||
var hasRequiredSax;
|
||
|
||
function requireSax () {
|
||
if (hasRequiredSax) return sax;
|
||
hasRequiredSax = 1;
|
||
(function (exports) {
|
||
(function (sax) { // wrapper for non-node envs
|
||
sax.parser = function (strict, opt) { return new SAXParser(strict, opt) };
|
||
sax.SAXParser = SAXParser;
|
||
sax.SAXStream = SAXStream;
|
||
sax.createStream = createStream;
|
||
|
||
// When we pass the MAX_BUFFER_LENGTH position, start checking for buffer overruns.
|
||
// When we check, schedule the next check for MAX_BUFFER_LENGTH - (max(buffer lengths)),
|
||
// since that's the earliest that a buffer overrun could occur. This way, checks are
|
||
// as rare as required, but as often as necessary to ensure never crossing this bound.
|
||
// Furthermore, buffers are only tested at most once per write(), so passing a very
|
||
// large string into write() might have undesirable effects, but this is manageable by
|
||
// the caller, so it is assumed to be safe. Thus, a call to write() may, in the extreme
|
||
// edge case, result in creating at most one complete copy of the string passed in.
|
||
// Set to Infinity to have unlimited buffers.
|
||
sax.MAX_BUFFER_LENGTH = 64 * 1024;
|
||
|
||
var buffers = [
|
||
'comment', 'sgmlDecl', 'textNode', 'tagName', 'doctype',
|
||
'procInstName', 'procInstBody', 'entity', 'attribName',
|
||
'attribValue', 'cdata', 'script'
|
||
];
|
||
|
||
sax.EVENTS = [
|
||
'text',
|
||
'processinginstruction',
|
||
'sgmldeclaration',
|
||
'doctype',
|
||
'comment',
|
||
'opentagstart',
|
||
'attribute',
|
||
'opentag',
|
||
'closetag',
|
||
'opencdata',
|
||
'cdata',
|
||
'closecdata',
|
||
'error',
|
||
'end',
|
||
'ready',
|
||
'script',
|
||
'opennamespace',
|
||
'closenamespace'
|
||
];
|
||
|
||
function SAXParser (strict, opt) {
|
||
if (!(this instanceof SAXParser)) {
|
||
return new SAXParser(strict, opt)
|
||
}
|
||
|
||
var parser = this;
|
||
clearBuffers(parser);
|
||
parser.q = parser.c = '';
|
||
parser.bufferCheckPosition = sax.MAX_BUFFER_LENGTH;
|
||
parser.opt = opt || {};
|
||
parser.opt.lowercase = parser.opt.lowercase || parser.opt.lowercasetags;
|
||
parser.looseCase = parser.opt.lowercase ? 'toLowerCase' : 'toUpperCase';
|
||
parser.tags = [];
|
||
parser.closed = parser.closedRoot = parser.sawRoot = false;
|
||
parser.tag = parser.error = null;
|
||
parser.strict = !!strict;
|
||
parser.noscript = !!(strict || parser.opt.noscript);
|
||
parser.state = S.BEGIN;
|
||
parser.strictEntities = parser.opt.strictEntities;
|
||
parser.ENTITIES = parser.strictEntities ? Object.create(sax.XML_ENTITIES) : Object.create(sax.ENTITIES);
|
||
parser.attribList = [];
|
||
|
||
// namespaces form a prototype chain.
|
||
// it always points at the current tag,
|
||
// which protos to its parent tag.
|
||
if (parser.opt.xmlns) {
|
||
parser.ns = Object.create(rootNS);
|
||
}
|
||
|
||
// mostly just for error reporting
|
||
parser.trackPosition = parser.opt.position !== false;
|
||
if (parser.trackPosition) {
|
||
parser.position = parser.line = parser.column = 0;
|
||
}
|
||
emit(parser, 'onready');
|
||
}
|
||
|
||
if (!Object.create) {
|
||
Object.create = function (o) {
|
||
function F () {}
|
||
F.prototype = o;
|
||
var newf = new F();
|
||
return newf
|
||
};
|
||
}
|
||
|
||
if (!Object.keys) {
|
||
Object.keys = function (o) {
|
||
var a = [];
|
||
for (var i in o) if (o.hasOwnProperty(i)) a.push(i);
|
||
return a
|
||
};
|
||
}
|
||
|
||
function checkBufferLength (parser) {
|
||
var maxAllowed = Math.max(sax.MAX_BUFFER_LENGTH, 10);
|
||
var maxActual = 0;
|
||
for (var i = 0, l = buffers.length; i < l; i++) {
|
||
var len = parser[buffers[i]].length;
|
||
if (len > maxAllowed) {
|
||
// Text/cdata nodes can get big, and since they're buffered,
|
||
// we can get here under normal conditions.
|
||
// Avoid issues by emitting the text node now,
|
||
// so at least it won't get any bigger.
|
||
switch (buffers[i]) {
|
||
case 'textNode':
|
||
closeText(parser);
|
||
break
|
||
|
||
case 'cdata':
|
||
emitNode(parser, 'oncdata', parser.cdata);
|
||
parser.cdata = '';
|
||
break
|
||
|
||
case 'script':
|
||
emitNode(parser, 'onscript', parser.script);
|
||
parser.script = '';
|
||
break
|
||
|
||
default:
|
||
error(parser, 'Max buffer length exceeded: ' + buffers[i]);
|
||
}
|
||
}
|
||
maxActual = Math.max(maxActual, len);
|
||
}
|
||
// schedule the next check for the earliest possible buffer overrun.
|
||
var m = sax.MAX_BUFFER_LENGTH - maxActual;
|
||
parser.bufferCheckPosition = m + parser.position;
|
||
}
|
||
|
||
function clearBuffers (parser) {
|
||
for (var i = 0, l = buffers.length; i < l; i++) {
|
||
parser[buffers[i]] = '';
|
||
}
|
||
}
|
||
|
||
function flushBuffers (parser) {
|
||
closeText(parser);
|
||
if (parser.cdata !== '') {
|
||
emitNode(parser, 'oncdata', parser.cdata);
|
||
parser.cdata = '';
|
||
}
|
||
if (parser.script !== '') {
|
||
emitNode(parser, 'onscript', parser.script);
|
||
parser.script = '';
|
||
}
|
||
}
|
||
|
||
SAXParser.prototype = {
|
||
end: function () { end(this); },
|
||
write: write,
|
||
resume: function () { this.error = null; return this },
|
||
close: function () { return this.write(null) },
|
||
flush: function () { flushBuffers(this); }
|
||
};
|
||
|
||
var Stream;
|
||
try {
|
||
Stream = require$$0.Stream;
|
||
} catch (ex) {
|
||
Stream = function () {};
|
||
}
|
||
|
||
var streamWraps = sax.EVENTS.filter(function (ev) {
|
||
return ev !== 'error' && ev !== 'end'
|
||
});
|
||
|
||
function createStream (strict, opt) {
|
||
return new SAXStream(strict, opt)
|
||
}
|
||
|
||
function SAXStream (strict, opt) {
|
||
if (!(this instanceof SAXStream)) {
|
||
return new SAXStream(strict, opt)
|
||
}
|
||
|
||
Stream.apply(this);
|
||
|
||
this._parser = new SAXParser(strict, opt);
|
||
this.writable = true;
|
||
this.readable = true;
|
||
|
||
var me = this;
|
||
|
||
this._parser.onend = function () {
|
||
me.emit('end');
|
||
};
|
||
|
||
this._parser.onerror = function (er) {
|
||
me.emit('error', er);
|
||
|
||
// if didn't throw, then means error was handled.
|
||
// go ahead and clear error, so we can write again.
|
||
me._parser.error = null;
|
||
};
|
||
|
||
this._decoder = null;
|
||
|
||
streamWraps.forEach(function (ev) {
|
||
Object.defineProperty(me, 'on' + ev, {
|
||
get: function () {
|
||
return me._parser['on' + ev]
|
||
},
|
||
set: function (h) {
|
||
if (!h) {
|
||
me.removeAllListeners(ev);
|
||
me._parser['on' + ev] = h;
|
||
return h
|
||
}
|
||
me.on(ev, h);
|
||
},
|
||
enumerable: true,
|
||
configurable: false
|
||
});
|
||
});
|
||
}
|
||
|
||
SAXStream.prototype = Object.create(Stream.prototype, {
|
||
constructor: {
|
||
value: SAXStream
|
||
}
|
||
});
|
||
|
||
SAXStream.prototype.write = function (data) {
|
||
if (typeof Buffer === 'function' &&
|
||
typeof Buffer.isBuffer === 'function' &&
|
||
Buffer.isBuffer(data)) {
|
||
if (!this._decoder) {
|
||
var SD = require$$1.StringDecoder;
|
||
this._decoder = new SD('utf8');
|
||
}
|
||
data = this._decoder.write(data);
|
||
}
|
||
|
||
this._parser.write(data.toString());
|
||
this.emit('data', data);
|
||
return true
|
||
};
|
||
|
||
SAXStream.prototype.end = function (chunk) {
|
||
if (chunk && chunk.length) {
|
||
this.write(chunk);
|
||
}
|
||
this._parser.end();
|
||
return true
|
||
};
|
||
|
||
SAXStream.prototype.on = function (ev, handler) {
|
||
var me = this;
|
||
if (!me._parser['on' + ev] && streamWraps.indexOf(ev) !== -1) {
|
||
me._parser['on' + ev] = function () {
|
||
var args = arguments.length === 1 ? [arguments[0]] : Array.apply(null, arguments);
|
||
args.splice(0, 0, ev);
|
||
me.emit.apply(me, args);
|
||
};
|
||
}
|
||
|
||
return Stream.prototype.on.call(me, ev, handler)
|
||
};
|
||
|
||
// this really needs to be replaced with character classes.
|
||
// XML allows all manner of ridiculous numbers and digits.
|
||
var CDATA = '[CDATA[';
|
||
var DOCTYPE = 'DOCTYPE';
|
||
var XML_NAMESPACE = 'http://www.w3.org/XML/1998/namespace';
|
||
var XMLNS_NAMESPACE = 'http://www.w3.org/2000/xmlns/';
|
||
var rootNS = { xml: XML_NAMESPACE, xmlns: XMLNS_NAMESPACE };
|
||
|
||
// http://www.w3.org/TR/REC-xml/#NT-NameStartChar
|
||
// This implementation works on strings, a single character at a time
|
||
// as such, it cannot ever support astral-plane characters (10000-EFFFF)
|
||
// without a significant breaking change to either this parser, or the
|
||
// JavaScript language. Implementation of an emoji-capable xml parser
|
||
// is left as an exercise for the reader.
|
||
var nameStart = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/;
|
||
|
||
var nameBody = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/;
|
||
|
||
var entityStart = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/;
|
||
var entityBody = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/;
|
||
|
||
function isWhitespace (c) {
|
||
return c === ' ' || c === '\n' || c === '\r' || c === '\t'
|
||
}
|
||
|
||
function isQuote (c) {
|
||
return c === '"' || c === '\''
|
||
}
|
||
|
||
function isAttribEnd (c) {
|
||
return c === '>' || isWhitespace(c)
|
||
}
|
||
|
||
function isMatch (regex, c) {
|
||
return regex.test(c)
|
||
}
|
||
|
||
function notMatch (regex, c) {
|
||
return !isMatch(regex, c)
|
||
}
|
||
|
||
var S = 0;
|
||
sax.STATE = {
|
||
BEGIN: S++, // leading byte order mark or whitespace
|
||
BEGIN_WHITESPACE: S++, // leading whitespace
|
||
TEXT: S++, // general stuff
|
||
TEXT_ENTITY: S++, // & and such.
|
||
OPEN_WAKA: S++, // <
|
||
SGML_DECL: S++, // <!BLARG
|
||
SGML_DECL_QUOTED: S++, // <!BLARG foo "bar
|
||
DOCTYPE: S++, // <!DOCTYPE
|
||
DOCTYPE_QUOTED: S++, // <!DOCTYPE "//blah
|
||
DOCTYPE_DTD: S++, // <!DOCTYPE "//blah" [ ...
|
||
DOCTYPE_DTD_QUOTED: S++, // <!DOCTYPE "//blah" [ "foo
|
||
COMMENT_STARTING: S++, // <!-
|
||
COMMENT: S++, // <!--
|
||
COMMENT_ENDING: S++, // <!-- blah -
|
||
COMMENT_ENDED: S++, // <!-- blah --
|
||
CDATA: S++, // <![CDATA[ something
|
||
CDATA_ENDING: S++, // ]
|
||
CDATA_ENDING_2: S++, // ]]
|
||
PROC_INST: S++, // <?hi
|
||
PROC_INST_BODY: S++, // <?hi there
|
||
PROC_INST_ENDING: S++, // <?hi "there" ?
|
||
OPEN_TAG: S++, // <strong
|
||
OPEN_TAG_SLASH: S++, // <strong /
|
||
ATTRIB: S++, // <a
|
||
ATTRIB_NAME: S++, // <a foo
|
||
ATTRIB_NAME_SAW_WHITE: S++, // <a foo _
|
||
ATTRIB_VALUE: S++, // <a foo=
|
||
ATTRIB_VALUE_QUOTED: S++, // <a foo="bar
|
||
ATTRIB_VALUE_CLOSED: S++, // <a foo="bar"
|
||
ATTRIB_VALUE_UNQUOTED: S++, // <a foo=bar
|
||
ATTRIB_VALUE_ENTITY_Q: S++, // <foo bar="""
|
||
ATTRIB_VALUE_ENTITY_U: S++, // <foo bar="
|
||
CLOSE_TAG: S++, // </a
|
||
CLOSE_TAG_SAW_WHITE: S++, // </a >
|
||
SCRIPT: S++, // <script> ...
|
||
SCRIPT_ENDING: S++ // <script> ... <
|
||
};
|
||
|
||
sax.XML_ENTITIES = {
|
||
'amp': '&',
|
||
'gt': '>',
|
||
'lt': '<',
|
||
'quot': '"',
|
||
'apos': "'"
|
||
};
|
||
|
||
sax.ENTITIES = {
|
||
'amp': '&',
|
||
'gt': '>',
|
||
'lt': '<',
|
||
'quot': '"',
|
||
'apos': "'",
|
||
'AElig': 198,
|
||
'Aacute': 193,
|
||
'Acirc': 194,
|
||
'Agrave': 192,
|
||
'Aring': 197,
|
||
'Atilde': 195,
|
||
'Auml': 196,
|
||
'Ccedil': 199,
|
||
'ETH': 208,
|
||
'Eacute': 201,
|
||
'Ecirc': 202,
|
||
'Egrave': 200,
|
||
'Euml': 203,
|
||
'Iacute': 205,
|
||
'Icirc': 206,
|
||
'Igrave': 204,
|
||
'Iuml': 207,
|
||
'Ntilde': 209,
|
||
'Oacute': 211,
|
||
'Ocirc': 212,
|
||
'Ograve': 210,
|
||
'Oslash': 216,
|
||
'Otilde': 213,
|
||
'Ouml': 214,
|
||
'THORN': 222,
|
||
'Uacute': 218,
|
||
'Ucirc': 219,
|
||
'Ugrave': 217,
|
||
'Uuml': 220,
|
||
'Yacute': 221,
|
||
'aacute': 225,
|
||
'acirc': 226,
|
||
'aelig': 230,
|
||
'agrave': 224,
|
||
'aring': 229,
|
||
'atilde': 227,
|
||
'auml': 228,
|
||
'ccedil': 231,
|
||
'eacute': 233,
|
||
'ecirc': 234,
|
||
'egrave': 232,
|
||
'eth': 240,
|
||
'euml': 235,
|
||
'iacute': 237,
|
||
'icirc': 238,
|
||
'igrave': 236,
|
||
'iuml': 239,
|
||
'ntilde': 241,
|
||
'oacute': 243,
|
||
'ocirc': 244,
|
||
'ograve': 242,
|
||
'oslash': 248,
|
||
'otilde': 245,
|
||
'ouml': 246,
|
||
'szlig': 223,
|
||
'thorn': 254,
|
||
'uacute': 250,
|
||
'ucirc': 251,
|
||
'ugrave': 249,
|
||
'uuml': 252,
|
||
'yacute': 253,
|
||
'yuml': 255,
|
||
'copy': 169,
|
||
'reg': 174,
|
||
'nbsp': 160,
|
||
'iexcl': 161,
|
||
'cent': 162,
|
||
'pound': 163,
|
||
'curren': 164,
|
||
'yen': 165,
|
||
'brvbar': 166,
|
||
'sect': 167,
|
||
'uml': 168,
|
||
'ordf': 170,
|
||
'laquo': 171,
|
||
'not': 172,
|
||
'shy': 173,
|
||
'macr': 175,
|
||
'deg': 176,
|
||
'plusmn': 177,
|
||
'sup1': 185,
|
||
'sup2': 178,
|
||
'sup3': 179,
|
||
'acute': 180,
|
||
'micro': 181,
|
||
'para': 182,
|
||
'middot': 183,
|
||
'cedil': 184,
|
||
'ordm': 186,
|
||
'raquo': 187,
|
||
'frac14': 188,
|
||
'frac12': 189,
|
||
'frac34': 190,
|
||
'iquest': 191,
|
||
'times': 215,
|
||
'divide': 247,
|
||
'OElig': 338,
|
||
'oelig': 339,
|
||
'Scaron': 352,
|
||
'scaron': 353,
|
||
'Yuml': 376,
|
||
'fnof': 402,
|
||
'circ': 710,
|
||
'tilde': 732,
|
||
'Alpha': 913,
|
||
'Beta': 914,
|
||
'Gamma': 915,
|
||
'Delta': 916,
|
||
'Epsilon': 917,
|
||
'Zeta': 918,
|
||
'Eta': 919,
|
||
'Theta': 920,
|
||
'Iota': 921,
|
||
'Kappa': 922,
|
||
'Lambda': 923,
|
||
'Mu': 924,
|
||
'Nu': 925,
|
||
'Xi': 926,
|
||
'Omicron': 927,
|
||
'Pi': 928,
|
||
'Rho': 929,
|
||
'Sigma': 931,
|
||
'Tau': 932,
|
||
'Upsilon': 933,
|
||
'Phi': 934,
|
||
'Chi': 935,
|
||
'Psi': 936,
|
||
'Omega': 937,
|
||
'alpha': 945,
|
||
'beta': 946,
|
||
'gamma': 947,
|
||
'delta': 948,
|
||
'epsilon': 949,
|
||
'zeta': 950,
|
||
'eta': 951,
|
||
'theta': 952,
|
||
'iota': 953,
|
||
'kappa': 954,
|
||
'lambda': 955,
|
||
'mu': 956,
|
||
'nu': 957,
|
||
'xi': 958,
|
||
'omicron': 959,
|
||
'pi': 960,
|
||
'rho': 961,
|
||
'sigmaf': 962,
|
||
'sigma': 963,
|
||
'tau': 964,
|
||
'upsilon': 965,
|
||
'phi': 966,
|
||
'chi': 967,
|
||
'psi': 968,
|
||
'omega': 969,
|
||
'thetasym': 977,
|
||
'upsih': 978,
|
||
'piv': 982,
|
||
'ensp': 8194,
|
||
'emsp': 8195,
|
||
'thinsp': 8201,
|
||
'zwnj': 8204,
|
||
'zwj': 8205,
|
||
'lrm': 8206,
|
||
'rlm': 8207,
|
||
'ndash': 8211,
|
||
'mdash': 8212,
|
||
'lsquo': 8216,
|
||
'rsquo': 8217,
|
||
'sbquo': 8218,
|
||
'ldquo': 8220,
|
||
'rdquo': 8221,
|
||
'bdquo': 8222,
|
||
'dagger': 8224,
|
||
'Dagger': 8225,
|
||
'bull': 8226,
|
||
'hellip': 8230,
|
||
'permil': 8240,
|
||
'prime': 8242,
|
||
'Prime': 8243,
|
||
'lsaquo': 8249,
|
||
'rsaquo': 8250,
|
||
'oline': 8254,
|
||
'frasl': 8260,
|
||
'euro': 8364,
|
||
'image': 8465,
|
||
'weierp': 8472,
|
||
'real': 8476,
|
||
'trade': 8482,
|
||
'alefsym': 8501,
|
||
'larr': 8592,
|
||
'uarr': 8593,
|
||
'rarr': 8594,
|
||
'darr': 8595,
|
||
'harr': 8596,
|
||
'crarr': 8629,
|
||
'lArr': 8656,
|
||
'uArr': 8657,
|
||
'rArr': 8658,
|
||
'dArr': 8659,
|
||
'hArr': 8660,
|
||
'forall': 8704,
|
||
'part': 8706,
|
||
'exist': 8707,
|
||
'empty': 8709,
|
||
'nabla': 8711,
|
||
'isin': 8712,
|
||
'notin': 8713,
|
||
'ni': 8715,
|
||
'prod': 8719,
|
||
'sum': 8721,
|
||
'minus': 8722,
|
||
'lowast': 8727,
|
||
'radic': 8730,
|
||
'prop': 8733,
|
||
'infin': 8734,
|
||
'ang': 8736,
|
||
'and': 8743,
|
||
'or': 8744,
|
||
'cap': 8745,
|
||
'cup': 8746,
|
||
'int': 8747,
|
||
'there4': 8756,
|
||
'sim': 8764,
|
||
'cong': 8773,
|
||
'asymp': 8776,
|
||
'ne': 8800,
|
||
'equiv': 8801,
|
||
'le': 8804,
|
||
'ge': 8805,
|
||
'sub': 8834,
|
||
'sup': 8835,
|
||
'nsub': 8836,
|
||
'sube': 8838,
|
||
'supe': 8839,
|
||
'oplus': 8853,
|
||
'otimes': 8855,
|
||
'perp': 8869,
|
||
'sdot': 8901,
|
||
'lceil': 8968,
|
||
'rceil': 8969,
|
||
'lfloor': 8970,
|
||
'rfloor': 8971,
|
||
'lang': 9001,
|
||
'rang': 9002,
|
||
'loz': 9674,
|
||
'spades': 9824,
|
||
'clubs': 9827,
|
||
'hearts': 9829,
|
||
'diams': 9830
|
||
};
|
||
|
||
Object.keys(sax.ENTITIES).forEach(function (key) {
|
||
var e = sax.ENTITIES[key];
|
||
var s = typeof e === 'number' ? String.fromCharCode(e) : e;
|
||
sax.ENTITIES[key] = s;
|
||
});
|
||
|
||
for (var s in sax.STATE) {
|
||
sax.STATE[sax.STATE[s]] = s;
|
||
}
|
||
|
||
// shorthand
|
||
S = sax.STATE;
|
||
|
||
function emit (parser, event, data) {
|
||
parser[event] && parser[event](data);
|
||
}
|
||
|
||
function emitNode (parser, nodeType, data) {
|
||
if (parser.textNode) closeText(parser);
|
||
emit(parser, nodeType, data);
|
||
}
|
||
|
||
function closeText (parser) {
|
||
parser.textNode = textopts(parser.opt, parser.textNode);
|
||
if (parser.textNode) emit(parser, 'ontext', parser.textNode);
|
||
parser.textNode = '';
|
||
}
|
||
|
||
function textopts (opt, text) {
|
||
if (opt.trim) text = text.trim();
|
||
if (opt.normalize) text = text.replace(/\s+/g, ' ');
|
||
return text
|
||
}
|
||
|
||
function error (parser, er) {
|
||
closeText(parser);
|
||
if (parser.trackPosition) {
|
||
er += '\nLine: ' + parser.line +
|
||
'\nColumn: ' + parser.column +
|
||
'\nChar: ' + parser.c;
|
||
}
|
||
er = new Error(er);
|
||
parser.error = er;
|
||
emit(parser, 'onerror', er);
|
||
return parser
|
||
}
|
||
|
||
function end (parser) {
|
||
if (parser.sawRoot && !parser.closedRoot) strictFail(parser, 'Unclosed root tag');
|
||
if ((parser.state !== S.BEGIN) &&
|
||
(parser.state !== S.BEGIN_WHITESPACE) &&
|
||
(parser.state !== S.TEXT)) {
|
||
error(parser, 'Unexpected end');
|
||
}
|
||
closeText(parser);
|
||
parser.c = '';
|
||
parser.closed = true;
|
||
emit(parser, 'onend');
|
||
SAXParser.call(parser, parser.strict, parser.opt);
|
||
return parser
|
||
}
|
||
|
||
function strictFail (parser, message) {
|
||
if (typeof parser !== 'object' || !(parser instanceof SAXParser)) {
|
||
throw new Error('bad call to strictFail')
|
||
}
|
||
if (parser.strict) {
|
||
error(parser, message);
|
||
}
|
||
}
|
||
|
||
function newTag (parser) {
|
||
if (!parser.strict) parser.tagName = parser.tagName[parser.looseCase]();
|
||
var parent = parser.tags[parser.tags.length - 1] || parser;
|
||
var tag = parser.tag = { name: parser.tagName, attributes: {} };
|
||
|
||
// will be overridden if tag contails an xmlns="foo" or xmlns:foo="bar"
|
||
if (parser.opt.xmlns) {
|
||
tag.ns = parent.ns;
|
||
}
|
||
parser.attribList.length = 0;
|
||
emitNode(parser, 'onopentagstart', tag);
|
||
}
|
||
|
||
function qname (name, attribute) {
|
||
var i = name.indexOf(':');
|
||
var qualName = i < 0 ? [ '', name ] : name.split(':');
|
||
var prefix = qualName[0];
|
||
var local = qualName[1];
|
||
|
||
// <x "xmlns"="http://foo">
|
||
if (attribute && name === 'xmlns') {
|
||
prefix = 'xmlns';
|
||
local = '';
|
||
}
|
||
|
||
return { prefix: prefix, local: local }
|
||
}
|
||
|
||
function attrib (parser) {
|
||
if (!parser.strict) {
|
||
parser.attribName = parser.attribName[parser.looseCase]();
|
||
}
|
||
|
||
if (parser.attribList.indexOf(parser.attribName) !== -1 ||
|
||
parser.tag.attributes.hasOwnProperty(parser.attribName)) {
|
||
parser.attribName = parser.attribValue = '';
|
||
return
|
||
}
|
||
|
||
if (parser.opt.xmlns) {
|
||
var qn = qname(parser.attribName, true);
|
||
var prefix = qn.prefix;
|
||
var local = qn.local;
|
||
|
||
if (prefix === 'xmlns') {
|
||
// namespace binding attribute. push the binding into scope
|
||
if (local === 'xml' && parser.attribValue !== XML_NAMESPACE) {
|
||
strictFail(parser,
|
||
'xml: prefix must be bound to ' + XML_NAMESPACE + '\n' +
|
||
'Actual: ' + parser.attribValue);
|
||
} else if (local === 'xmlns' && parser.attribValue !== XMLNS_NAMESPACE) {
|
||
strictFail(parser,
|
||
'xmlns: prefix must be bound to ' + XMLNS_NAMESPACE + '\n' +
|
||
'Actual: ' + parser.attribValue);
|
||
} else {
|
||
var tag = parser.tag;
|
||
var parent = parser.tags[parser.tags.length - 1] || parser;
|
||
if (tag.ns === parent.ns) {
|
||
tag.ns = Object.create(parent.ns);
|
||
}
|
||
tag.ns[local] = parser.attribValue;
|
||
}
|
||
}
|
||
|
||
// defer onattribute events until all attributes have been seen
|
||
// so any new bindings can take effect. preserve attribute order
|
||
// so deferred events can be emitted in document order
|
||
parser.attribList.push([parser.attribName, parser.attribValue]);
|
||
} else {
|
||
// in non-xmlns mode, we can emit the event right away
|
||
parser.tag.attributes[parser.attribName] = parser.attribValue;
|
||
emitNode(parser, 'onattribute', {
|
||
name: parser.attribName,
|
||
value: parser.attribValue
|
||
});
|
||
}
|
||
|
||
parser.attribName = parser.attribValue = '';
|
||
}
|
||
|
||
function openTag (parser, selfClosing) {
|
||
if (parser.opt.xmlns) {
|
||
// emit namespace binding events
|
||
var tag = parser.tag;
|
||
|
||
// add namespace info to tag
|
||
var qn = qname(parser.tagName);
|
||
tag.prefix = qn.prefix;
|
||
tag.local = qn.local;
|
||
tag.uri = tag.ns[qn.prefix] || '';
|
||
|
||
if (tag.prefix && !tag.uri) {
|
||
strictFail(parser, 'Unbound namespace prefix: ' +
|
||
JSON.stringify(parser.tagName));
|
||
tag.uri = qn.prefix;
|
||
}
|
||
|
||
var parent = parser.tags[parser.tags.length - 1] || parser;
|
||
if (tag.ns && parent.ns !== tag.ns) {
|
||
Object.keys(tag.ns).forEach(function (p) {
|
||
emitNode(parser, 'onopennamespace', {
|
||
prefix: p,
|
||
uri: tag.ns[p]
|
||
});
|
||
});
|
||
}
|
||
|
||
// handle deferred onattribute events
|
||
// Note: do not apply default ns to attributes:
|
||
// http://www.w3.org/TR/REC-xml-names/#defaulting
|
||
for (var i = 0, l = parser.attribList.length; i < l; i++) {
|
||
var nv = parser.attribList[i];
|
||
var name = nv[0];
|
||
var value = nv[1];
|
||
var qualName = qname(name, true);
|
||
var prefix = qualName.prefix;
|
||
var local = qualName.local;
|
||
var uri = prefix === '' ? '' : (tag.ns[prefix] || '');
|
||
var a = {
|
||
name: name,
|
||
value: value,
|
||
prefix: prefix,
|
||
local: local,
|
||
uri: uri
|
||
};
|
||
|
||
// if there's any attributes with an undefined namespace,
|
||
// then fail on them now.
|
||
if (prefix && prefix !== 'xmlns' && !uri) {
|
||
strictFail(parser, 'Unbound namespace prefix: ' +
|
||
JSON.stringify(prefix));
|
||
a.uri = prefix;
|
||
}
|
||
parser.tag.attributes[name] = a;
|
||
emitNode(parser, 'onattribute', a);
|
||
}
|
||
parser.attribList.length = 0;
|
||
}
|
||
|
||
parser.tag.isSelfClosing = !!selfClosing;
|
||
|
||
// process the tag
|
||
parser.sawRoot = true;
|
||
parser.tags.push(parser.tag);
|
||
emitNode(parser, 'onopentag', parser.tag);
|
||
if (!selfClosing) {
|
||
// special case for <script> in non-strict mode.
|
||
if (!parser.noscript && parser.tagName.toLowerCase() === 'script') {
|
||
parser.state = S.SCRIPT;
|
||
} else {
|
||
parser.state = S.TEXT;
|
||
}
|
||
parser.tag = null;
|
||
parser.tagName = '';
|
||
}
|
||
parser.attribName = parser.attribValue = '';
|
||
parser.attribList.length = 0;
|
||
}
|
||
|
||
function closeTag (parser) {
|
||
if (!parser.tagName) {
|
||
strictFail(parser, 'Weird empty close tag.');
|
||
parser.textNode += '</>';
|
||
parser.state = S.TEXT;
|
||
return
|
||
}
|
||
|
||
if (parser.script) {
|
||
if (parser.tagName !== 'script') {
|
||
parser.script += '</' + parser.tagName + '>';
|
||
parser.tagName = '';
|
||
parser.state = S.SCRIPT;
|
||
return
|
||
}
|
||
emitNode(parser, 'onscript', parser.script);
|
||
parser.script = '';
|
||
}
|
||
|
||
// first make sure that the closing tag actually exists.
|
||
// <a><b></c></b></a> will close everything, otherwise.
|
||
var t = parser.tags.length;
|
||
var tagName = parser.tagName;
|
||
if (!parser.strict) {
|
||
tagName = tagName[parser.looseCase]();
|
||
}
|
||
var closeTo = tagName;
|
||
while (t--) {
|
||
var close = parser.tags[t];
|
||
if (close.name !== closeTo) {
|
||
// fail the first time in strict mode
|
||
strictFail(parser, 'Unexpected close tag');
|
||
} else {
|
||
break
|
||
}
|
||
}
|
||
|
||
// didn't find it. we already failed for strict, so just abort.
|
||
if (t < 0) {
|
||
strictFail(parser, 'Unmatched closing tag: ' + parser.tagName);
|
||
parser.textNode += '</' + parser.tagName + '>';
|
||
parser.state = S.TEXT;
|
||
return
|
||
}
|
||
parser.tagName = tagName;
|
||
var s = parser.tags.length;
|
||
while (s-- > t) {
|
||
var tag = parser.tag = parser.tags.pop();
|
||
parser.tagName = parser.tag.name;
|
||
emitNode(parser, 'onclosetag', parser.tagName);
|
||
|
||
var x = {};
|
||
for (var i in tag.ns) {
|
||
x[i] = tag.ns[i];
|
||
}
|
||
|
||
var parent = parser.tags[parser.tags.length - 1] || parser;
|
||
if (parser.opt.xmlns && tag.ns !== parent.ns) {
|
||
// remove namespace bindings introduced by tag
|
||
Object.keys(tag.ns).forEach(function (p) {
|
||
var n = tag.ns[p];
|
||
emitNode(parser, 'onclosenamespace', { prefix: p, uri: n });
|
||
});
|
||
}
|
||
}
|
||
if (t === 0) parser.closedRoot = true;
|
||
parser.tagName = parser.attribValue = parser.attribName = '';
|
||
parser.attribList.length = 0;
|
||
parser.state = S.TEXT;
|
||
}
|
||
|
||
function parseEntity (parser) {
|
||
var entity = parser.entity;
|
||
var entityLC = entity.toLowerCase();
|
||
var num;
|
||
var numStr = '';
|
||
|
||
if (parser.ENTITIES[entity]) {
|
||
return parser.ENTITIES[entity]
|
||
}
|
||
if (parser.ENTITIES[entityLC]) {
|
||
return parser.ENTITIES[entityLC]
|
||
}
|
||
entity = entityLC;
|
||
if (entity.charAt(0) === '#') {
|
||
if (entity.charAt(1) === 'x') {
|
||
entity = entity.slice(2);
|
||
num = parseInt(entity, 16);
|
||
numStr = num.toString(16);
|
||
} else {
|
||
entity = entity.slice(1);
|
||
num = parseInt(entity, 10);
|
||
numStr = num.toString(10);
|
||
}
|
||
}
|
||
entity = entity.replace(/^0+/, '');
|
||
if (isNaN(num) || numStr.toLowerCase() !== entity) {
|
||
strictFail(parser, 'Invalid character entity');
|
||
return '&' + parser.entity + ';'
|
||
}
|
||
|
||
return String.fromCodePoint(num)
|
||
}
|
||
|
||
function beginWhiteSpace (parser, c) {
|
||
if (c === '<') {
|
||
parser.state = S.OPEN_WAKA;
|
||
parser.startTagPosition = parser.position;
|
||
} else if (!isWhitespace(c)) {
|
||
// have to process this as a text node.
|
||
// weird, but happens.
|
||
strictFail(parser, 'Non-whitespace before first tag.');
|
||
parser.textNode = c;
|
||
parser.state = S.TEXT;
|
||
}
|
||
}
|
||
|
||
function charAt (chunk, i) {
|
||
var result = '';
|
||
if (i < chunk.length) {
|
||
result = chunk.charAt(i);
|
||
}
|
||
return result
|
||
}
|
||
|
||
function write (chunk) {
|
||
var parser = this;
|
||
if (this.error) {
|
||
throw this.error
|
||
}
|
||
if (parser.closed) {
|
||
return error(parser,
|
||
'Cannot write after close. Assign an onready handler.')
|
||
}
|
||
if (chunk === null) {
|
||
return end(parser)
|
||
}
|
||
if (typeof chunk === 'object') {
|
||
chunk = chunk.toString();
|
||
}
|
||
var i = 0;
|
||
var c = '';
|
||
while (true) {
|
||
c = charAt(chunk, i++);
|
||
parser.c = c;
|
||
|
||
if (!c) {
|
||
break
|
||
}
|
||
|
||
if (parser.trackPosition) {
|
||
parser.position++;
|
||
if (c === '\n') {
|
||
parser.line++;
|
||
parser.column = 0;
|
||
} else {
|
||
parser.column++;
|
||
}
|
||
}
|
||
|
||
switch (parser.state) {
|
||
case S.BEGIN:
|
||
parser.state = S.BEGIN_WHITESPACE;
|
||
if (c === '\uFEFF') {
|
||
continue
|
||
}
|
||
beginWhiteSpace(parser, c);
|
||
continue
|
||
|
||
case S.BEGIN_WHITESPACE:
|
||
beginWhiteSpace(parser, c);
|
||
continue
|
||
|
||
case S.TEXT:
|
||
if (parser.sawRoot && !parser.closedRoot) {
|
||
var starti = i - 1;
|
||
while (c && c !== '<' && c !== '&') {
|
||
c = charAt(chunk, i++);
|
||
if (c && parser.trackPosition) {
|
||
parser.position++;
|
||
if (c === '\n') {
|
||
parser.line++;
|
||
parser.column = 0;
|
||
} else {
|
||
parser.column++;
|
||
}
|
||
}
|
||
}
|
||
parser.textNode += chunk.substring(starti, i - 1);
|
||
}
|
||
if (c === '<' && !(parser.sawRoot && parser.closedRoot && !parser.strict)) {
|
||
parser.state = S.OPEN_WAKA;
|
||
parser.startTagPosition = parser.position;
|
||
} else {
|
||
if (!isWhitespace(c) && (!parser.sawRoot || parser.closedRoot)) {
|
||
strictFail(parser, 'Text data outside of root node.');
|
||
}
|
||
if (c === '&') {
|
||
parser.state = S.TEXT_ENTITY;
|
||
} else {
|
||
parser.textNode += c;
|
||
}
|
||
}
|
||
continue
|
||
|
||
case S.SCRIPT:
|
||
// only non-strict
|
||
if (c === '<') {
|
||
parser.state = S.SCRIPT_ENDING;
|
||
} else {
|
||
parser.script += c;
|
||
}
|
||
continue
|
||
|
||
case S.SCRIPT_ENDING:
|
||
if (c === '/') {
|
||
parser.state = S.CLOSE_TAG;
|
||
} else {
|
||
parser.script += '<' + c;
|
||
parser.state = S.SCRIPT;
|
||
}
|
||
continue
|
||
|
||
case S.OPEN_WAKA:
|
||
// either a /, ?, !, or text is coming next.
|
||
if (c === '!') {
|
||
parser.state = S.SGML_DECL;
|
||
parser.sgmlDecl = '';
|
||
} else if (isWhitespace(c)) ; else if (isMatch(nameStart, c)) {
|
||
parser.state = S.OPEN_TAG;
|
||
parser.tagName = c;
|
||
} else if (c === '/') {
|
||
parser.state = S.CLOSE_TAG;
|
||
parser.tagName = '';
|
||
} else if (c === '?') {
|
||
parser.state = S.PROC_INST;
|
||
parser.procInstName = parser.procInstBody = '';
|
||
} else {
|
||
strictFail(parser, 'Unencoded <');
|
||
// if there was some whitespace, then add that in.
|
||
if (parser.startTagPosition + 1 < parser.position) {
|
||
var pad = parser.position - parser.startTagPosition;
|
||
c = new Array(pad).join(' ') + c;
|
||
}
|
||
parser.textNode += '<' + c;
|
||
parser.state = S.TEXT;
|
||
}
|
||
continue
|
||
|
||
case S.SGML_DECL:
|
||
if ((parser.sgmlDecl + c).toUpperCase() === CDATA) {
|
||
emitNode(parser, 'onopencdata');
|
||
parser.state = S.CDATA;
|
||
parser.sgmlDecl = '';
|
||
parser.cdata = '';
|
||
} else if (parser.sgmlDecl + c === '--') {
|
||
parser.state = S.COMMENT;
|
||
parser.comment = '';
|
||
parser.sgmlDecl = '';
|
||
} else if ((parser.sgmlDecl + c).toUpperCase() === DOCTYPE) {
|
||
parser.state = S.DOCTYPE;
|
||
if (parser.doctype || parser.sawRoot) {
|
||
strictFail(parser,
|
||
'Inappropriately located doctype declaration');
|
||
}
|
||
parser.doctype = '';
|
||
parser.sgmlDecl = '';
|
||
} else if (c === '>') {
|
||
emitNode(parser, 'onsgmldeclaration', parser.sgmlDecl);
|
||
parser.sgmlDecl = '';
|
||
parser.state = S.TEXT;
|
||
} else if (isQuote(c)) {
|
||
parser.state = S.SGML_DECL_QUOTED;
|
||
parser.sgmlDecl += c;
|
||
} else {
|
||
parser.sgmlDecl += c;
|
||
}
|
||
continue
|
||
|
||
case S.SGML_DECL_QUOTED:
|
||
if (c === parser.q) {
|
||
parser.state = S.SGML_DECL;
|
||
parser.q = '';
|
||
}
|
||
parser.sgmlDecl += c;
|
||
continue
|
||
|
||
case S.DOCTYPE:
|
||
if (c === '>') {
|
||
parser.state = S.TEXT;
|
||
emitNode(parser, 'ondoctype', parser.doctype);
|
||
parser.doctype = true; // just remember that we saw it.
|
||
} else {
|
||
parser.doctype += c;
|
||
if (c === '[') {
|
||
parser.state = S.DOCTYPE_DTD;
|
||
} else if (isQuote(c)) {
|
||
parser.state = S.DOCTYPE_QUOTED;
|
||
parser.q = c;
|
||
}
|
||
}
|
||
continue
|
||
|
||
case S.DOCTYPE_QUOTED:
|
||
parser.doctype += c;
|
||
if (c === parser.q) {
|
||
parser.q = '';
|
||
parser.state = S.DOCTYPE;
|
||
}
|
||
continue
|
||
|
||
case S.DOCTYPE_DTD:
|
||
parser.doctype += c;
|
||
if (c === ']') {
|
||
parser.state = S.DOCTYPE;
|
||
} else if (isQuote(c)) {
|
||
parser.state = S.DOCTYPE_DTD_QUOTED;
|
||
parser.q = c;
|
||
}
|
||
continue
|
||
|
||
case S.DOCTYPE_DTD_QUOTED:
|
||
parser.doctype += c;
|
||
if (c === parser.q) {
|
||
parser.state = S.DOCTYPE_DTD;
|
||
parser.q = '';
|
||
}
|
||
continue
|
||
|
||
case S.COMMENT:
|
||
if (c === '-') {
|
||
parser.state = S.COMMENT_ENDING;
|
||
} else {
|
||
parser.comment += c;
|
||
}
|
||
continue
|
||
|
||
case S.COMMENT_ENDING:
|
||
if (c === '-') {
|
||
parser.state = S.COMMENT_ENDED;
|
||
parser.comment = textopts(parser.opt, parser.comment);
|
||
if (parser.comment) {
|
||
emitNode(parser, 'oncomment', parser.comment);
|
||
}
|
||
parser.comment = '';
|
||
} else {
|
||
parser.comment += '-' + c;
|
||
parser.state = S.COMMENT;
|
||
}
|
||
continue
|
||
|
||
case S.COMMENT_ENDED:
|
||
if (c !== '>') {
|
||
strictFail(parser, 'Malformed comment');
|
||
// allow <!-- blah -- bloo --> in non-strict mode,
|
||
// which is a comment of " blah -- bloo "
|
||
parser.comment += '--' + c;
|
||
parser.state = S.COMMENT;
|
||
} else {
|
||
parser.state = S.TEXT;
|
||
}
|
||
continue
|
||
|
||
case S.CDATA:
|
||
if (c === ']') {
|
||
parser.state = S.CDATA_ENDING;
|
||
} else {
|
||
parser.cdata += c;
|
||
}
|
||
continue
|
||
|
||
case S.CDATA_ENDING:
|
||
if (c === ']') {
|
||
parser.state = S.CDATA_ENDING_2;
|
||
} else {
|
||
parser.cdata += ']' + c;
|
||
parser.state = S.CDATA;
|
||
}
|
||
continue
|
||
|
||
case S.CDATA_ENDING_2:
|
||
if (c === '>') {
|
||
if (parser.cdata) {
|
||
emitNode(parser, 'oncdata', parser.cdata);
|
||
}
|
||
emitNode(parser, 'onclosecdata');
|
||
parser.cdata = '';
|
||
parser.state = S.TEXT;
|
||
} else if (c === ']') {
|
||
parser.cdata += ']';
|
||
} else {
|
||
parser.cdata += ']]' + c;
|
||
parser.state = S.CDATA;
|
||
}
|
||
continue
|
||
|
||
case S.PROC_INST:
|
||
if (c === '?') {
|
||
parser.state = S.PROC_INST_ENDING;
|
||
} else if (isWhitespace(c)) {
|
||
parser.state = S.PROC_INST_BODY;
|
||
} else {
|
||
parser.procInstName += c;
|
||
}
|
||
continue
|
||
|
||
case S.PROC_INST_BODY:
|
||
if (!parser.procInstBody && isWhitespace(c)) {
|
||
continue
|
||
} else if (c === '?') {
|
||
parser.state = S.PROC_INST_ENDING;
|
||
} else {
|
||
parser.procInstBody += c;
|
||
}
|
||
continue
|
||
|
||
case S.PROC_INST_ENDING:
|
||
if (c === '>') {
|
||
emitNode(parser, 'onprocessinginstruction', {
|
||
name: parser.procInstName,
|
||
body: parser.procInstBody
|
||
});
|
||
parser.procInstName = parser.procInstBody = '';
|
||
parser.state = S.TEXT;
|
||
} else {
|
||
parser.procInstBody += '?' + c;
|
||
parser.state = S.PROC_INST_BODY;
|
||
}
|
||
continue
|
||
|
||
case S.OPEN_TAG:
|
||
if (isMatch(nameBody, c)) {
|
||
parser.tagName += c;
|
||
} else {
|
||
newTag(parser);
|
||
if (c === '>') {
|
||
openTag(parser);
|
||
} else if (c === '/') {
|
||
parser.state = S.OPEN_TAG_SLASH;
|
||
} else {
|
||
if (!isWhitespace(c)) {
|
||
strictFail(parser, 'Invalid character in tag name');
|
||
}
|
||
parser.state = S.ATTRIB;
|
||
}
|
||
}
|
||
continue
|
||
|
||
case S.OPEN_TAG_SLASH:
|
||
if (c === '>') {
|
||
openTag(parser, true);
|
||
closeTag(parser);
|
||
} else {
|
||
strictFail(parser, 'Forward-slash in opening tag not followed by >');
|
||
parser.state = S.ATTRIB;
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB:
|
||
// haven't read the attribute name yet.
|
||
if (isWhitespace(c)) {
|
||
continue
|
||
} else if (c === '>') {
|
||
openTag(parser);
|
||
} else if (c === '/') {
|
||
parser.state = S.OPEN_TAG_SLASH;
|
||
} else if (isMatch(nameStart, c)) {
|
||
parser.attribName = c;
|
||
parser.attribValue = '';
|
||
parser.state = S.ATTRIB_NAME;
|
||
} else {
|
||
strictFail(parser, 'Invalid attribute name');
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB_NAME:
|
||
if (c === '=') {
|
||
parser.state = S.ATTRIB_VALUE;
|
||
} else if (c === '>') {
|
||
strictFail(parser, 'Attribute without value');
|
||
parser.attribValue = parser.attribName;
|
||
attrib(parser);
|
||
openTag(parser);
|
||
} else if (isWhitespace(c)) {
|
||
parser.state = S.ATTRIB_NAME_SAW_WHITE;
|
||
} else if (isMatch(nameBody, c)) {
|
||
parser.attribName += c;
|
||
} else {
|
||
strictFail(parser, 'Invalid attribute name');
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB_NAME_SAW_WHITE:
|
||
if (c === '=') {
|
||
parser.state = S.ATTRIB_VALUE;
|
||
} else if (isWhitespace(c)) {
|
||
continue
|
||
} else {
|
||
strictFail(parser, 'Attribute without value');
|
||
parser.tag.attributes[parser.attribName] = '';
|
||
parser.attribValue = '';
|
||
emitNode(parser, 'onattribute', {
|
||
name: parser.attribName,
|
||
value: ''
|
||
});
|
||
parser.attribName = '';
|
||
if (c === '>') {
|
||
openTag(parser);
|
||
} else if (isMatch(nameStart, c)) {
|
||
parser.attribName = c;
|
||
parser.state = S.ATTRIB_NAME;
|
||
} else {
|
||
strictFail(parser, 'Invalid attribute name');
|
||
parser.state = S.ATTRIB;
|
||
}
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB_VALUE:
|
||
if (isWhitespace(c)) {
|
||
continue
|
||
} else if (isQuote(c)) {
|
||
parser.q = c;
|
||
parser.state = S.ATTRIB_VALUE_QUOTED;
|
||
} else {
|
||
strictFail(parser, 'Unquoted attribute value');
|
||
parser.state = S.ATTRIB_VALUE_UNQUOTED;
|
||
parser.attribValue = c;
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB_VALUE_QUOTED:
|
||
if (c !== parser.q) {
|
||
if (c === '&') {
|
||
parser.state = S.ATTRIB_VALUE_ENTITY_Q;
|
||
} else {
|
||
parser.attribValue += c;
|
||
}
|
||
continue
|
||
}
|
||
attrib(parser);
|
||
parser.q = '';
|
||
parser.state = S.ATTRIB_VALUE_CLOSED;
|
||
continue
|
||
|
||
case S.ATTRIB_VALUE_CLOSED:
|
||
if (isWhitespace(c)) {
|
||
parser.state = S.ATTRIB;
|
||
} else if (c === '>') {
|
||
openTag(parser);
|
||
} else if (c === '/') {
|
||
parser.state = S.OPEN_TAG_SLASH;
|
||
} else if (isMatch(nameStart, c)) {
|
||
strictFail(parser, 'No whitespace between attributes');
|
||
parser.attribName = c;
|
||
parser.attribValue = '';
|
||
parser.state = S.ATTRIB_NAME;
|
||
} else {
|
||
strictFail(parser, 'Invalid attribute name');
|
||
}
|
||
continue
|
||
|
||
case S.ATTRIB_VALUE_UNQUOTED:
|
||
if (!isAttribEnd(c)) {
|
||
if (c === '&') {
|
||
parser.state = S.ATTRIB_VALUE_ENTITY_U;
|
||
} else {
|
||
parser.attribValue += c;
|
||
}
|
||
continue
|
||
}
|
||
attrib(parser);
|
||
if (c === '>') {
|
||
openTag(parser);
|
||
} else {
|
||
parser.state = S.ATTRIB;
|
||
}
|
||
continue
|
||
|
||
case S.CLOSE_TAG:
|
||
if (!parser.tagName) {
|
||
if (isWhitespace(c)) {
|
||
continue
|
||
} else if (notMatch(nameStart, c)) {
|
||
if (parser.script) {
|
||
parser.script += '</' + c;
|
||
parser.state = S.SCRIPT;
|
||
} else {
|
||
strictFail(parser, 'Invalid tagname in closing tag.');
|
||
}
|
||
} else {
|
||
parser.tagName = c;
|
||
}
|
||
} else if (c === '>') {
|
||
closeTag(parser);
|
||
} else if (isMatch(nameBody, c)) {
|
||
parser.tagName += c;
|
||
} else if (parser.script) {
|
||
parser.script += '</' + parser.tagName;
|
||
parser.tagName = '';
|
||
parser.state = S.SCRIPT;
|
||
} else {
|
||
if (!isWhitespace(c)) {
|
||
strictFail(parser, 'Invalid tagname in closing tag');
|
||
}
|
||
parser.state = S.CLOSE_TAG_SAW_WHITE;
|
||
}
|
||
continue
|
||
|
||
case S.CLOSE_TAG_SAW_WHITE:
|
||
if (isWhitespace(c)) {
|
||
continue
|
||
}
|
||
if (c === '>') {
|
||
closeTag(parser);
|
||
} else {
|
||
strictFail(parser, 'Invalid characters in closing tag');
|
||
}
|
||
continue
|
||
|
||
case S.TEXT_ENTITY:
|
||
case S.ATTRIB_VALUE_ENTITY_Q:
|
||
case S.ATTRIB_VALUE_ENTITY_U:
|
||
var returnState;
|
||
var buffer;
|
||
switch (parser.state) {
|
||
case S.TEXT_ENTITY:
|
||
returnState = S.TEXT;
|
||
buffer = 'textNode';
|
||
break
|
||
|
||
case S.ATTRIB_VALUE_ENTITY_Q:
|
||
returnState = S.ATTRIB_VALUE_QUOTED;
|
||
buffer = 'attribValue';
|
||
break
|
||
|
||
case S.ATTRIB_VALUE_ENTITY_U:
|
||
returnState = S.ATTRIB_VALUE_UNQUOTED;
|
||
buffer = 'attribValue';
|
||
break
|
||
}
|
||
|
||
if (c === ';') {
|
||
parser[buffer] += parseEntity(parser);
|
||
parser.entity = '';
|
||
parser.state = returnState;
|
||
} else if (isMatch(parser.entity.length ? entityBody : entityStart, c)) {
|
||
parser.entity += c;
|
||
} else {
|
||
strictFail(parser, 'Invalid character in entity name');
|
||
parser[buffer] += '&' + parser.entity + c;
|
||
parser.entity = '';
|
||
parser.state = returnState;
|
||
}
|
||
|
||
continue
|
||
|
||
default:
|
||
throw new Error(parser, 'Unknown state: ' + parser.state)
|
||
}
|
||
} // while
|
||
|
||
if (parser.position >= parser.bufferCheckPosition) {
|
||
checkBufferLength(parser);
|
||
}
|
||
return parser
|
||
}
|
||
|
||
/*! http://mths.be/fromcodepoint v0.1.0 by @mathias */
|
||
/* istanbul ignore next */
|
||
if (!String.fromCodePoint) {
|
||
(function () {
|
||
var stringFromCharCode = String.fromCharCode;
|
||
var floor = Math.floor;
|
||
var fromCodePoint = function () {
|
||
var MAX_SIZE = 0x4000;
|
||
var codeUnits = [];
|
||
var highSurrogate;
|
||
var lowSurrogate;
|
||
var index = -1;
|
||
var length = arguments.length;
|
||
if (!length) {
|
||
return ''
|
||
}
|
||
var result = '';
|
||
while (++index < length) {
|
||
var codePoint = Number(arguments[index]);
|
||
if (
|
||
!isFinite(codePoint) || // `NaN`, `+Infinity`, or `-Infinity`
|
||
codePoint < 0 || // not a valid Unicode code point
|
||
codePoint > 0x10FFFF || // not a valid Unicode code point
|
||
floor(codePoint) !== codePoint // not an integer
|
||
) {
|
||
throw RangeError('Invalid code point: ' + codePoint)
|
||
}
|
||
if (codePoint <= 0xFFFF) { // BMP code point
|
||
codeUnits.push(codePoint);
|
||
} else { // Astral code point; split in surrogate halves
|
||
// http://mathiasbynens.be/notes/javascript-encoding#surrogate-formulae
|
||
codePoint -= 0x10000;
|
||
highSurrogate = (codePoint >> 10) + 0xD800;
|
||
lowSurrogate = (codePoint % 0x400) + 0xDC00;
|
||
codeUnits.push(highSurrogate, lowSurrogate);
|
||
}
|
||
if (index + 1 === length || codeUnits.length > MAX_SIZE) {
|
||
result += stringFromCharCode.apply(null, codeUnits);
|
||
codeUnits.length = 0;
|
||
}
|
||
}
|
||
return result
|
||
};
|
||
/* istanbul ignore next */
|
||
if (Object.defineProperty) {
|
||
Object.defineProperty(String, 'fromCodePoint', {
|
||
value: fromCodePoint,
|
||
configurable: true,
|
||
writable: true
|
||
});
|
||
} else {
|
||
String.fromCodePoint = fromCodePoint;
|
||
}
|
||
}());
|
||
}
|
||
})(exports);
|
||
} (sax));
|
||
return sax;
|
||
}
|
||
|
||
var arrayHelper;
|
||
var hasRequiredArrayHelper;
|
||
|
||
function requireArrayHelper () {
|
||
if (hasRequiredArrayHelper) return arrayHelper;
|
||
hasRequiredArrayHelper = 1;
|
||
arrayHelper = {
|
||
|
||
isArray: function(value) {
|
||
if (Array.isArray) {
|
||
return Array.isArray(value);
|
||
}
|
||
// fallback for older browsers like IE 8
|
||
return Object.prototype.toString.call( value ) === '[object Array]';
|
||
}
|
||
|
||
};
|
||
return arrayHelper;
|
||
}
|
||
|
||
var optionsHelper;
|
||
var hasRequiredOptionsHelper;
|
||
|
||
function requireOptionsHelper () {
|
||
if (hasRequiredOptionsHelper) return optionsHelper;
|
||
hasRequiredOptionsHelper = 1;
|
||
var isArray = requireArrayHelper().isArray;
|
||
|
||
optionsHelper = {
|
||
|
||
copyOptions: function (options) {
|
||
var key, copy = {};
|
||
for (key in options) {
|
||
if (options.hasOwnProperty(key)) {
|
||
copy[key] = options[key];
|
||
}
|
||
}
|
||
return copy;
|
||
},
|
||
|
||
ensureFlagExists: function (item, options) {
|
||
if (!(item in options) || typeof options[item] !== 'boolean') {
|
||
options[item] = false;
|
||
}
|
||
},
|
||
|
||
ensureSpacesExists: function (options) {
|
||
if (!('spaces' in options) || (typeof options.spaces !== 'number' && typeof options.spaces !== 'string')) {
|
||
options.spaces = 0;
|
||
}
|
||
},
|
||
|
||
ensureAlwaysArrayExists: function (options) {
|
||
if (!('alwaysArray' in options) || (typeof options.alwaysArray !== 'boolean' && !isArray(options.alwaysArray))) {
|
||
options.alwaysArray = false;
|
||
}
|
||
},
|
||
|
||
ensureKeyExists: function (key, options) {
|
||
if (!(key + 'Key' in options) || typeof options[key + 'Key'] !== 'string') {
|
||
options[key + 'Key'] = options.compact ? '_' + key : key;
|
||
}
|
||
},
|
||
|
||
checkFnExists: function (key, options) {
|
||
return key + 'Fn' in options;
|
||
}
|
||
|
||
};
|
||
return optionsHelper;
|
||
}
|
||
|
||
var xml2js;
|
||
var hasRequiredXml2js;
|
||
|
||
function requireXml2js () {
|
||
if (hasRequiredXml2js) return xml2js;
|
||
hasRequiredXml2js = 1;
|
||
var sax = requireSax();
|
||
var helper = requireOptionsHelper();
|
||
var isArray = requireArrayHelper().isArray;
|
||
|
||
var options;
|
||
var currentElement;
|
||
|
||
function validateOptions(userOptions) {
|
||
options = helper.copyOptions(userOptions);
|
||
helper.ensureFlagExists('ignoreDeclaration', options);
|
||
helper.ensureFlagExists('ignoreInstruction', options);
|
||
helper.ensureFlagExists('ignoreAttributes', options);
|
||
helper.ensureFlagExists('ignoreText', options);
|
||
helper.ensureFlagExists('ignoreComment', options);
|
||
helper.ensureFlagExists('ignoreCdata', options);
|
||
helper.ensureFlagExists('ignoreDoctype', options);
|
||
helper.ensureFlagExists('compact', options);
|
||
helper.ensureFlagExists('alwaysChildren', options);
|
||
helper.ensureFlagExists('addParent', options);
|
||
helper.ensureFlagExists('trim', options);
|
||
helper.ensureFlagExists('nativeType', options);
|
||
helper.ensureFlagExists('nativeTypeAttributes', options);
|
||
helper.ensureFlagExists('sanitize', options);
|
||
helper.ensureFlagExists('instructionHasAttributes', options);
|
||
helper.ensureFlagExists('captureSpacesBetweenElements', options);
|
||
helper.ensureAlwaysArrayExists(options);
|
||
helper.ensureKeyExists('declaration', options);
|
||
helper.ensureKeyExists('instruction', options);
|
||
helper.ensureKeyExists('attributes', options);
|
||
helper.ensureKeyExists('text', options);
|
||
helper.ensureKeyExists('comment', options);
|
||
helper.ensureKeyExists('cdata', options);
|
||
helper.ensureKeyExists('doctype', options);
|
||
helper.ensureKeyExists('type', options);
|
||
helper.ensureKeyExists('name', options);
|
||
helper.ensureKeyExists('elements', options);
|
||
helper.ensureKeyExists('parent', options);
|
||
helper.checkFnExists('doctype', options);
|
||
helper.checkFnExists('instruction', options);
|
||
helper.checkFnExists('cdata', options);
|
||
helper.checkFnExists('comment', options);
|
||
helper.checkFnExists('text', options);
|
||
helper.checkFnExists('instructionName', options);
|
||
helper.checkFnExists('elementName', options);
|
||
helper.checkFnExists('attributeName', options);
|
||
helper.checkFnExists('attributeValue', options);
|
||
helper.checkFnExists('attributes', options);
|
||
return options;
|
||
}
|
||
|
||
function nativeType(value) {
|
||
var nValue = Number(value);
|
||
if (!isNaN(nValue)) {
|
||
return nValue;
|
||
}
|
||
var bValue = value.toLowerCase();
|
||
if (bValue === 'true') {
|
||
return true;
|
||
} else if (bValue === 'false') {
|
||
return false;
|
||
}
|
||
return value;
|
||
}
|
||
|
||
function addField(type, value) {
|
||
var key;
|
||
if (options.compact) {
|
||
if (
|
||
!currentElement[options[type + 'Key']] &&
|
||
(isArray(options.alwaysArray) ? options.alwaysArray.indexOf(options[type + 'Key']) !== -1 : options.alwaysArray)
|
||
) {
|
||
currentElement[options[type + 'Key']] = [];
|
||
}
|
||
if (currentElement[options[type + 'Key']] && !isArray(currentElement[options[type + 'Key']])) {
|
||
currentElement[options[type + 'Key']] = [currentElement[options[type + 'Key']]];
|
||
}
|
||
if (type + 'Fn' in options && typeof value === 'string') {
|
||
value = options[type + 'Fn'](value, currentElement);
|
||
}
|
||
if (type === 'instruction' && ('instructionFn' in options || 'instructionNameFn' in options)) {
|
||
for (key in value) {
|
||
if (value.hasOwnProperty(key)) {
|
||
if ('instructionFn' in options) {
|
||
value[key] = options.instructionFn(value[key], key, currentElement);
|
||
} else {
|
||
var temp = value[key];
|
||
delete value[key];
|
||
value[options.instructionNameFn(key, temp, currentElement)] = temp;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
if (isArray(currentElement[options[type + 'Key']])) {
|
||
currentElement[options[type + 'Key']].push(value);
|
||
} else {
|
||
currentElement[options[type + 'Key']] = value;
|
||
}
|
||
} else {
|
||
if (!currentElement[options.elementsKey]) {
|
||
currentElement[options.elementsKey] = [];
|
||
}
|
||
var element = {};
|
||
element[options.typeKey] = type;
|
||
if (type === 'instruction') {
|
||
for (key in value) {
|
||
if (value.hasOwnProperty(key)) {
|
||
break;
|
||
}
|
||
}
|
||
element[options.nameKey] = 'instructionNameFn' in options ? options.instructionNameFn(key, value, currentElement) : key;
|
||
if (options.instructionHasAttributes) {
|
||
element[options.attributesKey] = value[key][options.attributesKey];
|
||
if ('instructionFn' in options) {
|
||
element[options.attributesKey] = options.instructionFn(element[options.attributesKey], key, currentElement);
|
||
}
|
||
} else {
|
||
if ('instructionFn' in options) {
|
||
value[key] = options.instructionFn(value[key], key, currentElement);
|
||
}
|
||
element[options.instructionKey] = value[key];
|
||
}
|
||
} else {
|
||
if (type + 'Fn' in options) {
|
||
value = options[type + 'Fn'](value, currentElement);
|
||
}
|
||
element[options[type + 'Key']] = value;
|
||
}
|
||
if (options.addParent) {
|
||
element[options.parentKey] = currentElement;
|
||
}
|
||
currentElement[options.elementsKey].push(element);
|
||
}
|
||
}
|
||
|
||
function manipulateAttributes(attributes) {
|
||
if ('attributesFn' in options && attributes) {
|
||
attributes = options.attributesFn(attributes, currentElement);
|
||
}
|
||
if ((options.trim || 'attributeValueFn' in options || 'attributeNameFn' in options || options.nativeTypeAttributes) && attributes) {
|
||
var key;
|
||
for (key in attributes) {
|
||
if (attributes.hasOwnProperty(key)) {
|
||
if (options.trim) attributes[key] = attributes[key].trim();
|
||
if (options.nativeTypeAttributes) {
|
||
attributes[key] = nativeType(attributes[key]);
|
||
}
|
||
if ('attributeValueFn' in options) attributes[key] = options.attributeValueFn(attributes[key], key, currentElement);
|
||
if ('attributeNameFn' in options) {
|
||
var temp = attributes[key];
|
||
delete attributes[key];
|
||
attributes[options.attributeNameFn(key, attributes[key], currentElement)] = temp;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
return attributes;
|
||
}
|
||
|
||
function onInstruction(instruction) {
|
||
var attributes = {};
|
||
if (instruction.body && (instruction.name.toLowerCase() === 'xml' || options.instructionHasAttributes)) {
|
||
var attrsRegExp = /([\w:-]+)\s*=\s*(?:"([^"]*)"|'([^']*)'|(\w+))\s*/g;
|
||
var match;
|
||
while ((match = attrsRegExp.exec(instruction.body)) !== null) {
|
||
attributes[match[1]] = match[2] || match[3] || match[4];
|
||
}
|
||
attributes = manipulateAttributes(attributes);
|
||
}
|
||
if (instruction.name.toLowerCase() === 'xml') {
|
||
if (options.ignoreDeclaration) {
|
||
return;
|
||
}
|
||
currentElement[options.declarationKey] = {};
|
||
if (Object.keys(attributes).length) {
|
||
currentElement[options.declarationKey][options.attributesKey] = attributes;
|
||
}
|
||
if (options.addParent) {
|
||
currentElement[options.declarationKey][options.parentKey] = currentElement;
|
||
}
|
||
} else {
|
||
if (options.ignoreInstruction) {
|
||
return;
|
||
}
|
||
if (options.trim) {
|
||
instruction.body = instruction.body.trim();
|
||
}
|
||
var value = {};
|
||
if (options.instructionHasAttributes && Object.keys(attributes).length) {
|
||
value[instruction.name] = {};
|
||
value[instruction.name][options.attributesKey] = attributes;
|
||
} else {
|
||
value[instruction.name] = instruction.body;
|
||
}
|
||
addField('instruction', value);
|
||
}
|
||
}
|
||
|
||
function onStartElement(name, attributes) {
|
||
var element;
|
||
if (typeof name === 'object') {
|
||
attributes = name.attributes;
|
||
name = name.name;
|
||
}
|
||
attributes = manipulateAttributes(attributes);
|
||
if ('elementNameFn' in options) {
|
||
name = options.elementNameFn(name, currentElement);
|
||
}
|
||
if (options.compact) {
|
||
element = {};
|
||
if (!options.ignoreAttributes && attributes && Object.keys(attributes).length) {
|
||
element[options.attributesKey] = {};
|
||
var key;
|
||
for (key in attributes) {
|
||
if (attributes.hasOwnProperty(key)) {
|
||
element[options.attributesKey][key] = attributes[key];
|
||
}
|
||
}
|
||
}
|
||
if (
|
||
!(name in currentElement) &&
|
||
(isArray(options.alwaysArray) ? options.alwaysArray.indexOf(name) !== -1 : options.alwaysArray)
|
||
) {
|
||
currentElement[name] = [];
|
||
}
|
||
if (currentElement[name] && !isArray(currentElement[name])) {
|
||
currentElement[name] = [currentElement[name]];
|
||
}
|
||
if (isArray(currentElement[name])) {
|
||
currentElement[name].push(element);
|
||
} else {
|
||
currentElement[name] = element;
|
||
}
|
||
} else {
|
||
if (!currentElement[options.elementsKey]) {
|
||
currentElement[options.elementsKey] = [];
|
||
}
|
||
element = {};
|
||
element[options.typeKey] = 'element';
|
||
element[options.nameKey] = name;
|
||
if (!options.ignoreAttributes && attributes && Object.keys(attributes).length) {
|
||
element[options.attributesKey] = attributes;
|
||
}
|
||
if (options.alwaysChildren) {
|
||
element[options.elementsKey] = [];
|
||
}
|
||
currentElement[options.elementsKey].push(element);
|
||
}
|
||
element[options.parentKey] = currentElement; // will be deleted in onEndElement() if !options.addParent
|
||
currentElement = element;
|
||
}
|
||
|
||
function onText(text) {
|
||
if (options.ignoreText) {
|
||
return;
|
||
}
|
||
if (!text.trim() && !options.captureSpacesBetweenElements) {
|
||
return;
|
||
}
|
||
if (options.trim) {
|
||
text = text.trim();
|
||
}
|
||
if (options.nativeType) {
|
||
text = nativeType(text);
|
||
}
|
||
if (options.sanitize) {
|
||
text = text.replace(/&/g, '&').replace(/</g, '<').replace(/>/g, '>');
|
||
}
|
||
addField('text', text);
|
||
}
|
||
|
||
function onComment(comment) {
|
||
if (options.ignoreComment) {
|
||
return;
|
||
}
|
||
if (options.trim) {
|
||
comment = comment.trim();
|
||
}
|
||
addField('comment', comment);
|
||
}
|
||
|
||
function onEndElement(name) {
|
||
var parentElement = currentElement[options.parentKey];
|
||
if (!options.addParent) {
|
||
delete currentElement[options.parentKey];
|
||
}
|
||
currentElement = parentElement;
|
||
}
|
||
|
||
function onCdata(cdata) {
|
||
if (options.ignoreCdata) {
|
||
return;
|
||
}
|
||
if (options.trim) {
|
||
cdata = cdata.trim();
|
||
}
|
||
addField('cdata', cdata);
|
||
}
|
||
|
||
function onDoctype(doctype) {
|
||
if (options.ignoreDoctype) {
|
||
return;
|
||
}
|
||
doctype = doctype.replace(/^ /, '');
|
||
if (options.trim) {
|
||
doctype = doctype.trim();
|
||
}
|
||
addField('doctype', doctype);
|
||
}
|
||
|
||
function onError(error) {
|
||
error.note = error; //console.error(error);
|
||
}
|
||
|
||
xml2js = function (xml, userOptions) {
|
||
|
||
var parser = sax.parser(true, {}) ;
|
||
var result = {};
|
||
currentElement = result;
|
||
|
||
options = validateOptions(userOptions);
|
||
|
||
{
|
||
parser.opt = {strictEntities: true};
|
||
parser.onopentag = onStartElement;
|
||
parser.ontext = onText;
|
||
parser.oncomment = onComment;
|
||
parser.onclosetag = onEndElement;
|
||
parser.onerror = onError;
|
||
parser.oncdata = onCdata;
|
||
parser.ondoctype = onDoctype;
|
||
parser.onprocessinginstruction = onInstruction;
|
||
}
|
||
|
||
{
|
||
parser.write(xml).close();
|
||
}
|
||
|
||
if (result[options.elementsKey]) {
|
||
var temp = result[options.elementsKey];
|
||
delete result[options.elementsKey];
|
||
result[options.elementsKey] = temp;
|
||
delete result.text;
|
||
}
|
||
|
||
return result;
|
||
|
||
};
|
||
return xml2js;
|
||
}
|
||
|
||
var xml2json;
|
||
var hasRequiredXml2json;
|
||
|
||
function requireXml2json () {
|
||
if (hasRequiredXml2json) return xml2json;
|
||
hasRequiredXml2json = 1;
|
||
var helper = requireOptionsHelper();
|
||
var xml2js = requireXml2js();
|
||
|
||
function validateOptions (userOptions) {
|
||
var options = helper.copyOptions(userOptions);
|
||
helper.ensureSpacesExists(options);
|
||
return options;
|
||
}
|
||
|
||
xml2json = function(xml, userOptions) {
|
||
var options, js, json, parentKey;
|
||
options = validateOptions(userOptions);
|
||
js = xml2js(xml, options);
|
||
parentKey = 'compact' in options && options.compact ? '_parent' : 'parent';
|
||
// parentKey = ptions.compact ? '_parent' : 'parent'; // consider this
|
||
if ('addParent' in options && options.addParent) {
|
||
json = JSON.stringify(js, function (k, v) { return k === parentKey? '_' : v; }, options.spaces);
|
||
} else {
|
||
json = JSON.stringify(js, null, options.spaces);
|
||
}
|
||
return json.replace(/\u2028/g, '\\u2028').replace(/\u2029/g, '\\u2029');
|
||
};
|
||
return xml2json;
|
||
}
|
||
|
||
var js2xml;
|
||
var hasRequiredJs2xml;
|
||
|
||
function requireJs2xml () {
|
||
if (hasRequiredJs2xml) return js2xml;
|
||
hasRequiredJs2xml = 1;
|
||
var helper = requireOptionsHelper();
|
||
var isArray = requireArrayHelper().isArray;
|
||
|
||
var currentElement, currentElementName;
|
||
|
||
function validateOptions(userOptions) {
|
||
var options = helper.copyOptions(userOptions);
|
||
helper.ensureFlagExists('ignoreDeclaration', options);
|
||
helper.ensureFlagExists('ignoreInstruction', options);
|
||
helper.ensureFlagExists('ignoreAttributes', options);
|
||
helper.ensureFlagExists('ignoreText', options);
|
||
helper.ensureFlagExists('ignoreComment', options);
|
||
helper.ensureFlagExists('ignoreCdata', options);
|
||
helper.ensureFlagExists('ignoreDoctype', options);
|
||
helper.ensureFlagExists('compact', options);
|
||
helper.ensureFlagExists('indentText', options);
|
||
helper.ensureFlagExists('indentCdata', options);
|
||
helper.ensureFlagExists('indentAttributes', options);
|
||
helper.ensureFlagExists('indentInstruction', options);
|
||
helper.ensureFlagExists('fullTagEmptyElement', options);
|
||
helper.ensureFlagExists('noQuotesForNativeAttributes', options);
|
||
helper.ensureSpacesExists(options);
|
||
if (typeof options.spaces === 'number') {
|
||
options.spaces = Array(options.spaces + 1).join(' ');
|
||
}
|
||
helper.ensureKeyExists('declaration', options);
|
||
helper.ensureKeyExists('instruction', options);
|
||
helper.ensureKeyExists('attributes', options);
|
||
helper.ensureKeyExists('text', options);
|
||
helper.ensureKeyExists('comment', options);
|
||
helper.ensureKeyExists('cdata', options);
|
||
helper.ensureKeyExists('doctype', options);
|
||
helper.ensureKeyExists('type', options);
|
||
helper.ensureKeyExists('name', options);
|
||
helper.ensureKeyExists('elements', options);
|
||
helper.checkFnExists('doctype', options);
|
||
helper.checkFnExists('instruction', options);
|
||
helper.checkFnExists('cdata', options);
|
||
helper.checkFnExists('comment', options);
|
||
helper.checkFnExists('text', options);
|
||
helper.checkFnExists('instructionName', options);
|
||
helper.checkFnExists('elementName', options);
|
||
helper.checkFnExists('attributeName', options);
|
||
helper.checkFnExists('attributeValue', options);
|
||
helper.checkFnExists('attributes', options);
|
||
helper.checkFnExists('fullTagEmptyElement', options);
|
||
return options;
|
||
}
|
||
|
||
function writeIndentation(options, depth, firstLine) {
|
||
return (!firstLine && options.spaces ? '\n' : '') + Array(depth + 1).join(options.spaces);
|
||
}
|
||
|
||
function writeAttributes(attributes, options, depth) {
|
||
if (options.ignoreAttributes) {
|
||
return '';
|
||
}
|
||
if ('attributesFn' in options) {
|
||
attributes = options.attributesFn(attributes, currentElementName, currentElement);
|
||
}
|
||
var key, attr, attrName, quote, result = [];
|
||
for (key in attributes) {
|
||
if (attributes.hasOwnProperty(key) && attributes[key] !== null && attributes[key] !== undefined) {
|
||
quote = options.noQuotesForNativeAttributes && typeof attributes[key] !== 'string' ? '' : '"';
|
||
attr = '' + attributes[key]; // ensure number and boolean are converted to String
|
||
attr = attr.replace(/"/g, '"');
|
||
attrName = 'attributeNameFn' in options ? options.attributeNameFn(key, attr, currentElementName, currentElement) : key;
|
||
result.push((options.spaces && options.indentAttributes? writeIndentation(options, depth+1, false) : ' '));
|
||
result.push(attrName + '=' + quote + ('attributeValueFn' in options ? options.attributeValueFn(attr, key, currentElementName, currentElement) : attr) + quote);
|
||
}
|
||
}
|
||
if (attributes && Object.keys(attributes).length && options.spaces && options.indentAttributes) {
|
||
result.push(writeIndentation(options, depth, false));
|
||
}
|
||
return result.join('');
|
||
}
|
||
|
||
function writeDeclaration(declaration, options, depth) {
|
||
currentElement = declaration;
|
||
currentElementName = 'xml';
|
||
return options.ignoreDeclaration ? '' : '<?' + 'xml' + writeAttributes(declaration[options.attributesKey], options, depth) + '?>';
|
||
}
|
||
|
||
function writeInstruction(instruction, options, depth) {
|
||
if (options.ignoreInstruction) {
|
||
return '';
|
||
}
|
||
var key;
|
||
for (key in instruction) {
|
||
if (instruction.hasOwnProperty(key)) {
|
||
break;
|
||
}
|
||
}
|
||
var instructionName = 'instructionNameFn' in options ? options.instructionNameFn(key, instruction[key], currentElementName, currentElement) : key;
|
||
if (typeof instruction[key] === 'object') {
|
||
currentElement = instruction;
|
||
currentElementName = instructionName;
|
||
return '<?' + instructionName + writeAttributes(instruction[key][options.attributesKey], options, depth) + '?>';
|
||
} else {
|
||
var instructionValue = instruction[key] ? instruction[key] : '';
|
||
if ('instructionFn' in options) instructionValue = options.instructionFn(instructionValue, key, currentElementName, currentElement);
|
||
return '<?' + instructionName + (instructionValue ? ' ' + instructionValue : '') + '?>';
|
||
}
|
||
}
|
||
|
||
function writeComment(comment, options) {
|
||
return options.ignoreComment ? '' : '<!--' + ('commentFn' in options ? options.commentFn(comment, currentElementName, currentElement) : comment) + '-->';
|
||
}
|
||
|
||
function writeCdata(cdata, options) {
|
||
return options.ignoreCdata ? '' : '<![CDATA[' + ('cdataFn' in options ? options.cdataFn(cdata, currentElementName, currentElement) : cdata.replace(']]>', ']]]]><![CDATA[>')) + ']]>';
|
||
}
|
||
|
||
function writeDoctype(doctype, options) {
|
||
return options.ignoreDoctype ? '' : '<!DOCTYPE ' + ('doctypeFn' in options ? options.doctypeFn(doctype, currentElementName, currentElement) : doctype) + '>';
|
||
}
|
||
|
||
function writeText(text, options) {
|
||
if (options.ignoreText) return '';
|
||
text = '' + text; // ensure Number and Boolean are converted to String
|
||
text = text.replace(/&/g, '&'); // desanitize to avoid double sanitization
|
||
text = text.replace(/&/g, '&').replace(/</g, '<').replace(/>/g, '>');
|
||
return 'textFn' in options ? options.textFn(text, currentElementName, currentElement) : text;
|
||
}
|
||
|
||
function hasContent(element, options) {
|
||
var i;
|
||
if (element.elements && element.elements.length) {
|
||
for (i = 0; i < element.elements.length; ++i) {
|
||
switch (element.elements[i][options.typeKey]) {
|
||
case 'text':
|
||
if (options.indentText) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case 'cdata':
|
||
if (options.indentCdata) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case 'instruction':
|
||
if (options.indentInstruction) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case 'doctype':
|
||
case 'comment':
|
||
case 'element':
|
||
return true;
|
||
default:
|
||
return true;
|
||
}
|
||
}
|
||
}
|
||
return false;
|
||
}
|
||
|
||
function writeElement(element, options, depth) {
|
||
currentElement = element;
|
||
currentElementName = element.name;
|
||
var xml = [], elementName = 'elementNameFn' in options ? options.elementNameFn(element.name, element) : element.name;
|
||
xml.push('<' + elementName);
|
||
if (element[options.attributesKey]) {
|
||
xml.push(writeAttributes(element[options.attributesKey], options, depth));
|
||
}
|
||
var withClosingTag = element[options.elementsKey] && element[options.elementsKey].length || element[options.attributesKey] && element[options.attributesKey]['xml:space'] === 'preserve';
|
||
if (!withClosingTag) {
|
||
if ('fullTagEmptyElementFn' in options) {
|
||
withClosingTag = options.fullTagEmptyElementFn(element.name, element);
|
||
} else {
|
||
withClosingTag = options.fullTagEmptyElement;
|
||
}
|
||
}
|
||
if (withClosingTag) {
|
||
xml.push('>');
|
||
if (element[options.elementsKey] && element[options.elementsKey].length) {
|
||
xml.push(writeElements(element[options.elementsKey], options, depth + 1));
|
||
currentElement = element;
|
||
currentElementName = element.name;
|
||
}
|
||
xml.push(options.spaces && hasContent(element, options) ? '\n' + Array(depth + 1).join(options.spaces) : '');
|
||
xml.push('</' + elementName + '>');
|
||
} else {
|
||
xml.push('/>');
|
||
}
|
||
return xml.join('');
|
||
}
|
||
|
||
function writeElements(elements, options, depth, firstLine) {
|
||
return elements.reduce(function (xml, element) {
|
||
var indent = writeIndentation(options, depth, firstLine && !xml);
|
||
switch (element.type) {
|
||
case 'element': return xml + indent + writeElement(element, options, depth);
|
||
case 'comment': return xml + indent + writeComment(element[options.commentKey], options);
|
||
case 'doctype': return xml + indent + writeDoctype(element[options.doctypeKey], options);
|
||
case 'cdata': return xml + (options.indentCdata ? indent : '') + writeCdata(element[options.cdataKey], options);
|
||
case 'text': return xml + (options.indentText ? indent : '') + writeText(element[options.textKey], options);
|
||
case 'instruction':
|
||
var instruction = {};
|
||
instruction[element[options.nameKey]] = element[options.attributesKey] ? element : element[options.instructionKey];
|
||
return xml + (options.indentInstruction ? indent : '') + writeInstruction(instruction, options, depth);
|
||
}
|
||
}, '');
|
||
}
|
||
|
||
function hasContentCompact(element, options, anyContent) {
|
||
var key;
|
||
for (key in element) {
|
||
if (element.hasOwnProperty(key)) {
|
||
switch (key) {
|
||
case options.parentKey:
|
||
case options.attributesKey:
|
||
break; // skip to next key
|
||
case options.textKey:
|
||
if (options.indentText || anyContent) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case options.cdataKey:
|
||
if (options.indentCdata || anyContent) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case options.instructionKey:
|
||
if (options.indentInstruction || anyContent) {
|
||
return true;
|
||
}
|
||
break; // skip to next key
|
||
case options.doctypeKey:
|
||
case options.commentKey:
|
||
return true;
|
||
default:
|
||
return true;
|
||
}
|
||
}
|
||
}
|
||
return false;
|
||
}
|
||
|
||
function writeElementCompact(element, name, options, depth, indent) {
|
||
currentElement = element;
|
||
currentElementName = name;
|
||
var elementName = 'elementNameFn' in options ? options.elementNameFn(name, element) : name;
|
||
if (typeof element === 'undefined' || element === null || element === '') {
|
||
return 'fullTagEmptyElementFn' in options && options.fullTagEmptyElementFn(name, element) || options.fullTagEmptyElement ? '<' + elementName + '></' + elementName + '>' : '<' + elementName + '/>';
|
||
}
|
||
var xml = [];
|
||
if (name) {
|
||
xml.push('<' + elementName);
|
||
if (typeof element !== 'object') {
|
||
xml.push('>' + writeText(element,options) + '</' + elementName + '>');
|
||
return xml.join('');
|
||
}
|
||
if (element[options.attributesKey]) {
|
||
xml.push(writeAttributes(element[options.attributesKey], options, depth));
|
||
}
|
||
var withClosingTag = hasContentCompact(element, options, true) || element[options.attributesKey] && element[options.attributesKey]['xml:space'] === 'preserve';
|
||
if (!withClosingTag) {
|
||
if ('fullTagEmptyElementFn' in options) {
|
||
withClosingTag = options.fullTagEmptyElementFn(name, element);
|
||
} else {
|
||
withClosingTag = options.fullTagEmptyElement;
|
||
}
|
||
}
|
||
if (withClosingTag) {
|
||
xml.push('>');
|
||
} else {
|
||
xml.push('/>');
|
||
return xml.join('');
|
||
}
|
||
}
|
||
xml.push(writeElementsCompact(element, options, depth + 1, false));
|
||
currentElement = element;
|
||
currentElementName = name;
|
||
if (name) {
|
||
xml.push((indent ? writeIndentation(options, depth, false) : '') + '</' + elementName + '>');
|
||
}
|
||
return xml.join('');
|
||
}
|
||
|
||
function writeElementsCompact(element, options, depth, firstLine) {
|
||
var i, key, nodes, xml = [];
|
||
for (key in element) {
|
||
if (element.hasOwnProperty(key)) {
|
||
nodes = isArray(element[key]) ? element[key] : [element[key]];
|
||
for (i = 0; i < nodes.length; ++i) {
|
||
switch (key) {
|
||
case options.declarationKey: xml.push(writeDeclaration(nodes[i], options, depth)); break;
|
||
case options.instructionKey: xml.push((options.indentInstruction ? writeIndentation(options, depth, firstLine) : '') + writeInstruction(nodes[i], options, depth)); break;
|
||
case options.attributesKey: case options.parentKey: break; // skip
|
||
case options.textKey: xml.push((options.indentText ? writeIndentation(options, depth, firstLine) : '') + writeText(nodes[i], options)); break;
|
||
case options.cdataKey: xml.push((options.indentCdata ? writeIndentation(options, depth, firstLine) : '') + writeCdata(nodes[i], options)); break;
|
||
case options.doctypeKey: xml.push(writeIndentation(options, depth, firstLine) + writeDoctype(nodes[i], options)); break;
|
||
case options.commentKey: xml.push(writeIndentation(options, depth, firstLine) + writeComment(nodes[i], options)); break;
|
||
default: xml.push(writeIndentation(options, depth, firstLine) + writeElementCompact(nodes[i], key, options, depth, hasContentCompact(nodes[i], options)));
|
||
}
|
||
firstLine = firstLine && !xml.length;
|
||
}
|
||
}
|
||
}
|
||
return xml.join('');
|
||
}
|
||
|
||
js2xml = function (js, options) {
|
||
options = validateOptions(options);
|
||
var xml = [];
|
||
currentElement = js;
|
||
currentElementName = '_root_';
|
||
if (options.compact) {
|
||
xml.push(writeElementsCompact(js, options, 0, true));
|
||
} else {
|
||
if (js[options.declarationKey]) {
|
||
xml.push(writeDeclaration(js[options.declarationKey], options, 0));
|
||
}
|
||
if (js[options.elementsKey] && js[options.elementsKey].length) {
|
||
xml.push(writeElements(js[options.elementsKey], options, 0, !xml.length));
|
||
}
|
||
}
|
||
return xml.join('');
|
||
};
|
||
return js2xml;
|
||
}
|
||
|
||
var json2xml;
|
||
var hasRequiredJson2xml;
|
||
|
||
function requireJson2xml () {
|
||
if (hasRequiredJson2xml) return json2xml;
|
||
hasRequiredJson2xml = 1;
|
||
var js2xml = requireJs2xml();
|
||
|
||
json2xml = function (json, options) {
|
||
if (json instanceof Buffer) {
|
||
json = json.toString();
|
||
}
|
||
var js = null;
|
||
if (typeof (json) === 'string') {
|
||
try {
|
||
js = JSON.parse(json);
|
||
} catch (e) {
|
||
throw new Error('The JSON structure is invalid');
|
||
}
|
||
} else {
|
||
js = json;
|
||
}
|
||
return js2xml(js, options);
|
||
};
|
||
return json2xml;
|
||
}
|
||
|
||
/*jslint node:true */
|
||
|
||
var lib;
|
||
var hasRequiredLib;
|
||
|
||
function requireLib () {
|
||
if (hasRequiredLib) return lib;
|
||
hasRequiredLib = 1;
|
||
var xml2js = requireXml2js();
|
||
var xml2json = requireXml2json();
|
||
var js2xml = requireJs2xml();
|
||
var json2xml = requireJson2xml();
|
||
|
||
lib = {
|
||
xml2js: xml2js,
|
||
xml2json: xml2json,
|
||
js2xml: js2xml,
|
||
json2xml: json2xml
|
||
};
|
||
return lib;
|
||
}
|
||
|
||
var libExports = requireLib();
|
||
var convert = /*@__PURE__*/getDefaultExportFromCjs(libExports);
|
||
|
||
const camelCase = (str) => str.replace(/([-_]\w)/g, (g) => g[1].toUpperCase());
|
||
|
||
const nativeType = (value) => {
|
||
const nValue = Number(value);
|
||
if (!Number.isNaN(nValue)) {
|
||
return nValue;
|
||
}
|
||
const bValue = value.toLowerCase();
|
||
if (bValue === 'true') {
|
||
return true;
|
||
}
|
||
if (bValue === 'false') {
|
||
return false;
|
||
}
|
||
return value;
|
||
};
|
||
|
||
const urlEquals = (urlA, urlB) => {
|
||
if (!urlA && !urlB) {
|
||
return true;
|
||
}
|
||
if (!urlA || !urlB) {
|
||
return false;
|
||
}
|
||
const trimmedUrlA = urlA.trim();
|
||
const trimmedUrlB = urlB.trim();
|
||
if (Math.abs(trimmedUrlA.length - trimmedUrlB.length) > 1) {
|
||
return false;
|
||
}
|
||
const strippedUrlA = trimmedUrlA.slice(-1) === '/' ? trimmedUrlA.slice(0, -1) : trimmedUrlA;
|
||
const strippedUrlB = trimmedUrlB.slice(-1) === '/' ? trimmedUrlB.slice(0, -1) : trimmedUrlB;
|
||
return urlA.includes(strippedUrlB) || urlB.includes(strippedUrlA);
|
||
};
|
||
const urlContains = (urlA, urlB) => {
|
||
if (!urlA && !urlB) {
|
||
return true;
|
||
}
|
||
if (!urlA || !urlB) {
|
||
return false;
|
||
}
|
||
const trimmedUrlA = urlA.trim();
|
||
const trimmedUrlB = urlB.trim();
|
||
const strippedUrlA = trimmedUrlA.slice(-1) === '/' ? trimmedUrlA.slice(0, -1) : trimmedUrlA;
|
||
const strippedUrlB = trimmedUrlB.slice(-1) === '/' ? trimmedUrlB.slice(0, -1) : trimmedUrlB;
|
||
return urlA.includes(strippedUrlB) || urlB.includes(strippedUrlA);
|
||
};
|
||
const getDAVAttribute = (nsArr) => nsArr.reduce((prev, curr) => ({ ...prev, [DAVAttributeMap[curr]]: curr }), {});
|
||
const cleanupFalsy = (obj) => Object.entries(obj).reduce((prev, [key, value]) => {
|
||
if (value)
|
||
return { ...prev, [key]: value };
|
||
return prev;
|
||
}, {});
|
||
const conditionalParam = (key, param) => {
|
||
if (param) {
|
||
return {
|
||
[key]: param,
|
||
};
|
||
}
|
||
return {};
|
||
};
|
||
const excludeHeaders = (headers, headersToExclude) => {
|
||
if (!headers) {
|
||
return {};
|
||
}
|
||
if (!headersToExclude || headersToExclude.length === 0) {
|
||
return headers;
|
||
}
|
||
return Object.fromEntries(Object.entries(headers).filter(([key]) => !headersToExclude.includes(key)));
|
||
};
|
||
|
||
var requestHelpers = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
cleanupFalsy: cleanupFalsy,
|
||
conditionalParam: conditionalParam,
|
||
excludeHeaders: excludeHeaders,
|
||
getDAVAttribute: getDAVAttribute,
|
||
urlContains: urlContains,
|
||
urlEquals: urlEquals
|
||
});
|
||
|
||
const debug$5 = getLogger('tsdav:request');
|
||
const davRequest = async (params) => {
|
||
var _a;
|
||
const { url, init, convertIncoming = true, parseOutgoing = true } = params;
|
||
const { headers = {}, body, namespace, method, attributes } = init;
|
||
const xmlBody = convertIncoming
|
||
? convert.js2xml({
|
||
_declaration: { _attributes: { version: '1.0', encoding: 'utf-8' } },
|
||
...body,
|
||
_attributes: attributes,
|
||
}, {
|
||
compact: true,
|
||
spaces: 2,
|
||
elementNameFn: (name) => {
|
||
// add namespace to all keys without namespace
|
||
if (namespace && !/^.+:.+/.test(name)) {
|
||
return `${namespace}:${name}`;
|
||
}
|
||
return name;
|
||
},
|
||
})
|
||
: body;
|
||
// debug('outgoing xml:');
|
||
// debug(`${method} ${url}`);
|
||
// debug(
|
||
// `headers: ${JSON.stringify(
|
||
// {
|
||
// 'Content-Type': 'text/xml;charset=UTF-8',
|
||
// ...cleanupFalsy(headers),
|
||
// },
|
||
// null,
|
||
// 2
|
||
// )}`
|
||
// );
|
||
// debug(xmlBody);
|
||
const davResponse = await browserPonyfillExports.fetch(url, {
|
||
headers: {
|
||
'Content-Type': 'text/xml;charset=UTF-8',
|
||
...cleanupFalsy(headers),
|
||
},
|
||
body: xmlBody,
|
||
method,
|
||
});
|
||
const resText = await davResponse.text();
|
||
// filter out invalid responses
|
||
// debug('response xml:');
|
||
// debug(resText);
|
||
// debug(davResponse);
|
||
if (!davResponse.ok ||
|
||
!((_a = davResponse.headers.get('content-type')) === null || _a === void 0 ? void 0 : _a.includes('xml')) ||
|
||
!parseOutgoing) {
|
||
return [
|
||
{
|
||
href: davResponse.url,
|
||
ok: davResponse.ok,
|
||
status: davResponse.status,
|
||
statusText: davResponse.statusText,
|
||
raw: resText,
|
||
},
|
||
];
|
||
}
|
||
const result = convert.xml2js(resText, {
|
||
compact: true,
|
||
trim: true,
|
||
textFn: (value, parentElement) => {
|
||
try {
|
||
// This is needed for xml-js design reasons
|
||
const parentOfParent = parentElement._parent;
|
||
const pOpKeys = Object.keys(parentOfParent);
|
||
const keyNo = pOpKeys.length;
|
||
const keyName = pOpKeys[keyNo - 1];
|
||
const arrOfKey = parentOfParent[keyName];
|
||
const arrOfKeyLen = arrOfKey.length;
|
||
if (arrOfKeyLen > 0) {
|
||
const arr = arrOfKey;
|
||
const arrIndex = arrOfKey.length - 1;
|
||
arr[arrIndex] = nativeType(value);
|
||
}
|
||
else {
|
||
parentOfParent[keyName] = nativeType(value);
|
||
}
|
||
}
|
||
catch (e) {
|
||
debug$5(e.stack);
|
||
}
|
||
},
|
||
// remove namespace & camelCase
|
||
elementNameFn: (attributeName) => camelCase(attributeName.replace(/^.+:/, '')),
|
||
attributesFn: (value) => {
|
||
const newVal = { ...value };
|
||
delete newVal.xmlns;
|
||
return newVal;
|
||
},
|
||
ignoreDeclaration: true,
|
||
});
|
||
const responseBodies = Array.isArray(result.multistatus.response)
|
||
? result.multistatus.response
|
||
: [result.multistatus.response];
|
||
return responseBodies.map((responseBody) => {
|
||
var _a, _b;
|
||
const statusRegex = /^\S+\s(?<status>\d+)\s(?<statusText>.+)$/;
|
||
if (!responseBody) {
|
||
return {
|
||
status: davResponse.status,
|
||
statusText: davResponse.statusText,
|
||
ok: davResponse.ok,
|
||
};
|
||
}
|
||
const matchArr = statusRegex.exec(responseBody.status);
|
||
return {
|
||
raw: result,
|
||
href: responseBody.href,
|
||
status: (matchArr === null || matchArr === void 0 ? void 0 : matchArr.groups) ? Number.parseInt(matchArr === null || matchArr === void 0 ? void 0 : matchArr.groups.status, 10) : davResponse.status,
|
||
statusText: (_b = (_a = matchArr === null || matchArr === void 0 ? void 0 : matchArr.groups) === null || _a === void 0 ? void 0 : _a.statusText) !== null && _b !== void 0 ? _b : davResponse.statusText,
|
||
ok: !responseBody.error,
|
||
error: responseBody.error,
|
||
responsedescription: responseBody.responsedescription,
|
||
props: (Array.isArray(responseBody.propstat)
|
||
? responseBody.propstat
|
||
: [responseBody.propstat]).reduce((prev, curr) => {
|
||
return {
|
||
...prev,
|
||
...curr === null || curr === void 0 ? void 0 : curr.prop,
|
||
};
|
||
}, {}),
|
||
};
|
||
});
|
||
};
|
||
const propfind = async (params) => {
|
||
const { url, props, depth, headers, headersToExclude } = params;
|
||
return davRequest({
|
||
url,
|
||
init: {
|
||
method: 'PROPFIND',
|
||
headers: excludeHeaders(cleanupFalsy({ depth, ...headers }), headersToExclude),
|
||
namespace: DAVNamespaceShort.DAV,
|
||
body: {
|
||
propfind: {
|
||
_attributes: getDAVAttribute([
|
||
DAVNamespace.CALDAV,
|
||
DAVNamespace.CALDAV_APPLE,
|
||
DAVNamespace.CALENDAR_SERVER,
|
||
DAVNamespace.CARDDAV,
|
||
DAVNamespace.DAV,
|
||
]),
|
||
prop: props,
|
||
},
|
||
},
|
||
},
|
||
});
|
||
};
|
||
const createObject = async (params) => {
|
||
const { url, data, headers, headersToExclude } = params;
|
||
return browserPonyfillExports.fetch(url, {
|
||
method: 'PUT',
|
||
body: data,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
const updateObject = async (params) => {
|
||
const { url, data, etag, headers, headersToExclude } = params;
|
||
return browserPonyfillExports.fetch(url, {
|
||
method: 'PUT',
|
||
body: data,
|
||
headers: excludeHeaders(cleanupFalsy({ 'If-Match': etag, ...headers }), headersToExclude),
|
||
});
|
||
};
|
||
const deleteObject = async (params) => {
|
||
const { url, headers, etag, headersToExclude } = params;
|
||
return browserPonyfillExports.fetch(url, {
|
||
method: 'DELETE',
|
||
headers: excludeHeaders(cleanupFalsy({ 'If-Match': etag, ...headers }), headersToExclude),
|
||
});
|
||
};
|
||
|
||
var request = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
createObject: createObject,
|
||
davRequest: davRequest,
|
||
deleteObject: deleteObject,
|
||
propfind: propfind,
|
||
updateObject: updateObject
|
||
});
|
||
|
||
function hasFields(obj, fields) {
|
||
const inObj = (object) => fields.every((f) => object[f]);
|
||
if (Array.isArray(obj)) {
|
||
return obj.every((o) => inObj(o));
|
||
}
|
||
return inObj(obj);
|
||
}
|
||
const findMissingFieldNames = (obj, fields) => fields.reduce((prev, curr) => (obj[curr] ? prev : `${prev.length ? `${prev},` : ''}${curr.toString()}`), '');
|
||
|
||
const debug$4 = getLogger('tsdav:collection');
|
||
const collectionQuery = async (params) => {
|
||
const { url, body, depth, defaultNamespace = DAVNamespaceShort.DAV, headers, headersToExclude, } = params;
|
||
const queryResults = await davRequest({
|
||
url,
|
||
init: {
|
||
method: 'REPORT',
|
||
headers: excludeHeaders(cleanupFalsy({ depth, ...headers }), headersToExclude),
|
||
namespace: defaultNamespace,
|
||
body,
|
||
},
|
||
});
|
||
// empty query result
|
||
if (queryResults.length === 1 && !queryResults[0].raw) {
|
||
return [];
|
||
}
|
||
return queryResults;
|
||
};
|
||
const makeCollection = async (params) => {
|
||
const { url, props, depth, headers, headersToExclude } = params;
|
||
return davRequest({
|
||
url,
|
||
init: {
|
||
method: 'MKCOL',
|
||
headers: excludeHeaders(cleanupFalsy({ depth, ...headers }), headersToExclude),
|
||
namespace: DAVNamespaceShort.DAV,
|
||
body: props
|
||
? {
|
||
mkcol: {
|
||
set: {
|
||
prop: props,
|
||
},
|
||
},
|
||
}
|
||
: undefined,
|
||
},
|
||
});
|
||
};
|
||
const supportedReportSet = async (params) => {
|
||
var _a, _b, _c, _d, _e;
|
||
const { collection, headers, headersToExclude } = params;
|
||
const res = await propfind({
|
||
url: collection.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:supported-report-set`]: {},
|
||
},
|
||
depth: '0',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
return ((_e = (_d = (_c = (_b = (_a = res[0]) === null || _a === void 0 ? void 0 : _a.props) === null || _b === void 0 ? void 0 : _b.supportedReportSet) === null || _c === void 0 ? void 0 : _c.supportedReport) === null || _d === void 0 ? void 0 : _d.map((sr) => Object.keys(sr.report)[0])) !== null && _e !== void 0 ? _e : []);
|
||
};
|
||
const isCollectionDirty = async (params) => {
|
||
var _a, _b, _c;
|
||
const { collection, headers, headersToExclude } = params;
|
||
const responses = await propfind({
|
||
url: collection.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.CALENDAR_SERVER}:getctag`]: {},
|
||
},
|
||
depth: '0',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
const res = responses.filter((r) => urlContains(collection.url, r.href))[0];
|
||
if (!res) {
|
||
throw new Error('Collection does not exist on server');
|
||
}
|
||
return {
|
||
isDirty: collection.ctag !== ((_a = res.props) === null || _a === void 0 ? void 0 : _a.getctag),
|
||
newCtag: (_c = (_b = res.props) === null || _b === void 0 ? void 0 : _b.getctag) === null || _c === void 0 ? void 0 : _c.toString(),
|
||
};
|
||
};
|
||
/**
|
||
* This is for webdav sync-collection only
|
||
*/
|
||
const syncCollection = (params) => {
|
||
const { url, props, headers, syncLevel, syncToken, headersToExclude } = params;
|
||
return davRequest({
|
||
url,
|
||
init: {
|
||
method: 'REPORT',
|
||
namespace: DAVNamespaceShort.DAV,
|
||
headers: excludeHeaders({ ...headers }, headersToExclude),
|
||
body: {
|
||
'sync-collection': {
|
||
_attributes: getDAVAttribute([
|
||
DAVNamespace.CALDAV,
|
||
DAVNamespace.CARDDAV,
|
||
DAVNamespace.DAV,
|
||
]),
|
||
'sync-level': syncLevel,
|
||
'sync-token': syncToken,
|
||
[`${DAVNamespaceShort.DAV}:prop`]: props,
|
||
},
|
||
},
|
||
},
|
||
});
|
||
};
|
||
/** remote collection to local */
|
||
const smartCollectionSync = async (params) => {
|
||
var _a, _b, _c, _d, _e, _f, _g, _h, _j, _k, _l;
|
||
const { collection, method, headers, headersToExclude, account, detailedResult } = params;
|
||
const requiredFields = ['accountType', 'homeUrl'];
|
||
if (!account || !hasFields(account, requiredFields)) {
|
||
if (!account) {
|
||
throw new Error('no account for smartCollectionSync');
|
||
}
|
||
throw new Error(`account must have ${findMissingFieldNames(account, requiredFields)} before smartCollectionSync`);
|
||
}
|
||
const syncMethod = method !== null && method !== void 0 ? method : (((_a = collection.reports) === null || _a === void 0 ? void 0 : _a.includes('syncCollection')) ? 'webdav' : 'basic');
|
||
debug$4(`smart collection sync with type ${account.accountType} and method ${syncMethod}`);
|
||
if (syncMethod === 'webdav') {
|
||
const result = await syncCollection({
|
||
url: collection.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${account.accountType === 'caldav' ? DAVNamespaceShort.CALDAV : DAVNamespaceShort.CARDDAV}:${account.accountType === 'caldav' ? 'calendar-data' : 'address-data'}`]: {},
|
||
[`${DAVNamespaceShort.DAV}:displayname`]: {},
|
||
},
|
||
syncLevel: 1,
|
||
syncToken: collection.syncToken,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
const objectResponses = result.filter((r) => {
|
||
var _a;
|
||
const extName = account.accountType === 'caldav' ? '.ics' : '.vcf';
|
||
return ((_a = r.href) === null || _a === void 0 ? void 0 : _a.slice(-4)) === extName;
|
||
});
|
||
const changedObjectUrls = objectResponses.filter((o) => o.status !== 404).map((r) => r.href);
|
||
const deletedObjectUrls = objectResponses.filter((o) => o.status === 404).map((r) => r.href);
|
||
const multiGetObjectResponse = changedObjectUrls.length
|
||
? ((_c = (await ((_b = collection === null || collection === void 0 ? void 0 : collection.objectMultiGet) === null || _b === void 0 ? void 0 : _b.call(collection, {
|
||
url: collection.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${account.accountType === 'caldav'
|
||
? DAVNamespaceShort.CALDAV
|
||
: DAVNamespaceShort.CARDDAV}:${account.accountType === 'caldav' ? 'calendar-data' : 'address-data'}`]: {},
|
||
},
|
||
objectUrls: changedObjectUrls,
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
})))) !== null && _c !== void 0 ? _c : [])
|
||
: [];
|
||
const remoteObjects = multiGetObjectResponse.map((res) => {
|
||
var _a, _b, _c, _d, _e, _f, _g, _h, _j, _k;
|
||
return {
|
||
url: (_a = res.href) !== null && _a !== void 0 ? _a : '',
|
||
etag: (_b = res.props) === null || _b === void 0 ? void 0 : _b.getetag,
|
||
data: (account === null || account === void 0 ? void 0 : account.accountType) === 'caldav'
|
||
? ((_e = (_d = (_c = res.props) === null || _c === void 0 ? void 0 : _c.calendarData) === null || _d === void 0 ? void 0 : _d._cdata) !== null && _e !== void 0 ? _e : (_f = res.props) === null || _f === void 0 ? void 0 : _f.calendarData)
|
||
: ((_j = (_h = (_g = res.props) === null || _g === void 0 ? void 0 : _g.addressData) === null || _h === void 0 ? void 0 : _h._cdata) !== null && _j !== void 0 ? _j : (_k = res.props) === null || _k === void 0 ? void 0 : _k.addressData),
|
||
};
|
||
});
|
||
const localObjects = (_d = collection.objects) !== null && _d !== void 0 ? _d : [];
|
||
// no existing url
|
||
const created = remoteObjects.filter((o) => localObjects.every((lo) => !urlContains(lo.url, o.url)));
|
||
// debug(`created objects: ${created.map((o) => o.url).join('\n')}`);
|
||
// have same url, but etag different
|
||
const updated = localObjects.reduce((prev, curr) => {
|
||
const found = remoteObjects.find((ro) => urlContains(ro.url, curr.url));
|
||
if (found && found.etag && found.etag !== curr.etag) {
|
||
return [...prev, found];
|
||
}
|
||
return prev;
|
||
}, []);
|
||
// debug(`updated objects: ${updated.map((o) => o.url).join('\n')}`);
|
||
const deleted = deletedObjectUrls.map((o) => ({
|
||
url: o,
|
||
etag: '',
|
||
}));
|
||
// debug(`deleted objects: ${deleted.map((o) => o.url).join('\n')}`);
|
||
const unchanged = localObjects.filter((lo) => remoteObjects.some((ro) => urlContains(lo.url, ro.url) && ro.etag === lo.etag));
|
||
return {
|
||
...collection,
|
||
objects: detailedResult
|
||
? { created, updated, deleted }
|
||
: [...unchanged, ...created, ...updated],
|
||
// all syncToken in the results are the same so we use the first one here
|
||
syncToken: (_h = (_g = (_f = (_e = result[0]) === null || _e === void 0 ? void 0 : _e.raw) === null || _f === void 0 ? void 0 : _f.multistatus) === null || _g === void 0 ? void 0 : _g.syncToken) !== null && _h !== void 0 ? _h : collection.syncToken,
|
||
};
|
||
}
|
||
if (syncMethod === 'basic') {
|
||
const { isDirty, newCtag } = await isCollectionDirty({
|
||
collection,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
const localObjects = (_j = collection.objects) !== null && _j !== void 0 ? _j : [];
|
||
const remoteObjects = (_l = (await ((_k = collection.fetchObjects) === null || _k === void 0 ? void 0 : _k.call(collection, {
|
||
collection,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
})))) !== null && _l !== void 0 ? _l : [];
|
||
// no existing url
|
||
const created = remoteObjects.filter((ro) => localObjects.every((lo) => !urlContains(lo.url, ro.url)));
|
||
// debug(`created objects: ${created.map((o) => o.url).join('\n')}`);
|
||
// have same url, but etag different
|
||
const updated = localObjects.reduce((prev, curr) => {
|
||
const found = remoteObjects.find((ro) => urlContains(ro.url, curr.url));
|
||
if (found && found.etag && found.etag !== curr.etag) {
|
||
return [...prev, found];
|
||
}
|
||
return prev;
|
||
}, []);
|
||
// debug(`updated objects: ${updated.map((o) => o.url).join('\n')}`);
|
||
// does not present in remote
|
||
const deleted = localObjects.filter((cal) => remoteObjects.every((ro) => !urlContains(ro.url, cal.url)));
|
||
// debug(`deleted objects: ${deleted.map((o) => o.url).join('\n')}`);
|
||
const unchanged = localObjects.filter((lo) => remoteObjects.some((ro) => urlContains(lo.url, ro.url) && ro.etag === lo.etag));
|
||
if (isDirty) {
|
||
return {
|
||
...collection,
|
||
objects: detailedResult
|
||
? { created, updated, deleted }
|
||
: [...unchanged, ...created, ...updated],
|
||
ctag: newCtag,
|
||
};
|
||
}
|
||
}
|
||
return detailedResult
|
||
? {
|
||
...collection,
|
||
objects: {
|
||
created: [],
|
||
updated: [],
|
||
deleted: [],
|
||
},
|
||
}
|
||
: collection;
|
||
};
|
||
|
||
var collection = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
collectionQuery: collectionQuery,
|
||
isCollectionDirty: isCollectionDirty,
|
||
makeCollection: makeCollection,
|
||
smartCollectionSync: smartCollectionSync,
|
||
supportedReportSet: supportedReportSet,
|
||
syncCollection: syncCollection
|
||
});
|
||
|
||
const debug$3 = getLogger('tsdav:addressBook');
|
||
const addressBookQuery = async (params) => {
|
||
const { url, props, filters, depth, headers, headersToExclude } = params;
|
||
return collectionQuery({
|
||
url,
|
||
body: {
|
||
'addressbook-query': {
|
||
_attributes: getDAVAttribute([DAVNamespace.CARDDAV, DAVNamespace.DAV]),
|
||
[`${DAVNamespaceShort.DAV}:prop`]: props,
|
||
filter: filters !== null && filters !== void 0 ? filters : {
|
||
'prop-filter': {
|
||
_attributes: {
|
||
name: 'FN',
|
||
},
|
||
},
|
||
},
|
||
},
|
||
},
|
||
defaultNamespace: DAVNamespaceShort.CARDDAV,
|
||
depth,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
const addressBookMultiGet = async (params) => {
|
||
const { url, props, objectUrls, depth, headers } = params;
|
||
return collectionQuery({
|
||
url,
|
||
body: {
|
||
'addressbook-multiget': {
|
||
_attributes: getDAVAttribute([DAVNamespace.DAV, DAVNamespace.CARDDAV]),
|
||
[`${DAVNamespaceShort.DAV}:prop`]: props,
|
||
[`${DAVNamespaceShort.DAV}:href`]: objectUrls,
|
||
},
|
||
},
|
||
defaultNamespace: DAVNamespaceShort.CARDDAV,
|
||
depth,
|
||
headers,
|
||
});
|
||
};
|
||
const fetchAddressBooks = async (params) => {
|
||
const { account, headers, props: customProps, headersToExclude } = params !== null && params !== void 0 ? params : {};
|
||
const requiredFields = ['homeUrl', 'rootUrl'];
|
||
if (!account || !hasFields(account, requiredFields)) {
|
||
if (!account) {
|
||
throw new Error('no account for fetchAddressBooks');
|
||
}
|
||
throw new Error(`account must have ${findMissingFieldNames(account, requiredFields)} before fetchAddressBooks`);
|
||
}
|
||
const res = await propfind({
|
||
url: account.homeUrl,
|
||
props: customProps !== null && customProps !== void 0 ? customProps : {
|
||
[`${DAVNamespaceShort.DAV}:displayname`]: {},
|
||
[`${DAVNamespaceShort.CALENDAR_SERVER}:getctag`]: {},
|
||
[`${DAVNamespaceShort.DAV}:resourcetype`]: {},
|
||
[`${DAVNamespaceShort.DAV}:sync-token`]: {},
|
||
},
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
return Promise.all(res
|
||
.filter((r) => { var _a, _b; return Object.keys((_b = (_a = r.props) === null || _a === void 0 ? void 0 : _a.resourcetype) !== null && _b !== void 0 ? _b : {}).includes('addressbook'); })
|
||
.map((rs) => {
|
||
var _a, _b, _c, _d, _e, _f, _g, _h, _j;
|
||
const displayName = (_c = (_b = (_a = rs.props) === null || _a === void 0 ? void 0 : _a.displayname) === null || _b === void 0 ? void 0 : _b._cdata) !== null && _c !== void 0 ? _c : (_d = rs.props) === null || _d === void 0 ? void 0 : _d.displayname;
|
||
debug$3(`Found address book named ${typeof displayName === 'string' ? displayName : ''},
|
||
props: ${JSON.stringify(rs.props)}`);
|
||
return {
|
||
url: new URL((_e = rs.href) !== null && _e !== void 0 ? _e : '', (_f = account.rootUrl) !== null && _f !== void 0 ? _f : '').href,
|
||
ctag: (_g = rs.props) === null || _g === void 0 ? void 0 : _g.getctag,
|
||
displayName: typeof displayName === 'string' ? displayName : '',
|
||
resourcetype: Object.keys((_h = rs.props) === null || _h === void 0 ? void 0 : _h.resourcetype),
|
||
syncToken: (_j = rs.props) === null || _j === void 0 ? void 0 : _j.syncToken,
|
||
};
|
||
})
|
||
.map(async (addr) => ({
|
||
...addr,
|
||
reports: await supportedReportSet({ collection: addr, headers }),
|
||
})));
|
||
};
|
||
const fetchVCards = async (params) => {
|
||
const { addressBook, headers, objectUrls, headersToExclude, urlFilter = (url) => url, useMultiGet = true, } = params;
|
||
debug$3(`Fetching vcards from ${addressBook === null || addressBook === void 0 ? void 0 : addressBook.url}`);
|
||
const requiredFields = ['url'];
|
||
if (!addressBook || !hasFields(addressBook, requiredFields)) {
|
||
if (!addressBook) {
|
||
throw new Error('cannot fetchVCards for undefined addressBook');
|
||
}
|
||
throw new Error(`addressBook must have ${findMissingFieldNames(addressBook, requiredFields)} before fetchVCards`);
|
||
}
|
||
const vcardUrls = (objectUrls !== null && objectUrls !== void 0 ? objectUrls :
|
||
// fetch all objects of the calendar
|
||
(await addressBookQuery({
|
||
url: addressBook.url,
|
||
props: { [`${DAVNamespaceShort.DAV}:getetag`]: {} },
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
})).map((res) => { var _a; return (res.ok ? ((_a = res.href) !== null && _a !== void 0 ? _a : '') : ''); }))
|
||
.map((url) => (url.startsWith('http') || !url ? url : new URL(url, addressBook.url).href))
|
||
.filter(urlFilter)
|
||
.map((url) => new URL(url).pathname);
|
||
let vCardResults = [];
|
||
if (vcardUrls.length > 0) {
|
||
if (useMultiGet) {
|
||
vCardResults = await addressBookMultiGet({
|
||
url: addressBook.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${DAVNamespaceShort.CARDDAV}:address-data`]: {},
|
||
},
|
||
objectUrls: vcardUrls,
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
}
|
||
else {
|
||
vCardResults = await addressBookQuery({
|
||
url: addressBook.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${DAVNamespaceShort.CARDDAV}:address-data`]: {},
|
||
},
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
}
|
||
}
|
||
return vCardResults.map((res) => {
|
||
var _a, _b, _c, _d, _e, _f;
|
||
return ({
|
||
url: new URL((_a = res.href) !== null && _a !== void 0 ? _a : '', addressBook.url).href,
|
||
etag: (_b = res.props) === null || _b === void 0 ? void 0 : _b.getetag,
|
||
data: (_e = (_d = (_c = res.props) === null || _c === void 0 ? void 0 : _c.addressData) === null || _d === void 0 ? void 0 : _d._cdata) !== null && _e !== void 0 ? _e : (_f = res.props) === null || _f === void 0 ? void 0 : _f.addressData,
|
||
});
|
||
});
|
||
};
|
||
const createVCard = async (params) => {
|
||
const { addressBook, vCardString, filename, headers, headersToExclude } = params;
|
||
return createObject({
|
||
url: new URL(filename, addressBook.url).href,
|
||
data: vCardString,
|
||
headers: excludeHeaders({
|
||
'content-type': 'text/vcard; charset=utf-8',
|
||
'If-None-Match': '*',
|
||
...headers,
|
||
}, headersToExclude),
|
||
});
|
||
};
|
||
const updateVCard = async (params) => {
|
||
const { vCard, headers, headersToExclude } = params;
|
||
return updateObject({
|
||
url: vCard.url,
|
||
data: vCard.data,
|
||
etag: vCard.etag,
|
||
headers: excludeHeaders({
|
||
'content-type': 'text/vcard; charset=utf-8',
|
||
...headers,
|
||
}, headersToExclude),
|
||
});
|
||
};
|
||
const deleteVCard = async (params) => {
|
||
const { vCard, headers, headersToExclude } = params;
|
||
return deleteObject({
|
||
url: vCard.url,
|
||
etag: vCard.etag,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
|
||
var addressBook = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
addressBookMultiGet: addressBookMultiGet,
|
||
addressBookQuery: addressBookQuery,
|
||
createVCard: createVCard,
|
||
deleteVCard: deleteVCard,
|
||
fetchAddressBooks: fetchAddressBooks,
|
||
fetchVCards: fetchVCards,
|
||
updateVCard: updateVCard
|
||
});
|
||
|
||
const debug$2 = getLogger('tsdav:calendar');
|
||
const calendarQuery = async (params) => {
|
||
const { url, props, filters, timezone, depth, headers, headersToExclude } = params;
|
||
return collectionQuery({
|
||
url,
|
||
body: {
|
||
'calendar-query': cleanupFalsy({
|
||
_attributes: getDAVAttribute([
|
||
DAVNamespace.CALDAV,
|
||
DAVNamespace.CALENDAR_SERVER,
|
||
DAVNamespace.CALDAV_APPLE,
|
||
DAVNamespace.DAV,
|
||
]),
|
||
[`${DAVNamespaceShort.DAV}:prop`]: props,
|
||
filter: filters,
|
||
timezone,
|
||
}),
|
||
},
|
||
defaultNamespace: DAVNamespaceShort.CALDAV,
|
||
depth,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
const calendarMultiGet = async (params) => {
|
||
const { url, props, objectUrls, filters, timezone, depth, headers, headersToExclude } = params;
|
||
return collectionQuery({
|
||
url,
|
||
body: {
|
||
'calendar-multiget': {
|
||
_attributes: getDAVAttribute([DAVNamespace.DAV, DAVNamespace.CALDAV]),
|
||
[`${DAVNamespaceShort.DAV}:prop`]: props,
|
||
[`${DAVNamespaceShort.DAV}:href`]: objectUrls,
|
||
...conditionalParam('filter', filters),
|
||
timezone,
|
||
},
|
||
},
|
||
defaultNamespace: DAVNamespaceShort.CALDAV,
|
||
depth,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
const makeCalendar = async (params) => {
|
||
const { url, props, depth, headers, headersToExclude } = params;
|
||
return davRequest({
|
||
url,
|
||
init: {
|
||
method: 'MKCALENDAR',
|
||
headers: excludeHeaders(cleanupFalsy({ depth, ...headers }), headersToExclude),
|
||
namespace: DAVNamespaceShort.DAV,
|
||
body: {
|
||
[`${DAVNamespaceShort.CALDAV}:mkcalendar`]: {
|
||
_attributes: getDAVAttribute([
|
||
DAVNamespace.DAV,
|
||
DAVNamespace.CALDAV,
|
||
DAVNamespace.CALDAV_APPLE,
|
||
]),
|
||
set: {
|
||
prop: props,
|
||
},
|
||
},
|
||
},
|
||
},
|
||
});
|
||
};
|
||
const fetchCalendars = async (params) => {
|
||
const { headers, account, props: customProps, projectedProps, headersToExclude } = params !== null && params !== void 0 ? params : {};
|
||
const requiredFields = ['homeUrl', 'rootUrl'];
|
||
if (!account || !hasFields(account, requiredFields)) {
|
||
if (!account) {
|
||
throw new Error('no account for fetchCalendars');
|
||
}
|
||
throw new Error(`account must have ${findMissingFieldNames(account, requiredFields)} before fetchCalendars`);
|
||
}
|
||
const res = await propfind({
|
||
url: account.homeUrl,
|
||
props: customProps !== null && customProps !== void 0 ? customProps : {
|
||
[`${DAVNamespaceShort.CALDAV}:calendar-description`]: {},
|
||
[`${DAVNamespaceShort.CALDAV}:calendar-timezone`]: {},
|
||
[`${DAVNamespaceShort.DAV}:displayname`]: {},
|
||
[`${DAVNamespaceShort.CALDAV_APPLE}:calendar-color`]: {},
|
||
[`${DAVNamespaceShort.CALENDAR_SERVER}:getctag`]: {},
|
||
[`${DAVNamespaceShort.DAV}:resourcetype`]: {},
|
||
[`${DAVNamespaceShort.CALDAV}:supported-calendar-component-set`]: {},
|
||
[`${DAVNamespaceShort.DAV}:sync-token`]: {},
|
||
},
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
return Promise.all(res
|
||
.filter((r) => { var _a, _b; return Object.keys((_b = (_a = r.props) === null || _a === void 0 ? void 0 : _a.resourcetype) !== null && _b !== void 0 ? _b : {}).includes('calendar'); })
|
||
.filter((rc) => {
|
||
var _a, _b, _c, _d;
|
||
// filter out none iCal format calendars.
|
||
const components = (Array.isArray((_a = rc.props) === null || _a === void 0 ? void 0 : _a.supportedCalendarComponentSet.comp)
|
||
? (_b = rc.props) === null || _b === void 0 ? void 0 : _b.supportedCalendarComponentSet.comp.map((sc) => sc._attributes.name)
|
||
: [(_d = (_c = rc.props) === null || _c === void 0 ? void 0 : _c.supportedCalendarComponentSet.comp) === null || _d === void 0 ? void 0 : _d._attributes.name]) || [];
|
||
return components.some((c) => Object.values(ICALObjects).includes(c));
|
||
})
|
||
.map((rs) => {
|
||
var _a, _b, _c, _d, _e, _f, _g, _h, _j, _k, _l, _m, _o, _p, _q, _r;
|
||
// debug(`Found calendar ${rs.props?.displayname}`);
|
||
const description = (_a = rs.props) === null || _a === void 0 ? void 0 : _a.calendarDescription;
|
||
const timezone = (_b = rs.props) === null || _b === void 0 ? void 0 : _b.calendarTimezone;
|
||
return {
|
||
description: typeof description === 'string' ? description : '',
|
||
timezone: typeof timezone === 'string' ? timezone : '',
|
||
url: new URL((_c = rs.href) !== null && _c !== void 0 ? _c : '', (_d = account.rootUrl) !== null && _d !== void 0 ? _d : '').href,
|
||
ctag: (_e = rs.props) === null || _e === void 0 ? void 0 : _e.getctag,
|
||
calendarColor: (_f = rs.props) === null || _f === void 0 ? void 0 : _f.calendarColor,
|
||
displayName: (_h = (_g = rs.props) === null || _g === void 0 ? void 0 : _g.displayname._cdata) !== null && _h !== void 0 ? _h : (_j = rs.props) === null || _j === void 0 ? void 0 : _j.displayname,
|
||
components: Array.isArray((_k = rs.props) === null || _k === void 0 ? void 0 : _k.supportedCalendarComponentSet.comp)
|
||
? (_l = rs.props) === null || _l === void 0 ? void 0 : _l.supportedCalendarComponentSet.comp.map((sc) => sc._attributes.name)
|
||
: [(_o = (_m = rs.props) === null || _m === void 0 ? void 0 : _m.supportedCalendarComponentSet.comp) === null || _o === void 0 ? void 0 : _o._attributes.name],
|
||
resourcetype: Object.keys((_p = rs.props) === null || _p === void 0 ? void 0 : _p.resourcetype),
|
||
syncToken: (_q = rs.props) === null || _q === void 0 ? void 0 : _q.syncToken,
|
||
...conditionalParam('projectedProps', Object.fromEntries(Object.entries((_r = rs.props) !== null && _r !== void 0 ? _r : {}).filter(([key]) => projectedProps === null || projectedProps === void 0 ? void 0 : projectedProps[key]))),
|
||
};
|
||
})
|
||
.map(async (cal) => ({
|
||
...cal,
|
||
reports: await supportedReportSet({
|
||
collection: cal,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}),
|
||
})));
|
||
};
|
||
const fetchCalendarObjects = async (params) => {
|
||
const { calendar, objectUrls, filters: customFilters, timeRange, headers, expand, urlFilter = (url) => Boolean(url === null || url === void 0 ? void 0 : url.includes('.ics')), useMultiGet = true, headersToExclude, } = params;
|
||
if (timeRange) {
|
||
// validate timeRange
|
||
const ISO_8601 = /^\d{4}(-\d\d(-\d\d(T\d\d:\d\d(:\d\d)?(\.\d+)?(([+-]\d\d:\d\d)|Z)?)?)?)?$/i;
|
||
const ISO_8601_FULL = /^\d{4}-\d\d-\d\dT\d\d:\d\d:\d\d(\.\d+)?(([+-]\d\d:\d\d)|Z)?$/i;
|
||
if ((!ISO_8601.test(timeRange.start) || !ISO_8601.test(timeRange.end)) &&
|
||
(!ISO_8601_FULL.test(timeRange.start) || !ISO_8601_FULL.test(timeRange.end))) {
|
||
throw new Error('invalid timeRange format, not in ISO8601');
|
||
}
|
||
}
|
||
debug$2(`Fetching calendar objects from ${calendar === null || calendar === void 0 ? void 0 : calendar.url}`);
|
||
const requiredFields = ['url'];
|
||
if (!calendar || !hasFields(calendar, requiredFields)) {
|
||
if (!calendar) {
|
||
throw new Error('cannot fetchCalendarObjects for undefined calendar');
|
||
}
|
||
throw new Error(`calendar must have ${findMissingFieldNames(calendar, requiredFields)} before fetchCalendarObjects`);
|
||
}
|
||
// default to fetch all
|
||
const filters = customFilters !== null && customFilters !== void 0 ? customFilters : [
|
||
{
|
||
'comp-filter': {
|
||
_attributes: {
|
||
name: 'VCALENDAR',
|
||
},
|
||
'comp-filter': {
|
||
_attributes: {
|
||
name: 'VEVENT',
|
||
},
|
||
...(timeRange
|
||
? {
|
||
'time-range': {
|
||
_attributes: {
|
||
start: `${new Date(timeRange.start)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
end: `${new Date(timeRange.end)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
},
|
||
},
|
||
}
|
||
: {}),
|
||
},
|
||
},
|
||
},
|
||
];
|
||
const calendarObjectUrls = (objectUrls !== null && objectUrls !== void 0 ? objectUrls :
|
||
// fetch all objects of the calendar
|
||
(await calendarQuery({
|
||
url: calendar.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {
|
||
...(expand && timeRange
|
||
? {
|
||
[`${DAVNamespaceShort.CALDAV}:expand`]: {
|
||
_attributes: {
|
||
start: `${new Date(timeRange.start)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
end: `${new Date(timeRange.end)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
},
|
||
},
|
||
}
|
||
: {}),
|
||
},
|
||
},
|
||
filters,
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
})).map((res) => { var _a; return (_a = res.href) !== null && _a !== void 0 ? _a : ''; }))
|
||
.map((url) => (url.startsWith('http') || !url ? url : new URL(url, calendar.url).href)) // patch up to full url if url is not full
|
||
.filter(urlFilter) // custom filter function on calendar objects
|
||
.map((url) => new URL(url).pathname); // obtain pathname of the url
|
||
let calendarObjectResults = [];
|
||
if (calendarObjectUrls.length > 0) {
|
||
if (!useMultiGet || expand) {
|
||
calendarObjectResults = await calendarQuery({
|
||
url: calendar.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${DAVNamespaceShort.CALDAV}:calendar-data`]: {
|
||
...(expand && timeRange
|
||
? {
|
||
[`${DAVNamespaceShort.CALDAV}:expand`]: {
|
||
_attributes: {
|
||
start: `${new Date(timeRange.start)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
end: `${new Date(timeRange.end)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
},
|
||
},
|
||
}
|
||
: {}),
|
||
},
|
||
},
|
||
filters,
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
}
|
||
else {
|
||
calendarObjectResults = await calendarMultiGet({
|
||
url: calendar.url,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:getetag`]: {},
|
||
[`${DAVNamespaceShort.CALDAV}:calendar-data`]: {
|
||
...(expand && timeRange
|
||
? {
|
||
[`${DAVNamespaceShort.CALDAV}:expand`]: {
|
||
_attributes: {
|
||
start: `${new Date(timeRange.start)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
end: `${new Date(timeRange.end)
|
||
.toISOString()
|
||
.slice(0, 19)
|
||
.replace(/[-:.]/g, '')}Z`,
|
||
},
|
||
},
|
||
}
|
||
: {}),
|
||
},
|
||
},
|
||
objectUrls: calendarObjectUrls,
|
||
depth: '1',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
}
|
||
}
|
||
return calendarObjectResults.map((res) => {
|
||
var _a, _b, _c, _d, _e, _f;
|
||
return ({
|
||
url: new URL((_a = res.href) !== null && _a !== void 0 ? _a : '', calendar.url).href,
|
||
etag: `${(_b = res.props) === null || _b === void 0 ? void 0 : _b.getetag}`,
|
||
data: (_e = (_d = (_c = res.props) === null || _c === void 0 ? void 0 : _c.calendarData) === null || _d === void 0 ? void 0 : _d._cdata) !== null && _e !== void 0 ? _e : (_f = res.props) === null || _f === void 0 ? void 0 : _f.calendarData,
|
||
});
|
||
});
|
||
};
|
||
const createCalendarObject = async (params) => {
|
||
const { calendar, iCalString, filename, headers, headersToExclude } = params;
|
||
return createObject({
|
||
url: new URL(filename, calendar.url).href,
|
||
data: iCalString,
|
||
headers: excludeHeaders({
|
||
'content-type': 'text/calendar; charset=utf-8',
|
||
'If-None-Match': '*',
|
||
...headers,
|
||
}, headersToExclude),
|
||
});
|
||
};
|
||
const updateCalendarObject = async (params) => {
|
||
const { calendarObject, headers, headersToExclude } = params;
|
||
return updateObject({
|
||
url: calendarObject.url,
|
||
data: calendarObject.data,
|
||
etag: calendarObject.etag,
|
||
headers: excludeHeaders({
|
||
'content-type': 'text/calendar; charset=utf-8',
|
||
...headers,
|
||
}, headersToExclude),
|
||
});
|
||
};
|
||
const deleteCalendarObject = async (params) => {
|
||
const { calendarObject, headers, headersToExclude } = params;
|
||
return deleteObject({
|
||
url: calendarObject.url,
|
||
etag: calendarObject.etag,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
};
|
||
/**
|
||
* Sync remote calendars to local
|
||
*/
|
||
const syncCalendars = async (params) => {
|
||
var _a;
|
||
const { oldCalendars, account, detailedResult, headers, headersToExclude } = params;
|
||
if (!account) {
|
||
throw new Error('Must have account before syncCalendars');
|
||
}
|
||
const localCalendars = (_a = oldCalendars !== null && oldCalendars !== void 0 ? oldCalendars : account.calendars) !== null && _a !== void 0 ? _a : [];
|
||
const remoteCalendars = await fetchCalendars({
|
||
account,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
// no existing url
|
||
const created = remoteCalendars.filter((rc) => localCalendars.every((lc) => !urlContains(lc.url, rc.url)));
|
||
debug$2(`new calendars: ${created.map((cc) => cc.displayName)}`);
|
||
// have same url, but syncToken/ctag different
|
||
const updated = localCalendars.reduce((prev, curr) => {
|
||
const found = remoteCalendars.find((rc) => urlContains(rc.url, curr.url));
|
||
if (found &&
|
||
((found.syncToken && found.syncToken !== curr.syncToken) ||
|
||
(found.ctag && found.ctag !== curr.ctag))) {
|
||
return [...prev, found];
|
||
}
|
||
return prev;
|
||
}, []);
|
||
debug$2(`updated calendars: ${updated.map((cc) => cc.displayName)}`);
|
||
const updatedWithObjects = await Promise.all(updated.map(async (u) => {
|
||
const result = await smartCollectionSync({
|
||
collection: { ...u, objectMultiGet: calendarMultiGet },
|
||
method: 'webdav',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
account,
|
||
});
|
||
return result;
|
||
}));
|
||
// does not present in remote
|
||
const deleted = localCalendars.filter((cal) => remoteCalendars.every((rc) => !urlContains(rc.url, cal.url)));
|
||
debug$2(`deleted calendars: ${deleted.map((cc) => cc.displayName)}`);
|
||
const unchanged = localCalendars.filter((cal) => remoteCalendars.some((rc) => urlContains(rc.url, cal.url) &&
|
||
((rc.syncToken && rc.syncToken !== cal.syncToken) || (rc.ctag && rc.ctag !== cal.ctag))));
|
||
// debug(`unchanged calendars: ${unchanged.map((cc) => cc.displayName)}`);
|
||
return detailedResult
|
||
? {
|
||
created,
|
||
updated,
|
||
deleted,
|
||
}
|
||
: [...unchanged, ...created, ...updatedWithObjects];
|
||
};
|
||
const freeBusyQuery = async (params) => {
|
||
const { url, timeRange, depth, headers, headersToExclude } = params;
|
||
if (timeRange) {
|
||
// validate timeRange
|
||
const ISO_8601 = /^\d{4}(-\d\d(-\d\d(T\d\d:\d\d(:\d\d)?(\.\d+)?(([+-]\d\d:\d\d)|Z)?)?)?)?$/i;
|
||
const ISO_8601_FULL = /^\d{4}-\d\d-\d\dT\d\d:\d\d:\d\d(\.\d+)?(([+-]\d\d:\d\d)|Z)?$/i;
|
||
if ((!ISO_8601.test(timeRange.start) || !ISO_8601.test(timeRange.end)) &&
|
||
(!ISO_8601_FULL.test(timeRange.start) || !ISO_8601_FULL.test(timeRange.end))) {
|
||
throw new Error('invalid timeRange format, not in ISO8601');
|
||
}
|
||
}
|
||
else {
|
||
throw new Error('timeRange is required');
|
||
}
|
||
const result = await collectionQuery({
|
||
url,
|
||
body: {
|
||
'free-busy-query': cleanupFalsy({
|
||
_attributes: getDAVAttribute([DAVNamespace.CALDAV]),
|
||
[`${DAVNamespaceShort.CALDAV}:time-range`]: {
|
||
_attributes: {
|
||
start: `${new Date(timeRange.start).toISOString().slice(0, 19).replace(/[-:.]/g, '')}Z`,
|
||
end: `${new Date(timeRange.end).toISOString().slice(0, 19).replace(/[-:.]/g, '')}Z`,
|
||
},
|
||
},
|
||
}),
|
||
},
|
||
defaultNamespace: DAVNamespaceShort.CALDAV,
|
||
depth,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
return result[0];
|
||
};
|
||
|
||
var calendar = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
calendarMultiGet: calendarMultiGet,
|
||
calendarQuery: calendarQuery,
|
||
createCalendarObject: createCalendarObject,
|
||
deleteCalendarObject: deleteCalendarObject,
|
||
fetchCalendarObjects: fetchCalendarObjects,
|
||
fetchCalendars: fetchCalendars,
|
||
freeBusyQuery: freeBusyQuery,
|
||
makeCalendar: makeCalendar,
|
||
syncCalendars: syncCalendars,
|
||
updateCalendarObject: updateCalendarObject
|
||
});
|
||
|
||
const debug$1 = getLogger('tsdav:account');
|
||
const serviceDiscovery = async (params) => {
|
||
var _a, _b;
|
||
debug$1('Service discovery...');
|
||
const { account, headers, headersToExclude } = params;
|
||
const endpoint = new URL(account.serverUrl);
|
||
const uri = new URL(`/.well-known/${account.accountType}`, endpoint);
|
||
uri.protocol = (_a = endpoint.protocol) !== null && _a !== void 0 ? _a : 'http';
|
||
try {
|
||
const response = await browserPonyfillExports.fetch(uri.href, {
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
method: 'PROPFIND',
|
||
redirect: 'manual',
|
||
});
|
||
if (response.status >= 300 && response.status < 400) {
|
||
// http redirect.
|
||
const location = response.headers.get('Location');
|
||
if (typeof location === 'string' && location.length) {
|
||
debug$1(`Service discovery redirected to ${location}`);
|
||
const serviceURL = new URL(location, endpoint);
|
||
if (serviceURL.hostname === uri.hostname && uri.port && !serviceURL.port) {
|
||
serviceURL.port = uri.port;
|
||
}
|
||
serviceURL.protocol = (_b = endpoint.protocol) !== null && _b !== void 0 ? _b : 'http';
|
||
return serviceURL.href;
|
||
}
|
||
}
|
||
}
|
||
catch (err) {
|
||
debug$1(`Service discovery failed: ${err.stack}`);
|
||
}
|
||
return endpoint.href;
|
||
};
|
||
const fetchPrincipalUrl = async (params) => {
|
||
var _a, _b, _c, _d, _e;
|
||
const { account, headers, headersToExclude } = params;
|
||
const requiredFields = ['rootUrl'];
|
||
if (!hasFields(account, requiredFields)) {
|
||
throw new Error(`account must have ${findMissingFieldNames(account, requiredFields)} before fetchPrincipalUrl`);
|
||
}
|
||
debug$1(`Fetching principal url from path ${account.rootUrl}`);
|
||
const [response] = await propfind({
|
||
url: account.rootUrl,
|
||
props: {
|
||
[`${DAVNamespaceShort.DAV}:current-user-principal`]: {},
|
||
},
|
||
depth: '0',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
if (!response.ok) {
|
||
debug$1(`Fetch principal url failed: ${response.statusText}`);
|
||
if (response.status === 401) {
|
||
throw new Error('Invalid credentials');
|
||
}
|
||
}
|
||
debug$1(`Fetched principal url ${(_b = (_a = response.props) === null || _a === void 0 ? void 0 : _a.currentUserPrincipal) === null || _b === void 0 ? void 0 : _b.href}`);
|
||
return new URL((_e = (_d = (_c = response.props) === null || _c === void 0 ? void 0 : _c.currentUserPrincipal) === null || _d === void 0 ? void 0 : _d.href) !== null && _e !== void 0 ? _e : '', account.rootUrl).href;
|
||
};
|
||
const fetchHomeUrl = async (params) => {
|
||
var _a, _b;
|
||
const { account, headers, headersToExclude } = params;
|
||
const requiredFields = ['principalUrl', 'rootUrl'];
|
||
if (!hasFields(account, requiredFields)) {
|
||
throw new Error(`account must have ${findMissingFieldNames(account, requiredFields)} before fetchHomeUrl`);
|
||
}
|
||
debug$1(`Fetch home url from ${account.principalUrl}`);
|
||
const responses = await propfind({
|
||
url: account.principalUrl,
|
||
props: account.accountType === 'caldav'
|
||
? { [`${DAVNamespaceShort.CALDAV}:calendar-home-set`]: {} }
|
||
: { [`${DAVNamespaceShort.CARDDAV}:addressbook-home-set`]: {} },
|
||
depth: '0',
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
});
|
||
const matched = responses.find((r) => urlContains(account.principalUrl, r.href));
|
||
if (!matched || !matched.ok) {
|
||
throw new Error('cannot find homeUrl');
|
||
}
|
||
const result = new URL(account.accountType === 'caldav'
|
||
? (_a = matched === null || matched === void 0 ? void 0 : matched.props) === null || _a === void 0 ? void 0 : _a.calendarHomeSet.href
|
||
: (_b = matched === null || matched === void 0 ? void 0 : matched.props) === null || _b === void 0 ? void 0 : _b.addressbookHomeSet.href, account.rootUrl).href;
|
||
debug$1(`Fetched home url ${result}`);
|
||
return result;
|
||
};
|
||
const createAccount = async (params) => {
|
||
const { account, headers, loadCollections = false, loadObjects = false, headersToExclude, } = params;
|
||
const newAccount = { ...account };
|
||
newAccount.rootUrl =
|
||
newAccount.rootUrl ||
|
||
(await serviceDiscovery({
|
||
account,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}));
|
||
newAccount.principalUrl =
|
||
newAccount.principalUrl ||
|
||
(await fetchPrincipalUrl({
|
||
account: newAccount,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}));
|
||
newAccount.homeUrl =
|
||
newAccount.homeUrl ||
|
||
(await fetchHomeUrl({
|
||
account: newAccount,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}));
|
||
// to load objects you must first load collections
|
||
if (loadCollections || loadObjects) {
|
||
if (account.accountType === 'caldav') {
|
||
newAccount.calendars = await fetchCalendars({
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
account: newAccount,
|
||
});
|
||
}
|
||
else if (account.accountType === 'carddav') {
|
||
newAccount.addressBooks = await fetchAddressBooks({
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
account: newAccount,
|
||
});
|
||
}
|
||
}
|
||
if (loadObjects) {
|
||
if (account.accountType === 'caldav' && newAccount.calendars) {
|
||
newAccount.calendars = await Promise.all(newAccount.calendars.map(async (cal) => ({
|
||
...cal,
|
||
objects: await fetchCalendarObjects({
|
||
calendar: cal,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}),
|
||
})));
|
||
}
|
||
else if (account.accountType === 'carddav' && newAccount.addressBooks) {
|
||
newAccount.addressBooks = await Promise.all(newAccount.addressBooks.map(async (addr) => ({
|
||
...addr,
|
||
objects: await fetchVCards({
|
||
addressBook: addr,
|
||
headers: excludeHeaders(headers, headersToExclude),
|
||
}),
|
||
})));
|
||
}
|
||
}
|
||
return newAccount;
|
||
};
|
||
|
||
var account = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
createAccount: createAccount,
|
||
fetchHomeUrl: fetchHomeUrl,
|
||
fetchPrincipalUrl: fetchPrincipalUrl,
|
||
serviceDiscovery: serviceDiscovery
|
||
});
|
||
|
||
var base64$1 = {exports: {}};
|
||
|
||
/*! https://mths.be/base64 v1.0.0 by @mathias | MIT license */
|
||
var base64 = base64$1.exports;
|
||
|
||
var hasRequiredBase64;
|
||
|
||
function requireBase64 () {
|
||
if (hasRequiredBase64) return base64$1.exports;
|
||
hasRequiredBase64 = 1;
|
||
(function (module, exports) {
|
||
(function(root) {
|
||
|
||
// Detect free variables `exports`.
|
||
var freeExports = exports;
|
||
|
||
// Detect free variable `module`.
|
||
var freeModule = module &&
|
||
module.exports == freeExports && module;
|
||
|
||
// Detect free variable `global`, from Node.js or Browserified code, and use
|
||
// it as `root`.
|
||
var freeGlobal = typeof commonjsGlobal == 'object' && commonjsGlobal;
|
||
if (freeGlobal.global === freeGlobal || freeGlobal.window === freeGlobal) {
|
||
root = freeGlobal;
|
||
}
|
||
|
||
/*--------------------------------------------------------------------------*/
|
||
|
||
var InvalidCharacterError = function(message) {
|
||
this.message = message;
|
||
};
|
||
InvalidCharacterError.prototype = new Error;
|
||
InvalidCharacterError.prototype.name = 'InvalidCharacterError';
|
||
|
||
var error = function(message) {
|
||
// Note: the error messages used throughout this file match those used by
|
||
// the native `atob`/`btoa` implementation in Chromium.
|
||
throw new InvalidCharacterError(message);
|
||
};
|
||
|
||
var TABLE = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/';
|
||
// http://whatwg.org/html/common-microsyntaxes.html#space-character
|
||
var REGEX_SPACE_CHARACTERS = /[\t\n\f\r ]/g;
|
||
|
||
// `decode` is designed to be fully compatible with `atob` as described in the
|
||
// HTML Standard. http://whatwg.org/html/webappapis.html#dom-windowbase64-atob
|
||
// The optimized base64-decoding algorithm used is based on @atk’s excellent
|
||
// implementation. https://gist.github.com/atk/1020396
|
||
var decode = function(input) {
|
||
input = String(input)
|
||
.replace(REGEX_SPACE_CHARACTERS, '');
|
||
var length = input.length;
|
||
if (length % 4 == 0) {
|
||
input = input.replace(/==?$/, '');
|
||
length = input.length;
|
||
}
|
||
if (
|
||
length % 4 == 1 ||
|
||
// http://whatwg.org/C#alphanumeric-ascii-characters
|
||
/[^+a-zA-Z0-9/]/.test(input)
|
||
) {
|
||
error(
|
||
'Invalid character: the string to be decoded is not correctly encoded.'
|
||
);
|
||
}
|
||
var bitCounter = 0;
|
||
var bitStorage;
|
||
var buffer;
|
||
var output = '';
|
||
var position = -1;
|
||
while (++position < length) {
|
||
buffer = TABLE.indexOf(input.charAt(position));
|
||
bitStorage = bitCounter % 4 ? bitStorage * 64 + buffer : buffer;
|
||
// Unless this is the first of a group of 4 characters…
|
||
if (bitCounter++ % 4) {
|
||
// …convert the first 8 bits to a single ASCII character.
|
||
output += String.fromCharCode(
|
||
0xFF & bitStorage >> (-2 * bitCounter & 6)
|
||
);
|
||
}
|
||
}
|
||
return output;
|
||
};
|
||
|
||
// `encode` is designed to be fully compatible with `btoa` as described in the
|
||
// HTML Standard: http://whatwg.org/html/webappapis.html#dom-windowbase64-btoa
|
||
var encode = function(input) {
|
||
input = String(input);
|
||
if (/[^\0-\xFF]/.test(input)) {
|
||
// Note: no need to special-case astral symbols here, as surrogates are
|
||
// matched, and the input is supposed to only contain ASCII anyway.
|
||
error(
|
||
'The string to be encoded contains characters outside of the ' +
|
||
'Latin1 range.'
|
||
);
|
||
}
|
||
var padding = input.length % 3;
|
||
var output = '';
|
||
var position = -1;
|
||
var a;
|
||
var b;
|
||
var c;
|
||
var buffer;
|
||
// Make sure any padding is handled outside of the loop.
|
||
var length = input.length - padding;
|
||
|
||
while (++position < length) {
|
||
// Read three bytes, i.e. 24 bits.
|
||
a = input.charCodeAt(position) << 16;
|
||
b = input.charCodeAt(++position) << 8;
|
||
c = input.charCodeAt(++position);
|
||
buffer = a + b + c;
|
||
// Turn the 24 bits into four chunks of 6 bits each, and append the
|
||
// matching character for each of them to the output.
|
||
output += (
|
||
TABLE.charAt(buffer >> 18 & 0x3F) +
|
||
TABLE.charAt(buffer >> 12 & 0x3F) +
|
||
TABLE.charAt(buffer >> 6 & 0x3F) +
|
||
TABLE.charAt(buffer & 0x3F)
|
||
);
|
||
}
|
||
|
||
if (padding == 2) {
|
||
a = input.charCodeAt(position) << 8;
|
||
b = input.charCodeAt(++position);
|
||
buffer = a + b;
|
||
output += (
|
||
TABLE.charAt(buffer >> 10) +
|
||
TABLE.charAt((buffer >> 4) & 0x3F) +
|
||
TABLE.charAt((buffer << 2) & 0x3F) +
|
||
'='
|
||
);
|
||
} else if (padding == 1) {
|
||
buffer = input.charCodeAt(position);
|
||
output += (
|
||
TABLE.charAt(buffer >> 2) +
|
||
TABLE.charAt((buffer << 4) & 0x3F) +
|
||
'=='
|
||
);
|
||
}
|
||
|
||
return output;
|
||
};
|
||
|
||
var base64 = {
|
||
'encode': encode,
|
||
'decode': decode,
|
||
'version': '1.0.0'
|
||
};
|
||
|
||
// Some AMD build optimizers, like r.js, check for specific condition patterns
|
||
// like the following:
|
||
if (freeExports && !freeExports.nodeType) {
|
||
if (freeModule) { // in Node.js or RingoJS v0.8.0+
|
||
freeModule.exports = base64;
|
||
} else { // in Narwhal or RingoJS v0.7.0-
|
||
for (var key in base64) {
|
||
base64.hasOwnProperty(key) && (freeExports[key] = base64[key]);
|
||
}
|
||
}
|
||
} else { // in Rhino or a web browser
|
||
root.base64 = base64;
|
||
}
|
||
|
||
}(base64));
|
||
} (base64$1, base64$1.exports));
|
||
return base64$1.exports;
|
||
}
|
||
|
||
var base64Exports = requireBase64();
|
||
|
||
const debug = getLogger('tsdav:authHelper');
|
||
/**
|
||
* Provide given params as default params to given function with optional params.
|
||
*
|
||
* suitable only for one param functions
|
||
* params are shallow merged
|
||
*/
|
||
const defaultParam = (fn, params) => (...args) => {
|
||
return fn({ ...params, ...args[0] });
|
||
};
|
||
const getBasicAuthHeaders = (credentials) => {
|
||
debug(`Basic auth token generated: ${base64Exports.encode(`${credentials.username}:${credentials.password}`)}`);
|
||
return {
|
||
authorization: `Basic ${base64Exports.encode(`${credentials.username}:${credentials.password}`)}`,
|
||
};
|
||
};
|
||
const fetchOauthTokens = async (credentials) => {
|
||
const requireFields = [
|
||
'authorizationCode',
|
||
'redirectUrl',
|
||
'clientId',
|
||
'clientSecret',
|
||
'tokenUrl',
|
||
];
|
||
if (!hasFields(credentials, requireFields)) {
|
||
throw new Error(`Oauth credentials missing: ${findMissingFieldNames(credentials, requireFields)}`);
|
||
}
|
||
const param = new URLSearchParams({
|
||
grant_type: 'authorization_code',
|
||
code: credentials.authorizationCode,
|
||
redirect_uri: credentials.redirectUrl,
|
||
client_id: credentials.clientId,
|
||
client_secret: credentials.clientSecret,
|
||
});
|
||
debug(credentials.tokenUrl);
|
||
debug(param.toString());
|
||
const response = await browserPonyfillExports.fetch(credentials.tokenUrl, {
|
||
method: 'POST',
|
||
body: param.toString(),
|
||
headers: {
|
||
'content-length': `${param.toString().length}`,
|
||
'content-type': 'application/x-www-form-urlencoded',
|
||
},
|
||
});
|
||
if (response.ok) {
|
||
const tokens = await response.json();
|
||
return tokens;
|
||
}
|
||
debug(`Fetch Oauth tokens failed: ${await response.text()}`);
|
||
return {};
|
||
};
|
||
const refreshAccessToken = async (credentials) => {
|
||
const requireFields = [
|
||
'refreshToken',
|
||
'clientId',
|
||
'clientSecret',
|
||
'tokenUrl',
|
||
];
|
||
if (!hasFields(credentials, requireFields)) {
|
||
throw new Error(`Oauth credentials missing: ${findMissingFieldNames(credentials, requireFields)}`);
|
||
}
|
||
const param = new URLSearchParams({
|
||
client_id: credentials.clientId,
|
||
client_secret: credentials.clientSecret,
|
||
refresh_token: credentials.refreshToken,
|
||
grant_type: 'refresh_token',
|
||
});
|
||
const response = await browserPonyfillExports.fetch(credentials.tokenUrl, {
|
||
method: 'POST',
|
||
body: param.toString(),
|
||
headers: {
|
||
'Content-Type': 'application/x-www-form-urlencoded',
|
||
},
|
||
});
|
||
if (response.ok) {
|
||
const tokens = await response.json();
|
||
return tokens;
|
||
}
|
||
debug(`Refresh access token failed: ${await response.text()}`);
|
||
return {};
|
||
};
|
||
const getOauthHeaders = async (credentials) => {
|
||
var _a;
|
||
debug('Fetching oauth headers');
|
||
let tokens = {};
|
||
if (!credentials.refreshToken) {
|
||
// No refresh token, fetch new tokens
|
||
tokens = await fetchOauthTokens(credentials);
|
||
}
|
||
else if ((credentials.refreshToken && !credentials.accessToken) ||
|
||
Date.now() > ((_a = credentials.expiration) !== null && _a !== void 0 ? _a : 0)) {
|
||
// have refresh token, but no accessToken, fetch access token only
|
||
// or have both, but accessToken was expired
|
||
tokens = await refreshAccessToken(credentials);
|
||
}
|
||
// now we should have valid access token
|
||
debug(`Oauth tokens fetched: ${tokens.access_token}`);
|
||
return {
|
||
tokens,
|
||
headers: {
|
||
authorization: `Bearer ${tokens.access_token}`,
|
||
},
|
||
};
|
||
};
|
||
|
||
var authHelpers = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
defaultParam: defaultParam,
|
||
fetchOauthTokens: fetchOauthTokens,
|
||
getBasicAuthHeaders: getBasicAuthHeaders,
|
||
getOauthHeaders: getOauthHeaders,
|
||
refreshAccessToken: refreshAccessToken
|
||
});
|
||
|
||
const createDAVClient = async (params) => {
|
||
var _a;
|
||
const { serverUrl, credentials, authMethod, defaultAccountType, authFunction, rootUrl, principalUrl, homeUrl, } = params;
|
||
let authHeaders = {};
|
||
switch (authMethod) {
|
||
case 'Basic':
|
||
authHeaders = getBasicAuthHeaders(credentials);
|
||
break;
|
||
case 'Oauth':
|
||
authHeaders = (await getOauthHeaders(credentials)).headers;
|
||
break;
|
||
case 'Digest':
|
||
authHeaders = {
|
||
Authorization: `Digest ${credentials.digestString}`,
|
||
};
|
||
break;
|
||
case 'Custom':
|
||
authHeaders = (_a = (await (authFunction === null || authFunction === void 0 ? void 0 : authFunction(credentials)))) !== null && _a !== void 0 ? _a : {};
|
||
break;
|
||
default:
|
||
throw new Error('Invalid auth method');
|
||
}
|
||
const defaultAccount = defaultAccountType
|
||
? await createAccount({
|
||
account: {
|
||
serverUrl,
|
||
credentials,
|
||
accountType: defaultAccountType,
|
||
rootUrl,
|
||
principalUrl,
|
||
homeUrl,
|
||
},
|
||
headers: authHeaders,
|
||
})
|
||
: undefined;
|
||
const davRequest$1 = async (params0) => {
|
||
const { init, ...rest } = params0;
|
||
const { headers, ...restInit } = init;
|
||
return davRequest({
|
||
...rest,
|
||
init: {
|
||
...restInit,
|
||
headers: {
|
||
...authHeaders,
|
||
...headers,
|
||
},
|
||
},
|
||
});
|
||
};
|
||
const createObject$1 = defaultParam(createObject, {
|
||
url: serverUrl,
|
||
headers: authHeaders,
|
||
});
|
||
const updateObject$1 = defaultParam(updateObject, { headers: authHeaders, url: serverUrl });
|
||
const deleteObject$1 = defaultParam(deleteObject, { headers: authHeaders, url: serverUrl });
|
||
const propfind$1 = defaultParam(propfind, { headers: authHeaders });
|
||
// account
|
||
const createAccount$1 = async (params0) => {
|
||
const { account, headers, loadCollections, loadObjects } = params0;
|
||
return createAccount({
|
||
account: { serverUrl, credentials, ...account },
|
||
headers: { ...authHeaders, ...headers },
|
||
loadCollections,
|
||
loadObjects,
|
||
});
|
||
};
|
||
// collection
|
||
const collectionQuery$1 = defaultParam(collectionQuery, { headers: authHeaders });
|
||
const makeCollection$1 = defaultParam(makeCollection, { headers: authHeaders });
|
||
const syncCollection$1 = defaultParam(syncCollection, { headers: authHeaders });
|
||
const supportedReportSet$1 = defaultParam(supportedReportSet, {
|
||
headers: authHeaders,
|
||
});
|
||
const isCollectionDirty$1 = defaultParam(isCollectionDirty, {
|
||
headers: authHeaders,
|
||
});
|
||
const smartCollectionSync$1 = defaultParam(smartCollectionSync, {
|
||
headers: authHeaders,
|
||
account: defaultAccount,
|
||
});
|
||
// calendar
|
||
const calendarQuery$1 = defaultParam(calendarQuery, { headers: authHeaders });
|
||
const calendarMultiGet$1 = defaultParam(calendarMultiGet, { headers: authHeaders });
|
||
const makeCalendar$1 = defaultParam(makeCalendar, { headers: authHeaders });
|
||
const fetchCalendars$1 = defaultParam(fetchCalendars, {
|
||
headers: authHeaders,
|
||
account: defaultAccount,
|
||
});
|
||
const fetchCalendarObjects$1 = defaultParam(fetchCalendarObjects, {
|
||
headers: authHeaders,
|
||
});
|
||
const createCalendarObject$1 = defaultParam(createCalendarObject, {
|
||
headers: authHeaders,
|
||
});
|
||
const updateCalendarObject$1 = defaultParam(updateCalendarObject, {
|
||
headers: authHeaders,
|
||
});
|
||
const deleteCalendarObject$1 = defaultParam(deleteCalendarObject, {
|
||
headers: authHeaders,
|
||
});
|
||
const syncCalendars$1 = defaultParam(syncCalendars, {
|
||
account: defaultAccount,
|
||
headers: authHeaders,
|
||
});
|
||
// addressBook
|
||
const addressBookQuery$1 = defaultParam(addressBookQuery, { headers: authHeaders });
|
||
const addressBookMultiGet$1 = defaultParam(addressBookMultiGet, { headers: authHeaders });
|
||
const fetchAddressBooks$1 = defaultParam(fetchAddressBooks, {
|
||
account: defaultAccount,
|
||
headers: authHeaders,
|
||
});
|
||
const fetchVCards$1 = defaultParam(fetchVCards, { headers: authHeaders });
|
||
const createVCard$1 = defaultParam(createVCard, { headers: authHeaders });
|
||
const updateVCard$1 = defaultParam(updateVCard, { headers: authHeaders });
|
||
const deleteVCard$1 = defaultParam(deleteVCard, { headers: authHeaders });
|
||
return {
|
||
davRequest: davRequest$1,
|
||
propfind: propfind$1,
|
||
createAccount: createAccount$1,
|
||
createObject: createObject$1,
|
||
updateObject: updateObject$1,
|
||
deleteObject: deleteObject$1,
|
||
calendarQuery: calendarQuery$1,
|
||
addressBookQuery: addressBookQuery$1,
|
||
collectionQuery: collectionQuery$1,
|
||
makeCollection: makeCollection$1,
|
||
calendarMultiGet: calendarMultiGet$1,
|
||
makeCalendar: makeCalendar$1,
|
||
syncCollection: syncCollection$1,
|
||
supportedReportSet: supportedReportSet$1,
|
||
isCollectionDirty: isCollectionDirty$1,
|
||
smartCollectionSync: smartCollectionSync$1,
|
||
fetchCalendars: fetchCalendars$1,
|
||
fetchCalendarObjects: fetchCalendarObjects$1,
|
||
createCalendarObject: createCalendarObject$1,
|
||
updateCalendarObject: updateCalendarObject$1,
|
||
deleteCalendarObject: deleteCalendarObject$1,
|
||
syncCalendars: syncCalendars$1,
|
||
fetchAddressBooks: fetchAddressBooks$1,
|
||
addressBookMultiGet: addressBookMultiGet$1,
|
||
fetchVCards: fetchVCards$1,
|
||
createVCard: createVCard$1,
|
||
updateVCard: updateVCard$1,
|
||
deleteVCard: deleteVCard$1,
|
||
defaultAccount,
|
||
};
|
||
};
|
||
class DAVClient {
|
||
constructor(params) {
|
||
var _a, _b;
|
||
this.serverUrl = params.serverUrl;
|
||
this.credentials = params.credentials;
|
||
this.authMethod = (_a = params.authMethod) !== null && _a !== void 0 ? _a : 'Basic';
|
||
this.accountType = (_b = params.defaultAccountType) !== null && _b !== void 0 ? _b : 'caldav';
|
||
}
|
||
async login() {
|
||
var _a;
|
||
switch (this.authMethod) {
|
||
case 'Basic':
|
||
this.authHeaders = getBasicAuthHeaders(this.credentials);
|
||
break;
|
||
case 'Oauth':
|
||
this.authHeaders = (await getOauthHeaders(this.credentials)).headers;
|
||
break;
|
||
case 'Digest':
|
||
this.authHeaders = {
|
||
Authorization: `Digest ${this.credentials.digestString}`,
|
||
};
|
||
break;
|
||
case 'Custom':
|
||
this.authHeaders = await ((_a = this.authFunction) === null || _a === void 0 ? void 0 : _a.call(this, this.credentials));
|
||
break;
|
||
default:
|
||
throw new Error('Invalid auth method');
|
||
}
|
||
this.account = this.accountType
|
||
? await createAccount({
|
||
account: {
|
||
serverUrl: this.serverUrl,
|
||
credentials: this.credentials,
|
||
accountType: this.accountType,
|
||
},
|
||
headers: this.authHeaders,
|
||
})
|
||
: undefined;
|
||
}
|
||
async davRequest(params0) {
|
||
const { init, ...rest } = params0;
|
||
const { headers, ...restInit } = init;
|
||
return davRequest({
|
||
...rest,
|
||
init: {
|
||
...restInit,
|
||
headers: {
|
||
...this.authHeaders,
|
||
...headers,
|
||
},
|
||
},
|
||
});
|
||
}
|
||
async createObject(...params) {
|
||
return defaultParam(createObject, {
|
||
url: this.serverUrl,
|
||
headers: this.authHeaders,
|
||
})(params[0]);
|
||
}
|
||
async updateObject(...params) {
|
||
return defaultParam(updateObject, { headers: this.authHeaders, url: this.serverUrl })(params[0]);
|
||
}
|
||
async deleteObject(...params) {
|
||
return defaultParam(deleteObject, { headers: this.authHeaders, url: this.serverUrl })(params[0]);
|
||
}
|
||
async propfind(...params) {
|
||
return defaultParam(propfind, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async createAccount(params0) {
|
||
const { account, headers, loadCollections, loadObjects } = params0;
|
||
return createAccount({
|
||
account: { serverUrl: this.serverUrl, credentials: this.credentials, ...account },
|
||
headers: { ...this.authHeaders, ...headers },
|
||
loadCollections,
|
||
loadObjects,
|
||
});
|
||
}
|
||
async collectionQuery(...params) {
|
||
return defaultParam(collectionQuery, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async makeCollection(...params) {
|
||
return defaultParam(makeCollection, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async syncCollection(...params) {
|
||
return defaultParam(syncCollection, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async supportedReportSet(...params) {
|
||
return defaultParam(supportedReportSet, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async isCollectionDirty(...params) {
|
||
return defaultParam(isCollectionDirty, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async smartCollectionSync(...params) {
|
||
return defaultParam(smartCollectionSync, {
|
||
headers: this.authHeaders,
|
||
account: this.account,
|
||
})(params[0]);
|
||
}
|
||
async calendarQuery(...params) {
|
||
return defaultParam(calendarQuery, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async makeCalendar(...params) {
|
||
return defaultParam(makeCalendar, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async calendarMultiGet(...params) {
|
||
return defaultParam(calendarMultiGet, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async fetchCalendars(...params) {
|
||
return defaultParam(fetchCalendars, { headers: this.authHeaders, account: this.account })(params === null || params === void 0 ? void 0 : params[0]);
|
||
}
|
||
async fetchCalendarObjects(...params) {
|
||
return defaultParam(fetchCalendarObjects, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async createCalendarObject(...params) {
|
||
return defaultParam(createCalendarObject, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async updateCalendarObject(...params) {
|
||
return defaultParam(updateCalendarObject, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async deleteCalendarObject(...params) {
|
||
return defaultParam(deleteCalendarObject, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async syncCalendars(...params) {
|
||
return defaultParam(syncCalendars, {
|
||
headers: this.authHeaders,
|
||
account: this.account,
|
||
})(params[0]);
|
||
}
|
||
async addressBookQuery(...params) {
|
||
return defaultParam(addressBookQuery, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async addressBookMultiGet(...params) {
|
||
return defaultParam(addressBookMultiGet, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async fetchAddressBooks(...params) {
|
||
return defaultParam(fetchAddressBooks, { headers: this.authHeaders, account: this.account })(params === null || params === void 0 ? void 0 : params[0]);
|
||
}
|
||
async fetchVCards(...params) {
|
||
return defaultParam(fetchVCards, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async createVCard(...params) {
|
||
return defaultParam(createVCard, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async updateVCard(...params) {
|
||
return defaultParam(updateVCard, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
async deleteVCard(...params) {
|
||
return defaultParam(deleteVCard, { headers: this.authHeaders })(params[0]);
|
||
}
|
||
}
|
||
|
||
var client = /*#__PURE__*/Object.freeze({
|
||
__proto__: null,
|
||
DAVClient: DAVClient,
|
||
createDAVClient: createDAVClient
|
||
});
|
||
|
||
var index = {
|
||
DAVNamespace,
|
||
DAVNamespaceShort,
|
||
DAVAttributeMap,
|
||
...client,
|
||
...request,
|
||
...collection,
|
||
...account,
|
||
...addressBook,
|
||
...calendar,
|
||
...authHelpers,
|
||
...requestHelpers,
|
||
};
|
||
|
||
export { DAVAttributeMap, DAVClient, DAVNamespace, DAVNamespaceShort, addressBookQuery, calendarMultiGet, calendarQuery, cleanupFalsy, collectionQuery, createAccount, createCalendarObject, createDAVClient, createObject, createVCard, davRequest, index as default, deleteCalendarObject, deleteObject, deleteVCard, fetchAddressBooks, fetchCalendarObjects, fetchCalendars, fetchOauthTokens, fetchVCards, freeBusyQuery, getBasicAuthHeaders, getDAVAttribute, getOauthHeaders, isCollectionDirty, makeCalendar, propfind, refreshAccessToken, smartCollectionSync, supportedReportSet, syncCalendars, syncCollection, updateCalendarObject, updateObject, updateVCard, urlContains, urlEquals };
|