Update package dependencies and remove unused files from node_modules
This commit is contained in:
File diff suppressed because one or more lines are too long
@@ -1,11 +1,19 @@
|
||||
import { Q as commonjsGlobal, P as getDefaultExportFromCjs } from './dep-DBxKXgDP.js';
|
||||
import require$$0$2 from 'fs';
|
||||
import { D as commonjsGlobal, C as getDefaultExportFromCjs } from './dep-C6uTJdX2.js';
|
||||
import require$$0__default from 'fs';
|
||||
import require$$0 from 'postcss';
|
||||
import require$$0$1 from 'path';
|
||||
import require$$3 from 'crypto';
|
||||
import require$$1 from 'util';
|
||||
import { l as lib } from './dep-3RmXg9uo.js';
|
||||
import require$$0$2 from 'util';
|
||||
import { l as lib } from './dep-IQS-Za7F.js';
|
||||
|
||||
import { fileURLToPath as __cjs_fileURLToPath } from 'node:url';
|
||||
import { dirname as __cjs_dirname } from 'node:path';
|
||||
import { createRequire as __cjs_createRequire } from 'node:module';
|
||||
|
||||
const __filename = __cjs_fileURLToPath(import.meta.url);
|
||||
const __dirname = __cjs_dirname(__filename);
|
||||
const require = __cjs_createRequire(import.meta.url);
|
||||
const __require = require;
|
||||
function _mergeNamespaces(n, m) {
|
||||
for (var i = 0; i < m.length; i++) {
|
||||
var e = m[i];
|
||||
@@ -388,6 +396,9 @@ var localsConvention = {};
|
||||
* Copyright Jeremy Ashkenas, DocumentCloud and Investigative Reporters & Editors
|
||||
*/
|
||||
|
||||
/** Used as references for various `Number` constants. */
|
||||
var INFINITY = 1 / 0;
|
||||
|
||||
/** `Object#toString` result references. */
|
||||
var symbolTag = '[object Symbol]';
|
||||
|
||||
@@ -722,7 +733,7 @@ function baseToString(value) {
|
||||
return symbolToString ? symbolToString.call(value) : '';
|
||||
}
|
||||
var result = (value + '');
|
||||
return (result == '0' && (1 / value) == -Infinity) ? '-0' : result;
|
||||
return (result == '0' && (1 / value) == -INFINITY) ? '-0' : result;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -737,7 +748,7 @@ function baseToString(value) {
|
||||
function castSlice(array, start, end) {
|
||||
var length = array.length;
|
||||
end = end === undefined ? length : end;
|
||||
return (false && end >= length) ? array : baseSlice(array, start, end);
|
||||
return (!start && end >= length) ? array : baseSlice(array, start, end);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -1459,7 +1470,7 @@ function requireWasmHash () {
|
||||
// 64 is the maximum chunk size for every possible wasm hash implementation
|
||||
// 4 is the maximum number of bytes per char for string encoding (max is utf-8)
|
||||
// ~3 makes sure that it's always a block of 4 chars, so avoid partially encoded bytes for base64
|
||||
const MAX_SHORT_STRING = Math.floor((65536 - 64) / 4) & -4;
|
||||
const MAX_SHORT_STRING = Math.floor((65536 - 64) / 4) & ~3;
|
||||
|
||||
class WasmHash {
|
||||
/**
|
||||
@@ -2024,17 +2035,12 @@ function getHashDigest$1(buffer, algorithm, digestType, maxLength) {
|
||||
digestType === "base49" ||
|
||||
digestType === "base52" ||
|
||||
digestType === "base58" ||
|
||||
digestType === "base62" ||
|
||||
digestType === "base64safe"
|
||||
digestType === "base62"
|
||||
) {
|
||||
return encodeBufferToBase(
|
||||
hash.digest(),
|
||||
digestType === "base64safe" ? 64 : digestType.substr(4),
|
||||
maxLength
|
||||
);
|
||||
return encodeBufferToBase(hash.digest(), digestType.substr(4), maxLength);
|
||||
} else {
|
||||
return hash.digest(digestType || "hex").substr(0, maxLength);
|
||||
}
|
||||
|
||||
return hash.digest(digestType || "hex").substr(0, maxLength);
|
||||
}
|
||||
|
||||
var getHashDigest_1 = getHashDigest$1;
|
||||
@@ -2090,10 +2096,9 @@ function interpolateName$1(loaderContext, name, options = {}) {
|
||||
directory = resourcePath.replace(/\\/g, "/").replace(/\.\.(\/)?/g, "_$1");
|
||||
}
|
||||
|
||||
if (directory.length <= 1) {
|
||||
if (directory.length === 1) {
|
||||
directory = "";
|
||||
} else {
|
||||
// directory.length > 1
|
||||
} else if (directory.length > 1) {
|
||||
folder = path$1.basename(directory);
|
||||
}
|
||||
}
|
||||
@@ -2116,7 +2121,7 @@ function interpolateName$1(loaderContext, name, options = {}) {
|
||||
// `hash` and `contenthash` are same in `loader-utils` context
|
||||
// let's keep `hash` for backward compatibility
|
||||
.replace(
|
||||
/\[(?:([^[:\]]+):)?(?:hash|contenthash)(?::([a-z]+\d*(?:safe)?))?(?::(\d+))?\]/gi,
|
||||
/\[(?:([^[:\]]+):)?(?:hash|contenthash)(?::([a-z]+\d*))?(?::(\d+))?\]/gi,
|
||||
(all, hashType, digestType, maxLength) =>
|
||||
getHashDigest(content, hashType, digestType, parseInt(maxLength, 10))
|
||||
);
|
||||
@@ -2645,9 +2650,6 @@ types.UNIVERSAL = UNIVERSAL;
|
||||
_proto.prepend = function prepend(selector) {
|
||||
selector.parent = this;
|
||||
this.nodes.unshift(selector);
|
||||
for (var id in this.indexes) {
|
||||
this.indexes[id]++;
|
||||
}
|
||||
return this;
|
||||
};
|
||||
_proto.at = function at(index) {
|
||||
@@ -2684,39 +2686,29 @@ types.UNIVERSAL = UNIVERSAL;
|
||||
return this.removeAll();
|
||||
};
|
||||
_proto.insertAfter = function insertAfter(oldNode, newNode) {
|
||||
var _this$nodes;
|
||||
newNode.parent = this;
|
||||
var oldIndex = this.index(oldNode);
|
||||
var resetNode = [];
|
||||
for (var i = 2; i < arguments.length; i++) {
|
||||
resetNode.push(arguments[i]);
|
||||
}
|
||||
(_this$nodes = this.nodes).splice.apply(_this$nodes, [oldIndex + 1, 0, newNode].concat(resetNode));
|
||||
this.nodes.splice(oldIndex + 1, 0, newNode);
|
||||
newNode.parent = this;
|
||||
var index;
|
||||
for (var id in this.indexes) {
|
||||
index = this.indexes[id];
|
||||
if (oldIndex < index) {
|
||||
this.indexes[id] = index + arguments.length - 1;
|
||||
if (oldIndex <= index) {
|
||||
this.indexes[id] = index + 1;
|
||||
}
|
||||
}
|
||||
return this;
|
||||
};
|
||||
_proto.insertBefore = function insertBefore(oldNode, newNode) {
|
||||
var _this$nodes2;
|
||||
newNode.parent = this;
|
||||
var oldIndex = this.index(oldNode);
|
||||
var resetNode = [];
|
||||
for (var i = 2; i < arguments.length; i++) {
|
||||
resetNode.push(arguments[i]);
|
||||
}
|
||||
(_this$nodes2 = this.nodes).splice.apply(_this$nodes2, [oldIndex, 0, newNode].concat(resetNode));
|
||||
this.nodes.splice(oldIndex, 0, newNode);
|
||||
newNode.parent = this;
|
||||
var index;
|
||||
for (var id in this.indexes) {
|
||||
index = this.indexes[id];
|
||||
if (index >= oldIndex) {
|
||||
this.indexes[id] = index + arguments.length - 1;
|
||||
if (index <= oldIndex) {
|
||||
this.indexes[id] = index + 1;
|
||||
}
|
||||
}
|
||||
return this;
|
||||
@@ -2742,7 +2734,7 @@ types.UNIVERSAL = UNIVERSAL;
|
||||
* Return the most specific node at the line and column number given.
|
||||
* The source location is based on the original parsed location, locations aren't
|
||||
* updated as selector nodes are mutated.
|
||||
*
|
||||
*
|
||||
* Note that this location is relative to the location of the first character
|
||||
* of the selector, and not the location of the selector in the overall document
|
||||
* when used in conjunction with postcss.
|
||||
@@ -3410,7 +3402,7 @@ var attribute$1 = {};
|
||||
* For Node.js, simply re-export the core `util.deprecate` function.
|
||||
*/
|
||||
|
||||
var node = require$$1.deprecate;
|
||||
var node = require$$0$2.deprecate;
|
||||
|
||||
(function (exports) {
|
||||
|
||||
@@ -4760,7 +4752,7 @@ tokenTypes.combinator = combinator$1;
|
||||
}
|
||||
// We need to decide between a space that's a descendant combinator and meaningless whitespace at the end of a selector.
|
||||
var nextSigTokenPos = this.locateNextMeaningfulToken(this.position);
|
||||
if (nextSigTokenPos < 0 || this.tokens[nextSigTokenPos][_tokenize.FIELDS.TYPE] === tokens.comma || this.tokens[nextSigTokenPos][_tokenize.FIELDS.TYPE] === tokens.closeParenthesis) {
|
||||
if (nextSigTokenPos < 0 || this.tokens[nextSigTokenPos][_tokenize.FIELDS.TYPE] === tokens.comma) {
|
||||
var nodes = this.parseWhitespaceEquivalentTokens(nextSigTokenPos);
|
||||
if (nodes.length > 0) {
|
||||
var last = this.current.last;
|
||||
@@ -5639,40 +5631,8 @@ const selectorParser$1 = distExports;
|
||||
const valueParser = lib;
|
||||
const { extractICSS } = src$4;
|
||||
|
||||
const IGNORE_FILE_MARKER = "cssmodules-pure-no-check";
|
||||
const IGNORE_NEXT_LINE_MARKER = "cssmodules-pure-ignore";
|
||||
|
||||
const isSpacing = (node) => node.type === "combinator" && node.value === " ";
|
||||
|
||||
const isPureCheckDisabled = (root) => {
|
||||
for (const node of root.nodes) {
|
||||
if (node.type !== "comment") {
|
||||
return false;
|
||||
}
|
||||
if (node.text.trim().startsWith(IGNORE_FILE_MARKER)) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
};
|
||||
|
||||
function getIgnoreComment(node) {
|
||||
if (!node.parent) {
|
||||
return;
|
||||
}
|
||||
const indexInParent = node.parent.index(node);
|
||||
for (let i = indexInParent - 1; i >= 0; i--) {
|
||||
const prevNode = node.parent.nodes[i];
|
||||
if (prevNode.type === "comment") {
|
||||
if (prevNode.text.trimStart().startsWith(IGNORE_NEXT_LINE_MARKER)) {
|
||||
return prevNode;
|
||||
}
|
||||
} else {
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function normalizeNodeArray(nodes) {
|
||||
const array = [];
|
||||
|
||||
@@ -5692,8 +5652,6 @@ function normalizeNodeArray(nodes) {
|
||||
return array;
|
||||
}
|
||||
|
||||
const isPureSelectorSymbol = Symbol("is-pure-selector");
|
||||
|
||||
function localizeNode(rule, mode, localAliasMap) {
|
||||
const transform = (node, context) => {
|
||||
if (context.ignoreNextSpacing && !isSpacing(node)) {
|
||||
@@ -5901,7 +5859,7 @@ function localizeNode(rule, mode, localAliasMap) {
|
||||
}
|
||||
case "nesting": {
|
||||
if (node.value === "&") {
|
||||
context.hasLocals = rule.parent[isPureSelectorSymbol];
|
||||
context.hasLocals = true;
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -6007,7 +5965,7 @@ function localizeDeclarationValues(localize, declaration, context) {
|
||||
const subContext = {
|
||||
options: context.options,
|
||||
global: context.global,
|
||||
localizeNextItem: localize,
|
||||
localizeNextItem: localize && !context.global,
|
||||
localAliasMap: context.localAliasMap,
|
||||
};
|
||||
nodes[index] = localizeDeclNode(node, subContext);
|
||||
@@ -6016,20 +5974,24 @@ function localizeDeclarationValues(localize, declaration, context) {
|
||||
declaration.value = valueNodes.toString();
|
||||
}
|
||||
|
||||
// letter
|
||||
// An uppercase letter or a lowercase letter.
|
||||
//
|
||||
// ident-start code point
|
||||
// A letter, a non-ASCII code point, or U+005F LOW LINE (_).
|
||||
//
|
||||
// ident code point
|
||||
// An ident-start code point, a digit, or U+002D HYPHEN-MINUS (-).
|
||||
function localizeDeclaration(declaration, context) {
|
||||
const isAnimation = /animation$/i.test(declaration.prop);
|
||||
|
||||
// We don't validate `hex digits`, because we don't need it, it is work of linters.
|
||||
const validIdent =
|
||||
/^-?([a-z\u0080-\uFFFF_]|(\\[^\r\n\f])|-(?![0-9]))((\\[^\r\n\f])|[a-z\u0080-\uFFFF_0-9-])*$/i;
|
||||
if (isAnimation) {
|
||||
// letter
|
||||
// An uppercase letter or a lowercase letter.
|
||||
//
|
||||
// ident-start code point
|
||||
// A letter, a non-ASCII code point, or U+005F LOW LINE (_).
|
||||
//
|
||||
// ident code point
|
||||
// An ident-start code point, a digit, or U+002D HYPHEN-MINUS (-).
|
||||
|
||||
/*
|
||||
// We don't validate `hex digits`, because we don't need it, it is work of linters.
|
||||
const validIdent =
|
||||
/^-?([a-z\u0080-\uFFFF_]|(\\[^\r\n\f])|-(?![0-9]))((\\[^\r\n\f])|[a-z\u0080-\uFFFF_0-9-])*$/i;
|
||||
|
||||
/*
|
||||
The spec defines some keywords that you can use to describe properties such as the timing
|
||||
function. These are still valid animation names, so as long as there is a property that accepts
|
||||
a keyword, it is given priority. Only when all the properties that can take a keyword are
|
||||
@@ -6040,43 +6002,38 @@ const validIdent =
|
||||
The animation will repeat an infinite number of times from the first argument, and will have an
|
||||
animation name of infinite from the second.
|
||||
*/
|
||||
const animationKeywords = {
|
||||
// animation-direction
|
||||
$normal: 1,
|
||||
$reverse: 1,
|
||||
$alternate: 1,
|
||||
"$alternate-reverse": 1,
|
||||
// animation-fill-mode
|
||||
$forwards: 1,
|
||||
$backwards: 1,
|
||||
$both: 1,
|
||||
// animation-iteration-count
|
||||
$infinite: 1,
|
||||
// animation-play-state
|
||||
$paused: 1,
|
||||
$running: 1,
|
||||
// animation-timing-function
|
||||
$ease: 1,
|
||||
"$ease-in": 1,
|
||||
"$ease-out": 1,
|
||||
"$ease-in-out": 1,
|
||||
$linear: 1,
|
||||
"$step-end": 1,
|
||||
"$step-start": 1,
|
||||
// Special
|
||||
$none: Infinity, // No matter how many times you write none, it will never be an animation name
|
||||
// Global values
|
||||
$initial: Infinity,
|
||||
$inherit: Infinity,
|
||||
$unset: Infinity,
|
||||
$revert: Infinity,
|
||||
"$revert-layer": Infinity,
|
||||
};
|
||||
|
||||
function localizeDeclaration(declaration, context) {
|
||||
const isAnimation = /animation(-name)?$/i.test(declaration.prop);
|
||||
|
||||
if (isAnimation) {
|
||||
const animationKeywords = {
|
||||
// animation-direction
|
||||
$normal: 1,
|
||||
$reverse: 1,
|
||||
$alternate: 1,
|
||||
"$alternate-reverse": 1,
|
||||
// animation-fill-mode
|
||||
$forwards: 1,
|
||||
$backwards: 1,
|
||||
$both: 1,
|
||||
// animation-iteration-count
|
||||
$infinite: 1,
|
||||
// animation-play-state
|
||||
$paused: 1,
|
||||
$running: 1,
|
||||
// animation-timing-function
|
||||
$ease: 1,
|
||||
"$ease-in": 1,
|
||||
"$ease-out": 1,
|
||||
"$ease-in-out": 1,
|
||||
$linear: 1,
|
||||
"$step-end": 1,
|
||||
"$step-start": 1,
|
||||
// Special
|
||||
$none: Infinity, // No matter how many times you write none, it will never be an animation name
|
||||
// Global values
|
||||
$initial: Infinity,
|
||||
$inherit: Infinity,
|
||||
$unset: Infinity,
|
||||
$revert: Infinity,
|
||||
"$revert-layer": Infinity,
|
||||
};
|
||||
let parsedAnimationKeywords = {};
|
||||
const valueNodes = valueParser(declaration.value).walk((node) => {
|
||||
// If div-token appeared (represents as comma ','), a possibility of an animation-keywords should be reflesh.
|
||||
@@ -6084,28 +6041,9 @@ function localizeDeclaration(declaration, context) {
|
||||
parsedAnimationKeywords = {};
|
||||
|
||||
return;
|
||||
} else if (
|
||||
node.type === "function" &&
|
||||
node.value.toLowerCase() === "local" &&
|
||||
node.nodes.length === 1
|
||||
) {
|
||||
node.type = "word";
|
||||
node.value = node.nodes[0].value;
|
||||
|
||||
return localizeDeclNode(node, {
|
||||
options: context.options,
|
||||
global: context.global,
|
||||
localizeNextItem: true,
|
||||
localAliasMap: context.localAliasMap,
|
||||
});
|
||||
} else if (node.type === "function") {
|
||||
// replace `animation: global(example)` with `animation-name: example`
|
||||
if (node.value.toLowerCase() === "global" && node.nodes.length === 1) {
|
||||
node.type = "word";
|
||||
node.value = node.nodes[0].value;
|
||||
}
|
||||
|
||||
// Do not handle nested functions
|
||||
}
|
||||
// Do not handle nested functions
|
||||
else if (node.type === "function") {
|
||||
return false;
|
||||
}
|
||||
// Ignore all except word
|
||||
@@ -6132,12 +6070,14 @@ function localizeDeclaration(declaration, context) {
|
||||
}
|
||||
}
|
||||
|
||||
return localizeDeclNode(node, {
|
||||
const subContext = {
|
||||
options: context.options,
|
||||
global: context.global,
|
||||
localizeNextItem: shouldParseAnimationName && !context.global,
|
||||
localAliasMap: context.localAliasMap,
|
||||
});
|
||||
};
|
||||
|
||||
return localizeDeclNode(node, subContext);
|
||||
});
|
||||
|
||||
declaration.value = valueNodes.toString();
|
||||
@@ -6145,35 +6085,19 @@ function localizeDeclaration(declaration, context) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (/url\(/i.test(declaration.value)) {
|
||||
const isAnimationName = /animation(-name)?$/i.test(declaration.prop);
|
||||
|
||||
if (isAnimationName) {
|
||||
return localizeDeclarationValues(true, declaration, context);
|
||||
}
|
||||
|
||||
const hasUrl = /url\(/i.test(declaration.value);
|
||||
|
||||
if (hasUrl) {
|
||||
return localizeDeclarationValues(false, declaration, context);
|
||||
}
|
||||
}
|
||||
|
||||
const isPureSelector = (context, rule) => {
|
||||
if (!rule.parent || rule.type === "root") {
|
||||
return !context.hasPureGlobals;
|
||||
}
|
||||
|
||||
if (rule.type === "rule" && rule[isPureSelectorSymbol]) {
|
||||
return rule[isPureSelectorSymbol] || isPureSelector(context, rule.parent);
|
||||
}
|
||||
|
||||
return !context.hasPureGlobals || isPureSelector(context, rule.parent);
|
||||
};
|
||||
|
||||
const isNodeWithoutDeclarations = (rule) => {
|
||||
if (rule.nodes.length > 0) {
|
||||
return !rule.nodes.every(
|
||||
(item) =>
|
||||
item.type === "rule" ||
|
||||
(item.type === "atrule" && !isNodeWithoutDeclarations(item))
|
||||
);
|
||||
}
|
||||
|
||||
return true;
|
||||
};
|
||||
|
||||
src$2.exports = (options = {}) => {
|
||||
if (
|
||||
options &&
|
||||
@@ -6198,7 +6122,6 @@ src$2.exports = (options = {}) => {
|
||||
return {
|
||||
Once(root) {
|
||||
const { icssImports } = extractICSS(root, false);
|
||||
const enforcePureMode = pureMode && !isPureCheckDisabled(root);
|
||||
|
||||
Object.keys(icssImports).forEach((key) => {
|
||||
Object.keys(icssImports[key]).forEach((prop) => {
|
||||
@@ -6218,15 +6141,10 @@ src$2.exports = (options = {}) => {
|
||||
let globalKeyframes = globalMode;
|
||||
|
||||
if (globalMatch) {
|
||||
if (enforcePureMode) {
|
||||
const ignoreComment = getIgnoreComment(atRule);
|
||||
if (!ignoreComment) {
|
||||
throw atRule.error(
|
||||
"@keyframes :global(...) is not allowed in pure mode"
|
||||
);
|
||||
} else {
|
||||
ignoreComment.remove();
|
||||
}
|
||||
if (pureMode) {
|
||||
throw atRule.error(
|
||||
"@keyframes :global(...) is not allowed in pure mode"
|
||||
);
|
||||
}
|
||||
atRule.params = globalMatch[1];
|
||||
globalKeyframes = true;
|
||||
@@ -6250,14 +6168,6 @@ src$2.exports = (options = {}) => {
|
||||
});
|
||||
} else if (/scope$/i.test(atRule.name)) {
|
||||
if (atRule.params) {
|
||||
const ignoreComment = pureMode
|
||||
? getIgnoreComment(atRule)
|
||||
: undefined;
|
||||
|
||||
if (ignoreComment) {
|
||||
ignoreComment.remove();
|
||||
}
|
||||
|
||||
atRule.params = atRule.params
|
||||
.split("to")
|
||||
.map((item) => {
|
||||
@@ -6271,11 +6181,7 @@ src$2.exports = (options = {}) => {
|
||||
context.options = options;
|
||||
context.localAliasMap = localAliasMap;
|
||||
|
||||
if (
|
||||
enforcePureMode &&
|
||||
context.hasPureGlobals &&
|
||||
!ignoreComment
|
||||
) {
|
||||
if (pureMode && context.hasPureGlobals) {
|
||||
throw atRule.error(
|
||||
'Selector in at-rule"' +
|
||||
selector +
|
||||
@@ -6326,28 +6232,13 @@ src$2.exports = (options = {}) => {
|
||||
context.options = options;
|
||||
context.localAliasMap = localAliasMap;
|
||||
|
||||
const ignoreComment = enforcePureMode
|
||||
? getIgnoreComment(rule)
|
||||
: undefined;
|
||||
const isNotPure = enforcePureMode && !isPureSelector(context, rule);
|
||||
|
||||
if (
|
||||
isNotPure &&
|
||||
isNodeWithoutDeclarations(rule) &&
|
||||
!ignoreComment
|
||||
) {
|
||||
if (pureMode && context.hasPureGlobals) {
|
||||
throw rule.error(
|
||||
'Selector "' +
|
||||
rule.selector +
|
||||
'" is not pure ' +
|
||||
"(pure selectors must contain at least one local class or id)"
|
||||
);
|
||||
} else if (ignoreComment) {
|
||||
ignoreComment.remove();
|
||||
}
|
||||
|
||||
if (pureMode) {
|
||||
rule[isPureSelectorSymbol] = !isNotPure;
|
||||
}
|
||||
|
||||
rule.selector = context.selector;
|
||||
@@ -7086,7 +6977,7 @@ function makePlugin(opts) {
|
||||
};
|
||||
}
|
||||
|
||||
var _fs = require$$0$2;
|
||||
var _fs = require$$0__default;
|
||||
|
||||
var _fs2 = fs;
|
||||
|
@@ -1,10 +1,16 @@
|
||||
import { P as getDefaultExportFromCjs } from './dep-DBxKXgDP.js';
|
||||
import { C as getDefaultExportFromCjs } from './dep-C6uTJdX2.js';
|
||||
import require$$0 from 'path';
|
||||
import { l as lib } from './dep-3RmXg9uo.js';
|
||||
import require$$0__default from 'fs';
|
||||
import { l as lib } from './dep-IQS-Za7F.js';
|
||||
|
||||
import { fileURLToPath as __cjs_fileURLToPath } from 'node:url';
|
||||
import { dirname as __cjs_dirname } from 'node:path';
|
||||
import { createRequire as __cjs_createRequire } from 'node:module';
|
||||
|
||||
const __require = __cjs_createRequire(import.meta.url);
|
||||
const __filename = __cjs_fileURLToPath(import.meta.url);
|
||||
const __dirname = __cjs_dirname(__filename);
|
||||
const require = __cjs_createRequire(import.meta.url);
|
||||
const __require = require;
|
||||
function _mergeNamespaces(n, m) {
|
||||
for (var i = 0; i < m.length; i++) {
|
||||
var e = m[i];
|
||||
@@ -184,6 +190,160 @@ var applyStyles$1 = function applyStyles(bundle, styles) {
|
||||
});
|
||||
};
|
||||
|
||||
var readCache$1 = {exports: {}};
|
||||
|
||||
var pify$2 = {exports: {}};
|
||||
|
||||
var processFn = function (fn, P, opts) {
|
||||
return function () {
|
||||
var that = this;
|
||||
var args = new Array(arguments.length);
|
||||
|
||||
for (var i = 0; i < arguments.length; i++) {
|
||||
args[i] = arguments[i];
|
||||
}
|
||||
|
||||
return new P(function (resolve, reject) {
|
||||
args.push(function (err, result) {
|
||||
if (err) {
|
||||
reject(err);
|
||||
} else if (opts.multiArgs) {
|
||||
var results = new Array(arguments.length - 1);
|
||||
|
||||
for (var i = 1; i < arguments.length; i++) {
|
||||
results[i - 1] = arguments[i];
|
||||
}
|
||||
|
||||
resolve(results);
|
||||
} else {
|
||||
resolve(result);
|
||||
}
|
||||
});
|
||||
|
||||
fn.apply(that, args);
|
||||
});
|
||||
};
|
||||
};
|
||||
|
||||
var pify$1 = pify$2.exports = function (obj, P, opts) {
|
||||
if (typeof P !== 'function') {
|
||||
opts = P;
|
||||
P = Promise;
|
||||
}
|
||||
|
||||
opts = opts || {};
|
||||
opts.exclude = opts.exclude || [/.+Sync$/];
|
||||
|
||||
var filter = function (key) {
|
||||
var match = function (pattern) {
|
||||
return typeof pattern === 'string' ? key === pattern : pattern.test(key);
|
||||
};
|
||||
|
||||
return opts.include ? opts.include.some(match) : !opts.exclude.some(match);
|
||||
};
|
||||
|
||||
var ret = typeof obj === 'function' ? function () {
|
||||
if (opts.excludeMain) {
|
||||
return obj.apply(this, arguments);
|
||||
}
|
||||
|
||||
return processFn(obj, P, opts).apply(this, arguments);
|
||||
} : {};
|
||||
|
||||
return Object.keys(obj).reduce(function (ret, key) {
|
||||
var x = obj[key];
|
||||
|
||||
ret[key] = typeof x === 'function' && filter(key) ? processFn(x, P, opts) : x;
|
||||
|
||||
return ret;
|
||||
}, ret);
|
||||
};
|
||||
|
||||
pify$1.all = pify$1;
|
||||
|
||||
var pifyExports = pify$2.exports;
|
||||
|
||||
var fs = require$$0__default;
|
||||
var path$3 = require$$0;
|
||||
var pify = pifyExports;
|
||||
|
||||
var stat = pify(fs.stat);
|
||||
var readFile = pify(fs.readFile);
|
||||
var resolve = path$3.resolve;
|
||||
|
||||
var cache = Object.create(null);
|
||||
|
||||
function convert(content, encoding) {
|
||||
if (Buffer.isEncoding(encoding)) {
|
||||
return content.toString(encoding);
|
||||
}
|
||||
return content;
|
||||
}
|
||||
|
||||
readCache$1.exports = function (path, encoding) {
|
||||
path = resolve(path);
|
||||
|
||||
return stat(path).then(function (stats) {
|
||||
var item = cache[path];
|
||||
|
||||
if (item && item.mtime.getTime() === stats.mtime.getTime()) {
|
||||
return convert(item.content, encoding);
|
||||
}
|
||||
|
||||
return readFile(path).then(function (data) {
|
||||
cache[path] = {
|
||||
mtime: stats.mtime,
|
||||
content: data
|
||||
};
|
||||
|
||||
return convert(data, encoding);
|
||||
});
|
||||
}).catch(function (err) {
|
||||
cache[path] = null;
|
||||
return Promise.reject(err);
|
||||
});
|
||||
};
|
||||
|
||||
readCache$1.exports.sync = function (path, encoding) {
|
||||
path = resolve(path);
|
||||
|
||||
try {
|
||||
var stats = fs.statSync(path);
|
||||
var item = cache[path];
|
||||
|
||||
if (item && item.mtime.getTime() === stats.mtime.getTime()) {
|
||||
return convert(item.content, encoding);
|
||||
}
|
||||
|
||||
var data = fs.readFileSync(path);
|
||||
|
||||
cache[path] = {
|
||||
mtime: stats.mtime,
|
||||
content: data
|
||||
};
|
||||
|
||||
return convert(data, encoding);
|
||||
} catch (err) {
|
||||
cache[path] = null;
|
||||
throw err;
|
||||
}
|
||||
|
||||
};
|
||||
|
||||
readCache$1.exports.get = function (path, encoding) {
|
||||
path = resolve(path);
|
||||
if (cache[path]) {
|
||||
return convert(cache[path].content, encoding);
|
||||
}
|
||||
return null;
|
||||
};
|
||||
|
||||
readCache$1.exports.clear = function () {
|
||||
cache = Object.create(null);
|
||||
};
|
||||
|
||||
var readCacheExports = readCache$1.exports;
|
||||
|
||||
const anyDataURLRegexp = /^data:text\/css(?:;(base64|plain))?,/i;
|
||||
const base64DataURLRegexp = /^data:text\/css;base64,/i;
|
||||
const plainDataURLRegexp = /^data:text\/css;plain,/i;
|
||||
@@ -212,6 +372,17 @@ var dataUrl = {
|
||||
contents,
|
||||
};
|
||||
|
||||
const readCache = readCacheExports;
|
||||
const dataURL$1 = dataUrl;
|
||||
|
||||
var loadContent$1 = function loadContent(filename) {
|
||||
if (dataURL$1.isValid(filename)) {
|
||||
return dataURL$1.contents(filename)
|
||||
}
|
||||
|
||||
return readCache(filename, "utf-8")
|
||||
};
|
||||
|
||||
// external tooling
|
||||
const valueParser = lib;
|
||||
|
||||
@@ -749,7 +920,7 @@ const path = require$$0;
|
||||
const applyConditions = applyConditions$1;
|
||||
const applyRaws = applyRaws$1;
|
||||
const applyStyles = applyStyles$1;
|
||||
const loadContent = () => "";
|
||||
const loadContent = loadContent$1;
|
||||
const parseStyles = parseStyles_1;
|
||||
const resolveId = (id) => id;
|
||||
|
File diff suppressed because it is too large
Load Diff
@@ -1,3 +1,11 @@
|
||||
import { fileURLToPath as __cjs_fileURLToPath } from 'node:url';
|
||||
import { dirname as __cjs_dirname } from 'node:path';
|
||||
import { createRequire as __cjs_createRequire } from 'node:module';
|
||||
|
||||
const __filename = __cjs_fileURLToPath(import.meta.url);
|
||||
const __dirname = __cjs_dirname(__filename);
|
||||
const require = __cjs_createRequire(import.meta.url);
|
||||
const __require = require;
|
||||
var openParentheses = "(".charCodeAt(0);
|
||||
var closeParentheses = ")".charCodeAt(0);
|
||||
var singleQuote = "'".charCodeAt(0);
|
Reference in New Issue
Block a user