- Introduced blog routing using Next.js App Router - Implemented dynamic [slug] pages for blog posts - Added MDX-based content loading via lib/posts - Integrated shared TopBar layout with navigation - Established clear content, lib and component separation
1860 lines
73 KiB
JavaScript
1860 lines
73 KiB
JavaScript
(globalThis.TURBOPACK || (globalThis.TURBOPACK = [])).push([
|
|
typeof document === "object" ? document.currentScript : undefined,
|
|
{"otherChunks":["static/chunks/[turbopack]_browser_dev_hmr-client_hmr-client_ts_956a0d3a._.js","static/chunks/node_modules_next_dist_compiled_react-dom_1e674e59._.js","static/chunks/node_modules_next_dist_compiled_react-server-dom-turbopack_9212ccad._.js","static/chunks/node_modules_next_dist_compiled_next-devtools_index_1dd7fb59.js","static/chunks/node_modules_next_dist_compiled_a0e4c7b4._.js","static/chunks/node_modules_next_dist_client_17643121._.js","static/chunks/node_modules_next_dist_f3530cac._.js","static/chunks/node_modules_@swc_helpers_cjs_d80fb378._.js"],"runtimeModuleIds":["[project]/node_modules/next/dist/compiled/@next/react-refresh-utils/dist/runtime.js [app-client] (ecmascript)","[project]/node_modules/next/dist/client/app-next-turbopack.js [app-client] (ecmascript)"]}
|
|
]);
|
|
(() => {
|
|
if (!Array.isArray(globalThis.TURBOPACK)) {
|
|
return;
|
|
}
|
|
|
|
const CHUNK_BASE_PATH = "/_next/";
|
|
const RELATIVE_ROOT_PATH = "/ROOT";
|
|
const RUNTIME_PUBLIC_PATH = "/_next/";
|
|
const CHUNK_SUFFIX = (self.TURBOPACK_CHUNK_SUFFIX ?? document?.currentScript?.getAttribute?.('src')?.replace(/^(.*(?=\?)|^.*$)/, "")) || "";
|
|
/**
|
|
* This file contains runtime types and functions that are shared between all
|
|
* TurboPack ECMAScript runtimes.
|
|
*
|
|
* It will be prepended to the runtime code of each runtime.
|
|
*/ /* eslint-disable @typescript-eslint/no-unused-vars */ /// <reference path="./runtime-types.d.ts" />
|
|
const REEXPORTED_OBJECTS = new WeakMap();
|
|
/**
|
|
* Constructs the `__turbopack_context__` object for a module.
|
|
*/ function Context(module, exports) {
|
|
this.m = module;
|
|
// We need to store this here instead of accessing it from the module object to:
|
|
// 1. Make it available to factories directly, since we rewrite `this` to
|
|
// `__turbopack_context__.e` in CJS modules.
|
|
// 2. Support async modules which rewrite `module.exports` to a promise, so we
|
|
// can still access the original exports object from functions like
|
|
// `esmExport`
|
|
// Ideally we could find a new approach for async modules and drop this property altogether.
|
|
this.e = exports;
|
|
}
|
|
const contextPrototype = Context.prototype;
|
|
const hasOwnProperty = Object.prototype.hasOwnProperty;
|
|
const toStringTag = typeof Symbol !== 'undefined' && Symbol.toStringTag;
|
|
function defineProp(obj, name, options) {
|
|
if (!hasOwnProperty.call(obj, name)) Object.defineProperty(obj, name, options);
|
|
}
|
|
function getOverwrittenModule(moduleCache, id) {
|
|
let module = moduleCache[id];
|
|
if (!module) {
|
|
// This is invoked when a module is merged into another module, thus it wasn't invoked via
|
|
// instantiateModule and the cache entry wasn't created yet.
|
|
module = createModuleObject(id);
|
|
moduleCache[id] = module;
|
|
}
|
|
return module;
|
|
}
|
|
/**
|
|
* Creates the module object. Only done here to ensure all module objects have the same shape.
|
|
*/ function createModuleObject(id) {
|
|
return {
|
|
exports: {},
|
|
error: undefined,
|
|
id,
|
|
namespaceObject: undefined
|
|
};
|
|
}
|
|
const BindingTag_Value = 0;
|
|
/**
|
|
* Adds the getters to the exports object.
|
|
*/ function esm(exports, bindings) {
|
|
defineProp(exports, '__esModule', {
|
|
value: true
|
|
});
|
|
if (toStringTag) defineProp(exports, toStringTag, {
|
|
value: 'Module'
|
|
});
|
|
let i = 0;
|
|
while(i < bindings.length){
|
|
const propName = bindings[i++];
|
|
const tagOrFunction = bindings[i++];
|
|
if (typeof tagOrFunction === 'number') {
|
|
if (tagOrFunction === BindingTag_Value) {
|
|
defineProp(exports, propName, {
|
|
value: bindings[i++],
|
|
enumerable: true,
|
|
writable: false
|
|
});
|
|
} else {
|
|
throw new Error(`unexpected tag: ${tagOrFunction}`);
|
|
}
|
|
} else {
|
|
const getterFn = tagOrFunction;
|
|
if (typeof bindings[i] === 'function') {
|
|
const setterFn = bindings[i++];
|
|
defineProp(exports, propName, {
|
|
get: getterFn,
|
|
set: setterFn,
|
|
enumerable: true
|
|
});
|
|
} else {
|
|
defineProp(exports, propName, {
|
|
get: getterFn,
|
|
enumerable: true
|
|
});
|
|
}
|
|
}
|
|
}
|
|
Object.seal(exports);
|
|
}
|
|
/**
|
|
* Makes the module an ESM with exports
|
|
*/ function esmExport(bindings, id) {
|
|
let module;
|
|
let exports;
|
|
if (id != null) {
|
|
module = getOverwrittenModule(this.c, id);
|
|
exports = module.exports;
|
|
} else {
|
|
module = this.m;
|
|
exports = this.e;
|
|
}
|
|
module.namespaceObject = exports;
|
|
esm(exports, bindings);
|
|
}
|
|
contextPrototype.s = esmExport;
|
|
function ensureDynamicExports(module, exports) {
|
|
let reexportedObjects = REEXPORTED_OBJECTS.get(module);
|
|
if (!reexportedObjects) {
|
|
REEXPORTED_OBJECTS.set(module, reexportedObjects = []);
|
|
module.exports = module.namespaceObject = new Proxy(exports, {
|
|
get (target, prop) {
|
|
if (hasOwnProperty.call(target, prop) || prop === 'default' || prop === '__esModule') {
|
|
return Reflect.get(target, prop);
|
|
}
|
|
for (const obj of reexportedObjects){
|
|
const value = Reflect.get(obj, prop);
|
|
if (value !== undefined) return value;
|
|
}
|
|
return undefined;
|
|
},
|
|
ownKeys (target) {
|
|
const keys = Reflect.ownKeys(target);
|
|
for (const obj of reexportedObjects){
|
|
for (const key of Reflect.ownKeys(obj)){
|
|
if (key !== 'default' && !keys.includes(key)) keys.push(key);
|
|
}
|
|
}
|
|
return keys;
|
|
}
|
|
});
|
|
}
|
|
return reexportedObjects;
|
|
}
|
|
/**
|
|
* Dynamically exports properties from an object
|
|
*/ function dynamicExport(object, id) {
|
|
let module;
|
|
let exports;
|
|
if (id != null) {
|
|
module = getOverwrittenModule(this.c, id);
|
|
exports = module.exports;
|
|
} else {
|
|
module = this.m;
|
|
exports = this.e;
|
|
}
|
|
const reexportedObjects = ensureDynamicExports(module, exports);
|
|
if (typeof object === 'object' && object !== null) {
|
|
reexportedObjects.push(object);
|
|
}
|
|
}
|
|
contextPrototype.j = dynamicExport;
|
|
function exportValue(value, id) {
|
|
let module;
|
|
if (id != null) {
|
|
module = getOverwrittenModule(this.c, id);
|
|
} else {
|
|
module = this.m;
|
|
}
|
|
module.exports = value;
|
|
}
|
|
contextPrototype.v = exportValue;
|
|
function exportNamespace(namespace, id) {
|
|
let module;
|
|
if (id != null) {
|
|
module = getOverwrittenModule(this.c, id);
|
|
} else {
|
|
module = this.m;
|
|
}
|
|
module.exports = module.namespaceObject = namespace;
|
|
}
|
|
contextPrototype.n = exportNamespace;
|
|
function createGetter(obj, key) {
|
|
return ()=>obj[key];
|
|
}
|
|
/**
|
|
* @returns prototype of the object
|
|
*/ const getProto = Object.getPrototypeOf ? (obj)=>Object.getPrototypeOf(obj) : (obj)=>obj.__proto__;
|
|
/** Prototypes that are not expanded for exports */ const LEAF_PROTOTYPES = [
|
|
null,
|
|
getProto({}),
|
|
getProto([]),
|
|
getProto(getProto)
|
|
];
|
|
/**
|
|
* @param raw
|
|
* @param ns
|
|
* @param allowExportDefault
|
|
* * `false`: will have the raw module as default export
|
|
* * `true`: will have the default property as default export
|
|
*/ function interopEsm(raw, ns, allowExportDefault) {
|
|
const bindings = [];
|
|
let defaultLocation = -1;
|
|
for(let current = raw; (typeof current === 'object' || typeof current === 'function') && !LEAF_PROTOTYPES.includes(current); current = getProto(current)){
|
|
for (const key of Object.getOwnPropertyNames(current)){
|
|
bindings.push(key, createGetter(raw, key));
|
|
if (defaultLocation === -1 && key === 'default') {
|
|
defaultLocation = bindings.length - 1;
|
|
}
|
|
}
|
|
}
|
|
// this is not really correct
|
|
// we should set the `default` getter if the imported module is a `.cjs file`
|
|
if (!(allowExportDefault && defaultLocation >= 0)) {
|
|
// Replace the binding with one for the namespace itself in order to preserve iteration order.
|
|
if (defaultLocation >= 0) {
|
|
// Replace the getter with the value
|
|
bindings.splice(defaultLocation, 1, BindingTag_Value, raw);
|
|
} else {
|
|
bindings.push('default', BindingTag_Value, raw);
|
|
}
|
|
}
|
|
esm(ns, bindings);
|
|
return ns;
|
|
}
|
|
function createNS(raw) {
|
|
if (typeof raw === 'function') {
|
|
return function(...args) {
|
|
return raw.apply(this, args);
|
|
};
|
|
} else {
|
|
return Object.create(null);
|
|
}
|
|
}
|
|
function esmImport(id) {
|
|
const module = getOrInstantiateModuleFromParent(id, this.m);
|
|
// any ES module has to have `module.namespaceObject` defined.
|
|
if (module.namespaceObject) return module.namespaceObject;
|
|
// only ESM can be an async module, so we don't need to worry about exports being a promise here.
|
|
const raw = module.exports;
|
|
return module.namespaceObject = interopEsm(raw, createNS(raw), raw && raw.__esModule);
|
|
}
|
|
contextPrototype.i = esmImport;
|
|
function asyncLoader(moduleId) {
|
|
const loader = this.r(moduleId);
|
|
return loader(esmImport.bind(this));
|
|
}
|
|
contextPrototype.A = asyncLoader;
|
|
// Add a simple runtime require so that environments without one can still pass
|
|
// `typeof require` CommonJS checks so that exports are correctly registered.
|
|
const runtimeRequire = // @ts-ignore
|
|
typeof require === 'function' ? require : function require1() {
|
|
throw new Error('Unexpected use of runtime require');
|
|
};
|
|
contextPrototype.t = runtimeRequire;
|
|
function commonJsRequire(id) {
|
|
return getOrInstantiateModuleFromParent(id, this.m).exports;
|
|
}
|
|
contextPrototype.r = commonJsRequire;
|
|
/**
|
|
* Remove fragments and query parameters since they are never part of the context map keys
|
|
*
|
|
* This matches how we parse patterns at resolving time. Arguably we should only do this for
|
|
* strings passed to `import` but the resolve does it for `import` and `require` and so we do
|
|
* here as well.
|
|
*/ function parseRequest(request) {
|
|
// Per the URI spec fragments can contain `?` characters, so we should trim it off first
|
|
// https://datatracker.ietf.org/doc/html/rfc3986#section-3.5
|
|
const hashIndex = request.indexOf('#');
|
|
if (hashIndex !== -1) {
|
|
request = request.substring(0, hashIndex);
|
|
}
|
|
const queryIndex = request.indexOf('?');
|
|
if (queryIndex !== -1) {
|
|
request = request.substring(0, queryIndex);
|
|
}
|
|
return request;
|
|
}
|
|
/**
|
|
* `require.context` and require/import expression runtime.
|
|
*/ function moduleContext(map) {
|
|
function moduleContext(id) {
|
|
id = parseRequest(id);
|
|
if (hasOwnProperty.call(map, id)) {
|
|
return map[id].module();
|
|
}
|
|
const e = new Error(`Cannot find module '${id}'`);
|
|
e.code = 'MODULE_NOT_FOUND';
|
|
throw e;
|
|
}
|
|
moduleContext.keys = ()=>{
|
|
return Object.keys(map);
|
|
};
|
|
moduleContext.resolve = (id)=>{
|
|
id = parseRequest(id);
|
|
if (hasOwnProperty.call(map, id)) {
|
|
return map[id].id();
|
|
}
|
|
const e = new Error(`Cannot find module '${id}'`);
|
|
e.code = 'MODULE_NOT_FOUND';
|
|
throw e;
|
|
};
|
|
moduleContext.import = async (id)=>{
|
|
return await moduleContext(id);
|
|
};
|
|
return moduleContext;
|
|
}
|
|
contextPrototype.f = moduleContext;
|
|
/**
|
|
* Returns the path of a chunk defined by its data.
|
|
*/ function getChunkPath(chunkData) {
|
|
return typeof chunkData === 'string' ? chunkData : chunkData.path;
|
|
}
|
|
function isPromise(maybePromise) {
|
|
return maybePromise != null && typeof maybePromise === 'object' && 'then' in maybePromise && typeof maybePromise.then === 'function';
|
|
}
|
|
function isAsyncModuleExt(obj) {
|
|
return turbopackQueues in obj;
|
|
}
|
|
function createPromise() {
|
|
let resolve;
|
|
let reject;
|
|
const promise = new Promise((res, rej)=>{
|
|
reject = rej;
|
|
resolve = res;
|
|
});
|
|
return {
|
|
promise,
|
|
resolve: resolve,
|
|
reject: reject
|
|
};
|
|
}
|
|
// Load the CompressedmoduleFactories of a chunk into the `moduleFactories` Map.
|
|
// The CompressedModuleFactories format is
|
|
// - 1 or more module ids
|
|
// - a module factory function
|
|
// So walking this is a little complex but the flat structure is also fast to
|
|
// traverse, we can use `typeof` operators to distinguish the two cases.
|
|
function installCompressedModuleFactories(chunkModules, offset, moduleFactories, newModuleId) {
|
|
let i = offset;
|
|
while(i < chunkModules.length){
|
|
let moduleId = chunkModules[i];
|
|
let end = i + 1;
|
|
// Find our factory function
|
|
while(end < chunkModules.length && typeof chunkModules[end] !== 'function'){
|
|
end++;
|
|
}
|
|
if (end === chunkModules.length) {
|
|
throw new Error('malformed chunk format, expected a factory function');
|
|
}
|
|
// Each chunk item has a 'primary id' and optional additional ids. If the primary id is already
|
|
// present we know all the additional ids are also present, so we don't need to check.
|
|
if (!moduleFactories.has(moduleId)) {
|
|
const moduleFactoryFn = chunkModules[end];
|
|
applyModuleFactoryName(moduleFactoryFn);
|
|
newModuleId?.(moduleId);
|
|
for(; i < end; i++){
|
|
moduleId = chunkModules[i];
|
|
moduleFactories.set(moduleId, moduleFactoryFn);
|
|
}
|
|
}
|
|
i = end + 1; // end is pointing at the last factory advance to the next id or the end of the array.
|
|
}
|
|
}
|
|
// everything below is adapted from webpack
|
|
// https://github.com/webpack/webpack/blob/6be4065ade1e252c1d8dcba4af0f43e32af1bdc1/lib/runtime/AsyncModuleRuntimeModule.js#L13
|
|
const turbopackQueues = Symbol('turbopack queues');
|
|
const turbopackExports = Symbol('turbopack exports');
|
|
const turbopackError = Symbol('turbopack error');
|
|
function resolveQueue(queue) {
|
|
if (queue && queue.status !== 1) {
|
|
queue.status = 1;
|
|
queue.forEach((fn)=>fn.queueCount--);
|
|
queue.forEach((fn)=>fn.queueCount-- ? fn.queueCount++ : fn());
|
|
}
|
|
}
|
|
function wrapDeps(deps) {
|
|
return deps.map((dep)=>{
|
|
if (dep !== null && typeof dep === 'object') {
|
|
if (isAsyncModuleExt(dep)) return dep;
|
|
if (isPromise(dep)) {
|
|
const queue = Object.assign([], {
|
|
status: 0
|
|
});
|
|
const obj = {
|
|
[turbopackExports]: {},
|
|
[turbopackQueues]: (fn)=>fn(queue)
|
|
};
|
|
dep.then((res)=>{
|
|
obj[turbopackExports] = res;
|
|
resolveQueue(queue);
|
|
}, (err)=>{
|
|
obj[turbopackError] = err;
|
|
resolveQueue(queue);
|
|
});
|
|
return obj;
|
|
}
|
|
}
|
|
return {
|
|
[turbopackExports]: dep,
|
|
[turbopackQueues]: ()=>{}
|
|
};
|
|
});
|
|
}
|
|
function asyncModule(body, hasAwait) {
|
|
const module = this.m;
|
|
const queue = hasAwait ? Object.assign([], {
|
|
status: -1
|
|
}) : undefined;
|
|
const depQueues = new Set();
|
|
const { resolve, reject, promise: rawPromise } = createPromise();
|
|
const promise = Object.assign(rawPromise, {
|
|
[turbopackExports]: module.exports,
|
|
[turbopackQueues]: (fn)=>{
|
|
queue && fn(queue);
|
|
depQueues.forEach(fn);
|
|
promise['catch'](()=>{});
|
|
}
|
|
});
|
|
const attributes = {
|
|
get () {
|
|
return promise;
|
|
},
|
|
set (v) {
|
|
// Calling `esmExport` leads to this.
|
|
if (v !== promise) {
|
|
promise[turbopackExports] = v;
|
|
}
|
|
}
|
|
};
|
|
Object.defineProperty(module, 'exports', attributes);
|
|
Object.defineProperty(module, 'namespaceObject', attributes);
|
|
function handleAsyncDependencies(deps) {
|
|
const currentDeps = wrapDeps(deps);
|
|
const getResult = ()=>currentDeps.map((d)=>{
|
|
if (d[turbopackError]) throw d[turbopackError];
|
|
return d[turbopackExports];
|
|
});
|
|
const { promise, resolve } = createPromise();
|
|
const fn = Object.assign(()=>resolve(getResult), {
|
|
queueCount: 0
|
|
});
|
|
function fnQueue(q) {
|
|
if (q !== queue && !depQueues.has(q)) {
|
|
depQueues.add(q);
|
|
if (q && q.status === 0) {
|
|
fn.queueCount++;
|
|
q.push(fn);
|
|
}
|
|
}
|
|
}
|
|
currentDeps.map((dep)=>dep[turbopackQueues](fnQueue));
|
|
return fn.queueCount ? promise : getResult();
|
|
}
|
|
function asyncResult(err) {
|
|
if (err) {
|
|
reject(promise[turbopackError] = err);
|
|
} else {
|
|
resolve(promise[turbopackExports]);
|
|
}
|
|
resolveQueue(queue);
|
|
}
|
|
body(handleAsyncDependencies, asyncResult);
|
|
if (queue && queue.status === -1) {
|
|
queue.status = 0;
|
|
}
|
|
}
|
|
contextPrototype.a = asyncModule;
|
|
/**
|
|
* A pseudo "fake" URL object to resolve to its relative path.
|
|
*
|
|
* When UrlRewriteBehavior is set to relative, calls to the `new URL()` will construct url without base using this
|
|
* runtime function to generate context-agnostic urls between different rendering context, i.e ssr / client to avoid
|
|
* hydration mismatch.
|
|
*
|
|
* This is based on webpack's existing implementation:
|
|
* https://github.com/webpack/webpack/blob/87660921808566ef3b8796f8df61bd79fc026108/lib/runtime/RelativeUrlRuntimeModule.js
|
|
*/ const relativeURL = function relativeURL(inputUrl) {
|
|
const realUrl = new URL(inputUrl, 'x:/');
|
|
const values = {};
|
|
for(const key in realUrl)values[key] = realUrl[key];
|
|
values.href = inputUrl;
|
|
values.pathname = inputUrl.replace(/[?#].*/, '');
|
|
values.origin = values.protocol = '';
|
|
values.toString = values.toJSON = (..._args)=>inputUrl;
|
|
for(const key in values)Object.defineProperty(this, key, {
|
|
enumerable: true,
|
|
configurable: true,
|
|
value: values[key]
|
|
});
|
|
};
|
|
relativeURL.prototype = URL.prototype;
|
|
contextPrototype.U = relativeURL;
|
|
/**
|
|
* Utility function to ensure all variants of an enum are handled.
|
|
*/ function invariant(never, computeMessage) {
|
|
throw new Error(`Invariant: ${computeMessage(never)}`);
|
|
}
|
|
/**
|
|
* A stub function to make `require` available but non-functional in ESM.
|
|
*/ function requireStub(_moduleId) {
|
|
throw new Error('dynamic usage of require is not supported');
|
|
}
|
|
contextPrototype.z = requireStub;
|
|
// Make `globalThis` available to the module in a way that cannot be shadowed by a local variable.
|
|
contextPrototype.g = globalThis;
|
|
function applyModuleFactoryName(factory) {
|
|
// Give the module factory a nice name to improve stack traces.
|
|
Object.defineProperty(factory, 'name', {
|
|
value: 'module evaluation'
|
|
});
|
|
}
|
|
/**
|
|
* This file contains runtime types and functions that are shared between all
|
|
* Turbopack *development* ECMAScript runtimes.
|
|
*
|
|
* It will be appended to the runtime code of each runtime right after the
|
|
* shared runtime utils.
|
|
*/ /* eslint-disable @typescript-eslint/no-unused-vars */ /// <reference path="../base/globals.d.ts" />
|
|
/// <reference path="../../../shared/runtime-utils.ts" />
|
|
// Used in WebWorkers to tell the runtime about the chunk base path
|
|
const browserContextPrototype = Context.prototype;
|
|
var SourceType = /*#__PURE__*/ function(SourceType) {
|
|
/**
|
|
* The module was instantiated because it was included in an evaluated chunk's
|
|
* runtime.
|
|
* SourceData is a ChunkPath.
|
|
*/ SourceType[SourceType["Runtime"] = 0] = "Runtime";
|
|
/**
|
|
* The module was instantiated because a parent module imported it.
|
|
* SourceData is a ModuleId.
|
|
*/ SourceType[SourceType["Parent"] = 1] = "Parent";
|
|
/**
|
|
* The module was instantiated because it was included in a chunk's hot module
|
|
* update.
|
|
* SourceData is an array of ModuleIds or undefined.
|
|
*/ SourceType[SourceType["Update"] = 2] = "Update";
|
|
return SourceType;
|
|
}(SourceType || {});
|
|
const moduleFactories = new Map();
|
|
contextPrototype.M = moduleFactories;
|
|
const availableModules = new Map();
|
|
const availableModuleChunks = new Map();
|
|
function factoryNotAvailableMessage(moduleId, sourceType, sourceData) {
|
|
let instantiationReason;
|
|
switch(sourceType){
|
|
case 0:
|
|
instantiationReason = `as a runtime entry of chunk ${sourceData}`;
|
|
break;
|
|
case 1:
|
|
instantiationReason = `because it was required from module ${sourceData}`;
|
|
break;
|
|
case 2:
|
|
instantiationReason = 'because of an HMR update';
|
|
break;
|
|
default:
|
|
invariant(sourceType, (sourceType)=>`Unknown source type: ${sourceType}`);
|
|
}
|
|
return `Module ${moduleId} was instantiated ${instantiationReason}, but the module factory is not available.`;
|
|
}
|
|
function loadChunk(chunkData) {
|
|
return loadChunkInternal(1, this.m.id, chunkData);
|
|
}
|
|
browserContextPrototype.l = loadChunk;
|
|
function loadInitialChunk(chunkPath, chunkData) {
|
|
return loadChunkInternal(0, chunkPath, chunkData);
|
|
}
|
|
async function loadChunkInternal(sourceType, sourceData, chunkData) {
|
|
if (typeof chunkData === 'string') {
|
|
return loadChunkPath(sourceType, sourceData, chunkData);
|
|
}
|
|
const includedList = chunkData.included || [];
|
|
const modulesPromises = includedList.map((included)=>{
|
|
if (moduleFactories.has(included)) return true;
|
|
return availableModules.get(included);
|
|
});
|
|
if (modulesPromises.length > 0 && modulesPromises.every((p)=>p)) {
|
|
// When all included items are already loaded or loading, we can skip loading ourselves
|
|
await Promise.all(modulesPromises);
|
|
return;
|
|
}
|
|
const includedModuleChunksList = chunkData.moduleChunks || [];
|
|
const moduleChunksPromises = includedModuleChunksList.map((included)=>{
|
|
// TODO(alexkirsz) Do we need this check?
|
|
// if (moduleFactories[included]) return true;
|
|
return availableModuleChunks.get(included);
|
|
}).filter((p)=>p);
|
|
let promise;
|
|
if (moduleChunksPromises.length > 0) {
|
|
// Some module chunks are already loaded or loading.
|
|
if (moduleChunksPromises.length === includedModuleChunksList.length) {
|
|
// When all included module chunks are already loaded or loading, we can skip loading ourselves
|
|
await Promise.all(moduleChunksPromises);
|
|
return;
|
|
}
|
|
const moduleChunksToLoad = new Set();
|
|
for (const moduleChunk of includedModuleChunksList){
|
|
if (!availableModuleChunks.has(moduleChunk)) {
|
|
moduleChunksToLoad.add(moduleChunk);
|
|
}
|
|
}
|
|
for (const moduleChunkToLoad of moduleChunksToLoad){
|
|
const promise = loadChunkPath(sourceType, sourceData, moduleChunkToLoad);
|
|
availableModuleChunks.set(moduleChunkToLoad, promise);
|
|
moduleChunksPromises.push(promise);
|
|
}
|
|
promise = Promise.all(moduleChunksPromises);
|
|
} else {
|
|
promise = loadChunkPath(sourceType, sourceData, chunkData.path);
|
|
// Mark all included module chunks as loading if they are not already loaded or loading.
|
|
for (const includedModuleChunk of includedModuleChunksList){
|
|
if (!availableModuleChunks.has(includedModuleChunk)) {
|
|
availableModuleChunks.set(includedModuleChunk, promise);
|
|
}
|
|
}
|
|
}
|
|
for (const included of includedList){
|
|
if (!availableModules.has(included)) {
|
|
// It might be better to race old and new promises, but it's rare that the new promise will be faster than a request started earlier.
|
|
// In production it's even more rare, because the chunk optimization tries to deduplicate modules anyway.
|
|
availableModules.set(included, promise);
|
|
}
|
|
}
|
|
await promise;
|
|
}
|
|
const loadedChunk = Promise.resolve(undefined);
|
|
const instrumentedBackendLoadChunks = new WeakMap();
|
|
// Do not make this async. React relies on referential equality of the returned Promise.
|
|
function loadChunkByUrl(chunkUrl) {
|
|
return loadChunkByUrlInternal(1, this.m.id, chunkUrl);
|
|
}
|
|
browserContextPrototype.L = loadChunkByUrl;
|
|
// Do not make this async. React relies on referential equality of the returned Promise.
|
|
function loadChunkByUrlInternal(sourceType, sourceData, chunkUrl) {
|
|
const thenable = BACKEND.loadChunkCached(sourceType, chunkUrl);
|
|
let entry = instrumentedBackendLoadChunks.get(thenable);
|
|
if (entry === undefined) {
|
|
const resolve = instrumentedBackendLoadChunks.set.bind(instrumentedBackendLoadChunks, thenable, loadedChunk);
|
|
entry = thenable.then(resolve).catch((cause)=>{
|
|
let loadReason;
|
|
switch(sourceType){
|
|
case 0:
|
|
loadReason = `as a runtime dependency of chunk ${sourceData}`;
|
|
break;
|
|
case 1:
|
|
loadReason = `from module ${sourceData}`;
|
|
break;
|
|
case 2:
|
|
loadReason = 'from an HMR update';
|
|
break;
|
|
default:
|
|
invariant(sourceType, (sourceType)=>`Unknown source type: ${sourceType}`);
|
|
}
|
|
let error = new Error(`Failed to load chunk ${chunkUrl} ${loadReason}${cause ? `: ${cause}` : ''}`, cause ? {
|
|
cause
|
|
} : undefined);
|
|
error.name = 'ChunkLoadError';
|
|
throw error;
|
|
});
|
|
instrumentedBackendLoadChunks.set(thenable, entry);
|
|
}
|
|
return entry;
|
|
}
|
|
// Do not make this async. React relies on referential equality of the returned Promise.
|
|
function loadChunkPath(sourceType, sourceData, chunkPath) {
|
|
const url = getChunkRelativeUrl(chunkPath);
|
|
return loadChunkByUrlInternal(sourceType, sourceData, url);
|
|
}
|
|
/**
|
|
* Returns an absolute url to an asset.
|
|
*/ function resolvePathFromModule(moduleId) {
|
|
const exported = this.r(moduleId);
|
|
return exported?.default ?? exported;
|
|
}
|
|
browserContextPrototype.R = resolvePathFromModule;
|
|
/**
|
|
* no-op for browser
|
|
* @param modulePath
|
|
*/ function resolveAbsolutePath(modulePath) {
|
|
return `/ROOT/${modulePath ?? ''}`;
|
|
}
|
|
browserContextPrototype.P = resolveAbsolutePath;
|
|
/**
|
|
* Returns a blob URL for the worker.
|
|
* @param chunks list of chunks to load
|
|
*/ function getWorkerBlobURL(chunks) {
|
|
// It is important to reverse the array so when bootstrapping we can infer what chunk is being
|
|
// evaluated by poping urls off of this array. See `getPathFromScript`
|
|
let bootstrap = `self.TURBOPACK_WORKER_LOCATION = ${JSON.stringify(location.origin)};
|
|
self.TURBOPACK_CHUNK_SUFFIX = ${JSON.stringify(CHUNK_SUFFIX)};
|
|
self.TURBOPACK_NEXT_CHUNK_URLS = ${JSON.stringify(chunks.reverse().map(getChunkRelativeUrl), null, 2)};
|
|
importScripts(...self.TURBOPACK_NEXT_CHUNK_URLS.map(c => self.TURBOPACK_WORKER_LOCATION + c).reverse());`;
|
|
let blob = new Blob([
|
|
bootstrap
|
|
], {
|
|
type: 'text/javascript'
|
|
});
|
|
return URL.createObjectURL(blob);
|
|
}
|
|
browserContextPrototype.b = getWorkerBlobURL;
|
|
/**
|
|
* Instantiates a runtime module.
|
|
*/ function instantiateRuntimeModule(moduleId, chunkPath) {
|
|
return instantiateModule(moduleId, 0, chunkPath);
|
|
}
|
|
/**
|
|
* Returns the URL relative to the origin where a chunk can be fetched from.
|
|
*/ function getChunkRelativeUrl(chunkPath) {
|
|
return `${CHUNK_BASE_PATH}${chunkPath.split('/').map((p)=>encodeURIComponent(p)).join('/')}${CHUNK_SUFFIX}`;
|
|
}
|
|
function getPathFromScript(chunkScript) {
|
|
if (typeof chunkScript === 'string') {
|
|
return chunkScript;
|
|
}
|
|
const chunkUrl = typeof TURBOPACK_NEXT_CHUNK_URLS !== 'undefined' ? TURBOPACK_NEXT_CHUNK_URLS.pop() : chunkScript.getAttribute('src');
|
|
const src = decodeURIComponent(chunkUrl.replace(/[?#].*$/, ''));
|
|
const path = src.startsWith(CHUNK_BASE_PATH) ? src.slice(CHUNK_BASE_PATH.length) : src;
|
|
return path;
|
|
}
|
|
const regexJsUrl = /\.js(?:\?[^#]*)?(?:#.*)?$/;
|
|
/**
|
|
* Checks if a given path/URL ends with .js, optionally followed by ?query or #fragment.
|
|
*/ function isJs(chunkUrlOrPath) {
|
|
return regexJsUrl.test(chunkUrlOrPath);
|
|
}
|
|
const regexCssUrl = /\.css(?:\?[^#]*)?(?:#.*)?$/;
|
|
/**
|
|
* Checks if a given path/URL ends with .css, optionally followed by ?query or #fragment.
|
|
*/ function isCss(chunkUrl) {
|
|
return regexCssUrl.test(chunkUrl);
|
|
}
|
|
function loadWebAssembly(chunkPath, edgeModule, importsObj) {
|
|
return BACKEND.loadWebAssembly(1, this.m.id, chunkPath, edgeModule, importsObj);
|
|
}
|
|
contextPrototype.w = loadWebAssembly;
|
|
function loadWebAssemblyModule(chunkPath, edgeModule) {
|
|
return BACKEND.loadWebAssemblyModule(1, this.m.id, chunkPath, edgeModule);
|
|
}
|
|
contextPrototype.u = loadWebAssemblyModule;
|
|
/// <reference path="./dev-globals.d.ts" />
|
|
/// <reference path="./dev-protocol.d.ts" />
|
|
/// <reference path="./dev-extensions.ts" />
|
|
const devContextPrototype = Context.prototype;
|
|
/**
|
|
* This file contains runtime types and functions that are shared between all
|
|
* Turbopack *development* ECMAScript runtimes.
|
|
*
|
|
* It will be appended to the runtime code of each runtime right after the
|
|
* shared runtime utils.
|
|
*/ /* eslint-disable @typescript-eslint/no-unused-vars */ const devModuleCache = Object.create(null);
|
|
devContextPrototype.c = devModuleCache;
|
|
class UpdateApplyError extends Error {
|
|
name = 'UpdateApplyError';
|
|
dependencyChain;
|
|
constructor(message, dependencyChain){
|
|
super(message);
|
|
this.dependencyChain = dependencyChain;
|
|
}
|
|
}
|
|
/**
|
|
* Module IDs that are instantiated as part of the runtime of a chunk.
|
|
*/ const runtimeModules = new Set();
|
|
/**
|
|
* Map from module ID to the chunks that contain this module.
|
|
*
|
|
* In HMR, we need to keep track of which modules are contained in which so
|
|
* chunks. This is so we don't eagerly dispose of a module when it is removed
|
|
* from chunk A, but still exists in chunk B.
|
|
*/ const moduleChunksMap = new Map();
|
|
/**
|
|
* Map from a chunk path to all modules it contains.
|
|
*/ const chunkModulesMap = new Map();
|
|
/**
|
|
* Chunk lists that contain a runtime. When these chunk lists receive an update
|
|
* that can't be reconciled with the current state of the page, we need to
|
|
* reload the runtime entirely.
|
|
*/ const runtimeChunkLists = new Set();
|
|
/**
|
|
* Map from a chunk list to the chunk paths it contains.
|
|
*/ const chunkListChunksMap = new Map();
|
|
/**
|
|
* Map from a chunk path to the chunk lists it belongs to.
|
|
*/ const chunkChunkListsMap = new Map();
|
|
/**
|
|
* Maps module IDs to persisted data between executions of their hot module
|
|
* implementation (`hot.data`).
|
|
*/ const moduleHotData = new Map();
|
|
/**
|
|
* Maps module instances to their hot module state.
|
|
*/ const moduleHotState = new Map();
|
|
/**
|
|
* Modules that call `module.hot.invalidate()` (while being updated).
|
|
*/ const queuedInvalidatedModules = new Set();
|
|
/**
|
|
* Gets or instantiates a runtime module.
|
|
*/ // @ts-ignore
|
|
function getOrInstantiateRuntimeModule(chunkPath, moduleId) {
|
|
const module = devModuleCache[moduleId];
|
|
if (module) {
|
|
if (module.error) {
|
|
throw module.error;
|
|
}
|
|
return module;
|
|
}
|
|
// @ts-ignore
|
|
return instantiateModule(moduleId, SourceType.Runtime, chunkPath);
|
|
}
|
|
/**
|
|
* Retrieves a module from the cache, or instantiate it if it is not cached.
|
|
*/ // @ts-ignore Defined in `runtime-utils.ts`
|
|
const getOrInstantiateModuleFromParent = (id, sourceModule)=>{
|
|
if (!sourceModule.hot.active) {
|
|
console.warn(`Unexpected import of module ${id} from module ${sourceModule.id}, which was deleted by an HMR update`);
|
|
}
|
|
const module = devModuleCache[id];
|
|
if (sourceModule.children.indexOf(id) === -1) {
|
|
sourceModule.children.push(id);
|
|
}
|
|
if (module) {
|
|
if (module.error) {
|
|
throw module.error;
|
|
}
|
|
if (module.parents.indexOf(sourceModule.id) === -1) {
|
|
module.parents.push(sourceModule.id);
|
|
}
|
|
return module;
|
|
}
|
|
return instantiateModule(id, SourceType.Parent, sourceModule.id);
|
|
};
|
|
function DevContext(module, exports, refresh) {
|
|
Context.call(this, module, exports);
|
|
this.k = refresh;
|
|
}
|
|
DevContext.prototype = Context.prototype;
|
|
function instantiateModule(moduleId, sourceType, sourceData) {
|
|
// We are in development, this is always a string.
|
|
let id = moduleId;
|
|
const moduleFactory = moduleFactories.get(id);
|
|
if (typeof moduleFactory !== 'function') {
|
|
// This can happen if modules incorrectly handle HMR disposes/updates,
|
|
// e.g. when they keep a `setTimeout` around which still executes old code
|
|
// and contains e.g. a `require("something")` call.
|
|
throw new Error(factoryNotAvailableMessage(id, sourceType, sourceData) + ' It might have been deleted in an HMR update.');
|
|
}
|
|
const hotData = moduleHotData.get(id);
|
|
const { hot, hotState } = createModuleHot(id, hotData);
|
|
let parents;
|
|
switch(sourceType){
|
|
case SourceType.Runtime:
|
|
runtimeModules.add(id);
|
|
parents = [];
|
|
break;
|
|
case SourceType.Parent:
|
|
// No need to add this module as a child of the parent module here, this
|
|
// has already been taken care of in `getOrInstantiateModuleFromParent`.
|
|
parents = [
|
|
sourceData
|
|
];
|
|
break;
|
|
case SourceType.Update:
|
|
parents = sourceData || [];
|
|
break;
|
|
default:
|
|
invariant(sourceType, (sourceType)=>`Unknown source type: ${sourceType}`);
|
|
}
|
|
const module = createModuleObject(id);
|
|
const exports = module.exports;
|
|
module.parents = parents;
|
|
module.children = [];
|
|
module.hot = hot;
|
|
devModuleCache[id] = module;
|
|
moduleHotState.set(module, hotState);
|
|
// NOTE(alexkirsz) This can fail when the module encounters a runtime error.
|
|
try {
|
|
runModuleExecutionHooks(module, (refresh)=>{
|
|
const context = new DevContext(module, exports, refresh);
|
|
moduleFactory(context, module, exports);
|
|
});
|
|
} catch (error) {
|
|
module.error = error;
|
|
throw error;
|
|
}
|
|
if (module.namespaceObject && module.exports !== module.namespaceObject) {
|
|
// in case of a circular dependency: cjs1 -> esm2 -> cjs1
|
|
interopEsm(module.exports, module.namespaceObject);
|
|
}
|
|
return module;
|
|
}
|
|
const DUMMY_REFRESH_CONTEXT = {
|
|
register: (_type, _id)=>{},
|
|
signature: ()=>(_type)=>{},
|
|
registerExports: (_module, _helpers)=>{}
|
|
};
|
|
/**
|
|
* NOTE(alexkirsz) Webpack has a "module execution" interception hook that
|
|
* Next.js' React Refresh runtime hooks into to add module context to the
|
|
* refresh registry.
|
|
*/ function runModuleExecutionHooks(module, executeModule) {
|
|
if (typeof globalThis.$RefreshInterceptModuleExecution$ === 'function') {
|
|
const cleanupReactRefreshIntercept = globalThis.$RefreshInterceptModuleExecution$(module.id);
|
|
try {
|
|
executeModule({
|
|
register: globalThis.$RefreshReg$,
|
|
signature: globalThis.$RefreshSig$,
|
|
registerExports: registerExportsAndSetupBoundaryForReactRefresh
|
|
});
|
|
} finally{
|
|
// Always cleanup the intercept, even if module execution failed.
|
|
cleanupReactRefreshIntercept();
|
|
}
|
|
} else {
|
|
// If the react refresh hooks are not installed we need to bind dummy functions.
|
|
// This is expected when running in a Web Worker. It is also common in some of
|
|
// our test environments.
|
|
executeModule(DUMMY_REFRESH_CONTEXT);
|
|
}
|
|
}
|
|
/**
|
|
* This is adapted from https://github.com/vercel/next.js/blob/3466862d9dc9c8bb3131712134d38757b918d1c0/packages/react-refresh-utils/internal/ReactRefreshModule.runtime.ts
|
|
*/ function registerExportsAndSetupBoundaryForReactRefresh(module, helpers) {
|
|
const currentExports = module.exports;
|
|
const prevExports = module.hot.data.prevExports ?? null;
|
|
helpers.registerExportsForReactRefresh(currentExports, module.id);
|
|
// A module can be accepted automatically based on its exports, e.g. when
|
|
// it is a Refresh Boundary.
|
|
if (helpers.isReactRefreshBoundary(currentExports)) {
|
|
// Save the previous exports on update, so we can compare the boundary
|
|
// signatures.
|
|
module.hot.dispose((data)=>{
|
|
data.prevExports = currentExports;
|
|
});
|
|
// Unconditionally accept an update to this module, we'll check if it's
|
|
// still a Refresh Boundary later.
|
|
module.hot.accept();
|
|
// This field is set when the previous version of this module was a
|
|
// Refresh Boundary, letting us know we need to check for invalidation or
|
|
// enqueue an update.
|
|
if (prevExports !== null) {
|
|
// A boundary can become ineligible if its exports are incompatible
|
|
// with the previous exports.
|
|
//
|
|
// For example, if you add/remove/change exports, we'll want to
|
|
// re-execute the importing modules, and force those components to
|
|
// re-render. Similarly, if you convert a class component to a
|
|
// function, we want to invalidate the boundary.
|
|
if (helpers.shouldInvalidateReactRefreshBoundary(helpers.getRefreshBoundarySignature(prevExports), helpers.getRefreshBoundarySignature(currentExports))) {
|
|
module.hot.invalidate();
|
|
} else {
|
|
helpers.scheduleUpdate();
|
|
}
|
|
}
|
|
} else {
|
|
// Since we just executed the code for the module, it's possible that the
|
|
// new exports made it ineligible for being a boundary.
|
|
// We only care about the case when we were _previously_ a boundary,
|
|
// because we already accepted this update (accidental side effect).
|
|
const isNoLongerABoundary = prevExports !== null;
|
|
if (isNoLongerABoundary) {
|
|
module.hot.invalidate();
|
|
}
|
|
}
|
|
}
|
|
function formatDependencyChain(dependencyChain) {
|
|
return `Dependency chain: ${dependencyChain.join(' -> ')}`;
|
|
}
|
|
function computeOutdatedModules(added, modified) {
|
|
const newModuleFactories = new Map();
|
|
for (const [moduleId, entry] of added){
|
|
if (entry != null) {
|
|
newModuleFactories.set(moduleId, _eval(entry));
|
|
}
|
|
}
|
|
const outdatedModules = computedInvalidatedModules(modified.keys());
|
|
for (const [moduleId, entry] of modified){
|
|
newModuleFactories.set(moduleId, _eval(entry));
|
|
}
|
|
return {
|
|
outdatedModules,
|
|
newModuleFactories
|
|
};
|
|
}
|
|
function computedInvalidatedModules(invalidated) {
|
|
const outdatedModules = new Set();
|
|
for (const moduleId of invalidated){
|
|
const effect = getAffectedModuleEffects(moduleId);
|
|
switch(effect.type){
|
|
case 'unaccepted':
|
|
throw new UpdateApplyError(`cannot apply update: unaccepted module. ${formatDependencyChain(effect.dependencyChain)}.`, effect.dependencyChain);
|
|
case 'self-declined':
|
|
throw new UpdateApplyError(`cannot apply update: self-declined module. ${formatDependencyChain(effect.dependencyChain)}.`, effect.dependencyChain);
|
|
case 'accepted':
|
|
for (const outdatedModuleId of effect.outdatedModules){
|
|
outdatedModules.add(outdatedModuleId);
|
|
}
|
|
break;
|
|
// TODO(alexkirsz) Dependencies: handle dependencies effects.
|
|
default:
|
|
invariant(effect, (effect)=>`Unknown effect type: ${effect?.type}`);
|
|
}
|
|
}
|
|
return outdatedModules;
|
|
}
|
|
function computeOutdatedSelfAcceptedModules(outdatedModules) {
|
|
const outdatedSelfAcceptedModules = [];
|
|
for (const moduleId of outdatedModules){
|
|
const module = devModuleCache[moduleId];
|
|
const hotState = moduleHotState.get(module);
|
|
if (module && hotState.selfAccepted && !hotState.selfInvalidated) {
|
|
outdatedSelfAcceptedModules.push({
|
|
moduleId,
|
|
errorHandler: hotState.selfAccepted
|
|
});
|
|
}
|
|
}
|
|
return outdatedSelfAcceptedModules;
|
|
}
|
|
/**
|
|
* Adds, deletes, and moves modules between chunks. This must happen before the
|
|
* dispose phase as it needs to know which modules were removed from all chunks,
|
|
* which we can only compute *after* taking care of added and moved modules.
|
|
*/ function updateChunksPhase(chunksAddedModules, chunksDeletedModules) {
|
|
for (const [chunkPath, addedModuleIds] of chunksAddedModules){
|
|
for (const moduleId of addedModuleIds){
|
|
addModuleToChunk(moduleId, chunkPath);
|
|
}
|
|
}
|
|
const disposedModules = new Set();
|
|
for (const [chunkPath, addedModuleIds] of chunksDeletedModules){
|
|
for (const moduleId of addedModuleIds){
|
|
if (removeModuleFromChunk(moduleId, chunkPath)) {
|
|
disposedModules.add(moduleId);
|
|
}
|
|
}
|
|
}
|
|
return {
|
|
disposedModules
|
|
};
|
|
}
|
|
function disposePhase(outdatedModules, disposedModules) {
|
|
for (const moduleId of outdatedModules){
|
|
disposeModule(moduleId, 'replace');
|
|
}
|
|
for (const moduleId of disposedModules){
|
|
disposeModule(moduleId, 'clear');
|
|
}
|
|
// Removing modules from the module cache is a separate step.
|
|
// We also want to keep track of previous parents of the outdated modules.
|
|
const outdatedModuleParents = new Map();
|
|
for (const moduleId of outdatedModules){
|
|
const oldModule = devModuleCache[moduleId];
|
|
outdatedModuleParents.set(moduleId, oldModule?.parents);
|
|
delete devModuleCache[moduleId];
|
|
}
|
|
// TODO(alexkirsz) Dependencies: remove outdated dependency from module
|
|
// children.
|
|
return {
|
|
outdatedModuleParents
|
|
};
|
|
}
|
|
/**
|
|
* Disposes of an instance of a module.
|
|
*
|
|
* Returns the persistent hot data that should be kept for the next module
|
|
* instance.
|
|
*
|
|
* NOTE: mode = "replace" will not remove modules from the devModuleCache
|
|
* This must be done in a separate step afterwards.
|
|
* This is important because all modules need to be disposed to update the
|
|
* parent/child relationships before they are actually removed from the devModuleCache.
|
|
* If this was done in this method, the following disposeModule calls won't find
|
|
* the module from the module id in the cache.
|
|
*/ function disposeModule(moduleId, mode) {
|
|
const module = devModuleCache[moduleId];
|
|
if (!module) {
|
|
return;
|
|
}
|
|
const hotState = moduleHotState.get(module);
|
|
const data = {};
|
|
// Run the `hot.dispose` handler, if any, passing in the persistent
|
|
// `hot.data` object.
|
|
for (const disposeHandler of hotState.disposeHandlers){
|
|
disposeHandler(data);
|
|
}
|
|
// This used to warn in `getOrInstantiateModuleFromParent` when a disposed
|
|
// module is still importing other modules.
|
|
module.hot.active = false;
|
|
moduleHotState.delete(module);
|
|
// TODO(alexkirsz) Dependencies: delete the module from outdated deps.
|
|
// Remove the disposed module from its children's parent list.
|
|
// It will be added back once the module re-instantiates and imports its
|
|
// children again.
|
|
for (const childId of module.children){
|
|
const child = devModuleCache[childId];
|
|
if (!child) {
|
|
continue;
|
|
}
|
|
const idx = child.parents.indexOf(module.id);
|
|
if (idx >= 0) {
|
|
child.parents.splice(idx, 1);
|
|
}
|
|
}
|
|
switch(mode){
|
|
case 'clear':
|
|
delete devModuleCache[module.id];
|
|
moduleHotData.delete(module.id);
|
|
break;
|
|
case 'replace':
|
|
moduleHotData.set(module.id, data);
|
|
break;
|
|
default:
|
|
invariant(mode, (mode)=>`invalid mode: ${mode}`);
|
|
}
|
|
}
|
|
function applyPhase(outdatedSelfAcceptedModules, newModuleFactories, outdatedModuleParents, reportError) {
|
|
// Update module factories.
|
|
for (const [moduleId, factory] of newModuleFactories.entries()){
|
|
applyModuleFactoryName(factory);
|
|
moduleFactories.set(moduleId, factory);
|
|
}
|
|
// TODO(alexkirsz) Run new runtime entries here.
|
|
// TODO(alexkirsz) Dependencies: call accept handlers for outdated deps.
|
|
// Re-instantiate all outdated self-accepted modules.
|
|
for (const { moduleId, errorHandler } of outdatedSelfAcceptedModules){
|
|
try {
|
|
instantiateModule(moduleId, SourceType.Update, outdatedModuleParents.get(moduleId));
|
|
} catch (err) {
|
|
if (typeof errorHandler === 'function') {
|
|
try {
|
|
errorHandler(err, {
|
|
moduleId,
|
|
module: devModuleCache[moduleId]
|
|
});
|
|
} catch (err2) {
|
|
reportError(err2);
|
|
reportError(err);
|
|
}
|
|
} else {
|
|
reportError(err);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
function applyUpdate(update) {
|
|
switch(update.type){
|
|
case 'ChunkListUpdate':
|
|
applyChunkListUpdate(update);
|
|
break;
|
|
default:
|
|
invariant(update, (update)=>`Unknown update type: ${update.type}`);
|
|
}
|
|
}
|
|
function applyChunkListUpdate(update) {
|
|
if (update.merged != null) {
|
|
for (const merged of update.merged){
|
|
switch(merged.type){
|
|
case 'EcmascriptMergedUpdate':
|
|
applyEcmascriptMergedUpdate(merged);
|
|
break;
|
|
default:
|
|
invariant(merged, (merged)=>`Unknown merged type: ${merged.type}`);
|
|
}
|
|
}
|
|
}
|
|
if (update.chunks != null) {
|
|
for (const [chunkPath, chunkUpdate] of Object.entries(update.chunks)){
|
|
const chunkUrl = getChunkRelativeUrl(chunkPath);
|
|
switch(chunkUpdate.type){
|
|
case 'added':
|
|
BACKEND.loadChunkCached(SourceType.Update, chunkUrl);
|
|
break;
|
|
case 'total':
|
|
DEV_BACKEND.reloadChunk?.(chunkUrl);
|
|
break;
|
|
case 'deleted':
|
|
DEV_BACKEND.unloadChunk?.(chunkUrl);
|
|
break;
|
|
case 'partial':
|
|
invariant(chunkUpdate.instruction, (instruction)=>`Unknown partial instruction: ${JSON.stringify(instruction)}.`);
|
|
break;
|
|
default:
|
|
invariant(chunkUpdate, (chunkUpdate)=>`Unknown chunk update type: ${chunkUpdate.type}`);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
function applyEcmascriptMergedUpdate(update) {
|
|
const { entries = {}, chunks = {} } = update;
|
|
const { added, modified, chunksAdded, chunksDeleted } = computeChangedModules(entries, chunks);
|
|
const { outdatedModules, newModuleFactories } = computeOutdatedModules(added, modified);
|
|
const { disposedModules } = updateChunksPhase(chunksAdded, chunksDeleted);
|
|
applyInternal(outdatedModules, disposedModules, newModuleFactories);
|
|
}
|
|
function applyInvalidatedModules(outdatedModules) {
|
|
if (queuedInvalidatedModules.size > 0) {
|
|
computedInvalidatedModules(queuedInvalidatedModules).forEach((moduleId)=>{
|
|
outdatedModules.add(moduleId);
|
|
});
|
|
queuedInvalidatedModules.clear();
|
|
}
|
|
return outdatedModules;
|
|
}
|
|
function applyInternal(outdatedModules, disposedModules, newModuleFactories) {
|
|
outdatedModules = applyInvalidatedModules(outdatedModules);
|
|
const outdatedSelfAcceptedModules = computeOutdatedSelfAcceptedModules(outdatedModules);
|
|
const { outdatedModuleParents } = disposePhase(outdatedModules, disposedModules);
|
|
// we want to continue on error and only throw the error after we tried applying all updates
|
|
let error;
|
|
function reportError(err) {
|
|
if (!error) error = err;
|
|
}
|
|
applyPhase(outdatedSelfAcceptedModules, newModuleFactories, outdatedModuleParents, reportError);
|
|
if (error) {
|
|
throw error;
|
|
}
|
|
if (queuedInvalidatedModules.size > 0) {
|
|
applyInternal(new Set(), [], new Map());
|
|
}
|
|
}
|
|
function computeChangedModules(entries, updates) {
|
|
const chunksAdded = new Map();
|
|
const chunksDeleted = new Map();
|
|
const added = new Map();
|
|
const modified = new Map();
|
|
const deleted = new Set();
|
|
for (const [chunkPath, mergedChunkUpdate] of Object.entries(updates)){
|
|
switch(mergedChunkUpdate.type){
|
|
case 'added':
|
|
{
|
|
const updateAdded = new Set(mergedChunkUpdate.modules);
|
|
for (const moduleId of updateAdded){
|
|
added.set(moduleId, entries[moduleId]);
|
|
}
|
|
chunksAdded.set(chunkPath, updateAdded);
|
|
break;
|
|
}
|
|
case 'deleted':
|
|
{
|
|
// We could also use `mergedChunkUpdate.modules` here.
|
|
const updateDeleted = new Set(chunkModulesMap.get(chunkPath));
|
|
for (const moduleId of updateDeleted){
|
|
deleted.add(moduleId);
|
|
}
|
|
chunksDeleted.set(chunkPath, updateDeleted);
|
|
break;
|
|
}
|
|
case 'partial':
|
|
{
|
|
const updateAdded = new Set(mergedChunkUpdate.added);
|
|
const updateDeleted = new Set(mergedChunkUpdate.deleted);
|
|
for (const moduleId of updateAdded){
|
|
added.set(moduleId, entries[moduleId]);
|
|
}
|
|
for (const moduleId of updateDeleted){
|
|
deleted.add(moduleId);
|
|
}
|
|
chunksAdded.set(chunkPath, updateAdded);
|
|
chunksDeleted.set(chunkPath, updateDeleted);
|
|
break;
|
|
}
|
|
default:
|
|
invariant(mergedChunkUpdate, (mergedChunkUpdate)=>`Unknown merged chunk update type: ${mergedChunkUpdate.type}`);
|
|
}
|
|
}
|
|
// If a module was added from one chunk and deleted from another in the same update,
|
|
// consider it to be modified, as it means the module was moved from one chunk to another
|
|
// AND has new code in a single update.
|
|
for (const moduleId of added.keys()){
|
|
if (deleted.has(moduleId)) {
|
|
added.delete(moduleId);
|
|
deleted.delete(moduleId);
|
|
}
|
|
}
|
|
for (const [moduleId, entry] of Object.entries(entries)){
|
|
// Modules that haven't been added to any chunk but have new code are considered
|
|
// to be modified.
|
|
// This needs to be under the previous loop, as we need it to get rid of modules
|
|
// that were added and deleted in the same update.
|
|
if (!added.has(moduleId)) {
|
|
modified.set(moduleId, entry);
|
|
}
|
|
}
|
|
return {
|
|
added,
|
|
deleted,
|
|
modified,
|
|
chunksAdded,
|
|
chunksDeleted
|
|
};
|
|
}
|
|
function getAffectedModuleEffects(moduleId) {
|
|
const outdatedModules = new Set();
|
|
const queue = [
|
|
{
|
|
moduleId,
|
|
dependencyChain: []
|
|
}
|
|
];
|
|
let nextItem;
|
|
while(nextItem = queue.shift()){
|
|
const { moduleId, dependencyChain } = nextItem;
|
|
if (moduleId != null) {
|
|
if (outdatedModules.has(moduleId)) {
|
|
continue;
|
|
}
|
|
outdatedModules.add(moduleId);
|
|
}
|
|
// We've arrived at the runtime of the chunk, which means that nothing
|
|
// else above can accept this update.
|
|
if (moduleId === undefined) {
|
|
return {
|
|
type: 'unaccepted',
|
|
dependencyChain
|
|
};
|
|
}
|
|
const module = devModuleCache[moduleId];
|
|
const hotState = moduleHotState.get(module);
|
|
if (// The module is not in the cache. Since this is a "modified" update,
|
|
// it means that the module was never instantiated before.
|
|
!module || hotState.selfAccepted && !hotState.selfInvalidated) {
|
|
continue;
|
|
}
|
|
if (hotState.selfDeclined) {
|
|
return {
|
|
type: 'self-declined',
|
|
dependencyChain,
|
|
moduleId
|
|
};
|
|
}
|
|
if (runtimeModules.has(moduleId)) {
|
|
queue.push({
|
|
moduleId: undefined,
|
|
dependencyChain: [
|
|
...dependencyChain,
|
|
moduleId
|
|
]
|
|
});
|
|
continue;
|
|
}
|
|
for (const parentId of module.parents){
|
|
const parent = devModuleCache[parentId];
|
|
if (!parent) {
|
|
continue;
|
|
}
|
|
// TODO(alexkirsz) Dependencies: check accepted and declined
|
|
// dependencies here.
|
|
queue.push({
|
|
moduleId: parentId,
|
|
dependencyChain: [
|
|
...dependencyChain,
|
|
moduleId
|
|
]
|
|
});
|
|
}
|
|
}
|
|
return {
|
|
type: 'accepted',
|
|
moduleId,
|
|
outdatedModules
|
|
};
|
|
}
|
|
function handleApply(chunkListPath, update) {
|
|
switch(update.type){
|
|
case 'partial':
|
|
{
|
|
// This indicates that the update is can be applied to the current state of the application.
|
|
applyUpdate(update.instruction);
|
|
break;
|
|
}
|
|
case 'restart':
|
|
{
|
|
// This indicates that there is no way to apply the update to the
|
|
// current state of the application, and that the application must be
|
|
// restarted.
|
|
DEV_BACKEND.restart();
|
|
break;
|
|
}
|
|
case 'notFound':
|
|
{
|
|
// This indicates that the chunk list no longer exists: either the dynamic import which created it was removed,
|
|
// or the page itself was deleted.
|
|
// If it is a dynamic import, we simply discard all modules that the chunk has exclusive access to.
|
|
// If it is a runtime chunk list, we restart the application.
|
|
if (runtimeChunkLists.has(chunkListPath)) {
|
|
DEV_BACKEND.restart();
|
|
} else {
|
|
disposeChunkList(chunkListPath);
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
throw new Error(`Unknown update type: ${update.type}`);
|
|
}
|
|
}
|
|
function createModuleHot(moduleId, hotData) {
|
|
const hotState = {
|
|
selfAccepted: false,
|
|
selfDeclined: false,
|
|
selfInvalidated: false,
|
|
disposeHandlers: []
|
|
};
|
|
const hot = {
|
|
// TODO(alexkirsz) This is not defined in the HMR API. It was used to
|
|
// decide whether to warn whenever an HMR-disposed module required other
|
|
// modules. We might want to remove it.
|
|
active: true,
|
|
data: hotData ?? {},
|
|
// TODO(alexkirsz) Support full (dep, callback, errorHandler) form.
|
|
accept: (modules, _callback, _errorHandler)=>{
|
|
if (modules === undefined) {
|
|
hotState.selfAccepted = true;
|
|
} else if (typeof modules === 'function') {
|
|
hotState.selfAccepted = modules;
|
|
} else {
|
|
throw new Error('unsupported `accept` signature');
|
|
}
|
|
},
|
|
decline: (dep)=>{
|
|
if (dep === undefined) {
|
|
hotState.selfDeclined = true;
|
|
} else {
|
|
throw new Error('unsupported `decline` signature');
|
|
}
|
|
},
|
|
dispose: (callback)=>{
|
|
hotState.disposeHandlers.push(callback);
|
|
},
|
|
addDisposeHandler: (callback)=>{
|
|
hotState.disposeHandlers.push(callback);
|
|
},
|
|
removeDisposeHandler: (callback)=>{
|
|
const idx = hotState.disposeHandlers.indexOf(callback);
|
|
if (idx >= 0) {
|
|
hotState.disposeHandlers.splice(idx, 1);
|
|
}
|
|
},
|
|
invalidate: ()=>{
|
|
hotState.selfInvalidated = true;
|
|
queuedInvalidatedModules.add(moduleId);
|
|
},
|
|
// NOTE(alexkirsz) This is part of the management API, which we don't
|
|
// implement, but the Next.js React Refresh runtime uses this to decide
|
|
// whether to schedule an update.
|
|
status: ()=>'idle',
|
|
// NOTE(alexkirsz) Since we always return "idle" for now, these are no-ops.
|
|
addStatusHandler: (_handler)=>{},
|
|
removeStatusHandler: (_handler)=>{},
|
|
// NOTE(jridgewell) Check returns the list of updated modules, but we don't
|
|
// want the webpack code paths to ever update (the turbopack paths handle
|
|
// this already).
|
|
check: ()=>Promise.resolve(null)
|
|
};
|
|
return {
|
|
hot,
|
|
hotState
|
|
};
|
|
}
|
|
/**
|
|
* Removes a module from a chunk.
|
|
* Returns `true` if there are no remaining chunks including this module.
|
|
*/ function removeModuleFromChunk(moduleId, chunkPath) {
|
|
const moduleChunks = moduleChunksMap.get(moduleId);
|
|
moduleChunks.delete(chunkPath);
|
|
const chunkModules = chunkModulesMap.get(chunkPath);
|
|
chunkModules.delete(moduleId);
|
|
const noRemainingModules = chunkModules.size === 0;
|
|
if (noRemainingModules) {
|
|
chunkModulesMap.delete(chunkPath);
|
|
}
|
|
const noRemainingChunks = moduleChunks.size === 0;
|
|
if (noRemainingChunks) {
|
|
moduleChunksMap.delete(moduleId);
|
|
}
|
|
return noRemainingChunks;
|
|
}
|
|
/**
|
|
* Disposes of a chunk list and its corresponding exclusive chunks.
|
|
*/ function disposeChunkList(chunkListPath) {
|
|
const chunkPaths = chunkListChunksMap.get(chunkListPath);
|
|
if (chunkPaths == null) {
|
|
return false;
|
|
}
|
|
chunkListChunksMap.delete(chunkListPath);
|
|
for (const chunkPath of chunkPaths){
|
|
const chunkChunkLists = chunkChunkListsMap.get(chunkPath);
|
|
chunkChunkLists.delete(chunkListPath);
|
|
if (chunkChunkLists.size === 0) {
|
|
chunkChunkListsMap.delete(chunkPath);
|
|
disposeChunk(chunkPath);
|
|
}
|
|
}
|
|
// We must also dispose of the chunk list's chunk itself to ensure it may
|
|
// be reloaded properly in the future.
|
|
const chunkListUrl = getChunkRelativeUrl(chunkListPath);
|
|
DEV_BACKEND.unloadChunk?.(chunkListUrl);
|
|
return true;
|
|
}
|
|
/**
|
|
* Disposes of a chunk and its corresponding exclusive modules.
|
|
*
|
|
* @returns Whether the chunk was disposed of.
|
|
*/ function disposeChunk(chunkPath) {
|
|
const chunkUrl = getChunkRelativeUrl(chunkPath);
|
|
// This should happen whether the chunk has any modules in it or not.
|
|
// For instance, CSS chunks have no modules in them, but they still need to be unloaded.
|
|
DEV_BACKEND.unloadChunk?.(chunkUrl);
|
|
const chunkModules = chunkModulesMap.get(chunkPath);
|
|
if (chunkModules == null) {
|
|
return false;
|
|
}
|
|
chunkModules.delete(chunkPath);
|
|
for (const moduleId of chunkModules){
|
|
const moduleChunks = moduleChunksMap.get(moduleId);
|
|
moduleChunks.delete(chunkPath);
|
|
const noRemainingChunks = moduleChunks.size === 0;
|
|
if (noRemainingChunks) {
|
|
moduleChunksMap.delete(moduleId);
|
|
disposeModule(moduleId, 'clear');
|
|
availableModules.delete(moduleId);
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
/**
|
|
* Adds a module to a chunk.
|
|
*/ function addModuleToChunk(moduleId, chunkPath) {
|
|
let moduleChunks = moduleChunksMap.get(moduleId);
|
|
if (!moduleChunks) {
|
|
moduleChunks = new Set([
|
|
chunkPath
|
|
]);
|
|
moduleChunksMap.set(moduleId, moduleChunks);
|
|
} else {
|
|
moduleChunks.add(chunkPath);
|
|
}
|
|
let chunkModules = chunkModulesMap.get(chunkPath);
|
|
if (!chunkModules) {
|
|
chunkModules = new Set([
|
|
moduleId
|
|
]);
|
|
chunkModulesMap.set(chunkPath, chunkModules);
|
|
} else {
|
|
chunkModules.add(moduleId);
|
|
}
|
|
}
|
|
/**
|
|
* Marks a chunk list as a runtime chunk list. There can be more than one
|
|
* runtime chunk list. For instance, integration tests can have multiple chunk
|
|
* groups loaded at runtime, each with its own chunk list.
|
|
*/ function markChunkListAsRuntime(chunkListPath) {
|
|
runtimeChunkLists.add(chunkListPath);
|
|
}
|
|
function registerChunk(registration) {
|
|
const chunkPath = getPathFromScript(registration[0]);
|
|
let runtimeParams;
|
|
// When bootstrapping we are passed a single runtimeParams object so we can distinguish purely based on length
|
|
if (registration.length === 2) {
|
|
runtimeParams = registration[1];
|
|
} else {
|
|
runtimeParams = undefined;
|
|
installCompressedModuleFactories(registration, /* offset= */ 1, moduleFactories, (id)=>addModuleToChunk(id, chunkPath));
|
|
}
|
|
return BACKEND.registerChunk(chunkPath, runtimeParams);
|
|
}
|
|
/**
|
|
* Subscribes to chunk list updates from the update server and applies them.
|
|
*/ function registerChunkList(chunkList) {
|
|
const chunkListScript = chunkList.script;
|
|
const chunkListPath = getPathFromScript(chunkListScript);
|
|
// The "chunk" is also registered to finish the loading in the backend
|
|
BACKEND.registerChunk(chunkListPath);
|
|
globalThis.TURBOPACK_CHUNK_UPDATE_LISTENERS.push([
|
|
chunkListPath,
|
|
handleApply.bind(null, chunkListPath)
|
|
]);
|
|
// Adding chunks to chunk lists and vice versa.
|
|
const chunkPaths = new Set(chunkList.chunks.map(getChunkPath));
|
|
chunkListChunksMap.set(chunkListPath, chunkPaths);
|
|
for (const chunkPath of chunkPaths){
|
|
let chunkChunkLists = chunkChunkListsMap.get(chunkPath);
|
|
if (!chunkChunkLists) {
|
|
chunkChunkLists = new Set([
|
|
chunkListPath
|
|
]);
|
|
chunkChunkListsMap.set(chunkPath, chunkChunkLists);
|
|
} else {
|
|
chunkChunkLists.add(chunkListPath);
|
|
}
|
|
}
|
|
if (chunkList.source === 'entry') {
|
|
markChunkListAsRuntime(chunkListPath);
|
|
}
|
|
}
|
|
globalThis.TURBOPACK_CHUNK_UPDATE_LISTENERS ??= [];
|
|
/**
|
|
* This file contains the runtime code specific to the Turbopack development
|
|
* ECMAScript DOM runtime.
|
|
*
|
|
* It will be appended to the base development runtime code.
|
|
*/ /* eslint-disable @typescript-eslint/no-unused-vars */ /// <reference path="../../../browser/runtime/base/runtime-base.ts" />
|
|
/// <reference path="../../../shared/runtime-types.d.ts" />
|
|
let BACKEND;
|
|
/**
|
|
* Maps chunk paths to the corresponding resolver.
|
|
*/ const chunkResolvers = new Map();
|
|
(()=>{
|
|
BACKEND = {
|
|
async registerChunk (chunkPath, params) {
|
|
const chunkUrl = getChunkRelativeUrl(chunkPath);
|
|
const resolver = getOrCreateResolver(chunkUrl);
|
|
resolver.resolve();
|
|
if (params == null) {
|
|
return;
|
|
}
|
|
for (const otherChunkData of params.otherChunks){
|
|
const otherChunkPath = getChunkPath(otherChunkData);
|
|
const otherChunkUrl = getChunkRelativeUrl(otherChunkPath);
|
|
// Chunk might have started loading, so we want to avoid triggering another load.
|
|
getOrCreateResolver(otherChunkUrl);
|
|
}
|
|
// This waits for chunks to be loaded, but also marks included items as available.
|
|
await Promise.all(params.otherChunks.map((otherChunkData)=>loadInitialChunk(chunkPath, otherChunkData)));
|
|
if (params.runtimeModuleIds.length > 0) {
|
|
for (const moduleId of params.runtimeModuleIds){
|
|
getOrInstantiateRuntimeModule(chunkPath, moduleId);
|
|
}
|
|
}
|
|
},
|
|
/**
|
|
* Loads the given chunk, and returns a promise that resolves once the chunk
|
|
* has been loaded.
|
|
*/ loadChunkCached (sourceType, chunkUrl) {
|
|
return doLoadChunk(sourceType, chunkUrl);
|
|
},
|
|
async loadWebAssembly (_sourceType, _sourceData, wasmChunkPath, _edgeModule, importsObj) {
|
|
const req = fetchWebAssembly(wasmChunkPath);
|
|
const { instance } = await WebAssembly.instantiateStreaming(req, importsObj);
|
|
return instance.exports;
|
|
},
|
|
async loadWebAssemblyModule (_sourceType, _sourceData, wasmChunkPath, _edgeModule) {
|
|
const req = fetchWebAssembly(wasmChunkPath);
|
|
return await WebAssembly.compileStreaming(req);
|
|
}
|
|
};
|
|
function getOrCreateResolver(chunkUrl) {
|
|
let resolver = chunkResolvers.get(chunkUrl);
|
|
if (!resolver) {
|
|
let resolve;
|
|
let reject;
|
|
const promise = new Promise((innerResolve, innerReject)=>{
|
|
resolve = innerResolve;
|
|
reject = innerReject;
|
|
});
|
|
resolver = {
|
|
resolved: false,
|
|
loadingStarted: false,
|
|
promise,
|
|
resolve: ()=>{
|
|
resolver.resolved = true;
|
|
resolve();
|
|
},
|
|
reject: reject
|
|
};
|
|
chunkResolvers.set(chunkUrl, resolver);
|
|
}
|
|
return resolver;
|
|
}
|
|
/**
|
|
* Loads the given chunk, and returns a promise that resolves once the chunk
|
|
* has been loaded.
|
|
*/ function doLoadChunk(sourceType, chunkUrl) {
|
|
const resolver = getOrCreateResolver(chunkUrl);
|
|
if (resolver.loadingStarted) {
|
|
return resolver.promise;
|
|
}
|
|
if (sourceType === SourceType.Runtime) {
|
|
// We don't need to load chunks references from runtime code, as they're already
|
|
// present in the DOM.
|
|
resolver.loadingStarted = true;
|
|
if (isCss(chunkUrl)) {
|
|
// CSS chunks do not register themselves, and as such must be marked as
|
|
// loaded instantly.
|
|
resolver.resolve();
|
|
}
|
|
// We need to wait for JS chunks to register themselves within `registerChunk`
|
|
// before we can start instantiating runtime modules, hence the absence of
|
|
// `resolver.resolve()` in this branch.
|
|
return resolver.promise;
|
|
}
|
|
if (typeof importScripts === 'function') {
|
|
// We're in a web worker
|
|
if (isCss(chunkUrl)) {
|
|
// ignore
|
|
} else if (isJs(chunkUrl)) {
|
|
self.TURBOPACK_NEXT_CHUNK_URLS.push(chunkUrl);
|
|
importScripts(TURBOPACK_WORKER_LOCATION + chunkUrl);
|
|
} else {
|
|
throw new Error(`can't infer type of chunk from URL ${chunkUrl} in worker`);
|
|
}
|
|
} else {
|
|
// TODO(PACK-2140): remove this once all filenames are guaranteed to be escaped.
|
|
const decodedChunkUrl = decodeURI(chunkUrl);
|
|
if (isCss(chunkUrl)) {
|
|
const previousLinks = document.querySelectorAll(`link[rel=stylesheet][href="${chunkUrl}"],link[rel=stylesheet][href^="${chunkUrl}?"],link[rel=stylesheet][href="${decodedChunkUrl}"],link[rel=stylesheet][href^="${decodedChunkUrl}?"]`);
|
|
if (previousLinks.length > 0) {
|
|
// CSS chunks do not register themselves, and as such must be marked as
|
|
// loaded instantly.
|
|
resolver.resolve();
|
|
} else {
|
|
const link = document.createElement('link');
|
|
link.rel = 'stylesheet';
|
|
link.href = chunkUrl;
|
|
link.onerror = ()=>{
|
|
resolver.reject();
|
|
};
|
|
link.onload = ()=>{
|
|
// CSS chunks do not register themselves, and as such must be marked as
|
|
// loaded instantly.
|
|
resolver.resolve();
|
|
};
|
|
// Append to the `head` for webpack compatibility.
|
|
document.head.appendChild(link);
|
|
}
|
|
} else if (isJs(chunkUrl)) {
|
|
const previousScripts = document.querySelectorAll(`script[src="${chunkUrl}"],script[src^="${chunkUrl}?"],script[src="${decodedChunkUrl}"],script[src^="${decodedChunkUrl}?"]`);
|
|
if (previousScripts.length > 0) {
|
|
// There is this edge where the script already failed loading, but we
|
|
// can't detect that. The Promise will never resolve in this case.
|
|
for (const script of Array.from(previousScripts)){
|
|
script.addEventListener('error', ()=>{
|
|
resolver.reject();
|
|
});
|
|
}
|
|
} else {
|
|
const script = document.createElement('script');
|
|
script.src = chunkUrl;
|
|
// We'll only mark the chunk as loaded once the script has been executed,
|
|
// which happens in `registerChunk`. Hence the absence of `resolve()` in
|
|
// this branch.
|
|
script.onerror = ()=>{
|
|
resolver.reject();
|
|
};
|
|
// Append to the `head` for webpack compatibility.
|
|
document.head.appendChild(script);
|
|
}
|
|
} else {
|
|
throw new Error(`can't infer type of chunk from URL ${chunkUrl}`);
|
|
}
|
|
}
|
|
resolver.loadingStarted = true;
|
|
return resolver.promise;
|
|
}
|
|
function fetchWebAssembly(wasmChunkPath) {
|
|
return fetch(getChunkRelativeUrl(wasmChunkPath));
|
|
}
|
|
})();
|
|
/**
|
|
* This file contains the runtime code specific to the Turbopack development
|
|
* ECMAScript DOM runtime.
|
|
*
|
|
* It will be appended to the base development runtime code.
|
|
*/ /* eslint-disable @typescript-eslint/no-unused-vars */ /// <reference path="../base/runtime-base.ts" />
|
|
/// <reference path="../base/dev-base.ts" />
|
|
/// <reference path="./runtime-backend-dom.ts" />
|
|
/// <reference path="../../../shared/require-type.d.ts" />
|
|
let DEV_BACKEND;
|
|
(()=>{
|
|
DEV_BACKEND = {
|
|
unloadChunk (chunkUrl) {
|
|
deleteResolver(chunkUrl);
|
|
// TODO(PACK-2140): remove this once all filenames are guaranteed to be escaped.
|
|
const decodedChunkUrl = decodeURI(chunkUrl);
|
|
if (isCss(chunkUrl)) {
|
|
const links = document.querySelectorAll(`link[href="${chunkUrl}"],link[href^="${chunkUrl}?"],link[href="${decodedChunkUrl}"],link[href^="${decodedChunkUrl}?"]`);
|
|
for (const link of Array.from(links)){
|
|
link.remove();
|
|
}
|
|
} else if (isJs(chunkUrl)) {
|
|
// Unloading a JS chunk would have no effect, as it lives in the JS
|
|
// runtime once evaluated.
|
|
// However, we still want to remove the script tag from the DOM to keep
|
|
// the HTML somewhat consistent from the user's perspective.
|
|
const scripts = document.querySelectorAll(`script[src="${chunkUrl}"],script[src^="${chunkUrl}?"],script[src="${decodedChunkUrl}"],script[src^="${decodedChunkUrl}?"]`);
|
|
for (const script of Array.from(scripts)){
|
|
script.remove();
|
|
}
|
|
} else {
|
|
throw new Error(`can't infer type of chunk from URL ${chunkUrl}`);
|
|
}
|
|
},
|
|
reloadChunk (chunkUrl) {
|
|
return new Promise((resolve, reject)=>{
|
|
if (!isCss(chunkUrl)) {
|
|
reject(new Error('The DOM backend can only reload CSS chunks'));
|
|
return;
|
|
}
|
|
const decodedChunkUrl = decodeURI(chunkUrl);
|
|
const previousLinks = document.querySelectorAll(`link[rel=stylesheet][href="${chunkUrl}"],link[rel=stylesheet][href^="${chunkUrl}?"],link[rel=stylesheet][href="${decodedChunkUrl}"],link[rel=stylesheet][href^="${decodedChunkUrl}?"]`);
|
|
if (previousLinks.length === 0) {
|
|
reject(new Error(`No link element found for chunk ${chunkUrl}`));
|
|
return;
|
|
}
|
|
const link = document.createElement('link');
|
|
link.rel = 'stylesheet';
|
|
if (navigator.userAgent.includes('Firefox')) {
|
|
// Firefox won't reload CSS files that were previously loaded on the current page,
|
|
// we need to add a query param to make sure CSS is actually reloaded from the server.
|
|
//
|
|
// I believe this is this issue: https://bugzilla.mozilla.org/show_bug.cgi?id=1037506
|
|
//
|
|
// Safari has a similar issue, but only if you have a `<link rel=preload ... />` tag
|
|
// pointing to the same URL as the stylesheet: https://bugs.webkit.org/show_bug.cgi?id=187726
|
|
link.href = `${chunkUrl}?ts=${Date.now()}`;
|
|
} else {
|
|
link.href = chunkUrl;
|
|
}
|
|
link.onerror = ()=>{
|
|
reject();
|
|
};
|
|
link.onload = ()=>{
|
|
// First load the new CSS, then remove the old ones. This prevents visible
|
|
// flickering that would happen in-between removing the previous CSS and
|
|
// loading the new one.
|
|
for (const previousLink of Array.from(previousLinks))previousLink.remove();
|
|
// CSS chunks do not register themselves, and as such must be marked as
|
|
// loaded instantly.
|
|
resolve();
|
|
};
|
|
// Make sure to insert the new CSS right after the previous one, so that
|
|
// its precedence is higher.
|
|
previousLinks[0].parentElement.insertBefore(link, previousLinks[0].nextSibling);
|
|
});
|
|
},
|
|
restart: ()=>self.location.reload()
|
|
};
|
|
function deleteResolver(chunkUrl) {
|
|
chunkResolvers.delete(chunkUrl);
|
|
}
|
|
})();
|
|
function _eval({ code, url, map }) {
|
|
code += `\n\n//# sourceURL=${encodeURI(location.origin + CHUNK_BASE_PATH + url + CHUNK_SUFFIX)}`;
|
|
if (map) {
|
|
code += `\n//# sourceMappingURL=data:application/json;charset=utf-8;base64,${btoa(// btoa doesn't handle nonlatin characters, so escape them as \x sequences
|
|
// See https://stackoverflow.com/a/26603875
|
|
unescape(encodeURIComponent(map)))}`;
|
|
}
|
|
// eslint-disable-next-line no-eval
|
|
return eval(code);
|
|
}
|
|
const chunksToRegister = globalThis.TURBOPACK;
|
|
globalThis.TURBOPACK = { push: registerChunk };
|
|
chunksToRegister.forEach(registerChunk);
|
|
const chunkListsToRegister = globalThis.TURBOPACK_CHUNK_LISTS || [];
|
|
globalThis.TURBOPACK_CHUNK_LISTS = { push: registerChunkList };
|
|
chunkListsToRegister.forEach(registerChunkList);
|
|
})();
|
|
|
|
|
|
//# sourceMappingURL=_23a915ee._.js.map
|