aboutsummaryrefslogtreecommitdiff
path: root/build/assets/js/693.a33d4660.js
diff options
context:
space:
mode:
author安竹洋平 <61961825+yasutakeyohei@users.noreply.github.com>2024-01-25 00:14:42 +0900
committer安竹洋平 <61961825+yasutakeyohei@users.noreply.github.com>2024-01-25 00:14:42 +0900
commit5970320e4994df7652eedea9e16770e1720342f8 (patch)
treedca76a41f1e16f794c18ad61bcacf3c79ec46a78 /build/assets/js/693.a33d4660.js
parent25ce656bd27f13305ba30e9978625b0117642c08 (diff)
build
Diffstat (limited to 'build/assets/js/693.a33d4660.js')
-rw-r--r--build/assets/js/693.a33d4660.js14570
1 files changed, 0 insertions, 14570 deletions
diff --git a/build/assets/js/693.a33d4660.js b/build/assets/js/693.a33d4660.js
deleted file mode 100644
index 5aaa56e5..00000000
--- a/build/assets/js/693.a33d4660.js
+++ /dev/null
@@ -1,14570 +0,0 @@
-"use strict";
-exports.id = 693;
-exports.ids = [693];
-exports.modules = {
-
-/***/ 69583:
-/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
-
-
-// EXPORTS
-__webpack_require__.d(__webpack_exports__, {
- a: () => (/* binding */ createText),
- c: () => (/* binding */ computeDimensionOfText)
-});
-
-// NAMESPACE OBJECT: ./node_modules/mermaid/node_modules/micromark/lib/constructs.js
-var constructs_namespaceObject = {};
-__webpack_require__.r(constructs_namespaceObject);
-__webpack_require__.d(constructs_namespaceObject, {
- attentionMarkers: () => (attentionMarkers),
- contentInitial: () => (contentInitial),
- disable: () => (disable),
- document: () => (constructs_document),
- flow: () => (constructs_flow),
- flowInitial: () => (flowInitial),
- insideSpan: () => (insideSpan),
- string: () => (constructs_string),
- text: () => (constructs_text)
-});
-
-// EXTERNAL MODULE: ./node_modules/mermaid/dist/mermaid-934d9bea.js + 8 modules
-var mermaid_934d9bea = __webpack_require__(85322);
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/mdast-util-to-string/lib/index.js
-/**
- * @typedef {import('mdast').Root|import('mdast').Content} Node
- *
- * @typedef Options
- * Configuration (optional).
- * @property {boolean | null | undefined} [includeImageAlt=true]
- * Whether to use `alt` for `image`s.
- * @property {boolean | null | undefined} [includeHtml=true]
- * Whether to use `value` of HTML.
- */
-
-/** @type {Options} */
-const emptyOptions = {}
-
-/**
- * Get the text content of a node or list of nodes.
- *
- * Prefers the node’s plain-text fields, otherwise serializes its children,
- * and if the given value is an array, serialize the nodes in it.
- *
- * @param {unknown} value
- * Thing to serialize, typically `Node`.
- * @param {Options | null | undefined} [options]
- * Configuration (optional).
- * @returns {string}
- * Serialized `value`.
- */
-function lib_toString(value, options) {
- const settings = options || emptyOptions
- const includeImageAlt =
- typeof settings.includeImageAlt === 'boolean'
- ? settings.includeImageAlt
- : true
- const includeHtml =
- typeof settings.includeHtml === 'boolean' ? settings.includeHtml : true
-
- return one(value, includeImageAlt, includeHtml)
-}
-
-/**
- * One node or several nodes.
- *
- * @param {unknown} value
- * Thing to serialize.
- * @param {boolean} includeImageAlt
- * Include image `alt`s.
- * @param {boolean} includeHtml
- * Include HTML.
- * @returns {string}
- * Serialized node.
- */
-function one(value, includeImageAlt, includeHtml) {
- if (node(value)) {
- if ('value' in value) {
- return value.type === 'html' && !includeHtml ? '' : value.value
- }
-
- if (includeImageAlt && 'alt' in value && value.alt) {
- return value.alt
- }
-
- if ('children' in value) {
- return lib_all(value.children, includeImageAlt, includeHtml)
- }
- }
-
- if (Array.isArray(value)) {
- return lib_all(value, includeImageAlt, includeHtml)
- }
-
- return ''
-}
-
-/**
- * Serialize a list of nodes.
- *
- * @param {Array<unknown>} values
- * Thing to serialize.
- * @param {boolean} includeImageAlt
- * Include image `alt`s.
- * @param {boolean} includeHtml
- * Include HTML.
- * @returns {string}
- * Serialized nodes.
- */
-function lib_all(values, includeImageAlt, includeHtml) {
- /** @type {Array<string>} */
- const result = []
- let index = -1
-
- while (++index < values.length) {
- result[index] = one(values[index], includeImageAlt, includeHtml)
- }
-
- return result.join('')
-}
-
-/**
- * Check if `value` looks like a node.
- *
- * @param {unknown} value
- * Thing.
- * @returns {value is Node}
- * Whether `value` is a node.
- */
-function node(value) {
- return Boolean(value && typeof value === 'object')
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-chunked/index.js
-/**
- * Like `Array#splice`, but smarter for giant arrays.
- *
- * `Array#splice` takes all items to be inserted as individual argument which
- * causes a stack overflow in V8 when trying to insert 100k items for instance.
- *
- * Otherwise, this does not return the removed items, and takes `items` as an
- * array instead of rest parameters.
- *
- * @template {unknown} T
- * Item type.
- * @param {Array<T>} list
- * List to operate on.
- * @param {number} start
- * Index to remove/insert at (can be negative).
- * @param {number} remove
- * Number of items to remove.
- * @param {Array<T>} items
- * Items to inject into `list`.
- * @returns {void}
- * Nothing.
- */
-function splice(list, start, remove, items) {
- const end = list.length
- let chunkStart = 0
- /** @type {Array<unknown>} */
- let parameters
-
- // Make start between zero and `end` (included).
- if (start < 0) {
- start = -start > end ? 0 : end + start
- } else {
- start = start > end ? end : start
- }
- remove = remove > 0 ? remove : 0
-
- // No need to chunk the items if there’s only a couple (10k) items.
- if (items.length < 10000) {
- parameters = Array.from(items)
- parameters.unshift(start, remove)
- // @ts-expect-error Hush, it’s fine.
- list.splice(...parameters)
- } else {
- // Delete `remove` items starting from `start`
- if (remove) list.splice(start, remove)
-
- // Insert the items in chunks to not cause stack overflows.
- while (chunkStart < items.length) {
- parameters = items.slice(chunkStart, chunkStart + 10000)
- parameters.unshift(start, 0)
- // @ts-expect-error Hush, it’s fine.
- list.splice(...parameters)
- chunkStart += 10000
- start += 10000
- }
- }
-}
-
-/**
- * Append `items` (an array) at the end of `list` (another array).
- * When `list` was empty, returns `items` instead.
- *
- * This prevents a potentially expensive operation when `list` is empty,
- * and adds items in batches to prevent V8 from hanging.
- *
- * @template {unknown} T
- * Item type.
- * @param {Array<T>} list
- * List to operate on.
- * @param {Array<T>} items
- * Items to add to `list`.
- * @returns {Array<T>}
- * Either `list` or `items`.
- */
-function push(list, items) {
- if (list.length > 0) {
- splice(list, list.length, 0, items)
- return list
- }
- return items
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-combine-extensions/index.js
-/**
- * @typedef {import('micromark-util-types').Extension} Extension
- * @typedef {import('micromark-util-types').Handles} Handles
- * @typedef {import('micromark-util-types').HtmlExtension} HtmlExtension
- * @typedef {import('micromark-util-types').NormalizedExtension} NormalizedExtension
- */
-
-
-
-const micromark_util_combine_extensions_hasOwnProperty = {}.hasOwnProperty
-
-/**
- * Combine multiple syntax extensions into one.
- *
- * @param {Array<Extension>} extensions
- * List of syntax extensions.
- * @returns {NormalizedExtension}
- * A single combined extension.
- */
-function combineExtensions(extensions) {
- /** @type {NormalizedExtension} */
- const all = {}
- let index = -1
-
- while (++index < extensions.length) {
- syntaxExtension(all, extensions[index])
- }
-
- return all
-}
-
-/**
- * Merge `extension` into `all`.
- *
- * @param {NormalizedExtension} all
- * Extension to merge into.
- * @param {Extension} extension
- * Extension to merge.
- * @returns {void}
- */
-function syntaxExtension(all, extension) {
- /** @type {keyof Extension} */
- let hook
-
- for (hook in extension) {
- const maybe = micromark_util_combine_extensions_hasOwnProperty.call(all, hook) ? all[hook] : undefined
- /** @type {Record<string, unknown>} */
- const left = maybe || (all[hook] = {})
- /** @type {Record<string, unknown> | undefined} */
- const right = extension[hook]
- /** @type {string} */
- let code
-
- if (right) {
- for (code in right) {
- if (!micromark_util_combine_extensions_hasOwnProperty.call(left, code)) left[code] = []
- const value = right[code]
- constructs(
- // @ts-expect-error Looks like a list.
- left[code],
- Array.isArray(value) ? value : value ? [value] : []
- )
- }
- }
- }
-}
-
-/**
- * Merge `list` into `existing` (both lists of constructs).
- * Mutates `existing`.
- *
- * @param {Array<unknown>} existing
- * @param {Array<unknown>} list
- * @returns {void}
- */
-function constructs(existing, list) {
- let index = -1
- /** @type {Array<unknown>} */
- const before = []
-
- while (++index < list.length) {
- // @ts-expect-error Looks like an object.
- ;(list[index].add === 'after' ? existing : before).push(list[index])
- }
-
- splice(existing, 0, 0, before)
-}
-
-/**
- * Combine multiple HTML extensions into one.
- *
- * @param {Array<HtmlExtension>} htmlExtensions
- * List of HTML extensions.
- * @returns {HtmlExtension}
- * A single combined HTML extension.
- */
-function combineHtmlExtensions(htmlExtensions) {
- /** @type {HtmlExtension} */
- const handlers = {}
- let index = -1
-
- while (++index < htmlExtensions.length) {
- htmlExtension(handlers, htmlExtensions[index])
- }
-
- return handlers
-}
-
-/**
- * Merge `extension` into `all`.
- *
- * @param {HtmlExtension} all
- * Extension to merge into.
- * @param {HtmlExtension} extension
- * Extension to merge.
- * @returns {void}
- */
-function htmlExtension(all, extension) {
- /** @type {keyof HtmlExtension} */
- let hook
-
- for (hook in extension) {
- const maybe = micromark_util_combine_extensions_hasOwnProperty.call(all, hook) ? all[hook] : undefined
- const left = maybe || (all[hook] = {})
- const right = extension[hook]
- /** @type {keyof Handles} */
- let type
-
- if (right) {
- for (type in right) {
- // @ts-expect-error assume document vs regular handler are managed correctly.
- left[type] = right[type]
- }
- }
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/micromark-util-character/lib/unicode-punctuation-regex.js
-// This module is generated by `script/`.
-//
-// CommonMark handles attention (emphasis, strong) markers based on what comes
-// before or after them.
-// One such difference is if those characters are Unicode punctuation.
-// This script is generated from the Unicode data.
-
-/**
- * Regular expression that matches a unicode punctuation character.
- */
-const unicodePunctuationRegex =
- /[!-\/:-@\[-`\{-~\xA1\xA7\xAB\xB6\xB7\xBB\xBF\u037E\u0387\u055A-\u055F\u0589\u058A\u05BE\u05C0\u05C3\u05C6\u05F3\u05F4\u0609\u060A\u060C\u060D\u061B\u061D-\u061F\u066A-\u066D\u06D4\u0700-\u070D\u07F7-\u07F9\u0830-\u083E\u085E\u0964\u0965\u0970\u09FD\u0A76\u0AF0\u0C77\u0C84\u0DF4\u0E4F\u0E5A\u0E5B\u0F04-\u0F12\u0F14\u0F3A-\u0F3D\u0F85\u0FD0-\u0FD4\u0FD9\u0FDA\u104A-\u104F\u10FB\u1360-\u1368\u1400\u166E\u169B\u169C\u16EB-\u16ED\u1735\u1736\u17D4-\u17D6\u17D8-\u17DA\u1800-\u180A\u1944\u1945\u1A1E\u1A1F\u1AA0-\u1AA6\u1AA8-\u1AAD\u1B5A-\u1B60\u1B7D\u1B7E\u1BFC-\u1BFF\u1C3B-\u1C3F\u1C7E\u1C7F\u1CC0-\u1CC7\u1CD3\u2010-\u2027\u2030-\u2043\u2045-\u2051\u2053-\u205E\u207D\u207E\u208D\u208E\u2308-\u230B\u2329\u232A\u2768-\u2775\u27C5\u27C6\u27E6-\u27EF\u2983-\u2998\u29D8-\u29DB\u29FC\u29FD\u2CF9-\u2CFC\u2CFE\u2CFF\u2D70\u2E00-\u2E2E\u2E30-\u2E4F\u2E52-\u2E5D\u3001-\u3003\u3008-\u3011\u3014-\u301F\u3030\u303D\u30A0\u30FB\uA4FE\uA4FF\uA60D-\uA60F\uA673\uA67E\uA6F2-\uA6F7\uA874-\uA877\uA8CE\uA8CF\uA8F8-\uA8FA\uA8FC\uA92E\uA92F\uA95F\uA9C1-\uA9CD\uA9DE\uA9DF\uAA5C-\uAA5F\uAADE\uAADF\uAAF0\uAAF1\uABEB\uFD3E\uFD3F\uFE10-\uFE19\uFE30-\uFE52\uFE54-\uFE61\uFE63\uFE68\uFE6A\uFE6B\uFF01-\uFF03\uFF05-\uFF0A\uFF0C-\uFF0F\uFF1A\uFF1B\uFF1F\uFF20\uFF3B-\uFF3D\uFF3F\uFF5B\uFF5D\uFF5F-\uFF65]/
-
-;// CONCATENATED MODULE: ./node_modules/micromark-util-character/index.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- */
-
-
-
-/**
- * Check whether the character code represents an ASCII alpha (`a` through `z`,
- * case insensitive).
- *
- * An **ASCII alpha** is an ASCII upper alpha or ASCII lower alpha.
- *
- * An **ASCII upper alpha** is a character in the inclusive range U+0041 (`A`)
- * to U+005A (`Z`).
- *
- * An **ASCII lower alpha** is a character in the inclusive range U+0061 (`a`)
- * to U+007A (`z`).
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiAlpha = regexCheck(/[A-Za-z]/)
-
-/**
- * Check whether the character code represents an ASCII alphanumeric (`a`
- * through `z`, case insensitive, or `0` through `9`).
- *
- * An **ASCII alphanumeric** is an ASCII digit (see `asciiDigit`) or ASCII alpha
- * (see `asciiAlpha`).
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiAlphanumeric = regexCheck(/[\dA-Za-z]/)
-
-/**
- * Check whether the character code represents an ASCII atext.
- *
- * atext is an ASCII alphanumeric (see `asciiAlphanumeric`), or a character in
- * the inclusive ranges U+0023 NUMBER SIGN (`#`) to U+0027 APOSTROPHE (`'`),
- * U+002A ASTERISK (`*`), U+002B PLUS SIGN (`+`), U+002D DASH (`-`), U+002F
- * SLASH (`/`), U+003D EQUALS TO (`=`), U+003F QUESTION MARK (`?`), U+005E
- * CARET (`^`) to U+0060 GRAVE ACCENT (`` ` ``), or U+007B LEFT CURLY BRACE
- * (`{`) to U+007E TILDE (`~`).
- *
- * See:
- * **\[RFC5322]**:
- * [Internet Message Format](https://tools.ietf.org/html/rfc5322).
- * P. Resnick.
- * IETF.
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiAtext = regexCheck(/[#-'*+\--9=?A-Z^-~]/)
-
-/**
- * Check whether a character code is an ASCII control character.
- *
- * An **ASCII control** is a character in the inclusive range U+0000 NULL (NUL)
- * to U+001F (US), or U+007F (DEL).
- *
- * @param {Code} code
- * Code.
- * @returns {boolean}
- * Whether it matches.
- */
-function asciiControl(code) {
- return (
- // Special whitespace codes (which have negative values), C0 and Control
- // character DEL
- code !== null && (code < 32 || code === 127)
- )
-}
-
-/**
- * Check whether the character code represents an ASCII digit (`0` through `9`).
- *
- * An **ASCII digit** is a character in the inclusive range U+0030 (`0`) to
- * U+0039 (`9`).
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiDigit = regexCheck(/\d/)
-
-/**
- * Check whether the character code represents an ASCII hex digit (`a` through
- * `f`, case insensitive, or `0` through `9`).
- *
- * An **ASCII hex digit** is an ASCII digit (see `asciiDigit`), ASCII upper hex
- * digit, or an ASCII lower hex digit.
- *
- * An **ASCII upper hex digit** is a character in the inclusive range U+0041
- * (`A`) to U+0046 (`F`).
- *
- * An **ASCII lower hex digit** is a character in the inclusive range U+0061
- * (`a`) to U+0066 (`f`).
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiHexDigit = regexCheck(/[\dA-Fa-f]/)
-
-/**
- * Check whether the character code represents ASCII punctuation.
- *
- * An **ASCII punctuation** is a character in the inclusive ranges U+0021
- * EXCLAMATION MARK (`!`) to U+002F SLASH (`/`), U+003A COLON (`:`) to U+0040 AT
- * SIGN (`@`), U+005B LEFT SQUARE BRACKET (`[`) to U+0060 GRAVE ACCENT
- * (`` ` ``), or U+007B LEFT CURLY BRACE (`{`) to U+007E TILDE (`~`).
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const asciiPunctuation = regexCheck(/[!-/:-@[-`{-~]/)
-
-/**
- * Check whether a character code is a markdown line ending.
- *
- * A **markdown line ending** is the virtual characters M-0003 CARRIAGE RETURN
- * LINE FEED (CRLF), M-0004 LINE FEED (LF) and M-0005 CARRIAGE RETURN (CR).
- *
- * In micromark, the actual character U+000A LINE FEED (LF) and U+000D CARRIAGE
- * RETURN (CR) are replaced by these virtual characters depending on whether
- * they occurred together.
- *
- * @param {Code} code
- * Code.
- * @returns {boolean}
- * Whether it matches.
- */
-function markdownLineEnding(code) {
- return code !== null && code < -2
-}
-
-/**
- * Check whether a character code is a markdown line ending (see
- * `markdownLineEnding`) or markdown space (see `markdownSpace`).
- *
- * @param {Code} code
- * Code.
- * @returns {boolean}
- * Whether it matches.
- */
-function markdownLineEndingOrSpace(code) {
- return code !== null && (code < 0 || code === 32)
-}
-
-/**
- * Check whether a character code is a markdown space.
- *
- * A **markdown space** is the concrete character U+0020 SPACE (SP) and the
- * virtual characters M-0001 VIRTUAL SPACE (VS) and M-0002 HORIZONTAL TAB (HT).
- *
- * In micromark, the actual character U+0009 CHARACTER TABULATION (HT) is
- * replaced by one M-0002 HORIZONTAL TAB (HT) and between 0 and 3 M-0001 VIRTUAL
- * SPACE (VS) characters, depending on the column at which the tab occurred.
- *
- * @param {Code} code
- * Code.
- * @returns {boolean}
- * Whether it matches.
- */
-function markdownSpace(code) {
- return code === -2 || code === -1 || code === 32
-}
-
-// Size note: removing ASCII from the regex and using `asciiPunctuation` here
-// In fact adds to the bundle size.
-/**
- * Check whether the character code represents Unicode punctuation.
- *
- * A **Unicode punctuation** is a character in the Unicode `Pc` (Punctuation,
- * Connector), `Pd` (Punctuation, Dash), `Pe` (Punctuation, Close), `Pf`
- * (Punctuation, Final quote), `Pi` (Punctuation, Initial quote), `Po`
- * (Punctuation, Other), or `Ps` (Punctuation, Open) categories, or an ASCII
- * punctuation (see `asciiPunctuation`).
- *
- * See:
- * **\[UNICODE]**:
- * [The Unicode Standard](https://www.unicode.org/versions/).
- * Unicode Consortium.
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const unicodePunctuation = regexCheck(unicodePunctuationRegex)
-
-/**
- * Check whether the character code represents Unicode whitespace.
- *
- * Note that this does handle micromark specific markdown whitespace characters.
- * See `markdownLineEndingOrSpace` to check that.
- *
- * A **Unicode whitespace** is a character in the Unicode `Zs` (Separator,
- * Space) category, or U+0009 CHARACTER TABULATION (HT), U+000A LINE FEED (LF),
- * U+000C (FF), or U+000D CARRIAGE RETURN (CR) (**\[UNICODE]**).
- *
- * See:
- * **\[UNICODE]**:
- * [The Unicode Standard](https://www.unicode.org/versions/).
- * Unicode Consortium.
- *
- * @param code
- * Code.
- * @returns
- * Whether it matches.
- */
-const unicodeWhitespace = regexCheck(/\s/)
-
-/**
- * Create a code check from a regex.
- *
- * @param {RegExp} regex
- * @returns {(code: Code) => boolean}
- */
-function regexCheck(regex) {
- return check
-
- /**
- * Check whether a code matches the bound regex.
- *
- * @param {Code} code
- * Character code.
- * @returns {boolean}
- * Whether the character code matches the bound regex.
- */
- function check(code) {
- return code !== null && regex.test(String.fromCharCode(code))
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/micromark-factory-space/index.js
-/**
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenType} TokenType
- */
-
-
-
-// To do: implement `spaceOrTab`, `spaceOrTabMinMax`, `spaceOrTabWithOptions`.
-
-/**
- * Parse spaces and tabs.
- *
- * There is no `nok` parameter:
- *
- * * spaces in markdown are often optional, in which case this factory can be
- * used and `ok` will be switched to whether spaces were found or not
- * * one line ending or space can be detected with `markdownSpace(code)` right
- * before using `factorySpace`
- *
- * ###### Examples
- *
- * Where `␉` represents a tab (plus how much it expands) and `␠` represents a
- * single space.
- *
- * ```markdown
- * ␉
- * ␠␠␠␠
- * ␉␠
- * ```
- *
- * @param {Effects} effects
- * Context.
- * @param {State} ok
- * State switched to when successful.
- * @param {TokenType} type
- * Type (`' \t'`).
- * @param {number | undefined} [max=Infinity]
- * Max (exclusive).
- * @returns
- * Start state.
- */
-function factorySpace(effects, ok, type, max) {
- const limit = max ? max - 1 : Number.POSITIVE_INFINITY
- let size = 0
- return start
-
- /** @type {State} */
- function start(code) {
- if (markdownSpace(code)) {
- effects.enter(type)
- return prefix(code)
- }
- return ok(code)
- }
-
- /** @type {State} */
- function prefix(code) {
- if (markdownSpace(code) && size++ < limit) {
- effects.consume(code)
- return prefix
- }
- effects.exit(type)
- return ok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/initialize/content.js
-/**
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').Initializer} Initializer
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- */
-
-
-
-/** @type {InitialConstruct} */
-const content = {
- tokenize: initializeContent
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Initializer}
- */
-function initializeContent(effects) {
- const contentStart = effects.attempt(
- this.parser.constructs.contentInitial,
- afterContentStartConstruct,
- paragraphInitial
- )
- /** @type {Token} */
- let previous
- return contentStart
-
- /** @type {State} */
- function afterContentStartConstruct(code) {
- if (code === null) {
- effects.consume(code)
- return
- }
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return factorySpace(effects, contentStart, 'linePrefix')
- }
-
- /** @type {State} */
- function paragraphInitial(code) {
- effects.enter('paragraph')
- return lineStart(code)
- }
-
- /** @type {State} */
- function lineStart(code) {
- const token = effects.enter('chunkText', {
- contentType: 'text',
- previous
- })
- if (previous) {
- previous.next = token
- }
- previous = token
- return data(code)
- }
-
- /** @type {State} */
- function data(code) {
- if (code === null) {
- effects.exit('chunkText')
- effects.exit('paragraph')
- effects.consume(code)
- return
- }
- if (markdownLineEnding(code)) {
- effects.consume(code)
- effects.exit('chunkText')
- return lineStart
- }
-
- // Data.
- effects.consume(code)
- return data
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/initialize/document.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').ContainerState} ContainerState
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').Initializer} Initializer
- * @typedef {import('micromark-util-types').Point} Point
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-/**
- * @typedef {[Construct, ContainerState]} StackItem
- */
-
-
-
-
-/** @type {InitialConstruct} */
-const document_document = {
- tokenize: initializeDocument
-}
-
-/** @type {Construct} */
-const containerConstruct = {
- tokenize: tokenizeContainer
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Initializer}
- */
-function initializeDocument(effects) {
- const self = this
- /** @type {Array<StackItem>} */
- const stack = []
- let continued = 0
- /** @type {TokenizeContext | undefined} */
- let childFlow
- /** @type {Token | undefined} */
- let childToken
- /** @type {number} */
- let lineStartOffset
- return start
-
- /** @type {State} */
- function start(code) {
- // First we iterate through the open blocks, starting with the root
- // document, and descending through last children down to the last open
- // block.
- // Each block imposes a condition that the line must satisfy if the block is
- // to remain open.
- // For example, a block quote requires a `>` character.
- // A paragraph requires a non-blank line.
- // In this phase we may match all or just some of the open blocks.
- // But we cannot close unmatched blocks yet, because we may have a lazy
- // continuation line.
- if (continued < stack.length) {
- const item = stack[continued]
- self.containerState = item[1]
- return effects.attempt(
- item[0].continuation,
- documentContinue,
- checkNewContainers
- )(code)
- }
-
- // Done.
- return checkNewContainers(code)
- }
-
- /** @type {State} */
- function documentContinue(code) {
- continued++
-
- // Note: this field is called `_closeFlow` but it also closes containers.
- // Perhaps a good idea to rename it but it’s already used in the wild by
- // extensions.
- if (self.containerState._closeFlow) {
- self.containerState._closeFlow = undefined
- if (childFlow) {
- closeFlow()
- }
-
- // Note: this algorithm for moving events around is similar to the
- // algorithm when dealing with lazy lines in `writeToChild`.
- const indexBeforeExits = self.events.length
- let indexBeforeFlow = indexBeforeExits
- /** @type {Point | undefined} */
- let point
-
- // Find the flow chunk.
- while (indexBeforeFlow--) {
- if (
- self.events[indexBeforeFlow][0] === 'exit' &&
- self.events[indexBeforeFlow][1].type === 'chunkFlow'
- ) {
- point = self.events[indexBeforeFlow][1].end
- break
- }
- }
- exitContainers(continued)
-
- // Fix positions.
- let index = indexBeforeExits
- while (index < self.events.length) {
- self.events[index][1].end = Object.assign({}, point)
- index++
- }
-
- // Inject the exits earlier (they’re still also at the end).
- splice(
- self.events,
- indexBeforeFlow + 1,
- 0,
- self.events.slice(indexBeforeExits)
- )
-
- // Discard the duplicate exits.
- self.events.length = index
- return checkNewContainers(code)
- }
- return start(code)
- }
-
- /** @type {State} */
- function checkNewContainers(code) {
- // Next, after consuming the continuation markers for existing blocks, we
- // look for new block starts (e.g. `>` for a block quote).
- // If we encounter a new block start, we close any blocks unmatched in
- // step 1 before creating the new block as a child of the last matched
- // block.
- if (continued === stack.length) {
- // No need to `check` whether there’s a container, of `exitContainers`
- // would be moot.
- // We can instead immediately `attempt` to parse one.
- if (!childFlow) {
- return documentContinued(code)
- }
-
- // If we have concrete content, such as block HTML or fenced code,
- // we can’t have containers “pierce” into them, so we can immediately
- // start.
- if (childFlow.currentConstruct && childFlow.currentConstruct.concrete) {
- return flowStart(code)
- }
-
- // If we do have flow, it could still be a blank line,
- // but we’d be interrupting it w/ a new container if there’s a current
- // construct.
- // To do: next major: remove `_gfmTableDynamicInterruptHack` (no longer
- // needed in micromark-extension-gfm-table@1.0.6).
- self.interrupt = Boolean(
- childFlow.currentConstruct && !childFlow._gfmTableDynamicInterruptHack
- )
- }
-
- // Check if there is a new container.
- self.containerState = {}
- return effects.check(
- containerConstruct,
- thereIsANewContainer,
- thereIsNoNewContainer
- )(code)
- }
-
- /** @type {State} */
- function thereIsANewContainer(code) {
- if (childFlow) closeFlow()
- exitContainers(continued)
- return documentContinued(code)
- }
-
- /** @type {State} */
- function thereIsNoNewContainer(code) {
- self.parser.lazy[self.now().line] = continued !== stack.length
- lineStartOffset = self.now().offset
- return flowStart(code)
- }
-
- /** @type {State} */
- function documentContinued(code) {
- // Try new containers.
- self.containerState = {}
- return effects.attempt(
- containerConstruct,
- containerContinue,
- flowStart
- )(code)
- }
-
- /** @type {State} */
- function containerContinue(code) {
- continued++
- stack.push([self.currentConstruct, self.containerState])
- // Try another.
- return documentContinued(code)
- }
-
- /** @type {State} */
- function flowStart(code) {
- if (code === null) {
- if (childFlow) closeFlow()
- exitContainers(0)
- effects.consume(code)
- return
- }
- childFlow = childFlow || self.parser.flow(self.now())
- effects.enter('chunkFlow', {
- contentType: 'flow',
- previous: childToken,
- _tokenizer: childFlow
- })
- return flowContinue(code)
- }
-
- /** @type {State} */
- function flowContinue(code) {
- if (code === null) {
- writeToChild(effects.exit('chunkFlow'), true)
- exitContainers(0)
- effects.consume(code)
- return
- }
- if (markdownLineEnding(code)) {
- effects.consume(code)
- writeToChild(effects.exit('chunkFlow'))
- // Get ready for the next line.
- continued = 0
- self.interrupt = undefined
- return start
- }
- effects.consume(code)
- return flowContinue
- }
-
- /**
- * @param {Token} token
- * @param {boolean | undefined} [eof]
- * @returns {void}
- */
- function writeToChild(token, eof) {
- const stream = self.sliceStream(token)
- if (eof) stream.push(null)
- token.previous = childToken
- if (childToken) childToken.next = token
- childToken = token
- childFlow.defineSkip(token.start)
- childFlow.write(stream)
-
- // Alright, so we just added a lazy line:
- //
- // ```markdown
- // > a
- // b.
- //
- // Or:
- //
- // > ~~~c
- // d
- //
- // Or:
- //
- // > | e |
- // f
- // ```
- //
- // The construct in the second example (fenced code) does not accept lazy
- // lines, so it marked itself as done at the end of its first line, and
- // then the content construct parses `d`.
- // Most constructs in markdown match on the first line: if the first line
- // forms a construct, a non-lazy line can’t “unmake” it.
- //
- // The construct in the third example is potentially a GFM table, and
- // those are *weird*.
- // It *could* be a table, from the first line, if the following line
- // matches a condition.
- // In this case, that second line is lazy, which “unmakes” the first line
- // and turns the whole into one content block.
- //
- // We’ve now parsed the non-lazy and the lazy line, and can figure out
- // whether the lazy line started a new flow block.
- // If it did, we exit the current containers between the two flow blocks.
- if (self.parser.lazy[token.start.line]) {
- let index = childFlow.events.length
- while (index--) {
- if (
- // The token starts before the line ending…
- childFlow.events[index][1].start.offset < lineStartOffset &&
- // …and either is not ended yet…
- (!childFlow.events[index][1].end ||
- // …or ends after it.
- childFlow.events[index][1].end.offset > lineStartOffset)
- ) {
- // Exit: there’s still something open, which means it’s a lazy line
- // part of something.
- return
- }
- }
-
- // Note: this algorithm for moving events around is similar to the
- // algorithm when closing flow in `documentContinue`.
- const indexBeforeExits = self.events.length
- let indexBeforeFlow = indexBeforeExits
- /** @type {boolean | undefined} */
- let seen
- /** @type {Point | undefined} */
- let point
-
- // Find the previous chunk (the one before the lazy line).
- while (indexBeforeFlow--) {
- if (
- self.events[indexBeforeFlow][0] === 'exit' &&
- self.events[indexBeforeFlow][1].type === 'chunkFlow'
- ) {
- if (seen) {
- point = self.events[indexBeforeFlow][1].end
- break
- }
- seen = true
- }
- }
- exitContainers(continued)
-
- // Fix positions.
- index = indexBeforeExits
- while (index < self.events.length) {
- self.events[index][1].end = Object.assign({}, point)
- index++
- }
-
- // Inject the exits earlier (they’re still also at the end).
- splice(
- self.events,
- indexBeforeFlow + 1,
- 0,
- self.events.slice(indexBeforeExits)
- )
-
- // Discard the duplicate exits.
- self.events.length = index
- }
- }
-
- /**
- * @param {number} size
- * @returns {void}
- */
- function exitContainers(size) {
- let index = stack.length
-
- // Exit open containers.
- while (index-- > size) {
- const entry = stack[index]
- self.containerState = entry[1]
- entry[0].exit.call(self, effects)
- }
- stack.length = size
- }
- function closeFlow() {
- childFlow.write([null])
- childToken = undefined
- childFlow = undefined
- self.containerState._closeFlow = undefined
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeContainer(effects, ok, nok) {
- // Always populated by defaults.
-
- return factorySpace(
- effects,
- effects.attempt(this.parser.constructs.document, ok, nok),
- 'linePrefix',
- this.parser.constructs.disable.null.includes('codeIndented') ? undefined : 4
- )
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/blank-line.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const blankLine = {
- tokenize: tokenizeBlankLine,
- partial: true
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeBlankLine(effects, ok, nok) {
- return start
-
- /**
- * Start of blank line.
- *
- * > 👉 **Note**: `␠` represents a space character.
- *
- * ```markdown
- * > | ␠␠␊
- * ^
- * > | ␊
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- return markdownSpace(code)
- ? factorySpace(effects, after, 'linePrefix')(code)
- : after(code)
- }
-
- /**
- * At eof/eol, after optional whitespace.
- *
- * > 👉 **Note**: `␠` represents a space character.
- *
- * ```markdown
- * > | ␠␠␊
- * ^
- * > | ␊
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- return code === null || markdownLineEnding(code) ? ok(code) : nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-subtokenize/index.js
-/**
- * @typedef {import('micromark-util-types').Chunk} Chunk
- * @typedef {import('micromark-util-types').Event} Event
- * @typedef {import('micromark-util-types').Token} Token
- */
-
-
-/**
- * Tokenize subcontent.
- *
- * @param {Array<Event>} events
- * List of events.
- * @returns {boolean}
- * Whether subtokens were found.
- */
-function subtokenize(events) {
- /** @type {Record<string, number>} */
- const jumps = {}
- let index = -1
- /** @type {Event} */
- let event
- /** @type {number | undefined} */
- let lineIndex
- /** @type {number} */
- let otherIndex
- /** @type {Event} */
- let otherEvent
- /** @type {Array<Event>} */
- let parameters
- /** @type {Array<Event>} */
- let subevents
- /** @type {boolean | undefined} */
- let more
- while (++index < events.length) {
- while (index in jumps) {
- index = jumps[index]
- }
- event = events[index]
-
- // Add a hook for the GFM tasklist extension, which needs to know if text
- // is in the first content of a list item.
- if (
- index &&
- event[1].type === 'chunkFlow' &&
- events[index - 1][1].type === 'listItemPrefix'
- ) {
- subevents = event[1]._tokenizer.events
- otherIndex = 0
- if (
- otherIndex < subevents.length &&
- subevents[otherIndex][1].type === 'lineEndingBlank'
- ) {
- otherIndex += 2
- }
- if (
- otherIndex < subevents.length &&
- subevents[otherIndex][1].type === 'content'
- ) {
- while (++otherIndex < subevents.length) {
- if (subevents[otherIndex][1].type === 'content') {
- break
- }
- if (subevents[otherIndex][1].type === 'chunkText') {
- subevents[otherIndex][1]._isInFirstContentOfListItem = true
- otherIndex++
- }
- }
- }
- }
-
- // Enter.
- if (event[0] === 'enter') {
- if (event[1].contentType) {
- Object.assign(jumps, subcontent(events, index))
- index = jumps[index]
- more = true
- }
- }
- // Exit.
- else if (event[1]._container) {
- otherIndex = index
- lineIndex = undefined
- while (otherIndex--) {
- otherEvent = events[otherIndex]
- if (
- otherEvent[1].type === 'lineEnding' ||
- otherEvent[1].type === 'lineEndingBlank'
- ) {
- if (otherEvent[0] === 'enter') {
- if (lineIndex) {
- events[lineIndex][1].type = 'lineEndingBlank'
- }
- otherEvent[1].type = 'lineEnding'
- lineIndex = otherIndex
- }
- } else {
- break
- }
- }
- if (lineIndex) {
- // Fix position.
- event[1].end = Object.assign({}, events[lineIndex][1].start)
-
- // Switch container exit w/ line endings.
- parameters = events.slice(lineIndex, index)
- parameters.unshift(event)
- splice(events, lineIndex, index - lineIndex + 1, parameters)
- }
- }
- }
- return !more
-}
-
-/**
- * Tokenize embedded tokens.
- *
- * @param {Array<Event>} events
- * @param {number} eventIndex
- * @returns {Record<string, number>}
- */
-function subcontent(events, eventIndex) {
- const token = events[eventIndex][1]
- const context = events[eventIndex][2]
- let startPosition = eventIndex - 1
- /** @type {Array<number>} */
- const startPositions = []
- const tokenizer =
- token._tokenizer || context.parser[token.contentType](token.start)
- const childEvents = tokenizer.events
- /** @type {Array<[number, number]>} */
- const jumps = []
- /** @type {Record<string, number>} */
- const gaps = {}
- /** @type {Array<Chunk>} */
- let stream
- /** @type {Token | undefined} */
- let previous
- let index = -1
- /** @type {Token | undefined} */
- let current = token
- let adjust = 0
- let start = 0
- const breaks = [start]
-
- // Loop forward through the linked tokens to pass them in order to the
- // subtokenizer.
- while (current) {
- // Find the position of the event for this token.
- while (events[++startPosition][1] !== current) {
- // Empty.
- }
- startPositions.push(startPosition)
- if (!current._tokenizer) {
- stream = context.sliceStream(current)
- if (!current.next) {
- stream.push(null)
- }
- if (previous) {
- tokenizer.defineSkip(current.start)
- }
- if (current._isInFirstContentOfListItem) {
- tokenizer._gfmTasklistFirstContentOfListItem = true
- }
- tokenizer.write(stream)
- if (current._isInFirstContentOfListItem) {
- tokenizer._gfmTasklistFirstContentOfListItem = undefined
- }
- }
-
- // Unravel the next token.
- previous = current
- current = current.next
- }
-
- // Now, loop back through all events (and linked tokens), to figure out which
- // parts belong where.
- current = token
- while (++index < childEvents.length) {
- if (
- // Find a void token that includes a break.
- childEvents[index][0] === 'exit' &&
- childEvents[index - 1][0] === 'enter' &&
- childEvents[index][1].type === childEvents[index - 1][1].type &&
- childEvents[index][1].start.line !== childEvents[index][1].end.line
- ) {
- start = index + 1
- breaks.push(start)
- // Help GC.
- current._tokenizer = undefined
- current.previous = undefined
- current = current.next
- }
- }
-
- // Help GC.
- tokenizer.events = []
-
- // If there’s one more token (which is the cases for lines that end in an
- // EOF), that’s perfect: the last point we found starts it.
- // If there isn’t then make sure any remaining content is added to it.
- if (current) {
- // Help GC.
- current._tokenizer = undefined
- current.previous = undefined
- } else {
- breaks.pop()
- }
-
- // Now splice the events from the subtokenizer into the current events,
- // moving back to front so that splice indices aren’t affected.
- index = breaks.length
- while (index--) {
- const slice = childEvents.slice(breaks[index], breaks[index + 1])
- const start = startPositions.pop()
- jumps.unshift([start, start + slice.length - 1])
- splice(events, start, 2, slice)
- }
- index = -1
- while (++index < jumps.length) {
- gaps[adjust + jumps[index][0]] = adjust + jumps[index][1]
- adjust += jumps[index][1] - jumps[index][0] - 1
- }
- return gaps
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/content.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-/**
- * No name because it must not be turned off.
- * @type {Construct}
- */
-const content_content = {
- tokenize: tokenizeContent,
- resolve: resolveContent
-}
-
-/** @type {Construct} */
-const continuationConstruct = {
- tokenize: tokenizeContinuation,
- partial: true
-}
-
-/**
- * Content is transparent: it’s parsed right now. That way, definitions are also
- * parsed right now: before text in paragraphs (specifically, media) are parsed.
- *
- * @type {Resolver}
- */
-function resolveContent(events) {
- subtokenize(events)
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeContent(effects, ok) {
- /** @type {Token | undefined} */
- let previous
- return chunkStart
-
- /**
- * Before a content chunk.
- *
- * ```markdown
- * > | abc
- * ^
- * ```
- *
- * @type {State}
- */
- function chunkStart(code) {
- effects.enter('content')
- previous = effects.enter('chunkContent', {
- contentType: 'content'
- })
- return chunkInside(code)
- }
-
- /**
- * In a content chunk.
- *
- * ```markdown
- * > | abc
- * ^^^
- * ```
- *
- * @type {State}
- */
- function chunkInside(code) {
- if (code === null) {
- return contentEnd(code)
- }
-
- // To do: in `markdown-rs`, each line is parsed on its own, and everything
- // is stitched together resolving.
- if (markdownLineEnding(code)) {
- return effects.check(
- continuationConstruct,
- contentContinue,
- contentEnd
- )(code)
- }
-
- // Data.
- effects.consume(code)
- return chunkInside
- }
-
- /**
- *
- *
- * @type {State}
- */
- function contentEnd(code) {
- effects.exit('chunkContent')
- effects.exit('content')
- return ok(code)
- }
-
- /**
- *
- *
- * @type {State}
- */
- function contentContinue(code) {
- effects.consume(code)
- effects.exit('chunkContent')
- previous.next = effects.enter('chunkContent', {
- contentType: 'content',
- previous
- })
- previous = previous.next
- return chunkInside
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeContinuation(effects, ok, nok) {
- const self = this
- return startLookahead
-
- /**
- *
- *
- * @type {State}
- */
- function startLookahead(code) {
- effects.exit('chunkContent')
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return factorySpace(effects, prefixed, 'linePrefix')
- }
-
- /**
- *
- *
- * @type {State}
- */
- function prefixed(code) {
- if (code === null || markdownLineEnding(code)) {
- return nok(code)
- }
-
- // Always populated by defaults.
-
- const tail = self.events[self.events.length - 1]
- if (
- !self.parser.constructs.disable.null.includes('codeIndented') &&
- tail &&
- tail[1].type === 'linePrefix' &&
- tail[2].sliceSerialize(tail[1], true).length >= 4
- ) {
- return ok(code)
- }
- return effects.interrupt(self.parser.constructs.flow, nok, ok)(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/initialize/flow.js
-/**
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').Initializer} Initializer
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- */
-
-
-
-
-/** @type {InitialConstruct} */
-const flow = {
- tokenize: initializeFlow
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Initializer}
- */
-function initializeFlow(effects) {
- const self = this
- const initial = effects.attempt(
- // Try to parse a blank line.
- blankLine,
- atBlankEnding,
- // Try to parse initial flow (essentially, only code).
- effects.attempt(
- this.parser.constructs.flowInitial,
- afterConstruct,
- factorySpace(
- effects,
- effects.attempt(
- this.parser.constructs.flow,
- afterConstruct,
- effects.attempt(content_content, afterConstruct)
- ),
- 'linePrefix'
- )
- )
- )
- return initial
-
- /** @type {State} */
- function atBlankEnding(code) {
- if (code === null) {
- effects.consume(code)
- return
- }
- effects.enter('lineEndingBlank')
- effects.consume(code)
- effects.exit('lineEndingBlank')
- self.currentConstruct = undefined
- return initial
- }
-
- /** @type {State} */
- function afterConstruct(code) {
- if (code === null) {
- effects.consume(code)
- return
- }
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- self.currentConstruct = undefined
- return initial
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/initialize/text.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').Initializer} Initializer
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- */
-
-const resolver = {
- resolveAll: createResolver()
-}
-const string = initializeFactory('string')
-const text_text = initializeFactory('text')
-
-/**
- * @param {'string' | 'text'} field
- * @returns {InitialConstruct}
- */
-function initializeFactory(field) {
- return {
- tokenize: initializeText,
- resolveAll: createResolver(
- field === 'text' ? resolveAllLineSuffixes : undefined
- )
- }
-
- /**
- * @this {TokenizeContext}
- * @type {Initializer}
- */
- function initializeText(effects) {
- const self = this
- const constructs = this.parser.constructs[field]
- const text = effects.attempt(constructs, start, notText)
- return start
-
- /** @type {State} */
- function start(code) {
- return atBreak(code) ? text(code) : notText(code)
- }
-
- /** @type {State} */
- function notText(code) {
- if (code === null) {
- effects.consume(code)
- return
- }
- effects.enter('data')
- effects.consume(code)
- return data
- }
-
- /** @type {State} */
- function data(code) {
- if (atBreak(code)) {
- effects.exit('data')
- return text(code)
- }
-
- // Data.
- effects.consume(code)
- return data
- }
-
- /**
- * @param {Code} code
- * @returns {boolean}
- */
- function atBreak(code) {
- if (code === null) {
- return true
- }
- const list = constructs[code]
- let index = -1
- if (list) {
- // Always populated by defaults.
-
- while (++index < list.length) {
- const item = list[index]
- if (!item.previous || item.previous.call(self, self.previous)) {
- return true
- }
- }
- }
- return false
- }
- }
-}
-
-/**
- * @param {Resolver | undefined} [extraResolver]
- * @returns {Resolver}
- */
-function createResolver(extraResolver) {
- return resolveAllText
-
- /** @type {Resolver} */
- function resolveAllText(events, context) {
- let index = -1
- /** @type {number | undefined} */
- let enter
-
- // A rather boring computation (to merge adjacent `data` events) which
- // improves mm performance by 29%.
- while (++index <= events.length) {
- if (enter === undefined) {
- if (events[index] && events[index][1].type === 'data') {
- enter = index
- index++
- }
- } else if (!events[index] || events[index][1].type !== 'data') {
- // Don’t do anything if there is one data token.
- if (index !== enter + 2) {
- events[enter][1].end = events[index - 1][1].end
- events.splice(enter + 2, index - enter - 2)
- index = enter + 2
- }
- enter = undefined
- }
- }
- return extraResolver ? extraResolver(events, context) : events
- }
-}
-
-/**
- * A rather ugly set of instructions which again looks at chunks in the input
- * stream.
- * The reason to do this here is that it is *much* faster to parse in reverse.
- * And that we can’t hook into `null` to split the line suffix before an EOF.
- * To do: figure out if we can make this into a clean utility, or even in core.
- * As it will be useful for GFMs literal autolink extension (and maybe even
- * tables?)
- *
- * @type {Resolver}
- */
-function resolveAllLineSuffixes(events, context) {
- let eventIndex = 0 // Skip first.
-
- while (++eventIndex <= events.length) {
- if (
- (eventIndex === events.length ||
- events[eventIndex][1].type === 'lineEnding') &&
- events[eventIndex - 1][1].type === 'data'
- ) {
- const data = events[eventIndex - 1][1]
- const chunks = context.sliceStream(data)
- let index = chunks.length
- let bufferIndex = -1
- let size = 0
- /** @type {boolean | undefined} */
- let tabs
- while (index--) {
- const chunk = chunks[index]
- if (typeof chunk === 'string') {
- bufferIndex = chunk.length
- while (chunk.charCodeAt(bufferIndex - 1) === 32) {
- size++
- bufferIndex--
- }
- if (bufferIndex) break
- bufferIndex = -1
- }
- // Number
- else if (chunk === -2) {
- tabs = true
- size++
- } else if (chunk === -1) {
- // Empty
- } else {
- // Replacement character, exit.
- index++
- break
- }
- }
- if (size) {
- const token = {
- type:
- eventIndex === events.length || tabs || size < 2
- ? 'lineSuffix'
- : 'hardBreakTrailing',
- start: {
- line: data.end.line,
- column: data.end.column - size,
- offset: data.end.offset - size,
- _index: data.start._index + index,
- _bufferIndex: index
- ? bufferIndex
- : data.start._bufferIndex + bufferIndex
- },
- end: Object.assign({}, data.end)
- }
- data.end = Object.assign({}, token.start)
- if (data.start.offset === data.end.offset) {
- Object.assign(data, token)
- } else {
- events.splice(
- eventIndex,
- 0,
- ['enter', token, context],
- ['exit', token, context]
- )
- eventIndex += 2
- }
- }
- eventIndex++
- }
- }
- return events
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-resolve-all/index.js
-/**
- * @typedef {import('micromark-util-types').Event} Event
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- */
-
-/**
- * Call all `resolveAll`s.
- *
- * @param {Array<{resolveAll?: Resolver | undefined}>} constructs
- * List of constructs, optionally with `resolveAll`s.
- * @param {Array<Event>} events
- * List of events.
- * @param {TokenizeContext} context
- * Context used by `tokenize`.
- * @returns {Array<Event>}
- * Changed events.
- */
-function resolveAll(constructs, events, context) {
- /** @type {Array<Resolver>} */
- const called = []
- let index = -1
-
- while (++index < constructs.length) {
- const resolve = constructs[index].resolveAll
-
- if (resolve && !called.includes(resolve)) {
- events = resolve(events, context)
- called.push(resolve)
- }
- }
-
- return events
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/create-tokenizer.js
-/**
- * @typedef {import('micromark-util-types').Chunk} Chunk
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').ConstructRecord} ConstructRecord
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').ParseContext} ParseContext
- * @typedef {import('micromark-util-types').Point} Point
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenType} TokenType
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- */
-
-/**
- * @callback Restore
- * @returns {void}
- *
- * @typedef Info
- * @property {Restore} restore
- * @property {number} from
- *
- * @callback ReturnHandle
- * Handle a successful run.
- * @param {Construct} construct
- * @param {Info} info
- * @returns {void}
- */
-
-
-
-
-/**
- * Create a tokenizer.
- * Tokenizers deal with one type of data (e.g., containers, flow, text).
- * The parser is the object dealing with it all.
- * `initialize` works like other constructs, except that only its `tokenize`
- * function is used, in which case it doesn’t receive an `ok` or `nok`.
- * `from` can be given to set the point before the first character, although
- * when further lines are indented, they must be set with `defineSkip`.
- *
- * @param {ParseContext} parser
- * @param {InitialConstruct} initialize
- * @param {Omit<Point, '_bufferIndex' | '_index'> | undefined} [from]
- * @returns {TokenizeContext}
- */
-function createTokenizer(parser, initialize, from) {
- /** @type {Point} */
- let point = Object.assign(
- from
- ? Object.assign({}, from)
- : {
- line: 1,
- column: 1,
- offset: 0
- },
- {
- _index: 0,
- _bufferIndex: -1
- }
- )
- /** @type {Record<string, number>} */
- const columnStart = {}
- /** @type {Array<Construct>} */
- const resolveAllConstructs = []
- /** @type {Array<Chunk>} */
- let chunks = []
- /** @type {Array<Token>} */
- let stack = []
- /** @type {boolean | undefined} */
- let consumed = true
-
- /**
- * Tools used for tokenizing.
- *
- * @type {Effects}
- */
- const effects = {
- consume,
- enter,
- exit,
- attempt: constructFactory(onsuccessfulconstruct),
- check: constructFactory(onsuccessfulcheck),
- interrupt: constructFactory(onsuccessfulcheck, {
- interrupt: true
- })
- }
-
- /**
- * State and tools for resolving and serializing.
- *
- * @type {TokenizeContext}
- */
- const context = {
- previous: null,
- code: null,
- containerState: {},
- events: [],
- parser,
- sliceStream,
- sliceSerialize,
- now,
- defineSkip,
- write
- }
-
- /**
- * The state function.
- *
- * @type {State | void}
- */
- let state = initialize.tokenize.call(context, effects)
-
- /**
- * Track which character we expect to be consumed, to catch bugs.
- *
- * @type {Code}
- */
- let expectedCode
- if (initialize.resolveAll) {
- resolveAllConstructs.push(initialize)
- }
- return context
-
- /** @type {TokenizeContext['write']} */
- function write(slice) {
- chunks = push(chunks, slice)
- main()
-
- // Exit if we’re not done, resolve might change stuff.
- if (chunks[chunks.length - 1] !== null) {
- return []
- }
- addResult(initialize, 0)
-
- // Otherwise, resolve, and exit.
- context.events = resolveAll(resolveAllConstructs, context.events, context)
- return context.events
- }
-
- //
- // Tools.
- //
-
- /** @type {TokenizeContext['sliceSerialize']} */
- function sliceSerialize(token, expandTabs) {
- return serializeChunks(sliceStream(token), expandTabs)
- }
-
- /** @type {TokenizeContext['sliceStream']} */
- function sliceStream(token) {
- return sliceChunks(chunks, token)
- }
-
- /** @type {TokenizeContext['now']} */
- function now() {
- // This is a hot path, so we clone manually instead of `Object.assign({}, point)`
- const {line, column, offset, _index, _bufferIndex} = point
- return {
- line,
- column,
- offset,
- _index,
- _bufferIndex
- }
- }
-
- /** @type {TokenizeContext['defineSkip']} */
- function defineSkip(value) {
- columnStart[value.line] = value.column
- accountForPotentialSkip()
- }
-
- //
- // State management.
- //
-
- /**
- * Main loop (note that `_index` and `_bufferIndex` in `point` are modified by
- * `consume`).
- * Here is where we walk through the chunks, which either include strings of
- * several characters, or numerical character codes.
- * The reason to do this in a loop instead of a call is so the stack can
- * drain.
- *
- * @returns {void}
- */
- function main() {
- /** @type {number} */
- let chunkIndex
- while (point._index < chunks.length) {
- const chunk = chunks[point._index]
-
- // If we’re in a buffer chunk, loop through it.
- if (typeof chunk === 'string') {
- chunkIndex = point._index
- if (point._bufferIndex < 0) {
- point._bufferIndex = 0
- }
- while (
- point._index === chunkIndex &&
- point._bufferIndex < chunk.length
- ) {
- go(chunk.charCodeAt(point._bufferIndex))
- }
- } else {
- go(chunk)
- }
- }
- }
-
- /**
- * Deal with one code.
- *
- * @param {Code} code
- * @returns {void}
- */
- function go(code) {
- consumed = undefined
- expectedCode = code
- state = state(code)
- }
-
- /** @type {Effects['consume']} */
- function consume(code) {
- if (markdownLineEnding(code)) {
- point.line++
- point.column = 1
- point.offset += code === -3 ? 2 : 1
- accountForPotentialSkip()
- } else if (code !== -1) {
- point.column++
- point.offset++
- }
-
- // Not in a string chunk.
- if (point._bufferIndex < 0) {
- point._index++
- } else {
- point._bufferIndex++
-
- // At end of string chunk.
- // @ts-expect-error Points w/ non-negative `_bufferIndex` reference
- // strings.
- if (point._bufferIndex === chunks[point._index].length) {
- point._bufferIndex = -1
- point._index++
- }
- }
-
- // Expose the previous character.
- context.previous = code
-
- // Mark as consumed.
- consumed = true
- }
-
- /** @type {Effects['enter']} */
- function enter(type, fields) {
- /** @type {Token} */
- // @ts-expect-error Patch instead of assign required fields to help GC.
- const token = fields || {}
- token.type = type
- token.start = now()
- context.events.push(['enter', token, context])
- stack.push(token)
- return token
- }
-
- /** @type {Effects['exit']} */
- function exit(type) {
- const token = stack.pop()
- token.end = now()
- context.events.push(['exit', token, context])
- return token
- }
-
- /**
- * Use results.
- *
- * @type {ReturnHandle}
- */
- function onsuccessfulconstruct(construct, info) {
- addResult(construct, info.from)
- }
-
- /**
- * Discard results.
- *
- * @type {ReturnHandle}
- */
- function onsuccessfulcheck(_, info) {
- info.restore()
- }
-
- /**
- * Factory to attempt/check/interrupt.
- *
- * @param {ReturnHandle} onreturn
- * @param {{interrupt?: boolean | undefined} | undefined} [fields]
- */
- function constructFactory(onreturn, fields) {
- return hook
-
- /**
- * Handle either an object mapping codes to constructs, a list of
- * constructs, or a single construct.
- *
- * @param {Array<Construct> | Construct | ConstructRecord} constructs
- * @param {State} returnState
- * @param {State | undefined} [bogusState]
- * @returns {State}
- */
- function hook(constructs, returnState, bogusState) {
- /** @type {Array<Construct>} */
- let listOfConstructs
- /** @type {number} */
- let constructIndex
- /** @type {Construct} */
- let currentConstruct
- /** @type {Info} */
- let info
- return Array.isArray(constructs) /* c8 ignore next 1 */
- ? handleListOfConstructs(constructs)
- : 'tokenize' in constructs
- ? // @ts-expect-error Looks like a construct.
- handleListOfConstructs([constructs])
- : handleMapOfConstructs(constructs)
-
- /**
- * Handle a list of construct.
- *
- * @param {ConstructRecord} map
- * @returns {State}
- */
- function handleMapOfConstructs(map) {
- return start
-
- /** @type {State} */
- function start(code) {
- const def = code !== null && map[code]
- const all = code !== null && map.null
- const list = [
- // To do: add more extension tests.
- /* c8 ignore next 2 */
- ...(Array.isArray(def) ? def : def ? [def] : []),
- ...(Array.isArray(all) ? all : all ? [all] : [])
- ]
- return handleListOfConstructs(list)(code)
- }
- }
-
- /**
- * Handle a list of construct.
- *
- * @param {Array<Construct>} list
- * @returns {State}
- */
- function handleListOfConstructs(list) {
- listOfConstructs = list
- constructIndex = 0
- if (list.length === 0) {
- return bogusState
- }
- return handleConstruct(list[constructIndex])
- }
-
- /**
- * Handle a single construct.
- *
- * @param {Construct} construct
- * @returns {State}
- */
- function handleConstruct(construct) {
- return start
-
- /** @type {State} */
- function start(code) {
- // To do: not needed to store if there is no bogus state, probably?
- // Currently doesn’t work because `inspect` in document does a check
- // w/o a bogus, which doesn’t make sense. But it does seem to help perf
- // by not storing.
- info = store()
- currentConstruct = construct
- if (!construct.partial) {
- context.currentConstruct = construct
- }
-
- // Always populated by defaults.
-
- if (
- construct.name &&
- context.parser.constructs.disable.null.includes(construct.name)
- ) {
- return nok(code)
- }
- return construct.tokenize.call(
- // If we do have fields, create an object w/ `context` as its
- // prototype.
- // This allows a “live binding”, which is needed for `interrupt`.
- fields ? Object.assign(Object.create(context), fields) : context,
- effects,
- ok,
- nok
- )(code)
- }
- }
-
- /** @type {State} */
- function ok(code) {
- consumed = true
- onreturn(currentConstruct, info)
- return returnState
- }
-
- /** @type {State} */
- function nok(code) {
- consumed = true
- info.restore()
- if (++constructIndex < listOfConstructs.length) {
- return handleConstruct(listOfConstructs[constructIndex])
- }
- return bogusState
- }
- }
- }
-
- /**
- * @param {Construct} construct
- * @param {number} from
- * @returns {void}
- */
- function addResult(construct, from) {
- if (construct.resolveAll && !resolveAllConstructs.includes(construct)) {
- resolveAllConstructs.push(construct)
- }
- if (construct.resolve) {
- splice(
- context.events,
- from,
- context.events.length - from,
- construct.resolve(context.events.slice(from), context)
- )
- }
- if (construct.resolveTo) {
- context.events = construct.resolveTo(context.events, context)
- }
- }
-
- /**
- * Store state.
- *
- * @returns {Info}
- */
- function store() {
- const startPoint = now()
- const startPrevious = context.previous
- const startCurrentConstruct = context.currentConstruct
- const startEventsIndex = context.events.length
- const startStack = Array.from(stack)
- return {
- restore,
- from: startEventsIndex
- }
-
- /**
- * Restore state.
- *
- * @returns {void}
- */
- function restore() {
- point = startPoint
- context.previous = startPrevious
- context.currentConstruct = startCurrentConstruct
- context.events.length = startEventsIndex
- stack = startStack
- accountForPotentialSkip()
- }
- }
-
- /**
- * Move the current point a bit forward in the line when it’s on a column
- * skip.
- *
- * @returns {void}
- */
- function accountForPotentialSkip() {
- if (point.line in columnStart && point.column < 2) {
- point.column = columnStart[point.line]
- point.offset += columnStart[point.line] - 1
- }
- }
-}
-
-/**
- * Get the chunks from a slice of chunks in the range of a token.
- *
- * @param {Array<Chunk>} chunks
- * @param {Pick<Token, 'end' | 'start'>} token
- * @returns {Array<Chunk>}
- */
-function sliceChunks(chunks, token) {
- const startIndex = token.start._index
- const startBufferIndex = token.start._bufferIndex
- const endIndex = token.end._index
- const endBufferIndex = token.end._bufferIndex
- /** @type {Array<Chunk>} */
- let view
- if (startIndex === endIndex) {
- // @ts-expect-error `_bufferIndex` is used on string chunks.
- view = [chunks[startIndex].slice(startBufferIndex, endBufferIndex)]
- } else {
- view = chunks.slice(startIndex, endIndex)
- if (startBufferIndex > -1) {
- const head = view[0]
- if (typeof head === 'string') {
- view[0] = head.slice(startBufferIndex)
- } else {
- view.shift()
- }
- }
- if (endBufferIndex > 0) {
- // @ts-expect-error `_bufferIndex` is used on string chunks.
- view.push(chunks[endIndex].slice(0, endBufferIndex))
- }
- }
- return view
-}
-
-/**
- * Get the string value of a slice of chunks.
- *
- * @param {Array<Chunk>} chunks
- * @param {boolean | undefined} [expandTabs=false]
- * @returns {string}
- */
-function serializeChunks(chunks, expandTabs) {
- let index = -1
- /** @type {Array<string>} */
- const result = []
- /** @type {boolean | undefined} */
- let atTab
- while (++index < chunks.length) {
- const chunk = chunks[index]
- /** @type {string} */
- let value
- if (typeof chunk === 'string') {
- value = chunk
- } else
- switch (chunk) {
- case -5: {
- value = '\r'
- break
- }
- case -4: {
- value = '\n'
- break
- }
- case -3: {
- value = '\r' + '\n'
- break
- }
- case -2: {
- value = expandTabs ? ' ' : '\t'
- break
- }
- case -1: {
- if (!expandTabs && atTab) continue
- value = ' '
- break
- }
- default: {
- // Currently only replacement character.
- value = String.fromCharCode(chunk)
- }
- }
- atTab = chunk === -2
- result.push(value)
- }
- return result.join('')
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/thematic-break.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const thematicBreak = {
- name: 'thematicBreak',
- tokenize: tokenizeThematicBreak
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeThematicBreak(effects, ok, nok) {
- let size = 0
- /** @type {NonNullable<Code>} */
- let marker
- return start
-
- /**
- * Start of thematic break.
- *
- * ```markdown
- * > | ***
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('thematicBreak')
- // To do: parse indent like `markdown-rs`.
- return before(code)
- }
-
- /**
- * After optional whitespace, at marker.
- *
- * ```markdown
- * > | ***
- * ^
- * ```
- *
- * @type {State}
- */
- function before(code) {
- marker = code
- return atBreak(code)
- }
-
- /**
- * After something, before something else.
- *
- * ```markdown
- * > | ***
- * ^
- * ```
- *
- * @type {State}
- */
- function atBreak(code) {
- if (code === marker) {
- effects.enter('thematicBreakSequence')
- return sequence(code)
- }
- if (size >= 3 && (code === null || markdownLineEnding(code))) {
- effects.exit('thematicBreak')
- return ok(code)
- }
- return nok(code)
- }
-
- /**
- * In sequence.
- *
- * ```markdown
- * > | ***
- * ^
- * ```
- *
- * @type {State}
- */
- function sequence(code) {
- if (code === marker) {
- effects.consume(code)
- size++
- return sequence
- }
- effects.exit('thematicBreakSequence')
- return markdownSpace(code)
- ? factorySpace(effects, atBreak, 'whitespace')(code)
- : atBreak(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/list.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').ContainerState} ContainerState
- * @typedef {import('micromark-util-types').Exiter} Exiter
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-
-
-/** @type {Construct} */
-const list = {
- name: 'list',
- tokenize: tokenizeListStart,
- continuation: {
- tokenize: tokenizeListContinuation
- },
- exit: tokenizeListEnd
-}
-
-/** @type {Construct} */
-const listItemPrefixWhitespaceConstruct = {
- tokenize: tokenizeListItemPrefixWhitespace,
- partial: true
-}
-
-/** @type {Construct} */
-const indentConstruct = {
- tokenize: tokenizeIndent,
- partial: true
-}
-
-// To do: `markdown-rs` parses list items on their own and later stitches them
-// together.
-
-/**
- * @type {Tokenizer}
- * @this {TokenizeContext}
- */
-function tokenizeListStart(effects, ok, nok) {
- const self = this
- const tail = self.events[self.events.length - 1]
- let initialSize =
- tail && tail[1].type === 'linePrefix'
- ? tail[2].sliceSerialize(tail[1], true).length
- : 0
- let size = 0
- return start
-
- /** @type {State} */
- function start(code) {
- const kind =
- self.containerState.type ||
- (code === 42 || code === 43 || code === 45
- ? 'listUnordered'
- : 'listOrdered')
- if (
- kind === 'listUnordered'
- ? !self.containerState.marker || code === self.containerState.marker
- : asciiDigit(code)
- ) {
- if (!self.containerState.type) {
- self.containerState.type = kind
- effects.enter(kind, {
- _container: true
- })
- }
- if (kind === 'listUnordered') {
- effects.enter('listItemPrefix')
- return code === 42 || code === 45
- ? effects.check(thematicBreak, nok, atMarker)(code)
- : atMarker(code)
- }
- if (!self.interrupt || code === 49) {
- effects.enter('listItemPrefix')
- effects.enter('listItemValue')
- return inside(code)
- }
- }
- return nok(code)
- }
-
- /** @type {State} */
- function inside(code) {
- if (asciiDigit(code) && ++size < 10) {
- effects.consume(code)
- return inside
- }
- if (
- (!self.interrupt || size < 2) &&
- (self.containerState.marker
- ? code === self.containerState.marker
- : code === 41 || code === 46)
- ) {
- effects.exit('listItemValue')
- return atMarker(code)
- }
- return nok(code)
- }
-
- /**
- * @type {State}
- **/
- function atMarker(code) {
- effects.enter('listItemMarker')
- effects.consume(code)
- effects.exit('listItemMarker')
- self.containerState.marker = self.containerState.marker || code
- return effects.check(
- blankLine,
- // Can’t be empty when interrupting.
- self.interrupt ? nok : onBlank,
- effects.attempt(
- listItemPrefixWhitespaceConstruct,
- endOfPrefix,
- otherPrefix
- )
- )
- }
-
- /** @type {State} */
- function onBlank(code) {
- self.containerState.initialBlankLine = true
- initialSize++
- return endOfPrefix(code)
- }
-
- /** @type {State} */
- function otherPrefix(code) {
- if (markdownSpace(code)) {
- effects.enter('listItemPrefixWhitespace')
- effects.consume(code)
- effects.exit('listItemPrefixWhitespace')
- return endOfPrefix
- }
- return nok(code)
- }
-
- /** @type {State} */
- function endOfPrefix(code) {
- self.containerState.size =
- initialSize +
- self.sliceSerialize(effects.exit('listItemPrefix'), true).length
- return ok(code)
- }
-}
-
-/**
- * @type {Tokenizer}
- * @this {TokenizeContext}
- */
-function tokenizeListContinuation(effects, ok, nok) {
- const self = this
- self.containerState._closeFlow = undefined
- return effects.check(blankLine, onBlank, notBlank)
-
- /** @type {State} */
- function onBlank(code) {
- self.containerState.furtherBlankLines =
- self.containerState.furtherBlankLines ||
- self.containerState.initialBlankLine
-
- // We have a blank line.
- // Still, try to consume at most the items size.
- return factorySpace(
- effects,
- ok,
- 'listItemIndent',
- self.containerState.size + 1
- )(code)
- }
-
- /** @type {State} */
- function notBlank(code) {
- if (self.containerState.furtherBlankLines || !markdownSpace(code)) {
- self.containerState.furtherBlankLines = undefined
- self.containerState.initialBlankLine = undefined
- return notInCurrentItem(code)
- }
- self.containerState.furtherBlankLines = undefined
- self.containerState.initialBlankLine = undefined
- return effects.attempt(indentConstruct, ok, notInCurrentItem)(code)
- }
-
- /** @type {State} */
- function notInCurrentItem(code) {
- // While we do continue, we signal that the flow should be closed.
- self.containerState._closeFlow = true
- // As we’re closing flow, we’re no longer interrupting.
- self.interrupt = undefined
- // Always populated by defaults.
-
- return factorySpace(
- effects,
- effects.attempt(list, ok, nok),
- 'linePrefix',
- self.parser.constructs.disable.null.includes('codeIndented')
- ? undefined
- : 4
- )(code)
- }
-}
-
-/**
- * @type {Tokenizer}
- * @this {TokenizeContext}
- */
-function tokenizeIndent(effects, ok, nok) {
- const self = this
- return factorySpace(
- effects,
- afterPrefix,
- 'listItemIndent',
- self.containerState.size + 1
- )
-
- /** @type {State} */
- function afterPrefix(code) {
- const tail = self.events[self.events.length - 1]
- return tail &&
- tail[1].type === 'listItemIndent' &&
- tail[2].sliceSerialize(tail[1], true).length === self.containerState.size
- ? ok(code)
- : nok(code)
- }
-}
-
-/**
- * @type {Exiter}
- * @this {TokenizeContext}
- */
-function tokenizeListEnd(effects) {
- effects.exit(this.containerState.type)
-}
-
-/**
- * @type {Tokenizer}
- * @this {TokenizeContext}
- */
-function tokenizeListItemPrefixWhitespace(effects, ok, nok) {
- const self = this
-
- // Always populated by defaults.
-
- return factorySpace(
- effects,
- afterPrefix,
- 'listItemPrefixWhitespace',
- self.parser.constructs.disable.null.includes('codeIndented')
- ? undefined
- : 4 + 1
- )
-
- /** @type {State} */
- function afterPrefix(code) {
- const tail = self.events[self.events.length - 1]
- return !markdownSpace(code) &&
- tail &&
- tail[1].type === 'listItemPrefixWhitespace'
- ? ok(code)
- : nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/block-quote.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Exiter} Exiter
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const blockQuote = {
- name: 'blockQuote',
- tokenize: tokenizeBlockQuoteStart,
- continuation: {
- tokenize: tokenizeBlockQuoteContinuation
- },
- exit
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeBlockQuoteStart(effects, ok, nok) {
- const self = this
- return start
-
- /**
- * Start of block quote.
- *
- * ```markdown
- * > | > a
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- if (code === 62) {
- const state = self.containerState
- if (!state.open) {
- effects.enter('blockQuote', {
- _container: true
- })
- state.open = true
- }
- effects.enter('blockQuotePrefix')
- effects.enter('blockQuoteMarker')
- effects.consume(code)
- effects.exit('blockQuoteMarker')
- return after
- }
- return nok(code)
- }
-
- /**
- * After `>`, before optional whitespace.
- *
- * ```markdown
- * > | > a
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- if (markdownSpace(code)) {
- effects.enter('blockQuotePrefixWhitespace')
- effects.consume(code)
- effects.exit('blockQuotePrefixWhitespace')
- effects.exit('blockQuotePrefix')
- return ok
- }
- effects.exit('blockQuotePrefix')
- return ok(code)
- }
-}
-
-/**
- * Start of block quote continuation.
- *
- * ```markdown
- * | > a
- * > | > b
- * ^
- * ```
- *
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeBlockQuoteContinuation(effects, ok, nok) {
- const self = this
- return contStart
-
- /**
- * Start of block quote continuation.
- *
- * Also used to parse the first block quote opening.
- *
- * ```markdown
- * | > a
- * > | > b
- * ^
- * ```
- *
- * @type {State}
- */
- function contStart(code) {
- if (markdownSpace(code)) {
- // Always populated by defaults.
-
- return factorySpace(
- effects,
- contBefore,
- 'linePrefix',
- self.parser.constructs.disable.null.includes('codeIndented')
- ? undefined
- : 4
- )(code)
- }
- return contBefore(code)
- }
-
- /**
- * At `>`, after optional whitespace.
- *
- * Also used to parse the first block quote opening.
- *
- * ```markdown
- * | > a
- * > | > b
- * ^
- * ```
- *
- * @type {State}
- */
- function contBefore(code) {
- return effects.attempt(blockQuote, ok, nok)(code)
- }
-}
-
-/** @type {Exiter} */
-function exit(effects) {
- effects.exit('blockQuote')
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-factory-destination/index.js
-/**
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenType} TokenType
- */
-
-
-/**
- * Parse destinations.
- *
- * ###### Examples
- *
- * ```markdown
- * <a>
- * <a\>b>
- * <a b>
- * <a)>
- * a
- * a\)b
- * a(b)c
- * a(b)
- * ```
- *
- * @param {Effects} effects
- * Context.
- * @param {State} ok
- * State switched to when successful.
- * @param {State} nok
- * State switched to when unsuccessful.
- * @param {TokenType} type
- * Type for whole (`<a>` or `b`).
- * @param {TokenType} literalType
- * Type when enclosed (`<a>`).
- * @param {TokenType} literalMarkerType
- * Type for enclosing (`<` and `>`).
- * @param {TokenType} rawType
- * Type when not enclosed (`b`).
- * @param {TokenType} stringType
- * Type for the value (`a` or `b`).
- * @param {number | undefined} [max=Infinity]
- * Depth of nested parens (inclusive).
- * @returns {State}
- * Start state.
- */ // eslint-disable-next-line max-params
-function factoryDestination(
- effects,
- ok,
- nok,
- type,
- literalType,
- literalMarkerType,
- rawType,
- stringType,
- max
-) {
- const limit = max || Number.POSITIVE_INFINITY
- let balance = 0
- return start
-
- /**
- * Start of destination.
- *
- * ```markdown
- * > | <aa>
- * ^
- * > | aa
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- if (code === 60) {
- effects.enter(type)
- effects.enter(literalType)
- effects.enter(literalMarkerType)
- effects.consume(code)
- effects.exit(literalMarkerType)
- return enclosedBefore
- }
-
- // ASCII control, space, closing paren.
- if (code === null || code === 32 || code === 41 || asciiControl(code)) {
- return nok(code)
- }
- effects.enter(type)
- effects.enter(rawType)
- effects.enter(stringType)
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return raw(code)
- }
-
- /**
- * After `<`, at an enclosed destination.
- *
- * ```markdown
- * > | <aa>
- * ^
- * ```
- *
- * @type {State}
- */
- function enclosedBefore(code) {
- if (code === 62) {
- effects.enter(literalMarkerType)
- effects.consume(code)
- effects.exit(literalMarkerType)
- effects.exit(literalType)
- effects.exit(type)
- return ok
- }
- effects.enter(stringType)
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return enclosed(code)
- }
-
- /**
- * In enclosed destination.
- *
- * ```markdown
- * > | <aa>
- * ^
- * ```
- *
- * @type {State}
- */
- function enclosed(code) {
- if (code === 62) {
- effects.exit('chunkString')
- effects.exit(stringType)
- return enclosedBefore(code)
- }
- if (code === null || code === 60 || markdownLineEnding(code)) {
- return nok(code)
- }
- effects.consume(code)
- return code === 92 ? enclosedEscape : enclosed
- }
-
- /**
- * After `\`, at a special character.
- *
- * ```markdown
- * > | <a\*a>
- * ^
- * ```
- *
- * @type {State}
- */
- function enclosedEscape(code) {
- if (code === 60 || code === 62 || code === 92) {
- effects.consume(code)
- return enclosed
- }
- return enclosed(code)
- }
-
- /**
- * In raw destination.
- *
- * ```markdown
- * > | aa
- * ^
- * ```
- *
- * @type {State}
- */
- function raw(code) {
- if (
- !balance &&
- (code === null || code === 41 || markdownLineEndingOrSpace(code))
- ) {
- effects.exit('chunkString')
- effects.exit(stringType)
- effects.exit(rawType)
- effects.exit(type)
- return ok(code)
- }
- if (balance < limit && code === 40) {
- effects.consume(code)
- balance++
- return raw
- }
- if (code === 41) {
- effects.consume(code)
- balance--
- return raw
- }
-
- // ASCII control (but *not* `\0`) and space and `(`.
- // Note: in `markdown-rs`, `\0` exists in codes, in `micromark-js` it
- // doesn’t.
- if (code === null || code === 32 || code === 40 || asciiControl(code)) {
- return nok(code)
- }
- effects.consume(code)
- return code === 92 ? rawEscape : raw
- }
-
- /**
- * After `\`, at special character.
- *
- * ```markdown
- * > | a\*a
- * ^
- * ```
- *
- * @type {State}
- */
- function rawEscape(code) {
- if (code === 40 || code === 41 || code === 92) {
- effects.consume(code)
- return raw
- }
- return raw(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-factory-label/index.js
-/**
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').TokenType} TokenType
- */
-
-
-/**
- * Parse labels.
- *
- * > 👉 **Note**: labels in markdown are capped at 999 characters in the string.
- *
- * ###### Examples
- *
- * ```markdown
- * [a]
- * [a
- * b]
- * [a\]b]
- * ```
- *
- * @this {TokenizeContext}
- * Tokenize context.
- * @param {Effects} effects
- * Context.
- * @param {State} ok
- * State switched to when successful.
- * @param {State} nok
- * State switched to when unsuccessful.
- * @param {TokenType} type
- * Type of the whole label (`[a]`).
- * @param {TokenType} markerType
- * Type for the markers (`[` and `]`).
- * @param {TokenType} stringType
- * Type for the identifier (`a`).
- * @returns {State}
- * Start state.
- */ // eslint-disable-next-line max-params
-function factoryLabel(effects, ok, nok, type, markerType, stringType) {
- const self = this
- let size = 0
- /** @type {boolean} */
- let seen
- return start
-
- /**
- * Start of label.
- *
- * ```markdown
- * > | [a]
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter(type)
- effects.enter(markerType)
- effects.consume(code)
- effects.exit(markerType)
- effects.enter(stringType)
- return atBreak
- }
-
- /**
- * In label, at something, before something else.
- *
- * ```markdown
- * > | [a]
- * ^
- * ```
- *
- * @type {State}
- */
- function atBreak(code) {
- if (
- size > 999 ||
- code === null ||
- code === 91 ||
- (code === 93 && !seen) ||
- // To do: remove in the future once we’ve switched from
- // `micromark-extension-footnote` to `micromark-extension-gfm-footnote`,
- // which doesn’t need this.
- // Hidden footnotes hook.
- /* c8 ignore next 3 */
- (code === 94 &&
- !size &&
- '_hiddenFootnoteSupport' in self.parser.constructs)
- ) {
- return nok(code)
- }
- if (code === 93) {
- effects.exit(stringType)
- effects.enter(markerType)
- effects.consume(code)
- effects.exit(markerType)
- effects.exit(type)
- return ok
- }
-
- // To do: indent? Link chunks and EOLs together?
- if (markdownLineEnding(code)) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return atBreak
- }
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return labelInside(code)
- }
-
- /**
- * In label, in text.
- *
- * ```markdown
- * > | [a]
- * ^
- * ```
- *
- * @type {State}
- */
- function labelInside(code) {
- if (
- code === null ||
- code === 91 ||
- code === 93 ||
- markdownLineEnding(code) ||
- size++ > 999
- ) {
- effects.exit('chunkString')
- return atBreak(code)
- }
- effects.consume(code)
- if (!seen) seen = !markdownSpace(code)
- return code === 92 ? labelEscape : labelInside
- }
-
- /**
- * After `\`, at a special character.
- *
- * ```markdown
- * > | [a\*a]
- * ^
- * ```
- *
- * @type {State}
- */
- function labelEscape(code) {
- if (code === 91 || code === 92 || code === 93) {
- effects.consume(code)
- size++
- return labelInside
- }
- return labelInside(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-factory-title/index.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenType} TokenType
- */
-
-
-
-/**
- * Parse titles.
- *
- * ###### Examples
- *
- * ```markdown
- * "a"
- * 'b'
- * (c)
- * "a
- * b"
- * 'a
- * b'
- * (a\)b)
- * ```
- *
- * @param {Effects} effects
- * Context.
- * @param {State} ok
- * State switched to when successful.
- * @param {State} nok
- * State switched to when unsuccessful.
- * @param {TokenType} type
- * Type of the whole title (`"a"`, `'b'`, `(c)`).
- * @param {TokenType} markerType
- * Type for the markers (`"`, `'`, `(`, and `)`).
- * @param {TokenType} stringType
- * Type for the value (`a`).
- * @returns {State}
- * Start state.
- */ // eslint-disable-next-line max-params
-function factoryTitle(effects, ok, nok, type, markerType, stringType) {
- /** @type {NonNullable<Code>} */
- let marker
- return start
-
- /**
- * Start of title.
- *
- * ```markdown
- * > | "a"
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- if (code === 34 || code === 39 || code === 40) {
- effects.enter(type)
- effects.enter(markerType)
- effects.consume(code)
- effects.exit(markerType)
- marker = code === 40 ? 41 : code
- return begin
- }
- return nok(code)
- }
-
- /**
- * After opening marker.
- *
- * This is also used at the closing marker.
- *
- * ```markdown
- * > | "a"
- * ^
- * ```
- *
- * @type {State}
- */
- function begin(code) {
- if (code === marker) {
- effects.enter(markerType)
- effects.consume(code)
- effects.exit(markerType)
- effects.exit(type)
- return ok
- }
- effects.enter(stringType)
- return atBreak(code)
- }
-
- /**
- * At something, before something else.
- *
- * ```markdown
- * > | "a"
- * ^
- * ```
- *
- * @type {State}
- */
- function atBreak(code) {
- if (code === marker) {
- effects.exit(stringType)
- return begin(marker)
- }
- if (code === null) {
- return nok(code)
- }
-
- // Note: blank lines can’t exist in content.
- if (markdownLineEnding(code)) {
- // To do: use `space_or_tab_eol_with_options`, connect.
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return factorySpace(effects, atBreak, 'linePrefix')
- }
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return inside(code)
- }
-
- /**
- *
- *
- * @type {State}
- */
- function inside(code) {
- if (code === marker || code === null || markdownLineEnding(code)) {
- effects.exit('chunkString')
- return atBreak(code)
- }
- effects.consume(code)
- return code === 92 ? escape : inside
- }
-
- /**
- * After `\`, at a special character.
- *
- * ```markdown
- * > | "a\*b"
- * ^
- * ```
- *
- * @type {State}
- */
- function escape(code) {
- if (code === marker || code === 92) {
- effects.consume(code)
- return inside
- }
- return inside(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-factory-whitespace/index.js
-/**
- * @typedef {import('micromark-util-types').Effects} Effects
- * @typedef {import('micromark-util-types').State} State
- */
-
-
-
-/**
- * Parse spaces and tabs.
- *
- * There is no `nok` parameter:
- *
- * * line endings or spaces in markdown are often optional, in which case this
- * factory can be used and `ok` will be switched to whether spaces were found
- * or not
- * * one line ending or space can be detected with
- * `markdownLineEndingOrSpace(code)` right before using `factoryWhitespace`
- *
- * @param {Effects} effects
- * Context.
- * @param {State} ok
- * State switched to when successful.
- * @returns
- * Start state.
- */
-function factoryWhitespace(effects, ok) {
- /** @type {boolean} */
- let seen
- return start
-
- /** @type {State} */
- function start(code) {
- if (markdownLineEnding(code)) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- seen = true
- return start
- }
- if (markdownSpace(code)) {
- return factorySpace(
- effects,
- start,
- seen ? 'linePrefix' : 'lineSuffix'
- )(code)
- }
- return ok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-normalize-identifier/index.js
-/**
- * Normalize an identifier (as found in references, definitions).
- *
- * Collapses markdown whitespace, trim, and then lower- and uppercase.
- *
- * Some characters are considered “uppercase”, such as U+03F4 (`ϴ`), but if their
- * lowercase counterpart (U+03B8 (`θ`)) is uppercased will result in a different
- * uppercase character (U+0398 (`Θ`)).
- * So, to get a canonical form, we perform both lower- and uppercase.
- *
- * Using uppercase last makes sure keys will never interact with default
- * prototypal values (such as `constructor`): nothing in the prototype of
- * `Object` is uppercase.
- *
- * @param {string} value
- * Identifier to normalize.
- * @returns {string}
- * Normalized identifier.
- */
-function normalizeIdentifier(value) {
- return (
- value
- // Collapse markdown whitespace.
- .replace(/[\t\n\r ]+/g, ' ')
- // Trim.
- .replace(/^ | $/g, '')
- // Some characters are considered “uppercase”, but if their lowercase
- // counterpart is uppercased will result in a different uppercase
- // character.
- // Hence, to get that form, we perform both lower- and uppercase.
- // Upper case makes sure keys will not interact with default prototypal
- // methods: no method is uppercase.
- .toLowerCase()
- .toUpperCase()
- )
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/definition.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-
-
-
-
-/** @type {Construct} */
-const definition = {
- name: 'definition',
- tokenize: tokenizeDefinition
-}
-
-/** @type {Construct} */
-const titleBefore = {
- tokenize: tokenizeTitleBefore,
- partial: true
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeDefinition(effects, ok, nok) {
- const self = this
- /** @type {string} */
- let identifier
- return start
-
- /**
- * At start of a definition.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // Do not interrupt paragraphs (but do follow definitions).
- // To do: do `interrupt` the way `markdown-rs` does.
- // To do: parse whitespace the way `markdown-rs` does.
- effects.enter('definition')
- return before(code)
- }
-
- /**
- * After optional whitespace, at `[`.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function before(code) {
- // To do: parse whitespace the way `markdown-rs` does.
-
- return factoryLabel.call(
- self,
- effects,
- labelAfter,
- // Note: we don’t need to reset the way `markdown-rs` does.
- nok,
- 'definitionLabel',
- 'definitionLabelMarker',
- 'definitionLabelString'
- )(code)
- }
-
- /**
- * After label.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function labelAfter(code) {
- identifier = normalizeIdentifier(
- self.sliceSerialize(self.events[self.events.length - 1][1]).slice(1, -1)
- )
- if (code === 58) {
- effects.enter('definitionMarker')
- effects.consume(code)
- effects.exit('definitionMarker')
- return markerAfter
- }
- return nok(code)
- }
-
- /**
- * After marker.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function markerAfter(code) {
- // Note: whitespace is optional.
- return markdownLineEndingOrSpace(code)
- ? factoryWhitespace(effects, destinationBefore)(code)
- : destinationBefore(code)
- }
-
- /**
- * Before destination.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function destinationBefore(code) {
- return factoryDestination(
- effects,
- destinationAfter,
- // Note: we don’t need to reset the way `markdown-rs` does.
- nok,
- 'definitionDestination',
- 'definitionDestinationLiteral',
- 'definitionDestinationLiteralMarker',
- 'definitionDestinationRaw',
- 'definitionDestinationString'
- )(code)
- }
-
- /**
- * After destination.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function destinationAfter(code) {
- return effects.attempt(titleBefore, after, after)(code)
- }
-
- /**
- * After definition.
- *
- * ```markdown
- * > | [a]: b
- * ^
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- return markdownSpace(code)
- ? factorySpace(effects, afterWhitespace, 'whitespace')(code)
- : afterWhitespace(code)
- }
-
- /**
- * After definition, after optional whitespace.
- *
- * ```markdown
- * > | [a]: b
- * ^
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function afterWhitespace(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('definition')
-
- // Note: we don’t care about uniqueness.
- // It’s likely that that doesn’t happen very frequently.
- // It is more likely that it wastes precious time.
- self.parser.defined.push(identifier)
-
- // To do: `markdown-rs` interrupt.
- // // You’d be interrupting.
- // tokenizer.interrupt = true
- return ok(code)
- }
- return nok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeTitleBefore(effects, ok, nok) {
- return titleBefore
-
- /**
- * After destination, at whitespace.
- *
- * ```markdown
- * > | [a]: b
- * ^
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function titleBefore(code) {
- return markdownLineEndingOrSpace(code)
- ? factoryWhitespace(effects, beforeMarker)(code)
- : nok(code)
- }
-
- /**
- * At title.
- *
- * ```markdown
- * | [a]: b
- * > | "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function beforeMarker(code) {
- return factoryTitle(
- effects,
- titleAfter,
- nok,
- 'definitionTitle',
- 'definitionTitleMarker',
- 'definitionTitleString'
- )(code)
- }
-
- /**
- * After title.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function titleAfter(code) {
- return markdownSpace(code)
- ? factorySpace(effects, titleAfterOptionalWhitespace, 'whitespace')(code)
- : titleAfterOptionalWhitespace(code)
- }
-
- /**
- * After title, after optional whitespace.
- *
- * ```markdown
- * > | [a]: b "c"
- * ^
- * ```
- *
- * @type {State}
- */
- function titleAfterOptionalWhitespace(code) {
- return code === null || markdownLineEnding(code) ? ok(code) : nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/code-indented.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const codeIndented = {
- name: 'codeIndented',
- tokenize: tokenizeCodeIndented
-}
-
-/** @type {Construct} */
-const furtherStart = {
- tokenize: tokenizeFurtherStart,
- partial: true
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeCodeIndented(effects, ok, nok) {
- const self = this
- return start
-
- /**
- * Start of code (indented).
- *
- * > **Parsing note**: it is not needed to check if this first line is a
- * > filled line (that it has a non-whitespace character), because blank lines
- * > are parsed already, so we never run into that.
- *
- * ```markdown
- * > | aaa
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // To do: manually check if interrupting like `markdown-rs`.
-
- effects.enter('codeIndented')
- // To do: use an improved `space_or_tab` function like `markdown-rs`,
- // so that we can drop the next state.
- return factorySpace(effects, afterPrefix, 'linePrefix', 4 + 1)(code)
- }
-
- /**
- * At start, after 1 or 4 spaces.
- *
- * ```markdown
- * > | aaa
- * ^
- * ```
- *
- * @type {State}
- */
- function afterPrefix(code) {
- const tail = self.events[self.events.length - 1]
- return tail &&
- tail[1].type === 'linePrefix' &&
- tail[2].sliceSerialize(tail[1], true).length >= 4
- ? atBreak(code)
- : nok(code)
- }
-
- /**
- * At a break.
- *
- * ```markdown
- * > | aaa
- * ^ ^
- * ```
- *
- * @type {State}
- */
- function atBreak(code) {
- if (code === null) {
- return after(code)
- }
- if (markdownLineEnding(code)) {
- return effects.attempt(furtherStart, atBreak, after)(code)
- }
- effects.enter('codeFlowValue')
- return inside(code)
- }
-
- /**
- * In code content.
- *
- * ```markdown
- * > | aaa
- * ^^^^
- * ```
- *
- * @type {State}
- */
- function inside(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('codeFlowValue')
- return atBreak(code)
- }
- effects.consume(code)
- return inside
- }
-
- /** @type {State} */
- function after(code) {
- effects.exit('codeIndented')
- // To do: allow interrupting like `markdown-rs`.
- // Feel free to interrupt.
- // tokenizer.interrupt = false
- return ok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeFurtherStart(effects, ok, nok) {
- const self = this
- return furtherStart
-
- /**
- * At eol, trying to parse another indent.
- *
- * ```markdown
- * > | aaa
- * ^
- * | bbb
- * ```
- *
- * @type {State}
- */
- function furtherStart(code) {
- // To do: improve `lazy` / `pierce` handling.
- // If this is a lazy line, it can’t be code.
- if (self.parser.lazy[self.now().line]) {
- return nok(code)
- }
- if (markdownLineEnding(code)) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return furtherStart
- }
-
- // To do: the code here in `micromark-js` is a bit different from
- // `markdown-rs` because there it can attempt spaces.
- // We can’t yet.
- //
- // To do: use an improved `space_or_tab` function like `markdown-rs`,
- // so that we can drop the next state.
- return factorySpace(effects, afterPrefix, 'linePrefix', 4 + 1)(code)
- }
-
- /**
- * At start, after 1 or 4 spaces.
- *
- * ```markdown
- * > | aaa
- * ^
- * ```
- *
- * @type {State}
- */
- function afterPrefix(code) {
- const tail = self.events[self.events.length - 1]
- return tail &&
- tail[1].type === 'linePrefix' &&
- tail[2].sliceSerialize(tail[1], true).length >= 4
- ? ok(code)
- : markdownLineEnding(code)
- ? furtherStart(code)
- : nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/heading-atx.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-/** @type {Construct} */
-const headingAtx = {
- name: 'headingAtx',
- tokenize: tokenizeHeadingAtx,
- resolve: resolveHeadingAtx
-}
-
-/** @type {Resolver} */
-function resolveHeadingAtx(events, context) {
- let contentEnd = events.length - 2
- let contentStart = 3
- /** @type {Token} */
- let content
- /** @type {Token} */
- let text
-
- // Prefix whitespace, part of the opening.
- if (events[contentStart][1].type === 'whitespace') {
- contentStart += 2
- }
-
- // Suffix whitespace, part of the closing.
- if (
- contentEnd - 2 > contentStart &&
- events[contentEnd][1].type === 'whitespace'
- ) {
- contentEnd -= 2
- }
- if (
- events[contentEnd][1].type === 'atxHeadingSequence' &&
- (contentStart === contentEnd - 1 ||
- (contentEnd - 4 > contentStart &&
- events[contentEnd - 2][1].type === 'whitespace'))
- ) {
- contentEnd -= contentStart + 1 === contentEnd ? 2 : 4
- }
- if (contentEnd > contentStart) {
- content = {
- type: 'atxHeadingText',
- start: events[contentStart][1].start,
- end: events[contentEnd][1].end
- }
- text = {
- type: 'chunkText',
- start: events[contentStart][1].start,
- end: events[contentEnd][1].end,
- contentType: 'text'
- }
- splice(events, contentStart, contentEnd - contentStart + 1, [
- ['enter', content, context],
- ['enter', text, context],
- ['exit', text, context],
- ['exit', content, context]
- ])
- }
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeHeadingAtx(effects, ok, nok) {
- let size = 0
- return start
-
- /**
- * Start of a heading (atx).
- *
- * ```markdown
- * > | ## aa
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // To do: parse indent like `markdown-rs`.
- effects.enter('atxHeading')
- return before(code)
- }
-
- /**
- * After optional whitespace, at `#`.
- *
- * ```markdown
- * > | ## aa
- * ^
- * ```
- *
- * @type {State}
- */
- function before(code) {
- effects.enter('atxHeadingSequence')
- return sequenceOpen(code)
- }
-
- /**
- * In opening sequence.
- *
- * ```markdown
- * > | ## aa
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceOpen(code) {
- if (code === 35 && size++ < 6) {
- effects.consume(code)
- return sequenceOpen
- }
-
- // Always at least one `#`.
- if (code === null || markdownLineEndingOrSpace(code)) {
- effects.exit('atxHeadingSequence')
- return atBreak(code)
- }
- return nok(code)
- }
-
- /**
- * After something, before something else.
- *
- * ```markdown
- * > | ## aa
- * ^
- * ```
- *
- * @type {State}
- */
- function atBreak(code) {
- if (code === 35) {
- effects.enter('atxHeadingSequence')
- return sequenceFurther(code)
- }
- if (code === null || markdownLineEnding(code)) {
- effects.exit('atxHeading')
- // To do: interrupt like `markdown-rs`.
- // // Feel free to interrupt.
- // tokenizer.interrupt = false
- return ok(code)
- }
- if (markdownSpace(code)) {
- return factorySpace(effects, atBreak, 'whitespace')(code)
- }
-
- // To do: generate `data` tokens, add the `text` token later.
- // Needs edit map, see: `markdown.rs`.
- effects.enter('atxHeadingText')
- return data(code)
- }
-
- /**
- * In further sequence (after whitespace).
- *
- * Could be normal “visible” hashes in the heading or a final sequence.
- *
- * ```markdown
- * > | ## aa ##
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceFurther(code) {
- if (code === 35) {
- effects.consume(code)
- return sequenceFurther
- }
- effects.exit('atxHeadingSequence')
- return atBreak(code)
- }
-
- /**
- * In text.
- *
- * ```markdown
- * > | ## aa
- * ^
- * ```
- *
- * @type {State}
- */
- function data(code) {
- if (code === null || code === 35 || markdownLineEndingOrSpace(code)) {
- effects.exit('atxHeadingText')
- return atBreak(code)
- }
- effects.consume(code)
- return data
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/setext-underline.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const setextUnderline = {
- name: 'setextUnderline',
- tokenize: tokenizeSetextUnderline,
- resolveTo: resolveToSetextUnderline
-}
-
-/** @type {Resolver} */
-function resolveToSetextUnderline(events, context) {
- // To do: resolve like `markdown-rs`.
- let index = events.length
- /** @type {number | undefined} */
- let content
- /** @type {number | undefined} */
- let text
- /** @type {number | undefined} */
- let definition
-
- // Find the opening of the content.
- // It’ll always exist: we don’t tokenize if it isn’t there.
- while (index--) {
- if (events[index][0] === 'enter') {
- if (events[index][1].type === 'content') {
- content = index
- break
- }
- if (events[index][1].type === 'paragraph') {
- text = index
- }
- }
- // Exit
- else {
- if (events[index][1].type === 'content') {
- // Remove the content end (if needed we’ll add it later)
- events.splice(index, 1)
- }
- if (!definition && events[index][1].type === 'definition') {
- definition = index
- }
- }
- }
- const heading = {
- type: 'setextHeading',
- start: Object.assign({}, events[text][1].start),
- end: Object.assign({}, events[events.length - 1][1].end)
- }
-
- // Change the paragraph to setext heading text.
- events[text][1].type = 'setextHeadingText'
-
- // If we have definitions in the content, we’ll keep on having content,
- // but we need move it.
- if (definition) {
- events.splice(text, 0, ['enter', heading, context])
- events.splice(definition + 1, 0, ['exit', events[content][1], context])
- events[content][1].end = Object.assign({}, events[definition][1].end)
- } else {
- events[content][1] = heading
- }
-
- // Add the heading exit at the end.
- events.push(['exit', heading, context])
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeSetextUnderline(effects, ok, nok) {
- const self = this
- /** @type {NonNullable<Code>} */
- let marker
- return start
-
- /**
- * At start of heading (setext) underline.
- *
- * ```markdown
- * | aa
- * > | ==
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- let index = self.events.length
- /** @type {boolean | undefined} */
- let paragraph
- // Find an opening.
- while (index--) {
- // Skip enter/exit of line ending, line prefix, and content.
- // We can now either have a definition or a paragraph.
- if (
- self.events[index][1].type !== 'lineEnding' &&
- self.events[index][1].type !== 'linePrefix' &&
- self.events[index][1].type !== 'content'
- ) {
- paragraph = self.events[index][1].type === 'paragraph'
- break
- }
- }
-
- // To do: handle lazy/pierce like `markdown-rs`.
- // To do: parse indent like `markdown-rs`.
- if (!self.parser.lazy[self.now().line] && (self.interrupt || paragraph)) {
- effects.enter('setextHeadingLine')
- marker = code
- return before(code)
- }
- return nok(code)
- }
-
- /**
- * After optional whitespace, at `-` or `=`.
- *
- * ```markdown
- * | aa
- * > | ==
- * ^
- * ```
- *
- * @type {State}
- */
- function before(code) {
- effects.enter('setextHeadingLineSequence')
- return inside(code)
- }
-
- /**
- * In sequence.
- *
- * ```markdown
- * | aa
- * > | ==
- * ^
- * ```
- *
- * @type {State}
- */
- function inside(code) {
- if (code === marker) {
- effects.consume(code)
- return inside
- }
- effects.exit('setextHeadingLineSequence')
- return markdownSpace(code)
- ? factorySpace(effects, after, 'lineSuffix')(code)
- : after(code)
- }
-
- /**
- * After sequence, after optional whitespace.
- *
- * ```markdown
- * | aa
- * > | ==
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('setextHeadingLine')
- return ok(code)
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-html-tag-name/index.js
-/**
- * List of lowercase HTML “block” tag names.
- *
- * The list, when parsing HTML (flow), results in more relaxed rules (condition
- * 6).
- * Because they are known blocks, the HTML-like syntax doesn’t have to be
- * strictly parsed.
- * For tag names not in this list, a more strict algorithm (condition 7) is used
- * to detect whether the HTML-like syntax is seen as HTML (flow) or not.
- *
- * This is copied from:
- * <https://spec.commonmark.org/0.30/#html-blocks>.
- *
- * > 👉 **Note**: `search` was added in `CommonMark@0.31`.
- */
-const htmlBlockNames = [
- 'address',
- 'article',
- 'aside',
- 'base',
- 'basefont',
- 'blockquote',
- 'body',
- 'caption',
- 'center',
- 'col',
- 'colgroup',
- 'dd',
- 'details',
- 'dialog',
- 'dir',
- 'div',
- 'dl',
- 'dt',
- 'fieldset',
- 'figcaption',
- 'figure',
- 'footer',
- 'form',
- 'frame',
- 'frameset',
- 'h1',
- 'h2',
- 'h3',
- 'h4',
- 'h5',
- 'h6',
- 'head',
- 'header',
- 'hr',
- 'html',
- 'iframe',
- 'legend',
- 'li',
- 'link',
- 'main',
- 'menu',
- 'menuitem',
- 'nav',
- 'noframes',
- 'ol',
- 'optgroup',
- 'option',
- 'p',
- 'param',
- 'search',
- 'section',
- 'summary',
- 'table',
- 'tbody',
- 'td',
- 'tfoot',
- 'th',
- 'thead',
- 'title',
- 'tr',
- 'track',
- 'ul'
-]
-
-/**
- * List of lowercase HTML “raw” tag names.
- *
- * The list, when parsing HTML (flow), results in HTML that can include lines
- * without exiting, until a closing tag also in this list is found (condition
- * 1).
- *
- * This module is copied from:
- * <https://spec.commonmark.org/0.30/#html-blocks>.
- *
- * > 👉 **Note**: `textarea` was added in `CommonMark@0.30`.
- */
-const htmlRawNames = ['pre', 'script', 'style', 'textarea']
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/html-flow.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-
-/** @type {Construct} */
-const htmlFlow = {
- name: 'htmlFlow',
- tokenize: tokenizeHtmlFlow,
- resolveTo: resolveToHtmlFlow,
- concrete: true
-}
-
-/** @type {Construct} */
-const blankLineBefore = {
- tokenize: tokenizeBlankLineBefore,
- partial: true
-}
-const nonLazyContinuationStart = {
- tokenize: tokenizeNonLazyContinuationStart,
- partial: true
-}
-
-/** @type {Resolver} */
-function resolveToHtmlFlow(events) {
- let index = events.length
- while (index--) {
- if (events[index][0] === 'enter' && events[index][1].type === 'htmlFlow') {
- break
- }
- }
- if (index > 1 && events[index - 2][1].type === 'linePrefix') {
- // Add the prefix start to the HTML token.
- events[index][1].start = events[index - 2][1].start
- // Add the prefix start to the HTML line token.
- events[index + 1][1].start = events[index - 2][1].start
- // Remove the line prefix.
- events.splice(index - 2, 2)
- }
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeHtmlFlow(effects, ok, nok) {
- const self = this
- /** @type {number} */
- let marker
- /** @type {boolean} */
- let closingTag
- /** @type {string} */
- let buffer
- /** @type {number} */
- let index
- /** @type {Code} */
- let markerB
- return start
-
- /**
- * Start of HTML (flow).
- *
- * ```markdown
- * > | <x />
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // To do: parse indent like `markdown-rs`.
- return before(code)
- }
-
- /**
- * At `<`, after optional whitespace.
- *
- * ```markdown
- * > | <x />
- * ^
- * ```
- *
- * @type {State}
- */
- function before(code) {
- effects.enter('htmlFlow')
- effects.enter('htmlFlowData')
- effects.consume(code)
- return open
- }
-
- /**
- * After `<`, at tag name or other stuff.
- *
- * ```markdown
- * > | <x />
- * ^
- * > | <!doctype>
- * ^
- * > | <!--xxx-->
- * ^
- * ```
- *
- * @type {State}
- */
- function open(code) {
- if (code === 33) {
- effects.consume(code)
- return declarationOpen
- }
- if (code === 47) {
- effects.consume(code)
- closingTag = true
- return tagCloseStart
- }
- if (code === 63) {
- effects.consume(code)
- marker = 3
- // To do:
- // tokenizer.concrete = true
- // To do: use `markdown-rs` style interrupt.
- // While we’re in an instruction instead of a declaration, we’re on a `?`
- // right now, so we do need to search for `>`, similar to declarations.
- return self.interrupt ? ok : continuationDeclarationInside
- }
-
- // ASCII alphabetical.
- if (asciiAlpha(code)) {
- effects.consume(code)
- // @ts-expect-error: not null.
- buffer = String.fromCharCode(code)
- return tagName
- }
- return nok(code)
- }
-
- /**
- * After `<!`, at declaration, comment, or CDATA.
- *
- * ```markdown
- * > | <!doctype>
- * ^
- * > | <!--xxx-->
- * ^
- * > | <![CDATA[>&<]]>
- * ^
- * ```
- *
- * @type {State}
- */
- function declarationOpen(code) {
- if (code === 45) {
- effects.consume(code)
- marker = 2
- return commentOpenInside
- }
- if (code === 91) {
- effects.consume(code)
- marker = 5
- index = 0
- return cdataOpenInside
- }
-
- // ASCII alphabetical.
- if (asciiAlpha(code)) {
- effects.consume(code)
- marker = 4
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok : continuationDeclarationInside
- }
- return nok(code)
- }
-
- /**
- * After `<!-`, inside a comment, at another `-`.
- *
- * ```markdown
- * > | <!--xxx-->
- * ^
- * ```
- *
- * @type {State}
- */
- function commentOpenInside(code) {
- if (code === 45) {
- effects.consume(code)
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok : continuationDeclarationInside
- }
- return nok(code)
- }
-
- /**
- * After `<![`, inside CDATA, expecting `CDATA[`.
- *
- * ```markdown
- * > | <![CDATA[>&<]]>
- * ^^^^^^
- * ```
- *
- * @type {State}
- */
- function cdataOpenInside(code) {
- const value = 'CDATA['
- if (code === value.charCodeAt(index++)) {
- effects.consume(code)
- if (index === value.length) {
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok : continuation
- }
- return cdataOpenInside
- }
- return nok(code)
- }
-
- /**
- * After `</`, in closing tag, at tag name.
- *
- * ```markdown
- * > | </x>
- * ^
- * ```
- *
- * @type {State}
- */
- function tagCloseStart(code) {
- if (asciiAlpha(code)) {
- effects.consume(code)
- // @ts-expect-error: not null.
- buffer = String.fromCharCode(code)
- return tagName
- }
- return nok(code)
- }
-
- /**
- * In tag name.
- *
- * ```markdown
- * > | <ab>
- * ^^
- * > | </ab>
- * ^^
- * ```
- *
- * @type {State}
- */
- function tagName(code) {
- if (
- code === null ||
- code === 47 ||
- code === 62 ||
- markdownLineEndingOrSpace(code)
- ) {
- const slash = code === 47
- const name = buffer.toLowerCase()
- if (!slash && !closingTag && htmlRawNames.includes(name)) {
- marker = 1
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok(code) : continuation(code)
- }
- if (htmlBlockNames.includes(buffer.toLowerCase())) {
- marker = 6
- if (slash) {
- effects.consume(code)
- return basicSelfClosing
- }
-
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok(code) : continuation(code)
- }
- marker = 7
- // Do not support complete HTML when interrupting.
- return self.interrupt && !self.parser.lazy[self.now().line]
- ? nok(code)
- : closingTag
- ? completeClosingTagAfter(code)
- : completeAttributeNameBefore(code)
- }
-
- // ASCII alphanumerical and `-`.
- if (code === 45 || asciiAlphanumeric(code)) {
- effects.consume(code)
- buffer += String.fromCharCode(code)
- return tagName
- }
- return nok(code)
- }
-
- /**
- * After closing slash of a basic tag name.
- *
- * ```markdown
- * > | <div/>
- * ^
- * ```
- *
- * @type {State}
- */
- function basicSelfClosing(code) {
- if (code === 62) {
- effects.consume(code)
- // // Do not form containers.
- // tokenizer.concrete = true
- return self.interrupt ? ok : continuation
- }
- return nok(code)
- }
-
- /**
- * After closing slash of a complete tag name.
- *
- * ```markdown
- * > | <x/>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeClosingTagAfter(code) {
- if (markdownSpace(code)) {
- effects.consume(code)
- return completeClosingTagAfter
- }
- return completeEnd(code)
- }
-
- /**
- * At an attribute name.
- *
- * At first, this state is used after a complete tag name, after whitespace,
- * where it expects optional attributes or the end of the tag.
- * It is also reused after attributes, when expecting more optional
- * attributes.
- *
- * ```markdown
- * > | <a />
- * ^
- * > | <a :b>
- * ^
- * > | <a _b>
- * ^
- * > | <a b>
- * ^
- * > | <a >
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeNameBefore(code) {
- if (code === 47) {
- effects.consume(code)
- return completeEnd
- }
-
- // ASCII alphanumerical and `:` and `_`.
- if (code === 58 || code === 95 || asciiAlpha(code)) {
- effects.consume(code)
- return completeAttributeName
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return completeAttributeNameBefore
- }
- return completeEnd(code)
- }
-
- /**
- * In attribute name.
- *
- * ```markdown
- * > | <a :b>
- * ^
- * > | <a _b>
- * ^
- * > | <a b>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeName(code) {
- // ASCII alphanumerical and `-`, `.`, `:`, and `_`.
- if (
- code === 45 ||
- code === 46 ||
- code === 58 ||
- code === 95 ||
- asciiAlphanumeric(code)
- ) {
- effects.consume(code)
- return completeAttributeName
- }
- return completeAttributeNameAfter(code)
- }
-
- /**
- * After attribute name, at an optional initializer, the end of the tag, or
- * whitespace.
- *
- * ```markdown
- * > | <a b>
- * ^
- * > | <a b=c>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeNameAfter(code) {
- if (code === 61) {
- effects.consume(code)
- return completeAttributeValueBefore
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return completeAttributeNameAfter
- }
- return completeAttributeNameBefore(code)
- }
-
- /**
- * Before unquoted, double quoted, or single quoted attribute value, allowing
- * whitespace.
- *
- * ```markdown
- * > | <a b=c>
- * ^
- * > | <a b="c">
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeValueBefore(code) {
- if (
- code === null ||
- code === 60 ||
- code === 61 ||
- code === 62 ||
- code === 96
- ) {
- return nok(code)
- }
- if (code === 34 || code === 39) {
- effects.consume(code)
- markerB = code
- return completeAttributeValueQuoted
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return completeAttributeValueBefore
- }
- return completeAttributeValueUnquoted(code)
- }
-
- /**
- * In double or single quoted attribute value.
- *
- * ```markdown
- * > | <a b="c">
- * ^
- * > | <a b='c'>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeValueQuoted(code) {
- if (code === markerB) {
- effects.consume(code)
- markerB = null
- return completeAttributeValueQuotedAfter
- }
- if (code === null || markdownLineEnding(code)) {
- return nok(code)
- }
- effects.consume(code)
- return completeAttributeValueQuoted
- }
-
- /**
- * In unquoted attribute value.
- *
- * ```markdown
- * > | <a b=c>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeValueUnquoted(code) {
- if (
- code === null ||
- code === 34 ||
- code === 39 ||
- code === 47 ||
- code === 60 ||
- code === 61 ||
- code === 62 ||
- code === 96 ||
- markdownLineEndingOrSpace(code)
- ) {
- return completeAttributeNameAfter(code)
- }
- effects.consume(code)
- return completeAttributeValueUnquoted
- }
-
- /**
- * After double or single quoted attribute value, before whitespace or the
- * end of the tag.
- *
- * ```markdown
- * > | <a b="c">
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAttributeValueQuotedAfter(code) {
- if (code === 47 || code === 62 || markdownSpace(code)) {
- return completeAttributeNameBefore(code)
- }
- return nok(code)
- }
-
- /**
- * In certain circumstances of a complete tag where only an `>` is allowed.
- *
- * ```markdown
- * > | <a b="c">
- * ^
- * ```
- *
- * @type {State}
- */
- function completeEnd(code) {
- if (code === 62) {
- effects.consume(code)
- return completeAfter
- }
- return nok(code)
- }
-
- /**
- * After `>` in a complete tag.
- *
- * ```markdown
- * > | <x>
- * ^
- * ```
- *
- * @type {State}
- */
- function completeAfter(code) {
- if (code === null || markdownLineEnding(code)) {
- // // Do not form containers.
- // tokenizer.concrete = true
- return continuation(code)
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return completeAfter
- }
- return nok(code)
- }
-
- /**
- * In continuation of any HTML kind.
- *
- * ```markdown
- * > | <!--xxx-->
- * ^
- * ```
- *
- * @type {State}
- */
- function continuation(code) {
- if (code === 45 && marker === 2) {
- effects.consume(code)
- return continuationCommentInside
- }
- if (code === 60 && marker === 1) {
- effects.consume(code)
- return continuationRawTagOpen
- }
- if (code === 62 && marker === 4) {
- effects.consume(code)
- return continuationClose
- }
- if (code === 63 && marker === 3) {
- effects.consume(code)
- return continuationDeclarationInside
- }
- if (code === 93 && marker === 5) {
- effects.consume(code)
- return continuationCdataInside
- }
- if (markdownLineEnding(code) && (marker === 6 || marker === 7)) {
- effects.exit('htmlFlowData')
- return effects.check(
- blankLineBefore,
- continuationAfter,
- continuationStart
- )(code)
- }
- if (code === null || markdownLineEnding(code)) {
- effects.exit('htmlFlowData')
- return continuationStart(code)
- }
- effects.consume(code)
- return continuation
- }
-
- /**
- * In continuation, at eol.
- *
- * ```markdown
- * > | <x>
- * ^
- * | asd
- * ```
- *
- * @type {State}
- */
- function continuationStart(code) {
- return effects.check(
- nonLazyContinuationStart,
- continuationStartNonLazy,
- continuationAfter
- )(code)
- }
-
- /**
- * In continuation, at eol, before non-lazy content.
- *
- * ```markdown
- * > | <x>
- * ^
- * | asd
- * ```
- *
- * @type {State}
- */
- function continuationStartNonLazy(code) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return continuationBefore
- }
-
- /**
- * In continuation, before non-lazy content.
- *
- * ```markdown
- * | <x>
- * > | asd
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationBefore(code) {
- if (code === null || markdownLineEnding(code)) {
- return continuationStart(code)
- }
- effects.enter('htmlFlowData')
- return continuation(code)
- }
-
- /**
- * In comment continuation, after one `-`, expecting another.
- *
- * ```markdown
- * > | <!--xxx-->
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationCommentInside(code) {
- if (code === 45) {
- effects.consume(code)
- return continuationDeclarationInside
- }
- return continuation(code)
- }
-
- /**
- * In raw continuation, after `<`, at `/`.
- *
- * ```markdown
- * > | <script>console.log(1)</script>
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationRawTagOpen(code) {
- if (code === 47) {
- effects.consume(code)
- buffer = ''
- return continuationRawEndTag
- }
- return continuation(code)
- }
-
- /**
- * In raw continuation, after `</`, in a raw tag name.
- *
- * ```markdown
- * > | <script>console.log(1)</script>
- * ^^^^^^
- * ```
- *
- * @type {State}
- */
- function continuationRawEndTag(code) {
- if (code === 62) {
- const name = buffer.toLowerCase()
- if (htmlRawNames.includes(name)) {
- effects.consume(code)
- return continuationClose
- }
- return continuation(code)
- }
- if (asciiAlpha(code) && buffer.length < 8) {
- effects.consume(code)
- // @ts-expect-error: not null.
- buffer += String.fromCharCode(code)
- return continuationRawEndTag
- }
- return continuation(code)
- }
-
- /**
- * In cdata continuation, after `]`, expecting `]>`.
- *
- * ```markdown
- * > | <![CDATA[>&<]]>
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationCdataInside(code) {
- if (code === 93) {
- effects.consume(code)
- return continuationDeclarationInside
- }
- return continuation(code)
- }
-
- /**
- * In declaration or instruction continuation, at `>`.
- *
- * ```markdown
- * > | <!-->
- * ^
- * > | <?>
- * ^
- * > | <!q>
- * ^
- * > | <!--ab-->
- * ^
- * > | <![CDATA[>&<]]>
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationDeclarationInside(code) {
- if (code === 62) {
- effects.consume(code)
- return continuationClose
- }
-
- // More dashes.
- if (code === 45 && marker === 2) {
- effects.consume(code)
- return continuationDeclarationInside
- }
- return continuation(code)
- }
-
- /**
- * In closed continuation: everything we get until the eol/eof is part of it.
- *
- * ```markdown
- * > | <!doctype>
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationClose(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('htmlFlowData')
- return continuationAfter(code)
- }
- effects.consume(code)
- return continuationClose
- }
-
- /**
- * Done.
- *
- * ```markdown
- * > | <!doctype>
- * ^
- * ```
- *
- * @type {State}
- */
- function continuationAfter(code) {
- effects.exit('htmlFlow')
- // // Feel free to interrupt.
- // tokenizer.interrupt = false
- // // No longer concrete.
- // tokenizer.concrete = false
- return ok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeNonLazyContinuationStart(effects, ok, nok) {
- const self = this
- return start
-
- /**
- * At eol, before continuation.
- *
- * ```markdown
- * > | * ```js
- * ^
- * | b
- * ```
- *
- * @type {State}
- */
- function start(code) {
- if (markdownLineEnding(code)) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return after
- }
- return nok(code)
- }
-
- /**
- * A continuation.
- *
- * ```markdown
- * | * ```js
- * > | b
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- return self.parser.lazy[self.now().line] ? nok(code) : ok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeBlankLineBefore(effects, ok, nok) {
- return start
-
- /**
- * Before eol, expecting blank line.
- *
- * ```markdown
- * > | <div>
- * ^
- * |
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return effects.attempt(blankLine, ok, nok)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/code-fenced.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const nonLazyContinuation = {
- tokenize: tokenizeNonLazyContinuation,
- partial: true
-}
-
-/** @type {Construct} */
-const codeFenced = {
- name: 'codeFenced',
- tokenize: tokenizeCodeFenced,
- concrete: true
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeCodeFenced(effects, ok, nok) {
- const self = this
- /** @type {Construct} */
- const closeStart = {
- tokenize: tokenizeCloseStart,
- partial: true
- }
- let initialPrefix = 0
- let sizeOpen = 0
- /** @type {NonNullable<Code>} */
- let marker
- return start
-
- /**
- * Start of code.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // To do: parse whitespace like `markdown-rs`.
- return beforeSequenceOpen(code)
- }
-
- /**
- * In opening fence, after prefix, at sequence.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function beforeSequenceOpen(code) {
- const tail = self.events[self.events.length - 1]
- initialPrefix =
- tail && tail[1].type === 'linePrefix'
- ? tail[2].sliceSerialize(tail[1], true).length
- : 0
- marker = code
- effects.enter('codeFenced')
- effects.enter('codeFencedFence')
- effects.enter('codeFencedFenceSequence')
- return sequenceOpen(code)
- }
-
- /**
- * In opening fence sequence.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function sequenceOpen(code) {
- if (code === marker) {
- sizeOpen++
- effects.consume(code)
- return sequenceOpen
- }
- if (sizeOpen < 3) {
- return nok(code)
- }
- effects.exit('codeFencedFenceSequence')
- return markdownSpace(code)
- ? factorySpace(effects, infoBefore, 'whitespace')(code)
- : infoBefore(code)
- }
-
- /**
- * In opening fence, after the sequence (and optional whitespace), before info.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function infoBefore(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('codeFencedFence')
- return self.interrupt
- ? ok(code)
- : effects.check(nonLazyContinuation, atNonLazyBreak, after)(code)
- }
- effects.enter('codeFencedFenceInfo')
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return info(code)
- }
-
- /**
- * In info.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function info(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('chunkString')
- effects.exit('codeFencedFenceInfo')
- return infoBefore(code)
- }
- if (markdownSpace(code)) {
- effects.exit('chunkString')
- effects.exit('codeFencedFenceInfo')
- return factorySpace(effects, metaBefore, 'whitespace')(code)
- }
- if (code === 96 && code === marker) {
- return nok(code)
- }
- effects.consume(code)
- return info
- }
-
- /**
- * In opening fence, after info and whitespace, before meta.
- *
- * ```markdown
- * > | ~~~js eval
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function metaBefore(code) {
- if (code === null || markdownLineEnding(code)) {
- return infoBefore(code)
- }
- effects.enter('codeFencedFenceMeta')
- effects.enter('chunkString', {
- contentType: 'string'
- })
- return meta(code)
- }
-
- /**
- * In meta.
- *
- * ```markdown
- * > | ~~~js eval
- * ^
- * | alert(1)
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function meta(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('chunkString')
- effects.exit('codeFencedFenceMeta')
- return infoBefore(code)
- }
- if (code === 96 && code === marker) {
- return nok(code)
- }
- effects.consume(code)
- return meta
- }
-
- /**
- * At eol/eof in code, before a non-lazy closing fence or content.
- *
- * ```markdown
- * > | ~~~js
- * ^
- * > | alert(1)
- * ^
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function atNonLazyBreak(code) {
- return effects.attempt(closeStart, after, contentBefore)(code)
- }
-
- /**
- * Before code content, not a closing fence, at eol.
- *
- * ```markdown
- * | ~~~js
- * > | alert(1)
- * ^
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function contentBefore(code) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return contentStart
- }
-
- /**
- * Before code content, not a closing fence.
- *
- * ```markdown
- * | ~~~js
- * > | alert(1)
- * ^
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function contentStart(code) {
- return initialPrefix > 0 && markdownSpace(code)
- ? factorySpace(
- effects,
- beforeContentChunk,
- 'linePrefix',
- initialPrefix + 1
- )(code)
- : beforeContentChunk(code)
- }
-
- /**
- * Before code content, after optional prefix.
- *
- * ```markdown
- * | ~~~js
- * > | alert(1)
- * ^
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function beforeContentChunk(code) {
- if (code === null || markdownLineEnding(code)) {
- return effects.check(nonLazyContinuation, atNonLazyBreak, after)(code)
- }
- effects.enter('codeFlowValue')
- return contentChunk(code)
- }
-
- /**
- * In code content.
- *
- * ```markdown
- * | ~~~js
- * > | alert(1)
- * ^^^^^^^^
- * | ~~~
- * ```
- *
- * @type {State}
- */
- function contentChunk(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('codeFlowValue')
- return beforeContentChunk(code)
- }
- effects.consume(code)
- return contentChunk
- }
-
- /**
- * After code.
- *
- * ```markdown
- * | ~~~js
- * | alert(1)
- * > | ~~~
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- effects.exit('codeFenced')
- return ok(code)
- }
-
- /**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
- function tokenizeCloseStart(effects, ok, nok) {
- let size = 0
- return startBefore
-
- /**
- *
- *
- * @type {State}
- */
- function startBefore(code) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return start
- }
-
- /**
- * Before closing fence, at optional whitespace.
- *
- * ```markdown
- * | ~~~js
- * | alert(1)
- * > | ~~~
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // Always populated by defaults.
-
- // To do: `enter` here or in next state?
- effects.enter('codeFencedFence')
- return markdownSpace(code)
- ? factorySpace(
- effects,
- beforeSequenceClose,
- 'linePrefix',
- self.parser.constructs.disable.null.includes('codeIndented')
- ? undefined
- : 4
- )(code)
- : beforeSequenceClose(code)
- }
-
- /**
- * In closing fence, after optional whitespace, at sequence.
- *
- * ```markdown
- * | ~~~js
- * | alert(1)
- * > | ~~~
- * ^
- * ```
- *
- * @type {State}
- */
- function beforeSequenceClose(code) {
- if (code === marker) {
- effects.enter('codeFencedFenceSequence')
- return sequenceClose(code)
- }
- return nok(code)
- }
-
- /**
- * In closing fence sequence.
- *
- * ```markdown
- * | ~~~js
- * | alert(1)
- * > | ~~~
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceClose(code) {
- if (code === marker) {
- size++
- effects.consume(code)
- return sequenceClose
- }
- if (size >= sizeOpen) {
- effects.exit('codeFencedFenceSequence')
- return markdownSpace(code)
- ? factorySpace(effects, sequenceCloseAfter, 'whitespace')(code)
- : sequenceCloseAfter(code)
- }
- return nok(code)
- }
-
- /**
- * After closing fence sequence, after optional whitespace.
- *
- * ```markdown
- * | ~~~js
- * | alert(1)
- * > | ~~~
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceCloseAfter(code) {
- if (code === null || markdownLineEnding(code)) {
- effects.exit('codeFencedFence')
- return ok(code)
- }
- return nok(code)
- }
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeNonLazyContinuation(effects, ok, nok) {
- const self = this
- return start
-
- /**
- *
- *
- * @type {State}
- */
- function start(code) {
- if (code === null) {
- return nok(code)
- }
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return lineStart
- }
-
- /**
- *
- *
- * @type {State}
- */
- function lineStart(code) {
- return self.parser.lazy[self.now().line] ? nok(code) : ok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/character-entities/index.js
-/**
- * Map of named character references.
- *
- * @type {Record<string, string>}
- */
-const characterEntities = {
- AElig: 'Æ',
- AMP: '&',
- Aacute: 'Á',
- Abreve: 'Ă',
- Acirc: 'Â',
- Acy: 'А',
- Afr: '𝔄',
- Agrave: 'À',
- Alpha: 'Α',
- Amacr: 'Ā',
- And: '⩓',
- Aogon: 'Ą',
- Aopf: '𝔸',
- ApplyFunction: '⁡',
- Aring: 'Å',
- Ascr: '𝒜',
- Assign: '≔',
- Atilde: 'Ã',
- Auml: 'Ä',
- Backslash: '∖',
- Barv: '⫧',
- Barwed: '⌆',
- Bcy: 'Б',
- Because: '∵',
- Bernoullis: 'ℬ',
- Beta: 'Β',
- Bfr: '𝔅',
- Bopf: '𝔹',
- Breve: '˘',
- Bscr: 'ℬ',
- Bumpeq: '≎',
- CHcy: 'Ч',
- COPY: '©',
- Cacute: 'Ć',
- Cap: '⋒',
- CapitalDifferentialD: 'ⅅ',
- Cayleys: 'ℭ',
- Ccaron: 'Č',
- Ccedil: 'Ç',
- Ccirc: 'Ĉ',
- Cconint: '∰',
- Cdot: 'Ċ',
- Cedilla: '¸',
- CenterDot: '·',
- Cfr: 'ℭ',
- Chi: 'Χ',
- CircleDot: '⊙',
- CircleMinus: '⊖',
- CirclePlus: '⊕',
- CircleTimes: '⊗',
- ClockwiseContourIntegral: '∲',
- CloseCurlyDoubleQuote: '”',
- CloseCurlyQuote: '’',
- Colon: '∷',
- Colone: '⩴',
- Congruent: '≡',
- Conint: '∯',
- ContourIntegral: '∮',
- Copf: 'ℂ',
- Coproduct: '∐',
- CounterClockwiseContourIntegral: '∳',
- Cross: '⨯',
- Cscr: '𝒞',
- Cup: '⋓',
- CupCap: '≍',
- DD: 'ⅅ',
- DDotrahd: '⤑',
- DJcy: 'Ђ',
- DScy: 'Ѕ',
- DZcy: 'Џ',
- Dagger: '‡',
- Darr: '↡',
- Dashv: '⫤',
- Dcaron: 'Ď',
- Dcy: 'Д',
- Del: '∇',
- Delta: 'Δ',
- Dfr: '𝔇',
- DiacriticalAcute: '´',
- DiacriticalDot: '˙',
- DiacriticalDoubleAcute: '˝',
- DiacriticalGrave: '`',
- DiacriticalTilde: '˜',
- Diamond: '⋄',
- DifferentialD: 'ⅆ',
- Dopf: '𝔻',
- Dot: '¨',
- DotDot: '⃜',
- DotEqual: '≐',
- DoubleContourIntegral: '∯',
- DoubleDot: '¨',
- DoubleDownArrow: '⇓',
- DoubleLeftArrow: '⇐',
- DoubleLeftRightArrow: '⇔',
- DoubleLeftTee: '⫤',
- DoubleLongLeftArrow: '⟸',
- DoubleLongLeftRightArrow: '⟺',
- DoubleLongRightArrow: '⟹',
- DoubleRightArrow: '⇒',
- DoubleRightTee: '⊨',
- DoubleUpArrow: '⇑',
- DoubleUpDownArrow: '⇕',
- DoubleVerticalBar: '∥',
- DownArrow: '↓',
- DownArrowBar: '⤓',
- DownArrowUpArrow: '⇵',
- DownBreve: '̑',
- DownLeftRightVector: '⥐',
- DownLeftTeeVector: '⥞',
- DownLeftVector: '↽',
- DownLeftVectorBar: '⥖',
- DownRightTeeVector: '⥟',
- DownRightVector: '⇁',
- DownRightVectorBar: '⥗',
- DownTee: '⊤',
- DownTeeArrow: '↧',
- Downarrow: '⇓',
- Dscr: '𝒟',
- Dstrok: 'Đ',
- ENG: 'Ŋ',
- ETH: 'Ð',
- Eacute: 'É',
- Ecaron: 'Ě',
- Ecirc: 'Ê',
- Ecy: 'Э',
- Edot: 'Ė',
- Efr: '𝔈',
- Egrave: 'È',
- Element: '∈',
- Emacr: 'Ē',
- EmptySmallSquare: '◻',
- EmptyVerySmallSquare: '▫',
- Eogon: 'Ę',
- Eopf: '𝔼',
- Epsilon: 'Ε',
- Equal: '⩵',
- EqualTilde: '≂',
- Equilibrium: '⇌',
- Escr: 'ℰ',
- Esim: '⩳',
- Eta: 'Η',
- Euml: 'Ë',
- Exists: '∃',
- ExponentialE: 'ⅇ',
- Fcy: 'Ф',
- Ffr: '𝔉',
- FilledSmallSquare: '◼',
- FilledVerySmallSquare: '▪',
- Fopf: '𝔽',
- ForAll: '∀',
- Fouriertrf: 'ℱ',
- Fscr: 'ℱ',
- GJcy: 'Ѓ',
- GT: '>',
- Gamma: 'Γ',
- Gammad: 'Ϝ',
- Gbreve: 'Ğ',
- Gcedil: 'Ģ',
- Gcirc: 'Ĝ',
- Gcy: 'Г',
- Gdot: 'Ġ',
- Gfr: '𝔊',
- Gg: '⋙',
- Gopf: '𝔾',
- GreaterEqual: '≥',
- GreaterEqualLess: '⋛',
- GreaterFullEqual: '≧',
- GreaterGreater: '⪢',
- GreaterLess: '≷',
- GreaterSlantEqual: '⩾',
- GreaterTilde: '≳',
- Gscr: '𝒢',
- Gt: '≫',
- HARDcy: 'Ъ',
- Hacek: 'ˇ',
- Hat: '^',
- Hcirc: 'Ĥ',
- Hfr: 'ℌ',
- HilbertSpace: 'ℋ',
- Hopf: 'ℍ',
- HorizontalLine: '─',
- Hscr: 'ℋ',
- Hstrok: 'Ħ',
- HumpDownHump: '≎',
- HumpEqual: '≏',
- IEcy: 'Е',
- IJlig: 'IJ',
- IOcy: 'Ё',
- Iacute: 'Í',
- Icirc: 'Î',
- Icy: 'И',
- Idot: 'İ',
- Ifr: 'ℑ',
- Igrave: 'Ì',
- Im: 'ℑ',
- Imacr: 'Ī',
- ImaginaryI: 'ⅈ',
- Implies: '⇒',
- Int: '∬',
- Integral: '∫',
- Intersection: '⋂',
- InvisibleComma: '⁣',
- InvisibleTimes: '⁢',
- Iogon: 'Į',
- Iopf: '𝕀',
- Iota: 'Ι',
- Iscr: 'ℐ',
- Itilde: 'Ĩ',
- Iukcy: 'І',
- Iuml: 'Ï',
- Jcirc: 'Ĵ',
- Jcy: 'Й',
- Jfr: '𝔍',
- Jopf: '𝕁',
- Jscr: '𝒥',
- Jsercy: 'Ј',
- Jukcy: 'Є',
- KHcy: 'Х',
- KJcy: 'Ќ',
- Kappa: 'Κ',
- Kcedil: 'Ķ',
- Kcy: 'К',
- Kfr: '𝔎',
- Kopf: '𝕂',
- Kscr: '𝒦',
- LJcy: 'Љ',
- LT: '<',
- Lacute: 'Ĺ',
- Lambda: 'Λ',
- Lang: '⟪',
- Laplacetrf: 'ℒ',
- Larr: '↞',
- Lcaron: 'Ľ',
- Lcedil: 'Ļ',
- Lcy: 'Л',
- LeftAngleBracket: '⟨',
- LeftArrow: '←',
- LeftArrowBar: '⇤',
- LeftArrowRightArrow: '⇆',
- LeftCeiling: '⌈',
- LeftDoubleBracket: '⟦',
- LeftDownTeeVector: '⥡',
- LeftDownVector: '⇃',
- LeftDownVectorBar: '⥙',
- LeftFloor: '⌊',
- LeftRightArrow: '↔',
- LeftRightVector: '⥎',
- LeftTee: '⊣',
- LeftTeeArrow: '↤',
- LeftTeeVector: '⥚',
- LeftTriangle: '⊲',
- LeftTriangleBar: '⧏',
- LeftTriangleEqual: '⊴',
- LeftUpDownVector: '⥑',
- LeftUpTeeVector: '⥠',
- LeftUpVector: '↿',
- LeftUpVectorBar: '⥘',
- LeftVector: '↼',
- LeftVectorBar: '⥒',
- Leftarrow: '⇐',
- Leftrightarrow: '⇔',
- LessEqualGreater: '⋚',
- LessFullEqual: '≦',
- LessGreater: '≶',
- LessLess: '⪡',
- LessSlantEqual: '⩽',
- LessTilde: '≲',
- Lfr: '𝔏',
- Ll: '⋘',
- Lleftarrow: '⇚',
- Lmidot: 'Ŀ',
- LongLeftArrow: '⟵',
- LongLeftRightArrow: '⟷',
- LongRightArrow: '⟶',
- Longleftarrow: '⟸',
- Longleftrightarrow: '⟺',
- Longrightarrow: '⟹',
- Lopf: '𝕃',
- LowerLeftArrow: '↙',
- LowerRightArrow: '↘',
- Lscr: 'ℒ',
- Lsh: '↰',
- Lstrok: 'Ł',
- Lt: '≪',
- Map: '⤅',
- Mcy: 'М',
- MediumSpace: ' ',
- Mellintrf: 'ℳ',
- Mfr: '𝔐',
- MinusPlus: '∓',
- Mopf: '𝕄',
- Mscr: 'ℳ',
- Mu: 'Μ',
- NJcy: 'Њ',
- Nacute: 'Ń',
- Ncaron: 'Ň',
- Ncedil: 'Ņ',
- Ncy: 'Н',
- NegativeMediumSpace: '​',
- NegativeThickSpace: '​',
- NegativeThinSpace: '​',
- NegativeVeryThinSpace: '​',
- NestedGreaterGreater: '≫',
- NestedLessLess: '≪',
- NewLine: '\n',
- Nfr: '𝔑',
- NoBreak: '⁠',
- NonBreakingSpace: ' ',
- Nopf: 'ℕ',
- Not: '⫬',
- NotCongruent: '≢',
- NotCupCap: '≭',
- NotDoubleVerticalBar: '∦',
- NotElement: '∉',
- NotEqual: '≠',
- NotEqualTilde: '≂̸',
- NotExists: '∄',
- NotGreater: '≯',
- NotGreaterEqual: '≱',
- NotGreaterFullEqual: '≧̸',
- NotGreaterGreater: '≫̸',
- NotGreaterLess: '≹',
- NotGreaterSlantEqual: '⩾̸',
- NotGreaterTilde: '≵',
- NotHumpDownHump: '≎̸',
- NotHumpEqual: '≏̸',
- NotLeftTriangle: '⋪',
- NotLeftTriangleBar: '⧏̸',
- NotLeftTriangleEqual: '⋬',
- NotLess: '≮',
- NotLessEqual: '≰',
- NotLessGreater: '≸',
- NotLessLess: '≪̸',
- NotLessSlantEqual: '⩽̸',
- NotLessTilde: '≴',
- NotNestedGreaterGreater: '⪢̸',
- NotNestedLessLess: '⪡̸',
- NotPrecedes: '⊀',
- NotPrecedesEqual: '⪯̸',
- NotPrecedesSlantEqual: '⋠',
- NotReverseElement: '∌',
- NotRightTriangle: '⋫',
- NotRightTriangleBar: '⧐̸',
- NotRightTriangleEqual: '⋭',
- NotSquareSubset: '⊏̸',
- NotSquareSubsetEqual: '⋢',
- NotSquareSuperset: '⊐̸',
- NotSquareSupersetEqual: '⋣',
- NotSubset: '⊂⃒',
- NotSubsetEqual: '⊈',
- NotSucceeds: '⊁',
- NotSucceedsEqual: '⪰̸',
- NotSucceedsSlantEqual: '⋡',
- NotSucceedsTilde: '≿̸',
- NotSuperset: '⊃⃒',
- NotSupersetEqual: '⊉',
- NotTilde: '≁',
- NotTildeEqual: '≄',
- NotTildeFullEqual: '≇',
- NotTildeTilde: '≉',
- NotVerticalBar: '∤',
- Nscr: '𝒩',
- Ntilde: 'Ñ',
- Nu: 'Ν',
- OElig: 'Œ',
- Oacute: 'Ó',
- Ocirc: 'Ô',
- Ocy: 'О',
- Odblac: 'Ő',
- Ofr: '𝔒',
- Ograve: 'Ò',
- Omacr: 'Ō',
- Omega: 'Ω',
- Omicron: 'Ο',
- Oopf: '𝕆',
- OpenCurlyDoubleQuote: '“',
- OpenCurlyQuote: '‘',
- Or: '⩔',
- Oscr: '𝒪',
- Oslash: 'Ø',
- Otilde: 'Õ',
- Otimes: '⨷',
- Ouml: 'Ö',
- OverBar: '‾',
- OverBrace: '⏞',
- OverBracket: '⎴',
- OverParenthesis: '⏜',
- PartialD: '∂',
- Pcy: 'П',
- Pfr: '𝔓',
- Phi: 'Φ',
- Pi: 'Π',
- PlusMinus: '±',
- Poincareplane: 'ℌ',
- Popf: 'ℙ',
- Pr: '⪻',
- Precedes: '≺',
- PrecedesEqual: '⪯',
- PrecedesSlantEqual: '≼',
- PrecedesTilde: '≾',
- Prime: '″',
- Product: '∏',
- Proportion: '∷',
- Proportional: '∝',
- Pscr: '𝒫',
- Psi: 'Ψ',
- QUOT: '"',
- Qfr: '𝔔',
- Qopf: 'ℚ',
- Qscr: '𝒬',
- RBarr: '⤐',
- REG: '®',
- Racute: 'Ŕ',
- Rang: '⟫',
- Rarr: '↠',
- Rarrtl: '⤖',
- Rcaron: 'Ř',
- Rcedil: 'Ŗ',
- Rcy: 'Р',
- Re: 'ℜ',
- ReverseElement: '∋',
- ReverseEquilibrium: '⇋',
- ReverseUpEquilibrium: '⥯',
- Rfr: 'ℜ',
- Rho: 'Ρ',
- RightAngleBracket: '⟩',
- RightArrow: '→',
- RightArrowBar: '⇥',
- RightArrowLeftArrow: '⇄',
- RightCeiling: '⌉',
- RightDoubleBracket: '⟧',
- RightDownTeeVector: '⥝',
- RightDownVector: '⇂',
- RightDownVectorBar: '⥕',
- RightFloor: '⌋',
- RightTee: '⊢',
- RightTeeArrow: '↦',
- RightTeeVector: '⥛',
- RightTriangle: '⊳',
- RightTriangleBar: '⧐',
- RightTriangleEqual: '⊵',
- RightUpDownVector: '⥏',
- RightUpTeeVector: '⥜',
- RightUpVector: '↾',
- RightUpVectorBar: '⥔',
- RightVector: '⇀',
- RightVectorBar: '⥓',
- Rightarrow: '⇒',
- Ropf: 'ℝ',
- RoundImplies: '⥰',
- Rrightarrow: '⇛',
- Rscr: 'ℛ',
- Rsh: '↱',
- RuleDelayed: '⧴',
- SHCHcy: 'Щ',
- SHcy: 'Ш',
- SOFTcy: 'Ь',
- Sacute: 'Ś',
- Sc: '⪼',
- Scaron: 'Š',
- Scedil: 'Ş',
- Scirc: 'Ŝ',
- Scy: 'С',
- Sfr: '𝔖',
- ShortDownArrow: '↓',
- ShortLeftArrow: '←',
- ShortRightArrow: '→',
- ShortUpArrow: '↑',
- Sigma: 'Σ',
- SmallCircle: '∘',
- Sopf: '𝕊',
- Sqrt: '√',
- Square: '□',
- SquareIntersection: '⊓',
- SquareSubset: '⊏',
- SquareSubsetEqual: '⊑',
- SquareSuperset: '⊐',
- SquareSupersetEqual: '⊒',
- SquareUnion: '⊔',
- Sscr: '𝒮',
- Star: '⋆',
- Sub: '⋐',
- Subset: '⋐',
- SubsetEqual: '⊆',
- Succeeds: '≻',
- SucceedsEqual: '⪰',
- SucceedsSlantEqual: '≽',
- SucceedsTilde: '≿',
- SuchThat: '∋',
- Sum: '∑',
- Sup: '⋑',
- Superset: '⊃',
- SupersetEqual: '⊇',
- Supset: '⋑',
- THORN: 'Þ',
- TRADE: '™',
- TSHcy: 'Ћ',
- TScy: 'Ц',
- Tab: '\t',
- Tau: 'Τ',
- Tcaron: 'Ť',
- Tcedil: 'Ţ',
- Tcy: 'Т',
- Tfr: '𝔗',
- Therefore: '∴',
- Theta: 'Θ',
- ThickSpace: '  ',
- ThinSpace: ' ',
- Tilde: '∼',
- TildeEqual: '≃',
- TildeFullEqual: '≅',
- TildeTilde: '≈',
- Topf: '𝕋',
- TripleDot: '⃛',
- Tscr: '𝒯',
- Tstrok: 'Ŧ',
- Uacute: 'Ú',
- Uarr: '↟',
- Uarrocir: '⥉',
- Ubrcy: 'Ў',
- Ubreve: 'Ŭ',
- Ucirc: 'Û',
- Ucy: 'У',
- Udblac: 'Ű',
- Ufr: '𝔘',
- Ugrave: 'Ù',
- Umacr: 'Ū',
- UnderBar: '_',
- UnderBrace: '⏟',
- UnderBracket: '⎵',
- UnderParenthesis: '⏝',
- Union: '⋃',
- UnionPlus: '⊎',
- Uogon: 'Ų',
- Uopf: '𝕌',
- UpArrow: '↑',
- UpArrowBar: '⤒',
- UpArrowDownArrow: '⇅',
- UpDownArrow: '↕',
- UpEquilibrium: '⥮',
- UpTee: '⊥',
- UpTeeArrow: '↥',
- Uparrow: '⇑',
- Updownarrow: '⇕',
- UpperLeftArrow: '↖',
- UpperRightArrow: '↗',
- Upsi: 'ϒ',
- Upsilon: 'Υ',
- Uring: 'Ů',
- Uscr: '𝒰',
- Utilde: 'Ũ',
- Uuml: 'Ü',
- VDash: '⊫',
- Vbar: '⫫',
- Vcy: 'В',
- Vdash: '⊩',
- Vdashl: '⫦',
- Vee: '⋁',
- Verbar: '‖',
- Vert: '‖',
- VerticalBar: '∣',
- VerticalLine: '|',
- VerticalSeparator: '❘',
- VerticalTilde: '≀',
- VeryThinSpace: ' ',
- Vfr: '𝔙',
- Vopf: '𝕍',
- Vscr: '𝒱',
- Vvdash: '⊪',
- Wcirc: 'Ŵ',
- Wedge: '⋀',
- Wfr: '𝔚',
- Wopf: '𝕎',
- Wscr: '𝒲',
- Xfr: '𝔛',
- Xi: 'Ξ',
- Xopf: '𝕏',
- Xscr: '𝒳',
- YAcy: 'Я',
- YIcy: 'Ї',
- YUcy: 'Ю',
- Yacute: 'Ý',
- Ycirc: 'Ŷ',
- Ycy: 'Ы',
- Yfr: '𝔜',
- Yopf: '𝕐',
- Yscr: '𝒴',
- Yuml: 'Ÿ',
- ZHcy: 'Ж',
- Zacute: 'Ź',
- Zcaron: 'Ž',
- Zcy: 'З',
- Zdot: 'Ż',
- ZeroWidthSpace: '​',
- Zeta: 'Ζ',
- Zfr: 'ℨ',
- Zopf: 'ℤ',
- Zscr: '𝒵',
- aacute: 'á',
- abreve: 'ă',
- ac: '∾',
- acE: '∾̳',
- acd: '∿',
- acirc: 'â',
- acute: '´',
- acy: 'а',
- aelig: 'æ',
- af: '⁡',
- afr: '𝔞',
- agrave: 'à',
- alefsym: 'ℵ',
- aleph: 'ℵ',
- alpha: 'α',
- amacr: 'ā',
- amalg: '⨿',
- amp: '&',
- and: '∧',
- andand: '⩕',
- andd: '⩜',
- andslope: '⩘',
- andv: '⩚',
- ang: '∠',
- ange: '⦤',
- angle: '∠',
- angmsd: '∡',
- angmsdaa: '⦨',
- angmsdab: '⦩',
- angmsdac: '⦪',
- angmsdad: '⦫',
- angmsdae: '⦬',
- angmsdaf: '⦭',
- angmsdag: '⦮',
- angmsdah: '⦯',
- angrt: '∟',
- angrtvb: '⊾',
- angrtvbd: '⦝',
- angsph: '∢',
- angst: 'Å',
- angzarr: '⍼',
- aogon: 'ą',
- aopf: '𝕒',
- ap: '≈',
- apE: '⩰',
- apacir: '⩯',
- ape: '≊',
- apid: '≋',
- apos: "'",
- approx: '≈',
- approxeq: '≊',
- aring: 'å',
- ascr: '𝒶',
- ast: '*',
- asymp: '≈',
- asympeq: '≍',
- atilde: 'ã',
- auml: 'ä',
- awconint: '∳',
- awint: '⨑',
- bNot: '⫭',
- backcong: '≌',
- backepsilon: '϶',
- backprime: '‵',
- backsim: '∽',
- backsimeq: '⋍',
- barvee: '⊽',
- barwed: '⌅',
- barwedge: '⌅',
- bbrk: '⎵',
- bbrktbrk: '⎶',
- bcong: '≌',
- bcy: 'б',
- bdquo: '„',
- becaus: '∵',
- because: '∵',
- bemptyv: '⦰',
- bepsi: '϶',
- bernou: 'ℬ',
- beta: 'β',
- beth: 'ℶ',
- between: '≬',
- bfr: '𝔟',
- bigcap: '⋂',
- bigcirc: '◯',
- bigcup: '⋃',
- bigodot: '⨀',
- bigoplus: '⨁',
- bigotimes: '⨂',
- bigsqcup: '⨆',
- bigstar: '★',
- bigtriangledown: '▽',
- bigtriangleup: '△',
- biguplus: '⨄',
- bigvee: '⋁',
- bigwedge: '⋀',
- bkarow: '⤍',
- blacklozenge: '⧫',
- blacksquare: '▪',
- blacktriangle: '▴',
- blacktriangledown: '▾',
- blacktriangleleft: '◂',
- blacktriangleright: '▸',
- blank: '␣',
- blk12: '▒',
- blk14: '░',
- blk34: '▓',
- block: '█',
- bne: '=⃥',
- bnequiv: '≡⃥',
- bnot: '⌐',
- bopf: '𝕓',
- bot: '⊥',
- bottom: '⊥',
- bowtie: '⋈',
- boxDL: '╗',
- boxDR: '╔',
- boxDl: '╖',
- boxDr: '╓',
- boxH: '═',
- boxHD: '╦',
- boxHU: '╩',
- boxHd: '╤',
- boxHu: '╧',
- boxUL: '╝',
- boxUR: '╚',
- boxUl: '╜',
- boxUr: '╙',
- boxV: '║',
- boxVH: '╬',
- boxVL: '╣',
- boxVR: '╠',
- boxVh: '╫',
- boxVl: '╢',
- boxVr: '╟',
- boxbox: '⧉',
- boxdL: '╕',
- boxdR: '╒',
- boxdl: '┐',
- boxdr: '┌',
- boxh: '─',
- boxhD: '╥',
- boxhU: '╨',
- boxhd: '┬',
- boxhu: '┴',
- boxminus: '⊟',
- boxplus: '⊞',
- boxtimes: '⊠',
- boxuL: '╛',
- boxuR: '╘',
- boxul: '┘',
- boxur: '└',
- boxv: '│',
- boxvH: '╪',
- boxvL: '╡',
- boxvR: '╞',
- boxvh: '┼',
- boxvl: '┤',
- boxvr: '├',
- bprime: '‵',
- breve: '˘',
- brvbar: '¦',
- bscr: '𝒷',
- bsemi: '⁏',
- bsim: '∽',
- bsime: '⋍',
- bsol: '\\',
- bsolb: '⧅',
- bsolhsub: '⟈',
- bull: '•',
- bullet: '•',
- bump: '≎',
- bumpE: '⪮',
- bumpe: '≏',
- bumpeq: '≏',
- cacute: 'ć',
- cap: '∩',
- capand: '⩄',
- capbrcup: '⩉',
- capcap: '⩋',
- capcup: '⩇',
- capdot: '⩀',
- caps: '∩︀',
- caret: '⁁',
- caron: 'ˇ',
- ccaps: '⩍',
- ccaron: 'č',
- ccedil: 'ç',
- ccirc: 'ĉ',
- ccups: '⩌',
- ccupssm: '⩐',
- cdot: 'ċ',
- cedil: '¸',
- cemptyv: '⦲',
- cent: '¢',
- centerdot: '·',
- cfr: '𝔠',
- chcy: 'ч',
- check: '✓',
- checkmark: '✓',
- chi: 'χ',
- cir: '○',
- cirE: '⧃',
- circ: 'ˆ',
- circeq: '≗',
- circlearrowleft: '↺',
- circlearrowright: '↻',
- circledR: '®',
- circledS: 'Ⓢ',
- circledast: '⊛',
- circledcirc: '⊚',
- circleddash: '⊝',
- cire: '≗',
- cirfnint: '⨐',
- cirmid: '⫯',
- cirscir: '⧂',
- clubs: '♣',
- clubsuit: '♣',
- colon: ':',
- colone: '≔',
- coloneq: '≔',
- comma: ',',
- commat: '@',
- comp: '∁',
- compfn: '∘',
- complement: '∁',
- complexes: 'ℂ',
- cong: '≅',
- congdot: '⩭',
- conint: '∮',
- copf: '𝕔',
- coprod: '∐',
- copy: '©',
- copysr: '℗',
- crarr: '↵',
- cross: '✗',
- cscr: '𝒸',
- csub: '⫏',
- csube: '⫑',
- csup: '⫐',
- csupe: '⫒',
- ctdot: '⋯',
- cudarrl: '⤸',
- cudarrr: '⤵',
- cuepr: '⋞',
- cuesc: '⋟',
- cularr: '↶',
- cularrp: '⤽',
- cup: '∪',
- cupbrcap: '⩈',
- cupcap: '⩆',
- cupcup: '⩊',
- cupdot: '⊍',
- cupor: '⩅',
- cups: '∪︀',
- curarr: '↷',
- curarrm: '⤼',
- curlyeqprec: '⋞',
- curlyeqsucc: '⋟',
- curlyvee: '⋎',
- curlywedge: '⋏',
- curren: '¤',
- curvearrowleft: '↶',
- curvearrowright: '↷',
- cuvee: '⋎',
- cuwed: '⋏',
- cwconint: '∲',
- cwint: '∱',
- cylcty: '⌭',
- dArr: '⇓',
- dHar: '⥥',
- dagger: '†',
- daleth: 'ℸ',
- darr: '↓',
- dash: '‐',
- dashv: '⊣',
- dbkarow: '⤏',
- dblac: '˝',
- dcaron: 'ď',
- dcy: 'д',
- dd: 'ⅆ',
- ddagger: '‡',
- ddarr: '⇊',
- ddotseq: '⩷',
- deg: '°',
- delta: 'δ',
- demptyv: '⦱',
- dfisht: '⥿',
- dfr: '𝔡',
- dharl: '⇃',
- dharr: '⇂',
- diam: '⋄',
- diamond: '⋄',
- diamondsuit: '♦',
- diams: '♦',
- die: '¨',
- digamma: 'ϝ',
- disin: '⋲',
- div: '÷',
- divide: '÷',
- divideontimes: '⋇',
- divonx: '⋇',
- djcy: 'ђ',
- dlcorn: '⌞',
- dlcrop: '⌍',
- dollar: '$',
- dopf: '𝕕',
- dot: '˙',
- doteq: '≐',
- doteqdot: '≑',
- dotminus: '∸',
- dotplus: '∔',
- dotsquare: '⊡',
- doublebarwedge: '⌆',
- downarrow: '↓',
- downdownarrows: '⇊',
- downharpoonleft: '⇃',
- downharpoonright: '⇂',
- drbkarow: '⤐',
- drcorn: '⌟',
- drcrop: '⌌',
- dscr: '𝒹',
- dscy: 'ѕ',
- dsol: '⧶',
- dstrok: 'đ',
- dtdot: '⋱',
- dtri: '▿',
- dtrif: '▾',
- duarr: '⇵',
- duhar: '⥯',
- dwangle: '⦦',
- dzcy: 'џ',
- dzigrarr: '⟿',
- eDDot: '⩷',
- eDot: '≑',
- eacute: 'é',
- easter: '⩮',
- ecaron: 'ě',
- ecir: '≖',
- ecirc: 'ê',
- ecolon: '≕',
- ecy: 'э',
- edot: 'ė',
- ee: 'ⅇ',
- efDot: '≒',
- efr: '𝔢',
- eg: '⪚',
- egrave: 'è',
- egs: '⪖',
- egsdot: '⪘',
- el: '⪙',
- elinters: '⏧',
- ell: 'ℓ',
- els: '⪕',
- elsdot: '⪗',
- emacr: 'ē',
- empty: '∅',
- emptyset: '∅',
- emptyv: '∅',
- emsp13: ' ',
- emsp14: ' ',
- emsp: ' ',
- eng: 'ŋ',
- ensp: ' ',
- eogon: 'ę',
- eopf: '𝕖',
- epar: '⋕',
- eparsl: '⧣',
- eplus: '⩱',
- epsi: 'ε',
- epsilon: 'ε',
- epsiv: 'ϵ',
- eqcirc: '≖',
- eqcolon: '≕',
- eqsim: '≂',
- eqslantgtr: '⪖',
- eqslantless: '⪕',
- equals: '=',
- equest: '≟',
- equiv: '≡',
- equivDD: '⩸',
- eqvparsl: '⧥',
- erDot: '≓',
- erarr: '⥱',
- escr: 'ℯ',
- esdot: '≐',
- esim: '≂',
- eta: 'η',
- eth: 'ð',
- euml: 'ë',
- euro: '€',
- excl: '!',
- exist: '∃',
- expectation: 'ℰ',
- exponentiale: 'ⅇ',
- fallingdotseq: '≒',
- fcy: 'ф',
- female: '♀',
- ffilig: 'ffi',
- fflig: 'ff',
- ffllig: 'ffl',
- ffr: '𝔣',
- filig: 'fi',
- fjlig: 'fj',
- flat: '♭',
- fllig: 'fl',
- fltns: '▱',
- fnof: 'ƒ',
- fopf: '𝕗',
- forall: '∀',
- fork: '⋔',
- forkv: '⫙',
- fpartint: '⨍',
- frac12: '½',
- frac13: '⅓',
- frac14: '¼',
- frac15: '⅕',
- frac16: '⅙',
- frac18: '⅛',
- frac23: '⅔',
- frac25: '⅖',
- frac34: '¾',
- frac35: '⅗',
- frac38: '⅜',
- frac45: '⅘',
- frac56: '⅚',
- frac58: '⅝',
- frac78: '⅞',
- frasl: '⁄',
- frown: '⌢',
- fscr: '𝒻',
- gE: '≧',
- gEl: '⪌',
- gacute: 'ǵ',
- gamma: 'γ',
- gammad: 'ϝ',
- gap: '⪆',
- gbreve: 'ğ',
- gcirc: 'ĝ',
- gcy: 'г',
- gdot: 'ġ',
- ge: '≥',
- gel: '⋛',
- geq: '≥',
- geqq: '≧',
- geqslant: '⩾',
- ges: '⩾',
- gescc: '⪩',
- gesdot: '⪀',
- gesdoto: '⪂',
- gesdotol: '⪄',
- gesl: '⋛︀',
- gesles: '⪔',
- gfr: '𝔤',
- gg: '≫',
- ggg: '⋙',
- gimel: 'ℷ',
- gjcy: 'ѓ',
- gl: '≷',
- glE: '⪒',
- gla: '⪥',
- glj: '⪤',
- gnE: '≩',
- gnap: '⪊',
- gnapprox: '⪊',
- gne: '⪈',
- gneq: '⪈',
- gneqq: '≩',
- gnsim: '⋧',
- gopf: '𝕘',
- grave: '`',
- gscr: 'ℊ',
- gsim: '≳',
- gsime: '⪎',
- gsiml: '⪐',
- gt: '>',
- gtcc: '⪧',
- gtcir: '⩺',
- gtdot: '⋗',
- gtlPar: '⦕',
- gtquest: '⩼',
- gtrapprox: '⪆',
- gtrarr: '⥸',
- gtrdot: '⋗',
- gtreqless: '⋛',
- gtreqqless: '⪌',
- gtrless: '≷',
- gtrsim: '≳',
- gvertneqq: '≩︀',
- gvnE: '≩︀',
- hArr: '⇔',
- hairsp: ' ',
- half: '½',
- hamilt: 'ℋ',
- hardcy: 'ъ',
- harr: '↔',
- harrcir: '⥈',
- harrw: '↭',
- hbar: 'ℏ',
- hcirc: 'ĥ',
- hearts: '♥',
- heartsuit: '♥',
- hellip: '…',
- hercon: '⊹',
- hfr: '𝔥',
- hksearow: '⤥',
- hkswarow: '⤦',
- hoarr: '⇿',
- homtht: '∻',
- hookleftarrow: '↩',
- hookrightarrow: '↪',
- hopf: '𝕙',
- horbar: '―',
- hscr: '𝒽',
- hslash: 'ℏ',
- hstrok: 'ħ',
- hybull: '⁃',
- hyphen: '‐',
- iacute: 'í',
- ic: '⁣',
- icirc: 'î',
- icy: 'и',
- iecy: 'е',
- iexcl: '¡',
- iff: '⇔',
- ifr: '𝔦',
- igrave: 'ì',
- ii: 'ⅈ',
- iiiint: '⨌',
- iiint: '∭',
- iinfin: '⧜',
- iiota: '℩',
- ijlig: 'ij',
- imacr: 'ī',
- image: 'ℑ',
- imagline: 'ℐ',
- imagpart: 'ℑ',
- imath: 'ı',
- imof: '⊷',
- imped: 'Ƶ',
- in: '∈',
- incare: '℅',
- infin: '∞',
- infintie: '⧝',
- inodot: 'ı',
- int: '∫',
- intcal: '⊺',
- integers: 'ℤ',
- intercal: '⊺',
- intlarhk: '⨗',
- intprod: '⨼',
- iocy: 'ё',
- iogon: 'į',
- iopf: '𝕚',
- iota: 'ι',
- iprod: '⨼',
- iquest: '¿',
- iscr: '𝒾',
- isin: '∈',
- isinE: '⋹',
- isindot: '⋵',
- isins: '⋴',
- isinsv: '⋳',
- isinv: '∈',
- it: '⁢',
- itilde: 'ĩ',
- iukcy: 'і',
- iuml: 'ï',
- jcirc: 'ĵ',
- jcy: 'й',
- jfr: '𝔧',
- jmath: 'ȷ',
- jopf: '𝕛',
- jscr: '𝒿',
- jsercy: 'ј',
- jukcy: 'є',
- kappa: 'κ',
- kappav: 'ϰ',
- kcedil: 'ķ',
- kcy: 'к',
- kfr: '𝔨',
- kgreen: 'ĸ',
- khcy: 'х',
- kjcy: 'ќ',
- kopf: '𝕜',
- kscr: '𝓀',
- lAarr: '⇚',
- lArr: '⇐',
- lAtail: '⤛',
- lBarr: '⤎',
- lE: '≦',
- lEg: '⪋',
- lHar: '⥢',
- lacute: 'ĺ',
- laemptyv: '⦴',
- lagran: 'ℒ',
- lambda: 'λ',
- lang: '⟨',
- langd: '⦑',
- langle: '⟨',
- lap: '⪅',
- laquo: '«',
- larr: '←',
- larrb: '⇤',
- larrbfs: '⤟',
- larrfs: '⤝',
- larrhk: '↩',
- larrlp: '↫',
- larrpl: '⤹',
- larrsim: '⥳',
- larrtl: '↢',
- lat: '⪫',
- latail: '⤙',
- late: '⪭',
- lates: '⪭︀',
- lbarr: '⤌',
- lbbrk: '❲',
- lbrace: '{',
- lbrack: '[',
- lbrke: '⦋',
- lbrksld: '⦏',
- lbrkslu: '⦍',
- lcaron: 'ľ',
- lcedil: 'ļ',
- lceil: '⌈',
- lcub: '{',
- lcy: 'л',
- ldca: '⤶',
- ldquo: '“',
- ldquor: '„',
- ldrdhar: '⥧',
- ldrushar: '⥋',
- ldsh: '↲',
- le: '≤',
- leftarrow: '←',
- leftarrowtail: '↢',
- leftharpoondown: '↽',
- leftharpoonup: '↼',
- leftleftarrows: '⇇',
- leftrightarrow: '↔',
- leftrightarrows: '⇆',
- leftrightharpoons: '⇋',
- leftrightsquigarrow: '↭',
- leftthreetimes: '⋋',
- leg: '⋚',
- leq: '≤',
- leqq: '≦',
- leqslant: '⩽',
- les: '⩽',
- lescc: '⪨',
- lesdot: '⩿',
- lesdoto: '⪁',
- lesdotor: '⪃',
- lesg: '⋚︀',
- lesges: '⪓',
- lessapprox: '⪅',
- lessdot: '⋖',
- lesseqgtr: '⋚',
- lesseqqgtr: '⪋',
- lessgtr: '≶',
- lesssim: '≲',
- lfisht: '⥼',
- lfloor: '⌊',
- lfr: '𝔩',
- lg: '≶',
- lgE: '⪑',
- lhard: '↽',
- lharu: '↼',
- lharul: '⥪',
- lhblk: '▄',
- ljcy: 'љ',
- ll: '≪',
- llarr: '⇇',
- llcorner: '⌞',
- llhard: '⥫',
- lltri: '◺',
- lmidot: 'ŀ',
- lmoust: '⎰',
- lmoustache: '⎰',
- lnE: '≨',
- lnap: '⪉',
- lnapprox: '⪉',
- lne: '⪇',
- lneq: '⪇',
- lneqq: '≨',
- lnsim: '⋦',
- loang: '⟬',
- loarr: '⇽',
- lobrk: '⟦',
- longleftarrow: '⟵',
- longleftrightarrow: '⟷',
- longmapsto: '⟼',
- longrightarrow: '⟶',
- looparrowleft: '↫',
- looparrowright: '↬',
- lopar: '⦅',
- lopf: '𝕝',
- loplus: '⨭',
- lotimes: '⨴',
- lowast: '∗',
- lowbar: '_',
- loz: '◊',
- lozenge: '◊',
- lozf: '⧫',
- lpar: '(',
- lparlt: '⦓',
- lrarr: '⇆',
- lrcorner: '⌟',
- lrhar: '⇋',
- lrhard: '⥭',
- lrm: '‎',
- lrtri: '⊿',
- lsaquo: '‹',
- lscr: '𝓁',
- lsh: '↰',
- lsim: '≲',
- lsime: '⪍',
- lsimg: '⪏',
- lsqb: '[',
- lsquo: '‘',
- lsquor: '‚',
- lstrok: 'ł',
- lt: '<',
- ltcc: '⪦',
- ltcir: '⩹',
- ltdot: '⋖',
- lthree: '⋋',
- ltimes: '⋉',
- ltlarr: '⥶',
- ltquest: '⩻',
- ltrPar: '⦖',
- ltri: '◃',
- ltrie: '⊴',
- ltrif: '◂',
- lurdshar: '⥊',
- luruhar: '⥦',
- lvertneqq: '≨︀',
- lvnE: '≨︀',
- mDDot: '∺',
- macr: '¯',
- male: '♂',
- malt: '✠',
- maltese: '✠',
- map: '↦',
- mapsto: '↦',
- mapstodown: '↧',
- mapstoleft: '↤',
- mapstoup: '↥',
- marker: '▮',
- mcomma: '⨩',
- mcy: 'м',
- mdash: '—',
- measuredangle: '∡',
- mfr: '𝔪',
- mho: '℧',
- micro: 'µ',
- mid: '∣',
- midast: '*',
- midcir: '⫰',
- middot: '·',
- minus: '−',
- minusb: '⊟',
- minusd: '∸',
- minusdu: '⨪',
- mlcp: '⫛',
- mldr: '…',
- mnplus: '∓',
- models: '⊧',
- mopf: '𝕞',
- mp: '∓',
- mscr: '𝓂',
- mstpos: '∾',
- mu: 'μ',
- multimap: '⊸',
- mumap: '⊸',
- nGg: '⋙̸',
- nGt: '≫⃒',
- nGtv: '≫̸',
- nLeftarrow: '⇍',
- nLeftrightarrow: '⇎',
- nLl: '⋘̸',
- nLt: '≪⃒',
- nLtv: '≪̸',
- nRightarrow: '⇏',
- nVDash: '⊯',
- nVdash: '⊮',
- nabla: '∇',
- nacute: 'ń',
- nang: '∠⃒',
- nap: '≉',
- napE: '⩰̸',
- napid: '≋̸',
- napos: 'ʼn',
- napprox: '≉',
- natur: '♮',
- natural: '♮',
- naturals: 'ℕ',
- nbsp: ' ',
- nbump: '≎̸',
- nbumpe: '≏̸',
- ncap: '⩃',
- ncaron: 'ň',
- ncedil: 'ņ',
- ncong: '≇',
- ncongdot: '⩭̸',
- ncup: '⩂',
- ncy: 'н',
- ndash: '–',
- ne: '≠',
- neArr: '⇗',
- nearhk: '⤤',
- nearr: '↗',
- nearrow: '↗',
- nedot: '≐̸',
- nequiv: '≢',
- nesear: '⤨',
- nesim: '≂̸',
- nexist: '∄',
- nexists: '∄',
- nfr: '𝔫',
- ngE: '≧̸',
- nge: '≱',
- ngeq: '≱',
- ngeqq: '≧̸',
- ngeqslant: '⩾̸',
- nges: '⩾̸',
- ngsim: '≵',
- ngt: '≯',
- ngtr: '≯',
- nhArr: '⇎',
- nharr: '↮',
- nhpar: '⫲',
- ni: '∋',
- nis: '⋼',
- nisd: '⋺',
- niv: '∋',
- njcy: 'њ',
- nlArr: '⇍',
- nlE: '≦̸',
- nlarr: '↚',
- nldr: '‥',
- nle: '≰',
- nleftarrow: '↚',
- nleftrightarrow: '↮',
- nleq: '≰',
- nleqq: '≦̸',
- nleqslant: '⩽̸',
- nles: '⩽̸',
- nless: '≮',
- nlsim: '≴',
- nlt: '≮',
- nltri: '⋪',
- nltrie: '⋬',
- nmid: '∤',
- nopf: '𝕟',
- not: '¬',
- notin: '∉',
- notinE: '⋹̸',
- notindot: '⋵̸',
- notinva: '∉',
- notinvb: '⋷',
- notinvc: '⋶',
- notni: '∌',
- notniva: '∌',
- notnivb: '⋾',
- notnivc: '⋽',
- npar: '∦',
- nparallel: '∦',
- nparsl: '⫽⃥',
- npart: '∂̸',
- npolint: '⨔',
- npr: '⊀',
- nprcue: '⋠',
- npre: '⪯̸',
- nprec: '⊀',
- npreceq: '⪯̸',
- nrArr: '⇏',
- nrarr: '↛',
- nrarrc: '⤳̸',
- nrarrw: '↝̸',
- nrightarrow: '↛',
- nrtri: '⋫',
- nrtrie: '⋭',
- nsc: '⊁',
- nsccue: '⋡',
- nsce: '⪰̸',
- nscr: '𝓃',
- nshortmid: '∤',
- nshortparallel: '∦',
- nsim: '≁',
- nsime: '≄',
- nsimeq: '≄',
- nsmid: '∤',
- nspar: '∦',
- nsqsube: '⋢',
- nsqsupe: '⋣',
- nsub: '⊄',
- nsubE: '⫅̸',
- nsube: '⊈',
- nsubset: '⊂⃒',
- nsubseteq: '⊈',
- nsubseteqq: '⫅̸',
- nsucc: '⊁',
- nsucceq: '⪰̸',
- nsup: '⊅',
- nsupE: '⫆̸',
- nsupe: '⊉',
- nsupset: '⊃⃒',
- nsupseteq: '⊉',
- nsupseteqq: '⫆̸',
- ntgl: '≹',
- ntilde: 'ñ',
- ntlg: '≸',
- ntriangleleft: '⋪',
- ntrianglelefteq: '⋬',
- ntriangleright: '⋫',
- ntrianglerighteq: '⋭',
- nu: 'ν',
- num: '#',
- numero: '№',
- numsp: ' ',
- nvDash: '⊭',
- nvHarr: '⤄',
- nvap: '≍⃒',
- nvdash: '⊬',
- nvge: '≥⃒',
- nvgt: '>⃒',
- nvinfin: '⧞',
- nvlArr: '⤂',
- nvle: '≤⃒',
- nvlt: '<⃒',
- nvltrie: '⊴⃒',
- nvrArr: '⤃',
- nvrtrie: '⊵⃒',
- nvsim: '∼⃒',
- nwArr: '⇖',
- nwarhk: '⤣',
- nwarr: '↖',
- nwarrow: '↖',
- nwnear: '⤧',
- oS: 'Ⓢ',
- oacute: 'ó',
- oast: '⊛',
- ocir: '⊚',
- ocirc: 'ô',
- ocy: 'о',
- odash: '⊝',
- odblac: 'ő',
- odiv: '⨸',
- odot: '⊙',
- odsold: '⦼',
- oelig: 'œ',
- ofcir: '⦿',
- ofr: '𝔬',
- ogon: '˛',
- ograve: 'ò',
- ogt: '⧁',
- ohbar: '⦵',
- ohm: 'Ω',
- oint: '∮',
- olarr: '↺',
- olcir: '⦾',
- olcross: '⦻',
- oline: '‾',
- olt: '⧀',
- omacr: 'ō',
- omega: 'ω',
- omicron: 'ο',
- omid: '⦶',
- ominus: '⊖',
- oopf: '𝕠',
- opar: '⦷',
- operp: '⦹',
- oplus: '⊕',
- or: '∨',
- orarr: '↻',
- ord: '⩝',
- order: 'ℴ',
- orderof: 'ℴ',
- ordf: 'ª',
- ordm: 'º',
- origof: '⊶',
- oror: '⩖',
- orslope: '⩗',
- orv: '⩛',
- oscr: 'ℴ',
- oslash: 'ø',
- osol: '⊘',
- otilde: 'õ',
- otimes: '⊗',
- otimesas: '⨶',
- ouml: 'ö',
- ovbar: '⌽',
- par: '∥',
- para: '¶',
- parallel: '∥',
- parsim: '⫳',
- parsl: '⫽',
- part: '∂',
- pcy: 'п',
- percnt: '%',
- period: '.',
- permil: '‰',
- perp: '⊥',
- pertenk: '‱',
- pfr: '𝔭',
- phi: 'φ',
- phiv: 'ϕ',
- phmmat: 'ℳ',
- phone: '☎',
- pi: 'π',
- pitchfork: '⋔',
- piv: 'ϖ',
- planck: 'ℏ',
- planckh: 'ℎ',
- plankv: 'ℏ',
- plus: '+',
- plusacir: '⨣',
- plusb: '⊞',
- pluscir: '⨢',
- plusdo: '∔',
- plusdu: '⨥',
- pluse: '⩲',
- plusmn: '±',
- plussim: '⨦',
- plustwo: '⨧',
- pm: '±',
- pointint: '⨕',
- popf: '𝕡',
- pound: '£',
- pr: '≺',
- prE: '⪳',
- prap: '⪷',
- prcue: '≼',
- pre: '⪯',
- prec: '≺',
- precapprox: '⪷',
- preccurlyeq: '≼',
- preceq: '⪯',
- precnapprox: '⪹',
- precneqq: '⪵',
- precnsim: '⋨',
- precsim: '≾',
- prime: '′',
- primes: 'ℙ',
- prnE: '⪵',
- prnap: '⪹',
- prnsim: '⋨',
- prod: '∏',
- profalar: '⌮',
- profline: '⌒',
- profsurf: '⌓',
- prop: '∝',
- propto: '∝',
- prsim: '≾',
- prurel: '⊰',
- pscr: '𝓅',
- psi: 'ψ',
- puncsp: ' ',
- qfr: '𝔮',
- qint: '⨌',
- qopf: '𝕢',
- qprime: '⁗',
- qscr: '𝓆',
- quaternions: 'ℍ',
- quatint: '⨖',
- quest: '?',
- questeq: '≟',
- quot: '"',
- rAarr: '⇛',
- rArr: '⇒',
- rAtail: '⤜',
- rBarr: '⤏',
- rHar: '⥤',
- race: '∽̱',
- racute: 'ŕ',
- radic: '√',
- raemptyv: '⦳',
- rang: '⟩',
- rangd: '⦒',
- range: '⦥',
- rangle: '⟩',
- raquo: '»',
- rarr: '→',
- rarrap: '⥵',
- rarrb: '⇥',
- rarrbfs: '⤠',
- rarrc: '⤳',
- rarrfs: '⤞',
- rarrhk: '↪',
- rarrlp: '↬',
- rarrpl: '⥅',
- rarrsim: '⥴',
- rarrtl: '↣',
- rarrw: '↝',
- ratail: '⤚',
- ratio: '∶',
- rationals: 'ℚ',
- rbarr: '⤍',
- rbbrk: '❳',
- rbrace: '}',
- rbrack: ']',
- rbrke: '⦌',
- rbrksld: '⦎',
- rbrkslu: '⦐',
- rcaron: 'ř',
- rcedil: 'ŗ',
- rceil: '⌉',
- rcub: '}',
- rcy: 'р',
- rdca: '⤷',
- rdldhar: '⥩',
- rdquo: '”',
- rdquor: '”',
- rdsh: '↳',
- real: 'ℜ',
- realine: 'ℛ',
- realpart: 'ℜ',
- reals: 'ℝ',
- rect: '▭',
- reg: '®',
- rfisht: '⥽',
- rfloor: '⌋',
- rfr: '𝔯',
- rhard: '⇁',
- rharu: '⇀',
- rharul: '⥬',
- rho: 'ρ',
- rhov: 'ϱ',
- rightarrow: '→',
- rightarrowtail: '↣',
- rightharpoondown: '⇁',
- rightharpoonup: '⇀',
- rightleftarrows: '⇄',
- rightleftharpoons: '⇌',
- rightrightarrows: '⇉',
- rightsquigarrow: '↝',
- rightthreetimes: '⋌',
- ring: '˚',
- risingdotseq: '≓',
- rlarr: '⇄',
- rlhar: '⇌',
- rlm: '‏',
- rmoust: '⎱',
- rmoustache: '⎱',
- rnmid: '⫮',
- roang: '⟭',
- roarr: '⇾',
- robrk: '⟧',
- ropar: '⦆',
- ropf: '𝕣',
- roplus: '⨮',
- rotimes: '⨵',
- rpar: ')',
- rpargt: '⦔',
- rppolint: '⨒',
- rrarr: '⇉',
- rsaquo: '›',
- rscr: '𝓇',
- rsh: '↱',
- rsqb: ']',
- rsquo: '’',
- rsquor: '’',
- rthree: '⋌',
- rtimes: '⋊',
- rtri: '▹',
- rtrie: '⊵',
- rtrif: '▸',
- rtriltri: '⧎',
- ruluhar: '⥨',
- rx: '℞',
- sacute: 'ś',
- sbquo: '‚',
- sc: '≻',
- scE: '⪴',
- scap: '⪸',
- scaron: 'š',
- sccue: '≽',
- sce: '⪰',
- scedil: 'ş',
- scirc: 'ŝ',
- scnE: '⪶',
- scnap: '⪺',
- scnsim: '⋩',
- scpolint: '⨓',
- scsim: '≿',
- scy: 'с',
- sdot: '⋅',
- sdotb: '⊡',
- sdote: '⩦',
- seArr: '⇘',
- searhk: '⤥',
- searr: '↘',
- searrow: '↘',
- sect: '§',
- semi: ';',
- seswar: '⤩',
- setminus: '∖',
- setmn: '∖',
- sext: '✶',
- sfr: '𝔰',
- sfrown: '⌢',
- sharp: '♯',
- shchcy: 'щ',
- shcy: 'ш',
- shortmid: '∣',
- shortparallel: '∥',
- shy: '­',
- sigma: 'σ',
- sigmaf: 'ς',
- sigmav: 'ς',
- sim: '∼',
- simdot: '⩪',
- sime: '≃',
- simeq: '≃',
- simg: '⪞',
- simgE: '⪠',
- siml: '⪝',
- simlE: '⪟',
- simne: '≆',
- simplus: '⨤',
- simrarr: '⥲',
- slarr: '←',
- smallsetminus: '∖',
- smashp: '⨳',
- smeparsl: '⧤',
- smid: '∣',
- smile: '⌣',
- smt: '⪪',
- smte: '⪬',
- smtes: '⪬︀',
- softcy: 'ь',
- sol: '/',
- solb: '⧄',
- solbar: '⌿',
- sopf: '𝕤',
- spades: '♠',
- spadesuit: '♠',
- spar: '∥',
- sqcap: '⊓',
- sqcaps: '⊓︀',
- sqcup: '⊔',
- sqcups: '⊔︀',
- sqsub: '⊏',
- sqsube: '⊑',
- sqsubset: '⊏',
- sqsubseteq: '⊑',
- sqsup: '⊐',
- sqsupe: '⊒',
- sqsupset: '⊐',
- sqsupseteq: '⊒',
- squ: '□',
- square: '□',
- squarf: '▪',
- squf: '▪',
- srarr: '→',
- sscr: '𝓈',
- ssetmn: '∖',
- ssmile: '⌣',
- sstarf: '⋆',
- star: '☆',
- starf: '★',
- straightepsilon: 'ϵ',
- straightphi: 'ϕ',
- strns: '¯',
- sub: '⊂',
- subE: '⫅',
- subdot: '⪽',
- sube: '⊆',
- subedot: '⫃',
- submult: '⫁',
- subnE: '⫋',
- subne: '⊊',
- subplus: '⪿',
- subrarr: '⥹',
- subset: '⊂',
- subseteq: '⊆',
- subseteqq: '⫅',
- subsetneq: '⊊',
- subsetneqq: '⫋',
- subsim: '⫇',
- subsub: '⫕',
- subsup: '⫓',
- succ: '≻',
- succapprox: '⪸',
- succcurlyeq: '≽',
- succeq: '⪰',
- succnapprox: '⪺',
- succneqq: '⪶',
- succnsim: '⋩',
- succsim: '≿',
- sum: '∑',
- sung: '♪',
- sup1: '¹',
- sup2: '²',
- sup3: '³',
- sup: '⊃',
- supE: '⫆',
- supdot: '⪾',
- supdsub: '⫘',
- supe: '⊇',
- supedot: '⫄',
- suphsol: '⟉',
- suphsub: '⫗',
- suplarr: '⥻',
- supmult: '⫂',
- supnE: '⫌',
- supne: '⊋',
- supplus: '⫀',
- supset: '⊃',
- supseteq: '⊇',
- supseteqq: '⫆',
- supsetneq: '⊋',
- supsetneqq: '⫌',
- supsim: '⫈',
- supsub: '⫔',
- supsup: '⫖',
- swArr: '⇙',
- swarhk: '⤦',
- swarr: '↙',
- swarrow: '↙',
- swnwar: '⤪',
- szlig: 'ß',
- target: '⌖',
- tau: 'τ',
- tbrk: '⎴',
- tcaron: 'ť',
- tcedil: 'ţ',
- tcy: 'т',
- tdot: '⃛',
- telrec: '⌕',
- tfr: '𝔱',
- there4: '∴',
- therefore: '∴',
- theta: 'θ',
- thetasym: 'ϑ',
- thetav: 'ϑ',
- thickapprox: '≈',
- thicksim: '∼',
- thinsp: ' ',
- thkap: '≈',
- thksim: '∼',
- thorn: 'þ',
- tilde: '˜',
- times: '×',
- timesb: '⊠',
- timesbar: '⨱',
- timesd: '⨰',
- tint: '∭',
- toea: '⤨',
- top: '⊤',
- topbot: '⌶',
- topcir: '⫱',
- topf: '𝕥',
- topfork: '⫚',
- tosa: '⤩',
- tprime: '‴',
- trade: '™',
- triangle: '▵',
- triangledown: '▿',
- triangleleft: '◃',
- trianglelefteq: '⊴',
- triangleq: '≜',
- triangleright: '▹',
- trianglerighteq: '⊵',
- tridot: '◬',
- trie: '≜',
- triminus: '⨺',
- triplus: '⨹',
- trisb: '⧍',
- tritime: '⨻',
- trpezium: '⏢',
- tscr: '𝓉',
- tscy: 'ц',
- tshcy: 'ћ',
- tstrok: 'ŧ',
- twixt: '≬',
- twoheadleftarrow: '↞',
- twoheadrightarrow: '↠',
- uArr: '⇑',
- uHar: '⥣',
- uacute: 'ú',
- uarr: '↑',
- ubrcy: 'ў',
- ubreve: 'ŭ',
- ucirc: 'û',
- ucy: 'у',
- udarr: '⇅',
- udblac: 'ű',
- udhar: '⥮',
- ufisht: '⥾',
- ufr: '𝔲',
- ugrave: 'ù',
- uharl: '↿',
- uharr: '↾',
- uhblk: '▀',
- ulcorn: '⌜',
- ulcorner: '⌜',
- ulcrop: '⌏',
- ultri: '◸',
- umacr: 'ū',
- uml: '¨',
- uogon: 'ų',
- uopf: '𝕦',
- uparrow: '↑',
- updownarrow: '↕',
- upharpoonleft: '↿',
- upharpoonright: '↾',
- uplus: '⊎',
- upsi: 'υ',
- upsih: 'ϒ',
- upsilon: 'υ',
- upuparrows: '⇈',
- urcorn: '⌝',
- urcorner: '⌝',
- urcrop: '⌎',
- uring: 'ů',
- urtri: '◹',
- uscr: '𝓊',
- utdot: '⋰',
- utilde: 'ũ',
- utri: '▵',
- utrif: '▴',
- uuarr: '⇈',
- uuml: 'ü',
- uwangle: '⦧',
- vArr: '⇕',
- vBar: '⫨',
- vBarv: '⫩',
- vDash: '⊨',
- vangrt: '⦜',
- varepsilon: 'ϵ',
- varkappa: 'ϰ',
- varnothing: '∅',
- varphi: 'ϕ',
- varpi: 'ϖ',
- varpropto: '∝',
- varr: '↕',
- varrho: 'ϱ',
- varsigma: 'ς',
- varsubsetneq: '⊊︀',
- varsubsetneqq: '⫋︀',
- varsupsetneq: '⊋︀',
- varsupsetneqq: '⫌︀',
- vartheta: 'ϑ',
- vartriangleleft: '⊲',
- vartriangleright: '⊳',
- vcy: 'в',
- vdash: '⊢',
- vee: '∨',
- veebar: '⊻',
- veeeq: '≚',
- vellip: '⋮',
- verbar: '|',
- vert: '|',
- vfr: '𝔳',
- vltri: '⊲',
- vnsub: '⊂⃒',
- vnsup: '⊃⃒',
- vopf: '𝕧',
- vprop: '∝',
- vrtri: '⊳',
- vscr: '𝓋',
- vsubnE: '⫋︀',
- vsubne: '⊊︀',
- vsupnE: '⫌︀',
- vsupne: '⊋︀',
- vzigzag: '⦚',
- wcirc: 'ŵ',
- wedbar: '⩟',
- wedge: '∧',
- wedgeq: '≙',
- weierp: '℘',
- wfr: '𝔴',
- wopf: '𝕨',
- wp: '℘',
- wr: '≀',
- wreath: '≀',
- wscr: '𝓌',
- xcap: '⋂',
- xcirc: '◯',
- xcup: '⋃',
- xdtri: '▽',
- xfr: '𝔵',
- xhArr: '⟺',
- xharr: '⟷',
- xi: 'ξ',
- xlArr: '⟸',
- xlarr: '⟵',
- xmap: '⟼',
- xnis: '⋻',
- xodot: '⨀',
- xopf: '𝕩',
- xoplus: '⨁',
- xotime: '⨂',
- xrArr: '⟹',
- xrarr: '⟶',
- xscr: '𝓍',
- xsqcup: '⨆',
- xuplus: '⨄',
- xutri: '△',
- xvee: '⋁',
- xwedge: '⋀',
- yacute: 'ý',
- yacy: 'я',
- ycirc: 'ŷ',
- ycy: 'ы',
- yen: '¥',
- yfr: '𝔶',
- yicy: 'ї',
- yopf: '𝕪',
- yscr: '𝓎',
- yucy: 'ю',
- yuml: 'ÿ',
- zacute: 'ź',
- zcaron: 'ž',
- zcy: 'з',
- zdot: 'ż',
- zeetrf: 'ℨ',
- zeta: 'ζ',
- zfr: '𝔷',
- zhcy: 'ж',
- zigrarr: '⇝',
- zopf: '𝕫',
- zscr: '𝓏',
- zwj: '‍',
- zwnj: '‌'
-}
-
-;// CONCATENATED MODULE: ./node_modules/decode-named-character-reference/index.js
-
-
-const own = {}.hasOwnProperty
-
-/**
- * Decode a single character reference (without the `&` or `;`).
- * You probably only need this when you’re building parsers yourself that follow
- * different rules compared to HTML.
- * This is optimized to be tiny in browsers.
- *
- * @param {string} value
- * `notin` (named), `#123` (deci), `#x123` (hexa).
- * @returns {string|false}
- * Decoded reference.
- */
-function decodeNamedCharacterReference(value) {
- return own.call(characterEntities, value) ? characterEntities[value] : false
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/character-reference.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const characterReference = {
- name: 'characterReference',
- tokenize: tokenizeCharacterReference
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeCharacterReference(effects, ok, nok) {
- const self = this
- let size = 0
- /** @type {number} */
- let max
- /** @type {(code: Code) => boolean} */
- let test
- return start
-
- /**
- * Start of character reference.
- *
- * ```markdown
- * > | a&amp;b
- * ^
- * > | a&#123;b
- * ^
- * > | a&#x9;b
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('characterReference')
- effects.enter('characterReferenceMarker')
- effects.consume(code)
- effects.exit('characterReferenceMarker')
- return open
- }
-
- /**
- * After `&`, at `#` for numeric references or alphanumeric for named
- * references.
- *
- * ```markdown
- * > | a&amp;b
- * ^
- * > | a&#123;b
- * ^
- * > | a&#x9;b
- * ^
- * ```
- *
- * @type {State}
- */
- function open(code) {
- if (code === 35) {
- effects.enter('characterReferenceMarkerNumeric')
- effects.consume(code)
- effects.exit('characterReferenceMarkerNumeric')
- return numeric
- }
- effects.enter('characterReferenceValue')
- max = 31
- test = asciiAlphanumeric
- return value(code)
- }
-
- /**
- * After `#`, at `x` for hexadecimals or digit for decimals.
- *
- * ```markdown
- * > | a&#123;b
- * ^
- * > | a&#x9;b
- * ^
- * ```
- *
- * @type {State}
- */
- function numeric(code) {
- if (code === 88 || code === 120) {
- effects.enter('characterReferenceMarkerHexadecimal')
- effects.consume(code)
- effects.exit('characterReferenceMarkerHexadecimal')
- effects.enter('characterReferenceValue')
- max = 6
- test = asciiHexDigit
- return value
- }
- effects.enter('characterReferenceValue')
- max = 7
- test = asciiDigit
- return value(code)
- }
-
- /**
- * After markers (`&#x`, `&#`, or `&`), in value, before `;`.
- *
- * The character reference kind defines what and how many characters are
- * allowed.
- *
- * ```markdown
- * > | a&amp;b
- * ^^^
- * > | a&#123;b
- * ^^^
- * > | a&#x9;b
- * ^
- * ```
- *
- * @type {State}
- */
- function value(code) {
- if (code === 59 && size) {
- const token = effects.exit('characterReferenceValue')
- if (
- test === asciiAlphanumeric &&
- !decodeNamedCharacterReference(self.sliceSerialize(token))
- ) {
- return nok(code)
- }
-
- // To do: `markdown-rs` uses a different name:
- // `CharacterReferenceMarkerSemi`.
- effects.enter('characterReferenceMarker')
- effects.consume(code)
- effects.exit('characterReferenceMarker')
- effects.exit('characterReference')
- return ok
- }
- if (test(code) && size++ < max) {
- effects.consume(code)
- return value
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/character-escape.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-/** @type {Construct} */
-const characterEscape = {
- name: 'characterEscape',
- tokenize: tokenizeCharacterEscape
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeCharacterEscape(effects, ok, nok) {
- return start
-
- /**
- * Start of character escape.
- *
- * ```markdown
- * > | a\*b
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('characterEscape')
- effects.enter('escapeMarker')
- effects.consume(code)
- effects.exit('escapeMarker')
- return inside
- }
-
- /**
- * After `\`, at punctuation.
- *
- * ```markdown
- * > | a\*b
- * ^
- * ```
- *
- * @type {State}
- */
- function inside(code) {
- // ASCII punctuation.
- if (asciiPunctuation(code)) {
- effects.enter('characterEscapeValue')
- effects.consume(code)
- effects.exit('characterEscapeValue')
- effects.exit('characterEscape')
- return ok
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/line-ending.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const lineEnding = {
- name: 'lineEnding',
- tokenize: tokenizeLineEnding
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeLineEnding(effects, ok) {
- return start
-
- /** @type {State} */
- function start(code) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return factorySpace(effects, ok, 'linePrefix')
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/label-end.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Event} Event
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-
-
-
-
-
-/** @type {Construct} */
-const labelEnd = {
- name: 'labelEnd',
- tokenize: tokenizeLabelEnd,
- resolveTo: resolveToLabelEnd,
- resolveAll: resolveAllLabelEnd
-}
-
-/** @type {Construct} */
-const resourceConstruct = {
- tokenize: tokenizeResource
-}
-/** @type {Construct} */
-const referenceFullConstruct = {
- tokenize: tokenizeReferenceFull
-}
-/** @type {Construct} */
-const referenceCollapsedConstruct = {
- tokenize: tokenizeReferenceCollapsed
-}
-
-/** @type {Resolver} */
-function resolveAllLabelEnd(events) {
- let index = -1
- while (++index < events.length) {
- const token = events[index][1]
- if (
- token.type === 'labelImage' ||
- token.type === 'labelLink' ||
- token.type === 'labelEnd'
- ) {
- // Remove the marker.
- events.splice(index + 1, token.type === 'labelImage' ? 4 : 2)
- token.type = 'data'
- index++
- }
- }
- return events
-}
-
-/** @type {Resolver} */
-function resolveToLabelEnd(events, context) {
- let index = events.length
- let offset = 0
- /** @type {Token} */
- let token
- /** @type {number | undefined} */
- let open
- /** @type {number | undefined} */
- let close
- /** @type {Array<Event>} */
- let media
-
- // Find an opening.
- while (index--) {
- token = events[index][1]
- if (open) {
- // If we see another link, or inactive link label, we’ve been here before.
- if (
- token.type === 'link' ||
- (token.type === 'labelLink' && token._inactive)
- ) {
- break
- }
-
- // Mark other link openings as inactive, as we can’t have links in
- // links.
- if (events[index][0] === 'enter' && token.type === 'labelLink') {
- token._inactive = true
- }
- } else if (close) {
- if (
- events[index][0] === 'enter' &&
- (token.type === 'labelImage' || token.type === 'labelLink') &&
- !token._balanced
- ) {
- open = index
- if (token.type !== 'labelLink') {
- offset = 2
- break
- }
- }
- } else if (token.type === 'labelEnd') {
- close = index
- }
- }
- const group = {
- type: events[open][1].type === 'labelLink' ? 'link' : 'image',
- start: Object.assign({}, events[open][1].start),
- end: Object.assign({}, events[events.length - 1][1].end)
- }
- const label = {
- type: 'label',
- start: Object.assign({}, events[open][1].start),
- end: Object.assign({}, events[close][1].end)
- }
- const text = {
- type: 'labelText',
- start: Object.assign({}, events[open + offset + 2][1].end),
- end: Object.assign({}, events[close - 2][1].start)
- }
- media = [
- ['enter', group, context],
- ['enter', label, context]
- ]
-
- // Opening marker.
- media = push(media, events.slice(open + 1, open + offset + 3))
-
- // Text open.
- media = push(media, [['enter', text, context]])
-
- // Always populated by defaults.
-
- // Between.
- media = push(
- media,
- resolveAll(
- context.parser.constructs.insideSpan.null,
- events.slice(open + offset + 4, close - 3),
- context
- )
- )
-
- // Text close, marker close, label close.
- media = push(media, [
- ['exit', text, context],
- events[close - 2],
- events[close - 1],
- ['exit', label, context]
- ])
-
- // Reference, resource, or so.
- media = push(media, events.slice(close + 1))
-
- // Media close.
- media = push(media, [['exit', group, context]])
- splice(events, open, events.length, media)
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeLabelEnd(effects, ok, nok) {
- const self = this
- let index = self.events.length
- /** @type {Token} */
- let labelStart
- /** @type {boolean} */
- let defined
-
- // Find an opening.
- while (index--) {
- if (
- (self.events[index][1].type === 'labelImage' ||
- self.events[index][1].type === 'labelLink') &&
- !self.events[index][1]._balanced
- ) {
- labelStart = self.events[index][1]
- break
- }
- }
- return start
-
- /**
- * Start of label end.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * > | [a][b] c
- * ^
- * > | [a][] b
- * ^
- * > | [a] b
- * ```
- *
- * @type {State}
- */
- function start(code) {
- // If there is not an okay opening.
- if (!labelStart) {
- return nok(code)
- }
-
- // If the corresponding label (link) start is marked as inactive,
- // it means we’d be wrapping a link, like this:
- //
- // ```markdown
- // > | a [b [c](d) e](f) g.
- // ^
- // ```
- //
- // We can’t have that, so it’s just balanced brackets.
- if (labelStart._inactive) {
- return labelEndNok(code)
- }
- defined = self.parser.defined.includes(
- normalizeIdentifier(
- self.sliceSerialize({
- start: labelStart.end,
- end: self.now()
- })
- )
- )
- effects.enter('labelEnd')
- effects.enter('labelMarker')
- effects.consume(code)
- effects.exit('labelMarker')
- effects.exit('labelEnd')
- return after
- }
-
- /**
- * After `]`.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * > | [a][b] c
- * ^
- * > | [a][] b
- * ^
- * > | [a] b
- * ^
- * ```
- *
- * @type {State}
- */
- function after(code) {
- // Note: `markdown-rs` also parses GFM footnotes here, which for us is in
- // an extension.
-
- // Resource (`[asd](fgh)`)?
- if (code === 40) {
- return effects.attempt(
- resourceConstruct,
- labelEndOk,
- defined ? labelEndOk : labelEndNok
- )(code)
- }
-
- // Full (`[asd][fgh]`) or collapsed (`[asd][]`) reference?
- if (code === 91) {
- return effects.attempt(
- referenceFullConstruct,
- labelEndOk,
- defined ? referenceNotFull : labelEndNok
- )(code)
- }
-
- // Shortcut (`[asd]`) reference?
- return defined ? labelEndOk(code) : labelEndNok(code)
- }
-
- /**
- * After `]`, at `[`, but not at a full reference.
- *
- * > 👉 **Note**: we only get here if the label is defined.
- *
- * ```markdown
- * > | [a][] b
- * ^
- * > | [a] b
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceNotFull(code) {
- return effects.attempt(
- referenceCollapsedConstruct,
- labelEndOk,
- labelEndNok
- )(code)
- }
-
- /**
- * Done, we found something.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * > | [a][b] c
- * ^
- * > | [a][] b
- * ^
- * > | [a] b
- * ^
- * ```
- *
- * @type {State}
- */
- function labelEndOk(code) {
- // Note: `markdown-rs` does a bunch of stuff here.
- return ok(code)
- }
-
- /**
- * Done, it’s nothing.
- *
- * There was an okay opening, but we didn’t match anything.
- *
- * ```markdown
- * > | [a](b c
- * ^
- * > | [a][b c
- * ^
- * > | [a] b
- * ^
- * ```
- *
- * @type {State}
- */
- function labelEndNok(code) {
- labelStart._balanced = true
- return nok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeResource(effects, ok, nok) {
- return resourceStart
-
- /**
- * At a resource.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceStart(code) {
- effects.enter('resource')
- effects.enter('resourceMarker')
- effects.consume(code)
- effects.exit('resourceMarker')
- return resourceBefore
- }
-
- /**
- * In resource, after `(`, at optional whitespace.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceBefore(code) {
- return markdownLineEndingOrSpace(code)
- ? factoryWhitespace(effects, resourceOpen)(code)
- : resourceOpen(code)
- }
-
- /**
- * In resource, after optional whitespace, at `)` or a destination.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceOpen(code) {
- if (code === 41) {
- return resourceEnd(code)
- }
- return factoryDestination(
- effects,
- resourceDestinationAfter,
- resourceDestinationMissing,
- 'resourceDestination',
- 'resourceDestinationLiteral',
- 'resourceDestinationLiteralMarker',
- 'resourceDestinationRaw',
- 'resourceDestinationString',
- 32
- )(code)
- }
-
- /**
- * In resource, after destination, at optional whitespace.
- *
- * ```markdown
- * > | [a](b) c
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceDestinationAfter(code) {
- return markdownLineEndingOrSpace(code)
- ? factoryWhitespace(effects, resourceBetween)(code)
- : resourceEnd(code)
- }
-
- /**
- * At invalid destination.
- *
- * ```markdown
- * > | [a](<<) b
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceDestinationMissing(code) {
- return nok(code)
- }
-
- /**
- * In resource, after destination and whitespace, at `(` or title.
- *
- * ```markdown
- * > | [a](b ) c
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceBetween(code) {
- if (code === 34 || code === 39 || code === 40) {
- return factoryTitle(
- effects,
- resourceTitleAfter,
- nok,
- 'resourceTitle',
- 'resourceTitleMarker',
- 'resourceTitleString'
- )(code)
- }
- return resourceEnd(code)
- }
-
- /**
- * In resource, after title, at optional whitespace.
- *
- * ```markdown
- * > | [a](b "c") d
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceTitleAfter(code) {
- return markdownLineEndingOrSpace(code)
- ? factoryWhitespace(effects, resourceEnd)(code)
- : resourceEnd(code)
- }
-
- /**
- * In resource, at `)`.
- *
- * ```markdown
- * > | [a](b) d
- * ^
- * ```
- *
- * @type {State}
- */
- function resourceEnd(code) {
- if (code === 41) {
- effects.enter('resourceMarker')
- effects.consume(code)
- effects.exit('resourceMarker')
- effects.exit('resource')
- return ok
- }
- return nok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeReferenceFull(effects, ok, nok) {
- const self = this
- return referenceFull
-
- /**
- * In a reference (full), at the `[`.
- *
- * ```markdown
- * > | [a][b] d
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceFull(code) {
- return factoryLabel.call(
- self,
- effects,
- referenceFullAfter,
- referenceFullMissing,
- 'reference',
- 'referenceMarker',
- 'referenceString'
- )(code)
- }
-
- /**
- * In a reference (full), after `]`.
- *
- * ```markdown
- * > | [a][b] d
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceFullAfter(code) {
- return self.parser.defined.includes(
- normalizeIdentifier(
- self.sliceSerialize(self.events[self.events.length - 1][1]).slice(1, -1)
- )
- )
- ? ok(code)
- : nok(code)
- }
-
- /**
- * In reference (full) that was missing.
- *
- * ```markdown
- * > | [a][b d
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceFullMissing(code) {
- return nok(code)
- }
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeReferenceCollapsed(effects, ok, nok) {
- return referenceCollapsedStart
-
- /**
- * In reference (collapsed), at `[`.
- *
- * > 👉 **Note**: we only get here if the label is defined.
- *
- * ```markdown
- * > | [a][] d
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceCollapsedStart(code) {
- // We only attempt a collapsed label if there’s a `[`.
-
- effects.enter('reference')
- effects.enter('referenceMarker')
- effects.consume(code)
- effects.exit('referenceMarker')
- return referenceCollapsedOpen
- }
-
- /**
- * In reference (collapsed), at `]`.
- *
- * > 👉 **Note**: we only get here if the label is defined.
- *
- * ```markdown
- * > | [a][] d
- * ^
- * ```
- *
- * @type {State}
- */
- function referenceCollapsedOpen(code) {
- if (code === 93) {
- effects.enter('referenceMarker')
- effects.consume(code)
- effects.exit('referenceMarker')
- effects.exit('reference')
- return ok
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/label-start-image.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const labelStartImage = {
- name: 'labelStartImage',
- tokenize: tokenizeLabelStartImage,
- resolveAll: labelEnd.resolveAll
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeLabelStartImage(effects, ok, nok) {
- const self = this
- return start
-
- /**
- * Start of label (image) start.
- *
- * ```markdown
- * > | a ![b] c
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('labelImage')
- effects.enter('labelImageMarker')
- effects.consume(code)
- effects.exit('labelImageMarker')
- return open
- }
-
- /**
- * After `!`, at `[`.
- *
- * ```markdown
- * > | a ![b] c
- * ^
- * ```
- *
- * @type {State}
- */
- function open(code) {
- if (code === 91) {
- effects.enter('labelMarker')
- effects.consume(code)
- effects.exit('labelMarker')
- effects.exit('labelImage')
- return after
- }
- return nok(code)
- }
-
- /**
- * After `![`.
- *
- * ```markdown
- * > | a ![b] c
- * ^
- * ```
- *
- * This is needed in because, when GFM footnotes are enabled, images never
- * form when started with a `^`.
- * Instead, links form:
- *
- * ```markdown
- * ![^a](b)
- *
- * ![^a][b]
- *
- * [b]: c
- * ```
- *
- * ```html
- * <p>!<a href=\"b\">^a</a></p>
- * <p>!<a href=\"c\">^a</a></p>
- * ```
- *
- * @type {State}
- */
- function after(code) {
- // To do: use a new field to do this, this is still needed for
- // `micromark-extension-gfm-footnote`, but the `label-start-link`
- // behavior isn’t.
- // Hidden footnotes hook.
- /* c8 ignore next 3 */
- return code === 94 && '_hiddenFootnoteSupport' in self.parser.constructs
- ? nok(code)
- : ok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-classify-character/index.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- */
-
-
-/**
- * Classify whether a code represents whitespace, punctuation, or something
- * else.
- *
- * Used for attention (emphasis, strong), whose sequences can open or close
- * based on the class of surrounding characters.
- *
- * > 👉 **Note**: eof (`null`) is seen as whitespace.
- *
- * @param {Code} code
- * Code.
- * @returns {typeof constants.characterGroupWhitespace | typeof constants.characterGroupPunctuation | undefined}
- * Group.
- */
-function classifyCharacter(code) {
- if (
- code === null ||
- markdownLineEndingOrSpace(code) ||
- unicodeWhitespace(code)
- ) {
- return 1
- }
- if (unicodePunctuation(code)) {
- return 2
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/attention.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Event} Event
- * @typedef {import('micromark-util-types').Point} Point
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-
-/** @type {Construct} */
-const attention = {
- name: 'attention',
- tokenize: tokenizeAttention,
- resolveAll: resolveAllAttention
-}
-
-/**
- * Take all events and resolve attention to emphasis or strong.
- *
- * @type {Resolver}
- */
-function resolveAllAttention(events, context) {
- let index = -1
- /** @type {number} */
- let open
- /** @type {Token} */
- let group
- /** @type {Token} */
- let text
- /** @type {Token} */
- let openingSequence
- /** @type {Token} */
- let closingSequence
- /** @type {number} */
- let use
- /** @type {Array<Event>} */
- let nextEvents
- /** @type {number} */
- let offset
-
- // Walk through all events.
- //
- // Note: performance of this is fine on an mb of normal markdown, but it’s
- // a bottleneck for malicious stuff.
- while (++index < events.length) {
- // Find a token that can close.
- if (
- events[index][0] === 'enter' &&
- events[index][1].type === 'attentionSequence' &&
- events[index][1]._close
- ) {
- open = index
-
- // Now walk back to find an opener.
- while (open--) {
- // Find a token that can open the closer.
- if (
- events[open][0] === 'exit' &&
- events[open][1].type === 'attentionSequence' &&
- events[open][1]._open &&
- // If the markers are the same:
- context.sliceSerialize(events[open][1]).charCodeAt(0) ===
- context.sliceSerialize(events[index][1]).charCodeAt(0)
- ) {
- // If the opening can close or the closing can open,
- // and the close size *is not* a multiple of three,
- // but the sum of the opening and closing size *is* multiple of three,
- // then don’t match.
- if (
- (events[open][1]._close || events[index][1]._open) &&
- (events[index][1].end.offset - events[index][1].start.offset) % 3 &&
- !(
- (events[open][1].end.offset -
- events[open][1].start.offset +
- events[index][1].end.offset -
- events[index][1].start.offset) %
- 3
- )
- ) {
- continue
- }
-
- // Number of markers to use from the sequence.
- use =
- events[open][1].end.offset - events[open][1].start.offset > 1 &&
- events[index][1].end.offset - events[index][1].start.offset > 1
- ? 2
- : 1
- const start = Object.assign({}, events[open][1].end)
- const end = Object.assign({}, events[index][1].start)
- movePoint(start, -use)
- movePoint(end, use)
- openingSequence = {
- type: use > 1 ? 'strongSequence' : 'emphasisSequence',
- start,
- end: Object.assign({}, events[open][1].end)
- }
- closingSequence = {
- type: use > 1 ? 'strongSequence' : 'emphasisSequence',
- start: Object.assign({}, events[index][1].start),
- end
- }
- text = {
- type: use > 1 ? 'strongText' : 'emphasisText',
- start: Object.assign({}, events[open][1].end),
- end: Object.assign({}, events[index][1].start)
- }
- group = {
- type: use > 1 ? 'strong' : 'emphasis',
- start: Object.assign({}, openingSequence.start),
- end: Object.assign({}, closingSequence.end)
- }
- events[open][1].end = Object.assign({}, openingSequence.start)
- events[index][1].start = Object.assign({}, closingSequence.end)
- nextEvents = []
-
- // If there are more markers in the opening, add them before.
- if (events[open][1].end.offset - events[open][1].start.offset) {
- nextEvents = push(nextEvents, [
- ['enter', events[open][1], context],
- ['exit', events[open][1], context]
- ])
- }
-
- // Opening.
- nextEvents = push(nextEvents, [
- ['enter', group, context],
- ['enter', openingSequence, context],
- ['exit', openingSequence, context],
- ['enter', text, context]
- ])
-
- // Always populated by defaults.
-
- // Between.
- nextEvents = push(
- nextEvents,
- resolveAll(
- context.parser.constructs.insideSpan.null,
- events.slice(open + 1, index),
- context
- )
- )
-
- // Closing.
- nextEvents = push(nextEvents, [
- ['exit', text, context],
- ['enter', closingSequence, context],
- ['exit', closingSequence, context],
- ['exit', group, context]
- ])
-
- // If there are more markers in the closing, add them after.
- if (events[index][1].end.offset - events[index][1].start.offset) {
- offset = 2
- nextEvents = push(nextEvents, [
- ['enter', events[index][1], context],
- ['exit', events[index][1], context]
- ])
- } else {
- offset = 0
- }
- splice(events, open - 1, index - open + 3, nextEvents)
- index = open + nextEvents.length - offset - 2
- break
- }
- }
- }
- }
-
- // Remove remaining sequences.
- index = -1
- while (++index < events.length) {
- if (events[index][1].type === 'attentionSequence') {
- events[index][1].type = 'data'
- }
- }
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeAttention(effects, ok) {
- const attentionMarkers = this.parser.constructs.attentionMarkers.null
- const previous = this.previous
- const before = classifyCharacter(previous)
-
- /** @type {NonNullable<Code>} */
- let marker
- return start
-
- /**
- * Before a sequence.
- *
- * ```markdown
- * > | **
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- marker = code
- effects.enter('attentionSequence')
- return inside(code)
- }
-
- /**
- * In a sequence.
- *
- * ```markdown
- * > | **
- * ^^
- * ```
- *
- * @type {State}
- */
- function inside(code) {
- if (code === marker) {
- effects.consume(code)
- return inside
- }
- const token = effects.exit('attentionSequence')
-
- // To do: next major: move this to resolver, just like `markdown-rs`.
- const after = classifyCharacter(code)
-
- // Always populated by defaults.
-
- const open =
- !after || (after === 2 && before) || attentionMarkers.includes(code)
- const close =
- !before || (before === 2 && after) || attentionMarkers.includes(previous)
- token._open = Boolean(marker === 42 ? open : open && (before || !close))
- token._close = Boolean(marker === 42 ? close : close && (after || !open))
- return ok(code)
- }
-}
-
-/**
- * Move a point a bit.
- *
- * Note: `move` only works inside lines! It’s not possible to move past other
- * chunks (replacement characters, tabs, or line endings).
- *
- * @param {Point} point
- * @param {number} offset
- * @returns {void}
- */
-function movePoint(point, offset) {
- point.column += offset
- point.offset += offset
- point._bufferIndex += offset
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/autolink.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-/** @type {Construct} */
-const autolink = {
- name: 'autolink',
- tokenize: tokenizeAutolink
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeAutolink(effects, ok, nok) {
- let size = 0
- return start
-
- /**
- * Start of an autolink.
- *
- * ```markdown
- * > | a<https://example.com>b
- * ^
- * > | a<user@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('autolink')
- effects.enter('autolinkMarker')
- effects.consume(code)
- effects.exit('autolinkMarker')
- effects.enter('autolinkProtocol')
- return open
- }
-
- /**
- * After `<`, at protocol or atext.
- *
- * ```markdown
- * > | a<https://example.com>b
- * ^
- * > | a<user@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function open(code) {
- if (asciiAlpha(code)) {
- effects.consume(code)
- return schemeOrEmailAtext
- }
- return emailAtext(code)
- }
-
- /**
- * At second byte of protocol or atext.
- *
- * ```markdown
- * > | a<https://example.com>b
- * ^
- * > | a<user@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function schemeOrEmailAtext(code) {
- // ASCII alphanumeric and `+`, `-`, and `.`.
- if (code === 43 || code === 45 || code === 46 || asciiAlphanumeric(code)) {
- // Count the previous alphabetical from `open` too.
- size = 1
- return schemeInsideOrEmailAtext(code)
- }
- return emailAtext(code)
- }
-
- /**
- * In ambiguous protocol or atext.
- *
- * ```markdown
- * > | a<https://example.com>b
- * ^
- * > | a<user@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function schemeInsideOrEmailAtext(code) {
- if (code === 58) {
- effects.consume(code)
- size = 0
- return urlInside
- }
-
- // ASCII alphanumeric and `+`, `-`, and `.`.
- if (
- (code === 43 || code === 45 || code === 46 || asciiAlphanumeric(code)) &&
- size++ < 32
- ) {
- effects.consume(code)
- return schemeInsideOrEmailAtext
- }
- size = 0
- return emailAtext(code)
- }
-
- /**
- * After protocol, in URL.
- *
- * ```markdown
- * > | a<https://example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function urlInside(code) {
- if (code === 62) {
- effects.exit('autolinkProtocol')
- effects.enter('autolinkMarker')
- effects.consume(code)
- effects.exit('autolinkMarker')
- effects.exit('autolink')
- return ok
- }
-
- // ASCII control, space, or `<`.
- if (code === null || code === 32 || code === 60 || asciiControl(code)) {
- return nok(code)
- }
- effects.consume(code)
- return urlInside
- }
-
- /**
- * In email atext.
- *
- * ```markdown
- * > | a<user.name@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function emailAtext(code) {
- if (code === 64) {
- effects.consume(code)
- return emailAtSignOrDot
- }
- if (asciiAtext(code)) {
- effects.consume(code)
- return emailAtext
- }
- return nok(code)
- }
-
- /**
- * In label, after at-sign or dot.
- *
- * ```markdown
- * > | a<user.name@example.com>b
- * ^ ^
- * ```
- *
- * @type {State}
- */
- function emailAtSignOrDot(code) {
- return asciiAlphanumeric(code) ? emailLabel(code) : nok(code)
- }
-
- /**
- * In label, where `.` and `>` are allowed.
- *
- * ```markdown
- * > | a<user.name@example.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function emailLabel(code) {
- if (code === 46) {
- effects.consume(code)
- size = 0
- return emailAtSignOrDot
- }
- if (code === 62) {
- // Exit, then change the token type.
- effects.exit('autolinkProtocol').type = 'autolinkEmail'
- effects.enter('autolinkMarker')
- effects.consume(code)
- effects.exit('autolinkMarker')
- effects.exit('autolink')
- return ok
- }
- return emailValue(code)
- }
-
- /**
- * In label, where `.` and `>` are *not* allowed.
- *
- * Though, this is also used in `emailLabel` to parse other values.
- *
- * ```markdown
- * > | a<user.name@ex-ample.com>b
- * ^
- * ```
- *
- * @type {State}
- */
- function emailValue(code) {
- // ASCII alphanumeric or `-`.
- if ((code === 45 || asciiAlphanumeric(code)) && size++ < 63) {
- const next = code === 45 ? emailValue : emailLabel
- effects.consume(code)
- return next
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/html-text.js
-/**
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const htmlText = {
- name: 'htmlText',
- tokenize: tokenizeHtmlText
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeHtmlText(effects, ok, nok) {
- const self = this
- /** @type {NonNullable<Code> | undefined} */
- let marker
- /** @type {number} */
- let index
- /** @type {State} */
- let returnState
- return start
-
- /**
- * Start of HTML (text).
- *
- * ```markdown
- * > | a <b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('htmlText')
- effects.enter('htmlTextData')
- effects.consume(code)
- return open
- }
-
- /**
- * After `<`, at tag name or other stuff.
- *
- * ```markdown
- * > | a <b> c
- * ^
- * > | a <!doctype> c
- * ^
- * > | a <!--b--> c
- * ^
- * ```
- *
- * @type {State}
- */
- function open(code) {
- if (code === 33) {
- effects.consume(code)
- return declarationOpen
- }
- if (code === 47) {
- effects.consume(code)
- return tagCloseStart
- }
- if (code === 63) {
- effects.consume(code)
- return instruction
- }
-
- // ASCII alphabetical.
- if (asciiAlpha(code)) {
- effects.consume(code)
- return tagOpen
- }
- return nok(code)
- }
-
- /**
- * After `<!`, at declaration, comment, or CDATA.
- *
- * ```markdown
- * > | a <!doctype> c
- * ^
- * > | a <!--b--> c
- * ^
- * > | a <![CDATA[>&<]]> c
- * ^
- * ```
- *
- * @type {State}
- */
- function declarationOpen(code) {
- if (code === 45) {
- effects.consume(code)
- return commentOpenInside
- }
- if (code === 91) {
- effects.consume(code)
- index = 0
- return cdataOpenInside
- }
- if (asciiAlpha(code)) {
- effects.consume(code)
- return declaration
- }
- return nok(code)
- }
-
- /**
- * In a comment, after `<!-`, at another `-`.
- *
- * ```markdown
- * > | a <!--b--> c
- * ^
- * ```
- *
- * @type {State}
- */
- function commentOpenInside(code) {
- if (code === 45) {
- effects.consume(code)
- return commentEnd
- }
- return nok(code)
- }
-
- /**
- * In comment.
- *
- * ```markdown
- * > | a <!--b--> c
- * ^
- * ```
- *
- * @type {State}
- */
- function comment(code) {
- if (code === null) {
- return nok(code)
- }
- if (code === 45) {
- effects.consume(code)
- return commentClose
- }
- if (markdownLineEnding(code)) {
- returnState = comment
- return lineEndingBefore(code)
- }
- effects.consume(code)
- return comment
- }
-
- /**
- * In comment, after `-`.
- *
- * ```markdown
- * > | a <!--b--> c
- * ^
- * ```
- *
- * @type {State}
- */
- function commentClose(code) {
- if (code === 45) {
- effects.consume(code)
- return commentEnd
- }
- return comment(code)
- }
-
- /**
- * In comment, after `--`.
- *
- * ```markdown
- * > | a <!--b--> c
- * ^
- * ```
- *
- * @type {State}
- */
- function commentEnd(code) {
- return code === 62
- ? end(code)
- : code === 45
- ? commentClose(code)
- : comment(code)
- }
-
- /**
- * After `<![`, in CDATA, expecting `CDATA[`.
- *
- * ```markdown
- * > | a <![CDATA[>&<]]> b
- * ^^^^^^
- * ```
- *
- * @type {State}
- */
- function cdataOpenInside(code) {
- const value = 'CDATA['
- if (code === value.charCodeAt(index++)) {
- effects.consume(code)
- return index === value.length ? cdata : cdataOpenInside
- }
- return nok(code)
- }
-
- /**
- * In CDATA.
- *
- * ```markdown
- * > | a <![CDATA[>&<]]> b
- * ^^^
- * ```
- *
- * @type {State}
- */
- function cdata(code) {
- if (code === null) {
- return nok(code)
- }
- if (code === 93) {
- effects.consume(code)
- return cdataClose
- }
- if (markdownLineEnding(code)) {
- returnState = cdata
- return lineEndingBefore(code)
- }
- effects.consume(code)
- return cdata
- }
-
- /**
- * In CDATA, after `]`, at another `]`.
- *
- * ```markdown
- * > | a <![CDATA[>&<]]> b
- * ^
- * ```
- *
- * @type {State}
- */
- function cdataClose(code) {
- if (code === 93) {
- effects.consume(code)
- return cdataEnd
- }
- return cdata(code)
- }
-
- /**
- * In CDATA, after `]]`, at `>`.
- *
- * ```markdown
- * > | a <![CDATA[>&<]]> b
- * ^
- * ```
- *
- * @type {State}
- */
- function cdataEnd(code) {
- if (code === 62) {
- return end(code)
- }
- if (code === 93) {
- effects.consume(code)
- return cdataEnd
- }
- return cdata(code)
- }
-
- /**
- * In declaration.
- *
- * ```markdown
- * > | a <!b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function declaration(code) {
- if (code === null || code === 62) {
- return end(code)
- }
- if (markdownLineEnding(code)) {
- returnState = declaration
- return lineEndingBefore(code)
- }
- effects.consume(code)
- return declaration
- }
-
- /**
- * In instruction.
- *
- * ```markdown
- * > | a <?b?> c
- * ^
- * ```
- *
- * @type {State}
- */
- function instruction(code) {
- if (code === null) {
- return nok(code)
- }
- if (code === 63) {
- effects.consume(code)
- return instructionClose
- }
- if (markdownLineEnding(code)) {
- returnState = instruction
- return lineEndingBefore(code)
- }
- effects.consume(code)
- return instruction
- }
-
- /**
- * In instruction, after `?`, at `>`.
- *
- * ```markdown
- * > | a <?b?> c
- * ^
- * ```
- *
- * @type {State}
- */
- function instructionClose(code) {
- return code === 62 ? end(code) : instruction(code)
- }
-
- /**
- * After `</`, in closing tag, at tag name.
- *
- * ```markdown
- * > | a </b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function tagCloseStart(code) {
- // ASCII alphabetical.
- if (asciiAlpha(code)) {
- effects.consume(code)
- return tagClose
- }
- return nok(code)
- }
-
- /**
- * After `</x`, in a tag name.
- *
- * ```markdown
- * > | a </b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function tagClose(code) {
- // ASCII alphanumerical and `-`.
- if (code === 45 || asciiAlphanumeric(code)) {
- effects.consume(code)
- return tagClose
- }
- return tagCloseBetween(code)
- }
-
- /**
- * In closing tag, after tag name.
- *
- * ```markdown
- * > | a </b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function tagCloseBetween(code) {
- if (markdownLineEnding(code)) {
- returnState = tagCloseBetween
- return lineEndingBefore(code)
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return tagCloseBetween
- }
- return end(code)
- }
-
- /**
- * After `<x`, in opening tag name.
- *
- * ```markdown
- * > | a <b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpen(code) {
- // ASCII alphanumerical and `-`.
- if (code === 45 || asciiAlphanumeric(code)) {
- effects.consume(code)
- return tagOpen
- }
- if (code === 47 || code === 62 || markdownLineEndingOrSpace(code)) {
- return tagOpenBetween(code)
- }
- return nok(code)
- }
-
- /**
- * In opening tag, after tag name.
- *
- * ```markdown
- * > | a <b> c
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenBetween(code) {
- if (code === 47) {
- effects.consume(code)
- return end
- }
-
- // ASCII alphabetical and `:` and `_`.
- if (code === 58 || code === 95 || asciiAlpha(code)) {
- effects.consume(code)
- return tagOpenAttributeName
- }
- if (markdownLineEnding(code)) {
- returnState = tagOpenBetween
- return lineEndingBefore(code)
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return tagOpenBetween
- }
- return end(code)
- }
-
- /**
- * In attribute name.
- *
- * ```markdown
- * > | a <b c> d
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeName(code) {
- // ASCII alphabetical and `-`, `.`, `:`, and `_`.
- if (
- code === 45 ||
- code === 46 ||
- code === 58 ||
- code === 95 ||
- asciiAlphanumeric(code)
- ) {
- effects.consume(code)
- return tagOpenAttributeName
- }
- return tagOpenAttributeNameAfter(code)
- }
-
- /**
- * After attribute name, before initializer, the end of the tag, or
- * whitespace.
- *
- * ```markdown
- * > | a <b c> d
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeNameAfter(code) {
- if (code === 61) {
- effects.consume(code)
- return tagOpenAttributeValueBefore
- }
- if (markdownLineEnding(code)) {
- returnState = tagOpenAttributeNameAfter
- return lineEndingBefore(code)
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return tagOpenAttributeNameAfter
- }
- return tagOpenBetween(code)
- }
-
- /**
- * Before unquoted, double quoted, or single quoted attribute value, allowing
- * whitespace.
- *
- * ```markdown
- * > | a <b c=d> e
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeValueBefore(code) {
- if (
- code === null ||
- code === 60 ||
- code === 61 ||
- code === 62 ||
- code === 96
- ) {
- return nok(code)
- }
- if (code === 34 || code === 39) {
- effects.consume(code)
- marker = code
- return tagOpenAttributeValueQuoted
- }
- if (markdownLineEnding(code)) {
- returnState = tagOpenAttributeValueBefore
- return lineEndingBefore(code)
- }
- if (markdownSpace(code)) {
- effects.consume(code)
- return tagOpenAttributeValueBefore
- }
- effects.consume(code)
- return tagOpenAttributeValueUnquoted
- }
-
- /**
- * In double or single quoted attribute value.
- *
- * ```markdown
- * > | a <b c="d"> e
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeValueQuoted(code) {
- if (code === marker) {
- effects.consume(code)
- marker = undefined
- return tagOpenAttributeValueQuotedAfter
- }
- if (code === null) {
- return nok(code)
- }
- if (markdownLineEnding(code)) {
- returnState = tagOpenAttributeValueQuoted
- return lineEndingBefore(code)
- }
- effects.consume(code)
- return tagOpenAttributeValueQuoted
- }
-
- /**
- * In unquoted attribute value.
- *
- * ```markdown
- * > | a <b c=d> e
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeValueUnquoted(code) {
- if (
- code === null ||
- code === 34 ||
- code === 39 ||
- code === 60 ||
- code === 61 ||
- code === 96
- ) {
- return nok(code)
- }
- if (code === 47 || code === 62 || markdownLineEndingOrSpace(code)) {
- return tagOpenBetween(code)
- }
- effects.consume(code)
- return tagOpenAttributeValueUnquoted
- }
-
- /**
- * After double or single quoted attribute value, before whitespace or the end
- * of the tag.
- *
- * ```markdown
- * > | a <b c="d"> e
- * ^
- * ```
- *
- * @type {State}
- */
- function tagOpenAttributeValueQuotedAfter(code) {
- if (code === 47 || code === 62 || markdownLineEndingOrSpace(code)) {
- return tagOpenBetween(code)
- }
- return nok(code)
- }
-
- /**
- * In certain circumstances of a tag where only an `>` is allowed.
- *
- * ```markdown
- * > | a <b c="d"> e
- * ^
- * ```
- *
- * @type {State}
- */
- function end(code) {
- if (code === 62) {
- effects.consume(code)
- effects.exit('htmlTextData')
- effects.exit('htmlText')
- return ok
- }
- return nok(code)
- }
-
- /**
- * At eol.
- *
- * > 👉 **Note**: we can’t have blank lines in text, so no need to worry about
- * > empty tokens.
- *
- * ```markdown
- * > | a <!--a
- * ^
- * | b-->
- * ```
- *
- * @type {State}
- */
- function lineEndingBefore(code) {
- effects.exit('htmlTextData')
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return lineEndingAfter
- }
-
- /**
- * After eol, at optional whitespace.
- *
- * > 👉 **Note**: we can’t have blank lines in text, so no need to worry about
- * > empty tokens.
- *
- * ```markdown
- * | a <!--a
- * > | b-->
- * ^
- * ```
- *
- * @type {State}
- */
- function lineEndingAfter(code) {
- // Always populated by defaults.
-
- return markdownSpace(code)
- ? factorySpace(
- effects,
- lineEndingAfterPrefix,
- 'linePrefix',
- self.parser.constructs.disable.null.includes('codeIndented')
- ? undefined
- : 4
- )(code)
- : lineEndingAfterPrefix(code)
- }
-
- /**
- * After eol, after optional whitespace.
- *
- * > 👉 **Note**: we can’t have blank lines in text, so no need to worry about
- * > empty tokens.
- *
- * ```markdown
- * | a <!--a
- * > | b-->
- * ^
- * ```
- *
- * @type {State}
- */
- function lineEndingAfterPrefix(code) {
- effects.enter('htmlTextData')
- return returnState(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/label-start-link.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-
-/** @type {Construct} */
-const labelStartLink = {
- name: 'labelStartLink',
- tokenize: tokenizeLabelStartLink,
- resolveAll: labelEnd.resolveAll
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeLabelStartLink(effects, ok, nok) {
- const self = this
- return start
-
- /**
- * Start of label (link) start.
- *
- * ```markdown
- * > | a [b] c
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('labelLink')
- effects.enter('labelMarker')
- effects.consume(code)
- effects.exit('labelMarker')
- effects.exit('labelLink')
- return after
- }
-
- /** @type {State} */
- function after(code) {
- // To do: this isn’t needed in `micromark-extension-gfm-footnote`,
- // remove.
- // Hidden footnotes hook.
- /* c8 ignore next 3 */
- return code === 94 && '_hiddenFootnoteSupport' in self.parser.constructs
- ? nok(code)
- : ok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/hard-break-escape.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-/** @type {Construct} */
-const hardBreakEscape = {
- name: 'hardBreakEscape',
- tokenize: tokenizeHardBreakEscape
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeHardBreakEscape(effects, ok, nok) {
- return start
-
- /**
- * Start of a hard break (escape).
- *
- * ```markdown
- * > | a\
- * ^
- * | b
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('hardBreakEscape')
- effects.consume(code)
- return after
- }
-
- /**
- * After `\`, at eol.
- *
- * ```markdown
- * > | a\
- * ^
- * | b
- * ```
- *
- * @type {State}
- */
- function after(code) {
- if (markdownLineEnding(code)) {
- effects.exit('hardBreakEscape')
- return ok(code)
- }
- return nok(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-core-commonmark/lib/code-text.js
-/**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Previous} Previous
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
-
-
-/** @type {Construct} */
-const codeText = {
- name: 'codeText',
- tokenize: tokenizeCodeText,
- resolve: resolveCodeText,
- previous
-}
-
-// To do: next major: don’t resolve, like `markdown-rs`.
-/** @type {Resolver} */
-function resolveCodeText(events) {
- let tailExitIndex = events.length - 4
- let headEnterIndex = 3
- /** @type {number} */
- let index
- /** @type {number | undefined} */
- let enter
-
- // If we start and end with an EOL or a space.
- if (
- (events[headEnterIndex][1].type === 'lineEnding' ||
- events[headEnterIndex][1].type === 'space') &&
- (events[tailExitIndex][1].type === 'lineEnding' ||
- events[tailExitIndex][1].type === 'space')
- ) {
- index = headEnterIndex
-
- // And we have data.
- while (++index < tailExitIndex) {
- if (events[index][1].type === 'codeTextData') {
- // Then we have padding.
- events[headEnterIndex][1].type = 'codeTextPadding'
- events[tailExitIndex][1].type = 'codeTextPadding'
- headEnterIndex += 2
- tailExitIndex -= 2
- break
- }
- }
- }
-
- // Merge adjacent spaces and data.
- index = headEnterIndex - 1
- tailExitIndex++
- while (++index <= tailExitIndex) {
- if (enter === undefined) {
- if (index !== tailExitIndex && events[index][1].type !== 'lineEnding') {
- enter = index
- }
- } else if (
- index === tailExitIndex ||
- events[index][1].type === 'lineEnding'
- ) {
- events[enter][1].type = 'codeTextData'
- if (index !== enter + 2) {
- events[enter][1].end = events[index - 1][1].end
- events.splice(enter + 2, index - enter - 2)
- tailExitIndex -= index - enter - 2
- index = enter + 2
- }
- enter = undefined
- }
- }
- return events
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Previous}
- */
-function previous(code) {
- // If there is a previous code, there will always be a tail.
- return (
- code !== 96 ||
- this.events[this.events.length - 1][1].type === 'characterEscape'
- )
-}
-
-/**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
-function tokenizeCodeText(effects, ok, nok) {
- const self = this
- let sizeOpen = 0
- /** @type {number} */
- let size
- /** @type {Token} */
- let token
- return start
-
- /**
- * Start of code (text).
- *
- * ```markdown
- * > | `a`
- * ^
- * > | \`a`
- * ^
- * ```
- *
- * @type {State}
- */
- function start(code) {
- effects.enter('codeText')
- effects.enter('codeTextSequence')
- return sequenceOpen(code)
- }
-
- /**
- * In opening sequence.
- *
- * ```markdown
- * > | `a`
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceOpen(code) {
- if (code === 96) {
- effects.consume(code)
- sizeOpen++
- return sequenceOpen
- }
- effects.exit('codeTextSequence')
- return between(code)
- }
-
- /**
- * Between something and something else.
- *
- * ```markdown
- * > | `a`
- * ^^
- * ```
- *
- * @type {State}
- */
- function between(code) {
- // EOF.
- if (code === null) {
- return nok(code)
- }
-
- // To do: next major: don’t do spaces in resolve, but when compiling,
- // like `markdown-rs`.
- // Tabs don’t work, and virtual spaces don’t make sense.
- if (code === 32) {
- effects.enter('space')
- effects.consume(code)
- effects.exit('space')
- return between
- }
-
- // Closing fence? Could also be data.
- if (code === 96) {
- token = effects.enter('codeTextSequence')
- size = 0
- return sequenceClose(code)
- }
- if (markdownLineEnding(code)) {
- effects.enter('lineEnding')
- effects.consume(code)
- effects.exit('lineEnding')
- return between
- }
-
- // Data.
- effects.enter('codeTextData')
- return data(code)
- }
-
- /**
- * In data.
- *
- * ```markdown
- * > | `a`
- * ^
- * ```
- *
- * @type {State}
- */
- function data(code) {
- if (
- code === null ||
- code === 32 ||
- code === 96 ||
- markdownLineEnding(code)
- ) {
- effects.exit('codeTextData')
- return between(code)
- }
- effects.consume(code)
- return data
- }
-
- /**
- * In closing sequence.
- *
- * ```markdown
- * > | `a`
- * ^
- * ```
- *
- * @type {State}
- */
- function sequenceClose(code) {
- // More.
- if (code === 96) {
- effects.consume(code)
- size++
- return sequenceClose
- }
-
- // Done!
- if (size === sizeOpen) {
- effects.exit('codeTextSequence')
- effects.exit('codeText')
- return ok(code)
- }
-
- // More or less accents: mark as data.
- token.type = 'codeTextData'
- return data(code)
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/constructs.js
-/**
- * @typedef {import('micromark-util-types').Extension} Extension
- */
-
-
-
-
-/** @satisfies {Extension['document']} */
-const constructs_document = {
- [42]: list,
- [43]: list,
- [45]: list,
- [48]: list,
- [49]: list,
- [50]: list,
- [51]: list,
- [52]: list,
- [53]: list,
- [54]: list,
- [55]: list,
- [56]: list,
- [57]: list,
- [62]: blockQuote
-}
-
-/** @satisfies {Extension['contentInitial']} */
-const contentInitial = {
- [91]: definition
-}
-
-/** @satisfies {Extension['flowInitial']} */
-const flowInitial = {
- [-2]: codeIndented,
- [-1]: codeIndented,
- [32]: codeIndented
-}
-
-/** @satisfies {Extension['flow']} */
-const constructs_flow = {
- [35]: headingAtx,
- [42]: thematicBreak,
- [45]: [setextUnderline, thematicBreak],
- [60]: htmlFlow,
- [61]: setextUnderline,
- [95]: thematicBreak,
- [96]: codeFenced,
- [126]: codeFenced
-}
-
-/** @satisfies {Extension['string']} */
-const constructs_string = {
- [38]: characterReference,
- [92]: characterEscape
-}
-
-/** @satisfies {Extension['text']} */
-const constructs_text = {
- [-5]: lineEnding,
- [-4]: lineEnding,
- [-3]: lineEnding,
- [33]: labelStartImage,
- [38]: characterReference,
- [42]: attention,
- [60]: [autolink, htmlText],
- [91]: labelStartLink,
- [92]: [hardBreakEscape, characterEscape],
- [93]: labelEnd,
- [95]: attention,
- [96]: codeText
-}
-
-/** @satisfies {Extension['insideSpan']} */
-const insideSpan = {
- null: [attention, resolver]
-}
-
-/** @satisfies {Extension['attentionMarkers']} */
-const attentionMarkers = {
- null: [42, 95]
-}
-
-/** @satisfies {Extension['disable']} */
-const disable = {
- null: []
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/parse.js
-/**
- * @typedef {import('micromark-util-types').Create} Create
- * @typedef {import('micromark-util-types').FullNormalizedExtension} FullNormalizedExtension
- * @typedef {import('micromark-util-types').InitialConstruct} InitialConstruct
- * @typedef {import('micromark-util-types').ParseContext} ParseContext
- * @typedef {import('micromark-util-types').ParseOptions} ParseOptions
- */
-
-
-
-
-
-
-
-
-
-/**
- * @param {ParseOptions | null | undefined} [options]
- * @returns {ParseContext}
- */
-function parse(options) {
- const settings = options || {}
- const constructs =
- /** @type {FullNormalizedExtension} */
- combineExtensions([constructs_namespaceObject, ...(settings.extensions || [])])
-
- /** @type {ParseContext} */
- const parser = {
- defined: [],
- lazy: {},
- constructs,
- content: create(content),
- document: create(document_document),
- flow: create(flow),
- string: create(string),
- text: create(text_text)
- }
- return parser
-
- /**
- * @param {InitialConstruct} initial
- */
- function create(initial) {
- return creator
- /** @type {Create} */
- function creator(from) {
- return createTokenizer(parser, initial, from)
- }
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/preprocess.js
-/**
- * @typedef {import('micromark-util-types').Chunk} Chunk
- * @typedef {import('micromark-util-types').Code} Code
- * @typedef {import('micromark-util-types').Encoding} Encoding
- * @typedef {import('micromark-util-types').Value} Value
- */
-
-/**
- * @callback Preprocessor
- * @param {Value} value
- * @param {Encoding | null | undefined} [encoding]
- * @param {boolean | null | undefined} [end=false]
- * @returns {Array<Chunk>}
- */
-
-const search = /[\0\t\n\r]/g
-
-/**
- * @returns {Preprocessor}
- */
-function preprocess() {
- let column = 1
- let buffer = ''
- /** @type {boolean | undefined} */
- let start = true
- /** @type {boolean | undefined} */
- let atCarriageReturn
- return preprocessor
-
- /** @type {Preprocessor} */
- function preprocessor(value, encoding, end) {
- /** @type {Array<Chunk>} */
- const chunks = []
- /** @type {RegExpMatchArray | null} */
- let match
- /** @type {number} */
- let next
- /** @type {number} */
- let startPosition
- /** @type {number} */
- let endPosition
- /** @type {Code} */
- let code
-
- // @ts-expect-error `Buffer` does allow an encoding.
- value = buffer + value.toString(encoding)
- startPosition = 0
- buffer = ''
- if (start) {
- // To do: `markdown-rs` actually parses BOMs (byte order mark).
- if (value.charCodeAt(0) === 65279) {
- startPosition++
- }
- start = undefined
- }
- while (startPosition < value.length) {
- search.lastIndex = startPosition
- match = search.exec(value)
- endPosition =
- match && match.index !== undefined ? match.index : value.length
- code = value.charCodeAt(endPosition)
- if (!match) {
- buffer = value.slice(startPosition)
- break
- }
- if (code === 10 && startPosition === endPosition && atCarriageReturn) {
- chunks.push(-3)
- atCarriageReturn = undefined
- } else {
- if (atCarriageReturn) {
- chunks.push(-5)
- atCarriageReturn = undefined
- }
- if (startPosition < endPosition) {
- chunks.push(value.slice(startPosition, endPosition))
- column += endPosition - startPosition
- }
- switch (code) {
- case 0: {
- chunks.push(65533)
- column++
- break
- }
- case 9: {
- next = Math.ceil(column / 4) * 4
- chunks.push(-2)
- while (column++ < next) chunks.push(-1)
- break
- }
- case 10: {
- chunks.push(-4)
- column = 1
- break
- }
- default: {
- atCarriageReturn = true
- column = 1
- }
- }
- }
- startPosition = endPosition + 1
- }
- if (end) {
- if (atCarriageReturn) chunks.push(-5)
- if (buffer) chunks.push(buffer)
- chunks.push(null)
- }
- return chunks
- }
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark/lib/postprocess.js
-/**
- * @typedef {import('micromark-util-types').Event} Event
- */
-
-
-
-/**
- * @param {Array<Event>} events
- * @returns {Array<Event>}
- */
-function postprocess(events) {
- while (!subtokenize(events)) {
- // Empty
- }
- return events
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-decode-numeric-character-reference/index.js
-/**
- * Turn the number (in string form as either hexa- or plain decimal) coming from
- * a numeric character reference into a character.
- *
- * Sort of like `String.fromCharCode(Number.parseInt(value, base))`, but makes
- * non-characters and control characters safe.
- *
- * @param {string} value
- * Value to decode.
- * @param {number} base
- * Numeric base.
- * @returns {string}
- * Character.
- */
-function decodeNumericCharacterReference(value, base) {
- const code = Number.parseInt(value, base)
- if (
- // C0 except for HT, LF, FF, CR, space.
- code < 9 ||
- code === 11 ||
- (code > 13 && code < 32) ||
- // Control character (DEL) of C0, and C1 controls.
- (code > 126 && code < 160) ||
- // Lone high surrogates and low surrogates.
- (code > 55295 && code < 57344) ||
- // Noncharacters.
- (code > 64975 && code < 65008) /* eslint-disable no-bitwise */ ||
- (code & 65535) === 65535 ||
- (code & 65535) === 65534 /* eslint-enable no-bitwise */ ||
- // Out of range
- code > 1114111
- ) {
- return '\uFFFD'
- }
- return String.fromCharCode(code)
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/micromark-util-decode-string/index.js
-
-
-const characterEscapeOrReference =
- /\\([!-/:-@[-`{-~])|&(#(?:\d{1,7}|x[\da-f]{1,6})|[\da-z]{1,31});/gi
-
-/**
- * Decode markdown strings (which occur in places such as fenced code info
- * strings, destinations, labels, and titles).
- *
- * The “string” content type allows character escapes and -references.
- * This decodes those.
- *
- * @param {string} value
- * Value to decode.
- * @returns {string}
- * Decoded value.
- */
-function decodeString(value) {
- return value.replace(characterEscapeOrReference, decode)
-}
-
-/**
- * @param {string} $0
- * @param {string} $1
- * @param {string} $2
- * @returns {string}
- */
-function decode($0, $1, $2) {
- if ($1) {
- // Escape.
- return $1
- }
-
- // Reference.
- const head = $2.charCodeAt(0)
- if (head === 35) {
- const head = $2.charCodeAt(1)
- const hex = head === 120 || head === 88
- return decodeNumericCharacterReference($2.slice(hex ? 2 : 1), hex ? 16 : 10)
- }
- return decodeNamedCharacterReference($2) || $0
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/unist-util-stringify-position/lib/index.js
-/**
- * @typedef {import('unist').Node} Node
- * @typedef {import('unist').Point} Point
- * @typedef {import('unist').Position} Position
- */
-
-/**
- * @typedef NodeLike
- * @property {string} type
- * @property {PositionLike | null | undefined} [position]
- *
- * @typedef PositionLike
- * @property {PointLike | null | undefined} [start]
- * @property {PointLike | null | undefined} [end]
- *
- * @typedef PointLike
- * @property {number | null | undefined} [line]
- * @property {number | null | undefined} [column]
- * @property {number | null | undefined} [offset]
- */
-
-/**
- * Serialize the positional info of a point, position (start and end points),
- * or node.
- *
- * @param {Node | NodeLike | Position | PositionLike | Point | PointLike | null | undefined} [value]
- * Node, position, or point.
- * @returns {string}
- * Pretty printed positional info of a node (`string`).
- *
- * In the format of a range `ls:cs-le:ce` (when given `node` or `position`)
- * or a point `l:c` (when given `point`), where `l` stands for line, `c` for
- * column, `s` for `start`, and `e` for end.
- * An empty string (`''`) is returned if the given value is neither `node`,
- * `position`, nor `point`.
- */
-function stringifyPosition(value) {
- // Nothing.
- if (!value || typeof value !== 'object') {
- return ''
- }
-
- // Node.
- if ('position' in value || 'type' in value) {
- return position(value.position)
- }
-
- // Position.
- if ('start' in value || 'end' in value) {
- return position(value)
- }
-
- // Point.
- if ('line' in value || 'column' in value) {
- return point(value)
- }
-
- // ?
- return ''
-}
-
-/**
- * @param {Point | PointLike | null | undefined} point
- * @returns {string}
- */
-function point(point) {
- return index(point && point.line) + ':' + index(point && point.column)
-}
-
-/**
- * @param {Position | PositionLike | null | undefined} pos
- * @returns {string}
- */
-function position(pos) {
- return point(pos && pos.start) + '-' + point(pos && pos.end)
-}
-
-/**
- * @param {number | null | undefined} value
- * @returns {number}
- */
-function index(value) {
- return value && typeof value === 'number' ? value : 1
-}
-
-;// CONCATENATED MODULE: ./node_modules/mermaid/node_modules/mdast-util-from-markdown/lib/index.js
-/**
- * @typedef {import('micromark-util-types').Encoding} Encoding
- * @typedef {import('micromark-util-types').Event} Event
- * @typedef {import('micromark-util-types').ParseOptions} ParseOptions
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Value} Value
- *
- * @typedef {import('unist').Parent} UnistParent
- * @typedef {import('unist').Point} Point
- *
- * @typedef {import('mdast').PhrasingContent} PhrasingContent
- * @typedef {import('mdast').StaticPhrasingContent} StaticPhrasingContent
- * @typedef {import('mdast').Content} Content
- * @typedef {import('mdast').Break} Break
- * @typedef {import('mdast').Blockquote} Blockquote
- * @typedef {import('mdast').Code} Code
- * @typedef {import('mdast').Definition} Definition
- * @typedef {import('mdast').Emphasis} Emphasis
- * @typedef {import('mdast').Heading} Heading
- * @typedef {import('mdast').HTML} HTML
- * @typedef {import('mdast').Image} Image
- * @typedef {import('mdast').ImageReference} ImageReference
- * @typedef {import('mdast').InlineCode} InlineCode
- * @typedef {import('mdast').Link} Link
- * @typedef {import('mdast').LinkReference} LinkReference
- * @typedef {import('mdast').List} List
- * @typedef {import('mdast').ListItem} ListItem
- * @typedef {import('mdast').Paragraph} Paragraph
- * @typedef {import('mdast').Root} Root
- * @typedef {import('mdast').Strong} Strong
- * @typedef {import('mdast').Text} Text
- * @typedef {import('mdast').ThematicBreak} ThematicBreak
- * @typedef {import('mdast').ReferenceType} ReferenceType
- * @typedef {import('../index.js').CompileData} CompileData
- */
-
-/**
- * @typedef {Root | Content} Node
- * @typedef {Extract<Node, UnistParent>} Parent
- *
- * @typedef {Omit<UnistParent, 'type' | 'children'> & {type: 'fragment', children: Array<PhrasingContent>}} Fragment
- */
-
-/**
- * @callback Transform
- * Extra transform, to change the AST afterwards.
- * @param {Root} tree
- * Tree to transform.
- * @returns {Root | undefined | null | void}
- * New tree or nothing (in which case the current tree is used).
- *
- * @callback Handle
- * Handle a token.
- * @param {CompileContext} this
- * Context.
- * @param {Token} token
- * Current token.
- * @returns {void}
- * Nothing.
- *
- * @typedef {Record<string, Handle>} Handles
- * Token types mapping to handles
- *
- * @callback OnEnterError
- * Handle the case where the `right` token is open, but it is closed (by the
- * `left` token) or because we reached the end of the document.
- * @param {Omit<CompileContext, 'sliceSerialize'>} this
- * Context.
- * @param {Token | undefined} left
- * Left token.
- * @param {Token} right
- * Right token.
- * @returns {void}
- * Nothing.
- *
- * @callback OnExitError
- * Handle the case where the `right` token is open but it is closed by
- * exiting the `left` token.
- * @param {Omit<CompileContext, 'sliceSerialize'>} this
- * Context.
- * @param {Token} left
- * Left token.
- * @param {Token} right
- * Right token.
- * @returns {void}
- * Nothing.
- *
- * @typedef {[Token, OnEnterError | undefined]} TokenTuple
- * Open token on the stack, with an optional error handler for when
- * that token isn’t closed properly.
- */
-
-/**
- * @typedef Config
- * Configuration.
- *
- * We have our defaults, but extensions will add more.
- * @property {Array<string>} canContainEols
- * Token types where line endings are used.
- * @property {Handles} enter
- * Opening handles.
- * @property {Handles} exit
- * Closing handles.
- * @property {Array<Transform>} transforms
- * Tree transforms.
- *
- * @typedef {Partial<Config>} Extension
- * Change how markdown tokens from micromark are turned into mdast.
- *
- * @typedef CompileContext
- * mdast compiler context.
- * @property {Array<Node | Fragment>} stack
- * Stack of nodes.
- * @property {Array<TokenTuple>} tokenStack
- * Stack of tokens.
- * @property {<Key extends keyof CompileData>(key: Key) => CompileData[Key]} getData
- * Get data from the key/value store.
- * @property {<Key extends keyof CompileData>(key: Key, value?: CompileData[Key]) => void} setData
- * Set data into the key/value store.
- * @property {(this: CompileContext) => void} buffer
- * Capture some of the output data.
- * @property {(this: CompileContext) => string} resume
- * Stop capturing and access the output data.
- * @property {<Kind extends Node>(this: CompileContext, node: Kind, token: Token, onError?: OnEnterError) => Kind} enter
- * Enter a token.
- * @property {(this: CompileContext, token: Token, onError?: OnExitError) => Node} exit
- * Exit a token.
- * @property {TokenizeContext['sliceSerialize']} sliceSerialize
- * Get the string value of a token.
- * @property {Config} config
- * Configuration.
- *
- * @typedef FromMarkdownOptions
- * Configuration for how to build mdast.
- * @property {Array<Extension | Array<Extension>> | null | undefined} [mdastExtensions]
- * Extensions for this utility to change how tokens are turned into a tree.
- *
- * @typedef {ParseOptions & FromMarkdownOptions} Options
- * Configuration.
- */
-
-// To do: micromark: create a registry of tokens?
-// To do: next major: don’t return given `Node` from `enter`.
-// To do: next major: remove setter/getter.
-
-
-
-
-
-
-
-
-
-
-const lib_own = {}.hasOwnProperty
-
-/**
- * @param value
- * Markdown to parse.
- * @param encoding
- * Character encoding for when `value` is `Buffer`.
- * @param options
- * Configuration.
- * @returns
- * mdast tree.
- */
-const fromMarkdown =
- /**
- * @type {(
- * ((value: Value, encoding: Encoding, options?: Options | null | undefined) => Root) &
- * ((value: Value, options?: Options | null | undefined) => Root)
- * )}
- */
-
- /**
- * @param {Value} value
- * @param {Encoding | Options | null | undefined} [encoding]
- * @param {Options | null | undefined} [options]
- * @returns {Root}
- */
- function (value, encoding, options) {
- if (typeof encoding !== 'string') {
- options = encoding
- encoding = undefined
- }
- return compiler(options)(
- postprocess(
- parse(options).document().write(preprocess()(value, encoding, true))
- )
- )
- }
-
-/**
- * Note this compiler only understand complete buffering, not streaming.
- *
- * @param {Options | null | undefined} [options]
- */
-function compiler(options) {
- /** @type {Config} */
- const config = {
- transforms: [],
- canContainEols: ['emphasis', 'fragment', 'heading', 'paragraph', 'strong'],
- enter: {
- autolink: opener(link),
- autolinkProtocol: onenterdata,
- autolinkEmail: onenterdata,
- atxHeading: opener(heading),
- blockQuote: opener(blockQuote),
- characterEscape: onenterdata,
- characterReference: onenterdata,
- codeFenced: opener(codeFlow),
- codeFencedFenceInfo: buffer,
- codeFencedFenceMeta: buffer,
- codeIndented: opener(codeFlow, buffer),
- codeText: opener(codeText, buffer),
- codeTextData: onenterdata,
- data: onenterdata,
- codeFlowValue: onenterdata,
- definition: opener(definition),
- definitionDestinationString: buffer,
- definitionLabelString: buffer,
- definitionTitleString: buffer,
- emphasis: opener(emphasis),
- hardBreakEscape: opener(hardBreak),
- hardBreakTrailing: opener(hardBreak),
- htmlFlow: opener(html, buffer),
- htmlFlowData: onenterdata,
- htmlText: opener(html, buffer),
- htmlTextData: onenterdata,
- image: opener(image),
- label: buffer,
- link: opener(link),
- listItem: opener(listItem),
- listItemValue: onenterlistitemvalue,
- listOrdered: opener(list, onenterlistordered),
- listUnordered: opener(list),
- paragraph: opener(paragraph),
- reference: onenterreference,
- referenceString: buffer,
- resourceDestinationString: buffer,
- resourceTitleString: buffer,
- setextHeading: opener(heading),
- strong: opener(strong),
- thematicBreak: opener(thematicBreak)
- },
- exit: {
- atxHeading: closer(),
- atxHeadingSequence: onexitatxheadingsequence,
- autolink: closer(),
- autolinkEmail: onexitautolinkemail,
- autolinkProtocol: onexitautolinkprotocol,
- blockQuote: closer(),
- characterEscapeValue: onexitdata,
- characterReferenceMarkerHexadecimal: onexitcharacterreferencemarker,
- characterReferenceMarkerNumeric: onexitcharacterreferencemarker,
- characterReferenceValue: onexitcharacterreferencevalue,
- codeFenced: closer(onexitcodefenced),
- codeFencedFence: onexitcodefencedfence,
- codeFencedFenceInfo: onexitcodefencedfenceinfo,
- codeFencedFenceMeta: onexitcodefencedfencemeta,
- codeFlowValue: onexitdata,
- codeIndented: closer(onexitcodeindented),
- codeText: closer(onexitcodetext),
- codeTextData: onexitdata,
- data: onexitdata,
- definition: closer(),
- definitionDestinationString: onexitdefinitiondestinationstring,
- definitionLabelString: onexitdefinitionlabelstring,
- definitionTitleString: onexitdefinitiontitlestring,
- emphasis: closer(),
- hardBreakEscape: closer(onexithardbreak),
- hardBreakTrailing: closer(onexithardbreak),
- htmlFlow: closer(onexithtmlflow),
- htmlFlowData: onexitdata,
- htmlText: closer(onexithtmltext),
- htmlTextData: onexitdata,
- image: closer(onexitimage),
- label: onexitlabel,
- labelText: onexitlabeltext,
- lineEnding: onexitlineending,
- link: closer(onexitlink),
- listItem: closer(),
- listOrdered: closer(),
- listUnordered: closer(),
- paragraph: closer(),
- referenceString: onexitreferencestring,
- resourceDestinationString: onexitresourcedestinationstring,
- resourceTitleString: onexitresourcetitlestring,
- resource: onexitresource,
- setextHeading: closer(onexitsetextheading),
- setextHeadingLineSequence: onexitsetextheadinglinesequence,
- setextHeadingText: onexitsetextheadingtext,
- strong: closer(),
- thematicBreak: closer()
- }
- }
- configure(config, (options || {}).mdastExtensions || [])
-
- /** @type {CompileData} */
- const data = {}
- return compile
-
- /**
- * Turn micromark events into an mdast tree.
- *
- * @param {Array<Event>} events
- * Events.
- * @returns {Root}
- * mdast tree.
- */
- function compile(events) {
- /** @type {Root} */
- let tree = {
- type: 'root',
- children: []
- }
- /** @type {Omit<CompileContext, 'sliceSerialize'>} */
- const context = {
- stack: [tree],
- tokenStack: [],
- config,
- enter,
- exit,
- buffer,
- resume,
- setData,
- getData
- }
- /** @type {Array<number>} */
- const listStack = []
- let index = -1
- while (++index < events.length) {
- // We preprocess lists to add `listItem` tokens, and to infer whether
- // items the list itself are spread out.
- if (
- events[index][1].type === 'listOrdered' ||
- events[index][1].type === 'listUnordered'
- ) {
- if (events[index][0] === 'enter') {
- listStack.push(index)
- } else {
- const tail = listStack.pop()
- index = prepareList(events, tail, index)
- }
- }
- }
- index = -1
- while (++index < events.length) {
- const handler = config[events[index][0]]
- if (lib_own.call(handler, events[index][1].type)) {
- handler[events[index][1].type].call(
- Object.assign(
- {
- sliceSerialize: events[index][2].sliceSerialize
- },
- context
- ),
- events[index][1]
- )
- }
- }
-
- // Handle tokens still being open.
- if (context.tokenStack.length > 0) {
- const tail = context.tokenStack[context.tokenStack.length - 1]
- const handler = tail[1] || defaultOnError
- handler.call(context, undefined, tail[0])
- }
-
- // Figure out `root` position.
- tree.position = {
- start: lib_point(
- events.length > 0
- ? events[0][1].start
- : {
- line: 1,
- column: 1,
- offset: 0
- }
- ),
- end: lib_point(
- events.length > 0
- ? events[events.length - 2][1].end
- : {
- line: 1,
- column: 1,
- offset: 0
- }
- )
- }
-
- // Call transforms.
- index = -1
- while (++index < config.transforms.length) {
- tree = config.transforms[index](tree) || tree
- }
- return tree
- }
-
- /**
- * @param {Array<Event>} events
- * @param {number} start
- * @param {number} length
- * @returns {number}
- */
- function prepareList(events, start, length) {
- let index = start - 1
- let containerBalance = -1
- let listSpread = false
- /** @type {Token | undefined} */
- let listItem
- /** @type {number | undefined} */
- let lineIndex
- /** @type {number | undefined} */
- let firstBlankLineIndex
- /** @type {boolean | undefined} */
- let atMarker
- while (++index <= length) {
- const event = events[index]
- if (
- event[1].type === 'listUnordered' ||
- event[1].type === 'listOrdered' ||
- event[1].type === 'blockQuote'
- ) {
- if (event[0] === 'enter') {
- containerBalance++
- } else {
- containerBalance--
- }
- atMarker = undefined
- } else if (event[1].type === 'lineEndingBlank') {
- if (event[0] === 'enter') {
- if (
- listItem &&
- !atMarker &&
- !containerBalance &&
- !firstBlankLineIndex
- ) {
- firstBlankLineIndex = index
- }
- atMarker = undefined
- }
- } else if (
- event[1].type === 'linePrefix' ||
- event[1].type === 'listItemValue' ||
- event[1].type === 'listItemMarker' ||
- event[1].type === 'listItemPrefix' ||
- event[1].type === 'listItemPrefixWhitespace'
- ) {
- // Empty.
- } else {
- atMarker = undefined
- }
- if (
- (!containerBalance &&
- event[0] === 'enter' &&
- event[1].type === 'listItemPrefix') ||
- (containerBalance === -1 &&
- event[0] === 'exit' &&
- (event[1].type === 'listUnordered' ||
- event[1].type === 'listOrdered'))
- ) {
- if (listItem) {
- let tailIndex = index
- lineIndex = undefined
- while (tailIndex--) {
- const tailEvent = events[tailIndex]
- if (
- tailEvent[1].type === 'lineEnding' ||
- tailEvent[1].type === 'lineEndingBlank'
- ) {
- if (tailEvent[0] === 'exit') continue
- if (lineIndex) {
- events[lineIndex][1].type = 'lineEndingBlank'
- listSpread = true
- }
- tailEvent[1].type = 'lineEnding'
- lineIndex = tailIndex
- } else if (
- tailEvent[1].type === 'linePrefix' ||
- tailEvent[1].type === 'blockQuotePrefix' ||
- tailEvent[1].type === 'blockQuotePrefixWhitespace' ||
- tailEvent[1].type === 'blockQuoteMarker' ||
- tailEvent[1].type === 'listItemIndent'
- ) {
- // Empty
- } else {
- break
- }
- }
- if (
- firstBlankLineIndex &&
- (!lineIndex || firstBlankLineIndex < lineIndex)
- ) {
- listItem._spread = true
- }
-
- // Fix position.
- listItem.end = Object.assign(
- {},
- lineIndex ? events[lineIndex][1].start : event[1].end
- )
- events.splice(lineIndex || index, 0, ['exit', listItem, event[2]])
- index++
- length++
- }
-
- // Create a new list item.
- if (event[1].type === 'listItemPrefix') {
- listItem = {
- type: 'listItem',
- _spread: false,
- start: Object.assign({}, event[1].start),
- // @ts-expect-error: we’ll add `end` in a second.
- end: undefined
- }
- // @ts-expect-error: `listItem` is most definitely defined, TS...
- events.splice(index, 0, ['enter', listItem, event[2]])
- index++
- length++
- firstBlankLineIndex = undefined
- atMarker = true
- }
- }
- }
- events[start][1]._spread = listSpread
- return length
- }
-
- /**
- * Set data.
- *
- * @template {keyof CompileData} Key
- * Field type.
- * @param {Key} key
- * Key of field.
- * @param {CompileData[Key]} [value]
- * New value.
- * @returns {void}
- * Nothing.
- */
- function setData(key, value) {
- data[key] = value
- }
-
- /**
- * Get data.
- *
- * @template {keyof CompileData} Key
- * Field type.
- * @param {Key} key
- * Key of field.
- * @returns {CompileData[Key]}
- * Value.
- */
- function getData(key) {
- return data[key]
- }
-
- /**
- * Create an opener handle.
- *
- * @param {(token: Token) => Node} create
- * Create a node.
- * @param {Handle} [and]
- * Optional function to also run.
- * @returns {Handle}
- * Handle.
- */
- function opener(create, and) {
- return open
-
- /**
- * @this {CompileContext}
- * @param {Token} token
- * @returns {void}
- */
- function open(token) {
- enter.call(this, create(token), token)
- if (and) and.call(this, token)
- }
- }
-
- /**
- * @this {CompileContext}
- * @returns {void}
- */
- function buffer() {
- this.stack.push({
- type: 'fragment',
- children: []
- })
- }
-
- /**
- * @template {Node} Kind
- * Node type.
- * @this {CompileContext}
- * Context.
- * @param {Kind} node
- * Node to enter.
- * @param {Token} token
- * Corresponding token.
- * @param {OnEnterError | undefined} [errorHandler]
- * Handle the case where this token is open, but it is closed by something else.
- * @returns {Kind}
- * The given node.
- */
- function enter(node, token, errorHandler) {
- const parent = this.stack[this.stack.length - 1]
- // @ts-expect-error: Assume `Node` can exist as a child of `parent`.
- parent.children.push(node)
- this.stack.push(node)
- this.tokenStack.push([token, errorHandler])
- // @ts-expect-error: `end` will be patched later.
- node.position = {
- start: lib_point(token.start)
- }
- return node
- }
-
- /**
- * Create a closer handle.
- *
- * @param {Handle} [and]
- * Optional function to also run.
- * @returns {Handle}
- * Handle.
- */
- function closer(and) {
- return close
-
- /**
- * @this {CompileContext}
- * @param {Token} token
- * @returns {void}
- */
- function close(token) {
- if (and) and.call(this, token)
- exit.call(this, token)
- }
- }
-
- /**
- * @this {CompileContext}
- * Context.
- * @param {Token} token
- * Corresponding token.
- * @param {OnExitError | undefined} [onExitError]
- * Handle the case where another token is open.
- * @returns {Node}
- * The closed node.
- */
- function exit(token, onExitError) {
- const node = this.stack.pop()
- const open = this.tokenStack.pop()
- if (!open) {
- throw new Error(
- 'Cannot close `' +
- token.type +
- '` (' +
- stringifyPosition({
- start: token.start,
- end: token.end
- }) +
- '): it’s not open'
- )
- } else if (open[0].type !== token.type) {
- if (onExitError) {
- onExitError.call(this, token, open[0])
- } else {
- const handler = open[1] || defaultOnError
- handler.call(this, token, open[0])
- }
- }
- node.position.end = lib_point(token.end)
- return node
- }
-
- /**
- * @this {CompileContext}
- * @returns {string}
- */
- function resume() {
- return lib_toString(this.stack.pop())
- }
-
- //
- // Handlers.
- //
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onenterlistordered() {
- setData('expectingFirstListItemValue', true)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onenterlistitemvalue(token) {
- if (getData('expectingFirstListItemValue')) {
- const ancestor = this.stack[this.stack.length - 2]
- ancestor.start = Number.parseInt(this.sliceSerialize(token), 10)
- setData('expectingFirstListItemValue')
- }
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcodefencedfenceinfo() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.lang = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcodefencedfencemeta() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.meta = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcodefencedfence() {
- // Exit if this is the closing fence.
- if (getData('flowCodeInside')) return
- this.buffer()
- setData('flowCodeInside', true)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcodefenced() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.value = data.replace(/^(\r?\n|\r)|(\r?\n|\r)$/g, '')
- setData('flowCodeInside')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcodeindented() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.value = data.replace(/(\r?\n|\r)$/g, '')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitdefinitionlabelstring(token) {
- const label = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.label = label
- node.identifier = normalizeIdentifier(
- this.sliceSerialize(token)
- ).toLowerCase()
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitdefinitiontitlestring() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.title = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitdefinitiondestinationstring() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.url = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitatxheadingsequence(token) {
- const node = this.stack[this.stack.length - 1]
- if (!node.depth) {
- const depth = this.sliceSerialize(token).length
- node.depth = depth
- }
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitsetextheadingtext() {
- setData('setextHeadingSlurpLineEnding', true)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitsetextheadinglinesequence(token) {
- const node = this.stack[this.stack.length - 1]
- node.depth = this.sliceSerialize(token).charCodeAt(0) === 61 ? 1 : 2
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitsetextheading() {
- setData('setextHeadingSlurpLineEnding')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onenterdata(token) {
- const node = this.stack[this.stack.length - 1]
- let tail = node.children[node.children.length - 1]
- if (!tail || tail.type !== 'text') {
- // Add a new text node.
- tail = text()
- // @ts-expect-error: we’ll add `end` later.
- tail.position = {
- start: lib_point(token.start)
- }
- // @ts-expect-error: Assume `parent` accepts `text`.
- node.children.push(tail)
- }
- this.stack.push(tail)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitdata(token) {
- const tail = this.stack.pop()
- tail.value += this.sliceSerialize(token)
- tail.position.end = lib_point(token.end)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitlineending(token) {
- const context = this.stack[this.stack.length - 1]
- // If we’re at a hard break, include the line ending in there.
- if (getData('atHardBreak')) {
- const tail = context.children[context.children.length - 1]
- tail.position.end = lib_point(token.end)
- setData('atHardBreak')
- return
- }
- if (
- !getData('setextHeadingSlurpLineEnding') &&
- config.canContainEols.includes(context.type)
- ) {
- onenterdata.call(this, token)
- onexitdata.call(this, token)
- }
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexithardbreak() {
- setData('atHardBreak', true)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexithtmlflow() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.value = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexithtmltext() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.value = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitcodetext() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.value = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitlink() {
- const node = this.stack[this.stack.length - 1]
- // Note: there are also `identifier` and `label` fields on this link node!
- // These are used / cleaned here.
- // To do: clean.
- if (getData('inReference')) {
- /** @type {ReferenceType} */
- const referenceType = getData('referenceType') || 'shortcut'
- node.type += 'Reference'
- // @ts-expect-error: mutate.
- node.referenceType = referenceType
- // @ts-expect-error: mutate.
- delete node.url
- delete node.title
- } else {
- // @ts-expect-error: mutate.
- delete node.identifier
- // @ts-expect-error: mutate.
- delete node.label
- }
- setData('referenceType')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitimage() {
- const node = this.stack[this.stack.length - 1]
- // Note: there are also `identifier` and `label` fields on this link node!
- // These are used / cleaned here.
- // To do: clean.
- if (getData('inReference')) {
- /** @type {ReferenceType} */
- const referenceType = getData('referenceType') || 'shortcut'
- node.type += 'Reference'
- // @ts-expect-error: mutate.
- node.referenceType = referenceType
- // @ts-expect-error: mutate.
- delete node.url
- delete node.title
- } else {
- // @ts-expect-error: mutate.
- delete node.identifier
- // @ts-expect-error: mutate.
- delete node.label
- }
- setData('referenceType')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitlabeltext(token) {
- const string = this.sliceSerialize(token)
- const ancestor = this.stack[this.stack.length - 2]
- // @ts-expect-error: stash this on the node, as it might become a reference
- // later.
- ancestor.label = decodeString(string)
- // @ts-expect-error: same as above.
- ancestor.identifier = normalizeIdentifier(string).toLowerCase()
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitlabel() {
- const fragment = this.stack[this.stack.length - 1]
- const value = this.resume()
- const node = this.stack[this.stack.length - 1]
- // Assume a reference.
- setData('inReference', true)
- if (node.type === 'link') {
- /** @type {Array<StaticPhrasingContent>} */
- // @ts-expect-error: Assume static phrasing content.
- const children = fragment.children
- node.children = children
- } else {
- node.alt = value
- }
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitresourcedestinationstring() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.url = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitresourcetitlestring() {
- const data = this.resume()
- const node = this.stack[this.stack.length - 1]
- node.title = data
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitresource() {
- setData('inReference')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onenterreference() {
- setData('referenceType', 'collapsed')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitreferencestring(token) {
- const label = this.resume()
- const node = this.stack[this.stack.length - 1]
- // @ts-expect-error: stash this on the node, as it might become a reference
- // later.
- node.label = label
- // @ts-expect-error: same as above.
- node.identifier = normalizeIdentifier(
- this.sliceSerialize(token)
- ).toLowerCase()
- setData('referenceType', 'full')
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
-
- function onexitcharacterreferencemarker(token) {
- setData('characterReferenceType', token.type)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitcharacterreferencevalue(token) {
- const data = this.sliceSerialize(token)
- const type = getData('characterReferenceType')
- /** @type {string} */
- let value
- if (type) {
- value = decodeNumericCharacterReference(
- data,
- type === 'characterReferenceMarkerNumeric' ? 10 : 16
- )
- setData('characterReferenceType')
- } else {
- const result = decodeNamedCharacterReference(data)
- value = result
- }
- const tail = this.stack.pop()
- tail.value += value
- tail.position.end = lib_point(token.end)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitautolinkprotocol(token) {
- onexitdata.call(this, token)
- const node = this.stack[this.stack.length - 1]
- node.url = this.sliceSerialize(token)
- }
-
- /**
- * @this {CompileContext}
- * @type {Handle}
- */
- function onexitautolinkemail(token) {
- onexitdata.call(this, token)
- const node = this.stack[this.stack.length - 1]
- node.url = 'mailto:' + this.sliceSerialize(token)
- }
-
- //
- // Creaters.
- //
-
- /** @returns {Blockquote} */
- function blockQuote() {
- return {
- type: 'blockquote',
- children: []
- }
- }
-
- /** @returns {Code} */
- function codeFlow() {
- return {
- type: 'code',
- lang: null,
- meta: null,
- value: ''
- }
- }
-
- /** @returns {InlineCode} */
- function codeText() {
- return {
- type: 'inlineCode',
- value: ''
- }
- }
-
- /** @returns {Definition} */
- function definition() {
- return {
- type: 'definition',
- identifier: '',
- label: null,
- title: null,
- url: ''
- }
- }
-
- /** @returns {Emphasis} */
- function emphasis() {
- return {
- type: 'emphasis',
- children: []
- }
- }
-
- /** @returns {Heading} */
- function heading() {
- // @ts-expect-error `depth` will be set later.
- return {
- type: 'heading',
- depth: undefined,
- children: []
- }
- }
-
- /** @returns {Break} */
- function hardBreak() {
- return {
- type: 'break'
- }
- }
-
- /** @returns {HTML} */
- function html() {
- return {
- type: 'html',
- value: ''
- }
- }
-
- /** @returns {Image} */
- function image() {
- return {
- type: 'image',
- title: null,
- url: '',
- alt: null
- }
- }
-
- /** @returns {Link} */
- function link() {
- return {
- type: 'link',
- title: null,
- url: '',
- children: []
- }
- }
-
- /**
- * @param {Token} token
- * @returns {List}
- */
- function list(token) {
- return {
- type: 'list',
- ordered: token.type === 'listOrdered',
- start: null,
- spread: token._spread,
- children: []
- }
- }
-
- /**
- * @param {Token} token
- * @returns {ListItem}
- */
- function listItem(token) {
- return {
- type: 'listItem',
- spread: token._spread,
- checked: null,
- children: []
- }
- }
-
- /** @returns {Paragraph} */
- function paragraph() {
- return {
- type: 'paragraph',
- children: []
- }
- }
-
- /** @returns {Strong} */
- function strong() {
- return {
- type: 'strong',
- children: []
- }
- }
-
- /** @returns {Text} */
- function text() {
- return {
- type: 'text',
- value: ''
- }
- }
-
- /** @returns {ThematicBreak} */
- function thematicBreak() {
- return {
- type: 'thematicBreak'
- }
- }
-}
-
-/**
- * Copy a point-like value.
- *
- * @param {Point} d
- * Point-like value.
- * @returns {Point}
- * unist point.
- */
-function lib_point(d) {
- return {
- line: d.line,
- column: d.column,
- offset: d.offset
- }
-}
-
-/**
- * @param {Config} combined
- * @param {Array<Extension | Array<Extension>>} extensions
- * @returns {void}
- */
-function configure(combined, extensions) {
- let index = -1
- while (++index < extensions.length) {
- const value = extensions[index]
- if (Array.isArray(value)) {
- configure(combined, value)
- } else {
- extension(combined, value)
- }
- }
-}
-
-/**
- * @param {Config} combined
- * @param {Extension} extension
- * @returns {void}
- */
-function extension(combined, extension) {
- /** @type {keyof Extension} */
- let key
- for (key in extension) {
- if (lib_own.call(extension, key)) {
- if (key === 'canContainEols') {
- const right = extension[key]
- if (right) {
- combined[key].push(...right)
- }
- } else if (key === 'transforms') {
- const right = extension[key]
- if (right) {
- combined[key].push(...right)
- }
- } else if (key === 'enter' || key === 'exit') {
- const right = extension[key]
- if (right) {
- Object.assign(combined[key], right)
- }
- }
- }
- }
-}
-
-/** @type {OnEnterError} */
-function defaultOnError(left, right) {
- if (left) {
- throw new Error(
- 'Cannot close `' +
- left.type +
- '` (' +
- stringifyPosition({
- start: left.start,
- end: left.end
- }) +
- '): a different token (`' +
- right.type +
- '`, ' +
- stringifyPosition({
- start: right.start,
- end: right.end
- }) +
- ') is open'
- )
- } else {
- throw new Error(
- 'Cannot close document, a token (`' +
- right.type +
- '`, ' +
- stringifyPosition({
- start: right.start,
- end: right.end
- }) +
- ') is still open'
- )
- }
-}
-
-// EXTERNAL MODULE: ./node_modules/ts-dedent/esm/index.js
-var esm = __webpack_require__(18464);
-;// CONCATENATED MODULE: ./node_modules/mermaid/dist/createText-aebacdfe.js
-
-
-
-function preprocessMarkdown(markdown) {
- const withoutMultipleNewlines = markdown.replace(/\n{2,}/g, "\n");
- const withoutExtraSpaces = (0,esm/* dedent */.Z)(withoutMultipleNewlines);
- return withoutExtraSpaces;
-}
-function markdownToLines(markdown) {
- const preprocessedMarkdown = preprocessMarkdown(markdown);
- const { children } = fromMarkdown(preprocessedMarkdown);
- const lines = [[]];
- let currentLine = 0;
- function processNode(node, parentType = "normal") {
- if (node.type === "text") {
- const textLines = node.value.split("\n");
- textLines.forEach((textLine, index) => {
- if (index !== 0) {
- currentLine++;
- lines.push([]);
- }
- textLine.split(" ").forEach((word) => {
- if (word) {
- lines[currentLine].push({ content: word, type: parentType });
- }
- });
- });
- } else if (node.type === "strong" || node.type === "emphasis") {
- node.children.forEach((contentNode) => {
- processNode(contentNode, node.type);
- });
- }
- }
- children.forEach((treeNode) => {
- if (treeNode.type === "paragraph") {
- treeNode.children.forEach((contentNode) => {
- processNode(contentNode);
- });
- }
- });
- return lines;
-}
-function markdownToHTML(markdown) {
- const { children } = fromMarkdown(markdown);
- function output(node) {
- if (node.type === "text") {
- return node.value.replace(/\n/g, "<br/>");
- } else if (node.type === "strong") {
- return `<strong>${node.children.map(output).join("")}</strong>`;
- } else if (node.type === "emphasis") {
- return `<em>${node.children.map(output).join("")}</em>`;
- } else if (node.type === "paragraph") {
- return `<p>${node.children.map(output).join("")}</p>`;
- }
- return `Unsupported markdown: ${node.type}`;
- }
- return children.map(output).join("");
-}
-function splitTextToChars(text) {
- if (Intl.Segmenter) {
- return [...new Intl.Segmenter().segment(text)].map((s) => s.segment);
- }
- return [...text];
-}
-function splitWordToFitWidth(checkFit, word) {
- const characters = splitTextToChars(word.content);
- return splitWordToFitWidthRecursion(checkFit, [], characters, word.type);
-}
-function splitWordToFitWidthRecursion(checkFit, usedChars, remainingChars, type) {
- if (remainingChars.length === 0) {
- return [
- { content: usedChars.join(""), type },
- { content: "", type }
- ];
- }
- const [nextChar, ...rest] = remainingChars;
- const newWord = [...usedChars, nextChar];
- if (checkFit([{ content: newWord.join(""), type }])) {
- return splitWordToFitWidthRecursion(checkFit, newWord, rest, type);
- }
- if (usedChars.length === 0 && nextChar) {
- usedChars.push(nextChar);
- remainingChars.shift();
- }
- return [
- { content: usedChars.join(""), type },
- { content: remainingChars.join(""), type }
- ];
-}
-function splitLineToFitWidth(line, checkFit) {
- if (line.some(({ content }) => content.includes("\n"))) {
- throw new Error("splitLineToFitWidth does not support newlines in the line");
- }
- return splitLineToFitWidthRecursion(line, checkFit);
-}
-function splitLineToFitWidthRecursion(words, checkFit, lines = [], newLine = []) {
- if (words.length === 0) {
- if (newLine.length > 0) {
- lines.push(newLine);
- }
- return lines.length > 0 ? lines : [];
- }
- let joiner = "";
- if (words[0].content === " ") {
- joiner = " ";
- words.shift();
- }
- const nextWord = words.shift() ?? { content: " ", type: "normal" };
- const lineWithNextWord = [...newLine];
- if (joiner !== "") {
- lineWithNextWord.push({ content: joiner, type: "normal" });
- }
- lineWithNextWord.push(nextWord);
- if (checkFit(lineWithNextWord)) {
- return splitLineToFitWidthRecursion(words, checkFit, lines, lineWithNextWord);
- }
- if (newLine.length > 0) {
- lines.push(newLine);
- words.unshift(nextWord);
- } else if (nextWord.content) {
- const [line, rest] = splitWordToFitWidth(checkFit, nextWord);
- lines.push([line]);
- if (rest.content) {
- words.unshift(rest);
- }
- }
- return splitLineToFitWidthRecursion(words, checkFit, lines);
-}
-function applyStyle(dom, styleFn) {
- if (styleFn) {
- dom.attr("style", styleFn);
- }
-}
-function addHtmlSpan(element, node, width, classes, addBackground = false) {
- const fo = element.append("foreignObject");
- const div = fo.append("xhtml:div");
- const label = node.label;
- const labelClass = node.isNode ? "nodeLabel" : "edgeLabel";
- div.html(
- `
- <span class="${labelClass} ${classes}" ` + (node.labelStyle ? 'style="' + node.labelStyle + '"' : "") + ">" + label + "</span>"
- );
- applyStyle(div, node.labelStyle);
- div.style("display", "table-cell");
- div.style("white-space", "nowrap");
- div.style("max-width", width + "px");
- div.attr("xmlns", "http://www.w3.org/1999/xhtml");
- if (addBackground) {
- div.attr("class", "labelBkg");
- }
- let bbox = div.node().getBoundingClientRect();
- if (bbox.width === width) {
- div.style("display", "table");
- div.style("white-space", "break-spaces");
- div.style("width", width + "px");
- bbox = div.node().getBoundingClientRect();
- }
- fo.style("width", bbox.width);
- fo.style("height", bbox.height);
- return fo.node();
-}
-function createTspan(textElement, lineIndex, lineHeight) {
- return textElement.append("tspan").attr("class", "text-outer-tspan").attr("x", 0).attr("y", lineIndex * lineHeight - 0.1 + "em").attr("dy", lineHeight + "em");
-}
-function computeWidthOfText(parentNode, lineHeight, line) {
- const testElement = parentNode.append("text");
- const testSpan = createTspan(testElement, 1, lineHeight);
- updateTextContentAndStyles(testSpan, line);
- const textLength = testSpan.node().getComputedTextLength();
- testElement.remove();
- return textLength;
-}
-function computeDimensionOfText(parentNode, lineHeight, text) {
- var _a;
- const testElement = parentNode.append("text");
- const testSpan = createTspan(testElement, 1, lineHeight);
- updateTextContentAndStyles(testSpan, [{ content: text, type: "normal" }]);
- const textDimension = (_a = testSpan.node()) == null ? void 0 : _a.getBoundingClientRect();
- if (textDimension) {
- testElement.remove();
- }
- return textDimension;
-}
-function createFormattedText(width, g, structuredText, addBackground = false) {
- const lineHeight = 1.1;
- const labelGroup = g.append("g");
- const bkg = labelGroup.insert("rect").attr("class", "background");
- const textElement = labelGroup.append("text").attr("y", "-10.1");
- let lineIndex = 0;
- for (const line of structuredText) {
- const checkWidth = (line2) => computeWidthOfText(labelGroup, lineHeight, line2) <= width;
- const linesUnderWidth = checkWidth(line) ? [line] : splitLineToFitWidth(line, checkWidth);
- for (const preparedLine of linesUnderWidth) {
- const tspan = createTspan(textElement, lineIndex, lineHeight);
- updateTextContentAndStyles(tspan, preparedLine);
- lineIndex++;
- }
- }
- if (addBackground) {
- const bbox = textElement.node().getBBox();
- const padding = 2;
- bkg.attr("x", -padding).attr("y", -padding).attr("width", bbox.width + 2 * padding).attr("height", bbox.height + 2 * padding);
- return labelGroup.node();
- } else {
- return textElement.node();
- }
-}
-function updateTextContentAndStyles(tspan, wrappedLine) {
- tspan.text("");
- wrappedLine.forEach((word, index) => {
- const innerTspan = tspan.append("tspan").attr("font-style", word.type === "emphasis" ? "italic" : "normal").attr("class", "text-inner-tspan").attr("font-weight", word.type === "strong" ? "bold" : "normal");
- if (index === 0) {
- innerTspan.text(word.content);
- } else {
- innerTspan.text(" " + word.content);
- }
- });
-}
-const createText = (el, text = "", {
- style = "",
- isTitle = false,
- classes = "",
- useHtmlLabels = true,
- isNode = true,
- width = 200,
- addSvgBackground = false
-} = {}) => {
- mermaid_934d9bea.l.info("createText", text, style, isTitle, classes, useHtmlLabels, isNode, addSvgBackground);
- if (useHtmlLabels) {
- const htmlText = markdownToHTML(text);
- const node = {
- isNode,
- label: (0,mermaid_934d9bea.J)(htmlText).replace(
- /fa[blrs]?:fa-[\w-]+/g,
- (s) => `<i class='${s.replace(":", " ")}'></i>`
- ),
- labelStyle: style.replace("fill:", "color:")
- };
- const vertexNode = addHtmlSpan(el, node, width, classes, addSvgBackground);
- return vertexNode;
- } else {
- const structuredText = markdownToLines(text);
- const svgLabel = createFormattedText(width, el, structuredText, addSvgBackground);
- return svgLabel;
- }
-};
-
-
-
-/***/ }),
-
-/***/ 32693:
-/***/ ((__unused_webpack___webpack_module__, __webpack_exports__, __webpack_require__) => {
-
-/* harmony export */ __webpack_require__.d(__webpack_exports__, {
-/* harmony export */ diagram: () => (/* binding */ diagram)
-/* harmony export */ });
-/* harmony import */ var _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__ = __webpack_require__(85322);
-/* harmony import */ var _createText_aebacdfe_js__WEBPACK_IMPORTED_MODULE_4__ = __webpack_require__(69583);
-/* harmony import */ var d3__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(64218);
-/* harmony import */ var dayjs__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(27484);
-/* harmony import */ var _braintree_sanitize_url__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(17967);
-/* harmony import */ var dompurify__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(20683);
-
-
-
-
-
-
-
-
-
-
-
-
-
-var parser = function() {
- var o = function(k, v, o2, l) {
- for (o2 = o2 || {}, l = k.length; l--; o2[k[l]] = v)
- ;
- return o2;
- }, $V0 = [1, 10, 12, 14, 16, 18, 19, 21, 23], $V1 = [2, 6], $V2 = [1, 3], $V3 = [1, 5], $V4 = [1, 6], $V5 = [1, 7], $V6 = [1, 5, 10, 12, 14, 16, 18, 19, 21, 23, 34, 35, 36], $V7 = [1, 25], $V8 = [1, 26], $V9 = [1, 28], $Va = [1, 29], $Vb = [1, 30], $Vc = [1, 31], $Vd = [1, 32], $Ve = [1, 33], $Vf = [1, 34], $Vg = [1, 35], $Vh = [1, 36], $Vi = [1, 37], $Vj = [1, 43], $Vk = [1, 42], $Vl = [1, 47], $Vm = [1, 50], $Vn = [1, 10, 12, 14, 16, 18, 19, 21, 23, 34, 35, 36], $Vo = [1, 10, 12, 14, 16, 18, 19, 21, 23, 24, 26, 27, 28, 34, 35, 36], $Vp = [1, 10, 12, 14, 16, 18, 19, 21, 23, 24, 26, 27, 28, 34, 35, 36, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50], $Vq = [1, 64];
- var parser2 = {
- trace: function trace() {
- },
- yy: {},
- symbols_: { "error": 2, "start": 3, "eol": 4, "XYCHART": 5, "chartConfig": 6, "document": 7, "CHART_ORIENTATION": 8, "statement": 9, "title": 10, "text": 11, "X_AXIS": 12, "parseXAxis": 13, "Y_AXIS": 14, "parseYAxis": 15, "LINE": 16, "plotData": 17, "BAR": 18, "acc_title": 19, "acc_title_value": 20, "acc_descr": 21, "acc_descr_value": 22, "acc_descr_multiline_value": 23, "SQUARE_BRACES_START": 24, "commaSeparatedNumbers": 25, "SQUARE_BRACES_END": 26, "NUMBER_WITH_DECIMAL": 27, "COMMA": 28, "xAxisData": 29, "bandData": 30, "ARROW_DELIMITER": 31, "commaSeparatedTexts": 32, "yAxisData": 33, "NEWLINE": 34, "SEMI": 35, "EOF": 36, "alphaNum": 37, "STR": 38, "MD_STR": 39, "alphaNumToken": 40, "AMP": 41, "NUM": 42, "ALPHA": 43, "PLUS": 44, "EQUALS": 45, "MULT": 46, "DOT": 47, "BRKT": 48, "MINUS": 49, "UNDERSCORE": 50, "$accept": 0, "$end": 1 },
- terminals_: { 2: "error", 5: "XYCHART", 8: "CHART_ORIENTATION", 10: "title", 12: "X_AXIS", 14: "Y_AXIS", 16: "LINE", 18: "BAR", 19: "acc_title", 20: "acc_title_value", 21: "acc_descr", 22: "acc_descr_value", 23: "acc_descr_multiline_value", 24: "SQUARE_BRACES_START", 26: "SQUARE_BRACES_END", 27: "NUMBER_WITH_DECIMAL", 28: "COMMA", 31: "ARROW_DELIMITER", 34: "NEWLINE", 35: "SEMI", 36: "EOF", 38: "STR", 39: "MD_STR", 41: "AMP", 42: "NUM", 43: "ALPHA", 44: "PLUS", 45: "EQUALS", 46: "MULT", 47: "DOT", 48: "BRKT", 49: "MINUS", 50: "UNDERSCORE" },
- productions_: [0, [3, 2], [3, 3], [3, 2], [3, 1], [6, 1], [7, 0], [7, 2], [9, 2], [9, 2], [9, 2], [9, 2], [9, 2], [9, 3], [9, 2], [9, 3], [9, 2], [9, 2], [9, 1], [17, 3], [25, 3], [25, 1], [13, 1], [13, 2], [13, 1], [29, 1], [29, 3], [30, 3], [32, 3], [32, 1], [15, 1], [15, 2], [15, 1], [33, 3], [4, 1], [4, 1], [4, 1], [11, 1], [11, 1], [11, 1], [37, 1], [37, 2], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1], [40, 1]],
- performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate, $$, _$) {
- var $0 = $$.length - 1;
- switch (yystate) {
- case 5:
- yy.setOrientation($$[$0]);
- break;
- case 9:
- yy.setDiagramTitle($$[$0].text.trim());
- break;
- case 12:
- yy.setLineData({ text: "", type: "text" }, $$[$0]);
- break;
- case 13:
- yy.setLineData($$[$0 - 1], $$[$0]);
- break;
- case 14:
- yy.setBarData({ text: "", type: "text" }, $$[$0]);
- break;
- case 15:
- yy.setBarData($$[$0 - 1], $$[$0]);
- break;
- case 16:
- this.$ = $$[$0].trim();
- yy.setAccTitle(this.$);
- break;
- case 17:
- case 18:
- this.$ = $$[$0].trim();
- yy.setAccDescription(this.$);
- break;
- case 19:
- this.$ = $$[$0 - 1];
- break;
- case 20:
- this.$ = [Number($$[$0 - 2]), ...$$[$0]];
- break;
- case 21:
- this.$ = [Number($$[$0])];
- break;
- case 22:
- yy.setXAxisTitle($$[$0]);
- break;
- case 23:
- yy.setXAxisTitle($$[$0 - 1]);
- break;
- case 24:
- yy.setXAxisTitle({ type: "text", text: "" });
- break;
- case 25:
- yy.setXAxisBand($$[$0]);
- break;
- case 26:
- yy.setXAxisRangeData(Number($$[$0 - 2]), Number($$[$0]));
- break;
- case 27:
- this.$ = $$[$0 - 1];
- break;
- case 28:
- this.$ = [$$[$0 - 2], ...$$[$0]];
- break;
- case 29:
- this.$ = [$$[$0]];
- break;
- case 30:
- yy.setYAxisTitle($$[$0]);
- break;
- case 31:
- yy.setYAxisTitle($$[$0 - 1]);
- break;
- case 32:
- yy.setYAxisTitle({ type: "text", text: "" });
- break;
- case 33:
- yy.setYAxisRangeData(Number($$[$0 - 2]), Number($$[$0]));
- break;
- case 37:
- this.$ = { text: $$[$0], type: "text" };
- break;
- case 38:
- this.$ = { text: $$[$0], type: "text" };
- break;
- case 39:
- this.$ = { text: $$[$0], type: "markdown" };
- break;
- case 40:
- this.$ = $$[$0];
- break;
- case 41:
- this.$ = $$[$0 - 1] + "" + $$[$0];
- break;
- }
- },
- table: [o($V0, $V1, { 3: 1, 4: 2, 7: 4, 5: $V2, 34: $V3, 35: $V4, 36: $V5 }), { 1: [3] }, o($V0, $V1, { 4: 2, 7: 4, 3: 8, 5: $V2, 34: $V3, 35: $V4, 36: $V5 }), o($V0, $V1, { 4: 2, 7: 4, 6: 9, 3: 10, 5: $V2, 8: [1, 11], 34: $V3, 35: $V4, 36: $V5 }), { 1: [2, 4], 9: 12, 10: [1, 13], 12: [1, 14], 14: [1, 15], 16: [1, 16], 18: [1, 17], 19: [1, 18], 21: [1, 19], 23: [1, 20] }, o($V6, [2, 34]), o($V6, [2, 35]), o($V6, [2, 36]), { 1: [2, 1] }, o($V0, $V1, { 4: 2, 7: 4, 3: 21, 5: $V2, 34: $V3, 35: $V4, 36: $V5 }), { 1: [2, 3] }, o($V6, [2, 5]), o($V0, [2, 7], { 4: 22, 34: $V3, 35: $V4, 36: $V5 }), { 11: 23, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, { 11: 39, 13: 38, 24: $Vj, 27: $Vk, 29: 40, 30: 41, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, { 11: 45, 15: 44, 27: $Vl, 33: 46, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, { 11: 49, 17: 48, 24: $Vm, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, { 11: 52, 17: 51, 24: $Vm, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, { 20: [1, 53] }, { 22: [1, 54] }, o($Vn, [2, 18]), { 1: [2, 2] }, o($Vn, [2, 8]), o($Vn, [2, 9]), o($Vo, [2, 37], { 40: 55, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }), o($Vo, [2, 38]), o($Vo, [2, 39]), o($Vp, [2, 40]), o($Vp, [2, 42]), o($Vp, [2, 43]), o($Vp, [2, 44]), o($Vp, [2, 45]), o($Vp, [2, 46]), o($Vp, [2, 47]), o($Vp, [2, 48]), o($Vp, [2, 49]), o($Vp, [2, 50]), o($Vp, [2, 51]), o($Vn, [2, 10]), o($Vn, [2, 22], { 30: 41, 29: 56, 24: $Vj, 27: $Vk }), o($Vn, [2, 24]), o($Vn, [2, 25]), { 31: [1, 57] }, { 11: 59, 32: 58, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, o($Vn, [2, 11]), o($Vn, [2, 30], { 33: 60, 27: $Vl }), o($Vn, [2, 32]), { 31: [1, 61] }, o($Vn, [2, 12]), { 17: 62, 24: $Vm }, { 25: 63, 27: $Vq }, o($Vn, [2, 14]), { 17: 65, 24: $Vm }, o($Vn, [2, 16]), o($Vn, [2, 17]), o($Vp, [2, 41]), o($Vn, [2, 23]), { 27: [1, 66] }, { 26: [1, 67] }, { 26: [2, 29], 28: [1, 68] }, o($Vn, [2, 31]), { 27: [1, 69] }, o($Vn, [2, 13]), { 26: [1, 70] }, { 26: [2, 21], 28: [1, 71] }, o($Vn, [2, 15]), o($Vn, [2, 26]), o($Vn, [2, 27]), { 11: 59, 32: 72, 37: 24, 38: $V7, 39: $V8, 40: 27, 41: $V9, 42: $Va, 43: $Vb, 44: $Vc, 45: $Vd, 46: $Ve, 47: $Vf, 48: $Vg, 49: $Vh, 50: $Vi }, o($Vn, [2, 33]), o($Vn, [2, 19]), { 25: 73, 27: $Vq }, { 26: [2, 28] }, { 26: [2, 20] }],
- defaultActions: { 8: [2, 1], 10: [2, 3], 21: [2, 2], 72: [2, 28], 73: [2, 20] },
- parseError: function parseError(str, hash) {
- if (hash.recoverable) {
- this.trace(str);
- } else {
- var error = new Error(str);
- error.hash = hash;
- throw error;
- }
- },
- parse: function parse(input) {
- var self = this, stack = [0], tstack = [], vstack = [null], lstack = [], table = this.table, yytext = "", yylineno = 0, yyleng = 0, TERROR = 2, EOF = 1;
- var args = lstack.slice.call(arguments, 1);
- var lexer2 = Object.create(this.lexer);
- var sharedState = { yy: {} };
- for (var k in this.yy) {
- if (Object.prototype.hasOwnProperty.call(this.yy, k)) {
- sharedState.yy[k] = this.yy[k];
- }
- }
- lexer2.setInput(input, sharedState.yy);
- sharedState.yy.lexer = lexer2;
- sharedState.yy.parser = this;
- if (typeof lexer2.yylloc == "undefined") {
- lexer2.yylloc = {};
- }
- var yyloc = lexer2.yylloc;
- lstack.push(yyloc);
- var ranges = lexer2.options && lexer2.options.ranges;
- if (typeof sharedState.yy.parseError === "function") {
- this.parseError = sharedState.yy.parseError;
- } else {
- this.parseError = Object.getPrototypeOf(this).parseError;
- }
- function lex() {
- var token;
- token = tstack.pop() || lexer2.lex() || EOF;
- if (typeof token !== "number") {
- if (token instanceof Array) {
- tstack = token;
- token = tstack.pop();
- }
- token = self.symbols_[token] || token;
- }
- return token;
- }
- var symbol, state, action, r, yyval = {}, p, len, newState, expected;
- while (true) {
- state = stack[stack.length - 1];
- if (this.defaultActions[state]) {
- action = this.defaultActions[state];
- } else {
- if (symbol === null || typeof symbol == "undefined") {
- symbol = lex();
- }
- action = table[state] && table[state][symbol];
- }
- if (typeof action === "undefined" || !action.length || !action[0]) {
- var errStr = "";
- expected = [];
- for (p in table[state]) {
- if (this.terminals_[p] && p > TERROR) {
- expected.push("'" + this.terminals_[p] + "'");
- }
- }
- if (lexer2.showPosition) {
- errStr = "Parse error on line " + (yylineno + 1) + ":\n" + lexer2.showPosition() + "\nExpecting " + expected.join(", ") + ", got '" + (this.terminals_[symbol] || symbol) + "'";
- } else {
- errStr = "Parse error on line " + (yylineno + 1) + ": Unexpected " + (symbol == EOF ? "end of input" : "'" + (this.terminals_[symbol] || symbol) + "'");
- }
- this.parseError(errStr, {
- text: lexer2.match,
- token: this.terminals_[symbol] || symbol,
- line: lexer2.yylineno,
- loc: yyloc,
- expected
- });
- }
- if (action[0] instanceof Array && action.length > 1) {
- throw new Error("Parse Error: multiple actions possible at state: " + state + ", token: " + symbol);
- }
- switch (action[0]) {
- case 1:
- stack.push(symbol);
- vstack.push(lexer2.yytext);
- lstack.push(lexer2.yylloc);
- stack.push(action[1]);
- symbol = null;
- {
- yyleng = lexer2.yyleng;
- yytext = lexer2.yytext;
- yylineno = lexer2.yylineno;
- yyloc = lexer2.yylloc;
- }
- break;
- case 2:
- len = this.productions_[action[1]][1];
- yyval.$ = vstack[vstack.length - len];
- yyval._$ = {
- first_line: lstack[lstack.length - (len || 1)].first_line,
- last_line: lstack[lstack.length - 1].last_line,
- first_column: lstack[lstack.length - (len || 1)].first_column,
- last_column: lstack[lstack.length - 1].last_column
- };
- if (ranges) {
- yyval._$.range = [
- lstack[lstack.length - (len || 1)].range[0],
- lstack[lstack.length - 1].range[1]
- ];
- }
- r = this.performAction.apply(yyval, [
- yytext,
- yyleng,
- yylineno,
- sharedState.yy,
- action[1],
- vstack,
- lstack
- ].concat(args));
- if (typeof r !== "undefined") {
- return r;
- }
- if (len) {
- stack = stack.slice(0, -1 * len * 2);
- vstack = vstack.slice(0, -1 * len);
- lstack = lstack.slice(0, -1 * len);
- }
- stack.push(this.productions_[action[1]][0]);
- vstack.push(yyval.$);
- lstack.push(yyval._$);
- newState = table[stack[stack.length - 2]][stack[stack.length - 1]];
- stack.push(newState);
- break;
- case 3:
- return true;
- }
- }
- return true;
- }
- };
- var lexer = function() {
- var lexer2 = {
- EOF: 1,
- parseError: function parseError(str, hash) {
- if (this.yy.parser) {
- this.yy.parser.parseError(str, hash);
- } else {
- throw new Error(str);
- }
- },
- // resets the lexer, sets new input
- setInput: function(input, yy) {
- this.yy = yy || this.yy || {};
- this._input = input;
- this._more = this._backtrack = this.done = false;
- this.yylineno = this.yyleng = 0;
- this.yytext = this.matched = this.match = "";
- this.conditionStack = ["INITIAL"];
- this.yylloc = {
- first_line: 1,
- first_column: 0,
- last_line: 1,
- last_column: 0
- };
- if (this.options.ranges) {
- this.yylloc.range = [0, 0];
- }
- this.offset = 0;
- return this;
- },
- // consumes and returns one char from the input
- input: function() {
- var ch = this._input[0];
- this.yytext += ch;
- this.yyleng++;
- this.offset++;
- this.match += ch;
- this.matched += ch;
- var lines = ch.match(/(?:\r\n?|\n).*/g);
- if (lines) {
- this.yylineno++;
- this.yylloc.last_line++;
- } else {
- this.yylloc.last_column++;
- }
- if (this.options.ranges) {
- this.yylloc.range[1]++;
- }
- this._input = this._input.slice(1);
- return ch;
- },
- // unshifts one char (or a string) into the input
- unput: function(ch) {
- var len = ch.length;
- var lines = ch.split(/(?:\r\n?|\n)/g);
- this._input = ch + this._input;
- this.yytext = this.yytext.substr(0, this.yytext.length - len);
- this.offset -= len;
- var oldLines = this.match.split(/(?:\r\n?|\n)/g);
- this.match = this.match.substr(0, this.match.length - 1);
- this.matched = this.matched.substr(0, this.matched.length - 1);
- if (lines.length - 1) {
- this.yylineno -= lines.length - 1;
- }
- var r = this.yylloc.range;
- this.yylloc = {
- first_line: this.yylloc.first_line,
- last_line: this.yylineno + 1,
- first_column: this.yylloc.first_column,
- last_column: lines ? (lines.length === oldLines.length ? this.yylloc.first_column : 0) + oldLines[oldLines.length - lines.length].length - lines[0].length : this.yylloc.first_column - len
- };
- if (this.options.ranges) {
- this.yylloc.range = [r[0], r[0] + this.yyleng - len];
- }
- this.yyleng = this.yytext.length;
- return this;
- },
- // When called from action, caches matched text and appends it on next action
- more: function() {
- this._more = true;
- return this;
- },
- // When called from action, signals the lexer that this rule fails to match the input, so the next matching rule (regex) should be tested instead.
- reject: function() {
- if (this.options.backtrack_lexer) {
- this._backtrack = true;
- } else {
- return this.parseError("Lexical error on line " + (this.yylineno + 1) + ". You can only invoke reject() in the lexer when the lexer is of the backtracking persuasion (options.backtrack_lexer = true).\n" + this.showPosition(), {
- text: "",
- token: null,
- line: this.yylineno
- });
- }
- return this;
- },
- // retain first n characters of the match
- less: function(n) {
- this.unput(this.match.slice(n));
- },
- // displays already matched input, i.e. for error messages
- pastInput: function() {
- var past = this.matched.substr(0, this.matched.length - this.match.length);
- return (past.length > 20 ? "..." : "") + past.substr(-20).replace(/\n/g, "");
- },
- // displays upcoming input, i.e. for error messages
- upcomingInput: function() {
- var next = this.match;
- if (next.length < 20) {
- next += this._input.substr(0, 20 - next.length);
- }
- return (next.substr(0, 20) + (next.length > 20 ? "..." : "")).replace(/\n/g, "");
- },
- // displays the character position where the lexing error occurred, i.e. for error messages
- showPosition: function() {
- var pre = this.pastInput();
- var c = new Array(pre.length + 1).join("-");
- return pre + this.upcomingInput() + "\n" + c + "^";
- },
- // test the lexed token: return FALSE when not a match, otherwise return token
- test_match: function(match, indexed_rule) {
- var token, lines, backup;
- if (this.options.backtrack_lexer) {
- backup = {
- yylineno: this.yylineno,
- yylloc: {
- first_line: this.yylloc.first_line,
- last_line: this.last_line,
- first_column: this.yylloc.first_column,
- last_column: this.yylloc.last_column
- },
- yytext: this.yytext,
- match: this.match,
- matches: this.matches,
- matched: this.matched,
- yyleng: this.yyleng,
- offset: this.offset,
- _more: this._more,
- _input: this._input,
- yy: this.yy,
- conditionStack: this.conditionStack.slice(0),
- done: this.done
- };
- if (this.options.ranges) {
- backup.yylloc.range = this.yylloc.range.slice(0);
- }
- }
- lines = match[0].match(/(?:\r\n?|\n).*/g);
- if (lines) {
- this.yylineno += lines.length;
- }
- this.yylloc = {
- first_line: this.yylloc.last_line,
- last_line: this.yylineno + 1,
- first_column: this.yylloc.last_column,
- last_column: lines ? lines[lines.length - 1].length - lines[lines.length - 1].match(/\r?\n?/)[0].length : this.yylloc.last_column + match[0].length
- };
- this.yytext += match[0];
- this.match += match[0];
- this.matches = match;
- this.yyleng = this.yytext.length;
- if (this.options.ranges) {
- this.yylloc.range = [this.offset, this.offset += this.yyleng];
- }
- this._more = false;
- this._backtrack = false;
- this._input = this._input.slice(match[0].length);
- this.matched += match[0];
- token = this.performAction.call(this, this.yy, this, indexed_rule, this.conditionStack[this.conditionStack.length - 1]);
- if (this.done && this._input) {
- this.done = false;
- }
- if (token) {
- return token;
- } else if (this._backtrack) {
- for (var k in backup) {
- this[k] = backup[k];
- }
- return false;
- }
- return false;
- },
- // return next match in input
- next: function() {
- if (this.done) {
- return this.EOF;
- }
- if (!this._input) {
- this.done = true;
- }
- var token, match, tempMatch, index;
- if (!this._more) {
- this.yytext = "";
- this.match = "";
- }
- var rules = this._currentRules();
- for (var i = 0; i < rules.length; i++) {
- tempMatch = this._input.match(this.rules[rules[i]]);
- if (tempMatch && (!match || tempMatch[0].length > match[0].length)) {
- match = tempMatch;
- index = i;
- if (this.options.backtrack_lexer) {
- token = this.test_match(tempMatch, rules[i]);
- if (token !== false) {
- return token;
- } else if (this._backtrack) {
- match = false;
- continue;
- } else {
- return false;
- }
- } else if (!this.options.flex) {
- break;
- }
- }
- }
- if (match) {
- token = this.test_match(match, rules[index]);
- if (token !== false) {
- return token;
- }
- return false;
- }
- if (this._input === "") {
- return this.EOF;
- } else {
- return this.parseError("Lexical error on line " + (this.yylineno + 1) + ". Unrecognized text.\n" + this.showPosition(), {
- text: "",
- token: null,
- line: this.yylineno
- });
- }
- },
- // return next match that has a token
- lex: function lex() {
- var r = this.next();
- if (r) {
- return r;
- } else {
- return this.lex();
- }
- },
- // activates a new lexer condition state (pushes the new lexer condition state onto the condition stack)
- begin: function begin(condition) {
- this.conditionStack.push(condition);
- },
- // pop the previously active lexer condition state off the condition stack
- popState: function popState() {
- var n = this.conditionStack.length - 1;
- if (n > 0) {
- return this.conditionStack.pop();
- } else {
- return this.conditionStack[0];
- }
- },
- // produce the lexer rule set which is active for the currently active lexer condition state
- _currentRules: function _currentRules() {
- if (this.conditionStack.length && this.conditionStack[this.conditionStack.length - 1]) {
- return this.conditions[this.conditionStack[this.conditionStack.length - 1]].rules;
- } else {
- return this.conditions["INITIAL"].rules;
- }
- },
- // return the currently active lexer condition state; when an index argument is provided it produces the N-th previous condition state, if available
- topState: function topState(n) {
- n = this.conditionStack.length - 1 - Math.abs(n || 0);
- if (n >= 0) {
- return this.conditionStack[n];
- } else {
- return "INITIAL";
- }
- },
- // alias for begin(condition)
- pushState: function pushState(condition) {
- this.begin(condition);
- },
- // return the number of states currently on the stack
- stateStackSize: function stateStackSize() {
- return this.conditionStack.length;
- },
- options: { "case-insensitive": true },
- performAction: function anonymous(yy, yy_, $avoiding_name_collisions, YY_START) {
- switch ($avoiding_name_collisions) {
- case 0:
- break;
- case 1:
- break;
- case 2:
- this.popState();
- return 34;
- case 3:
- this.popState();
- return 34;
- case 4:
- return 34;
- case 5:
- break;
- case 6:
- return 10;
- case 7:
- this.pushState("acc_title");
- return 19;
- case 8:
- this.popState();
- return "acc_title_value";
- case 9:
- this.pushState("acc_descr");
- return 21;
- case 10:
- this.popState();
- return "acc_descr_value";
- case 11:
- this.pushState("acc_descr_multiline");
- break;
- case 12:
- this.popState();
- break;
- case 13:
- return "acc_descr_multiline_value";
- case 14:
- return 5;
- case 15:
- return 8;
- case 16:
- this.pushState("axis_data");
- return "X_AXIS";
- case 17:
- this.pushState("axis_data");
- return "Y_AXIS";
- case 18:
- this.pushState("axis_band_data");
- return 24;
- case 19:
- return 31;
- case 20:
- this.pushState("data");
- return 16;
- case 21:
- this.pushState("data");
- return 18;
- case 22:
- this.pushState("data_inner");
- return 24;
- case 23:
- return 27;
- case 24:
- this.popState();
- return 26;
- case 25:
- this.popState();
- break;
- case 26:
- this.pushState("string");
- break;
- case 27:
- this.popState();
- break;
- case 28:
- return "STR";
- case 29:
- return 24;
- case 30:
- return 26;
- case 31:
- return 43;
- case 32:
- return "COLON";
- case 33:
- return 44;
- case 34:
- return 28;
- case 35:
- return 45;
- case 36:
- return 46;
- case 37:
- return 48;
- case 38:
- return 50;
- case 39:
- return 47;
- case 40:
- return 41;
- case 41:
- return 49;
- case 42:
- return 42;
- case 43:
- break;
- case 44:
- return 35;
- case 45:
- return 36;
- }
- },
- rules: [/^(?:%%(?!\{)[^\n]*)/i, /^(?:[^\}]%%[^\n]*)/i, /^(?:(\r?\n))/i, /^(?:(\r?\n))/i, /^(?:[\n\r]+)/i, /^(?:%%[^\n]*)/i, /^(?:title\b)/i, /^(?:accTitle\s*:\s*)/i, /^(?:(?!\n||)*[^\n]*)/i, /^(?:accDescr\s*:\s*)/i, /^(?:(?!\n||)*[^\n]*)/i, /^(?:accDescr\s*\{\s*)/i, /^(?:\{)/i, /^(?:[^\}]*)/i, /^(?:xychart-beta\b)/i, /^(?:(?:vertical|horizontal))/i, /^(?:x-axis\b)/i, /^(?:y-axis\b)/i, /^(?:\[)/i, /^(?:-->)/i, /^(?:line\b)/i, /^(?:bar\b)/i, /^(?:\[)/i, /^(?:[+-]?(?:\d+(?:\.\d+)?|\.\d+))/i, /^(?:\])/i, /^(?:(?:`\) \{ this\.pushState\(md_string\); \}\n<md_string>\(\?:\(\?!`"\)\.\)\+ \{ return MD_STR; \}\n<md_string>\(\?:`))/i, /^(?:["])/i, /^(?:["])/i, /^(?:[^"]*)/i, /^(?:\[)/i, /^(?:\])/i, /^(?:[A-Za-z]+)/i, /^(?::)/i, /^(?:\+)/i, /^(?:,)/i, /^(?:=)/i, /^(?:\*)/i, /^(?:#)/i, /^(?:[\_])/i, /^(?:\.)/i, /^(?:&)/i, /^(?:-)/i, /^(?:[0-9]+)/i, /^(?:\s+)/i, /^(?:;)/i, /^(?:$)/i],
- conditions: { "data_inner": { "rules": [0, 1, 4, 5, 6, 7, 9, 11, 14, 15, 16, 17, 20, 21, 23, 24, 25, 26, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45], "inclusive": true }, "data": { "rules": [0, 1, 3, 4, 5, 6, 7, 9, 11, 14, 15, 16, 17, 20, 21, 22, 25, 26, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45], "inclusive": true }, "axis_band_data": { "rules": [0, 1, 4, 5, 6, 7, 9, 11, 14, 15, 16, 17, 20, 21, 24, 25, 26, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45], "inclusive": true }, "axis_data": { "rules": [0, 1, 2, 4, 5, 6, 7, 9, 11, 14, 15, 16, 17, 18, 19, 20, 21, 23, 25, 26, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45], "inclusive": true }, "acc_descr_multiline": { "rules": [12, 13], "inclusive": false }, "acc_descr": { "rules": [10], "inclusive": false }, "acc_title": { "rules": [8], "inclusive": false }, "title": { "rules": [], "inclusive": false }, "md_string": { "rules": [], "inclusive": false }, "string": { "rules": [27, 28], "inclusive": false }, "INITIAL": { "rules": [0, 1, 4, 5, 6, 7, 9, 11, 14, 15, 16, 17, 20, 21, 25, 26, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45], "inclusive": true } }
- };
- return lexer2;
- }();
- parser2.lexer = lexer;
- function Parser() {
- this.yy = {};
- }
- Parser.prototype = parser2;
- parser2.Parser = Parser;
- return new Parser();
-}();
-parser.parser = parser;
-const parser$1 = parser;
-function isBarPlot(data) {
- return data.type === "bar";
-}
-function isBandAxisData(data) {
- return data.type === "band";
-}
-function isLinearAxisData(data) {
- return data.type === "linear";
-}
-class TextDimensionCalculatorWithFont {
- constructor(parentGroup) {
- this.parentGroup = parentGroup;
- }
- getMaxDimension(texts, fontSize) {
- if (!this.parentGroup) {
- return {
- width: texts.reduce((acc, cur) => Math.max(cur.length, acc), 0) * fontSize,
- height: fontSize
- };
- }
- const dimension = {
- width: 0,
- height: 0
- };
- const elem = this.parentGroup.append("g").attr("visibility", "hidden").attr("font-size", fontSize);
- for (const t of texts) {
- const bbox = (0,_createText_aebacdfe_js__WEBPACK_IMPORTED_MODULE_4__.c)(elem, 1, t);
- const width = bbox ? bbox.width : t.length * fontSize;
- const height = bbox ? bbox.height : fontSize;
- dimension.width = Math.max(dimension.width, width);
- dimension.height = Math.max(dimension.height, height);
- }
- elem.remove();
- return dimension;
- }
-}
-const BAR_WIDTH_TO_TICK_WIDTH_RATIO = 0.7;
-const MAX_OUTER_PADDING_PERCENT_FOR_WRT_LABEL = 0.2;
-class BaseAxis {
- constructor(axisConfig, title, textDimensionCalculator, axisThemeConfig) {
- this.axisConfig = axisConfig;
- this.title = title;
- this.textDimensionCalculator = textDimensionCalculator;
- this.axisThemeConfig = axisThemeConfig;
- this.boundingRect = { x: 0, y: 0, width: 0, height: 0 };
- this.axisPosition = "left";
- this.showTitle = false;
- this.showLabel = false;
- this.showTick = false;
- this.showAxisLine = false;
- this.outerPadding = 0;
- this.titleTextHeight = 0;
- this.labelTextHeight = 0;
- this.range = [0, 10];
- this.boundingRect = { x: 0, y: 0, width: 0, height: 0 };
- this.axisPosition = "left";
- }
- setRange(range) {
- this.range = range;
- if (this.axisPosition === "left" || this.axisPosition === "right") {
- this.boundingRect.height = range[1] - range[0];
- } else {
- this.boundingRect.width = range[1] - range[0];
- }
- this.recalculateScale();
- }
- getRange() {
- return [this.range[0] + this.outerPadding, this.range[1] - this.outerPadding];
- }
- setAxisPosition(axisPosition) {
- this.axisPosition = axisPosition;
- this.setRange(this.range);
- }
- getTickDistance() {
- const range = this.getRange();
- return Math.abs(range[0] - range[1]) / this.getTickValues().length;
- }
- getAxisOuterPadding() {
- return this.outerPadding;
- }
- getLabelDimension() {
- return this.textDimensionCalculator.getMaxDimension(
- this.getTickValues().map((tick) => tick.toString()),
- this.axisConfig.labelFontSize
- );
- }
- recalculateOuterPaddingToDrawBar() {
- if (BAR_WIDTH_TO_TICK_WIDTH_RATIO * this.getTickDistance() > this.outerPadding * 2) {
- this.outerPadding = Math.floor(BAR_WIDTH_TO_TICK_WIDTH_RATIO * this.getTickDistance() / 2);
- }
- this.recalculateScale();
- }
- calculateSpaceIfDrawnHorizontally(availableSpace) {
- let availableHeight = availableSpace.height;
- if (this.axisConfig.showAxisLine && availableHeight > this.axisConfig.axisLineWidth) {
- availableHeight -= this.axisConfig.axisLineWidth;
- this.showAxisLine = true;
- }
- if (this.axisConfig.showLabel) {
- const spaceRequired = this.getLabelDimension();
- const maxPadding = MAX_OUTER_PADDING_PERCENT_FOR_WRT_LABEL * availableSpace.width;
- this.outerPadding = Math.min(spaceRequired.width / 2, maxPadding);
- const heightRequired = spaceRequired.height + this.axisConfig.labelPadding * 2;
- this.labelTextHeight = spaceRequired.height;
- if (heightRequired <= availableHeight) {
- availableHeight -= heightRequired;
- this.showLabel = true;
- }
- }
- if (this.axisConfig.showTick && availableHeight >= this.axisConfig.tickLength) {
- this.showTick = true;
- availableHeight -= this.axisConfig.tickLength;
- }
- if (this.axisConfig.showTitle && this.title) {
- const spaceRequired = this.textDimensionCalculator.getMaxDimension(
- [this.title],
- this.axisConfig.titleFontSize
- );
- const heightRequired = spaceRequired.height + this.axisConfig.titlePadding * 2;
- this.titleTextHeight = spaceRequired.height;
- if (heightRequired <= availableHeight) {
- availableHeight -= heightRequired;
- this.showTitle = true;
- }
- }
- this.boundingRect.width = availableSpace.width;
- this.boundingRect.height = availableSpace.height - availableHeight;
- }
- calculateSpaceIfDrawnVertical(availableSpace) {
- let availableWidth = availableSpace.width;
- if (this.axisConfig.showAxisLine && availableWidth > this.axisConfig.axisLineWidth) {
- availableWidth -= this.axisConfig.axisLineWidth;
- this.showAxisLine = true;
- }
- if (this.axisConfig.showLabel) {
- const spaceRequired = this.getLabelDimension();
- const maxPadding = MAX_OUTER_PADDING_PERCENT_FOR_WRT_LABEL * availableSpace.height;
- this.outerPadding = Math.min(spaceRequired.height / 2, maxPadding);
- const widthRequired = spaceRequired.width + this.axisConfig.labelPadding * 2;
- if (widthRequired <= availableWidth) {
- availableWidth -= widthRequired;
- this.showLabel = true;
- }
- }
- if (this.axisConfig.showTick && availableWidth >= this.axisConfig.tickLength) {
- this.showTick = true;
- availableWidth -= this.axisConfig.tickLength;
- }
- if (this.axisConfig.showTitle && this.title) {
- const spaceRequired = this.textDimensionCalculator.getMaxDimension(
- [this.title],
- this.axisConfig.titleFontSize
- );
- const widthRequired = spaceRequired.height + this.axisConfig.titlePadding * 2;
- this.titleTextHeight = spaceRequired.height;
- if (widthRequired <= availableWidth) {
- availableWidth -= widthRequired;
- this.showTitle = true;
- }
- }
- this.boundingRect.width = availableSpace.width - availableWidth;
- this.boundingRect.height = availableSpace.height;
- }
- calculateSpace(availableSpace) {
- if (this.axisPosition === "left" || this.axisPosition === "right") {
- this.calculateSpaceIfDrawnVertical(availableSpace);
- } else {
- this.calculateSpaceIfDrawnHorizontally(availableSpace);
- }
- this.recalculateScale();
- return {
- width: this.boundingRect.width,
- height: this.boundingRect.height
- };
- }
- setBoundingBoxXY(point) {
- this.boundingRect.x = point.x;
- this.boundingRect.y = point.y;
- }
- getDrawableElementsForLeftAxis() {
- const drawableElement = [];
- if (this.showAxisLine) {
- const x = this.boundingRect.x + this.boundingRect.width - this.axisConfig.axisLineWidth / 2;
- drawableElement.push({
- type: "path",
- groupTexts: ["left-axis", "axisl-line"],
- data: [
- {
- path: `M ${x},${this.boundingRect.y} L ${x},${this.boundingRect.y + this.boundingRect.height} `,
- strokeFill: this.axisThemeConfig.axisLineColor,
- strokeWidth: this.axisConfig.axisLineWidth
- }
- ]
- });
- }
- if (this.showLabel) {
- drawableElement.push({
- type: "text",
- groupTexts: ["left-axis", "label"],
- data: this.getTickValues().map((tick) => ({
- text: tick.toString(),
- x: this.boundingRect.x + this.boundingRect.width - (this.showLabel ? this.axisConfig.labelPadding : 0) - (this.showTick ? this.axisConfig.tickLength : 0) - (this.showAxisLine ? this.axisConfig.axisLineWidth : 0),
- y: this.getScaleValue(tick),
- fill: this.axisThemeConfig.labelColor,
- fontSize: this.axisConfig.labelFontSize,
- rotation: 0,
- verticalPos: "middle",
- horizontalPos: "right"
- }))
- });
- }
- if (this.showTick) {
- const x = this.boundingRect.x + this.boundingRect.width - (this.showAxisLine ? this.axisConfig.axisLineWidth : 0);
- drawableElement.push({
- type: "path",
- groupTexts: ["left-axis", "ticks"],
- data: this.getTickValues().map((tick) => ({
- path: `M ${x},${this.getScaleValue(tick)} L ${x - this.axisConfig.tickLength},${this.getScaleValue(tick)}`,
- strokeFill: this.axisThemeConfig.tickColor,
- strokeWidth: this.axisConfig.tickWidth
- }))
- });
- }
- if (this.showTitle) {
- drawableElement.push({
- type: "text",
- groupTexts: ["left-axis", "title"],
- data: [
- {
- text: this.title,
- x: this.boundingRect.x + this.axisConfig.titlePadding,
- y: this.boundingRect.y + this.boundingRect.height / 2,
- fill: this.axisThemeConfig.titleColor,
- fontSize: this.axisConfig.titleFontSize,
- rotation: 270,
- verticalPos: "top",
- horizontalPos: "center"
- }
- ]
- });
- }
- return drawableElement;
- }
- getDrawableElementsForBottomAxis() {
- const drawableElement = [];
- if (this.showAxisLine) {
- const y = this.boundingRect.y + this.axisConfig.axisLineWidth / 2;
- drawableElement.push({
- type: "path",
- groupTexts: ["bottom-axis", "axis-line"],
- data: [
- {
- path: `M ${this.boundingRect.x},${y} L ${this.boundingRect.x + this.boundingRect.width},${y}`,
- strokeFill: this.axisThemeConfig.axisLineColor,
- strokeWidth: this.axisConfig.axisLineWidth
- }
- ]
- });
- }
- if (this.showLabel) {
- drawableElement.push({
- type: "text",
- groupTexts: ["bottom-axis", "label"],
- data: this.getTickValues().map((tick) => ({
- text: tick.toString(),
- x: this.getScaleValue(tick),
- y: this.boundingRect.y + this.axisConfig.labelPadding + (this.showTick ? this.axisConfig.tickLength : 0) + (this.showAxisLine ? this.axisConfig.axisLineWidth : 0),
- fill: this.axisThemeConfig.labelColor,
- fontSize: this.axisConfig.labelFontSize,
- rotation: 0,
- verticalPos: "top",
- horizontalPos: "center"
- }))
- });
- }
- if (this.showTick) {
- const y = this.boundingRect.y + (this.showAxisLine ? this.axisConfig.axisLineWidth : 0);
- drawableElement.push({
- type: "path",
- groupTexts: ["bottom-axis", "ticks"],
- data: this.getTickValues().map((tick) => ({
- path: `M ${this.getScaleValue(tick)},${y} L ${this.getScaleValue(tick)},${y + this.axisConfig.tickLength}`,
- strokeFill: this.axisThemeConfig.tickColor,
- strokeWidth: this.axisConfig.tickWidth
- }))
- });
- }
- if (this.showTitle) {
- drawableElement.push({
- type: "text",
- groupTexts: ["bottom-axis", "title"],
- data: [
- {
- text: this.title,
- x: this.range[0] + (this.range[1] - this.range[0]) / 2,
- y: this.boundingRect.y + this.boundingRect.height - this.axisConfig.titlePadding - this.titleTextHeight,
- fill: this.axisThemeConfig.titleColor,
- fontSize: this.axisConfig.titleFontSize,
- rotation: 0,
- verticalPos: "top",
- horizontalPos: "center"
- }
- ]
- });
- }
- return drawableElement;
- }
- getDrawableElementsForTopAxis() {
- const drawableElement = [];
- if (this.showAxisLine) {
- const y = this.boundingRect.y + this.boundingRect.height - this.axisConfig.axisLineWidth / 2;
- drawableElement.push({
- type: "path",
- groupTexts: ["top-axis", "axis-line"],
- data: [
- {
- path: `M ${this.boundingRect.x},${y} L ${this.boundingRect.x + this.boundingRect.width},${y}`,
- strokeFill: this.axisThemeConfig.axisLineColor,
- strokeWidth: this.axisConfig.axisLineWidth
- }
- ]
- });
- }
- if (this.showLabel) {
- drawableElement.push({
- type: "text",
- groupTexts: ["top-axis", "label"],
- data: this.getTickValues().map((tick) => ({
- text: tick.toString(),
- x: this.getScaleValue(tick),
- y: this.boundingRect.y + (this.showTitle ? this.titleTextHeight + this.axisConfig.titlePadding * 2 : 0) + this.axisConfig.labelPadding,
- fill: this.axisThemeConfig.labelColor,
- fontSize: this.axisConfig.labelFontSize,
- rotation: 0,
- verticalPos: "top",
- horizontalPos: "center"
- }))
- });
- }
- if (this.showTick) {
- const y = this.boundingRect.y;
- drawableElement.push({
- type: "path",
- groupTexts: ["top-axis", "ticks"],
- data: this.getTickValues().map((tick) => ({
- path: `M ${this.getScaleValue(tick)},${y + this.boundingRect.height - (this.showAxisLine ? this.axisConfig.axisLineWidth : 0)} L ${this.getScaleValue(tick)},${y + this.boundingRect.height - this.axisConfig.tickLength - (this.showAxisLine ? this.axisConfig.axisLineWidth : 0)}`,
- strokeFill: this.axisThemeConfig.tickColor,
- strokeWidth: this.axisConfig.tickWidth
- }))
- });
- }
- if (this.showTitle) {
- drawableElement.push({
- type: "text",
- groupTexts: ["top-axis", "title"],
- data: [
- {
- text: this.title,
- x: this.boundingRect.x + this.boundingRect.width / 2,
- y: this.boundingRect.y + this.axisConfig.titlePadding,
- fill: this.axisThemeConfig.titleColor,
- fontSize: this.axisConfig.titleFontSize,
- rotation: 0,
- verticalPos: "top",
- horizontalPos: "center"
- }
- ]
- });
- }
- return drawableElement;
- }
- getDrawableElements() {
- if (this.axisPosition === "left") {
- return this.getDrawableElementsForLeftAxis();
- }
- if (this.axisPosition === "right") {
- throw Error("Drawing of right axis is not implemented");
- }
- if (this.axisPosition === "bottom") {
- return this.getDrawableElementsForBottomAxis();
- }
- if (this.axisPosition === "top") {
- return this.getDrawableElementsForTopAxis();
- }
- return [];
- }
-}
-class BandAxis extends BaseAxis {
- constructor(axisConfig, axisThemeConfig, categories, title, textDimensionCalculator) {
- super(axisConfig, title, textDimensionCalculator, axisThemeConfig);
- this.categories = categories;
- this.scale = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .scaleBand */ .tiA)().domain(this.categories).range(this.getRange());
- }
- setRange(range) {
- super.setRange(range);
- }
- recalculateScale() {
- this.scale = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .scaleBand */ .tiA)().domain(this.categories).range(this.getRange()).paddingInner(1).paddingOuter(0).align(0.5);
- _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.l.trace("BandAxis axis final categories, range: ", this.categories, this.getRange());
- }
- getTickValues() {
- return this.categories;
- }
- getScaleValue(value) {
- return this.scale(value) || this.getRange()[0];
- }
-}
-class LinearAxis extends BaseAxis {
- constructor(axisConfig, axisThemeConfig, domain, title, textDimensionCalculator) {
- super(axisConfig, title, textDimensionCalculator, axisThemeConfig);
- this.domain = domain;
- this.scale = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .scaleLinear */ .BYU)().domain(this.domain).range(this.getRange());
- }
- getTickValues() {
- return this.scale.ticks();
- }
- recalculateScale() {
- const domain = [...this.domain];
- if (this.axisPosition === "left") {
- domain.reverse();
- }
- this.scale = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .scaleLinear */ .BYU)().domain(domain).range(this.getRange());
- }
- getScaleValue(value) {
- return this.scale(value);
- }
-}
-function getAxis(data, axisConfig, axisThemeConfig, tmpSVGGroup2) {
- const textDimansionCalculator = new TextDimensionCalculatorWithFont(tmpSVGGroup2);
- if (isBandAxisData(data)) {
- return new BandAxis(
- axisConfig,
- axisThemeConfig,
- data.categories,
- data.title,
- textDimansionCalculator
- );
- }
- return new LinearAxis(
- axisConfig,
- axisThemeConfig,
- [data.min, data.max],
- data.title,
- textDimansionCalculator
- );
-}
-class ChartTitle {
- constructor(textDimensionCalculator, chartConfig, chartData, chartThemeConfig) {
- this.textDimensionCalculator = textDimensionCalculator;
- this.chartConfig = chartConfig;
- this.chartData = chartData;
- this.chartThemeConfig = chartThemeConfig;
- this.boundingRect = {
- x: 0,
- y: 0,
- width: 0,
- height: 0
- };
- this.showChartTitle = false;
- }
- setBoundingBoxXY(point) {
- this.boundingRect.x = point.x;
- this.boundingRect.y = point.y;
- }
- calculateSpace(availableSpace) {
- const titleDimension = this.textDimensionCalculator.getMaxDimension(
- [this.chartData.title],
- this.chartConfig.titleFontSize
- );
- const widthRequired = Math.max(titleDimension.width, availableSpace.width);
- const heightRequired = titleDimension.height + 2 * this.chartConfig.titlePadding;
- if (titleDimension.width <= widthRequired && titleDimension.height <= heightRequired && this.chartConfig.showTitle && this.chartData.title) {
- this.boundingRect.width = widthRequired;
- this.boundingRect.height = heightRequired;
- this.showChartTitle = true;
- }
- return {
- width: this.boundingRect.width,
- height: this.boundingRect.height
- };
- }
- getDrawableElements() {
- const drawableElem = [];
- if (this.showChartTitle) {
- drawableElem.push({
- groupTexts: ["chart-title"],
- type: "text",
- data: [
- {
- fontSize: this.chartConfig.titleFontSize,
- text: this.chartData.title,
- verticalPos: "middle",
- horizontalPos: "center",
- x: this.boundingRect.x + this.boundingRect.width / 2,
- y: this.boundingRect.y + this.boundingRect.height / 2,
- fill: this.chartThemeConfig.titleColor,
- rotation: 0
- }
- ]
- });
- }
- return drawableElem;
- }
-}
-function getChartTitleComponent(chartConfig, chartData, chartThemeConfig, tmpSVGGroup2) {
- const textDimensionCalculator = new TextDimensionCalculatorWithFont(tmpSVGGroup2);
- return new ChartTitle(textDimensionCalculator, chartConfig, chartData, chartThemeConfig);
-}
-class LinePlot {
- constructor(plotData, xAxis, yAxis, orientation, plotIndex2) {
- this.plotData = plotData;
- this.xAxis = xAxis;
- this.yAxis = yAxis;
- this.orientation = orientation;
- this.plotIndex = plotIndex2;
- }
- getDrawableElement() {
- const finalData = this.plotData.data.map((d) => [
- this.xAxis.getScaleValue(d[0]),
- this.yAxis.getScaleValue(d[1])
- ]);
- let path;
- if (this.orientation === "horizontal") {
- path = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .line */ .jvg)().y((d) => d[0]).x((d) => d[1])(finalData);
- } else {
- path = (0,d3__WEBPACK_IMPORTED_MODULE_0__/* .line */ .jvg)().x((d) => d[0]).y((d) => d[1])(finalData);
- }
- if (!path) {
- return [];
- }
- return [
- {
- groupTexts: ["plot", `line-plot-${this.plotIndex}`],
- type: "path",
- data: [
- {
- path,
- strokeFill: this.plotData.strokeFill,
- strokeWidth: this.plotData.strokeWidth
- }
- ]
- }
- ];
- }
-}
-class BarPlot {
- constructor(barData, boundingRect, xAxis, yAxis, orientation, plotIndex2) {
- this.barData = barData;
- this.boundingRect = boundingRect;
- this.xAxis = xAxis;
- this.yAxis = yAxis;
- this.orientation = orientation;
- this.plotIndex = plotIndex2;
- }
- getDrawableElement() {
- const finalData = this.barData.data.map((d) => [
- this.xAxis.getScaleValue(d[0]),
- this.yAxis.getScaleValue(d[1])
- ]);
- const barPaddingPercent = 0.05;
- const barWidth = Math.min(this.xAxis.getAxisOuterPadding() * 2, this.xAxis.getTickDistance()) * (1 - barPaddingPercent);
- const barWidthHalf = barWidth / 2;
- if (this.orientation === "horizontal") {
- return [
- {
- groupTexts: ["plot", `bar-plot-${this.plotIndex}`],
- type: "rect",
- data: finalData.map((data) => ({
- x: this.boundingRect.x,
- y: data[0] - barWidthHalf,
- height: barWidth,
- width: data[1] - this.boundingRect.x,
- fill: this.barData.fill,
- strokeWidth: 0,
- strokeFill: this.barData.fill
- }))
- }
- ];
- }
- return [
- {
- groupTexts: ["plot", `bar-plot-${this.plotIndex}`],
- type: "rect",
- data: finalData.map((data) => ({
- x: data[0] - barWidthHalf,
- y: data[1],
- width: barWidth,
- height: this.boundingRect.y + this.boundingRect.height - data[1],
- fill: this.barData.fill,
- strokeWidth: 0,
- strokeFill: this.barData.fill
- }))
- }
- ];
- }
-}
-class BasePlot {
- constructor(chartConfig, chartData, chartThemeConfig) {
- this.chartConfig = chartConfig;
- this.chartData = chartData;
- this.chartThemeConfig = chartThemeConfig;
- this.boundingRect = {
- x: 0,
- y: 0,
- width: 0,
- height: 0
- };
- }
- setAxes(xAxis, yAxis) {
- this.xAxis = xAxis;
- this.yAxis = yAxis;
- }
- setBoundingBoxXY(point) {
- this.boundingRect.x = point.x;
- this.boundingRect.y = point.y;
- }
- calculateSpace(availableSpace) {
- this.boundingRect.width = availableSpace.width;
- this.boundingRect.height = availableSpace.height;
- return {
- width: this.boundingRect.width,
- height: this.boundingRect.height
- };
- }
- getDrawableElements() {
- if (!(this.xAxis && this.yAxis)) {
- throw Error("Axes must be passed to render Plots");
- }
- const drawableElem = [];
- for (const [i, plot] of this.chartData.plots.entries()) {
- switch (plot.type) {
- case "line":
- {
- const linePlot = new LinePlot(
- plot,
- this.xAxis,
- this.yAxis,
- this.chartConfig.chartOrientation,
- i
- );
- drawableElem.push(...linePlot.getDrawableElement());
- }
- break;
- case "bar":
- {
- const barPlot = new BarPlot(
- plot,
- this.boundingRect,
- this.xAxis,
- this.yAxis,
- this.chartConfig.chartOrientation,
- i
- );
- drawableElem.push(...barPlot.getDrawableElement());
- }
- break;
- }
- }
- return drawableElem;
- }
-}
-function getPlotComponent(chartConfig, chartData, chartThemeConfig) {
- return new BasePlot(chartConfig, chartData, chartThemeConfig);
-}
-class Orchestrator {
- constructor(chartConfig, chartData, chartThemeConfig, tmpSVGGroup2) {
- this.chartConfig = chartConfig;
- this.chartData = chartData;
- this.componentStore = {
- title: getChartTitleComponent(chartConfig, chartData, chartThemeConfig, tmpSVGGroup2),
- plot: getPlotComponent(chartConfig, chartData, chartThemeConfig),
- xAxis: getAxis(
- chartData.xAxis,
- chartConfig.xAxis,
- {
- titleColor: chartThemeConfig.xAxisTitleColor,
- labelColor: chartThemeConfig.xAxisLabelColor,
- tickColor: chartThemeConfig.xAxisTickColor,
- axisLineColor: chartThemeConfig.xAxisLineColor
- },
- tmpSVGGroup2
- ),
- yAxis: getAxis(
- chartData.yAxis,
- chartConfig.yAxis,
- {
- titleColor: chartThemeConfig.yAxisTitleColor,
- labelColor: chartThemeConfig.yAxisLabelColor,
- tickColor: chartThemeConfig.yAxisTickColor,
- axisLineColor: chartThemeConfig.yAxisLineColor
- },
- tmpSVGGroup2
- )
- };
- }
- calculateVerticalSpace() {
- let availableWidth = this.chartConfig.width;
- let availableHeight = this.chartConfig.height;
- let plotX = 0;
- let plotY = 0;
- let chartWidth = Math.floor(availableWidth * this.chartConfig.plotReservedSpacePercent / 100);
- let chartHeight = Math.floor(
- availableHeight * this.chartConfig.plotReservedSpacePercent / 100
- );
- let spaceUsed = this.componentStore.plot.calculateSpace({
- width: chartWidth,
- height: chartHeight
- });
- availableWidth -= spaceUsed.width;
- availableHeight -= spaceUsed.height;
- spaceUsed = this.componentStore.title.calculateSpace({
- width: this.chartConfig.width,
- height: availableHeight
- });
- plotY = spaceUsed.height;
- availableHeight -= spaceUsed.height;
- this.componentStore.xAxis.setAxisPosition("bottom");
- spaceUsed = this.componentStore.xAxis.calculateSpace({
- width: availableWidth,
- height: availableHeight
- });
- availableHeight -= spaceUsed.height;
- this.componentStore.yAxis.setAxisPosition("left");
- spaceUsed = this.componentStore.yAxis.calculateSpace({
- width: availableWidth,
- height: availableHeight
- });
- plotX = spaceUsed.width;
- availableWidth -= spaceUsed.width;
- if (availableWidth > 0) {
- chartWidth += availableWidth;
- availableWidth = 0;
- }
- if (availableHeight > 0) {
- chartHeight += availableHeight;
- availableHeight = 0;
- }
- this.componentStore.plot.calculateSpace({
- width: chartWidth,
- height: chartHeight
- });
- this.componentStore.plot.setBoundingBoxXY({ x: plotX, y: plotY });
- this.componentStore.xAxis.setRange([plotX, plotX + chartWidth]);
- this.componentStore.xAxis.setBoundingBoxXY({ x: plotX, y: plotY + chartHeight });
- this.componentStore.yAxis.setRange([plotY, plotY + chartHeight]);
- this.componentStore.yAxis.setBoundingBoxXY({ x: 0, y: plotY });
- if (this.chartData.plots.some((p) => isBarPlot(p))) {
- this.componentStore.xAxis.recalculateOuterPaddingToDrawBar();
- }
- }
- calculateHorizonatalSpace() {
- let availableWidth = this.chartConfig.width;
- let availableHeight = this.chartConfig.height;
- let titleYEnd = 0;
- let plotX = 0;
- let plotY = 0;
- let chartWidth = Math.floor(availableWidth * this.chartConfig.plotReservedSpacePercent / 100);
- let chartHeight = Math.floor(
- availableHeight * this.chartConfig.plotReservedSpacePercent / 100
- );
- let spaceUsed = this.componentStore.plot.calculateSpace({
- width: chartWidth,
- height: chartHeight
- });
- availableWidth -= spaceUsed.width;
- availableHeight -= spaceUsed.height;
- spaceUsed = this.componentStore.title.calculateSpace({
- width: this.chartConfig.width,
- height: availableHeight
- });
- titleYEnd = spaceUsed.height;
- availableHeight -= spaceUsed.height;
- this.componentStore.xAxis.setAxisPosition("left");
- spaceUsed = this.componentStore.xAxis.calculateSpace({
- width: availableWidth,
- height: availableHeight
- });
- availableWidth -= spaceUsed.width;
- plotX = spaceUsed.width;
- this.componentStore.yAxis.setAxisPosition("top");
- spaceUsed = this.componentStore.yAxis.calculateSpace({
- width: availableWidth,
- height: availableHeight
- });
- availableHeight -= spaceUsed.height;
- plotY = titleYEnd + spaceUsed.height;
- if (availableWidth > 0) {
- chartWidth += availableWidth;
- availableWidth = 0;
- }
- if (availableHeight > 0) {
- chartHeight += availableHeight;
- availableHeight = 0;
- }
- this.componentStore.plot.calculateSpace({
- width: chartWidth,
- height: chartHeight
- });
- this.componentStore.plot.setBoundingBoxXY({ x: plotX, y: plotY });
- this.componentStore.yAxis.setRange([plotX, plotX + chartWidth]);
- this.componentStore.yAxis.setBoundingBoxXY({ x: plotX, y: titleYEnd });
- this.componentStore.xAxis.setRange([plotY, plotY + chartHeight]);
- this.componentStore.xAxis.setBoundingBoxXY({ x: 0, y: plotY });
- if (this.chartData.plots.some((p) => isBarPlot(p))) {
- this.componentStore.xAxis.recalculateOuterPaddingToDrawBar();
- }
- }
- calculateSpace() {
- if (this.chartConfig.chartOrientation === "horizontal") {
- this.calculateHorizonatalSpace();
- } else {
- this.calculateVerticalSpace();
- }
- }
- getDrawableElement() {
- this.calculateSpace();
- const drawableElem = [];
- this.componentStore.plot.setAxes(this.componentStore.xAxis, this.componentStore.yAxis);
- for (const component of Object.values(this.componentStore)) {
- drawableElem.push(...component.getDrawableElements());
- }
- return drawableElem;
- }
-}
-class XYChartBuilder {
- static build(config, chartData, chartThemeConfig, tmpSVGGroup2) {
- const orchestrator = new Orchestrator(config, chartData, chartThemeConfig, tmpSVGGroup2);
- return orchestrator.getDrawableElement();
- }
-}
-let plotIndex = 0;
-let tmpSVGGroup;
-let xyChartConfig = getChartDefaultConfig();
-let xyChartThemeConfig = getChartDefaultThemeConfig();
-let xyChartData = getChartDefaultData();
-let plotColorPalette = xyChartThemeConfig.plotColorPalette.split(",").map((color) => color.trim());
-let hasSetXAxis = false;
-let hasSetYAxis = false;
-function getChartDefaultThemeConfig() {
- const defaultThemeVariables = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.D)();
- const config = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.E)();
- return (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.B)(defaultThemeVariables.xyChart, config.themeVariables.xyChart);
-}
-function getChartDefaultConfig() {
- const config = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.E)();
- return (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.B)(
- _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.A.xyChart,
- config.xyChart
- );
-}
-function getChartDefaultData() {
- return {
- yAxis: {
- type: "linear",
- title: "",
- min: Infinity,
- max: -Infinity
- },
- xAxis: {
- type: "band",
- title: "",
- categories: []
- },
- title: "",
- plots: []
- };
-}
-function textSanitizer(text) {
- const config = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.E)();
- return (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.d)(text.trim(), config);
-}
-function setTmpSVGG(SVGG) {
- tmpSVGGroup = SVGG;
-}
-function setOrientation(orientation) {
- if (orientation === "horizontal") {
- xyChartConfig.chartOrientation = "horizontal";
- } else {
- xyChartConfig.chartOrientation = "vertical";
- }
-}
-function setXAxisTitle(title) {
- xyChartData.xAxis.title = textSanitizer(title.text);
-}
-function setXAxisRangeData(min, max) {
- xyChartData.xAxis = { type: "linear", title: xyChartData.xAxis.title, min, max };
- hasSetXAxis = true;
-}
-function setXAxisBand(categories) {
- xyChartData.xAxis = {
- type: "band",
- title: xyChartData.xAxis.title,
- categories: categories.map((c) => textSanitizer(c.text))
- };
- hasSetXAxis = true;
-}
-function setYAxisTitle(title) {
- xyChartData.yAxis.title = textSanitizer(title.text);
-}
-function setYAxisRangeData(min, max) {
- xyChartData.yAxis = { type: "linear", title: xyChartData.yAxis.title, min, max };
- hasSetYAxis = true;
-}
-function setYAxisRangeFromPlotData(data) {
- const minValue = Math.min(...data);
- const maxValue = Math.max(...data);
- const prevMinValue = isLinearAxisData(xyChartData.yAxis) ? xyChartData.yAxis.min : Infinity;
- const prevMaxValue = isLinearAxisData(xyChartData.yAxis) ? xyChartData.yAxis.max : -Infinity;
- xyChartData.yAxis = {
- type: "linear",
- title: xyChartData.yAxis.title,
- min: Math.min(prevMinValue, minValue),
- max: Math.max(prevMaxValue, maxValue)
- };
-}
-function transformDataWithoutCategory(data) {
- let retData = [];
- if (data.length === 0) {
- return retData;
- }
- if (!hasSetXAxis) {
- const prevMinValue = isLinearAxisData(xyChartData.xAxis) ? xyChartData.xAxis.min : Infinity;
- const prevMaxValue = isLinearAxisData(xyChartData.xAxis) ? xyChartData.xAxis.max : -Infinity;
- setXAxisRangeData(Math.min(prevMinValue, 1), Math.max(prevMaxValue, data.length));
- }
- if (!hasSetYAxis) {
- setYAxisRangeFromPlotData(data);
- }
- if (isBandAxisData(xyChartData.xAxis)) {
- retData = xyChartData.xAxis.categories.map((c, i) => [c, data[i]]);
- }
- if (isLinearAxisData(xyChartData.xAxis)) {
- const min = xyChartData.xAxis.min;
- const max = xyChartData.xAxis.max;
- const step = (max - min + 1) / data.length;
- const categories = [];
- for (let i = min; i <= max; i += step) {
- categories.push(`${i}`);
- }
- retData = categories.map((c, i) => [c, data[i]]);
- }
- return retData;
-}
-function getPlotColorFromPalette(plotIndex2) {
- return plotColorPalette[plotIndex2 === 0 ? 0 : plotIndex2 % plotColorPalette.length];
-}
-function setLineData(title, data) {
- const plotData = transformDataWithoutCategory(data);
- xyChartData.plots.push({
- type: "line",
- strokeFill: getPlotColorFromPalette(plotIndex),
- strokeWidth: 2,
- data: plotData
- });
- plotIndex++;
-}
-function setBarData(title, data) {
- const plotData = transformDataWithoutCategory(data);
- xyChartData.plots.push({
- type: "bar",
- fill: getPlotColorFromPalette(plotIndex),
- data: plotData
- });
- plotIndex++;
-}
-function getDrawableElem() {
- if (xyChartData.plots.length === 0) {
- throw Error("No Plot to render, please provide a plot with some data");
- }
- xyChartData.title = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.r)();
- return XYChartBuilder.build(xyChartConfig, xyChartData, xyChartThemeConfig, tmpSVGGroup);
-}
-function getChartThemeConfig() {
- return xyChartThemeConfig;
-}
-function getChartConfig() {
- return xyChartConfig;
-}
-const clear = function() {
- (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.t)();
- plotIndex = 0;
- xyChartConfig = getChartDefaultConfig();
- xyChartData = getChartDefaultData();
- xyChartThemeConfig = getChartDefaultThemeConfig();
- plotColorPalette = xyChartThemeConfig.plotColorPalette.split(",").map((color) => color.trim());
- hasSetXAxis = false;
- hasSetYAxis = false;
-};
-const db = {
- getDrawableElem,
- clear,
- setAccTitle: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.s,
- getAccTitle: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.g,
- setDiagramTitle: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.q,
- getDiagramTitle: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.r,
- getAccDescription: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.a,
- setAccDescription: _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.b,
- setOrientation,
- setXAxisTitle,
- setXAxisRangeData,
- setXAxisBand,
- setYAxisTitle,
- setYAxisRangeData,
- setLineData,
- setBarData,
- setTmpSVGG,
- getChartThemeConfig,
- getChartConfig
-};
-const draw = (txt, id, _version, diagObj) => {
- const db2 = diagObj.db;
- const themeConfig = db2.getChartThemeConfig();
- const chartConfig = db2.getChartConfig();
- function getDominantBaseLine(horizontalPos) {
- return horizontalPos === "top" ? "text-before-edge" : "middle";
- }
- function getTextAnchor(verticalPos) {
- return verticalPos === "left" ? "start" : verticalPos === "right" ? "end" : "middle";
- }
- function getTextTransformation(data) {
- return `translate(${data.x}, ${data.y}) rotate(${data.rotation || 0})`;
- }
- _mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.l.debug("Rendering xychart chart\n" + txt);
- const svg = (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.z)(id);
- const group = svg.append("g").attr("class", "main");
- const background = group.append("rect").attr("width", chartConfig.width).attr("height", chartConfig.height).attr("class", "background");
- (0,_mermaid_934d9bea_js__WEBPACK_IMPORTED_MODULE_5__.i)(svg, chartConfig.height, chartConfig.width, true);
- svg.attr("viewBox", `0 0 ${chartConfig.width} ${chartConfig.height}`);
- background.attr("fill", themeConfig.backgroundColor);
- db2.setTmpSVGG(svg.append("g").attr("class", "mermaid-tmp-group"));
- const shapes = db2.getDrawableElem();
- const groups = {};
- function getGroup(gList) {
- let elem = group;
- let prefix = "";
- for (const [i] of gList.entries()) {
- let parent = group;
- if (i > 0 && groups[prefix]) {
- parent = groups[prefix];
- }
- prefix += gList[i];
- elem = groups[prefix];
- if (!elem) {
- elem = groups[prefix] = parent.append("g").attr("class", gList[i]);
- }
- }
- return elem;
- }
- for (const shape of shapes) {
- if (shape.data.length === 0) {
- continue;
- }
- const shapeGroup = getGroup(shape.groupTexts);
- switch (shape.type) {
- case "rect":
- shapeGroup.selectAll("rect").data(shape.data).enter().append("rect").attr("x", (data) => data.x).attr("y", (data) => data.y).attr("width", (data) => data.width).attr("height", (data) => data.height).attr("fill", (data) => data.fill).attr("stroke", (data) => data.strokeFill).attr("stroke-width", (data) => data.strokeWidth);
- break;
- case "text":
- shapeGroup.selectAll("text").data(shape.data).enter().append("text").attr("x", 0).attr("y", 0).attr("fill", (data) => data.fill).attr("font-size", (data) => data.fontSize).attr("dominant-baseline", (data) => getDominantBaseLine(data.verticalPos)).attr("text-anchor", (data) => getTextAnchor(data.horizontalPos)).attr("transform", (data) => getTextTransformation(data)).text((data) => data.text);
- break;
- case "path":
- shapeGroup.selectAll("path").data(shape.data).enter().append("path").attr("d", (data) => data.path).attr("fill", (data) => data.fill ? data.fill : "none").attr("stroke", (data) => data.strokeFill).attr("stroke-width", (data) => data.strokeWidth);
- break;
- }
- }
-};
-const renderer = {
- draw
-};
-const diagram = {
- parser: parser$1,
- db,
- renderer
-};
-
-
-
-/***/ })
-
-};
-; \ No newline at end of file