summaryrefslogtreecommitdiff
path: root/node_modules/markdown-it/dist/markdown-it.js
diff options
context:
space:
mode:
authorMinteck <contact@minteck.org>2022-01-20 13:43:34 +0100
committerMinteck <contact@minteck.org>2022-01-20 13:43:34 +0100
commitc2aa7bf38fb30de2d04f87f8e7780e4c768ae6b1 (patch)
tree226598e8d17d20e3721358f7c60b1cc6b851163a /node_modules/markdown-it/dist/markdown-it.js
downloadcobalt-c2aa7bf38fb30de2d04f87f8e7780e4c768ae6b1.tar.gz
cobalt-c2aa7bf38fb30de2d04f87f8e7780e4c768ae6b1.tar.bz2
cobalt-c2aa7bf38fb30de2d04f87f8e7780e4c768ae6b1.zip
Initial commit
Diffstat (limited to 'node_modules/markdown-it/dist/markdown-it.js')
-rw-r--r--node_modules/markdown-it/dist/markdown-it.js8356
1 files changed, 8356 insertions, 0 deletions
diff --git a/node_modules/markdown-it/dist/markdown-it.js b/node_modules/markdown-it/dist/markdown-it.js
new file mode 100644
index 0000000..1ef5530
--- /dev/null
+++ b/node_modules/markdown-it/dist/markdown-it.js
@@ -0,0 +1,8356 @@
+/*! markdown-it 12.3.2 https://github.com/markdown-it/markdown-it @license MIT */
+(function(global, factory) {
+ typeof exports === "object" && typeof module !== "undefined" ? module.exports = factory() : typeof define === "function" && define.amd ? define(factory) : (global = typeof globalThis !== "undefined" ? globalThis : global || self,
+ global.markdownit = factory());
+})(this, (function() {
+ "use strict";
+ function createCommonjsModule(fn, basedir, module) {
+ return module = {
+ path: basedir,
+ exports: {},
+ require: function(path, base) {
+ return commonjsRequire(path, base === undefined || base === null ? module.path : base);
+ }
+ }, fn(module, module.exports), module.exports;
+ }
+ function getAugmentedNamespace(n) {
+ if (n.__esModule) return n;
+ var a = Object.defineProperty({}, "__esModule", {
+ value: true
+ });
+ Object.keys(n).forEach((function(k) {
+ var d = Object.getOwnPropertyDescriptor(n, k);
+ Object.defineProperty(a, k, d.get ? d : {
+ enumerable: true,
+ get: function() {
+ return n[k];
+ }
+ });
+ }));
+ return a;
+ }
+ function commonjsRequire() {
+ throw new Error("Dynamic requires are not currently supported by @rollup/plugin-commonjs");
+ }
+ var require$$0 = {
+ Aacute: "\xc1",
+ aacute: "\xe1",
+ Abreve: "\u0102",
+ abreve: "\u0103",
+ ac: "\u223e",
+ acd: "\u223f",
+ acE: "\u223e\u0333",
+ Acirc: "\xc2",
+ acirc: "\xe2",
+ acute: "\xb4",
+ Acy: "\u0410",
+ acy: "\u0430",
+ AElig: "\xc6",
+ aelig: "\xe6",
+ af: "\u2061",
+ Afr: "\ud835\udd04",
+ afr: "\ud835\udd1e",
+ Agrave: "\xc0",
+ agrave: "\xe0",
+ alefsym: "\u2135",
+ aleph: "\u2135",
+ Alpha: "\u0391",
+ alpha: "\u03b1",
+ Amacr: "\u0100",
+ amacr: "\u0101",
+ amalg: "\u2a3f",
+ amp: "&",
+ AMP: "&",
+ andand: "\u2a55",
+ And: "\u2a53",
+ and: "\u2227",
+ andd: "\u2a5c",
+ andslope: "\u2a58",
+ andv: "\u2a5a",
+ ang: "\u2220",
+ ange: "\u29a4",
+ angle: "\u2220",
+ angmsdaa: "\u29a8",
+ angmsdab: "\u29a9",
+ angmsdac: "\u29aa",
+ angmsdad: "\u29ab",
+ angmsdae: "\u29ac",
+ angmsdaf: "\u29ad",
+ angmsdag: "\u29ae",
+ angmsdah: "\u29af",
+ angmsd: "\u2221",
+ angrt: "\u221f",
+ angrtvb: "\u22be",
+ angrtvbd: "\u299d",
+ angsph: "\u2222",
+ angst: "\xc5",
+ angzarr: "\u237c",
+ Aogon: "\u0104",
+ aogon: "\u0105",
+ Aopf: "\ud835\udd38",
+ aopf: "\ud835\udd52",
+ apacir: "\u2a6f",
+ ap: "\u2248",
+ apE: "\u2a70",
+ ape: "\u224a",
+ apid: "\u224b",
+ apos: "'",
+ ApplyFunction: "\u2061",
+ approx: "\u2248",
+ approxeq: "\u224a",
+ Aring: "\xc5",
+ aring: "\xe5",
+ Ascr: "\ud835\udc9c",
+ ascr: "\ud835\udcb6",
+ Assign: "\u2254",
+ ast: "*",
+ asymp: "\u2248",
+ asympeq: "\u224d",
+ Atilde: "\xc3",
+ atilde: "\xe3",
+ Auml: "\xc4",
+ auml: "\xe4",
+ awconint: "\u2233",
+ awint: "\u2a11",
+ backcong: "\u224c",
+ backepsilon: "\u03f6",
+ backprime: "\u2035",
+ backsim: "\u223d",
+ backsimeq: "\u22cd",
+ Backslash: "\u2216",
+ Barv: "\u2ae7",
+ barvee: "\u22bd",
+ barwed: "\u2305",
+ Barwed: "\u2306",
+ barwedge: "\u2305",
+ bbrk: "\u23b5",
+ bbrktbrk: "\u23b6",
+ bcong: "\u224c",
+ Bcy: "\u0411",
+ bcy: "\u0431",
+ bdquo: "\u201e",
+ becaus: "\u2235",
+ because: "\u2235",
+ Because: "\u2235",
+ bemptyv: "\u29b0",
+ bepsi: "\u03f6",
+ bernou: "\u212c",
+ Bernoullis: "\u212c",
+ Beta: "\u0392",
+ beta: "\u03b2",
+ beth: "\u2136",
+ between: "\u226c",
+ Bfr: "\ud835\udd05",
+ bfr: "\ud835\udd1f",
+ bigcap: "\u22c2",
+ bigcirc: "\u25ef",
+ bigcup: "\u22c3",
+ bigodot: "\u2a00",
+ bigoplus: "\u2a01",
+ bigotimes: "\u2a02",
+ bigsqcup: "\u2a06",
+ bigstar: "\u2605",
+ bigtriangledown: "\u25bd",
+ bigtriangleup: "\u25b3",
+ biguplus: "\u2a04",
+ bigvee: "\u22c1",
+ bigwedge: "\u22c0",
+ bkarow: "\u290d",
+ blacklozenge: "\u29eb",
+ blacksquare: "\u25aa",
+ blacktriangle: "\u25b4",
+ blacktriangledown: "\u25be",
+ blacktriangleleft: "\u25c2",
+ blacktriangleright: "\u25b8",
+ blank: "\u2423",
+ blk12: "\u2592",
+ blk14: "\u2591",
+ blk34: "\u2593",
+ block: "\u2588",
+ bne: "=\u20e5",
+ bnequiv: "\u2261\u20e5",
+ bNot: "\u2aed",
+ bnot: "\u2310",
+ Bopf: "\ud835\udd39",
+ bopf: "\ud835\udd53",
+ bot: "\u22a5",
+ bottom: "\u22a5",
+ bowtie: "\u22c8",
+ boxbox: "\u29c9",
+ boxdl: "\u2510",
+ boxdL: "\u2555",
+ boxDl: "\u2556",
+ boxDL: "\u2557",
+ boxdr: "\u250c",
+ boxdR: "\u2552",
+ boxDr: "\u2553",
+ boxDR: "\u2554",
+ boxh: "\u2500",
+ boxH: "\u2550",
+ boxhd: "\u252c",
+ boxHd: "\u2564",
+ boxhD: "\u2565",
+ boxHD: "\u2566",
+ boxhu: "\u2534",
+ boxHu: "\u2567",
+ boxhU: "\u2568",
+ boxHU: "\u2569",
+ boxminus: "\u229f",
+ boxplus: "\u229e",
+ boxtimes: "\u22a0",
+ boxul: "\u2518",
+ boxuL: "\u255b",
+ boxUl: "\u255c",
+ boxUL: "\u255d",
+ boxur: "\u2514",
+ boxuR: "\u2558",
+ boxUr: "\u2559",
+ boxUR: "\u255a",
+ boxv: "\u2502",
+ boxV: "\u2551",
+ boxvh: "\u253c",
+ boxvH: "\u256a",
+ boxVh: "\u256b",
+ boxVH: "\u256c",
+ boxvl: "\u2524",
+ boxvL: "\u2561",
+ boxVl: "\u2562",
+ boxVL: "\u2563",
+ boxvr: "\u251c",
+ boxvR: "\u255e",
+ boxVr: "\u255f",
+ boxVR: "\u2560",
+ bprime: "\u2035",
+ breve: "\u02d8",
+ Breve: "\u02d8",
+ brvbar: "\xa6",
+ bscr: "\ud835\udcb7",
+ Bscr: "\u212c",
+ bsemi: "\u204f",
+ bsim: "\u223d",
+ bsime: "\u22cd",
+ bsolb: "\u29c5",
+ bsol: "\\",
+ bsolhsub: "\u27c8",
+ bull: "\u2022",
+ bullet: "\u2022",
+ bump: "\u224e",
+ bumpE: "\u2aae",
+ bumpe: "\u224f",
+ Bumpeq: "\u224e",
+ bumpeq: "\u224f",
+ Cacute: "\u0106",
+ cacute: "\u0107",
+ capand: "\u2a44",
+ capbrcup: "\u2a49",
+ capcap: "\u2a4b",
+ cap: "\u2229",
+ Cap: "\u22d2",
+ capcup: "\u2a47",
+ capdot: "\u2a40",
+ CapitalDifferentialD: "\u2145",
+ caps: "\u2229\ufe00",
+ caret: "\u2041",
+ caron: "\u02c7",
+ Cayleys: "\u212d",
+ ccaps: "\u2a4d",
+ Ccaron: "\u010c",
+ ccaron: "\u010d",
+ Ccedil: "\xc7",
+ ccedil: "\xe7",
+ Ccirc: "\u0108",
+ ccirc: "\u0109",
+ Cconint: "\u2230",
+ ccups: "\u2a4c",
+ ccupssm: "\u2a50",
+ Cdot: "\u010a",
+ cdot: "\u010b",
+ cedil: "\xb8",
+ Cedilla: "\xb8",
+ cemptyv: "\u29b2",
+ cent: "\xa2",
+ centerdot: "\xb7",
+ CenterDot: "\xb7",
+ cfr: "\ud835\udd20",
+ Cfr: "\u212d",
+ CHcy: "\u0427",
+ chcy: "\u0447",
+ check: "\u2713",
+ checkmark: "\u2713",
+ Chi: "\u03a7",
+ chi: "\u03c7",
+ circ: "\u02c6",
+ circeq: "\u2257",
+ circlearrowleft: "\u21ba",
+ circlearrowright: "\u21bb",
+ circledast: "\u229b",
+ circledcirc: "\u229a",
+ circleddash: "\u229d",
+ CircleDot: "\u2299",
+ circledR: "\xae",
+ circledS: "\u24c8",
+ CircleMinus: "\u2296",
+ CirclePlus: "\u2295",
+ CircleTimes: "\u2297",
+ cir: "\u25cb",
+ cirE: "\u29c3",
+ cire: "\u2257",
+ cirfnint: "\u2a10",
+ cirmid: "\u2aef",
+ cirscir: "\u29c2",
+ ClockwiseContourIntegral: "\u2232",
+ CloseCurlyDoubleQuote: "\u201d",
+ CloseCurlyQuote: "\u2019",
+ clubs: "\u2663",
+ clubsuit: "\u2663",
+ colon: ":",
+ Colon: "\u2237",
+ Colone: "\u2a74",
+ colone: "\u2254",
+ coloneq: "\u2254",
+ comma: ",",
+ commat: "@",
+ comp: "\u2201",
+ compfn: "\u2218",
+ complement: "\u2201",
+ complexes: "\u2102",
+ cong: "\u2245",
+ congdot: "\u2a6d",
+ Congruent: "\u2261",
+ conint: "\u222e",
+ Conint: "\u222f",
+ ContourIntegral: "\u222e",
+ copf: "\ud835\udd54",
+ Copf: "\u2102",
+ coprod: "\u2210",
+ Coproduct: "\u2210",
+ copy: "\xa9",
+ COPY: "\xa9",
+ copysr: "\u2117",
+ CounterClockwiseContourIntegral: "\u2233",
+ crarr: "\u21b5",
+ cross: "\u2717",
+ Cross: "\u2a2f",
+ Cscr: "\ud835\udc9e",
+ cscr: "\ud835\udcb8",
+ csub: "\u2acf",
+ csube: "\u2ad1",
+ csup: "\u2ad0",
+ csupe: "\u2ad2",
+ ctdot: "\u22ef",
+ cudarrl: "\u2938",
+ cudarrr: "\u2935",
+ cuepr: "\u22de",
+ cuesc: "\u22df",
+ cularr: "\u21b6",
+ cularrp: "\u293d",
+ cupbrcap: "\u2a48",
+ cupcap: "\u2a46",
+ CupCap: "\u224d",
+ cup: "\u222a",
+ Cup: "\u22d3",
+ cupcup: "\u2a4a",
+ cupdot: "\u228d",
+ cupor: "\u2a45",
+ cups: "\u222a\ufe00",
+ curarr: "\u21b7",
+ curarrm: "\u293c",
+ curlyeqprec: "\u22de",
+ curlyeqsucc: "\u22df",
+ curlyvee: "\u22ce",
+ curlywedge: "\u22cf",
+ curren: "\xa4",
+ curvearrowleft: "\u21b6",
+ curvearrowright: "\u21b7",
+ cuvee: "\u22ce",
+ cuwed: "\u22cf",
+ cwconint: "\u2232",
+ cwint: "\u2231",
+ cylcty: "\u232d",
+ dagger: "\u2020",
+ Dagger: "\u2021",
+ daleth: "\u2138",
+ darr: "\u2193",
+ Darr: "\u21a1",
+ dArr: "\u21d3",
+ dash: "\u2010",
+ Dashv: "\u2ae4",
+ dashv: "\u22a3",
+ dbkarow: "\u290f",
+ dblac: "\u02dd",
+ Dcaron: "\u010e",
+ dcaron: "\u010f",
+ Dcy: "\u0414",
+ dcy: "\u0434",
+ ddagger: "\u2021",
+ ddarr: "\u21ca",
+ DD: "\u2145",
+ dd: "\u2146",
+ DDotrahd: "\u2911",
+ ddotseq: "\u2a77",
+ deg: "\xb0",
+ Del: "\u2207",
+ Delta: "\u0394",
+ delta: "\u03b4",
+ demptyv: "\u29b1",
+ dfisht: "\u297f",
+ Dfr: "\ud835\udd07",
+ dfr: "\ud835\udd21",
+ dHar: "\u2965",
+ dharl: "\u21c3",
+ dharr: "\u21c2",
+ DiacriticalAcute: "\xb4",
+ DiacriticalDot: "\u02d9",
+ DiacriticalDoubleAcute: "\u02dd",
+ DiacriticalGrave: "`",
+ DiacriticalTilde: "\u02dc",
+ diam: "\u22c4",
+ diamond: "\u22c4",
+ Diamond: "\u22c4",
+ diamondsuit: "\u2666",
+ diams: "\u2666",
+ die: "\xa8",
+ DifferentialD: "\u2146",
+ digamma: "\u03dd",
+ disin: "\u22f2",
+ div: "\xf7",
+ divide: "\xf7",
+ divideontimes: "\u22c7",
+ divonx: "\u22c7",
+ DJcy: "\u0402",
+ djcy: "\u0452",
+ dlcorn: "\u231e",
+ dlcrop: "\u230d",
+ dollar: "$",
+ Dopf: "\ud835\udd3b",
+ dopf: "\ud835\udd55",
+ Dot: "\xa8",
+ dot: "\u02d9",
+ DotDot: "\u20dc",
+ doteq: "\u2250",
+ doteqdot: "\u2251",
+ DotEqual: "\u2250",
+ dotminus: "\u2238",
+ dotplus: "\u2214",
+ dotsquare: "\u22a1",
+ doublebarwedge: "\u2306",
+ DoubleContourIntegral: "\u222f",
+ DoubleDot: "\xa8",
+ DoubleDownArrow: "\u21d3",
+ DoubleLeftArrow: "\u21d0",
+ DoubleLeftRightArrow: "\u21d4",
+ DoubleLeftTee: "\u2ae4",
+ DoubleLongLeftArrow: "\u27f8",
+ DoubleLongLeftRightArrow: "\u27fa",
+ DoubleLongRightArrow: "\u27f9",
+ DoubleRightArrow: "\u21d2",
+ DoubleRightTee: "\u22a8",
+ DoubleUpArrow: "\u21d1",
+ DoubleUpDownArrow: "\u21d5",
+ DoubleVerticalBar: "\u2225",
+ DownArrowBar: "\u2913",
+ downarrow: "\u2193",
+ DownArrow: "\u2193",
+ Downarrow: "\u21d3",
+ DownArrowUpArrow: "\u21f5",
+ DownBreve: "\u0311",
+ downdownarrows: "\u21ca",
+ downharpoonleft: "\u21c3",
+ downharpoonright: "\u21c2",
+ DownLeftRightVector: "\u2950",
+ DownLeftTeeVector: "\u295e",
+ DownLeftVectorBar: "\u2956",
+ DownLeftVector: "\u21bd",
+ DownRightTeeVector: "\u295f",
+ DownRightVectorBar: "\u2957",
+ DownRightVector: "\u21c1",
+ DownTeeArrow: "\u21a7",
+ DownTee: "\u22a4",
+ drbkarow: "\u2910",
+ drcorn: "\u231f",
+ drcrop: "\u230c",
+ Dscr: "\ud835\udc9f",
+ dscr: "\ud835\udcb9",
+ DScy: "\u0405",
+ dscy: "\u0455",
+ dsol: "\u29f6",
+ Dstrok: "\u0110",
+ dstrok: "\u0111",
+ dtdot: "\u22f1",
+ dtri: "\u25bf",
+ dtrif: "\u25be",
+ duarr: "\u21f5",
+ duhar: "\u296f",
+ dwangle: "\u29a6",
+ DZcy: "\u040f",
+ dzcy: "\u045f",
+ dzigrarr: "\u27ff",
+ Eacute: "\xc9",
+ eacute: "\xe9",
+ easter: "\u2a6e",
+ Ecaron: "\u011a",
+ ecaron: "\u011b",
+ Ecirc: "\xca",
+ ecirc: "\xea",
+ ecir: "\u2256",
+ ecolon: "\u2255",
+ Ecy: "\u042d",
+ ecy: "\u044d",
+ eDDot: "\u2a77",
+ Edot: "\u0116",
+ edot: "\u0117",
+ eDot: "\u2251",
+ ee: "\u2147",
+ efDot: "\u2252",
+ Efr: "\ud835\udd08",
+ efr: "\ud835\udd22",
+ eg: "\u2a9a",
+ Egrave: "\xc8",
+ egrave: "\xe8",
+ egs: "\u2a96",
+ egsdot: "\u2a98",
+ el: "\u2a99",
+ Element: "\u2208",
+ elinters: "\u23e7",
+ ell: "\u2113",
+ els: "\u2a95",
+ elsdot: "\u2a97",
+ Emacr: "\u0112",
+ emacr: "\u0113",
+ empty: "\u2205",
+ emptyset: "\u2205",
+ EmptySmallSquare: "\u25fb",
+ emptyv: "\u2205",
+ EmptyVerySmallSquare: "\u25ab",
+ emsp13: "\u2004",
+ emsp14: "\u2005",
+ emsp: "\u2003",
+ ENG: "\u014a",
+ eng: "\u014b",
+ ensp: "\u2002",
+ Eogon: "\u0118",
+ eogon: "\u0119",
+ Eopf: "\ud835\udd3c",
+ eopf: "\ud835\udd56",
+ epar: "\u22d5",
+ eparsl: "\u29e3",
+ eplus: "\u2a71",
+ epsi: "\u03b5",
+ Epsilon: "\u0395",
+ epsilon: "\u03b5",
+ epsiv: "\u03f5",
+ eqcirc: "\u2256",
+ eqcolon: "\u2255",
+ eqsim: "\u2242",
+ eqslantgtr: "\u2a96",
+ eqslantless: "\u2a95",
+ Equal: "\u2a75",
+ equals: "=",
+ EqualTilde: "\u2242",
+ equest: "\u225f",
+ Equilibrium: "\u21cc",
+ equiv: "\u2261",
+ equivDD: "\u2a78",
+ eqvparsl: "\u29e5",
+ erarr: "\u2971",
+ erDot: "\u2253",
+ escr: "\u212f",
+ Escr: "\u2130",
+ esdot: "\u2250",
+ Esim: "\u2a73",
+ esim: "\u2242",
+ Eta: "\u0397",
+ eta: "\u03b7",
+ ETH: "\xd0",
+ eth: "\xf0",
+ Euml: "\xcb",
+ euml: "\xeb",
+ euro: "\u20ac",
+ excl: "!",
+ exist: "\u2203",
+ Exists: "\u2203",
+ expectation: "\u2130",
+ exponentiale: "\u2147",
+ ExponentialE: "\u2147",
+ fallingdotseq: "\u2252",
+ Fcy: "\u0424",
+ fcy: "\u0444",
+ female: "\u2640",
+ ffilig: "\ufb03",
+ fflig: "\ufb00",
+ ffllig: "\ufb04",
+ Ffr: "\ud835\udd09",
+ ffr: "\ud835\udd23",
+ filig: "\ufb01",
+ FilledSmallSquare: "\u25fc",
+ FilledVerySmallSquare: "\u25aa",
+ fjlig: "fj",
+ flat: "\u266d",
+ fllig: "\ufb02",
+ fltns: "\u25b1",
+ fnof: "\u0192",
+ Fopf: "\ud835\udd3d",
+ fopf: "\ud835\udd57",
+ forall: "\u2200",
+ ForAll: "\u2200",
+ fork: "\u22d4",
+ forkv: "\u2ad9",
+ Fouriertrf: "\u2131",
+ fpartint: "\u2a0d",
+ frac12: "\xbd",
+ frac13: "\u2153",
+ frac14: "\xbc",
+ frac15: "\u2155",
+ frac16: "\u2159",
+ frac18: "\u215b",
+ frac23: "\u2154",
+ frac25: "\u2156",
+ frac34: "\xbe",
+ frac35: "\u2157",
+ frac38: "\u215c",
+ frac45: "\u2158",
+ frac56: "\u215a",
+ frac58: "\u215d",
+ frac78: "\u215e",
+ frasl: "\u2044",
+ frown: "\u2322",
+ fscr: "\ud835\udcbb",
+ Fscr: "\u2131",
+ gacute: "\u01f5",
+ Gamma: "\u0393",
+ gamma: "\u03b3",
+ Gammad: "\u03dc",
+ gammad: "\u03dd",
+ gap: "\u2a86",
+ Gbreve: "\u011e",
+ gbreve: "\u011f",
+ Gcedil: "\u0122",
+ Gcirc: "\u011c",
+ gcirc: "\u011d",
+ Gcy: "\u0413",
+ gcy: "\u0433",
+ Gdot: "\u0120",
+ gdot: "\u0121",
+ ge: "\u2265",
+ gE: "\u2267",
+ gEl: "\u2a8c",
+ gel: "\u22db",
+ geq: "\u2265",
+ geqq: "\u2267",
+ geqslant: "\u2a7e",
+ gescc: "\u2aa9",
+ ges: "\u2a7e",
+ gesdot: "\u2a80",
+ gesdoto: "\u2a82",
+ gesdotol: "\u2a84",
+ gesl: "\u22db\ufe00",
+ gesles: "\u2a94",
+ Gfr: "\ud835\udd0a",
+ gfr: "\ud835\udd24",
+ gg: "\u226b",
+ Gg: "\u22d9",
+ ggg: "\u22d9",
+ gimel: "\u2137",
+ GJcy: "\u0403",
+ gjcy: "\u0453",
+ gla: "\u2aa5",
+ gl: "\u2277",
+ glE: "\u2a92",
+ glj: "\u2aa4",
+ gnap: "\u2a8a",
+ gnapprox: "\u2a8a",
+ gne: "\u2a88",
+ gnE: "\u2269",
+ gneq: "\u2a88",
+ gneqq: "\u2269",
+ gnsim: "\u22e7",
+ Gopf: "\ud835\udd3e",
+ gopf: "\ud835\udd58",
+ grave: "`",
+ GreaterEqual: "\u2265",
+ GreaterEqualLess: "\u22db",
+ GreaterFullEqual: "\u2267",
+ GreaterGreater: "\u2aa2",
+ GreaterLess: "\u2277",
+ GreaterSlantEqual: "\u2a7e",
+ GreaterTilde: "\u2273",
+ Gscr: "\ud835\udca2",
+ gscr: "\u210a",
+ gsim: "\u2273",
+ gsime: "\u2a8e",
+ gsiml: "\u2a90",
+ gtcc: "\u2aa7",
+ gtcir: "\u2a7a",
+ gt: ">",
+ GT: ">",
+ Gt: "\u226b",
+ gtdot: "\u22d7",
+ gtlPar: "\u2995",
+ gtquest: "\u2a7c",
+ gtrapprox: "\u2a86",
+ gtrarr: "\u2978",
+ gtrdot: "\u22d7",
+ gtreqless: "\u22db",
+ gtreqqless: "\u2a8c",
+ gtrless: "\u2277",
+ gtrsim: "\u2273",
+ gvertneqq: "\u2269\ufe00",
+ gvnE: "\u2269\ufe00",
+ Hacek: "\u02c7",
+ hairsp: "\u200a",
+ half: "\xbd",
+ hamilt: "\u210b",
+ HARDcy: "\u042a",
+ hardcy: "\u044a",
+ harrcir: "\u2948",
+ harr: "\u2194",
+ hArr: "\u21d4",
+ harrw: "\u21ad",
+ Hat: "^",
+ hbar: "\u210f",
+ Hcirc: "\u0124",
+ hcirc: "\u0125",
+ hearts: "\u2665",
+ heartsuit: "\u2665",
+ hellip: "\u2026",
+ hercon: "\u22b9",
+ hfr: "\ud835\udd25",
+ Hfr: "\u210c",
+ HilbertSpace: "\u210b",
+ hksearow: "\u2925",
+ hkswarow: "\u2926",
+ hoarr: "\u21ff",
+ homtht: "\u223b",
+ hookleftarrow: "\u21a9",
+ hookrightarrow: "\u21aa",
+ hopf: "\ud835\udd59",
+ Hopf: "\u210d",
+ horbar: "\u2015",
+ HorizontalLine: "\u2500",
+ hscr: "\ud835\udcbd",
+ Hscr: "\u210b",
+ hslash: "\u210f",
+ Hstrok: "\u0126",
+ hstrok: "\u0127",
+ HumpDownHump: "\u224e",
+ HumpEqual: "\u224f",
+ hybull: "\u2043",
+ hyphen: "\u2010",
+ Iacute: "\xcd",
+ iacute: "\xed",
+ ic: "\u2063",
+ Icirc: "\xce",
+ icirc: "\xee",
+ Icy: "\u0418",
+ icy: "\u0438",
+ Idot: "\u0130",
+ IEcy: "\u0415",
+ iecy: "\u0435",
+ iexcl: "\xa1",
+ iff: "\u21d4",
+ ifr: "\ud835\udd26",
+ Ifr: "\u2111",
+ Igrave: "\xcc",
+ igrave: "\xec",
+ ii: "\u2148",
+ iiiint: "\u2a0c",
+ iiint: "\u222d",
+ iinfin: "\u29dc",
+ iiota: "\u2129",
+ IJlig: "\u0132",
+ ijlig: "\u0133",
+ Imacr: "\u012a",
+ imacr: "\u012b",
+ image: "\u2111",
+ ImaginaryI: "\u2148",
+ imagline: "\u2110",
+ imagpart: "\u2111",
+ imath: "\u0131",
+ Im: "\u2111",
+ imof: "\u22b7",
+ imped: "\u01b5",
+ Implies: "\u21d2",
+ incare: "\u2105",
+ in: "\u2208",
+ infin: "\u221e",
+ infintie: "\u29dd",
+ inodot: "\u0131",
+ intcal: "\u22ba",
+ int: "\u222b",
+ Int: "\u222c",
+ integers: "\u2124",
+ Integral: "\u222b",
+ intercal: "\u22ba",
+ Intersection: "\u22c2",
+ intlarhk: "\u2a17",
+ intprod: "\u2a3c",
+ InvisibleComma: "\u2063",
+ InvisibleTimes: "\u2062",
+ IOcy: "\u0401",
+ iocy: "\u0451",
+ Iogon: "\u012e",
+ iogon: "\u012f",
+ Iopf: "\ud835\udd40",
+ iopf: "\ud835\udd5a",
+ Iota: "\u0399",
+ iota: "\u03b9",
+ iprod: "\u2a3c",
+ iquest: "\xbf",
+ iscr: "\ud835\udcbe",
+ Iscr: "\u2110",
+ isin: "\u2208",
+ isindot: "\u22f5",
+ isinE: "\u22f9",
+ isins: "\u22f4",
+ isinsv: "\u22f3",
+ isinv: "\u2208",
+ it: "\u2062",
+ Itilde: "\u0128",
+ itilde: "\u0129",
+ Iukcy: "\u0406",
+ iukcy: "\u0456",
+ Iuml: "\xcf",
+ iuml: "\xef",
+ Jcirc: "\u0134",
+ jcirc: "\u0135",
+ Jcy: "\u0419",
+ jcy: "\u0439",
+ Jfr: "\ud835\udd0d",
+ jfr: "\ud835\udd27",
+ jmath: "\u0237",
+ Jopf: "\ud835\udd41",
+ jopf: "\ud835\udd5b",
+ Jscr: "\ud835\udca5",
+ jscr: "\ud835\udcbf",
+ Jsercy: "\u0408",
+ jsercy: "\u0458",
+ Jukcy: "\u0404",
+ jukcy: "\u0454",
+ Kappa: "\u039a",
+ kappa: "\u03ba",
+ kappav: "\u03f0",
+ Kcedil: "\u0136",
+ kcedil: "\u0137",
+ Kcy: "\u041a",
+ kcy: "\u043a",
+ Kfr: "\ud835\udd0e",
+ kfr: "\ud835\udd28",
+ kgreen: "\u0138",
+ KHcy: "\u0425",
+ khcy: "\u0445",
+ KJcy: "\u040c",
+ kjcy: "\u045c",
+ Kopf: "\ud835\udd42",
+ kopf: "\ud835\udd5c",
+ Kscr: "\ud835\udca6",
+ kscr: "\ud835\udcc0",
+ lAarr: "\u21da",
+ Lacute: "\u0139",
+ lacute: "\u013a",
+ laemptyv: "\u29b4",
+ lagran: "\u2112",
+ Lambda: "\u039b",
+ lambda: "\u03bb",
+ lang: "\u27e8",
+ Lang: "\u27ea",
+ langd: "\u2991",
+ langle: "\u27e8",
+ lap: "\u2a85",
+ Laplacetrf: "\u2112",
+ laquo: "\xab",
+ larrb: "\u21e4",
+ larrbfs: "\u291f",
+ larr: "\u2190",
+ Larr: "\u219e",
+ lArr: "\u21d0",
+ larrfs: "\u291d",
+ larrhk: "\u21a9",
+ larrlp: "\u21ab",
+ larrpl: "\u2939",
+ larrsim: "\u2973",
+ larrtl: "\u21a2",
+ latail: "\u2919",
+ lAtail: "\u291b",
+ lat: "\u2aab",
+ late: "\u2aad",
+ lates: "\u2aad\ufe00",
+ lbarr: "\u290c",
+ lBarr: "\u290e",
+ lbbrk: "\u2772",
+ lbrace: "{",
+ lbrack: "[",
+ lbrke: "\u298b",
+ lbrksld: "\u298f",
+ lbrkslu: "\u298d",
+ Lcaron: "\u013d",
+ lcaron: "\u013e",
+ Lcedil: "\u013b",
+ lcedil: "\u013c",
+ lceil: "\u2308",
+ lcub: "{",
+ Lcy: "\u041b",
+ lcy: "\u043b",
+ ldca: "\u2936",
+ ldquo: "\u201c",
+ ldquor: "\u201e",
+ ldrdhar: "\u2967",
+ ldrushar: "\u294b",
+ ldsh: "\u21b2",
+ le: "\u2264",
+ lE: "\u2266",
+ LeftAngleBracket: "\u27e8",
+ LeftArrowBar: "\u21e4",
+ leftarrow: "\u2190",
+ LeftArrow: "\u2190",
+ Leftarrow: "\u21d0",
+ LeftArrowRightArrow: "\u21c6",
+ leftarrowtail: "\u21a2",
+ LeftCeiling: "\u2308",
+ LeftDoubleBracket: "\u27e6",
+ LeftDownTeeVector: "\u2961",
+ LeftDownVectorBar: "\u2959",
+ LeftDownVector: "\u21c3",
+ LeftFloor: "\u230a",
+ leftharpoondown: "\u21bd",
+ leftharpoonup: "\u21bc",
+ leftleftarrows: "\u21c7",
+ leftrightarrow: "\u2194",
+ LeftRightArrow: "\u2194",
+ Leftrightarrow: "\u21d4",
+ leftrightarrows: "\u21c6",
+ leftrightharpoons: "\u21cb",
+ leftrightsquigarrow: "\u21ad",
+ LeftRightVector: "\u294e",
+ LeftTeeArrow: "\u21a4",
+ LeftTee: "\u22a3",
+ LeftTeeVector: "\u295a",
+ leftthreetimes: "\u22cb",
+ LeftTriangleBar: "\u29cf",
+ LeftTriangle: "\u22b2",
+ LeftTriangleEqual: "\u22b4",
+ LeftUpDownVector: "\u2951",
+ LeftUpTeeVector: "\u2960",
+ LeftUpVectorBar: "\u2958",
+ LeftUpVector: "\u21bf",
+ LeftVectorBar: "\u2952",
+ LeftVector: "\u21bc",
+ lEg: "\u2a8b",
+ leg: "\u22da",
+ leq: "\u2264",
+ leqq: "\u2266",
+ leqslant: "\u2a7d",
+ lescc: "\u2aa8",
+ les: "\u2a7d",
+ lesdot: "\u2a7f",
+ lesdoto: "\u2a81",
+ lesdotor: "\u2a83",
+ lesg: "\u22da\ufe00",
+ lesges: "\u2a93",
+ lessapprox: "\u2a85",
+ lessdot: "\u22d6",
+ lesseqgtr: "\u22da",
+ lesseqqgtr: "\u2a8b",
+ LessEqualGreater: "\u22da",
+ LessFullEqual: "\u2266",
+ LessGreater: "\u2276",
+ lessgtr: "\u2276",
+ LessLess: "\u2aa1",
+ lesssim: "\u2272",
+ LessSlantEqual: "\u2a7d",
+ LessTilde: "\u2272",
+ lfisht: "\u297c",
+ lfloor: "\u230a",
+ Lfr: "\ud835\udd0f",
+ lfr: "\ud835\udd29",
+ lg: "\u2276",
+ lgE: "\u2a91",
+ lHar: "\u2962",
+ lhard: "\u21bd",
+ lharu: "\u21bc",
+ lharul: "\u296a",
+ lhblk: "\u2584",
+ LJcy: "\u0409",
+ ljcy: "\u0459",
+ llarr: "\u21c7",
+ ll: "\u226a",
+ Ll: "\u22d8",
+ llcorner: "\u231e",
+ Lleftarrow: "\u21da",
+ llhard: "\u296b",
+ lltri: "\u25fa",
+ Lmidot: "\u013f",
+ lmidot: "\u0140",
+ lmoustache: "\u23b0",
+ lmoust: "\u23b0",
+ lnap: "\u2a89",
+ lnapprox: "\u2a89",
+ lne: "\u2a87",
+ lnE: "\u2268",
+ lneq: "\u2a87",
+ lneqq: "\u2268",
+ lnsim: "\u22e6",
+ loang: "\u27ec",
+ loarr: "\u21fd",
+ lobrk: "\u27e6",
+ longleftarrow: "\u27f5",
+ LongLeftArrow: "\u27f5",
+ Longleftarrow: "\u27f8",
+ longleftrightarrow: "\u27f7",
+ LongLeftRightArrow: "\u27f7",
+ Longleftrightarrow: "\u27fa",
+ longmapsto: "\u27fc",
+ longrightarrow: "\u27f6",
+ LongRightArrow: "\u27f6",
+ Longrightarrow: "\u27f9",
+ looparrowleft: "\u21ab",
+ looparrowright: "\u21ac",
+ lopar: "\u2985",
+ Lopf: "\ud835\udd43",
+ lopf: "\ud835\udd5d",
+ loplus: "\u2a2d",
+ lotimes: "\u2a34",
+ lowast: "\u2217",
+ lowbar: "_",
+ LowerLeftArrow: "\u2199",
+ LowerRightArrow: "\u2198",
+ loz: "\u25ca",
+ lozenge: "\u25ca",
+ lozf: "\u29eb",
+ lpar: "(",
+ lparlt: "\u2993",
+ lrarr: "\u21c6",
+ lrcorner: "\u231f",
+ lrhar: "\u21cb",
+ lrhard: "\u296d",
+ lrm: "\u200e",
+ lrtri: "\u22bf",
+ lsaquo: "\u2039",
+ lscr: "\ud835\udcc1",
+ Lscr: "\u2112",
+ lsh: "\u21b0",
+ Lsh: "\u21b0",
+ lsim: "\u2272",
+ lsime: "\u2a8d",
+ lsimg: "\u2a8f",
+ lsqb: "[",
+ lsquo: "\u2018",
+ lsquor: "\u201a",
+ Lstrok: "\u0141",
+ lstrok: "\u0142",
+ ltcc: "\u2aa6",
+ ltcir: "\u2a79",
+ lt: "<",
+ LT: "<",
+ Lt: "\u226a",
+ ltdot: "\u22d6",
+ lthree: "\u22cb",
+ ltimes: "\u22c9",
+ ltlarr: "\u2976",
+ ltquest: "\u2a7b",
+ ltri: "\u25c3",
+ ltrie: "\u22b4",
+ ltrif: "\u25c2",
+ ltrPar: "\u2996",
+ lurdshar: "\u294a",
+ luruhar: "\u2966",
+ lvertneqq: "\u2268\ufe00",
+ lvnE: "\u2268\ufe00",
+ macr: "\xaf",
+ male: "\u2642",
+ malt: "\u2720",
+ maltese: "\u2720",
+ Map: "\u2905",
+ map: "\u21a6",
+ mapsto: "\u21a6",
+ mapstodown: "\u21a7",
+ mapstoleft: "\u21a4",
+ mapstoup: "\u21a5",
+ marker: "\u25ae",
+ mcomma: "\u2a29",
+ Mcy: "\u041c",
+ mcy: "\u043c",
+ mdash: "\u2014",
+ mDDot: "\u223a",
+ measuredangle: "\u2221",
+ MediumSpace: "\u205f",
+ Mellintrf: "\u2133",
+ Mfr: "\ud835\udd10",
+ mfr: "\ud835\udd2a",
+ mho: "\u2127",
+ micro: "\xb5",
+ midast: "*",
+ midcir: "\u2af0",
+ mid: "\u2223",
+ middot: "\xb7",
+ minusb: "\u229f",
+ minus: "\u2212",
+ minusd: "\u2238",
+ minusdu: "\u2a2a",
+ MinusPlus: "\u2213",
+ mlcp: "\u2adb",
+ mldr: "\u2026",
+ mnplus: "\u2213",
+ models: "\u22a7",
+ Mopf: "\ud835\udd44",
+ mopf: "\ud835\udd5e",
+ mp: "\u2213",
+ mscr: "\ud835\udcc2",
+ Mscr: "\u2133",
+ mstpos: "\u223e",
+ Mu: "\u039c",
+ mu: "\u03bc",
+ multimap: "\u22b8",
+ mumap: "\u22b8",
+ nabla: "\u2207",
+ Nacute: "\u0143",
+ nacute: "\u0144",
+ nang: "\u2220\u20d2",
+ nap: "\u2249",
+ napE: "\u2a70\u0338",
+ napid: "\u224b\u0338",
+ napos: "\u0149",
+ napprox: "\u2249",
+ natural: "\u266e",
+ naturals: "\u2115",
+ natur: "\u266e",
+ nbsp: "\xa0",
+ nbump: "\u224e\u0338",
+ nbumpe: "\u224f\u0338",
+ ncap: "\u2a43",
+ Ncaron: "\u0147",
+ ncaron: "\u0148",
+ Ncedil: "\u0145",
+ ncedil: "\u0146",
+ ncong: "\u2247",
+ ncongdot: "\u2a6d\u0338",
+ ncup: "\u2a42",
+ Ncy: "\u041d",
+ ncy: "\u043d",
+ ndash: "\u2013",
+ nearhk: "\u2924",
+ nearr: "\u2197",
+ neArr: "\u21d7",
+ nearrow: "\u2197",
+ ne: "\u2260",
+ nedot: "\u2250\u0338",
+ NegativeMediumSpace: "\u200b",
+ NegativeThickSpace: "\u200b",
+ NegativeThinSpace: "\u200b",
+ NegativeVeryThinSpace: "\u200b",
+ nequiv: "\u2262",
+ nesear: "\u2928",
+ nesim: "\u2242\u0338",
+ NestedGreaterGreater: "\u226b",
+ NestedLessLess: "\u226a",
+ NewLine: "\n",
+ nexist: "\u2204",
+ nexists: "\u2204",
+ Nfr: "\ud835\udd11",
+ nfr: "\ud835\udd2b",
+ ngE: "\u2267\u0338",
+ nge: "\u2271",
+ ngeq: "\u2271",
+ ngeqq: "\u2267\u0338",
+ ngeqslant: "\u2a7e\u0338",
+ nges: "\u2a7e\u0338",
+ nGg: "\u22d9\u0338",
+ ngsim: "\u2275",
+ nGt: "\u226b\u20d2",
+ ngt: "\u226f",
+ ngtr: "\u226f",
+ nGtv: "\u226b\u0338",
+ nharr: "\u21ae",
+ nhArr: "\u21ce",
+ nhpar: "\u2af2",
+ ni: "\u220b",
+ nis: "\u22fc",
+ nisd: "\u22fa",
+ niv: "\u220b",
+ NJcy: "\u040a",
+ njcy: "\u045a",
+ nlarr: "\u219a",
+ nlArr: "\u21cd",
+ nldr: "\u2025",
+ nlE: "\u2266\u0338",
+ nle: "\u2270",
+ nleftarrow: "\u219a",
+ nLeftarrow: "\u21cd",
+ nleftrightarrow: "\u21ae",
+ nLeftrightarrow: "\u21ce",
+ nleq: "\u2270",
+ nleqq: "\u2266\u0338",
+ nleqslant: "\u2a7d\u0338",
+ nles: "\u2a7d\u0338",
+ nless: "\u226e",
+ nLl: "\u22d8\u0338",
+ nlsim: "\u2274",
+ nLt: "\u226a\u20d2",
+ nlt: "\u226e",
+ nltri: "\u22ea",
+ nltrie: "\u22ec",
+ nLtv: "\u226a\u0338",
+ nmid: "\u2224",
+ NoBreak: "\u2060",
+ NonBreakingSpace: "\xa0",
+ nopf: "\ud835\udd5f",
+ Nopf: "\u2115",
+ Not: "\u2aec",
+ not: "\xac",
+ NotCongruent: "\u2262",
+ NotCupCap: "\u226d",
+ NotDoubleVerticalBar: "\u2226",
+ NotElement: "\u2209",
+ NotEqual: "\u2260",
+ NotEqualTilde: "\u2242\u0338",
+ NotExists: "\u2204",
+ NotGreater: "\u226f",
+ NotGreaterEqual: "\u2271",
+ NotGreaterFullEqual: "\u2267\u0338",
+ NotGreaterGreater: "\u226b\u0338",
+ NotGreaterLess: "\u2279",
+ NotGreaterSlantEqual: "\u2a7e\u0338",
+ NotGreaterTilde: "\u2275",
+ NotHumpDownHump: "\u224e\u0338",
+ NotHumpEqual: "\u224f\u0338",
+ notin: "\u2209",
+ notindot: "\u22f5\u0338",
+ notinE: "\u22f9\u0338",
+ notinva: "\u2209",
+ notinvb: "\u22f7",
+ notinvc: "\u22f6",
+ NotLeftTriangleBar: "\u29cf\u0338",
+ NotLeftTriangle: "\u22ea",
+ NotLeftTriangleEqual: "\u22ec",
+ NotLess: "\u226e",
+ NotLessEqual: "\u2270",
+ NotLessGreater: "\u2278",
+ NotLessLess: "\u226a\u0338",
+ NotLessSlantEqual: "\u2a7d\u0338",
+ NotLessTilde: "\u2274",
+ NotNestedGreaterGreater: "\u2aa2\u0338",
+ NotNestedLessLess: "\u2aa1\u0338",
+ notni: "\u220c",
+ notniva: "\u220c",
+ notnivb: "\u22fe",
+ notnivc: "\u22fd",
+ NotPrecedes: "\u2280",
+ NotPrecedesEqual: "\u2aaf\u0338",
+ NotPrecedesSlantEqual: "\u22e0",
+ NotReverseElement: "\u220c",
+ NotRightTriangleBar: "\u29d0\u0338",
+ NotRightTriangle: "\u22eb",
+ NotRightTriangleEqual: "\u22ed",
+ NotSquareSubset: "\u228f\u0338",
+ NotSquareSubsetEqual: "\u22e2",
+ NotSquareSuperset: "\u2290\u0338",
+ NotSquareSupersetEqual: "\u22e3",
+ NotSubset: "\u2282\u20d2",
+ NotSubsetEqual: "\u2288",
+ NotSucceeds: "\u2281",
+ NotSucceedsEqual: "\u2ab0\u0338",
+ NotSucceedsSlantEqual: "\u22e1",
+ NotSucceedsTilde: "\u227f\u0338",
+ NotSuperset: "\u2283\u20d2",
+ NotSupersetEqual: "\u2289",
+ NotTilde: "\u2241",
+ NotTildeEqual: "\u2244",
+ NotTildeFullEqual: "\u2247",
+ NotTildeTilde: "\u2249",
+ NotVerticalBar: "\u2224",
+ nparallel: "\u2226",
+ npar: "\u2226",
+ nparsl: "\u2afd\u20e5",
+ npart: "\u2202\u0338",
+ npolint: "\u2a14",
+ npr: "\u2280",
+ nprcue: "\u22e0",
+ nprec: "\u2280",
+ npreceq: "\u2aaf\u0338",
+ npre: "\u2aaf\u0338",
+ nrarrc: "\u2933\u0338",
+ nrarr: "\u219b",
+ nrArr: "\u21cf",
+ nrarrw: "\u219d\u0338",
+ nrightarrow: "\u219b",
+ nRightarrow: "\u21cf",
+ nrtri: "\u22eb",
+ nrtrie: "\u22ed",
+ nsc: "\u2281",
+ nsccue: "\u22e1",
+ nsce: "\u2ab0\u0338",
+ Nscr: "\ud835\udca9",
+ nscr: "\ud835\udcc3",
+ nshortmid: "\u2224",
+ nshortparallel: "\u2226",
+ nsim: "\u2241",
+ nsime: "\u2244",
+ nsimeq: "\u2244",
+ nsmid: "\u2224",
+ nspar: "\u2226",
+ nsqsube: "\u22e2",
+ nsqsupe: "\u22e3",
+ nsub: "\u2284",
+ nsubE: "\u2ac5\u0338",
+ nsube: "\u2288",
+ nsubset: "\u2282\u20d2",
+ nsubseteq: "\u2288",
+ nsubseteqq: "\u2ac5\u0338",
+ nsucc: "\u2281",
+ nsucceq: "\u2ab0\u0338",
+ nsup: "\u2285",
+ nsupE: "\u2ac6\u0338",
+ nsupe: "\u2289",
+ nsupset: "\u2283\u20d2",
+ nsupseteq: "\u2289",
+ nsupseteqq: "\u2ac6\u0338",
+ ntgl: "\u2279",
+ Ntilde: "\xd1",
+ ntilde: "\xf1",
+ ntlg: "\u2278",
+ ntriangleleft: "\u22ea",
+ ntrianglelefteq: "\u22ec",
+ ntriangleright: "\u22eb",
+ ntrianglerighteq: "\u22ed",
+ Nu: "\u039d",
+ nu: "\u03bd",
+ num: "#",
+ numero: "\u2116",
+ numsp: "\u2007",
+ nvap: "\u224d\u20d2",
+ nvdash: "\u22ac",
+ nvDash: "\u22ad",
+ nVdash: "\u22ae",
+ nVDash: "\u22af",
+ nvge: "\u2265\u20d2",
+ nvgt: ">\u20d2",
+ nvHarr: "\u2904",
+ nvinfin: "\u29de",
+ nvlArr: "\u2902",
+ nvle: "\u2264\u20d2",
+ nvlt: "<\u20d2",
+ nvltrie: "\u22b4\u20d2",
+ nvrArr: "\u2903",
+ nvrtrie: "\u22b5\u20d2",
+ nvsim: "\u223c\u20d2",
+ nwarhk: "\u2923",
+ nwarr: "\u2196",
+ nwArr: "\u21d6",
+ nwarrow: "\u2196",
+ nwnear: "\u2927",
+ Oacute: "\xd3",
+ oacute: "\xf3",
+ oast: "\u229b",
+ Ocirc: "\xd4",
+ ocirc: "\xf4",
+ ocir: "\u229a",
+ Ocy: "\u041e",
+ ocy: "\u043e",
+ odash: "\u229d",
+ Odblac: "\u0150",
+ odblac: "\u0151",
+ odiv: "\u2a38",
+ odot: "\u2299",
+ odsold: "\u29bc",
+ OElig: "\u0152",
+ oelig: "\u0153",
+ ofcir: "\u29bf",
+ Ofr: "\ud835\udd12",
+ ofr: "\ud835\udd2c",
+ ogon: "\u02db",
+ Ograve: "\xd2",
+ ograve: "\xf2",
+ ogt: "\u29c1",
+ ohbar: "\u29b5",
+ ohm: "\u03a9",
+ oint: "\u222e",
+ olarr: "\u21ba",
+ olcir: "\u29be",
+ olcross: "\u29bb",
+ oline: "\u203e",
+ olt: "\u29c0",
+ Omacr: "\u014c",
+ omacr: "\u014d",
+ Omega: "\u03a9",
+ omega: "\u03c9",
+ Omicron: "\u039f",
+ omicron: "\u03bf",
+ omid: "\u29b6",
+ ominus: "\u2296",
+ Oopf: "\ud835\udd46",
+ oopf: "\ud835\udd60",
+ opar: "\u29b7",
+ OpenCurlyDoubleQuote: "\u201c",
+ OpenCurlyQuote: "\u2018",
+ operp: "\u29b9",
+ oplus: "\u2295",
+ orarr: "\u21bb",
+ Or: "\u2a54",
+ or: "\u2228",
+ ord: "\u2a5d",
+ order: "\u2134",
+ orderof: "\u2134",
+ ordf: "\xaa",
+ ordm: "\xba",
+ origof: "\u22b6",
+ oror: "\u2a56",
+ orslope: "\u2a57",
+ orv: "\u2a5b",
+ oS: "\u24c8",
+ Oscr: "\ud835\udcaa",
+ oscr: "\u2134",
+ Oslash: "\xd8",
+ oslash: "\xf8",
+ osol: "\u2298",
+ Otilde: "\xd5",
+ otilde: "\xf5",
+ otimesas: "\u2a36",
+ Otimes: "\u2a37",
+ otimes: "\u2297",
+ Ouml: "\xd6",
+ ouml: "\xf6",
+ ovbar: "\u233d",
+ OverBar: "\u203e",
+ OverBrace: "\u23de",
+ OverBracket: "\u23b4",
+ OverParenthesis: "\u23dc",
+ para: "\xb6",
+ parallel: "\u2225",
+ par: "\u2225",
+ parsim: "\u2af3",
+ parsl: "\u2afd",
+ part: "\u2202",
+ PartialD: "\u2202",
+ Pcy: "\u041f",
+ pcy: "\u043f",
+ percnt: "%",
+ period: ".",
+ permil: "\u2030",
+ perp: "\u22a5",
+ pertenk: "\u2031",
+ Pfr: "\ud835\udd13",
+ pfr: "\ud835\udd2d",
+ Phi: "\u03a6",
+ phi: "\u03c6",
+ phiv: "\u03d5",
+ phmmat: "\u2133",
+ phone: "\u260e",
+ Pi: "\u03a0",
+ pi: "\u03c0",
+ pitchfork: "\u22d4",
+ piv: "\u03d6",
+ planck: "\u210f",
+ planckh: "\u210e",
+ plankv: "\u210f",
+ plusacir: "\u2a23",
+ plusb: "\u229e",
+ pluscir: "\u2a22",
+ plus: "+",
+ plusdo: "\u2214",
+ plusdu: "\u2a25",
+ pluse: "\u2a72",
+ PlusMinus: "\xb1",
+ plusmn: "\xb1",
+ plussim: "\u2a26",
+ plustwo: "\u2a27",
+ pm: "\xb1",
+ Poincareplane: "\u210c",
+ pointint: "\u2a15",
+ popf: "\ud835\udd61",
+ Popf: "\u2119",
+ pound: "\xa3",
+ prap: "\u2ab7",
+ Pr: "\u2abb",
+ pr: "\u227a",
+ prcue: "\u227c",
+ precapprox: "\u2ab7",
+ prec: "\u227a",
+ preccurlyeq: "\u227c",
+ Precedes: "\u227a",
+ PrecedesEqual: "\u2aaf",
+ PrecedesSlantEqual: "\u227c",
+ PrecedesTilde: "\u227e",
+ preceq: "\u2aaf",
+ precnapprox: "\u2ab9",
+ precneqq: "\u2ab5",
+ precnsim: "\u22e8",
+ pre: "\u2aaf",
+ prE: "\u2ab3",
+ precsim: "\u227e",
+ prime: "\u2032",
+ Prime: "\u2033",
+ primes: "\u2119",
+ prnap: "\u2ab9",
+ prnE: "\u2ab5",
+ prnsim: "\u22e8",
+ prod: "\u220f",
+ Product: "\u220f",
+ profalar: "\u232e",
+ profline: "\u2312",
+ profsurf: "\u2313",
+ prop: "\u221d",
+ Proportional: "\u221d",
+ Proportion: "\u2237",
+ propto: "\u221d",
+ prsim: "\u227e",
+ prurel: "\u22b0",
+ Pscr: "\ud835\udcab",
+ pscr: "\ud835\udcc5",
+ Psi: "\u03a8",
+ psi: "\u03c8",
+ puncsp: "\u2008",
+ Qfr: "\ud835\udd14",
+ qfr: "\ud835\udd2e",
+ qint: "\u2a0c",
+ qopf: "\ud835\udd62",
+ Qopf: "\u211a",
+ qprime: "\u2057",
+ Qscr: "\ud835\udcac",
+ qscr: "\ud835\udcc6",
+ quaternions: "\u210d",
+ quatint: "\u2a16",
+ quest: "?",
+ questeq: "\u225f",
+ quot: '"',
+ QUOT: '"',
+ rAarr: "\u21db",
+ race: "\u223d\u0331",
+ Racute: "\u0154",
+ racute: "\u0155",
+ radic: "\u221a",
+ raemptyv: "\u29b3",
+ rang: "\u27e9",
+ Rang: "\u27eb",
+ rangd: "\u2992",
+ range: "\u29a5",
+ rangle: "\u27e9",
+ raquo: "\xbb",
+ rarrap: "\u2975",
+ rarrb: "\u21e5",
+ rarrbfs: "\u2920",
+ rarrc: "\u2933",
+ rarr: "\u2192",
+ Rarr: "\u21a0",
+ rArr: "\u21d2",
+ rarrfs: "\u291e",
+ rarrhk: "\u21aa",
+ rarrlp: "\u21ac",
+ rarrpl: "\u2945",
+ rarrsim: "\u2974",
+ Rarrtl: "\u2916",
+ rarrtl: "\u21a3",
+ rarrw: "\u219d",
+ ratail: "\u291a",
+ rAtail: "\u291c",
+ ratio: "\u2236",
+ rationals: "\u211a",
+ rbarr: "\u290d",
+ rBarr: "\u290f",
+ RBarr: "\u2910",
+ rbbrk: "\u2773",
+ rbrace: "}",
+ rbrack: "]",
+ rbrke: "\u298c",
+ rbrksld: "\u298e",
+ rbrkslu: "\u2990",
+ Rcaron: "\u0158",
+ rcaron: "\u0159",
+ Rcedil: "\u0156",
+ rcedil: "\u0157",
+ rceil: "\u2309",
+ rcub: "}",
+ Rcy: "\u0420",
+ rcy: "\u0440",
+ rdca: "\u2937",
+ rdldhar: "\u2969",
+ rdquo: "\u201d",
+ rdquor: "\u201d",
+ rdsh: "\u21b3",
+ real: "\u211c",
+ realine: "\u211b",
+ realpart: "\u211c",
+ reals: "\u211d",
+ Re: "\u211c",
+ rect: "\u25ad",
+ reg: "\xae",
+ REG: "\xae",
+ ReverseElement: "\u220b",
+ ReverseEquilibrium: "\u21cb",
+ ReverseUpEquilibrium: "\u296f",
+ rfisht: "\u297d",
+ rfloor: "\u230b",
+ rfr: "\ud835\udd2f",
+ Rfr: "\u211c",
+ rHar: "\u2964",
+ rhard: "\u21c1",
+ rharu: "\u21c0",
+ rharul: "\u296c",
+ Rho: "\u03a1",
+ rho: "\u03c1",
+ rhov: "\u03f1",
+ RightAngleBracket: "\u27e9",
+ RightArrowBar: "\u21e5",
+ rightarrow: "\u2192",
+ RightArrow: "\u2192",
+ Rightarrow: "\u21d2",
+ RightArrowLeftArrow: "\u21c4",
+ rightarrowtail: "\u21a3",
+ RightCeiling: "\u2309",
+ RightDoubleBracket: "\u27e7",
+ RightDownTeeVector: "\u295d",
+ RightDownVectorBar: "\u2955",
+ RightDownVector: "\u21c2",
+ RightFloor: "\u230b",
+ rightharpoondown: "\u21c1",
+ rightharpoonup: "\u21c0",
+ rightleftarrows: "\u21c4",
+ rightleftharpoons: "\u21cc",
+ rightrightarrows: "\u21c9",
+ rightsquigarrow: "\u219d",
+ RightTeeArrow: "\u21a6",
+ RightTee: "\u22a2",
+ RightTeeVector: "\u295b",
+ rightthreetimes: "\u22cc",
+ RightTriangleBar: "\u29d0",
+ RightTriangle: "\u22b3",
+ RightTriangleEqual: "\u22b5",
+ RightUpDownVector: "\u294f",
+ RightUpTeeVector: "\u295c",
+ RightUpVectorBar: "\u2954",
+ RightUpVector: "\u21be",
+ RightVectorBar: "\u2953",
+ RightVector: "\u21c0",
+ ring: "\u02da",
+ risingdotseq: "\u2253",
+ rlarr: "\u21c4",
+ rlhar: "\u21cc",
+ rlm: "\u200f",
+ rmoustache: "\u23b1",
+ rmoust: "\u23b1",
+ rnmid: "\u2aee",
+ roang: "\u27ed",
+ roarr: "\u21fe",
+ robrk: "\u27e7",
+ ropar: "\u2986",
+ ropf: "\ud835\udd63",
+ Ropf: "\u211d",
+ roplus: "\u2a2e",
+ rotimes: "\u2a35",
+ RoundImplies: "\u2970",
+ rpar: ")",
+ rpargt: "\u2994",
+ rppolint: "\u2a12",
+ rrarr: "\u21c9",
+ Rrightarrow: "\u21db",
+ rsaquo: "\u203a",
+ rscr: "\ud835\udcc7",
+ Rscr: "\u211b",
+ rsh: "\u21b1",
+ Rsh: "\u21b1",
+ rsqb: "]",
+ rsquo: "\u2019",
+ rsquor: "\u2019",
+ rthree: "\u22cc",
+ rtimes: "\u22ca",
+ rtri: "\u25b9",
+ rtrie: "\u22b5",
+ rtrif: "\u25b8",
+ rtriltri: "\u29ce",
+ RuleDelayed: "\u29f4",
+ ruluhar: "\u2968",
+ rx: "\u211e",
+ Sacute: "\u015a",
+ sacute: "\u015b",
+ sbquo: "\u201a",
+ scap: "\u2ab8",
+ Scaron: "\u0160",
+ scaron: "\u0161",
+ Sc: "\u2abc",
+ sc: "\u227b",
+ sccue: "\u227d",
+ sce: "\u2ab0",
+ scE: "\u2ab4",
+ Scedil: "\u015e",
+ scedil: "\u015f",
+ Scirc: "\u015c",
+ scirc: "\u015d",
+ scnap: "\u2aba",
+ scnE: "\u2ab6",
+ scnsim: "\u22e9",
+ scpolint: "\u2a13",
+ scsim: "\u227f",
+ Scy: "\u0421",
+ scy: "\u0441",
+ sdotb: "\u22a1",
+ sdot: "\u22c5",
+ sdote: "\u2a66",
+ searhk: "\u2925",
+ searr: "\u2198",
+ seArr: "\u21d8",
+ searrow: "\u2198",
+ sect: "\xa7",
+ semi: ";",
+ seswar: "\u2929",
+ setminus: "\u2216",
+ setmn: "\u2216",
+ sext: "\u2736",
+ Sfr: "\ud835\udd16",
+ sfr: "\ud835\udd30",
+ sfrown: "\u2322",
+ sharp: "\u266f",
+ SHCHcy: "\u0429",
+ shchcy: "\u0449",
+ SHcy: "\u0428",
+ shcy: "\u0448",
+ ShortDownArrow: "\u2193",
+ ShortLeftArrow: "\u2190",
+ shortmid: "\u2223",
+ shortparallel: "\u2225",
+ ShortRightArrow: "\u2192",
+ ShortUpArrow: "\u2191",
+ shy: "\xad",
+ Sigma: "\u03a3",
+ sigma: "\u03c3",
+ sigmaf: "\u03c2",
+ sigmav: "\u03c2",
+ sim: "\u223c",
+ simdot: "\u2a6a",
+ sime: "\u2243",
+ simeq: "\u2243",
+ simg: "\u2a9e",
+ simgE: "\u2aa0",
+ siml: "\u2a9d",
+ simlE: "\u2a9f",
+ simne: "\u2246",
+ simplus: "\u2a24",
+ simrarr: "\u2972",
+ slarr: "\u2190",
+ SmallCircle: "\u2218",
+ smallsetminus: "\u2216",
+ smashp: "\u2a33",
+ smeparsl: "\u29e4",
+ smid: "\u2223",
+ smile: "\u2323",
+ smt: "\u2aaa",
+ smte: "\u2aac",
+ smtes: "\u2aac\ufe00",
+ SOFTcy: "\u042c",
+ softcy: "\u044c",
+ solbar: "\u233f",
+ solb: "\u29c4",
+ sol: "/",
+ Sopf: "\ud835\udd4a",
+ sopf: "\ud835\udd64",
+ spades: "\u2660",
+ spadesuit: "\u2660",
+ spar: "\u2225",
+ sqcap: "\u2293",
+ sqcaps: "\u2293\ufe00",
+ sqcup: "\u2294",
+ sqcups: "\u2294\ufe00",
+ Sqrt: "\u221a",
+ sqsub: "\u228f",
+ sqsube: "\u2291",
+ sqsubset: "\u228f",
+ sqsubseteq: "\u2291",
+ sqsup: "\u2290",
+ sqsupe: "\u2292",
+ sqsupset: "\u2290",
+ sqsupseteq: "\u2292",
+ square: "\u25a1",
+ Square: "\u25a1",
+ SquareIntersection: "\u2293",
+ SquareSubset: "\u228f",
+ SquareSubsetEqual: "\u2291",
+ SquareSuperset: "\u2290",
+ SquareSupersetEqual: "\u2292",
+ SquareUnion: "\u2294",
+ squarf: "\u25aa",
+ squ: "\u25a1",
+ squf: "\u25aa",
+ srarr: "\u2192",
+ Sscr: "\ud835\udcae",
+ sscr: "\ud835\udcc8",
+ ssetmn: "\u2216",
+ ssmile: "\u2323",
+ sstarf: "\u22c6",
+ Star: "\u22c6",
+ star: "\u2606",
+ starf: "\u2605",
+ straightepsilon: "\u03f5",
+ straightphi: "\u03d5",
+ strns: "\xaf",
+ sub: "\u2282",
+ Sub: "\u22d0",
+ subdot: "\u2abd",
+ subE: "\u2ac5",
+ sube: "\u2286",
+ subedot: "\u2ac3",
+ submult: "\u2ac1",
+ subnE: "\u2acb",
+ subne: "\u228a",
+ subplus: "\u2abf",
+ subrarr: "\u2979",
+ subset: "\u2282",
+ Subset: "\u22d0",
+ subseteq: "\u2286",
+ subseteqq: "\u2ac5",
+ SubsetEqual: "\u2286",
+ subsetneq: "\u228a",
+ subsetneqq: "\u2acb",
+ subsim: "\u2ac7",
+ subsub: "\u2ad5",
+ subsup: "\u2ad3",
+ succapprox: "\u2ab8",
+ succ: "\u227b",
+ succcurlyeq: "\u227d",
+ Succeeds: "\u227b",
+ SucceedsEqual: "\u2ab0",
+ SucceedsSlantEqual: "\u227d",
+ SucceedsTilde: "\u227f",
+ succeq: "\u2ab0",
+ succnapprox: "\u2aba",
+ succneqq: "\u2ab6",
+ succnsim: "\u22e9",
+ succsim: "\u227f",
+ SuchThat: "\u220b",
+ sum: "\u2211",
+ Sum: "\u2211",
+ sung: "\u266a",
+ sup1: "\xb9",
+ sup2: "\xb2",
+ sup3: "\xb3",
+ sup: "\u2283",
+ Sup: "\u22d1",
+ supdot: "\u2abe",
+ supdsub: "\u2ad8",
+ supE: "\u2ac6",
+ supe: "\u2287",
+ supedot: "\u2ac4",
+ Superset: "\u2283",
+ SupersetEqual: "\u2287",
+ suphsol: "\u27c9",
+ suphsub: "\u2ad7",
+ suplarr: "\u297b",
+ supmult: "\u2ac2",
+ supnE: "\u2acc",
+ supne: "\u228b",
+ supplus: "\u2ac0",
+ supset: "\u2283",
+ Supset: "\u22d1",
+ supseteq: "\u2287",
+ supseteqq: "\u2ac6",
+ supsetneq: "\u228b",
+ supsetneqq: "\u2acc",
+ supsim: "\u2ac8",
+ supsub: "\u2ad4",
+ supsup: "\u2ad6",
+ swarhk: "\u2926",
+ swarr: "\u2199",
+ swArr: "\u21d9",
+ swarrow: "\u2199",
+ swnwar: "\u292a",
+ szlig: "\xdf",
+ Tab: "\t",
+ target: "\u2316",
+ Tau: "\u03a4",
+ tau: "\u03c4",
+ tbrk: "\u23b4",
+ Tcaron: "\u0164",
+ tcaron: "\u0165",
+ Tcedil: "\u0162",
+ tcedil: "\u0163",
+ Tcy: "\u0422",
+ tcy: "\u0442",
+ tdot: "\u20db",
+ telrec: "\u2315",
+ Tfr: "\ud835\udd17",
+ tfr: "\ud835\udd31",
+ there4: "\u2234",
+ therefore: "\u2234",
+ Therefore: "\u2234",
+ Theta: "\u0398",
+ theta: "\u03b8",
+ thetasym: "\u03d1",
+ thetav: "\u03d1",
+ thickapprox: "\u2248",
+ thicksim: "\u223c",
+ ThickSpace: "\u205f\u200a",
+ ThinSpace: "\u2009",
+ thinsp: "\u2009",
+ thkap: "\u2248",
+ thksim: "\u223c",
+ THORN: "\xde",
+ thorn: "\xfe",
+ tilde: "\u02dc",
+ Tilde: "\u223c",
+ TildeEqual: "\u2243",
+ TildeFullEqual: "\u2245",
+ TildeTilde: "\u2248",
+ timesbar: "\u2a31",
+ timesb: "\u22a0",
+ times: "\xd7",
+ timesd: "\u2a30",
+ tint: "\u222d",
+ toea: "\u2928",
+ topbot: "\u2336",
+ topcir: "\u2af1",
+ top: "\u22a4",
+ Topf: "\ud835\udd4b",
+ topf: "\ud835\udd65",
+ topfork: "\u2ada",
+ tosa: "\u2929",
+ tprime: "\u2034",
+ trade: "\u2122",
+ TRADE: "\u2122",
+ triangle: "\u25b5",
+ triangledown: "\u25bf",
+ triangleleft: "\u25c3",
+ trianglelefteq: "\u22b4",
+ triangleq: "\u225c",
+ triangleright: "\u25b9",
+ trianglerighteq: "\u22b5",
+ tridot: "\u25ec",
+ trie: "\u225c",
+ triminus: "\u2a3a",
+ TripleDot: "\u20db",
+ triplus: "\u2a39",
+ trisb: "\u29cd",
+ tritime: "\u2a3b",
+ trpezium: "\u23e2",
+ Tscr: "\ud835\udcaf",
+ tscr: "\ud835\udcc9",
+ TScy: "\u0426",
+ tscy: "\u0446",
+ TSHcy: "\u040b",
+ tshcy: "\u045b",
+ Tstrok: "\u0166",
+ tstrok: "\u0167",
+ twixt: "\u226c",
+ twoheadleftarrow: "\u219e",
+ twoheadrightarrow: "\u21a0",
+ Uacute: "\xda",
+ uacute: "\xfa",
+ uarr: "\u2191",
+ Uarr: "\u219f",
+ uArr: "\u21d1",
+ Uarrocir: "\u2949",
+ Ubrcy: "\u040e",
+ ubrcy: "\u045e",
+ Ubreve: "\u016c",
+ ubreve: "\u016d",
+ Ucirc: "\xdb",
+ ucirc: "\xfb",
+ Ucy: "\u0423",
+ ucy: "\u0443",
+ udarr: "\u21c5",
+ Udblac: "\u0170",
+ udblac: "\u0171",
+ udhar: "\u296e",
+ ufisht: "\u297e",
+ Ufr: "\ud835\udd18",
+ ufr: "\ud835\udd32",
+ Ugrave: "\xd9",
+ ugrave: "\xf9",
+ uHar: "\u2963",
+ uharl: "\u21bf",
+ uharr: "\u21be",
+ uhblk: "\u2580",
+ ulcorn: "\u231c",
+ ulcorner: "\u231c",
+ ulcrop: "\u230f",
+ ultri: "\u25f8",
+ Umacr: "\u016a",
+ umacr: "\u016b",
+ uml: "\xa8",
+ UnderBar: "_",
+ UnderBrace: "\u23df",
+ UnderBracket: "\u23b5",
+ UnderParenthesis: "\u23dd",
+ Union: "\u22c3",
+ UnionPlus: "\u228e",
+ Uogon: "\u0172",
+ uogon: "\u0173",
+ Uopf: "\ud835\udd4c",
+ uopf: "\ud835\udd66",
+ UpArrowBar: "\u2912",
+ uparrow: "\u2191",
+ UpArrow: "\u2191",
+ Uparrow: "\u21d1",
+ UpArrowDownArrow: "\u21c5",
+ updownarrow: "\u2195",
+ UpDownArrow: "\u2195",
+ Updownarrow: "\u21d5",
+ UpEquilibrium: "\u296e",
+ upharpoonleft: "\u21bf",
+ upharpoonright: "\u21be",
+ uplus: "\u228e",
+ UpperLeftArrow: "\u2196",
+ UpperRightArrow: "\u2197",
+ upsi: "\u03c5",
+ Upsi: "\u03d2",
+ upsih: "\u03d2",
+ Upsilon: "\u03a5",
+ upsilon: "\u03c5",
+ UpTeeArrow: "\u21a5",
+ UpTee: "\u22a5",
+ upuparrows: "\u21c8",
+ urcorn: "\u231d",
+ urcorner: "\u231d",
+ urcrop: "\u230e",
+ Uring: "\u016e",
+ uring: "\u016f",
+ urtri: "\u25f9",
+ Uscr: "\ud835\udcb0",
+ uscr: "\ud835\udcca",
+ utdot: "\u22f0",
+ Utilde: "\u0168",
+ utilde: "\u0169",
+ utri: "\u25b5",
+ utrif: "\u25b4",
+ uuarr: "\u21c8",
+ Uuml: "\xdc",
+ uuml: "\xfc",
+ uwangle: "\u29a7",
+ vangrt: "\u299c",
+ varepsilon: "\u03f5",
+ varkappa: "\u03f0",
+ varnothing: "\u2205",
+ varphi: "\u03d5",
+ varpi: "\u03d6",
+ varpropto: "\u221d",
+ varr: "\u2195",
+ vArr: "\u21d5",
+ varrho: "\u03f1",
+ varsigma: "\u03c2",
+ varsubsetneq: "\u228a\ufe00",
+ varsubsetneqq: "\u2acb\ufe00",
+ varsupsetneq: "\u228b\ufe00",
+ varsupsetneqq: "\u2acc\ufe00",
+ vartheta: "\u03d1",
+ vartriangleleft: "\u22b2",
+ vartriangleright: "\u22b3",
+ vBar: "\u2ae8",
+ Vbar: "\u2aeb",
+ vBarv: "\u2ae9",
+ Vcy: "\u0412",
+ vcy: "\u0432",
+ vdash: "\u22a2",
+ vDash: "\u22a8",
+ Vdash: "\u22a9",
+ VDash: "\u22ab",
+ Vdashl: "\u2ae6",
+ veebar: "\u22bb",
+ vee: "\u2228",
+ Vee: "\u22c1",
+ veeeq: "\u225a",
+ vellip: "\u22ee",
+ verbar: "|",
+ Verbar: "\u2016",
+ vert: "|",
+ Vert: "\u2016",
+ VerticalBar: "\u2223",
+ VerticalLine: "|",
+ VerticalSeparator: "\u2758",
+ VerticalTilde: "\u2240",
+ VeryThinSpace: "\u200a",
+ Vfr: "\ud835\udd19",
+ vfr: "\ud835\udd33",
+ vltri: "\u22b2",
+ vnsub: "\u2282\u20d2",
+ vnsup: "\u2283\u20d2",
+ Vopf: "\ud835\udd4d",
+ vopf: "\ud835\udd67",
+ vprop: "\u221d",
+ vrtri: "\u22b3",
+ Vscr: "\ud835\udcb1",
+ vscr: "\ud835\udccb",
+ vsubnE: "\u2acb\ufe00",
+ vsubne: "\u228a\ufe00",
+ vsupnE: "\u2acc\ufe00",
+ vsupne: "\u228b\ufe00",
+ Vvdash: "\u22aa",
+ vzigzag: "\u299a",
+ Wcirc: "\u0174",
+ wcirc: "\u0175",
+ wedbar: "\u2a5f",
+ wedge: "\u2227",
+ Wedge: "\u22c0",
+ wedgeq: "\u2259",
+ weierp: "\u2118",
+ Wfr: "\ud835\udd1a",
+ wfr: "\ud835\udd34",
+ Wopf: "\ud835\udd4e",
+ wopf: "\ud835\udd68",
+ wp: "\u2118",
+ wr: "\u2240",
+ wreath: "\u2240",
+ Wscr: "\ud835\udcb2",
+ wscr: "\ud835\udccc",
+ xcap: "\u22c2",
+ xcirc: "\u25ef",
+ xcup: "\u22c3",
+ xdtri: "\u25bd",
+ Xfr: "\ud835\udd1b",
+ xfr: "\ud835\udd35",
+ xharr: "\u27f7",
+ xhArr: "\u27fa",
+ Xi: "\u039e",
+ xi: "\u03be",
+ xlarr: "\u27f5",
+ xlArr: "\u27f8",
+ xmap: "\u27fc",
+ xnis: "\u22fb",
+ xodot: "\u2a00",
+ Xopf: "\ud835\udd4f",
+ xopf: "\ud835\udd69",
+ xoplus: "\u2a01",
+ xotime: "\u2a02",
+ xrarr: "\u27f6",
+ xrArr: "\u27f9",
+ Xscr: "\ud835\udcb3",
+ xscr: "\ud835\udccd",
+ xsqcup: "\u2a06",
+ xuplus: "\u2a04",
+ xutri: "\u25b3",
+ xvee: "\u22c1",
+ xwedge: "\u22c0",
+ Yacute: "\xdd",
+ yacute: "\xfd",
+ YAcy: "\u042f",
+ yacy: "\u044f",
+ Ycirc: "\u0176",
+ ycirc: "\u0177",
+ Ycy: "\u042b",
+ ycy: "\u044b",
+ yen: "\xa5",
+ Yfr: "\ud835\udd1c",
+ yfr: "\ud835\udd36",
+ YIcy: "\u0407",
+ yicy: "\u0457",
+ Yopf: "\ud835\udd50",
+ yopf: "\ud835\udd6a",
+ Yscr: "\ud835\udcb4",
+ yscr: "\ud835\udcce",
+ YUcy: "\u042e",
+ yucy: "\u044e",
+ yuml: "\xff",
+ Yuml: "\u0178",
+ Zacute: "\u0179",
+ zacute: "\u017a",
+ Zcaron: "\u017d",
+ zcaron: "\u017e",
+ Zcy: "\u0417",
+ zcy: "\u0437",
+ Zdot: "\u017b",
+ zdot: "\u017c",
+ zeetrf: "\u2128",
+ ZeroWidthSpace: "\u200b",
+ Zeta: "\u0396",
+ zeta: "\u03b6",
+ zfr: "\ud835\udd37",
+ Zfr: "\u2128",
+ ZHcy: "\u0416",
+ zhcy: "\u0436",
+ zigrarr: "\u21dd",
+ zopf: "\ud835\udd6b",
+ Zopf: "\u2124",
+ Zscr: "\ud835\udcb5",
+ zscr: "\ud835\udccf",
+ zwj: "\u200d",
+ zwnj: "\u200c"
+ };
+ /*eslint quotes:0*/ var entities = require$$0;
+ var regex$4 = /[!-#%-\*,-\/:;\?@\[-\]_\{\}\xA1\xA7\xAB\xB6\xB7\xBB\xBF\u037E\u0387\u055A-\u055F\u0589\u058A\u05BE\u05C0\u05C3\u05C6\u05F3\u05F4\u0609\u060A\u060C\u060D\u061B\u061E\u061F\u066A-\u066D\u06D4\u0700-\u070D\u07F7-\u07F9\u0830-\u083E\u085E\u0964\u0965\u0970\u09FD\u0A76\u0AF0\u0C84\u0DF4\u0E4F\u0E5A\u0E5B\u0F04-\u0F12\u0F14\u0F3A-\u0F3D\u0F85\u0FD0-\u0FD4\u0FD9\u0FDA\u104A-\u104F\u10FB\u1360-\u1368\u1400\u166D\u166E\u169B\u169C\u16EB-\u16ED\u1735\u1736\u17D4-\u17D6\u17D8-\u17DA\u1800-\u180A\u1944\u1945\u1A1E\u1A1F\u1AA0-\u1AA6\u1AA8-\u1AAD\u1B5A-\u1B60\u1BFC-\u1BFF\u1C3B-\u1C3F\u1C7E\u1C7F\u1CC0-\u1CC7\u1CD3\u2010-\u2027\u2030-\u2043\u2045-\u2051\u2053-\u205E\u207D\u207E\u208D\u208E\u2308-\u230B\u2329\u232A\u2768-\u2775\u27C5\u27C6\u27E6-\u27EF\u2983-\u2998\u29D8-\u29DB\u29FC\u29FD\u2CF9-\u2CFC\u2CFE\u2CFF\u2D70\u2E00-\u2E2E\u2E30-\u2E4E\u3001-\u3003\u3008-\u3011\u3014-\u301F\u3030\u303D\u30A0\u30FB\uA4FE\uA4FF\uA60D-\uA60F\uA673\uA67E\uA6F2-\uA6F7\uA874-\uA877\uA8CE\uA8CF\uA8F8-\uA8FA\uA8FC\uA92E\uA92F\uA95F\uA9C1-\uA9CD\uA9DE\uA9DF\uAA5C-\uAA5F\uAADE\uAADF\uAAF0\uAAF1\uABEB\uFD3E\uFD3F\uFE10-\uFE19\uFE30-\uFE52\uFE54-\uFE61\uFE63\uFE68\uFE6A\uFE6B\uFF01-\uFF03\uFF05-\uFF0A\uFF0C-\uFF0F\uFF1A\uFF1B\uFF1F\uFF20\uFF3B-\uFF3D\uFF3F\uFF5B\uFF5D\uFF5F-\uFF65]|\uD800[\uDD00-\uDD02\uDF9F\uDFD0]|\uD801\uDD6F|\uD802[\uDC57\uDD1F\uDD3F\uDE50-\uDE58\uDE7F\uDEF0-\uDEF6\uDF39-\uDF3F\uDF99-\uDF9C]|\uD803[\uDF55-\uDF59]|\uD804[\uDC47-\uDC4D\uDCBB\uDCBC\uDCBE-\uDCC1\uDD40-\uDD43\uDD74\uDD75\uDDC5-\uDDC8\uDDCD\uDDDB\uDDDD-\uDDDF\uDE38-\uDE3D\uDEA9]|\uD805[\uDC4B-\uDC4F\uDC5B\uDC5D\uDCC6\uDDC1-\uDDD7\uDE41-\uDE43\uDE60-\uDE6C\uDF3C-\uDF3E]|\uD806[\uDC3B\uDE3F-\uDE46\uDE9A-\uDE9C\uDE9E-\uDEA2]|\uD807[\uDC41-\uDC45\uDC70\uDC71\uDEF7\uDEF8]|\uD809[\uDC70-\uDC74]|\uD81A[\uDE6E\uDE6F\uDEF5\uDF37-\uDF3B\uDF44]|\uD81B[\uDE97-\uDE9A]|\uD82F\uDC9F|\uD836[\uDE87-\uDE8B]|\uD83A[\uDD5E\uDD5F]/;
+ var encodeCache = {};
+ // Create a lookup array where anything but characters in `chars` string
+ // and alphanumeric chars is percent-encoded.
+
+ function getEncodeCache(exclude) {
+ var i, ch, cache = encodeCache[exclude];
+ if (cache) {
+ return cache;
+ }
+ cache = encodeCache[exclude] = [];
+ for (i = 0; i < 128; i++) {
+ ch = String.fromCharCode(i);
+ if (/^[0-9a-z]$/i.test(ch)) {
+ // always allow unencoded alphanumeric characters
+ cache.push(ch);
+ } else {
+ cache.push("%" + ("0" + i.toString(16).toUpperCase()).slice(-2));
+ }
+ }
+ for (i = 0; i < exclude.length; i++) {
+ cache[exclude.charCodeAt(i)] = exclude[i];
+ }
+ return cache;
+ }
+ // Encode unsafe characters with percent-encoding, skipping already
+ // encoded sequences.
+
+ // - string - string to encode
+ // - exclude - list of characters to ignore (in addition to a-zA-Z0-9)
+ // - keepEscaped - don't encode '%' in a correct escape sequence (default: true)
+
+ function encode$2(string, exclude, keepEscaped) {
+ var i, l, code, nextCode, cache, result = "";
+ if (typeof exclude !== "string") {
+ // encode(string, keepEscaped)
+ keepEscaped = exclude;
+ exclude = encode$2.defaultChars;
+ }
+ if (typeof keepEscaped === "undefined") {
+ keepEscaped = true;
+ }
+ cache = getEncodeCache(exclude);
+ for (i = 0, l = string.length; i < l; i++) {
+ code = string.charCodeAt(i);
+ if (keepEscaped && code === 37 /* % */ && i + 2 < l) {
+ if (/^[0-9a-f]{2}$/i.test(string.slice(i + 1, i + 3))) {
+ result += string.slice(i, i + 3);
+ i += 2;
+ continue;
+ }
+ }
+ if (code < 128) {
+ result += cache[code];
+ continue;
+ }
+ if (code >= 55296 && code <= 57343) {
+ if (code >= 55296 && code <= 56319 && i + 1 < l) {
+ nextCode = string.charCodeAt(i + 1);
+ if (nextCode >= 56320 && nextCode <= 57343) {
+ result += encodeURIComponent(string[i] + string[i + 1]);
+ i++;
+ continue;
+ }
+ }
+ result += "%EF%BF%BD";
+ continue;
+ }
+ result += encodeURIComponent(string[i]);
+ }
+ return result;
+ }
+ encode$2.defaultChars = ";/?:@&=+$,-_.!~*'()#";
+ encode$2.componentChars = "-_.!~*'()";
+ var encode_1 = encode$2;
+ /* eslint-disable no-bitwise */ var decodeCache = {};
+ function getDecodeCache(exclude) {
+ var i, ch, cache = decodeCache[exclude];
+ if (cache) {
+ return cache;
+ }
+ cache = decodeCache[exclude] = [];
+ for (i = 0; i < 128; i++) {
+ ch = String.fromCharCode(i);
+ cache.push(ch);
+ }
+ for (i = 0; i < exclude.length; i++) {
+ ch = exclude.charCodeAt(i);
+ cache[ch] = "%" + ("0" + ch.toString(16).toUpperCase()).slice(-2);
+ }
+ return cache;
+ }
+ // Decode percent-encoded string.
+
+ function decode$2(string, exclude) {
+ var cache;
+ if (typeof exclude !== "string") {
+ exclude = decode$2.defaultChars;
+ }
+ cache = getDecodeCache(exclude);
+ return string.replace(/(%[a-f0-9]{2})+/gi, (function(seq) {
+ var i, l, b1, b2, b3, b4, chr, result = "";
+ for (i = 0, l = seq.length; i < l; i += 3) {
+ b1 = parseInt(seq.slice(i + 1, i + 3), 16);
+ if (b1 < 128) {
+ result += cache[b1];
+ continue;
+ }
+ if ((b1 & 224) === 192 && i + 3 < l) {
+ // 110xxxxx 10xxxxxx
+ b2 = parseInt(seq.slice(i + 4, i + 6), 16);
+ if ((b2 & 192) === 128) {
+ chr = b1 << 6 & 1984 | b2 & 63;
+ if (chr < 128) {
+ result += "\ufffd\ufffd";
+ } else {
+ result += String.fromCharCode(chr);
+ }
+ i += 3;
+ continue;
+ }
+ }
+ if ((b1 & 240) === 224 && i + 6 < l) {
+ // 1110xxxx 10xxxxxx 10xxxxxx
+ b2 = parseInt(seq.slice(i + 4, i + 6), 16);
+ b3 = parseInt(seq.slice(i + 7, i + 9), 16);
+ if ((b2 & 192) === 128 && (b3 & 192) === 128) {
+ chr = b1 << 12 & 61440 | b2 << 6 & 4032 | b3 & 63;
+ if (chr < 2048 || chr >= 55296 && chr <= 57343) {
+ result += "\ufffd\ufffd\ufffd";
+ } else {
+ result += String.fromCharCode(chr);
+ }
+ i += 6;
+ continue;
+ }
+ }
+ if ((b1 & 248) === 240 && i + 9 < l) {
+ // 111110xx 10xxxxxx 10xxxxxx 10xxxxxx
+ b2 = parseInt(seq.slice(i + 4, i + 6), 16);
+ b3 = parseInt(seq.slice(i + 7, i + 9), 16);
+ b4 = parseInt(seq.slice(i + 10, i + 12), 16);
+ if ((b2 & 192) === 128 && (b3 & 192) === 128 && (b4 & 192) === 128) {
+ chr = b1 << 18 & 1835008 | b2 << 12 & 258048 | b3 << 6 & 4032 | b4 & 63;
+ if (chr < 65536 || chr > 1114111) {
+ result += "\ufffd\ufffd\ufffd\ufffd";
+ } else {
+ chr -= 65536;
+ result += String.fromCharCode(55296 + (chr >> 10), 56320 + (chr & 1023));
+ }
+ i += 9;
+ continue;
+ }
+ }
+ result += "\ufffd";
+ }
+ return result;
+ }));
+ }
+ decode$2.defaultChars = ";/?:@&=+$,#";
+ decode$2.componentChars = "";
+ var decode_1 = decode$2;
+ var format$1 = function format(url) {
+ var result = "";
+ result += url.protocol || "";
+ result += url.slashes ? "//" : "";
+ result += url.auth ? url.auth + "@" : "";
+ if (url.hostname && url.hostname.indexOf(":") !== -1) {
+ // ipv6 address
+ result += "[" + url.hostname + "]";
+ } else {
+ result += url.hostname || "";
+ }
+ result += url.port ? ":" + url.port : "";
+ result += url.pathname || "";
+ result += url.search || "";
+ result += url.hash || "";
+ return result;
+ };
+ // Copyright Joyent, Inc. and other Node contributors.
+
+ // Changes from joyent/node:
+
+ // 1. No leading slash in paths,
+ // e.g. in `url.parse('http://foo?bar')` pathname is ``, not `/`
+
+ // 2. Backslashes are not replaced with slashes,
+ // so `http:\\example.org\` is treated like a relative path
+
+ // 3. Trailing colon is treated like a part of the path,
+ // i.e. in `http://example.org:foo` pathname is `:foo`
+
+ // 4. Nothing is URL-encoded in the resulting object,
+ // (in joyent/node some chars in auth and paths are encoded)
+
+ // 5. `url.parse()` does not have `parseQueryString` argument
+
+ // 6. Removed extraneous result properties: `host`, `path`, `query`, etc.,
+ // which can be constructed using other parts of the url.
+
+ function Url() {
+ this.protocol = null;
+ this.slashes = null;
+ this.auth = null;
+ this.port = null;
+ this.hostname = null;
+ this.hash = null;
+ this.search = null;
+ this.pathname = null;
+ }
+ // Reference: RFC 3986, RFC 1808, RFC 2396
+ // define these here so at least they only have to be
+ // compiled once on the first module load.
+ var protocolPattern = /^([a-z0-9.+-]+:)/i, portPattern = /:[0-9]*$/,
+ // Special case for a simple path URL
+ simplePathPattern = /^(\/\/?(?!\/)[^\?\s]*)(\?[^\s]*)?$/,
+ // RFC 2396: characters reserved for delimiting URLs.
+ // We actually just auto-escape these.
+ delims = [ "<", ">", '"', "`", " ", "\r", "\n", "\t" ],
+ // RFC 2396: characters not allowed for various reasons.
+ unwise = [ "{", "}", "|", "\\", "^", "`" ].concat(delims),
+ // Allowed by RFCs, but cause of XSS attacks. Always escape these.
+ autoEscape = [ "'" ].concat(unwise),
+ // Characters that are never ever allowed in a hostname.
+ // Note that any invalid chars are also handled, but these
+ // are the ones that are *expected* to be seen, so we fast-path
+ // them.
+ nonHostChars = [ "%", "/", "?", ";", "#" ].concat(autoEscape), hostEndingChars = [ "/", "?", "#" ], hostnameMaxLen = 255, hostnamePartPattern = /^[+a-z0-9A-Z_-]{0,63}$/, hostnamePartStart = /^([+a-z0-9A-Z_-]{0,63})(.*)$/,
+ // protocols that can allow "unsafe" and "unwise" chars.
+ /* eslint-disable no-script-url */
+ // protocols that never have a hostname.
+ hostlessProtocol = {
+ javascript: true,
+ "javascript:": true
+ },
+ // protocols that always contain a // bit.
+ slashedProtocol = {
+ http: true,
+ https: true,
+ ftp: true,
+ gopher: true,
+ file: true,
+ "http:": true,
+ "https:": true,
+ "ftp:": true,
+ "gopher:": true,
+ "file:": true
+ };
+ /* eslint-enable no-script-url */ function urlParse(url, slashesDenoteHost) {
+ if (url && url instanceof Url) {
+ return url;
+ }
+ var u = new Url;
+ u.parse(url, slashesDenoteHost);
+ return u;
+ }
+ Url.prototype.parse = function(url, slashesDenoteHost) {
+ var i, l, lowerProto, hec, slashes, rest = url;
+ // trim before proceeding.
+ // This is to support parse stuff like " http://foo.com \n"
+ rest = rest.trim();
+ if (!slashesDenoteHost && url.split("#").length === 1) {
+ // Try fast path regexp
+ var simplePath = simplePathPattern.exec(rest);
+ if (simplePath) {
+ this.pathname = simplePath[1];
+ if (simplePath[2]) {
+ this.search = simplePath[2];
+ }
+ return this;
+ }
+ }
+ var proto = protocolPattern.exec(rest);
+ if (proto) {
+ proto = proto[0];
+ lowerProto = proto.toLowerCase();
+ this.protocol = proto;
+ rest = rest.substr(proto.length);
+ }
+ // figure out if it's got a host
+ // user@server is *always* interpreted as a hostname, and url
+ // resolution will treat //foo/bar as host=foo,path=bar because that's
+ // how the browser resolves relative URLs.
+ if (slashesDenoteHost || proto || rest.match(/^\/\/[^@\/]+@[^@\/]+/)) {
+ slashes = rest.substr(0, 2) === "//";
+ if (slashes && !(proto && hostlessProtocol[proto])) {
+ rest = rest.substr(2);
+ this.slashes = true;
+ }
+ }
+ if (!hostlessProtocol[proto] && (slashes || proto && !slashedProtocol[proto])) {
+ // there's a hostname.
+ // the first instance of /, ?, ;, or # ends the host.
+ // If there is an @ in the hostname, then non-host chars *are* allowed
+ // to the left of the last @ sign, unless some host-ending character
+ // comes *before* the @-sign.
+ // URLs are obnoxious.
+ // ex:
+ // http://a@b@c/ => user:a@b host:c
+ // http://a@b?@c => user:a host:c path:/?@c
+ // v0.12 TODO(isaacs): This is not quite how Chrome does things.
+ // Review our test case against browsers more comprehensively.
+ // find the first instance of any hostEndingChars
+ var hostEnd = -1;
+ for (i = 0; i < hostEndingChars.length; i++) {
+ hec = rest.indexOf(hostEndingChars[i]);
+ if (hec !== -1 && (hostEnd === -1 || hec < hostEnd)) {
+ hostEnd = hec;
+ }
+ }
+ // at this point, either we have an explicit point where the
+ // auth portion cannot go past, or the last @ char is the decider.
+ var auth, atSign;
+ if (hostEnd === -1) {
+ // atSign can be anywhere.
+ atSign = rest.lastIndexOf("@");
+ } else {
+ // atSign must be in auth portion.
+ // http://a@b/c@d => host:b auth:a path:/c@d
+ atSign = rest.lastIndexOf("@", hostEnd);
+ }
+ // Now we have a portion which is definitely the auth.
+ // Pull that off.
+ if (atSign !== -1) {
+ auth = rest.slice(0, atSign);
+ rest = rest.slice(atSign + 1);
+ this.auth = auth;
+ }
+ // the host is the remaining to the left of the first non-host char
+ hostEnd = -1;
+ for (i = 0; i < nonHostChars.length; i++) {
+ hec = rest.indexOf(nonHostChars[i]);
+ if (hec !== -1 && (hostEnd === -1 || hec < hostEnd)) {
+ hostEnd = hec;
+ }
+ }
+ // if we still have not hit it, then the entire thing is a host.
+ if (hostEnd === -1) {
+ hostEnd = rest.length;
+ }
+ if (rest[hostEnd - 1] === ":") {
+ hostEnd--;
+ }
+ var host = rest.slice(0, hostEnd);
+ rest = rest.slice(hostEnd);
+ // pull out port.
+ this.parseHost(host);
+ // we've indicated that there is a hostname,
+ // so even if it's empty, it has to be present.
+ this.hostname = this.hostname || "";
+ // if hostname begins with [ and ends with ]
+ // assume that it's an IPv6 address.
+ var ipv6Hostname = this.hostname[0] === "[" && this.hostname[this.hostname.length - 1] === "]";
+ // validate a little.
+ if (!ipv6Hostname) {
+ var hostparts = this.hostname.split(/\./);
+ for (i = 0, l = hostparts.length; i < l; i++) {
+ var part = hostparts[i];
+ if (!part) {
+ continue;
+ }
+ if (!part.match(hostnamePartPattern)) {
+ var newpart = "";
+ for (var j = 0, k = part.length; j < k; j++) {
+ if (part.charCodeAt(j) > 127) {
+ // we replace non-ASCII char with a temporary placeholder
+ // we need this to make sure size of hostname is not
+ // broken by replacing non-ASCII by nothing
+ newpart += "x";
+ } else {
+ newpart += part[j];
+ }
+ }
+ // we test again with ASCII char only
+ if (!newpart.match(hostnamePartPattern)) {
+ var validParts = hostparts.slice(0, i);
+ var notHost = hostparts.slice(i + 1);
+ var bit = part.match(hostnamePartStart);
+ if (bit) {
+ validParts.push(bit[1]);
+ notHost.unshift(bit[2]);
+ }
+ if (notHost.length) {
+ rest = notHost.join(".") + rest;
+ }
+ this.hostname = validParts.join(".");
+ break;
+ }
+ }
+ }
+ }
+ if (this.hostname.length > hostnameMaxLen) {
+ this.hostname = "";
+ }
+ // strip [ and ] from the hostname
+ // the host field still retains them, though
+ if (ipv6Hostname) {
+ this.hostname = this.hostname.substr(1, this.hostname.length - 2);
+ }
+ }
+ // chop off from the tail first.
+ var hash = rest.indexOf("#");
+ if (hash !== -1) {
+ // got a fragment string.
+ this.hash = rest.substr(hash);
+ rest = rest.slice(0, hash);
+ }
+ var qm = rest.indexOf("?");
+ if (qm !== -1) {
+ this.search = rest.substr(qm);
+ rest = rest.slice(0, qm);
+ }
+ if (rest) {
+ this.pathname = rest;
+ }
+ if (slashedProtocol[lowerProto] && this.hostname && !this.pathname) {
+ this.pathname = "";
+ }
+ return this;
+ };
+ Url.prototype.parseHost = function(host) {
+ var port = portPattern.exec(host);
+ if (port) {
+ port = port[0];
+ if (port !== ":") {
+ this.port = port.substr(1);
+ }
+ host = host.substr(0, host.length - port.length);
+ }
+ if (host) {
+ this.hostname = host;
+ }
+ };
+ var parse$1 = urlParse;
+ var encode$1 = encode_1;
+ var decode$1 = decode_1;
+ var format = format$1;
+ var parse = parse$1;
+ var mdurl = {
+ encode: encode$1,
+ decode: decode$1,
+ format: format,
+ parse: parse
+ };
+ var regex$3 = /[\0-\uD7FF\uE000-\uFFFF]|[\uD800-\uDBFF][\uDC00-\uDFFF]|[\uD800-\uDBFF](?![\uDC00-\uDFFF])|(?:[^\uD800-\uDBFF]|^)[\uDC00-\uDFFF]/;
+ var regex$2 = /[\0-\x1F\x7F-\x9F]/;
+ var regex$1 = /[\xAD\u0600-\u0605\u061C\u06DD\u070F\u08E2\u180E\u200B-\u200F\u202A-\u202E\u2060-\u2064\u2066-\u206F\uFEFF\uFFF9-\uFFFB]|\uD804[\uDCBD\uDCCD]|\uD82F[\uDCA0-\uDCA3]|\uD834[\uDD73-\uDD7A]|\uDB40[\uDC01\uDC20-\uDC7F]/;
+ var regex = /[ \xA0\u1680\u2000-\u200A\u2028\u2029\u202F\u205F\u3000]/;
+ var Any = regex$3;
+ var Cc = regex$2;
+ var Cf = regex$1;
+ var P = regex$4;
+ var Z = regex;
+ var uc_micro = {
+ Any: Any,
+ Cc: Cc,
+ Cf: Cf,
+ P: P,
+ Z: Z
+ };
+ var utils = createCommonjsModule((function(module, exports) {
+ function _class(obj) {
+ return Object.prototype.toString.call(obj);
+ }
+ function isString(obj) {
+ return _class(obj) === "[object String]";
+ }
+ var _hasOwnProperty = Object.prototype.hasOwnProperty;
+ function has(object, key) {
+ return _hasOwnProperty.call(object, key);
+ }
+ // Merge objects
+
+ function assign(obj /*from1, from2, from3, ...*/) {
+ var sources = Array.prototype.slice.call(arguments, 1);
+ sources.forEach((function(source) {
+ if (!source) {
+ return;
+ }
+ if (typeof source !== "object") {
+ throw new TypeError(source + "must be object");
+ }
+ Object.keys(source).forEach((function(key) {
+ obj[key] = source[key];
+ }));
+ }));
+ return obj;
+ }
+ // Remove element from array and put another array at those position.
+ // Useful for some operations with tokens
+ function arrayReplaceAt(src, pos, newElements) {
+ return [].concat(src.slice(0, pos), newElements, src.slice(pos + 1));
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ function isValidEntityCode(c) {
+ /*eslint no-bitwise:0*/
+ // broken sequence
+ if (c >= 55296 && c <= 57343) {
+ return false;
+ }
+ // never used
+ if (c >= 64976 && c <= 65007) {
+ return false;
+ }
+ if ((c & 65535) === 65535 || (c & 65535) === 65534) {
+ return false;
+ }
+ // control codes
+ if (c >= 0 && c <= 8) {
+ return false;
+ }
+ if (c === 11) {
+ return false;
+ }
+ if (c >= 14 && c <= 31) {
+ return false;
+ }
+ if (c >= 127 && c <= 159) {
+ return false;
+ }
+ // out of range
+ if (c > 1114111) {
+ return false;
+ }
+ return true;
+ }
+ function fromCodePoint(c) {
+ /*eslint no-bitwise:0*/
+ if (c > 65535) {
+ c -= 65536;
+ var surrogate1 = 55296 + (c >> 10), surrogate2 = 56320 + (c & 1023);
+ return String.fromCharCode(surrogate1, surrogate2);
+ }
+ return String.fromCharCode(c);
+ }
+ var UNESCAPE_MD_RE = /\\([!"#$%&'()*+,\-.\/:;<=>?@[\\\]^_`{|}~])/g;
+ var ENTITY_RE = /&([a-z#][a-z0-9]{1,31});/gi;
+ var UNESCAPE_ALL_RE = new RegExp(UNESCAPE_MD_RE.source + "|" + ENTITY_RE.source, "gi");
+ var DIGITAL_ENTITY_TEST_RE = /^#((?:x[a-f0-9]{1,8}|[0-9]{1,8}))/i;
+ function replaceEntityPattern(match, name) {
+ var code = 0;
+ if (has(entities, name)) {
+ return entities[name];
+ }
+ if (name.charCodeAt(0) === 35 /* # */ && DIGITAL_ENTITY_TEST_RE.test(name)) {
+ code = name[1].toLowerCase() === "x" ? parseInt(name.slice(2), 16) : parseInt(name.slice(1), 10);
+ if (isValidEntityCode(code)) {
+ return fromCodePoint(code);
+ }
+ }
+ return match;
+ }
+ /*function replaceEntities(str) {
+ if (str.indexOf('&') < 0) { return str; }
+
+ return str.replace(ENTITY_RE, replaceEntityPattern);
+ }*/ function unescapeMd(str) {
+ if (str.indexOf("\\") < 0) {
+ return str;
+ }
+ return str.replace(UNESCAPE_MD_RE, "$1");
+ }
+ function unescapeAll(str) {
+ if (str.indexOf("\\") < 0 && str.indexOf("&") < 0) {
+ return str;
+ }
+ return str.replace(UNESCAPE_ALL_RE, (function(match, escaped, entity) {
+ if (escaped) {
+ return escaped;
+ }
+ return replaceEntityPattern(match, entity);
+ }));
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ var HTML_ESCAPE_TEST_RE = /[&<>"]/;
+ var HTML_ESCAPE_REPLACE_RE = /[&<>"]/g;
+ var HTML_REPLACEMENTS = {
+ "&": "&amp;",
+ "<": "&lt;",
+ ">": "&gt;",
+ '"': "&quot;"
+ };
+ function replaceUnsafeChar(ch) {
+ return HTML_REPLACEMENTS[ch];
+ }
+ function escapeHtml(str) {
+ if (HTML_ESCAPE_TEST_RE.test(str)) {
+ return str.replace(HTML_ESCAPE_REPLACE_RE, replaceUnsafeChar);
+ }
+ return str;
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ var REGEXP_ESCAPE_RE = /[.?*+^$[\]\\(){}|-]/g;
+ function escapeRE(str) {
+ return str.replace(REGEXP_ESCAPE_RE, "\\$&");
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ function isSpace(code) {
+ switch (code) {
+ case 9:
+ case 32:
+ return true;
+ }
+ return false;
+ }
+ // Zs (unicode class) || [\t\f\v\r\n]
+ function isWhiteSpace(code) {
+ if (code >= 8192 && code <= 8202) {
+ return true;
+ }
+ switch (code) {
+ case 9:
+ // \t
+ case 10:
+ // \n
+ case 11:
+ // \v
+ case 12:
+ // \f
+ case 13:
+ // \r
+ case 32:
+ case 160:
+ case 5760:
+ case 8239:
+ case 8287:
+ case 12288:
+ return true;
+ }
+ return false;
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ /*eslint-disable max-len*/
+ // Currently without astral characters support.
+ function isPunctChar(ch) {
+ return regex$4.test(ch);
+ }
+ // Markdown ASCII punctuation characters.
+
+ // !, ", #, $, %, &, ', (, ), *, +, ,, -, ., /, :, ;, <, =, >, ?, @, [, \, ], ^, _, `, {, |, }, or ~
+ // http://spec.commonmark.org/0.15/#ascii-punctuation-character
+
+ // Don't confuse with unicode punctuation !!! It lacks some chars in ascii range.
+
+ function isMdAsciiPunct(ch) {
+ switch (ch) {
+ case 33 /* ! */ :
+ case 34 /* " */ :
+ case 35 /* # */ :
+ case 36 /* $ */ :
+ case 37 /* % */ :
+ case 38 /* & */ :
+ case 39 /* ' */ :
+ case 40 /* ( */ :
+ case 41 /* ) */ :
+ case 42 /* * */ :
+ case 43 /* + */ :
+ case 44 /* , */ :
+ case 45 /* - */ :
+ case 46 /* . */ :
+ case 47 /* / */ :
+ case 58 /* : */ :
+ case 59 /* ; */ :
+ case 60 /* < */ :
+ case 61 /* = */ :
+ case 62 /* > */ :
+ case 63 /* ? */ :
+ case 64 /* @ */ :
+ case 91 /* [ */ :
+ case 92 /* \ */ :
+ case 93 /* ] */ :
+ case 94 /* ^ */ :
+ case 95 /* _ */ :
+ case 96 /* ` */ :
+ case 123 /* { */ :
+ case 124 /* | */ :
+ case 125 /* } */ :
+ case 126 /* ~ */ :
+ return true;
+
+ default:
+ return false;
+ }
+ }
+ // Hepler to unify [reference labels].
+
+ function normalizeReference(str) {
+ // Trim and collapse whitespace
+ str = str.trim().replace(/\s+/g, " ");
+ // In node v10 'ẞ'.toLowerCase() === 'Ṿ', which is presumed to be a bug
+ // fixed in v12 (couldn't find any details).
+
+ // So treat this one as a special case
+ // (remove this when node v10 is no longer supported).
+
+ if ("\u1e9e".toLowerCase() === "\u1e7e") {
+ str = str.replace(/\u1e9e/g, "\xdf");
+ }
+ // .toLowerCase().toUpperCase() should get rid of all differences
+ // between letter variants.
+
+ // Simple .toLowerCase() doesn't normalize 125 code points correctly,
+ // and .toUpperCase doesn't normalize 6 of them (list of exceptions:
+ // İ, ϴ, ẞ, Ω, K, Å - those are already uppercased, but have differently
+ // uppercased versions).
+
+ // Here's an example showing how it happens. Lets take greek letter omega:
+ // uppercase U+0398 (Θ), U+03f4 (ϴ) and lowercase U+03b8 (θ), U+03d1 (ϑ)
+
+ // Unicode entries:
+ // 0398;GREEK CAPITAL LETTER THETA;Lu;0;L;;;;;N;;;;03B8;
+ // 03B8;GREEK SMALL LETTER THETA;Ll;0;L;;;;;N;;;0398;;0398
+ // 03D1;GREEK THETA SYMBOL;Ll;0;L;<compat> 03B8;;;;N;GREEK SMALL LETTER SCRIPT THETA;;0398;;0398
+ // 03F4;GREEK CAPITAL THETA SYMBOL;Lu;0;L;<compat> 0398;;;;N;;;;03B8;
+
+ // Case-insensitive comparison should treat all of them as equivalent.
+
+ // But .toLowerCase() doesn't change ϑ (it's already lowercase),
+ // and .toUpperCase() doesn't change ϴ (already uppercase).
+
+ // Applying first lower then upper case normalizes any character:
+ // '\u0398\u03f4\u03b8\u03d1'.toLowerCase().toUpperCase() === '\u0398\u0398\u0398\u0398'
+
+ // Note: this is equivalent to unicode case folding; unicode normalization
+ // is a different step that is not required here.
+
+ // Final result should be uppercased, because it's later stored in an object
+ // (this avoid a conflict with Object.prototype members,
+ // most notably, `__proto__`)
+
+ return str.toLowerCase().toUpperCase();
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ // Re-export libraries commonly used in both markdown-it and its plugins,
+ // so plugins won't have to depend on them explicitly, which reduces their
+ // bundled size (e.g. a browser build).
+
+ exports.lib = {};
+ exports.lib.mdurl = mdurl;
+ exports.lib.ucmicro = uc_micro;
+ exports.assign = assign;
+ exports.isString = isString;
+ exports.has = has;
+ exports.unescapeMd = unescapeMd;
+ exports.unescapeAll = unescapeAll;
+ exports.isValidEntityCode = isValidEntityCode;
+ exports.fromCodePoint = fromCodePoint;
+ // exports.replaceEntities = replaceEntities;
+ exports.escapeHtml = escapeHtml;
+ exports.arrayReplaceAt = arrayReplaceAt;
+ exports.isSpace = isSpace;
+ exports.isWhiteSpace = isWhiteSpace;
+ exports.isMdAsciiPunct = isMdAsciiPunct;
+ exports.isPunctChar = isPunctChar;
+ exports.escapeRE = escapeRE;
+ exports.normalizeReference = normalizeReference;
+ }));
+ // Parse link label
+ var parse_link_label = function parseLinkLabel(state, start, disableNested) {
+ var level, found, marker, prevPos, labelEnd = -1, max = state.posMax, oldPos = state.pos;
+ state.pos = start + 1;
+ level = 1;
+ while (state.pos < max) {
+ marker = state.src.charCodeAt(state.pos);
+ if (marker === 93 /* ] */) {
+ level--;
+ if (level === 0) {
+ found = true;
+ break;
+ }
+ }
+ prevPos = state.pos;
+ state.md.inline.skipToken(state);
+ if (marker === 91 /* [ */) {
+ if (prevPos === state.pos - 1) {
+ // increase level if we find text `[`, which is not a part of any token
+ level++;
+ } else if (disableNested) {
+ state.pos = oldPos;
+ return -1;
+ }
+ }
+ }
+ if (found) {
+ labelEnd = state.pos;
+ }
+ // restore old state
+ state.pos = oldPos;
+ return labelEnd;
+ };
+ var unescapeAll$2 = utils.unescapeAll;
+ var parse_link_destination = function parseLinkDestination(str, pos, max) {
+ var code, level, lines = 0, start = pos, result = {
+ ok: false,
+ pos: 0,
+ lines: 0,
+ str: ""
+ };
+ if (str.charCodeAt(pos) === 60 /* < */) {
+ pos++;
+ while (pos < max) {
+ code = str.charCodeAt(pos);
+ if (code === 10 /* \n */) {
+ return result;
+ }
+ if (code === 60 /* < */) {
+ return result;
+ }
+ if (code === 62 /* > */) {
+ result.pos = pos + 1;
+ result.str = unescapeAll$2(str.slice(start + 1, pos));
+ result.ok = true;
+ return result;
+ }
+ if (code === 92 /* \ */ && pos + 1 < max) {
+ pos += 2;
+ continue;
+ }
+ pos++;
+ }
+ // no closing '>'
+ return result;
+ }
+ // this should be ... } else { ... branch
+ level = 0;
+ while (pos < max) {
+ code = str.charCodeAt(pos);
+ if (code === 32) {
+ break;
+ }
+ // ascii control characters
+ if (code < 32 || code === 127) {
+ break;
+ }
+ if (code === 92 /* \ */ && pos + 1 < max) {
+ if (str.charCodeAt(pos + 1) === 32) {
+ break;
+ }
+ pos += 2;
+ continue;
+ }
+ if (code === 40 /* ( */) {
+ level++;
+ if (level > 32) {
+ return result;
+ }
+ }
+ if (code === 41 /* ) */) {
+ if (level === 0) {
+ break;
+ }
+ level--;
+ }
+ pos++;
+ }
+ if (start === pos) {
+ return result;
+ }
+ if (level !== 0) {
+ return result;
+ }
+ result.str = unescapeAll$2(str.slice(start, pos));
+ result.lines = lines;
+ result.pos = pos;
+ result.ok = true;
+ return result;
+ };
+ var unescapeAll$1 = utils.unescapeAll;
+ var parse_link_title = function parseLinkTitle(str, pos, max) {
+ var code, marker, lines = 0, start = pos, result = {
+ ok: false,
+ pos: 0,
+ lines: 0,
+ str: ""
+ };
+ if (pos >= max) {
+ return result;
+ }
+ marker = str.charCodeAt(pos);
+ if (marker !== 34 /* " */ && marker !== 39 /* ' */ && marker !== 40 /* ( */) {
+ return result;
+ }
+ pos++;
+ // if opening marker is "(", switch it to closing marker ")"
+ if (marker === 40) {
+ marker = 41;
+ }
+ while (pos < max) {
+ code = str.charCodeAt(pos);
+ if (code === marker) {
+ result.pos = pos + 1;
+ result.lines = lines;
+ result.str = unescapeAll$1(str.slice(start + 1, pos));
+ result.ok = true;
+ return result;
+ } else if (code === 40 /* ( */ && marker === 41 /* ) */) {
+ return result;
+ } else if (code === 10) {
+ lines++;
+ } else if (code === 92 /* \ */ && pos + 1 < max) {
+ pos++;
+ if (str.charCodeAt(pos) === 10) {
+ lines++;
+ }
+ }
+ pos++;
+ }
+ return result;
+ };
+ var parseLinkLabel = parse_link_label;
+ var parseLinkDestination = parse_link_destination;
+ var parseLinkTitle = parse_link_title;
+ var helpers = {
+ parseLinkLabel: parseLinkLabel,
+ parseLinkDestination: parseLinkDestination,
+ parseLinkTitle: parseLinkTitle
+ };
+ var assign$1 = utils.assign;
+ var unescapeAll = utils.unescapeAll;
+ var escapeHtml = utils.escapeHtml;
+ ////////////////////////////////////////////////////////////////////////////////
+ var default_rules = {};
+ default_rules.code_inline = function(tokens, idx, options, env, slf) {
+ var token = tokens[idx];
+ return "<code" + slf.renderAttrs(token) + ">" + escapeHtml(tokens[idx].content) + "</code>";
+ };
+ default_rules.code_block = function(tokens, idx, options, env, slf) {
+ var token = tokens[idx];
+ return "<pre" + slf.renderAttrs(token) + "><code>" + escapeHtml(tokens[idx].content) + "</code></pre>\n";
+ };
+ default_rules.fence = function(tokens, idx, options, env, slf) {
+ var token = tokens[idx], info = token.info ? unescapeAll(token.info).trim() : "", langName = "", langAttrs = "", highlighted, i, arr, tmpAttrs, tmpToken;
+ if (info) {
+ arr = info.split(/(\s+)/g);
+ langName = arr[0];
+ langAttrs = arr.slice(2).join("");
+ }
+ if (options.highlight) {
+ highlighted = options.highlight(token.content, langName, langAttrs) || escapeHtml(token.content);
+ } else {
+ highlighted = escapeHtml(token.content);
+ }
+ if (highlighted.indexOf("<pre") === 0) {
+ return highlighted + "\n";
+ }
+ // If language exists, inject class gently, without modifying original token.
+ // May be, one day we will add .deepClone() for token and simplify this part, but
+ // now we prefer to keep things local.
+ if (info) {
+ i = token.attrIndex("class");
+ tmpAttrs = token.attrs ? token.attrs.slice() : [];
+ if (i < 0) {
+ tmpAttrs.push([ "class", options.langPrefix + langName ]);
+ } else {
+ tmpAttrs[i] = tmpAttrs[i].slice();
+ tmpAttrs[i][1] += " " + options.langPrefix + langName;
+ }
+ // Fake token just to render attributes
+ tmpToken = {
+ attrs: tmpAttrs
+ };
+ return "<pre><code" + slf.renderAttrs(tmpToken) + ">" + highlighted + "</code></pre>\n";
+ }
+ return "<pre><code" + slf.renderAttrs(token) + ">" + highlighted + "</code></pre>\n";
+ };
+ default_rules.image = function(tokens, idx, options, env, slf) {
+ var token = tokens[idx];
+ // "alt" attr MUST be set, even if empty. Because it's mandatory and
+ // should be placed on proper position for tests.
+
+ // Replace content with actual value
+ token.attrs[token.attrIndex("alt")][1] = slf.renderInlineAsText(token.children, options, env);
+ return slf.renderToken(tokens, idx, options);
+ };
+ default_rules.hardbreak = function(tokens, idx, options /*, env */) {
+ return options.xhtmlOut ? "<br />\n" : "<br>\n";
+ };
+ default_rules.softbreak = function(tokens, idx, options /*, env */) {
+ return options.breaks ? options.xhtmlOut ? "<br />\n" : "<br>\n" : "\n";
+ };
+ default_rules.text = function(tokens, idx /*, options, env */) {
+ return escapeHtml(tokens[idx].content);
+ };
+ default_rules.html_block = function(tokens, idx /*, options, env */) {
+ return tokens[idx].content;
+ };
+ default_rules.html_inline = function(tokens, idx /*, options, env */) {
+ return tokens[idx].content;
+ };
+ /**
+ * new Renderer()
+ *
+ * Creates new [[Renderer]] instance and fill [[Renderer#rules]] with defaults.
+ **/ function Renderer() {
+ /**
+ * Renderer#rules -> Object
+ *
+ * Contains render rules for tokens. Can be updated and extended.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * md.renderer.rules.strong_open = function () { return '<b>'; };
+ * md.renderer.rules.strong_close = function () { return '</b>'; };
+ *
+ * var result = md.renderInline(...);
+ * ```
+ *
+ * Each rule is called as independent static function with fixed signature:
+ *
+ * ```javascript
+ * function my_token_render(tokens, idx, options, env, renderer) {
+ * // ...
+ * return renderedHTML;
+ * }
+ * ```
+ *
+ * See [source code](https://github.com/markdown-it/markdown-it/blob/master/lib/renderer.js)
+ * for more details and examples.
+ **/
+ this.rules = assign$1({}, default_rules);
+ }
+ /**
+ * Renderer.renderAttrs(token) -> String
+ *
+ * Render token attributes to string.
+ **/ Renderer.prototype.renderAttrs = function renderAttrs(token) {
+ var i, l, result;
+ if (!token.attrs) {
+ return "";
+ }
+ result = "";
+ for (i = 0, l = token.attrs.length; i < l; i++) {
+ result += " " + escapeHtml(token.attrs[i][0]) + '="' + escapeHtml(token.attrs[i][1]) + '"';
+ }
+ return result;
+ };
+ /**
+ * Renderer.renderToken(tokens, idx, options) -> String
+ * - tokens (Array): list of tokens
+ * - idx (Numbed): token index to render
+ * - options (Object): params of parser instance
+ *
+ * Default token renderer. Can be overriden by custom function
+ * in [[Renderer#rules]].
+ **/ Renderer.prototype.renderToken = function renderToken(tokens, idx, options) {
+ var nextToken, result = "", needLf = false, token = tokens[idx];
+ // Tight list paragraphs
+ if (token.hidden) {
+ return "";
+ }
+ // Insert a newline between hidden paragraph and subsequent opening
+ // block-level tag.
+
+ // For example, here we should insert a newline before blockquote:
+ // - a
+ // >
+
+ if (token.block && token.nesting !== -1 && idx && tokens[idx - 1].hidden) {
+ result += "\n";
+ }
+ // Add token name, e.g. `<img`
+ result += (token.nesting === -1 ? "</" : "<") + token.tag;
+ // Encode attributes, e.g. `<img src="foo"`
+ result += this.renderAttrs(token);
+ // Add a slash for self-closing tags, e.g. `<img src="foo" /`
+ if (token.nesting === 0 && options.xhtmlOut) {
+ result += " /";
+ }
+ // Check if we need to add a newline after this tag
+ if (token.block) {
+ needLf = true;
+ if (token.nesting === 1) {
+ if (idx + 1 < tokens.length) {
+ nextToken = tokens[idx + 1];
+ if (nextToken.type === "inline" || nextToken.hidden) {
+ // Block-level tag containing an inline tag.
+ needLf = false;
+ } else if (nextToken.nesting === -1 && nextToken.tag === token.tag) {
+ // Opening tag + closing tag of the same type. E.g. `<li></li>`.
+ needLf = false;
+ }
+ }
+ }
+ }
+ result += needLf ? ">\n" : ">";
+ return result;
+ };
+ /**
+ * Renderer.renderInline(tokens, options, env) -> String
+ * - tokens (Array): list on block tokens to render
+ * - options (Object): params of parser instance
+ * - env (Object): additional data from parsed input (references, for example)
+ *
+ * The same as [[Renderer.render]], but for single token of `inline` type.
+ **/ Renderer.prototype.renderInline = function(tokens, options, env) {
+ var type, result = "", rules = this.rules;
+ for (var i = 0, len = tokens.length; i < len; i++) {
+ type = tokens[i].type;
+ if (typeof rules[type] !== "undefined") {
+ result += rules[type](tokens, i, options, env, this);
+ } else {
+ result += this.renderToken(tokens, i, options);
+ }
+ }
+ return result;
+ };
+ /** internal
+ * Renderer.renderInlineAsText(tokens, options, env) -> String
+ * - tokens (Array): list on block tokens to render
+ * - options (Object): params of parser instance
+ * - env (Object): additional data from parsed input (references, for example)
+ *
+ * Special kludge for image `alt` attributes to conform CommonMark spec.
+ * Don't try to use it! Spec requires to show `alt` content with stripped markup,
+ * instead of simple escaping.
+ **/ Renderer.prototype.renderInlineAsText = function(tokens, options, env) {
+ var result = "";
+ for (var i = 0, len = tokens.length; i < len; i++) {
+ if (tokens[i].type === "text") {
+ result += tokens[i].content;
+ } else if (tokens[i].type === "image") {
+ result += this.renderInlineAsText(tokens[i].children, options, env);
+ } else if (tokens[i].type === "softbreak") {
+ result += "\n";
+ }
+ }
+ return result;
+ };
+ /**
+ * Renderer.render(tokens, options, env) -> String
+ * - tokens (Array): list on block tokens to render
+ * - options (Object): params of parser instance
+ * - env (Object): additional data from parsed input (references, for example)
+ *
+ * Takes token stream and generates HTML. Probably, you will never need to call
+ * this method directly.
+ **/ Renderer.prototype.render = function(tokens, options, env) {
+ var i, len, type, result = "", rules = this.rules;
+ for (i = 0, len = tokens.length; i < len; i++) {
+ type = tokens[i].type;
+ if (type === "inline") {
+ result += this.renderInline(tokens[i].children, options, env);
+ } else if (typeof rules[type] !== "undefined") {
+ result += rules[tokens[i].type](tokens, i, options, env, this);
+ } else {
+ result += this.renderToken(tokens, i, options, env);
+ }
+ }
+ return result;
+ };
+ var renderer = Renderer;
+ /**
+ * class Ruler
+ *
+ * Helper class, used by [[MarkdownIt#core]], [[MarkdownIt#block]] and
+ * [[MarkdownIt#inline]] to manage sequences of functions (rules):
+ *
+ * - keep rules in defined order
+ * - assign the name to each rule
+ * - enable/disable rules
+ * - add/replace rules
+ * - allow assign rules to additional named chains (in the same)
+ * - cacheing lists of active rules
+ *
+ * You will not need use this class directly until write plugins. For simple
+ * rules control use [[MarkdownIt.disable]], [[MarkdownIt.enable]] and
+ * [[MarkdownIt.use]].
+ **/
+ /**
+ * new Ruler()
+ **/ function Ruler() {
+ // List of added rules. Each element is:
+ // {
+ // name: XXX,
+ // enabled: Boolean,
+ // fn: Function(),
+ // alt: [ name2, name3 ]
+ // }
+ this.__rules__ = [];
+ // Cached rule chains.
+
+ // First level - chain name, '' for default.
+ // Second level - diginal anchor for fast filtering by charcodes.
+
+ this.__cache__ = null;
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ // Helper methods, should not be used directly
+ // Find rule index by name
+
+ Ruler.prototype.__find__ = function(name) {
+ for (var i = 0; i < this.__rules__.length; i++) {
+ if (this.__rules__[i].name === name) {
+ return i;
+ }
+ }
+ return -1;
+ };
+ // Build rules lookup cache
+
+ Ruler.prototype.__compile__ = function() {
+ var self = this;
+ var chains = [ "" ];
+ // collect unique names
+ self.__rules__.forEach((function(rule) {
+ if (!rule.enabled) {
+ return;
+ }
+ rule.alt.forEach((function(altName) {
+ if (chains.indexOf(altName) < 0) {
+ chains.push(altName);
+ }
+ }));
+ }));
+ self.__cache__ = {};
+ chains.forEach((function(chain) {
+ self.__cache__[chain] = [];
+ self.__rules__.forEach((function(rule) {
+ if (!rule.enabled) {
+ return;
+ }
+ if (chain && rule.alt.indexOf(chain) < 0) {
+ return;
+ }
+ self.__cache__[chain].push(rule.fn);
+ }));
+ }));
+ };
+ /**
+ * Ruler.at(name, fn [, options])
+ * - name (String): rule name to replace.
+ * - fn (Function): new rule function.
+ * - options (Object): new rule options (not mandatory).
+ *
+ * Replace rule by name with new function & options. Throws error if name not
+ * found.
+ *
+ * ##### Options:
+ *
+ * - __alt__ - array with names of "alternate" chains.
+ *
+ * ##### Example
+ *
+ * Replace existing typographer replacement rule with new one:
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * md.core.ruler.at('replacements', function replace(state) {
+ * //...
+ * });
+ * ```
+ **/ Ruler.prototype.at = function(name, fn, options) {
+ var index = this.__find__(name);
+ var opt = options || {};
+ if (index === -1) {
+ throw new Error("Parser rule not found: " + name);
+ }
+ this.__rules__[index].fn = fn;
+ this.__rules__[index].alt = opt.alt || [];
+ this.__cache__ = null;
+ };
+ /**
+ * Ruler.before(beforeName, ruleName, fn [, options])
+ * - beforeName (String): new rule will be added before this one.
+ * - ruleName (String): name of added rule.
+ * - fn (Function): rule function.
+ * - options (Object): rule options (not mandatory).
+ *
+ * Add new rule to chain before one with given name. See also
+ * [[Ruler.after]], [[Ruler.push]].
+ *
+ * ##### Options:
+ *
+ * - __alt__ - array with names of "alternate" chains.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * md.block.ruler.before('paragraph', 'my_rule', function replace(state) {
+ * //...
+ * });
+ * ```
+ **/ Ruler.prototype.before = function(beforeName, ruleName, fn, options) {
+ var index = this.__find__(beforeName);
+ var opt = options || {};
+ if (index === -1) {
+ throw new Error("Parser rule not found: " + beforeName);
+ }
+ this.__rules__.splice(index, 0, {
+ name: ruleName,
+ enabled: true,
+ fn: fn,
+ alt: opt.alt || []
+ });
+ this.__cache__ = null;
+ };
+ /**
+ * Ruler.after(afterName, ruleName, fn [, options])
+ * - afterName (String): new rule will be added after this one.
+ * - ruleName (String): name of added rule.
+ * - fn (Function): rule function.
+ * - options (Object): rule options (not mandatory).
+ *
+ * Add new rule to chain after one with given name. See also
+ * [[Ruler.before]], [[Ruler.push]].
+ *
+ * ##### Options:
+ *
+ * - __alt__ - array with names of "alternate" chains.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * md.inline.ruler.after('text', 'my_rule', function replace(state) {
+ * //...
+ * });
+ * ```
+ **/ Ruler.prototype.after = function(afterName, ruleName, fn, options) {
+ var index = this.__find__(afterName);
+ var opt = options || {};
+ if (index === -1) {
+ throw new Error("Parser rule not found: " + afterName);
+ }
+ this.__rules__.splice(index + 1, 0, {
+ name: ruleName,
+ enabled: true,
+ fn: fn,
+ alt: opt.alt || []
+ });
+ this.__cache__ = null;
+ };
+ /**
+ * Ruler.push(ruleName, fn [, options])
+ * - ruleName (String): name of added rule.
+ * - fn (Function): rule function.
+ * - options (Object): rule options (not mandatory).
+ *
+ * Push new rule to the end of chain. See also
+ * [[Ruler.before]], [[Ruler.after]].
+ *
+ * ##### Options:
+ *
+ * - __alt__ - array with names of "alternate" chains.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * md.core.ruler.push('my_rule', function replace(state) {
+ * //...
+ * });
+ * ```
+ **/ Ruler.prototype.push = function(ruleName, fn, options) {
+ var opt = options || {};
+ this.__rules__.push({
+ name: ruleName,
+ enabled: true,
+ fn: fn,
+ alt: opt.alt || []
+ });
+ this.__cache__ = null;
+ };
+ /**
+ * Ruler.enable(list [, ignoreInvalid]) -> Array
+ * - list (String|Array): list of rule names to enable.
+ * - ignoreInvalid (Boolean): set `true` to ignore errors when rule not found.
+ *
+ * Enable rules with given names. If any rule name not found - throw Error.
+ * Errors can be disabled by second param.
+ *
+ * Returns list of found rule names (if no exception happened).
+ *
+ * See also [[Ruler.disable]], [[Ruler.enableOnly]].
+ **/ Ruler.prototype.enable = function(list, ignoreInvalid) {
+ if (!Array.isArray(list)) {
+ list = [ list ];
+ }
+ var result = [];
+ // Search by name and enable
+ list.forEach((function(name) {
+ var idx = this.__find__(name);
+ if (idx < 0) {
+ if (ignoreInvalid) {
+ return;
+ }
+ throw new Error("Rules manager: invalid rule name " + name);
+ }
+ this.__rules__[idx].enabled = true;
+ result.push(name);
+ }), this);
+ this.__cache__ = null;
+ return result;
+ };
+ /**
+ * Ruler.enableOnly(list [, ignoreInvalid])
+ * - list (String|Array): list of rule names to enable (whitelist).
+ * - ignoreInvalid (Boolean): set `true` to ignore errors when rule not found.
+ *
+ * Enable rules with given names, and disable everything else. If any rule name
+ * not found - throw Error. Errors can be disabled by second param.
+ *
+ * See also [[Ruler.disable]], [[Ruler.enable]].
+ **/ Ruler.prototype.enableOnly = function(list, ignoreInvalid) {
+ if (!Array.isArray(list)) {
+ list = [ list ];
+ }
+ this.__rules__.forEach((function(rule) {
+ rule.enabled = false;
+ }));
+ this.enable(list, ignoreInvalid);
+ };
+ /**
+ * Ruler.disable(list [, ignoreInvalid]) -> Array
+ * - list (String|Array): list of rule names to disable.
+ * - ignoreInvalid (Boolean): set `true` to ignore errors when rule not found.
+ *
+ * Disable rules with given names. If any rule name not found - throw Error.
+ * Errors can be disabled by second param.
+ *
+ * Returns list of found rule names (if no exception happened).
+ *
+ * See also [[Ruler.enable]], [[Ruler.enableOnly]].
+ **/ Ruler.prototype.disable = function(list, ignoreInvalid) {
+ if (!Array.isArray(list)) {
+ list = [ list ];
+ }
+ var result = [];
+ // Search by name and disable
+ list.forEach((function(name) {
+ var idx = this.__find__(name);
+ if (idx < 0) {
+ if (ignoreInvalid) {
+ return;
+ }
+ throw new Error("Rules manager: invalid rule name " + name);
+ }
+ this.__rules__[idx].enabled = false;
+ result.push(name);
+ }), this);
+ this.__cache__ = null;
+ return result;
+ };
+ /**
+ * Ruler.getRules(chainName) -> Array
+ *
+ * Return array of active functions (rules) for given chain name. It analyzes
+ * rules configuration, compiles caches if not exists and returns result.
+ *
+ * Default chain name is `''` (empty string). It can't be skipped. That's
+ * done intentionally, to keep signature monomorphic for high speed.
+ **/ Ruler.prototype.getRules = function(chainName) {
+ if (this.__cache__ === null) {
+ this.__compile__();
+ }
+ // Chain can be empty, if rules disabled. But we still have to return Array.
+ return this.__cache__[chainName] || [];
+ };
+ var ruler = Ruler;
+ // Normalize input string
+ // https://spec.commonmark.org/0.29/#line-ending
+ var NEWLINES_RE = /\r\n?|\n/g;
+ var NULL_RE = /\0/g;
+ var normalize = function normalize(state) {
+ var str;
+ // Normalize newlines
+ str = state.src.replace(NEWLINES_RE, "\n");
+ // Replace NULL characters
+ str = str.replace(NULL_RE, "\ufffd");
+ state.src = str;
+ };
+ var block = function block(state) {
+ var token;
+ if (state.inlineMode) {
+ token = new state.Token("inline", "", 0);
+ token.content = state.src;
+ token.map = [ 0, 1 ];
+ token.children = [];
+ state.tokens.push(token);
+ } else {
+ state.md.block.parse(state.src, state.md, state.env, state.tokens);
+ }
+ };
+ var inline = function inline(state) {
+ var tokens = state.tokens, tok, i, l;
+ // Parse inlines
+ for (i = 0, l = tokens.length; i < l; i++) {
+ tok = tokens[i];
+ if (tok.type === "inline") {
+ state.md.inline.parse(tok.content, state.md, state.env, tok.children);
+ }
+ }
+ };
+ var arrayReplaceAt = utils.arrayReplaceAt;
+ function isLinkOpen(str) {
+ return /^<a[>\s]/i.test(str);
+ }
+ function isLinkClose(str) {
+ return /^<\/a\s*>/i.test(str);
+ }
+ var linkify = function linkify(state) {
+ var i, j, l, tokens, token, currentToken, nodes, ln, text, pos, lastPos, level, htmlLinkLevel, url, fullUrl, urlText, blockTokens = state.tokens, links;
+ if (!state.md.options.linkify) {
+ return;
+ }
+ for (j = 0, l = blockTokens.length; j < l; j++) {
+ if (blockTokens[j].type !== "inline" || !state.md.linkify.pretest(blockTokens[j].content)) {
+ continue;
+ }
+ tokens = blockTokens[j].children;
+ htmlLinkLevel = 0;
+ // We scan from the end, to keep position when new tags added.
+ // Use reversed logic in links start/end match
+ for (i = tokens.length - 1; i >= 0; i--) {
+ currentToken = tokens[i];
+ // Skip content of markdown links
+ if (currentToken.type === "link_close") {
+ i--;
+ while (tokens[i].level !== currentToken.level && tokens[i].type !== "link_open") {
+ i--;
+ }
+ continue;
+ }
+ // Skip content of html tag links
+ if (currentToken.type === "html_inline") {
+ if (isLinkOpen(currentToken.content) && htmlLinkLevel > 0) {
+ htmlLinkLevel--;
+ }
+ if (isLinkClose(currentToken.content)) {
+ htmlLinkLevel++;
+ }
+ }
+ if (htmlLinkLevel > 0) {
+ continue;
+ }
+ if (currentToken.type === "text" && state.md.linkify.test(currentToken.content)) {
+ text = currentToken.content;
+ links = state.md.linkify.match(text);
+ // Now split string to nodes
+ nodes = [];
+ level = currentToken.level;
+ lastPos = 0;
+ for (ln = 0; ln < links.length; ln++) {
+ url = links[ln].url;
+ fullUrl = state.md.normalizeLink(url);
+ if (!state.md.validateLink(fullUrl)) {
+ continue;
+ }
+ urlText = links[ln].text;
+ // Linkifier might send raw hostnames like "example.com", where url
+ // starts with domain name. So we prepend http:// in those cases,
+ // and remove it afterwards.
+
+ if (!links[ln].schema) {
+ urlText = state.md.normalizeLinkText("http://" + urlText).replace(/^http:\/\//, "");
+ } else if (links[ln].schema === "mailto:" && !/^mailto:/i.test(urlText)) {
+ urlText = state.md.normalizeLinkText("mailto:" + urlText).replace(/^mailto:/, "");
+ } else {
+ urlText = state.md.normalizeLinkText(urlText);
+ }
+ pos = links[ln].index;
+ if (pos > lastPos) {
+ token = new state.Token("text", "", 0);
+ token.content = text.slice(lastPos, pos);
+ token.level = level;
+ nodes.push(token);
+ }
+ token = new state.Token("link_open", "a", 1);
+ token.attrs = [ [ "href", fullUrl ] ];
+ token.level = level++;
+ token.markup = "linkify";
+ token.info = "auto";
+ nodes.push(token);
+ token = new state.Token("text", "", 0);
+ token.content = urlText;
+ token.level = level;
+ nodes.push(token);
+ token = new state.Token("link_close", "a", -1);
+ token.level = --level;
+ token.markup = "linkify";
+ token.info = "auto";
+ nodes.push(token);
+ lastPos = links[ln].lastIndex;
+ }
+ if (lastPos < text.length) {
+ token = new state.Token("text", "", 0);
+ token.content = text.slice(lastPos);
+ token.level = level;
+ nodes.push(token);
+ }
+ // replace current node
+ blockTokens[j].children = tokens = arrayReplaceAt(tokens, i, nodes);
+ }
+ }
+ }
+ };
+ // Simple typographic replacements
+ // TODO:
+ // - fractionals 1/2, 1/4, 3/4 -> ½, ¼, ¾
+ // - miltiplication 2 x 4 -> 2 × 4
+ var RARE_RE = /\+-|\.\.|\?\?\?\?|!!!!|,,|--/;
+ // Workaround for phantomjs - need regex without /g flag,
+ // or root check will fail every second time
+ var SCOPED_ABBR_TEST_RE = /\((c|tm|r|p)\)/i;
+ var SCOPED_ABBR_RE = /\((c|tm|r|p)\)/gi;
+ var SCOPED_ABBR = {
+ c: "\xa9",
+ r: "\xae",
+ p: "\xa7",
+ tm: "\u2122"
+ };
+ function replaceFn(match, name) {
+ return SCOPED_ABBR[name.toLowerCase()];
+ }
+ function replace_scoped(inlineTokens) {
+ var i, token, inside_autolink = 0;
+ for (i = inlineTokens.length - 1; i >= 0; i--) {
+ token = inlineTokens[i];
+ if (token.type === "text" && !inside_autolink) {
+ token.content = token.content.replace(SCOPED_ABBR_RE, replaceFn);
+ }
+ if (token.type === "link_open" && token.info === "auto") {
+ inside_autolink--;
+ }
+ if (token.type === "link_close" && token.info === "auto") {
+ inside_autolink++;
+ }
+ }
+ }
+ function replace_rare(inlineTokens) {
+ var i, token, inside_autolink = 0;
+ for (i = inlineTokens.length - 1; i >= 0; i--) {
+ token = inlineTokens[i];
+ if (token.type === "text" && !inside_autolink) {
+ if (RARE_RE.test(token.content)) {
+ token.content = token.content.replace(/\+-/g, "\xb1").replace(/\.{2,}/g, "\u2026").replace(/([?!])\u2026/g, "$1..").replace(/([?!]){4,}/g, "$1$1$1").replace(/,{2,}/g, ",").replace(/(^|[^-])---(?=[^-]|$)/gm, "$1\u2014").replace(/(^|\s)--(?=\s|$)/gm, "$1\u2013").replace(/(^|[^-\s])--(?=[^-\s]|$)/gm, "$1\u2013");
+ }
+ }
+ if (token.type === "link_open" && token.info === "auto") {
+ inside_autolink--;
+ }
+ if (token.type === "link_close" && token.info === "auto") {
+ inside_autolink++;
+ }
+ }
+ }
+ var replacements = function replace(state) {
+ var blkIdx;
+ if (!state.md.options.typographer) {
+ return;
+ }
+ for (blkIdx = state.tokens.length - 1; blkIdx >= 0; blkIdx--) {
+ if (state.tokens[blkIdx].type !== "inline") {
+ continue;
+ }
+ if (SCOPED_ABBR_TEST_RE.test(state.tokens[blkIdx].content)) {
+ replace_scoped(state.tokens[blkIdx].children);
+ }
+ if (RARE_RE.test(state.tokens[blkIdx].content)) {
+ replace_rare(state.tokens[blkIdx].children);
+ }
+ }
+ };
+ var isWhiteSpace$1 = utils.isWhiteSpace;
+ var isPunctChar$1 = utils.isPunctChar;
+ var isMdAsciiPunct$1 = utils.isMdAsciiPunct;
+ var QUOTE_TEST_RE = /['"]/;
+ var QUOTE_RE = /['"]/g;
+ var APOSTROPHE = "\u2019";
+ /* ’ */ function replaceAt(str, index, ch) {
+ return str.substr(0, index) + ch + str.substr(index + 1);
+ }
+ function process_inlines(tokens, state) {
+ var i, token, text, t, pos, max, thisLevel, item, lastChar, nextChar, isLastPunctChar, isNextPunctChar, isLastWhiteSpace, isNextWhiteSpace, canOpen, canClose, j, isSingle, stack, openQuote, closeQuote;
+ stack = [];
+ for (i = 0; i < tokens.length; i++) {
+ token = tokens[i];
+ thisLevel = tokens[i].level;
+ for (j = stack.length - 1; j >= 0; j--) {
+ if (stack[j].level <= thisLevel) {
+ break;
+ }
+ }
+ stack.length = j + 1;
+ if (token.type !== "text") {
+ continue;
+ }
+ text = token.content;
+ pos = 0;
+ max = text.length;
+ /*eslint no-labels:0,block-scoped-var:0*/ OUTER: while (pos < max) {
+ QUOTE_RE.lastIndex = pos;
+ t = QUOTE_RE.exec(text);
+ if (!t) {
+ break;
+ }
+ canOpen = canClose = true;
+ pos = t.index + 1;
+ isSingle = t[0] === "'";
+ // Find previous character,
+ // default to space if it's the beginning of the line
+
+ lastChar = 32;
+ if (t.index - 1 >= 0) {
+ lastChar = text.charCodeAt(t.index - 1);
+ } else {
+ for (j = i - 1; j >= 0; j--) {
+ if (tokens[j].type === "softbreak" || tokens[j].type === "hardbreak") break;
+ // lastChar defaults to 0x20
+ if (!tokens[j].content) continue;
+ // should skip all tokens except 'text', 'html_inline' or 'code_inline'
+ lastChar = tokens[j].content.charCodeAt(tokens[j].content.length - 1);
+ break;
+ }
+ }
+ // Find next character,
+ // default to space if it's the end of the line
+
+ nextChar = 32;
+ if (pos < max) {
+ nextChar = text.charCodeAt(pos);
+ } else {
+ for (j = i + 1; j < tokens.length; j++) {
+ if (tokens[j].type === "softbreak" || tokens[j].type === "hardbreak") break;
+ // nextChar defaults to 0x20
+ if (!tokens[j].content) continue;
+ // should skip all tokens except 'text', 'html_inline' or 'code_inline'
+ nextChar = tokens[j].content.charCodeAt(0);
+ break;
+ }
+ }
+ isLastPunctChar = isMdAsciiPunct$1(lastChar) || isPunctChar$1(String.fromCharCode(lastChar));
+ isNextPunctChar = isMdAsciiPunct$1(nextChar) || isPunctChar$1(String.fromCharCode(nextChar));
+ isLastWhiteSpace = isWhiteSpace$1(lastChar);
+ isNextWhiteSpace = isWhiteSpace$1(nextChar);
+ if (isNextWhiteSpace) {
+ canOpen = false;
+ } else if (isNextPunctChar) {
+ if (!(isLastWhiteSpace || isLastPunctChar)) {
+ canOpen = false;
+ }
+ }
+ if (isLastWhiteSpace) {
+ canClose = false;
+ } else if (isLastPunctChar) {
+ if (!(isNextWhiteSpace || isNextPunctChar)) {
+ canClose = false;
+ }
+ }
+ if (nextChar === 34 /* " */ && t[0] === '"') {
+ if (lastChar >= 48 /* 0 */ && lastChar <= 57 /* 9 */) {
+ // special case: 1"" - count first quote as an inch
+ canClose = canOpen = false;
+ }
+ }
+ if (canOpen && canClose) {
+ // Replace quotes in the middle of punctuation sequence, but not
+ // in the middle of the words, i.e.:
+ // 1. foo " bar " baz - not replaced
+ // 2. foo-"-bar-"-baz - replaced
+ // 3. foo"bar"baz - not replaced
+ canOpen = isLastPunctChar;
+ canClose = isNextPunctChar;
+ }
+ if (!canOpen && !canClose) {
+ // middle of word
+ if (isSingle) {
+ token.content = replaceAt(token.content, t.index, APOSTROPHE);
+ }
+ continue;
+ }
+ if (canClose) {
+ // this could be a closing quote, rewind the stack to get a match
+ for (j = stack.length - 1; j >= 0; j--) {
+ item = stack[j];
+ if (stack[j].level < thisLevel) {
+ break;
+ }
+ if (item.single === isSingle && stack[j].level === thisLevel) {
+ item = stack[j];
+ if (isSingle) {
+ openQuote = state.md.options.quotes[2];
+ closeQuote = state.md.options.quotes[3];
+ } else {
+ openQuote = state.md.options.quotes[0];
+ closeQuote = state.md.options.quotes[1];
+ }
+ // replace token.content *before* tokens[item.token].content,
+ // because, if they are pointing at the same token, replaceAt
+ // could mess up indices when quote length != 1
+ token.content = replaceAt(token.content, t.index, closeQuote);
+ tokens[item.token].content = replaceAt(tokens[item.token].content, item.pos, openQuote);
+ pos += closeQuote.length - 1;
+ if (item.token === i) {
+ pos += openQuote.length - 1;
+ }
+ text = token.content;
+ max = text.length;
+ stack.length = j;
+ continue OUTER;
+ }
+ }
+ }
+ if (canOpen) {
+ stack.push({
+ token: i,
+ pos: t.index,
+ single: isSingle,
+ level: thisLevel
+ });
+ } else if (canClose && isSingle) {
+ token.content = replaceAt(token.content, t.index, APOSTROPHE);
+ }
+ }
+ }
+ }
+ var smartquotes = function smartquotes(state) {
+ /*eslint max-depth:0*/
+ var blkIdx;
+ if (!state.md.options.typographer) {
+ return;
+ }
+ for (blkIdx = state.tokens.length - 1; blkIdx >= 0; blkIdx--) {
+ if (state.tokens[blkIdx].type !== "inline" || !QUOTE_TEST_RE.test(state.tokens[blkIdx].content)) {
+ continue;
+ }
+ process_inlines(state.tokens[blkIdx].children, state);
+ }
+ };
+ // Token class
+ /**
+ * class Token
+ **/
+ /**
+ * new Token(type, tag, nesting)
+ *
+ * Create new token and fill passed properties.
+ **/ function Token(type, tag, nesting) {
+ /**
+ * Token#type -> String
+ *
+ * Type of the token (string, e.g. "paragraph_open")
+ **/
+ this.type = type;
+ /**
+ * Token#tag -> String
+ *
+ * html tag name, e.g. "p"
+ **/ this.tag = tag;
+ /**
+ * Token#attrs -> Array
+ *
+ * Html attributes. Format: `[ [ name1, value1 ], [ name2, value2 ] ]`
+ **/ this.attrs = null;
+ /**
+ * Token#map -> Array
+ *
+ * Source map info. Format: `[ line_begin, line_end ]`
+ **/ this.map = null;
+ /**
+ * Token#nesting -> Number
+ *
+ * Level change (number in {-1, 0, 1} set), where:
+ *
+ * - `1` means the tag is opening
+ * - `0` means the tag is self-closing
+ * - `-1` means the tag is closing
+ **/ this.nesting = nesting;
+ /**
+ * Token#level -> Number
+ *
+ * nesting level, the same as `state.level`
+ **/ this.level = 0;
+ /**
+ * Token#children -> Array
+ *
+ * An array of child nodes (inline and img tokens)
+ **/ this.children = null;
+ /**
+ * Token#content -> String
+ *
+ * In a case of self-closing tag (code, html, fence, etc.),
+ * it has contents of this tag.
+ **/ this.content = "";
+ /**
+ * Token#markup -> String
+ *
+ * '*' or '_' for emphasis, fence string for fence, etc.
+ **/ this.markup = "";
+ /**
+ * Token#info -> String
+ *
+ * Additional information:
+ *
+ * - Info string for "fence" tokens
+ * - The value "auto" for autolink "link_open" and "link_close" tokens
+ * - The string value of the item marker for ordered-list "list_item_open" tokens
+ **/ this.info = "";
+ /**
+ * Token#meta -> Object
+ *
+ * A place for plugins to store an arbitrary data
+ **/ this.meta = null;
+ /**
+ * Token#block -> Boolean
+ *
+ * True for block-level tokens, false for inline tokens.
+ * Used in renderer to calculate line breaks
+ **/ this.block = false;
+ /**
+ * Token#hidden -> Boolean
+ *
+ * If it's true, ignore this element when rendering. Used for tight lists
+ * to hide paragraphs.
+ **/ this.hidden = false;
+ }
+ /**
+ * Token.attrIndex(name) -> Number
+ *
+ * Search attribute index by name.
+ **/ Token.prototype.attrIndex = function attrIndex(name) {
+ var attrs, i, len;
+ if (!this.attrs) {
+ return -1;
+ }
+ attrs = this.attrs;
+ for (i = 0, len = attrs.length; i < len; i++) {
+ if (attrs[i][0] === name) {
+ return i;
+ }
+ }
+ return -1;
+ };
+ /**
+ * Token.attrPush(attrData)
+ *
+ * Add `[ name, value ]` attribute to list. Init attrs if necessary
+ **/ Token.prototype.attrPush = function attrPush(attrData) {
+ if (this.attrs) {
+ this.attrs.push(attrData);
+ } else {
+ this.attrs = [ attrData ];
+ }
+ };
+ /**
+ * Token.attrSet(name, value)
+ *
+ * Set `name` attribute to `value`. Override old value if exists.
+ **/ Token.prototype.attrSet = function attrSet(name, value) {
+ var idx = this.attrIndex(name), attrData = [ name, value ];
+ if (idx < 0) {
+ this.attrPush(attrData);
+ } else {
+ this.attrs[idx] = attrData;
+ }
+ };
+ /**
+ * Token.attrGet(name)
+ *
+ * Get the value of attribute `name`, or null if it does not exist.
+ **/ Token.prototype.attrGet = function attrGet(name) {
+ var idx = this.attrIndex(name), value = null;
+ if (idx >= 0) {
+ value = this.attrs[idx][1];
+ }
+ return value;
+ };
+ /**
+ * Token.attrJoin(name, value)
+ *
+ * Join value to existing attribute via space. Or create new attribute if not
+ * exists. Useful to operate with token classes.
+ **/ Token.prototype.attrJoin = function attrJoin(name, value) {
+ var idx = this.attrIndex(name);
+ if (idx < 0) {
+ this.attrPush([ name, value ]);
+ } else {
+ this.attrs[idx][1] = this.attrs[idx][1] + " " + value;
+ }
+ };
+ var token = Token;
+ function StateCore(src, md, env) {
+ this.src = src;
+ this.env = env;
+ this.tokens = [];
+ this.inlineMode = false;
+ this.md = md;
+ // link to parser instance
+ }
+ // re-export Token class to use in core rules
+ StateCore.prototype.Token = token;
+ var state_core = StateCore;
+ var _rules$2 = [ [ "normalize", normalize ], [ "block", block ], [ "inline", inline ], [ "linkify", linkify ], [ "replacements", replacements ], [ "smartquotes", smartquotes ] ];
+ /**
+ * new Core()
+ **/ function Core() {
+ /**
+ * Core#ruler -> Ruler
+ *
+ * [[Ruler]] instance. Keep configuration of core rules.
+ **/
+ this.ruler = new ruler;
+ for (var i = 0; i < _rules$2.length; i++) {
+ this.ruler.push(_rules$2[i][0], _rules$2[i][1]);
+ }
+ }
+ /**
+ * Core.process(state)
+ *
+ * Executes core chain rules.
+ **/ Core.prototype.process = function(state) {
+ var i, l, rules;
+ rules = this.ruler.getRules("");
+ for (i = 0, l = rules.length; i < l; i++) {
+ rules[i](state);
+ }
+ };
+ Core.prototype.State = state_core;
+ var parser_core = Core;
+ var isSpace$a = utils.isSpace;
+ function getLine(state, line) {
+ var pos = state.bMarks[line] + state.tShift[line], max = state.eMarks[line];
+ return state.src.substr(pos, max - pos);
+ }
+ function escapedSplit(str) {
+ var result = [], pos = 0, max = str.length, ch, isEscaped = false, lastPos = 0, current = "";
+ ch = str.charCodeAt(pos);
+ while (pos < max) {
+ if (ch === 124 /* | */) {
+ if (!isEscaped) {
+ // pipe separating cells, '|'
+ result.push(current + str.substring(lastPos, pos));
+ current = "";
+ lastPos = pos + 1;
+ } else {
+ // escaped pipe, '\|'
+ current += str.substring(lastPos, pos - 1);
+ lastPos = pos;
+ }
+ }
+ isEscaped = ch === 92 /* \ */;
+ pos++;
+ ch = str.charCodeAt(pos);
+ }
+ result.push(current + str.substring(lastPos));
+ return result;
+ }
+ var table = function table(state, startLine, endLine, silent) {
+ var ch, lineText, pos, i, l, nextLine, columns, columnCount, token, aligns, t, tableLines, tbodyLines, oldParentType, terminate, terminatorRules, firstCh, secondCh;
+ // should have at least two lines
+ if (startLine + 2 > endLine) {
+ return false;
+ }
+ nextLine = startLine + 1;
+ if (state.sCount[nextLine] < state.blkIndent) {
+ return false;
+ }
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[nextLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ // first character of the second line should be '|', '-', ':',
+ // and no other characters are allowed but spaces;
+ // basically, this is the equivalent of /^[-:|][-:|\s]*$/ regexp
+ pos = state.bMarks[nextLine] + state.tShift[nextLine];
+ if (pos >= state.eMarks[nextLine]) {
+ return false;
+ }
+ firstCh = state.src.charCodeAt(pos++);
+ if (firstCh !== 124 /* | */ && firstCh !== 45 /* - */ && firstCh !== 58 /* : */) {
+ return false;
+ }
+ if (pos >= state.eMarks[nextLine]) {
+ return false;
+ }
+ secondCh = state.src.charCodeAt(pos++);
+ if (secondCh !== 124 /* | */ && secondCh !== 45 /* - */ && secondCh !== 58 /* : */ && !isSpace$a(secondCh)) {
+ return false;
+ }
+ // if first character is '-', then second character must not be a space
+ // (due to parsing ambiguity with list)
+ if (firstCh === 45 /* - */ && isSpace$a(secondCh)) {
+ return false;
+ }
+ while (pos < state.eMarks[nextLine]) {
+ ch = state.src.charCodeAt(pos);
+ if (ch !== 124 /* | */ && ch !== 45 /* - */ && ch !== 58 /* : */ && !isSpace$a(ch)) {
+ return false;
+ }
+ pos++;
+ }
+ lineText = getLine(state, startLine + 1);
+ columns = lineText.split("|");
+ aligns = [];
+ for (i = 0; i < columns.length; i++) {
+ t = columns[i].trim();
+ if (!t) {
+ // allow empty columns before and after table, but not in between columns;
+ // e.g. allow ` |---| `, disallow ` ---||--- `
+ if (i === 0 || i === columns.length - 1) {
+ continue;
+ } else {
+ return false;
+ }
+ }
+ if (!/^:?-+:?$/.test(t)) {
+ return false;
+ }
+ if (t.charCodeAt(t.length - 1) === 58 /* : */) {
+ aligns.push(t.charCodeAt(0) === 58 /* : */ ? "center" : "right");
+ } else if (t.charCodeAt(0) === 58 /* : */) {
+ aligns.push("left");
+ } else {
+ aligns.push("");
+ }
+ }
+ lineText = getLine(state, startLine).trim();
+ if (lineText.indexOf("|") === -1) {
+ return false;
+ }
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ columns = escapedSplit(lineText);
+ if (columns.length && columns[0] === "") columns.shift();
+ if (columns.length && columns[columns.length - 1] === "") columns.pop();
+ // header row will define an amount of columns in the entire table,
+ // and align row should be exactly the same (the rest of the rows can differ)
+ columnCount = columns.length;
+ if (columnCount === 0 || columnCount !== aligns.length) {
+ return false;
+ }
+ if (silent) {
+ return true;
+ }
+ oldParentType = state.parentType;
+ state.parentType = "table";
+ // use 'blockquote' lists for termination because it's
+ // the most similar to tables
+ terminatorRules = state.md.block.ruler.getRules("blockquote");
+ token = state.push("table_open", "table", 1);
+ token.map = tableLines = [ startLine, 0 ];
+ token = state.push("thead_open", "thead", 1);
+ token.map = [ startLine, startLine + 1 ];
+ token = state.push("tr_open", "tr", 1);
+ token.map = [ startLine, startLine + 1 ];
+ for (i = 0; i < columns.length; i++) {
+ token = state.push("th_open", "th", 1);
+ if (aligns[i]) {
+ token.attrs = [ [ "style", "text-align:" + aligns[i] ] ];
+ }
+ token = state.push("inline", "", 0);
+ token.content = columns[i].trim();
+ token.children = [];
+ token = state.push("th_close", "th", -1);
+ }
+ token = state.push("tr_close", "tr", -1);
+ token = state.push("thead_close", "thead", -1);
+ for (nextLine = startLine + 2; nextLine < endLine; nextLine++) {
+ if (state.sCount[nextLine] < state.blkIndent) {
+ break;
+ }
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ break;
+ }
+ lineText = getLine(state, nextLine).trim();
+ if (!lineText) {
+ break;
+ }
+ if (state.sCount[nextLine] - state.blkIndent >= 4) {
+ break;
+ }
+ columns = escapedSplit(lineText);
+ if (columns.length && columns[0] === "") columns.shift();
+ if (columns.length && columns[columns.length - 1] === "") columns.pop();
+ if (nextLine === startLine + 2) {
+ token = state.push("tbody_open", "tbody", 1);
+ token.map = tbodyLines = [ startLine + 2, 0 ];
+ }
+ token = state.push("tr_open", "tr", 1);
+ token.map = [ nextLine, nextLine + 1 ];
+ for (i = 0; i < columnCount; i++) {
+ token = state.push("td_open", "td", 1);
+ if (aligns[i]) {
+ token.attrs = [ [ "style", "text-align:" + aligns[i] ] ];
+ }
+ token = state.push("inline", "", 0);
+ token.content = columns[i] ? columns[i].trim() : "";
+ token.children = [];
+ token = state.push("td_close", "td", -1);
+ }
+ token = state.push("tr_close", "tr", -1);
+ }
+ if (tbodyLines) {
+ token = state.push("tbody_close", "tbody", -1);
+ tbodyLines[1] = nextLine;
+ }
+ token = state.push("table_close", "table", -1);
+ tableLines[1] = nextLine;
+ state.parentType = oldParentType;
+ state.line = nextLine;
+ return true;
+ };
+ // Code block (4 spaces padded)
+ var code = function code(state, startLine, endLine /*, silent*/) {
+ var nextLine, last, token;
+ if (state.sCount[startLine] - state.blkIndent < 4) {
+ return false;
+ }
+ last = nextLine = startLine + 1;
+ while (nextLine < endLine) {
+ if (state.isEmpty(nextLine)) {
+ nextLine++;
+ continue;
+ }
+ if (state.sCount[nextLine] - state.blkIndent >= 4) {
+ nextLine++;
+ last = nextLine;
+ continue;
+ }
+ break;
+ }
+ state.line = last;
+ token = state.push("code_block", "code", 0);
+ token.content = state.getLines(startLine, last, 4 + state.blkIndent, false) + "\n";
+ token.map = [ startLine, state.line ];
+ return true;
+ };
+ // fences (``` lang, ~~~ lang)
+ var fence = function fence(state, startLine, endLine, silent) {
+ var marker, len, params, nextLine, mem, token, markup, haveEndMarker = false, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine];
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ if (pos + 3 > max) {
+ return false;
+ }
+ marker = state.src.charCodeAt(pos);
+ if (marker !== 126 /* ~ */ && marker !== 96 /* ` */) {
+ return false;
+ }
+ // scan marker length
+ mem = pos;
+ pos = state.skipChars(pos, marker);
+ len = pos - mem;
+ if (len < 3) {
+ return false;
+ }
+ markup = state.src.slice(mem, pos);
+ params = state.src.slice(pos, max);
+ if (marker === 96 /* ` */) {
+ if (params.indexOf(String.fromCharCode(marker)) >= 0) {
+ return false;
+ }
+ }
+ // Since start is found, we can report success here in validation mode
+ if (silent) {
+ return true;
+ }
+ // search end of block
+ nextLine = startLine;
+ for (;;) {
+ nextLine++;
+ if (nextLine >= endLine) {
+ // unclosed block should be autoclosed by end of document.
+ // also block seems to be autoclosed by end of parent
+ break;
+ }
+ pos = mem = state.bMarks[nextLine] + state.tShift[nextLine];
+ max = state.eMarks[nextLine];
+ if (pos < max && state.sCount[nextLine] < state.blkIndent) {
+ // non-empty line with negative indent should stop the list:
+ // - ```
+ // test
+ break;
+ }
+ if (state.src.charCodeAt(pos) !== marker) {
+ continue;
+ }
+ if (state.sCount[nextLine] - state.blkIndent >= 4) {
+ // closing fence should be indented less than 4 spaces
+ continue;
+ }
+ pos = state.skipChars(pos, marker);
+ // closing code fence must be at least as long as the opening one
+ if (pos - mem < len) {
+ continue;
+ }
+ // make sure tail has spaces only
+ pos = state.skipSpaces(pos);
+ if (pos < max) {
+ continue;
+ }
+ haveEndMarker = true;
+ // found!
+ break;
+ }
+ // If a fence has heading spaces, they should be removed from its inner block
+ len = state.sCount[startLine];
+ state.line = nextLine + (haveEndMarker ? 1 : 0);
+ token = state.push("fence", "code", 0);
+ token.info = params;
+ token.content = state.getLines(startLine + 1, nextLine, len, true);
+ token.markup = markup;
+ token.map = [ startLine, state.line ];
+ return true;
+ };
+ var isSpace$9 = utils.isSpace;
+ var blockquote = function blockquote(state, startLine, endLine, silent) {
+ var adjustTab, ch, i, initial, l, lastLineEmpty, lines, nextLine, offset, oldBMarks, oldBSCount, oldIndent, oldParentType, oldSCount, oldTShift, spaceAfterMarker, terminate, terminatorRules, token, isOutdented, oldLineMax = state.lineMax, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine];
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ // check the block quote marker
+ if (state.src.charCodeAt(pos++) !== 62 /* > */) {
+ return false;
+ }
+ // we know that it's going to be a valid blockquote,
+ // so no point trying to find the end of it in silent mode
+ if (silent) {
+ return true;
+ }
+ // set offset past spaces and ">"
+ initial = offset = state.sCount[startLine] + 1;
+ // skip one optional space after '>'
+ if (state.src.charCodeAt(pos) === 32 /* space */) {
+ // ' > test '
+ // ^ -- position start of line here:
+ pos++;
+ initial++;
+ offset++;
+ adjustTab = false;
+ spaceAfterMarker = true;
+ } else if (state.src.charCodeAt(pos) === 9 /* tab */) {
+ spaceAfterMarker = true;
+ if ((state.bsCount[startLine] + offset) % 4 === 3) {
+ // ' >\t test '
+ // ^ -- position start of line here (tab has width===1)
+ pos++;
+ initial++;
+ offset++;
+ adjustTab = false;
+ } else {
+ // ' >\t test '
+ // ^ -- position start of line here + shift bsCount slightly
+ // to make extra space appear
+ adjustTab = true;
+ }
+ } else {
+ spaceAfterMarker = false;
+ }
+ oldBMarks = [ state.bMarks[startLine] ];
+ state.bMarks[startLine] = pos;
+ while (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (isSpace$9(ch)) {
+ if (ch === 9) {
+ offset += 4 - (offset + state.bsCount[startLine] + (adjustTab ? 1 : 0)) % 4;
+ } else {
+ offset++;
+ }
+ } else {
+ break;
+ }
+ pos++;
+ }
+ oldBSCount = [ state.bsCount[startLine] ];
+ state.bsCount[startLine] = state.sCount[startLine] + 1 + (spaceAfterMarker ? 1 : 0);
+ lastLineEmpty = pos >= max;
+ oldSCount = [ state.sCount[startLine] ];
+ state.sCount[startLine] = offset - initial;
+ oldTShift = [ state.tShift[startLine] ];
+ state.tShift[startLine] = pos - state.bMarks[startLine];
+ terminatorRules = state.md.block.ruler.getRules("blockquote");
+ oldParentType = state.parentType;
+ state.parentType = "blockquote";
+ // Search the end of the block
+
+ // Block ends with either:
+ // 1. an empty line outside:
+ // ```
+ // > test
+
+ // ```
+ // 2. an empty line inside:
+ // ```
+ // >
+ // test
+ // ```
+ // 3. another tag:
+ // ```
+ // > test
+ // - - -
+ // ```
+ for (nextLine = startLine + 1; nextLine < endLine; nextLine++) {
+ // check if it's outdented, i.e. it's inside list item and indented
+ // less than said list item:
+ // ```
+ // 1. anything
+ // > current blockquote
+ // 2. checking this line
+ // ```
+ isOutdented = state.sCount[nextLine] < state.blkIndent;
+ pos = state.bMarks[nextLine] + state.tShift[nextLine];
+ max = state.eMarks[nextLine];
+ if (pos >= max) {
+ // Case 1: line is not inside the blockquote, and this line is empty.
+ break;
+ }
+ if (state.src.charCodeAt(pos++) === 62 /* > */ && !isOutdented) {
+ // This line is inside the blockquote.
+ // set offset past spaces and ">"
+ initial = offset = state.sCount[nextLine] + 1;
+ // skip one optional space after '>'
+ if (state.src.charCodeAt(pos) === 32 /* space */) {
+ // ' > test '
+ // ^ -- position start of line here:
+ pos++;
+ initial++;
+ offset++;
+ adjustTab = false;
+ spaceAfterMarker = true;
+ } else if (state.src.charCodeAt(pos) === 9 /* tab */) {
+ spaceAfterMarker = true;
+ if ((state.bsCount[nextLine] + offset) % 4 === 3) {
+ // ' >\t test '
+ // ^ -- position start of line here (tab has width===1)
+ pos++;
+ initial++;
+ offset++;
+ adjustTab = false;
+ } else {
+ // ' >\t test '
+ // ^ -- position start of line here + shift bsCount slightly
+ // to make extra space appear
+ adjustTab = true;
+ }
+ } else {
+ spaceAfterMarker = false;
+ }
+ oldBMarks.push(state.bMarks[nextLine]);
+ state.bMarks[nextLine] = pos;
+ while (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (isSpace$9(ch)) {
+ if (ch === 9) {
+ offset += 4 - (offset + state.bsCount[nextLine] + (adjustTab ? 1 : 0)) % 4;
+ } else {
+ offset++;
+ }
+ } else {
+ break;
+ }
+ pos++;
+ }
+ lastLineEmpty = pos >= max;
+ oldBSCount.push(state.bsCount[nextLine]);
+ state.bsCount[nextLine] = state.sCount[nextLine] + 1 + (spaceAfterMarker ? 1 : 0);
+ oldSCount.push(state.sCount[nextLine]);
+ state.sCount[nextLine] = offset - initial;
+ oldTShift.push(state.tShift[nextLine]);
+ state.tShift[nextLine] = pos - state.bMarks[nextLine];
+ continue;
+ }
+ // Case 2: line is not inside the blockquote, and the last line was empty.
+ if (lastLineEmpty) {
+ break;
+ }
+ // Case 3: another tag found.
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ // Quirk to enforce "hard termination mode" for paragraphs;
+ // normally if you call `tokenize(state, startLine, nextLine)`,
+ // paragraphs will look below nextLine for paragraph continuation,
+ // but if blockquote is terminated by another tag, they shouldn't
+ state.lineMax = nextLine;
+ if (state.blkIndent !== 0) {
+ // state.blkIndent was non-zero, we now set it to zero,
+ // so we need to re-calculate all offsets to appear as
+ // if indent wasn't changed
+ oldBMarks.push(state.bMarks[nextLine]);
+ oldBSCount.push(state.bsCount[nextLine]);
+ oldTShift.push(state.tShift[nextLine]);
+ oldSCount.push(state.sCount[nextLine]);
+ state.sCount[nextLine] -= state.blkIndent;
+ }
+ break;
+ }
+ oldBMarks.push(state.bMarks[nextLine]);
+ oldBSCount.push(state.bsCount[nextLine]);
+ oldTShift.push(state.tShift[nextLine]);
+ oldSCount.push(state.sCount[nextLine]);
+ // A negative indentation means that this is a paragraph continuation
+
+ state.sCount[nextLine] = -1;
+ }
+ oldIndent = state.blkIndent;
+ state.blkIndent = 0;
+ token = state.push("blockquote_open", "blockquote", 1);
+ token.markup = ">";
+ token.map = lines = [ startLine, 0 ];
+ state.md.block.tokenize(state, startLine, nextLine);
+ token = state.push("blockquote_close", "blockquote", -1);
+ token.markup = ">";
+ state.lineMax = oldLineMax;
+ state.parentType = oldParentType;
+ lines[1] = state.line;
+ // Restore original tShift; this might not be necessary since the parser
+ // has already been here, but just to make sure we can do that.
+ for (i = 0; i < oldTShift.length; i++) {
+ state.bMarks[i + startLine] = oldBMarks[i];
+ state.tShift[i + startLine] = oldTShift[i];
+ state.sCount[i + startLine] = oldSCount[i];
+ state.bsCount[i + startLine] = oldBSCount[i];
+ }
+ state.blkIndent = oldIndent;
+ return true;
+ };
+ var isSpace$8 = utils.isSpace;
+ var hr = function hr(state, startLine, endLine, silent) {
+ var marker, cnt, ch, token, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine];
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ marker = state.src.charCodeAt(pos++);
+ // Check hr marker
+ if (marker !== 42 /* * */ && marker !== 45 /* - */ && marker !== 95 /* _ */) {
+ return false;
+ }
+ // markers can be mixed with spaces, but there should be at least 3 of them
+ cnt = 1;
+ while (pos < max) {
+ ch = state.src.charCodeAt(pos++);
+ if (ch !== marker && !isSpace$8(ch)) {
+ return false;
+ }
+ if (ch === marker) {
+ cnt++;
+ }
+ }
+ if (cnt < 3) {
+ return false;
+ }
+ if (silent) {
+ return true;
+ }
+ state.line = startLine + 1;
+ token = state.push("hr", "hr", 0);
+ token.map = [ startLine, state.line ];
+ token.markup = Array(cnt + 1).join(String.fromCharCode(marker));
+ return true;
+ };
+ var isSpace$7 = utils.isSpace;
+ // Search `[-+*][\n ]`, returns next pos after marker on success
+ // or -1 on fail.
+ function skipBulletListMarker(state, startLine) {
+ var marker, pos, max, ch;
+ pos = state.bMarks[startLine] + state.tShift[startLine];
+ max = state.eMarks[startLine];
+ marker = state.src.charCodeAt(pos++);
+ // Check bullet
+ if (marker !== 42 /* * */ && marker !== 45 /* - */ && marker !== 43 /* + */) {
+ return -1;
+ }
+ if (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (!isSpace$7(ch)) {
+ // " -test " - is not a list item
+ return -1;
+ }
+ }
+ return pos;
+ }
+ // Search `\d+[.)][\n ]`, returns next pos after marker on success
+ // or -1 on fail.
+ function skipOrderedListMarker(state, startLine) {
+ var ch, start = state.bMarks[startLine] + state.tShift[startLine], pos = start, max = state.eMarks[startLine];
+ // List marker should have at least 2 chars (digit + dot)
+ if (pos + 1 >= max) {
+ return -1;
+ }
+ ch = state.src.charCodeAt(pos++);
+ if (ch < 48 /* 0 */ || ch > 57 /* 9 */) {
+ return -1;
+ }
+ for (;;) {
+ // EOL -> fail
+ if (pos >= max) {
+ return -1;
+ }
+ ch = state.src.charCodeAt(pos++);
+ if (ch >= 48 /* 0 */ && ch <= 57 /* 9 */) {
+ // List marker should have no more than 9 digits
+ // (prevents integer overflow in browsers)
+ if (pos - start >= 10) {
+ return -1;
+ }
+ continue;
+ }
+ // found valid marker
+ if (ch === 41 /* ) */ || ch === 46 /* . */) {
+ break;
+ }
+ return -1;
+ }
+ if (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (!isSpace$7(ch)) {
+ // " 1.test " - is not a list item
+ return -1;
+ }
+ }
+ return pos;
+ }
+ function markTightParagraphs(state, idx) {
+ var i, l, level = state.level + 2;
+ for (i = idx + 2, l = state.tokens.length - 2; i < l; i++) {
+ if (state.tokens[i].level === level && state.tokens[i].type === "paragraph_open") {
+ state.tokens[i + 2].hidden = true;
+ state.tokens[i].hidden = true;
+ i += 2;
+ }
+ }
+ }
+ var list = function list(state, startLine, endLine, silent) {
+ var ch, contentStart, i, indent, indentAfterMarker, initial, isOrdered, itemLines, l, listLines, listTokIdx, markerCharCode, markerValue, max, nextLine, offset, oldListIndent, oldParentType, oldSCount, oldTShift, oldTight, pos, posAfterMarker, prevEmptyEnd, start, terminate, terminatorRules, token, isTerminatingParagraph = false, tight = true;
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ // Special case:
+ // - item 1
+ // - item 2
+ // - item 3
+ // - item 4
+ // - this one is a paragraph continuation
+ if (state.listIndent >= 0 && state.sCount[startLine] - state.listIndent >= 4 && state.sCount[startLine] < state.blkIndent) {
+ return false;
+ }
+ // limit conditions when list can interrupt
+ // a paragraph (validation mode only)
+ if (silent && state.parentType === "paragraph") {
+ // Next list item should still terminate previous list item;
+ // This code can fail if plugins use blkIndent as well as lists,
+ // but I hope the spec gets fixed long before that happens.
+ if (state.sCount[startLine] >= state.blkIndent) {
+ isTerminatingParagraph = true;
+ }
+ }
+ // Detect list type and position after marker
+ if ((posAfterMarker = skipOrderedListMarker(state, startLine)) >= 0) {
+ isOrdered = true;
+ start = state.bMarks[startLine] + state.tShift[startLine];
+ markerValue = Number(state.src.slice(start, posAfterMarker - 1));
+ // If we're starting a new ordered list right after
+ // a paragraph, it should start with 1.
+ if (isTerminatingParagraph && markerValue !== 1) return false;
+ } else if ((posAfterMarker = skipBulletListMarker(state, startLine)) >= 0) {
+ isOrdered = false;
+ } else {
+ return false;
+ }
+ // If we're starting a new unordered list right after
+ // a paragraph, first line should not be empty.
+ if (isTerminatingParagraph) {
+ if (state.skipSpaces(posAfterMarker) >= state.eMarks[startLine]) return false;
+ }
+ // We should terminate list on style change. Remember first one to compare.
+ markerCharCode = state.src.charCodeAt(posAfterMarker - 1);
+ // For validation mode we can terminate immediately
+ if (silent) {
+ return true;
+ }
+ // Start list
+ listTokIdx = state.tokens.length;
+ if (isOrdered) {
+ token = state.push("ordered_list_open", "ol", 1);
+ if (markerValue !== 1) {
+ token.attrs = [ [ "start", markerValue ] ];
+ }
+ } else {
+ token = state.push("bullet_list_open", "ul", 1);
+ }
+ token.map = listLines = [ startLine, 0 ];
+ token.markup = String.fromCharCode(markerCharCode);
+
+ // Iterate list items
+
+ nextLine = startLine;
+ prevEmptyEnd = false;
+ terminatorRules = state.md.block.ruler.getRules("list");
+ oldParentType = state.parentType;
+ state.parentType = "list";
+ while (nextLine < endLine) {
+ pos = posAfterMarker;
+ max = state.eMarks[nextLine];
+ initial = offset = state.sCount[nextLine] + posAfterMarker - (state.bMarks[startLine] + state.tShift[startLine]);
+ while (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (ch === 9) {
+ offset += 4 - (offset + state.bsCount[nextLine]) % 4;
+ } else if (ch === 32) {
+ offset++;
+ } else {
+ break;
+ }
+ pos++;
+ }
+ contentStart = pos;
+ if (contentStart >= max) {
+ // trimming space in "- \n 3" case, indent is 1 here
+ indentAfterMarker = 1;
+ } else {
+ indentAfterMarker = offset - initial;
+ }
+ // If we have more than 4 spaces, the indent is 1
+ // (the rest is just indented code block)
+ if (indentAfterMarker > 4) {
+ indentAfterMarker = 1;
+ }
+ // " - test"
+ // ^^^^^ - calculating total length of this thing
+ indent = initial + indentAfterMarker;
+ // Run subparser & write tokens
+ token = state.push("list_item_open", "li", 1);
+ token.markup = String.fromCharCode(markerCharCode);
+ token.map = itemLines = [ startLine, 0 ];
+ if (isOrdered) {
+ token.info = state.src.slice(start, posAfterMarker - 1);
+ }
+ // change current state, then restore it after parser subcall
+ oldTight = state.tight;
+ oldTShift = state.tShift[startLine];
+ oldSCount = state.sCount[startLine];
+ // - example list
+ // ^ listIndent position will be here
+ // ^ blkIndent position will be here
+
+ oldListIndent = state.listIndent;
+ state.listIndent = state.blkIndent;
+ state.blkIndent = indent;
+ state.tight = true;
+ state.tShift[startLine] = contentStart - state.bMarks[startLine];
+ state.sCount[startLine] = offset;
+ if (contentStart >= max && state.isEmpty(startLine + 1)) {
+ // workaround for this case
+ // (list item is empty, list terminates before "foo"):
+ // ~~~~~~~~
+ // -
+ // foo
+ // ~~~~~~~~
+ state.line = Math.min(state.line + 2, endLine);
+ } else {
+ state.md.block.tokenize(state, startLine, endLine, true);
+ }
+ // If any of list item is tight, mark list as tight
+ if (!state.tight || prevEmptyEnd) {
+ tight = false;
+ }
+ // Item become loose if finish with empty line,
+ // but we should filter last element, because it means list finish
+ prevEmptyEnd = state.line - startLine > 1 && state.isEmpty(state.line - 1);
+ state.blkIndent = state.listIndent;
+ state.listIndent = oldListIndent;
+ state.tShift[startLine] = oldTShift;
+ state.sCount[startLine] = oldSCount;
+ state.tight = oldTight;
+ token = state.push("list_item_close", "li", -1);
+ token.markup = String.fromCharCode(markerCharCode);
+ nextLine = startLine = state.line;
+ itemLines[1] = nextLine;
+ contentStart = state.bMarks[startLine];
+ if (nextLine >= endLine) {
+ break;
+ }
+
+ // Try to check if list is terminated or continued.
+
+ if (state.sCount[nextLine] < state.blkIndent) {
+ break;
+ }
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ break;
+ }
+ // fail if terminating block found
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ break;
+ }
+ // fail if list has another type
+ if (isOrdered) {
+ posAfterMarker = skipOrderedListMarker(state, nextLine);
+ if (posAfterMarker < 0) {
+ break;
+ }
+ start = state.bMarks[nextLine] + state.tShift[nextLine];
+ } else {
+ posAfterMarker = skipBulletListMarker(state, nextLine);
+ if (posAfterMarker < 0) {
+ break;
+ }
+ }
+ if (markerCharCode !== state.src.charCodeAt(posAfterMarker - 1)) {
+ break;
+ }
+ }
+ // Finalize list
+ if (isOrdered) {
+ token = state.push("ordered_list_close", "ol", -1);
+ } else {
+ token = state.push("bullet_list_close", "ul", -1);
+ }
+ token.markup = String.fromCharCode(markerCharCode);
+ listLines[1] = nextLine;
+ state.line = nextLine;
+ state.parentType = oldParentType;
+ // mark paragraphs tight if needed
+ if (tight) {
+ markTightParagraphs(state, listTokIdx);
+ }
+ return true;
+ };
+ var normalizeReference$2 = utils.normalizeReference;
+ var isSpace$6 = utils.isSpace;
+ var reference = function reference(state, startLine, _endLine, silent) {
+ var ch, destEndPos, destEndLineNo, endLine, href, i, l, label, labelEnd, oldParentType, res, start, str, terminate, terminatorRules, title, lines = 0, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine], nextLine = startLine + 1;
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ if (state.src.charCodeAt(pos) !== 91 /* [ */) {
+ return false;
+ }
+ // Simple check to quickly interrupt scan on [link](url) at the start of line.
+ // Can be useful on practice: https://github.com/markdown-it/markdown-it/issues/54
+ while (++pos < max) {
+ if (state.src.charCodeAt(pos) === 93 /* ] */ && state.src.charCodeAt(pos - 1) !== 92 /* \ */) {
+ if (pos + 1 === max) {
+ return false;
+ }
+ if (state.src.charCodeAt(pos + 1) !== 58 /* : */) {
+ return false;
+ }
+ break;
+ }
+ }
+ endLine = state.lineMax;
+ // jump line-by-line until empty one or EOF
+ terminatorRules = state.md.block.ruler.getRules("reference");
+ oldParentType = state.parentType;
+ state.parentType = "reference";
+ for (;nextLine < endLine && !state.isEmpty(nextLine); nextLine++) {
+ // this would be a code block normally, but after paragraph
+ // it's considered a lazy continuation regardless of what's there
+ if (state.sCount[nextLine] - state.blkIndent > 3) {
+ continue;
+ }
+ // quirk for blockquotes, this line should already be checked by that rule
+ if (state.sCount[nextLine] < 0) {
+ continue;
+ }
+ // Some tags can terminate paragraph without empty line.
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ break;
+ }
+ }
+ str = state.getLines(startLine, nextLine, state.blkIndent, false).trim();
+ max = str.length;
+ for (pos = 1; pos < max; pos++) {
+ ch = str.charCodeAt(pos);
+ if (ch === 91 /* [ */) {
+ return false;
+ } else if (ch === 93 /* ] */) {
+ labelEnd = pos;
+ break;
+ } else if (ch === 10 /* \n */) {
+ lines++;
+ } else if (ch === 92 /* \ */) {
+ pos++;
+ if (pos < max && str.charCodeAt(pos) === 10) {
+ lines++;
+ }
+ }
+ }
+ if (labelEnd < 0 || str.charCodeAt(labelEnd + 1) !== 58 /* : */) {
+ return false;
+ }
+ // [label]: destination 'title'
+ // ^^^ skip optional whitespace here
+ for (pos = labelEnd + 2; pos < max; pos++) {
+ ch = str.charCodeAt(pos);
+ if (ch === 10) {
+ lines++;
+ } else if (isSpace$6(ch)) ; else {
+ break;
+ }
+ }
+ // [label]: destination 'title'
+ // ^^^^^^^^^^^ parse this
+ res = state.md.helpers.parseLinkDestination(str, pos, max);
+ if (!res.ok) {
+ return false;
+ }
+ href = state.md.normalizeLink(res.str);
+ if (!state.md.validateLink(href)) {
+ return false;
+ }
+ pos = res.pos;
+ lines += res.lines;
+ // save cursor state, we could require to rollback later
+ destEndPos = pos;
+ destEndLineNo = lines;
+ // [label]: destination 'title'
+ // ^^^ skipping those spaces
+ start = pos;
+ for (;pos < max; pos++) {
+ ch = str.charCodeAt(pos);
+ if (ch === 10) {
+ lines++;
+ } else if (isSpace$6(ch)) ; else {
+ break;
+ }
+ }
+ // [label]: destination 'title'
+ // ^^^^^^^ parse this
+ res = state.md.helpers.parseLinkTitle(str, pos, max);
+ if (pos < max && start !== pos && res.ok) {
+ title = res.str;
+ pos = res.pos;
+ lines += res.lines;
+ } else {
+ title = "";
+ pos = destEndPos;
+ lines = destEndLineNo;
+ }
+ // skip trailing spaces until the rest of the line
+ while (pos < max) {
+ ch = str.charCodeAt(pos);
+ if (!isSpace$6(ch)) {
+ break;
+ }
+ pos++;
+ }
+ if (pos < max && str.charCodeAt(pos) !== 10) {
+ if (title) {
+ // garbage at the end of the line after title,
+ // but it could still be a valid reference if we roll back
+ title = "";
+ pos = destEndPos;
+ lines = destEndLineNo;
+ while (pos < max) {
+ ch = str.charCodeAt(pos);
+ if (!isSpace$6(ch)) {
+ break;
+ }
+ pos++;
+ }
+ }
+ }
+ if (pos < max && str.charCodeAt(pos) !== 10) {
+ // garbage at the end of the line
+ return false;
+ }
+ label = normalizeReference$2(str.slice(1, labelEnd));
+ if (!label) {
+ // CommonMark 0.20 disallows empty labels
+ return false;
+ }
+ // Reference can not terminate anything. This check is for safety only.
+ /*istanbul ignore if*/ if (silent) {
+ return true;
+ }
+ if (typeof state.env.references === "undefined") {
+ state.env.references = {};
+ }
+ if (typeof state.env.references[label] === "undefined") {
+ state.env.references[label] = {
+ title: title,
+ href: href
+ };
+ }
+ state.parentType = oldParentType;
+ state.line = startLine + lines + 1;
+ return true;
+ };
+ // List of valid html blocks names, accorting to commonmark spec
+ var html_blocks = [ "address", "article", "aside", "base", "basefont", "blockquote", "body", "caption", "center", "col", "colgroup", "dd", "details", "dialog", "dir", "div", "dl", "dt", "fieldset", "figcaption", "figure", "footer", "form", "frame", "frameset", "h1", "h2", "h3", "h4", "h5", "h6", "head", "header", "hr", "html", "iframe", "legend", "li", "link", "main", "menu", "menuitem", "nav", "noframes", "ol", "optgroup", "option", "p", "param", "section", "source", "summary", "table", "tbody", "td", "tfoot", "th", "thead", "title", "tr", "track", "ul" ];
+ // Regexps to match html elements
+ var attr_name = "[a-zA-Z_:][a-zA-Z0-9:._-]*";
+ var unquoted = "[^\"'=<>`\\x00-\\x20]+";
+ var single_quoted = "'[^']*'";
+ var double_quoted = '"[^"]*"';
+ var attr_value = "(?:" + unquoted + "|" + single_quoted + "|" + double_quoted + ")";
+ var attribute = "(?:\\s+" + attr_name + "(?:\\s*=\\s*" + attr_value + ")?)";
+ var open_tag = "<[A-Za-z][A-Za-z0-9\\-]*" + attribute + "*\\s*\\/?>";
+ var close_tag = "<\\/[A-Za-z][A-Za-z0-9\\-]*\\s*>";
+ var comment = "\x3c!----\x3e|\x3c!--(?:-?[^>-])(?:-?[^-])*--\x3e";
+ var processing = "<[?][\\s\\S]*?[?]>";
+ var declaration = "<![A-Z]+\\s+[^>]*>";
+ var cdata = "<!\\[CDATA\\[[\\s\\S]*?\\]\\]>";
+ var HTML_TAG_RE$1 = new RegExp("^(?:" + open_tag + "|" + close_tag + "|" + comment + "|" + processing + "|" + declaration + "|" + cdata + ")");
+ var HTML_OPEN_CLOSE_TAG_RE$1 = new RegExp("^(?:" + open_tag + "|" + close_tag + ")");
+ var HTML_TAG_RE_1 = HTML_TAG_RE$1;
+ var HTML_OPEN_CLOSE_TAG_RE_1 = HTML_OPEN_CLOSE_TAG_RE$1;
+ var html_re = {
+ HTML_TAG_RE: HTML_TAG_RE_1,
+ HTML_OPEN_CLOSE_TAG_RE: HTML_OPEN_CLOSE_TAG_RE_1
+ };
+ var HTML_OPEN_CLOSE_TAG_RE = html_re.HTML_OPEN_CLOSE_TAG_RE;
+ // An array of opening and corresponding closing sequences for html tags,
+ // last argument defines whether it can terminate a paragraph or not
+
+ var HTML_SEQUENCES = [ [ /^<(script|pre|style|textarea)(?=(\s|>|$))/i, /<\/(script|pre|style|textarea)>/i, true ], [ /^<!--/, /-->/, true ], [ /^<\?/, /\?>/, true ], [ /^<![A-Z]/, />/, true ], [ /^<!\[CDATA\[/, /\]\]>/, true ], [ new RegExp("^</?(" + html_blocks.join("|") + ")(?=(\\s|/?>|$))", "i"), /^$/, true ], [ new RegExp(HTML_OPEN_CLOSE_TAG_RE.source + "\\s*$"), /^$/, false ] ];
+ var html_block = function html_block(state, startLine, endLine, silent) {
+ var i, nextLine, token, lineText, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine];
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ if (!state.md.options.html) {
+ return false;
+ }
+ if (state.src.charCodeAt(pos) !== 60 /* < */) {
+ return false;
+ }
+ lineText = state.src.slice(pos, max);
+ for (i = 0; i < HTML_SEQUENCES.length; i++) {
+ if (HTML_SEQUENCES[i][0].test(lineText)) {
+ break;
+ }
+ }
+ if (i === HTML_SEQUENCES.length) {
+ return false;
+ }
+ if (silent) {
+ // true if this sequence can be a terminator, false otherwise
+ return HTML_SEQUENCES[i][2];
+ }
+ nextLine = startLine + 1;
+ // If we are here - we detected HTML block.
+ // Let's roll down till block end.
+ if (!HTML_SEQUENCES[i][1].test(lineText)) {
+ for (;nextLine < endLine; nextLine++) {
+ if (state.sCount[nextLine] < state.blkIndent) {
+ break;
+ }
+ pos = state.bMarks[nextLine] + state.tShift[nextLine];
+ max = state.eMarks[nextLine];
+ lineText = state.src.slice(pos, max);
+ if (HTML_SEQUENCES[i][1].test(lineText)) {
+ if (lineText.length !== 0) {
+ nextLine++;
+ }
+ break;
+ }
+ }
+ }
+ state.line = nextLine;
+ token = state.push("html_block", "", 0);
+ token.map = [ startLine, nextLine ];
+ token.content = state.getLines(startLine, nextLine, state.blkIndent, true);
+ return true;
+ };
+ var isSpace$5 = utils.isSpace;
+ var heading = function heading(state, startLine, endLine, silent) {
+ var ch, level, tmp, token, pos = state.bMarks[startLine] + state.tShift[startLine], max = state.eMarks[startLine];
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ ch = state.src.charCodeAt(pos);
+ if (ch !== 35 /* # */ || pos >= max) {
+ return false;
+ }
+ // count heading level
+ level = 1;
+ ch = state.src.charCodeAt(++pos);
+ while (ch === 35 /* # */ && pos < max && level <= 6) {
+ level++;
+ ch = state.src.charCodeAt(++pos);
+ }
+ if (level > 6 || pos < max && !isSpace$5(ch)) {
+ return false;
+ }
+ if (silent) {
+ return true;
+ }
+ // Let's cut tails like ' ### ' from the end of string
+ max = state.skipSpacesBack(max, pos);
+ tmp = state.skipCharsBack(max, 35, pos);
+ // #
+ if (tmp > pos && isSpace$5(state.src.charCodeAt(tmp - 1))) {
+ max = tmp;
+ }
+ state.line = startLine + 1;
+ token = state.push("heading_open", "h" + String(level), 1);
+ token.markup = "########".slice(0, level);
+ token.map = [ startLine, state.line ];
+ token = state.push("inline", "", 0);
+ token.content = state.src.slice(pos, max).trim();
+ token.map = [ startLine, state.line ];
+ token.children = [];
+ token = state.push("heading_close", "h" + String(level), -1);
+ token.markup = "########".slice(0, level);
+ return true;
+ };
+ // lheading (---, ===)
+ var lheading = function lheading(state, startLine, endLine /*, silent*/) {
+ var content, terminate, i, l, token, pos, max, level, marker, nextLine = startLine + 1, oldParentType, terminatorRules = state.md.block.ruler.getRules("paragraph");
+ // if it's indented more than 3 spaces, it should be a code block
+ if (state.sCount[startLine] - state.blkIndent >= 4) {
+ return false;
+ }
+ oldParentType = state.parentType;
+ state.parentType = "paragraph";
+ // use paragraph to match terminatorRules
+ // jump line-by-line until empty one or EOF
+ for (;nextLine < endLine && !state.isEmpty(nextLine); nextLine++) {
+ // this would be a code block normally, but after paragraph
+ // it's considered a lazy continuation regardless of what's there
+ if (state.sCount[nextLine] - state.blkIndent > 3) {
+ continue;
+ }
+
+ // Check for underline in setext header
+
+ if (state.sCount[nextLine] >= state.blkIndent) {
+ pos = state.bMarks[nextLine] + state.tShift[nextLine];
+ max = state.eMarks[nextLine];
+ if (pos < max) {
+ marker = state.src.charCodeAt(pos);
+ if (marker === 45 /* - */ || marker === 61 /* = */) {
+ pos = state.skipChars(pos, marker);
+ pos = state.skipSpaces(pos);
+ if (pos >= max) {
+ level = marker === 61 /* = */ ? 1 : 2;
+ break;
+ }
+ }
+ }
+ }
+ // quirk for blockquotes, this line should already be checked by that rule
+ if (state.sCount[nextLine] < 0) {
+ continue;
+ }
+ // Some tags can terminate paragraph without empty line.
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ break;
+ }
+ }
+ if (!level) {
+ // Didn't find valid underline
+ return false;
+ }
+ content = state.getLines(startLine, nextLine, state.blkIndent, false).trim();
+ state.line = nextLine + 1;
+ token = state.push("heading_open", "h" + String(level), 1);
+ token.markup = String.fromCharCode(marker);
+ token.map = [ startLine, state.line ];
+ token = state.push("inline", "", 0);
+ token.content = content;
+ token.map = [ startLine, state.line - 1 ];
+ token.children = [];
+ token = state.push("heading_close", "h" + String(level), -1);
+ token.markup = String.fromCharCode(marker);
+ state.parentType = oldParentType;
+ return true;
+ };
+ // Paragraph
+ var paragraph = function paragraph(state, startLine /*, endLine*/) {
+ var content, terminate, i, l, token, oldParentType, nextLine = startLine + 1, terminatorRules = state.md.block.ruler.getRules("paragraph"), endLine = state.lineMax;
+ oldParentType = state.parentType;
+ state.parentType = "paragraph";
+ // jump line-by-line until empty one or EOF
+ for (;nextLine < endLine && !state.isEmpty(nextLine); nextLine++) {
+ // this would be a code block normally, but after paragraph
+ // it's considered a lazy continuation regardless of what's there
+ if (state.sCount[nextLine] - state.blkIndent > 3) {
+ continue;
+ }
+ // quirk for blockquotes, this line should already be checked by that rule
+ if (state.sCount[nextLine] < 0) {
+ continue;
+ }
+ // Some tags can terminate paragraph without empty line.
+ terminate = false;
+ for (i = 0, l = terminatorRules.length; i < l; i++) {
+ if (terminatorRules[i](state, nextLine, endLine, true)) {
+ terminate = true;
+ break;
+ }
+ }
+ if (terminate) {
+ break;
+ }
+ }
+ content = state.getLines(startLine, nextLine, state.blkIndent, false).trim();
+ state.line = nextLine;
+ token = state.push("paragraph_open", "p", 1);
+ token.map = [ startLine, state.line ];
+ token = state.push("inline", "", 0);
+ token.content = content;
+ token.map = [ startLine, state.line ];
+ token.children = [];
+ token = state.push("paragraph_close", "p", -1);
+ state.parentType = oldParentType;
+ return true;
+ };
+ var isSpace$4 = utils.isSpace;
+ function StateBlock(src, md, env, tokens) {
+ var ch, s, start, pos, len, indent, offset, indent_found;
+ this.src = src;
+ // link to parser instance
+ this.md = md;
+ this.env = env;
+
+ // Internal state vartiables
+
+ this.tokens = tokens;
+ this.bMarks = [];
+ // line begin offsets for fast jumps
+ this.eMarks = [];
+ // line end offsets for fast jumps
+ this.tShift = [];
+ // offsets of the first non-space characters (tabs not expanded)
+ this.sCount = [];
+ // indents for each line (tabs expanded)
+ // An amount of virtual spaces (tabs expanded) between beginning
+ // of each line (bMarks) and real beginning of that line.
+
+ // It exists only as a hack because blockquotes override bMarks
+ // losing information in the process.
+
+ // It's used only when expanding tabs, you can think about it as
+ // an initial tab length, e.g. bsCount=21 applied to string `\t123`
+ // means first tab should be expanded to 4-21%4 === 3 spaces.
+
+ this.bsCount = [];
+ // block parser variables
+ this.blkIndent = 0;
+ // required block content indent (for example, if we are
+ // inside a list, it would be positioned after list marker)
+ this.line = 0;
+ // line index in src
+ this.lineMax = 0;
+ // lines count
+ this.tight = false;
+ // loose/tight mode for lists
+ this.ddIndent = -1;
+ // indent of the current dd block (-1 if there isn't any)
+ this.listIndent = -1;
+ // indent of the current list block (-1 if there isn't any)
+ // can be 'blockquote', 'list', 'root', 'paragraph' or 'reference'
+ // used in lists to determine if they interrupt a paragraph
+ this.parentType = "root";
+ this.level = 0;
+ // renderer
+ this.result = "";
+ // Create caches
+ // Generate markers.
+ s = this.src;
+ indent_found = false;
+ for (start = pos = indent = offset = 0, len = s.length; pos < len; pos++) {
+ ch = s.charCodeAt(pos);
+ if (!indent_found) {
+ if (isSpace$4(ch)) {
+ indent++;
+ if (ch === 9) {
+ offset += 4 - offset % 4;
+ } else {
+ offset++;
+ }
+ continue;
+ } else {
+ indent_found = true;
+ }
+ }
+ if (ch === 10 || pos === len - 1) {
+ if (ch !== 10) {
+ pos++;
+ }
+ this.bMarks.push(start);
+ this.eMarks.push(pos);
+ this.tShift.push(indent);
+ this.sCount.push(offset);
+ this.bsCount.push(0);
+ indent_found = false;
+ indent = 0;
+ offset = 0;
+ start = pos + 1;
+ }
+ }
+ // Push fake entry to simplify cache bounds checks
+ this.bMarks.push(s.length);
+ this.eMarks.push(s.length);
+ this.tShift.push(0);
+ this.sCount.push(0);
+ this.bsCount.push(0);
+ this.lineMax = this.bMarks.length - 1;
+ // don't count last fake line
+ }
+ // Push new token to "stream".
+
+ StateBlock.prototype.push = function(type, tag, nesting) {
+ var token$1 = new token(type, tag, nesting);
+ token$1.block = true;
+ if (nesting < 0) this.level--;
+ // closing tag
+ token$1.level = this.level;
+ if (nesting > 0) this.level++;
+ // opening tag
+ this.tokens.push(token$1);
+ return token$1;
+ };
+ StateBlock.prototype.isEmpty = function isEmpty(line) {
+ return this.bMarks[line] + this.tShift[line] >= this.eMarks[line];
+ };
+ StateBlock.prototype.skipEmptyLines = function skipEmptyLines(from) {
+ for (var max = this.lineMax; from < max; from++) {
+ if (this.bMarks[from] + this.tShift[from] < this.eMarks[from]) {
+ break;
+ }
+ }
+ return from;
+ };
+ // Skip spaces from given position.
+ StateBlock.prototype.skipSpaces = function skipSpaces(pos) {
+ var ch;
+ for (var max = this.src.length; pos < max; pos++) {
+ ch = this.src.charCodeAt(pos);
+ if (!isSpace$4(ch)) {
+ break;
+ }
+ }
+ return pos;
+ };
+ // Skip spaces from given position in reverse.
+ StateBlock.prototype.skipSpacesBack = function skipSpacesBack(pos, min) {
+ if (pos <= min) {
+ return pos;
+ }
+ while (pos > min) {
+ if (!isSpace$4(this.src.charCodeAt(--pos))) {
+ return pos + 1;
+ }
+ }
+ return pos;
+ };
+ // Skip char codes from given position
+ StateBlock.prototype.skipChars = function skipChars(pos, code) {
+ for (var max = this.src.length; pos < max; pos++) {
+ if (this.src.charCodeAt(pos) !== code) {
+ break;
+ }
+ }
+ return pos;
+ };
+ // Skip char codes reverse from given position - 1
+ StateBlock.prototype.skipCharsBack = function skipCharsBack(pos, code, min) {
+ if (pos <= min) {
+ return pos;
+ }
+ while (pos > min) {
+ if (code !== this.src.charCodeAt(--pos)) {
+ return pos + 1;
+ }
+ }
+ return pos;
+ };
+ // cut lines range from source.
+ StateBlock.prototype.getLines = function getLines(begin, end, indent, keepLastLF) {
+ var i, lineIndent, ch, first, last, queue, lineStart, line = begin;
+ if (begin >= end) {
+ return "";
+ }
+ queue = new Array(end - begin);
+ for (i = 0; line < end; line++, i++) {
+ lineIndent = 0;
+ lineStart = first = this.bMarks[line];
+ if (line + 1 < end || keepLastLF) {
+ // No need for bounds check because we have fake entry on tail.
+ last = this.eMarks[line] + 1;
+ } else {
+ last = this.eMarks[line];
+ }
+ while (first < last && lineIndent < indent) {
+ ch = this.src.charCodeAt(first);
+ if (isSpace$4(ch)) {
+ if (ch === 9) {
+ lineIndent += 4 - (lineIndent + this.bsCount[line]) % 4;
+ } else {
+ lineIndent++;
+ }
+ } else if (first - lineStart < this.tShift[line]) {
+ // patched tShift masked characters to look like spaces (blockquotes, list markers)
+ lineIndent++;
+ } else {
+ break;
+ }
+ first++;
+ }
+ if (lineIndent > indent) {
+ // partially expanding tabs in code blocks, e.g '\t\tfoobar'
+ // with indent=2 becomes ' \tfoobar'
+ queue[i] = new Array(lineIndent - indent + 1).join(" ") + this.src.slice(first, last);
+ } else {
+ queue[i] = this.src.slice(first, last);
+ }
+ }
+ return queue.join("");
+ };
+ // re-export Token class to use in block rules
+ StateBlock.prototype.Token = token;
+ var state_block = StateBlock;
+ var _rules$1 = [
+ // First 2 params - rule name & source. Secondary array - list of rules,
+ // which can be terminated by this one.
+ [ "table", table, [ "paragraph", "reference" ] ], [ "code", code ], [ "fence", fence, [ "paragraph", "reference", "blockquote", "list" ] ], [ "blockquote", blockquote, [ "paragraph", "reference", "blockquote", "list" ] ], [ "hr", hr, [ "paragraph", "reference", "blockquote", "list" ] ], [ "list", list, [ "paragraph", "reference", "blockquote" ] ], [ "reference", reference ], [ "html_block", html_block, [ "paragraph", "reference", "blockquote" ] ], [ "heading", heading, [ "paragraph", "reference", "blockquote" ] ], [ "lheading", lheading ], [ "paragraph", paragraph ] ];
+ /**
+ * new ParserBlock()
+ **/ function ParserBlock() {
+ /**
+ * ParserBlock#ruler -> Ruler
+ *
+ * [[Ruler]] instance. Keep configuration of block rules.
+ **/
+ this.ruler = new ruler;
+ for (var i = 0; i < _rules$1.length; i++) {
+ this.ruler.push(_rules$1[i][0], _rules$1[i][1], {
+ alt: (_rules$1[i][2] || []).slice()
+ });
+ }
+ }
+ // Generate tokens for input range
+
+ ParserBlock.prototype.tokenize = function(state, startLine, endLine) {
+ var ok, i, rules = this.ruler.getRules(""), len = rules.length, line = startLine, hasEmptyLines = false, maxNesting = state.md.options.maxNesting;
+ while (line < endLine) {
+ state.line = line = state.skipEmptyLines(line);
+ if (line >= endLine) {
+ break;
+ }
+ // Termination condition for nested calls.
+ // Nested calls currently used for blockquotes & lists
+ if (state.sCount[line] < state.blkIndent) {
+ break;
+ }
+ // If nesting level exceeded - skip tail to the end. That's not ordinary
+ // situation and we should not care about content.
+ if (state.level >= maxNesting) {
+ state.line = endLine;
+ break;
+ }
+ // Try all possible rules.
+ // On success, rule should:
+
+ // - update `state.line`
+ // - update `state.tokens`
+ // - return true
+ for (i = 0; i < len; i++) {
+ ok = rules[i](state, line, endLine, false);
+ if (ok) {
+ break;
+ }
+ }
+ // set state.tight if we had an empty line before current tag
+ // i.e. latest empty line should not count
+ state.tight = !hasEmptyLines;
+ // paragraph might "eat" one newline after it in nested lists
+ if (state.isEmpty(state.line - 1)) {
+ hasEmptyLines = true;
+ }
+ line = state.line;
+ if (line < endLine && state.isEmpty(line)) {
+ hasEmptyLines = true;
+ line++;
+ state.line = line;
+ }
+ }
+ };
+ /**
+ * ParserBlock.parse(str, md, env, outTokens)
+ *
+ * Process input string and push block tokens into `outTokens`
+ **/ ParserBlock.prototype.parse = function(src, md, env, outTokens) {
+ var state;
+ if (!src) {
+ return;
+ }
+ state = new this.State(src, md, env, outTokens);
+ this.tokenize(state, state.line, state.lineMax);
+ };
+ ParserBlock.prototype.State = state_block;
+ var parser_block = ParserBlock;
+ // Skip text characters for text token, place those to pending buffer
+ // Rule to skip pure text
+ // '{}$%@~+=:' reserved for extentions
+ // !, ", #, $, %, &, ', (, ), *, +, ,, -, ., /, :, ;, <, =, >, ?, @, [, \, ], ^, _, `, {, |, }, or ~
+ // !!!! Don't confuse with "Markdown ASCII Punctuation" chars
+ // http://spec.commonmark.org/0.15/#ascii-punctuation-character
+ function isTerminatorChar(ch) {
+ switch (ch) {
+ case 10 /* \n */ :
+ case 33 /* ! */ :
+ case 35 /* # */ :
+ case 36 /* $ */ :
+ case 37 /* % */ :
+ case 38 /* & */ :
+ case 42 /* * */ :
+ case 43 /* + */ :
+ case 45 /* - */ :
+ case 58 /* : */ :
+ case 60 /* < */ :
+ case 61 /* = */ :
+ case 62 /* > */ :
+ case 64 /* @ */ :
+ case 91 /* [ */ :
+ case 92 /* \ */ :
+ case 93 /* ] */ :
+ case 94 /* ^ */ :
+ case 95 /* _ */ :
+ case 96 /* ` */ :
+ case 123 /* { */ :
+ case 125 /* } */ :
+ case 126 /* ~ */ :
+ return true;
+
+ default:
+ return false;
+ }
+ }
+ var text = function text(state, silent) {
+ var pos = state.pos;
+ while (pos < state.posMax && !isTerminatorChar(state.src.charCodeAt(pos))) {
+ pos++;
+ }
+ if (pos === state.pos) {
+ return false;
+ }
+ if (!silent) {
+ state.pending += state.src.slice(state.pos, pos);
+ }
+ state.pos = pos;
+ return true;
+ };
+ var isSpace$3 = utils.isSpace;
+ var newline = function newline(state, silent) {
+ var pmax, max, ws, pos = state.pos;
+ if (state.src.charCodeAt(pos) !== 10 /* \n */) {
+ return false;
+ }
+ pmax = state.pending.length - 1;
+ max = state.posMax;
+ // ' \n' -> hardbreak
+ // Lookup in pending chars is bad practice! Don't copy to other rules!
+ // Pending string is stored in concat mode, indexed lookups will cause
+ // convertion to flat mode.
+ if (!silent) {
+ if (pmax >= 0 && state.pending.charCodeAt(pmax) === 32) {
+ if (pmax >= 1 && state.pending.charCodeAt(pmax - 1) === 32) {
+ // Find whitespaces tail of pending chars.
+ ws = pmax - 1;
+ while (ws >= 1 && state.pending.charCodeAt(ws - 1) === 32) ws--;
+ state.pending = state.pending.slice(0, ws);
+ state.push("hardbreak", "br", 0);
+ } else {
+ state.pending = state.pending.slice(0, -1);
+ state.push("softbreak", "br", 0);
+ }
+ } else {
+ state.push("softbreak", "br", 0);
+ }
+ }
+ pos++;
+ // skip heading spaces for next line
+ while (pos < max && isSpace$3(state.src.charCodeAt(pos))) {
+ pos++;
+ }
+ state.pos = pos;
+ return true;
+ };
+ var isSpace$2 = utils.isSpace;
+ var ESCAPED = [];
+ for (var i = 0; i < 256; i++) {
+ ESCAPED.push(0);
+ }
+ "\\!\"#$%&'()*+,./:;<=>?@[]^_`{|}~-".split("").forEach((function(ch) {
+ ESCAPED[ch.charCodeAt(0)] = 1;
+ }));
+ var _escape = function escape(state, silent) {
+ var ch, pos = state.pos, max = state.posMax;
+ if (state.src.charCodeAt(pos) !== 92 /* \ */) {
+ return false;
+ }
+ pos++;
+ if (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (ch < 256 && ESCAPED[ch] !== 0) {
+ if (!silent) {
+ state.pending += state.src[pos];
+ }
+ state.pos += 2;
+ return true;
+ }
+ if (ch === 10) {
+ if (!silent) {
+ state.push("hardbreak", "br", 0);
+ }
+ pos++;
+ // skip leading whitespaces from next line
+ while (pos < max) {
+ ch = state.src.charCodeAt(pos);
+ if (!isSpace$2(ch)) {
+ break;
+ }
+ pos++;
+ }
+ state.pos = pos;
+ return true;
+ }
+ }
+ if (!silent) {
+ state.pending += "\\";
+ }
+ state.pos++;
+ return true;
+ };
+ // Parse backticks
+ var backticks = function backtick(state, silent) {
+ var start, max, marker, token, matchStart, matchEnd, openerLength, closerLength, pos = state.pos, ch = state.src.charCodeAt(pos);
+ if (ch !== 96 /* ` */) {
+ return false;
+ }
+ start = pos;
+ pos++;
+ max = state.posMax;
+ // scan marker length
+ while (pos < max && state.src.charCodeAt(pos) === 96 /* ` */) {
+ pos++;
+ }
+ marker = state.src.slice(start, pos);
+ openerLength = marker.length;
+ if (state.backticksScanned && (state.backticks[openerLength] || 0) <= start) {
+ if (!silent) state.pending += marker;
+ state.pos += openerLength;
+ return true;
+ }
+ matchStart = matchEnd = pos;
+ // Nothing found in the cache, scan until the end of the line (or until marker is found)
+ while ((matchStart = state.src.indexOf("`", matchEnd)) !== -1) {
+ matchEnd = matchStart + 1;
+ // scan marker length
+ while (matchEnd < max && state.src.charCodeAt(matchEnd) === 96 /* ` */) {
+ matchEnd++;
+ }
+ closerLength = matchEnd - matchStart;
+ if (closerLength === openerLength) {
+ // Found matching closer length.
+ if (!silent) {
+ token = state.push("code_inline", "code", 0);
+ token.markup = marker;
+ token.content = state.src.slice(pos, matchStart).replace(/\n/g, " ").replace(/^ (.+) $/, "$1");
+ }
+ state.pos = matchEnd;
+ return true;
+ }
+ // Some different length found, put it in cache as upper limit of where closer can be found
+ state.backticks[closerLength] = matchStart;
+ }
+ // Scanned through the end, didn't find anything
+ state.backticksScanned = true;
+ if (!silent) state.pending += marker;
+ state.pos += openerLength;
+ return true;
+ };
+ // ~~strike through~~
+ // Insert each marker as a separate text token, and add it to delimiter list
+
+ var tokenize$1 = function strikethrough(state, silent) {
+ var i, scanned, token, len, ch, start = state.pos, marker = state.src.charCodeAt(start);
+ if (silent) {
+ return false;
+ }
+ if (marker !== 126 /* ~ */) {
+ return false;
+ }
+ scanned = state.scanDelims(state.pos, true);
+ len = scanned.length;
+ ch = String.fromCharCode(marker);
+ if (len < 2) {
+ return false;
+ }
+ if (len % 2) {
+ token = state.push("text", "", 0);
+ token.content = ch;
+ len--;
+ }
+ for (i = 0; i < len; i += 2) {
+ token = state.push("text", "", 0);
+ token.content = ch + ch;
+ state.delimiters.push({
+ marker: marker,
+ length: 0,
+ // disable "rule of 3" length checks meant for emphasis
+ token: state.tokens.length - 1,
+ end: -1,
+ open: scanned.can_open,
+ close: scanned.can_close
+ });
+ }
+ state.pos += scanned.length;
+ return true;
+ };
+ function postProcess$1(state, delimiters) {
+ var i, j, startDelim, endDelim, token, loneMarkers = [], max = delimiters.length;
+ for (i = 0; i < max; i++) {
+ startDelim = delimiters[i];
+ if (startDelim.marker !== 126 /* ~ */) {
+ continue;
+ }
+ if (startDelim.end === -1) {
+ continue;
+ }
+ endDelim = delimiters[startDelim.end];
+ token = state.tokens[startDelim.token];
+ token.type = "s_open";
+ token.tag = "s";
+ token.nesting = 1;
+ token.markup = "~~";
+ token.content = "";
+ token = state.tokens[endDelim.token];
+ token.type = "s_close";
+ token.tag = "s";
+ token.nesting = -1;
+ token.markup = "~~";
+ token.content = "";
+ if (state.tokens[endDelim.token - 1].type === "text" && state.tokens[endDelim.token - 1].content === "~") {
+ loneMarkers.push(endDelim.token - 1);
+ }
+ }
+ // If a marker sequence has an odd number of characters, it's splitted
+ // like this: `~~~~~` -> `~` + `~~` + `~~`, leaving one marker at the
+ // start of the sequence.
+
+ // So, we have to move all those markers after subsequent s_close tags.
+
+ while (loneMarkers.length) {
+ i = loneMarkers.pop();
+ j = i + 1;
+ while (j < state.tokens.length && state.tokens[j].type === "s_close") {
+ j++;
+ }
+ j--;
+ if (i !== j) {
+ token = state.tokens[j];
+ state.tokens[j] = state.tokens[i];
+ state.tokens[i] = token;
+ }
+ }
+ }
+ // Walk through delimiter list and replace text tokens with tags
+
+ var postProcess_1$1 = function strikethrough(state) {
+ var curr, tokens_meta = state.tokens_meta, max = state.tokens_meta.length;
+ postProcess$1(state, state.delimiters);
+ for (curr = 0; curr < max; curr++) {
+ if (tokens_meta[curr] && tokens_meta[curr].delimiters) {
+ postProcess$1(state, tokens_meta[curr].delimiters);
+ }
+ }
+ };
+ var strikethrough = {
+ tokenize: tokenize$1,
+ postProcess: postProcess_1$1
+ };
+ // Process *this* and _that_
+ // Insert each marker as a separate text token, and add it to delimiter list
+
+ var tokenize = function emphasis(state, silent) {
+ var i, scanned, token, start = state.pos, marker = state.src.charCodeAt(start);
+ if (silent) {
+ return false;
+ }
+ if (marker !== 95 /* _ */ && marker !== 42 /* * */) {
+ return false;
+ }
+ scanned = state.scanDelims(state.pos, marker === 42);
+ for (i = 0; i < scanned.length; i++) {
+ token = state.push("text", "", 0);
+ token.content = String.fromCharCode(marker);
+ state.delimiters.push({
+ // Char code of the starting marker (number).
+ marker: marker,
+ // Total length of these series of delimiters.
+ length: scanned.length,
+ // A position of the token this delimiter corresponds to.
+ token: state.tokens.length - 1,
+ // If this delimiter is matched as a valid opener, `end` will be
+ // equal to its position, otherwise it's `-1`.
+ end: -1,
+ // Boolean flags that determine if this delimiter could open or close
+ // an emphasis.
+ open: scanned.can_open,
+ close: scanned.can_close
+ });
+ }
+ state.pos += scanned.length;
+ return true;
+ };
+ function postProcess(state, delimiters) {
+ var i, startDelim, endDelim, token, ch, isStrong, max = delimiters.length;
+ for (i = max - 1; i >= 0; i--) {
+ startDelim = delimiters[i];
+ if (startDelim.marker !== 95 /* _ */ && startDelim.marker !== 42 /* * */) {
+ continue;
+ }
+ // Process only opening markers
+ if (startDelim.end === -1) {
+ continue;
+ }
+ endDelim = delimiters[startDelim.end];
+ // If the previous delimiter has the same marker and is adjacent to this one,
+ // merge those into one strong delimiter.
+
+ // `<em><em>whatever</em></em>` -> `<strong>whatever</strong>`
+
+ isStrong = i > 0 && delimiters[i - 1].end === startDelim.end + 1 &&
+ // check that first two markers match and adjacent
+ delimiters[i - 1].marker === startDelim.marker && delimiters[i - 1].token === startDelim.token - 1 &&
+ // check that last two markers are adjacent (we can safely assume they match)
+ delimiters[startDelim.end + 1].token === endDelim.token + 1;
+ ch = String.fromCharCode(startDelim.marker);
+ token = state.tokens[startDelim.token];
+ token.type = isStrong ? "strong_open" : "em_open";
+ token.tag = isStrong ? "strong" : "em";
+ token.nesting = 1;
+ token.markup = isStrong ? ch + ch : ch;
+ token.content = "";
+ token = state.tokens[endDelim.token];
+ token.type = isStrong ? "strong_close" : "em_close";
+ token.tag = isStrong ? "strong" : "em";
+ token.nesting = -1;
+ token.markup = isStrong ? ch + ch : ch;
+ token.content = "";
+ if (isStrong) {
+ state.tokens[delimiters[i - 1].token].content = "";
+ state.tokens[delimiters[startDelim.end + 1].token].content = "";
+ i--;
+ }
+ }
+ }
+ // Walk through delimiter list and replace text tokens with tags
+
+ var postProcess_1 = function emphasis(state) {
+ var curr, tokens_meta = state.tokens_meta, max = state.tokens_meta.length;
+ postProcess(state, state.delimiters);
+ for (curr = 0; curr < max; curr++) {
+ if (tokens_meta[curr] && tokens_meta[curr].delimiters) {
+ postProcess(state, tokens_meta[curr].delimiters);
+ }
+ }
+ };
+ var emphasis = {
+ tokenize: tokenize,
+ postProcess: postProcess_1
+ };
+ var normalizeReference$1 = utils.normalizeReference;
+ var isSpace$1 = utils.isSpace;
+ var link = function link(state, silent) {
+ var attrs, code, label, labelEnd, labelStart, pos, res, ref, token, href = "", title = "", oldPos = state.pos, max = state.posMax, start = state.pos, parseReference = true;
+ if (state.src.charCodeAt(state.pos) !== 91 /* [ */) {
+ return false;
+ }
+ labelStart = state.pos + 1;
+ labelEnd = state.md.helpers.parseLinkLabel(state, state.pos, true);
+ // parser failed to find ']', so it's not a valid link
+ if (labelEnd < 0) {
+ return false;
+ }
+ pos = labelEnd + 1;
+ if (pos < max && state.src.charCodeAt(pos) === 40 /* ( */) {
+ // Inline link
+ // might have found a valid shortcut link, disable reference parsing
+ parseReference = false;
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ pos++;
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace$1(code) && code !== 10) {
+ break;
+ }
+ }
+ if (pos >= max) {
+ return false;
+ }
+ // [link]( <href> "title" )
+ // ^^^^^^ parsing link destination
+ start = pos;
+ res = state.md.helpers.parseLinkDestination(state.src, pos, state.posMax);
+ if (res.ok) {
+ href = state.md.normalizeLink(res.str);
+ if (state.md.validateLink(href)) {
+ pos = res.pos;
+ } else {
+ href = "";
+ }
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ start = pos;
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace$1(code) && code !== 10) {
+ break;
+ }
+ }
+ // [link]( <href> "title" )
+ // ^^^^^^^ parsing link title
+ res = state.md.helpers.parseLinkTitle(state.src, pos, state.posMax);
+ if (pos < max && start !== pos && res.ok) {
+ title = res.str;
+ pos = res.pos;
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace$1(code) && code !== 10) {
+ break;
+ }
+ }
+ }
+ }
+ if (pos >= max || state.src.charCodeAt(pos) !== 41 /* ) */) {
+ // parsing a valid shortcut link failed, fallback to reference
+ parseReference = true;
+ }
+ pos++;
+ }
+ if (parseReference) {
+ // Link reference
+ if (typeof state.env.references === "undefined") {
+ return false;
+ }
+ if (pos < max && state.src.charCodeAt(pos) === 91 /* [ */) {
+ start = pos + 1;
+ pos = state.md.helpers.parseLinkLabel(state, pos);
+ if (pos >= 0) {
+ label = state.src.slice(start, pos++);
+ } else {
+ pos = labelEnd + 1;
+ }
+ } else {
+ pos = labelEnd + 1;
+ }
+ // covers label === '' and label === undefined
+ // (collapsed reference link and shortcut reference link respectively)
+ if (!label) {
+ label = state.src.slice(labelStart, labelEnd);
+ }
+ ref = state.env.references[normalizeReference$1(label)];
+ if (!ref) {
+ state.pos = oldPos;
+ return false;
+ }
+ href = ref.href;
+ title = ref.title;
+ }
+
+ // We found the end of the link, and know for a fact it's a valid link;
+ // so all that's left to do is to call tokenizer.
+
+ if (!silent) {
+ state.pos = labelStart;
+ state.posMax = labelEnd;
+ token = state.push("link_open", "a", 1);
+ token.attrs = attrs = [ [ "href", href ] ];
+ if (title) {
+ attrs.push([ "title", title ]);
+ }
+ state.md.inline.tokenize(state);
+ token = state.push("link_close", "a", -1);
+ }
+ state.pos = pos;
+ state.posMax = max;
+ return true;
+ };
+ var normalizeReference = utils.normalizeReference;
+ var isSpace = utils.isSpace;
+ var image = function image(state, silent) {
+ var attrs, code, content, label, labelEnd, labelStart, pos, ref, res, title, token, tokens, start, href = "", oldPos = state.pos, max = state.posMax;
+ if (state.src.charCodeAt(state.pos) !== 33 /* ! */) {
+ return false;
+ }
+ if (state.src.charCodeAt(state.pos + 1) !== 91 /* [ */) {
+ return false;
+ }
+ labelStart = state.pos + 2;
+ labelEnd = state.md.helpers.parseLinkLabel(state, state.pos + 1, false);
+ // parser failed to find ']', so it's not a valid link
+ if (labelEnd < 0) {
+ return false;
+ }
+ pos = labelEnd + 1;
+ if (pos < max && state.src.charCodeAt(pos) === 40 /* ( */) {
+ // Inline link
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ pos++;
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace(code) && code !== 10) {
+ break;
+ }
+ }
+ if (pos >= max) {
+ return false;
+ }
+ // [link]( <href> "title" )
+ // ^^^^^^ parsing link destination
+ start = pos;
+ res = state.md.helpers.parseLinkDestination(state.src, pos, state.posMax);
+ if (res.ok) {
+ href = state.md.normalizeLink(res.str);
+ if (state.md.validateLink(href)) {
+ pos = res.pos;
+ } else {
+ href = "";
+ }
+ }
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ start = pos;
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace(code) && code !== 10) {
+ break;
+ }
+ }
+ // [link]( <href> "title" )
+ // ^^^^^^^ parsing link title
+ res = state.md.helpers.parseLinkTitle(state.src, pos, state.posMax);
+ if (pos < max && start !== pos && res.ok) {
+ title = res.str;
+ pos = res.pos;
+ // [link]( <href> "title" )
+ // ^^ skipping these spaces
+ for (;pos < max; pos++) {
+ code = state.src.charCodeAt(pos);
+ if (!isSpace(code) && code !== 10) {
+ break;
+ }
+ }
+ } else {
+ title = "";
+ }
+ if (pos >= max || state.src.charCodeAt(pos) !== 41 /* ) */) {
+ state.pos = oldPos;
+ return false;
+ }
+ pos++;
+ } else {
+ // Link reference
+ if (typeof state.env.references === "undefined") {
+ return false;
+ }
+ if (pos < max && state.src.charCodeAt(pos) === 91 /* [ */) {
+ start = pos + 1;
+ pos = state.md.helpers.parseLinkLabel(state, pos);
+ if (pos >= 0) {
+ label = state.src.slice(start, pos++);
+ } else {
+ pos = labelEnd + 1;
+ }
+ } else {
+ pos = labelEnd + 1;
+ }
+ // covers label === '' and label === undefined
+ // (collapsed reference link and shortcut reference link respectively)
+ if (!label) {
+ label = state.src.slice(labelStart, labelEnd);
+ }
+ ref = state.env.references[normalizeReference(label)];
+ if (!ref) {
+ state.pos = oldPos;
+ return false;
+ }
+ href = ref.href;
+ title = ref.title;
+ }
+
+ // We found the end of the link, and know for a fact it's a valid link;
+ // so all that's left to do is to call tokenizer.
+
+ if (!silent) {
+ content = state.src.slice(labelStart, labelEnd);
+ state.md.inline.parse(content, state.md, state.env, tokens = []);
+ token = state.push("image", "img", 0);
+ token.attrs = attrs = [ [ "src", href ], [ "alt", "" ] ];
+ token.children = tokens;
+ token.content = content;
+ if (title) {
+ attrs.push([ "title", title ]);
+ }
+ }
+ state.pos = pos;
+ state.posMax = max;
+ return true;
+ };
+ // Process autolinks '<protocol:...>'
+ /*eslint max-len:0*/ var EMAIL_RE = /^([a-zA-Z0-9.!#$%&'*+\/=?^_`{|}~-]+@[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*)$/;
+ var AUTOLINK_RE = /^([a-zA-Z][a-zA-Z0-9+.\-]{1,31}):([^<>\x00-\x20]*)$/;
+ var autolink = function autolink(state, silent) {
+ var url, fullUrl, token, ch, start, max, pos = state.pos;
+ if (state.src.charCodeAt(pos) !== 60 /* < */) {
+ return false;
+ }
+ start = state.pos;
+ max = state.posMax;
+ for (;;) {
+ if (++pos >= max) return false;
+ ch = state.src.charCodeAt(pos);
+ if (ch === 60 /* < */) return false;
+ if (ch === 62 /* > */) break;
+ }
+ url = state.src.slice(start + 1, pos);
+ if (AUTOLINK_RE.test(url)) {
+ fullUrl = state.md.normalizeLink(url);
+ if (!state.md.validateLink(fullUrl)) {
+ return false;
+ }
+ if (!silent) {
+ token = state.push("link_open", "a", 1);
+ token.attrs = [ [ "href", fullUrl ] ];
+ token.markup = "autolink";
+ token.info = "auto";
+ token = state.push("text", "", 0);
+ token.content = state.md.normalizeLinkText(url);
+ token = state.push("link_close", "a", -1);
+ token.markup = "autolink";
+ token.info = "auto";
+ }
+ state.pos += url.length + 2;
+ return true;
+ }
+ if (EMAIL_RE.test(url)) {
+ fullUrl = state.md.normalizeLink("mailto:" + url);
+ if (!state.md.validateLink(fullUrl)) {
+ return false;
+ }
+ if (!silent) {
+ token = state.push("link_open", "a", 1);
+ token.attrs = [ [ "href", fullUrl ] ];
+ token.markup = "autolink";
+ token.info = "auto";
+ token = state.push("text", "", 0);
+ token.content = state.md.normalizeLinkText(url);
+ token = state.push("link_close", "a", -1);
+ token.markup = "autolink";
+ token.info = "auto";
+ }
+ state.pos += url.length + 2;
+ return true;
+ }
+ return false;
+ };
+ var HTML_TAG_RE = html_re.HTML_TAG_RE;
+ function isLetter(ch) {
+ /*eslint no-bitwise:0*/
+ var lc = ch | 32;
+ // to lower case
+ return lc >= 97 /* a */ && lc <= 122 /* z */;
+ }
+ var html_inline = function html_inline(state, silent) {
+ var ch, match, max, token, pos = state.pos;
+ if (!state.md.options.html) {
+ return false;
+ }
+ // Check start
+ max = state.posMax;
+ if (state.src.charCodeAt(pos) !== 60 /* < */ || pos + 2 >= max) {
+ return false;
+ }
+ // Quick fail on second char
+ ch = state.src.charCodeAt(pos + 1);
+ if (ch !== 33 /* ! */ && ch !== 63 /* ? */ && ch !== 47 /* / */ && !isLetter(ch)) {
+ return false;
+ }
+ match = state.src.slice(pos).match(HTML_TAG_RE);
+ if (!match) {
+ return false;
+ }
+ if (!silent) {
+ token = state.push("html_inline", "", 0);
+ token.content = state.src.slice(pos, pos + match[0].length);
+ }
+ state.pos += match[0].length;
+ return true;
+ };
+ var has = utils.has;
+ var isValidEntityCode = utils.isValidEntityCode;
+ var fromCodePoint = utils.fromCodePoint;
+ var DIGITAL_RE = /^&#((?:x[a-f0-9]{1,6}|[0-9]{1,7}));/i;
+ var NAMED_RE = /^&([a-z][a-z0-9]{1,31});/i;
+ var entity = function entity(state, silent) {
+ var ch, code, match, pos = state.pos, max = state.posMax;
+ if (state.src.charCodeAt(pos) !== 38 /* & */) {
+ return false;
+ }
+ if (pos + 1 < max) {
+ ch = state.src.charCodeAt(pos + 1);
+ if (ch === 35 /* # */) {
+ match = state.src.slice(pos).match(DIGITAL_RE);
+ if (match) {
+ if (!silent) {
+ code = match[1][0].toLowerCase() === "x" ? parseInt(match[1].slice(1), 16) : parseInt(match[1], 10);
+ state.pending += isValidEntityCode(code) ? fromCodePoint(code) : fromCodePoint(65533);
+ }
+ state.pos += match[0].length;
+ return true;
+ }
+ } else {
+ match = state.src.slice(pos).match(NAMED_RE);
+ if (match) {
+ if (has(entities, match[1])) {
+ if (!silent) {
+ state.pending += entities[match[1]];
+ }
+ state.pos += match[0].length;
+ return true;
+ }
+ }
+ }
+ }
+ if (!silent) {
+ state.pending += "&";
+ }
+ state.pos++;
+ return true;
+ };
+ // For each opening emphasis-like marker find a matching closing one
+ function processDelimiters(state, delimiters) {
+ var closerIdx, openerIdx, closer, opener, minOpenerIdx, newMinOpenerIdx, isOddMatch, lastJump, openersBottom = {}, max = delimiters.length;
+ if (!max) return;
+ // headerIdx is the first delimiter of the current (where closer is) delimiter run
+ var headerIdx = 0;
+ var lastTokenIdx = -2;
+ // needs any value lower than -1
+ var jumps = [];
+ for (closerIdx = 0; closerIdx < max; closerIdx++) {
+ closer = delimiters[closerIdx];
+ jumps.push(0);
+ // markers belong to same delimiter run if:
+ // - they have adjacent tokens
+ // - AND markers are the same
+
+ if (delimiters[headerIdx].marker !== closer.marker || lastTokenIdx !== closer.token - 1) {
+ headerIdx = closerIdx;
+ }
+ lastTokenIdx = closer.token;
+ // Length is only used for emphasis-specific "rule of 3",
+ // if it's not defined (in strikethrough or 3rd party plugins),
+ // we can default it to 0 to disable those checks.
+
+ closer.length = closer.length || 0;
+ if (!closer.close) continue;
+ // Previously calculated lower bounds (previous fails)
+ // for each marker, each delimiter length modulo 3,
+ // and for whether this closer can be an opener;
+ // https://github.com/commonmark/cmark/commit/34250e12ccebdc6372b8b49c44fab57c72443460
+ if (!openersBottom.hasOwnProperty(closer.marker)) {
+ openersBottom[closer.marker] = [ -1, -1, -1, -1, -1, -1 ];
+ }
+ minOpenerIdx = openersBottom[closer.marker][(closer.open ? 3 : 0) + closer.length % 3];
+ openerIdx = headerIdx - jumps[headerIdx] - 1;
+ newMinOpenerIdx = openerIdx;
+ for (;openerIdx > minOpenerIdx; openerIdx -= jumps[openerIdx] + 1) {
+ opener = delimiters[openerIdx];
+ if (opener.marker !== closer.marker) continue;
+ if (opener.open && opener.end < 0) {
+ isOddMatch = false;
+ // from spec:
+
+ // If one of the delimiters can both open and close emphasis, then the
+ // sum of the lengths of the delimiter runs containing the opening and
+ // closing delimiters must not be a multiple of 3 unless both lengths
+ // are multiples of 3.
+
+ if (opener.close || closer.open) {
+ if ((opener.length + closer.length) % 3 === 0) {
+ if (opener.length % 3 !== 0 || closer.length % 3 !== 0) {
+ isOddMatch = true;
+ }
+ }
+ }
+ if (!isOddMatch) {
+ // If previous delimiter cannot be an opener, we can safely skip
+ // the entire sequence in future checks. This is required to make
+ // sure algorithm has linear complexity (see *_*_*_*_*_... case).
+ lastJump = openerIdx > 0 && !delimiters[openerIdx - 1].open ? jumps[openerIdx - 1] + 1 : 0;
+ jumps[closerIdx] = closerIdx - openerIdx + lastJump;
+ jumps[openerIdx] = lastJump;
+ closer.open = false;
+ opener.end = closerIdx;
+ opener.close = false;
+ newMinOpenerIdx = -1;
+ // treat next token as start of run,
+ // it optimizes skips in **<...>**a**<...>** pathological case
+ lastTokenIdx = -2;
+ break;
+ }
+ }
+ }
+ if (newMinOpenerIdx !== -1) {
+ // If match for this delimiter run failed, we want to set lower bound for
+ // future lookups. This is required to make sure algorithm has linear
+ // complexity.
+ // See details here:
+ // https://github.com/commonmark/cmark/issues/178#issuecomment-270417442
+ openersBottom[closer.marker][(closer.open ? 3 : 0) + (closer.length || 0) % 3] = newMinOpenerIdx;
+ }
+ }
+ }
+ var balance_pairs = function link_pairs(state) {
+ var curr, tokens_meta = state.tokens_meta, max = state.tokens_meta.length;
+ processDelimiters(state, state.delimiters);
+ for (curr = 0; curr < max; curr++) {
+ if (tokens_meta[curr] && tokens_meta[curr].delimiters) {
+ processDelimiters(state, tokens_meta[curr].delimiters);
+ }
+ }
+ };
+ // Clean up tokens after emphasis and strikethrough postprocessing:
+ var text_collapse = function text_collapse(state) {
+ var curr, last, level = 0, tokens = state.tokens, max = state.tokens.length;
+ for (curr = last = 0; curr < max; curr++) {
+ // re-calculate levels after emphasis/strikethrough turns some text nodes
+ // into opening/closing tags
+ if (tokens[curr].nesting < 0) level--;
+ // closing tag
+ tokens[curr].level = level;
+ if (tokens[curr].nesting > 0) level++;
+ // opening tag
+ if (tokens[curr].type === "text" && curr + 1 < max && tokens[curr + 1].type === "text") {
+ // collapse two adjacent text nodes
+ tokens[curr + 1].content = tokens[curr].content + tokens[curr + 1].content;
+ } else {
+ if (curr !== last) {
+ tokens[last] = tokens[curr];
+ }
+ last++;
+ }
+ }
+ if (curr !== last) {
+ tokens.length = last;
+ }
+ };
+ var isWhiteSpace = utils.isWhiteSpace;
+ var isPunctChar = utils.isPunctChar;
+ var isMdAsciiPunct = utils.isMdAsciiPunct;
+ function StateInline(src, md, env, outTokens) {
+ this.src = src;
+ this.env = env;
+ this.md = md;
+ this.tokens = outTokens;
+ this.tokens_meta = Array(outTokens.length);
+ this.pos = 0;
+ this.posMax = this.src.length;
+ this.level = 0;
+ this.pending = "";
+ this.pendingLevel = 0;
+ // Stores { start: end } pairs. Useful for backtrack
+ // optimization of pairs parse (emphasis, strikes).
+ this.cache = {};
+ // List of emphasis-like delimiters for current tag
+ this.delimiters = [];
+ // Stack of delimiter lists for upper level tags
+ this._prev_delimiters = [];
+ // backtick length => last seen position
+ this.backticks = {};
+ this.backticksScanned = false;
+ }
+ // Flush pending text
+
+ StateInline.prototype.pushPending = function() {
+ var token$1 = new token("text", "", 0);
+ token$1.content = this.pending;
+ token$1.level = this.pendingLevel;
+ this.tokens.push(token$1);
+ this.pending = "";
+ return token$1;
+ };
+ // Push new token to "stream".
+ // If pending text exists - flush it as text token
+
+ StateInline.prototype.push = function(type, tag, nesting) {
+ if (this.pending) {
+ this.pushPending();
+ }
+ var token$1 = new token(type, tag, nesting);
+ var token_meta = null;
+ if (nesting < 0) {
+ // closing tag
+ this.level--;
+ this.delimiters = this._prev_delimiters.pop();
+ }
+ token$1.level = this.level;
+ if (nesting > 0) {
+ // opening tag
+ this.level++;
+ this._prev_delimiters.push(this.delimiters);
+ this.delimiters = [];
+ token_meta = {
+ delimiters: this.delimiters
+ };
+ }
+ this.pendingLevel = this.level;
+ this.tokens.push(token$1);
+ this.tokens_meta.push(token_meta);
+ return token$1;
+ };
+ // Scan a sequence of emphasis-like markers, and determine whether
+ // it can start an emphasis sequence or end an emphasis sequence.
+
+ // - start - position to scan from (it should point at a valid marker);
+ // - canSplitWord - determine if these markers can be found inside a word
+
+ StateInline.prototype.scanDelims = function(start, canSplitWord) {
+ var pos = start, lastChar, nextChar, count, can_open, can_close, isLastWhiteSpace, isLastPunctChar, isNextWhiteSpace, isNextPunctChar, left_flanking = true, right_flanking = true, max = this.posMax, marker = this.src.charCodeAt(start);
+ // treat beginning of the line as a whitespace
+ lastChar = start > 0 ? this.src.charCodeAt(start - 1) : 32;
+ while (pos < max && this.src.charCodeAt(pos) === marker) {
+ pos++;
+ }
+ count = pos - start;
+ // treat end of the line as a whitespace
+ nextChar = pos < max ? this.src.charCodeAt(pos) : 32;
+ isLastPunctChar = isMdAsciiPunct(lastChar) || isPunctChar(String.fromCharCode(lastChar));
+ isNextPunctChar = isMdAsciiPunct(nextChar) || isPunctChar(String.fromCharCode(nextChar));
+ isLastWhiteSpace = isWhiteSpace(lastChar);
+ isNextWhiteSpace = isWhiteSpace(nextChar);
+ if (isNextWhiteSpace) {
+ left_flanking = false;
+ } else if (isNextPunctChar) {
+ if (!(isLastWhiteSpace || isLastPunctChar)) {
+ left_flanking = false;
+ }
+ }
+ if (isLastWhiteSpace) {
+ right_flanking = false;
+ } else if (isLastPunctChar) {
+ if (!(isNextWhiteSpace || isNextPunctChar)) {
+ right_flanking = false;
+ }
+ }
+ if (!canSplitWord) {
+ can_open = left_flanking && (!right_flanking || isLastPunctChar);
+ can_close = right_flanking && (!left_flanking || isNextPunctChar);
+ } else {
+ can_open = left_flanking;
+ can_close = right_flanking;
+ }
+ return {
+ can_open: can_open,
+ can_close: can_close,
+ length: count
+ };
+ };
+ // re-export Token class to use in block rules
+ StateInline.prototype.Token = token;
+ var state_inline = StateInline;
+ ////////////////////////////////////////////////////////////////////////////////
+ // Parser rules
+ var _rules = [ [ "text", text ], [ "newline", newline ], [ "escape", _escape ], [ "backticks", backticks ], [ "strikethrough", strikethrough.tokenize ], [ "emphasis", emphasis.tokenize ], [ "link", link ], [ "image", image ], [ "autolink", autolink ], [ "html_inline", html_inline ], [ "entity", entity ] ];
+ var _rules2 = [ [ "balance_pairs", balance_pairs ], [ "strikethrough", strikethrough.postProcess ], [ "emphasis", emphasis.postProcess ], [ "text_collapse", text_collapse ] ];
+ /**
+ * new ParserInline()
+ **/ function ParserInline() {
+ var i;
+ /**
+ * ParserInline#ruler -> Ruler
+ *
+ * [[Ruler]] instance. Keep configuration of inline rules.
+ **/ this.ruler = new ruler;
+ for (i = 0; i < _rules.length; i++) {
+ this.ruler.push(_rules[i][0], _rules[i][1]);
+ }
+ /**
+ * ParserInline#ruler2 -> Ruler
+ *
+ * [[Ruler]] instance. Second ruler used for post-processing
+ * (e.g. in emphasis-like rules).
+ **/ this.ruler2 = new ruler;
+ for (i = 0; i < _rules2.length; i++) {
+ this.ruler2.push(_rules2[i][0], _rules2[i][1]);
+ }
+ }
+ // Skip single token by running all rules in validation mode;
+ // returns `true` if any rule reported success
+
+ ParserInline.prototype.skipToken = function(state) {
+ var ok, i, pos = state.pos, rules = this.ruler.getRules(""), len = rules.length, maxNesting = state.md.options.maxNesting, cache = state.cache;
+ if (typeof cache[pos] !== "undefined") {
+ state.pos = cache[pos];
+ return;
+ }
+ if (state.level < maxNesting) {
+ for (i = 0; i < len; i++) {
+ // Increment state.level and decrement it later to limit recursion.
+ // It's harmless to do here, because no tokens are created. But ideally,
+ // we'd need a separate private state variable for this purpose.
+ state.level++;
+ ok = rules[i](state, true);
+ state.level--;
+ if (ok) {
+ break;
+ }
+ }
+ } else {
+ // Too much nesting, just skip until the end of the paragraph.
+ // NOTE: this will cause links to behave incorrectly in the following case,
+ // when an amount of `[` is exactly equal to `maxNesting + 1`:
+ // [[[[[[[[[[[[[[[[[[[[[foo]()
+ // TODO: remove this workaround when CM standard will allow nested links
+ // (we can replace it by preventing links from being parsed in
+ // validation mode)
+ state.pos = state.posMax;
+ }
+ if (!ok) {
+ state.pos++;
+ }
+ cache[pos] = state.pos;
+ };
+ // Generate tokens for input range
+
+ ParserInline.prototype.tokenize = function(state) {
+ var ok, i, rules = this.ruler.getRules(""), len = rules.length, end = state.posMax, maxNesting = state.md.options.maxNesting;
+ while (state.pos < end) {
+ // Try all possible rules.
+ // On success, rule should:
+ // - update `state.pos`
+ // - update `state.tokens`
+ // - return true
+ if (state.level < maxNesting) {
+ for (i = 0; i < len; i++) {
+ ok = rules[i](state, false);
+ if (ok) {
+ break;
+ }
+ }
+ }
+ if (ok) {
+ if (state.pos >= end) {
+ break;
+ }
+ continue;
+ }
+ state.pending += state.src[state.pos++];
+ }
+ if (state.pending) {
+ state.pushPending();
+ }
+ };
+ /**
+ * ParserInline.parse(str, md, env, outTokens)
+ *
+ * Process input string and push inline tokens into `outTokens`
+ **/ ParserInline.prototype.parse = function(str, md, env, outTokens) {
+ var i, rules, len;
+ var state = new this.State(str, md, env, outTokens);
+ this.tokenize(state);
+ rules = this.ruler2.getRules("");
+ len = rules.length;
+ for (i = 0; i < len; i++) {
+ rules[i](state);
+ }
+ };
+ ParserInline.prototype.State = state_inline;
+ var parser_inline = ParserInline;
+ var re = function(opts) {
+ var re = {};
+ // Use direct extract instead of `regenerate` to reduse browserified size
+ re.src_Any = regex$3.source;
+ re.src_Cc = regex$2.source;
+ re.src_Z = regex.source;
+ re.src_P = regex$4.source;
+ // \p{\Z\P\Cc\CF} (white spaces + control + format + punctuation)
+ re.src_ZPCc = [ re.src_Z, re.src_P, re.src_Cc ].join("|");
+ // \p{\Z\Cc} (white spaces + control)
+ re.src_ZCc = [ re.src_Z, re.src_Cc ].join("|");
+ // Experimental. List of chars, completely prohibited in links
+ // because can separate it from other part of text
+ var text_separators = "[><\uff5c]";
+ // All possible word characters (everything without punctuation, spaces & controls)
+ // Defined via punctuation & spaces to save space
+ // Should be something like \p{\L\N\S\M} (\w but without `_`)
+ re.src_pseudo_letter = "(?:(?!" + text_separators + "|" + re.src_ZPCc + ")" + re.src_Any + ")";
+ // The same as abothe but without [0-9]
+ // var src_pseudo_letter_non_d = '(?:(?![0-9]|' + src_ZPCc + ')' + src_Any + ')';
+ ////////////////////////////////////////////////////////////////////////////////
+ re.src_ip4 = "(?:(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.){3}(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)";
+ // Prohibit any of "@/[]()" in user/pass to avoid wrong domain fetch.
+ re.src_auth = "(?:(?:(?!" + re.src_ZCc + "|[@/\\[\\]()]).)+@)?";
+ re.src_port = "(?::(?:6(?:[0-4]\\d{3}|5(?:[0-4]\\d{2}|5(?:[0-2]\\d|3[0-5])))|[1-5]?\\d{1,4}))?";
+ re.src_host_terminator = "(?=$|" + text_separators + "|" + re.src_ZPCc + ")(?!-|_|:\\d|\\.-|\\.(?!$|" + re.src_ZPCc + "))";
+ re.src_path = "(?:" + "[/?#]" + "(?:" + "(?!" + re.src_ZCc + "|" + text_separators + "|[()[\\]{}.,\"'?!\\-;]).|" + "\\[(?:(?!" + re.src_ZCc + "|\\]).)*\\]|" + "\\((?:(?!" + re.src_ZCc + "|[)]).)*\\)|" + "\\{(?:(?!" + re.src_ZCc + "|[}]).)*\\}|" + '\\"(?:(?!' + re.src_ZCc + '|["]).)+\\"|' + "\\'(?:(?!" + re.src_ZCc + "|[']).)+\\'|" + "\\'(?=" + re.src_pseudo_letter + "|[-]).|" + // allow `I'm_king` if no pair found
+ "\\.{2,}[a-zA-Z0-9%/&]|" + // google has many dots in "google search" links (#66, #81).
+ // github has ... in commit range links,
+ // Restrict to
+ // - english
+ // - percent-encoded
+ // - parts of file path
+ // - params separator
+ // until more examples found.
+ "\\.(?!" + re.src_ZCc + "|[.]).|" + (opts && opts["---"] ? "\\-(?!--(?:[^-]|$))(?:-*)|" : "\\-+|") + ",(?!" + re.src_ZCc + ").|" + // allow `,,,` in paths
+ ";(?!" + re.src_ZCc + ").|" + // allow `;` if not followed by space-like char
+ "\\!+(?!" + re.src_ZCc + "|[!]).|" + // allow `!!!` in paths, but not at the end
+ "\\?(?!" + re.src_ZCc + "|[?])." + ")+" + "|\\/" + ")?";
+ // Allow anything in markdown spec, forbid quote (") at the first position
+ // because emails enclosed in quotes are far more common
+ re.src_email_name = '[\\-;:&=\\+\\$,\\.a-zA-Z0-9_][\\-;:&=\\+\\$,\\"\\.a-zA-Z0-9_]*';
+ re.src_xn = "xn--[a-z0-9\\-]{1,59}";
+ // More to read about domain names
+ // http://serverfault.com/questions/638260/
+ re.src_domain_root =
+ // Allow letters & digits (http://test1)
+ "(?:" + re.src_xn + "|" + re.src_pseudo_letter + "{1,63}" + ")";
+ re.src_domain = "(?:" + re.src_xn + "|" + "(?:" + re.src_pseudo_letter + ")" + "|" + "(?:" + re.src_pseudo_letter + "(?:-|" + re.src_pseudo_letter + "){0,61}" + re.src_pseudo_letter + ")" + ")";
+ re.src_host = "(?:" +
+ // Don't need IP check, because digits are already allowed in normal domain names
+ // src_ip4 +
+ // '|' +
+ "(?:(?:(?:" + re.src_domain + ")\\.)*" + re.src_domain /*_root*/ + ")" + ")";
+ re.tpl_host_fuzzy = "(?:" + re.src_ip4 + "|" + "(?:(?:(?:" + re.src_domain + ")\\.)+(?:%TLDS%))" + ")";
+ re.tpl_host_no_ip_fuzzy = "(?:(?:(?:" + re.src_domain + ")\\.)+(?:%TLDS%))";
+ re.src_host_strict = re.src_host + re.src_host_terminator;
+ re.tpl_host_fuzzy_strict = re.tpl_host_fuzzy + re.src_host_terminator;
+ re.src_host_port_strict = re.src_host + re.src_port + re.src_host_terminator;
+ re.tpl_host_port_fuzzy_strict = re.tpl_host_fuzzy + re.src_port + re.src_host_terminator;
+ re.tpl_host_port_no_ip_fuzzy_strict = re.tpl_host_no_ip_fuzzy + re.src_port + re.src_host_terminator;
+ ////////////////////////////////////////////////////////////////////////////////
+ // Main rules
+ // Rude test fuzzy links by host, for quick deny
+ re.tpl_host_fuzzy_test = "localhost|www\\.|\\.\\d{1,3}\\.|(?:\\.(?:%TLDS%)(?:" + re.src_ZPCc + "|>|$))";
+ re.tpl_email_fuzzy = "(^|" + text_separators + '|"|\\(|' + re.src_ZCc + ")" + "(" + re.src_email_name + "@" + re.tpl_host_fuzzy_strict + ")";
+ re.tpl_link_fuzzy =
+ // Fuzzy link can't be prepended with .:/\- and non punctuation.
+ // but can start with > (markdown blockquote)
+ "(^|(?![.:/\\-_@])(?:[$+<=>^`|\uff5c]|" + re.src_ZPCc + "))" + "((?![$+<=>^`|\uff5c])" + re.tpl_host_port_fuzzy_strict + re.src_path + ")";
+ re.tpl_link_no_ip_fuzzy =
+ // Fuzzy link can't be prepended with .:/\- and non punctuation.
+ // but can start with > (markdown blockquote)
+ "(^|(?![.:/\\-_@])(?:[$+<=>^`|\uff5c]|" + re.src_ZPCc + "))" + "((?![$+<=>^`|\uff5c])" + re.tpl_host_port_no_ip_fuzzy_strict + re.src_path + ")";
+ return re;
+ };
+ ////////////////////////////////////////////////////////////////////////////////
+ // Helpers
+ // Merge objects
+
+ function assign(obj /*from1, from2, from3, ...*/) {
+ var sources = Array.prototype.slice.call(arguments, 1);
+ sources.forEach((function(source) {
+ if (!source) {
+ return;
+ }
+ Object.keys(source).forEach((function(key) {
+ obj[key] = source[key];
+ }));
+ }));
+ return obj;
+ }
+ function _class(obj) {
+ return Object.prototype.toString.call(obj);
+ }
+ function isString(obj) {
+ return _class(obj) === "[object String]";
+ }
+ function isObject(obj) {
+ return _class(obj) === "[object Object]";
+ }
+ function isRegExp(obj) {
+ return _class(obj) === "[object RegExp]";
+ }
+ function isFunction(obj) {
+ return _class(obj) === "[object Function]";
+ }
+ function escapeRE(str) {
+ return str.replace(/[.?*+^$[\]\\(){}|-]/g, "\\$&");
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ var defaultOptions = {
+ fuzzyLink: true,
+ fuzzyEmail: true,
+ fuzzyIP: false
+ };
+ function isOptionsObj(obj) {
+ return Object.keys(obj || {}).reduce((function(acc, k) {
+ return acc || defaultOptions.hasOwnProperty(k);
+ }), false);
+ }
+ var defaultSchemas = {
+ "http:": {
+ validate: function(text, pos, self) {
+ var tail = text.slice(pos);
+ if (!self.re.http) {
+ // compile lazily, because "host"-containing variables can change on tlds update.
+ self.re.http = new RegExp("^\\/\\/" + self.re.src_auth + self.re.src_host_port_strict + self.re.src_path, "i");
+ }
+ if (self.re.http.test(tail)) {
+ return tail.match(self.re.http)[0].length;
+ }
+ return 0;
+ }
+ },
+ "https:": "http:",
+ "ftp:": "http:",
+ "//": {
+ validate: function(text, pos, self) {
+ var tail = text.slice(pos);
+ if (!self.re.no_http) {
+ // compile lazily, because "host"-containing variables can change on tlds update.
+ self.re.no_http = new RegExp("^" + self.re.src_auth +
+ // Don't allow single-level domains, because of false positives like '//test'
+ // with code comments
+ "(?:localhost|(?:(?:" + self.re.src_domain + ")\\.)+" + self.re.src_domain_root + ")" + self.re.src_port + self.re.src_host_terminator + self.re.src_path, "i");
+ }
+ if (self.re.no_http.test(tail)) {
+ // should not be `://` & `///`, that protects from errors in protocol name
+ if (pos >= 3 && text[pos - 3] === ":") {
+ return 0;
+ }
+ if (pos >= 3 && text[pos - 3] === "/") {
+ return 0;
+ }
+ return tail.match(self.re.no_http)[0].length;
+ }
+ return 0;
+ }
+ },
+ "mailto:": {
+ validate: function(text, pos, self) {
+ var tail = text.slice(pos);
+ if (!self.re.mailto) {
+ self.re.mailto = new RegExp("^" + self.re.src_email_name + "@" + self.re.src_host_strict, "i");
+ }
+ if (self.re.mailto.test(tail)) {
+ return tail.match(self.re.mailto)[0].length;
+ }
+ return 0;
+ }
+ }
+ };
+ /*eslint-disable max-len*/
+ // RE pattern for 2-character tlds (autogenerated by ./support/tlds_2char_gen.js)
+ var tlds_2ch_src_re = "a[cdefgilmnoqrstuwxz]|b[abdefghijmnorstvwyz]|c[acdfghiklmnoruvwxyz]|d[ejkmoz]|e[cegrstu]|f[ijkmor]|g[abdefghilmnpqrstuwy]|h[kmnrtu]|i[delmnoqrst]|j[emop]|k[eghimnprwyz]|l[abcikrstuvy]|m[acdeghklmnopqrstuvwxyz]|n[acefgilopruz]|om|p[aefghklmnrstwy]|qa|r[eosuw]|s[abcdeghijklmnortuvxyz]|t[cdfghjklmnortvwz]|u[agksyz]|v[aceginu]|w[fs]|y[et]|z[amw]";
+ // DON'T try to make PRs with changes. Extend TLDs with LinkifyIt.tlds() instead
+ var tlds_default = "biz|com|edu|gov|net|org|pro|web|xxx|aero|asia|coop|info|museum|name|shop|\u0440\u0444".split("|");
+ /*eslint-enable max-len*/
+ ////////////////////////////////////////////////////////////////////////////////
+ function resetScanCache(self) {
+ self.__index__ = -1;
+ self.__text_cache__ = "";
+ }
+ function createValidator(re) {
+ return function(text, pos) {
+ var tail = text.slice(pos);
+ if (re.test(tail)) {
+ return tail.match(re)[0].length;
+ }
+ return 0;
+ };
+ }
+ function createNormalizer() {
+ return function(match, self) {
+ self.normalize(match);
+ };
+ }
+ // Schemas compiler. Build regexps.
+
+ function compile(self) {
+ // Load & clone RE patterns.
+ var re$1 = self.re = re(self.__opts__);
+ // Define dynamic patterns
+ var tlds = self.__tlds__.slice();
+ self.onCompile();
+ if (!self.__tlds_replaced__) {
+ tlds.push(tlds_2ch_src_re);
+ }
+ tlds.push(re$1.src_xn);
+ re$1.src_tlds = tlds.join("|");
+ function untpl(tpl) {
+ return tpl.replace("%TLDS%", re$1.src_tlds);
+ }
+ re$1.email_fuzzy = RegExp(untpl(re$1.tpl_email_fuzzy), "i");
+ re$1.link_fuzzy = RegExp(untpl(re$1.tpl_link_fuzzy), "i");
+ re$1.link_no_ip_fuzzy = RegExp(untpl(re$1.tpl_link_no_ip_fuzzy), "i");
+ re$1.host_fuzzy_test = RegExp(untpl(re$1.tpl_host_fuzzy_test), "i");
+
+ // Compile each schema
+
+ var aliases = [];
+ self.__compiled__ = {};
+ // Reset compiled data
+ function schemaError(name, val) {
+ throw new Error('(LinkifyIt) Invalid schema "' + name + '": ' + val);
+ }
+ Object.keys(self.__schemas__).forEach((function(name) {
+ var val = self.__schemas__[name];
+ // skip disabled methods
+ if (val === null) {
+ return;
+ }
+ var compiled = {
+ validate: null,
+ link: null
+ };
+ self.__compiled__[name] = compiled;
+ if (isObject(val)) {
+ if (isRegExp(val.validate)) {
+ compiled.validate = createValidator(val.validate);
+ } else if (isFunction(val.validate)) {
+ compiled.validate = val.validate;
+ } else {
+ schemaError(name, val);
+ }
+ if (isFunction(val.normalize)) {
+ compiled.normalize = val.normalize;
+ } else if (!val.normalize) {
+ compiled.normalize = createNormalizer();
+ } else {
+ schemaError(name, val);
+ }
+ return;
+ }
+ if (isString(val)) {
+ aliases.push(name);
+ return;
+ }
+ schemaError(name, val);
+ }));
+
+ // Compile postponed aliases
+
+ aliases.forEach((function(alias) {
+ if (!self.__compiled__[self.__schemas__[alias]]) {
+ // Silently fail on missed schemas to avoid errons on disable.
+ // schemaError(alias, self.__schemas__[alias]);
+ return;
+ }
+ self.__compiled__[alias].validate = self.__compiled__[self.__schemas__[alias]].validate;
+ self.__compiled__[alias].normalize = self.__compiled__[self.__schemas__[alias]].normalize;
+ }));
+
+ // Fake record for guessed links
+
+ self.__compiled__[""] = {
+ validate: null,
+ normalize: createNormalizer()
+ };
+
+ // Build schema condition
+
+ var slist = Object.keys(self.__compiled__).filter((function(name) {
+ // Filter disabled & fake schemas
+ return name.length > 0 && self.__compiled__[name];
+ })).map(escapeRE).join("|");
+ // (?!_) cause 1.5x slowdown
+ self.re.schema_test = RegExp("(^|(?!_)(?:[><\uff5c]|" + re$1.src_ZPCc + "))(" + slist + ")", "i");
+ self.re.schema_search = RegExp("(^|(?!_)(?:[><\uff5c]|" + re$1.src_ZPCc + "))(" + slist + ")", "ig");
+ self.re.pretest = RegExp("(" + self.re.schema_test.source + ")|(" + self.re.host_fuzzy_test.source + ")|@", "i");
+
+ // Cleanup
+
+ resetScanCache(self);
+ }
+ /**
+ * class Match
+ *
+ * Match result. Single element of array, returned by [[LinkifyIt#match]]
+ **/ function Match(self, shift) {
+ var start = self.__index__, end = self.__last_index__, text = self.__text_cache__.slice(start, end);
+ /**
+ * Match#schema -> String
+ *
+ * Prefix (protocol) for matched string.
+ **/ this.schema = self.__schema__.toLowerCase();
+ /**
+ * Match#index -> Number
+ *
+ * First position of matched string.
+ **/ this.index = start + shift;
+ /**
+ * Match#lastIndex -> Number
+ *
+ * Next position after matched string.
+ **/ this.lastIndex = end + shift;
+ /**
+ * Match#raw -> String
+ *
+ * Matched string.
+ **/ this.raw = text;
+ /**
+ * Match#text -> String
+ *
+ * Notmalized text of matched string.
+ **/ this.text = text;
+ /**
+ * Match#url -> String
+ *
+ * Normalized url of matched string.
+ **/ this.url = text;
+ }
+ function createMatch(self, shift) {
+ var match = new Match(self, shift);
+ self.__compiled__[match.schema].normalize(match, self);
+ return match;
+ }
+ /**
+ * class LinkifyIt
+ **/
+ /**
+ * new LinkifyIt(schemas, options)
+ * - schemas (Object): Optional. Additional schemas to validate (prefix/validator)
+ * - options (Object): { fuzzyLink|fuzzyEmail|fuzzyIP: true|false }
+ *
+ * Creates new linkifier instance with optional additional schemas.
+ * Can be called without `new` keyword for convenience.
+ *
+ * By default understands:
+ *
+ * - `http(s)://...` , `ftp://...`, `mailto:...` & `//...` links
+ * - "fuzzy" links and emails (example.com, foo@bar.com).
+ *
+ * `schemas` is an object, where each key/value describes protocol/rule:
+ *
+ * - __key__ - link prefix (usually, protocol name with `:` at the end, `skype:`
+ * for example). `linkify-it` makes shure that prefix is not preceeded with
+ * alphanumeric char and symbols. Only whitespaces and punctuation allowed.
+ * - __value__ - rule to check tail after link prefix
+ * - _String_ - just alias to existing rule
+ * - _Object_
+ * - _validate_ - validator function (should return matched length on success),
+ * or `RegExp`.
+ * - _normalize_ - optional function to normalize text & url of matched result
+ * (for example, for @twitter mentions).
+ *
+ * `options`:
+ *
+ * - __fuzzyLink__ - recognige URL-s without `http(s):` prefix. Default `true`.
+ * - __fuzzyIP__ - allow IPs in fuzzy links above. Can conflict with some texts
+ * like version numbers. Default `false`.
+ * - __fuzzyEmail__ - recognize emails without `mailto:` prefix.
+ *
+ **/ function LinkifyIt(schemas, options) {
+ if (!(this instanceof LinkifyIt)) {
+ return new LinkifyIt(schemas, options);
+ }
+ if (!options) {
+ if (isOptionsObj(schemas)) {
+ options = schemas;
+ schemas = {};
+ }
+ }
+ this.__opts__ = assign({}, defaultOptions, options);
+ // Cache last tested result. Used to skip repeating steps on next `match` call.
+ this.__index__ = -1;
+ this.__last_index__ = -1;
+ // Next scan position
+ this.__schema__ = "";
+ this.__text_cache__ = "";
+ this.__schemas__ = assign({}, defaultSchemas, schemas);
+ this.__compiled__ = {};
+ this.__tlds__ = tlds_default;
+ this.__tlds_replaced__ = false;
+ this.re = {};
+ compile(this);
+ }
+ /** chainable
+ * LinkifyIt#add(schema, definition)
+ * - schema (String): rule name (fixed pattern prefix)
+ * - definition (String|RegExp|Object): schema definition
+ *
+ * Add new rule definition. See constructor description for details.
+ **/ LinkifyIt.prototype.add = function add(schema, definition) {
+ this.__schemas__[schema] = definition;
+ compile(this);
+ return this;
+ };
+ /** chainable
+ * LinkifyIt#set(options)
+ * - options (Object): { fuzzyLink|fuzzyEmail|fuzzyIP: true|false }
+ *
+ * Set recognition options for links without schema.
+ **/ LinkifyIt.prototype.set = function set(options) {
+ this.__opts__ = assign(this.__opts__, options);
+ return this;
+ };
+ /**
+ * LinkifyIt#test(text) -> Boolean
+ *
+ * Searches linkifiable pattern and returns `true` on success or `false` on fail.
+ **/ LinkifyIt.prototype.test = function test(text) {
+ // Reset scan cache
+ this.__text_cache__ = text;
+ this.__index__ = -1;
+ if (!text.length) {
+ return false;
+ }
+ var m, ml, me, len, shift, next, re, tld_pos, at_pos;
+ // try to scan for link with schema - that's the most simple rule
+ if (this.re.schema_test.test(text)) {
+ re = this.re.schema_search;
+ re.lastIndex = 0;
+ while ((m = re.exec(text)) !== null) {
+ len = this.testSchemaAt(text, m[2], re.lastIndex);
+ if (len) {
+ this.__schema__ = m[2];
+ this.__index__ = m.index + m[1].length;
+ this.__last_index__ = m.index + m[0].length + len;
+ break;
+ }
+ }
+ }
+ if (this.__opts__.fuzzyLink && this.__compiled__["http:"]) {
+ // guess schemaless links
+ tld_pos = text.search(this.re.host_fuzzy_test);
+ if (tld_pos >= 0) {
+ // if tld is located after found link - no need to check fuzzy pattern
+ if (this.__index__ < 0 || tld_pos < this.__index__) {
+ if ((ml = text.match(this.__opts__.fuzzyIP ? this.re.link_fuzzy : this.re.link_no_ip_fuzzy)) !== null) {
+ shift = ml.index + ml[1].length;
+ if (this.__index__ < 0 || shift < this.__index__) {
+ this.__schema__ = "";
+ this.__index__ = shift;
+ this.__last_index__ = ml.index + ml[0].length;
+ }
+ }
+ }
+ }
+ }
+ if (this.__opts__.fuzzyEmail && this.__compiled__["mailto:"]) {
+ // guess schemaless emails
+ at_pos = text.indexOf("@");
+ if (at_pos >= 0) {
+ // We can't skip this check, because this cases are possible:
+ // 192.168.1.1@gmail.com, my.in@example.com
+ if ((me = text.match(this.re.email_fuzzy)) !== null) {
+ shift = me.index + me[1].length;
+ next = me.index + me[0].length;
+ if (this.__index__ < 0 || shift < this.__index__ || shift === this.__index__ && next > this.__last_index__) {
+ this.__schema__ = "mailto:";
+ this.__index__ = shift;
+ this.__last_index__ = next;
+ }
+ }
+ }
+ }
+ return this.__index__ >= 0;
+ };
+ /**
+ * LinkifyIt#pretest(text) -> Boolean
+ *
+ * Very quick check, that can give false positives. Returns true if link MAY BE
+ * can exists. Can be used for speed optimization, when you need to check that
+ * link NOT exists.
+ **/ LinkifyIt.prototype.pretest = function pretest(text) {
+ return this.re.pretest.test(text);
+ };
+ /**
+ * LinkifyIt#testSchemaAt(text, name, position) -> Number
+ * - text (String): text to scan
+ * - name (String): rule (schema) name
+ * - position (Number): text offset to check from
+ *
+ * Similar to [[LinkifyIt#test]] but checks only specific protocol tail exactly
+ * at given position. Returns length of found pattern (0 on fail).
+ **/ LinkifyIt.prototype.testSchemaAt = function testSchemaAt(text, schema, pos) {
+ // If not supported schema check requested - terminate
+ if (!this.__compiled__[schema.toLowerCase()]) {
+ return 0;
+ }
+ return this.__compiled__[schema.toLowerCase()].validate(text, pos, this);
+ };
+ /**
+ * LinkifyIt#match(text) -> Array|null
+ *
+ * Returns array of found link descriptions or `null` on fail. We strongly
+ * recommend to use [[LinkifyIt#test]] first, for best speed.
+ *
+ * ##### Result match description
+ *
+ * - __schema__ - link schema, can be empty for fuzzy links, or `//` for
+ * protocol-neutral links.
+ * - __index__ - offset of matched text
+ * - __lastIndex__ - index of next char after mathch end
+ * - __raw__ - matched text
+ * - __text__ - normalized text
+ * - __url__ - link, generated from matched text
+ **/ LinkifyIt.prototype.match = function match(text) {
+ var shift = 0, result = [];
+ // Try to take previous element from cache, if .test() called before
+ if (this.__index__ >= 0 && this.__text_cache__ === text) {
+ result.push(createMatch(this, shift));
+ shift = this.__last_index__;
+ }
+ // Cut head if cache was used
+ var tail = shift ? text.slice(shift) : text;
+ // Scan string until end reached
+ while (this.test(tail)) {
+ result.push(createMatch(this, shift));
+ tail = tail.slice(this.__last_index__);
+ shift += this.__last_index__;
+ }
+ if (result.length) {
+ return result;
+ }
+ return null;
+ };
+ /** chainable
+ * LinkifyIt#tlds(list [, keepOld]) -> this
+ * - list (Array): list of tlds
+ * - keepOld (Boolean): merge with current list if `true` (`false` by default)
+ *
+ * Load (or merge) new tlds list. Those are user for fuzzy links (without prefix)
+ * to avoid false positives. By default this algorythm used:
+ *
+ * - hostname with any 2-letter root zones are ok.
+ * - biz|com|edu|gov|net|org|pro|web|xxx|aero|asia|coop|info|museum|name|shop|рф
+ * are ok.
+ * - encoded (`xn--...`) root zones are ok.
+ *
+ * If list is replaced, then exact match for 2-chars root zones will be checked.
+ **/ LinkifyIt.prototype.tlds = function tlds(list, keepOld) {
+ list = Array.isArray(list) ? list : [ list ];
+ if (!keepOld) {
+ this.__tlds__ = list.slice();
+ this.__tlds_replaced__ = true;
+ compile(this);
+ return this;
+ }
+ this.__tlds__ = this.__tlds__.concat(list).sort().filter((function(el, idx, arr) {
+ return el !== arr[idx - 1];
+ })).reverse();
+ compile(this);
+ return this;
+ };
+ /**
+ * LinkifyIt#normalize(match)
+ *
+ * Default normalizer (if schema does not define it's own).
+ **/ LinkifyIt.prototype.normalize = function normalize(match) {
+ // Do minimal possible changes by default. Need to collect feedback prior
+ // to move forward https://github.com/markdown-it/linkify-it/issues/1
+ if (!match.schema) {
+ match.url = "http://" + match.url;
+ }
+ if (match.schema === "mailto:" && !/^mailto:/i.test(match.url)) {
+ match.url = "mailto:" + match.url;
+ }
+ };
+ /**
+ * LinkifyIt#onCompile()
+ *
+ * Override to modify basic RegExp-s.
+ **/ LinkifyIt.prototype.onCompile = function onCompile() {};
+ var linkifyIt = LinkifyIt;
+ /*! https://mths.be/punycode v1.4.1 by @mathias */
+ /** Highest positive signed 32-bit float value */ var maxInt = 2147483647;
+ // aka. 0x7FFFFFFF or 2^31-1
+ /** Bootstring parameters */ var base = 36;
+ var tMin = 1;
+ var tMax = 26;
+ var skew = 38;
+ var damp = 700;
+ var initialBias = 72;
+ var initialN = 128;
+ // 0x80
+ var delimiter = "-";
+ // '\x2D'
+ /** Regular expressions */ var regexPunycode = /^xn--/;
+ var regexNonASCII = /[^\x20-\x7E]/;
+ // unprintable ASCII chars + non-ASCII chars
+ var regexSeparators = /[\x2E\u3002\uFF0E\uFF61]/g;
+ // RFC 3490 separators
+ /** Error messages */ var errors = {
+ overflow: "Overflow: input needs wider integers to process",
+ "not-basic": "Illegal input >= 0x80 (not a basic code point)",
+ "invalid-input": "Invalid input"
+ };
+ /** Convenience shortcuts */ var baseMinusTMin = base - tMin;
+ var floor = Math.floor;
+ var stringFromCharCode = String.fromCharCode;
+ /*--------------------------------------------------------------------------*/
+ /**
+ * A generic error utility function.
+ * @private
+ * @param {String} type The error type.
+ * @returns {Error} Throws a `RangeError` with the applicable error message.
+ */ function error(type) {
+ throw new RangeError(errors[type]);
+ }
+ /**
+ * A generic `Array#map` utility function.
+ * @private
+ * @param {Array} array The array to iterate over.
+ * @param {Function} callback The function that gets called for every array
+ * item.
+ * @returns {Array} A new array of values returned by the callback function.
+ */ function map(array, fn) {
+ var length = array.length;
+ var result = [];
+ while (length--) {
+ result[length] = fn(array[length]);
+ }
+ return result;
+ }
+ /**
+ * A simple `Array#map`-like wrapper to work with domain name strings or email
+ * addresses.
+ * @private
+ * @param {String} domain The domain name or email address.
+ * @param {Function} callback The function that gets called for every
+ * character.
+ * @returns {Array} A new string of characters returned by the callback
+ * function.
+ */ function mapDomain(string, fn) {
+ var parts = string.split("@");
+ var result = "";
+ if (parts.length > 1) {
+ // In email addresses, only the domain name should be punycoded. Leave
+ // the local part (i.e. everything up to `@`) intact.
+ result = parts[0] + "@";
+ string = parts[1];
+ }
+ // Avoid `split(regex)` for IE8 compatibility. See #17.
+ string = string.replace(regexSeparators, ".");
+ var labels = string.split(".");
+ var encoded = map(labels, fn).join(".");
+ return result + encoded;
+ }
+ /**
+ * Creates an array containing the numeric code points of each Unicode
+ * character in the string. While JavaScript uses UCS-2 internally,
+ * this function will convert a pair of surrogate halves (each of which
+ * UCS-2 exposes as separate characters) into a single code point,
+ * matching UTF-16.
+ * @see `punycode.ucs2.encode`
+ * @see <https://mathiasbynens.be/notes/javascript-encoding>
+ * @memberOf punycode.ucs2
+ * @name decode
+ * @param {String} string The Unicode input string (UCS-2).
+ * @returns {Array} The new array of code points.
+ */ function ucs2decode(string) {
+ var output = [], counter = 0, length = string.length, value, extra;
+ while (counter < length) {
+ value = string.charCodeAt(counter++);
+ if (value >= 55296 && value <= 56319 && counter < length) {
+ // high surrogate, and there is a next character
+ extra = string.charCodeAt(counter++);
+ if ((extra & 64512) == 56320) {
+ // low surrogate
+ output.push(((value & 1023) << 10) + (extra & 1023) + 65536);
+ } else {
+ // unmatched surrogate; only append this code unit, in case the next
+ // code unit is the high surrogate of a surrogate pair
+ output.push(value);
+ counter--;
+ }
+ } else {
+ output.push(value);
+ }
+ }
+ return output;
+ }
+ /**
+ * Creates a string based on an array of numeric code points.
+ * @see `punycode.ucs2.decode`
+ * @memberOf punycode.ucs2
+ * @name encode
+ * @param {Array} codePoints The array of numeric code points.
+ * @returns {String} The new Unicode string (UCS-2).
+ */ function ucs2encode(array) {
+ return map(array, (function(value) {
+ var output = "";
+ if (value > 65535) {
+ value -= 65536;
+ output += stringFromCharCode(value >>> 10 & 1023 | 55296);
+ value = 56320 | value & 1023;
+ }
+ output += stringFromCharCode(value);
+ return output;
+ })).join("");
+ }
+ /**
+ * Converts a basic code point into a digit/integer.
+ * @see `digitToBasic()`
+ * @private
+ * @param {Number} codePoint The basic numeric code point value.
+ * @returns {Number} The numeric value of a basic code point (for use in
+ * representing integers) in the range `0` to `base - 1`, or `base` if
+ * the code point does not represent a value.
+ */ function basicToDigit(codePoint) {
+ if (codePoint - 48 < 10) {
+ return codePoint - 22;
+ }
+ if (codePoint - 65 < 26) {
+ return codePoint - 65;
+ }
+ if (codePoint - 97 < 26) {
+ return codePoint - 97;
+ }
+ return base;
+ }
+ /**
+ * Converts a digit/integer into a basic code point.
+ * @see `basicToDigit()`
+ * @private
+ * @param {Number} digit The numeric value of a basic code point.
+ * @returns {Number} The basic code point whose value (when used for
+ * representing integers) is `digit`, which needs to be in the range
+ * `0` to `base - 1`. If `flag` is non-zero, the uppercase form is
+ * used; else, the lowercase form is used. The behavior is undefined
+ * if `flag` is non-zero and `digit` has no uppercase form.
+ */ function digitToBasic(digit, flag) {
+ // 0..25 map to ASCII a..z or A..Z
+ // 26..35 map to ASCII 0..9
+ return digit + 22 + 75 * (digit < 26) - ((flag != 0) << 5);
+ }
+ /**
+ * Bias adaptation function as per section 3.4 of RFC 3492.
+ * https://tools.ietf.org/html/rfc3492#section-3.4
+ * @private
+ */ function adapt(delta, numPoints, firstTime) {
+ var k = 0;
+ delta = firstTime ? floor(delta / damp) : delta >> 1;
+ delta += floor(delta / numPoints);
+ for (;delta > baseMinusTMin * tMax >> 1; k += base) {
+ delta = floor(delta / baseMinusTMin);
+ }
+ return floor(k + (baseMinusTMin + 1) * delta / (delta + skew));
+ }
+ /**
+ * Converts a Punycode string of ASCII-only symbols to a string of Unicode
+ * symbols.
+ * @memberOf punycode
+ * @param {String} input The Punycode string of ASCII-only symbols.
+ * @returns {String} The resulting string of Unicode symbols.
+ */ function decode(input) {
+ // Don't use UCS-2
+ var output = [], inputLength = input.length, out, i = 0, n = initialN, bias = initialBias, basic, j, index, oldi, w, k, digit, t,
+ /** Cached calculation results */
+ baseMinusT;
+ // Handle the basic code points: let `basic` be the number of input code
+ // points before the last delimiter, or `0` if there is none, then copy
+ // the first basic code points to the output.
+ basic = input.lastIndexOf(delimiter);
+ if (basic < 0) {
+ basic = 0;
+ }
+ for (j = 0; j < basic; ++j) {
+ // if it's not a basic code point
+ if (input.charCodeAt(j) >= 128) {
+ error("not-basic");
+ }
+ output.push(input.charCodeAt(j));
+ }
+ // Main decoding loop: start just after the last delimiter if any basic code
+ // points were copied; start at the beginning otherwise.
+ for (index = basic > 0 ? basic + 1 : 0; index < inputLength; ) {
+ // `index` is the index of the next character to be consumed.
+ // Decode a generalized variable-length integer into `delta`,
+ // which gets added to `i`. The overflow checking is easier
+ // if we increase `i` as we go, then subtract off its starting
+ // value at the end to obtain `delta`.
+ for (oldi = i, w = 1, k = base; ;k += base) {
+ if (index >= inputLength) {
+ error("invalid-input");
+ }
+ digit = basicToDigit(input.charCodeAt(index++));
+ if (digit >= base || digit > floor((maxInt - i) / w)) {
+ error("overflow");
+ }
+ i += digit * w;
+ t = k <= bias ? tMin : k >= bias + tMax ? tMax : k - bias;
+ if (digit < t) {
+ break;
+ }
+ baseMinusT = base - t;
+ if (w > floor(maxInt / baseMinusT)) {
+ error("overflow");
+ }
+ w *= baseMinusT;
+ }
+ out = output.length + 1;
+ bias = adapt(i - oldi, out, oldi == 0);
+ // `i` was supposed to wrap around from `out` to `0`,
+ // incrementing `n` each time, so we'll fix that now:
+ if (floor(i / out) > maxInt - n) {
+ error("overflow");
+ }
+ n += floor(i / out);
+ i %= out;
+ // Insert `n` at position `i` of the output
+ output.splice(i++, 0, n);
+ }
+ return ucs2encode(output);
+ }
+ /**
+ * Converts a string of Unicode symbols (e.g. a domain name label) to a
+ * Punycode string of ASCII-only symbols.
+ * @memberOf punycode
+ * @param {String} input The string of Unicode symbols.
+ * @returns {String} The resulting Punycode string of ASCII-only symbols.
+ */ function encode(input) {
+ var n, delta, handledCPCount, basicLength, bias, j, m, q, k, t, currentValue, output = [],
+ /** `inputLength` will hold the number of code points in `input`. */
+ inputLength,
+ /** Cached calculation results */
+ handledCPCountPlusOne, baseMinusT, qMinusT;
+ // Convert the input in UCS-2 to Unicode
+ input = ucs2decode(input);
+ // Cache the length
+ inputLength = input.length;
+ // Initialize the state
+ n = initialN;
+ delta = 0;
+ bias = initialBias;
+ // Handle the basic code points
+ for (j = 0; j < inputLength; ++j) {
+ currentValue = input[j];
+ if (currentValue < 128) {
+ output.push(stringFromCharCode(currentValue));
+ }
+ }
+ handledCPCount = basicLength = output.length;
+ // `handledCPCount` is the number of code points that have been handled;
+ // `basicLength` is the number of basic code points.
+ // Finish the basic string - if it is not empty - with a delimiter
+ if (basicLength) {
+ output.push(delimiter);
+ }
+ // Main encoding loop:
+ while (handledCPCount < inputLength) {
+ // All non-basic code points < n have been handled already. Find the next
+ // larger one:
+ for (m = maxInt, j = 0; j < inputLength; ++j) {
+ currentValue = input[j];
+ if (currentValue >= n && currentValue < m) {
+ m = currentValue;
+ }
+ }
+ // Increase `delta` enough to advance the decoder's <n,i> state to <m,0>,
+ // but guard against overflow
+ handledCPCountPlusOne = handledCPCount + 1;
+ if (m - n > floor((maxInt - delta) / handledCPCountPlusOne)) {
+ error("overflow");
+ }
+ delta += (m - n) * handledCPCountPlusOne;
+ n = m;
+ for (j = 0; j < inputLength; ++j) {
+ currentValue = input[j];
+ if (currentValue < n && ++delta > maxInt) {
+ error("overflow");
+ }
+ if (currentValue == n) {
+ // Represent delta as a generalized variable-length integer
+ for (q = delta, k = base; ;k += base) {
+ t = k <= bias ? tMin : k >= bias + tMax ? tMax : k - bias;
+ if (q < t) {
+ break;
+ }
+ qMinusT = q - t;
+ baseMinusT = base - t;
+ output.push(stringFromCharCode(digitToBasic(t + qMinusT % baseMinusT, 0)));
+ q = floor(qMinusT / baseMinusT);
+ }
+ output.push(stringFromCharCode(digitToBasic(q, 0)));
+ bias = adapt(delta, handledCPCountPlusOne, handledCPCount == basicLength);
+ delta = 0;
+ ++handledCPCount;
+ }
+ }
+ ++delta;
+ ++n;
+ }
+ return output.join("");
+ }
+ /**
+ * Converts a Punycode string representing a domain name or an email address
+ * to Unicode. Only the Punycoded parts of the input will be converted, i.e.
+ * it doesn't matter if you call it on a string that has already been
+ * converted to Unicode.
+ * @memberOf punycode
+ * @param {String} input The Punycoded domain name or email address to
+ * convert to Unicode.
+ * @returns {String} The Unicode representation of the given Punycode
+ * string.
+ */ function toUnicode(input) {
+ return mapDomain(input, (function(string) {
+ return regexPunycode.test(string) ? decode(string.slice(4).toLowerCase()) : string;
+ }));
+ }
+ /**
+ * Converts a Unicode string representing a domain name or an email address to
+ * Punycode. Only the non-ASCII parts of the domain name will be converted,
+ * i.e. it doesn't matter if you call it with a domain that's already in
+ * ASCII.
+ * @memberOf punycode
+ * @param {String} input The domain name or email address to convert, as a
+ * Unicode string.
+ * @returns {String} The Punycode representation of the given domain name or
+ * email address.
+ */ function toASCII(input) {
+ return mapDomain(input, (function(string) {
+ return regexNonASCII.test(string) ? "xn--" + encode(string) : string;
+ }));
+ }
+ var version = "1.4.1";
+ /**
+ * An object of methods to convert from JavaScript's internal character
+ * representation (UCS-2) to Unicode code points, and back.
+ * @see <https://mathiasbynens.be/notes/javascript-encoding>
+ * @memberOf punycode
+ * @type Object
+ */ var ucs2 = {
+ decode: ucs2decode,
+ encode: ucs2encode
+ };
+ var punycode$1 = {
+ version: version,
+ ucs2: ucs2,
+ toASCII: toASCII,
+ toUnicode: toUnicode,
+ encode: encode,
+ decode: decode
+ };
+ var punycode$2 = Object.freeze({
+ __proto__: null,
+ decode: decode,
+ encode: encode,
+ toUnicode: toUnicode,
+ toASCII: toASCII,
+ version: version,
+ ucs2: ucs2,
+ default: punycode$1
+ });
+ // markdown-it default options
+ var _default = {
+ options: {
+ html: false,
+ // Enable HTML tags in source
+ xhtmlOut: false,
+ // Use '/' to close single tags (<br />)
+ breaks: false,
+ // Convert '\n' in paragraphs into <br>
+ langPrefix: "language-",
+ // CSS language prefix for fenced blocks
+ linkify: false,
+ // autoconvert URL-like texts to links
+ // Enable some language-neutral replacements + quotes beautification
+ typographer: false,
+ // Double + single quotes replacement pairs, when typographer enabled,
+ // and smartquotes on. Could be either a String or an Array.
+ // For example, you can use '«»„“' for Russian, '„“‚‘' for German,
+ // and ['«\xA0', '\xA0»', '‹\xA0', '\xA0›'] for French (including nbsp).
+ quotes: "\u201c\u201d\u2018\u2019",
+ /* “”‘’ */
+ // Highlighter function. Should return escaped HTML,
+ // or '' if the source string is not changed and should be escaped externaly.
+ // If result starts with <pre... internal wrapper is skipped.
+ // function (/*str, lang*/) { return ''; }
+ highlight: null,
+ maxNesting: 100
+ },
+ components: {
+ core: {},
+ block: {},
+ inline: {}
+ }
+ };
+ // "Zero" preset, with nothing enabled. Useful for manual configuring of simple
+ var zero = {
+ options: {
+ html: false,
+ // Enable HTML tags in source
+ xhtmlOut: false,
+ // Use '/' to close single tags (<br />)
+ breaks: false,
+ // Convert '\n' in paragraphs into <br>
+ langPrefix: "language-",
+ // CSS language prefix for fenced blocks
+ linkify: false,
+ // autoconvert URL-like texts to links
+ // Enable some language-neutral replacements + quotes beautification
+ typographer: false,
+ // Double + single quotes replacement pairs, when typographer enabled,
+ // and smartquotes on. Could be either a String or an Array.
+ // For example, you can use '«»„“' for Russian, '„“‚‘' for German,
+ // and ['«\xA0', '\xA0»', '‹\xA0', '\xA0›'] for French (including nbsp).
+ quotes: "\u201c\u201d\u2018\u2019",
+ /* “”‘’ */
+ // Highlighter function. Should return escaped HTML,
+ // or '' if the source string is not changed and should be escaped externaly.
+ // If result starts with <pre... internal wrapper is skipped.
+ // function (/*str, lang*/) { return ''; }
+ highlight: null,
+ maxNesting: 20
+ },
+ components: {
+ core: {
+ rules: [ "normalize", "block", "inline" ]
+ },
+ block: {
+ rules: [ "paragraph" ]
+ },
+ inline: {
+ rules: [ "text" ],
+ rules2: [ "balance_pairs", "text_collapse" ]
+ }
+ }
+ };
+ // Commonmark default options
+ var commonmark = {
+ options: {
+ html: true,
+ // Enable HTML tags in source
+ xhtmlOut: true,
+ // Use '/' to close single tags (<br />)
+ breaks: false,
+ // Convert '\n' in paragraphs into <br>
+ langPrefix: "language-",
+ // CSS language prefix for fenced blocks
+ linkify: false,
+ // autoconvert URL-like texts to links
+ // Enable some language-neutral replacements + quotes beautification
+ typographer: false,
+ // Double + single quotes replacement pairs, when typographer enabled,
+ // and smartquotes on. Could be either a String or an Array.
+ // For example, you can use '«»„“' for Russian, '„“‚‘' for German,
+ // and ['«\xA0', '\xA0»', '‹\xA0', '\xA0›'] for French (including nbsp).
+ quotes: "\u201c\u201d\u2018\u2019",
+ /* “”‘’ */
+ // Highlighter function. Should return escaped HTML,
+ // or '' if the source string is not changed and should be escaped externaly.
+ // If result starts with <pre... internal wrapper is skipped.
+ // function (/*str, lang*/) { return ''; }
+ highlight: null,
+ maxNesting: 20
+ },
+ components: {
+ core: {
+ rules: [ "normalize", "block", "inline" ]
+ },
+ block: {
+ rules: [ "blockquote", "code", "fence", "heading", "hr", "html_block", "lheading", "list", "reference", "paragraph" ]
+ },
+ inline: {
+ rules: [ "autolink", "backticks", "emphasis", "entity", "escape", "html_inline", "image", "link", "newline", "text" ],
+ rules2: [ "balance_pairs", "emphasis", "text_collapse" ]
+ }
+ }
+ };
+ var punycode = getAugmentedNamespace(punycode$2);
+ var config = {
+ default: _default,
+ zero: zero,
+ commonmark: commonmark
+ };
+ ////////////////////////////////////////////////////////////////////////////////
+
+ // This validator can prohibit more than really needed to prevent XSS. It's a
+ // tradeoff to keep code simple and to be secure by default.
+
+ // If you need different setup - override validator method as you wish. Or
+ // replace it with dummy function and use external sanitizer.
+
+ var BAD_PROTO_RE = /^(vbscript|javascript|file|data):/;
+ var GOOD_DATA_RE = /^data:image\/(gif|png|jpeg|webp);/;
+ function validateLink(url) {
+ // url should be normalized at this point, and existing entities are decoded
+ var str = url.trim().toLowerCase();
+ return BAD_PROTO_RE.test(str) ? GOOD_DATA_RE.test(str) ? true : false : true;
+ }
+ ////////////////////////////////////////////////////////////////////////////////
+ var RECODE_HOSTNAME_FOR = [ "http:", "https:", "mailto:" ];
+ function normalizeLink(url) {
+ var parsed = mdurl.parse(url, true);
+ if (parsed.hostname) {
+ // Encode hostnames in urls like:
+ // `http://host/`, `https://host/`, `mailto:user@host`, `//host/`
+ // We don't encode unknown schemas, because it's likely that we encode
+ // something we shouldn't (e.g. `skype:name` treated as `skype:host`)
+ if (!parsed.protocol || RECODE_HOSTNAME_FOR.indexOf(parsed.protocol) >= 0) {
+ try {
+ parsed.hostname = punycode.toASCII(parsed.hostname);
+ } catch (er) {}
+ }
+ }
+ return mdurl.encode(mdurl.format(parsed));
+ }
+ function normalizeLinkText(url) {
+ var parsed = mdurl.parse(url, true);
+ if (parsed.hostname) {
+ // Encode hostnames in urls like:
+ // `http://host/`, `https://host/`, `mailto:user@host`, `//host/`
+ // We don't encode unknown schemas, because it's likely that we encode
+ // something we shouldn't (e.g. `skype:name` treated as `skype:host`)
+ if (!parsed.protocol || RECODE_HOSTNAME_FOR.indexOf(parsed.protocol) >= 0) {
+ try {
+ parsed.hostname = punycode.toUnicode(parsed.hostname);
+ } catch (er) {}
+ }
+ }
+ // add '%' to exclude list because of https://github.com/markdown-it/markdown-it/issues/720
+ return mdurl.decode(mdurl.format(parsed), mdurl.decode.defaultChars + "%");
+ }
+ /**
+ * class MarkdownIt
+ *
+ * Main parser/renderer class.
+ *
+ * ##### Usage
+ *
+ * ```javascript
+ * // node.js, "classic" way:
+ * var MarkdownIt = require('markdown-it'),
+ * md = new MarkdownIt();
+ * var result = md.render('# markdown-it rulezz!');
+ *
+ * // node.js, the same, but with sugar:
+ * var md = require('markdown-it')();
+ * var result = md.render('# markdown-it rulezz!');
+ *
+ * // browser without AMD, added to "window" on script load
+ * // Note, there are no dash.
+ * var md = window.markdownit();
+ * var result = md.render('# markdown-it rulezz!');
+ * ```
+ *
+ * Single line rendering, without paragraph wrap:
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ * var result = md.renderInline('__markdown-it__ rulezz!');
+ * ```
+ **/
+ /**
+ * new MarkdownIt([presetName, options])
+ * - presetName (String): optional, `commonmark` / `zero`
+ * - options (Object)
+ *
+ * Creates parser instanse with given config. Can be called without `new`.
+ *
+ * ##### presetName
+ *
+ * MarkdownIt provides named presets as a convenience to quickly
+ * enable/disable active syntax rules and options for common use cases.
+ *
+ * - ["commonmark"](https://github.com/markdown-it/markdown-it/blob/master/lib/presets/commonmark.js) -
+ * configures parser to strict [CommonMark](http://commonmark.org/) mode.
+ * - [default](https://github.com/markdown-it/markdown-it/blob/master/lib/presets/default.js) -
+ * similar to GFM, used when no preset name given. Enables all available rules,
+ * but still without html, typographer & autolinker.
+ * - ["zero"](https://github.com/markdown-it/markdown-it/blob/master/lib/presets/zero.js) -
+ * all rules disabled. Useful to quickly setup your config via `.enable()`.
+ * For example, when you need only `bold` and `italic` markup and nothing else.
+ *
+ * ##### options:
+ *
+ * - __html__ - `false`. Set `true` to enable HTML tags in source. Be careful!
+ * That's not safe! You may need external sanitizer to protect output from XSS.
+ * It's better to extend features via plugins, instead of enabling HTML.
+ * - __xhtmlOut__ - `false`. Set `true` to add '/' when closing single tags
+ * (`<br />`). This is needed only for full CommonMark compatibility. In real
+ * world you will need HTML output.
+ * - __breaks__ - `false`. Set `true` to convert `\n` in paragraphs into `<br>`.
+ * - __langPrefix__ - `language-`. CSS language class prefix for fenced blocks.
+ * Can be useful for external highlighters.
+ * - __linkify__ - `false`. Set `true` to autoconvert URL-like text to links.
+ * - __typographer__ - `false`. Set `true` to enable [some language-neutral
+ * replacement](https://github.com/markdown-it/markdown-it/blob/master/lib/rules_core/replacements.js) +
+ * quotes beautification (smartquotes).
+ * - __quotes__ - `“”‘’`, String or Array. Double + single quotes replacement
+ * pairs, when typographer enabled and smartquotes on. For example, you can
+ * use `'«»„“'` for Russian, `'„“‚‘'` for German, and
+ * `['«\xA0', '\xA0»', '‹\xA0', '\xA0›']` for French (including nbsp).
+ * - __highlight__ - `null`. Highlighter function for fenced code blocks.
+ * Highlighter `function (str, lang)` should return escaped HTML. It can also
+ * return empty string if the source was not changed and should be escaped
+ * externaly. If result starts with <pre... internal wrapper is skipped.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * // commonmark mode
+ * var md = require('markdown-it')('commonmark');
+ *
+ * // default mode
+ * var md = require('markdown-it')();
+ *
+ * // enable everything
+ * var md = require('markdown-it')({
+ * html: true,
+ * linkify: true,
+ * typographer: true
+ * });
+ * ```
+ *
+ * ##### Syntax highlighting
+ *
+ * ```js
+ * var hljs = require('highlight.js') // https://highlightjs.org/
+ *
+ * var md = require('markdown-it')({
+ * highlight: function (str, lang) {
+ * if (lang && hljs.getLanguage(lang)) {
+ * try {
+ * return hljs.highlight(str, { language: lang, ignoreIllegals: true }).value;
+ * } catch (__) {}
+ * }
+ *
+ * return ''; // use external default escaping
+ * }
+ * });
+ * ```
+ *
+ * Or with full wrapper override (if you need assign class to `<pre>`):
+ *
+ * ```javascript
+ * var hljs = require('highlight.js') // https://highlightjs.org/
+ *
+ * // Actual default values
+ * var md = require('markdown-it')({
+ * highlight: function (str, lang) {
+ * if (lang && hljs.getLanguage(lang)) {
+ * try {
+ * return '<pre class="hljs"><code>' +
+ * hljs.highlight(str, { language: lang, ignoreIllegals: true }).value +
+ * '</code></pre>';
+ * } catch (__) {}
+ * }
+ *
+ * return '<pre class="hljs"><code>' + md.utils.escapeHtml(str) + '</code></pre>';
+ * }
+ * });
+ * ```
+ *
+ **/ function MarkdownIt(presetName, options) {
+ if (!(this instanceof MarkdownIt)) {
+ return new MarkdownIt(presetName, options);
+ }
+ if (!options) {
+ if (!utils.isString(presetName)) {
+ options = presetName || {};
+ presetName = "default";
+ }
+ }
+ /**
+ * MarkdownIt#inline -> ParserInline
+ *
+ * Instance of [[ParserInline]]. You may need it to add new rules when
+ * writing plugins. For simple rules control use [[MarkdownIt.disable]] and
+ * [[MarkdownIt.enable]].
+ **/ this.inline = new parser_inline;
+ /**
+ * MarkdownIt#block -> ParserBlock
+ *
+ * Instance of [[ParserBlock]]. You may need it to add new rules when
+ * writing plugins. For simple rules control use [[MarkdownIt.disable]] and
+ * [[MarkdownIt.enable]].
+ **/ this.block = new parser_block;
+ /**
+ * MarkdownIt#core -> Core
+ *
+ * Instance of [[Core]] chain executor. You may need it to add new rules when
+ * writing plugins. For simple rules control use [[MarkdownIt.disable]] and
+ * [[MarkdownIt.enable]].
+ **/ this.core = new parser_core;
+ /**
+ * MarkdownIt#renderer -> Renderer
+ *
+ * Instance of [[Renderer]]. Use it to modify output look. Or to add rendering
+ * rules for new token types, generated by plugins.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ *
+ * function myToken(tokens, idx, options, env, self) {
+ * //...
+ * return result;
+ * };
+ *
+ * md.renderer.rules['my_token'] = myToken
+ * ```
+ *
+ * See [[Renderer]] docs and [source code](https://github.com/markdown-it/markdown-it/blob/master/lib/renderer.js).
+ **/ this.renderer = new renderer;
+ /**
+ * MarkdownIt#linkify -> LinkifyIt
+ *
+ * [linkify-it](https://github.com/markdown-it/linkify-it) instance.
+ * Used by [linkify](https://github.com/markdown-it/markdown-it/blob/master/lib/rules_core/linkify.js)
+ * rule.
+ **/ this.linkify = new linkifyIt;
+ /**
+ * MarkdownIt#validateLink(url) -> Boolean
+ *
+ * Link validation function. CommonMark allows too much in links. By default
+ * we disable `javascript:`, `vbscript:`, `file:` schemas, and almost all `data:...` schemas
+ * except some embedded image types.
+ *
+ * You can change this behaviour:
+ *
+ * ```javascript
+ * var md = require('markdown-it')();
+ * // enable everything
+ * md.validateLink = function () { return true; }
+ * ```
+ **/ this.validateLink = validateLink;
+ /**
+ * MarkdownIt#normalizeLink(url) -> String
+ *
+ * Function used to encode link url to a machine-readable format,
+ * which includes url-encoding, punycode, etc.
+ **/ this.normalizeLink = normalizeLink;
+ /**
+ * MarkdownIt#normalizeLinkText(url) -> String
+ *
+ * Function used to decode link url to a human-readable format`
+ **/ this.normalizeLinkText = normalizeLinkText;
+ // Expose utils & helpers for easy acces from plugins
+ /**
+ * MarkdownIt#utils -> utils
+ *
+ * Assorted utility functions, useful to write plugins. See details
+ * [here](https://github.com/markdown-it/markdown-it/blob/master/lib/common/utils.js).
+ **/ this.utils = utils;
+ /**
+ * MarkdownIt#helpers -> helpers
+ *
+ * Link components parser functions, useful to write plugins. See details
+ * [here](https://github.com/markdown-it/markdown-it/blob/master/lib/helpers).
+ **/ this.helpers = utils.assign({}, helpers);
+ this.options = {};
+ this.configure(presetName);
+ if (options) {
+ this.set(options);
+ }
+ }
+ /** chainable
+ * MarkdownIt.set(options)
+ *
+ * Set parser options (in the same format as in constructor). Probably, you
+ * will never need it, but you can change options after constructor call.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')()
+ * .set({ html: true, breaks: true })
+ * .set({ typographer, true });
+ * ```
+ *
+ * __Note:__ To achieve the best possible performance, don't modify a
+ * `markdown-it` instance options on the fly. If you need multiple configurations
+ * it's best to create multiple instances and initialize each with separate
+ * config.
+ **/ MarkdownIt.prototype.set = function(options) {
+ utils.assign(this.options, options);
+ return this;
+ };
+ /** chainable, internal
+ * MarkdownIt.configure(presets)
+ *
+ * Batch load of all options and compenent settings. This is internal method,
+ * and you probably will not need it. But if you will - see available presets
+ * and data structure [here](https://github.com/markdown-it/markdown-it/tree/master/lib/presets)
+ *
+ * We strongly recommend to use presets instead of direct config loads. That
+ * will give better compatibility with next versions.
+ **/ MarkdownIt.prototype.configure = function(presets) {
+ var self = this, presetName;
+ if (utils.isString(presets)) {
+ presetName = presets;
+ presets = config[presetName];
+ if (!presets) {
+ throw new Error('Wrong `markdown-it` preset "' + presetName + '", check name');
+ }
+ }
+ if (!presets) {
+ throw new Error("Wrong `markdown-it` preset, can't be empty");
+ }
+ if (presets.options) {
+ self.set(presets.options);
+ }
+ if (presets.components) {
+ Object.keys(presets.components).forEach((function(name) {
+ if (presets.components[name].rules) {
+ self[name].ruler.enableOnly(presets.components[name].rules);
+ }
+ if (presets.components[name].rules2) {
+ self[name].ruler2.enableOnly(presets.components[name].rules2);
+ }
+ }));
+ }
+ return this;
+ };
+ /** chainable
+ * MarkdownIt.enable(list, ignoreInvalid)
+ * - list (String|Array): rule name or list of rule names to enable
+ * - ignoreInvalid (Boolean): set `true` to ignore errors when rule not found.
+ *
+ * Enable list or rules. It will automatically find appropriate components,
+ * containing rules with given names. If rule not found, and `ignoreInvalid`
+ * not set - throws exception.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var md = require('markdown-it')()
+ * .enable(['sub', 'sup'])
+ * .disable('smartquotes');
+ * ```
+ **/ MarkdownIt.prototype.enable = function(list, ignoreInvalid) {
+ var result = [];
+ if (!Array.isArray(list)) {
+ list = [ list ];
+ }
+ [ "core", "block", "inline" ].forEach((function(chain) {
+ result = result.concat(this[chain].ruler.enable(list, true));
+ }), this);
+ result = result.concat(this.inline.ruler2.enable(list, true));
+ var missed = list.filter((function(name) {
+ return result.indexOf(name) < 0;
+ }));
+ if (missed.length && !ignoreInvalid) {
+ throw new Error("MarkdownIt. Failed to enable unknown rule(s): " + missed);
+ }
+ return this;
+ };
+ /** chainable
+ * MarkdownIt.disable(list, ignoreInvalid)
+ * - list (String|Array): rule name or list of rule names to disable.
+ * - ignoreInvalid (Boolean): set `true` to ignore errors when rule not found.
+ *
+ * The same as [[MarkdownIt.enable]], but turn specified rules off.
+ **/ MarkdownIt.prototype.disable = function(list, ignoreInvalid) {
+ var result = [];
+ if (!Array.isArray(list)) {
+ list = [ list ];
+ }
+ [ "core", "block", "inline" ].forEach((function(chain) {
+ result = result.concat(this[chain].ruler.disable(list, true));
+ }), this);
+ result = result.concat(this.inline.ruler2.disable(list, true));
+ var missed = list.filter((function(name) {
+ return result.indexOf(name) < 0;
+ }));
+ if (missed.length && !ignoreInvalid) {
+ throw new Error("MarkdownIt. Failed to disable unknown rule(s): " + missed);
+ }
+ return this;
+ };
+ /** chainable
+ * MarkdownIt.use(plugin, params)
+ *
+ * Load specified plugin with given params into current parser instance.
+ * It's just a sugar to call `plugin(md, params)` with curring.
+ *
+ * ##### Example
+ *
+ * ```javascript
+ * var iterator = require('markdown-it-for-inline');
+ * var md = require('markdown-it')()
+ * .use(iterator, 'foo_replace', 'text', function (tokens, idx) {
+ * tokens[idx].content = tokens[idx].content.replace(/foo/g, 'bar');
+ * });
+ * ```
+ **/ MarkdownIt.prototype.use = function(plugin /*, params, ... */) {
+ var args = [ this ].concat(Array.prototype.slice.call(arguments, 1));
+ plugin.apply(plugin, args);
+ return this;
+ };
+ /** internal
+ * MarkdownIt.parse(src, env) -> Array
+ * - src (String): source string
+ * - env (Object): environment sandbox
+ *
+ * Parse input string and return list of block tokens (special token type
+ * "inline" will contain list of inline tokens). You should not call this
+ * method directly, until you write custom renderer (for example, to produce
+ * AST).
+ *
+ * `env` is used to pass data between "distributed" rules and return additional
+ * metadata like reference info, needed for the renderer. It also can be used to
+ * inject data in specific cases. Usually, you will be ok to pass `{}`,
+ * and then pass updated object to renderer.
+ **/ MarkdownIt.prototype.parse = function(src, env) {
+ if (typeof src !== "string") {
+ throw new Error("Input data should be a String");
+ }
+ var state = new this.core.State(src, this, env);
+ this.core.process(state);
+ return state.tokens;
+ };
+ /**
+ * MarkdownIt.render(src [, env]) -> String
+ * - src (String): source string
+ * - env (Object): environment sandbox
+ *
+ * Render markdown string into html. It does all magic for you :).
+ *
+ * `env` can be used to inject additional metadata (`{}` by default).
+ * But you will not need it with high probability. See also comment
+ * in [[MarkdownIt.parse]].
+ **/ MarkdownIt.prototype.render = function(src, env) {
+ env = env || {};
+ return this.renderer.render(this.parse(src, env), this.options, env);
+ };
+ /** internal
+ * MarkdownIt.parseInline(src, env) -> Array
+ * - src (String): source string
+ * - env (Object): environment sandbox
+ *
+ * The same as [[MarkdownIt.parse]] but skip all block rules. It returns the
+ * block tokens list with the single `inline` element, containing parsed inline
+ * tokens in `children` property. Also updates `env` object.
+ **/ MarkdownIt.prototype.parseInline = function(src, env) {
+ var state = new this.core.State(src, this, env);
+ state.inlineMode = true;
+ this.core.process(state);
+ return state.tokens;
+ };
+ /**
+ * MarkdownIt.renderInline(src [, env]) -> String
+ * - src (String): source string
+ * - env (Object): environment sandbox
+ *
+ * Similar to [[MarkdownIt.render]] but for single paragraph content. Result
+ * will NOT be wrapped into `<p>` tags.
+ **/ MarkdownIt.prototype.renderInline = function(src, env) {
+ env = env || {};
+ return this.renderer.render(this.parseInline(src, env), this.options, env);
+ };
+ var lib = MarkdownIt;
+ var markdownIt = lib;
+ return markdownIt;
+}));