661 lines
24 KiB
JavaScript
661 lines
24 KiB
JavaScript
import { c as getConfig, g as getAccTitle, s as setAccTitle, a as getAccDescription, b as setAccDescription, r as getDiagramTitle, q as setDiagramTitle, t as clear$1, e as common, I as defaultConfig, i as configureSvgSize } from "./mermaid-0d192ec3.js";
|
|
import { select, scaleOrdinal, schemeTableau10 } from "d3";
|
|
import { sankey, sankeyLinkHorizontal, sankeyLeft, sankeyRight, sankeyCenter, sankeyJustify } from "d3-sankey";
|
|
import "ts-dedent";
|
|
import "dayjs";
|
|
import "@braintree/sanitize-url";
|
|
import "dompurify";
|
|
import "khroma";
|
|
import "lodash-es/memoize.js";
|
|
import "lodash-es/merge.js";
|
|
import "stylis";
|
|
import "lodash-es/isEmpty.js";
|
|
var parser = function() {
|
|
var o = function(k, v, o2, l) {
|
|
for (o2 = o2 || {}, l = k.length; l--; o2[k[l]] = v)
|
|
;
|
|
return o2;
|
|
}, $V0 = [1, 9], $V1 = [1, 10], $V2 = [1, 5, 10, 12];
|
|
var parser2 = {
|
|
trace: function trace() {
|
|
},
|
|
yy: {},
|
|
symbols_: { "error": 2, "start": 3, "SANKEY": 4, "NEWLINE": 5, "csv": 6, "opt_eof": 7, "record": 8, "csv_tail": 9, "EOF": 10, "field[source]": 11, "COMMA": 12, "field[target]": 13, "field[value]": 14, "field": 15, "escaped": 16, "non_escaped": 17, "DQUOTE": 18, "ESCAPED_TEXT": 19, "NON_ESCAPED_TEXT": 20, "$accept": 0, "$end": 1 },
|
|
terminals_: { 2: "error", 4: "SANKEY", 5: "NEWLINE", 10: "EOF", 11: "field[source]", 12: "COMMA", 13: "field[target]", 14: "field[value]", 18: "DQUOTE", 19: "ESCAPED_TEXT", 20: "NON_ESCAPED_TEXT" },
|
|
productions_: [0, [3, 4], [6, 2], [9, 2], [9, 0], [7, 1], [7, 0], [8, 5], [15, 1], [15, 1], [16, 3], [17, 1]],
|
|
performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate, $$, _$) {
|
|
var $0 = $$.length - 1;
|
|
switch (yystate) {
|
|
case 7:
|
|
const source = yy.findOrCreateNode($$[$0 - 4].trim().replaceAll('""', '"'));
|
|
const target = yy.findOrCreateNode($$[$0 - 2].trim().replaceAll('""', '"'));
|
|
const value = parseFloat($$[$0].trim());
|
|
yy.addLink(source, target, value);
|
|
break;
|
|
case 8:
|
|
case 9:
|
|
case 11:
|
|
this.$ = $$[$0];
|
|
break;
|
|
case 10:
|
|
this.$ = $$[$0 - 1];
|
|
break;
|
|
}
|
|
},
|
|
table: [{ 3: 1, 4: [1, 2] }, { 1: [3] }, { 5: [1, 3] }, { 6: 4, 8: 5, 15: 6, 16: 7, 17: 8, 18: $V0, 20: $V1 }, { 1: [2, 6], 7: 11, 10: [1, 12] }, o($V1, [2, 4], { 9: 13, 5: [1, 14] }), { 12: [1, 15] }, o($V2, [2, 8]), o($V2, [2, 9]), { 19: [1, 16] }, o($V2, [2, 11]), { 1: [2, 1] }, { 1: [2, 5] }, o($V1, [2, 2]), { 6: 17, 8: 5, 15: 6, 16: 7, 17: 8, 18: $V0, 20: $V1 }, { 15: 18, 16: 7, 17: 8, 18: $V0, 20: $V1 }, { 18: [1, 19] }, o($V1, [2, 3]), { 12: [1, 20] }, o($V2, [2, 10]), { 15: 21, 16: 7, 17: 8, 18: $V0, 20: $V1 }, o([1, 5, 10], [2, 7])],
|
|
defaultActions: { 11: [2, 1], 12: [2, 5] },
|
|
parseError: function parseError(str, hash) {
|
|
if (hash.recoverable) {
|
|
this.trace(str);
|
|
} else {
|
|
var error = new Error(str);
|
|
error.hash = hash;
|
|
throw error;
|
|
}
|
|
},
|
|
parse: function parse(input) {
|
|
var self = this, stack = [0], tstack = [], vstack = [null], lstack = [], table = this.table, yytext = "", yylineno = 0, yyleng = 0, TERROR = 2, EOF = 1;
|
|
var args = lstack.slice.call(arguments, 1);
|
|
var lexer2 = Object.create(this.lexer);
|
|
var sharedState = { yy: {} };
|
|
for (var k in this.yy) {
|
|
if (Object.prototype.hasOwnProperty.call(this.yy, k)) {
|
|
sharedState.yy[k] = this.yy[k];
|
|
}
|
|
}
|
|
lexer2.setInput(input, sharedState.yy);
|
|
sharedState.yy.lexer = lexer2;
|
|
sharedState.yy.parser = this;
|
|
if (typeof lexer2.yylloc == "undefined") {
|
|
lexer2.yylloc = {};
|
|
}
|
|
var yyloc = lexer2.yylloc;
|
|
lstack.push(yyloc);
|
|
var ranges = lexer2.options && lexer2.options.ranges;
|
|
if (typeof sharedState.yy.parseError === "function") {
|
|
this.parseError = sharedState.yy.parseError;
|
|
} else {
|
|
this.parseError = Object.getPrototypeOf(this).parseError;
|
|
}
|
|
function lex() {
|
|
var token;
|
|
token = tstack.pop() || lexer2.lex() || EOF;
|
|
if (typeof token !== "number") {
|
|
if (token instanceof Array) {
|
|
tstack = token;
|
|
token = tstack.pop();
|
|
}
|
|
token = self.symbols_[token] || token;
|
|
}
|
|
return token;
|
|
}
|
|
var symbol, state, action, r, yyval = {}, p, len, newState, expected;
|
|
while (true) {
|
|
state = stack[stack.length - 1];
|
|
if (this.defaultActions[state]) {
|
|
action = this.defaultActions[state];
|
|
} else {
|
|
if (symbol === null || typeof symbol == "undefined") {
|
|
symbol = lex();
|
|
}
|
|
action = table[state] && table[state][symbol];
|
|
}
|
|
if (typeof action === "undefined" || !action.length || !action[0]) {
|
|
var errStr = "";
|
|
expected = [];
|
|
for (p in table[state]) {
|
|
if (this.terminals_[p] && p > TERROR) {
|
|
expected.push("'" + this.terminals_[p] + "'");
|
|
}
|
|
}
|
|
if (lexer2.showPosition) {
|
|
errStr = "Parse error on line " + (yylineno + 1) + ":\n" + lexer2.showPosition() + "\nExpecting " + expected.join(", ") + ", got '" + (this.terminals_[symbol] || symbol) + "'";
|
|
} else {
|
|
errStr = "Parse error on line " + (yylineno + 1) + ": Unexpected " + (symbol == EOF ? "end of input" : "'" + (this.terminals_[symbol] || symbol) + "'");
|
|
}
|
|
this.parseError(errStr, {
|
|
text: lexer2.match,
|
|
token: this.terminals_[symbol] || symbol,
|
|
line: lexer2.yylineno,
|
|
loc: yyloc,
|
|
expected
|
|
});
|
|
}
|
|
if (action[0] instanceof Array && action.length > 1) {
|
|
throw new Error("Parse Error: multiple actions possible at state: " + state + ", token: " + symbol);
|
|
}
|
|
switch (action[0]) {
|
|
case 1:
|
|
stack.push(symbol);
|
|
vstack.push(lexer2.yytext);
|
|
lstack.push(lexer2.yylloc);
|
|
stack.push(action[1]);
|
|
symbol = null;
|
|
{
|
|
yyleng = lexer2.yyleng;
|
|
yytext = lexer2.yytext;
|
|
yylineno = lexer2.yylineno;
|
|
yyloc = lexer2.yylloc;
|
|
}
|
|
break;
|
|
case 2:
|
|
len = this.productions_[action[1]][1];
|
|
yyval.$ = vstack[vstack.length - len];
|
|
yyval._$ = {
|
|
first_line: lstack[lstack.length - (len || 1)].first_line,
|
|
last_line: lstack[lstack.length - 1].last_line,
|
|
first_column: lstack[lstack.length - (len || 1)].first_column,
|
|
last_column: lstack[lstack.length - 1].last_column
|
|
};
|
|
if (ranges) {
|
|
yyval._$.range = [
|
|
lstack[lstack.length - (len || 1)].range[0],
|
|
lstack[lstack.length - 1].range[1]
|
|
];
|
|
}
|
|
r = this.performAction.apply(yyval, [
|
|
yytext,
|
|
yyleng,
|
|
yylineno,
|
|
sharedState.yy,
|
|
action[1],
|
|
vstack,
|
|
lstack
|
|
].concat(args));
|
|
if (typeof r !== "undefined") {
|
|
return r;
|
|
}
|
|
if (len) {
|
|
stack = stack.slice(0, -1 * len * 2);
|
|
vstack = vstack.slice(0, -1 * len);
|
|
lstack = lstack.slice(0, -1 * len);
|
|
}
|
|
stack.push(this.productions_[action[1]][0]);
|
|
vstack.push(yyval.$);
|
|
lstack.push(yyval._$);
|
|
newState = table[stack[stack.length - 2]][stack[stack.length - 1]];
|
|
stack.push(newState);
|
|
break;
|
|
case 3:
|
|
return true;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
};
|
|
var lexer = function() {
|
|
var lexer2 = {
|
|
EOF: 1,
|
|
parseError: function parseError(str, hash) {
|
|
if (this.yy.parser) {
|
|
this.yy.parser.parseError(str, hash);
|
|
} else {
|
|
throw new Error(str);
|
|
}
|
|
},
|
|
// resets the lexer, sets new input
|
|
setInput: function(input, yy) {
|
|
this.yy = yy || this.yy || {};
|
|
this._input = input;
|
|
this._more = this._backtrack = this.done = false;
|
|
this.yylineno = this.yyleng = 0;
|
|
this.yytext = this.matched = this.match = "";
|
|
this.conditionStack = ["INITIAL"];
|
|
this.yylloc = {
|
|
first_line: 1,
|
|
first_column: 0,
|
|
last_line: 1,
|
|
last_column: 0
|
|
};
|
|
if (this.options.ranges) {
|
|
this.yylloc.range = [0, 0];
|
|
}
|
|
this.offset = 0;
|
|
return this;
|
|
},
|
|
// consumes and returns one char from the input
|
|
input: function() {
|
|
var ch = this._input[0];
|
|
this.yytext += ch;
|
|
this.yyleng++;
|
|
this.offset++;
|
|
this.match += ch;
|
|
this.matched += ch;
|
|
var lines = ch.match(/(?:\r\n?|\n).*/g);
|
|
if (lines) {
|
|
this.yylineno++;
|
|
this.yylloc.last_line++;
|
|
} else {
|
|
this.yylloc.last_column++;
|
|
}
|
|
if (this.options.ranges) {
|
|
this.yylloc.range[1]++;
|
|
}
|
|
this._input = this._input.slice(1);
|
|
return ch;
|
|
},
|
|
// unshifts one char (or a string) into the input
|
|
unput: function(ch) {
|
|
var len = ch.length;
|
|
var lines = ch.split(/(?:\r\n?|\n)/g);
|
|
this._input = ch + this._input;
|
|
this.yytext = this.yytext.substr(0, this.yytext.length - len);
|
|
this.offset -= len;
|
|
var oldLines = this.match.split(/(?:\r\n?|\n)/g);
|
|
this.match = this.match.substr(0, this.match.length - 1);
|
|
this.matched = this.matched.substr(0, this.matched.length - 1);
|
|
if (lines.length - 1) {
|
|
this.yylineno -= lines.length - 1;
|
|
}
|
|
var r = this.yylloc.range;
|
|
this.yylloc = {
|
|
first_line: this.yylloc.first_line,
|
|
last_line: this.yylineno + 1,
|
|
first_column: this.yylloc.first_column,
|
|
last_column: lines ? (lines.length === oldLines.length ? this.yylloc.first_column : 0) + oldLines[oldLines.length - lines.length].length - lines[0].length : this.yylloc.first_column - len
|
|
};
|
|
if (this.options.ranges) {
|
|
this.yylloc.range = [r[0], r[0] + this.yyleng - len];
|
|
}
|
|
this.yyleng = this.yytext.length;
|
|
return this;
|
|
},
|
|
// When called from action, caches matched text and appends it on next action
|
|
more: function() {
|
|
this._more = true;
|
|
return this;
|
|
},
|
|
// When called from action, signals the lexer that this rule fails to match the input, so the next matching rule (regex) should be tested instead.
|
|
reject: function() {
|
|
if (this.options.backtrack_lexer) {
|
|
this._backtrack = true;
|
|
} else {
|
|
return this.parseError("Lexical error on line " + (this.yylineno + 1) + ". You can only invoke reject() in the lexer when the lexer is of the backtracking persuasion (options.backtrack_lexer = true).\n" + this.showPosition(), {
|
|
text: "",
|
|
token: null,
|
|
line: this.yylineno
|
|
});
|
|
}
|
|
return this;
|
|
},
|
|
// retain first n characters of the match
|
|
less: function(n) {
|
|
this.unput(this.match.slice(n));
|
|
},
|
|
// displays already matched input, i.e. for error messages
|
|
pastInput: function() {
|
|
var past = this.matched.substr(0, this.matched.length - this.match.length);
|
|
return (past.length > 20 ? "..." : "") + past.substr(-20).replace(/\n/g, "");
|
|
},
|
|
// displays upcoming input, i.e. for error messages
|
|
upcomingInput: function() {
|
|
var next = this.match;
|
|
if (next.length < 20) {
|
|
next += this._input.substr(0, 20 - next.length);
|
|
}
|
|
return (next.substr(0, 20) + (next.length > 20 ? "..." : "")).replace(/\n/g, "");
|
|
},
|
|
// displays the character position where the lexing error occurred, i.e. for error messages
|
|
showPosition: function() {
|
|
var pre = this.pastInput();
|
|
var c = new Array(pre.length + 1).join("-");
|
|
return pre + this.upcomingInput() + "\n" + c + "^";
|
|
},
|
|
// test the lexed token: return FALSE when not a match, otherwise return token
|
|
test_match: function(match, indexed_rule) {
|
|
var token, lines, backup;
|
|
if (this.options.backtrack_lexer) {
|
|
backup = {
|
|
yylineno: this.yylineno,
|
|
yylloc: {
|
|
first_line: this.yylloc.first_line,
|
|
last_line: this.last_line,
|
|
first_column: this.yylloc.first_column,
|
|
last_column: this.yylloc.last_column
|
|
},
|
|
yytext: this.yytext,
|
|
match: this.match,
|
|
matches: this.matches,
|
|
matched: this.matched,
|
|
yyleng: this.yyleng,
|
|
offset: this.offset,
|
|
_more: this._more,
|
|
_input: this._input,
|
|
yy: this.yy,
|
|
conditionStack: this.conditionStack.slice(0),
|
|
done: this.done
|
|
};
|
|
if (this.options.ranges) {
|
|
backup.yylloc.range = this.yylloc.range.slice(0);
|
|
}
|
|
}
|
|
lines = match[0].match(/(?:\r\n?|\n).*/g);
|
|
if (lines) {
|
|
this.yylineno += lines.length;
|
|
}
|
|
this.yylloc = {
|
|
first_line: this.yylloc.last_line,
|
|
last_line: this.yylineno + 1,
|
|
first_column: this.yylloc.last_column,
|
|
last_column: lines ? lines[lines.length - 1].length - lines[lines.length - 1].match(/\r?\n?/)[0].length : this.yylloc.last_column + match[0].length
|
|
};
|
|
this.yytext += match[0];
|
|
this.match += match[0];
|
|
this.matches = match;
|
|
this.yyleng = this.yytext.length;
|
|
if (this.options.ranges) {
|
|
this.yylloc.range = [this.offset, this.offset += this.yyleng];
|
|
}
|
|
this._more = false;
|
|
this._backtrack = false;
|
|
this._input = this._input.slice(match[0].length);
|
|
this.matched += match[0];
|
|
token = this.performAction.call(this, this.yy, this, indexed_rule, this.conditionStack[this.conditionStack.length - 1]);
|
|
if (this.done && this._input) {
|
|
this.done = false;
|
|
}
|
|
if (token) {
|
|
return token;
|
|
} else if (this._backtrack) {
|
|
for (var k in backup) {
|
|
this[k] = backup[k];
|
|
}
|
|
return false;
|
|
}
|
|
return false;
|
|
},
|
|
// return next match in input
|
|
next: function() {
|
|
if (this.done) {
|
|
return this.EOF;
|
|
}
|
|
if (!this._input) {
|
|
this.done = true;
|
|
}
|
|
var token, match, tempMatch, index;
|
|
if (!this._more) {
|
|
this.yytext = "";
|
|
this.match = "";
|
|
}
|
|
var rules = this._currentRules();
|
|
for (var i = 0; i < rules.length; i++) {
|
|
tempMatch = this._input.match(this.rules[rules[i]]);
|
|
if (tempMatch && (!match || tempMatch[0].length > match[0].length)) {
|
|
match = tempMatch;
|
|
index = i;
|
|
if (this.options.backtrack_lexer) {
|
|
token = this.test_match(tempMatch, rules[i]);
|
|
if (token !== false) {
|
|
return token;
|
|
} else if (this._backtrack) {
|
|
match = false;
|
|
continue;
|
|
} else {
|
|
return false;
|
|
}
|
|
} else if (!this.options.flex) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (match) {
|
|
token = this.test_match(match, rules[index]);
|
|
if (token !== false) {
|
|
return token;
|
|
}
|
|
return false;
|
|
}
|
|
if (this._input === "") {
|
|
return this.EOF;
|
|
} else {
|
|
return this.parseError("Lexical error on line " + (this.yylineno + 1) + ". Unrecognized text.\n" + this.showPosition(), {
|
|
text: "",
|
|
token: null,
|
|
line: this.yylineno
|
|
});
|
|
}
|
|
},
|
|
// return next match that has a token
|
|
lex: function lex() {
|
|
var r = this.next();
|
|
if (r) {
|
|
return r;
|
|
} else {
|
|
return this.lex();
|
|
}
|
|
},
|
|
// activates a new lexer condition state (pushes the new lexer condition state onto the condition stack)
|
|
begin: function begin(condition) {
|
|
this.conditionStack.push(condition);
|
|
},
|
|
// pop the previously active lexer condition state off the condition stack
|
|
popState: function popState() {
|
|
var n = this.conditionStack.length - 1;
|
|
if (n > 0) {
|
|
return this.conditionStack.pop();
|
|
} else {
|
|
return this.conditionStack[0];
|
|
}
|
|
},
|
|
// produce the lexer rule set which is active for the currently active lexer condition state
|
|
_currentRules: function _currentRules() {
|
|
if (this.conditionStack.length && this.conditionStack[this.conditionStack.length - 1]) {
|
|
return this.conditions[this.conditionStack[this.conditionStack.length - 1]].rules;
|
|
} else {
|
|
return this.conditions["INITIAL"].rules;
|
|
}
|
|
},
|
|
// return the currently active lexer condition state; when an index argument is provided it produces the N-th previous condition state, if available
|
|
topState: function topState(n) {
|
|
n = this.conditionStack.length - 1 - Math.abs(n || 0);
|
|
if (n >= 0) {
|
|
return this.conditionStack[n];
|
|
} else {
|
|
return "INITIAL";
|
|
}
|
|
},
|
|
// alias for begin(condition)
|
|
pushState: function pushState(condition) {
|
|
this.begin(condition);
|
|
},
|
|
// return the number of states currently on the stack
|
|
stateStackSize: function stateStackSize() {
|
|
return this.conditionStack.length;
|
|
},
|
|
options: { "easy_keword_rules": true },
|
|
performAction: function anonymous(yy, yy_, $avoiding_name_collisions, YY_START) {
|
|
switch ($avoiding_name_collisions) {
|
|
case 0:
|
|
this.pushState("csv");
|
|
return 4;
|
|
case 1:
|
|
return 10;
|
|
case 2:
|
|
return 5;
|
|
case 3:
|
|
return 12;
|
|
case 4:
|
|
this.pushState("escaped_text");
|
|
return 18;
|
|
case 5:
|
|
return 20;
|
|
case 6:
|
|
this.popState("escaped_text");
|
|
return 18;
|
|
case 7:
|
|
return 19;
|
|
}
|
|
},
|
|
rules: [/^(?:sankey-beta\b)/, /^(?:$)/, /^(?:((\u000D\u000A)|(\u000A)))/, /^(?:(\u002C))/, /^(?:(\u0022))/, /^(?:([\u0020-\u0021\u0023-\u002B\u002D-\u007E])*)/, /^(?:(\u0022)(?!(\u0022)))/, /^(?:(([\u0020-\u0021\u0023-\u002B\u002D-\u007E])|(\u002C)|(\u000D)|(\u000A)|(\u0022)(\u0022))*)/],
|
|
conditions: { "csv": { "rules": [1, 2, 3, 4, 5, 6, 7], "inclusive": false }, "escaped_text": { "rules": [6, 7], "inclusive": false }, "INITIAL": { "rules": [0, 1, 2, 3, 4, 5, 6, 7], "inclusive": true } }
|
|
};
|
|
return lexer2;
|
|
}();
|
|
parser2.lexer = lexer;
|
|
function Parser() {
|
|
this.yy = {};
|
|
}
|
|
Parser.prototype = parser2;
|
|
parser2.Parser = Parser;
|
|
return new Parser();
|
|
}();
|
|
parser.parser = parser;
|
|
const parser$1 = parser;
|
|
let links = [];
|
|
let nodes = [];
|
|
let nodesMap = {};
|
|
const clear = () => {
|
|
links = [];
|
|
nodes = [];
|
|
nodesMap = {};
|
|
clear$1();
|
|
};
|
|
class SankeyLink {
|
|
constructor(source, target, value = 0) {
|
|
this.source = source;
|
|
this.target = target;
|
|
this.value = value;
|
|
}
|
|
}
|
|
const addLink = (source, target, value) => {
|
|
links.push(new SankeyLink(source, target, value));
|
|
};
|
|
class SankeyNode {
|
|
constructor(ID) {
|
|
this.ID = ID;
|
|
}
|
|
}
|
|
const findOrCreateNode = (ID) => {
|
|
ID = common.sanitizeText(ID, getConfig());
|
|
if (!nodesMap[ID]) {
|
|
nodesMap[ID] = new SankeyNode(ID);
|
|
nodes.push(nodesMap[ID]);
|
|
}
|
|
return nodesMap[ID];
|
|
};
|
|
const getNodes = () => nodes;
|
|
const getLinks = () => links;
|
|
const getGraph = () => ({
|
|
nodes: nodes.map((node) => ({ id: node.ID })),
|
|
links: links.map((link) => ({
|
|
source: link.source.ID,
|
|
target: link.target.ID,
|
|
value: link.value
|
|
}))
|
|
});
|
|
const db = {
|
|
nodesMap,
|
|
getConfig: () => getConfig().sankey,
|
|
getNodes,
|
|
getLinks,
|
|
getGraph,
|
|
addLink,
|
|
findOrCreateNode,
|
|
getAccTitle,
|
|
setAccTitle,
|
|
getAccDescription,
|
|
setAccDescription,
|
|
getDiagramTitle,
|
|
setDiagramTitle,
|
|
clear
|
|
};
|
|
const _Uid = class {
|
|
static next(name) {
|
|
return new _Uid(name + ++_Uid.count);
|
|
}
|
|
constructor(id) {
|
|
this.id = id;
|
|
this.href = `#${id}`;
|
|
}
|
|
toString() {
|
|
return "url(" + this.href + ")";
|
|
}
|
|
};
|
|
let Uid = _Uid;
|
|
Uid.count = 0;
|
|
const alignmentsMap = {
|
|
left: sankeyLeft,
|
|
right: sankeyRight,
|
|
center: sankeyCenter,
|
|
justify: sankeyJustify
|
|
};
|
|
const draw = function(text, id, _version, diagObj) {
|
|
const { securityLevel, sankey: conf } = getConfig();
|
|
const defaultSankeyConfig = defaultConfig.sankey;
|
|
let sandboxElement;
|
|
if (securityLevel === "sandbox") {
|
|
sandboxElement = select("#i" + id);
|
|
}
|
|
const root = securityLevel === "sandbox" ? select(sandboxElement.nodes()[0].contentDocument.body) : select("body");
|
|
const svg = securityLevel === "sandbox" ? root.select(`[id="${id}"]`) : select(`[id="${id}"]`);
|
|
const width = (conf == null ? void 0 : conf.width) ?? defaultSankeyConfig.width;
|
|
const height = (conf == null ? void 0 : conf.height) ?? defaultSankeyConfig.width;
|
|
const useMaxWidth = (conf == null ? void 0 : conf.useMaxWidth) ?? defaultSankeyConfig.useMaxWidth;
|
|
const nodeAlignment = (conf == null ? void 0 : conf.nodeAlignment) ?? defaultSankeyConfig.nodeAlignment;
|
|
const prefix = (conf == null ? void 0 : conf.prefix) ?? defaultSankeyConfig.prefix;
|
|
const suffix = (conf == null ? void 0 : conf.suffix) ?? defaultSankeyConfig.suffix;
|
|
const showValues = (conf == null ? void 0 : conf.showValues) ?? defaultSankeyConfig.showValues;
|
|
configureSvgSize(svg, height, width, useMaxWidth);
|
|
const graph = diagObj.db.getGraph();
|
|
const nodeAlign = alignmentsMap[nodeAlignment];
|
|
const nodeWidth = 10;
|
|
const sankey$1 = sankey().nodeId((d) => d.id).nodeWidth(nodeWidth).nodePadding(10 + (showValues ? 15 : 0)).nodeAlign(nodeAlign).extent([
|
|
[0, 0],
|
|
[width, height]
|
|
]);
|
|
sankey$1(graph);
|
|
const colorScheme = scaleOrdinal(schemeTableau10);
|
|
svg.append("g").attr("class", "nodes").selectAll(".node").data(graph.nodes).join("g").attr("class", "node").attr("id", (d) => (d.uid = Uid.next("node-")).id).attr("transform", function(d) {
|
|
return "translate(" + d.x0 + "," + d.y0 + ")";
|
|
}).attr("x", (d) => d.x0).attr("y", (d) => d.y0).append("rect").attr("height", (d) => {
|
|
return d.y1 - d.y0;
|
|
}).attr("width", (d) => d.x1 - d.x0).attr("fill", (d) => colorScheme(d.id));
|
|
const getText = ({ id: id2, value }) => {
|
|
if (!showValues) {
|
|
return id2;
|
|
}
|
|
return `${id2}
|
|
${prefix}${Math.round(value * 100) / 100}${suffix}`;
|
|
};
|
|
svg.append("g").attr("class", "node-labels").attr("font-family", "sans-serif").attr("font-size", 14).selectAll("text").data(graph.nodes).join("text").attr("x", (d) => d.x0 < width / 2 ? d.x1 + 6 : d.x0 - 6).attr("y", (d) => (d.y1 + d.y0) / 2).attr("dy", `${showValues ? "0" : "0.35"}em`).attr("text-anchor", (d) => d.x0 < width / 2 ? "start" : "end").text(getText);
|
|
const link = svg.append("g").attr("class", "links").attr("fill", "none").attr("stroke-opacity", 0.5).selectAll(".link").data(graph.links).join("g").attr("class", "link").style("mix-blend-mode", "multiply");
|
|
const linkColor = (conf == null ? void 0 : conf.linkColor) || "gradient";
|
|
if (linkColor === "gradient") {
|
|
const gradient = link.append("linearGradient").attr("id", (d) => (d.uid = Uid.next("linearGradient-")).id).attr("gradientUnits", "userSpaceOnUse").attr("x1", (d) => d.source.x1).attr("x2", (d) => d.target.x0);
|
|
gradient.append("stop").attr("offset", "0%").attr("stop-color", (d) => colorScheme(d.source.id));
|
|
gradient.append("stop").attr("offset", "100%").attr("stop-color", (d) => colorScheme(d.target.id));
|
|
}
|
|
let coloring;
|
|
switch (linkColor) {
|
|
case "gradient":
|
|
coloring = (d) => d.uid;
|
|
break;
|
|
case "source":
|
|
coloring = (d) => colorScheme(d.source.id);
|
|
break;
|
|
case "target":
|
|
coloring = (d) => colorScheme(d.target.id);
|
|
break;
|
|
default:
|
|
coloring = linkColor;
|
|
}
|
|
link.append("path").attr("d", sankeyLinkHorizontal()).attr("stroke", coloring).attr("stroke-width", (d) => Math.max(1, d.width));
|
|
};
|
|
const renderer = {
|
|
draw
|
|
};
|
|
const prepareTextForParsing = (text) => {
|
|
const textToParse = text.replaceAll(/^[^\S\n\r]+|[^\S\n\r]+$/g, "").replaceAll(/([\n\r])+/g, "\n").trim();
|
|
return textToParse;
|
|
};
|
|
const originalParse = parser$1.parse.bind(parser$1);
|
|
parser$1.parse = (text) => originalParse(prepareTextForParsing(text));
|
|
const diagram = {
|
|
parser: parser$1,
|
|
db,
|
|
renderer
|
|
};
|
|
export {
|
|
diagram
|
|
};
|