mirror of
https://github.com/mermaid-js/mermaid.git
synced 2025-08-22 17:56:43 +02:00
Style in place for edges
dependency on requirejs removed dist in place
This commit is contained in:
@@ -6,9 +6,6 @@ module.exports = function (config) {
|
||||
|
||||
// base path that will be used to resolve all patterns (eg. files, exclude)
|
||||
basePath: '',
|
||||
browserify: {
|
||||
watch: true
|
||||
},
|
||||
|
||||
// frameworks to use
|
||||
// available frameworks: https://npmjs.org/browse/keyword/karma-adapter
|
||||
@@ -17,15 +14,12 @@ module.exports = function (config) {
|
||||
|
||||
// list of files / patterns to load in the browser
|
||||
files: [
|
||||
'vendor/requirejs/require.js',
|
||||
'vendor/jasmine/lib/jasmine-core/jasmine.js',
|
||||
'src/**/*.spec.js',
|
||||
'test/test-main.js'
|
||||
'../src/**/*.js'
|
||||
],
|
||||
|
||||
|
||||
// list of files to exclude
|
||||
exclude: [],
|
||||
exclude: ['../src/backup/**/*.js'],
|
||||
|
||||
|
||||
// preprocess matching files before serving them to the browser
|
||||
|
31
gulpfile.js
31
gulpfile.js
@@ -1,6 +1,9 @@
|
||||
var gulp = require('gulp');
|
||||
var jison = require('gulp-jison');
|
||||
var shell = require('gulp-shell')
|
||||
var shell = require('gulp-shell');
|
||||
var concat = require('gulp-concat');
|
||||
var uglify = require('gulp-uglify');
|
||||
var extReplace = require('gulp-ext-replace');
|
||||
|
||||
gulp.task('jison2', function() {
|
||||
return gulp.src('./src/*.jison')
|
||||
@@ -9,11 +12,33 @@ gulp.task('jison2', function() {
|
||||
});
|
||||
|
||||
gulp.task('jison', shell.task([
|
||||
'jison src/parser/mermaid.jison -o src/parser/mermaid.js',
|
||||
'jison src/parser/flow.jison -o src/parser/flow.js',
|
||||
'source scripts/compileJison.sh'
|
||||
]))
|
||||
|
||||
gulp.task('jison2', shell.task([
|
||||
'jison src/parser/flow.jison -o src/parser/flow.js',
|
||||
'source scripts/compileFlow.sh'
|
||||
]))
|
||||
]))
|
||||
|
||||
gulp.task('distSlim', function() {
|
||||
gulp.src(['./src/parser/flow.js','./src/graph.js','./src/main.js'])
|
||||
.pipe(concat('mermaid.slim.js'))
|
||||
.pipe(gulp.dest('./dist/'))
|
||||
.pipe(uglify())
|
||||
.pipe(extReplace('.min.js'))
|
||||
.pipe(gulp.dest('./dist/'));
|
||||
|
||||
});
|
||||
|
||||
gulp.task('distFull', function() {
|
||||
gulp.src(['./lib/d3.v3.min.js', './lib/dagre-d3.min.js', './src/parser/flow.js','./src/graph.js','./src/main.js'])
|
||||
.pipe(concat('mermaid.full.js'))
|
||||
.pipe(gulp.dest('./dist/'))
|
||||
});
|
||||
|
||||
gulp.task('dist', ['distSlim', 'distFull'], function() {
|
||||
gulp.src(['./lib/d3.v3.min.js', './lib/dagre-d3.min.js', './dist/mermaid.slim.min.js'])
|
||||
.pipe(concat('mermaid.full.min.js'))
|
||||
.pipe(gulp.dest('./dist/'))
|
||||
});
|
@@ -23,7 +23,7 @@ module.exports = function(config) {
|
||||
|
||||
// list of files to exclude
|
||||
exclude: [
|
||||
//'src/parser/mermaid.js'
|
||||
//'src/parser/flow.js'
|
||||
],
|
||||
|
||||
|
||||
|
@@ -16,6 +16,9 @@
|
||||
"karma": "~0.12.20",
|
||||
"karma-jasmine": "~0.2.1",
|
||||
"karma-chrome-launcher": "~0.1.5",
|
||||
"karma-requirejs": "~0.2.2"
|
||||
"karma-requirejs": "~0.2.2",
|
||||
"gulp-concat": "~2.4.1",
|
||||
"gulp-uglify": "~1.0.1",
|
||||
"gulp-ext-replace": "~0.1.0"
|
||||
}
|
||||
}
|
||||
|
@@ -37,4 +37,13 @@ g.nodes().forEach(function(v) {
|
||||
});
|
||||
g.edges().forEach(function(e) {
|
||||
console.log("Edge " + e.v + " -> " + e.w + ": " + JSON.stringify(g.edge(e)));
|
||||
});
|
||||
});
|
||||
|
||||
var benv = require('benv');
|
||||
benv.setup();
|
||||
benv.require('./lib/d3.v3.min.js');
|
||||
benv.require('./lib/dagre-d3.min.js');
|
||||
benv.require('./lib/flow.js');
|
||||
|
||||
document.body.innerHTML='<div class="mermaid">a-->b;</div>;';
|
||||
benv.require('./src/main.js');
|
69
src/backup/flow.jison
Normal file
69
src/backup/flow.jison
Normal file
@@ -0,0 +1,69 @@
|
||||
/* description: Parses end executes mathematical expressions. */
|
||||
|
||||
/* lexical grammar */
|
||||
%lex
|
||||
|
||||
%%
|
||||
"style" return 'STYLE';
|
||||
[a-zåäöæøA-ZÅÄÖÆØ]+ return 'ALPHA';
|
||||
\#[a-f0-9]+ return 'HEX';
|
||||
[0-9]+ return 'NUM';
|
||||
"px" return 'UNIT';
|
||||
"pt" return 'UNIT';
|
||||
"dot" return 'UNIT';
|
||||
":" return 'COLON';
|
||||
\- return 'MINUS';
|
||||
";" return ';';
|
||||
"," return 'COMMA';
|
||||
[x] return 'ARROW_CROSS';
|
||||
">" return 'ARROW_POINT';
|
||||
[o] return 'ARROW_CIRCLE';
|
||||
"|" return 'PIPE';
|
||||
"(" return 'PS';
|
||||
")" return 'PE';
|
||||
"[" return 'SQS';
|
||||
"]" return 'SQE';
|
||||
"{" return 'DIAMOND_START'
|
||||
"}" return 'DIAMOND_STOP'
|
||||
\s return 'SPACE';
|
||||
\n return 'NEWLINE';
|
||||
<<EOF>> return 'EOF';
|
||||
|
||||
/lex
|
||||
|
||||
/* operator associations and precedence */
|
||||
|
||||
%left '^'
|
||||
|
||||
%start expressions
|
||||
|
||||
%% /* language grammar */
|
||||
|
||||
expressions
|
||||
: id EOF
|
||||
{return $1;}
|
||||
;
|
||||
|
||||
flow: id
|
||||
{$$='key';}
|
||||
| STYLE
|
||||
{$$=$1;}
|
||||
;
|
||||
|
||||
id: id MINUS word
|
||||
{$$=$1+'-'+$3}
|
||||
| word
|
||||
{$$=$1}
|
||||
;
|
||||
|
||||
word: ALPHA
|
||||
{$$=$1}
|
||||
;
|
||||
|
||||
|
||||
|
||||
%%
|
||||
define('parser/flow',function(){
|
||||
console.log('bcs123');
|
||||
return parser;
|
||||
});
|
@@ -72,12 +72,12 @@
|
||||
}
|
||||
*/
|
||||
var parser = (function(){
|
||||
var o=function(k,v,o,l){for(o=o||{},l=k.length;l--;o[k[l]]=v);return o},$V0=[1,3],$V1=[1,7],$V2=[1,8],$V3=[8,12,23],$V4=[1,23],$V5=[8,18,20,22],$V6=[1,46],$V7=[1,41],$V8=[1,43],$V9=[1,42],$Va=[1,44],$Vb=[1,45],$Vc=[1,47],$Vd=[8,28],$Ve=[6,8,15,23,28,30,31,32,33];
|
||||
var o=function(k,v,o,l){for(o=o||{},l=k.length;l--;o[k[l]]=v);return o},$V0=[1,4],$V1=[5,8];
|
||||
var parser = {trace: function trace() { },
|
||||
yy: {},
|
||||
symbols_: {"error":2,"expressions":3,"graph":4,"EOF":5,"SPACE":6,"edge":7,";":8,"styleStatement":9,"vertex":10,"link":11,"PIPE":12,"text":13,"STYLE":14,"ALPHA":15,"stylesOpt":16,"SQS":17,"SQE":18,"PS":19,"PE":20,"DIAMOND_START":21,"DIAMOND_STOP":22,"MINUS":23,"ARROW_POINT":24,"ARROW_CIRCLE":25,"ARROW_CROSS":26,"style":27,"COMMA":28,"styleComponent":29,"COLON":30,"NUM":31,"UNIT":32,"HEX":33,"colordef":34,"COLOR":35,"borderWidth":36,"borderStyle":37,"BORDER_STYLE":38,"$accept":0,"$end":1},
|
||||
terminals_: {2:"error",5:"EOF",6:"SPACE",8:";",12:"PIPE",14:"STYLE",15:"ALPHA",17:"SQS",18:"SQE",19:"PS",20:"PE",21:"DIAMOND_START",22:"DIAMOND_STOP",23:"MINUS",24:"ARROW_POINT",25:"ARROW_CIRCLE",26:"ARROW_CROSS",28:"COMMA",30:"COLON",31:"NUM",32:"UNIT",33:"HEX",35:"COLOR",38:"BORDER_STYLE"},
|
||||
productions_: [0,[3,2],[4,2],[4,3],[4,2],[7,1],[7,5],[7,3],[7,1],[9,5],[10,4],[10,4],[10,4],[10,1],[13,3],[13,3],[13,2],[13,1],[11,3],[11,3],[11,3],[11,3],[16,1],[16,3],[27,1],[27,2],[29,1],[29,1],[29,1],[29,1],[29,1],[29,1],[29,1],[34,1],[34,1],[36,2],[37,1]],
|
||||
symbols_: {"error":2,"expressions":3,"id":4,"EOF":5,"flow":6,"STYLE":7,"MINUS":8,"word":9,"ALPHA":10,"$accept":0,"$end":1},
|
||||
terminals_: {2:"error",5:"EOF",7:"STYLE",8:"MINUS",10:"ALPHA"},
|
||||
productions_: [0,[3,2],[6,1],[6,1],[4,3],[4,1],[9,1]],
|
||||
performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate /* action[1] */, $$ /* vstack */, _$ /* lstack */) {
|
||||
/* this == yyval */
|
||||
|
||||
@@ -86,88 +86,22 @@ switch (yystate) {
|
||||
case 1:
|
||||
return $$[$0-1];
|
||||
break;
|
||||
case 2: case 3:
|
||||
this.$ = $$[$0];
|
||||
case 2:
|
||||
this.$='key';
|
||||
break;
|
||||
case 4:
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 5:
|
||||
this.$ = 'ya';
|
||||
break;
|
||||
case 6:
|
||||
yy.addLink($$[$0-4],$$[$0-2],$$[$0-3],$$[$0]);this.$ = 'oy'
|
||||
break;
|
||||
case 7:
|
||||
yy.addLink($$[$0-2],$$[$0],$$[$0-1]);this.$ = 'oy'
|
||||
break;
|
||||
case 8:
|
||||
this.$ = 'yo';
|
||||
break;
|
||||
case 9:
|
||||
console.log('a4');this.$ = $$[$0-4];yy.addVertex($$[$0-2],undefined,undefined,$$[$0]);
|
||||
break;
|
||||
case 10:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'square');
|
||||
break;
|
||||
case 11:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'round');
|
||||
break;
|
||||
case 12:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'diamond');
|
||||
break;
|
||||
case 13:
|
||||
this.$ = $$[$0];yy.addVertex($$[$0]);
|
||||
break;
|
||||
case 14:
|
||||
this.$ = $$[$0-2] + ' ' +$$[$0];
|
||||
break;
|
||||
case 15:
|
||||
this.$ = $$[$0-2] + '-' +$$[$0];
|
||||
break;
|
||||
case 16:
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 17: case 36:
|
||||
this.$ = $$[$0];
|
||||
break;
|
||||
case 18:
|
||||
this.$ = {"type":"arrow"};
|
||||
break;
|
||||
case 19:
|
||||
this.$ = {"type":"arrow_circle"};
|
||||
break;
|
||||
case 20:
|
||||
this.$ = {"type":"arrow_cross"};
|
||||
break;
|
||||
case 21:
|
||||
this.$ = {"type":"arrow_open"};
|
||||
break;
|
||||
case 22:
|
||||
console.log('a:'+$$[$0]);this.$ = [$$[$0]]
|
||||
break;
|
||||
case 23:
|
||||
console.log('a3:'+$$[$0-2].length+','+$$[$0]);$$[$0-2].push($$[$0]);this.$ = $$[$0-2];
|
||||
break;
|
||||
case 24:
|
||||
case 3:
|
||||
this.$=$$[$0];
|
||||
break;
|
||||
case 25:
|
||||
console.log('b1:');this.$ = $$[$0-1] + $$[$0];
|
||||
case 4:
|
||||
this.$=$$[$0-2]+'-'+$$[$0]
|
||||
break;
|
||||
case 26: case 27: case 28: case 29: case 30: case 31: case 32:
|
||||
case 5: case 6:
|
||||
this.$=$$[$0]
|
||||
break;
|
||||
case 33: case 34:
|
||||
this.$ = yytext;
|
||||
break;
|
||||
case 35:
|
||||
this.$ = $$[$0-1]+''+$$[$0];
|
||||
break;
|
||||
}
|
||||
},
|
||||
table: [{3:1,4:2,6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{1:[3]},{5:[1,9]},{4:10,6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{8:[1,11]},{8:[2,5]},{8:[2,8],11:12,23:[1,13]},{6:[1,14]},o($V3,[2,13],{17:[1,15],19:[1,16],21:[1,17]}),{1:[2,1]},{5:[2,2]},{4:18,5:[2,4],6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{10:19,15:$V2},{23:[1,20]},{15:[1,21]},{13:22,15:$V4},{13:24,15:$V4},{13:25,15:$V4},{5:[2,3]},{8:[2,7],12:[1,26]},{23:[1,30],24:[1,27],25:[1,28],26:[1,29]},{6:[1,31]},{18:[1,32]},o($V5,[2,17],{6:[1,33],23:[1,34]}),{20:[1,35]},{22:[1,36]},{13:37,15:$V4},{15:[2,18]},{15:[2,19]},{15:[2,20]},{15:[2,21]},{6:$V6,15:$V7,16:38,23:$V8,27:39,29:40,30:$V9,31:$Va,32:$Vb,33:$Vc},o($V3,[2,10]),o($V5,[2,16],{13:48,15:$V4}),{13:49,15:$V4},o($V3,[2,11]),o($V3,[2,12]),{8:[2,6]},{8:[2,9],28:[1,50]},o($Vd,[2,22],{29:51,6:$V6,15:$V7,23:$V8,30:$V9,31:$Va,32:$Vb,33:$Vc}),o($Ve,[2,24]),o($Ve,[2,26]),o($Ve,[2,27]),o($Ve,[2,28]),o($Ve,[2,29]),o($Ve,[2,30]),o($Ve,[2,31]),o($Ve,[2,32]),o($V5,[2,14]),o($V5,[2,15]),{6:$V6,15:$V7,23:$V8,27:52,29:40,30:$V9,31:$Va,32:$Vb,33:$Vc},o($Ve,[2,25]),o($Vd,[2,23],{29:51,6:$V6,15:$V7,23:$V8,30:$V9,31:$Va,32:$Vb,33:$Vc})],
|
||||
defaultActions: {5:[2,5],9:[2,1],10:[2,2],18:[2,3],27:[2,18],28:[2,19],29:[2,20],30:[2,21],37:[2,6]},
|
||||
table: [{3:1,4:2,9:3,10:$V0},{1:[3]},{5:[1,5],8:[1,6]},o($V1,[2,5]),o($V1,[2,6]),{1:[2,1]},{9:7,10:$V0},o($V1,[2,4])],
|
||||
defaultActions: {5:[2,1]},
|
||||
parseError: function parseError(str, hash) {
|
||||
if (hash.recoverable) {
|
||||
this.trace(str);
|
||||
@@ -313,7 +247,8 @@ parse: function parse(input) {
|
||||
return true;
|
||||
}};
|
||||
|
||||
define('parser/mermaid',function(){
|
||||
define('parser/flow',function(){
|
||||
console.log('bcs123');
|
||||
return parser;
|
||||
});/* generated by jison-lex 0.3.4 */
|
||||
var lexer = (function(){
|
||||
@@ -643,49 +578,49 @@ options: {},
|
||||
performAction: function anonymous(yy,yy_,$avoiding_name_collisions,YY_START) {
|
||||
var YYSTATE=YY_START;
|
||||
switch($avoiding_name_collisions) {
|
||||
case 0:return 14;
|
||||
case 0:return 7;
|
||||
break;
|
||||
case 1:return 33;
|
||||
case 1:return 10;
|
||||
break;
|
||||
case 2:return 31;
|
||||
case 2:return 'HEX';
|
||||
break;
|
||||
case 3:return 32;
|
||||
case 3:return 'NUM';
|
||||
break;
|
||||
case 4:return 32;
|
||||
case 4:return 'UNIT';
|
||||
break;
|
||||
case 5:return 32;
|
||||
case 5:return 'UNIT';
|
||||
break;
|
||||
case 6:return 30;
|
||||
case 6:return 'UNIT';
|
||||
break;
|
||||
case 7:return 23;
|
||||
case 7:return 'COLON';
|
||||
break;
|
||||
case 8:return 8;
|
||||
break;
|
||||
case 9:return 28;
|
||||
case 9:return ';';
|
||||
break;
|
||||
case 10:return 26;
|
||||
case 10:return 'COMMA';
|
||||
break;
|
||||
case 11:return 24;
|
||||
case 11:return 'ARROW_CROSS';
|
||||
break;
|
||||
case 12:return 25;
|
||||
case 12:return 'ARROW_POINT';
|
||||
break;
|
||||
case 13:return 15;
|
||||
case 13:return 'ARROW_CIRCLE';
|
||||
break;
|
||||
case 14:return 12;
|
||||
case 14:return 'PIPE';
|
||||
break;
|
||||
case 15:return 19;
|
||||
case 15:return 'PS';
|
||||
break;
|
||||
case 16:return 20;
|
||||
case 16:return 'PE';
|
||||
break;
|
||||
case 17:return 17;
|
||||
case 17:return 'SQS';
|
||||
break;
|
||||
case 18:return 18;
|
||||
case 18:return 'SQE';
|
||||
break;
|
||||
case 19:return 21
|
||||
case 19:return 'DIAMOND_START'
|
||||
break;
|
||||
case 20:return 22
|
||||
case 20:return 'DIAMOND_STOP'
|
||||
break;
|
||||
case 21:return 6;
|
||||
case 21:return 'SPACE';
|
||||
break;
|
||||
case 22:return 'NEWLINE';
|
||||
break;
|
||||
@@ -693,7 +628,7 @@ case 23:return 5;
|
||||
break;
|
||||
}
|
||||
},
|
||||
rules: [/^(?:style\b)/,/^(?:#[a-f0-9]+)/,/^(?:[0-9]+)/,/^(?:px\b)/,/^(?:pt\b)/,/^(?:dot\b)/,/^(?::)/,/^(?:-)/,/^(?:;)/,/^(?:,)/,/^(?:[x])/,/^(?:>)/,/^(?:[o])/,/^(?:[a-zåäöæøA-ZÅÄÖÆØ]+)/,/^(?:\|)/,/^(?:\()/,/^(?:\))/,/^(?:\[)/,/^(?:\])/,/^(?:\{)/,/^(?:\})/,/^(?:\s)/,/^(?:\n)/,/^(?:$)/],
|
||||
rules: [/^(?:style\b)/,/^(?:[a-zåäöæøA-ZÅÄÖÆØ]+)/,/^(?:#[a-f0-9]+)/,/^(?:[0-9]+)/,/^(?:px\b)/,/^(?:pt\b)/,/^(?:dot\b)/,/^(?::)/,/^(?:-)/,/^(?:;)/,/^(?:,)/,/^(?:[x])/,/^(?:>)/,/^(?:[o])/,/^(?:\|)/,/^(?:\()/,/^(?:\))/,/^(?:\[)/,/^(?:\])/,/^(?:\{)/,/^(?:\})/,/^(?:\s)/,/^(?:\n)/,/^(?:$)/],
|
||||
conditions: {"INITIAL":{"rules":[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23],"inclusive":true}}
|
||||
});
|
||||
return lexer;
|
22
src/backup/flow.spec.js
Normal file
22
src/backup/flow.spec.js
Normal file
@@ -0,0 +1,22 @@
|
||||
/**
|
||||
* Created by knut on 14-11-03.
|
||||
*/
|
||||
define('parser/flow.spec',['parser/graph','parser/flow'],function(graph, p){
|
||||
|
||||
describe('when parsing ',function(){
|
||||
beforeEach(function(){
|
||||
graph.clear();
|
||||
p.yy = graph;
|
||||
/*p.parse.parseError= function parseError(str, hash) {
|
||||
console.log(str);
|
||||
}*/
|
||||
});
|
||||
|
||||
it('should handle a nodes and edges',function(){
|
||||
var res = p.parse('apa-apa-åäö');
|
||||
console.log('Done parsing:' + res);
|
||||
});
|
||||
});
|
||||
|
||||
});
|
||||
|
@@ -1,14 +1,14 @@
|
||||
/*
|
||||
|
||||
require.config({
|
||||
// Karma serves files from '/base'
|
||||
baseUrl: './'
|
||||
|
||||
});
|
||||
|
||||
require(['mermaid'],function(mermaid){
|
||||
require(['../mermaid'],function(mermaid){
|
||||
mermaid.init();
|
||||
});
|
||||
|
||||
*/
|
||||
|
||||
|
||||
console.log('Init running');
|
@@ -2,7 +2,7 @@
|
||||
* Created by knut on 14-11-06.
|
||||
*/
|
||||
// Now we're ready to require JointJS and write our application code.
|
||||
define('mermaid',['parser/graph','parser/mermaid'],function(graph,parser){
|
||||
define('mermaid',['../parser/graph','parser/mermaid'],function(graph,parser){
|
||||
var addVertices = function(vert,g){
|
||||
var keys = Object.keys(vert);
|
||||
|
||||
@@ -54,9 +54,10 @@ define('mermaid',['parser/graph','parser/mermaid'],function(graph,parser){
|
||||
g.setEdge(edge.start,edge.end);
|
||||
}
|
||||
else{
|
||||
g.setEdge(edge.start,edge.end,{ label: edge.text });
|
||||
g.setEdge(edge.start,edge.end,{ label: edge.text,style:"stroke: #f66; stroke-width: 1.5px;",
|
||||
arrowheadStyle: "fill: #f66" });
|
||||
}
|
||||
console.log('g.setEdge("'+edge.start+'","'+edge.end+'")');
|
||||
console.log('g.setEdge("'+edge.start+'","'+edge.end+'") ---');
|
||||
|
||||
});
|
||||
};
|
@@ -1,98 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html>
|
||||
<head>
|
||||
</head>
|
||||
<body>
|
||||
<div id="app"></div>
|
||||
|
||||
<svg width="100%" height="100">
|
||||
// make sure to give a distinct ID for the second graph placeholder
|
||||
<g id=DIAGRAM1 transform="translate(20,20)"/>
|
||||
</svg>
|
||||
|
||||
<scrpt src="../vendor/d3/d3.js"></scrpt>
|
||||
<script src="http://d3js.org/d3.v3.min.js" charset="utf-8"></script>
|
||||
<script src="../lib/dagre-d3.min.js"></script>
|
||||
<style id="css">
|
||||
/* This sets the color for "TK" nodes to a light blue green. */
|
||||
g.type-TK > rect {
|
||||
fill: #00ffd0;
|
||||
}
|
||||
|
||||
text {
|
||||
font-weight: 300;
|
||||
font-family: "Helvetica Neue", Helvetica, Arial, sans-serf;
|
||||
font-size: 14px;
|
||||
}
|
||||
|
||||
.node rect {
|
||||
stroke: #999;
|
||||
fill: #fff;
|
||||
stroke-width: 1.5px;
|
||||
}
|
||||
|
||||
.edgePath path {
|
||||
stroke: #333;
|
||||
stroke-width: 1.5px;
|
||||
}
|
||||
</style>
|
||||
<script>
|
||||
|
||||
// Create the renderer
|
||||
var render = new dagreD3.render();
|
||||
|
||||
|
||||
|
||||
// Create a new directed graph
|
||||
//var g = new dagre.graphlib.Graph();
|
||||
|
||||
// Create the input graph
|
||||
var g = new dagreD3.graphlib.Graph()
|
||||
.setGraph({})
|
||||
.setDefaultEdgeLabel(function() { return {}; });
|
||||
|
||||
// Add nodes to the graph. The first argument is the node id. The second is
|
||||
// metadata about the node. In this case we're going to add labels to each of
|
||||
// our nodes.
|
||||
g.setNode("kspacey", { label: "Kevin Spacey" });
|
||||
g.setNode("swilliams", { label: "Saul Williams som är en \nfinurlig \n åöä filivipp som gillar \natt hoppa och dansa"});
|
||||
g.setNode("bpitt", { label: "Brad Pitt", width: 78, height: 100 });
|
||||
g.setNode("hford", { label: "Harrison Ford", width: 138, height: 100 });
|
||||
g.setNode("lwilson", { label: "Luke Wilson", width: 114, height: 100 });
|
||||
g.setNode("kbacon", { label: "Kevin Bacon", width: 91, height: 100 });
|
||||
|
||||
// Add edges to the graph.
|
||||
g.setEdge("kspacey", "swilliams");
|
||||
g.setEdge("swilliams", "kbacon");
|
||||
g.setEdge("bpitt", "kbacon");
|
||||
g.setEdge("hford", "lwilson");
|
||||
g.setEdge("lwilson", "kbacon");
|
||||
|
||||
g.nodes().forEach(function(v) {
|
||||
var node = g.node(v);
|
||||
// Round the corners of the nodes
|
||||
node.rx = node.ry = 5;
|
||||
});
|
||||
|
||||
//dagre.layout(g);
|
||||
|
||||
// Create the renderer
|
||||
var render = new dagreD3.render();
|
||||
|
||||
// Set up an SVG group so that we can translate the final graph.
|
||||
var svg = d3.select("svg"),
|
||||
svgGroup = svg.append("g");
|
||||
|
||||
// Run the renderer. This is what draws the final graph.
|
||||
render(d3.select("svg g#DIAGRAM1"), g);
|
||||
|
||||
// Center the graph
|
||||
var xCenterOffset = (svg.attr("width") - g.graph().width) / 2;
|
||||
svgGroup.attr("transform", "translate(" + xCenterOffset + ", 20)");
|
||||
svg.attr("height", g.graph().height + 40);
|
||||
</script>
|
||||
|
||||
|
||||
|
||||
</body>
|
||||
</html>
|
@@ -1,78 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html>
|
||||
<head>
|
||||
</head>
|
||||
<body>
|
||||
<div id="app"></div>
|
||||
|
||||
<svg width="100%" height="100">
|
||||
// make sure to give a distinct ID for the second graph placeholder
|
||||
<g id=DIAGRAM1 transform="translate(20,20)"/>
|
||||
</svg>
|
||||
|
||||
<script src="http://d3js.org/d3.v3.min.js" charset="utf-8"></script>
|
||||
<script src="../lib/dagre-d3.min.js"></script>
|
||||
<style id="css">
|
||||
/* This sets the color for "TK" nodes to a light blue green. */
|
||||
g.type-TK > rect {
|
||||
fill: #00ffd0;
|
||||
}
|
||||
|
||||
text {
|
||||
font-weight: 300;
|
||||
font-family: "Helvetica Neue", Helvetica, Arial, sans-serf;
|
||||
font-size: 14px;
|
||||
}
|
||||
|
||||
.node rect {
|
||||
stroke: #999;
|
||||
fill: #fff;
|
||||
stroke-width: 1.5px;
|
||||
}
|
||||
|
||||
.edgePath path {
|
||||
stroke: #333;
|
||||
stroke-width: 1.5px;
|
||||
}
|
||||
</style>
|
||||
<script>
|
||||
// Create the input graph
|
||||
var g = new dagreD3.graphlib.Graph()
|
||||
.setGraph({})
|
||||
.setDefaultEdgeLabel(function() { return {}; });
|
||||
|
||||
console.log(g);
|
||||
|
||||
// Add nodes to the graph. The first argument is the node id. The second is
|
||||
// metadata about the node. In this case we're going to add labels to each of
|
||||
// our nodes.
|
||||
g.setNode("A", { label: "chimpansen hoppar" });
|
||||
g.setNode("B", { label: "gorillan hoppar" });
|
||||
g.setEdge("A","B");
|
||||
|
||||
g.nodes().forEach(function(v) {
|
||||
var node = g.node(v);
|
||||
// Round the corners of the nodes
|
||||
node.rx = node.ry = 5;
|
||||
});
|
||||
|
||||
// Create the renderer
|
||||
var render = new dagreD3.render();
|
||||
|
||||
// Set up an SVG group so that we can translate the final graph.
|
||||
var svg = d3.select("svg"),
|
||||
svgGroup = svg.append("g");
|
||||
|
||||
// Run the renderer. This is what draws the final graph.
|
||||
render(d3.select("svg g#DIAGRAM1"), g);
|
||||
|
||||
// Center the graph
|
||||
var xCenterOffset = (svg.attr("width") - g.graph().width) / 2;
|
||||
svgGroup.attr("transform", "translate(" + xCenterOffset + ", 20)");
|
||||
svg.attr("height", g.graph().height + 40);
|
||||
</script>
|
||||
|
||||
|
||||
|
||||
</body>
|
||||
</html>
|
@@ -5,9 +5,9 @@
|
||||
|
||||
%%
|
||||
"style" return 'STYLE';
|
||||
[a-zåäöæøA-ZÅÄÖÆØ]+ return 'ALPHA';
|
||||
\#[a-f0-9]+ return 'HEX';
|
||||
[0-9]+ return 'NUM';
|
||||
\# return 'BRKT';
|
||||
"px" return 'UNIT';
|
||||
"pt" return 'UNIT';
|
||||
"dot" return 'UNIT';
|
||||
@@ -18,6 +18,7 @@
|
||||
[x] return 'ARROW_CROSS';
|
||||
">" return 'ARROW_POINT';
|
||||
[o] return 'ARROW_CIRCLE';
|
||||
[a-zåäöæøA-ZÅÄÖÆØ]+ return 'ALPHA';
|
||||
"|" return 'PIPE';
|
||||
"(" return 'PS';
|
||||
")" return 'PE';
|
||||
@@ -40,30 +41,108 @@
|
||||
%% /* language grammar */
|
||||
|
||||
expressions
|
||||
: id EOF
|
||||
: graph EOF
|
||||
{return $1;}
|
||||
;
|
||||
|
||||
flow: id
|
||||
{$$='key';}
|
||||
| STYLE
|
||||
{$$=$1;}
|
||||
graph
|
||||
: SPACE graph
|
||||
{ $$ = $2;}
|
||||
| edge ';' graph
|
||||
{ $$ = $3;}
|
||||
| edge ';'
|
||||
{ $$ = $1;}
|
||||
;
|
||||
|
||||
id: id MINUS word
|
||||
{$$=$1+'-'+$3}
|
||||
| word
|
||||
{$$=$1}
|
||||
;
|
||||
edge: styleStatement
|
||||
{$$ = 'ya';}
|
||||
| vertex link vertex PIPE text
|
||||
{ yy.addLink($1,$3,$2,$5);$$ = 'oy'}
|
||||
| vertex link vertex
|
||||
{ yy.addLink($1,$3,$2);$$ = 'oy'}
|
||||
| vertex
|
||||
{$$ = 'yo';}
|
||||
;
|
||||
|
||||
word: ALPHA
|
||||
{$$=$1}
|
||||
;
|
||||
styleStatement:STYLE SPACE ALPHA SPACE stylesOpt
|
||||
{$$ = $1;yy.addVertex($3,undefined,undefined,$5);}
|
||||
| STYLE SPACE HEX SPACE stylesOpt
|
||||
{console.log('In parser - style: '+$5);$$ = $1;yy.updateLink($3,$5);}
|
||||
;
|
||||
|
||||
vertex: ALPHA SQS text SQE
|
||||
{$$ = $1;yy.addVertex($1,$3,'square');}
|
||||
| ALPHA PS text PE
|
||||
{$$ = $1;yy.addVertex($1,$3,'round');}
|
||||
| ALPHA DIAMOND_START text DIAMOND_STOP
|
||||
{$$ = $1;yy.addVertex($1,$3,'diamond');}
|
||||
| ALPHA
|
||||
{$$ = $1;yy.addVertex($1);}
|
||||
;
|
||||
// Characters and spaces
|
||||
text: ALPHA SPACE text
|
||||
{$$ = $1 + ' ' +$3;}
|
||||
| ALPHA MINUS text
|
||||
{$$ = $1 + '-' +$3;}
|
||||
| ALPHA SPACE
|
||||
{$$ = $1;}
|
||||
| ALPHA
|
||||
{$$ = $1;}
|
||||
;
|
||||
|
||||
link: MINUS MINUS ARROW_POINT
|
||||
{$$ = {"type":"arrow"};}
|
||||
| MINUS MINUS ARROW_CIRCLE
|
||||
{$$ = {"type":"arrow_circle"};}
|
||||
| MINUS MINUS ARROW_CROSS
|
||||
{$$ = {"type":"arrow_cross"};}
|
||||
| MINUS MINUS MINUS
|
||||
{$$ = {"type":"arrow_open"};}
|
||||
;
|
||||
|
||||
stylesOpt: style
|
||||
{$$ = [$1]}
|
||||
| stylesOpt COMMA style
|
||||
{$1.push($3);$$ = $1;}
|
||||
;
|
||||
|
||||
style: styleComponent
|
||||
{$$=$1;}
|
||||
|style styleComponent
|
||||
{$$ = $1 + $2;}
|
||||
;
|
||||
|
||||
styleComponent: ALPHA
|
||||
{$$=$1}
|
||||
| COLON
|
||||
{$$=$1}
|
||||
| MINUS
|
||||
{$$=$1}
|
||||
| NUM
|
||||
{$$=$1}
|
||||
| UNIT
|
||||
{$$=$1}
|
||||
| SPACE
|
||||
{$$=$1}
|
||||
| HEX
|
||||
{$$=$1}
|
||||
;
|
||||
|
||||
colordef: COLOR
|
||||
{$$ = yytext;}
|
||||
| HEX
|
||||
{$$ = yytext;}
|
||||
;
|
||||
|
||||
borderWidth: NUM UNIT
|
||||
{$$ = $1+''+$2;}
|
||||
;
|
||||
|
||||
borderStyle: BORDER_STYLE
|
||||
{$$ = $1;}
|
||||
;
|
||||
|
||||
%%
|
||||
define('parser/flow',function(){
|
||||
console.log('bcs123');
|
||||
/*define('parser/mermaid',function(){
|
||||
return parser;
|
||||
});
|
||||
});*/
|
@@ -72,12 +72,12 @@
|
||||
}
|
||||
*/
|
||||
var parser = (function(){
|
||||
var o=function(k,v,o,l){for(o=o||{},l=k.length;l--;o[k[l]]=v);return o},$V0=[1,4],$V1=[5,8];
|
||||
var o=function(k,v,o,l){for(o=o||{},l=k.length;l--;o[k[l]]=v);return o},$V0=[1,3],$V1=[1,7],$V2=[1,8],$V3=[8,12,24],$V4=[1,24],$V5=[8,19,21,23],$V6=[1,48],$V7=[1,43],$V8=[1,49],$V9=[1,45],$Va=[1,44],$Vb=[1,46],$Vc=[1,47],$Vd=[1,53],$Ve=[8,29],$Vf=[6,8,15,17,24,29,31,32,33];
|
||||
var parser = {trace: function trace() { },
|
||||
yy: {},
|
||||
symbols_: {"error":2,"expressions":3,"id":4,"EOF":5,"flow":6,"STYLE":7,"MINUS":8,"word":9,"ALPHA":10,"$accept":0,"$end":1},
|
||||
terminals_: {2:"error",5:"EOF",7:"STYLE",8:"MINUS",10:"ALPHA"},
|
||||
productions_: [0,[3,2],[6,1],[6,1],[4,3],[4,1],[9,1]],
|
||||
symbols_: {"error":2,"expressions":3,"graph":4,"EOF":5,"SPACE":6,"edge":7,";":8,"styleStatement":9,"vertex":10,"link":11,"PIPE":12,"text":13,"STYLE":14,"ALPHA":15,"stylesOpt":16,"HEX":17,"SQS":18,"SQE":19,"PS":20,"PE":21,"DIAMOND_START":22,"DIAMOND_STOP":23,"MINUS":24,"ARROW_POINT":25,"ARROW_CIRCLE":26,"ARROW_CROSS":27,"style":28,"COMMA":29,"styleComponent":30,"COLON":31,"NUM":32,"UNIT":33,"colordef":34,"COLOR":35,"borderWidth":36,"borderStyle":37,"BORDER_STYLE":38,"$accept":0,"$end":1},
|
||||
terminals_: {2:"error",5:"EOF",6:"SPACE",8:";",12:"PIPE",14:"STYLE",15:"ALPHA",17:"HEX",18:"SQS",19:"SQE",20:"PS",21:"PE",22:"DIAMOND_START",23:"DIAMOND_STOP",24:"MINUS",25:"ARROW_POINT",26:"ARROW_CIRCLE",27:"ARROW_CROSS",29:"COMMA",31:"COLON",32:"NUM",33:"UNIT",35:"COLOR",38:"BORDER_STYLE"},
|
||||
productions_: [0,[3,2],[4,2],[4,3],[4,2],[7,1],[7,5],[7,3],[7,1],[9,5],[9,5],[10,4],[10,4],[10,4],[10,1],[13,3],[13,3],[13,2],[13,1],[11,3],[11,3],[11,3],[11,3],[16,1],[16,3],[28,1],[28,2],[30,1],[30,1],[30,1],[30,1],[30,1],[30,1],[30,1],[34,1],[34,1],[36,2],[37,1]],
|
||||
performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate /* action[1] */, $$ /* vstack */, _$ /* lstack */) {
|
||||
/* this == yyval */
|
||||
|
||||
@@ -86,22 +86,91 @@ switch (yystate) {
|
||||
case 1:
|
||||
return $$[$0-1];
|
||||
break;
|
||||
case 2:
|
||||
this.$='key';
|
||||
break;
|
||||
case 3:
|
||||
this.$=$$[$0];
|
||||
case 2: case 3:
|
||||
this.$ = $$[$0];
|
||||
break;
|
||||
case 4:
|
||||
this.$=$$[$0-2]+'-'+$$[$0]
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 5: case 6:
|
||||
case 5:
|
||||
this.$ = 'ya';
|
||||
break;
|
||||
case 6:
|
||||
yy.addLink($$[$0-4],$$[$0-2],$$[$0-3],$$[$0]);this.$ = 'oy'
|
||||
break;
|
||||
case 7:
|
||||
yy.addLink($$[$0-2],$$[$0],$$[$0-1]);this.$ = 'oy'
|
||||
break;
|
||||
case 8:
|
||||
this.$ = 'yo';
|
||||
break;
|
||||
case 9:
|
||||
this.$ = $$[$0-4];yy.addVertex($$[$0-2],undefined,undefined,$$[$0]);
|
||||
break;
|
||||
case 10:
|
||||
console.log('In parser - style: '+$$[$0]);this.$ = $$[$0-4];yy.updateLink($$[$0-2],$$[$0]);
|
||||
break;
|
||||
case 11:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'square');
|
||||
break;
|
||||
case 12:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'round');
|
||||
break;
|
||||
case 13:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'diamond');
|
||||
break;
|
||||
case 14:
|
||||
this.$ = $$[$0];yy.addVertex($$[$0]);
|
||||
break;
|
||||
case 15:
|
||||
this.$ = $$[$0-2] + ' ' +$$[$0];
|
||||
break;
|
||||
case 16:
|
||||
this.$ = $$[$0-2] + '-' +$$[$0];
|
||||
break;
|
||||
case 17:
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 18: case 37:
|
||||
this.$ = $$[$0];
|
||||
break;
|
||||
case 19:
|
||||
this.$ = {"type":"arrow"};
|
||||
break;
|
||||
case 20:
|
||||
this.$ = {"type":"arrow_circle"};
|
||||
break;
|
||||
case 21:
|
||||
this.$ = {"type":"arrow_cross"};
|
||||
break;
|
||||
case 22:
|
||||
this.$ = {"type":"arrow_open"};
|
||||
break;
|
||||
case 23:
|
||||
this.$ = [$$[$0]]
|
||||
break;
|
||||
case 24:
|
||||
$$[$0-2].push($$[$0]);this.$ = $$[$0-2];
|
||||
break;
|
||||
case 25:
|
||||
this.$=$$[$0];
|
||||
break;
|
||||
case 26:
|
||||
this.$ = $$[$0-1] + $$[$0];
|
||||
break;
|
||||
case 27: case 28: case 29: case 30: case 31: case 32: case 33:
|
||||
this.$=$$[$0]
|
||||
break;
|
||||
case 34: case 35:
|
||||
this.$ = yytext;
|
||||
break;
|
||||
case 36:
|
||||
this.$ = $$[$0-1]+''+$$[$0];
|
||||
break;
|
||||
}
|
||||
},
|
||||
table: [{3:1,4:2,9:3,10:$V0},{1:[3]},{5:[1,5],8:[1,6]},o($V1,[2,5]),o($V1,[2,6]),{1:[2,1]},{9:7,10:$V0},o($V1,[2,4])],
|
||||
defaultActions: {5:[2,1]},
|
||||
table: [{3:1,4:2,6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{1:[3]},{5:[1,9]},{4:10,6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{8:[1,11]},{8:[2,5]},{8:[2,8],11:12,24:[1,13]},{6:[1,14]},o($V3,[2,14],{18:[1,15],20:[1,16],22:[1,17]}),{1:[2,1]},{5:[2,2]},{4:18,5:[2,4],6:$V0,7:4,9:5,10:6,14:$V1,15:$V2},{10:19,15:$V2},{24:[1,20]},{15:[1,21],17:[1,22]},{13:23,15:$V4},{13:25,15:$V4},{13:26,15:$V4},{5:[2,3]},{8:[2,7],12:[1,27]},{24:[1,31],25:[1,28],26:[1,29],27:[1,30]},{6:[1,32]},{6:[1,33]},{19:[1,34]},o($V5,[2,18],{6:[1,35],24:[1,36]}),{21:[1,37]},{23:[1,38]},{13:39,15:$V4},{15:[2,19]},{15:[2,20]},{15:[2,21]},{15:[2,22]},{6:$V6,15:$V7,16:40,17:$V8,24:$V9,28:41,30:42,31:$Va,32:$Vb,33:$Vc},{6:$V6,15:$V7,16:50,17:$V8,24:$V9,28:41,30:42,31:$Va,32:$Vb,33:$Vc},o($V3,[2,11]),o($V5,[2,17],{13:51,15:$V4}),{13:52,15:$V4},o($V3,[2,12]),o($V3,[2,13]),{8:[2,6]},{8:[2,9],29:$Vd},o($Ve,[2,23],{30:54,6:$V6,15:$V7,17:$V8,24:$V9,31:$Va,32:$Vb,33:$Vc}),o($Vf,[2,25]),o($Vf,[2,27]),o($Vf,[2,28]),o($Vf,[2,29]),o($Vf,[2,30]),o($Vf,[2,31]),o($Vf,[2,32]),o($Vf,[2,33]),{8:[2,10],29:$Vd},o($V5,[2,15]),o($V5,[2,16]),{6:$V6,15:$V7,17:$V8,24:$V9,28:55,30:42,31:$Va,32:$Vb,33:$Vc},o($Vf,[2,26]),o($Ve,[2,24],{30:54,6:$V6,15:$V7,17:$V8,24:$V9,31:$Va,32:$Vb,33:$Vc})],
|
||||
defaultActions: {5:[2,5],9:[2,1],10:[2,2],18:[2,3],28:[2,19],29:[2,20],30:[2,21],31:[2,22],39:[2,6]},
|
||||
parseError: function parseError(str, hash) {
|
||||
if (hash.recoverable) {
|
||||
this.trace(str);
|
||||
@@ -171,7 +240,7 @@ parse: function parse(input) {
|
||||
} else {
|
||||
errStr = 'Parse error on line ' + (yylineno + 1) + ': Unexpected ' + (symbol == EOF ? 'end of input' : '\'' + (this.terminals_[symbol] || symbol) + '\'');
|
||||
}
|
||||
console.log(errStr, {
|
||||
this.parseError(errStr, {
|
||||
text: lexer.match,
|
||||
token: this.terminals_[symbol] || symbol,
|
||||
line: lexer.yylineno,
|
||||
@@ -247,10 +316,9 @@ parse: function parse(input) {
|
||||
return true;
|
||||
}};
|
||||
|
||||
define('parser/flow',function(){
|
||||
console.log('bcs123');
|
||||
/*define('parser/mermaid',function(){
|
||||
return parser;
|
||||
});/* generated by jison-lex 0.3.4 */
|
||||
});*//* generated by jison-lex 0.3.4 */
|
||||
var lexer = (function(){
|
||||
var lexer = ({
|
||||
|
||||
@@ -578,58 +646,60 @@ options: {},
|
||||
performAction: function anonymous(yy,yy_,$avoiding_name_collisions,YY_START) {
|
||||
var YYSTATE=YY_START;
|
||||
switch($avoiding_name_collisions) {
|
||||
case 0:return 7;
|
||||
case 0:return 14;
|
||||
break;
|
||||
case 1:return 10;
|
||||
case 1:return 17;
|
||||
break;
|
||||
case 2:return 'HEX';
|
||||
case 2:return 32;
|
||||
break;
|
||||
case 3:return 'NUM';
|
||||
case 3:return 'BRKT';
|
||||
break;
|
||||
case 4:return 'UNIT';
|
||||
case 4:return 33;
|
||||
break;
|
||||
case 5:return 'UNIT';
|
||||
case 5:return 33;
|
||||
break;
|
||||
case 6:return 'UNIT';
|
||||
case 6:return 33;
|
||||
break;
|
||||
case 7:return 'COLON';
|
||||
case 7:return 31;
|
||||
break;
|
||||
case 8:return 8;
|
||||
case 8:return 24;
|
||||
break;
|
||||
case 9:return ';';
|
||||
case 9:return 8;
|
||||
break;
|
||||
case 10:return 'COMMA';
|
||||
case 10:return 29;
|
||||
break;
|
||||
case 11:return 'ARROW_CROSS';
|
||||
case 11:return 27;
|
||||
break;
|
||||
case 12:return 'ARROW_POINT';
|
||||
case 12:return 25;
|
||||
break;
|
||||
case 13:return 'ARROW_CIRCLE';
|
||||
case 13:return 26;
|
||||
break;
|
||||
case 14:return 'PIPE';
|
||||
case 14:return 15;
|
||||
break;
|
||||
case 15:return 'PS';
|
||||
case 15:return 12;
|
||||
break;
|
||||
case 16:return 'PE';
|
||||
case 16:return 20;
|
||||
break;
|
||||
case 17:return 'SQS';
|
||||
case 17:return 21;
|
||||
break;
|
||||
case 18:return 'SQE';
|
||||
case 18:return 18;
|
||||
break;
|
||||
case 19:return 'DIAMOND_START'
|
||||
case 19:return 19;
|
||||
break;
|
||||
case 20:return 'DIAMOND_STOP'
|
||||
case 20:return 22
|
||||
break;
|
||||
case 21:return 'SPACE';
|
||||
case 21:return 23
|
||||
break;
|
||||
case 22:return 'NEWLINE';
|
||||
case 22:return 6;
|
||||
break;
|
||||
case 23:return 5;
|
||||
case 23:return 'NEWLINE';
|
||||
break;
|
||||
case 24:return 5;
|
||||
break;
|
||||
}
|
||||
},
|
||||
rules: [/^(?:style\b)/,/^(?:[a-zåäöæøA-ZÅÄÖÆØ]+)/,/^(?:#[a-f0-9]+)/,/^(?:[0-9]+)/,/^(?:px\b)/,/^(?:pt\b)/,/^(?:dot\b)/,/^(?::)/,/^(?:-)/,/^(?:;)/,/^(?:,)/,/^(?:[x])/,/^(?:>)/,/^(?:[o])/,/^(?:\|)/,/^(?:\()/,/^(?:\))/,/^(?:\[)/,/^(?:\])/,/^(?:\{)/,/^(?:\})/,/^(?:\s)/,/^(?:\n)/,/^(?:$)/],
|
||||
conditions: {"INITIAL":{"rules":[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23],"inclusive":true}}
|
||||
rules: [/^(?:style\b)/,/^(?:#[a-f0-9]+)/,/^(?:[0-9]+)/,/^(?:#)/,/^(?:px\b)/,/^(?:pt\b)/,/^(?:dot\b)/,/^(?::)/,/^(?:-)/,/^(?:;)/,/^(?:,)/,/^(?:[x])/,/^(?:>)/,/^(?:[o])/,/^(?:[a-zåäöæøA-ZÅÄÖÆØ]+)/,/^(?:\|)/,/^(?:\()/,/^(?:\))/,/^(?:\[)/,/^(?:\])/,/^(?:\{)/,/^(?:\})/,/^(?:\s)/,/^(?:\n)/,/^(?:$)/],
|
||||
conditions: {"INITIAL":{"rules":[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24],"inclusive":true}}
|
||||
});
|
||||
return lexer;
|
||||
})();
|
||||
@@ -657,4 +727,4 @@ exports.main = function commonjsMain(args) {
|
||||
if (typeof module !== 'undefined' && require.main === module) {
|
||||
exports.main(process.argv.slice(1));
|
||||
}
|
||||
}
|
||||
}
|
@@ -1,22 +1,227 @@
|
||||
/**
|
||||
* Created by knut on 14-11-03.
|
||||
*/
|
||||
define('parser/flow.spec',['parser/graph','parser/flow'],function(graph, p){
|
||||
|
||||
describe('when parsing ',function(){
|
||||
beforeEach(function(){
|
||||
graph.clear();
|
||||
p.yy = graph;
|
||||
/*p.parse.parseError= function parseError(str, hash) {
|
||||
console.log(str);
|
||||
}*/
|
||||
});
|
||||
|
||||
it('should handle a nodes and edges',function(){
|
||||
var res = p.parse('apa-apa-åäö');
|
||||
console.log('Done parsing:' + res);
|
||||
});
|
||||
* Created by knut on 14-11-03.
|
||||
*/
|
||||
describe('when parsing ',function(){
|
||||
beforeEach(function(){
|
||||
mermaid.graph.clear();
|
||||
parser.yy = mermaid.graph;
|
||||
/*parser.parse.parseError= function parseError(str, hash) {
|
||||
console.log(str);
|
||||
}*/
|
||||
console.log('in mm spec');
|
||||
});
|
||||
|
||||
});
|
||||
it('should handle a nodes and edges',function(){
|
||||
var res = parser.parse('A-->B;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['A'].id).toBe('A');
|
||||
expect(vert['B'].id).toBe('B');
|
||||
expect(edges.length).toBe(1);
|
||||
expect(edges[0].start).toBe('A');
|
||||
expect(edges[0].end).toBe('B');
|
||||
expect(edges[0].type).toBe('arrow');
|
||||
expect(edges[0].text).toBe('');
|
||||
});
|
||||
|
||||
it('should handle open ended edges',function(){
|
||||
var res = parser.parse('A---B;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_open');
|
||||
});
|
||||
|
||||
it('should handle cross ended edges',function(){
|
||||
var res = parser.parse('A--xB;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle open ended edges',function(){
|
||||
var res = parser.parse('A--oB;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_circle');
|
||||
});
|
||||
|
||||
it('should handle text on edges without space',function(){
|
||||
var res = parser.parse('A--xB|textNoSpace;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle text on edges with space',function(){
|
||||
var res = parser.parse('A--xB|text including space;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle multi-line text',function(){
|
||||
var res = parser.parse('A--oB|text space;\nB-->C|more text with space;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(edges[0].type).toBe('arrow_circle');
|
||||
expect(edges[1].type).toBe('arrow');
|
||||
expect(vert['A'].id).toBe('A');
|
||||
expect(vert['B'].id).toBe('B');
|
||||
expect(vert['C'].id).toBe('C');
|
||||
expect(edges.length).toBe(2);
|
||||
expect(edges[0].start).toBe('A');
|
||||
expect(edges[0].end).toBe('B');
|
||||
expect(edges[0].text).toBe('text space');
|
||||
expect(edges[1].start).toBe('B');
|
||||
expect(edges[1].end).toBe('C');
|
||||
expect(edges[1].text).toBe('more text with space');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = parser.parse('A[chimpansen hoppar]-->C;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('square');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = parser.parse('A(chimpansen hoppar)-->C;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('round');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = parser.parse('A{chimpansen hoppar}-->C;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('diamond');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = parser.parse('A-->C{Chimpansen hoppar};');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['C'].type).toBe('diamond');
|
||||
expect(vert['C'].text).toBe('Chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with åäö and minus',function(){
|
||||
var res = parser.parse('A-->C{Chimpansen hoppar åäö-ÅÄÖ};');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['C'].type).toBe('diamond');
|
||||
expect(vert['C'].text).toBe('Chimpansen hoppar åäö-ÅÄÖ');
|
||||
});
|
||||
|
||||
it('should handle a single node',function(){
|
||||
// Silly but syntactically correct
|
||||
var res = parser.parse('A;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(edges.length).toBe(0);
|
||||
expect(vert['A'].styles.length).toBe(0);
|
||||
});
|
||||
|
||||
//console.log(parser.parse('style Q background:#fff;'));
|
||||
it('should handle styles for vertices',function(){
|
||||
var res = parser.parse('style Q background:#fff;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
var style = vert['Q'].styles[0];
|
||||
|
||||
expect(vert['Q'].styles.length).toBe(1);
|
||||
expect(vert['Q'].styles[0]).toBe('background:#fff');
|
||||
});
|
||||
|
||||
//console.log(parser.parse('style Q background:#fff;'));
|
||||
it('should handle styles for edges',function(){
|
||||
var res = parser.parse('a-->b;\nstyle #0 stroke: #f66;');
|
||||
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(edges.length).toBe(1);
|
||||
});
|
||||
|
||||
it('should handle multiple styles for a vortex',function(){
|
||||
var res = parser.parse('style R background:#fff,border:1px solid red;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['R'].styles.length).toBe(2);
|
||||
expect(vert['R'].styles[0]).toBe('background:#fff');
|
||||
expect(vert['R'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
|
||||
it('should handle multiple styles in a graph',function(){
|
||||
var res = parser.parse('style S background:#aaa;\nstyle T background:#bbb,border:1px solid red;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['S'].styles.length).toBe(1);
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['S'].styles[0]).toBe('background:#aaa');
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
|
||||
it('should handle styles and graph definitons in a graph',function(){
|
||||
var res = parser.parse('S-->T;\nstyle S background:#aaa;\nstyle T background:#bbb,border:1px solid red;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
var edges = parser.yy.getEdges();
|
||||
|
||||
expect(vert['S'].styles.length).toBe(1);
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['S'].styles[0]).toBe('background:#aaa');
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
it('should handle styles and graph definitons in a graph',function(){
|
||||
var res = parser.parse('style T background:#bbb,border:1px solid red;');
|
||||
//var res = parser.parse('style T background: #bbb;');
|
||||
|
||||
var vert = parser.yy.getVertices();
|
||||
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
});
|
||||
|
@@ -1,145 +0,0 @@
|
||||
/* description: Parses end executes mathematical expressions. */
|
||||
|
||||
/* lexical grammar */
|
||||
%lex
|
||||
|
||||
%%
|
||||
"style" return 'STYLE';
|
||||
\#[a-f0-9]+ return 'HEX';
|
||||
[0-9]+ return 'NUM';
|
||||
"px" return 'UNIT';
|
||||
"pt" return 'UNIT';
|
||||
"dot" return 'UNIT';
|
||||
":" return 'COLON';
|
||||
\- return 'MINUS';
|
||||
";" return ';';
|
||||
"," return 'COMMA';
|
||||
[x] return 'ARROW_CROSS';
|
||||
">" return 'ARROW_POINT';
|
||||
[o] return 'ARROW_CIRCLE';
|
||||
[a-zåäöæøA-ZÅÄÖÆØ]+ return 'ALPHA';
|
||||
"|" return 'PIPE';
|
||||
"(" return 'PS';
|
||||
")" return 'PE';
|
||||
"[" return 'SQS';
|
||||
"]" return 'SQE';
|
||||
"{" return 'DIAMOND_START'
|
||||
"}" return 'DIAMOND_STOP'
|
||||
\s return 'SPACE';
|
||||
\n return 'NEWLINE';
|
||||
<<EOF>> return 'EOF';
|
||||
|
||||
/lex
|
||||
|
||||
/* operator associations and precedence */
|
||||
|
||||
%left '^'
|
||||
|
||||
%start expressions
|
||||
|
||||
%% /* language grammar */
|
||||
|
||||
expressions
|
||||
: graph EOF
|
||||
{return $1;}
|
||||
;
|
||||
|
||||
graph
|
||||
: SPACE graph
|
||||
{ $$ = $2;}
|
||||
| edge ';' graph
|
||||
{ $$ = $3;}
|
||||
| edge ';'
|
||||
{ $$ = $1;}
|
||||
;
|
||||
|
||||
edge: styleStatement
|
||||
{$$ = 'ya';}
|
||||
| vertex link vertex PIPE text
|
||||
{ yy.addLink($1,$3,$2,$5);$$ = 'oy'}
|
||||
| vertex link vertex
|
||||
{ yy.addLink($1,$3,$2);$$ = 'oy'}
|
||||
| vertex
|
||||
{$$ = 'yo';}
|
||||
;
|
||||
|
||||
styleStatement:STYLE SPACE ALPHA SPACE stylesOpt
|
||||
{console.log('a4');$$ = $1;yy.addVertex($3,undefined,undefined,$5);}
|
||||
;
|
||||
|
||||
vertex: ALPHA SQS text SQE
|
||||
{$$ = $1;yy.addVertex($1,$3,'square');}
|
||||
| ALPHA PS text PE
|
||||
{$$ = $1;yy.addVertex($1,$3,'round');}
|
||||
| ALPHA DIAMOND_START text DIAMOND_STOP
|
||||
{$$ = $1;yy.addVertex($1,$3,'diamond');}
|
||||
| ALPHA
|
||||
{$$ = $1;yy.addVertex($1);}
|
||||
;
|
||||
// Characters and spaces
|
||||
text: ALPHA SPACE text
|
||||
{$$ = $1 + ' ' +$3;}
|
||||
| ALPHA MINUS text
|
||||
{$$ = $1 + '-' +$3;}
|
||||
| ALPHA SPACE
|
||||
{$$ = $1;}
|
||||
| ALPHA
|
||||
{$$ = $1;}
|
||||
;
|
||||
|
||||
link: MINUS MINUS ARROW_POINT
|
||||
{$$ = {"type":"arrow"};}
|
||||
| MINUS MINUS ARROW_CIRCLE
|
||||
{$$ = {"type":"arrow_circle"};}
|
||||
| MINUS MINUS ARROW_CROSS
|
||||
{$$ = {"type":"arrow_cross"};}
|
||||
| MINUS MINUS MINUS
|
||||
{$$ = {"type":"arrow_open"};}
|
||||
;
|
||||
|
||||
stylesOpt: style
|
||||
{console.log('a:'+$1);$$ = [$1]}
|
||||
| stylesOpt COMMA style
|
||||
{console.log('a3:'+$1.length+','+$3);$1.push($3);$$ = $1;}
|
||||
;
|
||||
|
||||
style: styleComponent
|
||||
{$$=$1;}
|
||||
|style styleComponent
|
||||
{console.log('b1:');$$ = $1 + $2;}
|
||||
;
|
||||
|
||||
styleComponent: ALPHA
|
||||
{$$=$1}
|
||||
| COLON
|
||||
{$$=$1}
|
||||
| MINUS
|
||||
{$$=$1}
|
||||
| NUM
|
||||
{$$=$1}
|
||||
| UNIT
|
||||
{$$=$1}
|
||||
| SPACE
|
||||
{$$=$1}
|
||||
| HEX
|
||||
{$$=$1}
|
||||
;
|
||||
|
||||
colordef: COLOR
|
||||
{$$ = yytext;}
|
||||
| HEX
|
||||
{$$ = yytext;}
|
||||
;
|
||||
|
||||
borderWidth: NUM UNIT
|
||||
{$$ = $1+''+$2;}
|
||||
;
|
||||
|
||||
borderStyle: BORDER_STYLE
|
||||
{$$ = $1;}
|
||||
;
|
||||
|
||||
%%
|
||||
define('parser/mermaid',function(){
|
||||
return parser;
|
||||
});
|
@@ -1,222 +0,0 @@
|
||||
/**
|
||||
* Created by knut on 14-11-03.
|
||||
*/
|
||||
define('parser/mermaid.spec',['parser/graph','parser/mermaid'],function(graph, p){
|
||||
|
||||
describe('when parsing ',function(){
|
||||
beforeEach(function(){
|
||||
graph.clear();
|
||||
p.yy = graph;
|
||||
/*p.parse.parseError= function parseError(str, hash) {
|
||||
console.log(str);
|
||||
}*/
|
||||
console.log('in mm spec');
|
||||
});
|
||||
|
||||
it('should handle a nodes and edges',function(){
|
||||
var res = p.parse('A-->B;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['A'].id).toBe('A');
|
||||
expect(vert['B'].id).toBe('B');
|
||||
expect(edges.length).toBe(1);
|
||||
expect(edges[0].start).toBe('A');
|
||||
expect(edges[0].end).toBe('B');
|
||||
expect(edges[0].type).toBe('arrow');
|
||||
expect(edges[0].text).toBe('');
|
||||
});
|
||||
|
||||
it('should handle open ended edges',function(){
|
||||
var res = p.parse('A---B;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_open');
|
||||
});
|
||||
|
||||
it('should handle cross ended edges',function(){
|
||||
var res = p.parse('A--xB;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle open ended edges',function(){
|
||||
var res = p.parse('A--oB;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_circle');
|
||||
});
|
||||
|
||||
it('should handle text on edges without space',function(){
|
||||
var res = p.parse('A--xB|textNoSpace;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle text on edges with space',function(){
|
||||
var res = p.parse('A--xB|text including space;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
|
||||
expect(edges[0].type).toBe('arrow_cross');
|
||||
});
|
||||
|
||||
it('should handle multi-line text',function(){
|
||||
var res = p.parse('A--oB|text space;\nB-->C|more text with space;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(edges[0].type).toBe('arrow_circle');
|
||||
expect(edges[1].type).toBe('arrow');
|
||||
expect(vert['A'].id).toBe('A');
|
||||
expect(vert['B'].id).toBe('B');
|
||||
expect(vert['C'].id).toBe('C');
|
||||
expect(edges.length).toBe(2);
|
||||
expect(edges[0].start).toBe('A');
|
||||
expect(edges[0].end).toBe('B');
|
||||
expect(edges[0].text).toBe('text space');
|
||||
expect(edges[1].start).toBe('B');
|
||||
expect(edges[1].end).toBe('C');
|
||||
expect(edges[1].text).toBe('more text with space');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = p.parse('A[chimpansen hoppar]-->C;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('square');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = p.parse('A(chimpansen hoppar)-->C;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('round');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = p.parse('A{chimpansen hoppar}-->C;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['A'].type).toBe('diamond');
|
||||
expect(vert['A'].text).toBe('chimpansen hoppar');
|
||||
});
|
||||
it('should handle text in vertices with space',function(){
|
||||
var res = p.parse('A-->C{Chimpansen hoppar};');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['C'].type).toBe('diamond');
|
||||
expect(vert['C'].text).toBe('Chimpansen hoppar');
|
||||
});
|
||||
|
||||
it('should handle text in vertices with åäö and minus',function(){
|
||||
var res = p.parse('A-->C{Chimpansen hoppar åäö-ÅÄÖ};');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['C'].type).toBe('diamond');
|
||||
expect(vert['C'].text).toBe('Chimpansen hoppar åäö-ÅÄÖ');
|
||||
});
|
||||
|
||||
it('should handle a single node',function(){
|
||||
// Silly but syntactically correct
|
||||
var res = p.parse('A;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(edges.length).toBe(0);
|
||||
expect(vert['A'].styles.length).toBe(0);
|
||||
});
|
||||
|
||||
//console.log(p.parse('style Q background:#fff;'));
|
||||
it('should handle styles for vertices',function(){
|
||||
var res = p.parse('style Q background:#fff;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
var style = vert['Q'].styles[0];
|
||||
|
||||
expect(vert['Q'].styles.length).toBe(1);
|
||||
expect(vert['Q'].styles[0]).toBe('background:#fff');
|
||||
});
|
||||
it('should handle multiple styles for a vortex',function(){
|
||||
var res = p.parse('style R background:#fff,border:1px solid red;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['R'].styles.length).toBe(2);
|
||||
expect(vert['R'].styles[0]).toBe('background:#fff');
|
||||
expect(vert['R'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
|
||||
it('should handle multiple styles in a graph',function(){
|
||||
var res = p.parse('style S background:#aaa;\nstyle T background:#bbb,border:1px solid red;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['S'].styles.length).toBe(1);
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['S'].styles[0]).toBe('background:#aaa');
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
|
||||
it('should handle styles and graph definitons in a graph',function(){
|
||||
var res = p.parse('S-->T;\nstyle S background:#aaa;\nstyle T background:#bbb,border:1px solid red;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
var edges = p.yy.getEdges();
|
||||
|
||||
expect(vert['S'].styles.length).toBe(1);
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['S'].styles[0]).toBe('background:#aaa');
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
it('should handle styles and graph definitons in a graph',function(){
|
||||
var res = p.parse('style T background:#bbb,border:1px solid red;');
|
||||
//var res = p.parse('style T background: #bbb;');
|
||||
|
||||
var vert = p.yy.getVertices();
|
||||
|
||||
expect(vert['T'].styles.length).toBe(2);
|
||||
expect(vert['T'].styles[0]).toBe('background:#bbb');
|
||||
expect(vert['T'].styles[1]).toBe('border:1px solid red');
|
||||
});
|
||||
});
|
||||
|
||||
});
|
||||
|
@@ -1,743 +0,0 @@
|
||||
/* parser generated by jison 0.4.15 */
|
||||
/*
|
||||
Returns a Parser object of the following structure:
|
||||
|
||||
Parser: {
|
||||
yy: {}
|
||||
}
|
||||
|
||||
Parser.prototype: {
|
||||
yy: {},
|
||||
trace: function(),
|
||||
symbols_: {associative list: name ==> number},
|
||||
terminals_: {associative list: number ==> name},
|
||||
productions_: [...],
|
||||
performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate, $$, _$),
|
||||
table: [...],
|
||||
defaultActions: {...},
|
||||
parseError: function(str, hash),
|
||||
parse: function(input),
|
||||
|
||||
lexer: {
|
||||
EOF: 1,
|
||||
parseError: function(str, hash),
|
||||
setInput: function(input),
|
||||
input: function(),
|
||||
unput: function(str),
|
||||
more: function(),
|
||||
less: function(n),
|
||||
pastInput: function(),
|
||||
upcomingInput: function(),
|
||||
showPosition: function(),
|
||||
test_match: function(regex_match_array, rule_index),
|
||||
next: function(),
|
||||
lex: function(),
|
||||
begin: function(condition),
|
||||
popState: function(),
|
||||
_currentRules: function(),
|
||||
topState: function(),
|
||||
pushState: function(condition),
|
||||
|
||||
options: {
|
||||
ranges: boolean (optional: true ==> token location info will include a .range[] member)
|
||||
flex: boolean (optional: true ==> flex-like lexing behaviour where the rules are tested exhaustively to find the longest match)
|
||||
backtrack_lexer: boolean (optional: true ==> lexer regexes are tested in order and for each matching regex the action code is invoked; the lexer terminates the scan when a token is returned by the action code)
|
||||
},
|
||||
|
||||
performAction: function(yy, yy_, $avoiding_name_collisions, YY_START),
|
||||
rules: [...],
|
||||
conditions: {associative list: name ==> set},
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
token location info (@$, _$, etc.): {
|
||||
first_line: n,
|
||||
last_line: n,
|
||||
first_column: n,
|
||||
last_column: n,
|
||||
range: [start_number, end_number] (where the numbers are indexes into the input string, regular zero-based)
|
||||
}
|
||||
|
||||
|
||||
the parseError function receives a 'hash' object with these members for lexer and parser errors: {
|
||||
text: (matched text)
|
||||
token: (the produced terminal token, if any)
|
||||
line: (yylineno)
|
||||
}
|
||||
while parser (grammar) errors will also provide these members, i.e. parser errors deliver a superset of attributes: {
|
||||
loc: (yylloc)
|
||||
expected: (string describing the set of expected tokens)
|
||||
recoverable: (boolean: TRUE when the parser has a error recovery rule available for this particular error)
|
||||
}
|
||||
*/
|
||||
var parser = (function(){
|
||||
var o=function(k,v,o,l){for(o=o||{},l=k.length;l--;o[k[l]]=v);return o},$V0=[1,3],$V1=[1,6],$V2=[1,7],$V3=[8,11,22,23,24,25],$V4=[13,14],$V5=[1,24],$V6=[8,17,19,21];
|
||||
var parser = {trace: function trace() { },
|
||||
yy: {},
|
||||
symbols_: {"error":2,"expressions":3,"graph":4,"EOF":5,"SPACE":6,"edge":7,";":8,"vertex":9,"link":10,"PIPE":11,"text":12,"STYLE":13,"ALPHA":14,"stylesOpt":15,"SQS":16,"SQE":17,"PS":18,"PE":19,"DIAMOND_START":20,"DIAMOND_STOP":21,"ARROW_POINT":22,"ARROW_CIRCLE":23,"ARROW_CROSS":24,"ARROW_OPEN":25,"styles":26,"styledef":27,"COMMA":28,"BKG":29,"COLON":30,"colordef":31,"COL":32,"COLORDEF":33,"BORDER":34,"borderWidth":35,"borderStyle":36,"styledef3":37,"styles2":38,"styledef2":39,"NUM":40,"UNIT":41,"HEX":42,"COLOR":43,"BORDER_STYLE":44,"$accept":0,"$end":1},
|
||||
terminals_: {2:"error",5:"EOF",6:"SPACE",8:";",11:"PIPE",13:"STYLE",14:"ALPHA",16:"SQS",17:"SQE",18:"PS",19:"PE",20:"DIAMOND_START",21:"DIAMOND_STOP",22:"ARROW_POINT",23:"ARROW_CIRCLE",24:"ARROW_CROSS",25:"ARROW_OPEN",28:"COMMA",29:"BKG",30:"COLON",32:"COL",33:"COLORDEF",34:"BORDER",40:"NUM",41:"UNIT",42:"HEX",43:"COLOR",44:"BORDER_STYLE"},
|
||||
productions_: [0,[3,2],[4,2],[4,3],[4,2],[7,5],[7,3],[7,1],[9,5],[9,4],[9,4],[9,4],[9,1],[12,3],[12,2],[12,1],[10,1],[10,1],[10,1],[10,1],[26,1],[26,3],[27,3],[27,3],[27,7],[15,1],[38,1],[38,3],[39,3],[39,3],[39,7],[37,1],[37,1],[37,1],[37,1],[37,1],[31,1],[31,1],[35,2],[36,1]],
|
||||
performAction: function anonymous(yytext, yyleng, yylineno, yy, yystate /* action[1] */, $$ /* vstack */, _$ /* lstack */) {
|
||||
/* this == yyval */
|
||||
|
||||
var $0 = $$.length - 1;
|
||||
switch (yystate) {
|
||||
case 1:
|
||||
return $$[$0-1];
|
||||
break;
|
||||
case 2: case 3:
|
||||
this.$ = $$[$0];
|
||||
break;
|
||||
case 4:
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 5:
|
||||
yy.addLink($$[$0-4],$$[$0-2],$$[$0-3],$$[$0]);this.$ = 'oy'
|
||||
break;
|
||||
case 6:
|
||||
yy.addLink($$[$0-2],$$[$0],$$[$0-1]);this.$ = 'oy'
|
||||
break;
|
||||
case 7:
|
||||
this.$ = 'yo';
|
||||
break;
|
||||
case 8:
|
||||
this.$ = $$[$0-4];yy.addVertex($$[$0-2],undefined,undefined,$$[$0]);
|
||||
break;
|
||||
case 9:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'square');
|
||||
break;
|
||||
case 10:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'round');
|
||||
break;
|
||||
case 11:
|
||||
this.$ = $$[$0-3];yy.addVertex($$[$0-3],$$[$0-1],'diamond');
|
||||
break;
|
||||
case 12:
|
||||
this.$ = $$[$0];yy.addVertex($$[$0]);
|
||||
break;
|
||||
case 13:
|
||||
this.$ = $$[$0-2] + ' ' +$$[$0];
|
||||
break;
|
||||
case 14:
|
||||
this.$ = $$[$0-1];
|
||||
break;
|
||||
case 15: case 39:
|
||||
this.$ = $$[$0];
|
||||
break;
|
||||
case 16:
|
||||
this.$ = {"type":"arrow"};
|
||||
break;
|
||||
case 17:
|
||||
this.$ = {"type":"arrow_circle"};
|
||||
break;
|
||||
case 18:
|
||||
this.$ = {"type":"arrow_cross"};
|
||||
break;
|
||||
case 19:
|
||||
this.$ = {"type":"arrow_open"};
|
||||
break;
|
||||
case 20: case 26:
|
||||
this.$ = [$$[$0]];
|
||||
break;
|
||||
case 21: case 27:
|
||||
$$[$0-2].push($$[$0]);this.$ = $$[$0-2];
|
||||
break;
|
||||
case 22: case 28:
|
||||
this.$={"background":$$[$0]}
|
||||
break;
|
||||
case 23: case 29:
|
||||
this.$={"color":$$[$0]}
|
||||
break;
|
||||
case 24: case 30:
|
||||
this.$={"border":$$[$0-4]+' '+$$[$0-2]+' '+$$[$0]}
|
||||
break;
|
||||
case 25:
|
||||
this.$ = $yytext
|
||||
break;
|
||||
case 31: case 32: case 33: case 34:
|
||||
this.$=$$[$0]
|
||||
break;
|
||||
case 35:
|
||||
this.$=$yytext
|
||||
break;
|
||||
case 36: case 37:
|
||||
this.$ = yytext;
|
||||
break;
|
||||
case 38:
|
||||
this.$ = $$[$0-1]+''+$$[$0];
|
||||
break;
|
||||
}
|
||||
},
|
||||
table: [{3:1,4:2,6:$V0,7:4,9:5,13:$V1,14:$V2},{1:[3]},{5:[1,8]},{4:9,6:$V0,7:4,9:5,13:$V1,14:$V2},{8:[1,10]},{8:[2,7],10:11,22:[1,12],23:[1,13],24:[1,14],25:[1,15]},{6:[1,16]},o($V3,[2,12],{16:[1,17],18:[1,18],20:[1,19]}),{1:[2,1]},{5:[2,2]},{4:20,5:[2,4],6:$V0,7:4,9:5,13:$V1,14:$V2},{9:21,13:$V1,14:$V2},o($V4,[2,16]),o($V4,[2,17]),o($V4,[2,18]),o($V4,[2,19]),{14:[1,22]},{12:23,14:$V5},{12:25,14:$V5},{12:26,14:$V5},{5:[2,3]},{8:[2,6],11:[1,27]},{6:[1,28]},{17:[1,29]},o($V6,[2,15],{6:[1,30]}),{19:[1,31]},{21:[1,32]},{12:33,14:$V5},{6:[1,39],14:[1,36],15:34,37:35,40:[1,37],41:[1,38],42:[1,40]},o($V3,[2,9]),o($V6,[2,14],{12:41,14:$V5}),o($V3,[2,10]),o($V3,[2,11]),{8:[2,5]},o($V3,[2,8]),o($V3,[2,25]),o($V3,[2,31]),o($V3,[2,32]),o($V3,[2,33]),o($V3,[2,34]),o($V3,[2,35]),o($V6,[2,13])],
|
||||
defaultActions: {8:[2,1],9:[2,2],20:[2,3],33:[2,5]},
|
||||
parseError: function parseError(str, hash) {
|
||||
if (hash.recoverable) {
|
||||
this.trace(str);
|
||||
} else {
|
||||
throw new Error(str);
|
||||
}
|
||||
},
|
||||
parse: function parse(input) {
|
||||
var self = this, stack = [0], tstack = [], vstack = [null], lstack = [], table = this.table, yytext = '', yylineno = 0, yyleng = 0, recovering = 0, TERROR = 2, EOF = 1;
|
||||
var args = lstack.slice.call(arguments, 1);
|
||||
var lexer = Object.create(this.lexer);
|
||||
var sharedState = { yy: {} };
|
||||
for (var k in this.yy) {
|
||||
if (Object.prototype.hasOwnProperty.call(this.yy, k)) {
|
||||
sharedState.yy[k] = this.yy[k];
|
||||
}
|
||||
}
|
||||
lexer.setInput(input, sharedState.yy);
|
||||
sharedState.yy.lexer = lexer;
|
||||
sharedState.yy.parser = this;
|
||||
if (typeof lexer.yylloc == 'undefined') {
|
||||
lexer.yylloc = {};
|
||||
}
|
||||
var yyloc = lexer.yylloc;
|
||||
lstack.push(yyloc);
|
||||
var ranges = lexer.options && lexer.options.ranges;
|
||||
if (typeof sharedState.yy.parseError === 'function') {
|
||||
this.parseError = sharedState.yy.parseError;
|
||||
} else {
|
||||
this.parseError = Object.getPrototypeOf(this).parseError;
|
||||
}
|
||||
function popStack(n) {
|
||||
stack.length = stack.length - 2 * n;
|
||||
vstack.length = vstack.length - n;
|
||||
lstack.length = lstack.length - n;
|
||||
}
|
||||
_token_stack:
|
||||
function lex() {
|
||||
var token;
|
||||
token = lexer.lex() || EOF;
|
||||
if (typeof token !== 'number') {
|
||||
token = self.symbols_[token] || token;
|
||||
}
|
||||
return token;
|
||||
}
|
||||
var symbol, preErrorSymbol, state, action, a, r, yyval = {}, p, len, newState, expected;
|
||||
while (true) {
|
||||
state = stack[stack.length - 1];
|
||||
if (this.defaultActions[state]) {
|
||||
action = this.defaultActions[state];
|
||||
} else {
|
||||
if (symbol === null || typeof symbol == 'undefined') {
|
||||
symbol = lex();
|
||||
}
|
||||
action = table[state] && table[state][symbol];
|
||||
}
|
||||
if (typeof action === 'undefined' || !action.length || !action[0]) {
|
||||
var errStr = '';
|
||||
expected = [];
|
||||
for (p in table[state]) {
|
||||
if (this.terminals_[p] && p > TERROR) {
|
||||
expected.push('\'' + this.terminals_[p] + '\'');
|
||||
}
|
||||
}
|
||||
if (lexer.showPosition) {
|
||||
errStr = 'Parse error on line ' + (yylineno + 1) + ':\n' + lexer.showPosition() + '\nExpecting ' + expected.join(', ') + ', got \'' + (this.terminals_[symbol] || symbol) + '\'';
|
||||
} else {
|
||||
errStr = 'Parse error on line ' + (yylineno + 1) + ': Unexpected ' + (symbol == EOF ? 'end of input' : '\'' + (this.terminals_[symbol] || symbol) + '\'');
|
||||
}
|
||||
console.log(errStr, {
|
||||
text: lexer.match,
|
||||
token: this.terminals_[symbol] || symbol,
|
||||
line: lexer.yylineno,
|
||||
loc: yyloc,
|
||||
expected: expected
|
||||
});
|
||||
}
|
||||
if (action[0] instanceof Array && action.length > 1) {
|
||||
throw new Error('Parse Error: multiple actions possible at state: ' + state + ', token: ' + symbol);
|
||||
}
|
||||
switch (action[0]) {
|
||||
case 1:
|
||||
stack.push(symbol);
|
||||
vstack.push(lexer.yytext);
|
||||
lstack.push(lexer.yylloc);
|
||||
stack.push(action[1]);
|
||||
symbol = null;
|
||||
if (!preErrorSymbol) {
|
||||
yyleng = lexer.yyleng;
|
||||
yytext = lexer.yytext;
|
||||
yylineno = lexer.yylineno;
|
||||
yyloc = lexer.yylloc;
|
||||
if (recovering > 0) {
|
||||
recovering--;
|
||||
}
|
||||
} else {
|
||||
symbol = preErrorSymbol;
|
||||
preErrorSymbol = null;
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
len = this.productions_[action[1]][1];
|
||||
yyval.$ = vstack[vstack.length - len];
|
||||
yyval._$ = {
|
||||
first_line: lstack[lstack.length - (len || 1)].first_line,
|
||||
last_line: lstack[lstack.length - 1].last_line,
|
||||
first_column: lstack[lstack.length - (len || 1)].first_column,
|
||||
last_column: lstack[lstack.length - 1].last_column
|
||||
};
|
||||
if (ranges) {
|
||||
yyval._$.range = [
|
||||
lstack[lstack.length - (len || 1)].range[0],
|
||||
lstack[lstack.length - 1].range[1]
|
||||
];
|
||||
}
|
||||
r = this.performAction.apply(yyval, [
|
||||
yytext,
|
||||
yyleng,
|
||||
yylineno,
|
||||
sharedState.yy,
|
||||
action[1],
|
||||
vstack,
|
||||
lstack
|
||||
].concat(args));
|
||||
if (typeof r !== 'undefined') {
|
||||
return r;
|
||||
}
|
||||
if (len) {
|
||||
stack = stack.slice(0, -1 * len * 2);
|
||||
vstack = vstack.slice(0, -1 * len);
|
||||
lstack = lstack.slice(0, -1 * len);
|
||||
}
|
||||
stack.push(this.productions_[action[1]][0]);
|
||||
vstack.push(yyval.$);
|
||||
lstack.push(yyval._$);
|
||||
newState = table[stack[stack.length - 2]][stack[stack.length - 1]];
|
||||
stack.push(newState);
|
||||
break;
|
||||
case 3:
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return true;
|
||||
}};
|
||||
|
||||
define('parser/mermaid',function(){
|
||||
console.log('abc123'+parser.parseError);
|
||||
return parser;
|
||||
});/* generated by jison-lex 0.3.4 */
|
||||
var lexer = (function(){
|
||||
var lexer = ({
|
||||
|
||||
EOF:1,
|
||||
|
||||
parseError:function parseError(str, hash) {
|
||||
if (this.yy.parser) {
|
||||
this.yy.parser.parseError(str, hash);
|
||||
} else {
|
||||
throw new Error(str);
|
||||
}
|
||||
},
|
||||
|
||||
// resets the lexer, sets new input
|
||||
setInput:function (input, yy) {
|
||||
this.yy = yy || this.yy || {};
|
||||
this._input = input;
|
||||
this._more = this._backtrack = this.done = false;
|
||||
this.yylineno = this.yyleng = 0;
|
||||
this.yytext = this.matched = this.match = '';
|
||||
this.conditionStack = ['INITIAL'];
|
||||
this.yylloc = {
|
||||
first_line: 1,
|
||||
first_column: 0,
|
||||
last_line: 1,
|
||||
last_column: 0
|
||||
};
|
||||
if (this.options.ranges) {
|
||||
this.yylloc.range = [0,0];
|
||||
}
|
||||
this.offset = 0;
|
||||
return this;
|
||||
},
|
||||
|
||||
// consumes and returns one char from the input
|
||||
input:function () {
|
||||
var ch = this._input[0];
|
||||
this.yytext += ch;
|
||||
this.yyleng++;
|
||||
this.offset++;
|
||||
this.match += ch;
|
||||
this.matched += ch;
|
||||
var lines = ch.match(/(?:\r\n?|\n).*/g);
|
||||
if (lines) {
|
||||
this.yylineno++;
|
||||
this.yylloc.last_line++;
|
||||
} else {
|
||||
this.yylloc.last_column++;
|
||||
}
|
||||
if (this.options.ranges) {
|
||||
this.yylloc.range[1]++;
|
||||
}
|
||||
|
||||
this._input = this._input.slice(1);
|
||||
return ch;
|
||||
},
|
||||
|
||||
// unshifts one char (or a string) into the input
|
||||
unput:function (ch) {
|
||||
var len = ch.length;
|
||||
var lines = ch.split(/(?:\r\n?|\n)/g);
|
||||
|
||||
this._input = ch + this._input;
|
||||
this.yytext = this.yytext.substr(0, this.yytext.length - len);
|
||||
//this.yyleng -= len;
|
||||
this.offset -= len;
|
||||
var oldLines = this.match.split(/(?:\r\n?|\n)/g);
|
||||
this.match = this.match.substr(0, this.match.length - 1);
|
||||
this.matched = this.matched.substr(0, this.matched.length - 1);
|
||||
|
||||
if (lines.length - 1) {
|
||||
this.yylineno -= lines.length - 1;
|
||||
}
|
||||
var r = this.yylloc.range;
|
||||
|
||||
this.yylloc = {
|
||||
first_line: this.yylloc.first_line,
|
||||
last_line: this.yylineno + 1,
|
||||
first_column: this.yylloc.first_column,
|
||||
last_column: lines ?
|
||||
(lines.length === oldLines.length ? this.yylloc.first_column : 0)
|
||||
+ oldLines[oldLines.length - lines.length].length - lines[0].length :
|
||||
this.yylloc.first_column - len
|
||||
};
|
||||
|
||||
if (this.options.ranges) {
|
||||
this.yylloc.range = [r[0], r[0] + this.yyleng - len];
|
||||
}
|
||||
this.yyleng = this.yytext.length;
|
||||
return this;
|
||||
},
|
||||
|
||||
// When called from action, caches matched text and appends it on next action
|
||||
more:function () {
|
||||
this._more = true;
|
||||
return this;
|
||||
},
|
||||
|
||||
// When called from action, signals the lexer that this rule fails to match the input, so the next matching rule (regex) should be tested instead.
|
||||
reject:function () {
|
||||
if (this.options.backtrack_lexer) {
|
||||
this._backtrack = true;
|
||||
} else {
|
||||
return this.parseError('Lexical error on line ' + (this.yylineno + 1) + '. You can only invoke reject() in the lexer when the lexer is of the backtracking persuasion (options.backtrack_lexer = true).\n' + this.showPosition(), {
|
||||
text: "",
|
||||
token: null,
|
||||
line: this.yylineno
|
||||
});
|
||||
|
||||
}
|
||||
return this;
|
||||
},
|
||||
|
||||
// retain first n characters of the match
|
||||
less:function (n) {
|
||||
this.unput(this.match.slice(n));
|
||||
},
|
||||
|
||||
// displays already matched input, i.e. for error messages
|
||||
pastInput:function () {
|
||||
var past = this.matched.substr(0, this.matched.length - this.match.length);
|
||||
return (past.length > 20 ? '...':'') + past.substr(-20).replace(/\n/g, "");
|
||||
},
|
||||
|
||||
// displays upcoming input, i.e. for error messages
|
||||
upcomingInput:function () {
|
||||
var next = this.match;
|
||||
if (next.length < 20) {
|
||||
next += this._input.substr(0, 20-next.length);
|
||||
}
|
||||
return (next.substr(0,20) + (next.length > 20 ? '...' : '')).replace(/\n/g, "");
|
||||
},
|
||||
|
||||
// displays the character position where the lexing error occurred, i.e. for error messages
|
||||
showPosition:function () {
|
||||
var pre = this.pastInput();
|
||||
var c = new Array(pre.length + 1).join("-");
|
||||
return pre + this.upcomingInput() + "\n" + c + "^";
|
||||
},
|
||||
|
||||
// test the lexed token: return FALSE when not a match, otherwise return token
|
||||
test_match:function (match, indexed_rule) {
|
||||
var token,
|
||||
lines,
|
||||
backup;
|
||||
|
||||
if (this.options.backtrack_lexer) {
|
||||
// save context
|
||||
backup = {
|
||||
yylineno: this.yylineno,
|
||||
yylloc: {
|
||||
first_line: this.yylloc.first_line,
|
||||
last_line: this.last_line,
|
||||
first_column: this.yylloc.first_column,
|
||||
last_column: this.yylloc.last_column
|
||||
},
|
||||
yytext: this.yytext,
|
||||
match: this.match,
|
||||
matches: this.matches,
|
||||
matched: this.matched,
|
||||
yyleng: this.yyleng,
|
||||
offset: this.offset,
|
||||
_more: this._more,
|
||||
_input: this._input,
|
||||
yy: this.yy,
|
||||
conditionStack: this.conditionStack.slice(0),
|
||||
done: this.done
|
||||
};
|
||||
if (this.options.ranges) {
|
||||
backup.yylloc.range = this.yylloc.range.slice(0);
|
||||
}
|
||||
}
|
||||
|
||||
lines = match[0].match(/(?:\r\n?|\n).*/g);
|
||||
if (lines) {
|
||||
this.yylineno += lines.length;
|
||||
}
|
||||
this.yylloc = {
|
||||
first_line: this.yylloc.last_line,
|
||||
last_line: this.yylineno + 1,
|
||||
first_column: this.yylloc.last_column,
|
||||
last_column: lines ?
|
||||
lines[lines.length - 1].length - lines[lines.length - 1].match(/\r?\n?/)[0].length :
|
||||
this.yylloc.last_column + match[0].length
|
||||
};
|
||||
this.yytext += match[0];
|
||||
this.match += match[0];
|
||||
this.matches = match;
|
||||
this.yyleng = this.yytext.length;
|
||||
if (this.options.ranges) {
|
||||
this.yylloc.range = [this.offset, this.offset += this.yyleng];
|
||||
}
|
||||
this._more = false;
|
||||
this._backtrack = false;
|
||||
this._input = this._input.slice(match[0].length);
|
||||
this.matched += match[0];
|
||||
token = this.performAction.call(this, this.yy, this, indexed_rule, this.conditionStack[this.conditionStack.length - 1]);
|
||||
if (this.done && this._input) {
|
||||
this.done = false;
|
||||
}
|
||||
if (token) {
|
||||
return token;
|
||||
} else if (this._backtrack) {
|
||||
// recover context
|
||||
for (var k in backup) {
|
||||
this[k] = backup[k];
|
||||
}
|
||||
return false; // rule action called reject() implying the next rule should be tested instead.
|
||||
}
|
||||
return false;
|
||||
},
|
||||
|
||||
// return next match in input
|
||||
next:function () {
|
||||
if (this.done) {
|
||||
return this.EOF;
|
||||
}
|
||||
if (!this._input) {
|
||||
this.done = true;
|
||||
}
|
||||
|
||||
var token,
|
||||
match,
|
||||
tempMatch,
|
||||
index;
|
||||
if (!this._more) {
|
||||
this.yytext = '';
|
||||
this.match = '';
|
||||
}
|
||||
var rules = this._currentRules();
|
||||
for (var i = 0; i < rules.length; i++) {
|
||||
tempMatch = this._input.match(this.rules[rules[i]]);
|
||||
if (tempMatch && (!match || tempMatch[0].length > match[0].length)) {
|
||||
match = tempMatch;
|
||||
index = i;
|
||||
if (this.options.backtrack_lexer) {
|
||||
token = this.test_match(tempMatch, rules[i]);
|
||||
if (token !== false) {
|
||||
return token;
|
||||
} else if (this._backtrack) {
|
||||
match = false;
|
||||
continue; // rule action called reject() implying a rule MISmatch.
|
||||
} else {
|
||||
// else: this is a lexer rule which consumes input without producing a token (e.g. whitespace)
|
||||
return false;
|
||||
}
|
||||
} else if (!this.options.flex) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
if (match) {
|
||||
token = this.test_match(match, rules[index]);
|
||||
if (token !== false) {
|
||||
return token;
|
||||
}
|
||||
// else: this is a lexer rule which consumes input without producing a token (e.g. whitespace)
|
||||
return false;
|
||||
}
|
||||
if (this._input === "") {
|
||||
return this.EOF;
|
||||
} else {
|
||||
return this.parseError('Lexical error on line ' + (this.yylineno + 1) + '. Unrecognized text.\n' + this.showPosition(), {
|
||||
text: "",
|
||||
token: null,
|
||||
line: this.yylineno
|
||||
});
|
||||
}
|
||||
},
|
||||
|
||||
// return next match that has a token
|
||||
lex:function lex() {
|
||||
var r = this.next();
|
||||
if (r) {
|
||||
return r;
|
||||
} else {
|
||||
return this.lex();
|
||||
}
|
||||
},
|
||||
|
||||
// activates a new lexer condition state (pushes the new lexer condition state onto the condition stack)
|
||||
begin:function begin(condition) {
|
||||
this.conditionStack.push(condition);
|
||||
},
|
||||
|
||||
// pop the previously active lexer condition state off the condition stack
|
||||
popState:function popState() {
|
||||
var n = this.conditionStack.length - 1;
|
||||
if (n > 0) {
|
||||
return this.conditionStack.pop();
|
||||
} else {
|
||||
return this.conditionStack[0];
|
||||
}
|
||||
},
|
||||
|
||||
// produce the lexer rule set which is active for the currently active lexer condition state
|
||||
_currentRules:function _currentRules() {
|
||||
if (this.conditionStack.length && this.conditionStack[this.conditionStack.length - 1]) {
|
||||
return this.conditions[this.conditionStack[this.conditionStack.length - 1]].rules;
|
||||
} else {
|
||||
return this.conditions["INITIAL"].rules;
|
||||
}
|
||||
},
|
||||
|
||||
// return the currently active lexer condition state; when an index argument is provided it produces the N-th previous condition state, if available
|
||||
topState:function topState(n) {
|
||||
n = this.conditionStack.length - 1 - Math.abs(n || 0);
|
||||
if (n >= 0) {
|
||||
return this.conditionStack[n];
|
||||
} else {
|
||||
return "INITIAL";
|
||||
}
|
||||
},
|
||||
|
||||
// alias for begin(condition)
|
||||
pushState:function pushState(condition) {
|
||||
this.begin(condition);
|
||||
},
|
||||
|
||||
// return the number of states currently on the stack
|
||||
stateStackSize:function stateStackSize() {
|
||||
return this.conditionStack.length;
|
||||
},
|
||||
options: {},
|
||||
performAction: function anonymous(yy,yy_,$avoiding_name_collisions,YY_START) {
|
||||
var YYSTATE=YY_START;
|
||||
switch($avoiding_name_collisions) {
|
||||
case 0:return 13;
|
||||
break;
|
||||
case 1:return 43;
|
||||
break;
|
||||
case 2:return 43;
|
||||
break;
|
||||
case 3:return 43;
|
||||
break;
|
||||
case 4:return 42;
|
||||
break;
|
||||
case 5:return 40;
|
||||
break;
|
||||
case 6:return 34;
|
||||
break;
|
||||
case 7:return 44;
|
||||
break;
|
||||
case 8:return 44;
|
||||
break;
|
||||
case 9:return 44;
|
||||
break;
|
||||
case 10:return 41;
|
||||
break;
|
||||
case 11:return 41;
|
||||
break;
|
||||
case 12:return 41;
|
||||
break;
|
||||
case 13:return 30;
|
||||
break;
|
||||
case 14:return 8;
|
||||
break;
|
||||
case 15:return 28;
|
||||
break;
|
||||
case 16:return 24;
|
||||
break;
|
||||
case 17:return 22;
|
||||
break;
|
||||
case 18:return 23;
|
||||
break;
|
||||
case 19:return 25;
|
||||
break;
|
||||
case 20:return 14;
|
||||
break;
|
||||
case 21:return 11;
|
||||
break;
|
||||
case 22:return 18;
|
||||
break;
|
||||
case 23:return 19;
|
||||
break;
|
||||
case 24:return 16;
|
||||
break;
|
||||
case 25:return 17;
|
||||
break;
|
||||
case 26:return 20
|
||||
break;
|
||||
case 27:return 21
|
||||
break;
|
||||
case 28:return 6;
|
||||
break;
|
||||
case 29:return 'NEWLINE';
|
||||
break;
|
||||
case 30:return 5;
|
||||
break;
|
||||
}
|
||||
},
|
||||
rules: [/^(?:style\b)/,/^(?:red\b)/,/^(?:blue\b)/,/^(?:black\b)/,/^(?:#[a-f0-9]+)/,/^(?:[0-9]+)/,/^(?:border\b)/,/^(?:dotted\b)/,/^(?:dashed\b)/,/^(?:solid\b)/,/^(?:px\b)/,/^(?:pt\b)/,/^(?:dot\b)/,/^(?::)/,/^(?:;)/,/^(?:,)/,/^(?:--[x])/,/^(?:-->)/,/^(?:--[o])/,/^(?:---)/,/^(?:[a-zA-Z]+)/,/^(?:\|)/,/^(?:\()/,/^(?:\))/,/^(?:\[)/,/^(?:\])/,/^(?:\{)/,/^(?:\})/,/^(?:\s)/,/^(?:\n)/,/^(?:$)/],
|
||||
conditions: {"INITIAL":{"rules":[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30],"inclusive":true}}
|
||||
});
|
||||
return lexer;
|
||||
})();
|
||||
parser.lexer = lexer;
|
||||
function Parser () {
|
||||
this.yy = {};
|
||||
}
|
||||
Parser.prototype = parser;parser.Parser = Parser;
|
||||
return new Parser;
|
||||
})();
|
||||
|
||||
|
||||
if (typeof require !== 'undefined' && typeof exports !== 'undefined') {
|
||||
exports.parser = parser;
|
||||
exports.Parser = parser.Parser;
|
||||
exports.parse = function () { return parser.parse.apply(parser, arguments); };
|
||||
exports.main = function commonjsMain(args) {
|
||||
if (!args[1]) {
|
||||
console.log('Usage: '+args[0]+' FILE');
|
||||
process.exit(1);
|
||||
}
|
||||
var source = require('fs').readFileSync(require('path').normalize(args[1]), "utf8");
|
||||
return exports.parser.parse(source);
|
||||
};
|
||||
if (typeof module !== 'undefined' && require.main === module) {
|
||||
exports.main(process.argv.slice(1));
|
||||
}
|
||||
}
|
@@ -25,21 +25,22 @@
|
||||
.edgePath path {
|
||||
stroke: #333;
|
||||
stroke-width: 1.5px;
|
||||
fill:#fff;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<div class="mermaid2">
|
||||
<div class="mermaid">
|
||||
A-->B;
|
||||
B-->D;
|
||||
A[AnnaStina]-->C|Fullgubbe;
|
||||
C-->D(Bjarke trummar pa bordet);
|
||||
B-->C{Obs};
|
||||
A[AnnaStina]-->C[Målar skåp];
|
||||
C-->D(Bjarke trummar på bordet);
|
||||
B-->C{Obs-gränby};
|
||||
</div>
|
||||
<div class="mermaid">
|
||||
a[Lasa bok]-->b;
|
||||
b{Fundera}-->c(Vidar)|Klocka;
|
||||
b-->d(Bjarke)|Lego;
|
||||
b-->d(Bjarke går)|Lego;
|
||||
style a fill:#6ff,stroke:#f66,stroke-width:2px,stroke-dasharray: 5, 5;
|
||||
</div>
|
||||
</body>
|
Reference in New Issue
Block a user