Move compiler checks and passes out of |PEG.Compiler| definition
This allows splitting them into separate files in the future. It also decreases indentation level in the code.
This commit is contained in:
parent
d7d1a0b28c
commit
2622f432bd
432
lib/compiler.js
432
lib/compiler.js
|
@ -240,218 +240,6 @@ PEG.Compiler = {
|
|||
this._uniqueIdentifierCounters = {};
|
||||
},
|
||||
|
||||
/*
|
||||
* Checks made on the grammar AST before compilation. Each check is a function
|
||||
* that is passed the AST and does not return anything. If the check passes,
|
||||
* the function does not do anything special, otherwise it throws
|
||||
* |PEG.GrammarError|. The checks are run in sequence in order of their
|
||||
* definition.
|
||||
*/
|
||||
_checks: [
|
||||
/* Checks that all referenced rules exist. */
|
||||
function(ast) {
|
||||
function nop() {}
|
||||
|
||||
function checkExpression(node) { check(node.expression); }
|
||||
|
||||
function checkSubnodes(propertyName) {
|
||||
return function(node) {
|
||||
PEG.ArrayUtils.each(node[propertyName], check);
|
||||
};
|
||||
}
|
||||
|
||||
var checkFunctions = {
|
||||
grammar:
|
||||
function(node) {
|
||||
for (var name in node.rules) {
|
||||
check(node.rules[name]);
|
||||
}
|
||||
},
|
||||
|
||||
rule: checkExpression,
|
||||
choice: checkSubnodes("alternatives"),
|
||||
sequence: checkSubnodes("elements"),
|
||||
labeled: checkExpression,
|
||||
simple_and: checkExpression,
|
||||
simple_not: checkExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: checkExpression,
|
||||
zero_or_more: checkExpression,
|
||||
one_or_more: checkExpression,
|
||||
action: checkExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node) {
|
||||
if (ast.rules[node.name] === undefined) {
|
||||
throw new PEG.GrammarError(
|
||||
"Referenced rule \"" + node.name + "\" does not exist."
|
||||
);
|
||||
}
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function check(node) { checkFunctions[node.type](node); }
|
||||
|
||||
check(ast);
|
||||
},
|
||||
|
||||
/* Checks that no left recursion is present. */
|
||||
function(ast) {
|
||||
function nop() {}
|
||||
|
||||
function checkExpression(node, appliedRules) {
|
||||
check(node.expression, appliedRules);
|
||||
}
|
||||
|
||||
var checkFunctions = {
|
||||
grammar:
|
||||
function(node, appliedRules) {
|
||||
for (var name in node.rules) {
|
||||
check(ast.rules[name], appliedRules);
|
||||
}
|
||||
},
|
||||
|
||||
rule:
|
||||
function(node, appliedRules) {
|
||||
check(node.expression, appliedRules.concat(node.name));
|
||||
},
|
||||
|
||||
choice:
|
||||
function(node, appliedRules) {
|
||||
PEG.ArrayUtils.each(node.alternatives, function(alternative) {
|
||||
check(alternative, appliedRules);
|
||||
});
|
||||
},
|
||||
|
||||
sequence:
|
||||
function(node, appliedRules) {
|
||||
if (node.elements.length > 0) {
|
||||
check(node.elements[0], appliedRules);
|
||||
}
|
||||
},
|
||||
|
||||
labeled: checkExpression,
|
||||
simple_and: checkExpression,
|
||||
simple_not: checkExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: checkExpression,
|
||||
zero_or_more: checkExpression,
|
||||
one_or_more: checkExpression,
|
||||
action: checkExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node, appliedRules) {
|
||||
if (PEG.ArrayUtils.contains(appliedRules, node.name)) {
|
||||
throw new PEG.GrammarError(
|
||||
"Left recursion detected for rule \"" + node.name + "\"."
|
||||
);
|
||||
}
|
||||
check(ast.rules[node.name], appliedRules);
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function check(node, appliedRules) {
|
||||
checkFunctions[node.type](node, appliedRules);
|
||||
}
|
||||
|
||||
check(ast, []);
|
||||
}
|
||||
],
|
||||
|
||||
/*
|
||||
* Optimalization passes made on the grammar AST before compilation. Each pass
|
||||
* is a function that is passed the AST and returns a new AST. The AST can be
|
||||
* modified in-place by the pass. The passes are run in sequence in order of
|
||||
* their definition.
|
||||
*/
|
||||
_passes: [
|
||||
/*
|
||||
* Removes proxy rules -- that is, rules that only delegate to other rule.
|
||||
*/
|
||||
function(ast) {
|
||||
function isProxyRule(node) {
|
||||
return node.type === "rule" && node.expression.type === "rule_ref";
|
||||
}
|
||||
|
||||
function replaceRuleRefs(ast, from, to) {
|
||||
function nop() {}
|
||||
|
||||
function replaceInExpression(node, from, to) {
|
||||
replace(node.expression, from, to);
|
||||
}
|
||||
|
||||
function replaceInSubnodes(propertyName) {
|
||||
return function(node, from, to) {
|
||||
PEG.ArrayUtils.each(node[propertyName], function(node) {
|
||||
replace(node, from, to);
|
||||
});
|
||||
};
|
||||
}
|
||||
|
||||
var replaceFunctions = {
|
||||
grammar:
|
||||
function(node, from, to) {
|
||||
for (var name in node.rules) {
|
||||
replace(ast.rules[name], from, to);
|
||||
}
|
||||
},
|
||||
|
||||
rule: replaceInExpression,
|
||||
choice: replaceInSubnodes("alternatives"),
|
||||
sequence: replaceInSubnodes("elements"),
|
||||
labeled: replaceInExpression,
|
||||
simple_and: replaceInExpression,
|
||||
simple_not: replaceInExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: replaceInExpression,
|
||||
zero_or_more: replaceInExpression,
|
||||
one_or_more: replaceInExpression,
|
||||
action: replaceInExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node, from, to) {
|
||||
if (node.name === from) {
|
||||
node.name = to;
|
||||
}
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function replace(node, from, to) {
|
||||
replaceFunctions[node.type](node, from, to);
|
||||
}
|
||||
|
||||
replace(ast, from, to);
|
||||
}
|
||||
|
||||
for (var name in ast.rules) {
|
||||
if (isProxyRule(ast.rules[name])) {
|
||||
replaceRuleRefs(ast, ast.rules[name].name, ast.rules[name].expression.name);
|
||||
if (name === ast.startRule) {
|
||||
ast.startRule = ast.rules[name].expression.name;
|
||||
}
|
||||
delete ast.rules[name];
|
||||
}
|
||||
}
|
||||
|
||||
return ast;
|
||||
}
|
||||
],
|
||||
|
||||
_compileFunctions: {
|
||||
grammar: function(node) {
|
||||
var initializerCode = node.initializer !== null
|
||||
|
@ -1090,12 +878,12 @@ PEG.Compiler = {
|
|||
* cause its malfunction.
|
||||
*/
|
||||
compileParser: function(ast) {
|
||||
for (var i = 0; i < this._checks.length; i++) {
|
||||
this._checks[i](ast);
|
||||
for (var i = 0; i < this.checks.length; i++) {
|
||||
this.checks[i](ast);
|
||||
}
|
||||
|
||||
for (var i = 0; i < this._passes.length; i++) {
|
||||
ast = this._passes[i](ast);
|
||||
for (var i = 0; i < this.passes.length; i++) {
|
||||
ast = this.passes[i](ast);
|
||||
}
|
||||
|
||||
var source = this.compileNode(ast);
|
||||
|
@ -1106,4 +894,216 @@ PEG.Compiler = {
|
|||
}
|
||||
};
|
||||
|
||||
/*
|
||||
* Checks made on the grammar AST before compilation. Each check is a function
|
||||
* that is passed the AST and does not return anything. If the check passes, the
|
||||
* function does not do anything special, otherwise it throws
|
||||
* |PEG.GrammarError|. The checks are run in sequence in order of their
|
||||
* definition.
|
||||
*/
|
||||
PEG.Compiler.checks = [
|
||||
/* Checks that all referenced rules exist. */
|
||||
function(ast) {
|
||||
function nop() {}
|
||||
|
||||
function checkExpression(node) { check(node.expression); }
|
||||
|
||||
function checkSubnodes(propertyName) {
|
||||
return function(node) {
|
||||
PEG.ArrayUtils.each(node[propertyName], check);
|
||||
};
|
||||
}
|
||||
|
||||
var checkFunctions = {
|
||||
grammar:
|
||||
function(node) {
|
||||
for (var name in node.rules) {
|
||||
check(node.rules[name]);
|
||||
}
|
||||
},
|
||||
|
||||
rule: checkExpression,
|
||||
choice: checkSubnodes("alternatives"),
|
||||
sequence: checkSubnodes("elements"),
|
||||
labeled: checkExpression,
|
||||
simple_and: checkExpression,
|
||||
simple_not: checkExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: checkExpression,
|
||||
zero_or_more: checkExpression,
|
||||
one_or_more: checkExpression,
|
||||
action: checkExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node) {
|
||||
if (ast.rules[node.name] === undefined) {
|
||||
throw new PEG.GrammarError(
|
||||
"Referenced rule \"" + node.name + "\" does not exist."
|
||||
);
|
||||
}
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function check(node) { checkFunctions[node.type](node); }
|
||||
|
||||
check(ast);
|
||||
},
|
||||
|
||||
/* Checks that no left recursion is present. */
|
||||
function(ast) {
|
||||
function nop() {}
|
||||
|
||||
function checkExpression(node, appliedRules) {
|
||||
check(node.expression, appliedRules);
|
||||
}
|
||||
|
||||
var checkFunctions = {
|
||||
grammar:
|
||||
function(node, appliedRules) {
|
||||
for (var name in node.rules) {
|
||||
check(ast.rules[name], appliedRules);
|
||||
}
|
||||
},
|
||||
|
||||
rule:
|
||||
function(node, appliedRules) {
|
||||
check(node.expression, appliedRules.concat(node.name));
|
||||
},
|
||||
|
||||
choice:
|
||||
function(node, appliedRules) {
|
||||
PEG.ArrayUtils.each(node.alternatives, function(alternative) {
|
||||
check(alternative, appliedRules);
|
||||
});
|
||||
},
|
||||
|
||||
sequence:
|
||||
function(node, appliedRules) {
|
||||
if (node.elements.length > 0) {
|
||||
check(node.elements[0], appliedRules);
|
||||
}
|
||||
},
|
||||
|
||||
labeled: checkExpression,
|
||||
simple_and: checkExpression,
|
||||
simple_not: checkExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: checkExpression,
|
||||
zero_or_more: checkExpression,
|
||||
one_or_more: checkExpression,
|
||||
action: checkExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node, appliedRules) {
|
||||
if (PEG.ArrayUtils.contains(appliedRules, node.name)) {
|
||||
throw new PEG.GrammarError(
|
||||
"Left recursion detected for rule \"" + node.name + "\"."
|
||||
);
|
||||
}
|
||||
check(ast.rules[node.name], appliedRules);
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function check(node, appliedRules) {
|
||||
checkFunctions[node.type](node, appliedRules);
|
||||
}
|
||||
|
||||
check(ast, []);
|
||||
}
|
||||
];
|
||||
|
||||
/*
|
||||
* Optimalization passes made on the grammar AST before compilation. Each pass
|
||||
* is a function that is passed the AST and returns a new AST. The AST can be
|
||||
* modified in-place by the pass. The passes are run in sequence in order of
|
||||
* their definition.
|
||||
*/
|
||||
PEG.Compiler.passes = [
|
||||
/*
|
||||
* Removes proxy rules -- that is, rules that only delegate to other rule.
|
||||
*/
|
||||
function(ast) {
|
||||
function isProxyRule(node) {
|
||||
return node.type === "rule" && node.expression.type === "rule_ref";
|
||||
}
|
||||
|
||||
function replaceRuleRefs(ast, from, to) {
|
||||
function nop() {}
|
||||
|
||||
function replaceInExpression(node, from, to) {
|
||||
replace(node.expression, from, to);
|
||||
}
|
||||
|
||||
function replaceInSubnodes(propertyName) {
|
||||
return function(node, from, to) {
|
||||
PEG.ArrayUtils.each(node[propertyName], function(node) {
|
||||
replace(node, from, to);
|
||||
});
|
||||
};
|
||||
}
|
||||
|
||||
var replaceFunctions = {
|
||||
grammar:
|
||||
function(node, from, to) {
|
||||
for (var name in node.rules) {
|
||||
replace(ast.rules[name], from, to);
|
||||
}
|
||||
},
|
||||
|
||||
rule: replaceInExpression,
|
||||
choice: replaceInSubnodes("alternatives"),
|
||||
sequence: replaceInSubnodes("elements"),
|
||||
labeled: replaceInExpression,
|
||||
simple_and: replaceInExpression,
|
||||
simple_not: replaceInExpression,
|
||||
semantic_and: nop,
|
||||
semantic_not: nop,
|
||||
optional: replaceInExpression,
|
||||
zero_or_more: replaceInExpression,
|
||||
one_or_more: replaceInExpression,
|
||||
action: replaceInExpression,
|
||||
|
||||
rule_ref:
|
||||
function(node, from, to) {
|
||||
if (node.name === from) {
|
||||
node.name = to;
|
||||
}
|
||||
},
|
||||
|
||||
literal: nop,
|
||||
any: nop,
|
||||
"class": nop
|
||||
};
|
||||
|
||||
function replace(node, from, to) {
|
||||
replaceFunctions[node.type](node, from, to);
|
||||
}
|
||||
|
||||
replace(ast, from, to);
|
||||
}
|
||||
|
||||
for (var name in ast.rules) {
|
||||
if (isProxyRule(ast.rules[name])) {
|
||||
replaceRuleRefs(ast, ast.rules[name].name, ast.rules[name].expression.name);
|
||||
if (name === ast.startRule) {
|
||||
ast.startRule = ast.rules[name].expression.name;
|
||||
}
|
||||
delete ast.rules[name];
|
||||
}
|
||||
}
|
||||
|
||||
return ast;
|
||||
}
|
||||
];
|
||||
|
||||
})();
|
||||
|
|
Loading…
Reference in a new issue