Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/DmitrySoshnikov/syntax
Syntactic analysis toolkit, language-agnostic parser generator.
https://github.com/DmitrySoshnikov/syntax
Last synced: 7 days ago
JSON representation
Syntactic analysis toolkit, language-agnostic parser generator.
- Host: GitHub
- URL: https://github.com/DmitrySoshnikov/syntax
- Owner: DmitrySoshnikov
- License: mit
- Created: 2015-11-27T05:30:04.000Z (almost 9 years ago)
- Default Branch: master
- Last Pushed: 2024-01-06T11:55:32.000Z (10 months ago)
- Last Synced: 2024-04-14T05:59:34.289Z (7 months ago)
- Language: JavaScript
- Homepage:
- Size: 1.3 MB
- Stars: 588
- Watchers: 15
- Forks: 83
- Open Issues: 48
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# syntax
[![Build Status](https://travis-ci.org/DmitrySoshnikov/syntax.svg?branch=master)](https://travis-ci.org/DmitrySoshnikov/syntax) [![npm version](https://badge.fury.io/js/syntax-cli.svg)](https://badge.fury.io/js/syntax-cli)
Syntactic analysis toolkit, language-agnostic parser generator.
Implements [LR](https://en.wikipedia.org/wiki/LR_parser) and [LL](https://en.wikipedia.org/wiki/LL_parser) parsing algorithms.
You can get an introductory overview of the tool in [this article](https://medium.com/@DmitrySoshnikov/syntax-language-agnostic-parser-generator-bd24468d7cfc).
### Table of Contents
- [Installation](#installation)
- [Development](#development)
- [CLI usage example](#cli-usage-example)
- [Parser generation](#parser-generation)
- [Language agnostic parser generator](#language-agnostic-parser-generator)
- [JavaScript default](#javascript-default)
- [Python plugin](#python-plugin)
- [PHP plugin](#php-plugin)
- [Ruby plugin](#ruby-plugin)
- [C++ plugin](#c-plugin)
- [C# plugin](#c-plugin-1)
- [Rust plugin](#rust-plugin)
- [Java plugin](#java-plugin)
- [Julia plugin](#julia-plugin)
- [Grammar format](#grammar-format)
- [JSON-like notation](#json-like-notation)
- [Yacc/Bison notation](#yaccbison-notation)
- [Grammar properties](#grammar-properties)
- [Lexical grammar and tokenizer](#lexical-grammar-and-tokenizer)
- [Getting list of tokens](#getting-list-of-tokens)
- [Using custom tokenizer](#using-custom-tokenizer)
- [Start conditions of lex rules, and tokenizer states](#start-conditions-of-lex-rules-and-tokenizer-states)
- [Access tokenizer from parser semantic actions](#access-tokenizer-from-parser-semantic-actions)
- [Case-insensitive match](#case-insensitive-match)
- [Working with precedence and associativity](#working-with-precedence-and-associativity)
- [Associative precedence](#associative-precedence)
- [Non-associative precedence](#non-associative-precedence)
- [Handler arguments notation](#handler-arguments-notation)
- [Positioned notation](#positioned-notation)
- [Named notation](#named-notation)
- [Capturing location objects](#capturing-location-objects)
- [Parsing modes](#parsing-modes)
- [LL parsing](#ll-parsing)
- [LR parsing](#lr-parsing)
- [LR conflicts](#lr-conflicts)
- [Conflicts resolution](#conflicts-resolution)
- [Validating grammar](#validating-grammar)
- [Module include, and parser events](#module-include-and-parser-events)
- [Debug mode](#debug-mode)### Installation
The tool can be installed as an [npm module](https://www.npmjs.com/package/syntax-cli) (notice, it's called `syntax-cli` there):
```
npm install -g syntax-clisyntax-cli --help
```### Development
1. Fork the https://github.com/DmitrySoshnikov/syntax repo
2. Make your changes
3. Make sure `npm test` passes (add new tests if needed)
4. Submit a PR> NOTE: If you need to implement a Syntax plugin for a new target programming language, address [this instruction](https://github.com/DmitrySoshnikov/syntax/blob/master/src/plugins/README.md).
For development from the github repository, run `build` command to transpile ES6 code:
```
git clone https://github.com//syntax.git
cd syntax
npm install
npm run build./bin/syntax --help
```Or for faster development cycle, one can also use `watch` command (notice though, it doesn't copy template files, but only transpiles ES6 code; for templates copying you have to use `build` command):
```
npm run watch
```### CLI usage example
```
./bin/syntax --grammar examples/grammar.lr0 --parse "aabb" --mode lr0 --table --collection
```### Parser generation
To generate a parser module, specify the `--output` option, which is a path to the output parser file. Once generated, the module can normally be required, and used for parsing strings based on a given grammar.
Example for the [JSON grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/json.ast.js):
```
./bin/syntax --grammar examples/json.ast.js --mode SLR1 --output json-parser.js✓ Successfully generated: json-parser.js
```Loading as a JS module:
```js
const JSONParser = require('./json-parser');let value = JSONParser.parse('{"x": 10, "y": [1, 2]}');
console.log(value); // JS object: {x: 10, y: [1, 2]}
```### Language agnostic parser generator
See [this instruction](https://github.com/DmitrySoshnikov/syntax/blob/master/src/plugins/README.md) how to implement a new plugin.
#### JavaScript default
Syntax is language agnostic when it comes to parser generation. The same grammar can be used for parser generation in different languages. Currently Syntax supports _JavaScript_, _Python_, _PHP_, _Ruby_, _C#_, _Rust_, and _Java_. The target language is determined by the output file extension.
#### Python plugin
For example, this is how to use the same [calculator grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc.py.g) example to generate parser module in Python:
```
./bin/syntax -g examples/calc.py.g -m lalr1 -o calcparser.py
```The `calcparser` module then can be required normally in Python for parsing:
```python
>>> import calcparser
>>> calcparser.parse('2 + 2 * 2')
>>> 6
```[Another example](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/module-include.py.g) shows how to use parser hooks (such as `on_parse_begin`, `on_parse_end`, and other) in Python. They are discussed below in the [module include](https://github.com/DmitrySoshnikov/syntax#module-include-and-parser-events) section.
#### PHP plugin
For PHP the procedure is pretty much the same, take a look at the similar [example](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc.php.g):
```
./bin/syntax -g examples/calc.php.g -m lalr1 -o CalcParser.php
```The output file contains the class name corresponding to the file name:
```php
NOTE: built-in tokenizer uses underlying regexp implementation to extract stream of tokens.It is possible to provide a custom tokenizer if a built-in isn't sufficient. For this pass the `--custom-tokenizer` option, which is a path to a file that implements a tokenizer. In this case the built-in tokenizer code won't be generated.
```
./bin/syntax --grammar examples/json.ast.js --mode SLR1 --output json-parser.js --custom-tokenizer './my-tokenizer.js'✓ Successfully generated: json-parser.js
```In the generated code, the custom tokenizer is just required as a module: `require('./my-tokenizer.js')`.
The tokenizer should implement the following iterator-like interface:
- `initString`: initializes a parsing string;
- `hasMoreTokens`: whether stream of tokens still has more tokens to consume;
- `getNextToken`: returns next token in the format `{type, value}`;For example:
```js
// File: ./my-tokenizer.jsconst MyTokenizer = {
initString(string) {
this._string = string;
this._cursor = 0;
},hasMoreTokens() {
return this._cursor < this._string.length;
},getNextToken() {
// Implement logic here.return {
// Basic data.
type: <>,
value: <>,// Location data.
startOffset: <>,
endOffset: <>,
startLine: <>,
endLine: <>,
startColumn: <>,
endColumn: <>,
}
},
};module.exports = MyTokenizer;
```#### Start conditions of lex rules, and tokenizer states
Built-in tokenizer supports _stateful tokenization_. This means the same lex rule can be applied in different states, and result to a different token. For lex rules it's known as _start conditions_.
Rules with explicit start conditions are executed _only_ when lexer enters the state corresponding to their names. Start conditions can be _inclusive_ (`%s`, 0), and _exclusive_ (`%x`, 1). Inclusive conditions also include rules _without_ any start conditions, and exclusive conditions do not include other rules when the parser enter their state. The rules with `*` start condition are always included.
```js
"lex": {
"startConditions": {
"comment": 1, // exclusive
"string": 1 // exclusive
},"rules": [
...// On `/*` enter `comment` state:
["\\/\\*", "this.pushState('comment');"],// The rule is executed only when tokenizers enters `comment`, or `string` state:
[["comment", "string"], "\\n", "lines++; return 'NL';"],...
],
}
```More information on the topic can be found in [this gist](https://gist.github.com/DmitrySoshnikov/f5e2583b37e8f758c789cea9dcdf238a).
As an example take a look at [this example grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/lexer-start-conditions.g.js), which calculates line numbers in a source file, including line numbers in comments. The comments themselves are skipped during tokenization, however the new lines are handled within comments separately to count those line numbers as well.
Another example is the [grammar for BNF](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/bnf.g) itself, which we use to parse BNF grammars represented as strings, rather than in JSON format. There we have `action` start condition to correctly parse `{` and `}` of JS code, being inside an actual handler for a grammar rule, which is itself surrounded by `{` and `}` braces.
#### Access tokenizer from parser semantic actions
It is also possible to access tokenizer instance from the parser semantic actions. It is exposed via the `yy.lexer` object (`yy.tokenizer` is an alias).
Having access to the lexer, it is possible, for example, to change its state, and yield different token types for the same characters.
As an example, differently parsing `{` and `}` being in an _expression_ or in a _statement_ position in ECMAScript language:
```
{x: 1} // BlockStatement
({x: 1}) // ObjectLiteral
```A simplified example for this can be found in the [parser-lexer-communication.g](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/parser-lexer-communication.g) grammar example.
#### Case-insensitive match
Lexical grammar rules can also be _case-insensitive_. From the command line it's control via the appropriate `--case-insensitive` (`-i`) option. It also can be specified in the lexical grammar itself -- for the whole grammar, or per each rule:
```js
// case-insensitive.lex{
"rules": [
// This rule is by default case-insensitive:
[`x`, `return "X"`],// This rule overrides global options:
[`y`, `return "Y"`, {"case-insensitive": false}],
],// Global options for the whole lexical grammar.
"options": {
"case-insensitive": true,
},
}
```See [this example](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/case-insensitive-lex.g) for details.
### Working with precedence and associativity
Precedence and associativity operators allow building more readable and elegant grammars, avoiding different kinds of conflicts, like "shift-reduce" conflicts.
Supported precedence operators are:
* `%left` -- left-associative;
* `%right` -- right-associative;
* `%nonassoc` -- non-associative.#### Associative precedence
Having the following snippet from the calculator grammar:
```
%%e
: e '+' e
| e '*' e
| '(' e ')'
| NUMBER
;
```we get a _"shift-reduce"_ conflict for the input like:
```
2 + 2 + 2
```Once have parsed the first `2 + 2`, and having `+` as the lookahead, the parser cannot decide, whether it should _reduce_ the parsed `2 + 2` value to `e`, _or_ it should to _shift_ further, since `2` itself can be `e` (by `NUMBER` rule), and parser can expect `+` after `e`.
Defining associativity for the `+` operator solves it easily and elegantly:
```
%left '+'
%left '*'%%
e
: e '+' e
| e '*' e
| '(' e ')'
| NUMBER
;
```Here by `%left '+'` we say that `+` is _left-associative_, which means that parser should parse `2 + 2 + 2` as `(2 + 2) + 2`, and not as `2 + (2 + 2)`. In other words, once the parser have parsed `2 + 2`, and still sees `+` as the lookahead, it chooses to _reduce_ instead of shift, and "shift-reduce" conflict is resolved.
Another example is having the snippet as:
```
2 + 2 * 2
```If parser would reduce in this case, we would get an _invalid mathematical expression_, since this expression, without any grouping parenthesis, should be parsed as `2 + (2 * 2)`, and not as `(2 + 2) * 2`.
To solve this we use `%left '*'` which in our grammar definition stays in order _after_ the `%left '+'`, which makes the `*` operator to have a _higher precedence_, than the `+` has. In this case parser chooses to _shift_ further instead of reducing.
Note, from [JSON-like notation](#json-like-notation) they are defined as:
```js
operators: [
['left', '+'],
['left', '*'],
// etc.
]
```#### Non-associative precedence
Sometimes we don't need any associativity, but just want to specify _precedence_ of some symbols. As a classic example, we can take the [dangling-else](https://en.wikipedia.org/wiki/Dangling_else) problem, for which we use `%nonassoc` operator:
```
%nonassoc THEN
%nonassoc 'else'%%
IfStatement
: 'if' '(' Expression ')' Statement %prec THEN
| 'if' '(' Expression ')' Statement 'else' Statement
;
```As we can see, `'else'` token has _higher precedence_ again, since goes after ("virtual") `THEN` token, so there is no "shift-reduce" conflict in this case.
Here `%prec` is used in production to specify which precedence to apply, using the "virtual" `THEN` symbol -- in this case it's not a real token (in contrast with `'else'`), but just _precedence name_ in order to refer it from the production.
You can find this problem handled in this [grammar example](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/lang.bnf).
### Handler arguments notation
The following notation is used for semantic action (handler) arguments:
- `yytext` -- a matched token value
- `yyleng` -- the length of the matched token
- Positioned arguments, e.g. `$1`, `$2`, etc.
- Positioned locations, e.g. `@1`, `@2`, etc.
- Named arguments, e.g. `$foo`, `$bar`, etc.
- Named locations, e.g. `@foo`, `@bar`, etc.
- `$$` -- result value
- `@$` -- result location#### Positioned notation
This is the simplest notation -- the semantic action arguments can be accessed via their number. For example, for the production:
```
exp : exp '+' term { $$ = $1 + $3 }
```The `exp` can be accessed as `$1`, the `$2` would contain `'+'`, and `$3` corresponds to the `term`.
#### Named notation
Sometimes using positioned arguments can be less readable, and may cause refactoring issues. E.g. if some symbol is removed from the production, the handler code should be updated:
```
exp : '+' exp term { $$ = $2 + $3 }
```In this case using _named arguments_ might be more suitable:
```
exp : exp '+' term { $$ = $exp + $term }
```Still the same, even if the production is changed:
```
exp : '+' exp term { $$ = $exp + $term }
```Notice though, that for _duplicated symbols_ named notation doesn't work, since causes ambiguity:
```
exp : exp '+' exp { $$ = $exp + $exp } /* ERROR! */
```In this case the positioned arguments should be used:
```
exp : exp '+' exp { $$ = $1 + $3 } /* OK! */
```### Capturing location objects
For some tools (e.g source-code transformation tools) it is important not only to produce AST nodes, but also to capture all the locations in the original source code. _Syntax_ supports `--loc` option for this. A default structure of a location object is the same as for a token:
```js
{
startOffset,
endOffset,
startLine,
endLine,
startColumn,
endColumn,
}
```However in actual AST nodes generation it is possible to build a custom location information based on this default location object.
The locations are accessed using `@1`, or `@foo` notation, the result location is in the `@$`:
```
exp : exp + exp
{
$$ = $1 + $2;// Default algorithm.
@$.startLine = @1.startLine;
@$.endLine = @3.endLine;
@$.startColumn = @1.startColumn;
@$.endColumn = @3.endColumn;
...
}
```By default _Syntax_ automatically calculates resulting location taking _start part_ from the _first symbol_ of a production, and the _end part_ -- from the _last symbol_ of the production. So the example above can actually omit manual result location calculation, and be just:
```
exp : exp + exp { $$ = $1 + $2; }
```It is possible to override though the default algorithm by just mutating the `@$`, and it's also possible to create a custom location:
```
exp : exp + exp { $$ = new AdditionNode($1, $3, Loc(@$)) }
```In this case function `Loc` can create custom location format. Here is [another example](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc-loc.bnf) of a grammar which uses location objects.
### Parsing modes
_Syntax_ supports several _LR_ parsing modes: _LR(0)_, _SLR(1)_, _LALR(1)_, _CLR(1)_ as well _LL(1)_ mode. The same grammar can be analyzed in different modes, from the CLI it's controlled via the `--mode` option, e.g. `--mode slr1`.
> Note: de facto standard for automatically generated parsers is usually the _LALR(1)_ parser. The _CLR(1)_ parser, being the most powerful, and able to parse wider grammar sets, can have much more states than LALR(1), and usually is suitable only for educational purposes. As well as its less powerful counterparts, _LR(0)_ and _SLR(1)_ which are less used on practice (although, some production-ready grammars can also normally be parsed by _SLR(1)_, e.g. [JSON grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/json.ast.js)).
Some grammars can be handled by one mode, but not by another. In this case a _conflict_ will be shown in the table.
#### LL parsing
Currently an LL(1) grammar is supposed to be already _left-factored_, and to be _non-left-recursive_. See section on [LL conflicts](https://en.wikipedia.org/wiki/LL_parser#Solutions_to_LL.281.29_Conflicts) for details.
> Note: left-recursion elimination, and left-factoring process can be automated for most of the cases (excluding some edge cases, which should be done manually), and implement a transformation to a non-left-recursive grammar.
A typical LL parsing table is less, than a corresponding LR-table. However, LR grammars cover more languages than LL grammars. In addition, an LL(1) grammar usually might look less elegant, or even less readable, than an LR grammar. As an example, take a look at the calculator grammar in the [non-left-recursive LL mode](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc.ll1), [left-recursive LR mode](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calculator.g), and also [left-recursive, and precedence-based LR-mode](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc.slr1).
At the moment, LL parser only implements syntax validation, not providing semantic actions (e.g. to construct an AST). For the semantic handlers, and actual AST construction see LR parsing.
#### LR parsing
LR parsing, and its the most practical version, the LALR(1), is widely used in automatically generated parsers. LR grammars usually look more readable, than corresponding LL grammars, since in contrast with the later, LR parser generators by default allow _left-recursion_, and do automatic conflict resolutions. The precedence and assoc operators allow building more elegant grammars with smaller parsing tables.
Take a look at the [example grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calc-eval.g) with a typical _syntax-directed translation (SDT)_, using semantic actions for AST construction, direct evaluation, and any other transformation.
Default algorithm used for the practical _LALR(1)_ mode, is the _"LALR(1) by SLR(1)"_. It is enabled by the default `--mode lalr1`, or explicit `--mode lalr1_by_slr1`. In addition, Syntax implements _"LALR(1) by compressing CLR(1)"_ algorithm (available via the `--mode lalr1_by_clr1`), which is slower in parser generation, and is suitable only for educational purposes.
> **NOTE:** prefer usage of the `--mode lalr1` as _the most practical_.
#### LR conflicts
In LR parsing there are two main types of conflicts: _"shift-reduce" (s/r)_ conflict, and _"reduce-reduce" (r/r)_ conflict. Taking as an example grammar from `examples/example1.slr1`, we see that the parsing table is normally constructed for `SLR(1)` mode, but has a "shift-reduce" conflict if ran in the `LR(0)` mode:
```
./bin/syntax --grammar examples/example1.slr1 --table
``````
./bin/syntax --grammar examples/example1.slr1 --table --mode lr0
```![sl1-grammar](http://dmitrysoshnikov.com/wp-content/uploads/2015/12/imageedit_2_9168334335.png)
![sl1-grammar-lr0-m](http://dmitrysoshnikov.com/wp-content/uploads/2015/12/imageedit_2_6530197571.png)#### Conflicts resolution
Sometimes changing parsing mode is not enough for fixing conflicts: for some grammars conflicts may stay and in the _LALR(1)_, and even the _CLR(1)_ modes. LR conflicts can be resolved though automatically and semi-automatically by specifying precedence and associativity of operators.
For example, the [following grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calculator-assoc-conflict.g) has a _shift-reduce_ conflict:
```
%token id%%
E
: E '+' E
| E '*' E
| id
;
```Therefore, a parsing is not possible using this grammar:
```
./bin/syntax -g examples/calculator-assoc-conflict.g -m lalr1 -w -p 'id * id + id'Parsing mode: LALR(1).
Parsing: id * id + id
Rejected: Parse error: Found "shift-reduce" conflict at state 6, terminal '+'.
```This can be fixed though using operators associativity and precedence:
```
%token id%left '+'
%left '*'%%
E
: E '+' E
| E '*' E
| id
;
```See detailed description of the conflicts resolution algorithm in [this example grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calculator-assoc.g), which is can be parsed normally:
```
./bin/syntax -g examples/calculator-assoc.g -m lalr1 -w -p 'id * id + id'Parsing mode: LALR(1).
Parsing: id * id + id
✓ Accepted
```### Validating grammar
By using `--validate` option, it is possible to check whether your grammar is free from different kinds of conflicts, and if it is not, to get needed information about which grammar rules conflict, and wich possible solutions can be applied to resolve them.
For example, discussed above [calculator-assoc-conflict](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calculator-assoc-conflict.g) grammar has _"shift-reduce"_ conflicts in two productions:
```
./bin/syntax -g examples/calculator-assoc-conflict.g -m slr1 --validateParsing mode: SLR(1).
Grammar has the following unresolved conflicts:
"Shift-reduce" conflicts:
1. Production: E -> E '+' E, conflicts with symbols '+', '*'.
2. Production: E -> E '*' E, conflicts with symbols '+', '*'.Possible solutions:
1. Conflicts possibly can be resolved by using "operators" section,
where you can specify precedence and associativity.2. By using different parsing mode, e.g. LALR1 instead of SLR1.
3. Restructuring grammar.
```As we can see, _Syntax_ showed which rules conflict with which lookahead symbols, and provided possible solutions.
In this case, rule `E -> E '+' E` conflicts with lookahead `'+'`, and `'*'`. This means, that if have `id + id + id`, which would expand to `E '+' E '+' E`, the parser wouldn't know whether to _reduce_ first `E '+' E` to `E`, or whether to _shift_ further when it sees the second `'+'` symbol.
By specifying operators precedence, and associativity, we can resolve this conflict, which is done in the [calculator-assoc](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/calculator-assoc.g) grammar:
```
./bin/syntax -g examples/calculator-assoc.g -m slr1 --validateParsing mode: SLR(1).
✓ Grammar doesn't have any conflicts!
```### Module include, and parser events
The `moduleInclude` directive allows injecting an arbitrary code to the generated parser file. This is usually code to require needed dependencies, or to define them inline. As an example, see [the corresponding example grammar](https://github.com/DmitrySoshnikov/syntax/blob/master/examples/module-include.g.js), which defines all classes for AST nodes inline, and then uses them in the rule handlers.
The code can also define handlers for some parse events (attaching them to `yyparse` object), in particular for `onParseBegin` and `onParseEnd`. This allow injecting a code which is executed when the parsing process starts, and ends accordingly.
```js
"moduleInclude": `
class Node {
constructor(type) {
this.type = type;
}
}class BinaryExpression extends Node {
...
}// Event handlers.
yyparse.onParseBegin = (string) => {
console.log('Parsing code:', string);
};yyparse.onParseEnd = (value) => {
console.log('Parsed value:', value);
};
`,...
["E + E", "$$ = new BinaryExpression($1, $3, $2)"],
```### Debug mode
Debug mode allows measuring timing of certain steps, and analyzing other debug information. From the CLI it's activated using `--debug` (`-d`) option:
```
./bin/syntax -g examples/calc-eval.g -m slr1 -p '2 + 2 * 2' --debugDEBUG mode is: ON
[DEBUG] Grammar (bnf) is in JS format
[DEBUG] Grammar loaded in: 2.255msParsing mode: SLR(1).
Parsing:
2 + 2 * 2
[DEBUG] Building canonical collection: 15.219ms
[DEBUG] Number of states in the collection: 22
[DEBUG] Building LR parsing table: 4.169ms
[DEBUG] LR parsing: 2.180ms✓ Accepted
Parsed value:
6
[DEBUG] Total time: 70.284ms
```