D-Scanner/std/d/parser.d

1037 lines
26 KiB
D
Executable File

// Written in the D programming language
/**
* This module contains a parser for D source code.
*/
module std.d.parser;
import std.d.lexer;
import std.d.ast;
version(unittest) import std.stdio;
Module parseModule(R)(R tokens) if (is (ElementType!R == Token))
{
auto parser = new Parser();
parser.tokens = tokens.array();
return parser.parseModule();
}
private:
struct Parser
{
Module parseModule()
{
Module m = new Module;
while (index < tokens.length)
{
switch (tokens[index].type)
{
case TokenType.module_:
if (m.declaration !is null)
m.declaration = parseModuleDeclaration();
else
error("Only one module declaration is allowed per module");
break;
default:
m.declDefs.insert(parseDeclDef());
}
}
return m;
}
ModuleDeclaration parseModuleDeclaration()
in
{
assert (expect(TokenType.module_));
}
body
{
ModuleDeclaration declaration = new ModuleDeclaration;
string recent;
loop: while (index < tokens.length)
{
if (tokens[index].type == TokenType.identifier)
{
recent = tokens[index++].value;
switch (tokens[index].type)
{
case TokenType.dot:
declaration.packageName ~= recent;
index++;
break;
case TokenType.semicolon:
declaration.moduleName = recent;
index++;
break loop;
default:
break;
}
}
else
error("Identifier expected");
}
return declaration;
}
DeclDef parseDeclDef()
{
switch (tokens[index].type)
{
default:
return null;
}
}
LabeledStatement parseLabeledStatement()
in
{
assert (tokens[index].type == TokenType.identifier);
}
body
{
auto ls = new LabeledStatement;
ls.label = tokens[index++].value;
ls.statement = parseNoScopeStatement();
return ls;
}
NoScopeStatement parseNoScopeStatement()
{
switch (tokens[index].type)
{
case TokenType.semicolon:
return new EmptyStatement;
case TokenType.lBrace:
return parseBlockStatement();
default:
return parseNonEmptyStatement();
}
}
NonEmptyStatement parseNonEmptyStatement()
{
switch (tokens[index].type)
{
case TokenType.case_:
return parseCaseStatement();
case TokenType.default_:
return parseDefaultStatement();
default:
return null;
}
}
CaseStatement parseCaseStatement()
{
return null;
}
DefaultStatement parseDefaultStatement()
{
return null;
}
Statement parseStatement()
{
return null;
}
BlockStatement parseBlockStatement()
{
auto statement = new BlockStatement();
expect(TokenType.lBrace);
switch (tokens[index].type)
{
case TokenType.rBrace:
break;
default:
statement.statements ~= parseStatement();
}
return statement;
}
GotoStatement parseGotoStatement()
{
expect(TokenType.goto_);
auto g = new GotoStatement;
switch (tokens[index].type)
{
case TokenType.identifier:
g.type = GotoStatement.GotoType.identifier;
g.identifier = tokens[index++].value;
break;
case TokenType.default_:
index++;
g.type = GotoStatement.GotoType.default_;
break;
case TokenType.case_:
g.type = GotoStatement.GotoType.case_;
index++;
break;
default:
error("Expected an identifier, \"default\", or \"case\" following \"goto\"");
return null;
}
return g;
}
ContinueStatement parseContinueStatement()
{
expect(TokenType.continue_);
return parseContinueBreakStatement!(ContinueStatement)();
}
BreakStatement parseBreakStatement()
{
expect(TokenType.break_);
return parseContinueBreakStatement!(BreakStatement)();
}
statementType parseContinueBreakStatement(alias statementType)()
{
index++;
auto c = new statementType;
switch (tokens[index].type)
{
case TokenType.identifier:
c.identifier = tokens[index++].value;
goto case;
case TokenType.semicolon:
return c;
default:
error("Identifier or semicolon expected");
return null;
}
}
ImportDeclaration parseImportDeclaration(Tokens)(ref Tokens tokens)
in
{
assert(tokens[i] == TokenType.import_);
}
body
{
auto declaration = new ImportDeclaration;
tokens.popFront();
Import im;
if (tokens[i].type != TokenType.identifier)
{
tokens.skipPastSemicolon();
return declaration;
}
void completeImport()
{
im.moduleName = tokens.moveFront().value;
tokens.popFront();
declaration.imports ~= im;
}
void parseImportBindings()
{
loop: while (!tokens.empty)
{
if (tokens[i].type != TokenType.identifier)
break;
switch (tokens.peek().type)
{
case TokenType.assign:
Import.ImportSymbol s;
s.alias_ = tokens.moveFront().value;
tokens.popFront();
if (tokens.empty || tokens[i].type != TokenType.identifier)
break loop;
s.symbolName = tokens.moveFront().value;
im.symbols ~= s;
if (!tokens.empty())
{
if (tokens[i].type == TokenType.comma)
tokens.popFront();
if (tokens[i].type == TokenType.semicolon)
{
tokens.popFront();
declaration.imports ~= im;
break loop;
}
}
break;
case TokenType.comma:
Import.ImportSymbol s;
s.symbolName = tokens.moveFront().value;
tokens.popFront();
im.symbols ~= s;
break;
case TokenType.semicolon:
Import.ImportSymbol s;
s.symbolName = tokens.moveFront().value;
tokens.popFront();
im.symbols ~= s;
declaration.imports ~= im;
break loop;
default:
break loop;
}
}
}
loop: while (!tokens.empty)
{
switch (tokens.peek().type)
{
case TokenType.dot:
im.packageParts ~= tokens.moveFront().value;
tokens.popFront();
break;
case TokenType.comma:
completeImport();
im = Import.init;
break;
case TokenType.semicolon:
completeImport();
break loop;
case TokenType.colon:
im.moduleName = tokens.moveFront().value;
tokens.popFront();
parseImportBindings();
break loop;
case TokenType.assign:
im.alias_ = tokens.moveFront().value;
tokens.popFront();
break;
default:
tokens.popFront();
break;
}
}
return declaration;
}
void error(string message)
{
import std.stdio;
stderr.writefln("%s(%d:%d): %s", fileName, tokens[index].line,
tokens[index].column, message);
while (index < tokens.length)
{
if (tokens[++index].type == TokenType.semicolon)
break;
}
}
Token* peekPast(alias O, alias C)()
in
{
assert (tokens[index].type == O);
}
body
{
int depth = 1;
auto i = index;
++i;
while (index < tokens.length)
{
if (tokens[i] == O)
++depth;
else if (tokens[i] == C)
{
--depth;
++i;
if (depth <= 0)
break;
}
++i;
}
return depth == 0 ? &tokens[i] : null;
}
Token* peekPastParens()
{
return peekPast!(TokenType.lParen, TokenType.rParen)();
}
Token* peekPastBrackets()
{
return peekPast!(TokenType.lBracket, TokenType.rBracket)();
}
Token* peekPastBraces()
{
return peekPast!(TokenType.lBrace, TokenType.rBrace)();
}
Token* expect(TokenType type)
{
if (tokens[index].type == type)
return &tokens[index++];
else
return null;
}
Token* peek()
{
return index + 1 < tokens.length ? &tokens[index + 1] : null;
}
bool nextIs(TokenType t)
{
return peek() && peek().type == t;
}
bool startsWith(TokenType types...)
{
for (size_t i = 0; i != types.length; ++i)
{
if (tokens[index + i].type != types[i])
return false;
}
return true;
}
bool moreTokens()
{
return index < tokens.length;
}
Token[] tokens;
size_t index;
string fileName;
}
unittest
{
auto source = cast(ubyte[]) q{import std.stdio;
import std.ascii: hexDigits;
import r = std.range;
import foo, bar;
import std.stdio : writefln, foo = writef;}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(4);
assert (tokens[i] == "import");
auto decl = parseImportDeclaration(tokens);
assert (decl.imports.length == 1);
assert (decl.imports[0].packageParts == ["std"]);
assert (decl.imports[0].moduleName == "stdio");
assert (tokens[i].value == "import", tokens.front.value);
assert (tokens.peek(3).value == "ascii", tokens.front.value);
decl = parseImportDeclaration(tokens);
assert (decl.imports.length == 1, "%d".format(decl.imports.length));
assert (decl.imports[0].packageParts == ["std"]);
assert (decl.imports[0].moduleName == "ascii", decl.imports[0].moduleName);
assert (decl.imports[0].symbols[0].symbolName == "hexDigits", decl.imports[0].symbols[0].symbolName);
assert (decl.imports[0].symbols[0].alias_.length == 0);
decl = parseImportDeclaration(tokens);
assert (decl.imports.length == 1, "%s".format(decl.imports.length));
assert (decl.imports[0].moduleName == "range");
assert (decl.imports[0].packageParts == ["std"]);
assert (decl.imports[0].alias_ == "r");
decl = parseImportDeclaration(tokens);
assert (decl.imports.length == 2);
assert (decl.imports[0].packageParts.length == 0);
assert (decl.imports[0].moduleName == "foo");
assert (decl.imports[1].packageParts.length == 0);
assert (decl.imports[1].moduleName == "bar");
decl = parseImportDeclaration(tokens);
assert (decl.imports.length == 1, "%s".format(decl.imports.length));
assert (decl.imports[0].packageParts == ["std"]);
assert (decl.imports[0].moduleName == "stdio");
assert (decl.imports[0].symbols.length == 2);
assert (decl.imports[0].symbols[0].symbolName == "writefln");
assert (decl.imports[0].symbols[1].symbolName == "writef");
assert (decl.imports[0].symbols[1].alias_ == "foo");
}
//
//unittest
//{
// auto a = cast(ubyte[]) q{/** */ module a.b.c;};
// LexerConfig config;
// auto ta = byToken(a, config);
// auto moda = parseModuleDeclaration(ta);
// assert (moda.packageName == ["a", "b"]);
// assert (moda.moduleName == "c");
//
// auto b = cast(ubyte[]) q{module a;};
// auto tb = byToken(b, config);
// auto modb = parseModuleDeclaration(tb);
// assert (modb.packageName.length == 0);
// assert (modb.moduleName == "a");
//}
//
//NonEmptyStatement parseNonEmptyStatement(Token[] tokens)
//{
// switch (tokens[i].type)
// {
// case TokenType.case_:
// return null;
// case TokenType.default_:
// return parseDefaultStatement(tokens);
// default:
// return parseNonEmptyStatementNoCaseNoDefault(tokens);
// }
//}
//
//NonEmptyStatementNoCaseNoDefault parseNonEmptyStatementNoCaseNoDefault(Token[] tokens)
//{
// switch (tokens[i].type)
// {
// case TokenType.identifier:
// case TokenType.if_:
// return parseIfStatement(tokens);
// case TokenType.while_:
// return parseWhileStatement(tokens);
// case TokenType.do_:
// return parseDoStatement(tokens);
// case TokenType.for_:
// return parseForStatement(tokens);
// case TokenType.foreach_:
// return parseForeachStatement(tokens);
// case TokenType.switch_:
// return parseSwitchStatement(tokens);
// case TokenType.final_:
// if (tokens.peek(1).type == TokenType.switch_)
// return parseFinalSwitchStatement(tokens);
// else
// goto default;
// case TokenType.continue_:
// return parseContinueStatement(tokens);
// case TokenType.break_:
// return parseBreakStatement(tokens);
// case TokenType.return_:
// return parseReturnStatement(tokens);
// case TokenType.goto_:
// return parseGotoStatement(tokens);
// case TokenType.with_:
// return parseWithStatement(tokens);
// case TokenType.synchronized_:
// return parseSynchronizedStatement(tokens);
// case TokenType.try_:
// return parseTryStatement(tokens);
// case TokenType.scope_:
// return parseScopeGuardStatement(tokens);
// case TokenType.throw_:
// return parseThrowStatement(tokens);
// case TokenType.asm_:
// return parseAsmStatement(tokens);
// case TokenType.pragma_:
// return parsePragmaStatement(tokens);
// case TokenType.mixin_:
// if (tokens.peek(1).type == TokenType.lParen)
// return parseMixinStatement(tokens);
// else if (tokens.peek(1).type == TokenType.identifier)
// return parseTemplateMixinStatement(tokens);
// else
// {
// error(tokens, "Expected identifier or ( following \"mixin\"");
// return null;
// }
// case TokenType.version_:
// if (tokens.peek(1).type == TokenType.lParen)
// return parseConditionalStatement(tokens);
// else
// {
// error(tokens, "Expected ( following \"version\"");
// return null;
// }
// case TokenType.debug_:
// return parseConditionalStatement(tokens);
// case TokenType.static_:
// if (tokens.peek(1).type == TokenType.if_)
// return parseConditionalStatement(tokens);
// else if (tokens.peek(1).type == TokenType.assert_)
// return parseStaticAssert(tokens);
// else
// {
// error(tokens, "Expected \"if\" or \"assert\" following \"static\"");
// return null;
// }
// case TokenType.import_:
// return parseImportDeclaration(tokens);
// default:
// auto d = parseDeclarationStatement(tokens);
// if (d is null)
// {
// auto e = parseExpressionStatement(tokens);
// if (e is null)
// {
// error(tokens, "OMGWTF");
// return null;
// }
// else
// return e;
// }
// else
// return d;
// }
//}
//
//
//
//T parseSingleTokenExpression(TokType, AstType, R)(ref R range)
//{
// auto node = new AstType;
// node.token = range.moveFront();
// return node;
//}
//
//AssignExpression parseAssignExpression(Tokens)(ref Tokens tokens)
//{
// auto expr = new AssignExpression;
// expr.left = parseConditionalExpression(tokens);
// switch (tokens[i].type)
// {
// case TokenType.assign:
// case TokenType.plusEqual:
// case TokenType.minusEqual:
// case TokenType.mulEqual:
// case TokenType.divEqual:
// case TokenType.modEqual:
// case TokenType.bitAndEqual:
// case TokenType.bitOrEqual:
// case TokenType.xorEqual:
// case TokenType.catEqual:
// case TokenType.shiftLeftEqual:
// case TokenType.shiftRightEqual:
// case TokenType.unsignedShiftRightEqual:
// case TokenType.powEqual:
// expr.operator = tokens.moveFront().type;
// expr.right = parseAssignExpression();
// default:
// break;
// }
// return expr;
//}
/+
bool isIdentifierOrBasicType(const TokenType type)
{
return isType(type) || type == TokenType.identifier;
}
string parseType(Tokens)(ref Tokens tokens)
{
if (tokens.front != TokenType.identifier && !isType(tokens.front.type))
return null;
auto app = appender!(ubyte[])();
switch (tokens.front.type)
{
case TokenType.const_:
case TokenType.immutable_:
case TokenType.shared_:
case TokenType.inout_:
case TokenType.typeof_:
app.put(cast(ubyte[]) tokens.moveFront().value);
if (tokens.empty) goto ret;
if (tokens.front.type == TokenType.lParen)
app.put(cast(ubyte[]) parenContent(tokens));
break;
case TokenType.bool_: .. case TokenType.wchar_:
case TokenType.identifier:
app.put(cast(ubyte[]) tokens.moveFront().value);
break;
default:
return null;
}
if (tokens.empty) goto ret;
if (tokens.front.type == TokenType.not)
{
app.put('!');
tokens.popFront();
if (tokens.empty) goto ret;
if (tokens.front.type == TokenType.lParen)
app.put(cast(ubyte[]) parenContent(tokens));
else if (isIdentifierOrBasicType(tokens.front.type))
app.put(cast(ubyte[]) tokens.moveFront().value);
else
goto ret;
}
else if (tokens.front.type == TokenType.function_ || tokens.front.type == TokenType.delegate_)
{
app.put(' ');
app.put(cast(ubyte[]) tokens.moveFront().value);
if (tokens.empty) goto ret;
if (tokens.front.type == TokenType.lParen)
{
app.put(cast(ubyte[]) parenContent(tokens));
goto ret;
}
}
loop: while (!tokens.empty)
{
switch (tokens.front.type)
{
case TokenType.star:
app.put('*');
tokens.popFront();
break;
case TokenType.lBracket:
app.put(cast(ubyte[]) bracketContent(tokens));
break;
default:
break loop;
}
}
ret:
return cast(string) app.data;
}
unittest
{
auto sources = [
q{int}c,
q{int function(int,int)}c,
q{void}c,
q{char*}c,
q{char*[]*[]}c,
q{Stuff!(int,double)*}c,
q{Template!a[]}c,
q{Template!(a)[]}c,
];
LexerConfig config;
foreach (source; sources)
{
auto tokens = (cast(ubyte[]) source).byToken(config).circularBuffer(4);
auto t = parseType(tokens);
assert (t == source, t);
}
}
Parameter parseParameter(Tokens)(ref Tokens tokens)
{
Parameter p;
p.type = parseType(tokens);
if (!tokens.empty && (tokens.front.type == TokenType.delegate_ || tokens.front.type == TokenType.function_))
{
p.type ~= " " ~ tokens.moveFront().value;
if (tokens.front.type == TokenType.lParen)
p.type ~= "(" ~ parenContent(tokens) ~")";
}
if (tokens.empty || tokens.front.type == TokenType.comma || tokens.front.type == TokenType.rParen)
{
p.name = p.type;
p.type = "";
}
else
p.name = tokens.moveFront().value;
return p;
}
unittest
{
auto source = cast(ubyte[]) q{int[] a}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(4);
auto p = parseParameter(tokens);
assert (p.name == "a", p.name);
assert (p.type == "int[]", p.type);
}
Parameter[] parseParameters(Tokens)(ref Tokens tokens)
in
{
assert (tokens.front == TokenType.lParen);
}
body
{
tokens.popFront();
if (tokens.front.type == TokenType.rParen)
{
tokens.popFront();
return [];
}
auto app = appender!(Parameter[])();
while (!tokens.empty)
{
app.put(parseParameter(tokens));
if (tokens.empty)
break;
else if (tokens.front.type == TokenType.rParen)
{
tokens.popFront();
break;
}
else if (tokens.front == TokenType.comma)
tokens.popFront();
else
{
tokens.popFront();
break;
}
}
return app.data;
}
unittest
{
auto source = cast(ubyte[]) q{(int[] a, double d, void*[] voids,
void function(int**, double*[][string]) vf, R);}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(4);
auto p = parseParameters(tokens);
assert (p.length == 5, "%d".format(p.length));
assert (p[0].name == "a", p[0].name);
assert (p[0].type == "int[]", p[0].type);
assert (p[1].name == "d", p[1].name);
assert (p[1].type == "double", p[1].type);
assert (p[2].name == "voids", p[2].name);
assert (p[2].type == "void*[]", p[2].type);
assert (p[3].name == "vf", p[3].name);
assert (p[3].type == "void function(int**,double*[][string])", p[3].type);
assert (p[4].name == "R", p[4].name);
assert (p[4].type == "", p[4].type);
assert (tokens.front.type == TokenType.semicolon, tokens.front.value);
}
unittest
{
auto source = cast(ubyte[]) q{()}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(2);
auto p = parseParameters(tokens);
assert (p.length == 0, "%s".format(p.length));
}
FunctionDeclaration parseFunctionDeclaration(Tokens)(ref Tokens tokens, string type)
{
FunctionDeclaration fun;
fun.returnType = type;
fun.name = tokens.moveFront().value;
fun.rtParameters = parseParameters(tokens);
if (tokens.front.type == TokenType.lParen)
{
fun.ctParameters = fun.rtParameters;
fun.rtParameters = parseParameters(tokens);
}
while (!tokens.empty && isAttribute(tokens.front.type))
fun.attributes.set(tokens.moveFront().type);
if (tokens.front.type == TokenType.lBrace)
skipBraces(tokens);
else if (tokens.front.type == TokenType.semicolon)
tokens.popFront();
return fun;
}
unittest
{
auto source = cast(ubyte[]) q{
void doStuff();
T calcSomething(T)(T input) { return input * 2; }
const(string)[] getStrings() const {}
}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(4);
auto decl = parseFunctionDeclaration(tokens, parseType(tokens));
assert (decl.name == "doStuff");
assert (decl.returnType == "void");
assert (decl.ctParameters.length == 0);
assert (decl.rtParameters.length == 0);
decl = parseFunctionDeclaration(tokens, parseType(tokens));
assert (decl.name == "calcSomething");
assert (decl.returnType == "T");
assert (decl.ctParameters[0].name == "T");
assert (decl.rtParameters[0].type == "T");
assert (decl.rtParameters[0].name == "input");
decl = parseFunctionDeclaration(tokens, parseType(tokens));
assert (decl.returnType == "const(string)[]", decl.returnType);
assert (decl.name == "getStrings", decl.name);
assert (decl.ctParameters.length == 0);
assert (decl.rtParameters.length == 0);
assert (decl.attributes.get() == [TokenType.const_]);
assert (tokens.empty);
}
VariableDeclaration parseVariableDeclaration(Tokens)(ref Tokens tokens, string type)
{
VariableDeclaration v;
v.line = tokens.front.line;
v.type = type;
v.name = tokens.front.value;
tokens.popFront();
if (tokens.front.type == TokenType.semicolon)
tokens.popFront();
else
skipPastSemicolon(tokens);
return v;
}
unittest
{
auto source = cast(ubyte[]) q{int c;}c;
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(2);
auto decl = parseVariableDeclaration(tokens, parseType(tokens));
assert (decl.name == "c");
assert (decl.type == "int", decl.type);
assert (tokens.empty);
}
ModuleDeclaration parseModuleDeclaration(Tokens)(ref Tokens tokens)
in
{
assert (tokens.front.type == TokenType.module_);
}
body
{
tokens.popFront();
ModuleDeclaration declaration;
string recent;
loop: while (!tokens.empty)
{
if (tokens.front.type == TokenType.identifier)
{
recent = tokens.moveFront().value;
switch (tokens.front.type)
{
case TokenType.dot:
declaration.package_ ~= recent;
tokens.popFront();
break;
case TokenType.semicolon:
declaration.name = recent;
tokens.popFront();
break loop;
default:
break;
}
}
else
skipPastSemicolon(tokens);
}
return declaration;
}
unittest
{
auto a = cast(ubyte[]) q{/** */ module a.b.c;};
LexerConfig config;
auto ta = byToken(a, config).circularBuffer(2);
auto moda = parseModuleDeclaration(ta);
assert (moda.package_ == ["a", "b"]);
assert (moda.name == "c");
auto b = cast(ubyte[]) q{module a;};
auto tb = byToken(b, config).circularBuffer(2);
auto modb = parseModuleDeclaration(tb);
assert (modb.package_.length == 0);
assert (modb.name == "a");
}
Module parseMod(Tokens)(ref Tokens tokens)
{
Module mod;
while (!tokens.empty)
{
switch (tokens.front.type)
{
// case TokenType.const_:
// case TokenType.immutable_:
// case TokenType.shared_:
// case TokenType.inout_:
// if (tokens.peek.type == TokenType.lParen)
// {
// auto
// }
// break;
case TokenType.rBrace:
return mod;
case TokenType.identifier:
case TokenType.bool_: .. case TokenType.wchar_:
auto type = parseType(tokens);
if (tokens.front.type == TokenType.identifier)
{
if (tokens.peek.type == TokenType.lParen)
mod.functions ~= parseFunctionDeclaration(tokens, type);
else
mod.variables ~= parseVariableDeclaration(tokens, type);
}
else
skipPastSemicolon(tokens);
break;
case TokenType.module_:
mod.moduleDeclaration = parseModuleDeclaration(tokens);
break;
case TokenType.class_:
mod.classes ~= parseClassDeclaration(tokens);
break;
// case TokenType.align_:
// case TokenType.deprecated_:
// case TokenType.extern_:
default:
tokens.popFront();
}
}
return mod;
}
Module parseModule(ref const(ubyte)[] source)
{
LexerConfig config;
auto tokens = source.byToken(config).circularBuffer(2);
AttributeList attributes;
return parseMod(tokens);
}
unittest
{
auto source = cast(const(ubyte)[]) q{
module a.b.c;
int x;
int doStuff();
int doOtherStuff() {}
class Point { int x; int y; }
}c;
auto mod = parseModule(source);
assert (mod.moduleDeclaration.name == "c");
assert (mod.moduleDeclaration.package_ == ["a", "b"]);
assert (mod.functions.length == 2);
assert (mod.variables.length == 1);
assert (mod.classes.length == 1);
}
ClassDeclaration parseClassDeclaration(Tokens)(ref Tokens tokens)
in
{
assert (tokens.front.type == TokenType.class_);
}
body
{
tokens.popFront();
ClassDeclaration decl;
if (tokens.front.type != TokenType.identifier)
return decl;
}
+/
void main(string[] args)
{
}