123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998 |
- /*
- * ECAL
- *
- * Copyright 2020 Matthias Ladkau. All rights reserved.
- *
- * This Source Code Form is subject to the terms of the MIT
- * License, If a copy of the MIT License was not distributed with this
- * file, You can obtain one at https://opensource.org/licenses/MIT.
- */
- package parser
- import (
- "fmt"
- )
- /*
- Map of AST nodes corresponding to lexer tokens. The map determines how a given
- sequence of lexer tokens are organized into an AST.
- */
- var astNodeMap map[LexTokenID]*ASTNode
- func init() {
- astNodeMap = map[LexTokenID]*ASTNode{
- TokenEOF: {NodeEOF, nil, nil, nil, nil, 0, ndTerm, nil},
- // Value tokens
- TokenSTRING: {NodeSTRING, nil, nil, nil, nil, 0, ndTerm, nil},
- TokenNUMBER: {NodeNUMBER, nil, nil, nil, nil, 0, ndTerm, nil},
- TokenIDENTIFIER: {NodeIDENTIFIER, nil, nil, nil, nil, 0, ndIdentifier, nil},
- // Constructed tokens
- TokenSTATEMENTS: {NodeSTATEMENTS, nil, nil, nil, nil, 0, nil, nil},
- TokenFUNCCALL: {NodeFUNCCALL, nil, nil, nil, nil, 0, nil, nil},
- TokenCOMPACCESS: {NodeCOMPACCESS, nil, nil, nil, nil, 0, nil, nil},
- TokenLIST: {NodeLIST, nil, nil, nil, nil, 0, nil, nil},
- TokenMAP: {NodeMAP, nil, nil, nil, nil, 0, nil, nil},
- TokenPARAMS: {NodePARAMS, nil, nil, nil, nil, 0, nil, nil},
- TokenGUARD: {NodeGUARD, nil, nil, nil, nil, 0, nil, nil},
- // Condition operators
- TokenGEQ: {NodeGEQ, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenLEQ: {NodeLEQ, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenNEQ: {NodeNEQ, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenEQ: {NodeEQ, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenGT: {NodeGT, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenLT: {NodeLT, nil, nil, nil, nil, 60, nil, ldInfix},
- // Grouping symbols
- TokenLPAREN: {"", nil, nil, nil, nil, 150, ndInner, nil},
- TokenRPAREN: {"", nil, nil, nil, nil, 0, nil, nil},
- TokenLBRACK: {"", nil, nil, nil, nil, 150, ndList, nil},
- TokenRBRACK: {"", nil, nil, nil, nil, 0, nil, nil},
- TokenLBRACE: {"", nil, nil, nil, nil, 150, ndMap, nil},
- TokenRBRACE: {"", nil, nil, nil, nil, 0, nil, nil},
- // Separators
- TokenDOT: {"", nil, nil, nil, nil, 0, nil, nil},
- TokenCOMMA: {"", nil, nil, nil, nil, 0, nil, nil},
- TokenSEMICOLON: {"", nil, nil, nil, nil, 0, nil, nil},
- // Grouping
- TokenCOLON: {NodeKVP, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenEQUAL: {NodePRESET, nil, nil, nil, nil, 60, nil, ldInfix},
- // Arithmetic operators
- TokenPLUS: {NodePLUS, nil, nil, nil, nil, 110, ndPrefix, ldInfix},
- TokenMINUS: {NodeMINUS, nil, nil, nil, nil, 110, ndPrefix, ldInfix},
- TokenTIMES: {NodeTIMES, nil, nil, nil, nil, 120, nil, ldInfix},
- TokenDIV: {NodeDIV, nil, nil, nil, nil, 120, nil, ldInfix},
- TokenDIVINT: {NodeDIVINT, nil, nil, nil, nil, 120, nil, ldInfix},
- TokenMODINT: {NodeMODINT, nil, nil, nil, nil, 120, nil, ldInfix},
- // Assignment statement
- TokenASSIGN: {NodeASSIGN, nil, nil, nil, nil, 10, nil, ldInfix},
- TokenLET: {NodeLET, nil, nil, nil, nil, 0, ndPrefix, nil},
- // Import statement
- TokenIMPORT: {NodeIMPORT, nil, nil, nil, nil, 0, ndImport, nil},
- TokenAS: {NodeAS, nil, nil, nil, nil, 0, nil, nil},
- // Sink definition
- TokenSINK: {NodeSINK, nil, nil, nil, nil, 0, ndSkink, nil},
- TokenKINDMATCH: {NodeKINDMATCH, nil, nil, nil, nil, 150, ndPrefix, nil},
- TokenSCOPEMATCH: {NodeSCOPEMATCH, nil, nil, nil, nil, 150, ndPrefix, nil},
- TokenSTATEMATCH: {NodeSTATEMATCH, nil, nil, nil, nil, 150, ndPrefix, nil},
- TokenPRIORITY: {NodePRIORITY, nil, nil, nil, nil, 150, ndPrefix, nil},
- TokenSUPPRESSES: {NodeSUPPRESSES, nil, nil, nil, nil, 150, ndPrefix, nil},
- // Function definition
- TokenFUNC: {NodeFUNC, nil, nil, nil, nil, 0, ndFunc, nil},
- TokenRETURN: {NodeRETURN, nil, nil, nil, nil, 0, ndReturn, nil},
- // Boolean operators
- TokenAND: {NodeAND, nil, nil, nil, nil, 40, nil, ldInfix},
- TokenOR: {NodeOR, nil, nil, nil, nil, 30, nil, ldInfix},
- TokenNOT: {NodeNOT, nil, nil, nil, nil, 20, ndPrefix, nil},
- // Condition operators
- TokenLIKE: {NodeLIKE, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenIN: {NodeIN, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenHASPREFIX: {NodeHASPREFIX, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenHASSUFFIX: {NodeHASSUFFIX, nil, nil, nil, nil, 60, nil, ldInfix},
- TokenNOTIN: {NodeNOTIN, nil, nil, nil, nil, 60, nil, ldInfix},
- // Constant terminals
- TokenFALSE: {NodeFALSE, nil, nil, nil, nil, 0, ndTerm, nil},
- TokenTRUE: {NodeTRUE, nil, nil, nil, nil, 0, ndTerm, nil},
- TokenNULL: {NodeNULL, nil, nil, nil, nil, 0, ndTerm, nil},
- // Conditional statements
- TokenIF: {NodeIF, nil, nil, nil, nil, 0, ndGuard, nil},
- TokenELIF: {"", nil, nil, nil, nil, 0, nil, nil},
- TokenELSE: {"", nil, nil, nil, nil, 0, nil, nil},
- // Loop statement
- TokenFOR: {NodeLOOP, nil, nil, nil, nil, 0, ndLoop, nil},
- TokenBREAK: {NodeBREAK, nil, nil, nil, nil, 0, ndTerm, nil},
- TokenCONTINUE: {NodeCONTINUE, nil, nil, nil, nil, 0, ndTerm, nil},
- // Try statement
- TokenTRY: {NodeTRY, nil, nil, nil, nil, 0, ndTry, nil},
- TokenEXCEPT: {NodeEXCEPT, nil, nil, nil, nil, 0, nil, nil},
- TokenFINALLY: {NodeFINALLY, nil, nil, nil, nil, 0, nil, nil},
- // Mutex statement
- TokenMUTEX: {NodeMUTEX, nil, nil, nil, nil, 0, ndMutex, nil},
- }
- }
- // Parser
- // ======
- /*
- Parser data structure
- */
- type parser struct {
- name string // Name to identify the input
- node *ASTNode // Current ast node
- tokens *LABuffer // Buffer which is connected to the channel which contains lex tokens
- rp RuntimeProvider // Runtime provider which creates runtime components
- }
- /*
- Parse parses a given input string and returns an AST.
- */
- func Parse(name string, input string) (*ASTNode, error) {
- return ParseWithRuntime(name, input, nil)
- }
- /*
- ParseWithRuntime parses a given input string and returns an AST decorated with
- runtime components.
- */
- func ParseWithRuntime(name string, input string, rp RuntimeProvider) (*ASTNode, error) {
- // Create a new parser with a look-ahead buffer of 3
- p := &parser{name, nil, NewLABuffer(Lex(name, input), 3), rp}
- // Read and set initial AST node
- node, err := p.next()
- if err != nil {
- return nil, err
- }
- p.node = node
- n, err := p.run(0)
- if err == nil && hasMoreStatements(p, n) {
- st := astNodeMap[TokenSTATEMENTS].instance(p, nil)
- st.Children = append(st.Children, n)
- for err == nil && hasMoreStatements(p, n) {
- // Skip semicolons
- if p.node.Token.ID == TokenSEMICOLON {
- skipToken(p, TokenSEMICOLON)
- }
- n, err = p.run(0)
- st.Children = append(st.Children, n)
- }
- n = st
- }
- if err == nil && p.node != nil && p.node.Token.ID != TokenEOF {
- token := *p.node.Token
- err = p.newParserError(ErrUnexpectedEnd, fmt.Sprintf("extra token id:%v (%v)",
- token.ID, token), token)
- }
- return n, err
- }
- /*
- run models the main parser function.
- */
- func (p *parser) run(rightBinding int) (*ASTNode, error) {
- var err error
- n := p.node
- p.node, err = p.next()
- if err != nil {
- return nil, err
- }
- // Start with the null denotation of this statement / expression
- if n.nullDenotation == nil {
- return nil, p.newParserError(ErrImpossibleNullDenotation,
- n.Token.String(), *n.Token)
- }
- left, err := n.nullDenotation(p, n)
- if err != nil {
- return nil, err
- }
- // Collect left denotations as long as the left binding power is greater
- // than the initial right one
- for rightBinding < p.node.binding {
- var nleft *ASTNode
- n = p.node
- if n.leftDenotation == nil {
- if left.Token.Lline < n.Token.Lline {
- // If the impossible left denotation is on a new line
- // we might be parsing a new statement
- return left, nil
- }
- return nil, p.newParserError(ErrImpossibleLeftDenotation,
- n.Token.String(), *n.Token)
- }
- p.node, err = p.next()
- if err != nil {
- return nil, err
- }
- // Get the next left denotation
- nleft, err = n.leftDenotation(p, n, left)
- left = nleft
- if err != nil {
- return nil, err
- }
- }
- return left, nil
- }
- /*
- next retrieves the next lexer token.
- */
- func (p *parser) next() (*ASTNode, error) {
- var preComments []MetaData
- var postComments []MetaData
- token, more := p.tokens.Next()
- for more && (token.ID == TokenPRECOMMENT || token.ID == TokenPOSTCOMMENT) {
- if token.ID == TokenPRECOMMENT {
- // Skip over pre comment token
- preComments = append(preComments, NewLexTokenInstance(token))
- token, more = p.tokens.Next()
- }
- if token.ID == TokenPOSTCOMMENT {
- // Skip over post comment token
- postComments = append(postComments, NewLexTokenInstance(token))
- token, more = p.tokens.Next()
- }
- }
- if !more {
- // Unexpected end of input - the associated token is an empty error token
- return nil, p.newParserError(ErrUnexpectedEnd, "", token)
- } else if token.ID == TokenError {
- // There was a lexer error wrap it in a parser error
- return nil, p.newParserError(ErrLexicalError, token.Val, token)
- } else if node, ok := astNodeMap[token.ID]; ok {
- // We got a normal AST component
- ret := node.instance(p, &token)
- ret.Meta = append(ret.Meta, preComments...) // Attach pre comments to the next AST node
- if len(postComments) > 0 && p.node != nil {
- p.node.Meta = append(p.node.Meta, postComments...) // Attach post comments to the previous AST node
- }
- return ret, nil
- }
- return nil, p.newParserError(ErrUnknownToken, fmt.Sprintf("id:%v (%v)", token.ID, token), token)
- }
- // Standard null denotation functions
- // ==================================
- /*
- ndTerm is used for terminals.
- */
- func ndTerm(p *parser, self *ASTNode) (*ASTNode, error) {
- return self, nil
- }
- /*
- ndInner returns the inner expression of an enclosed block and discard the
- block token. This method is used for brackets.
- */
- func ndInner(p *parser, self *ASTNode) (*ASTNode, error) {
- // Get the inner expression
- exp, err := p.run(0)
- if err != nil {
- return nil, err
- }
- // We return here the inner expression - discarding the bracket tokens
- return exp, skipToken(p, TokenRPAREN)
- }
- /*
- ndPrefix is used for prefix operators.
- */
- func ndPrefix(p *parser, self *ASTNode) (*ASTNode, error) {
- // Make sure a prefix will only prefix the next item
- val, err := p.run(self.binding + 20)
- if err != nil {
- return nil, err
- }
- self.Children = append(self.Children, val)
- return self, nil
- }
- // Null denotation functions for specific expressions
- // ==================================================
- /*
- ndImport is used to parse imports.
- */
- func ndImport(p *parser, self *ASTNode) (*ASTNode, error) {
- // Must specify a file path
- err := acceptChild(p, self, TokenSTRING)
- if err == nil {
- // Must specify AS
- if err = skipToken(p, TokenAS); err == nil {
- // Must specify an identifier
- err = acceptChild(p, self, TokenIDENTIFIER)
- }
- }
- return self, err
- }
- /*
- ndSink is used to parse sinks.
- */
- func ndSkink(p *parser, self *ASTNode) (*ASTNode, error) {
- var exp, ret *ASTNode
- // Must specify a name
- err := acceptChild(p, self, TokenIDENTIFIER)
- if err == nil {
- // Parse the rest of the parameters as children until we reach the body
- for err == nil && IsNotEndAndNotTokens(p, []LexTokenID{TokenLBRACE}) {
- if exp, err = p.run(150); err == nil {
- self.Children = append(self.Children, exp)
- // Skip commas
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- // Parse the body
- ret, err = parseInnerStatements(p, self)
- }
- }
- return ret, err
- }
- /*
- ndFunc is used to parse function definitions.
- */
- func ndFunc(p *parser, self *ASTNode) (*ASTNode, error) {
- var exp *ASTNode
- var err error
- // Might specify a function name
- if p.node.Token.ID == TokenIDENTIFIER {
- err = acceptChild(p, self, TokenIDENTIFIER)
- }
- // Read in parameters
- if err == nil {
- err = skipToken(p, TokenLPAREN)
- params := astNodeMap[TokenPARAMS].instance(p, nil)
- self.Children = append(self.Children, params)
- for err == nil && IsNotEndAndNotTokens(p, []LexTokenID{TokenRPAREN}) {
- // Parse all the expressions inside
- if exp, err = p.run(0); err == nil {
- params.Children = append(params.Children, exp)
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- err = skipToken(p, TokenRPAREN)
- }
- }
- if err == nil {
- // Parse the body
- self, err = parseInnerStatements(p, self)
- }
- return self, err
- }
- /*
- ndReturn is used to parse return statements.
- */
- func ndReturn(p *parser, self *ASTNode) (*ASTNode, error) {
- var err error
- if self.Token.Lline == p.node.Token.Lline {
- var val *ASTNode
- // Consume the next expression only if it is on the same line
- val, err = p.run(0)
- if err == nil {
- self.Children = append(self.Children, val)
- }
- }
- return self, err
- }
- /*
- ndIdentifier is to parse identifiers and function calls.
- */
- func ndIdentifier(p *parser, self *ASTNode) (*ASTNode, error) {
- var parseMore, parseSegment, parseFuncCall, parseCompositionAccess func(parent *ASTNode) error
- parseMore = func(current *ASTNode) error {
- var err error
- if p.node.Token.ID == TokenDOT {
- err = parseSegment(current)
- } else if p.node.Token.ID == TokenLPAREN {
- err = parseFuncCall(current)
- } else if p.node.Token.ID == TokenLBRACK && p.node.Token.Lline == self.Token.Lline {
- skipToken(p, TokenLBRACK)
- // Composition access needs to be on the same line as the identifier
- // as we might otherwise have a list
- err = parseCompositionAccess(current)
- }
- return err
- }
- parseSegment = func(current *ASTNode) error {
- var err error
- var next *ASTNode
- if err = skipToken(p, TokenDOT); err == nil {
- next = p.node
- if err = acceptChild(p, current, TokenIDENTIFIER); err == nil {
- err = parseMore(next)
- }
- }
- return err
- }
- parseFuncCall = func(current *ASTNode) error {
- var exp *ASTNode
- err := skipToken(p, TokenLPAREN)
- fc := astNodeMap[TokenFUNCCALL].instance(p, nil)
- current.Children = append(current.Children, fc)
- // Read in parameters
- for err == nil && IsNotEndAndNotTokens(p, []LexTokenID{TokenRPAREN}) {
- // Parse all the expressions inside the directives
- if exp, err = p.run(0); err == nil {
- fc.Children = append(fc.Children, exp)
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- if err = skipToken(p, TokenRPAREN); err == nil {
- err = parseMore(current)
- }
- }
- return err
- }
- parseCompositionAccess = func(current *ASTNode) error {
- var exp *ASTNode
- var err error
- ca := astNodeMap[TokenCOMPACCESS].instance(p, nil)
- current.Children = append(current.Children, ca)
- // Parse all the expressions inside the directives
- if exp, err = p.run(0); err == nil {
- ca.Children = append(ca.Children, exp)
- if err = skipToken(p, TokenRBRACK); err == nil {
- err = parseMore(current)
- }
- }
- return err
- }
- return self, parseMore(self)
- }
- /*
- ndList is used to collect elements of a list.
- */
- func ndList(p *parser, self *ASTNode) (*ASTNode, error) {
- var err error
- var exp *ASTNode
- // Create a list token
- st := astNodeMap[TokenLIST].instance(p, self.Token)
- // Get the inner expression
- for err == nil && IsNotEndAndNotTokens(p, []LexTokenID{TokenRBRACK}) {
- // Parse all the expressions inside
- if exp, err = p.run(0); err == nil {
- st.Children = append(st.Children, exp)
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- err = skipToken(p, TokenRBRACK)
- }
- // Must have a closing bracket
- return st, err
- }
- /*
- ndMap is used to collect elements of a map.
- */
- func ndMap(p *parser, self *ASTNode) (*ASTNode, error) {
- var err error
- var exp *ASTNode
- // Create a map token
- st := astNodeMap[TokenMAP].instance(p, self.Token)
- // Get the inner expression
- for err == nil && IsNotEndAndNotTokens(p, []LexTokenID{TokenRBRACE}) {
- // Parse all the expressions inside
- if exp, err = p.run(0); err == nil {
- st.Children = append(st.Children, exp)
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- err = skipToken(p, TokenRBRACE)
- }
- // Must have a closing brace
- return st, err
- }
- /*
- ndGuard is used to parse a conditional statement.
- */
- func ndGuard(p *parser, self *ASTNode) (*ASTNode, error) {
- var err error
- parseGuardAndStatements := func() error {
- // The brace starts statements while parsing the expression of an if statement
- nodeMapEntryBak := astNodeMap[TokenLBRACE]
- astNodeMap[TokenLBRACE] = &ASTNode{"", nil, nil, nil, nil, 0, parseInnerStatements, nil}
- exp, err := p.run(0)
- astNodeMap[TokenLBRACE] = nodeMapEntryBak
- if err == nil {
- g := astNodeMap[TokenGUARD].instance(p, nil)
- g.Children = append(g.Children, exp)
- self.Children = append(self.Children, g)
- _, err = parseInnerStatements(p, self)
- }
- return err
- }
- if err = parseGuardAndStatements(); err == nil {
- for err == nil && IsNotEndAndToken(p, TokenELIF) {
- // Parse an elif
- if err = skipToken(p, TokenELIF); err == nil {
- err = parseGuardAndStatements()
- }
- }
- if err == nil && p.node.Token.ID == TokenELSE {
- // Parse else
- if err = skipToken(p, TokenELSE); err == nil {
- g := astNodeMap[TokenGUARD].instance(p, nil)
- g.Children = append(g.Children, astNodeMap[TokenTRUE].instance(p, nil))
- self.Children = append(self.Children, g)
- _, err = parseInnerStatements(p, self)
- }
- }
- }
- return self, err
- }
- /*
- ndLoop is used to parse a loop statement.
- */
- func ndLoop(p *parser, self *ASTNode) (*ASTNode, error) {
- // The brace starts statements while parsing the expression of a for statement
- nodeMapEntryBak := astNodeMap[TokenLBRACE]
- astNodeMap[TokenLBRACE] = &ASTNode{"", nil, nil, nil, nil, 0, parseInnerStatements, nil}
- exp, err := p.run(0)
- astNodeMap[TokenLBRACE] = nodeMapEntryBak
- if err == nil {
- g := exp
- if exp.Token.ID != TokenIN {
- g = astNodeMap[TokenGUARD].instance(p, nil)
- g.Children = append(g.Children, exp)
- }
- self.Children = append(self.Children, g)
- _, err = parseInnerStatements(p, self)
- }
- return self, err
- }
- /*
- ndTry is used to parse a try block.
- */
- func ndTry(p *parser, self *ASTNode) (*ASTNode, error) {
- try, err := parseInnerStatements(p, self)
- for err == nil && IsNotEndAndToken(p, TokenEXCEPT) {
- except := p.node
- err = acceptChild(p, try, TokenEXCEPT)
- for err == nil &&
- IsNotEndAndNotTokens(p, []LexTokenID{TokenAS, TokenIDENTIFIER, TokenLBRACE}) {
- if err = acceptChild(p, except, TokenSTRING); err == nil {
- // Skip commas
- if p.node.Token.ID == TokenCOMMA {
- err = skipToken(p, TokenCOMMA)
- }
- }
- }
- if err == nil {
- if p.node.Token.ID == TokenAS {
- as := p.node
- if err = acceptChild(p, except, TokenAS); err == nil {
- err = acceptChild(p, as, TokenIDENTIFIER)
- }
- } else if p.node.Token.ID == TokenIDENTIFIER {
- err = acceptChild(p, except, TokenIDENTIFIER)
- }
- }
- if err == nil {
- _, err = parseInnerStatements(p, except)
- }
- }
- if err == nil && p.node.Token.ID == TokenFINALLY {
- finally := p.node
- if err = acceptChild(p, try, TokenFINALLY); err == nil {
- _, err = parseInnerStatements(p, finally)
- }
- }
- return try, err
- }
- /*
- ndMutex is used to parse a mutex block.
- */
- func ndMutex(p *parser, self *ASTNode) (*ASTNode, error) {
- var block *ASTNode
- err := acceptChild(p, self, TokenIDENTIFIER)
- if err == nil {
- block, err = parseInnerStatements(p, self)
- }
- return block, err
- }
- // Standard left denotation functions
- // ==================================
- /*
- ldInfix is used for infix operators.
- */
- func ldInfix(p *parser, self *ASTNode, left *ASTNode) (*ASTNode, error) {
- right, err := p.run(self.binding)
- if err != nil {
- return nil, err
- }
- self.Children = append(self.Children, left)
- self.Children = append(self.Children, right)
- return self, nil
- }
- // Helper functions
- // ================
- /*
- IsNotEndAndToken checks if the next token is of a specific type or the end has been reached.
- */
- func IsNotEndAndToken(p *parser, i LexTokenID) bool {
- return p.node != nil && p.node.Name != NodeEOF && p.node.Token.ID == i
- }
- /*
- IsNotEndAndNotTokens checks if the next token is not of a specific type or the end has been reached.
- */
- func IsNotEndAndNotTokens(p *parser, tokens []LexTokenID) bool {
- ret := p.node != nil && p.node.Name != NodeEOF
- for _, t := range tokens {
- ret = ret && p.node.Token.ID != t
- }
- return ret
- }
- /*
- hasMoreStatements returns true if there are more statements to parse.
- */
- func hasMoreStatements(p *parser, currentNode *ASTNode) bool {
- nextNode := p.node
- if nextNode == nil || nextNode.Token.ID == TokenEOF {
- return false
- } else if nextNode.Token.ID == TokenSEMICOLON {
- return true
- }
- return currentNode != nil && currentNode.Token.Lline < nextNode.Token.Lline
- }
- /*
- skipToken skips over a given token.
- */
- func skipToken(p *parser, ids ...LexTokenID) error {
- var err error
- canSkip := func(id LexTokenID) bool {
- for _, i := range ids {
- if i == id {
- return true
- }
- }
- return false
- }
- if !canSkip(p.node.Token.ID) {
- if p.node.Token.ID == TokenEOF {
- return p.newParserError(ErrUnexpectedEnd, "", *p.node.Token)
- }
- return p.newParserError(ErrUnexpectedToken, p.node.Token.Val, *p.node.Token)
- }
- // This should never return an error unless we skip over EOF or complex tokens
- // like values
- p.node, err = p.next()
- return err
- }
- /*
- acceptChild accepts the current token as a child.
- */
- func acceptChild(p *parser, self *ASTNode, id LexTokenID) error {
- var err error
- current := p.node
- if p.node, err = p.next(); err == nil {
- if current.Token.ID == id {
- self.Children = append(self.Children, current)
- } else {
- err = p.newParserError(ErrUnexpectedToken, current.Token.Val, *current.Token)
- }
- }
- return err
- }
- /*
- parseInnerStatements collects the inner statements of a block statement. It
- is assumed that a block statement starts with a left brace '{' and ends with
- a right brace '}'.
- */
- func parseInnerStatements(p *parser, self *ASTNode) (*ASTNode, error) {
- // Must start with an opening brace
- if err := skipToken(p, TokenLBRACE); err != nil {
- return nil, err
- }
- // Always create a statements node
- st := astNodeMap[TokenSTATEMENTS].instance(p, nil)
- self.Children = append(self.Children, st)
- // Check if there are actually children
- if p.node != nil && p.node.Token.ID != TokenRBRACE {
- n, err := p.run(0)
- if p.node != nil && p.node.Token.ID != TokenEOF {
- st.Children = append(st.Children, n)
- for hasMoreStatements(p, n) {
- if p.node.Token.ID == TokenSEMICOLON {
- skipToken(p, TokenSEMICOLON)
- } else if p.node.Token.ID == TokenRBRACE {
- break
- }
- n, err = p.run(0)
- st.Children = append(st.Children, n)
- }
- }
- if err != nil {
- return nil, err
- }
- }
- // Must end with a closing brace
- return self, skipToken(p, TokenRBRACE)
- }
|