一直以来对编译器/解释器等都较有兴趣。我非科班出身,当初还在大学时,只是马马虎虎看完了《编译原理》之类教材,上机非常少,对龙书之类圣经也只是浅尝辄止而已。工作至今,基本已将编译原理相关知识忘记得差不多了,可能也就还对譬如预处理词法分析语法分析 AST 生成等基础性的概念还有点印象罢。
约 1 年多前,我也有想法搞一套基于简化的 Pascal 语法的带类型的脚本语言“编译器”(PaxCompiler 之类可能太复杂了),并将此脚本语言编写的脚本与 Golang 交互起来。当然这只是我个人的业余兴趣而已,至于是否会付诸行动、能搞成怎样都是未知。而选择 Pascal 作为参考改编语言的原因,其一我比较喜欢它的语言设计,其二它曾是我某段时间内的工作语言所以感情成分使然,其三较之诸如 Python、Lua 我更喜欢带类型的脚本语言(TypeScript?我不太喜欢 JavaScript 的语法...),当然,Pascal 的语法形式也确实比较方便为之开发编译器/解释器。
而短期内,个人恐怕没有太多精力去啃龙书之类,于是索性,看点基础资料且按此系列教程之类慢慢温习并从 tokenizer 开始一步步实现自己的 EcoPascal——即便最终,它只是个玩具脚本语言而已。
近 2 天趁有空,粗略看了前文所述教程的前两章,并以 Golang 重写了这两章里的解释程序(代码写得有些粗放)。
第一章:
package interpreter import ( "fmt" ) // Token types // // EOF (end-of-file) token is used to indicate that // there is no more input left for lexical analysis type TokenType int const ( cTokenTypeOfNone TokenType = iota cTokenTypeOfInteger cTokenTypeOfPlusSign cTokenTypeOfEOF ) type token struct { t TokenType // token type: INTEGER, PLUS, or EOF v interface{} // token value: 0, 1, 2. 3, 4, 5, 6, 7, 8, 9, ‘+‘, or None } func newToken(t TokenType, v interface{}) token { return token{ t: t, v: v, } } type Interpreter struct { text []rune // client string input, e.g. "3+5" pos int // an index into text currToken token // current token instance } func New() *Interpreter { return &Interpreter{ text: []rune(""), pos: 0, currToken: newToken(cTokenTypeOfNone, nil), } } func convToDigit(c rune) (int, bool) { if ‘0‘ <= c && c <= ‘9‘ { return int(c - ‘0‘), true } return 0, false } // Lexical analyzer (also known as scanner or tokenizer) // // This method is responsible for breaking a sentence apart into tokens. // One token at a time. func (self *Interpreter) getNextToken() token { text := self.text // is self.pos index past the end of the self.text ? // if so, then return EOF token because there is no more // input left to convert into tokens if self.pos > len(text)-1 { return newToken(cTokenTypeOfEOF, nil) } // get a character at the position self.pos and decide // what token to create based on the single character // var currChar interface{} = text[self.pos] currChar := text[self.pos] // if the character is a digit then convert it to // integer, create an INTEGER token, increment self.pos // index to point to the next character after the digit, // and return the INTEGER token if v, ok := convToDigit(text[self.pos]); ok { self.pos += 1 return newToken(cTokenTypeOfInteger, v) } if currChar == ‘+‘ { self.pos += 1 return newToken(cTokenTypeOfPlusSign, ‘+‘) } panic(fmt.Sprintf("Error parsing input1: %s", string(self.text))) } // compare the current token type with the passed token type // and if they match then "eat" the current token // and assign the next token to the self.currToken, // otherwise raise an exception. func (self *Interpreter) eat(tokenType TokenType) { if self.currToken.t == tokenType { self.currToken = self.getNextToken() return } panic(fmt.Sprintf("Error parsing input: %s", self.text)) } // parse "INTEGER PLUS INTEGER" func (self *Interpreter) Parse(s string) int { self.text = []rune(s) self.pos = 0 // set current token to the first token taken from the input self.currToken = self.getNextToken() // we expect the current token to be a single-digit integer left := self.currToken self.eat(cTokenTypeOfInteger) // we expect the current token to be a ‘+‘ token // op := self.currToken self.eat(cTokenTypeOfPlusSign) // we expect the current token to be a single-digit integer right := self.currToken self.eat(cTokenTypeOfInteger) // after the above call the self.current_token is set to EOF token. // at this point INTEGER PLUS INTEGER sequence of tokens // has been successfully found and the method can just // return the result of adding two integers, thus // effectively interpreting client input return left.v.(int) + right.v.(int) }
第二章:
package interpreter import ( "fmt" "github.com/ecofast/rtl/sysutils" ) // Token types // // EOF (end-of-file) token is used to indicate that // there is no more input left for lexical analysis type TokenType int const ( cTokenTypeOfNone TokenType = iota cTokenTypeOfInteger cTokenTypeOfPlusSign cTokenTypeOfMinusSign cTokenTypeOfEOF ) type token struct { t TokenType // token type: INTEGER, PLUS, MINUS, or EOF v interface{} // token value: non-negative integer value, ‘+‘, ‘-‘, or None } func newToken(t TokenType, v interface{}) token { return token{ t: t, v: v, } } type Interpreter struct { text []rune // client string input, e.g. "3 + 5", "12 - 5", etc pos int // an index into text currToken token // current token instance currChar rune } func New() *Interpreter { return &Interpreter{ text: []rune(""), pos: 0, currToken: newToken(cTokenTypeOfNone, nil), currChar: 0, } } func isDigit(c rune) bool { if ‘0‘ <= c && c <= ‘9‘ { return true } return false } func convToDigit(c rune) (int, bool) { if ‘0‘ <= c && c <= ‘9‘ { return int(c - ‘0‘), true } return 0, false } func isSpace(c rune) bool { if ‘ ‘ == c { return true } return false } // Advance the ‘pos‘ pointer and set the ‘currChar‘ variable func (self *Interpreter) advance() { self.pos += 1 if self.pos > len(self.text)-1 { self.currChar = 0 } else { self.currChar = self.text[self.pos] } } func (self *Interpreter) skipWhiteSpace() { for self.currChar != 0 && isSpace(self.currChar) { self.advance() } } // Return a (multidigit) integer consumed from the input func (self *Interpreter) integer() int { ret := "" for self.currChar != 0 && isDigit(self.currChar) { ret += string(self.currChar) self.advance() } return sysutils.StrToInt(ret) } // Lexical analyzer (also known as scanner or tokenizer) // // This method is responsible for breaking a sentence apart into tokens. func (self *Interpreter) getNextToken() token { for self.currChar != 0 { if isSpace(self.currChar) { self.skipWhiteSpace() continue } if isDigit(self.currChar) { return newToken(cTokenTypeOfInteger, self.integer()) } if self.currChar == ‘+‘ { self.advance() return newToken(cTokenTypeOfPlusSign, ‘+‘) } if self.currChar == ‘-‘ { self.advance() return newToken(cTokenTypeOfMinusSign, ‘-‘) } panic(fmt.Sprintf("Error parsing input1: %s", string(self.text))) } return newToken(cTokenTypeOfEOF, nil) } // compare the current token type with the passed token type // and if they match then "eat" the current token // and assign the next token to the self.currToken, // otherwise raise an exception. func (self *Interpreter) eat(tokenType TokenType) { if self.currToken.t == tokenType { self.currToken = self.getNextToken() return } panic(fmt.Sprintf("Error parsing input: %s", self.text)) } // parse "INTEGER PLUS INTEGER" or "INTEGER MINUS INTEGER" func (self *Interpreter) Parse(s string) int { self.text = []rune(s) self.pos = 0 self.currChar = self.text[self.pos] // set current token to the first token taken from the input self.currToken = self.getNextToken() // we expect the current token to be an integer left := self.currToken self.eat(cTokenTypeOfInteger) // we expect the current token to be either a ‘+‘ or ‘-‘ op := self.currToken if op.t == cTokenTypeOfPlusSign { self.eat(cTokenTypeOfPlusSign) } else { self.eat(cTokenTypeOfMinusSign) } // we expect the current token to be an integer right := self.currToken self.eat(cTokenTypeOfInteger) // after the above call the self.current_token is set to EOF token. // at this point either the INTEGER PLUS INTEGER or // the INTEGER MINUS INTEGER sequence of tokens // has been successfully found and the method can just // return the result of adding or subtracting two integers, thus // effectively interpreting client input if op.t == cTokenTypeOfPlusSign { return left.v.(int) + right.v.(int) } return left.v.(int) - right.v.(int) }
有了“核心”解释程序,使用起来就很简单了:
package main import ( "fmt" "part1/interpreter" ) func main() { fmt.Println("Let‘s Build A Simple Interpreter - Part 1") parser := interpreter.New() s := "" for { if n, err := fmt.Scan(&s); n == 0 || err != nil { return } fmt.Println(parser.Parse(s)) } }
本兴趣项目已托管至 Github,比较可能会不定期慢慢更新。
《Let's Build A Simple Interpreter》之 Golang 版
原文地址:https://www.cnblogs.com/ecofast/p/8476689.html