100 lines
2.4 KiB
Go
100 lines
2.4 KiB
Go
package lexer
|
|
|
|
import (
|
|
"bytes"
|
|
"io"
|
|
"strings"
|
|
"text/scanner"
|
|
)
|
|
|
|
// TextScannerLexer is a lexer that uses the text/scanner module.
|
|
var (
|
|
TextScannerLexer Definition = &defaultDefinition{}
|
|
|
|
// DefaultDefinition defines properties for the default lexer.
|
|
DefaultDefinition = TextScannerLexer
|
|
)
|
|
|
|
type defaultDefinition struct{}
|
|
|
|
func (d *defaultDefinition) Lex(filename string, r io.Reader) (Lexer, error) {
|
|
return Lex(filename, r), nil
|
|
}
|
|
|
|
func (d *defaultDefinition) Symbols() map[string]TokenType {
|
|
return map[string]TokenType{
|
|
"EOF": EOF,
|
|
"Char": scanner.Char,
|
|
"Ident": scanner.Ident,
|
|
"Int": scanner.Int,
|
|
"Float": scanner.Float,
|
|
"String": scanner.String,
|
|
"RawString": scanner.RawString,
|
|
"Comment": scanner.Comment,
|
|
}
|
|
}
|
|
|
|
// textScannerLexer is a Lexer based on text/scanner.Scanner
|
|
type textScannerLexer struct {
|
|
scanner *scanner.Scanner
|
|
filename string
|
|
err error
|
|
}
|
|
|
|
// Lex an io.Reader with text/scanner.Scanner.
|
|
//
|
|
// This provides very fast lexing of source code compatible with Go tokens.
|
|
//
|
|
// Note that this differs from text/scanner.Scanner in that string tokens will be unquoted.
|
|
func Lex(filename string, r io.Reader) Lexer {
|
|
s := &scanner.Scanner{}
|
|
s.Init(r)
|
|
lexer := lexWithScanner(filename, s)
|
|
lexer.scanner.Error = func(s *scanner.Scanner, msg string) {
|
|
// This is to support single quoted strings. Hacky.
|
|
if !strings.HasSuffix(msg, "char literal") {
|
|
lexer.err = errorf(Position(lexer.scanner.Pos()), msg)
|
|
}
|
|
}
|
|
return lexer
|
|
}
|
|
|
|
// LexWithScanner creates a Lexer from a user-provided scanner.Scanner.
|
|
//
|
|
// Useful if you need to customise the Scanner.
|
|
func LexWithScanner(filename string, scan *scanner.Scanner) Lexer {
|
|
return lexWithScanner(filename, scan)
|
|
}
|
|
|
|
func lexWithScanner(filename string, scan *scanner.Scanner) *textScannerLexer {
|
|
lexer := &textScannerLexer{
|
|
filename: filename,
|
|
scanner: scan,
|
|
}
|
|
return lexer
|
|
}
|
|
|
|
// LexBytes returns a new default lexer over bytes.
|
|
func LexBytes(filename string, b []byte) Lexer {
|
|
return Lex(filename, bytes.NewReader(b))
|
|
}
|
|
|
|
// LexString returns a new default lexer over a string.
|
|
func LexString(filename, s string) Lexer {
|
|
return Lex(filename, strings.NewReader(s))
|
|
}
|
|
|
|
func (t *textScannerLexer) Next() (Token, error) {
|
|
typ := t.scanner.Scan()
|
|
text := t.scanner.TokenText()
|
|
pos := Position(t.scanner.Position)
|
|
pos.Filename = t.filename
|
|
if t.err != nil {
|
|
return Token{}, t.err
|
|
}
|
|
return Token{
|
|
Type: TokenType(typ),
|
|
Value: text,
|
|
Pos: pos,
|
|
}, nil
|
|
}
|