mirror of
https://github.com/golang/go
synced 2024-11-25 05:57:57 -07:00
exp/template: lexical scanner for new template package.
An unusual design using slice and a goroutine makes for a compact scanner with very little allocation. R=rsc, r CC=golang-dev, kevlar https://golang.org/cl/4610041
This commit is contained in:
parent
9ded954a3b
commit
7bc8e90e36
11
src/pkg/exp/template/Makefile
Normal file
11
src/pkg/exp/template/Makefile
Normal file
@ -0,0 +1,11 @@
|
||||
# Copyright 2011 The Go Authors. All rights reserved.
|
||||
# Use of this source code is governed by a BSD-style
|
||||
# license that can be found in the LICENSE file.
|
||||
|
||||
include ../../../Make.inc
|
||||
|
||||
TARG=template
|
||||
GOFILES=\
|
||||
lex.go\
|
||||
|
||||
include ../../../Make.pkg
|
302
src/pkg/exp/template/lex.go
Normal file
302
src/pkg/exp/template/lex.go
Normal file
@ -0,0 +1,302 @@
|
||||
// Copyright 2011 The Go Authors. All rights reserved.
|
||||
// Use of this source code is governed by a BSD-style
|
||||
// license that can be found in the LICENSE file.
|
||||
|
||||
package template
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"strings"
|
||||
"unicode"
|
||||
"utf8"
|
||||
)
|
||||
|
||||
// item represents a token or text string returned from the scanner.
|
||||
type item struct {
|
||||
typ itemType
|
||||
val string
|
||||
}
|
||||
|
||||
// itemType identifies the type of lex item.
|
||||
type itemType int
|
||||
|
||||
const (
|
||||
itemError itemType = iota // error occurred; value is text of error
|
||||
itemText // plain text
|
||||
itemLeftMeta // left meta-string
|
||||
itemRightMeta // right meta-string
|
||||
itemPipe // pipe symbol
|
||||
itemIdentifier // alphanumeric identifier
|
||||
itemNumber // number
|
||||
itemRawString // raw quoted string (includes quotes)
|
||||
itemString // quoted string (includes quotes)
|
||||
itemEOF
|
||||
)
|
||||
|
||||
const eof = -1
|
||||
|
||||
// stateFn represents the state of the scanner as a function that returns the next state.
|
||||
type stateFn func(*lexer) stateFn
|
||||
|
||||
// lexer holds the state of the scanner.
|
||||
type lexer struct {
|
||||
name string // the name of the input; used only for error reports.
|
||||
input string // the string being scanned.
|
||||
pos int // current position in the input.
|
||||
start int // start position of this item.
|
||||
width int // width of last rune read from input.
|
||||
items chan item // channel of scanned items.
|
||||
}
|
||||
|
||||
// next returns the next rune in the input.
|
||||
func (l *lexer) next() (rune int) {
|
||||
if l.pos >= len(l.input) {
|
||||
return eof
|
||||
}
|
||||
rune, l.width = utf8.DecodeRuneInString(l.input[l.pos:])
|
||||
l.pos += l.width
|
||||
return rune
|
||||
}
|
||||
|
||||
// peek returns but does not consume the next rune in the input.
|
||||
func (l *lexer) peek() int {
|
||||
rune := l.next()
|
||||
l.backup()
|
||||
return rune
|
||||
}
|
||||
|
||||
// backup steps back one rune. Can only be called once per call of next.
|
||||
func (l *lexer) backup() {
|
||||
l.pos -= l.width
|
||||
}
|
||||
|
||||
// emit passes an item back to the client.
|
||||
func (l *lexer) emit(t itemType) {
|
||||
start := l.start
|
||||
l.start = l.pos
|
||||
l.items <- item{t, l.input[start:l.pos]}
|
||||
}
|
||||
|
||||
// ignore discards whatever input is before this point.
|
||||
func (l *lexer) ignore() {
|
||||
l.start = l.pos
|
||||
}
|
||||
|
||||
// accept consumes the next rune if it's from the valid set.
|
||||
func (l *lexer) accept(valid string) bool {
|
||||
if strings.IndexRune(valid, l.next()) >= 0 {
|
||||
return true
|
||||
}
|
||||
l.backup()
|
||||
return false
|
||||
}
|
||||
|
||||
// acceptRun consumes a run of runes from the valid set.
|
||||
func (l *lexer) acceptRun(valid string) {
|
||||
for strings.IndexRune(valid, l.next()) >= 0 {
|
||||
}
|
||||
l.backup()
|
||||
}
|
||||
|
||||
// lineNumber reports which line we're on. Doing it this way
|
||||
// means we don't have to worry about peek double counting.
|
||||
func (l *lexer) lineNumber() int {
|
||||
return 1 + strings.Count(l.input[:l.pos], "\n")
|
||||
}
|
||||
|
||||
// error returns an error token and terminates the scan by passing
|
||||
// back a nil pointer that will be the next state, terminating l.run.
|
||||
func (l *lexer) error(format string, args ...interface{}) stateFn {
|
||||
format = fmt.Sprintf("%s:%d %s", l.name, l.lineNumber(), format)
|
||||
l.items <- item{itemError, fmt.Sprintf(format, args...)}
|
||||
return nil
|
||||
}
|
||||
|
||||
// run lexes the input by execute state functions until nil.
|
||||
func (l *lexer) run() {
|
||||
for state := lexText; state != nil; {
|
||||
state = state(l)
|
||||
}
|
||||
close(l.items)
|
||||
}
|
||||
|
||||
// lex launches a new scanner and returns the channel of items.
|
||||
func lex(name, input string) chan item {
|
||||
l := &lexer{
|
||||
name: name,
|
||||
input: input,
|
||||
items: make(chan item),
|
||||
}
|
||||
go l.run()
|
||||
return l.items
|
||||
}
|
||||
|
||||
// state functions
|
||||
|
||||
const leftMeta = "{{"
|
||||
const rightMeta = "}}"
|
||||
|
||||
// lexText scans until a metacharacter
|
||||
func lexText(l *lexer) stateFn {
|
||||
for {
|
||||
if strings.HasPrefix(l.input[l.pos:], leftMeta) {
|
||||
if l.pos > l.start {
|
||||
l.emit(itemText)
|
||||
}
|
||||
return lexLeftMeta
|
||||
}
|
||||
if l.next() == eof {
|
||||
break
|
||||
}
|
||||
}
|
||||
// Correctly reached EOF.
|
||||
if l.pos > l.start {
|
||||
l.emit(itemText)
|
||||
}
|
||||
l.emit(itemEOF)
|
||||
return nil
|
||||
}
|
||||
|
||||
// leftMeta scans the left "metacharacter", which is known to be present.
|
||||
func lexLeftMeta(l *lexer) stateFn {
|
||||
l.pos += len(leftMeta)
|
||||
l.emit(itemLeftMeta)
|
||||
return lexInsideAction
|
||||
}
|
||||
|
||||
// rightMeta scans the right "metacharacter", which is known to be present.
|
||||
func lexRightMeta(l *lexer) stateFn {
|
||||
l.pos += len(rightMeta)
|
||||
l.emit(itemRightMeta)
|
||||
return lexText
|
||||
}
|
||||
|
||||
// lexInsideAction scans the elements inside "metacharacters".
|
||||
func lexInsideAction(l *lexer) stateFn {
|
||||
// Either number, quoted string, or identifier.
|
||||
// Spaces separate and are ignored.
|
||||
// Pipe symbols separate and are emitted.
|
||||
for {
|
||||
if strings.HasPrefix(l.input[l.pos:], rightMeta) {
|
||||
return lexRightMeta
|
||||
}
|
||||
switch r := l.next(); {
|
||||
case r == eof || r == '\n':
|
||||
return l.error("unclosed action")
|
||||
case isSpace(r):
|
||||
l.ignore()
|
||||
case r == '|':
|
||||
l.emit(itemPipe)
|
||||
case r == '"':
|
||||
return lexQuote
|
||||
case r == '`':
|
||||
return lexRawQuote
|
||||
case r == '+' || r == '-' || r == '.' || ('0' <= r && r <= '9'):
|
||||
l.backup()
|
||||
return lexNumber
|
||||
case isAlphaNumeric(r):
|
||||
l.backup()
|
||||
return lexIdentifier
|
||||
default:
|
||||
return l.error("unrecognized character in action: %#U", r)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// lexIdentifier scans an alphanumeric.
|
||||
func lexIdentifier(l *lexer) stateFn {
|
||||
Loop:
|
||||
for {
|
||||
switch r := l.next(); {
|
||||
case isAlphaNumeric(r):
|
||||
// absorb
|
||||
default:
|
||||
l.backup()
|
||||
l.emit(itemIdentifier)
|
||||
break Loop
|
||||
}
|
||||
}
|
||||
return lexInsideAction
|
||||
}
|
||||
|
||||
// lexNumber scans a number: decimal, octal, hex, float, or imaginary. This
|
||||
// isn't a perfect number scanner - for instance it accepts "." and "0x0.2"
|
||||
// and "089" - but when it's wrong the input is invalid and the parser (via
|
||||
// strconv) will notice.
|
||||
// TODO: without expressions you can do imaginary but not complex.
|
||||
func lexNumber(l *lexer) stateFn {
|
||||
// Optional leading sign.
|
||||
l.accept("+-")
|
||||
// Is it hex?
|
||||
digits := "0123456789"
|
||||
if l.accept("0") && l.accept("xX") {
|
||||
digits = "0123456789abcdefABCDEF"
|
||||
}
|
||||
l.acceptRun(digits)
|
||||
if l.accept(".") {
|
||||
l.acceptRun(digits)
|
||||
}
|
||||
if l.accept("eE") {
|
||||
l.accept("+-")
|
||||
l.acceptRun("0123456789")
|
||||
}
|
||||
// Is it imaginary?
|
||||
l.accept("i")
|
||||
// Next thing mustn't be alphanumeric.
|
||||
if isAlphaNumeric(l.peek()) {
|
||||
l.next()
|
||||
return l.error("bad number syntax: %q", l.input[l.start:l.pos])
|
||||
}
|
||||
l.emit(itemNumber)
|
||||
return lexInsideAction
|
||||
}
|
||||
|
||||
// lexQuote scans a quoted string.
|
||||
func lexQuote(l *lexer) stateFn {
|
||||
Loop:
|
||||
for {
|
||||
switch l.next() {
|
||||
case '\\':
|
||||
if r := l.next(); r != eof && r != '\n' {
|
||||
break
|
||||
}
|
||||
fallthrough
|
||||
case eof, '\n':
|
||||
return l.error("unterminated quoted string")
|
||||
case '"':
|
||||
break Loop
|
||||
}
|
||||
}
|
||||
l.emit(itemString)
|
||||
return lexInsideAction
|
||||
}
|
||||
|
||||
// lexRawQuote scans a raw quoted string.
|
||||
func lexRawQuote(l *lexer) stateFn {
|
||||
Loop:
|
||||
for {
|
||||
switch l.next() {
|
||||
case eof, '\n':
|
||||
return l.error("unterminated raw quoted string")
|
||||
case '`':
|
||||
break Loop
|
||||
}
|
||||
}
|
||||
l.emit(itemRawString)
|
||||
return lexInsideAction
|
||||
}
|
||||
|
||||
// isSpace reports whether r is a space character.
|
||||
func isSpace(r int) bool {
|
||||
switch r {
|
||||
case ' ', '\t', '\n', '\r':
|
||||
return true
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
// isAlphaNumeric reports whether r is an alphabetic, digit, or underscore.
|
||||
func isAlphaNumeric(r int) bool {
|
||||
return r == '_' || unicode.IsLetter(r) || unicode.IsDigit(r)
|
||||
}
|
126
src/pkg/exp/template/lex_test.go
Normal file
126
src/pkg/exp/template/lex_test.go
Normal file
@ -0,0 +1,126 @@
|
||||
// Copyright 2011 The Go Authors. All rights reserved.
|
||||
// Use of this source code is governed by a BSD-style
|
||||
// license that can be found in the LICENSE file.
|
||||
|
||||
package template
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"reflect"
|
||||
"testing"
|
||||
)
|
||||
|
||||
// Make the types prettyprint.
|
||||
var itemName = map[itemType]string{
|
||||
itemError: "Error",
|
||||
itemText: "Text",
|
||||
itemLeftMeta: "LeftMeta",
|
||||
itemRightMeta: "RightMeta",
|
||||
itemPipe: "Pipe",
|
||||
itemIdentifier: "Identifier",
|
||||
itemNumber: "Number",
|
||||
itemRawString: "RawString",
|
||||
itemString: "String",
|
||||
itemEOF: "EOF",
|
||||
}
|
||||
|
||||
func (i itemType) String() string {
|
||||
s := itemName[i]
|
||||
if s == "" {
|
||||
return fmt.Sprintf("item%d", int(i))
|
||||
}
|
||||
return s
|
||||
}
|
||||
|
||||
type lexTest struct {
|
||||
name string
|
||||
input string
|
||||
items []item
|
||||
}
|
||||
|
||||
var (
|
||||
tEOF = item{itemEOF, ""}
|
||||
tLeft = item{itemLeftMeta, "{{"}
|
||||
tRight = item{itemRightMeta, "}}"}
|
||||
tPipe = item{itemPipe, "|"}
|
||||
tFor = item{itemIdentifier, "for"}
|
||||
tQuote = item{itemString, `"abc \n\t\" "`}
|
||||
raw = "`" + `abc\n\t\" ` + "`"
|
||||
tRawQuote = item{itemRawString, raw}
|
||||
)
|
||||
|
||||
var lexTests = []lexTest{
|
||||
{"empty", "", []item{tEOF}},
|
||||
{"spaces", " \t\n", []item{{itemText, " \t\n"}, tEOF}},
|
||||
{"text", `now is the time`, []item{{itemText, "now is the time"}, tEOF}},
|
||||
{"empty action", `{{}}`, []item{tLeft, tRight, tEOF}},
|
||||
{"for", `{{for }}`, []item{tLeft, tFor, tRight, tEOF}},
|
||||
{"quote", `{{"abc \n\t\" "}}`, []item{tLeft, tQuote, tRight, tEOF}},
|
||||
{"raw quote", "{{" + raw + "}}", []item{tLeft, tRawQuote, tRight, tEOF}},
|
||||
{"numbers", "{{1 02 0x14 -7.2i 1e3 +1.2e-4}}", []item{
|
||||
tLeft,
|
||||
{itemNumber, "1"},
|
||||
{itemNumber, "02"},
|
||||
{itemNumber, "0x14"},
|
||||
{itemNumber, "-7.2i"},
|
||||
{itemNumber, "1e3"},
|
||||
{itemNumber, "+1.2e-4"},
|
||||
tRight,
|
||||
tEOF,
|
||||
}},
|
||||
{"pipeline", `intro {{echo hi 1.2 |noargs|args 1 "hi"}} outro`, []item{
|
||||
{itemText, "intro "},
|
||||
tLeft,
|
||||
{itemIdentifier, "echo"},
|
||||
{itemIdentifier, "hi"},
|
||||
{itemNumber, "1.2"},
|
||||
tPipe,
|
||||
{itemIdentifier, "noargs"},
|
||||
tPipe,
|
||||
{itemIdentifier, "args"},
|
||||
{itemNumber, "1"},
|
||||
{itemString, `"hi"`},
|
||||
tRight,
|
||||
{itemText, " outro"},
|
||||
tEOF,
|
||||
}},
|
||||
// errors
|
||||
{"badchar", "#{{#}}", []item{
|
||||
{itemText, "#"},
|
||||
tLeft,
|
||||
{itemError, "badchar:1 unrecognized character in action: U+0023 '#'"},
|
||||
}},
|
||||
{"unclosed action", "{{\n}}", []item{
|
||||
tLeft,
|
||||
{itemError, "unclosed action:2 unclosed action"},
|
||||
}},
|
||||
{"unclosed quote", "{{\"\n\"}}", []item{
|
||||
tLeft,
|
||||
{itemError, "unclosed quote:2 unterminated quoted string"},
|
||||
}},
|
||||
{"unclosed raw quote", "{{`xx\n`}}", []item{
|
||||
tLeft,
|
||||
{itemError, "unclosed raw quote:2 unterminated raw quoted string"},
|
||||
}},
|
||||
{"bad number", "{{3k}}", []item{
|
||||
tLeft,
|
||||
{itemError, `bad number:1 bad number syntax: "3k"`},
|
||||
}},
|
||||
}
|
||||
|
||||
// collect gathers the emitted items into a slice.
|
||||
func collect(t *lexTest) (items []item) {
|
||||
for i := range lex(t.name, t.input) {
|
||||
items = append(items, i)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func TestLex(t *testing.T) {
|
||||
for _, test := range lexTests {
|
||||
items := collect(&test)
|
||||
if !reflect.DeepEqual(items, test.items) {
|
||||
t.Errorf("%s: got\n\t%v; expected\n\t%v", test.name, items, test.items)
|
||||
}
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue
Block a user