mirror of
https://github.com/golang/go
synced 2024-11-18 20:54:40 -07:00
d87e9bbada
We need to limit the concurrent file reads to prevent `too many open file` errors. The first attempt to solve this was done in change 189437, but this change turned out to be incorrect. So this change reverts the changes made in change 189437 and instead adds an new semaphore around the `nativeFileHandle.Read` method. The limit is currently set to half of the lowest open file system limit of *nix, Windows and MacOS system (their respective limits are 1024, 16384, and 256). Change-Id: I27d741e3a3c36eb11bfa7457575f5d5831c06ad7 Reviewed-on: https://go-review.googlesource.com/c/tools/+/190417 Reviewed-by: Ian Cottrell <iancottrell@google.com> Run-TryBot: Ian Cottrell <iancottrell@google.com> TryBot-Result: Gobot Gobot <gobot@golang.org>
306 lines
7.4 KiB
Go
306 lines
7.4 KiB
Go
// Copyright 2019 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package cache
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"go/ast"
|
|
"go/parser"
|
|
"go/scanner"
|
|
"go/token"
|
|
|
|
"golang.org/x/tools/internal/lsp/source"
|
|
"golang.org/x/tools/internal/lsp/telemetry"
|
|
"golang.org/x/tools/internal/lsp/telemetry/log"
|
|
"golang.org/x/tools/internal/lsp/telemetry/trace"
|
|
"golang.org/x/tools/internal/memoize"
|
|
errors "golang.org/x/xerrors"
|
|
)
|
|
|
|
// Limits the number of parallel parser calls per process.
|
|
var parseLimit = make(chan struct{}, 20)
|
|
|
|
// parseKey uniquely identifies a parsed Go file.
|
|
type parseKey struct {
|
|
file source.FileIdentity
|
|
mode source.ParseMode
|
|
}
|
|
|
|
type parseGoHandle struct {
|
|
handle *memoize.Handle
|
|
file source.FileHandle
|
|
mode source.ParseMode
|
|
}
|
|
|
|
type parseGoData struct {
|
|
memoize.NoCopy
|
|
|
|
ast *ast.File
|
|
err error
|
|
}
|
|
|
|
func (c *cache) ParseGoHandle(fh source.FileHandle, mode source.ParseMode) source.ParseGoHandle {
|
|
key := parseKey{
|
|
file: fh.Identity(),
|
|
mode: mode,
|
|
}
|
|
h := c.store.Bind(key, func(ctx context.Context) interface{} {
|
|
data := &parseGoData{}
|
|
data.ast, data.err = parseGo(ctx, c, fh, mode)
|
|
return data
|
|
})
|
|
return &parseGoHandle{
|
|
handle: h,
|
|
file: fh,
|
|
mode: mode,
|
|
}
|
|
}
|
|
|
|
func (h *parseGoHandle) File() source.FileHandle {
|
|
return h.file
|
|
}
|
|
|
|
func (h *parseGoHandle) Mode() source.ParseMode {
|
|
return h.mode
|
|
}
|
|
|
|
func (h *parseGoHandle) Parse(ctx context.Context) (*ast.File, error) {
|
|
v := h.handle.Get(ctx)
|
|
if v == nil {
|
|
return nil, ctx.Err()
|
|
}
|
|
data := v.(*parseGoData)
|
|
return data.ast, data.err
|
|
}
|
|
|
|
func (h *parseGoHandle) Cached(ctx context.Context) (*ast.File, error) {
|
|
v := h.handle.Cached()
|
|
if v == nil {
|
|
return nil, errors.Errorf("no cached value for %s", h.file.Identity().URI)
|
|
}
|
|
data := v.(*parseGoData)
|
|
return data.ast, data.err
|
|
}
|
|
|
|
func hashParseKey(ph source.ParseGoHandle) string {
|
|
b := bytes.NewBuffer(nil)
|
|
b.WriteString(ph.File().Identity().String())
|
|
b.WriteString(string(ph.Mode()))
|
|
return hashContents(b.Bytes())
|
|
}
|
|
|
|
func hashParseKeys(phs []source.ParseGoHandle) string {
|
|
b := bytes.NewBuffer(nil)
|
|
for _, ph := range phs {
|
|
b.WriteString(hashParseKey(ph))
|
|
}
|
|
return hashContents(b.Bytes())
|
|
}
|
|
|
|
func parseGo(ctx context.Context, c *cache, fh source.FileHandle, mode source.ParseMode) (*ast.File, error) {
|
|
ctx, done := trace.StartSpan(ctx, "cache.parseGo", telemetry.File.Of(fh.Identity().URI.Filename()))
|
|
defer done()
|
|
|
|
buf, _, err := fh.Read(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
parseLimit <- struct{}{}
|
|
defer func() { <-parseLimit }()
|
|
parserMode := parser.AllErrors | parser.ParseComments
|
|
if mode == source.ParseHeader {
|
|
parserMode = parser.ImportsOnly | parser.ParseComments
|
|
}
|
|
ast, err := parser.ParseFile(c.fset, fh.Identity().URI.Filename(), buf, parserMode)
|
|
if ast != nil {
|
|
if mode == source.ParseExported {
|
|
trimAST(ast)
|
|
}
|
|
// Fix any badly parsed parts of the AST.
|
|
tok := c.fset.File(ast.Pos())
|
|
if err := fix(ctx, ast, tok, buf); err != nil {
|
|
log.Error(ctx, "failed to fix AST", err)
|
|
}
|
|
}
|
|
if ast == nil {
|
|
return nil, err
|
|
}
|
|
return ast, err
|
|
}
|
|
|
|
// trimAST clears any part of the AST not relevant to type checking
|
|
// expressions at pos.
|
|
func trimAST(file *ast.File) {
|
|
ast.Inspect(file, func(n ast.Node) bool {
|
|
if n == nil {
|
|
return false
|
|
}
|
|
switch n := n.(type) {
|
|
case *ast.FuncDecl:
|
|
n.Body = nil
|
|
case *ast.BlockStmt:
|
|
n.List = nil
|
|
case *ast.CaseClause:
|
|
n.Body = nil
|
|
case *ast.CommClause:
|
|
n.Body = nil
|
|
case *ast.CompositeLit:
|
|
// Leave elts in place for [...]T
|
|
// array literals, because they can
|
|
// affect the expression's type.
|
|
if !isEllipsisArray(n.Type) {
|
|
n.Elts = nil
|
|
}
|
|
}
|
|
return true
|
|
})
|
|
}
|
|
|
|
func isEllipsisArray(n ast.Expr) bool {
|
|
at, ok := n.(*ast.ArrayType)
|
|
if !ok {
|
|
return false
|
|
}
|
|
_, ok = at.Len.(*ast.Ellipsis)
|
|
return ok
|
|
}
|
|
|
|
// fix inspects the AST and potentially modifies any *ast.BadStmts so that it can be
|
|
// type-checked more effectively.
|
|
func fix(ctx context.Context, file *ast.File, tok *token.File, src []byte) error {
|
|
var parent ast.Node
|
|
var err error
|
|
ast.Inspect(file, func(n ast.Node) bool {
|
|
if n == nil {
|
|
return false
|
|
}
|
|
switch n := n.(type) {
|
|
case *ast.BadStmt:
|
|
err = parseDeferOrGoStmt(n, parent, tok, src) // don't shadow err
|
|
if err != nil {
|
|
err = errors.Errorf("unable to parse defer or go from *ast.BadStmt: %v", err)
|
|
}
|
|
return false
|
|
default:
|
|
parent = n
|
|
return true
|
|
}
|
|
})
|
|
return err
|
|
}
|
|
|
|
// parseDeferOrGoStmt tries to parse an *ast.BadStmt into a defer or a go statement.
|
|
//
|
|
// go/parser packages a statement of the form "defer x." as an *ast.BadStmt because
|
|
// it does not include a call expression. This means that go/types skips type-checking
|
|
// this statement entirely, and we can't use the type information when completing.
|
|
// Here, we try to generate a fake *ast.DeferStmt or *ast.GoStmt to put into the AST,
|
|
// instead of the *ast.BadStmt.
|
|
func parseDeferOrGoStmt(bad *ast.BadStmt, parent ast.Node, tok *token.File, src []byte) error {
|
|
// Check if we have a bad statement containing either a "go" or "defer".
|
|
s := &scanner.Scanner{}
|
|
s.Init(tok, src, nil, 0)
|
|
|
|
var pos token.Pos
|
|
var tkn token.Token
|
|
var lit string
|
|
for {
|
|
if tkn == token.EOF {
|
|
return errors.Errorf("reached the end of the file")
|
|
}
|
|
if pos >= bad.From {
|
|
break
|
|
}
|
|
pos, tkn, lit = s.Scan()
|
|
}
|
|
var stmt ast.Stmt
|
|
switch lit {
|
|
case "defer":
|
|
stmt = &ast.DeferStmt{
|
|
Defer: pos,
|
|
}
|
|
case "go":
|
|
stmt = &ast.GoStmt{
|
|
Go: pos,
|
|
}
|
|
default:
|
|
return errors.Errorf("no defer or go statement found")
|
|
}
|
|
|
|
// The expression after the "defer" or "go" starts at this position.
|
|
from, _, _ := s.Scan()
|
|
var to, curr token.Pos
|
|
FindTo:
|
|
for {
|
|
curr, tkn, _ = s.Scan()
|
|
// TODO(rstambler): This still needs more handling to work correctly.
|
|
// We encounter a specific issue with code that looks like this:
|
|
//
|
|
// defer x.<>
|
|
// y := 1
|
|
//
|
|
// In this scenario, we parse it as "defer x.y", which then fails to
|
|
// type-check, and we don't get completions as expected.
|
|
switch tkn {
|
|
case token.COMMENT, token.EOF, token.SEMICOLON, token.DEFINE:
|
|
break FindTo
|
|
}
|
|
// to is the end of expression that should become the Fun part of the call.
|
|
to = curr
|
|
}
|
|
if !from.IsValid() || tok.Offset(from) >= len(src) {
|
|
return errors.Errorf("invalid from position")
|
|
}
|
|
if !to.IsValid() || tok.Offset(to)+1 >= len(src) {
|
|
return errors.Errorf("invalid to position")
|
|
}
|
|
exprstr := string(src[tok.Offset(from) : tok.Offset(to)+1])
|
|
expr, err := parser.ParseExpr(exprstr)
|
|
if expr == nil {
|
|
return errors.Errorf("no expr in %s: %v", exprstr, err)
|
|
}
|
|
// parser.ParseExpr returns undefined positions.
|
|
// Adjust them for the current file.
|
|
offsetPositions(expr, from-1)
|
|
|
|
// Package the expression into a fake *ast.CallExpr and re-insert into the function.
|
|
call := &ast.CallExpr{
|
|
Fun: expr,
|
|
Lparen: to,
|
|
Rparen: to,
|
|
}
|
|
switch stmt := stmt.(type) {
|
|
case *ast.DeferStmt:
|
|
stmt.Call = call
|
|
case *ast.GoStmt:
|
|
stmt.Call = call
|
|
}
|
|
switch parent := parent.(type) {
|
|
case *ast.BlockStmt:
|
|
for i, s := range parent.List {
|
|
if s == bad {
|
|
parent.List[i] = stmt
|
|
break
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// offsetPositions applies an offset to the positions in an ast.Node.
|
|
// TODO(rstambler): Add more cases here as they become necessary.
|
|
func offsetPositions(expr ast.Expr, offset token.Pos) {
|
|
ast.Inspect(expr, func(n ast.Node) bool {
|
|
switch n := n.(type) {
|
|
case *ast.Ident:
|
|
n.NamePos += offset
|
|
return false
|
|
default:
|
|
return true
|
|
}
|
|
})
|
|
}
|