--- /dev/null
+// Copyright 2018 The Go Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+// Package checker defines the implementation of the checker commands.
+// The same code drives the multi-analysis driver, the single-analysis
+// driver that is conventionally provided for convenience along with
+// each analysis package, and the test driver.
+package checker
+
+import (
+ "bytes"
+ "encoding/gob"
+ "flag"
+ "fmt"
+ "go/format"
+ "go/parser"
+ "go/token"
+ "go/types"
+ "io/ioutil"
+ "log"
+ "os"
+ "reflect"
+ "runtime"
+ "runtime/pprof"
+ "runtime/trace"
+ "sort"
+ "strings"
+ "sync"
+ "time"
+
+ "golang.org/x/tools/go/analysis"
+ "golang.org/x/tools/go/analysis/internal/analysisflags"
+ "golang.org/x/tools/go/packages"
+ "golang.org/x/tools/internal/analysisinternal"
+ "golang.org/x/tools/internal/span"
+)
+
+var (
+ // Debug is a set of single-letter flags:
+ //
+ // f show [f]acts as they are created
+ // p disable [p]arallel execution of analyzers
+ // s do additional [s]anity checks on fact types and serialization
+ // t show [t]iming info (NB: use 'p' flag to avoid GC/scheduler noise)
+ // v show [v]erbose logging
+ //
+ Debug = ""
+
+ // Log files for optional performance tracing.
+ CPUProfile, MemProfile, Trace string
+
+ // Fix determines whether to apply all suggested fixes.
+ Fix bool
+)
+
+// RegisterFlags registers command-line flags used by the analysis driver.
+func RegisterFlags() {
+ // When adding flags here, remember to update
+ // the list of suppressed flags in analysisflags.
+
+ flag.StringVar(&Debug, "debug", Debug, `debug flags, any subset of "fpstv"`)
+
+ flag.StringVar(&CPUProfile, "cpuprofile", "", "write CPU profile to this file")
+ flag.StringVar(&MemProfile, "memprofile", "", "write memory profile to this file")
+ flag.StringVar(&Trace, "trace", "", "write trace log to this file")
+
+ flag.BoolVar(&Fix, "fix", false, "apply all suggested fixes")
+}
+
+// Run loads the packages specified by args using go/packages,
+// then applies the specified analyzers to them.
+// Analysis flags must already have been set.
+// It provides most of the logic for the main functions of both the
+// singlechecker and the multi-analysis commands.
+// It returns the appropriate exit code.
+func Run(args []string, analyzers []*analysis.Analyzer) (exitcode int) {
+ if CPUProfile != "" {
+ f, err := os.Create(CPUProfile)
+ if err != nil {
+ log.Fatal(err)
+ }
+ if err := pprof.StartCPUProfile(f); err != nil {
+ log.Fatal(err)
+ }
+ // NB: profile won't be written in case of error.
+ defer pprof.StopCPUProfile()
+ }
+
+ if Trace != "" {
+ f, err := os.Create(Trace)
+ if err != nil {
+ log.Fatal(err)
+ }
+ if err := trace.Start(f); err != nil {
+ log.Fatal(err)
+ }
+ // NB: trace log won't be written in case of error.
+ defer func() {
+ trace.Stop()
+ log.Printf("To view the trace, run:\n$ go tool trace view %s", Trace)
+ }()
+ }
+
+ if MemProfile != "" {
+ f, err := os.Create(MemProfile)
+ if err != nil {
+ log.Fatal(err)
+ }
+ // NB: memprofile won't be written in case of error.
+ defer func() {
+ runtime.GC() // get up-to-date statistics
+ if err := pprof.WriteHeapProfile(f); err != nil {
+ log.Fatalf("Writing memory profile: %v", err)
+ }
+ f.Close()
+ }()
+ }
+
+ // Load the packages.
+ if dbg('v') {
+ log.SetPrefix("")
+ log.SetFlags(log.Lmicroseconds) // display timing
+ log.Printf("load %s", args)
+ }
+
+ // Optimization: if the selected analyzers don't produce/consume
+ // facts, we need source only for the initial packages.
+ allSyntax := needFacts(analyzers)
+ initial, err := load(args, allSyntax)
+ if err != nil {
+ log.Print(err)
+ return 1 // load errors
+ }
+
+ // Print the results.
+ roots := analyze(initial, analyzers)
+
+ if Fix {
+ applyFixes(roots)
+ }
+
+ return printDiagnostics(roots)
+}
+
+// load loads the initial packages.
+func load(patterns []string, allSyntax bool) ([]*packages.Package, error) {
+ mode := packages.LoadSyntax
+ if allSyntax {
+ mode = packages.LoadAllSyntax
+ }
+ conf := packages.Config{
+ Mode: mode,
+ Tests: true,
+ }
+ initial, err := packages.Load(&conf, patterns...)
+ if err == nil {
+ if n := packages.PrintErrors(initial); n > 1 {
+ err = fmt.Errorf("%d errors during loading", n)
+ } else if n == 1 {
+ err = fmt.Errorf("error during loading")
+ } else if len(initial) == 0 {
+ err = fmt.Errorf("%s matched no packages", strings.Join(patterns, " "))
+ }
+ }
+
+ return initial, err
+}
+
+// TestAnalyzer applies an analysis to a set of packages (and their
+// dependencies if necessary) and returns the results.
+//
+// Facts about pkg are returned in a map keyed by object; package facts
+// have a nil key.
+//
+// This entry point is used only by analysistest.
+func TestAnalyzer(a *analysis.Analyzer, pkgs []*packages.Package) []*TestAnalyzerResult {
+ var results []*TestAnalyzerResult
+ for _, act := range analyze(pkgs, []*analysis.Analyzer{a}) {
+ facts := make(map[types.Object][]analysis.Fact)
+ for key, fact := range act.objectFacts {
+ if key.obj.Pkg() == act.pass.Pkg {
+ facts[key.obj] = append(facts[key.obj], fact)
+ }
+ }
+ for key, fact := range act.packageFacts {
+ if key.pkg == act.pass.Pkg {
+ facts[nil] = append(facts[nil], fact)
+ }
+ }
+
+ results = append(results, &TestAnalyzerResult{act.pass, act.diagnostics, facts, act.result, act.err})
+ }
+ return results
+}
+
+type TestAnalyzerResult struct {
+ Pass *analysis.Pass
+ Diagnostics []analysis.Diagnostic
+ Facts map[types.Object][]analysis.Fact
+ Result interface{}
+ Err error
+}
+
+func analyze(pkgs []*packages.Package, analyzers []*analysis.Analyzer) []*action {
+ // Construct the action graph.
+ if dbg('v') {
+ log.Printf("building graph of analysis passes")
+ }
+
+ // Each graph node (action) is one unit of analysis.
+ // Edges express package-to-package (vertical) dependencies,
+ // and analysis-to-analysis (horizontal) dependencies.
+ type key struct {
+ *analysis.Analyzer
+ *packages.Package
+ }
+ actions := make(map[key]*action)
+
+ var mkAction func(a *analysis.Analyzer, pkg *packages.Package) *action
+ mkAction = func(a *analysis.Analyzer, pkg *packages.Package) *action {
+ k := key{a, pkg}
+ act, ok := actions[k]
+ if !ok {
+ act = &action{a: a, pkg: pkg}
+
+ // Add a dependency on each required analyzers.
+ for _, req := range a.Requires {
+ act.deps = append(act.deps, mkAction(req, pkg))
+ }
+
+ // An analysis that consumes/produces facts
+ // must run on the package's dependencies too.
+ if len(a.FactTypes) > 0 {
+ paths := make([]string, 0, len(pkg.Imports))
+ for path := range pkg.Imports {
+ paths = append(paths, path)
+ }
+ sort.Strings(paths) // for determinism
+ for _, path := range paths {
+ dep := mkAction(a, pkg.Imports[path])
+ act.deps = append(act.deps, dep)
+ }
+ }
+
+ actions[k] = act
+ }
+ return act
+ }
+
+ // Build nodes for initial packages.
+ var roots []*action
+ for _, a := range analyzers {
+ for _, pkg := range pkgs {
+ root := mkAction(a, pkg)
+ root.isroot = true
+ roots = append(roots, root)
+ }
+ }
+
+ // Execute the graph in parallel.
+ execAll(roots)
+
+ return roots
+}
+
+func applyFixes(roots []*action) {
+ visited := make(map[*action]bool)
+ var apply func(*action) error
+ var visitAll func(actions []*action) error
+ visitAll = func(actions []*action) error {
+ for _, act := range actions {
+ if !visited[act] {
+ visited[act] = true
+ visitAll(act.deps)
+ if err := apply(act); err != nil {
+ return err
+ }
+ }
+ }
+ return nil
+ }
+
+ // TODO(matloob): Is this tree business too complicated? (After all this is Go!)
+ // Just create a set (map) of edits, sort by pos and call it a day?
+ type offsetedit struct {
+ start, end int
+ newText []byte
+ } // TextEdit using byteOffsets instead of pos
+ type node struct {
+ edit offsetedit
+ left, right *node
+ }
+
+ var insert func(tree **node, edit offsetedit) error
+ insert = func(treeptr **node, edit offsetedit) error {
+ if *treeptr == nil {
+ *treeptr = &node{edit, nil, nil}
+ return nil
+ }
+ tree := *treeptr
+ if edit.end <= tree.edit.start {
+ return insert(&tree.left, edit)
+ } else if edit.start >= tree.edit.end {
+ return insert(&tree.right, edit)
+ }
+
+ // Overlapping text edit.
+ return fmt.Errorf("analyses applying overlapping text edits affecting pos range (%v, %v) and (%v, %v)",
+ edit.start, edit.end, tree.edit.start, tree.edit.end)
+
+ }
+
+ editsForFile := make(map[*token.File]*node)
+
+ apply = func(act *action) error {
+ for _, diag := range act.diagnostics {
+ for _, sf := range diag.SuggestedFixes {
+ for _, edit := range sf.TextEdits {
+ // Validate the edit.
+ if edit.Pos > edit.End {
+ return fmt.Errorf(
+ "diagnostic for analysis %v contains Suggested Fix with malformed edit: pos (%v) > end (%v)",
+ act.a.Name, edit.Pos, edit.End)
+ }
+ file, endfile := act.pkg.Fset.File(edit.Pos), act.pkg.Fset.File(edit.End)
+ if file == nil || endfile == nil || file != endfile {
+ return (fmt.Errorf(
+ "diagnostic for analysis %v contains Suggested Fix with malformed spanning files %v and %v",
+ act.a.Name, file.Name(), endfile.Name()))
+ }
+ start, end := file.Offset(edit.Pos), file.Offset(edit.End)
+
+ // TODO(matloob): Validate that edits do not affect other packages.
+ root := editsForFile[file]
+ if err := insert(&root, offsetedit{start, end, edit.NewText}); err != nil {
+ return err
+ }
+ editsForFile[file] = root // In case the root changed
+ }
+ }
+ }
+ return nil
+ }
+
+ visitAll(roots)
+
+ fset := token.NewFileSet() // Shared by parse calls below
+ // Now we've got a set of valid edits for each file. Get the new file contents.
+ for f, tree := range editsForFile {
+ contents, err := ioutil.ReadFile(f.Name())
+ if err != nil {
+ log.Fatal(err)
+ }
+
+ cur := 0 // current position in the file
+
+ var out bytes.Buffer
+
+ var recurse func(*node)
+ recurse = func(node *node) {
+ if node.left != nil {
+ recurse(node.left)
+ }
+
+ edit := node.edit
+ if edit.start > cur {
+ out.Write(contents[cur:edit.start])
+ out.Write(edit.newText)
+ }
+ cur = edit.end
+
+ if node.right != nil {
+ recurse(node.right)
+ }
+ }
+ recurse(tree)
+ // Write out the rest of the file.
+ if cur < len(contents) {
+ out.Write(contents[cur:])
+ }
+
+ // Try to format the file.
+ ff, err := parser.ParseFile(fset, f.Name(), out.Bytes(), parser.ParseComments)
+ if err == nil {
+ var buf bytes.Buffer
+ if err = format.Node(&buf, fset, ff); err == nil {
+ out = buf
+ }
+ }
+
+ ioutil.WriteFile(f.Name(), out.Bytes(), 0644)
+ }
+}
+
+// printDiagnostics prints the diagnostics for the root packages in either
+// plain text or JSON format. JSON format also includes errors for any
+// dependencies.
+//
+// It returns the exitcode: in plain mode, 0 for success, 1 for analysis
+// errors, and 3 for diagnostics. We avoid 2 since the flag package uses
+// it. JSON mode always succeeds at printing errors and diagnostics in a
+// structured form to stdout.
+func printDiagnostics(roots []*action) (exitcode int) {
+ // Print the output.
+ //
+ // Print diagnostics only for root packages,
+ // but errors for all packages.
+ printed := make(map[*action]bool)
+ var print func(*action)
+ var visitAll func(actions []*action)
+ visitAll = func(actions []*action) {
+ for _, act := range actions {
+ if !printed[act] {
+ printed[act] = true
+ visitAll(act.deps)
+ print(act)
+ }
+ }
+ }
+
+ if analysisflags.JSON {
+ // JSON output
+ tree := make(analysisflags.JSONTree)
+ print = func(act *action) {
+ var diags []analysis.Diagnostic
+ if act.isroot {
+ diags = act.diagnostics
+ }
+ tree.Add(act.pkg.Fset, act.pkg.ID, act.a.Name, diags, act.err)
+ }
+ visitAll(roots)
+ tree.Print()
+ } else {
+ // plain text output
+
+ // De-duplicate diagnostics by position (not token.Pos) to
+ // avoid double-reporting in source files that belong to
+ // multiple packages, such as foo and foo.test.
+ type key struct {
+ pos token.Position
+ end token.Position
+ *analysis.Analyzer
+ message string
+ }
+ seen := make(map[key]bool)
+
+ print = func(act *action) {
+ if act.err != nil {
+ fmt.Fprintf(os.Stderr, "%s: %v\n", act.a.Name, act.err)
+ exitcode = 1 // analysis failed, at least partially
+ return
+ }
+ if act.isroot {
+ for _, diag := range act.diagnostics {
+ // We don't display a.Name/f.Category
+ // as most users don't care.
+
+ posn := act.pkg.Fset.Position(diag.Pos)
+ end := act.pkg.Fset.Position(diag.End)
+ k := key{posn, end, act.a, diag.Message}
+ if seen[k] {
+ continue // duplicate
+ }
+ seen[k] = true
+
+ analysisflags.PrintPlain(act.pkg.Fset, diag)
+ }
+ }
+ }
+ visitAll(roots)
+
+ if exitcode == 0 && len(seen) > 0 {
+ exitcode = 3 // successfully produced diagnostics
+ }
+ }
+
+ // Print timing info.
+ if dbg('t') {
+ if !dbg('p') {
+ log.Println("Warning: times are mostly GC/scheduler noise; use -debug=tp to disable parallelism")
+ }
+ var all []*action
+ var total time.Duration
+ for act := range printed {
+ all = append(all, act)
+ total += act.duration
+ }
+ sort.Slice(all, func(i, j int) bool {
+ return all[i].duration > all[j].duration
+ })
+
+ // Print actions accounting for 90% of the total.
+ var sum time.Duration
+ for _, act := range all {
+ fmt.Fprintf(os.Stderr, "%s\t%s\n", act.duration, act)
+ sum += act.duration
+ if sum >= total*9/10 {
+ break
+ }
+ }
+ }
+
+ return exitcode
+}
+
+// needFacts reports whether any analysis required by the specified set
+// needs facts. If so, we must load the entire program from source.
+func needFacts(analyzers []*analysis.Analyzer) bool {
+ seen := make(map[*analysis.Analyzer]bool)
+ var q []*analysis.Analyzer // for BFS
+ q = append(q, analyzers...)
+ for len(q) > 0 {
+ a := q[0]
+ q = q[1:]
+ if !seen[a] {
+ seen[a] = true
+ if len(a.FactTypes) > 0 {
+ return true
+ }
+ q = append(q, a.Requires...)
+ }
+ }
+ return false
+}
+
+// An action represents one unit of analysis work: the application of
+// one analysis to one package. Actions form a DAG, both within a
+// package (as different analyzers are applied, either in sequence or
+// parallel), and across packages (as dependencies are analyzed).
+type action struct {
+ once sync.Once
+ a *analysis.Analyzer
+ pkg *packages.Package
+ pass *analysis.Pass
+ isroot bool
+ deps []*action
+ objectFacts map[objectFactKey]analysis.Fact
+ packageFacts map[packageFactKey]analysis.Fact
+ inputs map[*analysis.Analyzer]interface{}
+ result interface{}
+ diagnostics []analysis.Diagnostic
+ err error
+ duration time.Duration
+}
+
+type objectFactKey struct {
+ obj types.Object
+ typ reflect.Type
+}
+
+type packageFactKey struct {
+ pkg *types.Package
+ typ reflect.Type
+}
+
+func (act *action) String() string {
+ return fmt.Sprintf("%s@%s", act.a, act.pkg)
+}
+
+func execAll(actions []*action) {
+ sequential := dbg('p')
+ var wg sync.WaitGroup
+ for _, act := range actions {
+ wg.Add(1)
+ work := func(act *action) {
+ act.exec()
+ wg.Done()
+ }
+ if sequential {
+ work(act)
+ } else {
+ go work(act)
+ }
+ }
+ wg.Wait()
+}
+
+func (act *action) exec() { act.once.Do(act.execOnce) }
+
+func (act *action) execOnce() {
+ // Analyze dependencies.
+ execAll(act.deps)
+
+ // TODO(adonovan): uncomment this during profiling.
+ // It won't build pre-go1.11 but conditional compilation
+ // using build tags isn't warranted.
+ //
+ // ctx, task := trace.NewTask(context.Background(), "exec")
+ // trace.Log(ctx, "pass", act.String())
+ // defer task.End()
+
+ // Record time spent in this node but not its dependencies.
+ // In parallel mode, due to GC/scheduler contention, the
+ // time is 5x higher than in sequential mode, even with a
+ // semaphore limiting the number of threads here.
+ // So use -debug=tp.
+ if dbg('t') {
+ t0 := time.Now()
+ defer func() { act.duration = time.Since(t0) }()
+ }
+
+ // Report an error if any dependency failed.
+ var failed []string
+ for _, dep := range act.deps {
+ if dep.err != nil {
+ failed = append(failed, dep.String())
+ }
+ }
+ if failed != nil {
+ sort.Strings(failed)
+ act.err = fmt.Errorf("failed prerequisites: %s", strings.Join(failed, ", "))
+ return
+ }
+
+ // Plumb the output values of the dependencies
+ // into the inputs of this action. Also facts.
+ inputs := make(map[*analysis.Analyzer]interface{})
+ act.objectFacts = make(map[objectFactKey]analysis.Fact)
+ act.packageFacts = make(map[packageFactKey]analysis.Fact)
+ for _, dep := range act.deps {
+ if dep.pkg == act.pkg {
+ // Same package, different analysis (horizontal edge):
+ // in-memory outputs of prerequisite analyzers
+ // become inputs to this analysis pass.
+ inputs[dep.a] = dep.result
+
+ } else if dep.a == act.a { // (always true)
+ // Same analysis, different package (vertical edge):
+ // serialized facts produced by prerequisite analysis
+ // become available to this analysis pass.
+ inheritFacts(act, dep)
+ }
+ }
+
+ // Run the analysis.
+ pass := &analysis.Pass{
+ Analyzer: act.a,
+ Fset: act.pkg.Fset,
+ Files: act.pkg.Syntax,
+ OtherFiles: act.pkg.OtherFiles,
+ IgnoredFiles: act.pkg.IgnoredFiles,
+ Pkg: act.pkg.Types,
+ TypesInfo: act.pkg.TypesInfo,
+ TypesSizes: act.pkg.TypesSizes,
+ ResultOf: inputs,
+ Report: func(d analysis.Diagnostic) { act.diagnostics = append(act.diagnostics, d) },
+ ImportObjectFact: act.importObjectFact,
+ ExportObjectFact: act.exportObjectFact,
+ ImportPackageFact: act.importPackageFact,
+ ExportPackageFact: act.exportPackageFact,
+ AllObjectFacts: act.allObjectFacts,
+ AllPackageFacts: act.allPackageFacts,
+ }
+ act.pass = pass
+
+ var errors []types.Error
+ // Get any type errors that are attributed to the pkg.
+ // This is necessary to test analyzers that provide
+ // suggested fixes for compiler/type errors.
+ for _, err := range act.pkg.Errors {
+ if err.Kind != packages.TypeError {
+ continue
+ }
+ // err.Pos is a string of form: "file:line:col" or "file:line" or "" or "-"
+ spn := span.Parse(err.Pos)
+ // Extract the token positions from the error string.
+ line, col, offset := spn.Start().Line(), spn.Start().Column(), -1
+ act.pkg.Fset.Iterate(func(f *token.File) bool {
+ if f.Name() != spn.URI().Filename() {
+ return true
+ }
+ offset = int(f.LineStart(line)) + col - 1
+ return false
+ })
+ if offset == -1 {
+ continue
+ }
+ errors = append(errors, types.Error{
+ Fset: act.pkg.Fset,
+ Msg: err.Msg,
+ Pos: token.Pos(offset),
+ })
+ }
+ analysisinternal.SetTypeErrors(pass, errors)
+
+ var err error
+ if act.pkg.IllTyped && !pass.Analyzer.RunDespiteErrors {
+ err = fmt.Errorf("analysis skipped due to errors in package")
+ } else {
+ act.result, err = pass.Analyzer.Run(pass)
+ if err == nil {
+ if got, want := reflect.TypeOf(act.result), pass.Analyzer.ResultType; got != want {
+ err = fmt.Errorf(
+ "internal error: on package %s, analyzer %s returned a result of type %v, but declared ResultType %v",
+ pass.Pkg.Path(), pass.Analyzer, got, want)
+ }
+ }
+ }
+ act.err = err
+
+ // disallow calls after Run
+ pass.ExportObjectFact = nil
+ pass.ExportPackageFact = nil
+}
+
+// inheritFacts populates act.facts with
+// those it obtains from its dependency, dep.
+func inheritFacts(act, dep *action) {
+ serialize := dbg('s')
+
+ for key, fact := range dep.objectFacts {
+ // Filter out facts related to objects
+ // that are irrelevant downstream
+ // (equivalently: not in the compiler export data).
+ if !exportedFrom(key.obj, dep.pkg.Types) {
+ if false {
+ log.Printf("%v: discarding %T fact from %s for %s: %s", act, fact, dep, key.obj, fact)
+ }
+ continue
+ }
+
+ // Optionally serialize/deserialize fact
+ // to verify that it works across address spaces.
+ if serialize {
+ encodedFact, err := codeFact(fact)
+ if err != nil {
+ log.Panicf("internal error: encoding of %T fact failed in %v", fact, act)
+ }
+ fact = encodedFact
+ }
+
+ if false {
+ log.Printf("%v: inherited %T fact for %s: %s", act, fact, key.obj, fact)
+ }
+ act.objectFacts[key] = fact
+ }
+
+ for key, fact := range dep.packageFacts {
+ // TODO: filter out facts that belong to
+ // packages not mentioned in the export data
+ // to prevent side channels.
+
+ // Optionally serialize/deserialize fact
+ // to verify that it works across address spaces
+ // and is deterministic.
+ if serialize {
+ encodedFact, err := codeFact(fact)
+ if err != nil {
+ log.Panicf("internal error: encoding of %T fact failed in %v", fact, act)
+ }
+ fact = encodedFact
+ }
+
+ if false {
+ log.Printf("%v: inherited %T fact for %s: %s", act, fact, key.pkg.Path(), fact)
+ }
+ act.packageFacts[key] = fact
+ }
+}
+
+// codeFact encodes then decodes a fact,
+// just to exercise that logic.
+func codeFact(fact analysis.Fact) (analysis.Fact, error) {
+ // We encode facts one at a time.
+ // A real modular driver would emit all facts
+ // into one encoder to improve gob efficiency.
+ var buf bytes.Buffer
+ if err := gob.NewEncoder(&buf).Encode(fact); err != nil {
+ return nil, err
+ }
+
+ // Encode it twice and assert that we get the same bits.
+ // This helps detect nondeterministic Gob encoding (e.g. of maps).
+ var buf2 bytes.Buffer
+ if err := gob.NewEncoder(&buf2).Encode(fact); err != nil {
+ return nil, err
+ }
+ if !bytes.Equal(buf.Bytes(), buf2.Bytes()) {
+ return nil, fmt.Errorf("encoding of %T fact is nondeterministic", fact)
+ }
+
+ new := reflect.New(reflect.TypeOf(fact).Elem()).Interface().(analysis.Fact)
+ if err := gob.NewDecoder(&buf).Decode(new); err != nil {
+ return nil, err
+ }
+ return new, nil
+}
+
+// exportedFrom reports whether obj may be visible to a package that imports pkg.
+// This includes not just the exported members of pkg, but also unexported
+// constants, types, fields, and methods, perhaps belonging to oether packages,
+// that find there way into the API.
+// This is an overapproximation of the more accurate approach used by
+// gc export data, which walks the type graph, but it's much simpler.
+//
+// TODO(adonovan): do more accurate filtering by walking the type graph.
+func exportedFrom(obj types.Object, pkg *types.Package) bool {
+ switch obj := obj.(type) {
+ case *types.Func:
+ return obj.Exported() && obj.Pkg() == pkg ||
+ obj.Type().(*types.Signature).Recv() != nil
+ case *types.Var:
+ if obj.IsField() {
+ return true
+ }
+ // we can't filter more aggressively than this because we need
+ // to consider function parameters exported, but have no way
+ // of telling apart function parameters from local variables.
+ return obj.Pkg() == pkg
+ case *types.TypeName, *types.Const:
+ return true
+ }
+ return false // Nil, Builtin, Label, or PkgName
+}
+
+// importObjectFact implements Pass.ImportObjectFact.
+// Given a non-nil pointer ptr of type *T, where *T satisfies Fact,
+// importObjectFact copies the fact value to *ptr.
+func (act *action) importObjectFact(obj types.Object, ptr analysis.Fact) bool {
+ if obj == nil {
+ panic("nil object")
+ }
+ key := objectFactKey{obj, factType(ptr)}
+ if v, ok := act.objectFacts[key]; ok {
+ reflect.ValueOf(ptr).Elem().Set(reflect.ValueOf(v).Elem())
+ return true
+ }
+ return false
+}
+
+// exportObjectFact implements Pass.ExportObjectFact.
+func (act *action) exportObjectFact(obj types.Object, fact analysis.Fact) {
+ if act.pass.ExportObjectFact == nil {
+ log.Panicf("%s: Pass.ExportObjectFact(%s, %T) called after Run", act, obj, fact)
+ }
+
+ if obj.Pkg() != act.pkg.Types {
+ log.Panicf("internal error: in analysis %s of package %s: Fact.Set(%s, %T): can't set facts on objects belonging another package",
+ act.a, act.pkg, obj, fact)
+ }
+
+ key := objectFactKey{obj, factType(fact)}
+ act.objectFacts[key] = fact // clobber any existing entry
+ if dbg('f') {
+ objstr := types.ObjectString(obj, (*types.Package).Name)
+ fmt.Fprintf(os.Stderr, "%s: object %s has fact %s\n",
+ act.pkg.Fset.Position(obj.Pos()), objstr, fact)
+ }
+}
+
+// allObjectFacts implements Pass.AllObjectFacts.
+func (act *action) allObjectFacts() []analysis.ObjectFact {
+ facts := make([]analysis.ObjectFact, 0, len(act.objectFacts))
+ for k := range act.objectFacts {
+ facts = append(facts, analysis.ObjectFact{k.obj, act.objectFacts[k]})
+ }
+ return facts
+}
+
+// importPackageFact implements Pass.ImportPackageFact.
+// Given a non-nil pointer ptr of type *T, where *T satisfies Fact,
+// fact copies the fact value to *ptr.
+func (act *action) importPackageFact(pkg *types.Package, ptr analysis.Fact) bool {
+ if pkg == nil {
+ panic("nil package")
+ }
+ key := packageFactKey{pkg, factType(ptr)}
+ if v, ok := act.packageFacts[key]; ok {
+ reflect.ValueOf(ptr).Elem().Set(reflect.ValueOf(v).Elem())
+ return true
+ }
+ return false
+}
+
+// exportPackageFact implements Pass.ExportPackageFact.
+func (act *action) exportPackageFact(fact analysis.Fact) {
+ if act.pass.ExportPackageFact == nil {
+ log.Panicf("%s: Pass.ExportPackageFact(%T) called after Run", act, fact)
+ }
+
+ key := packageFactKey{act.pass.Pkg, factType(fact)}
+ act.packageFacts[key] = fact // clobber any existing entry
+ if dbg('f') {
+ fmt.Fprintf(os.Stderr, "%s: package %s has fact %s\n",
+ act.pkg.Fset.Position(act.pass.Files[0].Pos()), act.pass.Pkg.Path(), fact)
+ }
+}
+
+func factType(fact analysis.Fact) reflect.Type {
+ t := reflect.TypeOf(fact)
+ if t.Kind() != reflect.Ptr {
+ log.Fatalf("invalid Fact type: got %T, want pointer", t)
+ }
+ return t
+}
+
+// allObjectFacts implements Pass.AllObjectFacts.
+func (act *action) allPackageFacts() []analysis.PackageFact {
+ facts := make([]analysis.PackageFact, 0, len(act.packageFacts))
+ for k := range act.packageFacts {
+ facts = append(facts, analysis.PackageFact{k.pkg, act.packageFacts[k]})
+ }
+ return facts
+}
+
+func dbg(b byte) bool { return strings.IndexByte(Debug, b) >= 0 }