mirror of
https://github.com/nspcc-dev/neo-go.git
synced 2024-11-27 03:58:06 +00:00
2124 lines
56 KiB
Go
2124 lines
56 KiB
Go
package compiler
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"go/ast"
|
|
"go/constant"
|
|
"go/token"
|
|
"go/types"
|
|
"math"
|
|
"sort"
|
|
"strings"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/encoding/address"
|
|
"github.com/nspcc-dev/neo-go/pkg/io"
|
|
"github.com/nspcc-dev/neo-go/pkg/vm"
|
|
"github.com/nspcc-dev/neo-go/pkg/vm/emit"
|
|
"github.com/nspcc-dev/neo-go/pkg/vm/opcode"
|
|
"github.com/nspcc-dev/neo-go/pkg/vm/stackitem"
|
|
"golang.org/x/tools/go/loader"
|
|
)
|
|
|
|
type codegen struct {
|
|
// Information about the program with all its dependencies.
|
|
buildInfo *buildInfo
|
|
|
|
// prog holds the output buffer.
|
|
prog *io.BufBinWriter
|
|
|
|
// Type information.
|
|
typeInfo *types.Info
|
|
|
|
// A mapping of func identifiers with their scope.
|
|
funcs map[string]*funcScope
|
|
|
|
// A mapping of lambda functions into their scope.
|
|
lambda map[string]*funcScope
|
|
|
|
// Current funcScope being converted.
|
|
scope *funcScope
|
|
|
|
globals map[string]int
|
|
|
|
// A mapping from label's names to their ids.
|
|
labels map[labelWithType]uint16
|
|
// A list of nested label names together with evaluation stack depth.
|
|
labelList []labelWithStackSize
|
|
|
|
// A label for the for-loop being currently visited.
|
|
currentFor string
|
|
// A label for the switch statement being visited.
|
|
currentSwitch string
|
|
// A label to be used in the next statement.
|
|
nextLabel string
|
|
|
|
// sequencePoints is mapping from method name to a slice
|
|
// containing info about mapping from opcode's offset
|
|
// to a text span in the source file.
|
|
sequencePoints map[string][]DebugSeqPoint
|
|
|
|
// initEndOffset specifies the end of the initialization method.
|
|
initEndOffset int
|
|
// deployEndOffset specifies the end of the deployment method.
|
|
deployEndOffset int
|
|
|
|
// importMap contains mapping from package aliases to full package names for the current file.
|
|
importMap map[string]string
|
|
|
|
// constMap contains constants from foreign packages.
|
|
constMap map[string]types.TypeAndValue
|
|
|
|
// currPkg is current package being processed.
|
|
currPkg *types.Package
|
|
|
|
// mainPkg is a main package metadata.
|
|
mainPkg *loader.PackageInfo
|
|
|
|
// packages contains packages in the order they were loaded.
|
|
packages []string
|
|
|
|
// exceptionIndex is the index of static slot where exception is stored.
|
|
exceptionIndex int
|
|
|
|
// documents contains paths to all files used by the program.
|
|
documents []string
|
|
// docIndex maps file path to an index in documents array.
|
|
docIndex map[string]int
|
|
|
|
// Label table for recording jump destinations.
|
|
l []int
|
|
}
|
|
|
|
type labelOffsetType byte
|
|
|
|
const (
|
|
labelStart labelOffsetType = iota // labelStart is a default label type
|
|
labelEnd // labelEnd is a type for labels that are targets for break
|
|
labelPost // labelPost is a type for labels that are targets for continue
|
|
)
|
|
|
|
type labelWithType struct {
|
|
name string
|
|
typ labelOffsetType
|
|
}
|
|
|
|
type labelWithStackSize struct {
|
|
name string
|
|
sz int
|
|
}
|
|
|
|
type varType int
|
|
|
|
const (
|
|
varGlobal varType = iota
|
|
varLocal
|
|
varArgument
|
|
)
|
|
|
|
// newLabel creates a new label to jump to
|
|
func (c *codegen) newLabel() (l uint16) {
|
|
li := len(c.l)
|
|
if li > math.MaxUint16 {
|
|
c.prog.Err = errors.New("label number is too big")
|
|
return
|
|
}
|
|
l = uint16(li)
|
|
c.l = append(c.l, -1)
|
|
return
|
|
}
|
|
|
|
// newNamedLabel creates a new label with a specified name.
|
|
func (c *codegen) newNamedLabel(typ labelOffsetType, name string) (l uint16) {
|
|
l = c.newLabel()
|
|
lt := labelWithType{name: name, typ: typ}
|
|
c.labels[lt] = l
|
|
return
|
|
}
|
|
|
|
func (c *codegen) setLabel(l uint16) {
|
|
c.l[l] = c.pc() + 1
|
|
}
|
|
|
|
// pc returns the program offset off the last instruction.
|
|
func (c *codegen) pc() int {
|
|
return c.prog.Len() - 1
|
|
}
|
|
|
|
func (c *codegen) emitLoadConst(t types.TypeAndValue) {
|
|
if c.prog.Err != nil {
|
|
return
|
|
}
|
|
|
|
typ, ok := t.Type.Underlying().(*types.Basic)
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("compiler doesn't know how to convert this constant: %v", t)
|
|
return
|
|
}
|
|
|
|
switch typ.Kind() {
|
|
case types.Int, types.UntypedInt, types.Uint,
|
|
types.Int8, types.Uint8,
|
|
types.Int16, types.Uint16,
|
|
types.Int32, types.Uint32, types.Int64, types.Uint64:
|
|
val, _ := constant.Int64Val(t.Value)
|
|
emit.Int(c.prog.BinWriter, val)
|
|
case types.String, types.UntypedString:
|
|
val := constant.StringVal(t.Value)
|
|
emit.String(c.prog.BinWriter, val)
|
|
case types.Bool, types.UntypedBool:
|
|
val := constant.BoolVal(t.Value)
|
|
emit.Bool(c.prog.BinWriter, val)
|
|
default:
|
|
c.prog.Err = fmt.Errorf("compiler doesn't know how to convert this basic type: %v", t)
|
|
return
|
|
}
|
|
}
|
|
|
|
func (c *codegen) emitLoadField(i int) {
|
|
emit.Int(c.prog.BinWriter, int64(i))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PICKITEM)
|
|
}
|
|
|
|
func (c *codegen) emitStoreStructField(i int) {
|
|
emit.Int(c.prog.BinWriter, int64(i))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.ROT, opcode.SETITEM)
|
|
}
|
|
|
|
// getVarIndex returns variable type and position in corresponding slot,
|
|
// according to current scope.
|
|
func (c *codegen) getVarIndex(pkg string, name string) (varType, int) {
|
|
if pkg == "" {
|
|
if c.scope != nil {
|
|
vt, val := c.scope.vars.getVarIndex(name)
|
|
if val >= 0 {
|
|
return vt, val
|
|
}
|
|
}
|
|
}
|
|
if i, ok := c.globals[c.getIdentName(pkg, name)]; ok {
|
|
return varGlobal, i
|
|
}
|
|
|
|
return varLocal, c.scope.newVariable(varLocal, name)
|
|
}
|
|
|
|
func getBaseOpcode(t varType) (opcode.Opcode, opcode.Opcode) {
|
|
switch t {
|
|
case varGlobal:
|
|
return opcode.LDSFLD0, opcode.STSFLD0
|
|
case varLocal:
|
|
return opcode.LDLOC0, opcode.STLOC0
|
|
case varArgument:
|
|
return opcode.LDARG0, opcode.STARG0
|
|
default:
|
|
panic("invalid type")
|
|
}
|
|
}
|
|
|
|
// emitLoadVar loads specified variable to the evaluation stack.
|
|
func (c *codegen) emitLoadVar(pkg string, name string) {
|
|
t, i := c.getVarIndex(pkg, name)
|
|
c.emitLoadByIndex(t, i)
|
|
}
|
|
|
|
// emitLoadByIndex loads specified variable type with index i.
|
|
func (c *codegen) emitLoadByIndex(t varType, i int) {
|
|
base, _ := getBaseOpcode(t)
|
|
if i < 7 {
|
|
emit.Opcodes(c.prog.BinWriter, base+opcode.Opcode(i))
|
|
} else {
|
|
emit.Instruction(c.prog.BinWriter, base+7, []byte{byte(i)})
|
|
}
|
|
}
|
|
|
|
// emitStoreVar stores top value from the evaluation stack in the specified variable.
|
|
func (c *codegen) emitStoreVar(pkg string, name string) {
|
|
if name == "_" {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP)
|
|
return
|
|
}
|
|
t, i := c.getVarIndex(pkg, name)
|
|
c.emitStoreByIndex(t, i)
|
|
}
|
|
|
|
// emitLoadByIndex stores top value in the specified variable type with index i.
|
|
func (c *codegen) emitStoreByIndex(t varType, i int) {
|
|
_, base := getBaseOpcode(t)
|
|
if i < 7 {
|
|
emit.Opcodes(c.prog.BinWriter, base+opcode.Opcode(i))
|
|
} else {
|
|
emit.Instruction(c.prog.BinWriter, base+7, []byte{byte(i)})
|
|
}
|
|
}
|
|
|
|
func (c *codegen) emitDefault(t types.Type) {
|
|
switch t := t.Underlying().(type) {
|
|
case *types.Basic:
|
|
info := t.Info()
|
|
switch {
|
|
case info&types.IsInteger != 0:
|
|
emit.Int(c.prog.BinWriter, 0)
|
|
case info&types.IsString != 0:
|
|
emit.Bytes(c.prog.BinWriter, []byte{})
|
|
case info&types.IsBoolean != 0:
|
|
emit.Bool(c.prog.BinWriter, false)
|
|
default:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSHNULL)
|
|
}
|
|
case *types.Struct:
|
|
num := t.NumFields()
|
|
emit.Int(c.prog.BinWriter, int64(num))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWSTRUCT)
|
|
for i := 0; i < num; i++ {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
emit.Int(c.prog.BinWriter, int64(i))
|
|
c.emitDefault(t.Field(i).Type())
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SETITEM)
|
|
}
|
|
default:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSHNULL)
|
|
}
|
|
}
|
|
|
|
// convertGlobals traverses the AST and only converts global declarations.
|
|
// If we call this in convertFuncDecl then it will load all global variables
|
|
// into the scope of the function.
|
|
func (c *codegen) convertGlobals(f *ast.File, _ *types.Package) {
|
|
ast.Inspect(f, func(node ast.Node) bool {
|
|
switch n := node.(type) {
|
|
case *ast.FuncDecl:
|
|
return false
|
|
case *ast.GenDecl:
|
|
ast.Walk(c, n)
|
|
}
|
|
return true
|
|
})
|
|
}
|
|
|
|
func isInitFunc(decl *ast.FuncDecl) bool {
|
|
return decl.Name.Name == "init" && decl.Recv == nil &&
|
|
decl.Type.Params.NumFields() == 0 &&
|
|
decl.Type.Results.NumFields() == 0
|
|
}
|
|
|
|
func (c *codegen) clearSlots(n int) {
|
|
for i := 0; i < n; i++ {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSHNULL)
|
|
c.emitStoreByIndex(varLocal, i)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) convertInitFuncs(f *ast.File, pkg *types.Package, seenBefore bool) bool {
|
|
ast.Inspect(f, func(node ast.Node) bool {
|
|
switch n := node.(type) {
|
|
case *ast.FuncDecl:
|
|
if isInitFunc(n) {
|
|
if seenBefore {
|
|
cnt, _ := countLocals(n)
|
|
c.clearSlots(cnt)
|
|
seenBefore = true
|
|
}
|
|
c.convertFuncDecl(f, n, pkg)
|
|
}
|
|
case *ast.GenDecl:
|
|
return false
|
|
}
|
|
return true
|
|
})
|
|
return seenBefore
|
|
}
|
|
|
|
func isDeployFunc(decl *ast.FuncDecl) bool {
|
|
if decl.Name.Name != "_deploy" || decl.Recv != nil ||
|
|
decl.Type.Params.NumFields() != 1 ||
|
|
decl.Type.Results.NumFields() != 0 {
|
|
return false
|
|
}
|
|
typ, ok := decl.Type.Params.List[0].Type.(*ast.Ident)
|
|
return ok && typ.Name == "bool"
|
|
}
|
|
|
|
func (c *codegen) convertDeployFuncs() {
|
|
seenBefore := false
|
|
c.ForEachFile(func(f *ast.File, pkg *types.Package) {
|
|
ast.Inspect(f, func(node ast.Node) bool {
|
|
switch n := node.(type) {
|
|
case *ast.FuncDecl:
|
|
if isDeployFunc(n) {
|
|
if seenBefore {
|
|
cnt, _ := countLocals(n)
|
|
c.clearSlots(cnt)
|
|
}
|
|
c.convertFuncDecl(f, n, pkg)
|
|
seenBefore = true
|
|
}
|
|
case *ast.GenDecl:
|
|
return false
|
|
}
|
|
return true
|
|
})
|
|
})
|
|
}
|
|
|
|
func (c *codegen) convertFuncDecl(file ast.Node, decl *ast.FuncDecl, pkg *types.Package) {
|
|
var (
|
|
f *funcScope
|
|
ok, isLambda bool
|
|
)
|
|
isInit := isInitFunc(decl)
|
|
isDeploy := isDeployFunc(decl)
|
|
if isInit || isDeploy {
|
|
f = c.newFuncScope(decl, c.newLabel())
|
|
} else {
|
|
f, ok = c.funcs[c.getFuncNameFromDecl("", decl)]
|
|
if ok {
|
|
// If this function is a syscall or builtin we will not convert it to bytecode.
|
|
if isSyscall(f) || isCustomBuiltin(f) {
|
|
return
|
|
}
|
|
c.setLabel(f.label)
|
|
} else if f, ok = c.lambda[c.getIdentName("", decl.Name.Name)]; ok {
|
|
isLambda = ok
|
|
c.setLabel(f.label)
|
|
} else {
|
|
f = c.newFunc(decl)
|
|
}
|
|
}
|
|
|
|
f.rng.Start = uint16(c.prog.Len())
|
|
c.scope = f
|
|
ast.Inspect(decl, c.scope.analyzeVoidCalls) // @OPTIMIZE
|
|
|
|
// All globals copied into the scope of the function need to be added
|
|
// to the stack size of the function.
|
|
if !isInit && !isDeploy {
|
|
sizeLoc := f.countLocals()
|
|
if sizeLoc > 255 {
|
|
c.prog.Err = errors.New("maximum of 255 local variables is allowed")
|
|
}
|
|
sizeArg := f.countArgs()
|
|
if sizeArg > 255 {
|
|
c.prog.Err = errors.New("maximum of 255 local variables is allowed")
|
|
}
|
|
if sizeLoc != 0 || sizeArg != 0 {
|
|
emit.Instruction(c.prog.BinWriter, opcode.INITSLOT, []byte{byte(sizeLoc), byte(sizeArg)})
|
|
}
|
|
}
|
|
|
|
f.vars.newScope()
|
|
defer f.vars.dropScope()
|
|
|
|
// We need to handle methods, which in Go, is just syntactic sugar.
|
|
// The method receiver will be passed in as first argument.
|
|
// We check if this declaration has a receiver and load it into scope.
|
|
//
|
|
// FIXME: For now we will hard cast this to a struct. We can later fine tune this
|
|
// to support other types.
|
|
if decl.Recv != nil {
|
|
for _, arg := range decl.Recv.List {
|
|
// only create an argument here, it will be stored via INITSLOT
|
|
c.scope.newVariable(varArgument, arg.Names[0].Name)
|
|
}
|
|
}
|
|
|
|
// Load the arguments in scope.
|
|
for _, arg := range decl.Type.Params.List {
|
|
for _, id := range arg.Names {
|
|
// only create an argument here, it will be stored via INITSLOT
|
|
c.scope.newVariable(varArgument, id.Name)
|
|
}
|
|
}
|
|
|
|
ast.Walk(c, decl.Body)
|
|
|
|
// If we have reached the end of the function without encountering `return` statement,
|
|
// we should clean alt.stack manually.
|
|
// This can be the case with void and named-return functions.
|
|
if !isInit && !isDeploy && !lastStmtIsReturn(decl) {
|
|
c.saveSequencePoint(decl.Body)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.RET)
|
|
}
|
|
|
|
f.rng.End = uint16(c.prog.Len() - 1)
|
|
|
|
if !isLambda {
|
|
for _, f := range c.lambda {
|
|
c.convertFuncDecl(file, f.decl, pkg)
|
|
}
|
|
c.lambda = make(map[string]*funcScope)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) Visit(node ast.Node) ast.Visitor {
|
|
if c.prog.Err != nil {
|
|
return nil
|
|
}
|
|
switch n := node.(type) {
|
|
|
|
// General declarations.
|
|
// var (
|
|
// x = 2
|
|
// )
|
|
case *ast.GenDecl:
|
|
if n.Tok == token.VAR || n.Tok == token.CONST {
|
|
c.saveSequencePoint(n)
|
|
}
|
|
if n.Tok == token.CONST {
|
|
for _, spec := range n.Specs {
|
|
vs := spec.(*ast.ValueSpec)
|
|
for i := range vs.Names {
|
|
info := c.buildInfo.program.Package(c.currPkg.Path())
|
|
obj := info.Defs[vs.Names[i]]
|
|
c.constMap[c.getIdentName("", vs.Names[i].Name)] = types.TypeAndValue{
|
|
Type: obj.Type(),
|
|
Value: obj.(*types.Const).Val(),
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
for _, spec := range n.Specs {
|
|
switch t := spec.(type) {
|
|
case *ast.ValueSpec:
|
|
for _, id := range t.Names {
|
|
if id.Name != "_" {
|
|
if c.scope == nil {
|
|
// it is a global declaration
|
|
c.newGlobal("", id.Name)
|
|
} else {
|
|
c.scope.newLocal(id.Name)
|
|
}
|
|
c.registerDebugVariable(id.Name, t.Type)
|
|
}
|
|
}
|
|
for i := range t.Names {
|
|
if len(t.Values) != 0 {
|
|
ast.Walk(c, t.Values[i])
|
|
} else {
|
|
c.emitDefault(c.typeOf(t.Type))
|
|
}
|
|
c.emitStoreVar("", t.Names[i].Name)
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
|
|
case *ast.AssignStmt:
|
|
multiRet := len(n.Rhs) != len(n.Lhs)
|
|
c.saveSequencePoint(n)
|
|
// Assign operations are grouped https://github.com/golang/go/blob/master/src/go/types/stmt.go#L160
|
|
isAssignOp := token.ADD_ASSIGN <= n.Tok && n.Tok <= token.AND_NOT_ASSIGN
|
|
if isAssignOp {
|
|
// RHS can contain exactly one expression, thus there is no need to iterate.
|
|
ast.Walk(c, n.Lhs[0])
|
|
ast.Walk(c, n.Rhs[0])
|
|
c.emitToken(n.Tok, c.typeOf(n.Rhs[0]))
|
|
}
|
|
for i := 0; i < len(n.Lhs); i++ {
|
|
switch t := n.Lhs[i].(type) {
|
|
case *ast.Ident:
|
|
if n.Tok == token.DEFINE {
|
|
if !multiRet {
|
|
c.registerDebugVariable(t.Name, n.Rhs[i])
|
|
}
|
|
if t.Name != "_" {
|
|
c.scope.newLocal(t.Name)
|
|
}
|
|
}
|
|
if !isAssignOp && (i == 0 || !multiRet) {
|
|
ast.Walk(c, n.Rhs[i])
|
|
}
|
|
c.emitStoreVar("", t.Name)
|
|
|
|
case *ast.SelectorExpr:
|
|
if !isAssignOp {
|
|
ast.Walk(c, n.Rhs[i])
|
|
}
|
|
typ := c.typeOf(t.X)
|
|
if typ == nil {
|
|
// Store to other package global variable.
|
|
c.emitStoreVar(t.X.(*ast.Ident).Name, t.Sel.Name)
|
|
return nil
|
|
}
|
|
strct, ok := c.getStruct(typ)
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("nested selector assigns not supported yet")
|
|
return nil
|
|
}
|
|
ast.Walk(c, t.X) // load the struct
|
|
i := indexOfStruct(strct, t.Sel.Name) // get the index of the field
|
|
c.emitStoreStructField(i) // store the field
|
|
|
|
// Assignments to index expressions.
|
|
// slice[0] = 10
|
|
case *ast.IndexExpr:
|
|
if !isAssignOp {
|
|
ast.Walk(c, n.Rhs[i])
|
|
}
|
|
ast.Walk(c, t.X)
|
|
ast.Walk(c, t.Index)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.ROT, opcode.SETITEM)
|
|
}
|
|
}
|
|
return nil
|
|
|
|
case *ast.SliceExpr:
|
|
if isCompoundSlice(c.typeOf(n.X.(*ast.Ident)).Underlying()) {
|
|
c.prog.Err = errors.New("subslices are supported only for []byte")
|
|
return nil
|
|
}
|
|
name := n.X.(*ast.Ident).Name
|
|
c.emitLoadVar("", name)
|
|
|
|
if n.Low != nil {
|
|
ast.Walk(c, n.Low)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSH0)
|
|
}
|
|
|
|
if n.High != nil {
|
|
ast.Walk(c, n.High)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.OVER, opcode.SIZE)
|
|
}
|
|
|
|
emit.Opcodes(c.prog.BinWriter, opcode.OVER, opcode.SUB, opcode.SUBSTR)
|
|
|
|
return nil
|
|
|
|
case *ast.ReturnStmt:
|
|
l := c.newLabel()
|
|
c.setLabel(l)
|
|
|
|
cnt := 0
|
|
for i := range c.labelList {
|
|
cnt += c.labelList[i].sz
|
|
}
|
|
c.dropItems(cnt)
|
|
|
|
if len(n.Results) == 0 {
|
|
results := c.scope.decl.Type.Results
|
|
if results.NumFields() != 0 {
|
|
// function with named returns
|
|
for i := len(results.List) - 1; i >= 0; i-- {
|
|
names := results.List[i].Names
|
|
for j := len(names) - 1; j >= 0; j-- {
|
|
c.emitLoadVar("", names[j].Name)
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
// first result should be on top of the stack
|
|
for i := len(n.Results) - 1; i >= 0; i-- {
|
|
ast.Walk(c, n.Results[i])
|
|
}
|
|
}
|
|
|
|
c.processDefers()
|
|
|
|
c.saveSequencePoint(n)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.RET)
|
|
return nil
|
|
|
|
case *ast.IfStmt:
|
|
c.scope.vars.newScope()
|
|
defer c.scope.vars.dropScope()
|
|
|
|
lIf := c.newLabel()
|
|
lElse := c.newLabel()
|
|
lElseEnd := c.newLabel()
|
|
|
|
if n.Init != nil {
|
|
ast.Walk(c, n.Init)
|
|
}
|
|
if n.Cond != nil {
|
|
c.emitBoolExpr(n.Cond, true, false, lElse)
|
|
}
|
|
|
|
c.setLabel(lIf)
|
|
ast.Walk(c, n.Body)
|
|
if n.Else != nil {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, lElseEnd)
|
|
}
|
|
|
|
c.setLabel(lElse)
|
|
if n.Else != nil {
|
|
ast.Walk(c, n.Else)
|
|
}
|
|
c.setLabel(lElseEnd)
|
|
return nil
|
|
|
|
case *ast.SwitchStmt:
|
|
eqOpcode := opcode.EQUAL
|
|
if n.Tag != nil {
|
|
ast.Walk(c, n.Tag)
|
|
eqOpcode, _ = convertToken(token.EQL, c.typeOf(n.Tag))
|
|
} else {
|
|
emit.Bool(c.prog.BinWriter, true)
|
|
}
|
|
switchEnd, label := c.generateLabel(labelEnd)
|
|
|
|
lastSwitch := c.currentSwitch
|
|
c.currentSwitch = label
|
|
c.pushStackLabel(label, 1)
|
|
|
|
startLabels := make([]uint16, len(n.Body.List))
|
|
for i := range startLabels {
|
|
startLabels[i] = c.newLabel()
|
|
}
|
|
for i := range n.Body.List {
|
|
lEnd := c.newLabel()
|
|
lStart := startLabels[i]
|
|
cc := n.Body.List[i].(*ast.CaseClause)
|
|
|
|
if l := len(cc.List); l != 0 { // if not `default`
|
|
for j := range cc.List {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
ast.Walk(c, cc.List[j])
|
|
emit.Opcodes(c.prog.BinWriter, eqOpcode)
|
|
if j == l-1 {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFNOTL, lEnd)
|
|
} else {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFL, lStart)
|
|
}
|
|
}
|
|
}
|
|
|
|
c.scope.vars.newScope()
|
|
|
|
c.setLabel(lStart)
|
|
last := len(cc.Body) - 1
|
|
for j, stmt := range cc.Body {
|
|
if j == last && isFallthroughStmt(stmt) {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, startLabels[i+1])
|
|
break
|
|
}
|
|
ast.Walk(c, stmt)
|
|
}
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, switchEnd)
|
|
c.setLabel(lEnd)
|
|
|
|
c.scope.vars.dropScope()
|
|
}
|
|
|
|
c.setLabel(switchEnd)
|
|
c.dropStackLabel()
|
|
|
|
c.currentSwitch = lastSwitch
|
|
|
|
return nil
|
|
|
|
case *ast.FuncLit:
|
|
l := c.newLabel()
|
|
c.newLambda(l, n)
|
|
buf := make([]byte, 4)
|
|
binary.LittleEndian.PutUint16(buf, l)
|
|
emit.Instruction(c.prog.BinWriter, opcode.PUSHA, buf)
|
|
return nil
|
|
|
|
case *ast.BasicLit:
|
|
c.emitLoadConst(c.typeAndValueOf(n))
|
|
return nil
|
|
|
|
case *ast.StarExpr:
|
|
_, ok := c.getStruct(c.typeOf(n.X))
|
|
if !ok {
|
|
c.prog.Err = errors.New("dereferencing is only supported on structs")
|
|
return nil
|
|
}
|
|
ast.Walk(c, n.X)
|
|
c.emitConvert(stackitem.StructT)
|
|
return nil
|
|
|
|
case *ast.Ident:
|
|
if tv := c.typeAndValueOf(n); tv.Value != nil {
|
|
c.emitLoadConst(tv)
|
|
} else if n.Name == "nil" {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSHNULL)
|
|
} else {
|
|
c.emitLoadVar("", n.Name)
|
|
}
|
|
return nil
|
|
|
|
case *ast.CompositeLit:
|
|
switch typ := c.typeOf(n).Underlying().(type) {
|
|
case *types.Struct:
|
|
c.convertStruct(n, false)
|
|
case *types.Map:
|
|
c.convertMap(n)
|
|
default:
|
|
ln := len(n.Elts)
|
|
// ByteArrays needs a different approach than normal arrays.
|
|
if isByteSlice(typ) {
|
|
c.convertByteArray(n)
|
|
return nil
|
|
}
|
|
for i := ln - 1; i >= 0; i-- {
|
|
ast.Walk(c, n.Elts[i])
|
|
}
|
|
emit.Int(c.prog.BinWriter, int64(ln))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PACK)
|
|
}
|
|
|
|
return nil
|
|
|
|
case *ast.BinaryExpr:
|
|
c.emitBinaryExpr(n, false, false, 0)
|
|
return nil
|
|
|
|
case *ast.CallExpr:
|
|
var (
|
|
f *funcScope
|
|
ok bool
|
|
name string
|
|
numArgs = len(n.Args)
|
|
isBuiltin bool
|
|
isFunc bool
|
|
isLiteral bool
|
|
)
|
|
|
|
switch fun := n.Fun.(type) {
|
|
case *ast.Ident:
|
|
f, ok = c.funcs[c.getIdentName("", fun.Name)]
|
|
isBuiltin = isGoBuiltin(fun.Name)
|
|
if !ok && !isBuiltin {
|
|
name = fun.Name
|
|
}
|
|
// distinguish lambda invocations from type conversions
|
|
if fun.Obj != nil && fun.Obj.Kind == ast.Var {
|
|
isFunc = true
|
|
}
|
|
case *ast.SelectorExpr:
|
|
// If this is a method call we need to walk the AST to load the struct locally.
|
|
// Otherwise this is a function call from a imported package and we can call it
|
|
// directly.
|
|
name, isMethod := c.getFuncNameFromSelector(fun)
|
|
if isMethod {
|
|
ast.Walk(c, fun.X)
|
|
// Dont forget to add 1 extra argument when its a method.
|
|
numArgs++
|
|
}
|
|
|
|
f, ok = c.funcs[name]
|
|
// @FIXME this could cause runtime errors.
|
|
f.selector = fun.X.(*ast.Ident)
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("could not resolve function %s", fun.Sel.Name)
|
|
return nil
|
|
}
|
|
isBuiltin = isCustomBuiltin(f)
|
|
case *ast.ArrayType:
|
|
// For now we will assume that there are only byte slice conversions.
|
|
// E.g. []byte("foobar") or []byte(scriptHash).
|
|
ast.Walk(c, n.Args[0])
|
|
c.emitConvert(stackitem.BufferT)
|
|
return nil
|
|
case *ast.FuncLit:
|
|
isLiteral = true
|
|
}
|
|
|
|
c.saveSequencePoint(n)
|
|
|
|
args := transformArgs(n.Fun, n.Args)
|
|
|
|
// Handle the arguments
|
|
for _, arg := range args {
|
|
ast.Walk(c, arg)
|
|
typ := c.typeOf(arg)
|
|
_, ok := typ.Underlying().(*types.Struct)
|
|
if ok && !isInteropPath(typ.String()) {
|
|
// To clone struct fields we create a new array and append struct to it.
|
|
// This way even non-pointer struct fields will be copied.
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWARRAY0,
|
|
opcode.DUP, opcode.ROT, opcode.APPEND,
|
|
opcode.PUSH0, opcode.PICKITEM)
|
|
}
|
|
}
|
|
// Do not swap for builtin functions.
|
|
if !isBuiltin {
|
|
typ, ok := c.typeOf(n.Fun).(*types.Signature)
|
|
if ok && typ.Variadic() && !n.Ellipsis.IsValid() {
|
|
// pack variadic args into an array only if last argument is not of form `...`
|
|
varSize := len(n.Args) - typ.Params().Len() + 1
|
|
c.emitReverse(varSize)
|
|
emit.Int(c.prog.BinWriter, int64(varSize))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PACK)
|
|
numArgs -= varSize - 1
|
|
}
|
|
c.emitReverse(numArgs)
|
|
}
|
|
|
|
// Check builtin first to avoid nil pointer on funcScope!
|
|
switch {
|
|
case isBuiltin:
|
|
// Use the ident to check, builtins are not in func scopes.
|
|
// We can be sure builtins are of type *ast.Ident.
|
|
c.convertBuiltin(n)
|
|
case name != "":
|
|
// Function was not found thus is can be only an invocation of func-typed variable or type conversion.
|
|
// We care only about string conversions because all others are effectively no-op in NeoVM.
|
|
// E.g. one cannot write `bool(int(a))`, only `int32(int(a))`.
|
|
if isString(c.typeOf(n.Fun)) {
|
|
c.emitConvert(stackitem.ByteArrayT)
|
|
} else if isFunc {
|
|
c.emitLoadVar("", name)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.CALLA)
|
|
}
|
|
case isLiteral:
|
|
ast.Walk(c, n.Fun)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.CALLA)
|
|
case isSyscall(f):
|
|
c.convertSyscall(n, f.pkg.Name(), f.name)
|
|
default:
|
|
emit.Call(c.prog.BinWriter, opcode.CALLL, f.label)
|
|
}
|
|
|
|
if c.scope != nil && c.scope.voidCalls[n] {
|
|
var sz int
|
|
if f != nil {
|
|
sz = f.decl.Type.Results.NumFields()
|
|
} else if !isBuiltin {
|
|
// lambda invocation
|
|
f := c.typeOf(n.Fun).Underlying().(*types.Signature)
|
|
sz = f.Results().Len()
|
|
}
|
|
for i := 0; i < sz; i++ {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP)
|
|
}
|
|
}
|
|
|
|
return nil
|
|
|
|
case *ast.DeferStmt:
|
|
catch := c.newLabel()
|
|
finally := c.newLabel()
|
|
param := make([]byte, 8)
|
|
binary.LittleEndian.PutUint16(param[0:], catch)
|
|
binary.LittleEndian.PutUint16(param[4:], finally)
|
|
emit.Instruction(c.prog.BinWriter, opcode.TRYL, param)
|
|
c.scope.deferStack = append(c.scope.deferStack, deferInfo{
|
|
catchLabel: catch,
|
|
finallyLabel: finally,
|
|
expr: n.Call,
|
|
})
|
|
return nil
|
|
|
|
case *ast.SelectorExpr:
|
|
typ := c.typeOf(n.X)
|
|
if typ == nil {
|
|
// This is a global variable from a package.
|
|
pkgAlias := n.X.(*ast.Ident).Name
|
|
name := c.getIdentName(pkgAlias, n.Sel.Name)
|
|
if tv, ok := c.constMap[name]; ok {
|
|
c.emitLoadConst(tv)
|
|
} else {
|
|
c.emitLoadVar(pkgAlias, n.Sel.Name)
|
|
}
|
|
return nil
|
|
}
|
|
strct, ok := c.getStruct(typ)
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("selectors are supported only on structs")
|
|
return nil
|
|
}
|
|
ast.Walk(c, n.X) // load the struct
|
|
i := indexOfStruct(strct, n.Sel.Name)
|
|
c.emitLoadField(i) // load the field
|
|
return nil
|
|
|
|
case *ast.UnaryExpr:
|
|
if n.Op == token.AND {
|
|
// We support only taking address from struct literals.
|
|
// For identifiers we can't support "taking address" in a general way
|
|
// because both struct and array are reference types.
|
|
lit, ok := n.X.(*ast.CompositeLit)
|
|
if ok {
|
|
c.convertStruct(lit, true)
|
|
return nil
|
|
}
|
|
c.prog.Err = fmt.Errorf("'&' can be used only with struct literals")
|
|
return nil
|
|
}
|
|
|
|
ast.Walk(c, n.X)
|
|
// From https://golang.org/ref/spec#Operators
|
|
// there can be only following unary operators
|
|
// "+" | "-" | "!" | "^" | "*" | "&" | "<-" .
|
|
// of which last three are not used in SC
|
|
switch n.Op {
|
|
case token.ADD:
|
|
// +10 == 10, no need to do anything in this case
|
|
case token.SUB:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEGATE)
|
|
case token.NOT:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NOT)
|
|
case token.XOR:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.INVERT)
|
|
default:
|
|
c.prog.Err = fmt.Errorf("invalid unary operator: %s", n.Op)
|
|
return nil
|
|
}
|
|
return nil
|
|
|
|
case *ast.IncDecStmt:
|
|
ast.Walk(c, n.X)
|
|
c.emitToken(n.Tok, c.typeOf(n.X))
|
|
|
|
// For now only identifiers are supported for (post) for stmts.
|
|
// for i := 0; i < 10; i++ {}
|
|
// Where the post stmt is ( i++ )
|
|
if ident, ok := n.X.(*ast.Ident); ok {
|
|
c.emitStoreVar("", ident.Name)
|
|
}
|
|
return nil
|
|
|
|
case *ast.IndexExpr:
|
|
// Walk the expression, this could be either an Ident or SelectorExpr.
|
|
// This will load local whatever X is.
|
|
ast.Walk(c, n.X)
|
|
ast.Walk(c, n.Index)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PICKITEM) // just pickitem here
|
|
|
|
return nil
|
|
|
|
case *ast.BranchStmt:
|
|
var label string
|
|
if n.Label != nil {
|
|
label = n.Label.Name
|
|
} else if n.Tok == token.BREAK {
|
|
label = c.currentSwitch
|
|
} else if n.Tok == token.CONTINUE {
|
|
label = c.currentFor
|
|
}
|
|
|
|
cnt := 0
|
|
for i := len(c.labelList) - 1; i >= 0 && c.labelList[i].name != label; i-- {
|
|
cnt += c.labelList[i].sz
|
|
}
|
|
c.dropItems(cnt)
|
|
|
|
switch n.Tok {
|
|
case token.BREAK:
|
|
end := c.getLabelOffset(labelEnd, label)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, end)
|
|
case token.CONTINUE:
|
|
post := c.getLabelOffset(labelPost, label)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, post)
|
|
}
|
|
|
|
return nil
|
|
|
|
case *ast.LabeledStmt:
|
|
c.nextLabel = n.Label.Name
|
|
|
|
ast.Walk(c, n.Stmt)
|
|
|
|
return nil
|
|
|
|
case *ast.BlockStmt:
|
|
c.scope.vars.newScope()
|
|
defer c.scope.vars.dropScope()
|
|
|
|
for i := range n.List {
|
|
ast.Walk(c, n.List[i])
|
|
}
|
|
|
|
return nil
|
|
|
|
case *ast.ForStmt:
|
|
c.scope.vars.newScope()
|
|
defer c.scope.vars.dropScope()
|
|
|
|
fstart, label := c.generateLabel(labelStart)
|
|
fend := c.newNamedLabel(labelEnd, label)
|
|
fpost := c.newNamedLabel(labelPost, label)
|
|
|
|
lastLabel := c.currentFor
|
|
lastSwitch := c.currentSwitch
|
|
c.currentFor = label
|
|
c.currentSwitch = label
|
|
|
|
// Walk the initializer and condition.
|
|
if n.Init != nil {
|
|
ast.Walk(c, n.Init)
|
|
}
|
|
|
|
// Set label and walk the condition.
|
|
c.pushStackLabel(label, 0)
|
|
c.setLabel(fstart)
|
|
if n.Cond != nil {
|
|
ast.Walk(c, n.Cond)
|
|
|
|
// Jump if the condition is false
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFNOTL, fend)
|
|
}
|
|
|
|
// Walk body followed by the iterator (post stmt).
|
|
ast.Walk(c, n.Body)
|
|
c.setLabel(fpost)
|
|
if n.Post != nil {
|
|
ast.Walk(c, n.Post)
|
|
}
|
|
|
|
// Jump back to condition.
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, fstart)
|
|
c.setLabel(fend)
|
|
c.dropStackLabel()
|
|
|
|
c.currentFor = lastLabel
|
|
c.currentSwitch = lastSwitch
|
|
|
|
return nil
|
|
|
|
case *ast.RangeStmt:
|
|
c.scope.vars.newScope()
|
|
defer c.scope.vars.dropScope()
|
|
|
|
start, label := c.generateLabel(labelStart)
|
|
end := c.newNamedLabel(labelEnd, label)
|
|
post := c.newNamedLabel(labelPost, label)
|
|
|
|
lastFor := c.currentFor
|
|
lastSwitch := c.currentSwitch
|
|
c.currentFor = label
|
|
c.currentSwitch = label
|
|
|
|
ast.Walk(c, n.X)
|
|
|
|
// Implementation is a bit different for slices and maps:
|
|
// For slices we iterate index from 0 to len-1, storing array, len and index on stack.
|
|
// For maps we iterate index from 0 to len-1, storing map, keyarray, size and index on stack.
|
|
_, isMap := c.typeOf(n.X).Underlying().(*types.Map)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
if isMap {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.KEYS, opcode.DUP)
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SIZE, opcode.PUSH0)
|
|
|
|
stackSize := 3 // slice, len(slice), index
|
|
if isMap {
|
|
stackSize++ // map, keys, len(keys), index in keys
|
|
}
|
|
c.pushStackLabel(label, stackSize)
|
|
c.setLabel(start)
|
|
|
|
emit.Opcodes(c.prog.BinWriter, opcode.OVER, opcode.OVER)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPLEL, end)
|
|
|
|
var keyLoaded bool
|
|
needValue := n.Value != nil && n.Value.(*ast.Ident).Name != "_"
|
|
if n.Key != nil && n.Key.(*ast.Ident).Name != "_" {
|
|
if isMap {
|
|
c.rangeLoadKey()
|
|
if needValue {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
keyLoaded = true
|
|
}
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
}
|
|
c.emitStoreVar("", n.Key.(*ast.Ident).Name)
|
|
}
|
|
if needValue {
|
|
if !isMap || !keyLoaded {
|
|
c.rangeLoadKey()
|
|
}
|
|
if isMap {
|
|
// we have loaded only key from key array, now load value
|
|
emit.Int(c.prog.BinWriter, 4)
|
|
emit.Opcodes(c.prog.BinWriter,
|
|
opcode.PICK, // load map itself (+1 because key was pushed)
|
|
opcode.SWAP, // key should be on top
|
|
opcode.PICKITEM)
|
|
}
|
|
c.emitStoreVar("", n.Value.(*ast.Ident).Name)
|
|
}
|
|
|
|
ast.Walk(c, n.Body)
|
|
|
|
c.setLabel(post)
|
|
|
|
emit.Opcodes(c.prog.BinWriter, opcode.INC)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, start)
|
|
|
|
c.setLabel(end)
|
|
c.dropStackLabel()
|
|
|
|
c.currentFor = lastFor
|
|
c.currentSwitch = lastSwitch
|
|
|
|
return nil
|
|
|
|
// We dont really care about assertions for the core logic.
|
|
// The only thing we need is to please the compiler type checking.
|
|
// For this to work properly, we only need to walk the expression
|
|
// not the assertion type.
|
|
case *ast.TypeAssertExpr:
|
|
ast.Walk(c, n.X)
|
|
typ := toNeoType(c.typeOf(n.Type))
|
|
emit.Instruction(c.prog.BinWriter, opcode.CONVERT, []byte{byte(typ)})
|
|
return nil
|
|
}
|
|
return c
|
|
}
|
|
|
|
// processDefers emits code for `defer` statements.
|
|
// TRY-related opcodes handle exception as follows:
|
|
// 1. CATCH block is executed only if exception has occured.
|
|
// 2. FINALLY block is always executed, but after catch block.
|
|
// Go `defer` statements are a bit different:
|
|
// 1. `defer` is always executed irregardless of whether an exception has occured.
|
|
// 2. `recover` can or can not handle a possible exception.
|
|
// Thus we use the following approach:
|
|
// 1. Throwed exception is saved in a static field X, static fields Y and is set to true.
|
|
// 2. CATCH and FINALLY blocks are the same, and both contain the same CALLs.
|
|
// 3. In CATCH block we set Y to true and emit default return values if it is the last defer.
|
|
// 4. Execute FINALLY block only if Y is false.
|
|
func (c *codegen) processDefers() {
|
|
for i := len(c.scope.deferStack) - 1; i >= 0; i-- {
|
|
stmt := c.scope.deferStack[i]
|
|
after := c.newLabel()
|
|
emit.Jmp(c.prog.BinWriter, opcode.ENDTRYL, after)
|
|
|
|
c.setLabel(stmt.catchLabel)
|
|
c.emitStoreByIndex(varGlobal, c.exceptionIndex)
|
|
emit.Int(c.prog.BinWriter, 1)
|
|
c.emitStoreByIndex(varLocal, c.scope.finallyProcessedIndex)
|
|
ast.Walk(c, stmt.expr)
|
|
if i == 0 {
|
|
// After panic, default values must be returns, except for named returns,
|
|
// which we don't support here for now.
|
|
for i := len(c.scope.decl.Type.Results.List) - 1; i >= 0; i-- {
|
|
c.emitDefault(c.typeOf(c.scope.decl.Type.Results.List[i].Type))
|
|
}
|
|
}
|
|
emit.Jmp(c.prog.BinWriter, opcode.ENDTRYL, after)
|
|
|
|
c.setLabel(stmt.finallyLabel)
|
|
before := c.newLabel()
|
|
c.emitLoadByIndex(varLocal, c.scope.finallyProcessedIndex)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFL, before)
|
|
ast.Walk(c, stmt.expr)
|
|
c.setLabel(before)
|
|
emit.Int(c.prog.BinWriter, 0)
|
|
c.emitStoreByIndex(varLocal, c.scope.finallyProcessedIndex)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.ENDFINALLY)
|
|
c.setLabel(after)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) rangeLoadKey() {
|
|
emit.Int(c.prog.BinWriter, 2)
|
|
emit.Opcodes(c.prog.BinWriter,
|
|
opcode.PICK, // load keys
|
|
opcode.OVER, // load index in key array
|
|
opcode.PICKITEM)
|
|
}
|
|
|
|
func isFallthroughStmt(c ast.Node) bool {
|
|
s, ok := c.(*ast.BranchStmt)
|
|
return ok && s.Tok == token.FALLTHROUGH
|
|
}
|
|
|
|
func (c *codegen) getCompareWithNilArg(n *ast.BinaryExpr) ast.Expr {
|
|
if isExprNil(n.X) {
|
|
return n.Y
|
|
} else if isExprNil(n.Y) {
|
|
return n.X
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (c *codegen) emitJumpOnCondition(cond bool, jmpLabel uint16) {
|
|
if cond {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFL, jmpLabel)
|
|
} else {
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPIFNOTL, jmpLabel)
|
|
}
|
|
}
|
|
|
|
// emitBoolExpr emits boolean expression. If needJump is true and expression evaluates to `cond`,
|
|
// jump to jmpLabel is performed and no item is left on stack.
|
|
func (c *codegen) emitBoolExpr(n ast.Expr, needJump bool, cond bool, jmpLabel uint16) {
|
|
if be, ok := n.(*ast.BinaryExpr); ok {
|
|
c.emitBinaryExpr(be, needJump, cond, jmpLabel)
|
|
} else {
|
|
ast.Walk(c, n)
|
|
if needJump {
|
|
c.emitJumpOnCondition(cond, jmpLabel)
|
|
}
|
|
}
|
|
}
|
|
|
|
// emitBinaryExpr emits binary expression. If needJump is true and expression evaluates to `cond`,
|
|
// jump to jmpLabel is performed and no item is left on stack.
|
|
func (c *codegen) emitBinaryExpr(n *ast.BinaryExpr, needJump bool, cond bool, jmpLabel uint16) {
|
|
// The AST package will try to resolve all basic literals for us.
|
|
// If the typeinfo.Value is not nil we know that the expr is resolved
|
|
// and needs no further action. e.g. x := 2 + 2 + 2 will be resolved to 6.
|
|
// NOTE: Constants will also be automatically resolved be the AST parser.
|
|
// example:
|
|
// const x = 10
|
|
// x + 2 will results into 12
|
|
tinfo := c.typeAndValueOf(n)
|
|
if tinfo.Value != nil {
|
|
c.emitLoadConst(tinfo)
|
|
if needJump && isBool(tinfo.Type) {
|
|
c.emitJumpOnCondition(cond, jmpLabel)
|
|
}
|
|
return
|
|
} else if arg := c.getCompareWithNilArg(n); arg != nil {
|
|
ast.Walk(c, arg)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.ISNULL)
|
|
if needJump {
|
|
c.emitJumpOnCondition(cond == (n.Op == token.EQL), jmpLabel)
|
|
} else if n.Op == token.NEQ {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NOT)
|
|
}
|
|
return
|
|
}
|
|
|
|
switch n.Op {
|
|
case token.LAND, token.LOR:
|
|
end := c.newLabel()
|
|
|
|
// true || .. == true, false && .. == false
|
|
condShort := n.Op == token.LOR
|
|
if needJump {
|
|
l := end
|
|
if cond == condShort {
|
|
l = jmpLabel
|
|
}
|
|
c.emitBoolExpr(n.X, true, condShort, l)
|
|
c.emitBoolExpr(n.Y, true, cond, jmpLabel)
|
|
} else {
|
|
push := c.newLabel()
|
|
c.emitBoolExpr(n.X, true, condShort, push)
|
|
c.emitBoolExpr(n.Y, false, false, 0)
|
|
emit.Jmp(c.prog.BinWriter, opcode.JMPL, end)
|
|
c.setLabel(push)
|
|
emit.Bool(c.prog.BinWriter, condShort)
|
|
}
|
|
c.setLabel(end)
|
|
|
|
default:
|
|
ast.Walk(c, n.X)
|
|
ast.Walk(c, n.Y)
|
|
typ := c.typeOf(n.X)
|
|
if !needJump {
|
|
c.emitToken(n.Op, typ)
|
|
return
|
|
}
|
|
op, ok := getJumpForToken(n.Op, typ)
|
|
if !ok {
|
|
c.emitToken(n.Op, typ)
|
|
c.emitJumpOnCondition(cond, jmpLabel)
|
|
return
|
|
}
|
|
if !cond {
|
|
op = negateJmp(op)
|
|
}
|
|
emit.Jmp(c.prog.BinWriter, op, jmpLabel)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) pushStackLabel(name string, size int) {
|
|
c.labelList = append(c.labelList, labelWithStackSize{
|
|
name: name,
|
|
sz: size,
|
|
})
|
|
}
|
|
|
|
func (c *codegen) dropStackLabel() {
|
|
last := len(c.labelList) - 1
|
|
c.dropItems(c.labelList[last].sz)
|
|
c.labelList = c.labelList[:last]
|
|
}
|
|
|
|
func (c *codegen) dropItems(n int) {
|
|
if n < 4 {
|
|
for i := 0; i < n; i++ {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP)
|
|
}
|
|
return
|
|
}
|
|
|
|
emit.Int(c.prog.BinWriter, int64(n))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PACK, opcode.DROP)
|
|
}
|
|
|
|
// emitReverse reverses top num items of the stack.
|
|
func (c *codegen) emitReverse(num int) {
|
|
switch num {
|
|
case 0, 1:
|
|
case 2:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SWAP)
|
|
case 3:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REVERSE3)
|
|
case 4:
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REVERSE4)
|
|
default:
|
|
emit.Int(c.prog.BinWriter, int64(num))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REVERSEN)
|
|
}
|
|
}
|
|
|
|
// generateLabel returns a new label.
|
|
func (c *codegen) generateLabel(typ labelOffsetType) (uint16, string) {
|
|
name := c.nextLabel
|
|
if name == "" {
|
|
name = fmt.Sprintf("@%d", len(c.l))
|
|
}
|
|
|
|
c.nextLabel = ""
|
|
return c.newNamedLabel(typ, name), name
|
|
}
|
|
|
|
func (c *codegen) getLabelOffset(typ labelOffsetType, name string) uint16 {
|
|
return c.labels[labelWithType{name: name, typ: typ}]
|
|
}
|
|
|
|
// For `&&` and `||` it return an opcode which jumps only if result is known:
|
|
// false && .. == false, true || .. = true
|
|
func getJumpForToken(tok token.Token, typ types.Type) (opcode.Opcode, bool) {
|
|
switch tok {
|
|
case token.GTR:
|
|
return opcode.JMPGTL, true
|
|
case token.GEQ:
|
|
return opcode.JMPGEL, true
|
|
case token.LSS:
|
|
return opcode.JMPLTL, true
|
|
case token.LEQ:
|
|
return opcode.JMPLEL, true
|
|
case token.EQL, token.NEQ:
|
|
if isNumber(typ) {
|
|
if tok == token.EQL {
|
|
return opcode.JMPEQL, true
|
|
}
|
|
return opcode.JMPNEL, true
|
|
}
|
|
}
|
|
return 0, false
|
|
}
|
|
|
|
// getByteArray returns byte array value from constant expr.
|
|
// Only literals are supported.
|
|
func (c *codegen) getByteArray(expr ast.Expr) []byte {
|
|
switch t := expr.(type) {
|
|
case *ast.CompositeLit:
|
|
if !isByteSlice(c.typeOf(t.Type)) {
|
|
return nil
|
|
}
|
|
buf := make([]byte, len(t.Elts))
|
|
for i := 0; i < len(t.Elts); i++ {
|
|
t := c.typeAndValueOf(t.Elts[i])
|
|
val, _ := constant.Int64Val(t.Value)
|
|
buf[i] = byte(val)
|
|
}
|
|
return buf
|
|
case *ast.CallExpr:
|
|
if tv := c.typeAndValueOf(t.Args[0]); tv.Value != nil {
|
|
val := constant.StringVal(tv.Value)
|
|
return []byte(val)
|
|
}
|
|
|
|
return nil
|
|
default:
|
|
return nil
|
|
}
|
|
}
|
|
|
|
func (c *codegen) convertSyscall(expr *ast.CallExpr, api, name string) {
|
|
syscall, ok := syscalls[api][name]
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("unknown VM syscall api: %s.%s", api, name)
|
|
return
|
|
}
|
|
emit.Syscall(c.prog.BinWriter, syscall)
|
|
|
|
// This NOP instruction is basically not needed, but if we do, we have a
|
|
// one to one matching avm file with neo-python which is very nice for debugging.
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NOP)
|
|
}
|
|
|
|
// emitSliceHelper emits 3 items on stack: slice, its first index, and its size.
|
|
func (c *codegen) emitSliceHelper(e ast.Expr) {
|
|
if !isByteSlice(c.typeOf(e)) {
|
|
c.prog.Err = fmt.Errorf("copy is supported only for byte-slices")
|
|
return
|
|
}
|
|
var hasLowIndex bool
|
|
switch src := e.(type) {
|
|
case *ast.SliceExpr:
|
|
ast.Walk(c, src.X)
|
|
if src.High != nil {
|
|
ast.Walk(c, src.High)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.SIZE)
|
|
}
|
|
if src.Low != nil {
|
|
ast.Walk(c, src.Low)
|
|
hasLowIndex = true
|
|
} else {
|
|
emit.Int(c.prog.BinWriter, 0)
|
|
}
|
|
default:
|
|
ast.Walk(c, src)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.SIZE)
|
|
emit.Int(c.prog.BinWriter, 0)
|
|
}
|
|
if !hasLowIndex {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SWAP)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.ROT, opcode.SWAP, opcode.SUB)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) convertBuiltin(expr *ast.CallExpr) {
|
|
var name string
|
|
switch t := expr.Fun.(type) {
|
|
case *ast.Ident:
|
|
name = t.Name
|
|
case *ast.SelectorExpr:
|
|
name = t.Sel.Name
|
|
}
|
|
|
|
switch name {
|
|
case "copy":
|
|
// stack for MEMCPY is: dst, dst_index, src, src_index, count
|
|
c.emitSliceHelper(expr.Args[0])
|
|
c.emitSliceHelper(expr.Args[1])
|
|
emit.Int(c.prog.BinWriter, 3)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.ROLL, opcode.MIN)
|
|
if !c.scope.voidCalls[expr] {
|
|
// insert top item to the bottom of MEMCPY args, so that it is left on stack
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
emit.Int(c.prog.BinWriter, 6)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REVERSEN)
|
|
emit.Int(c.prog.BinWriter, 5)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REVERSEN)
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, opcode.MEMCPY)
|
|
case "make":
|
|
typ := c.typeOf(expr.Args[0])
|
|
switch {
|
|
case isMap(typ):
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWMAP)
|
|
default:
|
|
if len(expr.Args) == 3 {
|
|
c.prog.Err = fmt.Errorf("`make()` with a capacity argument is not supported")
|
|
return
|
|
}
|
|
ast.Walk(c, expr.Args[1])
|
|
if isByteSlice(typ) {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWBUFFER)
|
|
} else {
|
|
neoT := toNeoType(typ.(*types.Slice).Elem())
|
|
emit.Instruction(c.prog.BinWriter, opcode.NEWARRAYT, []byte{byte(neoT)})
|
|
}
|
|
}
|
|
case "len":
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.ISNULL)
|
|
emit.Instruction(c.prog.BinWriter, opcode.JMPIF, []byte{2 + 1 + 2})
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SIZE)
|
|
emit.Instruction(c.prog.BinWriter, opcode.JMP, []byte{2 + 1 + 1})
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP, opcode.PUSH0)
|
|
case "append":
|
|
arg := expr.Args[0]
|
|
typ := c.typeInfo.Types[arg].Type
|
|
c.emitReverse(len(expr.Args))
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.ISNULL)
|
|
emit.Instruction(c.prog.BinWriter, opcode.JMPIFNOT, []byte{2 + 3})
|
|
if isByteSlice(typ) {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP, opcode.PUSH0, opcode.NEWBUFFER)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DROP, opcode.NEWARRAY0, opcode.NOP)
|
|
}
|
|
// Jump target.
|
|
for range expr.Args[1:] {
|
|
if isByteSlice(typ) {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SWAP, opcode.CAT)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP, opcode.ROT, opcode.APPEND)
|
|
}
|
|
}
|
|
case "panic":
|
|
emit.Opcodes(c.prog.BinWriter, opcode.THROW)
|
|
case "recover":
|
|
if !c.scope.voidCalls[expr] {
|
|
c.emitLoadByIndex(varGlobal, c.exceptionIndex)
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, opcode.PUSHNULL)
|
|
c.emitStoreByIndex(varGlobal, c.exceptionIndex)
|
|
case "delete":
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REMOVE)
|
|
case "ToInteger", "ToByteArray", "ToBool":
|
|
typ := stackitem.IntegerT
|
|
switch name {
|
|
case "ToByteArray":
|
|
typ = stackitem.ByteArrayT
|
|
case "ToBool":
|
|
typ = stackitem.BooleanT
|
|
}
|
|
c.emitConvert(typ)
|
|
case "Remove":
|
|
if !isCompoundSlice(c.typeOf(expr.Args[0])) {
|
|
c.prog.Err = errors.New("`Remove` supports only non-byte slices")
|
|
return
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, opcode.REMOVE)
|
|
case "Equals":
|
|
emit.Opcodes(c.prog.BinWriter, opcode.EQUAL)
|
|
case "FromAddress":
|
|
// We can be sure that this is a ast.BasicLit just containing a simple
|
|
// address string. Note that the string returned from calling Value will
|
|
// contain double quotes that need to be stripped.
|
|
addressStr := expr.Args[0].(*ast.BasicLit).Value
|
|
addressStr = strings.Replace(addressStr, "\"", "", 2)
|
|
uint160, err := address.StringToUint160(addressStr)
|
|
if err != nil {
|
|
c.prog.Err = err
|
|
return
|
|
}
|
|
bytes := uint160.BytesBE()
|
|
emit.Bytes(c.prog.BinWriter, bytes)
|
|
c.emitConvert(stackitem.BufferT)
|
|
}
|
|
}
|
|
|
|
// transformArgs returns a list of function arguments
|
|
// which should be put on stack.
|
|
// There are special cases for builtins:
|
|
// 1. With FromAddress, parameter conversion is happening at compile-time
|
|
// so there is no need to push parameters on stack and perform an actual call
|
|
// 2. With panic, generated code depends on if argument was nil or a string so
|
|
// it should be handled accordingly.
|
|
func transformArgs(fun ast.Expr, args []ast.Expr) []ast.Expr {
|
|
switch f := fun.(type) {
|
|
case *ast.SelectorExpr:
|
|
if f.Sel.Name == "FromAddress" {
|
|
return args[1:]
|
|
}
|
|
case *ast.Ident:
|
|
switch f.Name {
|
|
case "make", "copy":
|
|
return nil
|
|
}
|
|
}
|
|
|
|
return args
|
|
}
|
|
|
|
// emitConvert converts top stack item to the specified type.
|
|
func (c *codegen) emitConvert(typ stackitem.Type) {
|
|
emit.Instruction(c.prog.BinWriter, opcode.CONVERT, []byte{byte(typ)})
|
|
}
|
|
|
|
func (c *codegen) convertByteArray(lit *ast.CompositeLit) {
|
|
buf := make([]byte, len(lit.Elts))
|
|
varIndices := []int{}
|
|
for i := 0; i < len(lit.Elts); i++ {
|
|
t := c.typeAndValueOf(lit.Elts[i])
|
|
if t.Value != nil {
|
|
val, _ := constant.Int64Val(t.Value)
|
|
buf[i] = byte(val)
|
|
} else {
|
|
varIndices = append(varIndices, i)
|
|
}
|
|
}
|
|
emit.Bytes(c.prog.BinWriter, buf)
|
|
c.emitConvert(stackitem.BufferT)
|
|
for _, i := range varIndices {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
emit.Int(c.prog.BinWriter, int64(i))
|
|
ast.Walk(c, lit.Elts[i])
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SETITEM)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) convertMap(lit *ast.CompositeLit) {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWMAP)
|
|
for i := range lit.Elts {
|
|
elem := lit.Elts[i].(*ast.KeyValueExpr)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
ast.Walk(c, elem.Key)
|
|
ast.Walk(c, elem.Value)
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SETITEM)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) getStruct(typ types.Type) (*types.Struct, bool) {
|
|
switch t := typ.Underlying().(type) {
|
|
case *types.Struct:
|
|
return t, true
|
|
case *types.Pointer:
|
|
strct, ok := t.Elem().Underlying().(*types.Struct)
|
|
return strct, ok
|
|
default:
|
|
return nil, false
|
|
}
|
|
}
|
|
|
|
func (c *codegen) convertStruct(lit *ast.CompositeLit, ptr bool) {
|
|
// Create a new structScope to initialize and store
|
|
// the positions of its variables.
|
|
strct, ok := c.typeOf(lit).Underlying().(*types.Struct)
|
|
if !ok {
|
|
c.prog.Err = fmt.Errorf("the given literal is not of type struct: %v", lit)
|
|
return
|
|
}
|
|
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NOP)
|
|
emit.Int(c.prog.BinWriter, int64(strct.NumFields()))
|
|
if ptr {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWARRAY)
|
|
} else {
|
|
emit.Opcodes(c.prog.BinWriter, opcode.NEWSTRUCT)
|
|
}
|
|
|
|
keyedLit := len(lit.Elts) > 0
|
|
if keyedLit {
|
|
_, ok := lit.Elts[0].(*ast.KeyValueExpr)
|
|
keyedLit = keyedLit && ok
|
|
}
|
|
// We need to locally store all the fields, even if they are not initialized.
|
|
// We will initialize all fields to their "zero" value.
|
|
for i := 0; i < strct.NumFields(); i++ {
|
|
sField := strct.Field(i)
|
|
var initialized bool
|
|
|
|
emit.Opcodes(c.prog.BinWriter, opcode.DUP)
|
|
emit.Int(c.prog.BinWriter, int64(i))
|
|
|
|
if !keyedLit {
|
|
if len(lit.Elts) > i {
|
|
ast.Walk(c, lit.Elts[i])
|
|
initialized = true
|
|
}
|
|
} else {
|
|
// Fields initialized by the program.
|
|
for _, field := range lit.Elts {
|
|
f := field.(*ast.KeyValueExpr)
|
|
fieldName := f.Key.(*ast.Ident).Name
|
|
|
|
if sField.Name() == fieldName {
|
|
ast.Walk(c, f.Value)
|
|
initialized = true
|
|
break
|
|
}
|
|
}
|
|
}
|
|
if !initialized {
|
|
c.emitDefault(sField.Type())
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, opcode.SETITEM)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) emitToken(tok token.Token, typ types.Type) {
|
|
op, err := convertToken(tok, typ)
|
|
if err != nil {
|
|
c.prog.Err = err
|
|
return
|
|
}
|
|
emit.Opcodes(c.prog.BinWriter, op)
|
|
}
|
|
|
|
func convertToken(tok token.Token, typ types.Type) (opcode.Opcode, error) {
|
|
switch tok {
|
|
case token.ADD_ASSIGN, token.ADD:
|
|
// VM has separate opcodes for number and string concatenation
|
|
if isString(typ) {
|
|
return opcode.CAT, nil
|
|
}
|
|
return opcode.ADD, nil
|
|
case token.SUB_ASSIGN:
|
|
return opcode.SUB, nil
|
|
case token.MUL_ASSIGN:
|
|
return opcode.MUL, nil
|
|
case token.QUO_ASSIGN:
|
|
return opcode.DIV, nil
|
|
case token.REM_ASSIGN:
|
|
return opcode.MOD, nil
|
|
case token.SUB:
|
|
return opcode.SUB, nil
|
|
case token.MUL:
|
|
return opcode.MUL, nil
|
|
case token.QUO:
|
|
return opcode.DIV, nil
|
|
case token.REM:
|
|
return opcode.MOD, nil
|
|
case token.LSS:
|
|
return opcode.LT, nil
|
|
case token.LEQ:
|
|
return opcode.LTE, nil
|
|
case token.GTR:
|
|
return opcode.GT, nil
|
|
case token.GEQ:
|
|
return opcode.GTE, nil
|
|
case token.EQL:
|
|
// VM has separate opcodes for number and string equality
|
|
if isNumber(typ) {
|
|
return opcode.NUMEQUAL, nil
|
|
}
|
|
return opcode.EQUAL, nil
|
|
case token.NEQ:
|
|
// VM has separate opcodes for number and string equality
|
|
if isNumber(typ) {
|
|
return opcode.NUMNOTEQUAL, nil
|
|
}
|
|
return opcode.NOTEQUAL, nil
|
|
case token.DEC:
|
|
return opcode.DEC, nil
|
|
case token.INC:
|
|
return opcode.INC, nil
|
|
case token.NOT:
|
|
return opcode.NOT, nil
|
|
case token.AND:
|
|
return opcode.AND, nil
|
|
case token.OR:
|
|
return opcode.OR, nil
|
|
case token.SHL:
|
|
return opcode.SHL, nil
|
|
case token.SHR:
|
|
return opcode.SHR, nil
|
|
case token.XOR:
|
|
return opcode.XOR, nil
|
|
default:
|
|
return 0, fmt.Errorf("compiler could not convert token: %s", tok)
|
|
}
|
|
}
|
|
|
|
func (c *codegen) newFunc(decl *ast.FuncDecl) *funcScope {
|
|
f := c.newFuncScope(decl, c.newLabel())
|
|
c.funcs[c.getFuncNameFromDecl("", decl)] = f
|
|
return f
|
|
}
|
|
|
|
// getFuncNameFromSelector returns fully-qualified function name from the selector expression.
|
|
// Second return value is true iff this was a method call, not foreign package call.
|
|
func (c *codegen) getFuncNameFromSelector(e *ast.SelectorExpr) (string, bool) {
|
|
ident := e.X.(*ast.Ident)
|
|
if c.typeInfo.Selections[e] != nil {
|
|
typ := c.typeInfo.Types[ident].Type.String()
|
|
return c.getIdentName(typ, e.Sel.Name), true
|
|
}
|
|
return c.getIdentName(ident.Name, e.Sel.Name), false
|
|
}
|
|
|
|
func (c *codegen) newLambda(u uint16, lit *ast.FuncLit) {
|
|
name := fmt.Sprintf("lambda@%d", u)
|
|
f := c.newFuncScope(&ast.FuncDecl{
|
|
Name: ast.NewIdent(name),
|
|
Type: lit.Type,
|
|
Body: lit.Body,
|
|
}, u)
|
|
c.lambda[c.getFuncNameFromDecl("", f.decl)] = f
|
|
}
|
|
|
|
func (c *codegen) compile(info *buildInfo, pkg *loader.PackageInfo) error {
|
|
c.mainPkg = pkg
|
|
c.analyzePkgOrder()
|
|
c.fillDocumentInfo()
|
|
funUsage := c.analyzeFuncUsage()
|
|
|
|
// Bring all imported functions into scope.
|
|
c.ForEachFile(c.resolveFuncDecls)
|
|
|
|
n, initLocals, deployLocals := c.traverseGlobals()
|
|
hasInit := initLocals > -1
|
|
if n > 0 || hasInit {
|
|
c.initEndOffset = c.prog.Len()
|
|
emit.Opcodes(c.prog.BinWriter, opcode.RET)
|
|
}
|
|
|
|
hasDeploy := deployLocals > -1
|
|
if hasDeploy {
|
|
emit.Instruction(c.prog.BinWriter, opcode.INITSLOT, []byte{byte(deployLocals), 1})
|
|
c.convertDeployFuncs()
|
|
c.deployEndOffset = c.prog.Len()
|
|
emit.Opcodes(c.prog.BinWriter, opcode.RET)
|
|
}
|
|
|
|
// sort map keys to generate code deterministically.
|
|
keys := make([]*types.Package, 0, len(info.program.AllPackages))
|
|
for p := range info.program.AllPackages {
|
|
keys = append(keys, p)
|
|
}
|
|
sort.Slice(keys, func(i, j int) bool { return keys[i].Path() < keys[j].Path() })
|
|
|
|
// Generate the code for the program.
|
|
c.ForEachFile(func(f *ast.File, pkg *types.Package) {
|
|
for _, decl := range f.Decls {
|
|
switch n := decl.(type) {
|
|
case *ast.FuncDecl:
|
|
// Don't convert the function if it's not used. This will save a lot
|
|
// of bytecode space.
|
|
name := c.getFuncNameFromDecl(pkg.Path(), n)
|
|
if !isInitFunc(n) && !isDeployFunc(n) && funUsage.funcUsed(name) && !isInteropPath(pkg.Path()) {
|
|
c.convertFuncDecl(f, n, pkg)
|
|
}
|
|
}
|
|
}
|
|
})
|
|
|
|
return c.prog.Err
|
|
}
|
|
|
|
func newCodegen(info *buildInfo, pkg *loader.PackageInfo) *codegen {
|
|
return &codegen{
|
|
buildInfo: info,
|
|
prog: io.NewBufBinWriter(),
|
|
l: []int{},
|
|
funcs: map[string]*funcScope{},
|
|
lambda: map[string]*funcScope{},
|
|
globals: map[string]int{},
|
|
labels: map[labelWithType]uint16{},
|
|
typeInfo: &pkg.Info,
|
|
constMap: map[string]types.TypeAndValue{},
|
|
docIndex: map[string]int{},
|
|
|
|
initEndOffset: -1,
|
|
deployEndOffset: -1,
|
|
|
|
sequencePoints: make(map[string][]DebugSeqPoint),
|
|
}
|
|
}
|
|
|
|
// CodeGen compiles the program to bytecode.
|
|
func CodeGen(info *buildInfo) ([]byte, *DebugInfo, error) {
|
|
pkg := info.program.Package(info.initialPackage)
|
|
c := newCodegen(info, pkg)
|
|
|
|
if err := c.compile(info, pkg); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
buf, err := c.writeJumps(c.prog.Bytes())
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
return buf, c.emitDebugInfo(buf), nil
|
|
}
|
|
|
|
func (c *codegen) resolveFuncDecls(f *ast.File, pkg *types.Package) {
|
|
for _, decl := range f.Decls {
|
|
switch n := decl.(type) {
|
|
case *ast.FuncDecl:
|
|
c.newFunc(n)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (c *codegen) writeJumps(b []byte) ([]byte, error) {
|
|
ctx := vm.NewContext(b)
|
|
var offsets []int
|
|
for op, _, err := ctx.Next(); err == nil && ctx.IP() < len(b); op, _, err = ctx.Next() {
|
|
switch op {
|
|
case opcode.JMP, opcode.JMPIFNOT, opcode.JMPIF, opcode.CALL,
|
|
opcode.JMPEQ, opcode.JMPNE,
|
|
opcode.JMPGT, opcode.JMPGE, opcode.JMPLE, opcode.JMPLT:
|
|
case opcode.TRYL:
|
|
nextIP := ctx.NextIP()
|
|
catchArg := b[nextIP-8:]
|
|
_, err := c.replaceLabelWithOffset(ctx.IP(), catchArg)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
finallyArg := b[nextIP-4:]
|
|
_, err = c.replaceLabelWithOffset(ctx.IP(), finallyArg)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
case opcode.JMPL, opcode.JMPIFL, opcode.JMPIFNOTL,
|
|
opcode.JMPEQL, opcode.JMPNEL,
|
|
opcode.JMPGTL, opcode.JMPGEL, opcode.JMPLEL, opcode.JMPLTL,
|
|
opcode.CALLL, opcode.PUSHA, opcode.ENDTRYL:
|
|
// we can't use arg returned by ctx.Next() because it is copied
|
|
nextIP := ctx.NextIP()
|
|
arg := b[nextIP-4:]
|
|
offset, err := c.replaceLabelWithOffset(ctx.IP(), arg)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if op != opcode.PUSHA && math.MinInt8 <= offset && offset <= math.MaxInt8 {
|
|
offsets = append(offsets, ctx.IP())
|
|
}
|
|
}
|
|
}
|
|
// Correct function ip range.
|
|
// Note: indices are sorted in increasing order.
|
|
for _, f := range c.funcs {
|
|
loop:
|
|
for _, ind := range offsets {
|
|
switch {
|
|
case ind > int(f.rng.End):
|
|
break loop
|
|
case ind < int(f.rng.Start):
|
|
f.rng.Start -= longToShortRemoveCount
|
|
f.rng.End -= longToShortRemoveCount
|
|
case ind >= int(f.rng.Start):
|
|
f.rng.End -= longToShortRemoveCount
|
|
}
|
|
}
|
|
}
|
|
return shortenJumps(b, offsets), nil
|
|
}
|
|
|
|
func (c *codegen) replaceLabelWithOffset(ip int, arg []byte) (int, error) {
|
|
index := binary.LittleEndian.Uint16(arg)
|
|
if int(index) > len(c.l) {
|
|
return 0, fmt.Errorf("unexpected label number: %d (max %d)", index, len(c.l))
|
|
}
|
|
offset := c.l[index] - ip
|
|
if offset > math.MaxInt32 || offset < math.MinInt32 {
|
|
return 0, fmt.Errorf("label offset is too big at the instruction %d: %d (max %d, min %d)",
|
|
ip, offset, math.MaxInt32, math.MinInt32)
|
|
}
|
|
binary.LittleEndian.PutUint32(arg, uint32(offset))
|
|
return offset, nil
|
|
}
|
|
|
|
// longToShortRemoveCount is a difference between short and long instruction sizes in bytes.
|
|
const longToShortRemoveCount = 3
|
|
|
|
// shortenJumps returns converts b to a program where all long JMP*/CALL* specified by absolute offsets,
|
|
// are replaced with their corresponding short counterparts. It panics if either b or offsets are invalid.
|
|
// This is done in 2 passes:
|
|
// 1. Alter jump offsets taking into account parts to be removed.
|
|
// 2. Perform actual removal of jump targets.
|
|
// Note: after jump offsets altering, there can appear new candidates for conversion.
|
|
// These are ignored for now.
|
|
func shortenJumps(b []byte, offsets []int) []byte {
|
|
if len(offsets) == 0 {
|
|
return b
|
|
}
|
|
|
|
// 1. Alter existing jump offsets.
|
|
ctx := vm.NewContext(b)
|
|
for op, _, err := ctx.Next(); err == nil && ctx.IP() < len(b); op, _, err = ctx.Next() {
|
|
// we can't use arg returned by ctx.Next() because it is copied
|
|
nextIP := ctx.NextIP()
|
|
ip := ctx.IP()
|
|
switch op {
|
|
case opcode.JMP, opcode.JMPIFNOT, opcode.JMPIF, opcode.CALL,
|
|
opcode.JMPEQ, opcode.JMPNE,
|
|
opcode.JMPGT, opcode.JMPGE, opcode.JMPLE, opcode.JMPLT, opcode.ENDTRY:
|
|
offset := int(int8(b[nextIP-1]))
|
|
offset += calcOffsetCorrection(ip, ip+offset, offsets)
|
|
b[nextIP-1] = byte(offset)
|
|
case opcode.TRY:
|
|
catchOffset := int(int8(b[nextIP-2]))
|
|
catchOffset += calcOffsetCorrection(ip, ip+catchOffset, offsets)
|
|
b[nextIP-1] = byte(catchOffset)
|
|
finallyOffset := int(int8(b[nextIP-1]))
|
|
finallyOffset += calcOffsetCorrection(ip, ip+finallyOffset, offsets)
|
|
b[nextIP-1] = byte(finallyOffset)
|
|
case opcode.JMPL, opcode.JMPIFL, opcode.JMPIFNOTL,
|
|
opcode.JMPEQL, opcode.JMPNEL,
|
|
opcode.JMPGTL, opcode.JMPGEL, opcode.JMPLEL, opcode.JMPLTL,
|
|
opcode.CALLL, opcode.PUSHA, opcode.ENDTRYL:
|
|
arg := b[nextIP-4:]
|
|
offset := int(int32(binary.LittleEndian.Uint32(arg)))
|
|
offset += calcOffsetCorrection(ip, ip+offset, offsets)
|
|
binary.LittleEndian.PutUint32(arg, uint32(offset))
|
|
case opcode.TRYL:
|
|
arg := b[nextIP-8:]
|
|
catchOffset := int(int32(binary.LittleEndian.Uint32(arg)))
|
|
catchOffset += calcOffsetCorrection(ip, ip+catchOffset, offsets)
|
|
binary.LittleEndian.PutUint32(arg, uint32(catchOffset))
|
|
arg = b[nextIP-4:]
|
|
finallyOffset := int(int32(binary.LittleEndian.Uint32(arg)))
|
|
finallyOffset += calcOffsetCorrection(ip, ip+finallyOffset, offsets)
|
|
binary.LittleEndian.PutUint32(arg, uint32(finallyOffset))
|
|
}
|
|
}
|
|
|
|
// 2. Convert instructions.
|
|
copyOffset := 0
|
|
l := len(offsets)
|
|
b[offsets[0]] = byte(toShortForm(opcode.Opcode(b[offsets[0]])))
|
|
for i := 0; i < l; i++ {
|
|
start := offsets[i] + 2
|
|
end := len(b)
|
|
if i != l-1 {
|
|
end = offsets[i+1] + 2
|
|
b[offsets[i+1]] = byte(toShortForm(opcode.Opcode(b[offsets[i+1]])))
|
|
}
|
|
copy(b[start-copyOffset:], b[start+3:end])
|
|
copyOffset += longToShortRemoveCount
|
|
}
|
|
return b[:len(b)-copyOffset]
|
|
}
|
|
|
|
func calcOffsetCorrection(ip, target int, offsets []int) int {
|
|
cnt := 0
|
|
start := sort.Search(len(offsets), func(i int) bool {
|
|
return offsets[i] >= ip || offsets[i] >= target
|
|
})
|
|
for i := start; i < len(offsets) && (offsets[i] < target || offsets[i] <= ip); i++ {
|
|
ind := offsets[i]
|
|
if ip <= ind && ind < target ||
|
|
ind != ip && target <= ind && ind <= ip {
|
|
cnt += longToShortRemoveCount
|
|
}
|
|
}
|
|
if ip < target {
|
|
return -cnt
|
|
}
|
|
return cnt
|
|
}
|
|
|
|
func negateJmp(op opcode.Opcode) opcode.Opcode {
|
|
switch op {
|
|
case opcode.JMPIFL:
|
|
return opcode.JMPIFNOTL
|
|
case opcode.JMPIFNOTL:
|
|
return opcode.JMPIFL
|
|
case opcode.JMPEQL:
|
|
return opcode.JMPNEL
|
|
case opcode.JMPNEL:
|
|
return opcode.JMPEQL
|
|
case opcode.JMPGTL:
|
|
return opcode.JMPLEL
|
|
case opcode.JMPGEL:
|
|
return opcode.JMPLTL
|
|
case opcode.JMPLEL:
|
|
return opcode.JMPGTL
|
|
case opcode.JMPLTL:
|
|
return opcode.JMPGEL
|
|
default:
|
|
panic(fmt.Errorf("invalid opcode in negateJmp: %s", op))
|
|
}
|
|
}
|
|
|
|
func toShortForm(op opcode.Opcode) opcode.Opcode {
|
|
switch op {
|
|
case opcode.JMPL:
|
|
return opcode.JMP
|
|
case opcode.JMPIFL:
|
|
return opcode.JMPIF
|
|
case opcode.JMPIFNOTL:
|
|
return opcode.JMPIFNOT
|
|
case opcode.JMPEQL:
|
|
return opcode.JMPEQ
|
|
case opcode.JMPNEL:
|
|
return opcode.JMPNE
|
|
case opcode.JMPGTL:
|
|
return opcode.JMPGT
|
|
case opcode.JMPGEL:
|
|
return opcode.JMPGE
|
|
case opcode.JMPLEL:
|
|
return opcode.JMPLE
|
|
case opcode.JMPLTL:
|
|
return opcode.JMPLT
|
|
case opcode.CALLL:
|
|
return opcode.CALL
|
|
case opcode.ENDTRYL:
|
|
return opcode.ENDTRY
|
|
default:
|
|
panic(fmt.Errorf("invalid opcode: %s", op))
|
|
}
|
|
}
|