|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"encoding/hex"
|
|
|
|
"flag"
|
|
|
|
"fmt"
|
|
|
|
"math/rand"
|
|
|
|
"os"
|
|
|
|
"path"
|
|
|
|
"runtime"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
|
|
"github.com/ethereum/go-ethereum/log"
|
|
|
|
"github.com/harmony-one/bls/ffi/go/bls"
|
|
|
|
|
|
|
|
"github.com/harmony-one/harmony/accounts"
|
|
|
|
"github.com/harmony-one/harmony/accounts/keystore"
|
|
|
|
"github.com/harmony-one/harmony/consensus"
|
|
|
|
"github.com/harmony-one/harmony/core"
|
|
|
|
"github.com/harmony-one/harmony/drand"
|
|
|
|
"github.com/harmony-one/harmony/internal/common"
|
|
|
|
nodeconfig "github.com/harmony-one/harmony/internal/configs/node"
|
|
|
|
"github.com/harmony-one/harmony/internal/ctxerror"
|
|
|
|
"github.com/harmony-one/harmony/internal/genesis"
|
|
|
|
hmykey "github.com/harmony-one/harmony/internal/keystore"
|
|
|
|
"github.com/harmony-one/harmony/internal/memprofiling"
|
|
|
|
"github.com/harmony-one/harmony/internal/profiler"
|
|
|
|
"github.com/harmony-one/harmony/internal/shardchain"
|
|
|
|
"github.com/harmony-one/harmony/internal/utils"
|
|
|
|
"github.com/harmony-one/harmony/node"
|
|
|
|
"github.com/harmony-one/harmony/p2p"
|
|
|
|
"github.com/harmony-one/harmony/p2p/p2pimpl"
|
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
version string
|
|
|
|
builtBy string
|
|
|
|
builtAt string
|
|
|
|
commit string
|
|
|
|
)
|
|
|
|
|
|
|
|
// InitLDBDatabase initializes a LDBDatabase. isGenesis=true will return the beacon chain database for normal shard nodes
|
|
|
|
func InitLDBDatabase(ip string, port string, freshDB bool, isBeacon bool) (*ethdb.LDBDatabase, error) {
|
|
|
|
var dbFileName string
|
|
|
|
if isBeacon {
|
|
|
|
dbFileName = fmt.Sprintf("./db/harmony_beacon_%s_%s", ip, port)
|
|
|
|
} else {
|
|
|
|
dbFileName = fmt.Sprintf("./db/harmony_%s_%s", ip, port)
|
|
|
|
}
|
|
|
|
if freshDB {
|
|
|
|
var err = os.RemoveAll(dbFileName)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Println(err.Error())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ethdb.NewLDBDatabase(dbFileName, 0, 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
func printVersion(me string) {
|
|
|
|
fmt.Fprintf(os.Stderr, "Harmony (C) 2018. %v, version %v-%v (%v %v)\n", path.Base(me), version, commit, builtBy, builtAt)
|
|
|
|
os.Exit(0)
|
|
|
|
}
|
|
|
|
|
|
|
|
func initLogFile(logFolder, role, ip, port string, onlyLogTps bool) {
|
|
|
|
// Setup a logger to stdout and log file.
|
|
|
|
if err := os.MkdirAll(logFolder, 0755); err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
logFileName := fmt.Sprintf("./%v/%s-%v-%v.log", logFolder, role, ip, port)
|
|
|
|
fileHandler := log.Must.FileHandler(logFileName, log.JSONFormat())
|
|
|
|
utils.AddLogHandler(fileHandler)
|
|
|
|
|
|
|
|
if onlyLogTps {
|
|
|
|
matchFilterHandler := log.MatchFilterHandler("msg", "TPS Report", utils.GetLogInstance().GetHandler())
|
|
|
|
utils.GetLogInstance().SetHandler(matchFilterHandler)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
ip = flag.String("ip", "127.0.0.1", "ip of the node")
|
|
|
|
port = flag.String("port", "9000", "port of the node.")
|
|
|
|
logFolder = flag.String("log_folder", "latest", "the folder collecting the logs of this execution")
|
|
|
|
freshDB = flag.Bool("fresh_db", false, "true means the existing disk based db will be removed")
|
|
|
|
profile = flag.Bool("profile", false, "Turn on profiling (CPU, Memory).")
|
|
|
|
metricsReportURL = flag.String("metrics_report_url", "", "If set, reports metrics to this URL.")
|
|
|
|
versionFlag = flag.Bool("version", false, "Output version info")
|
|
|
|
onlyLogTps = flag.Bool("only_log_tps", false, "Only log TPS if true")
|
|
|
|
//Leader needs to have a minimal number of peers to start consensus
|
|
|
|
minPeers = flag.Int("min_peers", 100, "Minimal number of Peers in shard")
|
|
|
|
// Key file to store the private key of staking account.
|
|
|
|
stakingKeyFile = flag.String("staking_key", "./.stakingkey", "the private key file of the harmony node")
|
|
|
|
// Key file to store the private key
|
|
|
|
keyFile = flag.String("key", "./.hmykey", "the p2p key file of the harmony node")
|
|
|
|
// isGenesis indicates this node is a genesis node
|
|
|
|
isGenesis = flag.Bool("is_genesis", true, "true means this node is a genesis node")
|
|
|
|
// isArchival indicates this node is an archival node that will save and archive current blockchain
|
|
|
|
isArchival = flag.Bool("is_archival", false, "true means this node is a archival node")
|
|
|
|
// delayCommit is the commit-delay timer, used by Harmony nodes
|
|
|
|
delayCommit = flag.String("delay_commit", "0ms", "how long to delay sending commit messages in consensus, ex: 500ms, 1s")
|
|
|
|
//isNewNode indicates this node is a new node
|
|
|
|
isNewNode = flag.Bool("is_newnode", false, "true means this node is a new node")
|
|
|
|
shardID = flag.Int("shard_id", -1, "the shard ID of this node")
|
|
|
|
enableMemProfiling = flag.Bool("enableMemProfiling", false, "Enable memsize logging.")
|
|
|
|
enableGC = flag.Bool("enableGC", true, "Enable calling garbage collector manually .")
|
|
|
|
|
|
|
|
// logConn logs incoming/outgoing connections
|
|
|
|
logConn = flag.Bool("log_conn", false, "log incoming/outgoing connections")
|
|
|
|
|
|
|
|
keystoreDir = flag.String("keystore", hmykey.DefaultKeyStoreDir, "The default keystore directory")
|
|
|
|
|
|
|
|
// -nopass is false by default. The keyfile must be encrypted.
|
|
|
|
hmyNoPass = flag.Bool("nopass", false, "No passphrase for the key (testing only)")
|
|
|
|
// -pass takes on "pass:password", "env:var", "file:pathname",
|
|
|
|
// "fd:number", or "stdin" form.
|
|
|
|
// See “PASS PHRASE ARGUMENTS” section of openssl(1) for details.
|
|
|
|
hmyPass = flag.String("pass", "", "how to get passphrase for the key")
|
|
|
|
|
|
|
|
stakingAccounts = flag.String("accounts", "", "account addresses of the node")
|
|
|
|
|
|
|
|
ks *keystore.KeyStore
|
|
|
|
myAccount accounts.Account
|
|
|
|
genesisAccount *genesis.DeployAccount
|
|
|
|
accountIndex int
|
|
|
|
|
|
|
|
// logging verbosity
|
|
|
|
verbosity = flag.Int("verbosity", 5, "Logging verbosity: 0=silent, 1=error, 2=warn, 3=info, 4=debug, 5=detail (default: 5)")
|
|
|
|
|
|
|
|
// dbDir is the database directory.
|
|
|
|
dbDir = flag.String("db_dir", "", "blockchain database directory")
|
|
|
|
|
|
|
|
// Disable view change.
|
|
|
|
disableViewChange = flag.Bool("disable_view_change", false,
|
|
|
|
"Do not propose view change (testing only)")
|
|
|
|
)
|
|
|
|
|
|
|
|
func initSetup() {
|
|
|
|
if *versionFlag {
|
|
|
|
printVersion(os.Args[0])
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set port and ip to global config.
|
|
|
|
nodeconfig.GetDefaultConfig().Port = *port
|
|
|
|
nodeconfig.GetDefaultConfig().IP = *ip
|
|
|
|
|
|
|
|
// Setup mem profiling.
|
|
|
|
memprofiling.GetMemProfiling().Config()
|
|
|
|
|
|
|
|
// Logging setup
|
|
|
|
utils.SetLogContext(*port, *ip)
|
|
|
|
utils.SetLogVerbosity(log.Lvl(*verbosity))
|
|
|
|
|
|
|
|
// Set default keystore Dir
|
|
|
|
hmykey.DefaultKeyStoreDir = *keystoreDir
|
|
|
|
|
|
|
|
// Add GOMAXPROCS to achieve max performance.
|
|
|
|
runtime.GOMAXPROCS(1024)
|
|
|
|
|
|
|
|
// Set up randomization seed.
|
|
|
|
rand.Seed(int64(time.Now().Nanosecond()))
|
|
|
|
|
|
|
|
if len(utils.BootNodes) == 0 {
|
|
|
|
bootNodeAddrs, err := utils.StringsToAddrs(utils.DefaultBootNodeAddrStrings)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
utils.BootNodes = bootNodeAddrs
|
|
|
|
}
|
|
|
|
|
|
|
|
ks = hmykey.GetHmyKeyStore()
|
|
|
|
|
|
|
|
allAccounts := ks.Accounts()
|
|
|
|
|
|
|
|
// TODO: lc try to enable multiple staking accounts per node
|
|
|
|
accountIndex, genesisAccount = genesis.FindAccount(*stakingAccounts)
|
|
|
|
|
|
|
|
if genesisAccount == nil {
|
|
|
|
fmt.Printf("Can't find the account address: %v!\n", *stakingAccounts)
|
|
|
|
os.Exit(100)
|
|
|
|
}
|
|
|
|
|
|
|
|
foundAccount := false
|
|
|
|
for _, account := range allAccounts {
|
|
|
|
if common.ParseAddr(genesisAccount.Address) == account.Address {
|
|
|
|
myAccount = account
|
|
|
|
foundAccount = true
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if !foundAccount {
|
|
|
|
fmt.Printf("Can't find the matching account key: %v!\n", genesisAccount.Address)
|
|
|
|
os.Exit(101)
|
|
|
|
}
|
|
|
|
|
|
|
|
genesisAccount.ShardID = uint32(accountIndex % core.GenesisShardNum)
|
|
|
|
|
|
|
|
fmt.Printf("My Account: %s\n", common.MustAddressToBech32(myAccount.Address))
|
|
|
|
fmt.Printf("Key URL: %s\n", myAccount.URL)
|
|
|
|
fmt.Printf("My Genesis Account: %v\n", *genesisAccount)
|
|
|
|
|
|
|
|
var myPass string
|
|
|
|
if !*hmyNoPass {
|
|
|
|
if *hmyPass == "" {
|
|
|
|
myPass = utils.AskForPassphrase("Passphrase: ")
|
|
|
|
} else if pass, err := utils.GetPassphraseFromSource(*hmyPass); err != nil {
|
|
|
|
fmt.Printf("Cannot read passphrase: %s\n", err)
|
|
|
|
os.Exit(3)
|
|
|
|
} else {
|
|
|
|
myPass = pass
|
|
|
|
}
|
|
|
|
err := ks.Unlock(myAccount, myPass)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Printf("Wrong Passphrase! Unable to unlock account key!\n")
|
|
|
|
os.Exit(3)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set up manual call for garbage collection.
|
|
|
|
if *enableGC {
|
|
|
|
memprofiling.MaybeCallGCPeriodically()
|
|
|
|
}
|
|
|
|
|
|
|
|
hmykey.SetHmyPass(myPass)
|
|
|
|
}
|
|
|
|
|
|
|
|
func createGlobalConfig() *nodeconfig.ConfigType {
|
|
|
|
var err error
|
|
|
|
var myShardID uint32
|
|
|
|
|
|
|
|
nodeConfig := nodeconfig.GetDefaultConfig()
|
|
|
|
|
|
|
|
// Specified Shard ID override calculated Shard ID
|
|
|
|
if *shardID >= 0 {
|
|
|
|
utils.GetLogInstance().Info("ShardID Override", "original", genesisAccount.ShardID, "override", *shardID)
|
|
|
|
genesisAccount.ShardID = uint32(*shardID)
|
|
|
|
}
|
|
|
|
|
|
|
|
if !*isNewNode {
|
|
|
|
nodeConfig = nodeconfig.GetShardConfig(uint32(genesisAccount.ShardID))
|
|
|
|
} else {
|
|
|
|
myShardID = 0 // This should be default value as new node doesn't belong to any shard.
|
|
|
|
if *shardID >= 0 {
|
|
|
|
utils.GetLogInstance().Info("ShardID Override", "original", myShardID, "override", *shardID)
|
|
|
|
myShardID = uint32(*shardID)
|
|
|
|
nodeConfig = nodeconfig.GetShardConfig(myShardID)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// The initial genesis nodes are sequentially put into genesis shards based on their accountIndex
|
|
|
|
nodeConfig.ShardID = uint32(genesisAccount.ShardID)
|
|
|
|
|
|
|
|
// Key Setup ================= [Start]
|
|
|
|
consensusPriKey := &bls.SecretKey{}
|
|
|
|
|
|
|
|
if *isGenesis {
|
|
|
|
err := consensusPriKey.DeserializeHexStr(genesisAccount.BlsPriKey)
|
|
|
|
if err != nil {
|
|
|
|
panic(fmt.Errorf("Failed to parse BLS private key: %s, %s", genesisAccount.BlsPriKey, err))
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// NewNode won't work
|
|
|
|
/*
|
|
|
|
err := consensusPriKey.DeserializeHexStr(genesis.NewNodeAccounts[])
|
|
|
|
if err != nil {
|
|
|
|
panic(fmt.Errorf("generate key error"))
|
|
|
|
}
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
|
|
|
|
// P2p private key is used for secure message transfer between p2p nodes.
|
|
|
|
nodeConfig.P2pPriKey, _, err = utils.LoadKeyFromFile(*keyFile)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Consensus keys are the BLS12-381 keys used to sign consensus messages
|
|
|
|
nodeConfig.ConsensusPriKey, nodeConfig.ConsensusPubKey = consensusPriKey, consensusPriKey.GetPublicKey()
|
|
|
|
if nodeConfig.ConsensusPriKey == nil || nodeConfig.ConsensusPubKey == nil {
|
|
|
|
panic(fmt.Errorf("Failed to initialize BLS keys: %s", consensusPriKey.SerializeToHexStr()))
|
|
|
|
}
|
|
|
|
// Key Setup ================= [End]
|
|
|
|
|
|
|
|
nodeConfig.SelfPeer = p2p.Peer{IP: *ip, Port: *port, ConsensusPubKey: nodeConfig.ConsensusPubKey}
|
|
|
|
|
|
|
|
if accountIndex < core.GenesisShardNum { // The first node in a shard is the leader at genesis
|
|
|
|
nodeConfig.Leader = nodeConfig.SelfPeer
|
|
|
|
nodeConfig.StringRole = "leader"
|
|
|
|
} else {
|
|
|
|
nodeConfig.StringRole = "validator"
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeConfig.Host, err = p2pimpl.NewHost(&nodeConfig.SelfPeer, nodeConfig.P2pPriKey)
|
|
|
|
if *logConn {
|
|
|
|
nodeConfig.Host.GetP2PHost().Network().Notify(utils.ConnLogger)
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
panic("unable to new host in harmony")
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := nodeConfig.Host.AddPeer(&nodeConfig.Leader); err != nil {
|
|
|
|
ctxerror.Warn(utils.GetLogger(), err, "(*p2p.Host).AddPeer failed",
|
|
|
|
"peer", &nodeConfig.Leader)
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeConfig.DBDir = *dbDir
|
|
|
|
|
|
|
|
return nodeConfig
|
|
|
|
}
|
|
|
|
|
|
|
|
func setUpConsensusAndNode(nodeConfig *nodeconfig.ConfigType) *node.Node {
|
|
|
|
// Consensus object.
|
|
|
|
// TODO: consensus object shouldn't start here
|
|
|
|
// TODO(minhdoan): During refactoring, found out that the peers list is actually empty. Need to clean up the logic of consensus later.
|
|
|
|
currentConsensus, err := consensus.New(nodeConfig.Host, nodeConfig.ShardID, nodeConfig.Leader, nodeConfig.ConsensusPriKey)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Fprintf(os.Stderr, "Error :%v \n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
|
|
|
commitDelay, err := time.ParseDuration(*delayCommit)
|
|
|
|
if err != nil || commitDelay < 0 {
|
|
|
|
_, _ = fmt.Fprintf(os.Stderr, "invalid commit delay %#v", *delayCommit)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
|
|
|
currentConsensus.SetCommitDelay(commitDelay)
|
|
|
|
currentConsensus.MinPeers = *minPeers
|
|
|
|
if *disableViewChange {
|
|
|
|
currentConsensus.DisableViewChangeForTestingOnly()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Current node.
|
|
|
|
chainDBFactory := &shardchain.LDBFactory{RootDir: nodeConfig.DBDir}
|
|
|
|
currentNode := node.New(nodeConfig.Host, currentConsensus, chainDBFactory, *isArchival)
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.NewNode)
|
|
|
|
currentNode.StakingAccount = myAccount
|
|
|
|
utils.GetLogInstance().Info("node account set",
|
|
|
|
"address", common.MustAddressToBech32(currentNode.StakingAccount.Address))
|
|
|
|
|
|
|
|
if gsif, err := consensus.NewGenesisStakeInfoFinder(); err == nil {
|
|
|
|
currentConsensus.SetStakeInfoFinder(gsif)
|
|
|
|
} else {
|
|
|
|
_, _ = fmt.Fprintf(os.Stderr, "Cannot initialize stake info: %v\n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: refactor the creation of blockchain out of node.New()
|
|
|
|
currentConsensus.ChainReader = currentNode.Blockchain()
|
|
|
|
|
|
|
|
// TODO: the setup should only based on shard state
|
|
|
|
if *isGenesis {
|
|
|
|
// TODO: need change config file and use switch instead of complicated "if else" condition
|
|
|
|
if nodeConfig.ShardID == 0 { // Beacon chain
|
|
|
|
nodeConfig.SetIsBeacon(true)
|
|
|
|
if nodeConfig.StringRole == "leader" {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.BeaconLeader)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(true)
|
|
|
|
} else {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.BeaconValidator)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(false)
|
|
|
|
}
|
|
|
|
currentNode.NodeConfig.SetShardGroupID(p2p.GroupIDBeacon)
|
|
|
|
currentNode.NodeConfig.SetClientGroupID(p2p.GroupIDBeaconClient)
|
|
|
|
} else {
|
|
|
|
if nodeConfig.StringRole == "leader" {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.ShardLeader)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(true)
|
|
|
|
} else {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.ShardValidator)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(false)
|
|
|
|
}
|
|
|
|
currentNode.NodeConfig.SetShardGroupID(p2p.NewGroupIDByShardID(p2p.ShardID(nodeConfig.ShardID)))
|
|
|
|
currentNode.NodeConfig.SetClientGroupID(p2p.NewClientGroupIDByShardID(p2p.ShardID(nodeConfig.ShardID)))
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if *isNewNode {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.NewNode)
|
|
|
|
currentNode.NodeConfig.SetClientGroupID(p2p.GroupIDBeaconClient)
|
|
|
|
currentNode.NodeConfig.SetBeaconGroupID(p2p.GroupIDBeacon)
|
|
|
|
if *shardID > -1 {
|
|
|
|
// I will be a validator (single leader is fixed for now)
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.ShardValidator)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(false)
|
|
|
|
currentNode.NodeConfig.SetShardGroupID(p2p.NewGroupIDByShardID(p2p.ShardID(nodeConfig.ShardID)))
|
|
|
|
currentNode.NodeConfig.SetClientGroupID(p2p.NewClientGroupIDByShardID(p2p.ShardID(nodeConfig.ShardID)))
|
|
|
|
}
|
|
|
|
} else if nodeConfig.StringRole == "leader" {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.ShardLeader)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(true)
|
|
|
|
currentNode.NodeConfig.SetShardGroupID(p2p.GroupIDUnknown)
|
|
|
|
} else {
|
|
|
|
currentNode.NodeConfig.SetRole(nodeconfig.ShardValidator)
|
|
|
|
currentNode.NodeConfig.SetIsLeader(false)
|
|
|
|
currentNode.NodeConfig.SetShardGroupID(p2p.GroupIDUnknown)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
currentNode.NodeConfig.ConsensusPubKey = nodeConfig.ConsensusPubKey
|
|
|
|
currentNode.NodeConfig.ConsensusPriKey = nodeConfig.ConsensusPriKey
|
|
|
|
|
|
|
|
// Add randomness protocol
|
|
|
|
// TODO: enable drand only for beacon chain
|
|
|
|
// TODO: put this in a better place other than main.
|
|
|
|
// TODO(minhdoan): During refactoring, found out that the peers list is actually empty. Need to clean up the logic of drand later.
|
|
|
|
dRand := drand.New(nodeConfig.Host, nodeConfig.ShardID, []p2p.Peer{}, nodeConfig.Leader, currentNode.ConfirmedBlockChannel, nodeConfig.ConsensusPriKey)
|
|
|
|
currentNode.Consensus.RegisterPRndChannel(dRand.PRndChannel)
|
|
|
|
currentNode.Consensus.RegisterRndChannel(dRand.RndChannel)
|
|
|
|
currentNode.DRand = dRand
|
|
|
|
|
|
|
|
// This needs to be executed after consensus and drand are setup
|
|
|
|
if !*isNewNode || *shardID > -1 { // initial staking new node doesn't need to initialize shard state
|
|
|
|
// TODO: Have a better why to distinguish non-genesis node
|
|
|
|
if err := currentNode.InitShardState(*shardID == -1 && !*isNewNode); err != nil {
|
|
|
|
ctxerror.Crit(utils.GetLogger(), err, "InitShardState failed",
|
|
|
|
"shardID", *shardID, "isNewNode", *isNewNode)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the consensus ID to be the current block number
|
|
|
|
height := currentNode.Blockchain().CurrentBlock().NumberU64()
|
|
|
|
|
|
|
|
currentConsensus.SetViewID(uint32(height))
|
|
|
|
utils.GetLogInstance().Info("Init Blockchain", "height", height)
|
|
|
|
|
|
|
|
// Assign closure functions to the consensus object
|
|
|
|
currentConsensus.BlockVerifier = currentNode.VerifyNewBlock
|
|
|
|
currentConsensus.OnConsensusDone = currentNode.PostConsensusProcessing
|
|
|
|
currentNode.State = node.NodeWaitToJoin
|
|
|
|
|
|
|
|
// Watching currentNode and currentConsensus.
|
|
|
|
memprofiling.GetMemProfiling().Add("currentNode", currentNode)
|
|
|
|
memprofiling.GetMemProfiling().Add("currentConsensus", currentConsensus)
|
|
|
|
return currentNode
|
|
|
|
}
|
|
|
|
|
|
|
|
func main() {
|
|
|
|
flag.Var(&utils.BootNodes, "bootnodes", "a list of bootnode multiaddress (delimited by ,)")
|
|
|
|
flag.Parse()
|
|
|
|
|
|
|
|
// Configure log parameters
|
|
|
|
utils.SetLogContext(*port, *ip)
|
|
|
|
utils.SetLogVerbosity(log.Lvl(*verbosity))
|
|
|
|
|
|
|
|
initSetup()
|
|
|
|
nodeConfig := createGlobalConfig()
|
|
|
|
initLogFile(*logFolder, nodeConfig.StringRole, *ip, *port, *onlyLogTps)
|
|
|
|
|
|
|
|
// Start Profiler for leader if profile argument is on
|
|
|
|
if nodeConfig.StringRole == "leader" && (*profile || *metricsReportURL != "") {
|
|
|
|
prof := profiler.GetProfiler()
|
|
|
|
prof.Config(nodeConfig.ShardID, *metricsReportURL)
|
|
|
|
if *profile {
|
|
|
|
prof.Start()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
currentNode := setUpConsensusAndNode(nodeConfig)
|
|
|
|
//if consensus.ShardID != 0 {
|
|
|
|
// go currentNode.SupportBeaconSyncing()
|
|
|
|
//}
|
|
|
|
|
|
|
|
utils.GetLogInstance().Info("==== New Harmony Node ====",
|
|
|
|
"BlsPubKey", hex.EncodeToString(nodeConfig.ConsensusPubKey.Serialize()),
|
|
|
|
"ShardID", nodeConfig.ShardID,
|
|
|
|
"ShardGroupID", nodeConfig.GetShardGroupID(),
|
|
|
|
"BeaconGroupID", nodeConfig.GetBeaconGroupID(),
|
|
|
|
"ClientGroupID", nodeConfig.GetClientGroupID(),
|
|
|
|
"Role", currentNode.NodeConfig.Role(),
|
|
|
|
"multiaddress", fmt.Sprintf("/ip4/%s/tcp/%s/p2p/%s",
|
|
|
|
*ip, *port, nodeConfig.Host.GetID().Pretty()))
|
|
|
|
|
|
|
|
if *enableMemProfiling {
|
|
|
|
memprofiling.GetMemProfiling().Start()
|
|
|
|
}
|
|
|
|
go currentNode.SupportSyncing()
|
|
|
|
currentNode.ServiceManagerSetup()
|
|
|
|
if err := currentNode.StartRPC(*port); err != nil {
|
|
|
|
ctxerror.Warn(utils.GetLogger(), err, "StartRPC failed")
|
|
|
|
}
|
|
|
|
currentNode.RunServices()
|
|
|
|
currentNode.StartServer()
|
|
|
|
}
|