You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
116 lines
3.2 KiB
116 lines
3.2 KiB
package main
|
|
|
|
import (
|
|
"flag"
|
|
"fmt"
|
|
"math/rand"
|
|
"os"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/simple-rules/harmony-benchmark/attack"
|
|
"github.com/simple-rules/harmony-benchmark/configr"
|
|
"github.com/simple-rules/harmony-benchmark/consensus"
|
|
"github.com/simple-rules/harmony-benchmark/log"
|
|
"github.com/simple-rules/harmony-benchmark/node"
|
|
"github.com/simple-rules/harmony-benchmark/utils"
|
|
)
|
|
|
|
const (
|
|
AttackProbability = 20
|
|
)
|
|
|
|
func attackDetermination(attackedMode int) bool {
|
|
switch attackedMode {
|
|
case 0:
|
|
return false
|
|
case 1:
|
|
return true
|
|
case 2:
|
|
return rand.Intn(100) < AttackProbability
|
|
}
|
|
return false
|
|
}
|
|
|
|
func startProfiler(shardID string, logFolder string) {
|
|
err := utils.RunCmd("./bin/profiler", "-pid", strconv.Itoa(os.Getpid()), "-shard_id", shardID, "-log_folder", logFolder)
|
|
if err != nil {
|
|
log.Error("Failed to start profiler")
|
|
}
|
|
}
|
|
|
|
func main() {
|
|
ip := flag.String("ip", "127.0.0.1", "IP of the node")
|
|
port := flag.String("port", "9000", "port of the node.")
|
|
configFile := flag.String("config_file", "config.txt", "file containing all ip addresses")
|
|
logFolder := flag.String("log_folder", "latest", "the folder collecting the logs of this execution")
|
|
attackedMode := flag.Int("attacked_mode", 0, "0 means not attacked, 1 means attacked, 2 means being open to be selected as attacked")
|
|
flag.Parse()
|
|
|
|
// Set up randomization seed.
|
|
rand.Seed(int64(time.Now().Nanosecond()))
|
|
|
|
// Attack determination.
|
|
attack.GetInstance().SetAttackEnabled(attackDetermination(*attackedMode))
|
|
|
|
configr := configr.NewConfigr()
|
|
configr.ReadConfigFile(*configFile)
|
|
shardID := configr.GetShardID(*ip, *port)
|
|
peers := configr.GetPeers(*ip, *port, shardID)
|
|
leader := configr.GetLeader(shardID)
|
|
|
|
var role string
|
|
if leader.Ip == *ip && leader.Port == *port {
|
|
role = "leader"
|
|
} else {
|
|
role = "validator"
|
|
}
|
|
|
|
// Setup a logger to stdout and log file.
|
|
logFileName := fmt.Sprintf("./%v/%s-%v-%v.log", *logFolder, role, *ip, *port)
|
|
h := log.MultiHandler(
|
|
log.StdoutHandler,
|
|
log.Must.FileHandler(logFileName, log.JSONFormat()), // Log to file
|
|
// log.Must.NetHandler("tcp", ":3000", log.JSONFormat()) // Log to remote
|
|
)
|
|
log.Root().SetHandler(h)
|
|
|
|
// Consensus object.
|
|
consensus := consensus.NewConsensus(*ip, *port, shardID, peers, leader)
|
|
|
|
// Start Profiler for leader
|
|
if role == "leader" {
|
|
startProfiler(shardID, *logFolder)
|
|
}
|
|
|
|
// Set logger to attack model.
|
|
attack.GetInstance().SetLogger(consensus.Log)
|
|
// Current node.
|
|
currentNode := node.New(consensus)
|
|
// Create client peer.
|
|
clientPeer := configr.GetClientPeer()
|
|
// If there is a client configured in the node list.
|
|
if clientPeer != nil {
|
|
currentNode.ClientPeer = clientPeer
|
|
}
|
|
|
|
// Assign closure functions to the consensus object
|
|
consensus.BlockVerifier = currentNode.VerifyNewBlock
|
|
consensus.OnConsensusDone = currentNode.PostConsensusProcessing
|
|
|
|
// Temporary testing code, to be removed.
|
|
currentNode.AddTestingAddresses(10000)
|
|
|
|
if consensus.IsLeader {
|
|
// Let consensus run
|
|
go func() {
|
|
consensus.WaitForNewBlock(currentNode.BlockChannel)
|
|
}()
|
|
// Node waiting for consensus readiness to create new block
|
|
go func() {
|
|
currentNode.WaitForConsensusReady(consensus.ReadySignal)
|
|
}()
|
|
}
|
|
|
|
currentNode.StartServer(*port)
|
|
}
|
|
|