package main import ( "fmt" "github.com/urfave/cli" "github.com/zrepl/zrepl/jobrun" "github.com/zrepl/zrepl/rpc" "github.com/zrepl/zrepl/sshbytestream" "github.com/zrepl/zrepl/zfs" "golang.org/x/sys/unix" "io" "log" "net/http" _ "net/http/pprof" "os" "runtime/debug" "sync" "time" ) type Logger interface { Printf(format string, v ...interface{}) } var conf Config var runner *jobrun.JobRunner var logFlags int = log.LUTC | log.Ldate | log.Ltime var logOut io.Writer var defaultLog Logger func main() { defer func() { e := recover() if e != nil { defaultLog.Printf("panic:\n%s\n\n", debug.Stack()) defaultLog.Printf("error: %t %s", e, e) os.Exit(1) } }() app := cli.NewApp() app.Name = "zrepl" app.Usage = "replicate zfs datasets" app.EnableBashCompletion = true app.Flags = []cli.Flag{ cli.StringFlag{Name: "config"}, cli.StringFlag{Name: "logfile"}, cli.StringFlag{Name: "debug.pprof.http"}, } app.Before = func(c *cli.Context) (err error) { // Logging if c.GlobalIsSet("logfile") { var logFile *os.File logFile, err = os.OpenFile(c.String("logfile"), os.O_CREATE|os.O_TRUNC|os.O_WRONLY, 0600) if err != nil { return } if err = unix.Dup2(int(logFile.Fd()), int(os.Stderr.Fd())); err != nil { logFile.WriteString(fmt.Sprintf("error duping logfile to stderr: %s\n", err)) return } logOut = logFile } else { logOut = os.Stderr } defaultLog = log.New(logOut, "", logFlags) // CPU profiling if c.GlobalIsSet("debug.pprof.http") { go func() { http.ListenAndServe(c.GlobalString("debug.pprof.http"), nil) }() } // Config if !c.GlobalIsSet("config") { return cli.NewExitError("config flag not set", 2) } if conf, err = ParseConfig(c.GlobalString("config")); err != nil { return cli.NewExitError(err, 2) } jobrunLogger := log.New(os.Stderr, "jobrun ", logFlags) runner = jobrun.NewJobRunner(jobrunLogger) return } app.Commands = []cli.Command{ { Name: "stdinserver", Aliases: []string{"s"}, Usage: "start in stdin server mode (from authorized keys)", Flags: []cli.Flag{ cli.StringFlag{Name: "identity"}, }, Action: cmdStdinServer, }, { Name: "run", Aliases: []string{"r"}, Usage: "do replication", Action: cmdRun, Flags: []cli.Flag{ cli.StringFlag{Name: "job"}, }, }, } app.Run(os.Args) } func cmdStdinServer(c *cli.Context) (err error) { if !c.IsSet("identity") { return cli.NewExitError("identity flag not set", 2) } identity := c.String("identity") var sshByteStream io.ReadWriteCloser if sshByteStream, err = sshbytestream.Incoming(); err != nil { return } findMapping := func(cm []ClientMapping, identity string) zfs.DatasetMapping { for i := range cm { if cm[i].From == identity { return cm[i].Mapping } } return nil } sinkMapping := func(identity string) (sink zfs.DatasetMapping, err error) { if sink = findMapping(conf.Sinks, identity); sink == nil { return nil, fmt.Errorf("could not find sink for dataset") } return } sinkLogger := log.New(logOut, fmt.Sprintf("sink[%s] ", identity), logFlags) handler := Handler{ Logger: sinkLogger, SinkMappingFunc: sinkMapping, PullACL: findMapping(conf.PullACLs, identity), } if err = rpc.ListenByteStreamRPC(sshByteStream, identity, handler, sinkLogger); err != nil { //os.Exit(1) err = cli.NewExitError(err, 1) defaultLog.Printf("listenbytestreamerror: %#v\n", err) } return } func cmdRun(c *cli.Context) error { var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() runner.Start() }() jobs := make([]jobrun.Job, len(conf.Pulls)+len(conf.Pushs)) i := 0 for _, pull := range conf.Pulls { jobs[i] = jobrun.Job{ Name: fmt.Sprintf("pull.%d", i), RepeatStrategy: pull.RepeatStrategy, RunFunc: func(log jobrun.Logger) error { log.Printf("doing pull: %v", pull) return jobPull(pull, c, log) }, } i++ } for _, push := range conf.Pushs { jobs[i] = jobrun.Job{ Name: fmt.Sprintf("push.%d", i), RepeatStrategy: push.RepeatStrategy, RunFunc: func(log jobrun.Logger) error { log.Printf("doing push: %v", push) return jobPush(push, c, log) }, } i++ } for _, j := range jobs { if c.IsSet("job") { if c.String("job") == j.Name { runner.AddJob(j) } continue } runner.AddJob(j) } for { select { case job := <-runner.NotificationChan(): log.Printf("job %s reported error: %v\n", job.Name, job.LastError) } } wg.Wait() return nil } func jobPull(pull Pull, c *cli.Context, log jobrun.Logger) (err error) { if lt, ok := pull.From.Transport.(LocalTransport); ok { lt.SetHandler(Handler{ Logger: log, PullACL: pull.Mapping, }) pull.From.Transport = lt log.Printf("fixing up local transport: %#v", pull.From.Transport) } var remote rpc.RPCRequester if remote, err = pull.From.Transport.Connect(log); err != nil { return } defer closeRPCWithTimeout(log, remote, time.Second*10, "") return doPull(PullContext{remote, log, pull.Mapping, pull.InitialReplPolicy}) } func jobPush(push Push, c *cli.Context, log jobrun.Logger) (err error) { if _, ok := push.To.Transport.(LocalTransport); ok { panic("no support for local pushs") } var remote rpc.RPCRequester if remote, err = push.To.Transport.Connect(log); err != nil { return err } defer closeRPCWithTimeout(log, remote, time.Second*10, "") log.Printf("building handler for PullMeRequest") handler := Handler{ Logger: log, PullACL: push.Filter, SinkMappingFunc: nil, // no need for that in the handler for PullMe } log.Printf("handler: %#v", handler) r := rpc.PullMeRequest{ InitialReplPolicy: push.InitialReplPolicy, } log.Printf("doing PullMeRequest: %#v", r) if err = remote.PullMeRequest(r, handler); err != nil { log.Printf("PullMeRequest failed: %s", err) return } log.Printf("push job finished") return }