rclone/rclone.go

638 lines
14 KiB
Go
Raw Normal View History

2013-06-27 21:00:01 +02:00
// Sync files and directories to and from local and remote object stores
2013-06-27 20:51:03 +02:00
//
// Nick Craig-Wood <nick@craig-wood.com>
package main
import (
"fmt"
"log"
"os"
"runtime"
"runtime/pprof"
"strings"
"sync"
"time"
2014-03-27 17:55:29 +01:00
"github.com/ogier/pflag"
"github.com/ncw/rclone/fs"
2013-06-27 21:13:07 +02:00
// Active file systems
_ "github.com/ncw/rclone/drive"
_ "github.com/ncw/rclone/local"
_ "github.com/ncw/rclone/s3"
_ "github.com/ncw/rclone/swift"
)
// Globals
var (
// Flags
2014-03-27 17:55:29 +01:00
cpuprofile = pflag.StringP("cpuprofile", "", "", "Write cpu profile to file")
dry_run = pflag.BoolP("dry-run", "n", false, "Do a trial run with no permanent changes")
statsInterval = pflag.DurationP("stats", "", time.Minute*1, "Interval to print stats")
)
2013-06-28 09:57:32 +02:00
// A pair of fs.Objects
type PairFsObjects struct {
2013-06-28 09:57:32 +02:00
src, dst fs.Object
}
type PairFsObjectsChan chan PairFsObjects
// Check to see if src needs to be copied to dst and if so puts it in out
2013-06-28 09:57:32 +02:00
func checkOne(src, dst fs.Object, out fs.ObjectsChan) {
if dst == nil {
2013-06-28 09:57:32 +02:00
fs.Debug(src, "Couldn't find local file - download")
out <- src
return
}
// Check to see if can store this
if !src.Storable() {
return
}
// Check to see if changed or not
2013-06-27 21:13:07 +02:00
if fs.Equal(src, dst) {
2013-06-28 09:57:32 +02:00
fs.Debug(src, "Unchanged skipping")
return
}
out <- src
}
// Read FsObjects~s on in send to out if they need uploading
//
// FIXME potentially doing lots of MD5SUMS at once
2013-06-28 09:57:32 +02:00
func PairChecker(in PairFsObjectsChan, out fs.ObjectsChan, wg *sync.WaitGroup) {
defer wg.Done()
for pair := range in {
src := pair.src
2013-06-27 21:13:07 +02:00
fs.Stats.Checking(src)
checkOne(src, pair.dst, out)
2013-06-27 21:13:07 +02:00
fs.Stats.DoneChecking(src)
}
}
// Read FsObjects~s on in send to out if they need uploading
//
// FIXME potentially doing lots of MD5SUMS at once
2013-06-28 09:57:32 +02:00
func Checker(in, out fs.ObjectsChan, fdst fs.Fs, wg *sync.WaitGroup) {
defer wg.Done()
for src := range in {
2013-06-27 21:13:07 +02:00
fs.Stats.Checking(src)
dst := fdst.NewFsObject(src.Remote())
checkOne(src, dst, out)
2013-06-27 21:13:07 +02:00
fs.Stats.DoneChecking(src)
}
}
// Read FsObjects on in and copy them
2013-06-28 09:57:32 +02:00
func Copier(in fs.ObjectsChan, fdst fs.Fs, wg *sync.WaitGroup) {
defer wg.Done()
for src := range in {
2013-06-27 21:13:07 +02:00
fs.Stats.Transferring(src)
fs.Copy(fdst, src)
fs.Stats.DoneTransferring(src)
}
}
// Copies fsrc into fdst
2013-06-27 21:13:07 +02:00
func CopyFs(fdst, fsrc fs.Fs) {
err := fdst.Mkdir()
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatal("Failed to make destination")
}
to_be_checked := fsrc.List()
to_be_uploaded := make(fs.ObjectsChan, fs.Config.Transfers)
var checkerWg sync.WaitGroup
checkerWg.Add(fs.Config.Checkers)
for i := 0; i < fs.Config.Checkers; i++ {
go Checker(to_be_checked, to_be_uploaded, fdst, &checkerWg)
}
var copierWg sync.WaitGroup
copierWg.Add(fs.Config.Transfers)
for i := 0; i < fs.Config.Transfers; i++ {
go Copier(to_be_uploaded, fdst, &copierWg)
}
log.Printf("Waiting for checks to finish")
checkerWg.Wait()
close(to_be_uploaded)
log.Printf("Waiting for transfers to finish")
copierWg.Wait()
}
// Delete all the files passed in the channel
2013-06-28 09:57:32 +02:00
func DeleteFiles(to_be_deleted fs.ObjectsChan) {
var wg sync.WaitGroup
wg.Add(fs.Config.Transfers)
for i := 0; i < fs.Config.Transfers; i++ {
go func() {
defer wg.Done()
for dst := range to_be_deleted {
if *dry_run {
2014-03-27 17:55:29 +01:00
fs.Debug(dst, "Not deleting as --dry-run")
} else {
2013-06-27 21:13:07 +02:00
fs.Stats.Checking(dst)
err := dst.Remove()
2013-06-27 21:13:07 +02:00
fs.Stats.DoneChecking(dst)
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2013-06-28 09:57:32 +02:00
fs.Log(dst, "Couldn't delete: %s", err)
} else {
2013-06-28 09:57:32 +02:00
fs.Debug(dst, "Deleted")
}
}
}
}()
}
log.Printf("Waiting for deletions to finish")
wg.Wait()
}
// Syncs fsrc into fdst
2013-06-27 21:13:07 +02:00
func Sync(fdst, fsrc fs.Fs) {
err := fdst.Mkdir()
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatal("Failed to make destination")
}
log.Printf("Building file list")
// Read the destination files first
// FIXME could do this in parallel and make it use less memory
2013-06-28 09:57:32 +02:00
delFiles := make(map[string]fs.Object)
2012-12-31 18:31:19 +01:00
for dst := range fdst.List() {
delFiles[dst.Remote()] = dst
}
// Read source files checking them off against dest files
to_be_checked := make(PairFsObjectsChan, fs.Config.Transfers)
to_be_uploaded := make(fs.ObjectsChan, fs.Config.Transfers)
var checkerWg sync.WaitGroup
checkerWg.Add(fs.Config.Checkers)
for i := 0; i < fs.Config.Checkers; i++ {
go PairChecker(to_be_checked, to_be_uploaded, &checkerWg)
}
var copierWg sync.WaitGroup
copierWg.Add(fs.Config.Transfers)
for i := 0; i < fs.Config.Transfers; i++ {
go Copier(to_be_uploaded, fdst, &copierWg)
}
go func() {
for src := range fsrc.List() {
remote := src.Remote()
dst, found := delFiles[remote]
if found {
delete(delFiles, remote)
to_be_checked <- PairFsObjects{src, dst}
} else {
// No need to check doesn't exist
to_be_uploaded <- src
}
}
close(to_be_checked)
}()
log.Printf("Waiting for checks to finish")
checkerWg.Wait()
close(to_be_uploaded)
log.Printf("Waiting for transfers to finish")
copierWg.Wait()
2013-06-27 21:13:07 +02:00
if fs.Stats.Errored() {
log.Printf("Not deleting files as there were IO errors")
return
}
// Delete the spare files
toDelete := make(fs.ObjectsChan, fs.Config.Transfers)
go func() {
for _, fs := range delFiles {
toDelete <- fs
}
close(toDelete)
}()
DeleteFiles(toDelete)
}
2012-12-31 18:31:19 +01:00
// Checks the files in fsrc and fdst according to Size and MD5SUM
2013-06-27 21:13:07 +02:00
func Check(fdst, fsrc fs.Fs) {
log.Printf("Building file list")
2012-12-31 18:31:19 +01:00
// Read the destination files first
// FIXME could do this in parallel and make it use less memory
2013-06-28 09:57:32 +02:00
dstFiles := make(map[string]fs.Object)
2012-12-31 18:31:19 +01:00
for dst := range fdst.List() {
dstFiles[dst.Remote()] = dst
}
// Read the source files checking them against dstFiles
// FIXME could do this in parallel and make it use less memory
2013-06-28 09:57:32 +02:00
srcFiles := make(map[string]fs.Object)
commonFiles := make(map[string][]fs.Object)
2012-12-31 18:31:19 +01:00
for src := range fsrc.List() {
remote := src.Remote()
if dst, ok := dstFiles[remote]; ok {
2013-06-28 09:57:32 +02:00
commonFiles[remote] = []fs.Object{dst, src}
2012-12-31 18:31:19 +01:00
delete(dstFiles, remote)
} else {
srcFiles[remote] = src
}
}
log.Printf("Files in %s but not in %s", fdst, fsrc)
for remote := range dstFiles {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2012-12-31 18:31:19 +01:00
log.Printf(remote)
}
log.Printf("Files in %s but not in %s", fsrc, fdst)
for remote := range srcFiles {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2012-12-31 18:31:19 +01:00
log.Printf(remote)
}
checks := make(chan []fs.Object, fs.Config.Transfers)
2012-12-31 18:31:19 +01:00
go func() {
for _, check := range commonFiles {
checks <- check
}
close(checks)
}()
var checkerWg sync.WaitGroup
checkerWg.Add(fs.Config.Checkers)
for i := 0; i < fs.Config.Checkers; i++ {
2012-12-31 18:31:19 +01:00
go func() {
defer checkerWg.Done()
for check := range checks {
dst, src := check[0], check[1]
2013-06-27 21:13:07 +02:00
fs.Stats.Checking(src)
2012-12-31 18:31:19 +01:00
if src.Size() != dst.Size() {
2013-06-27 21:13:07 +02:00
fs.Stats.DoneChecking(src)
fs.Stats.Error()
2013-06-28 09:57:32 +02:00
fs.Log(src, "Sizes differ")
2012-12-31 18:31:19 +01:00
continue
}
2013-06-27 21:13:07 +02:00
same, err := fs.CheckMd5sums(src, dst)
fs.Stats.DoneChecking(src)
2012-12-31 18:31:19 +01:00
if err != nil {
continue
}
if !same {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2013-06-28 09:57:32 +02:00
fs.Log(src, "Md5sums differ")
2012-12-31 18:31:19 +01:00
}
2013-06-28 09:57:32 +02:00
fs.Debug(src, "OK")
2012-12-31 18:31:19 +01:00
}
}()
}
log.Printf("Waiting for checks to finish")
checkerWg.Wait()
2013-07-02 09:33:17 +02:00
log.Printf("%d differences found", fs.Stats.GetErrors())
2012-12-31 18:31:19 +01:00
}
// List the Fs to stdout
//
// Lists in parallel which may get them out of order
2013-06-27 21:13:07 +02:00
func List(f, _ fs.Fs) {
in := f.List()
var wg sync.WaitGroup
wg.Add(fs.Config.Checkers)
for i := 0; i < fs.Config.Checkers; i++ {
go func() {
defer wg.Done()
2013-06-27 21:13:07 +02:00
for o := range in {
fs.Stats.Checking(o)
modTime := o.ModTime()
fs.Stats.DoneChecking(o)
fmt.Printf("%9d %19s %s\n", o.Size(), modTime.Format("2006-01-02 15:04:05.00000000"), o.Remote())
}
}()
}
wg.Wait()
}
2013-01-23 23:43:20 +01:00
// List the directories/buckets/containers in the Fs to stdout
2013-06-27 21:13:07 +02:00
func ListDir(f, _ fs.Fs) {
2013-01-23 23:43:20 +01:00
for dir := range f.ListDir() {
fmt.Printf("%12d %13s %9d %s\n", dir.Bytes, dir.When.Format("2006-01-02 15:04:05"), dir.Count, dir.Name)
2012-12-04 00:58:17 +01:00
}
}
// Makes a destination directory or container
2013-06-27 21:13:07 +02:00
func mkdir(fdst, fsrc fs.Fs) {
err := fdst.Mkdir()
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatalf("Mkdir failed: %s", err)
}
}
// Removes a container but not if not empty
2013-06-27 21:13:07 +02:00
func rmdir(fdst, fsrc fs.Fs) {
if *dry_run {
2014-03-27 17:55:29 +01:00
log.Printf("Not deleting %s as --dry-run", fdst)
} else {
err := fdst.Rmdir()
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatalf("Rmdir failed: %s", err)
}
}
}
// Removes a container and all of its contents
//
// FIXME doesn't delete local directories
2013-06-27 21:13:07 +02:00
func purge(fdst, fsrc fs.Fs) {
if f, ok := fdst.(fs.Purger); ok {
err := f.Purge()
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatalf("Purge failed: %s", err)
}
} else {
DeleteFiles(fdst.List())
log.Printf("Deleting path")
rmdir(fdst, fsrc)
}
}
// Edits the config file
func EditConfig(fdst, fsrc fs.Fs) {
fs.EditConfig()
}
2012-12-04 00:58:17 +01:00
type Command struct {
Name string
Help string
ArgsHelp string
Run func(fdst, fsrc fs.Fs)
MinArgs int
MaxArgs int
NoStats bool
2012-12-04 00:58:17 +01:00
}
// checkArgs checks there are enough arguments and prints a message if not
func (cmd *Command) checkArgs(args []string) {
if len(args) < cmd.MinArgs {
2012-12-04 00:58:17 +01:00
syntaxError()
fmt.Fprintf(os.Stderr, "Command %s needs %d arguments mininum\n", cmd.Name, cmd.MinArgs)
2012-12-04 00:58:17 +01:00
os.Exit(1)
} else if len(args) > cmd.MaxArgs {
2012-12-04 00:58:17 +01:00
syntaxError()
fmt.Fprintf(os.Stderr, "Command %s needs %d arguments maximum\n", cmd.Name, cmd.MaxArgs)
2012-12-04 00:58:17 +01:00
os.Exit(1)
}
}
var Commands = []Command{
{
Name: "copy",
ArgsHelp: "source://path dest://path",
Help: `
Copy the source to the destination. Doesn't transfer
unchanged files, testing first by modification time then by
MD5SUM. Doesn't delete files from the destination.`,
Run: CopyFs,
MinArgs: 2,
MaxArgs: 2,
},
{
Name: "sync",
ArgsHelp: "source://path dest://path",
Help: `
Sync the source to the destination. Doesn't transfer
unchanged files, testing first by modification time then by
MD5SUM. Deletes any files that exist in source that don't
exist in destination. Since this can cause data loss, test
2014-03-27 17:55:29 +01:00
first with the --dry-run flag.`,
Run: Sync,
MinArgs: 2,
MaxArgs: 2,
2012-12-04 00:58:17 +01:00
},
{
Name: "ls",
ArgsHelp: "[remote://path]",
Help: `
2013-01-23 23:43:20 +01:00
List all the objects in the the path.`,
Run: List,
MinArgs: 1,
MaxArgs: 1,
2013-01-23 23:43:20 +01:00
},
{
Name: "lsd",
ArgsHelp: "[remote://path]",
Help: `
2013-01-23 23:43:20 +01:00
List all directoryes/objects/buckets in the the path.`,
Run: ListDir,
MinArgs: 1,
MaxArgs: 1,
2012-12-04 00:58:17 +01:00
},
{
Name: "mkdir",
ArgsHelp: "remote://path",
Help: `
2012-12-31 18:31:19 +01:00
Make the path if it doesn't already exist`,
Run: mkdir,
MinArgs: 1,
MaxArgs: 1,
2012-12-04 00:58:17 +01:00
},
{
Name: "rmdir",
ArgsHelp: "remote://path",
Help: `
Remove the path. Note that you can't remove a path with
2012-12-31 18:31:19 +01:00
objects in it, use purge for that.`,
Run: rmdir,
MinArgs: 1,
MaxArgs: 1,
2012-12-04 00:58:17 +01:00
},
{
Name: "purge",
ArgsHelp: "remote://path",
Help: `
2012-12-31 18:31:19 +01:00
Remove the path and all of its contents.`,
Run: purge,
MinArgs: 1,
MaxArgs: 1,
},
2012-12-31 18:31:19 +01:00
{
Name: "check",
ArgsHelp: "source://path dest://path",
Help: `
2012-12-31 18:31:19 +01:00
Checks the files in the source and destination match. It
compares sizes and MD5SUMs and prints a report of files which
don't match. It doesn't alter the source or destination.`,
Run: Check,
MinArgs: 2,
MaxArgs: 2,
2012-12-31 18:31:19 +01:00
},
{
Name: "config",
Help: `
Enter an interactive configuration session.`,
Run: EditConfig,
NoStats: true,
},
{
Name: "help",
Help: `
2012-12-31 18:31:19 +01:00
This help.`,
NoStats: true,
2012-12-31 18:31:19 +01:00
},
2012-12-04 00:58:17 +01:00
}
// syntaxError prints the syntax
func syntaxError() {
2013-06-27 21:00:01 +02:00
fmt.Fprintf(os.Stderr, `Sync files and directories to and from local and remote object stores
2012-12-04 00:58:17 +01:00
Syntax: [options] subcommand <parameters> <parameters...>
Subcommands:
2012-12-04 00:58:17 +01:00
`)
for i := range Commands {
cmd := &Commands[i]
fmt.Fprintf(os.Stderr, " %s %s\n", cmd.Name, cmd.ArgsHelp)
fmt.Fprintf(os.Stderr, "%s\n\n", cmd.Help)
2012-12-04 00:58:17 +01:00
}
fmt.Fprintf(os.Stderr, "Options:\n")
2014-03-27 17:55:29 +01:00
pflag.PrintDefaults()
2012-12-04 00:58:17 +01:00
fmt.Fprintf(os.Stderr, `
It is only necessary to use a unique prefix of the subcommand, eg 'up' for 'upload'.
`)
}
// Exit with the message
func fatal(message string, args ...interface{}) {
syntaxError()
fmt.Fprintf(os.Stderr, message, args...)
os.Exit(1)
}
// Parse the command line flags
func ParseFlags() {
2014-03-27 17:55:29 +01:00
pflag.Usage = syntaxError
pflag.Parse()
runtime.GOMAXPROCS(runtime.NumCPU())
fs.LoadConfig()
2013-06-29 11:43:52 +02:00
// Setup profiling if desired
if *cpuprofile != "" {
f, err := os.Create(*cpuprofile)
if err != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
log.Fatal(err)
}
pprof.StartCPUProfile(f)
defer pprof.StopCPUProfile()
}
}
// Parse the command from the command line
func ParseCommand() (*Command, []string) {
args := pflag.Args()
if len(args) < 1 {
fatal("No command supplied\n")
}
2012-12-04 00:58:17 +01:00
cmd := strings.ToLower(args[0])
args = args[1:]
2012-12-04 00:58:17 +01:00
// Find the command doing a prefix match
var command *Command
2012-12-04 00:58:17 +01:00
for i := range Commands {
trialCommand := &Commands[i]
2012-12-04 00:58:17 +01:00
// exact command name found - use that
if trialCommand.Name == cmd {
command = trialCommand
2012-12-04 00:58:17 +01:00
break
} else if strings.HasPrefix(trialCommand.Name, cmd) {
if command != nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2012-12-04 00:58:17 +01:00
log.Fatalf("Not unique - matches multiple commands %q", cmd)
}
command = trialCommand
}
}
if command == nil {
2013-06-27 21:13:07 +02:00
fs.Stats.Error()
2012-12-04 00:58:17 +01:00
log.Fatalf("Unknown command %q", cmd)
}
if command.Run == nil {
syntaxError()
}
command.checkArgs(args)
return command, args
}
// Create a Fs from a name
func NewFs(remote string) fs.Fs {
f, err := fs.NewFs(remote)
if err != nil {
fs.Stats.Error()
log.Fatalf("Failed to create file system for %q: %v", remote, err)
}
return f
}
// Print the stats every statsInterval
func StartStats() {
go func() {
ch := time.Tick(*statsInterval)
for {
<-ch
fs.Stats.Log()
}
}()
}
func main() {
ParseFlags()
command, args := ParseCommand()
// Make source and destination fs
2013-06-27 21:13:07 +02:00
var fdst, fsrc fs.Fs
if len(args) >= 1 {
fdst = NewFs(args[0])
}
if len(args) >= 2 {
fsrc = fdst
fdst = NewFs(args[1])
}
fs.CalculateModifyWindow(fdst, fsrc)
if !command.NoStats {
StartStats()
}
// Run the actual command
if command.Run != nil {
command.Run(fdst, fsrc)
if !command.NoStats {
fmt.Println(fs.Stats)
}
if fs.Config.Verbose {
log.Printf("*** Go routines at exit %d\n", runtime.NumGoroutine())
}
2013-06-27 21:13:07 +02:00
if fs.Stats.Errored() {
os.Exit(1)
}
os.Exit(0)
2012-12-31 18:31:19 +01:00
} else {
}
}