mirror of
https://github.com/zrepl/zrepl.git
synced 2025-02-19 20:01:26 +01:00
parent
f76a0dec6d
commit
aa7a13a1c1
59
cmd/prune.go
59
cmd/prune.go
@ -3,7 +3,9 @@ package cmd
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"github.com/zrepl/zrepl/util"
|
||||||
"github.com/zrepl/zrepl/zfs"
|
"github.com/zrepl/zrepl/zfs"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -23,9 +25,8 @@ type PruneResult struct {
|
|||||||
Remove []zfs.FilesystemVersion
|
Remove []zfs.FilesystemVersion
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// FIXME must not call p.task.Enter because it runs in parallel
|
||||||
func (p *Pruner) filterFilesystems() (filesystems []*zfs.DatasetPath, stop bool) {
|
func (p *Pruner) filterFilesystems() (filesystems []*zfs.DatasetPath, stop bool) {
|
||||||
p.task.Enter("filter_fs")
|
|
||||||
defer p.task.Finish()
|
|
||||||
filesystems, err := zfs.ZFSListMapping(p.DatasetFilter)
|
filesystems, err := zfs.ZFSListMapping(p.DatasetFilter)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
p.task.Log().WithError(err).Error("error applying filesystem filter")
|
p.task.Log().WithError(err).Error("error applying filesystem filter")
|
||||||
@ -38,9 +39,8 @@ func (p *Pruner) filterFilesystems() (filesystems []*zfs.DatasetPath, stop bool)
|
|||||||
return filesystems, false
|
return filesystems, false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// FIXME must not call p.task.Enter because it runs in parallel
|
||||||
func (p *Pruner) filterVersions(fs *zfs.DatasetPath) (fsversions []zfs.FilesystemVersion, stop bool) {
|
func (p *Pruner) filterVersions(fs *zfs.DatasetPath) (fsversions []zfs.FilesystemVersion, stop bool) {
|
||||||
p.task.Enter("filter_versions")
|
|
||||||
defer p.task.Finish()
|
|
||||||
log := p.task.Log().WithField(logFSField, fs.ToString())
|
log := p.task.Log().WithField(logFSField, fs.ToString())
|
||||||
|
|
||||||
filter := NewPrefixFilter(p.SnapshotPrefix)
|
filter := NewPrefixFilter(p.SnapshotPrefix)
|
||||||
@ -56,9 +56,8 @@ func (p *Pruner) filterVersions(fs *zfs.DatasetPath) (fsversions []zfs.Filesyste
|
|||||||
return fsversions, false
|
return fsversions, false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Pruner) pruneFilesystem(fs *zfs.DatasetPath) (r PruneResult, valid bool) {
|
// FIXME must not call p.task.Enter because it runs in parallel
|
||||||
p.task.Enter("prune_fs")
|
func (p *Pruner) pruneFilesystem(fs *zfs.DatasetPath, destroySemaphore util.Semaphore) (r PruneResult, valid bool) {
|
||||||
defer p.task.Finish()
|
|
||||||
log := p.task.Log().WithField(logFSField, fs.ToString())
|
log := p.task.Log().WithField(logFSField, fs.ToString())
|
||||||
|
|
||||||
fsversions, stop := p.filterVersions(fs)
|
fsversions, stop := p.filterVersions(fs)
|
||||||
@ -66,9 +65,7 @@ func (p *Pruner) pruneFilesystem(fs *zfs.DatasetPath) (r PruneResult, valid bool
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
p.task.Enter("prune_policy")
|
|
||||||
keep, remove, err := p.PrunePolicy.Prune(fs, fsversions)
|
keep, remove, err := p.PrunePolicy.Prune(fs, fsversions)
|
||||||
p.task.Finish()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.WithError(err).Error("error evaluating prune policy")
|
log.WithError(err).Error("error evaluating prune policy")
|
||||||
return
|
return
|
||||||
@ -81,8 +78,13 @@ func (p *Pruner) pruneFilesystem(fs *zfs.DatasetPath) (r PruneResult, valid bool
|
|||||||
|
|
||||||
r = PruneResult{fs, fsversions, keep, remove}
|
r = PruneResult{fs, fsversions, keep, remove}
|
||||||
|
|
||||||
makeFields := func(v zfs.FilesystemVersion) (fields map[string]interface{}) {
|
var wg sync.WaitGroup
|
||||||
fields = make(map[string]interface{})
|
for v := range remove {
|
||||||
|
wg.Add(1)
|
||||||
|
go func(v zfs.FilesystemVersion) {
|
||||||
|
defer wg.Done()
|
||||||
|
// log fields
|
||||||
|
fields := make(map[string]interface{})
|
||||||
fields["version"] = v.ToAbsPath(fs)
|
fields["version"] = v.ToAbsPath(fs)
|
||||||
timeSince := v.Creation.Sub(p.Now)
|
timeSince := v.Creation.Sub(p.Now)
|
||||||
fields["age_ns"] = timeSince
|
fields["age_ns"] = timeSince
|
||||||
@ -90,24 +92,23 @@ func (p *Pruner) pruneFilesystem(fs *zfs.DatasetPath) (r PruneResult, valid bool
|
|||||||
days := timeSince / day
|
days := timeSince / day
|
||||||
remainder := timeSince % day
|
remainder := timeSince % day
|
||||||
fields["age_str"] = fmt.Sprintf("%dd%s", days, remainder)
|
fields["age_str"] = fmt.Sprintf("%dd%s", days, remainder)
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, v := range remove {
|
|
||||||
fields := makeFields(v)
|
|
||||||
log.WithFields(fields).Info("destroying version")
|
log.WithFields(fields).Info("destroying version")
|
||||||
// echo what we'll do and exec zfs destroy if not dry run
|
// echo what we'll do and exec zfs destroy if not dry run
|
||||||
// TODO special handling for EBUSY (zfs hold)
|
// TODO special handling for EBUSY (zfs hold)
|
||||||
// TODO error handling for clones? just echo to cli, skip over, and exit with non-zero status code (we're idempotent)
|
// TODO error handling for clones? just echo to cli, skip over, and exit with non-zero status code (we're idempotent)
|
||||||
if !p.DryRun {
|
if !p.DryRun {
|
||||||
p.task.Enter("destroy")
|
destroySemaphore.Down()
|
||||||
err := zfs.ZFSDestroyFilesystemVersion(fs, v)
|
err := zfs.ZFSDestroyFilesystemVersion(fs, v)
|
||||||
p.task.Finish()
|
destroySemaphore.Up()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.WithFields(fields).WithError(err).Error("error destroying version")
|
log.WithFields(fields).WithError(err).Error("error destroying version")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}(remove[v])
|
||||||
}
|
}
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
return r, true
|
return r, true
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -124,13 +125,31 @@ func (p *Pruner) Run(ctx context.Context) (r []PruneResult, err error) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
r = make([]PruneResult, 0, len(filesystems))
|
maxConcurrentDestroy := len(filesystems)
|
||||||
|
p.task.Log().WithField("max_concurrent_destroy", maxConcurrentDestroy).Info("begin concurrent destroy")
|
||||||
|
destroySem := util.NewSemaphore(maxConcurrentDestroy)
|
||||||
|
|
||||||
|
resChan := make(chan PruneResult, len(filesystems))
|
||||||
|
var wg sync.WaitGroup
|
||||||
for _, fs := range filesystems {
|
for _, fs := range filesystems {
|
||||||
res, ok := p.pruneFilesystem(fs)
|
wg.Add(1)
|
||||||
|
go func(fs *zfs.DatasetPath) {
|
||||||
|
defer wg.Done()
|
||||||
|
res, ok := p.pruneFilesystem(fs, destroySem)
|
||||||
if ok {
|
if ok {
|
||||||
r = append(r, res)
|
resChan <- res
|
||||||
}
|
}
|
||||||
|
|
||||||
|
}(fs)
|
||||||
|
}
|
||||||
|
wg.Wait()
|
||||||
|
close(resChan)
|
||||||
|
|
||||||
|
p.task.Log().Info("destroys done")
|
||||||
|
|
||||||
|
r = make([]PruneResult, 0, len(filesystems))
|
||||||
|
for res := range resChan {
|
||||||
|
r = append(r, res)
|
||||||
}
|
}
|
||||||
|
|
||||||
return
|
return
|
||||||
|
Loading…
Reference in New Issue
Block a user