2018-08-29 19:00:45 +02:00
|
|
|
package pruner
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
2018-08-30 11:49:06 +02:00
|
|
|
"fmt"
|
2018-08-30 17:40:45 +02:00
|
|
|
"github.com/pkg/errors"
|
2018-09-08 07:03:41 +02:00
|
|
|
"github.com/prometheus/client_golang/prometheus"
|
2018-08-30 17:40:45 +02:00
|
|
|
"github.com/zrepl/zrepl/config"
|
2018-08-30 11:49:06 +02:00
|
|
|
"github.com/zrepl/zrepl/logger"
|
2018-08-29 19:00:45 +02:00
|
|
|
"github.com/zrepl/zrepl/pruning"
|
|
|
|
"github.com/zrepl/zrepl/replication/pdu"
|
2018-08-30 11:49:06 +02:00
|
|
|
"net"
|
2018-09-06 03:24:15 +02:00
|
|
|
"sort"
|
2018-08-29 19:00:45 +02:00
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
)
|
|
|
|
|
|
|
|
// Try to keep it compatible with gitub.com/zrepl/zrepl/replication.Endpoint
|
2018-08-30 11:49:06 +02:00
|
|
|
type History interface {
|
2018-09-06 03:24:15 +02:00
|
|
|
ReplicationCursor(ctx context.Context, req *pdu.ReplicationCursorReq) (*pdu.ReplicationCursorRes, error)
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
type Target interface {
|
|
|
|
ListFilesystems(ctx context.Context) ([]*pdu.Filesystem, error)
|
|
|
|
ListFilesystemVersions(ctx context.Context, fs string) ([]*pdu.FilesystemVersion, error) // fix depS
|
2018-08-30 11:49:06 +02:00
|
|
|
DestroySnapshots(ctx context.Context, req *pdu.DestroySnapshotsReq) (*pdu.DestroySnapshotsRes, error)
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
type Logger = logger.Logger
|
|
|
|
|
|
|
|
type contextKey int
|
|
|
|
|
|
|
|
const contextKeyLogger contextKey = 0
|
|
|
|
|
|
|
|
func WithLogger(ctx context.Context, log Logger) context.Context {
|
|
|
|
return context.WithValue(ctx, contextKeyLogger, log)
|
|
|
|
}
|
|
|
|
|
2018-08-30 11:49:06 +02:00
|
|
|
func GetLogger(ctx context.Context) Logger {
|
2018-08-29 19:00:45 +02:00
|
|
|
if l, ok := ctx.Value(contextKeyLogger).(Logger); ok {
|
|
|
|
return l
|
|
|
|
}
|
|
|
|
return logger.NewNullLogger()
|
|
|
|
}
|
|
|
|
|
|
|
|
type args struct {
|
2018-09-06 03:24:15 +02:00
|
|
|
ctx context.Context
|
|
|
|
target Target
|
|
|
|
receiver History
|
|
|
|
rules []pruning.KeepRule
|
|
|
|
retryWait time.Duration
|
|
|
|
considerSnapAtCursorReplicated bool
|
2018-09-08 07:03:41 +02:00
|
|
|
promPruneSecs prometheus.Observer
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
type Pruner struct {
|
|
|
|
args args
|
|
|
|
|
|
|
|
mtx sync.RWMutex
|
|
|
|
|
|
|
|
state State
|
|
|
|
|
|
|
|
// State ErrWait|ErrPerm
|
|
|
|
sleepUntil time.Time
|
|
|
|
err error
|
|
|
|
|
|
|
|
// State Exec
|
2018-09-05 02:25:10 +02:00
|
|
|
prunePending []*fs
|
|
|
|
pruneCompleted []*fs
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
2018-08-30 17:40:45 +02:00
|
|
|
type PrunerFactory struct {
|
2018-09-06 03:24:15 +02:00
|
|
|
senderRules []pruning.KeepRule
|
|
|
|
receiverRules []pruning.KeepRule
|
|
|
|
retryWait time.Duration
|
|
|
|
considerSnapAtCursorReplicated bool
|
2018-09-08 07:03:41 +02:00
|
|
|
promPruneSecs *prometheus.HistogramVec
|
2018-08-30 17:40:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func checkContainsKeep1(rules []pruning.KeepRule) error {
|
|
|
|
if len(rules) == 0 {
|
|
|
|
return nil //No keep rules means keep all - ok
|
|
|
|
}
|
|
|
|
for _, e := range rules {
|
|
|
|
switch e.(type) {
|
|
|
|
case *pruning.KeepLastN:
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return errors.New("sender keep rules must contain last_n or be empty so that the last snapshot is definitely kept")
|
|
|
|
}
|
|
|
|
|
2018-09-08 07:03:41 +02:00
|
|
|
func NewPrunerFactory(in config.PruningSenderReceiver, promPruneSecs *prometheus.HistogramVec) (*PrunerFactory, error) {
|
2018-08-30 17:40:45 +02:00
|
|
|
keepRulesReceiver, err := pruning.RulesFromConfig(in.KeepReceiver)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "cannot build receiver pruning rules")
|
|
|
|
}
|
|
|
|
|
|
|
|
keepRulesSender, err := pruning.RulesFromConfig(in.KeepSender)
|
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "cannot build sender pruning rules")
|
|
|
|
}
|
|
|
|
|
2018-09-06 03:24:15 +02:00
|
|
|
considerSnapAtCursorReplicated := false
|
|
|
|
for _, r := range in.KeepSender {
|
|
|
|
knr, ok := r.Ret.(*config.PruneKeepNotReplicated)
|
|
|
|
if !ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
considerSnapAtCursorReplicated = considerSnapAtCursorReplicated || !knr.KeepSnapshotAtCursor
|
2018-08-30 17:40:45 +02:00
|
|
|
}
|
|
|
|
f := &PrunerFactory{
|
|
|
|
keepRulesSender,
|
|
|
|
keepRulesReceiver,
|
|
|
|
10 * time.Second, //FIXME constant
|
2018-09-06 03:24:15 +02:00
|
|
|
considerSnapAtCursorReplicated,
|
2018-09-08 07:03:41 +02:00
|
|
|
promPruneSecs,
|
2018-08-30 17:40:45 +02:00
|
|
|
}
|
|
|
|
return f, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *PrunerFactory) BuildSenderPruner(ctx context.Context, target Target, receiver History) *Pruner {
|
|
|
|
p := &Pruner{
|
|
|
|
args: args{
|
|
|
|
WithLogger(ctx, GetLogger(ctx).WithField("prune_side", "sender")),
|
|
|
|
target,
|
|
|
|
receiver,
|
|
|
|
f.senderRules,
|
|
|
|
f.retryWait,
|
2018-09-06 03:24:15 +02:00
|
|
|
f.considerSnapAtCursorReplicated,
|
2018-09-08 07:03:41 +02:00
|
|
|
f.promPruneSecs.WithLabelValues("sender"),
|
2018-08-30 17:40:45 +02:00
|
|
|
},
|
|
|
|
state: Plan,
|
|
|
|
}
|
|
|
|
return p
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *PrunerFactory) BuildReceiverPruner(ctx context.Context, target Target, receiver History) *Pruner {
|
2018-08-29 19:00:45 +02:00
|
|
|
p := &Pruner{
|
2018-08-30 17:40:45 +02:00
|
|
|
args: args{
|
|
|
|
WithLogger(ctx, GetLogger(ctx).WithField("prune_side", "receiver")),
|
|
|
|
target,
|
|
|
|
receiver,
|
|
|
|
f.receiverRules,
|
|
|
|
f.retryWait,
|
2018-09-06 03:24:15 +02:00
|
|
|
false, // senseless here anyways
|
2018-09-08 07:03:41 +02:00
|
|
|
f.promPruneSecs.WithLabelValues("receiver"),
|
2018-08-30 17:40:45 +02:00
|
|
|
},
|
2018-08-29 19:00:45 +02:00
|
|
|
state: Plan,
|
|
|
|
}
|
|
|
|
return p
|
|
|
|
}
|
|
|
|
|
2018-10-12 22:10:49 +02:00
|
|
|
//go:generate enumer -type=State
|
2018-08-29 19:00:45 +02:00
|
|
|
type State int
|
|
|
|
|
|
|
|
const (
|
|
|
|
Plan State = 1 << iota
|
|
|
|
PlanWait
|
|
|
|
Exec
|
|
|
|
ExecWait
|
|
|
|
ErrPerm
|
|
|
|
Done
|
|
|
|
)
|
|
|
|
|
|
|
|
func (s State) statefunc() state {
|
|
|
|
var statemap = map[State]state{
|
|
|
|
Plan: statePlan,
|
|
|
|
PlanWait: statePlanWait,
|
|
|
|
Exec: stateExec,
|
|
|
|
ExecWait: stateExecWait,
|
|
|
|
ErrPerm: nil,
|
|
|
|
Done: nil,
|
|
|
|
}
|
|
|
|
return statemap[s]
|
|
|
|
}
|
|
|
|
|
|
|
|
type updater func(func(*Pruner)) State
|
|
|
|
type state func(args *args, u updater) state
|
|
|
|
|
2018-08-30 17:40:45 +02:00
|
|
|
func (p *Pruner) Prune() {
|
2018-08-29 19:00:45 +02:00
|
|
|
p.prune(p.args)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Pruner) prune(args args) {
|
|
|
|
s := p.state.statefunc()
|
|
|
|
for s != nil {
|
|
|
|
pre := p.state
|
|
|
|
s = s(&args, func(f func(*Pruner)) State {
|
|
|
|
p.mtx.Lock()
|
|
|
|
defer p.mtx.Unlock()
|
|
|
|
f(p)
|
|
|
|
return p.state
|
|
|
|
})
|
|
|
|
post := p.state
|
2018-08-30 11:49:06 +02:00
|
|
|
GetLogger(args.ctx).
|
2018-08-29 19:00:45 +02:00
|
|
|
WithField("transition", fmt.Sprintf("%s=>%s", pre, post)).
|
|
|
|
Debug("state transition")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-09-24 19:22:44 +02:00
|
|
|
type Report struct {
|
|
|
|
State string
|
|
|
|
SleepUntil time.Time
|
|
|
|
Error string
|
|
|
|
Pending, Completed []FSReport
|
|
|
|
}
|
|
|
|
|
|
|
|
type FSReport struct {
|
|
|
|
Filesystem string
|
|
|
|
SnapshotList, DestroyList []SnapshotReport
|
|
|
|
Error string
|
|
|
|
}
|
|
|
|
|
|
|
|
type SnapshotReport struct {
|
|
|
|
Name string
|
|
|
|
Replicated bool
|
|
|
|
Date time.Time
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Pruner) Report() *Report {
|
|
|
|
p.mtx.Lock()
|
|
|
|
defer p.mtx.Unlock()
|
|
|
|
|
|
|
|
r := Report{State: p.state.String()}
|
|
|
|
|
|
|
|
if p.state & PlanWait|ExecWait != 0 {
|
|
|
|
r.SleepUntil = p.sleepUntil
|
|
|
|
}
|
|
|
|
if p.state & PlanWait|ExecWait|ErrPerm != 0 {
|
|
|
|
if p.err != nil {
|
|
|
|
r.Error = p.err.Error()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.state & Plan|PlanWait == 0 {
|
|
|
|
return &r
|
|
|
|
}
|
|
|
|
|
|
|
|
r.Pending = make([]FSReport, len(p.prunePending))
|
|
|
|
for i, fs := range p.prunePending{
|
|
|
|
r.Pending[i] = fs.Report()
|
|
|
|
}
|
|
|
|
r.Completed = make([]FSReport, len(p.pruneCompleted))
|
|
|
|
for i, fs := range p.pruneCompleted{
|
|
|
|
r.Completed[i] = fs.Report()
|
|
|
|
}
|
|
|
|
|
|
|
|
return &r
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
type fs struct {
|
2018-08-30 11:49:06 +02:00
|
|
|
path string
|
2018-09-24 19:22:44 +02:00
|
|
|
|
|
|
|
// snapshots presented by target
|
|
|
|
// (type snapshot)
|
2018-08-29 19:00:45 +02:00
|
|
|
snaps []pruning.Snapshot
|
2018-09-24 19:22:44 +02:00
|
|
|
// destroy list returned by pruning.PruneSnapshots(snaps)
|
|
|
|
// (type snapshot)
|
|
|
|
destroyList []pruning.Snapshot
|
2018-08-29 19:00:45 +02:00
|
|
|
|
|
|
|
mtx sync.RWMutex
|
|
|
|
// for Plan
|
|
|
|
err error
|
|
|
|
}
|
|
|
|
|
2018-08-30 11:49:06 +02:00
|
|
|
func (f *fs) Update(err error) {
|
2018-08-29 19:00:45 +02:00
|
|
|
f.mtx.Lock()
|
|
|
|
defer f.mtx.Unlock()
|
|
|
|
f.err = err
|
|
|
|
}
|
|
|
|
|
2018-09-24 19:22:44 +02:00
|
|
|
func (f *fs) Report() FSReport {
|
|
|
|
f.mtx.Lock()
|
|
|
|
defer f.mtx.Unlock()
|
|
|
|
|
|
|
|
r := FSReport{}
|
|
|
|
r.Filesystem = f.path
|
|
|
|
if f.err != nil {
|
|
|
|
r.Error = f.err.Error()
|
|
|
|
}
|
|
|
|
|
|
|
|
r.SnapshotList = make([]SnapshotReport, len(f.snaps))
|
|
|
|
for i, snap := range f.snaps {
|
|
|
|
r.SnapshotList[i] = snap.(snapshot).Report()
|
|
|
|
}
|
|
|
|
|
|
|
|
r.DestroyList = make([]SnapshotReport, len(f.destroyList))
|
|
|
|
for i, snap := range f.destroyList{
|
|
|
|
r.DestroyList[i] = snap.(snapshot).Report()
|
|
|
|
}
|
|
|
|
|
|
|
|
return r
|
|
|
|
}
|
|
|
|
|
2018-08-29 19:00:45 +02:00
|
|
|
type snapshot struct {
|
|
|
|
replicated bool
|
2018-08-30 11:49:06 +02:00
|
|
|
date time.Time
|
|
|
|
fsv *pdu.FilesystemVersion
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
2018-09-24 19:22:44 +02:00
|
|
|
func (s snapshot) Report() SnapshotReport {
|
|
|
|
return SnapshotReport{
|
|
|
|
Name: s.Name(),
|
|
|
|
Replicated: s.Replicated(),
|
|
|
|
Date: s.Date(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-08-29 19:00:45 +02:00
|
|
|
var _ pruning.Snapshot = snapshot{}
|
|
|
|
|
|
|
|
func (s snapshot) Name() string { return s.fsv.Name }
|
|
|
|
|
|
|
|
func (s snapshot) Replicated() bool { return s.replicated }
|
|
|
|
|
|
|
|
func (s snapshot) Date() time.Time { return s.date }
|
|
|
|
|
|
|
|
func shouldRetry(e error) bool {
|
|
|
|
switch e.(type) {
|
|
|
|
case nil:
|
|
|
|
return true
|
|
|
|
case net.Error:
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
func onErr(u updater, e error) state {
|
|
|
|
return u(func(p *Pruner) {
|
|
|
|
p.err = e
|
|
|
|
if !shouldRetry(e) {
|
|
|
|
p.state = ErrPerm
|
|
|
|
return
|
|
|
|
}
|
|
|
|
switch p.state {
|
2018-08-30 11:49:06 +02:00
|
|
|
case Plan:
|
|
|
|
p.state = PlanWait
|
|
|
|
case Exec:
|
|
|
|
p.state = ExecWait
|
|
|
|
default:
|
|
|
|
panic(p.state)
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
}).statefunc()
|
|
|
|
}
|
|
|
|
|
|
|
|
func statePlan(a *args, u updater) state {
|
|
|
|
|
|
|
|
ctx, target, receiver := a.ctx, a.target, a.receiver
|
|
|
|
|
|
|
|
tfss, err := target.ListFilesystems(ctx)
|
|
|
|
if err != nil {
|
|
|
|
return onErr(u, err)
|
|
|
|
}
|
|
|
|
|
2018-09-05 02:25:10 +02:00
|
|
|
pfss := make([]*fs, len(tfss))
|
2018-09-06 03:24:15 +02:00
|
|
|
fsloop:
|
2018-08-29 19:00:45 +02:00
|
|
|
for i, tfs := range tfss {
|
2018-09-06 03:24:15 +02:00
|
|
|
|
|
|
|
l := GetLogger(ctx).WithField("fs", tfs.Path)
|
|
|
|
l.Debug("plan filesystem")
|
|
|
|
|
2018-09-06 06:47:44 +02:00
|
|
|
|
|
|
|
pfs := &fs{
|
|
|
|
path: tfs.Path,
|
|
|
|
}
|
|
|
|
pfss[i] = pfs
|
|
|
|
|
2018-08-29 19:00:45 +02:00
|
|
|
tfsvs, err := target.ListFilesystemVersions(ctx, tfs.Path)
|
|
|
|
if err != nil {
|
2018-09-06 03:24:15 +02:00
|
|
|
l.WithError(err).Error("cannot list filesystem versions")
|
2018-09-06 06:47:44 +02:00
|
|
|
if shouldRetry(err) {
|
|
|
|
return onErr(u, err)
|
|
|
|
}
|
|
|
|
pfs.err = err
|
|
|
|
continue fsloop
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
2018-09-06 06:47:44 +02:00
|
|
|
pfs.snaps = make([]pruning.Snapshot, 0, len(tfsvs))
|
2018-08-29 19:00:45 +02:00
|
|
|
|
2018-09-06 03:24:15 +02:00
|
|
|
rcReq := &pdu.ReplicationCursorReq{
|
|
|
|
Filesystem: tfs.Path,
|
2018-09-24 12:26:55 +02:00
|
|
|
Op: &pdu.ReplicationCursorReq_Get{
|
|
|
|
Get: &pdu.ReplicationCursorReq_GetOp{},
|
|
|
|
},
|
2018-09-06 03:24:15 +02:00
|
|
|
}
|
|
|
|
rc, err := receiver.ReplicationCursor(ctx, rcReq)
|
|
|
|
if err != nil {
|
|
|
|
l.WithError(err).Error("cannot get replication cursor")
|
2018-09-06 06:47:44 +02:00
|
|
|
if shouldRetry(err) {
|
|
|
|
return onErr(u, err)
|
|
|
|
}
|
|
|
|
pfs.err = err
|
|
|
|
continue fsloop
|
2018-09-06 03:24:15 +02:00
|
|
|
}
|
|
|
|
if rc.GetError() != "" {
|
|
|
|
l.WithField("reqErr", rc.GetError()).Error("cannot get replication cursor")
|
2018-09-06 06:47:44 +02:00
|
|
|
pfs.err = fmt.Errorf("%s", rc.GetError())
|
|
|
|
continue fsloop
|
2018-09-06 03:24:15 +02:00
|
|
|
}
|
|
|
|
|
2018-08-29 19:00:45 +02:00
|
|
|
|
2018-09-06 03:24:15 +02:00
|
|
|
// scan from older to newer, all snapshots older than cursor are interpreted as replicated
|
|
|
|
sort.Slice(tfsvs, func(i, j int) bool {
|
|
|
|
return tfsvs[i].CreateTXG < tfsvs[j].CreateTXG
|
|
|
|
})
|
2018-10-12 15:29:07 +02:00
|
|
|
|
|
|
|
haveCursorSnapshot := false
|
|
|
|
for _, tfsv := range tfsvs {
|
|
|
|
if tfsv.Type != pdu.FilesystemVersion_Snapshot {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if tfsv.Guid == rc.GetGuid() {
|
|
|
|
haveCursorSnapshot = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
preCursor := haveCursorSnapshot
|
2018-08-29 19:00:45 +02:00
|
|
|
for _, tfsv := range tfsvs {
|
|
|
|
if tfsv.Type != pdu.FilesystemVersion_Snapshot {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
creation, err := tfsv.CreationAsTime()
|
|
|
|
if err != nil {
|
2018-09-06 03:24:15 +02:00
|
|
|
pfs.err = fmt.Errorf("%s%s has invalid creation date: %s", tfs, tfsv.RelName(), err)
|
|
|
|
l.WithError(pfs.err).Error("")
|
|
|
|
continue fsloop
|
2018-09-04 22:29:51 +02:00
|
|
|
}
|
2018-10-12 15:29:07 +02:00
|
|
|
// note that we cannot use CreateTXG because target and receiver could be on different pools
|
2018-09-06 03:24:15 +02:00
|
|
|
atCursor := tfsv.Guid == rc.GetGuid()
|
|
|
|
preCursor = preCursor && !atCursor
|
2018-08-29 19:00:45 +02:00
|
|
|
pfs.snaps = append(pfs.snaps, snapshot{
|
2018-09-06 03:24:15 +02:00
|
|
|
replicated: preCursor || (a.considerSnapAtCursorReplicated && atCursor),
|
2018-08-29 19:00:45 +02:00
|
|
|
date: creation,
|
2018-08-30 11:49:06 +02:00
|
|
|
fsv: tfsv,
|
2018-08-29 19:00:45 +02:00
|
|
|
})
|
|
|
|
}
|
2018-09-06 03:24:15 +02:00
|
|
|
if preCursor {
|
|
|
|
pfs.err = fmt.Errorf("replication cursor not found in prune target filesystem versions")
|
|
|
|
l.WithError(pfs.err).Error("")
|
|
|
|
continue fsloop
|
|
|
|
}
|
2018-08-29 19:00:45 +02:00
|
|
|
|
2018-09-24 19:22:44 +02:00
|
|
|
// Apply prune rules
|
|
|
|
pfs.destroyList = pruning.PruneSnapshots(pfs.snaps, a.rules)
|
|
|
|
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
return u(func(pruner *Pruner) {
|
|
|
|
for _, pfs := range pfss {
|
|
|
|
if pfs.err != nil {
|
|
|
|
pruner.pruneCompleted = append(pruner.pruneCompleted, pfs)
|
|
|
|
} else {
|
|
|
|
pruner.prunePending = append(pruner.prunePending, pfs)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pruner.state = Exec
|
|
|
|
}).statefunc()
|
|
|
|
}
|
|
|
|
|
|
|
|
func stateExec(a *args, u updater) state {
|
|
|
|
|
2018-09-05 02:25:10 +02:00
|
|
|
var pfs *fs
|
2018-08-29 19:00:45 +02:00
|
|
|
state := u(func(pruner *Pruner) {
|
|
|
|
if len(pruner.prunePending) == 0 {
|
2018-09-06 03:24:15 +02:00
|
|
|
nextState := Done
|
|
|
|
for _, pfs := range pruner.pruneCompleted {
|
|
|
|
if pfs.err != nil {
|
|
|
|
nextState = ErrPerm
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pruner.state = nextState
|
2018-08-29 19:00:45 +02:00
|
|
|
return
|
|
|
|
}
|
|
|
|
pfs = pruner.prunePending[0]
|
|
|
|
})
|
|
|
|
if state != Exec {
|
|
|
|
return state.statefunc()
|
|
|
|
}
|
|
|
|
|
2018-09-24 19:22:44 +02:00
|
|
|
destroyList := make([]*pdu.FilesystemVersion, len(pfs.destroyList))
|
2018-08-29 19:00:45 +02:00
|
|
|
for i := range destroyList {
|
2018-09-24 19:22:44 +02:00
|
|
|
destroyList[i] = pfs.destroyList[i].(snapshot).fsv
|
2018-08-30 11:49:06 +02:00
|
|
|
GetLogger(a.ctx).
|
|
|
|
WithField("fs", pfs.path).
|
|
|
|
WithField("destroy_snap", destroyList[i].Name).
|
|
|
|
Debug("policy destroys snapshot")
|
2018-08-29 19:00:45 +02:00
|
|
|
}
|
|
|
|
pfs.Update(nil)
|
2018-08-30 11:49:06 +02:00
|
|
|
req := pdu.DestroySnapshotsReq{
|
|
|
|
Filesystem: pfs.path,
|
|
|
|
Snapshots: destroyList,
|
|
|
|
}
|
|
|
|
_, err := a.target.DestroySnapshots(a.ctx, &req)
|
2018-08-29 19:00:45 +02:00
|
|
|
pfs.Update(err)
|
2018-08-30 11:49:06 +02:00
|
|
|
if err != nil && shouldRetry(err) {
|
2018-08-29 19:00:45 +02:00
|
|
|
return onErr(u, err)
|
|
|
|
}
|
|
|
|
// if it's not retryable, treat is like as being done
|
|
|
|
|
|
|
|
return u(func(pruner *Pruner) {
|
|
|
|
pruner.pruneCompleted = append(pruner.pruneCompleted, pfs)
|
|
|
|
pruner.prunePending = pruner.prunePending[1:]
|
|
|
|
}).statefunc()
|
|
|
|
}
|
|
|
|
|
|
|
|
func stateExecWait(a *args, u updater) state {
|
|
|
|
return doWait(Exec, a, u)
|
|
|
|
}
|
|
|
|
|
|
|
|
func statePlanWait(a *args, u updater) state {
|
|
|
|
return doWait(Plan, a, u)
|
|
|
|
}
|
|
|
|
|
|
|
|
func doWait(goback State, a *args, u updater) state {
|
|
|
|
timer := time.NewTimer(a.retryWait)
|
|
|
|
defer timer.Stop()
|
|
|
|
select {
|
|
|
|
case <-timer.C:
|
|
|
|
return u(func(pruner *Pruner) {
|
|
|
|
pruner.state = goback
|
|
|
|
}).statefunc()
|
|
|
|
case <-a.ctx.Done():
|
|
|
|
return onErr(u, a.ctx.Err())
|
|
|
|
}
|
|
|
|
}
|