[pve-devel] [PATCH pve-manager 05/18] PVE::Replication - use new calendar events instead of interval

Dietmar Maurer dietmar at proxmox.com
Mon May 29 11:59:35 CEST 2017


comment inline
> 
> On Tue, May 23, 2017 at 09:08:44AM +0200, Dietmar Maurer wrote:
> > And implement retry algorythm after failure:
> > 
> >   $next_sync = $state->{last_try} + 5*60*$fail_count;
> > 
> > and limit to 3 failures.
> > 
> > Signed-off-by: Dietmar Maurer <dietmar at proxmox.com>
> > ---
> >  PVE/API2/Replication.pm |  2 +-
> >  PVE/CLI/pvesr.pm        | 28 ++++++++++++++++++++--------
> >  PVE/Replication.pm      | 35 +++++++++++++++++++++--------------
> >  3 files changed, 42 insertions(+), 23 deletions(-)
> > 
> > diff --git a/PVE/API2/Replication.pm b/PVE/API2/Replication.pm
> > index c400a93c..977d3ec1 100644
> > --- a/PVE/API2/Replication.pm
> > +++ b/PVE/API2/Replication.pm
> > @@ -82,7 +82,7 @@ __PACKAGE__->register_method ({
> >  	    my $vmid = $d->{guest};
> >  	    next if !$rpcenv->check($authuser, "/vms/$vmid", [ 'VM.Audit' ]);
> >  	    $d->{id} = $id;
> > -	    foreach my $k (qw(last_sync fail_count error duration)) {
> > +	    foreach my $k (qw(last_sync last_try fail_count error duration)) {
> >  		$d->{$k} = $state->{$k} if defined($state->{$k});
> >  	    }
> >  	    if ($state->{pid} && $state->{ptime}) {
> > diff --git a/PVE/CLI/pvesr.pm b/PVE/CLI/pvesr.pm
> > index 38116f7b..115bc2c1 100644
> > --- a/PVE/CLI/pvesr.pm
> > +++ b/PVE/CLI/pvesr.pm
> > @@ -97,14 +97,14 @@ my $print_job_list = sub {
> >  
> >      my $format = "%-20s %10s %-20s %10s %5s %8s\n";
> >  
> > -    printf($format, "JobID", "GuestID", "Target", "Interval", "Rate",
> > "Enabled");
> > +    printf($format, "JobID", "GuestID", "Target", "Schedule", "Rate",
> > "Enabled");
> >  
> >      foreach my $job (sort { $a->{guest} <=> $b->{guest} } @$list) {
> >  	my $plugin = PVE::ReplicationConfig->lookup($job->{type});
> >  	my $tid = $plugin->get_unique_target_id($job);
> >  
> >  	printf($format, $job->{id}, $job->{guest}, $tid,
> > -	       defined($job->{interval}) ? $job->{interval} : '-',
> > +	       defined($job->{schedule}) ? $job->{schedule} : '*/15',
> >  	       defined($job->{rate}) ? $job->{rate} : '-',
> >  	       $job->{disable} ? 'no' : 'yes'
> >  	    );
> > @@ -114,21 +114,33 @@ my $print_job_list = sub {
> >  my $print_job_status = sub {
> >      my ($list) = @_;
> >  
> > -    my $format = "%-20s %10s %-20s %20s %10s %10s %s\n";
> > +    my $format = "%-20s %10s %-20s %20s %20s %10s %10s %s\n";
> >  
> > -    printf($format, "JobID", "GuestID", "Target", "LastSync", "Duration",
> > "FailCount", "State");
> > +    printf($format, "JobID", "GuestID", "Target", "LastSync", "NextSync",
> > "Duration", "FailCount", "State");
> >  
> >      foreach my $job (sort { $a->{guest} <=> $b->{guest} } @$list) {
> >  	my $plugin = PVE::ReplicationConfig->lookup($job->{type});
> >  	my $tid = $plugin->get_unique_target_id($job);
> >  
> > -	my $timestr = $job->{last_sync} ?
> > -	    strftime("%Y-%m-%d_%H:%M:%S", localtime($job->{last_sync})) : '-';
> > +	my $timestr = '-';
> > +	if ($job->{last_sync}) {
> > +	    strftime("%Y-%m-%d_%H:%M:%S", localtime($job->{last_sync}));
> > +	}
> > +
> > +	my $nextstr = '-';
> > +	if (my $next = $job->{next_sync}) {
> > +	    my $now = time();
> > +	    if ($next > $now) {
> > +		$nextstr = strftime("%Y-%m-%d_%H:%M:%S", localtime($job->{next_sync}));
> > +	    } else {
> > +		$nextstr = 'now'
> > +	    }
> > +	}
> >  
> >  	my $state = $job->{pid} ? "SYNCING" : $job->{error} // 'OK';
> >  
> >  	printf($format, $job->{id}, $job->{guest}, $tid,
> > -	       $timestr, $job->{duration} // '-',
> > +	       $timestr, $nextstr, $job->{duration} // '-',
> >  	       $job->{fail_count}, $state);
> >      }
> >  };
> > @@ -136,7 +148,7 @@ my $print_job_status = sub {
> >  our $cmddef = {
> >      status => [ 'PVE::API2::Replication', 'status', [], { node => $nodename
> > }, $print_job_status ],
> >  
> > -    jobs => [ 'PVE::API2::ReplicationConfig', 'index' , [], {},
> > $print_job_list ],
> > +    list => [ 'PVE::API2::ReplicationConfig', 'index' , [], {},
> > $print_job_list ],
> >      read => [ 'PVE::API2::ReplicationConfig', 'read' , ['id'], {},
> >  	     sub { my $res = shift; print to_json($res, { pretty => 1, canonical
> > => 1}); }],
> >      update => [ 'PVE::API2::ReplicationConfig', 'update' , ['id'], {} ],
> > diff --git a/PVE/Replication.pm b/PVE/Replication.pm
> > index ff4bbeb4..d878b44a 100644
> > --- a/PVE/Replication.pm
> > +++ b/PVE/Replication.pm
> > @@ -9,6 +9,7 @@ use Time::HiRes qw(gettimeofday tv_interval);
> >  use PVE::INotify;
> >  use PVE::ProcFSTools;
> >  use PVE::Tools;
> > +use PVE::CalendarEvent;
> >  use PVE::Cluster;
> >  use PVE::QemuConfig;
> >  use PVE::QemuServer;
> > @@ -46,7 +47,8 @@ my $get_job_state = sub {
> >      $state = {} if !$state;
> >  
> >      $state->{last_iteration} //= 0;
> > -    $state->{last_sync} //= 0;
> > +    $state->{last_try} //= 0; # last sync start time
> > +    $state->{last_sync} //= 0; # last successful sync start time
> >      $state->{fail_count} //= 0;
> >  
> >      return $state;
> > @@ -93,10 +95,23 @@ sub job_status {
> >  
> >  	next if $jobcfg->{disable};
> >  
> > -	$jobcfg->{state} = $get_job_state->($stateobj, $jobcfg);
> > +	my $state = $get_job_state->($stateobj, $jobcfg);
> > +	$jobcfg->{state} = $state;
> >  	$jobcfg->{id} = $jobid;
> >  	$jobcfg->{vmtype} = $vms->{ids}->{$vmid}->{type};
> >  
> > +	my $next_sync = 0;
> > +	if (my $fail_count = $state->{fail_count}) {
> > +	    if ($fail_count < 3) {
> > +		$next_sync = $state->{last_try} + 5*60*$fail_count;
> 
> both the 3 and the 5*60 constants seem awfully arbitrary.. maybe we
> could make them configurable? e.g., for a sync job that syncs once or
> twice a day, a window of 0+5+10+15=30 minutes of not being able to sync
> should probably not put that job into an error state?

yes, but I guess we we can add such things later.
 




More information about the pve-devel mailing list