Skip to content

Instantly share code, notes, and snippets.

@agarzon
Created November 18, 2011 21:24
Show Gist options
  • Save agarzon/1377797 to your computer and use it in GitHub Desktop.
Save agarzon/1377797 to your computer and use it in GitHub Desktop.
pt-table-sync improved
#!/usr/bin/env perl
# This program is part of Percona Toolkit: http://www.percona.com/software/
# See "COPYRIGHT, LICENSE, AND WARRANTY" at the end of this file for legal
# notices and disclaimers.
use strict;
use warnings FATAL => 'all';
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
# ###########################################################################
# OptionParser package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/OptionParser.pm
# t/lib/OptionParser.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package OptionParser;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use List::Util qw(max);
use Getopt::Long;
my $POD_link_re = '[LC]<"?([^">]+)"?>';
sub new {
my ( $class, %args ) = @_;
my @required_args = qw();
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($program_name) = $PROGRAM_NAME =~ m/([.A-Za-z-]+)$/;
$program_name ||= $PROGRAM_NAME;
my $home = $ENV{HOME} || $ENV{HOMEPATH} || $ENV{USERPROFILE} || '.';
my %attributes = (
'type' => 1,
'short form' => 1,
'group' => 1,
'default' => 1,
'cumulative' => 1,
'negatable' => 1,
);
my $self = {
head1 => 'OPTIONS', # These args are used internally
skip_rules => 0, # to instantiate another Option-
item => '--(.*)', # Parser obj that parses the
attributes => \%attributes, # DSN OPTIONS section. Tools
parse_attributes => \&_parse_attribs, # don't tinker with these args.
%args,
strict => 1, # disabled by a special rule
program_name => $program_name,
opts => {},
got_opts => 0,
short_opts => {},
defaults => {},
groups => {},
allowed_groups => {},
errors => [],
rules => [], # desc of rules for --help
mutex => [], # rule: opts are mutually exclusive
atleast1 => [], # rule: at least one opt is required
disables => {}, # rule: opt disables other opts
defaults_to => {}, # rule: opt defaults to value of other opt
DSNParser => undef,
default_files => [
"/etc/percona-toolkit/percona-toolkit.conf",
"/etc/percona-toolkit/$program_name.conf",
"$home/.percona-toolkit.conf",
"$home/.$program_name.conf",
],
types => {
string => 's', # standard Getopt type
int => 'i', # standard Getopt type
float => 'f', # standard Getopt type
Hash => 'H', # hash, formed from a comma-separated list
hash => 'h', # hash as above, but only if a value is given
Array => 'A', # array, similar to Hash
array => 'a', # array, similar to hash
DSN => 'd', # DSN
size => 'z', # size with kMG suffix (powers of 2^10)
time => 'm', # time, with an optional suffix of s/h/m/d
},
};
return bless $self, $class;
}
sub get_specs {
my ( $self, $file ) = @_;
$file ||= $self->{file} || __FILE__;
my @specs = $self->_pod_to_specs($file);
$self->_parse_specs(@specs);
open my $fh, "<", $file or die "Cannot open $file: $OS_ERROR";
my $contents = do { local $/ = undef; <$fh> };
close $fh;
if ( $contents =~ m/^=head1 DSN OPTIONS/m ) {
MKDEBUG && _d('Parsing DSN OPTIONS');
my $dsn_attribs = {
dsn => 1,
copy => 1,
};
my $parse_dsn_attribs = sub {
my ( $self, $option, $attribs ) = @_;
map {
my $val = $attribs->{$_};
if ( $val ) {
$val = $val eq 'yes' ? 1
: $val eq 'no' ? 0
: $val;
$attribs->{$_} = $val;
}
} keys %$attribs;
return {
key => $option,
%$attribs,
};
};
my $dsn_o = new OptionParser(
description => 'DSN OPTIONS',
head1 => 'DSN OPTIONS',
dsn => 0, # XXX don't infinitely recurse!
item => '\* (.)', # key opts are a single character
skip_rules => 1, # no rules before opts
attributes => $dsn_attribs,
parse_attributes => $parse_dsn_attribs,
);
my @dsn_opts = map {
my $opts = {
key => $_->{spec}->{key},
dsn => $_->{spec}->{dsn},
copy => $_->{spec}->{copy},
desc => $_->{desc},
};
$opts;
} $dsn_o->_pod_to_specs($file);
$self->{DSNParser} = DSNParser->new(opts => \@dsn_opts);
}
if ( $contents =~ m/^=head1 VERSION\n\n^(.+)$/m ) {
$self->{version} = $1;
MKDEBUG && _d($self->{version});
}
return;
}
sub DSNParser {
my ( $self ) = @_;
return $self->{DSNParser};
};
sub get_defaults_files {
my ( $self ) = @_;
return @{$self->{default_files}};
}
sub _pod_to_specs {
my ( $self, $file ) = @_;
$file ||= $self->{file} || __FILE__;
open my $fh, '<', $file or die "Cannot open $file: $OS_ERROR";
my @specs = ();
my @rules = ();
my $para;
local $INPUT_RECORD_SEPARATOR = '';
while ( $para = <$fh> ) {
next unless $para =~ m/^=head1 $self->{head1}/;
last;
}
while ( $para = <$fh> ) {
last if $para =~ m/^=over/;
next if $self->{skip_rules};
chomp $para;
$para =~ s/\s+/ /g;
$para =~ s/$POD_link_re/$1/go;
MKDEBUG && _d('Option rule:', $para);
push @rules, $para;
}
die "POD has no $self->{head1} section" unless $para;
do {
if ( my ($option) = $para =~ m/^=item $self->{item}/ ) {
chomp $para;
MKDEBUG && _d($para);
my %attribs;
$para = <$fh>; # read next paragraph, possibly attributes
if ( $para =~ m/: / ) { # attributes
$para =~ s/\s+\Z//g;
%attribs = map {
my ( $attrib, $val) = split(/: /, $_);
die "Unrecognized attribute for --$option: $attrib"
unless $self->{attributes}->{$attrib};
($attrib, $val);
} split(/; /, $para);
if ( $attribs{'short form'} ) {
$attribs{'short form'} =~ s/-//;
}
$para = <$fh>; # read next paragraph, probably short help desc
}
else {
MKDEBUG && _d('Option has no attributes');
}
$para =~ s/\s+\Z//g;
$para =~ s/\s+/ /g;
$para =~ s/$POD_link_re/$1/go;
$para =~ s/\.(?:\n.*| [A-Z].*|\Z)//s;
MKDEBUG && _d('Short help:', $para);
die "No description after option spec $option" if $para =~ m/^=item/;
if ( my ($base_option) = $option =~ m/^\[no\](.*)/ ) {
$option = $base_option;
$attribs{'negatable'} = 1;
}
push @specs, {
spec => $self->{parse_attributes}->($self, $option, \%attribs),
desc => $para
. (defined $attribs{default} ? " (default $attribs{default})" : ''),
group => ($attribs{'group'} ? $attribs{'group'} : 'default'),
};
}
while ( $para = <$fh> ) {
last unless $para;
if ( $para =~ m/^=head1/ ) {
$para = undef; # Can't 'last' out of a do {} block.
last;
}
last if $para =~ m/^=item /;
}
} while ( $para );
die "No valid specs in $self->{head1}" unless @specs;
close $fh;
return @specs, @rules;
}
sub _parse_specs {
my ( $self, @specs ) = @_;
my %disables; # special rule that requires deferred checking
foreach my $opt ( @specs ) {
if ( ref $opt ) { # It's an option spec, not a rule.
MKDEBUG && _d('Parsing opt spec:',
map { ($_, '=>', $opt->{$_}) } keys %$opt);
my ( $long, $short ) = $opt->{spec} =~ m/^([\w-]+)(?:\|([^!+=]*))?/;
if ( !$long ) {
die "Cannot parse long option from spec $opt->{spec}";
}
$opt->{long} = $long;
die "Duplicate long option --$long" if exists $self->{opts}->{$long};
$self->{opts}->{$long} = $opt;
if ( length $long == 1 ) {
MKDEBUG && _d('Long opt', $long, 'looks like short opt');
$self->{short_opts}->{$long} = $long;
}
if ( $short ) {
die "Duplicate short option -$short"
if exists $self->{short_opts}->{$short};
$self->{short_opts}->{$short} = $long;
$opt->{short} = $short;
}
else {
$opt->{short} = undef;
}
$opt->{is_negatable} = $opt->{spec} =~ m/!/ ? 1 : 0;
$opt->{is_cumulative} = $opt->{spec} =~ m/\+/ ? 1 : 0;
$opt->{is_required} = $opt->{desc} =~ m/required/ ? 1 : 0;
$opt->{group} ||= 'default';
$self->{groups}->{ $opt->{group} }->{$long} = 1;
$opt->{value} = undef;
$opt->{got} = 0;
my ( $type ) = $opt->{spec} =~ m/=(.)/;
$opt->{type} = $type;
MKDEBUG && _d($long, 'type:', $type);
$opt->{spec} =~ s/=./=s/ if ( $type && $type =~ m/[HhAadzm]/ );
if ( (my ($def) = $opt->{desc} =~ m/default\b(?: ([^)]+))?/) ) {
$self->{defaults}->{$long} = defined $def ? $def : 1;
MKDEBUG && _d($long, 'default:', $def);
}
if ( $long eq 'config' ) {
$self->{defaults}->{$long} = join(',', $self->get_defaults_files());
}
if ( (my ($dis) = $opt->{desc} =~ m/(disables .*)/) ) {
$disables{$long} = $dis;
MKDEBUG && _d('Deferring check of disables rule for', $opt, $dis);
}
$self->{opts}->{$long} = $opt;
}
else { # It's an option rule, not a spec.
MKDEBUG && _d('Parsing rule:', $opt);
push @{$self->{rules}}, $opt;
my @participants = $self->_get_participants($opt);
my $rule_ok = 0;
if ( $opt =~ m/mutually exclusive|one and only one/ ) {
$rule_ok = 1;
push @{$self->{mutex}}, \@participants;
MKDEBUG && _d(@participants, 'are mutually exclusive');
}
if ( $opt =~ m/at least one|one and only one/ ) {
$rule_ok = 1;
push @{$self->{atleast1}}, \@participants;
MKDEBUG && _d(@participants, 'require at least one');
}
if ( $opt =~ m/default to/ ) {
$rule_ok = 1;
$self->{defaults_to}->{$participants[0]} = $participants[1];
MKDEBUG && _d($participants[0], 'defaults to', $participants[1]);
}
if ( $opt =~ m/restricted to option groups/ ) {
$rule_ok = 1;
my ($groups) = $opt =~ m/groups ([\w\s\,]+)/;
my @groups = split(',', $groups);
%{$self->{allowed_groups}->{$participants[0]}} = map {
s/\s+//;
$_ => 1;
} @groups;
}
if( $opt =~ m/accepts additional command-line arguments/ ) {
$rule_ok = 1;
$self->{strict} = 0;
MKDEBUG && _d("Strict mode disabled by rule");
}
die "Unrecognized option rule: $opt" unless $rule_ok;
}
}
foreach my $long ( keys %disables ) {
my @participants = $self->_get_participants($disables{$long});
$self->{disables}->{$long} = \@participants;
MKDEBUG && _d('Option', $long, 'disables', @participants);
}
return;
}
sub _get_participants {
my ( $self, $str ) = @_;
my @participants;
foreach my $long ( $str =~ m/--(?:\[no\])?([\w-]+)/g ) {
die "Option --$long does not exist while processing rule $str"
unless exists $self->{opts}->{$long};
push @participants, $long;
}
MKDEBUG && _d('Participants for', $str, ':', @participants);
return @participants;
}
sub opts {
my ( $self ) = @_;
my %opts = %{$self->{opts}};
return %opts;
}
sub short_opts {
my ( $self ) = @_;
my %short_opts = %{$self->{short_opts}};
return %short_opts;
}
sub set_defaults {
my ( $self, %defaults ) = @_;
$self->{defaults} = {};
foreach my $long ( keys %defaults ) {
die "Cannot set default for nonexistent option $long"
unless exists $self->{opts}->{$long};
$self->{defaults}->{$long} = $defaults{$long};
MKDEBUG && _d('Default val for', $long, ':', $defaults{$long});
}
return;
}
sub get_defaults {
my ( $self ) = @_;
return $self->{defaults};
}
sub get_groups {
my ( $self ) = @_;
return $self->{groups};
}
sub _set_option {
my ( $self, $opt, $val ) = @_;
my $long = exists $self->{opts}->{$opt} ? $opt
: exists $self->{short_opts}->{$opt} ? $self->{short_opts}->{$opt}
: die "Getopt::Long gave a nonexistent option: $opt";
$opt = $self->{opts}->{$long};
if ( $opt->{is_cumulative} ) {
$opt->{value}++;
}
else {
$opt->{value} = $val;
}
$opt->{got} = 1;
MKDEBUG && _d('Got option', $long, '=', $val);
}
sub get_opts {
my ( $self ) = @_;
foreach my $long ( keys %{$self->{opts}} ) {
$self->{opts}->{$long}->{got} = 0;
$self->{opts}->{$long}->{value}
= exists $self->{defaults}->{$long} ? $self->{defaults}->{$long}
: $self->{opts}->{$long}->{is_cumulative} ? 0
: undef;
}
$self->{got_opts} = 0;
$self->{errors} = [];
if ( @ARGV && $ARGV[0] eq "--config" ) {
shift @ARGV;
$self->_set_option('config', shift @ARGV);
}
if ( $self->has('config') ) {
my @extra_args;
foreach my $filename ( split(',', $self->get('config')) ) {
eval {
push @extra_args, $self->_read_config_file($filename);
};
if ( $EVAL_ERROR ) {
if ( $self->got('config') ) {
die $EVAL_ERROR;
}
elsif ( MKDEBUG ) {
_d($EVAL_ERROR);
}
}
}
unshift @ARGV, @extra_args;
}
Getopt::Long::Configure('no_ignore_case', 'bundling');
GetOptions(
map { $_->{spec} => sub { $self->_set_option(@_); } }
grep { $_->{long} ne 'config' } # --config is handled specially above.
values %{$self->{opts}}
) or $self->save_error('Error parsing options');
if ( exists $self->{opts}->{version} && $self->{opts}->{version}->{got} ) {
if ( $self->{version} ) {
print $self->{version}, "\n";
}
else {
print "Error parsing version. See the VERSION section of the tool's documentation.\n";
}
exit 0;
}
if ( @ARGV && $self->{strict} ) {
$self->save_error("Unrecognized command-line options @ARGV");
}
foreach my $mutex ( @{$self->{mutex}} ) {
my @set = grep { $self->{opts}->{$_}->{got} } @$mutex;
if ( @set > 1 ) {
my $err = join(', ', map { "--$self->{opts}->{$_}->{long}" }
@{$mutex}[ 0 .. scalar(@$mutex) - 2] )
. ' and --'.$self->{opts}->{$mutex->[-1]}->{long}
. ' are mutually exclusive.';
$self->save_error($err);
}
}
foreach my $required ( @{$self->{atleast1}} ) {
my @set = grep { $self->{opts}->{$_}->{got} } @$required;
if ( @set == 0 ) {
my $err = join(', ', map { "--$self->{opts}->{$_}->{long}" }
@{$required}[ 0 .. scalar(@$required) - 2] )
.' or --'.$self->{opts}->{$required->[-1]}->{long};
$self->save_error("Specify at least one of $err");
}
}
$self->_check_opts( keys %{$self->{opts}} );
$self->{got_opts} = 1;
return;
}
sub _check_opts {
my ( $self, @long ) = @_;
my $long_last = scalar @long;
while ( @long ) {
foreach my $i ( 0..$#long ) {
my $long = $long[$i];
next unless $long;
my $opt = $self->{opts}->{$long};
if ( $opt->{got} ) {
if ( exists $self->{disables}->{$long} ) {
my @disable_opts = @{$self->{disables}->{$long}};
map { $self->{opts}->{$_}->{value} = undef; } @disable_opts;
MKDEBUG && _d('Unset options', @disable_opts,
'because', $long,'disables them');
}
if ( exists $self->{allowed_groups}->{$long} ) {
my @restricted_groups = grep {
!exists $self->{allowed_groups}->{$long}->{$_}
} keys %{$self->{groups}};
my @restricted_opts;
foreach my $restricted_group ( @restricted_groups ) {
RESTRICTED_OPT:
foreach my $restricted_opt (
keys %{$self->{groups}->{$restricted_group}} )
{
next RESTRICTED_OPT if $restricted_opt eq $long;
push @restricted_opts, $restricted_opt
if $self->{opts}->{$restricted_opt}->{got};
}
}
if ( @restricted_opts ) {
my $err;
if ( @restricted_opts == 1 ) {
$err = "--$restricted_opts[0]";
}
else {
$err = join(', ',
map { "--$self->{opts}->{$_}->{long}" }
grep { $_ }
@restricted_opts[0..scalar(@restricted_opts) - 2]
)
. ' or --'.$self->{opts}->{$restricted_opts[-1]}->{long};
}
$self->save_error("--$long is not allowed with $err");
}
}
}
elsif ( $opt->{is_required} ) {
$self->save_error("Required option --$long must be specified");
}
$self->_validate_type($opt);
if ( $opt->{parsed} ) {
delete $long[$i];
}
else {
MKDEBUG && _d('Temporarily failed to parse', $long);
}
}
die "Failed to parse options, possibly due to circular dependencies"
if @long == $long_last;
$long_last = @long;
}
return;
}
sub _validate_type {
my ( $self, $opt ) = @_;
return unless $opt;
if ( !$opt->{type} ) {
$opt->{parsed} = 1;
return;
}
my $val = $opt->{value};
if ( $val && $opt->{type} eq 'm' ) { # type time
MKDEBUG && _d('Parsing option', $opt->{long}, 'as a time value');
my ( $prefix, $num, $suffix ) = $val =~ m/([+-]?)(\d+)([a-z])?$/;
if ( !$suffix ) {
my ( $s ) = $opt->{desc} =~ m/\(suffix (.)\)/;
$suffix = $s || 's';
MKDEBUG && _d('No suffix given; using', $suffix, 'for',
$opt->{long}, '(value:', $val, ')');
}
if ( $suffix =~ m/[smhd]/ ) {
$val = $suffix eq 's' ? $num # Seconds
: $suffix eq 'm' ? $num * 60 # Minutes
: $suffix eq 'h' ? $num * 3600 # Hours
: $num * 86400; # Days
$opt->{value} = ($prefix || '') . $val;
MKDEBUG && _d('Setting option', $opt->{long}, 'to', $val);
}
else {
$self->save_error("Invalid time suffix for --$opt->{long}");
}
}
elsif ( $val && $opt->{type} eq 'd' ) { # type DSN
MKDEBUG && _d('Parsing option', $opt->{long}, 'as a DSN');
my $prev = {};
my $from_key = $self->{defaults_to}->{ $opt->{long} };
if ( $from_key ) {
MKDEBUG && _d($opt->{long}, 'DSN copies from', $from_key, 'DSN');
if ( $self->{opts}->{$from_key}->{parsed} ) {
$prev = $self->{opts}->{$from_key}->{value};
}
else {
MKDEBUG && _d('Cannot parse', $opt->{long}, 'until',
$from_key, 'parsed');
return;
}
}
my $defaults = $self->{DSNParser}->parse_options($self);
$opt->{value} = $self->{DSNParser}->parse($val, $prev, $defaults);
}
elsif ( $val && $opt->{type} eq 'z' ) { # type size
MKDEBUG && _d('Parsing option', $opt->{long}, 'as a size value');
$self->_parse_size($opt, $val);
}
elsif ( $opt->{type} eq 'H' || (defined $val && $opt->{type} eq 'h') ) {
$opt->{value} = { map { $_ => 1 } split(/(?<!\\),\s*/, ($val || '')) };
}
elsif ( $opt->{type} eq 'A' || (defined $val && $opt->{type} eq 'a') ) {
$opt->{value} = [ split(/(?<!\\),\s*/, ($val || '')) ];
}
else {
MKDEBUG && _d('Nothing to validate for option',
$opt->{long}, 'type', $opt->{type}, 'value', $val);
}
$opt->{parsed} = 1;
return;
}
sub get {
my ( $self, $opt ) = @_;
my $long = (length $opt == 1 ? $self->{short_opts}->{$opt} : $opt);
die "Option $opt does not exist"
unless $long && exists $self->{opts}->{$long};
return $self->{opts}->{$long}->{value};
}
sub got {
my ( $self, $opt ) = @_;
my $long = (length $opt == 1 ? $self->{short_opts}->{$opt} : $opt);
die "Option $opt does not exist"
unless $long && exists $self->{opts}->{$long};
return $self->{opts}->{$long}->{got};
}
sub has {
my ( $self, $opt ) = @_;
my $long = (length $opt == 1 ? $self->{short_opts}->{$opt} : $opt);
return defined $long ? exists $self->{opts}->{$long} : 0;
}
sub set {
my ( $self, $opt, $val ) = @_;
my $long = (length $opt == 1 ? $self->{short_opts}->{$opt} : $opt);
die "Option $opt does not exist"
unless $long && exists $self->{opts}->{$long};
$self->{opts}->{$long}->{value} = $val;
return;
}
sub save_error {
my ( $self, $error ) = @_;
push @{$self->{errors}}, $error;
return;
}
sub errors {
my ( $self ) = @_;
return $self->{errors};
}
sub usage {
my ( $self ) = @_;
warn "No usage string is set" unless $self->{usage}; # XXX
return "Usage: " . ($self->{usage} || '') . "\n";
}
sub descr {
my ( $self ) = @_;
warn "No description string is set" unless $self->{description}; # XXX
my $descr = ($self->{description} || $self->{program_name} || '')
. " For more details, please use the --help option, "
. "or try 'perldoc $PROGRAM_NAME' "
. "for complete documentation.";
$descr = join("\n", $descr =~ m/(.{0,80})(?:\s+|$)/g)
unless $ENV{DONT_BREAK_LINES};
$descr =~ s/ +$//mg;
return $descr;
}
sub usage_or_errors {
my ( $self, $file, $return ) = @_;
$file ||= $self->{file} || __FILE__;
if ( !$self->{description} || !$self->{usage} ) {
MKDEBUG && _d("Getting description and usage from SYNOPSIS in", $file);
my %synop = $self->_parse_synopsis($file);
$self->{description} ||= $synop{description};
$self->{usage} ||= $synop{usage};
MKDEBUG && _d("Description:", $self->{description},
"\nUsage:", $self->{usage});
}
if ( $self->{opts}->{help}->{got} ) {
print $self->print_usage() or die "Cannot print usage: $OS_ERROR";
exit 0 unless $return;
}
elsif ( scalar @{$self->{errors}} ) {
print $self->print_errors() or die "Cannot print errors: $OS_ERROR";
exit 0 unless $return;
}
return;
}
sub print_errors {
my ( $self ) = @_;
my $usage = $self->usage() . "\n";
if ( (my @errors = @{$self->{errors}}) ) {
$usage .= join("\n * ", 'Errors in command-line arguments:', @errors)
. "\n";
}
return $usage . "\n" . $self->descr();
}
sub print_usage {
my ( $self ) = @_;
die "Run get_opts() before print_usage()" unless $self->{got_opts};
my @opts = values %{$self->{opts}};
my $maxl = max(
map {
length($_->{long}) # option long name
+ ($_->{is_negatable} ? 4 : 0) # "[no]" if opt is negatable
+ ($_->{type} ? 2 : 0) # "=x" where x is the opt type
}
@opts);
my $maxs = max(0,
map {
length($_)
+ ($self->{opts}->{$_}->{is_negatable} ? 4 : 0)
+ ($self->{opts}->{$_}->{type} ? 2 : 0)
}
values %{$self->{short_opts}});
my $lcol = max($maxl, ($maxs + 3));
my $rcol = 80 - $lcol - 6;
my $rpad = ' ' x ( 80 - $rcol );
$maxs = max($lcol - 3, $maxs);
my $usage = $self->descr() . "\n" . $self->usage();
my @groups = reverse sort grep { $_ ne 'default'; } keys %{$self->{groups}};
push @groups, 'default';
foreach my $group ( reverse @groups ) {
$usage .= "\n".($group eq 'default' ? 'Options' : $group).":\n\n";
foreach my $opt (
sort { $a->{long} cmp $b->{long} }
grep { $_->{group} eq $group }
@opts )
{
my $long = $opt->{is_negatable} ? "[no]$opt->{long}" : $opt->{long};
my $short = $opt->{short};
my $desc = $opt->{desc};
$long .= $opt->{type} ? "=$opt->{type}" : "";
if ( $opt->{type} && $opt->{type} eq 'm' ) {
my ($s) = $desc =~ m/\(suffix (.)\)/;
$s ||= 's';
$desc =~ s/\s+\(suffix .\)//;
$desc .= ". Optional suffix s=seconds, m=minutes, h=hours, "
. "d=days; if no suffix, $s is used.";
}
$desc = join("\n$rpad", grep { $_ } $desc =~ m/(.{0,$rcol})(?:\s+|$)/g);
$desc =~ s/ +$//mg;
if ( $short ) {
$usage .= sprintf(" --%-${maxs}s -%s %s\n", $long, $short, $desc);
}
else {
$usage .= sprintf(" --%-${lcol}s %s\n", $long, $desc);
}
}
}
$usage .= "\nOption types: s=string, i=integer, f=float, h/H/a/A=comma-separated list, d=DSN, z=size, m=time\n";
if ( (my @rules = @{$self->{rules}}) ) {
$usage .= "\nRules:\n\n";
$usage .= join("\n", map { " $_" } @rules) . "\n";
}
if ( $self->{DSNParser} ) {
$usage .= "\n" . $self->{DSNParser}->usage();
}
$usage .= "\nOptions and values after processing arguments:\n\n";
foreach my $opt ( sort { $a->{long} cmp $b->{long} } @opts ) {
my $val = $opt->{value};
my $type = $opt->{type} || '';
my $bool = $opt->{spec} =~ m/^[\w-]+(?:\|[\w-])?!?$/;
$val = $bool ? ( $val ? 'TRUE' : 'FALSE' )
: !defined $val ? '(No value)'
: $type eq 'd' ? $self->{DSNParser}->as_string($val)
: $type =~ m/H|h/ ? join(',', sort keys %$val)
: $type =~ m/A|a/ ? join(',', @$val)
: $val;
$usage .= sprintf(" --%-${lcol}s %s\n", $opt->{long}, $val);
}
return $usage;
}
sub prompt_noecho {
shift @_ if ref $_[0] eq __PACKAGE__;
my ( $prompt ) = @_;
local $OUTPUT_AUTOFLUSH = 1;
print $prompt
or die "Cannot print: $OS_ERROR";
my $response;
eval {
require Term::ReadKey;
Term::ReadKey::ReadMode('noecho');
chomp($response = <STDIN>);
Term::ReadKey::ReadMode('normal');
print "\n"
or die "Cannot print: $OS_ERROR";
};
if ( $EVAL_ERROR ) {
die "Cannot read response; is Term::ReadKey installed? $EVAL_ERROR";
}
return $response;
}
sub _read_config_file {
my ( $self, $filename ) = @_;
open my $fh, "<", $filename or die "Cannot open $filename: $OS_ERROR\n";
my @args;
my $prefix = '--';
my $parse = 1;
LINE:
while ( my $line = <$fh> ) {
chomp $line;
next LINE if $line =~ m/^\s*(?:\#|\;|$)/;
$line =~ s/\s+#.*$//g;
$line =~ s/^\s+|\s+$//g;
if ( $line eq '--' ) {
$prefix = '';
$parse = 0;
next LINE;
}
if ( $parse
&& (my($opt, $arg) = $line =~ m/^\s*([^=\s]+?)(?:\s*=\s*(.*?)\s*)?$/)
) {
push @args, grep { defined $_ } ("$prefix$opt", $arg);
}
elsif ( $line =~ m/./ ) {
push @args, $line;
}
else {
die "Syntax error in file $filename at line $INPUT_LINE_NUMBER";
}
}
close $fh;
return @args;
}
sub read_para_after {
my ( $self, $file, $regex ) = @_;
open my $fh, "<", $file or die "Can't open $file: $OS_ERROR";
local $INPUT_RECORD_SEPARATOR = '';
my $para;
while ( $para = <$fh> ) {
next unless $para =~ m/^=pod$/m;
last;
}
while ( $para = <$fh> ) {
next unless $para =~ m/$regex/;
last;
}
$para = <$fh>;
chomp($para);
close $fh or die "Can't close $file: $OS_ERROR";
return $para;
}
sub clone {
my ( $self ) = @_;
my %clone = map {
my $hashref = $self->{$_};
my $val_copy = {};
foreach my $key ( keys %$hashref ) {
my $ref = ref $hashref->{$key};
$val_copy->{$key} = !$ref ? $hashref->{$key}
: $ref eq 'HASH' ? { %{$hashref->{$key}} }
: $ref eq 'ARRAY' ? [ @{$hashref->{$key}} ]
: $hashref->{$key};
}
$_ => $val_copy;
} qw(opts short_opts defaults);
foreach my $scalar ( qw(got_opts) ) {
$clone{$scalar} = $self->{$scalar};
}
return bless \%clone;
}
sub _parse_size {
my ( $self, $opt, $val ) = @_;
if ( lc($val || '') eq 'null' ) {
MKDEBUG && _d('NULL size for', $opt->{long});
$opt->{value} = 'null';
return;
}
my %factor_for = (k => 1_024, M => 1_048_576, G => 1_073_741_824);
my ($pre, $num, $factor) = $val =~ m/^([+-])?(\d+)([kMG])?$/;
if ( defined $num ) {
if ( $factor ) {
$num *= $factor_for{$factor};
MKDEBUG && _d('Setting option', $opt->{y},
'to num', $num, '* factor', $factor);
}
$opt->{value} = ($pre || '') . $num;
}
else {
$self->save_error("Invalid size for --$opt->{long}");
}
return;
}
sub _parse_attribs {
my ( $self, $option, $attribs ) = @_;
my $types = $self->{types};
return $option
. ($attribs->{'short form'} ? '|' . $attribs->{'short form'} : '' )
. ($attribs->{'negatable'} ? '!' : '' )
. ($attribs->{'cumulative'} ? '+' : '' )
. ($attribs->{'type'} ? '=' . $types->{$attribs->{type}} : '' );
}
sub _parse_synopsis {
my ( $self, $file ) = @_;
$file ||= $self->{file} || __FILE__;
MKDEBUG && _d("Parsing SYNOPSIS in", $file);
local $INPUT_RECORD_SEPARATOR = ''; # read paragraphs
open my $fh, "<", $file or die "Cannot open $file: $OS_ERROR";
my $para;
1 while defined($para = <$fh>) && $para !~ m/^=head1 SYNOPSIS/;
die "$file does not contain a SYNOPSIS section" unless $para;
my @synop;
for ( 1..2 ) { # 1 for the usage, 2 for the description
my $para = <$fh>;
push @synop, $para;
}
close $fh;
MKDEBUG && _d("Raw SYNOPSIS text:", @synop);
my ($usage, $desc) = @synop;
die "The SYNOPSIS section in $file is not formatted properly"
unless $usage && $desc;
$usage =~ s/^\s*Usage:\s+(.+)/$1/;
chomp $usage;
$desc =~ s/\n/ /g;
$desc =~ s/\s{2,}/ /g;
$desc =~ s/\. ([A-Z][a-z])/. $1/g;
$desc =~ s/\s+$//;
return (
description => $desc,
usage => $usage,
);
};
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
if ( MKDEBUG ) {
print '# ', $^X, ' ', $], "\n";
if ( my $uname = `uname -a` ) {
$uname =~ s/\s+/ /g;
print "# $uname\n";
}
print '# Arguments: ',
join(' ', map { my $a = "_[$_]_"; $a =~ s/\n/\n# /g; $a; } @ARGV), "\n";
}
1;
}
# ###########################################################################
# End OptionParser package
# ###########################################################################
# ###########################################################################
# Quoter package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/Quoter.pm
# t/lib/Quoter.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package Quoter;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
return bless {}, $class;
}
sub quote {
my ( $self, @vals ) = @_;
foreach my $val ( @vals ) {
$val =~ s/`/``/g;
}
return join('.', map { '`' . $_ . '`' } @vals);
}
sub quote_val {
my ( $self, $val ) = @_;
return 'NULL' unless defined $val; # undef = NULL
return "''" if $val eq ''; # blank string = ''
return $val if $val =~ m/^0x[0-9a-fA-F]+$/; # hex data
$val =~ s/(['\\])/\\$1/g;
return "'$val'";
}
sub split_unquote {
my ( $self, $db_tbl, $default_db ) = @_;
$db_tbl =~ s/`//g;
my ( $db, $tbl ) = split(/[.]/, $db_tbl);
if ( !$tbl ) {
$tbl = $db;
$db = $default_db;
}
return ($db, $tbl);
}
sub literal_like {
my ( $self, $like ) = @_;
return unless $like;
$like =~ s/([%_])/\\$1/g;
return "'$like'";
}
sub join_quote {
my ( $self, $default_db, $db_tbl ) = @_;
return unless $db_tbl;
my ($db, $tbl) = split(/[.]/, $db_tbl);
if ( !$tbl ) {
$tbl = $db;
$db = $default_db;
}
$db = "`$db`" if $db && $db !~ m/^`/;
$tbl = "`$tbl`" if $tbl && $tbl !~ m/^`/;
return $db ? "$db.$tbl" : $tbl;
}
1;
}
# ###########################################################################
# End Quoter package
# ###########################################################################
# ###########################################################################
# DSNParser package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/DSNParser.pm
# t/lib/DSNParser.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package DSNParser;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Data::Dumper;
$Data::Dumper::Indent = 0;
$Data::Dumper::Quotekeys = 0;
eval {
require DBI;
};
my $have_dbi = $EVAL_ERROR ? 0 : 1;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(opts) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = {
opts => {} # h, P, u, etc. Should come from DSN OPTIONS section in POD.
};
foreach my $opt ( @{$args{opts}} ) {
if ( !$opt->{key} || !$opt->{desc} ) {
die "Invalid DSN option: ", Dumper($opt);
}
MKDEBUG && _d('DSN option:',
join(', ',
map { "$_=" . (defined $opt->{$_} ? ($opt->{$_} || '') : 'undef') }
keys %$opt
)
);
$self->{opts}->{$opt->{key}} = {
dsn => $opt->{dsn},
desc => $opt->{desc},
copy => $opt->{copy} || 0,
};
}
return bless $self, $class;
}
sub prop {
my ( $self, $prop, $value ) = @_;
if ( @_ > 2 ) {
MKDEBUG && _d('Setting', $prop, 'property');
$self->{$prop} = $value;
}
return $self->{$prop};
}
sub parse {
my ( $self, $dsn, $prev, $defaults ) = @_;
if ( !$dsn ) {
MKDEBUG && _d('No DSN to parse');
return;
}
MKDEBUG && _d('Parsing', $dsn);
$prev ||= {};
$defaults ||= {};
my %given_props;
my %final_props;
my $opts = $self->{opts};
foreach my $dsn_part ( split(/,/, $dsn) ) {
if ( my ($prop_key, $prop_val) = $dsn_part =~ m/^(.)=(.*)$/ ) {
$given_props{$prop_key} = $prop_val;
}
else {
MKDEBUG && _d('Interpreting', $dsn_part, 'as h=', $dsn_part);
$given_props{h} = $dsn_part;
}
}
foreach my $key ( keys %$opts ) {
MKDEBUG && _d('Finding value for', $key);
$final_props{$key} = $given_props{$key};
if ( !defined $final_props{$key}
&& defined $prev->{$key} && $opts->{$key}->{copy} )
{
$final_props{$key} = $prev->{$key};
MKDEBUG && _d('Copying value for', $key, 'from previous DSN');
}
if ( !defined $final_props{$key} ) {
$final_props{$key} = $defaults->{$key};
MKDEBUG && _d('Copying value for', $key, 'from defaults');
}
}
foreach my $key ( keys %given_props ) {
die "Unknown DSN option '$key' in '$dsn'. For more details, "
. "please use the --help option, or try 'perldoc $PROGRAM_NAME' "
. "for complete documentation."
unless exists $opts->{$key};
}
if ( (my $required = $self->prop('required')) ) {
foreach my $key ( keys %$required ) {
die "Missing required DSN option '$key' in '$dsn'. For more details, "
. "please use the --help option, or try 'perldoc $PROGRAM_NAME' "
. "for complete documentation."
unless $final_props{$key};
}
}
return \%final_props;
}
sub parse_options {
my ( $self, $o ) = @_;
die 'I need an OptionParser object' unless ref $o eq 'OptionParser';
my $dsn_string
= join(',',
map { "$_=".$o->get($_); }
grep { $o->has($_) && $o->get($_) }
keys %{$self->{opts}}
);
MKDEBUG && _d('DSN string made from options:', $dsn_string);
return $self->parse($dsn_string);
}
sub as_string {
my ( $self, $dsn, $props ) = @_;
return $dsn unless ref $dsn;
my %allowed = $props ? map { $_=>1 } @$props : ();
return join(',',
map { "$_=" . ($_ eq 'p' ? '...' : $dsn->{$_}) }
grep { defined $dsn->{$_} && $self->{opts}->{$_} }
grep { !$props || $allowed{$_} }
sort keys %$dsn );
}
sub usage {
my ( $self ) = @_;
my $usage
= "DSN syntax is key=value[,key=value...] Allowable DSN keys:\n\n"
. " KEY COPY MEANING\n"
. " === ==== =============================================\n";
my %opts = %{$self->{opts}};
foreach my $key ( sort keys %opts ) {
$usage .= " $key "
. ($opts{$key}->{copy} ? 'yes ' : 'no ')
. ($opts{$key}->{desc} || '[No description]')
. "\n";
}
$usage .= "\n If the DSN is a bareword, the word is treated as the 'h' key.\n";
return $usage;
}
sub get_cxn_params {
my ( $self, $info ) = @_;
my $dsn;
my %opts = %{$self->{opts}};
my $driver = $self->prop('dbidriver') || '';
if ( $driver eq 'Pg' ) {
$dsn = 'DBI:Pg:dbname=' . ( $info->{D} || '' ) . ';'
. join(';', map { "$opts{$_}->{dsn}=$info->{$_}" }
grep { defined $info->{$_} }
qw(h P));
}
else {
$dsn = 'DBI:mysql:' . ( $info->{D} || '' ) . ';'
. join(';', map { "$opts{$_}->{dsn}=$info->{$_}" }
grep { defined $info->{$_} }
qw(F h P S A))
. ';mysql_read_default_group=client';
}
MKDEBUG && _d($dsn);
return ($dsn, $info->{u}, $info->{p});
}
sub fill_in_dsn {
my ( $self, $dbh, $dsn ) = @_;
my $vars = $dbh->selectall_hashref('SHOW VARIABLES', 'Variable_name');
my ($user, $db) = $dbh->selectrow_array('SELECT USER(), DATABASE()');
$user =~ s/@.*//;
$dsn->{h} ||= $vars->{hostname}->{Value};
$dsn->{S} ||= $vars->{'socket'}->{Value};
$dsn->{P} ||= $vars->{port}->{Value};
$dsn->{u} ||= $user;
$dsn->{D} ||= $db;
}
sub get_dbh {
my ( $self, $cxn_string, $user, $pass, $opts ) = @_;
$opts ||= {};
my $defaults = {
AutoCommit => 0,
RaiseError => 1,
PrintError => 0,
ShowErrorStatement => 1,
mysql_enable_utf8 => ($cxn_string =~ m/charset=utf8/i ? 1 : 0),
#Added by Alexander Garzon 2011-11-18
mysql_compression => 1,
};
@{$defaults}{ keys %$opts } = values %$opts;
if ( $opts->{mysql_use_result} ) {
$defaults->{mysql_use_result} = 1;
}
if ( !$have_dbi ) {
die "Cannot connect to MySQL because the Perl DBI module is not "
. "installed or not found. Run 'perl -MDBI' to see the directories "
. "that Perl searches for DBI. If DBI is not installed, try:\n"
. " Debian/Ubuntu apt-get install libdbi-perl\n"
. " RHEL/CentOS yum install perl-DBI\n"
. " OpenSolaris pgk install pkg:/SUNWpmdbi\n";
}
my $dbh;
my $tries = 2;
while ( !$dbh && $tries-- ) {
MKDEBUG && _d($cxn_string, ' ', $user, ' ', $pass,
join(', ', map { "$_=>$defaults->{$_}" } keys %$defaults ));
eval {
$dbh = DBI->connect($cxn_string, $user, $pass, $defaults);
if ( $cxn_string =~ m/mysql/i ) {
my $sql;
$sql = 'SELECT @@SQL_MODE';
MKDEBUG && _d($dbh, $sql);
my ($sql_mode) = $dbh->selectrow_array($sql);
$sql = 'SET @@SQL_QUOTE_SHOW_CREATE = 1'
. '/*!40101, @@SQL_MODE=\'NO_AUTO_VALUE_ON_ZERO'
. ($sql_mode ? ",$sql_mode" : '')
. '\'*/';
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
if ( my ($charset) = $cxn_string =~ m/charset=(\w+)/ ) {
$sql = "/*!40101 SET NAMES $charset*/";
MKDEBUG && _d($dbh, ':', $sql);
$dbh->do($sql);
MKDEBUG && _d('Enabling charset for STDOUT');
if ( $charset eq 'utf8' ) {
binmode(STDOUT, ':utf8')
or die "Can't binmode(STDOUT, ':utf8'): $OS_ERROR";
}
else {
binmode(STDOUT) or die "Can't binmode(STDOUT): $OS_ERROR";
}
}
if ( $self->prop('set-vars') ) {
$sql = "SET " . $self->prop('set-vars');
MKDEBUG && _d($dbh, ':', $sql);
$dbh->do($sql);
}
}
};
if ( !$dbh && $EVAL_ERROR ) {
MKDEBUG && _d($EVAL_ERROR);
if ( $EVAL_ERROR =~ m/not a compiled character set|character set utf8/ ) {
MKDEBUG && _d('Going to try again without utf8 support');
delete $defaults->{mysql_enable_utf8};
}
elsif ( $EVAL_ERROR =~ m/locate DBD\/mysql/i ) {
die "Cannot connect to MySQL because the Perl DBD::mysql module is "
. "not installed or not found. Run 'perl -MDBD::mysql' to see "
. "the directories that Perl searches for DBD::mysql. If "
. "DBD::mysql is not installed, try:\n"
. " Debian/Ubuntu apt-get install libdbd-mysql-perl\n"
. " RHEL/CentOS yum install perl-DBD-MySQL\n"
. " OpenSolaris pgk install pkg:/SUNWapu13dbd-mysql\n";
}
if ( !$tries ) {
die $EVAL_ERROR;
}
}
}
MKDEBUG && _d('DBH info: ',
$dbh,
Dumper($dbh->selectrow_hashref(
'SELECT DATABASE(), CONNECTION_ID(), VERSION()/*!50038 , @@hostname*/')),
'Connection info:', $dbh->{mysql_hostinfo},
'Character set info:', Dumper($dbh->selectall_arrayref(
'SHOW VARIABLES LIKE "character_set%"', { Slice => {}})),
'$DBD::mysql::VERSION:', $DBD::mysql::VERSION,
'$DBI::VERSION:', $DBI::VERSION,
);
return $dbh;
}
sub get_hostname {
my ( $self, $dbh ) = @_;
if ( my ($host) = ($dbh->{mysql_hostinfo} || '') =~ m/^(\w+) via/ ) {
return $host;
}
my ( $hostname, $one ) = $dbh->selectrow_array(
'SELECT /*!50038 @@hostname, */ 1');
return $hostname;
}
sub disconnect {
my ( $self, $dbh ) = @_;
MKDEBUG && $self->print_active_handles($dbh);
$dbh->disconnect;
}
sub print_active_handles {
my ( $self, $thing, $level ) = @_;
$level ||= 0;
printf("# Active %sh: %s %s %s\n", ($thing->{Type} || 'undef'), "\t" x $level,
$thing, (($thing->{Type} || '') eq 'st' ? $thing->{Statement} || '' : ''))
or die "Cannot print: $OS_ERROR";
foreach my $handle ( grep {defined} @{ $thing->{ChildHandles} } ) {
$self->print_active_handles( $handle, $level + 1 );
}
}
sub copy {
my ( $self, $dsn_1, $dsn_2, %args ) = @_;
die 'I need a dsn_1 argument' unless $dsn_1;
die 'I need a dsn_2 argument' unless $dsn_2;
my %new_dsn = map {
my $key = $_;
my $val;
if ( $args{overwrite} ) {
$val = defined $dsn_1->{$key} ? $dsn_1->{$key} : $dsn_2->{$key};
}
else {
$val = defined $dsn_2->{$key} ? $dsn_2->{$key} : $dsn_1->{$key};
}
$key => $val;
} keys %{$self->{opts}};
return \%new_dsn;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End DSNParser package
# ###########################################################################
# ###########################################################################
# VersionParser package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/VersionParser.pm
# t/lib/VersionParser.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package VersionParser;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class ) = @_;
bless {}, $class;
}
sub parse {
my ( $self, $str ) = @_;
my $result = sprintf('%03d%03d%03d', $str =~ m/(\d+)/g);
MKDEBUG && _d($str, 'parses to', $result);
return $result;
}
sub version_ge {
my ( $self, $dbh, $target ) = @_;
if ( !$self->{$dbh} ) {
$self->{$dbh} = $self->parse(
$dbh->selectrow_array('SELECT VERSION()'));
}
my $result = $self->{$dbh} ge $self->parse($target) ? 1 : 0;
MKDEBUG && _d($self->{$dbh}, 'ge', $target, ':', $result);
return $result;
}
sub innodb_version {
my ( $self, $dbh ) = @_;
return unless $dbh;
my $innodb_version = "NO";
my ($innodb) =
grep { $_->{engine} =~ m/InnoDB/i }
map {
my %hash;
@hash{ map { lc $_ } keys %$_ } = values %$_;
\%hash;
}
@{ $dbh->selectall_arrayref("SHOW ENGINES", {Slice=>{}}) };
if ( $innodb ) {
MKDEBUG && _d("InnoDB support:", $innodb->{support});
if ( $innodb->{support} =~ m/YES|DEFAULT/i ) {
my $vars = $dbh->selectrow_hashref(
"SHOW VARIABLES LIKE 'innodb_version'");
$innodb_version = !$vars ? "BUILTIN"
: ($vars->{Value} || $vars->{value});
}
else {
$innodb_version = $innodb->{support}; # probably DISABLED or NO
}
}
MKDEBUG && _d("InnoDB version:", $innodb_version);
return $innodb_version;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End VersionParser package
# ###########################################################################
# ###########################################################################
# TableSyncStream package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableSyncStream.pm
# t/lib/TableSyncStream.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableSyncStream;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub name {
return 'Stream';
}
sub can_sync {
return 1; # We can sync anything.
}
sub prepare_to_sync {
my ( $self, %args ) = @_;
my @required_args = qw(cols ChangeHandler);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
$self->{cols} = $args{cols};
$self->{buffer_in_mysql} = $args{buffer_in_mysql};
$self->{ChangeHandler} = $args{ChangeHandler};
$self->{done} = 0;
return;
}
sub uses_checksum {
return 0; # We don't need checksum queries.
}
sub set_checksum_queries {
return; # This shouldn't be called, but just in case.
}
sub prepare_sync_cycle {
my ( $self, $host ) = @_;
return;
}
sub get_sql {
my ( $self, %args ) = @_;
return "SELECT "
. ($self->{buffer_in_mysql} ? 'SQL_BUFFER_RESULT ' : '')
. join(', ', map { $self->{Quoter}->quote($_) } @{$self->{cols}})
. ' FROM ' . $self->{Quoter}->quote(@args{qw(database table)})
. ' WHERE ' . ( $args{where} || '1=1' );
}
sub same_row {
my ( $self, %args ) = @_;
return;
}
sub not_in_right {
my ( $self, %args ) = @_;
$self->{ChangeHandler}->change('INSERT', $args{lr}, $self->key_cols());
}
sub not_in_left {
my ( $self, %args ) = @_;
$self->{ChangeHandler}->change('DELETE', $args{rr}, $self->key_cols());
}
sub done_with_rows {
my ( $self ) = @_;
$self->{done} = 1;
}
sub done {
my ( $self ) = @_;
return $self->{done};
}
sub key_cols {
my ( $self ) = @_;
return $self->{cols};
}
sub pending_changes {
my ( $self ) = @_;
return;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableSyncStream package
# ###########################################################################
# ###########################################################################
# TableParser package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableParser.pm
# t/lib/TableParser.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableParser;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
sub new {
my ( $class, %args ) = @_;
my @required_args = qw(Quoter);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub parse {
my ( $self, $ddl, $opts ) = @_;
return unless $ddl;
if ( ref $ddl eq 'ARRAY' ) {
if ( lc $ddl->[0] eq 'table' ) {
$ddl = $ddl->[1];
}
else {
return {
engine => 'VIEW',
};
}
}
if ( $ddl !~ m/CREATE (?:TEMPORARY )?TABLE `/ ) {
die "Cannot parse table definition; is ANSI quoting "
. "enabled or SQL_QUOTE_SHOW_CREATE disabled?";
}
my ($name) = $ddl =~ m/CREATE (?:TEMPORARY )?TABLE\s+(`.+?`)/;
(undef, $name) = $self->{Quoter}->split_unquote($name) if $name;
$ddl =~ s/(`[^`]+`)/\L$1/g;
my $engine = $self->get_engine($ddl);
my @defs = $ddl =~ m/^(\s+`.*?),?$/gm;
my @cols = map { $_ =~ m/`([^`]+)`/ } @defs;
MKDEBUG && _d('Table cols:', join(', ', map { "`$_`" } @cols));
my %def_for;
@def_for{@cols} = @defs;
my (@nums, @null);
my (%type_for, %is_nullable, %is_numeric, %is_autoinc);
foreach my $col ( @cols ) {
my $def = $def_for{$col};
my ( $type ) = $def =~ m/`[^`]+`\s([a-z]+)/;
die "Can't determine column type for $def" unless $type;
$type_for{$col} = $type;
if ( $type =~ m/(?:(?:tiny|big|medium|small)?int|float|double|decimal|year)/ ) {
push @nums, $col;
$is_numeric{$col} = 1;
}
if ( $def !~ m/NOT NULL/ ) {
push @null, $col;
$is_nullable{$col} = 1;
}
$is_autoinc{$col} = $def =~ m/AUTO_INCREMENT/i ? 1 : 0;
}
my ($keys, $clustered_key) = $self->get_keys($ddl, $opts, \%is_nullable);
my ($charset) = $ddl =~ m/DEFAULT CHARSET=(\w+)/;
return {
name => $name,
cols => \@cols,
col_posn => { map { $cols[$_] => $_ } 0..$#cols },
is_col => { map { $_ => 1 } @cols },
null_cols => \@null,
is_nullable => \%is_nullable,
is_autoinc => \%is_autoinc,
clustered_key => $clustered_key,
keys => $keys,
defs => \%def_for,
numeric_cols => \@nums,
is_numeric => \%is_numeric,
engine => $engine,
type_for => \%type_for,
charset => $charset,
};
}
sub sort_indexes {
my ( $self, $tbl ) = @_;
my @indexes
= sort {
(($a ne 'PRIMARY') <=> ($b ne 'PRIMARY'))
|| ( !$tbl->{keys}->{$a}->{is_unique} <=> !$tbl->{keys}->{$b}->{is_unique} )
|| ( $tbl->{keys}->{$a}->{is_nullable} <=> $tbl->{keys}->{$b}->{is_nullable} )
|| ( scalar(@{$tbl->{keys}->{$a}->{cols}}) <=> scalar(@{$tbl->{keys}->{$b}->{cols}}) )
}
grep {
$tbl->{keys}->{$_}->{type} eq 'BTREE'
}
sort keys %{$tbl->{keys}};
MKDEBUG && _d('Indexes sorted best-first:', join(', ', @indexes));
return @indexes;
}
sub find_best_index {
my ( $self, $tbl, $index ) = @_;
my $best;
if ( $index ) {
($best) = grep { uc $_ eq uc $index } keys %{$tbl->{keys}};
}
if ( !$best ) {
if ( $index ) {
die "Index '$index' does not exist in table";
}
else {
($best) = $self->sort_indexes($tbl);
}
}
MKDEBUG && _d('Best index found is', $best);
return $best;
}
sub find_possible_keys {
my ( $self, $dbh, $database, $table, $quoter, $where ) = @_;
return () unless $where;
my $sql = 'EXPLAIN SELECT * FROM ' . $quoter->quote($database, $table)
. ' WHERE ' . $where;
MKDEBUG && _d($sql);
my $expl = $dbh->selectrow_hashref($sql);
$expl = { map { lc($_) => $expl->{$_} } keys %$expl };
if ( $expl->{possible_keys} ) {
MKDEBUG && _d('possible_keys =', $expl->{possible_keys});
my @candidates = split(',', $expl->{possible_keys});
my %possible = map { $_ => 1 } @candidates;
if ( $expl->{key} ) {
MKDEBUG && _d('MySQL chose', $expl->{key});
unshift @candidates, grep { $possible{$_} } split(',', $expl->{key});
MKDEBUG && _d('Before deduping:', join(', ', @candidates));
my %seen;
@candidates = grep { !$seen{$_}++ } @candidates;
}
MKDEBUG && _d('Final list:', join(', ', @candidates));
return @candidates;
}
else {
MKDEBUG && _d('No keys in possible_keys');
return ();
}
}
sub check_table {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db, $tbl) = @args{@required_args};
my $q = $self->{Quoter};
my $db_tbl = $q->quote($db, $tbl);
MKDEBUG && _d('Checking', $db_tbl);
my $sql = "SHOW TABLES FROM " . $q->quote($db)
. ' LIKE ' . $q->literal_like($tbl);
MKDEBUG && _d($sql);
my $row;
eval {
$row = $dbh->selectrow_arrayref($sql);
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d($EVAL_ERROR);
return 0;
}
if ( !$row->[0] || $row->[0] ne $tbl ) {
MKDEBUG && _d('Table does not exist');
return 0;
}
MKDEBUG && _d('Table exists; no privs to check');
return 1 unless $args{all_privs};
$sql = "SHOW FULL COLUMNS FROM $db_tbl";
MKDEBUG && _d($sql);
eval {
$row = $dbh->selectrow_hashref($sql);
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d($EVAL_ERROR);
return 0;
}
if ( !scalar keys %$row ) {
MKDEBUG && _d('Table has no columns:', Dumper($row));
return 0;
}
my $privs = $row->{privileges} || $row->{Privileges};
$sql = "DELETE FROM $db_tbl LIMIT 0";
MKDEBUG && _d($sql);
eval {
$dbh->do($sql);
};
my $can_delete = $EVAL_ERROR ? 0 : 1;
MKDEBUG && _d('User privs on', $db_tbl, ':', $privs,
($can_delete ? 'delete' : ''));
if ( !($privs =~ m/select/ && $privs =~ m/insert/ && $privs =~ m/update/
&& $can_delete) ) {
MKDEBUG && _d('User does not have all privs');
return 0;
}
MKDEBUG && _d('User has all privs');
return 1;
}
sub get_engine {
my ( $self, $ddl, $opts ) = @_;
my ( $engine ) = $ddl =~ m/\).*?(?:ENGINE|TYPE)=(\w+)/;
MKDEBUG && _d('Storage engine:', $engine);
return $engine || undef;
}
sub get_keys {
my ( $self, $ddl, $opts, $is_nullable ) = @_;
my $engine = $self->get_engine($ddl);
my $keys = {};
my $clustered_key = undef;
KEY:
foreach my $key ( $ddl =~ m/^ ((?:[A-Z]+ )?KEY .*)$/gm ) {
next KEY if $key =~ m/FOREIGN/;
my $key_ddl = $key;
MKDEBUG && _d('Parsed key:', $key_ddl);
if ( $engine !~ m/MEMORY|HEAP/ ) {
$key =~ s/USING HASH/USING BTREE/;
}
my ( $type, $cols ) = $key =~ m/(?:USING (\w+))? \((.+)\)/;
my ( $special ) = $key =~ m/(FULLTEXT|SPATIAL)/;
$type = $type || $special || 'BTREE';
if ( $opts->{mysql_version} && $opts->{mysql_version} lt '004001000'
&& $engine =~ m/HEAP|MEMORY/i )
{
$type = 'HASH'; # MySQL pre-4.1 supports only HASH indexes on HEAP
}
my ($name) = $key =~ m/(PRIMARY|`[^`]*`)/;
my $unique = $key =~ m/PRIMARY|UNIQUE/ ? 1 : 0;
my @cols;
my @col_prefixes;
foreach my $col_def ( $cols =~ m/`[^`]+`(?:\(\d+\))?/g ) {
my ($name, $prefix) = $col_def =~ m/`([^`]+)`(?:\((\d+)\))?/;
push @cols, $name;
push @col_prefixes, $prefix;
}
$name =~ s/`//g;
MKDEBUG && _d( $name, 'key cols:', join(', ', map { "`$_`" } @cols));
$keys->{$name} = {
name => $name,
type => $type,
colnames => $cols,
cols => \@cols,
col_prefixes => \@col_prefixes,
is_unique => $unique,
is_nullable => scalar(grep { $is_nullable->{$_} } @cols),
is_col => { map { $_ => 1 } @cols },
ddl => $key_ddl,
};
if ( $engine =~ m/InnoDB/i && !$clustered_key ) {
my $this_key = $keys->{$name};
if ( $this_key->{name} eq 'PRIMARY' ) {
$clustered_key = 'PRIMARY';
}
elsif ( $this_key->{is_unique} && !$this_key->{is_nullable} ) {
$clustered_key = $this_key->{name};
}
MKDEBUG && $clustered_key && _d('This key is the clustered key');
}
}
return $keys, $clustered_key;
}
sub get_fks {
my ( $self, $ddl, $opts ) = @_;
my $q = $self->{Quoter};
my $fks = {};
foreach my $fk (
$ddl =~ m/CONSTRAINT .* FOREIGN KEY .* REFERENCES [^\)]*\)/mg )
{
my ( $name ) = $fk =~ m/CONSTRAINT `(.*?)`/;
my ( $cols ) = $fk =~ m/FOREIGN KEY \(([^\)]+)\)/;
my ( $parent, $parent_cols ) = $fk =~ m/REFERENCES (\S+) \(([^\)]+)\)/;
my ($db, $tbl) = $q->split_unquote($parent, $opts->{database});
my %parent_tbl = (tbl => $tbl);
$parent_tbl{db} = $db if $db;
if ( $parent !~ m/\./ && $opts->{database} ) {
$parent = $q->quote($opts->{database}) . ".$parent";
}
$fks->{$name} = {
name => $name,
colnames => $cols,
cols => [ map { s/[ `]+//g; $_; } split(',', $cols) ],
parent_tbl => \%parent_tbl,
parent_tblname => $parent,
parent_cols => [ map { s/[ `]+//g; $_; } split(',', $parent_cols) ],
parent_colnames=> $parent_cols,
ddl => $fk,
};
}
return $fks;
}
sub remove_auto_increment {
my ( $self, $ddl ) = @_;
$ddl =~ s/(^\).*?) AUTO_INCREMENT=\d+\b/$1/m;
return $ddl;
}
sub remove_secondary_indexes {
my ( $self, $ddl ) = @_;
my $sec_indexes_ddl;
my $tbl_struct = $self->parse($ddl);
if ( ($tbl_struct->{engine} || '') =~ m/InnoDB/i ) {
my $clustered_key = $tbl_struct->{clustered_key};
$clustered_key ||= '';
my @sec_indexes = map {
my $key_def = $_->{ddl};
$key_def =~ s/([\(\)])/\\$1/g;
$ddl =~ s/\s+$key_def//i;
my $key_ddl = "ADD $_->{ddl}";
$key_ddl .= ',' unless $key_ddl =~ m/,$/;
$key_ddl;
}
grep { $_->{name} ne $clustered_key }
values %{$tbl_struct->{keys}};
MKDEBUG && _d('Secondary indexes:', Dumper(\@sec_indexes));
if ( @sec_indexes ) {
$sec_indexes_ddl = join(' ', @sec_indexes);
$sec_indexes_ddl =~ s/,$//;
}
$ddl =~ s/,(\n\) )/$1/s;
}
else {
MKDEBUG && _d('Not removing secondary indexes from',
$tbl_struct->{engine}, 'table');
}
return $ddl, $sec_indexes_ddl, $tbl_struct;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableParser package
# ###########################################################################
# ###########################################################################
# RowDiff package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/RowDiff.pm
# t/lib/RowDiff.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package RowDiff;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
die "I need a dbh" unless $args{dbh};
my $self = { %args };
return bless $self, $class;
}
sub compare_sets {
my ( $self, %args ) = @_;
my @required_args = qw(left_sth right_sth syncer tbl_struct);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $left_sth = $args{left_sth};
my $right_sth = $args{right_sth};
my $syncer = $args{syncer};
my $tbl_struct = $args{tbl_struct};
my ($lr, $rr); # Current row from the left/right sths.
$args{key_cols} = $syncer->key_cols(); # for key_cmp()
my $left_done = 0;
my $right_done = 0;
my $done = $self->{done};
do {
if ( !$lr && !$left_done ) {
MKDEBUG && _d('Fetching row from left');
eval { $lr = $left_sth->fetchrow_hashref(); };
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
$left_done = !$lr || $EVAL_ERROR ? 1 : 0;
}
elsif ( MKDEBUG ) {
_d('Left still has rows');
}
if ( !$rr && !$right_done ) {
MKDEBUG && _d('Fetching row from right');
eval { $rr = $right_sth->fetchrow_hashref(); };
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
$right_done = !$rr || $EVAL_ERROR ? 1 : 0;
}
elsif ( MKDEBUG ) {
_d('Right still has rows');
}
my $cmp;
if ( $lr && $rr ) {
$cmp = $self->key_cmp(%args, lr => $lr, rr => $rr);
MKDEBUG && _d('Key comparison on left and right:', $cmp);
}
if ( $lr || $rr ) {
if ( $lr && $rr && defined $cmp && $cmp == 0 ) {
MKDEBUG && _d('Left and right have the same key');
$syncer->same_row(%args, lr => $lr, rr => $rr);
$self->{same_row}->(%args, lr => $lr, rr => $rr)
if $self->{same_row};
$lr = $rr = undef; # Fetch another row from each side.
}
elsif ( !$rr || ( defined $cmp && $cmp < 0 ) ) {
MKDEBUG && _d('Left is not in right');
$syncer->not_in_right(%args, lr => $lr, rr => $rr);
$self->{not_in_right}->(%args, lr => $lr, rr => $rr)
if $self->{not_in_right};
$lr = undef;
}
else {
MKDEBUG && _d('Right is not in left');
$syncer->not_in_left(%args, lr => $lr, rr => $rr);
$self->{not_in_left}->(%args, lr => $lr, rr => $rr)
if $self->{not_in_left};
$rr = undef;
}
}
$left_done = $right_done = 1 if $done && $done->(%args);
} while ( !($left_done && $right_done) );
MKDEBUG && _d('No more rows');
$syncer->done_with_rows();
}
sub key_cmp {
my ( $self, %args ) = @_;
my @required_args = qw(lr rr key_cols tbl_struct);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless exists $args{$arg};
}
my ($lr, $rr, $key_cols, $tbl_struct) = @args{@required_args};
MKDEBUG && _d('Comparing keys using columns:', join(',', @$key_cols));
my $callback = $self->{key_cmp};
my $trf = $self->{trf};
foreach my $col ( @$key_cols ) {
my $l = $lr->{$col};
my $r = $rr->{$col};
if ( !defined $l || !defined $r ) {
MKDEBUG && _d($col, 'is not defined in both rows');
return defined $l ? 1 : defined $r ? -1 : 0;
}
else {
if ( $tbl_struct->{is_numeric}->{$col} ) { # Numeric column
MKDEBUG && _d($col, 'is numeric');
($l, $r) = $trf->($l, $r, $tbl_struct, $col) if $trf;
my $cmp = $l <=> $r;
if ( $cmp ) {
MKDEBUG && _d('Column', $col, 'differs:', $l, '!=', $r);
$callback->($col, $l, $r) if $callback;
return $cmp;
}
}
elsif ( $l ne $r ) {
my $cmp;
my $coll = $tbl_struct->{collation_for}->{$col};
if ( $coll && ( $coll ne 'latin1_swedish_ci'
|| $l =~ m/[^\040-\177]/ || $r =~ m/[^\040-\177]/) )
{
MKDEBUG && _d('Comparing', $col, 'via MySQL');
$cmp = $self->db_cmp($coll, $l, $r);
}
else {
MKDEBUG && _d('Comparing', $col, 'in lowercase');
$cmp = lc $l cmp lc $r;
}
if ( $cmp ) {
MKDEBUG && _d('Column', $col, 'differs:', $l, 'ne', $r);
$callback->($col, $l, $r) if $callback;
return $cmp;
}
}
}
}
return 0;
}
sub db_cmp {
my ( $self, $collation, $l, $r ) = @_;
if ( !$self->{sth}->{$collation} ) {
if ( !$self->{charset_for} ) {
MKDEBUG && _d('Fetching collations from MySQL');
my @collations = @{$self->{dbh}->selectall_arrayref(
'SHOW COLLATION', {Slice => { collation => 1, charset => 1 }})};
foreach my $collation ( @collations ) {
$self->{charset_for}->{$collation->{collation}}
= $collation->{charset};
}
}
my $sql = "SELECT STRCMP(_$self->{charset_for}->{$collation}? COLLATE $collation, "
. "_$self->{charset_for}->{$collation}? COLLATE $collation) AS res";
MKDEBUG && _d($sql);
$self->{sth}->{$collation} = $self->{dbh}->prepare($sql);
}
my $sth = $self->{sth}->{$collation};
$sth->execute($l, $r);
return $sth->fetchall_arrayref()->[0]->[0];
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End RowDiff package
# ###########################################################################
# ###########################################################################
# MySQLDump package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/MySQLDump.pm
# t/lib/MySQLDump.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package MySQLDump;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
( our $before = <<'EOF') =~ s/^ //gm;
/*!40101 SET @OLD_CHARACTER_SET_CLIENT=@@CHARACTER_SET_CLIENT */;
/*!40101 SET @OLD_CHARACTER_SET_RESULTS=@@CHARACTER_SET_RESULTS */;
/*!40101 SET @OLD_COLLATION_CONNECTION=@@COLLATION_CONNECTION */;
/*!40101 SET NAMES utf8 */;
/*!40103 SET @OLD_TIME_ZONE=@@TIME_ZONE */;
/*!40103 SET TIME_ZONE='+00:00' */;
/*!40014 SET @OLD_UNIQUE_CHECKS=@@UNIQUE_CHECKS, UNIQUE_CHECKS=0 */;
/*!40014 SET @OLD_FOREIGN_KEY_CHECKS=@@FOREIGN_KEY_CHECKS, FOREIGN_KEY_CHECKS=0 */;
/*!40101 SET @OLD_SQL_MODE=@@SQL_MODE, SQL_MODE='NO_AUTO_VALUE_ON_ZERO' */;
/*!40111 SET @OLD_SQL_NOTES=@@SQL_NOTES, SQL_NOTES=0 */;
EOF
( our $after = <<'EOF') =~ s/^ //gm;
/*!40103 SET TIME_ZONE=@OLD_TIME_ZONE */;
/*!40101 SET SQL_MODE=@OLD_SQL_MODE */;
/*!40014 SET FOREIGN_KEY_CHECKS=@OLD_FOREIGN_KEY_CHECKS */;
/*!40014 SET UNIQUE_CHECKS=@OLD_UNIQUE_CHECKS */;
/*!40101 SET CHARACTER_SET_CLIENT=@OLD_CHARACTER_SET_CLIENT */;
/*!40101 SET CHARACTER_SET_RESULTS=@OLD_CHARACTER_SET_RESULTS */;
/*!40101 SET COLLATION_CONNECTION=@OLD_COLLATION_CONNECTION */;
/*!40111 SET SQL_NOTES=@OLD_SQL_NOTES */;
EOF
sub new {
my ( $class, %args ) = @_;
my $self = {
cache => 0, # Afaik no script uses this cache any longer because
};
return bless $self, $class;
}
sub dump {
my ( $self, $dbh, $quoter, $db, $tbl, $what ) = @_;
if ( $what eq 'table' ) {
my $ddl = $self->get_create_table($dbh, $quoter, $db, $tbl);
return unless $ddl;
if ( $ddl->[0] eq 'table' ) {
return $before
. 'DROP TABLE IF EXISTS ' . $quoter->quote($tbl) . ";\n"
. $ddl->[1] . ";\n";
}
else {
return 'DROP TABLE IF EXISTS ' . $quoter->quote($tbl) . ";\n"
. '/*!50001 DROP VIEW IF EXISTS '
. $quoter->quote($tbl) . "*/;\n/*!50001 "
. $self->get_tmp_table($dbh, $quoter, $db, $tbl) . "*/;\n";
}
}
elsif ( $what eq 'triggers' ) {
my $trgs = $self->get_triggers($dbh, $quoter, $db, $tbl);
if ( $trgs && @$trgs ) {
my $result = $before . "\nDELIMITER ;;\n";
foreach my $trg ( @$trgs ) {
if ( $trg->{sql_mode} ) {
$result .= qq{/*!50003 SET SESSION SQL_MODE='$trg->{sql_mode}' */;;\n};
}
$result .= "/*!50003 CREATE */ ";
if ( $trg->{definer} ) {
my ( $user, $host )
= map { s/'/''/g; "'$_'"; }
split('@', $trg->{definer}, 2);
$result .= "/*!50017 DEFINER=$user\@$host */ ";
}
$result .= sprintf("/*!50003 TRIGGER %s %s %s ON %s\nFOR EACH ROW %s */;;\n\n",
$quoter->quote($trg->{trigger}),
@{$trg}{qw(timing event)},
$quoter->quote($trg->{table}),
$trg->{statement});
}
$result .= "DELIMITER ;\n\n/*!50003 SET SESSION SQL_MODE=\@OLD_SQL_MODE */;\n\n";
return $result;
}
else {
return undef;
}
}
elsif ( $what eq 'view' ) {
my $ddl = $self->get_create_table($dbh, $quoter, $db, $tbl);
return '/*!50001 DROP TABLE IF EXISTS ' . $quoter->quote($tbl) . "*/;\n"
. '/*!50001 DROP VIEW IF EXISTS ' . $quoter->quote($tbl) . "*/;\n"
. '/*!50001 ' . $ddl->[1] . "*/;\n";
}
else {
die "You didn't say what to dump.";
}
}
sub _use_db {
my ( $self, $dbh, $quoter, $new ) = @_;
if ( !$new ) {
MKDEBUG && _d('No new DB to use');
return;
}
my $sql = 'USE ' . $quoter->quote($new);
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
return;
}
sub get_create_table {
my ( $self, $dbh, $quoter, $db, $tbl ) = @_;
if ( !$self->{cache} || !$self->{tables}->{$db}->{$tbl} ) {
my $sql = '/*!40101 SET @OLD_SQL_MODE := @@SQL_MODE, '
. q{@@SQL_MODE := REPLACE(REPLACE(@@SQL_MODE, 'ANSI_QUOTES', ''), ',,', ','), }
. '@OLD_QUOTE := @@SQL_QUOTE_SHOW_CREATE, '
. '@@SQL_QUOTE_SHOW_CREATE := 1 */';
MKDEBUG && _d($sql);
eval { $dbh->do($sql); };
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
$self->_use_db($dbh, $quoter, $db);
$sql = "SHOW CREATE TABLE " . $quoter->quote($db, $tbl);
MKDEBUG && _d($sql);
my $href;
eval { $href = $dbh->selectrow_hashref($sql); };
if ( $EVAL_ERROR ) {
warn "Failed to $sql. The table may be damaged.\nError: $EVAL_ERROR";
return;
}
$sql = '/*!40101 SET @@SQL_MODE := @OLD_SQL_MODE, '
. '@@SQL_QUOTE_SHOW_CREATE := @OLD_QUOTE */';
MKDEBUG && _d($sql);
$dbh->do($sql);
my ($key) = grep { m/create table/i } keys %$href;
if ( $key ) {
MKDEBUG && _d('This table is a base table');
$self->{tables}->{$db}->{$tbl} = [ 'table', $href->{$key} ];
}
else {
MKDEBUG && _d('This table is a view');
($key) = grep { m/create view/i } keys %$href;
$self->{tables}->{$db}->{$tbl} = [ 'view', $href->{$key} ];
}
}
return $self->{tables}->{$db}->{$tbl};
}
sub get_columns {
my ( $self, $dbh, $quoter, $db, $tbl ) = @_;
MKDEBUG && _d('Get columns for', $db, $tbl);
if ( !$self->{cache} || !$self->{columns}->{$db}->{$tbl} ) {
$self->_use_db($dbh, $quoter, $db);
my $sql = "SHOW COLUMNS FROM " . $quoter->quote($db, $tbl);
MKDEBUG && _d($sql);
my $cols = $dbh->selectall_arrayref($sql, { Slice => {} });
$self->{columns}->{$db}->{$tbl} = [
map {
my %row;
@row{ map { lc $_ } keys %$_ } = values %$_;
\%row;
} @$cols
];
}
return $self->{columns}->{$db}->{$tbl};
}
sub get_tmp_table {
my ( $self, $dbh, $quoter, $db, $tbl ) = @_;
my $result = 'CREATE TABLE ' . $quoter->quote($tbl) . " (\n";
$result .= join(",\n",
map { ' ' . $quoter->quote($_->{field}) . ' ' . $_->{type} }
@{$self->get_columns($dbh, $quoter, $db, $tbl)});
$result .= "\n)";
MKDEBUG && _d($result);
return $result;
}
sub get_triggers {
my ( $self, $dbh, $quoter, $db, $tbl ) = @_;
if ( !$self->{cache} || !$self->{triggers}->{$db} ) {
$self->{triggers}->{$db} = {};
my $sql = '/*!40101 SET @OLD_SQL_MODE := @@SQL_MODE, '
. q{@@SQL_MODE := REPLACE(REPLACE(@@SQL_MODE, 'ANSI_QUOTES', ''), ',,', ','), }
. '@OLD_QUOTE := @@SQL_QUOTE_SHOW_CREATE, '
. '@@SQL_QUOTE_SHOW_CREATE := 1 */';
MKDEBUG && _d($sql);
eval { $dbh->do($sql); };
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
$sql = "SHOW TRIGGERS FROM " . $quoter->quote($db);
MKDEBUG && _d($sql);
my $sth = $dbh->prepare($sql);
$sth->execute();
if ( $sth->rows ) {
my $trgs = $sth->fetchall_arrayref({});
foreach my $trg (@$trgs) {
my %trg;
@trg{ map { lc $_ } keys %$trg } = values %$trg;
push @{ $self->{triggers}->{$db}->{ $trg{table} } }, \%trg;
}
}
$sql = '/*!40101 SET @@SQL_MODE := @OLD_SQL_MODE, '
. '@@SQL_QUOTE_SHOW_CREATE := @OLD_QUOTE */';
MKDEBUG && _d($sql);
$dbh->do($sql);
}
if ( $tbl ) {
return $self->{triggers}->{$db}->{$tbl};
}
return values %{$self->{triggers}->{$db}};
}
sub get_databases {
my ( $self, $dbh, $quoter, $like ) = @_;
if ( !$self->{cache} || !$self->{databases} || $like ) {
my $sql = 'SHOW DATABASES';
my @params;
if ( $like ) {
$sql .= ' LIKE ?';
push @params, $like;
}
my $sth = $dbh->prepare($sql);
MKDEBUG && _d($sql, @params);
$sth->execute( @params );
my @dbs = map { $_->[0] } @{$sth->fetchall_arrayref()};
$self->{databases} = \@dbs unless $like;
return @dbs;
}
return @{$self->{databases}};
}
sub get_table_status {
my ( $self, $dbh, $quoter, $db, $like ) = @_;
if ( !$self->{cache} || !$self->{table_status}->{$db} || $like ) {
my $sql = "SHOW TABLE STATUS FROM " . $quoter->quote($db);
my @params;
if ( $like ) {
$sql .= ' LIKE ?';
push @params, $like;
}
MKDEBUG && _d($sql, @params);
my $sth = $dbh->prepare($sql);
$sth->execute(@params);
my @tables = @{$sth->fetchall_arrayref({})};
@tables = map {
my %tbl; # Make a copy with lowercased keys
@tbl{ map { lc $_ } keys %$_ } = values %$_;
$tbl{engine} ||= $tbl{type} || $tbl{comment};
delete $tbl{type};
\%tbl;
} @tables;
$self->{table_status}->{$db} = \@tables unless $like;
return @tables;
}
return @{$self->{table_status}->{$db}};
}
sub get_table_list {
my ( $self, $dbh, $quoter, $db, $like ) = @_;
if ( !$self->{cache} || !$self->{table_list}->{$db} || $like ) {
my $sql = "SHOW /*!50002 FULL*/ TABLES FROM " . $quoter->quote($db);
my @params;
if ( $like ) {
$sql .= ' LIKE ?';
push @params, $like;
}
MKDEBUG && _d($sql, @params);
my $sth = $dbh->prepare($sql);
$sth->execute(@params);
my @tables = @{$sth->fetchall_arrayref()};
@tables = map {
my %tbl = (
name => $_->[0],
engine => ($_->[1] || '') eq 'VIEW' ? 'VIEW' : '',
);
\%tbl;
} @tables;
$self->{table_list}->{$db} = \@tables unless $like;
return @tables;
}
return @{$self->{table_list}->{$db}};
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End MySQLDump package
# ###########################################################################
# ###########################################################################
# ChangeHandler package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/ChangeHandler.pm
# t/lib/ChangeHandler.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package ChangeHandler;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
my $DUPE_KEY = qr/Duplicate entry/;
our @ACTIONS = qw(DELETE REPLACE INSERT UPDATE);
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter left_db left_tbl right_db right_tbl
replace queue) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $q = $args{Quoter};
my $self = {
hex_blob => 1,
%args,
left_db_tbl => $q->quote(@args{qw(left_db left_tbl)}),
right_db_tbl => $q->quote(@args{qw(right_db right_tbl)}),
};
$self->{src_db_tbl} = $self->{left_db_tbl};
$self->{dst_db_tbl} = $self->{right_db_tbl};
map { $self->{$_} = [] } @ACTIONS;
$self->{changes} = { map { $_ => 0 } @ACTIONS };
return bless $self, $class;
}
sub fetch_back {
my ( $self, $dbh ) = @_;
$self->{fetch_back} = $dbh;
MKDEBUG && _d('Set fetch back dbh', $dbh);
return;
}
sub set_src {
my ( $self, $src, $dbh ) = @_;
die "I need a src argument" unless $src;
if ( lc $src eq 'left' ) {
$self->{src_db_tbl} = $self->{left_db_tbl};
$self->{dst_db_tbl} = $self->{right_db_tbl};
}
elsif ( lc $src eq 'right' ) {
$self->{src_db_tbl} = $self->{right_db_tbl};
$self->{dst_db_tbl} = $self->{left_db_tbl};
}
else {
die "src argument must be either 'left' or 'right'"
}
MKDEBUG && _d('Set src to', $src);
$self->fetch_back($dbh) if $dbh;
return;
}
sub src {
my ( $self ) = @_;
return $self->{src_db_tbl};
}
sub dst {
my ( $self ) = @_;
return $self->{dst_db_tbl};
}
sub _take_action {
my ( $self, $sql, $dbh ) = @_;
MKDEBUG && _d('Calling subroutines on', $dbh, $sql);
foreach my $action ( @{$self->{actions}} ) {
$action->($sql, $dbh);
}
return;
}
sub change {
my ( $self, $action, $row, $cols, $dbh ) = @_;
MKDEBUG && _d($dbh, $action, 'where', $self->make_where_clause($row, $cols));
return unless $action;
$self->{changes}->{
$self->{replace} && $action ne 'DELETE' ? 'REPLACE' : $action
}++;
if ( $self->{queue} ) {
$self->__queue($action, $row, $cols, $dbh);
}
else {
eval {
my $func = "make_$action";
$self->_take_action($self->$func($row, $cols), $dbh);
};
if ( $EVAL_ERROR =~ m/$DUPE_KEY/ ) {
MKDEBUG && _d('Duplicate key violation; will queue and rewrite');
$self->{queue}++;
$self->{replace} = 1;
$self->__queue($action, $row, $cols, $dbh);
}
elsif ( $EVAL_ERROR ) {
die $EVAL_ERROR;
}
}
return;
}
sub __queue {
my ( $self, $action, $row, $cols, $dbh ) = @_;
MKDEBUG && _d('Queueing change for later');
if ( $self->{replace} ) {
$action = $action eq 'DELETE' ? $action : 'REPLACE';
}
push @{$self->{$action}}, [ $row, $cols, $dbh ];
}
sub process_rows {
my ( $self, $queue_level, $trace_msg ) = @_;
my $error_count = 0;
TRY: {
if ( $queue_level && $queue_level < $self->{queue} ) { # see redo below!
MKDEBUG && _d('Not processing now', $queue_level, '<', $self->{queue});
return;
}
MKDEBUG && _d('Processing rows:');
my ($row, $cur_act);
eval {
foreach my $action ( @ACTIONS ) {
my $func = "make_$action";
my $rows = $self->{$action};
MKDEBUG && _d(scalar(@$rows), 'to', $action);
$cur_act = $action;
while ( @$rows ) {
$row = shift @$rows;
my $sql = $self->$func(@$row);
$sql .= " /*percona-toolkit $trace_msg*/" if $trace_msg;
$self->_take_action($sql, $row->[2]);
}
}
$error_count = 0;
};
if ( !$error_count++ && $EVAL_ERROR =~ m/$DUPE_KEY/ ) {
MKDEBUG && _d('Duplicate key violation; re-queueing and rewriting');
$self->{queue}++; # Defer rows to the very end
$self->{replace} = 1;
$self->__queue($cur_act, @$row);
redo TRY;
}
elsif ( $EVAL_ERROR ) {
die $EVAL_ERROR;
}
}
}
sub make_DELETE {
my ( $self, $row, $cols ) = @_;
MKDEBUG && _d('Make DELETE');
return "DELETE FROM $self->{dst_db_tbl} WHERE "
. $self->make_where_clause($row, $cols)
. ' LIMIT 1';
}
sub make_UPDATE {
my ( $self, $row, $cols ) = @_;
MKDEBUG && _d('Make UPDATE');
if ( $self->{replace} ) {
return $self->make_row('REPLACE', $row, $cols);
}
my %in_where = map { $_ => 1 } @$cols;
my $where = $self->make_where_clause($row, $cols);
my @cols;
if ( my $dbh = $self->{fetch_back} ) {
my $sql = $self->make_fetch_back_query($where);
MKDEBUG && _d('Fetching data on dbh', $dbh, 'for UPDATE:', $sql);
my $res = $dbh->selectrow_hashref($sql);
@{$row}{keys %$res} = values %$res;
@cols = $self->sort_cols($res);
}
else {
@cols = $self->sort_cols($row);
}
return "UPDATE $self->{dst_db_tbl} SET "
. join(', ', map {
$self->{Quoter}->quote($_)
. '=' . $self->{Quoter}->quote_val($row->{$_})
} grep { !$in_where{$_} } @cols)
. " WHERE $where LIMIT 1";
}
sub make_INSERT {
my ( $self, $row, $cols ) = @_;
MKDEBUG && _d('Make INSERT');
if ( $self->{replace} ) {
return $self->make_row('REPLACE', $row, $cols);
}
return $self->make_row('INSERT', $row, $cols);
}
sub make_REPLACE {
my ( $self, $row, $cols ) = @_;
MKDEBUG && _d('Make REPLACE');
return $self->make_row('REPLACE', $row, $cols);
}
sub make_row {
my ( $self, $verb, $row, $cols ) = @_;
my @cols;
if ( my $dbh = $self->{fetch_back} ) {
my $where = $self->make_where_clause($row, $cols);
my $sql = $self->make_fetch_back_query($where);
MKDEBUG && _d('Fetching data on dbh', $dbh, 'for', $verb, ':', $sql);
my $res = $dbh->selectrow_hashref($sql);
@{$row}{keys %$res} = values %$res;
@cols = $self->sort_cols($res);
}
else {
@cols = $self->sort_cols($row);
}
my $q = $self->{Quoter};
return "$verb INTO $self->{dst_db_tbl}("
. join(', ', map { $q->quote($_) } @cols)
. ') VALUES ('
. join(', ', map { $q->quote_val($_) } @{$row}{@cols} )
. ')';
}
sub make_where_clause {
my ( $self, $row, $cols ) = @_;
my @clauses = map {
my $val = $row->{$_};
my $sep = defined $val ? '=' : ' IS ';
$self->{Quoter}->quote($_) . $sep . $self->{Quoter}->quote_val($val);
} @$cols;
return join(' AND ', @clauses);
}
sub get_changes {
my ( $self ) = @_;
return %{$self->{changes}};
}
sub sort_cols {
my ( $self, $row ) = @_;
my @cols;
if ( $self->{tbl_struct} ) {
my $pos = $self->{tbl_struct}->{col_posn};
my @not_in_tbl;
@cols = sort {
$pos->{$a} <=> $pos->{$b}
}
grep {
if ( !defined $pos->{$_} ) {
push @not_in_tbl, $_;
0;
}
else {
1;
}
}
keys %$row;
push @cols, @not_in_tbl if @not_in_tbl;
}
else {
@cols = sort keys %$row;
}
return @cols;
}
sub make_fetch_back_query {
my ( $self, $where ) = @_;
die "I need a where argument" unless $where;
my $cols = '*';
my $tbl_struct = $self->{tbl_struct};
if ( $tbl_struct ) {
$cols = join(', ',
map {
my $col = $_;
if ( $self->{hex_blob}
&& $tbl_struct->{type_for}->{$col} =~ m/blob|text|binary/ ) {
$col = "IF(`$col`='', '', CONCAT('0x', HEX(`$col`))) AS `$col`";
}
else {
$col = "`$col`";
}
$col;
} @{ $tbl_struct->{cols} }
);
if ( !$cols ) {
MKDEBUG && _d('Failed to make explicit columns list from tbl struct');
$cols = '*';
}
}
return "SELECT $cols FROM $self->{src_db_tbl} WHERE $where LIMIT 1";
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End ChangeHandler package
# ###########################################################################
# ###########################################################################
# TableChunker package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableChunker.pm
# t/lib/TableChunker.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableChunker;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use POSIX qw(floor ceil);
use List::Util qw(min max);
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter MySQLDump) ) {
die "I need a $arg argument" unless $args{$arg};
}
my %int_types = map { $_ => 1 } qw(bigint date datetime int mediumint smallint time timestamp tinyint year);
my %real_types = map { $_ => 1 } qw(decimal double float);
my $self = {
%args,
int_types => \%int_types,
real_types => \%real_types,
EPOCH => '1970-01-01',
};
return bless $self, $class;
}
sub find_chunk_columns {
my ( $self, %args ) = @_;
foreach my $arg ( qw(tbl_struct) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $tbl_struct = $args{tbl_struct};
my @possible_indexes;
foreach my $index ( values %{ $tbl_struct->{keys} } ) {
next unless $index->{type} eq 'BTREE';
next if grep { defined } @{$index->{col_prefixes}};
if ( $args{exact} ) {
next unless $index->{is_unique} && @{$index->{cols}} == 1;
}
push @possible_indexes, $index;
}
MKDEBUG && _d('Possible chunk indexes in order:',
join(', ', map { $_->{name} } @possible_indexes));
my $can_chunk_exact = 0;
my @candidate_cols;
foreach my $index ( @possible_indexes ) {
my $col = $index->{cols}->[0];
my $col_type = $tbl_struct->{type_for}->{$col};
next unless $self->{int_types}->{$col_type}
|| $self->{real_types}->{$col_type}
|| $col_type =~ m/char/;
push @candidate_cols, { column => $col, index => $index->{name} };
}
$can_chunk_exact = 1 if $args{exact} && scalar @candidate_cols;
if ( MKDEBUG ) {
my $chunk_type = $args{exact} ? 'Exact' : 'Inexact';
_d($chunk_type, 'chunkable:',
join(', ', map { "$_->{column} on $_->{index}" } @candidate_cols));
}
my @result;
MKDEBUG && _d('Ordering columns by order in tbl, PK first');
if ( $tbl_struct->{keys}->{PRIMARY} ) {
my $pk_first_col = $tbl_struct->{keys}->{PRIMARY}->{cols}->[0];
@result = grep { $_->{column} eq $pk_first_col } @candidate_cols;
@candidate_cols = grep { $_->{column} ne $pk_first_col } @candidate_cols;
}
my $i = 0;
my %col_pos = map { $_ => $i++ } @{$tbl_struct->{cols}};
push @result, sort { $col_pos{$a->{column}} <=> $col_pos{$b->{column}} }
@candidate_cols;
if ( MKDEBUG ) {
_d('Chunkable columns:',
join(', ', map { "$_->{column} on $_->{index}" } @result));
_d('Can chunk exactly:', $can_chunk_exact);
}
return ($can_chunk_exact, @result);
}
sub calculate_chunks {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl tbl_struct chunk_col rows_in_range chunk_size);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
MKDEBUG && _d('Calculate chunks for',
join(", ", map {"$_=".(defined $args{$_} ? $args{$_} : "undef")}
qw(db tbl chunk_col min max rows_in_range chunk_size zero_chunk exact)
));
if ( !$args{rows_in_range} ) {
MKDEBUG && _d("Empty table");
return '1=1';
}
if ( $args{rows_in_range} < $args{chunk_size} ) {
MKDEBUG && _d("Chunk size larger than rows in range");
return '1=1';
}
my $q = $self->{Quoter};
my $dbh = $args{dbh};
my $chunk_col = $args{chunk_col};
my $tbl_struct = $args{tbl_struct};
my $col_type = $tbl_struct->{type_for}->{$chunk_col};
MKDEBUG && _d('chunk col type:', $col_type);
my %chunker;
if ( $tbl_struct->{is_numeric}->{$chunk_col} || $col_type =~ /date|time/ ) {
%chunker = $self->_chunk_numeric(%args);
}
elsif ( $col_type =~ m/char/ ) {
%chunker = $self->_chunk_char(%args);
}
else {
die "Cannot chunk $col_type columns";
}
MKDEBUG && _d("Chunker:", Dumper(\%chunker));
my ($col, $start_point, $end_point, $interval, $range_func)
= @chunker{qw(col start_point end_point interval range_func)};
my @chunks;
if ( $start_point < $end_point ) {
push @chunks, "$col = 0" if $chunker{have_zero_chunk};
my ($beg, $end);
my $iter = 0;
for ( my $i = $start_point; $i < $end_point; $i += $interval ) {
($beg, $end) = $self->$range_func($dbh, $i, $interval, $end_point);
if ( $iter++ == 0 ) {
push @chunks,
($chunker{have_zero_chunk} ? "$col > 0 AND " : "")
."$col < " . $q->quote_val($end);
}
else {
push @chunks, "$col >= " . $q->quote_val($beg) . " AND $col < " . $q->quote_val($end);
}
}
# Alexander Garzon [email protected] 2011-11-17
#Bug : http://code.google.com/p/maatkit/issues/detail?id=1296#c2
#my $chunk_range = lc $args{chunk_range} || 'open';
my $chunk_range = 'open';
my $nullable = $args{tbl_struct}->{is_nullable}->{$args{chunk_col}};
pop @chunks;
if ( @chunks ) {
push @chunks, "$col >= " . $q->quote_val($beg)
. ($chunk_range eq 'openclosed'
? " AND $col <= " . $q->quote_val($args{max}) : "");
}
else {
push @chunks, $nullable ? "$col IS NOT NULL" : '1=1';
}
if ( $nullable ) {
push @chunks, "$col IS NULL";
}
}
else {
MKDEBUG && _d('No chunks; using single chunk 1=1');
push @chunks, '1=1';
}
return @chunks;
}
sub _chunk_numeric {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl tbl_struct chunk_col rows_in_range chunk_size);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $q = $self->{Quoter};
my $db_tbl = $q->quote($args{db}, $args{tbl});
my $col_type = $args{tbl_struct}->{type_for}->{$args{chunk_col}};
my $range_func;
if ( $col_type =~ m/(?:int|year|float|double|decimal)$/ ) {
$range_func = 'range_num';
}
elsif ( $col_type =~ m/^(?:timestamp|date|time)$/ ) {
$range_func = "range_$col_type";
}
elsif ( $col_type eq 'datetime' ) {
$range_func = 'range_datetime';
}
my ($start_point, $end_point);
eval {
$start_point = $self->value_to_number(
value => $args{min},
column_type => $col_type,
dbh => $args{dbh},
);
$end_point = $self->value_to_number(
value => $args{max},
column_type => $col_type,
dbh => $args{dbh},
);
};
if ( $EVAL_ERROR ) {
if ( $EVAL_ERROR =~ m/don't know how to chunk/ ) {
die $EVAL_ERROR;
}
else {
die "Error calculating chunk start and end points for table "
. "`$args{tbl_struct}->{name}` on column `$args{chunk_col}` "
. "with min/max values "
. join('/',
map { defined $args{$_} ? $args{$_} : 'undef' } qw(min max))
. ":\n\n"
. $EVAL_ERROR
. "\nVerify that the min and max values are valid for the column. "
. "If they are valid, this error could be caused by a bug in the "
. "tool.";
}
}
if ( !defined $start_point ) {
MKDEBUG && _d('Start point is undefined');
$start_point = 0;
}
if ( !defined $end_point || $end_point < $start_point ) {
MKDEBUG && _d('End point is undefined or before start point');
$end_point = 0;
}
MKDEBUG && _d("Actual chunk range:", $start_point, "to", $end_point);
my $have_zero_chunk = 0;
if ( $args{zero_chunk} ) {
if ( $start_point != $end_point && $start_point >= 0 ) {
MKDEBUG && _d('Zero chunking');
my $nonzero_val = $self->get_nonzero_value(
%args,
db_tbl => $db_tbl,
col => $args{chunk_col},
col_type => $col_type,
val => $args{min}
);
$start_point = $self->value_to_number(
value => $nonzero_val,
column_type => $col_type,
dbh => $args{dbh},
);
$have_zero_chunk = 1;
}
else {
MKDEBUG && _d("Cannot zero chunk");
}
}
MKDEBUG && _d("Using chunk range:", $start_point, "to", $end_point);
my $interval = $args{chunk_size}
* ($end_point - $start_point)
/ $args{rows_in_range};
if ( $self->{int_types}->{$col_type} ) {
$interval = ceil($interval);
}
$interval ||= $args{chunk_size};
if ( $args{exact} ) {
$interval = $args{chunk_size};
}
MKDEBUG && _d('Chunk interval:', $interval, 'units');
return (
col => $q->quote($args{chunk_col}),
start_point => $start_point,
end_point => $end_point,
interval => $interval,
range_func => $range_func,
have_zero_chunk => $have_zero_chunk,
);
}
sub _chunk_char {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl tbl_struct chunk_col rows_in_range chunk_size);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $q = $self->{Quoter};
my $db_tbl = $q->quote($args{db}, $args{tbl});
my $dbh = $args{dbh};
my $chunk_col = $args{chunk_col};
my $row;
my $sql;
$sql = "SELECT MIN($chunk_col), MAX($chunk_col) FROM $db_tbl "
. "ORDER BY `$chunk_col`";
MKDEBUG && _d($dbh, $sql);
$row = $dbh->selectrow_arrayref($sql);
my ($min_col, $max_col) = ($row->[0], $row->[1]);
$sql = "SELECT ORD(?) AS min_col_ord, ORD(?) AS max_col_ord";
MKDEBUG && _d($dbh, $sql);
my $ord_sth = $dbh->prepare($sql); # avoid quoting issues
$ord_sth->execute($min_col, $max_col);
$row = $ord_sth->fetchrow_arrayref();
my ($min_col_ord, $max_col_ord) = ($row->[0], $row->[1]);
MKDEBUG && _d("Min/max col char code:", $min_col_ord, $max_col_ord);
my $base;
my @chars;
MKDEBUG && _d("Table charset:", $args{tbl_struct}->{charset});
if ( ($args{tbl_struct}->{charset} || "") eq "latin1" ) {
my @sorted_latin1_chars = (
32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45,
46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59,
60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73,
74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87,
88, 89, 90, 91, 92, 93, 94, 95, 96, 123, 124, 125, 126, 161,
162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175,
176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189,
190, 191, 215, 216, 222, 223, 247, 255);
my ($first_char, $last_char);
for my $i ( 0..$#sorted_latin1_chars ) {
$first_char = $i and last if $sorted_latin1_chars[$i] >= $min_col_ord;
}
for my $i ( $first_char..$#sorted_latin1_chars ) {
$last_char = $i and last if $sorted_latin1_chars[$i] >= $max_col_ord;
};
@chars = map { chr $_; } @sorted_latin1_chars[$first_char..$last_char];
$base = scalar @chars;
}
else {
my $tmp_tbl = '__maatkit_char_chunking_map';
my $tmp_db_tbl = $q->quote($args{db}, $tmp_tbl);
$sql = "DROP TABLE IF EXISTS $tmp_db_tbl";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
my $col_def = $args{tbl_struct}->{defs}->{$chunk_col};
$sql = "CREATE TEMPORARY TABLE $tmp_db_tbl ($col_def) "
. "ENGINE=MEMORY";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
$sql = "INSERT INTO $tmp_db_tbl VALUE (CHAR(?))";
MKDEBUG && _d($dbh, $sql);
my $ins_char_sth = $dbh->prepare($sql); # avoid quoting issues
for my $char_code ( $min_col_ord..$max_col_ord ) {
$ins_char_sth->execute($char_code);
}
$sql = "SELECT `$chunk_col` FROM $tmp_db_tbl "
. "WHERE `$chunk_col` BETWEEN ? AND ? "
. "ORDER BY `$chunk_col`";
MKDEBUG && _d($dbh, $sql);
my $sel_char_sth = $dbh->prepare($sql);
$sel_char_sth->execute($min_col, $max_col);
@chars = map { $_->[0] } @{ $sel_char_sth->fetchall_arrayref() };
$base = scalar @chars;
$sql = "DROP TABLE $tmp_db_tbl";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
MKDEBUG && _d("Base", $base, "chars:", @chars);
$sql = "SELECT MAX(LENGTH($chunk_col)) FROM $db_tbl ORDER BY `$chunk_col`";
MKDEBUG && _d($dbh, $sql);
$row = $dbh->selectrow_arrayref($sql);
my $max_col_len = $row->[0];
MKDEBUG && _d("Max column value:", $max_col, $max_col_len);
my $n_values;
for my $n_chars ( 1..$max_col_len ) {
$n_values = $base**$n_chars;
if ( $n_values >= $args{chunk_size} ) {
MKDEBUG && _d($n_chars, "chars in base", $base, "expresses",
$n_values, "values");
last;
}
}
my $n_chunks = $args{rows_in_range} / $args{chunk_size};
my $interval = floor($n_values / $n_chunks) || 1;
my $range_func = sub {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $start_char = $self->base_count(
count_to => $start,
base => $base,
symbols => \@chars,
);
my $end_char = $self->base_count(
count_to => min($max, $start + $interval),
base => $base,
symbols => \@chars,
);
return $start_char, $end_char;
};
return (
col => $q->quote($chunk_col),
start_point => 0,
end_point => $n_values,
interval => $interval,
range_func => $range_func,
);
}
sub get_first_chunkable_column {
my ( $self, %args ) = @_;
foreach my $arg ( qw(tbl_struct) ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($exact, @cols) = $self->find_chunk_columns(%args);
my $col = $cols[0]->{column};
my $idx = $cols[0]->{index};
my $wanted_col = $args{chunk_column};
my $wanted_idx = $args{chunk_index};
MKDEBUG && _d("Preferred chunk col/idx:", $wanted_col, $wanted_idx);
if ( $wanted_col && $wanted_idx ) {
foreach my $chunkable_col ( @cols ) {
if ( $wanted_col eq $chunkable_col->{column}
&& $wanted_idx eq $chunkable_col->{index} ) {
$col = $wanted_col;
$idx = $wanted_idx;
last;
}
}
}
elsif ( $wanted_col ) {
foreach my $chunkable_col ( @cols ) {
if ( $wanted_col eq $chunkable_col->{column} ) {
$col = $wanted_col;
$idx = $chunkable_col->{index};
last;
}
}
}
elsif ( $wanted_idx ) {
foreach my $chunkable_col ( @cols ) {
if ( $wanted_idx eq $chunkable_col->{index} ) {
$col = $chunkable_col->{column};
$idx = $wanted_idx;
last;
}
}
}
MKDEBUG && _d('First chunkable col/index:', $col, $idx);
return $col, $idx;
}
sub size_to_rows {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl chunk_size);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db, $tbl, $chunk_size) = @args{@required_args};
my $q = $self->{Quoter};
my $du = $self->{MySQLDump};
my ($n_rows, $avg_row_length);
my ( $num, $suffix ) = $chunk_size =~ m/^(\d+)([MGk])?$/;
if ( $suffix ) { # Convert to bytes.
$chunk_size = $suffix eq 'k' ? $num * 1_024
: $suffix eq 'M' ? $num * 1_024 * 1_024
: $num * 1_024 * 1_024 * 1_024;
}
elsif ( $num ) {
$n_rows = $num;
}
else {
die "Invalid chunk size $chunk_size; must be an integer "
. "with optional suffix kMG";
}
if ( $suffix || $args{avg_row_length} ) {
my ($status) = $du->get_table_status($dbh, $q, $db, $tbl);
$avg_row_length = $status->{avg_row_length};
if ( !defined $n_rows ) {
$n_rows = $avg_row_length ? ceil($chunk_size / $avg_row_length) : undef;
}
}
return $n_rows, $avg_row_length;
}
sub get_range_statistics {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl chunk_col tbl_struct);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db, $tbl, $col) = @args{@required_args};
my $where = $args{where};
my $q = $self->{Quoter};
my $col_type = $args{tbl_struct}->{type_for}->{$col};
my $col_is_numeric = $args{tbl_struct}->{is_numeric}->{$col};
my $db_tbl = $q->quote($db, $tbl);
$col = $q->quote($col);
my ($min, $max);
eval {
my $sql = "SELECT MIN($col), MAX($col) FROM $db_tbl"
. ($args{index_hint} ? " $args{index_hint}" : "")
. ($where ? " WHERE ($where)" : '');
MKDEBUG && _d($dbh, $sql);
($min, $max) = $dbh->selectrow_array($sql);
MKDEBUG && _d("Actual end points:", $min, $max);
($min, $max) = $self->get_valid_end_points(
%args,
dbh => $dbh,
db_tbl => $db_tbl,
col => $col,
col_type => $col_type,
min => $min,
max => $max,
);
MKDEBUG && _d("Valid end points:", $min, $max);
};
if ( $EVAL_ERROR ) {
die "Error getting min and max values for table $db_tbl "
. "on column $col: $EVAL_ERROR";
}
my $sql = "EXPLAIN SELECT * FROM $db_tbl"
. ($args{index_hint} ? " $args{index_hint}" : "")
. ($where ? " WHERE $where" : '');
MKDEBUG && _d($sql);
my $expl = $dbh->selectrow_hashref($sql);
return (
min => $min,
max => $max,
rows_in_range => $expl->{rows},
);
}
sub inject_chunks {
my ( $self, %args ) = @_;
foreach my $arg ( qw(database table chunks chunk_num query) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
MKDEBUG && _d('Injecting chunk', $args{chunk_num});
my $query = $args{query};
my $comment = sprintf("/*%s.%s:%d/%d*/",
$args{database}, $args{table},
$args{chunk_num} + 1, scalar @{$args{chunks}});
$query =~ s!/\*PROGRESS_COMMENT\*/!$comment!;
my $where = "WHERE (" . $args{chunks}->[$args{chunk_num}] . ')';
if ( $args{where} && grep { $_ } @{$args{where}} ) {
$where .= " AND ("
. join(" AND ", map { "($_)" } grep { $_ } @{$args{where}} )
. ")";
}
my $db_tbl = $self->{Quoter}->quote(@args{qw(database table)});
my $index_hint = $args{index_hint} || '';
MKDEBUG && _d('Parameters:',
Dumper({WHERE => $where, DB_TBL => $db_tbl, INDEX_HINT => $index_hint}));
$query =~ s!/\*WHERE\*/! $where!;
$query =~ s!/\*DB_TBL\*/!$db_tbl!;
$query =~ s!/\*INDEX_HINT\*/! $index_hint!;
$query =~ s!/\*CHUNK_NUM\*/! $args{chunk_num} AS chunk_num,!;
return $query;
}
sub value_to_number {
my ( $self, %args ) = @_;
my @required_args = qw(column_type dbh);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $val = $args{value};
my ($col_type, $dbh) = @args{@required_args};
MKDEBUG && _d('Converting MySQL', $col_type, $val);
return unless defined $val; # value is NULL
my %mysql_conv_func_for = (
timestamp => 'UNIX_TIMESTAMP',
date => 'TO_DAYS',
time => 'TIME_TO_SEC',
datetime => 'TO_DAYS',
);
my $num;
if ( $col_type =~ m/(?:int|year|float|double|decimal)$/ ) {
$num = $val;
}
elsif ( $col_type =~ m/^(?:timestamp|date|time)$/ ) {
my $func = $mysql_conv_func_for{$col_type};
my $sql = "SELECT $func(?)";
MKDEBUG && _d($dbh, $sql, $val);
my $sth = $dbh->prepare($sql);
$sth->execute($val);
($num) = $sth->fetchrow_array();
}
elsif ( $col_type eq 'datetime' ) {
$num = $self->timestampdiff($dbh, $val);
}
else {
die "I don't know how to chunk $col_type\n";
}
MKDEBUG && _d('Converts to', $num);
return $num;
}
sub range_num {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $end = min($max, $start + $interval);
$start = sprintf('%.17f', $start) if $start =~ /e/;
$end = sprintf('%.17f', $end) if $end =~ /e/;
$start =~ s/\.(\d{5}).*$/.$1/;
$end =~ s/\.(\d{5}).*$/.$1/;
if ( $end > $start ) {
return ( $start, $end );
}
else {
die "Chunk size is too small: $end !> $start\n";
}
}
sub range_time {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $sql = "SELECT SEC_TO_TIME($start), SEC_TO_TIME(LEAST($max, $start + $interval))";
MKDEBUG && _d($sql);
return $dbh->selectrow_array($sql);
}
sub range_date {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $sql = "SELECT FROM_DAYS($start), FROM_DAYS(LEAST($max, $start + $interval))";
MKDEBUG && _d($sql);
return $dbh->selectrow_array($sql);
}
sub range_datetime {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $sql = "SELECT DATE_ADD('$self->{EPOCH}', INTERVAL $start SECOND), "
. "DATE_ADD('$self->{EPOCH}', INTERVAL LEAST($max, $start + $interval) SECOND)";
MKDEBUG && _d($sql);
return $dbh->selectrow_array($sql);
}
sub range_timestamp {
my ( $self, $dbh, $start, $interval, $max ) = @_;
my $sql = "SELECT FROM_UNIXTIME($start), FROM_UNIXTIME(LEAST($max, $start + $interval))";
MKDEBUG && _d($sql);
return $dbh->selectrow_array($sql);
}
sub timestampdiff {
my ( $self, $dbh, $time ) = @_;
my $sql = "SELECT (COALESCE(TO_DAYS('$time'), 0) * 86400 + TIME_TO_SEC('$time')) "
. "- TO_DAYS('$self->{EPOCH} 00:00:00') * 86400";
MKDEBUG && _d($sql);
my ( $diff ) = $dbh->selectrow_array($sql);
$sql = "SELECT DATE_ADD('$self->{EPOCH}', INTERVAL $diff SECOND)";
MKDEBUG && _d($sql);
my ( $check ) = $dbh->selectrow_array($sql);
die <<" EOF"
Incorrect datetime math: given $time, calculated $diff but checked to $check.
This could be due to a version of MySQL that overflows on large interval
values to DATE_ADD(), or the given datetime is not a valid date. If not,
please report this as a bug.
EOF
unless $check eq $time;
return $diff;
}
sub get_valid_end_points {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db_tbl col col_type);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db_tbl, $col, $col_type) = @args{@required_args};
my ($real_min, $real_max) = @args{qw(min max)};
my $err_fmt = "Error finding a valid %s value for table $db_tbl on "
. "column $col. The real %s value %s is invalid and "
. "no other valid values were found. Verify that the table "
. "has at least one valid value for this column"
. ($args{where} ? " where $args{where}." : ".");
my $valid_min = $real_min;
if ( defined $valid_min ) {
MKDEBUG && _d("Validating min end point:", $real_min);
$valid_min = $self->_get_valid_end_point(
%args,
val => $real_min,
endpoint => 'min',
);
die sprintf($err_fmt, 'minimum', 'minimum',
(defined $real_min ? $real_min : "NULL"))
unless defined $valid_min;
}
my $valid_max = $real_max;
if ( defined $valid_max ) {
MKDEBUG && _d("Validating max end point:", $real_min);
$valid_max = $self->_get_valid_end_point(
%args,
val => $real_max,
endpoint => 'max',
);
die sprintf($err_fmt, 'maximum', 'maximum',
(defined $real_max ? $real_max : "NULL"))
unless defined $valid_max;
}
return $valid_min, $valid_max;
}
sub _get_valid_end_point {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db_tbl col col_type);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db_tbl, $col, $col_type) = @args{@required_args};
my $val = $args{val};
return $val unless defined $val;
my $validate = $col_type =~ m/time|date/ ? \&_validate_temporal_value
: undef;
if ( !$validate ) {
MKDEBUG && _d("No validator for", $col_type, "values");
return $val;
}
return $val if defined $validate->($dbh, $val);
MKDEBUG && _d("Value is invalid, getting first valid value");
$val = $self->get_first_valid_value(
%args,
val => $val,
validate => $validate,
);
return $val;
}
sub get_first_valid_value {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db_tbl col validate endpoint);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db_tbl, $col, $validate, $endpoint) = @args{@required_args};
my $tries = defined $args{tries} ? $args{tries} : 5;
my $val = $args{val};
return unless defined $val;
my $cmp = $endpoint =~ m/min/i ? '>'
: $endpoint =~ m/max/i ? '<'
: die "Invalid endpoint arg: $endpoint";
my $sql = "SELECT $col FROM $db_tbl "
. ($args{index_hint} ? "$args{index_hint} " : "")
. "WHERE $col $cmp ? AND $col IS NOT NULL "
. ($args{where} ? "AND ($args{where}) " : "")
. "ORDER BY $col LIMIT 1";
MKDEBUG && _d($dbh, $sql);
my $sth = $dbh->prepare($sql);
my $last_val = $val;
while ( $tries-- ) {
$sth->execute($last_val);
my ($next_val) = $sth->fetchrow_array();
MKDEBUG && _d('Next value:', $next_val, '; tries left:', $tries);
if ( !defined $next_val ) {
MKDEBUG && _d('No more rows in table');
last;
}
if ( defined $validate->($dbh, $next_val) ) {
MKDEBUG && _d('First valid value:', $next_val);
$sth->finish();
return $next_val;
}
$last_val = $next_val;
}
$sth->finish();
$val = undef; # no valid value found
return $val;
}
sub _validate_temporal_value {
my ( $dbh, $val ) = @_;
my $sql = "SELECT IF(TIME_FORMAT(?,'%H:%i:%s')=?, TIME_TO_SEC(?), TO_DAYS(?))";
my $res;
eval {
MKDEBUG && _d($dbh, $sql, $val);
my $sth = $dbh->prepare($sql);
$sth->execute($val, $val, $val, $val);
($res) = $sth->fetchrow_array();
$sth->finish();
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d($EVAL_ERROR);
}
return $res;
}
sub get_nonzero_value {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db_tbl col col_type);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db_tbl, $col, $col_type) = @args{@required_args};
my $tries = defined $args{tries} ? $args{tries} : 5;
my $val = $args{val};
my $is_nonzero = $col_type =~ m/time|date/ ? \&_validate_temporal_value
: sub { return $_[1]; };
if ( !$is_nonzero->($dbh, $val) ) { # quasi-double-negative, sorry
MKDEBUG && _d('Discarding zero value:', $val);
my $sql = "SELECT $col FROM $db_tbl "
. ($args{index_hint} ? "$args{index_hint} " : "")
. "WHERE $col > ? AND $col IS NOT NULL "
. ($args{where} ? "AND ($args{where}) " : '')
. "ORDER BY $col LIMIT 1";
MKDEBUG && _d($sql);
my $sth = $dbh->prepare($sql);
my $last_val = $val;
while ( $tries-- ) {
$sth->execute($last_val);
my ($next_val) = $sth->fetchrow_array();
if ( $is_nonzero->($dbh, $next_val) ) {
MKDEBUG && _d('First non-zero value:', $next_val);
$sth->finish();
return $next_val;
}
$last_val = $next_val;
}
$sth->finish();
$val = undef; # no non-zero value found
}
return $val;
}
sub base_count {
my ( $self, %args ) = @_;
my @required_args = qw(count_to base symbols);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my ($n, $base, $symbols) = @args{@required_args};
return $symbols->[0] if $n == 0;
my $highest_power = floor(log($n)/log($base));
if ( $highest_power == 0 ){
return $symbols->[$n];
}
my @base_powers;
for my $power ( 0..$highest_power ) {
push @base_powers, ($base**$power) || 1;
}
my @base_multiples;
foreach my $base_power ( reverse @base_powers ) {
my $multiples = floor($n / $base_power);
push @base_multiples, $multiples;
$n -= $multiples * $base_power;
}
return join('', map { $symbols->[$_] } @base_multiples);
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableChunker package
# ###########################################################################
# ###########################################################################
# TableChecksum package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableChecksum.pm
# t/lib/TableChecksum.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableChecksum;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use List::Util qw(max);
our %ALGOS = (
CHECKSUM => { pref => 0, hash => 0 },
BIT_XOR => { pref => 2, hash => 1 },
ACCUM => { pref => 3, hash => 1 },
);
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter VersionParser) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub crc32 {
my ( $self, $string ) = @_;
my $poly = 0xEDB88320;
my $crc = 0xFFFFFFFF;
foreach my $char ( split(//, $string) ) {
my $comp = ($crc ^ ord($char)) & 0xFF;
for ( 1 .. 8 ) {
$comp = $comp & 1 ? $poly ^ ($comp >> 1) : $comp >> 1;
}
$crc = (($crc >> 8) & 0x00FFFFFF) ^ $comp;
}
return $crc ^ 0xFFFFFFFF;
}
sub get_crc_wid {
my ( $self, $dbh, $func ) = @_;
my $crc_wid = 16;
if ( uc $func ne 'FNV_64' && uc $func ne 'FNV1A_64' ) {
eval {
my ($val) = $dbh->selectrow_array("SELECT $func('a')");
$crc_wid = max(16, length($val));
};
}
return $crc_wid;
}
sub get_crc_type {
my ( $self, $dbh, $func ) = @_;
my $type = '';
my $length = 0;
my $sql = "SELECT $func('a')";
my $sth = $dbh->prepare($sql);
eval {
$sth->execute();
$type = $sth->{mysql_type_name}->[0];
$length = $sth->{mysql_length}->[0];
MKDEBUG && _d($sql, $type, $length);
if ( $type eq 'bigint' && $length < 20 ) {
$type = 'int';
}
};
$sth->finish;
MKDEBUG && _d('crc_type:', $type, 'length:', $length);
return ($type, $length);
}
sub best_algorithm {
my ( $self, %args ) = @_;
my ( $alg, $dbh ) = @args{ qw(algorithm dbh) };
my $vp = $self->{VersionParser};
my @choices = sort { $ALGOS{$a}->{pref} <=> $ALGOS{$b}->{pref} } keys %ALGOS;
die "Invalid checksum algorithm $alg"
if $alg && !$ALGOS{$alg};
if (
$args{where} || $args{chunk} # CHECKSUM does whole table
|| $args{replicate} # CHECKSUM can't do INSERT.. SELECT
|| !$vp->version_ge($dbh, '4.1.1')) # CHECKSUM doesn't exist
{
MKDEBUG && _d('Cannot use CHECKSUM algorithm');
@choices = grep { $_ ne 'CHECKSUM' } @choices;
}
if ( !$vp->version_ge($dbh, '4.1.1') ) {
MKDEBUG && _d('Cannot use BIT_XOR algorithm because MySQL < 4.1.1');
@choices = grep { $_ ne 'BIT_XOR' } @choices;
}
if ( $alg && grep { $_ eq $alg } @choices ) {
MKDEBUG && _d('User requested', $alg, 'algorithm');
return $alg;
}
if ( $args{count} && grep { $_ ne 'CHECKSUM' } @choices ) {
MKDEBUG && _d('Not using CHECKSUM algorithm because COUNT desired');
@choices = grep { $_ ne 'CHECKSUM' } @choices;
}
MKDEBUG && _d('Algorithms, in order:', @choices);
return $choices[0];
}
sub is_hash_algorithm {
my ( $self, $algorithm ) = @_;
return $ALGOS{$algorithm} && $ALGOS{$algorithm}->{hash};
}
sub choose_hash_func {
my ( $self, %args ) = @_;
my @funcs = qw(CRC32 FNV1A_64 FNV_64 MD5 SHA1);
if ( $args{function} ) {
unshift @funcs, $args{function};
}
my ($result, $error);
do {
my $func;
eval {
$func = shift(@funcs);
my $sql = "SELECT $func('test-string')";
MKDEBUG && _d($sql);
$args{dbh}->do($sql);
$result = $func;
};
if ( $EVAL_ERROR && $EVAL_ERROR =~ m/failed: (.*?) at \S+ line/ ) {
$error .= qq{$func cannot be used because "$1"\n};
MKDEBUG && _d($func, 'cannot be used because', $1);
}
} while ( @funcs && !$result );
die $error unless $result;
MKDEBUG && _d('Chosen hash func:', $result);
return $result;
}
sub optimize_xor {
my ( $self, %args ) = @_;
my ($dbh, $func) = @args{qw(dbh function)};
die "$func never needs the BIT_XOR optimization"
if $func =~ m/^(?:FNV1A_64|FNV_64|CRC32)$/i;
my $opt_slice = 0;
my $unsliced = uc $dbh->selectall_arrayref("SELECT $func('a')")->[0]->[0];
my $sliced = '';
my $start = 1;
my $crc_wid = length($unsliced) < 16 ? 16 : length($unsliced);
do { # Try different positions till sliced result equals non-sliced.
MKDEBUG && _d('Trying slice', $opt_slice);
$dbh->do('SET @crc := "", @cnt := 0');
my $slices = $self->make_xor_slices(
query => "\@crc := $func('a')",
crc_wid => $crc_wid,
opt_slice => $opt_slice,
);
my $sql = "SELECT CONCAT($slices) AS TEST FROM (SELECT NULL) AS x";
$sliced = ($dbh->selectrow_array($sql))[0];
if ( $sliced ne $unsliced ) {
MKDEBUG && _d('Slice', $opt_slice, 'does not work');
$start += 16;
++$opt_slice;
}
} while ( $start < $crc_wid && $sliced ne $unsliced );
if ( $sliced eq $unsliced ) {
MKDEBUG && _d('Slice', $opt_slice, 'works');
return $opt_slice;
}
else {
MKDEBUG && _d('No slice works');
return undef;
}
}
sub make_xor_slices {
my ( $self, %args ) = @_;
foreach my $arg ( qw(query crc_wid) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my ( $query, $crc_wid, $opt_slice ) = @args{qw(query crc_wid opt_slice)};
my @slices;
for ( my $start = 1; $start <= $crc_wid; $start += 16 ) {
my $len = $crc_wid - $start + 1;
if ( $len > 16 ) {
$len = 16;
}
push @slices,
"LPAD(CONV(BIT_XOR("
. "CAST(CONV(SUBSTRING(\@crc, $start, $len), 16, 10) AS UNSIGNED))"
. ", 10, 16), $len, '0')";
}
if ( defined $opt_slice && $opt_slice < @slices ) {
$slices[$opt_slice] =~ s/\@crc/\@crc := $query/;
}
else {
map { s/\@crc/$query/ } @slices;
}
return join(', ', @slices);
}
sub make_row_checksum {
my ( $self, %args ) = @_;
my ( $tbl_struct, $func ) = @args{ qw(tbl_struct function) };
my $q = $self->{Quoter};
my $sep = $args{sep} || '#';
$sep =~ s/'//g;
$sep ||= '#';
my $ignorecols = $args{ignorecols} || {};
my %cols = map { lc($_) => 1 }
grep { !exists $ignorecols->{$_} }
($args{cols} ? @{$args{cols}} : @{$tbl_struct->{cols}});
my %seen;
my @cols =
map {
my $type = $tbl_struct->{type_for}->{$_};
my $result = $q->quote($_);
if ( $type eq 'timestamp' ) {
$result .= ' + 0';
}
elsif ( $args{float_precision} && $type =~ m/float|double/ ) {
$result = "ROUND($result, $args{float_precision})";
}
elsif ( $args{trim} && $type =~ m/varchar/ ) {
$result = "TRIM($result)";
}
$result;
}
grep {
$cols{$_} && !$seen{$_}++
}
@{$tbl_struct->{cols}};
my $query;
if ( !$args{no_cols} ) {
$query = join(', ',
map {
my $col = $_;
if ( $col =~ m/\+ 0/ ) {
my ($real_col) = /^(\S+)/;
$col .= " AS $real_col";
}
elsif ( $col =~ m/TRIM/ ) {
my ($real_col) = m/TRIM\(([^\)]+)\)/;
$col .= " AS $real_col";
}
$col;
} @cols)
. ', ';
}
if ( uc $func ne 'FNV_64' && uc $func ne 'FNV1A_64' ) {
my @nulls = grep { $cols{$_} } @{$tbl_struct->{null_cols}};
if ( @nulls ) {
my $bitmap = "CONCAT("
. join(', ', map { 'ISNULL(' . $q->quote($_) . ')' } @nulls)
. ")";
push @cols, $bitmap;
}
$query .= @cols > 1
? "$func(CONCAT_WS('$sep', " . join(', ', @cols) . '))'
: "$func($cols[0])";
}
else {
my $fnv_func = uc $func;
$query .= "$fnv_func(" . join(', ', @cols) . ')';
}
return $query;
}
sub make_checksum_query {
my ( $self, %args ) = @_;
my @required_args = qw(db tbl tbl_struct algorithm crc_wid crc_type);
foreach my $arg( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ( $db, $tbl, $tbl_struct, $algorithm,
$crc_wid, $crc_type) = @args{@required_args};
my $func = $args{function};
my $q = $self->{Quoter};
my $result;
die "Invalid or missing checksum algorithm"
unless $algorithm && $ALGOS{$algorithm};
if ( $algorithm eq 'CHECKSUM' ) {
return "CHECKSUM TABLE " . $q->quote($db, $tbl);
}
my $expr = $self->make_row_checksum(%args, no_cols=>1);
if ( $algorithm eq 'BIT_XOR' ) {
if ( $crc_type =~ m/int$/ ) {
$result = "COALESCE(LOWER(CONV(BIT_XOR(CAST($expr AS UNSIGNED)), 10, 16)), 0) AS crc ";
}
else {
my $slices = $self->make_xor_slices( query => $expr, %args );
$result = "COALESCE(LOWER(CONCAT($slices)), 0) AS crc ";
}
}
else {
if ( $crc_type =~ m/int$/ ) {
$result = "COALESCE(RIGHT(MAX("
. "\@crc := CONCAT(LPAD(\@cnt := \@cnt + 1, 16, '0'), "
. "CONV(CAST($func(CONCAT(\@crc, $expr)) AS UNSIGNED), 10, 16))"
. "), $crc_wid), 0) AS crc ";
}
else {
$result = "COALESCE(RIGHT(MAX("
. "\@crc := CONCAT(LPAD(\@cnt := \@cnt + 1, 16, '0'), "
. "$func(CONCAT(\@crc, $expr)))"
. "), $crc_wid), 0) AS crc ";
}
}
if ( $args{replicate} ) {
$result = "REPLACE /*PROGRESS_COMMENT*/ INTO $args{replicate} "
. "(db, tbl, chunk, boundaries, this_cnt, this_crc) "
. "SELECT ?, ?, /*CHUNK_NUM*/ ?, COUNT(*) AS cnt, $result";
}
else {
$result = "SELECT "
. ($args{buffer} ? 'SQL_BUFFER_RESULT ' : '')
. "/*PROGRESS_COMMENT*//*CHUNK_NUM*/ COUNT(*) AS cnt, $result";
}
return $result . "FROM /*DB_TBL*//*INDEX_HINT*//*WHERE*/";
}
sub find_replication_differences {
my ( $self, $dbh, $table ) = @_;
(my $sql = <<" EOF") =~ s/\s+/ /gm;
SELECT db, tbl, chunk, boundaries,
COALESCE(this_cnt-master_cnt, 0) AS cnt_diff,
COALESCE(
this_crc <> master_crc OR ISNULL(master_crc) <> ISNULL(this_crc),
0
) AS crc_diff,
this_cnt, master_cnt, this_crc, master_crc
FROM $table
WHERE master_cnt <> this_cnt OR master_crc <> this_crc
OR ISNULL(master_crc) <> ISNULL(this_crc)
EOF
MKDEBUG && _d($sql);
my $diffs = $dbh->selectall_arrayref($sql, { Slice => {} });
return @$diffs;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableChecksum package
# ###########################################################################
# ###########################################################################
# TableSyncChunk package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableSyncChunk.pm
# t/lib/TableSyncChunk.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableSyncChunk;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(TableChunker Quoter) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub name {
return 'Chunk';
}
sub set_callback {
my ( $self, $callback, $code ) = @_;
$self->{$callback} = $code;
return;
}
sub can_sync {
my ( $self, %args ) = @_;
foreach my $arg ( qw(tbl_struct) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my ($exact, @chunkable_cols) = $self->{TableChunker}->find_chunk_columns(
%args,
exact => 1,
);
return unless $exact;
my $colno;
if ( $args{chunk_col} || $args{chunk_index} ) {
MKDEBUG && _d('Checking requested col', $args{chunk_col},
'and/or index', $args{chunk_index});
for my $i ( 0..$#chunkable_cols ) {
if ( $args{chunk_col} ) {
next unless $chunkable_cols[$i]->{column} eq $args{chunk_col};
}
if ( $args{chunk_index} ) {
next unless $chunkable_cols[$i]->{index} eq $args{chunk_index};
}
$colno = $i;
last;
}
if ( !$colno ) {
MKDEBUG && _d('Cannot chunk on column', $args{chunk_col},
'and/or using index', $args{chunk_index});
return;
}
}
else {
$colno = 0; # First, best chunkable column/index.
}
MKDEBUG && _d('Can chunk on column', $chunkable_cols[$colno]->{column},
'using index', $chunkable_cols[$colno]->{index});
return (
1,
chunk_col => $chunkable_cols[$colno]->{column},
chunk_index => $chunkable_cols[$colno]->{index},
),
}
sub prepare_to_sync {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl tbl_struct cols chunk_col
chunk_size crc_col ChangeHandler);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $chunker = $self->{TableChunker};
$self->{chunk_col} = $args{chunk_col};
$self->{crc_col} = $args{crc_col};
$self->{index_hint} = $args{index_hint};
$self->{buffer_in_mysql} = $args{buffer_in_mysql};
$self->{ChangeHandler} = $args{ChangeHandler};
$self->{ChangeHandler}->fetch_back($args{dbh});
push @{$args{cols}}, $args{chunk_col};
my @chunks;
my %range_params = $chunker->get_range_statistics(%args);
if ( !grep { !defined $range_params{$_} } qw(min max rows_in_range) ) {
($args{chunk_size}) = $chunker->size_to_rows(%args);
@chunks = $chunker->calculate_chunks(%args, %range_params);
}
else {
MKDEBUG && _d('No range statistics; using single chunk 1=1');
@chunks = '1=1';
}
$self->{chunks} = \@chunks;
$self->{chunk_num} = 0;
$self->{state} = 0;
return;
}
sub uses_checksum {
return 1;
}
sub set_checksum_queries {
my ( $self, $chunk_sql, $row_sql ) = @_;
die "I need a chunk_sql argument" unless $chunk_sql;
die "I need a row_sql argument" unless $row_sql;
$self->{chunk_sql} = $chunk_sql;
$self->{row_sql} = $row_sql;
return;
}
sub prepare_sync_cycle {
my ( $self, $host ) = @_;
my $sql = 'SET @crc := "", @cnt := 0';
MKDEBUG && _d($sql);
$host->{dbh}->do($sql);
return;
}
sub get_sql {
my ( $self, %args ) = @_;
if ( $self->{state} ) { # select rows in a chunk
my $q = $self->{Quoter};
return 'SELECT /*rows in chunk*/ '
. ($self->{buffer_in_mysql} ? 'SQL_BUFFER_RESULT ' : '')
. $self->{row_sql} . " AS $self->{crc_col}"
. ' FROM ' . $self->{Quoter}->quote(@args{qw(database table)})
. ' '. ($self->{index_hint} || '')
. ' WHERE (' . $self->{chunks}->[$self->{chunk_num}] . ')'
. ($args{where} ? " AND ($args{where})" : '')
. ' ORDER BY ' . join(', ', map {$q->quote($_) } @{$self->key_cols()});
}
else { # select a chunk of rows
return $self->{TableChunker}->inject_chunks(
database => $args{database},
table => $args{table},
chunks => $self->{chunks},
chunk_num => $self->{chunk_num},
query => $self->{chunk_sql},
index_hint => $self->{index_hint},
where => [ $args{where} ],
);
}
}
sub same_row {
my ( $self, %args ) = @_;
my ($lr, $rr) = @args{qw(lr rr)};
if ( $self->{state} ) { # checksumming rows
if ( $lr->{$self->{crc_col}} ne $rr->{$self->{crc_col}} ) {
my $action = 'UPDATE';
my $auth_row = $lr;
my $change_dbh;
if ( $self->{same_row} ) {
($action, $auth_row, $change_dbh) = $self->{same_row}->(%args);
}
$self->{ChangeHandler}->change(
$action, # Execute the action
$auth_row, # with these row values
$self->key_cols(), # identified by these key cols
$change_dbh, # on this dbh
);
}
}
elsif ( $lr->{cnt} != $rr->{cnt} || $lr->{crc} ne $rr->{crc} ) {
MKDEBUG && _d('Rows:', Dumper($lr, $rr));
MKDEBUG && _d('Will examine this chunk before moving to next');
$self->{state} = 1; # Must examine this chunk row-by-row
}
}
sub not_in_right {
my ( $self, %args ) = @_;
die "Called not_in_right in state 0" unless $self->{state};
my $action = 'INSERT';
my $auth_row = $args{lr};
my $change_dbh;
if ( $self->{not_in_right} ) {
($action, $auth_row, $change_dbh) = $self->{not_in_right}->(%args);
}
$self->{ChangeHandler}->change(
$action, # Execute the action
$auth_row, # with these row values
$self->key_cols(), # identified by these key cols
$change_dbh, # on this dbh
);
return;
}
sub not_in_left {
my ( $self, %args ) = @_;
die "Called not_in_left in state 0" unless $self->{state};
my $action = 'DELETE';
my $auth_row = $args{rr};
my $change_dbh;
if ( $self->{not_in_left} ) {
($action, $auth_row, $change_dbh) = $self->{not_in_left}->(%args);
}
$self->{ChangeHandler}->change(
$action, # Execute the action
$auth_row, # with these row values
$self->key_cols(), # identified by these key cols
$change_dbh, # on this dbh
);
return;
}
sub done_with_rows {
my ( $self ) = @_;
if ( $self->{state} == 1 ) {
$self->{state} = 2;
MKDEBUG && _d('Setting state =', $self->{state});
}
else {
$self->{state} = 0;
$self->{chunk_num}++;
MKDEBUG && _d('Setting state =', $self->{state},
'chunk_num =', $self->{chunk_num});
}
return;
}
sub done {
my ( $self ) = @_;
MKDEBUG && _d('Done with', $self->{chunk_num}, 'of',
scalar(@{$self->{chunks}}), 'chunks');
MKDEBUG && $self->{state} && _d('Chunk differs; must examine rows');
return $self->{state} == 0
&& $self->{chunk_num} >= scalar(@{$self->{chunks}})
}
sub pending_changes {
my ( $self ) = @_;
if ( $self->{state} ) {
MKDEBUG && _d('There are pending changes');
return 1;
}
else {
MKDEBUG && _d('No pending changes');
return 0;
}
}
sub key_cols {
my ( $self ) = @_;
my @cols;
if ( $self->{state} == 0 ) {
@cols = qw(chunk_num);
}
else {
@cols = $self->{chunk_col};
}
MKDEBUG && _d('State', $self->{state},',', 'key cols', join(', ', @cols));
return \@cols;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableSyncChunk package
# ###########################################################################
# ###########################################################################
# TableSyncNibble package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableSyncNibble.pm
# t/lib/TableSyncNibble.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableSyncNibble;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(TableNibbler TableChunker TableParser Quoter) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub name {
return 'Nibble';
}
sub can_sync {
my ( $self, %args ) = @_;
foreach my $arg ( qw(tbl_struct) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $nibble_index = $self->{TableParser}->find_best_index($args{tbl_struct});
if ( $nibble_index ) {
MKDEBUG && _d('Best nibble index:', Dumper($nibble_index));
if ( !$args{tbl_struct}->{keys}->{$nibble_index}->{is_unique} ) {
MKDEBUG && _d('Best nibble index is not unique');
return;
}
if ( $args{chunk_index} && $args{chunk_index} ne $nibble_index ) {
MKDEBUG && _d('Best nibble index is not requested index',
$args{chunk_index});
return;
}
}
else {
MKDEBUG && _d('No best nibble index returned');
return;
}
my $small_table = 0;
if ( $args{src} && $args{src}->{dbh} ) {
my $dbh = $args{src}->{dbh};
my $db = $args{src}->{db};
my $tbl = $args{src}->{tbl};
my $table_status;
eval {
my $sql = "SHOW TABLE STATUS FROM `$db` LIKE "
. $self->{Quoter}->literal_like($tbl);
MKDEBUG && _d($sql);
$table_status = $dbh->selectrow_hashref($sql);
};
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
if ( $table_status ) {
my $n_rows = defined $table_status->{Rows} ? $table_status->{Rows}
: defined $table_status->{rows} ? $table_status->{rows}
: undef;
$small_table = 1 if defined $n_rows && $n_rows <= 100;
}
}
MKDEBUG && _d('Small table:', $small_table);
MKDEBUG && _d('Can nibble using index', $nibble_index);
return (
1,
chunk_index => $nibble_index,
key_cols => $args{tbl_struct}->{keys}->{$nibble_index}->{cols},
small_table => $small_table,
);
}
sub prepare_to_sync {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db tbl tbl_struct chunk_index key_cols chunk_size
crc_col ChangeHandler);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
$self->{dbh} = $args{dbh};
$self->{tbl_struct} = $args{tbl_struct};
$self->{crc_col} = $args{crc_col};
$self->{index_hint} = $args{index_hint};
$self->{key_cols} = $args{key_cols};
($self->{chunk_size}) = $self->{TableChunker}->size_to_rows(%args);
$self->{buffer_in_mysql} = $args{buffer_in_mysql};
$self->{small_table} = $args{small_table};
$self->{ChangeHandler} = $args{ChangeHandler};
$self->{ChangeHandler}->fetch_back($args{dbh});
my %seen;
my @ucols = grep { !$seen{$_}++ } @{$args{cols}}, @{$args{key_cols}};
$args{cols} = \@ucols;
$self->{sel_stmt} = $self->{TableNibbler}->generate_asc_stmt(
%args,
index => $args{chunk_index}, # expects an index arg, not chunk_index
asc_only => 1,
);
$self->{nibble} = 0;
$self->{cached_row} = undef;
$self->{cached_nibble} = undef;
$self->{cached_boundaries} = undef;
$self->{state} = 0;
return;
}
sub uses_checksum {
return 1;
}
sub set_checksum_queries {
my ( $self, $nibble_sql, $row_sql ) = @_;
die "I need a nibble_sql argument" unless $nibble_sql;
die "I need a row_sql argument" unless $row_sql;
$self->{nibble_sql} = $nibble_sql;
$self->{row_sql} = $row_sql;
return;
}
sub prepare_sync_cycle {
my ( $self, $host ) = @_;
my $sql = 'SET @crc := "", @cnt := 0';
MKDEBUG && _d($sql);
$host->{dbh}->do($sql);
return;
}
sub get_sql {
my ( $self, %args ) = @_;
if ( $self->{state} ) {
my $q = $self->{Quoter};
return 'SELECT /*rows in nibble*/ '
. ($self->{buffer_in_mysql} ? 'SQL_BUFFER_RESULT ' : '')
. $self->{row_sql} . " AS $self->{crc_col}"
. ' FROM ' . $q->quote(@args{qw(database table)})
. ' ' . ($self->{index_hint} ? $self->{index_hint} : '')
. ' WHERE (' . $self->__get_boundaries(%args) . ')'
. ($args{where} ? " AND ($args{where})" : '')
. ' ORDER BY ' . join(', ', map {$q->quote($_) } @{$self->key_cols()});
}
else {
my $where = $self->__get_boundaries(%args);
return $self->{TableChunker}->inject_chunks(
database => $args{database},
table => $args{table},
chunks => [ $where ],
chunk_num => 0,
query => $self->{nibble_sql},
index_hint => $self->{index_hint},
where => [ $args{where} ],
);
}
}
sub __get_boundaries {
my ( $self, %args ) = @_;
my $q = $self->{Quoter};
my $s = $self->{sel_stmt};
my $lb; # Lower boundary part of WHERE
my $ub; # Upper boundary part of WHERE
my $row; # Next upper boundary row or cached_row
if ( $self->{cached_boundaries} ) {
MKDEBUG && _d('Using cached boundaries');
return $self->{cached_boundaries};
}
if ( $self->{cached_row} && $self->{cached_nibble} == $self->{nibble} ) {
MKDEBUG && _d('Using cached row for boundaries');
$row = $self->{cached_row};
}
else {
MKDEBUG && _d('Getting next upper boundary row');
my $sql;
($sql, $lb) = $self->__make_boundary_sql(%args); # $lb from outer scope!
if ( $self->{nibble} == 0 && !$self->{small_table} ) {
my $explain_index = $self->__get_explain_index($sql);
if ( lc($explain_index || '') ne lc($s->{index}) ) {
die 'Cannot nibble table '.$q->quote($args{database}, $args{table})
. " because MySQL chose "
. ($explain_index ? "the `$explain_index`" : 'no') . ' index'
. " instead of the `$s->{index}` index";
}
}
$row = $self->{dbh}->selectrow_hashref($sql);
MKDEBUG && _d($row ? 'Got a row' : "Didn't get a row");
}
if ( $row ) {
my $i = 0;
$ub = $s->{boundaries}->{'<='};
$ub =~ s/\?/$q->quote_val($row->{$s->{scols}->[$i]}, $self->{tbl_struct}->{is_numeric}->{$s->{scols}->[$i++]} || 0)/eg;
}
else {
MKDEBUG && _d('No upper boundary');
$ub = '1=1';
}
my $where = $lb ? "($lb AND $ub)" : $ub;
$self->{cached_row} = $row;
$self->{cached_nibble} = $self->{nibble};
$self->{cached_boundaries} = $where;
MKDEBUG && _d('WHERE clause:', $where);
return $where;
}
sub __make_boundary_sql {
my ( $self, %args ) = @_;
my $lb;
my $q = $self->{Quoter};
my $s = $self->{sel_stmt};
my $sql = "SELECT /*nibble boundary $self->{nibble}*/ "
. join(',', map { $q->quote($_) } @{$s->{cols}})
. " FROM " . $q->quote($args{database}, $args{table})
. ' ' . ($self->{index_hint} || '')
. ($args{where} ? " WHERE ($args{where})" : "");
if ( $self->{nibble} ) {
my $tmp = $self->{cached_row};
my $i = 0;
$lb = $s->{boundaries}->{'>'};
$lb =~ s/\?/$q->quote_val($tmp->{$s->{scols}->[$i]}, $self->{tbl_struct}->{is_numeric}->{$s->{scols}->[$i++]} || 0)/eg;
$sql .= $args{where} ? " AND $lb" : " WHERE $lb";
}
$sql .= " ORDER BY " . join(',', map { $q->quote($_) } @{$self->{key_cols}})
. ' LIMIT ' . ($self->{chunk_size} - 1) . ', 1';
MKDEBUG && _d('Lower boundary:', $lb);
MKDEBUG && _d('Next boundary sql:', $sql);
return $sql, $lb;
}
sub __get_explain_index {
my ( $self, $sql ) = @_;
return unless $sql;
my $explain;
eval {
$explain = $self->{dbh}->selectall_arrayref("EXPLAIN $sql",{Slice => {}});
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d($EVAL_ERROR);
return;
}
MKDEBUG && _d('EXPLAIN key:', $explain->[0]->{key});
return $explain->[0]->{key};
}
sub same_row {
my ( $self, %args ) = @_;
my ($lr, $rr) = @args{qw(lr rr)};
if ( $self->{state} ) {
if ( $lr->{$self->{crc_col}} ne $rr->{$self->{crc_col}} ) {
$self->{ChangeHandler}->change('UPDATE', $lr, $self->key_cols());
}
}
elsif ( $lr->{cnt} != $rr->{cnt} || $lr->{crc} ne $rr->{crc} ) {
MKDEBUG && _d('Rows:', Dumper($lr, $rr));
MKDEBUG && _d('Will examine this nibble before moving to next');
$self->{state} = 1; # Must examine this nibble row-by-row
}
}
sub not_in_right {
my ( $self, %args ) = @_;
die "Called not_in_right in state 0" unless $self->{state};
$self->{ChangeHandler}->change('INSERT', $args{lr}, $self->key_cols());
}
sub not_in_left {
my ( $self, %args ) = @_;
die "Called not_in_left in state 0" unless $self->{state};
$self->{ChangeHandler}->change('DELETE', $args{rr}, $self->key_cols());
}
sub done_with_rows {
my ( $self ) = @_;
if ( $self->{state} == 1 ) {
$self->{state} = 2;
MKDEBUG && _d('Setting state =', $self->{state});
}
else {
$self->{state} = 0;
$self->{nibble}++;
delete $self->{cached_boundaries};
MKDEBUG && _d('Setting state =', $self->{state},
', nibble =', $self->{nibble});
}
}
sub done {
my ( $self ) = @_;
MKDEBUG && _d('Done with nibble', $self->{nibble});
MKDEBUG && $self->{state} && _d('Nibble differs; must examine rows');
return $self->{state} == 0 && $self->{nibble} && !$self->{cached_row};
}
sub pending_changes {
my ( $self ) = @_;
if ( $self->{state} ) {
MKDEBUG && _d('There are pending changes');
return 1;
}
else {
MKDEBUG && _d('No pending changes');
return 0;
}
}
sub key_cols {
my ( $self ) = @_;
my @cols;
if ( $self->{state} == 0 ) {
@cols = qw(chunk_num);
}
else {
@cols = @{$self->{key_cols}};
}
MKDEBUG && _d('State', $self->{state},',', 'key cols', join(', ', @cols));
return \@cols;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableSyncNibble package
# ###########################################################################
# ###########################################################################
# TableSyncGroupBy package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableSyncGroupBy.pm
# t/lib/TableSyncGroupBy.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableSyncGroupBy;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub name {
return 'GroupBy';
}
sub can_sync {
return 1; # We can sync anything.
}
sub prepare_to_sync {
my ( $self, %args ) = @_;
my @required_args = qw(tbl_struct cols ChangeHandler);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
$self->{cols} = $args{cols};
$self->{buffer_in_mysql} = $args{buffer_in_mysql};
$self->{ChangeHandler} = $args{ChangeHandler};
$self->{count_col} = '__maatkit_count';
while ( $args{tbl_struct}->{is_col}->{$self->{count_col}} ) {
$self->{count_col} = "_$self->{count_col}";
}
MKDEBUG && _d('COUNT column will be named', $self->{count_col});
$self->{done} = 0;
return;
}
sub uses_checksum {
return 0; # We don't need checksum queries.
}
sub set_checksum_queries {
return; # This shouldn't be called, but just in case.
}
sub prepare_sync_cycle {
my ( $self, $host ) = @_;
return;
}
sub get_sql {
my ( $self, %args ) = @_;
my $cols = join(', ', map { $self->{Quoter}->quote($_) } @{$self->{cols}});
return "SELECT"
. ($self->{buffer_in_mysql} ? ' SQL_BUFFER_RESULT' : '')
. " $cols, COUNT(*) AS $self->{count_col}"
. ' FROM ' . $self->{Quoter}->quote(@args{qw(database table)})
. ' WHERE ' . ( $args{where} || '1=1' )
. " GROUP BY $cols ORDER BY $cols";
}
sub same_row {
my ( $self, %args ) = @_;
my ($lr, $rr) = @args{qw(lr rr)};
my $cc = $self->{count_col};
my $lc = $lr->{$cc};
my $rc = $rr->{$cc};
my $diff = abs($lc - $rc);
return unless $diff;
$lr = { %$lr };
delete $lr->{$cc};
$rr = { %$rr };
delete $rr->{$cc};
foreach my $i ( 1 .. $diff ) {
if ( $lc > $rc ) {
$self->{ChangeHandler}->change('INSERT', $lr, $self->key_cols());
}
else {
$self->{ChangeHandler}->change('DELETE', $rr, $self->key_cols());
}
}
}
sub not_in_right {
my ( $self, %args ) = @_;
my $lr = $args{lr};
$lr = { %$lr };
my $cnt = delete $lr->{$self->{count_col}};
foreach my $i ( 1 .. $cnt ) {
$self->{ChangeHandler}->change('INSERT', $lr, $self->key_cols());
}
}
sub not_in_left {
my ( $self, %args ) = @_;
my $rr = $args{rr};
$rr = { %$rr };
my $cnt = delete $rr->{$self->{count_col}};
foreach my $i ( 1 .. $cnt ) {
$self->{ChangeHandler}->change('DELETE', $rr, $self->key_cols());
}
}
sub done_with_rows {
my ( $self ) = @_;
$self->{done} = 1;
}
sub done {
my ( $self ) = @_;
return $self->{done};
}
sub key_cols {
my ( $self ) = @_;
return $self->{cols};
}
sub pending_changes {
my ( $self ) = @_;
return;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableSyncGroupBy package
# ###########################################################################
# ###########################################################################
# TableSyncer package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableSyncer.pm
# t/lib/TableSyncer.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableSyncer;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
sub new {
my ( $class, %args ) = @_;
my @required_args = qw(MasterSlave Quoter VersionParser TableChecksum Retry);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub get_best_plugin {
my ( $self, %args ) = @_;
foreach my $arg ( qw(plugins tbl_struct) ) {
die "I need a $arg argument" unless $args{$arg};
}
MKDEBUG && _d('Getting best plugin');
foreach my $plugin ( @{$args{plugins}} ) {
MKDEBUG && _d('Trying plugin', $plugin->name);
my ($can_sync, %plugin_args) = $plugin->can_sync(%args);
if ( $can_sync ) {
MKDEBUG && _d('Can sync with', $plugin->name, Dumper(\%plugin_args));
return $plugin, %plugin_args;
}
}
MKDEBUG && _d('No plugin can sync the table');
return;
}
sub sync_table {
my ( $self, %args ) = @_;
my @required_args = qw(plugins src dst tbl_struct cols chunk_size
RowDiff ChangeHandler);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
MKDEBUG && _d('Syncing table with args:',
map { "$_: " . Dumper($args{$_}) }
qw(plugins src dst tbl_struct cols chunk_size));
my ($plugins, $src, $dst, $tbl_struct, $cols, $chunk_size, $rd, $ch)
= @args{@required_args};
my $dp = $self->{DSNParser};
$args{trace} = 1 unless defined $args{trace};
if ( $args{bidirectional} && $args{ChangeHandler}->{queue} ) {
die "Queueing does not work with bidirectional syncing";
}
$args{index_hint} = 1 unless defined $args{index_hint};
$args{lock} ||= 0;
$args{wait} ||= 0;
$args{transaction} ||= 0;
$args{timeout_ok} ||= 0;
my $q = $self->{Quoter};
my $vp = $self->{VersionParser};
my ($plugin, %plugin_args) = $self->get_best_plugin(%args);
die "No plugin can sync $src->{db}.$src->{tbl}" unless $plugin;
my $crc_col = '__crc';
while ( $tbl_struct->{is_col}->{$crc_col} ) {
$crc_col = "_$crc_col"; # Prepend more _ until not a column.
}
MKDEBUG && _d('CRC column:', $crc_col);
my $index_hint;
my $hint = ($vp->version_ge($src->{dbh}, '4.0.9')
&& $vp->version_ge($dst->{dbh}, '4.0.9') ? 'FORCE' : 'USE')
. ' INDEX';
if ( $args{chunk_index} ) {
MKDEBUG && _d('Using given chunk index for index hint');
$index_hint = "$hint (" . $q->quote($args{chunk_index}) . ")";
}
elsif ( $plugin_args{chunk_index} && $args{index_hint} ) {
MKDEBUG && _d('Using chunk index chosen by plugin for index hint');
$index_hint = "$hint (" . $q->quote($plugin_args{chunk_index}) . ")";
}
MKDEBUG && _d('Index hint:', $index_hint);
eval {
$plugin->prepare_to_sync(
%args,
%plugin_args,
dbh => $src->{dbh},
db => $src->{db},
tbl => $src->{tbl},
crc_col => $crc_col,
index_hint => $index_hint,
);
};
if ( $EVAL_ERROR ) {
die 'Failed to prepare TableSync', $plugin->name, ' plugin: ',
$EVAL_ERROR;
}
if ( $plugin->uses_checksum() ) {
eval {
my ($chunk_sql, $row_sql) = $self->make_checksum_queries(%args);
$plugin->set_checksum_queries($chunk_sql, $row_sql);
};
if ( $EVAL_ERROR ) {
die "Failed to make checksum queries: $EVAL_ERROR";
}
}
if ( $args{dry_run} ) {
return $ch->get_changes(), ALGORITHM => $plugin->name;
}
eval {
$src->{dbh}->do("USE `$src->{db}`");
$dst->{dbh}->do("USE `$dst->{db}`");
};
if ( $EVAL_ERROR ) {
die "Failed to USE database on source or destination: $EVAL_ERROR";
}
MKDEBUG && _d('left dbh', $src->{dbh});
MKDEBUG && _d('right dbh', $dst->{dbh});
chomp(my $hostname = `hostname`);
my $trace_msg
= $args{trace} ? "src_db:$src->{db} src_tbl:$src->{tbl} "
. ($dp && $src->{dsn} ? "src_dsn:".$dp->as_string($src->{dsn}) : "")
. " dst_db:$dst->{db} dst_tbl:$dst->{tbl} "
. ($dp && $dst->{dsn} ? "dst_dsn:".$dp->as_string($dst->{dsn}) : "")
. " " . join(" ", map { "$_:" . ($args{$_} || 0) }
qw(lock transaction changing_src replicate bidirectional))
. " pid:$PID "
. ($ENV{USER} ? "user:$ENV{USER} " : "")
. ($hostname ? "host:$hostname" : "")
: "";
MKDEBUG && _d("Binlog trace message:", $trace_msg);
$self->lock_and_wait(%args, lock_level => 2); # per-table lock
my $callback = $args{callback};
my $cycle = 0;
while ( !$plugin->done() ) {
MKDEBUG && _d('Beginning sync cycle', $cycle);
my $src_sql = $plugin->get_sql(
database => $src->{db},
table => $src->{tbl},
where => $args{where},
);
my $dst_sql = $plugin->get_sql(
database => $dst->{db},
table => $dst->{tbl},
where => $args{where},
);
if ( $args{transaction} ) {
if ( $args{bidirectional} ) {
$src_sql .= ' FOR UPDATE';
$dst_sql .= ' FOR UPDATE';
}
elsif ( $args{changing_src} ) {
$src_sql .= ' FOR UPDATE';
$dst_sql .= ' LOCK IN SHARE MODE';
}
else {
$src_sql .= ' LOCK IN SHARE MODE';
$dst_sql .= ' FOR UPDATE';
}
}
MKDEBUG && _d('src:', $src_sql);
MKDEBUG && _d('dst:', $dst_sql);
$callback->($src_sql, $dst_sql) if $callback;
$plugin->prepare_sync_cycle($src);
$plugin->prepare_sync_cycle($dst);
my $src_sth = $src->{dbh}->prepare($src_sql);
my $dst_sth = $dst->{dbh}->prepare($dst_sql);
if ( $args{buffer_to_client} ) {
$src_sth->{mysql_use_result} = 1;
$dst_sth->{mysql_use_result} = 1;
}
my $executed_src = 0;
if ( !$cycle || !$plugin->pending_changes() ) {
$executed_src
= $self->lock_and_wait(%args, src_sth => $src_sth, lock_level => 1);
}
$src_sth->execute() unless $executed_src;
$dst_sth->execute();
$rd->compare_sets(
left_sth => $src_sth,
right_sth => $dst_sth,
left_dbh => $src->{dbh},
right_dbh => $dst->{dbh},
syncer => $plugin,
tbl_struct => $tbl_struct,
);
$ch->process_rows(1, $trace_msg);
MKDEBUG && _d('Finished sync cycle', $cycle);
$cycle++;
}
$ch->process_rows(0, $trace_msg);
$self->unlock(%args, lock_level => 2);
return $ch->get_changes(), ALGORITHM => $plugin->name;
}
sub make_checksum_queries {
my ( $self, %args ) = @_;
my @required_args = qw(src dst tbl_struct);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($src, $dst, $tbl_struct) = @args{@required_args};
my $checksum = $self->{TableChecksum};
my $src_algo = $checksum->best_algorithm(
algorithm => 'BIT_XOR',
dbh => $src->{dbh},
where => 1,
chunk => 1,
count => 1,
);
my $dst_algo = $checksum->best_algorithm(
algorithm => 'BIT_XOR',
dbh => $dst->{dbh},
where => 1,
chunk => 1,
count => 1,
);
if ( $src_algo ne $dst_algo ) {
die "Source and destination checksum algorithms are different: ",
"$src_algo on source, $dst_algo on destination"
}
MKDEBUG && _d('Chosen algo:', $src_algo);
my $src_func = $checksum->choose_hash_func(dbh => $src->{dbh}, %args);
my $dst_func = $checksum->choose_hash_func(dbh => $dst->{dbh}, %args);
if ( $src_func ne $dst_func ) {
die "Source and destination hash functions are different: ",
"$src_func on source, $dst_func on destination";
}
MKDEBUG && _d('Chosen hash func:', $src_func);
my $crc_wid = $checksum->get_crc_wid($src->{dbh}, $src_func);
my ($crc_type) = $checksum->get_crc_type($src->{dbh}, $src_func);
my $opt_slice;
if ( $src_algo eq 'BIT_XOR' && $crc_type !~ m/int$/ ) {
$opt_slice = $checksum->optimize_xor(
dbh => $src->{dbh},
function => $src_func
);
}
my $chunk_sql = $checksum->make_checksum_query(
%args,
db => $src->{db},
tbl => $src->{tbl},
algorithm => $src_algo,
function => $src_func,
crc_wid => $crc_wid,
crc_type => $crc_type,
opt_slice => $opt_slice,
replicate => undef, # replicate means something different to this sub
); # than what we use it for; do not pass it!
MKDEBUG && _d('Chunk sql:', $chunk_sql);
my $row_sql = $checksum->make_row_checksum(
%args,
function => $src_func,
);
MKDEBUG && _d('Row sql:', $row_sql);
return $chunk_sql, $row_sql;
}
sub lock_table {
my ( $self, $dbh, $where, $db_tbl, $mode ) = @_;
my $query = "LOCK TABLES $db_tbl $mode";
MKDEBUG && _d($query);
$dbh->do($query);
MKDEBUG && _d('Acquired table lock on', $where, 'in', $mode, 'mode');
}
sub unlock {
my ( $self, %args ) = @_;
foreach my $arg ( qw(src dst lock transaction lock_level) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $src = $args{src};
my $dst = $args{dst};
return unless $args{lock} && $args{lock} <= $args{lock_level};
foreach my $dbh ( $src->{dbh}, $dst->{dbh} ) {
if ( $args{transaction} ) {
MKDEBUG && _d('Committing', $dbh);
$dbh->commit();
}
else {
my $sql = 'UNLOCK TABLES';
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
}
return;
}
sub lock_and_wait {
my ( $self, %args ) = @_;
my $result = 0;
foreach my $arg ( qw(src dst lock lock_level) ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my $src = $args{src};
my $dst = $args{dst};
return unless $args{lock} && $args{lock} == $args{lock_level};
MKDEBUG && _d('lock and wait, lock level', $args{lock});
foreach my $dbh ( $src->{dbh}, $dst->{dbh} ) {
if ( $args{transaction} ) {
MKDEBUG && _d('Committing', $dbh);
$dbh->commit();
}
else {
my $sql = 'UNLOCK TABLES';
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
}
if ( $args{lock} == 3 ) {
my $sql = 'FLUSH TABLES WITH READ LOCK';
MKDEBUG && _d($src->{dbh}, $sql);
$src->{dbh}->do($sql);
}
else {
if ( $args{transaction} ) {
if ( $args{src_sth} ) {
MKDEBUG && _d('Executing statement on source to lock rows');
my $sql = "START TRANSACTION /*!40108 WITH CONSISTENT SNAPSHOT */";
MKDEBUG && _d($src->{dbh}, $sql);
$src->{dbh}->do($sql);
$args{src_sth}->execute();
$result = 1;
}
}
else {
$self->lock_table($src->{dbh}, 'source',
$self->{Quoter}->quote($src->{db}, $src->{tbl}),
$args{changing_src} ? 'WRITE' : 'READ');
}
}
eval {
if ( my $timeout = $args{wait} ) {
my $wait = $args{wait_retry_args}->{wait} || 10;
my $tries = $args{wait_retry_args}->{tries} || 3;
$self->{Retry}->retry(
wait => sub { sleep $wait; },
tries => $tries,
try => sub {
my ( %args ) = @_;
if ( $args{tryno} > 1 ) {
warn "Retrying MASTER_POS_WAIT() for --wait $timeout...";
}
my $ms = $self->{MasterSlave};
my $wait = $ms->wait_for_master(
master_status => $ms->get_master_status($src->{misc_dbh}),
slave_dbh => $dst->{dbh},
timeout => $timeout,
);
if ( !defined $wait->{result} ) {
my $msg;
if ( $wait->{waited} ) {
$msg = "The slave was stopped while waiting with "
. "MASTER_POS_WAIT().";
}
else {
$msg = "MASTER_POS_WAIT() returned NULL. Verify that "
. "the slave is running.";
}
if ( $tries - $args{tryno} ) {
$msg .= " Sleeping $wait seconds then retrying "
. ($tries - $args{tryno}) . " more times.";
}
warn $msg;
return;
}
elsif ( $wait->{result} == -1 ) {
die "Slave did not catch up to its master after waiting "
. "$timeout seconds with MASTER_POS_WAIT. Try inceasing "
. "the --wait time, or disable this feature by specifying "
. "--wait 0.";
}
else {
return $result; # slave caught up
}
},
on_failure => sub {
die "Slave did not catch up to its master after $tries attempts "
. "of waiting $timeout seconds with MASTER_POS_WAIT. "
. "Check that the slave is running, increase the --wait "
. "time, or disable this feature by specifying --wait 0.";
},
); # retry MasterSlave::wait_for_master()
}
if ( $args{changing_src} ) {
MKDEBUG && _d('Not locking destination because changing source ',
'(syncing via replication or sync-to-master)');
}
else {
if ( $args{lock} == 3 ) {
my $sql = 'FLUSH TABLES WITH READ LOCK';
MKDEBUG && _d($dst->{dbh}, ',', $sql);
$dst->{dbh}->do($sql);
}
elsif ( !$args{transaction} ) {
$self->lock_table($dst->{dbh}, 'dest',
$self->{Quoter}->quote($dst->{db}, $dst->{tbl}),
$args{execute} ? 'WRITE' : 'READ');
}
}
};
if ( $EVAL_ERROR ) {
if ( $args{src_sth}->{Active} ) {
$args{src_sth}->finish();
}
foreach my $dbh ( $src->{dbh}, $dst->{dbh}, $src->{misc_dbh} ) {
next unless $dbh;
MKDEBUG && _d('Caught error, unlocking/committing on', $dbh);
$dbh->do('UNLOCK TABLES');
$dbh->commit() unless $dbh->{AutoCommit};
}
die $EVAL_ERROR;
}
return $result;
}
sub have_all_privs {
my ( $self, $dbh, $db, $tbl ) = @_;
my $db_tbl = $self->{Quoter}->quote($db, $tbl);
my $sql = "SHOW FULL COLUMNS FROM $db_tbl";
MKDEBUG && _d('Permissions check:', $sql);
my $cols = $dbh->selectall_arrayref($sql, {Slice => {}});
my ($hdr_name) = grep { m/privileges/i } keys %{$cols->[0]};
my $privs = $cols->[0]->{$hdr_name};
$sql = "DELETE FROM $db_tbl LIMIT 0"; # FULL COLUMNS doesn't show all privs
MKDEBUG && _d('Permissions check:', $sql);
eval { $dbh->do($sql); };
my $can_delete = $EVAL_ERROR ? 0 : 1;
MKDEBUG && _d('User privs on', $db_tbl, ':', $privs,
($can_delete ? 'delete' : ''));
if ( $privs =~ m/select/ && $privs =~ m/insert/ && $privs =~ m/update/
&& $can_delete ) {
MKDEBUG && _d('User has all privs');
return 1;
}
MKDEBUG && _d('User does not have all privs');
return 0;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableSyncer package
# ###########################################################################
# ###########################################################################
# TableNibbler package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/TableNibbler.pm
# t/lib/TableNibbler.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package TableNibbler;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
my @required_args = qw(TableParser Quoter);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = { %args };
return bless $self, $class;
}
sub generate_asc_stmt {
my ( $self, %args ) = @_;
my @required_args = qw(tbl_struct index);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless defined $args{$arg};
}
my ($tbl_struct, $index) = @args{@required_args};
my @cols = $args{cols} ? @{$args{cols}} : @{$tbl_struct->{cols}};
my $q = $self->{Quoter};
die "Index '$index' does not exist in table"
unless exists $tbl_struct->{keys}->{$index};
my @asc_cols = @{$tbl_struct->{keys}->{$index}->{cols}};
my @asc_slice;
@asc_cols = @{$tbl_struct->{keys}->{$index}->{cols}};
MKDEBUG && _d('Will ascend index', $index);
MKDEBUG && _d('Will ascend columns', join(', ', @asc_cols));
if ( $args{asc_first} ) {
@asc_cols = $asc_cols[0];
MKDEBUG && _d('Ascending only first column');
}
my %col_posn = do { my $i = 0; map { $_ => $i++ } @cols };
foreach my $col ( @asc_cols ) {
if ( !exists $col_posn{$col} ) {
push @cols, $col;
$col_posn{$col} = $#cols;
}
push @asc_slice, $col_posn{$col};
}
MKDEBUG && _d('Will ascend, in ordinal position:', join(', ', @asc_slice));
my $asc_stmt = {
cols => \@cols,
index => $index,
where => '',
slice => [],
scols => [],
};
if ( @asc_slice ) {
my $cmp_where;
foreach my $cmp ( qw(< <= >= >) ) {
$cmp_where = $self->generate_cmp_where(
type => $cmp,
slice => \@asc_slice,
cols => \@cols,
quoter => $q,
is_nullable => $tbl_struct->{is_nullable},
);
$asc_stmt->{boundaries}->{$cmp} = $cmp_where->{where};
}
my $cmp = $args{asc_only} ? '>' : '>=';
$asc_stmt->{where} = $asc_stmt->{boundaries}->{$cmp};
$asc_stmt->{slice} = $cmp_where->{slice};
$asc_stmt->{scols} = $cmp_where->{scols};
}
return $asc_stmt;
}
sub generate_cmp_where {
my ( $self, %args ) = @_;
foreach my $arg ( qw(type slice cols is_nullable) ) {
die "I need a $arg arg" unless defined $args{$arg};
}
my @slice = @{$args{slice}};
my @cols = @{$args{cols}};
my $is_nullable = $args{is_nullable};
my $type = $args{type};
my $q = $self->{Quoter};
(my $cmp = $type) =~ s/=//;
my @r_slice; # Resulting slice columns, by ordinal
my @r_scols; # Ditto, by name
my @clauses;
foreach my $i ( 0 .. $#slice ) {
my @clause;
foreach my $j ( 0 .. $i - 1 ) {
my $ord = $slice[$j];
my $col = $cols[$ord];
my $quo = $q->quote($col);
if ( $is_nullable->{$col} ) {
push @clause, "((? IS NULL AND $quo IS NULL) OR ($quo = ?))";
push @r_slice, $ord, $ord;
push @r_scols, $col, $col;
}
else {
push @clause, "$quo = ?";
push @r_slice, $ord;
push @r_scols, $col;
}
}
my $ord = $slice[$i];
my $col = $cols[$ord];
my $quo = $q->quote($col);
my $end = $i == $#slice; # Last clause of the whole group.
if ( $is_nullable->{$col} ) {
if ( $type =~ m/=/ && $end ) {
push @clause, "(? IS NULL OR $quo $type ?)";
}
elsif ( $type =~ m/>/ ) {
push @clause, "((? IS NULL AND $quo IS NOT NULL) OR ($quo $cmp ?))";
}
else { # If $type =~ m/</ ) {
push @clause, "((? IS NOT NULL AND $quo IS NULL) OR ($quo $cmp ?))";
}
push @r_slice, $ord, $ord;
push @r_scols, $col, $col;
}
else {
push @r_slice, $ord;
push @r_scols, $col;
push @clause, ($type =~ m/=/ && $end ? "$quo $type ?" : "$quo $cmp ?");
}
push @clauses, '(' . join(' AND ', @clause) . ')';
}
my $result = '(' . join(' OR ', @clauses) . ')';
my $where = {
slice => \@r_slice,
scols => \@r_scols,
where => $result,
};
return $where;
}
sub generate_del_stmt {
my ( $self, %args ) = @_;
my $tbl = $args{tbl_struct};
my @cols = $args{cols} ? @{$args{cols}} : ();
my $tp = $self->{TableParser};
my $q = $self->{Quoter};
my @del_cols;
my @del_slice;
my $index = $tp->find_best_index($tbl, $args{index});
die "Cannot find an ascendable index in table" unless $index;
if ( $index ) {
@del_cols = @{$tbl->{keys}->{$index}->{cols}};
}
else {
@del_cols = @{$tbl->{cols}};
}
MKDEBUG && _d('Columns needed for DELETE:', join(', ', @del_cols));
my %col_posn = do { my $i = 0; map { $_ => $i++ } @cols };
foreach my $col ( @del_cols ) {
if ( !exists $col_posn{$col} ) {
push @cols, $col;
$col_posn{$col} = $#cols;
}
push @del_slice, $col_posn{$col};
}
MKDEBUG && _d('Ordinals needed for DELETE:', join(', ', @del_slice));
my $del_stmt = {
cols => \@cols,
index => $index,
where => '',
slice => [],
scols => [],
};
my @clauses;
foreach my $i ( 0 .. $#del_slice ) {
my $ord = $del_slice[$i];
my $col = $cols[$ord];
my $quo = $q->quote($col);
if ( $tbl->{is_nullable}->{$col} ) {
push @clauses, "((? IS NULL AND $quo IS NULL) OR ($quo = ?))";
push @{$del_stmt->{slice}}, $ord, $ord;
push @{$del_stmt->{scols}}, $col, $col;
}
else {
push @clauses, "$quo = ?";
push @{$del_stmt->{slice}}, $ord;
push @{$del_stmt->{scols}}, $col;
}
}
$del_stmt->{where} = '(' . join(' AND ', @clauses) . ')';
return $del_stmt;
}
sub generate_ins_stmt {
my ( $self, %args ) = @_;
foreach my $arg ( qw(ins_tbl sel_cols) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $ins_tbl = $args{ins_tbl};
my @sel_cols = @{$args{sel_cols}};
die "You didn't specify any SELECT columns" unless @sel_cols;
my @ins_cols;
my @ins_slice;
for my $i ( 0..$#sel_cols ) {
next unless $ins_tbl->{is_col}->{$sel_cols[$i]};
push @ins_cols, $sel_cols[$i];
push @ins_slice, $i;
}
return {
cols => \@ins_cols,
slice => \@ins_slice,
};
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End TableNibbler package
# ###########################################################################
# ###########################################################################
# MasterSlave package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/MasterSlave.pm
# t/lib/MasterSlave.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package MasterSlave;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
my $self = {
%args,
replication_thread => {},
};
return bless $self, $class;
}
sub recurse_to_slaves {
my ( $self, $args, $level ) = @_;
$level ||= 0;
my $dp = $args->{dsn_parser};
my $dsn = $args->{dsn};
my $dbh;
eval {
$dbh = $args->{dbh} || $dp->get_dbh(
$dp->get_cxn_params($dsn), { AutoCommit => 1 });
MKDEBUG && _d('Connected to', $dp->as_string($dsn));
};
if ( $EVAL_ERROR ) {
print STDERR "Cannot connect to ", $dp->as_string($dsn), "\n"
or die "Cannot print: $OS_ERROR";
return;
}
my $sql = 'SELECT @@SERVER_ID';
MKDEBUG && _d($sql);
my ($id) = $dbh->selectrow_array($sql);
MKDEBUG && _d('Working on server ID', $id);
my $master_thinks_i_am = $dsn->{server_id};
if ( !defined $id
|| ( defined $master_thinks_i_am && $master_thinks_i_am != $id )
|| $args->{server_ids_seen}->{$id}++
) {
MKDEBUG && _d('Server ID seen, or not what master said');
if ( $args->{skip_callback} ) {
$args->{skip_callback}->($dsn, $dbh, $level, $args->{parent});
}
return;
}
$args->{callback}->($dsn, $dbh, $level, $args->{parent});
if ( !defined $args->{recurse} || $level < $args->{recurse} ) {
my @slaves =
grep { !$_->{master_id} || $_->{master_id} == $id } # Only my slaves.
$self->find_slave_hosts($dp, $dbh, $dsn, $args->{method});
foreach my $slave ( @slaves ) {
MKDEBUG && _d('Recursing from',
$dp->as_string($dsn), 'to', $dp->as_string($slave));
$self->recurse_to_slaves(
{ %$args, dsn => $slave, dbh => undef, parent => $dsn }, $level + 1 );
}
}
}
sub find_slave_hosts {
my ( $self, $dsn_parser, $dbh, $dsn, $method ) = @_;
my @methods = qw(processlist hosts);
if ( $method ) {
@methods = grep { $_ ne $method } @methods;
unshift @methods, $method;
}
else {
if ( ($dsn->{P} || 3306) != 3306 ) {
MKDEBUG && _d('Port number is non-standard; using only hosts method');
@methods = qw(hosts);
}
}
MKDEBUG && _d('Looking for slaves on', $dsn_parser->as_string($dsn),
'using methods', @methods);
my @slaves;
METHOD:
foreach my $method ( @methods ) {
my $find_slaves = "_find_slaves_by_$method";
MKDEBUG && _d('Finding slaves with', $find_slaves);
@slaves = $self->$find_slaves($dsn_parser, $dbh, $dsn);
last METHOD if @slaves;
}
MKDEBUG && _d('Found', scalar(@slaves), 'slaves');
return @slaves;
}
sub _find_slaves_by_processlist {
my ( $self, $dsn_parser, $dbh, $dsn ) = @_;
my @slaves = map {
my $slave = $dsn_parser->parse("h=$_", $dsn);
$slave->{source} = 'processlist';
$slave;
}
grep { $_ }
map {
my ( $host ) = $_->{host} =~ m/^([^:]+):/;
if ( $host eq 'localhost' ) {
$host = '127.0.0.1'; # Replication never uses sockets.
}
$host;
} $self->get_connected_slaves($dbh);
return @slaves;
}
sub _find_slaves_by_hosts {
my ( $self, $dsn_parser, $dbh, $dsn ) = @_;
my @slaves;
my $sql = 'SHOW SLAVE HOSTS';
MKDEBUG && _d($dbh, $sql);
@slaves = @{$dbh->selectall_arrayref($sql, { Slice => {} })};
if ( @slaves ) {
MKDEBUG && _d('Found some SHOW SLAVE HOSTS info');
@slaves = map {
my %hash;
@hash{ map { lc $_ } keys %$_ } = values %$_;
my $spec = "h=$hash{host},P=$hash{port}"
. ( $hash{user} ? ",u=$hash{user}" : '')
. ( $hash{password} ? ",p=$hash{password}" : '');
my $dsn = $dsn_parser->parse($spec, $dsn);
$dsn->{server_id} = $hash{server_id};
$dsn->{master_id} = $hash{master_id};
$dsn->{source} = 'hosts';
$dsn;
} @slaves;
}
return @slaves;
}
sub get_connected_slaves {
my ( $self, $dbh ) = @_;
my $show = "SHOW GRANTS FOR ";
my $user = 'CURRENT_USER()';
my $vp = $self->{VersionParser};
if ( $vp && !$vp->version_ge($dbh, '4.1.2') ) {
$user = $dbh->selectrow_arrayref('SELECT USER()')->[0];
$user =~ s/([^@]+)@(.+)/'$1'\@'$2'/;
}
my $sql = $show . $user;
MKDEBUG && _d($dbh, $sql);
my $proc;
eval {
$proc = grep {
m/ALL PRIVILEGES.*?\*\.\*|PROCESS/
} @{$dbh->selectcol_arrayref($sql)};
};
if ( $EVAL_ERROR ) {
if ( $EVAL_ERROR =~ m/no such grant defined for user/ ) {
MKDEBUG && _d('Retrying SHOW GRANTS without host; error:',
$EVAL_ERROR);
($user) = split('@', $user);
$sql = $show . $user;
MKDEBUG && _d($sql);
eval {
$proc = grep {
m/ALL PRIVILEGES.*?\*\.\*|PROCESS/
} @{$dbh->selectcol_arrayref($sql)};
};
}
die "Failed to $sql: $EVAL_ERROR" if $EVAL_ERROR;
}
if ( !$proc ) {
die "You do not have the PROCESS privilege";
}
$sql = 'SHOW PROCESSLIST';
MKDEBUG && _d($dbh, $sql);
grep { $_->{command} =~ m/Binlog Dump/i }
map { # Lowercase the column names
my %hash;
@hash{ map { lc $_ } keys %$_ } = values %$_;
\%hash;
}
@{$dbh->selectall_arrayref($sql, { Slice => {} })};
}
sub is_master_of {
my ( $self, $master, $slave ) = @_;
my $master_status = $self->get_master_status($master)
or die "The server specified as a master is not a master";
my $slave_status = $self->get_slave_status($slave)
or die "The server specified as a slave is not a slave";
my @connected = $self->get_connected_slaves($master)
or die "The server specified as a master has no connected slaves";
my (undef, $port) = $master->selectrow_array('SHOW VARIABLES LIKE "port"');
if ( $port != $slave_status->{master_port} ) {
die "The slave is connected to $slave_status->{master_port} "
. "but the master's port is $port";
}
if ( !grep { $slave_status->{master_user} eq $_->{user} } @connected ) {
die "I don't see any slave I/O thread connected with user "
. $slave_status->{master_user};
}
if ( ($slave_status->{slave_io_state} || '')
eq 'Waiting for master to send event' )
{
my ( $master_log_name, $master_log_num )
= $master_status->{file} =~ m/^(.*?)\.0*([1-9][0-9]*)$/;
my ( $slave_log_name, $slave_log_num )
= $slave_status->{master_log_file} =~ m/^(.*?)\.0*([1-9][0-9]*)$/;
if ( $master_log_name ne $slave_log_name
|| abs($master_log_num - $slave_log_num) > 1 )
{
die "The slave thinks it is reading from "
. "$slave_status->{master_log_file}, but the "
. "master is writing to $master_status->{file}";
}
}
return 1;
}
sub get_master_dsn {
my ( $self, $dbh, $dsn, $dsn_parser ) = @_;
my $master = $self->get_slave_status($dbh) or return undef;
my $spec = "h=$master->{master_host},P=$master->{master_port}";
return $dsn_parser->parse($spec, $dsn);
}
sub get_slave_status {
my ( $self, $dbh ) = @_;
if ( !$self->{not_a_slave}->{$dbh} ) {
my $sth = $self->{sths}->{$dbh}->{SLAVE_STATUS}
||= $dbh->prepare('SHOW SLAVE STATUS');
MKDEBUG && _d($dbh, 'SHOW SLAVE STATUS');
$sth->execute();
my ($ss) = @{$sth->fetchall_arrayref({})};
if ( $ss && %$ss ) {
$ss = { map { lc($_) => $ss->{$_} } keys %$ss }; # lowercase the keys
return $ss;
}
MKDEBUG && _d('This server returns nothing for SHOW SLAVE STATUS');
$self->{not_a_slave}->{$dbh}++;
}
}
sub get_master_status {
my ( $self, $dbh ) = @_;
if ( $self->{not_a_master}->{$dbh} ) {
MKDEBUG && _d('Server on dbh', $dbh, 'is not a master');
return;
}
my $sth = $self->{sths}->{$dbh}->{MASTER_STATUS}
||= $dbh->prepare('SHOW MASTER STATUS');
MKDEBUG && _d($dbh, 'SHOW MASTER STATUS');
$sth->execute();
my ($ms) = @{$sth->fetchall_arrayref({})};
MKDEBUG && _d(
$ms ? map { "$_=" . (defined $ms->{$_} ? $ms->{$_} : '') } keys %$ms
: '');
if ( !$ms || scalar keys %$ms < 2 ) {
MKDEBUG && _d('Server on dbh', $dbh, 'does not seem to be a master');
$self->{not_a_master}->{$dbh}++;
}
return { map { lc($_) => $ms->{$_} } keys %$ms }; # lowercase the keys
}
sub wait_for_master {
my ( $self, %args ) = @_;
my @required_args = qw(master_status slave_dbh);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($master_status, $slave_dbh) = @args{@required_args};
my $timeout = $args{timeout} || 60;
my $result;
my $waited;
if ( $master_status ) {
my $sql = "SELECT MASTER_POS_WAIT('$master_status->{file}', "
. "$master_status->{position}, $timeout)";
MKDEBUG && _d($slave_dbh, $sql);
my $start = time;
($result) = $slave_dbh->selectrow_array($sql);
$waited = time - $start;
MKDEBUG && _d('Result of waiting:', $result);
MKDEBUG && _d("Waited", $waited, "seconds");
}
else {
MKDEBUG && _d('Not waiting: this server is not a master');
}
return {
result => $result,
waited => $waited,
};
}
sub stop_slave {
my ( $self, $dbh ) = @_;
my $sth = $self->{sths}->{$dbh}->{STOP_SLAVE}
||= $dbh->prepare('STOP SLAVE');
MKDEBUG && _d($dbh, $sth->{Statement});
$sth->execute();
}
sub start_slave {
my ( $self, $dbh, $pos ) = @_;
if ( $pos ) {
my $sql = "START SLAVE UNTIL MASTER_LOG_FILE='$pos->{file}', "
. "MASTER_LOG_POS=$pos->{position}";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
else {
my $sth = $self->{sths}->{$dbh}->{START_SLAVE}
||= $dbh->prepare('START SLAVE');
MKDEBUG && _d($dbh, $sth->{Statement});
$sth->execute();
}
}
sub catchup_to_master {
my ( $self, $slave, $master, $timeout ) = @_;
$self->stop_slave($master);
$self->stop_slave($slave);
my $slave_status = $self->get_slave_status($slave);
my $slave_pos = $self->repl_posn($slave_status);
my $master_status = $self->get_master_status($master);
my $master_pos = $self->repl_posn($master_status);
MKDEBUG && _d('Master position:', $self->pos_to_string($master_pos),
'Slave position:', $self->pos_to_string($slave_pos));
my $result;
if ( $self->pos_cmp($slave_pos, $master_pos) < 0 ) {
MKDEBUG && _d('Waiting for slave to catch up to master');
$self->start_slave($slave, $master_pos);
$result = $self->wait_for_master(
master_status => $master_status,
slave_dbh => $slave,
timeout => $timeout,
master_status => $master_status
);
if ( !defined $result->{result} ) {
$slave_status = $self->get_slave_status($slave);
if ( !$self->slave_is_running($slave_status) ) {
MKDEBUG && _d('Master position:',
$self->pos_to_string($master_pos),
'Slave position:', $self->pos_to_string($slave_pos));
$slave_pos = $self->repl_posn($slave_status);
if ( $self->pos_cmp($slave_pos, $master_pos) != 0 ) {
die "MASTER_POS_WAIT() returned NULL but slave has not "
. "caught up to master";
}
MKDEBUG && _d('Slave is caught up to master and stopped');
}
else {
die "Slave has not caught up to master and it is still running";
}
}
}
else {
MKDEBUG && _d("Slave is already caught up to master");
}
return $result;
}
sub catchup_to_same_pos {
my ( $self, $s1_dbh, $s2_dbh ) = @_;
$self->stop_slave($s1_dbh);
$self->stop_slave($s2_dbh);
my $s1_status = $self->get_slave_status($s1_dbh);
my $s2_status = $self->get_slave_status($s2_dbh);
my $s1_pos = $self->repl_posn($s1_status);
my $s2_pos = $self->repl_posn($s2_status);
if ( $self->pos_cmp($s1_pos, $s2_pos) < 0 ) {
$self->start_slave($s1_dbh, $s2_pos);
}
elsif ( $self->pos_cmp($s2_pos, $s1_pos) < 0 ) {
$self->start_slave($s2_dbh, $s1_pos);
}
$s1_status = $self->get_slave_status($s1_dbh);
$s2_status = $self->get_slave_status($s2_dbh);
$s1_pos = $self->repl_posn($s1_status);
$s2_pos = $self->repl_posn($s2_status);
if ( $self->slave_is_running($s1_status)
|| $self->slave_is_running($s2_status)
|| $self->pos_cmp($s1_pos, $s2_pos) != 0)
{
die "The servers aren't both stopped at the same position";
}
}
sub slave_is_running {
my ( $self, $slave_status ) = @_;
return ($slave_status->{slave_sql_running} || 'No') eq 'Yes';
}
sub has_slave_updates {
my ( $self, $dbh ) = @_;
my $sql = q{SHOW VARIABLES LIKE 'log_slave_updates'};
MKDEBUG && _d($dbh, $sql);
my ($name, $value) = $dbh->selectrow_array($sql);
return $value && $value =~ m/^(1|ON)$/;
}
sub repl_posn {
my ( $self, $status ) = @_;
if ( exists $status->{file} && exists $status->{position} ) {
return {
file => $status->{file},
position => $status->{position},
};
}
else {
return {
file => $status->{relay_master_log_file},
position => $status->{exec_master_log_pos},
};
}
}
sub get_slave_lag {
my ( $self, $dbh ) = @_;
my $stat = $self->get_slave_status($dbh);
return unless $stat; # server is not a slave
return $stat->{seconds_behind_master};
}
sub pos_cmp {
my ( $self, $a, $b ) = @_;
return $self->pos_to_string($a) cmp $self->pos_to_string($b);
}
sub short_host {
my ( $self, $dsn ) = @_;
my ($host, $port);
if ( $dsn->{master_host} ) {
$host = $dsn->{master_host};
$port = $dsn->{master_port};
}
else {
$host = $dsn->{h};
$port = $dsn->{P};
}
return ($host || '[default]') . ( ($port || 3306) == 3306 ? '' : ":$port" );
}
sub is_replication_thread {
my ( $self, $query, %args ) = @_;
return unless $query;
my $type = lc($args{type} || 'all');
die "Invalid type: $type"
unless $type =~ m/^binlog_dump|slave_io|slave_sql|all$/i;
my $match = 0;
if ( $type =~ m/binlog_dump|all/i ) {
$match = 1
if ($query->{Command} || $query->{command} || '') eq "Binlog Dump";
}
if ( !$match ) {
if ( ($query->{User} || $query->{user} || '') eq "system user" ) {
MKDEBUG && _d("Slave replication thread");
if ( $type ne 'all' ) {
my $state = $query->{State} || $query->{state} || '';
if ( $state =~ m/^init|end$/ ) {
MKDEBUG && _d("Special state:", $state);
$match = 1;
}
else {
my ($slave_sql) = $state =~ m/
^(Waiting\sfor\sthe\snext\sevent
|Reading\sevent\sfrom\sthe\srelay\slog
|Has\sread\sall\srelay\slog;\swaiting
|Making\stemp\sfile
|Waiting\sfor\sslave\smutex\son\sexit)/xi;
$match = $type eq 'slave_sql' && $slave_sql ? 1
: $type eq 'slave_io' && !$slave_sql ? 1
: 0;
}
}
else {
$match = 1;
}
}
else {
MKDEBUG && _d('Not system user');
}
if ( !defined $args{check_known_ids} || $args{check_known_ids} ) {
my $id = $query->{Id} || $query->{id};
if ( $match ) {
$self->{replication_thread}->{$id} = 1;
}
else {
if ( $self->{replication_thread}->{$id} ) {
MKDEBUG && _d("Thread ID is a known replication thread ID");
$match = 1;
}
}
}
}
MKDEBUG && _d('Matches', $type, 'replication thread:',
($match ? 'yes' : 'no'), '; match:', $match);
return $match;
}
sub get_replication_filters {
my ( $self, %args ) = @_;
my @required_args = qw(dbh);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh) = @args{@required_args};
my %filters = ();
my $status = $self->get_master_status($dbh);
if ( $status ) {
map { $filters{$_} = $status->{$_} }
grep { defined $status->{$_} && $status->{$_} ne '' }
qw(
binlog_do_db
binlog_ignore_db
);
}
$status = $self->get_slave_status($dbh);
if ( $status ) {
map { $filters{$_} = $status->{$_} }
grep { defined $status->{$_} && $status->{$_} ne '' }
qw(
replicate_do_db
replicate_ignore_db
replicate_do_table
replicate_ignore_table
replicate_wild_do_table
replicate_wild_ignore_table
);
my $sql = "SHOW VARIABLES LIKE 'slave_skip_errors'";
MKDEBUG && _d($dbh, $sql);
my $row = $dbh->selectrow_arrayref($sql);
$filters{slave_skip_errors} = $row->[1] if $row->[1] && $row->[1] ne 'OFF';
}
return \%filters;
}
sub pos_to_string {
my ( $self, $pos ) = @_;
my $fmt = '%s/%020d';
return sprintf($fmt, @{$pos}{qw(file position)});
}
sub reset_known_replication_threads {
my ( $self ) = @_;
$self->{replication_thread} = {};
return;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End MasterSlave package
# ###########################################################################
# ###########################################################################
# Daemon package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/Daemon.pm
# t/lib/Daemon.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package Daemon;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use POSIX qw(setsid);
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(o) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $o = $args{o};
my $self = {
o => $o,
log_file => $o->has('log') ? $o->get('log') : undef,
PID_file => $o->has('pid') ? $o->get('pid') : undef,
};
check_PID_file(undef, $self->{PID_file});
MKDEBUG && _d('Daemonized child will log to', $self->{log_file});
return bless $self, $class;
}
sub daemonize {
my ( $self ) = @_;
MKDEBUG && _d('About to fork and daemonize');
defined (my $pid = fork()) or die "Cannot fork: $OS_ERROR";
if ( $pid ) {
MKDEBUG && _d('I am the parent and now I die');
exit;
}
$self->{PID_owner} = $PID;
$self->{child} = 1;
POSIX::setsid() or die "Cannot start a new session: $OS_ERROR";
chdir '/' or die "Cannot chdir to /: $OS_ERROR";
$self->_make_PID_file();
$OUTPUT_AUTOFLUSH = 1;
if ( -t STDIN ) {
close STDIN;
open STDIN, '/dev/null'
or die "Cannot reopen STDIN to /dev/null: $OS_ERROR";
}
if ( $self->{log_file} ) {
close STDOUT;
open STDOUT, '>>', $self->{log_file}
or die "Cannot open log file $self->{log_file}: $OS_ERROR";
close STDERR;
open STDERR, ">&STDOUT"
or die "Cannot dupe STDERR to STDOUT: $OS_ERROR";
}
else {
if ( -t STDOUT ) {
close STDOUT;
open STDOUT, '>', '/dev/null'
or die "Cannot reopen STDOUT to /dev/null: $OS_ERROR";
}
if ( -t STDERR ) {
close STDERR;
open STDERR, '>', '/dev/null'
or die "Cannot reopen STDERR to /dev/null: $OS_ERROR";
}
}
MKDEBUG && _d('I am the child and now I live daemonized');
return;
}
sub check_PID_file {
my ( $self, $file ) = @_;
my $PID_file = $self ? $self->{PID_file} : $file;
MKDEBUG && _d('Checking PID file', $PID_file);
if ( $PID_file && -f $PID_file ) {
my $pid;
eval { chomp($pid = `cat $PID_file`); };
die "Cannot cat $PID_file: $OS_ERROR" if $EVAL_ERROR;
MKDEBUG && _d('PID file exists; it contains PID', $pid);
if ( $pid ) {
my $pid_is_alive = kill 0, $pid;
if ( $pid_is_alive ) {
die "The PID file $PID_file already exists "
. " and the PID that it contains, $pid, is running";
}
else {
warn "Overwriting PID file $PID_file because the PID that it "
. "contains, $pid, is not running";
}
}
else {
die "The PID file $PID_file already exists but it does not "
. "contain a PID";
}
}
else {
MKDEBUG && _d('No PID file');
}
return;
}
sub make_PID_file {
my ( $self ) = @_;
if ( exists $self->{child} ) {
die "Do not call Daemon::make_PID_file() for daemonized scripts";
}
$self->_make_PID_file();
$self->{PID_owner} = $PID;
return;
}
sub _make_PID_file {
my ( $self ) = @_;
my $PID_file = $self->{PID_file};
if ( !$PID_file ) {
MKDEBUG && _d('No PID file to create');
return;
}
$self->check_PID_file();
open my $PID_FH, '>', $PID_file
or die "Cannot open PID file $PID_file: $OS_ERROR";
print $PID_FH $PID
or die "Cannot print to PID file $PID_file: $OS_ERROR";
close $PID_FH
or die "Cannot close PID file $PID_file: $OS_ERROR";
MKDEBUG && _d('Created PID file:', $self->{PID_file});
return;
}
sub _remove_PID_file {
my ( $self ) = @_;
if ( $self->{PID_file} && -f $self->{PID_file} ) {
unlink $self->{PID_file}
or warn "Cannot remove PID file $self->{PID_file}: $OS_ERROR";
MKDEBUG && _d('Removed PID file');
}
else {
MKDEBUG && _d('No PID to remove');
}
return;
}
sub DESTROY {
my ( $self ) = @_;
$self->_remove_PID_file() if ($self->{PID_owner} || 0) == $PID;
return;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End Daemon package
# ###########################################################################
# ###########################################################################
# SchemaIterator r7141
# Don't update this package!
# ###########################################################################
package SchemaIterator;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use Data::Dumper;
$Data::Dumper::Indent = 1;
$Data::Dumper::Sortkeys = 1;
$Data::Dumper::Quotekeys = 0;
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
foreach my $arg ( qw(Quoter) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $self = {
%args,
filter => undef,
dbs => [],
};
return bless $self, $class;
}
sub make_filter {
my ( $self, $o ) = @_;
my @lines = (
'sub {',
' my ( $dbh, $db, $tbl ) = @_;',
' my $engine = undef;',
);
my @permit_dbs = _make_filter('unless', '$db', $o->get('databases'))
if $o->has('databases');
my @reject_dbs = _make_filter('if', '$db', $o->get('ignore-databases'))
if $o->has('ignore-databases');
my @dbs_regex;
if ( $o->has('databases-regex') && (my $p = $o->get('databases-regex')) ) {
push @dbs_regex, " return 0 unless \$db && (\$db =~ m/$p/o);";
}
my @reject_dbs_regex;
if ( $o->has('ignore-databases-regex')
&& (my $p = $o->get('ignore-databases-regex')) ) {
push @reject_dbs_regex, " return 0 if \$db && (\$db =~ m/$p/o);";
}
if ( @permit_dbs || @reject_dbs || @dbs_regex || @reject_dbs_regex ) {
push @lines,
' if ( $db ) {',
(@permit_dbs ? @permit_dbs : ()),
(@reject_dbs ? @reject_dbs : ()),
(@dbs_regex ? @dbs_regex : ()),
(@reject_dbs_regex ? @reject_dbs_regex : ()),
' }';
}
if ( $o->has('tables') || $o->has('ignore-tables')
|| $o->has('ignore-tables-regex') ) {
my $have_qtbl = 0;
my $have_only_qtbls = 0;
my %qtbls;
my @permit_tbls;
my @permit_qtbls;
my %permit_qtbls;
if ( $o->get('tables') ) {
my %tbls;
map {
if ( $_ =~ m/\./ ) {
$permit_qtbls{$_} = 1;
}
else {
$tbls{$_} = 1;
}
} keys %{ $o->get('tables') };
@permit_tbls = _make_filter('unless', '$tbl', \%tbls);
@permit_qtbls = _make_filter('unless', '$qtbl', \%permit_qtbls);
if ( @permit_qtbls ) {
push @lines,
' my $qtbl = ($db ? "$db." : "") . ($tbl ? $tbl : "");';
$have_qtbl = 1;
}
}
my @reject_tbls;
my @reject_qtbls;
my %reject_qtbls;
if ( $o->get('ignore-tables') ) {
my %tbls;
map {
if ( $_ =~ m/\./ ) {
$reject_qtbls{$_} = 1;
}
else {
$tbls{$_} = 1;
}
} keys %{ $o->get('ignore-tables') };
@reject_tbls= _make_filter('if', '$tbl', \%tbls);
@reject_qtbls = _make_filter('if', '$qtbl', \%reject_qtbls);
if ( @reject_qtbls && !$have_qtbl ) {
push @lines,
' my $qtbl = ($db ? "$db." : "") . ($tbl ? $tbl : "");';
}
}
if ( keys %permit_qtbls && !@permit_dbs ) {
my $dbs = {};
map {
my ($db, undef) = split(/\./, $_);
$dbs->{$db} = 1;
} keys %permit_qtbls;
MKDEBUG && _d('Adding restriction "--databases',
(join(',', keys %$dbs) . '"'));
if ( keys %$dbs ) {
$o->set('databases', $dbs);
return $self->make_filter($o);
}
}
my @tbls_regex;
if ( $o->has('tables-regex') && (my $p = $o->get('tables-regex')) ) {
push @tbls_regex, " return 0 unless \$tbl && (\$tbl =~ m/$p/o);";
}
my @reject_tbls_regex;
if ( $o->has('ignore-tables-regex')
&& (my $p = $o->get('ignore-tables-regex')) ) {
push @reject_tbls_regex,
" return 0 if \$tbl && (\$tbl =~ m/$p/o);";
}
my @get_eng;
my @permit_engs;
my @reject_engs;
if ( ($o->has('engines') && $o->get('engines'))
|| ($o->has('ignore-engines') && $o->get('ignore-engines')) ) {
push @get_eng,
' my $sql = "SHOW TABLE STATUS "',
' . ($db ? "FROM `$db`" : "")',
' . " LIKE \'$tbl\'";',
' MKDEBUG && _d($sql);',
' eval {',
' $engine = $dbh->selectrow_hashref($sql)->{engine};',
' };',
' MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);',
' MKDEBUG && _d($tbl, "uses engine", $engine);',
' $engine = lc $engine if $engine;',
@permit_engs
= _make_filter('unless', '$engine', $o->get('engines'), 1);
@reject_engs
= _make_filter('if', '$engine', $o->get('ignore-engines'), 1)
}
if ( @permit_tbls || @permit_qtbls || @reject_tbls || @tbls_regex
|| @reject_tbls_regex || @permit_engs || @reject_engs ) {
push @lines,
' if ( $tbl ) {',
(@permit_tbls ? @permit_tbls : ()),
(@reject_tbls ? @reject_tbls : ()),
(@tbls_regex ? @tbls_regex : ()),
(@reject_tbls_regex ? @reject_tbls_regex : ()),
(@permit_qtbls ? @permit_qtbls : ()),
(@reject_qtbls ? @reject_qtbls : ()),
(@get_eng ? @get_eng : ()),
(@permit_engs ? @permit_engs : ()),
(@reject_engs ? @reject_engs : ()),
' }';
}
}
push @lines,
' MKDEBUG && _d(\'Passes filters:\', $db, $tbl, $engine, $dbh);',
' return 1;', '}';
my $code = join("\n", @lines);
MKDEBUG && _d('filter sub:', $code);
my $filter_sub= eval $code
or die "Error compiling subroutine code:\n$code\n$EVAL_ERROR";
return $filter_sub;
}
sub set_filter {
my ( $self, $filter_sub ) = @_;
$self->{filter} = $filter_sub;
MKDEBUG && _d('Set filter sub');
return;
}
sub get_db_itr {
my ( $self, %args ) = @_;
my @required_args = qw(dbh);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh) = @args{@required_args};
my $filter = $self->{filter};
my @dbs;
eval {
my $sql = 'SHOW DATABASES';
MKDEBUG && _d($sql);
@dbs = grep {
my $ok = $filter ? $filter->($dbh, $_, undef) : 1;
$ok = 0 if $_ =~ m/information_schema|performance_schema|lost\+found/;
$ok;
} @{ $dbh->selectcol_arrayref($sql) };
MKDEBUG && _d('Found', scalar @dbs, 'databases');
};
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
my $iterator = sub {
return shift @dbs;
};
if (wantarray) {
return ($iterator, scalar @dbs);
}
else {
return $iterator;
}
}
sub get_tbl_itr {
my ( $self, %args ) = @_;
my @required_args = qw(dbh db);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($dbh, $db, $views) = @args{@required_args, 'views'};
my $filter = $self->{filter};
my @tbls;
if ( $db ) {
eval {
my $sql = 'SHOW /*!50002 FULL*/ TABLES FROM '
. $self->{Quoter}->quote($db);
MKDEBUG && _d($sql);
@tbls = map {
$_->[0]
}
grep {
my ($tbl, $type) = @$_;
my $ok = $filter ? $filter->($dbh, $db, $tbl) : 1;
if ( !$views ) {
$ok = 0 if ($type || '') eq 'VIEW';
}
$ok;
}
@{ $dbh->selectall_arrayref($sql) };
MKDEBUG && _d('Found', scalar @tbls, 'tables in', $db);
};
MKDEBUG && $EVAL_ERROR && _d($EVAL_ERROR);
}
else {
MKDEBUG && _d('No db given so no tables');
}
my $iterator = sub {
return shift @tbls;
};
if ( wantarray ) {
return ($iterator, scalar @tbls);
}
else {
return $iterator;
}
}
sub _make_filter {
my ( $cond, $var_name, $objs, $lc ) = @_;
my @lines;
if ( scalar keys %$objs ) {
my $test = join(' || ',
map { "$var_name eq '" . ($lc ? lc $_ : $_) ."'" } keys %$objs);
push @lines, " return 0 $cond $var_name && ($test);",
}
return @lines;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
# ###########################################################################
# End SchemaIterator package
# ###########################################################################
# ###########################################################################
# Transformers package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/Transformers.pm
# t/lib/Transformers.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package Transformers;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
use Time::Local qw(timegm timelocal);
use Digest::MD5 qw(md5_hex);
require Exporter;
our @ISA = qw(Exporter);
our %EXPORT_TAGS = ();
our @EXPORT = ();
our @EXPORT_OK = qw(
micro_t
percentage_of
secs_to_time
time_to_secs
shorten
ts
parse_timestamp
unix_timestamp
any_unix_timestamp
make_checksum
crc32
);
our $mysql_ts = qr/(\d\d)(\d\d)(\d\d) +(\d+):(\d+):(\d+)(\.\d+)?/;
our $proper_ts = qr/(\d\d\d\d)-(\d\d)-(\d\d)[T ](\d\d):(\d\d):(\d\d)(\.\d+)?/;
our $n_ts = qr/(\d{1,5})([shmd]?)/; # Limit \d{1,5} because \d{6} looks
sub micro_t {
my ( $t, %args ) = @_;
my $p_ms = defined $args{p_ms} ? $args{p_ms} : 0; # precision for ms vals
my $p_s = defined $args{p_s} ? $args{p_s} : 0; # precision for s vals
my $f;
$t = 0 if $t < 0;
$t = sprintf('%.17f', $t) if $t =~ /e/;
$t =~ s/\.(\d{1,6})\d*/\.$1/;
if ($t > 0 && $t <= 0.000999) {
$f = ($t * 1000000) . 'us';
}
elsif ($t >= 0.001000 && $t <= 0.999999) {
$f = sprintf("%.${p_ms}f", $t * 1000);
$f = ($f * 1) . 'ms'; # * 1 to remove insignificant zeros
}
elsif ($t >= 1) {
$f = sprintf("%.${p_s}f", $t);
$f = ($f * 1) . 's'; # * 1 to remove insignificant zeros
}
else {
$f = 0; # $t should = 0 at this point
}
return $f;
}
sub percentage_of {
my ( $is, $of, %args ) = @_;
my $p = $args{p} || 0; # float precision
my $fmt = $p ? "%.${p}f" : "%d";
return sprintf $fmt, ($is * 100) / ($of ||= 1);
}
sub secs_to_time {
my ( $secs, $fmt ) = @_;
$secs ||= 0;
return '00:00' unless $secs;
$fmt ||= $secs >= 86_400 ? 'd'
: $secs >= 3_600 ? 'h'
: 'm';
return
$fmt eq 'd' ? sprintf(
"%d+%02d:%02d:%02d",
int($secs / 86_400),
int(($secs % 86_400) / 3_600),
int(($secs % 3_600) / 60),
$secs % 60)
: $fmt eq 'h' ? sprintf(
"%02d:%02d:%02d",
int(($secs % 86_400) / 3_600),
int(($secs % 3_600) / 60),
$secs % 60)
: sprintf(
"%02d:%02d",
int(($secs % 3_600) / 60),
$secs % 60);
}
sub time_to_secs {
my ( $val, $default_suffix ) = @_;
die "I need a val argument" unless defined $val;
my $t = 0;
my ( $prefix, $num, $suffix ) = $val =~ m/([+-]?)(\d+)([a-z])?$/;
$suffix = $suffix || $default_suffix || 's';
if ( $suffix =~ m/[smhd]/ ) {
$t = $suffix eq 's' ? $num * 1 # Seconds
: $suffix eq 'm' ? $num * 60 # Minutes
: $suffix eq 'h' ? $num * 3600 # Hours
: $num * 86400; # Days
$t *= -1 if $prefix && $prefix eq '-';
}
else {
die "Invalid suffix for $val: $suffix";
}
return $t;
}
sub shorten {
my ( $num, %args ) = @_;
my $p = defined $args{p} ? $args{p} : 2; # float precision
my $d = defined $args{d} ? $args{d} : 1_024; # divisor
my $n = 0;
my @units = ('', qw(k M G T P E Z Y));
while ( $num >= $d && $n < @units - 1 ) {
$num /= $d;
++$n;
}
return sprintf(
$num =~ m/\./ || $n
? "%.${p}f%s"
: '%d',
$num, $units[$n]);
}
sub ts {
my ( $time, $gmt ) = @_;
my ( $sec, $min, $hour, $mday, $mon, $year )
= $gmt ? gmtime($time) : localtime($time);
$mon += 1;
$year += 1900;
my $val = sprintf("%d-%02d-%02dT%02d:%02d:%02d",
$year, $mon, $mday, $hour, $min, $sec);
if ( my ($us) = $time =~ m/(\.\d+)$/ ) {
$us = sprintf("%.6f", $us);
$us =~ s/^0\././;
$val .= $us;
}
return $val;
}
sub parse_timestamp {
my ( $val ) = @_;
if ( my($y, $m, $d, $h, $i, $s, $f)
= $val =~ m/^$mysql_ts$/ )
{
return sprintf "%d-%02d-%02d %02d:%02d:"
. (defined $f ? '%09.6f' : '%02d'),
$y + 2000, $m, $d, $h, $i, (defined $f ? $s + $f : $s);
}
return $val;
}
sub unix_timestamp {
my ( $val, $gmt ) = @_;
if ( my($y, $m, $d, $h, $i, $s, $us) = $val =~ m/^$proper_ts$/ ) {
$val = $gmt
? timegm($s, $i, $h, $d, $m - 1, $y)
: timelocal($s, $i, $h, $d, $m - 1, $y);
if ( defined $us ) {
$us = sprintf('%.6f', $us);
$us =~ s/^0\././;
$val .= $us;
}
}
return $val;
}
sub any_unix_timestamp {
my ( $val, $callback ) = @_;
if ( my ($n, $suffix) = $val =~ m/^$n_ts$/ ) {
$n = $suffix eq 's' ? $n # Seconds
: $suffix eq 'm' ? $n * 60 # Minutes
: $suffix eq 'h' ? $n * 3600 # Hours
: $suffix eq 'd' ? $n * 86400 # Days
: $n; # default: Seconds
MKDEBUG && _d('ts is now - N[shmd]:', $n);
return time - $n;
}
elsif ( $val =~ m/^\d{9,}/ ) {
MKDEBUG && _d('ts is already a unix timestamp');
return $val;
}
elsif ( my ($ymd, $hms) = $val =~ m/^(\d{6})(?:\s+(\d+:\d+:\d+))?/ ) {
MKDEBUG && _d('ts is MySQL slow log timestamp');
$val .= ' 00:00:00' unless $hms;
return unix_timestamp(parse_timestamp($val));
}
elsif ( ($ymd, $hms) = $val =~ m/^(\d{4}-\d\d-\d\d)(?:[T ](\d+:\d+:\d+))?/) {
MKDEBUG && _d('ts is properly formatted timestamp');
$val .= ' 00:00:00' unless $hms;
return unix_timestamp($val);
}
else {
MKDEBUG && _d('ts is MySQL expression');
return $callback->($val) if $callback && ref $callback eq 'CODE';
}
MKDEBUG && _d('Unknown ts type:', $val);
return;
}
sub make_checksum {
my ( $val ) = @_;
my $checksum = uc substr(md5_hex($val), -16);
MKDEBUG && _d($checksum, 'checksum for', $val);
return $checksum;
}
sub crc32 {
my ( $string ) = @_;
return unless $string;
my $poly = 0xEDB88320;
my $crc = 0xFFFFFFFF;
foreach my $char ( split(//, $string) ) {
my $comp = ($crc ^ ord($char)) & 0xFF;
for ( 1 .. 8 ) {
$comp = $comp & 1 ? $poly ^ ($comp >> 1) : $comp >> 1;
}
$crc = (($crc >> 8) & 0x00FFFFFF) ^ $comp;
}
return $crc ^ 0xFFFFFFFF;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End Transformers package
# ###########################################################################
# ###########################################################################
# Retry package
# This package is a copy without comments from the original. The original
# with comments and its test file can be found in the Bazaar repository at,
# lib/Retry.pm
# t/lib/Retry.t
# See https://launchpad.net/percona-toolkit for more information.
# ###########################################################################
{
package Retry;
use strict;
use warnings FATAL => 'all';
use English qw(-no_match_vars);
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
sub new {
my ( $class, %args ) = @_;
my $self = {
%args,
};
return bless $self, $class;
}
sub retry {
my ( $self, %args ) = @_;
my @required_args = qw(try wait);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
};
my ($try, $wait) = @args{@required_args};
my $tries = $args{tries} || 3;
my $tryno = 0;
while ( ++$tryno <= $tries ) {
MKDEBUG && _d("Retry", $tryno, "of", $tries);
my $result;
eval {
$result = $try->(tryno=>$tryno);
};
if ( defined $result ) {
MKDEBUG && _d("Try code succeeded");
if ( my $on_success = $args{on_success} ) {
MKDEBUG && _d("Calling on_success code");
$on_success->(tryno=>$tryno, result=>$result);
}
return $result;
}
if ( $EVAL_ERROR ) {
MKDEBUG && _d("Try code died:", $EVAL_ERROR);
die $EVAL_ERROR unless $args{retry_on_die};
}
if ( $tryno < $tries ) {
MKDEBUG && _d("Try code failed, calling wait code");
$wait->(tryno=>$tryno);
}
}
MKDEBUG && _d("Try code did not succeed");
if ( my $on_failure = $args{on_failure} ) {
MKDEBUG && _d("Calling on_failure code");
$on_failure->();
}
return;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
1;
}
# ###########################################################################
# End Retry package
# ###########################################################################
# ###########################################################################
# This is a combination of modules and programs in one -- a runnable module.
# http://www.perl.com/pub/a/2006/07/13/lightning-articles.html?page=last
# Or, look it up in the Camel book on pages 642 and 643 in the 3rd edition.
#
# Check at the end of this package for the call to main() which actually runs
# the program.
# ###########################################################################
package pt_table_sync;
use English qw(-no_match_vars);
use List::Util qw(sum max min);
use POSIX qw(ceil);
Transformers->import(qw(time_to_secs any_unix_timestamp));
use constant MKDEBUG => $ENV{MKDEBUG} || 0;
$OUTPUT_AUTOFLUSH = 1;
my %dsn_for;
sub main {
@ARGV = @_; # set global ARGV for this package
# Reset global vars else tests will have weird results.
%dsn_for = ();
# ########################################################################
# Get configuration information.
# ########################################################################
my $o = new OptionParser();
$o->get_specs();
$o->get_opts();
my $dp = $o->DSNParser();
$dp->prop('set-vars', $o->get('set-vars'));
if ( $o->get('replicate') || $o->get('sync-to-master') ) {
$o->set('wait', 60) unless $o->got('wait');
}
if ( $o->get('wait') ) {
$o->set('lock', 1) unless $o->got('lock');
}
if ( $o->get('dry-run') ) {
$o->set('verbose', 1);
}
# There's a conflict of interests: we added 't' and 'D' parts to dp,
# and there are -t and -D options (--tables, --databases), so parse_options()
# is going to return a DSN with the default values from -t and -D,
# but these are not actually be default dsn vals, they're filters.
# So we have to remove them from $dsn_defaults.
my $dsn_defaults = $dp->parse_options($o);
$dsn_defaults->{D} = undef;
$dsn_defaults->{t} = undef;
my @dsns;
while ( my $arg = shift(@ARGV) ) {
my $dsn = $dp->parse($arg, $dsns[0], $dsn_defaults);
die "You specified a t part, but not a D part in $arg"
if ($dsn->{t} && !$dsn->{D});
if ( $dsn->{D} && !$dsn->{t} ) {
die "You specified a database but not a table in $arg. Are you "
. "trying to sync only tables in the '$dsn->{D}' database? "
. "If so, use '--databases $dsn->{D}' instead.\n";
}
push @dsns, $dsn;
}
if ( !@dsns
|| (@dsns ==1 && !$o->get('replicate') && !$o->get('sync-to-master'))) {
$o->save_error('At least one DSN is required, and at least two are '
. 'required unless --sync-to-master or --replicate is specified');
}
if ( @dsns > 1 && $o->get('sync-to-master') && $o->get('replicate') ) {
$o->save_error('--sync-to-master and --replicate require only one DSN ',
' but ', scalar @dsns, ' where given');
}
if ( $o->get('lock-and-rename') ) {
if ( @dsns != 2 || !$dsns[0]->{t} || !$dsns[1]->{t} ) {
$o->save_error("--lock-and-rename requires exactly two DSNs and they "
. "must each specify a table.");
}
}
if ( $o->get('bidirectional') ) {
if ( $o->get('replicate') || $o->get('sync-to-master') ) {
$o->save_error('--bidirectional does not work with '
. '--replicate or --sync-to-master');
}
if ( @dsns < 2 ) {
$o->save_error('--bidirectional requires at least two DSNs');
}
if ( !$o->get('conflict-column') || !$o->get('conflict-comparison') ) {
$o->save_error('--bidirectional requires --conflict-column '
. 'and --conflict-comparison');
}
my $cc = $o->get('conflict-comparison');
my $cmp = $o->read_para_after(__FILE__, qr/MAGIC_comparisons/);
$cmp =~ s/ //g;
if ( $cc && $cc !~ m/$cmp/ ) {
$o->save_error("--conflict-comparison must be one of $cmp");
}
if ( $cc && $cc =~ m/equals|matches/ && !$o->get('conflict-value') ) {
$o->save_error("--conflict-comparison $cc requires --conflict-value")
}
# Override --algorithms becuase only TableSyncChunk works with
# bidirectional syncing.
$o->set('algorithms', 'Chunk');
$o->set('buffer-to-client', 0);
}
if ( $o->get('explain-hosts') ) {
foreach my $host ( @dsns ) {
print "# DSN: ", $dp->as_string($host), "\n";
}
return 0;
}
$o->usage_or_errors();
# ########################################################################
# If --pid, check it first since we'll die if it already exits.
# ########################################################################
my $daemon;
if ( $o->get('pid') ) {
# We're not daemoninzing, it just handles PID stuff. Keep $daemon
# in the the scope of main() because when it's destroyed it automatically
# removes the PID file.
$daemon = new Daemon(o=>$o);
$daemon->make_PID_file();
}
# ########################################################################
# Do the work.
# ########################################################################
my $q = new Quoter();
my $tp = new TableParser( Quoter => $q );
my $vp = new VersionParser();
my $ms = new MasterSlave(VersionParser => $vp);
my $du = new MySQLDump( cache => 0 );
my $rt = new Retry();
my $chunker = new TableChunker( Quoter => $q, MySQLDump => $du );
my $nibbler = new TableNibbler( Quoter => $q, TableParser => $tp );
my $checksum = new TableChecksum( Quoter => $q, VersionParser => $vp );
my $syncer = new TableSyncer(
Quoter => $q,
VersionParser => $vp,
MasterSlave => $ms,
TableChecksum => $checksum,
DSNParser => $dp,
Retry => $rt,
);
my %modules = (
OptionParser => $o,
DSNParser => $dp,
MySQLDump => $du,
TableParser => $tp,
Quoter => $q,
VersionParser => $vp,
TableChunker => $chunker,
TableNibbler => $nibbler,
TableChecksum => $checksum,
MasterSlave => $ms,
TableSyncer => $syncer,
);
# Create the sync plugins.
my $plugins = [];
my %have_plugin = get_plugins();
foreach my $algo ( split(',', $o->get('algorithms')) ) {
my $plugin_name = $have_plugin{lc $algo};
if ( !$plugin_name ) {
die "The $algo algorithm is not available. Available algorithms: "
. join(", ", sort keys %have_plugin);
}
MKDEBUG && _d('Loading', $plugin_name);
my $plugin;
eval {
$plugin = $plugin_name->new(%modules);
};
die "Error loading $plugin_name for $algo algorithm: $EVAL_ERROR"
if $EVAL_ERROR;
push @$plugins, $plugin;
}
# Create callbacks for bidirectional syncing. Currently, this only
# works with TableSyncChunk, so that should be the only plugin because
# --algorithms was overriden earlier.
if ( $o->get('bidirectional') ) {
set_bidirectional_callbacks(
plugin => $plugins->[0],
%modules,
);
}
my $exit_status = 0; # 1: internal error, 2: tables differed, 3: both
# dsn[0] is expected to be the master (i.e. the source). So if
# --sync-to-master, then dsn[0] is a slave. Find its master and
# make the master dsn[0] and the slave dsn[1].
if ( $o->get('sync-to-master') ) {
MKDEBUG && _d('Getting master of', $dp->as_string($dsns[0]));
$dsns[0]->{dbh} = get_cxn($dsns[0], %modules);
my $master = $ms->get_master_dsn($dsns[0]->{dbh}, $dsns[0], $dp)
or die "Can't determine master of " . $dp->as_string($dsns[0]);
unshift @dsns, $master; # dsn[0]=master, dsn[1]=slave
$dsns[0]->{dbh} = get_cxn($dsns[0], %modules);
if ( $o->get('check-master') ) {
$ms->is_master_of($dsns[0]->{dbh}, $dsns[1]->{dbh});
}
}
my %args = (
dsns => \@dsns,
plugins => $plugins,
%modules,
);
if ( $o->get('dry-run') ) {
print "# NOTE: --dry-run does not show if data needs to be synced because it\n"
. "# does not access, compare or sync data. --dry-run only shows\n"
. "# the work that would be done.\n";
}
if ( $o->get('lock-and-rename') ) {
$exit_status = lock_and_rename(%args);
}
elsif ( $dsns[0]->{t} ) {
$exit_status = sync_one_table(%args);
}
elsif ( $o->get('replicate') ) {
$exit_status = sync_via_replication(%args);
}
else {
$exit_status = sync_all(%args);
}
return $exit_status;
}
# ############################################################################
# Subroutines
# ############################################################################
# Sub: lock_and_rename
# Lock and rename a table.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# dsns - Arrayref of DSNs
# plugins - Arrayref of TableSync* objects
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# Quoter - <Quoter> object
#
# Returns:
# Exit status
sub lock_and_rename {
my ( %args ) = @_;
my @required_args = qw(dsns plugins OptionParser DSNParser Quoter
VersionParser);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my $dsns = $args{dsns};
my $o = $args{OptionParser};
my $dp = $args{DSNParser};
my $q = $args{Quoter};
MKDEBUG && _d('Locking and syncing ONE TABLE with rename');
my $src = {
dsn => $dsns->[0],
dbh => $dsns->[0]->{dbh} || get_cxn($dsns->[0], %args),
misc_dbh => get_cxn($dsns->[0], %args),
db => $dsns->[0]->{D},
tbl => $dsns->[0]->{t},
};
my $dst = {
dsn => $dsns->[1],
dbh => $dsns->[1]->{dbh} || get_cxn($dsns->[1], %args),
misc_dbh => get_cxn($dsns->[1], %args),
db => $dsns->[1]->{D},
tbl => $dsns->[1]->{t},
};
if ( $o->get('verbose') ) {
print_header("# Lock and rename " . $dp->as_string($src->{dsn}));
}
# We don't use lock_server() here because it does the usual stuff wrt
# waiting for slaves to catch up to master, etc, etc.
my $src_db_tbl = $q->quote($src->{db}, $src->{tbl});
my $dst_db_tbl = $q->quote($dst->{db}, $dst->{tbl});
my $tmp_db_tbl = $q->quote($src->{db}, $src->{tbl} . "_tmp_$PID");
my $sql = "LOCK TABLES $src_db_tbl WRITE";
MKDEBUG && _d($sql);
$src->{dbh}->do($sql);
$sql = "LOCK TABLES $dst_db_tbl WRITE";
MKDEBUG && _d($sql);
$dst->{dbh}->do($sql);
my $exit_status = sync_a_table(
src => $src,
dst => $dst,
%args,
);
# Now rename the tables to swap them.
$sql = "ALTER TABLE $src_db_tbl RENAME $tmp_db_tbl";
MKDEBUG && _d($sql);
$src->{dbh}->do($sql);
$sql = "ALTER TABLE $dst_db_tbl RENAME $src_db_tbl";
MKDEBUG && _d($sql);
$dst->{dbh}->do($sql);
$sql = "UNLOCK TABLES";
MKDEBUG && _d($sql);
$src->{dbh}->do($sql);
$sql = "ALTER TABLE $tmp_db_tbl RENAME $dst_db_tbl";
MKDEBUG && _d($sql);
$src->{dbh}->do($sql);
unlock_server(src => $src, dst => $dst, %args);
disconnect($src, $dst);
return $exit_status;
}
# Sub: sync_one_table
# Sync one table between one source host and multiple destination hosts.
# The first DSN in $args{dsns} specifies the source host, database (D),
# and table (t). The other DSNs are the destination hosts. If a destination
# DSN does not specify a database or table, the source database or table
# are used as defaults. Else, the destination-specific database or table
# are used. This allows you to sync tables with different names.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# dsns - Arrayref of DSNs
# plugins - Arrayref of TableSync* objects
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# Quoter - <Quoter> object
# VersionParser - <VersionParser> object
#
# Returns:
# Exit status
sub sync_one_table {
my ( %args ) = @_;
my @required_args = qw(dsns plugins OptionParser DSNParser Quoter
VersionParser);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my @dsns = @{$args{dsns}};
my $o = $args{OptionParser};
my $dp = $args{DSNParser};
MKDEBUG && _d('DSN has t part; syncing ONE TABLE between servers');
my $src = {
dsn => $dsns[0],
dbh => $dsns[0]->{dbh} || get_cxn($dsns[0], %args),
misc_dbh => get_cxn($dsns[0], %args),
db => $dsns[0]->{D},
tbl => $dsns[0]->{t},
};
my $exit_status = 0;
foreach my $dsn ( @dsns[1 .. $#dsns] ) {
my $dst = {
dsn => $dsn,
dbh => $dsn->{dbh} || get_cxn($dsn, %args),
misc_dbh => get_cxn($dsn, %args),
db => $dsn->{D} || $src->{db},
tbl => $dsn->{t} || $src->{tbl},
};
if ( $o->get('verbose') ) {
print_header("# Syncing " . $dp->as_string($dsn)
. ($o->get('dry-run')
? ' in dry-run mode, without accessing or comparing data'
: ''));
}
lock_server(src => $src, dst => $dst, %args);
$exit_status |= sync_a_table(
src => $src,
dst => $dst,
%args,
);
unlock_server(src => $src, dst => $dst, %args);
disconnect($dst);
}
disconnect($src);
return $exit_status;
}
# Sub: sync_via_replication
# Sync multiple destination hosts to one source host via replication.
# The first DSN in $args{dsns} specifies the source host.
# If --sync-to-master is specified, then the source host is a master
# and there is only one destination host which is its slave.
# Else, destination hosts are auto-discovered with
# <MasterSlave::recurse_to_slaves()>.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# dsns - Arrayref of DSNs
# plugins - Arrayref of TableSync* objects
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# Quoter - <Quoter> object
# VersionParser - <VersionParser> object
# TableChecksum - <TableChecksum> object
# MasterSlave - <MasterSlave> object
#
# Returns:
# Exit status
#
# See Also:
# <filter_diffs()>
sub sync_via_replication {
my ( %args ) = @_;
my @required_args = qw(dsns plugins OptionParser DSNParser Quoter
VersionParser TableChecksum MasterSlave);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my $dsns = $args{dsns};
my $o = $args{OptionParser};
my $dp = $args{DSNParser};
my $q = $args{Quoter};
my $checksum = $args{TableChecksum};
my $ms = $args{MasterSlave};
MKDEBUG && _d('Syncing via replication');
my $src = {
dsn => $dsns->[0],
dbh => $dsns->[0]->{dbh} || get_cxn($dsns->[0], %args),
misc_dbh => get_cxn($dsns->[0], %args),
db => undef, # set later
tbl => undef, # set later
};
# Filters for --databases and --tables. We have to do these manually
# since we don't use MySQLFind for --replicate.
my $databases = $o->get('databases');
my $tables = $o->get('tables');
my $exit_status = 0;
# Connect to the master and treat it as the source, then find
# differences on the slave and sync them.
if ( $o->get('sync-to-master') ) {
my $dst = {
dsn => $dsns->[1],
dbh => $dsns->[1]->{dbh} || get_cxn($dsns->[1], %args),
misc_dbh => get_cxn($dsns->[1], %args),
db => undef, # set later
tbl => undef, # set later
};
# First, check that the master (source) has no discrepancies itself,
# and ignore tables that do.
my %skip_table;
map { $skip_table{$_->{db}}->{$_->{tbl}}++ }
$checksum->find_replication_differences(
$src->{dbh}, $o->get('replicate'));
# Now check the slave for differences and sync them if necessary.
my @diffs = filter_diffs(
\%skip_table,
$databases,
$tables,
$checksum->find_replication_differences(
$dst->{dbh}, $o->get('replicate'))
);
if ( $o->get('verbose') ) {
print_header("# Syncing via replication " . $dp->as_string($dst->{dsn})
. ($o->get('dry-run') ?
' in dry-run mode, without accessing or comparing data' : ''));
}
if ( @diffs ) {
lock_server(src => $src, dst => $dst, %args);
foreach my $diff ( @diffs ) {
$src->{db} = $dst->{db} = $diff->{db};
$src->{tbl} = $dst->{tbl} = $diff->{tbl};
$exit_status |= sync_a_table(
src => $src,
dst => $dst,
where => $diff->{boundaries},
%args,
);
}
unlock_server(src => $src, dst => $dst, %args);
}
else {
MKDEBUG && _d('No checksum differences');
}
disconnect($dst);
} # sync-to-master
# The DSN is the master. Connect to each slave, find differences,
# then sync them.
else {
my %skip_table;
$ms->recurse_to_slaves(
{ dbh => $src->{dbh},
dsn => $src->{dsn},
dsn_parser => $dp,
recurse => 1,
callback => sub {
my ( $dsn, $dbh, $level, $parent ) = @_;
my @diffs = $checksum
->find_replication_differences($dbh, $o->get('replicate'));
if ( !$level ) {
# This is the master; don't sync any tables that are wrong
# here, for obvious reasons.
map { $skip_table{$_->{db}}->{$_->{tbl}}++ } @diffs;
}
else {
# This is a slave.
@diffs = filter_diffs(
\%skip_table,
$databases,
$tables,
@diffs
);
if ( $o->get('verbose') ) {
print_header("# Syncing via replication "
. $dp->as_string($dsn)
. ($o->get('dry-run')
? ' in dry-run mode, without '
. 'accessing or comparing data'
: ''));
}
if ( @diffs ) {
my $dst = {
dsn => $dsn,
dbh => $dbh,
misc_dbh => get_cxn($dsn, %args),
db => undef, # set later
tbl => undef, # set later
};
lock_server(src => $src, dst => $dst, %args);
foreach my $diff ( @diffs ) {
$src->{db} = $dst->{db} = $diff->{db};
$src->{tbl} = $dst->{tbl} = $diff->{tbl};
$exit_status |= sync_a_table(
src => $src,
dst => $dst,
where => $diff->{boundaries},
%args,
);
}
unlock_server(src => $src, dst => $dst, %args);
disconnect($dst);
}
else {
MKDEBUG && _d('No checksum differences');
}
} # this is a slave
return;
}, # recurse_to_slaves() callback
method => $o->get('recursion-method'),
},
);
} # DSN is master
disconnect($src);
return $exit_status;
}
# Sub: sync_all
# Sync every table between one source host and multiple destination hosts.
# The first DSN in $args{dsns} specifies the source host. The other DSNs
# are the destination hosts. Unlike <sync_one_table>, the database and
# table names must be the same on the source and destination hosts.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# dsns - Arrayref of DSNs
# plugins - Arrayref of TableSync* objects
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# Quoter - <Quoter> object
# VersionParser - <VersionParser> object
# TableParser - <TableParser> object
# MySQLDump - <MySQLDump> object
#
# Returns:
# Exit status
sub sync_all {
my ( %args ) = @_;
my @required_args = qw(dsns plugins OptionParser DSNParser Quoter
VersionParser TableParser MySQLDump);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my @dsns = @{$args{dsns}};
my $o = $args{OptionParser};
my $dp = $args{DSNParser};
MKDEBUG && _d('Syncing all dbs and tbls');
my $src = {
dsn => $dsns[0],
dbh => $dsns[0]->{dbh} || get_cxn($dsns[0], %args),
misc_dbh => get_cxn($dsns[0], %args),
db => undef, # set later
tbl => undef, # set later
};
my $si = new SchemaIterator(
Quoter => $args{Quoter},
);
$si->set_filter($si->make_filter($o));
# Make a list of all dbs.tbls on the source. It's more efficient this
# way because it avoids open/closing a dbh for each tbl and dsn, unless
# we pre-opened the dsn. It would also cause confusing verbose output.
my @dbs_tbls;
my $next_db = $si->get_db_itr(dbh => $src->{dbh});
while ( my $db = $next_db->() ) {
MKDEBUG && _d('Getting tables from', $db);
my $next_tbl = $si->get_tbl_itr(
dbh => $src->{dbh},
db => $db,
views => 0,
);
while ( my $tbl = $next_tbl->() ) {
MKDEBUG && _d('Got table', $tbl);
push @dbs_tbls, { db => $db, tbl => $tbl };
}
}
my $exit_status = 0;
foreach my $dsn ( @dsns[1 .. $#dsns] ) {
if ( $o->get('verbose') ) {
print_header("# Syncing " . $dp->as_string($dsn)
. ($o->get('dry-run')
? ' in dry-run mode, without accessing or comparing data' : ''));
}
my $dst = {
dsn => $dsn,
dbh => $dsn->{dbh} || get_cxn($dsn, %args),
misc_dbh => get_cxn($dsn, %args),
db => undef, # set later
tbl => undef, # set later
};
lock_server(src => $src, dst => $dst, %args);
foreach my $db_tbl ( @dbs_tbls ) {
$src->{db} = $dst->{db} = $db_tbl->{db};
$src->{tbl} = $dst->{tbl} = $db_tbl->{tbl};
$exit_status |= sync_a_table(
src => $src,
dst => $dst,
%args,
);
}
unlock_server(src => $src, dst => $dst, %args);
disconnect($dst);
}
disconnect($src);
return $exit_status;
}
# Sub: lock_server
# Lock a host with FLUSH TABLES WITH READ LOCK. This implements
# --lock 3 by calling <TableSyncer::lock_and_wait()>.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# src - Hashref with source host information
# dst - Hashref with destination host information
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# TableSyncer - <TableSyncer> object
sub lock_server {
my ( %args ) = @_;
foreach my $arg ( qw(src dst OptionParser DSNParser TableSyncer) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $o = $args{OptionParser};
return unless $o->get('lock') && $o->get('lock') == 3;
eval {
$args{TableSyncer}->lock_and_wait(
%args,
lock => 3,
lock_level => 3,
replicate => $o->get('replicate'),
timeout_ok => $o->get('timeout-ok'),
transaction => $o->get('transaction'),
wait => $o->get('wait'),
);
};
if ( $EVAL_ERROR ) {
die "Failed to lock server: $EVAL_ERROR";
}
return;
}
# Sub: unlock_server
# Unlock a host with UNLOCK TABLES. This implements
# --lock 3 by calling <TableSyncer::unlock()>.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# src - Hashref with source host information
# dst - Hashref with destination host information
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# TableSyncer - <TableSyncer> object
sub unlock_server {
my ( %args ) = @_;
my @required_args = qw(src dst OptionParser DSNParser TableSyncer);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($src, $dst, $o) = @args{@required_args};
return unless $o->get('lock') && $o->get('lock') == 3;
eval {
# Open connections as needed.
$src->{dbh} ||= get_cxn($src->{dsn}, %args);
$dst->{dbh} ||= get_cxn($dst->{dsn}, %args);
$src->{misc_dbh} ||= get_cxn($src->{dsn}, %args);
$args{TableSyncer}->unlock(
src_dbh => $src->{dbh},
src_db => '',
src_tbl => '',
dst_dbh => $dst->{dbh},
dst_db => '',
dst_tbl => '',
misc_dbh => $src->{misc_dbh},
replicate => $o->get('replicate') || 0,
timeout_ok => $o->get('timeout-ok') || 0,
transaction => $o->get('transaction') || 0,
wait => $o->get('wait') || 0,
lock => 3,
lock_level => 3,
);
};
if ( $EVAL_ERROR ) {
die "Failed to unlock server: $EVAL_ERROR";
}
return;
}
# Sub: sync_a_table
# Sync the destination host table to the source host table. This sub
# is not called directly but indirectly via the other sync_* subs.
# In turn, this sub calls <TableSyncer::sync_table()> which actually
# does the sync work. Calling sync_table() requires a fair amount of
# prep work that this sub does/simplifies. New <RowDiff> and <ChangeHandler>
# objects are created, so those packages need to be available.
#
# Parameters:
# $args - Arguments
#
# Required Arguments:
# src - Hashref with source host information
# dst - Hashref with destination host information
# plugins - Arrayref of TableSync* objects
# OptionParser - <OptionParser> object
# Quoter - <Quoter> object
# TableParser - <TableParser> object
# MySQLDump - <MySQLDump> object
# TableSyncer - <TableSyncer> object
#
# Returns:
# Exit status
sub sync_a_table {
my ( %args ) = @_;
my @required_args = qw(src dst plugins OptionParser Quoter TableParser
MySQLDump TableSyncer);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($src, $dst, undef, $o, $q, $tp, $du, $syncer) = @args{@required_args};
my ($start_ts, $end_ts);
my $exit_status = 0;
my %status;
eval {
$start_ts = get_server_time($src->{dbh}) if $o->get('verbose');
# This will either die if there's a problem or return the tbl struct.
my $tbl_struct = ok_to_sync($src, $dst, %args);
# If the table is InnoDB, prefer to sync it with transactions, unless
# the user explicitly said not to.
my $use_txn = $o->got('transaction') ? $o->get('transaction')
: $tbl_struct->{engine} eq 'InnoDB' ? 1
: 0;
# Turn off AutoCommit if we're using transactions.
$src->{dbh}->{AutoCommit} = !$use_txn;
$src->{misc_dbh}->{AutoCommit} = !$use_txn;
$dst->{dbh}->{AutoCommit} = !$use_txn;
$dst->{misc_dbh}->{AutoCommit} = !$use_txn;
# Determine which columns to compare.
my $ignore_columns = $o->get('ignore-columns');
my @compare_columns = grep {
!$ignore_columns->{lc $_};
} @{$o->get('columns') || $tbl_struct->{cols}};
# Make sure conflict col is in compare cols else conflicting
# rows won't have the col for --conflict-comparison.
if ( my $conflict_col = $o->get('conflict-column') ) {
push @compare_columns, $conflict_col
unless grep { $_ eq $conflict_col } @compare_columns;
}
# --print --verbose --verbose is the magic formula for having
# all src/dst sql printed so we can see the chunk/row sql.
my $callback;
if ( $o->get('print') && $o->get('verbose') >= 2 ) {
$callback = \&print_sql;
}
# get_change_dbh() may die if, for example, the destination is
# not a slave. Perhaps its work should be part of can_sync()?
my $change_dbh = get_change_dbh(tbl_struct => $tbl_struct, %args);
my $actions = make_action_subs(change_dbh => $change_dbh, %args);
my $rd = new RowDiff(dbh => $src->{misc_dbh});
my $ch = new ChangeHandler(
left_db => $src->{db},
left_tbl => $src->{tbl},
right_db => $dst->{db},
right_tbl => $dst->{tbl},
tbl_struct => $tbl_struct,
hex_blob => $o->get('hex-blob'),
queue => $o->get('buffer-to-client') ? 1 : 0,
replace => $o->get('replace')
|| $o->get('replicate')
|| $o->get('sync-to-master')
|| 0,
actions => $actions,
Quoter => $args{Quoter},
);
%status = $syncer->sync_table(
%args,
tbl_struct => $tbl_struct,
cols => \@compare_columns,
chunk_size => $o->get('chunk-size'),
RowDiff => $rd,
ChangeHandler => $ch,
transaction => $use_txn,
callback => $callback,
where => $args{where} || $o->get('where'),
bidirectional => $o->get('bidirectional'),
buffer_in_mysql => $o->get('buffer-in-mysql'),
buffer_to_client => $o->get('buffer-to-client'),
changing_src => $o->get('replicate')
|| $o->get('sync-to-master')
|| $o->get('bidirectional')
|| 0,
float_precision => $o->get('float-precision'),
index_hint => $o->get('index-hint'),
chunk_index => $o->get('chunk-index'),
chunk_col => $o->get('chunk-column'),
zero_chunk => $o->get('zero-chunk'),
lock => $o->get('lock'),
replace => $o->get('replace'),
replicate => $o->get('replicate'),
dry_run => $o->get('dry-run'),
timeout_ok => $o->get('timeout-ok'),
trim => $o->get('trim'),
wait => $o->get('wait'),
function => $o->get('function'),
);
if ( sum(@status{@ChangeHandler::ACTIONS}) ) {
$exit_status |= 2;
}
};
if ( $EVAL_ERROR ) {
print_err($EVAL_ERROR, $dst->{db}, $dst->{tbl}, $dst->{dsn}->{h});
$exit_status |= 1;
}
# Print this last so that the exit status is its final result.
if ( $o->get('verbose') ) {
$end_ts = get_server_time($src->{dbh}) || "";
print_results(
map { $_ || '0' } @status{@ChangeHandler::ACTIONS, 'ALGORITHM'},
$start_ts, $end_ts,
$exit_status, $src->{db}, $src->{tbl});
}
return $exit_status;
}
# Sub: get_change_dbh
# Return the dbh to write to for syncing changes. Write statements
# are executed on the "change dbh". If --sync-to-master or --replicate
# is specified, the source (master) dbh is the "change dbh". This means
# changes replicate to all slaves. Else, the destination dbh is the
# change dbh. This is the case when two independent servers (or perhaps
# one table on the same server) are synced. This sub implements
# --[no]check-slave because writing to a slave is generally a bad thing.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# src - Hashref with source host information
# dst - Hashref with destination host information
# tbl_struct - Hashref returned by <TableParser::parse()>
# OptionParser - <OptionParser> object
# DSNParser - <DSNParser> object
# MasterSlave - <MasterSlave> object
#
# Returns:
# Either $args{src}->{dbh} or $args{dst}->{dbh} if no checks fail.
#
# See Also:
# <make_action_subs()>
sub get_change_dbh {
my ( %args ) = @_;
my @required_args = qw(src dst tbl_struct OptionParser DSNParser
MasterSlave);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($src, $dst, $tbl_struct, $o, $dp, $ms) = @args{@required_args};
my $change_dbh = $dst->{dbh}; # The default case: making changes on dst.
if ( $o->get('sync-to-master') || $o->get('replicate') ) {
# Is it possible to make changes on the master (i.e. the source)?
# Only if REPLACE will work.
my $can_replace = grep { $_->{is_unique} } values %{$tbl_struct->{keys}};
MKDEBUG && _d("This table's replace-ability:", $can_replace);
die "Can't make changes on the master because no unique index exists"
unless $can_replace;
$change_dbh = $src->{dbh}; # The alternate case.
MKDEBUG && _d('Will make changes on source', $change_dbh);
}
elsif ( $o->get('check-slave') ) {
# Is it safe to change data on the destination? Only if it's *not*
# a slave. We don't change tables on slaves directly. If we are
# forced to change data on a slave, we require either that 1) binary
# logging is disabled, or 2) the check is bypassed. By the way, just
# because the server is a slave doesn't mean it's not also the master
# of the master (master-master replication).
my $slave_status = $ms->get_slave_status($dst->{dbh});
my (undef, $log_bin) = $dst->{dbh}->selectrow_array(
'SHOW VARIABLES LIKE "log_bin"');
my ($sql_log_bin) = $dst->{dbh}->selectrow_array(
'SELECT @@SQL_LOG_BIN');
MKDEBUG && _d('Variables on destination:',
'log_bin=', (defined $log_bin ? $log_bin : 'NULL'),
' @@SQL_LOG_BIN=', (defined $sql_log_bin ? $sql_log_bin : 'NULL'));
if ( $slave_status && $sql_log_bin && ($log_bin || 'OFF') eq 'ON' ) {
die "Can't make changes on ", $dp->as_string($dst->{dsn}),
" because it's a slave. See the documentation section",
" 'REPLICATION SAFETY' for solutions to this problem.";
}
MKDEBUG && _d('Will make changes on destination', $change_dbh);
}
return $change_dbh;
}
# Sub: make_action_subs
# Make callbacks for <ChangeHandler::new()> actions argument. This
# sub implements --print and --execute.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# change_dbh - dbh returned by <get_change_dbh>
# OptionParser - <OptionParser> object
#
# Returns:
# Arrayref of callbacks (coderefs)
sub make_action_subs {
my ( %args ) = @_;
my @required_args = qw(change_dbh OptionParser);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($change_dbh, $o) = @args{@required_args};
my @actions;
if ( $o->get('execute') ) {
push @actions, sub {
my ( $sql, $dbh ) = @_;
# Use $dbh if given. It's from a bidirectional callback.
$dbh ||= $change_dbh;
MKDEBUG && _d('Execute on dbh', $dbh, $sql);
$dbh->do($sql);
};
}
if ( $o->get('print') ) {
# Print AFTER executing, so the print isn't misleading in case of an
# index violation etc that doesn't actually get executed.
push @actions, sub {
my ( $sql, $dbh ) = @_;
# Append /*host:port*/ to the sql, if possible, so the user
# can see on which host it was/would be ran.
my $dsn = $dsn_for{$dbh} if $dbh;
if ( $dsn ) {
my $h = $dsn->{h} || $dsn->{S} || '';
my $p = $dsn->{P} || '';
$sql = "/*$h" . ($p ? ":$p" : '') . "*/ $sql";
}
print($sql, ";\n") or die "Cannot print: $OS_ERROR";
};
}
return \@actions;
}
# Sub: print_err
# Try to extract the MySQL error message and print it.
#
# Parameters:
# $msg - Error message
# $database - Database name being synced when error occurred
# $table - Table name being synced when error occurred
# $host - Host name error occurred on
sub print_err {
my ( $msg, $database, $table, $host ) = @_;
return if !defined $msg;
$msg =~ s/^.*?failed: (.*?) at \S+ line (\d+).*$/$1 at line $2/s;
$msg =~ s/\s+/ /g;
if ( $database && $table ) {
$msg .= " while doing $database.$table";
}
if ( $host ) {
$msg .= " on $host";
}
print STDERR $msg, "\n";
}
# Sub: get_cxn
# Connect to host specified by DSN.
#
# Parameters:
# $dsn - Host DSN
# %args - Arguments
#
# Required Arguments:
# OptionaParser - <OptionParser> object
# DSNParser - <DSNParser> object
#
# Returns:
# dbh
sub get_cxn {
my ( $dsn, %args ) = @_;
my @required_args = qw(OptionParser DSNParser);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($o, $dp) = @args{@required_args};
if ( !$dsn->{p} && $o->get('ask-pass') ) {
# Just "F=file" is a valid DSN but fill_in_dsn() can't help us
# because we haven't connected yet. If h is not specified,
# then user is relying on F or .my.cnf/system defaults.
# http://code.google.com/p/maatkit/issues/detail?id=947
my $host = $dsn->{h} ? $dsn->{h}
: "DSN ". $dp->as_string($dsn);
$dsn->{p} = OptionParser::prompt_noecho("Enter password for $host: ");
}
my $dbh = $dp->get_dbh(
$dp->get_cxn_params($dsn, {}) # get_cxn_params needs the 2nd arg
);
my $sql;
if ( !$o->get('bin-log') ) {
$sql = "/*!32316 SET SQL_LOG_BIN=0 */";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
if ( !$o->get('unique-checks') ) {
$sql = "/*!40014 SET UNIQUE_CHECKS=0 */";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
if ( !$o->get('foreign-key-checks') ) {
$sql = "/*!40014 SET FOREIGN_KEY_CHECKS=0 */";
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
}
# Disable auto-increment on zero (bug #1919897).
$sql = '/*!40101 SET @@SQL_MODE := CONCAT(@@SQL_MODE, '
. '",NO_AUTO_VALUE_ON_ZERO")*/';
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
# Ensure statement-based replication.
# http://code.google.com/p/maatkit/issues/detail?id=95
$sql = '/*!50105 SET @@binlog_format="STATEMENT"*/';
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
if ( $o->get('transaction') ) {
my $sql = "SET SESSION TRANSACTION ISOLATION LEVEL REPEATABLE READ";
eval {
MKDEBUG && _d($dbh, $sql);
$dbh->do($sql);
};
die "Failed to $sql: $EVAL_ERROR" if $EVAL_ERROR;
}
$dsn_for{$dbh} = $dsn;
MKDEBUG && _d('Opened dbh', $dbh);
return $dbh;
}
# Sub: ok_to_sync
# Check that the destination host table can be synced to the source host
# table. All sorts of sanity checks are performed to help ensure that
# syncing the table won't cause problems in <sync_a_table()> or
# <TableSyncer::sync_table()>.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# src - Hashref with source host information
# dst - Hashref with destination host information
# DSNParser - <DSNParser> object
# Quoter - <Quoter> object
# VersionParser - <VersionParser> object
# TableParser - <TableParser> object
# MySQLDump - <MySQLDump> object
# TableSyncer - <TableSyncer> object
# OptionParser - <OptionParser> object
#
# Returns:
# Table structure (from <TableParser::parse()>) if ok to sync, else it dies.
sub ok_to_sync {
my ( %args ) = @_;
my @required_args = qw(src dst DSNParser Quoter VersionParser TableParser
MySQLDump TableSyncer OptionParser);
foreach my $arg ( @required_args ) {
die "I need a $arg argument" unless $args{$arg};
}
my ($src, $dst, $dp, $q, $vp, $tp, $du, $syncer, $o) = @args{@required_args};
# First things first: check that the src and dst dbs and tbls exist.
# This can fail in cases like h=host,D=bad,t=also_bad (i.e. simple
# user error). It can also fail when syncing all dbs/tbls with sync_all()
# because the dst db/tbl is assumed to be the same as the src but
# this isn't always the case.
my $src_tbl_ddl;
eval {
# FYI: get_create_table() does USE db but doesn't eval it.
$src->{dbh}->do("USE `$src->{db}`");
$src_tbl_ddl = $du->get_create_table($src->{dbh}, $q,
$src->{db}, $src->{tbl});
};
die $EVAL_ERROR if $EVAL_ERROR;
my $dst_tbl_ddl;
eval {
# FYI: get_create_table() does USE db but doesn't eval it.
$dst->{dbh}->do("USE `$dst->{db}`");
$dst_tbl_ddl = $du->get_create_table($dst->{dbh}, $q,
$dst->{db}, $dst->{tbl});
};
die $EVAL_ERROR if $EVAL_ERROR;
# This doesn't work at the moment when syncing different table names.
# Check that src.db.tbl has the exact same schema as dst.db.tbl.
# if ( $o->get('check-schema') && ($src_tbl_ddl ne $dst_tbl_ddl) ) {
# die "Source and destination tables have different schemas";
# }
my $tbl_struct = $tp->parse($src_tbl_ddl);
# Check that the user has all the necessary privs on the tbls.
if ( $o->get('check-privileges') ) {
MKDEBUG && _d('Checking privileges');
if ( !$syncer->have_all_privs($src->{dbh}, $src->{db}, $src->{tbl}) ) {
my $user = get_current_user($src->{dbh}) || "";
die "User $user does not have all necessary privileges on ",
$q->quote($src->{db}, $src->{tbl});
}
if ( !$syncer->have_all_privs($dst->{dbh}, $dst->{db}, $dst->{tbl}) ) {
my $user = get_current_user($dst->{dbh}) || "";
die "User $user does not have all necessary privileges on ",
$q->quote($dst->{db}, $dst->{tbl});
}
}
# Check that no triggers are defined on the dst tbl.
if ( $o->get('check-triggers') ) {
MKDEBUG && _d('Checking for triggers');
if ( !defined $dst->{supports_triggers} ) {
$dst->{supports_triggers} = $vp->version_ge($dst->{dbh}, '5.0.2');
}
if ( $dst->{supports_triggers}
&& $du->get_triggers($dst->{dbh}, $q, $dst->{db}, $dst->{tbl}) ) {
die "Triggers are defined on the table";
}
else {
MKDEBUG && _d('Destination does not support triggers',
$dp->as_string($dst->{dsn}));
}
}
return $tbl_struct;
}
# Sub: filter_diffs
# Filter different slave tables according to the various schema object
# filters. This sub is called in <sync_via_replication()> to implement
# schema object filters like --databases and --tables.
#
# Parameters:
# $skip_table - Hashref of databases and tables to skip
# $databases - Hashref of databases to skip
# $tables - Hashref of tables to skip
# @diffs - Array of hashrefs, one for each different slave table
#
# Returns:
# Array of different slave tables that pass the filters
sub filter_diffs {
my ( $skip_table, $databases, $tables, @diffs ) = @_;
return grep {
!$skip_table->{$_->{db}}->{$_->{tbl}}
&& (!$databases || $databases->{$_->{db}})
&& (!$tables || ($tables->{$_->{tbl}} || $tables->{"$_->{db}.$_->{tbl}"}))
} @diffs;
}
# Sub: disconnect
# Disconnect host dbhs created by <get_cxn()>. To make sure all dbh
# are closed, pt-table-sync keeps track of the dbh it opens and this
# sub helps keep track of the dbh that are closed.
#
# Parameters:
# @hosts - Array of hashrefs with host information, one for each host
sub disconnect {
my ( @hosts ) = @_;
foreach my $host ( @hosts ) {
foreach my $thing ( qw(dbh misc_dbh) ) {
my $dbh = $host->{$thing};
next unless $dbh;
delete $dsn_for{$dbh};
$dbh->commit() unless $dbh->{AutoCommit};
$dbh->disconnect();
MKDEBUG && _d('Disconnected dbh', $dbh);
}
}
return;
}
# Sub: print_sql
# Callback for <TableSyncer::sync_table()> if --print --verbose --verbose
# is specified. The callback simply prints the SQL statements passed to
# it by sync_table(). They're usually (always?) identical statements.
#
# Parameters:
# $src_sql - SQL statement to be executed on the sourch host
# $dst_sql - SQL statement to be executed on the destination host
sub print_sql {
my ( $src_sql, $dst_sql ) = @_;
print "# $src_sql\n" if $src_sql;
print "# $dst_sql\n" if $dst_sql;
return;
}
use constant UPDATE_LEFT => -1;
use constant UPDATE_RIGHT => 1;
use constant UPDATE_NEITHER => 0; # neither value equals/matches
use constant FAILED_THRESHOLD => 2; # failed to exceed threshold
# Sub: cmd_conflict_col
# Compare --conflict-column values for --bidirectional. This sub is
# used as a callback in <set_bidirectional_callbacks()>.
#
# Parameters:
# $left_val - Column value from left (usually the source host)
# $right_val - Column value from right (usually the destination host)
# $cmp - Type of conflict comparison, --conflict-comparison
# $val - Value for certain types of comparisons, --conflict-value
# $thr - Threshold for certain types of comparisons,
# --conflict-threshold
#
# Returns:
# One of the constants above, UPDATE_* or FAILED_THRESHOLD
sub cmp_conflict_col {
my ( $left_val, $right_val, $cmp, $val, $thr ) = @_;
MKDEBUG && _d('Compare', @_);
my $res;
if ( $cmp eq 'newest' || $cmp eq 'oldest' ) {
$res = $cmp eq 'newest' ? ($left_val || '') cmp ($right_val || '')
: ($right_val || '') cmp ($left_val || '');
if ( $thr ) {
$thr = time_to_secs($thr);
my $lts = any_unix_timestamp($left_val);
my $rts = any_unix_timestamp($right_val);
my $diff = abs($lts - $rts);
MKDEBUG && _d('Check threshold, lts rts thr abs-diff:',
$lts, $rts, $thr, $diff);
if ( $diff < $thr ) {
MKDEBUG && _d("Failed threshold");
return FAILED_THRESHOLD;
}
}
}
elsif ( $cmp eq 'greatest' || $cmp eq 'least' ) {
$res = $cmp eq 'greatest' ? (($left_val ||0) > ($right_val ||0) ? 1 : -1)
: (($left_val ||0) < ($right_val ||0) ? 1 : -1);
$res = 0 if ($left_val || 0) == ($right_val || 0);
if ( $thr ) {
my $diff = abs($left_val - $right_val);
MKDEBUG && _d('Check threshold, abs-diff:', $diff);
if ( $diff < $thr ) {
MKDEBUG && _d("Failed threshold");
return FAILED_THRESHOLD;
}
}
}
elsif ( $cmp eq 'equals' ) {
$res = ($left_val || '') eq $val ? 1
: ($right_val || '') eq $val ? -1
: 0;
}
elsif ( $cmp eq 'matches' ) {
$res = ($left_val || '') =~ m/$val/ ? 1
: ($right_val || '') =~ m/$val/ ? -1
: 0;
}
else {
# Should happen; caller should have verified this.
die "Invalid comparison: $cmp";
}
return $res;
}
# Sub: set_bidirectional_callbacks
# Set syncer plugin callbacks for --bidirectional.
#
# Parameters:
# %args - Arguments
#
# Required Arguments:
# plugin - TableSync* object
# OptionParser - <OptionParser> object
sub set_bidirectional_callbacks {
my ( %args ) = @_;
foreach my $arg ( qw(plugin OptionParser) ) {
die "I need a $arg argument" unless $args{$arg};
}
my $o = $args{OptionParser};
my $plugin = $args{plugin};
my $col = $o->get('conflict-column');
my $cmp = $o->get('conflict-comparison');
my $val = $o->get('conflict-value');
my $thr = $o->get('conflict-threshold');
# plugin and syncer are actually the same module. For clarity we
# name them differently.
$plugin->set_callback('same_row', sub {
my ( %args ) = @_;
my ($lr, $rr, $syncer) = @args{qw(lr rr syncer)};
my $ch = $syncer->{ChangeHandler};
my $action = 'UPDATE';
my $change_dbh;
my $auth_row;
my $err;
my $left_val = $lr->{$col} || '';
my $right_val = $rr->{$col} || '';
MKDEBUG && _d('left', $col, 'value:', $left_val);
MKDEBUG && _d('right', $col, 'value:', $right_val);
my $res = cmp_conflict_col($left_val, $right_val, $cmp, $val, $thr);
if ( $res == UPDATE_LEFT ) {
MKDEBUG && _d("right dbh $args{right_dbh} $cmp; "
. "update left dbh $args{left_dbh}");
$ch->set_src('right', $args{right_dbh});
$auth_row = $args{rr};
$change_dbh = $args{left_dbh};
}
elsif ( $res == UPDATE_RIGHT ) {
MKDEBUG && _d("left dbh $args{left_dbh} $cmp; "
. "update right dbh $args{right_dbh}");
$ch->set_src('left', $args{left_dbh});
$auth_row = $args{lr};
$change_dbh = $args{right_dbh};
}
elsif ( $res == UPDATE_NEITHER ) {
if ( $cmp eq 'equals' || $cmp eq 'matches' ) {
$err = "neither `$col` value $cmp $val";
}
else {
$err = "`$col` values are the same"
}
}
elsif ( $res == FAILED_THRESHOLD ) {
$err = "`$col` values do not differ by the threhold, $thr."
}
else {
# Shouldn't happen.
die "cmp_conflict_col() returned an invalid result: $res."
}
if ( $err ) {
$action = undef; # skip change in case we just warn
my $where = $ch->make_where_clause($lr, $syncer->key_cols());
$err = "# Cannot resolve conflict WHERE $where: $err\n";
# die here is caught in sync_a_table(). We're deeply nested:
# sync_a_table > sync_table > compare_sets > syncer > here
$o->get('conflict-error') eq 'warn' ? warn $err : die $err;
}
return $action, $auth_row, $change_dbh;
});
$plugin->set_callback('not_in_right', sub {
my ( %args ) = @_;
$args{syncer}->{ChangeHandler}->set_src('left', $args{left_dbh});
return 'INSERT', $args{lr}, $args{right_dbh};
});
$plugin->set_callback('not_in_left', sub {
my ( %args ) = @_;
$args{syncer}->{ChangeHandler}->set_src('right', $args{right_dbh});
return 'INSERT', $args{rr}, $args{left_dbh};
});
return;
}
# Sub: get_plugins
# Get internal TableSync* plugins.
#
# Returns:
# Hash of available algoritms and the plugin/module names that
# implement them, like "chunk => TableSyncChunk".
sub get_plugins {
my ( %args ) = @_;
my $file = __FILE__;
open my $fh, "<", $file or die "Cannot open $file: $OS_ERROR";
my $contents = do { local $/ = undef; <$fh> };
close $fh;
my %local_plugins = map {
my $package = $_;
my ($module, $algo) = $package =~ m/(TableSync(\w+))/;
lc $algo => $module;
} $contents =~ m/^package TableSync\w{3,};/gm;
return %local_plugins;
}
{
# DELETE REPLACE INSERT UPDATE ALGORITHM START END EXIT DATABASE.TABLE
my $hdr = "# %6s %7s %6s %6s %-9s %-8s %-8s %-4s %s.%s\n";
sub print_header {
my ( $title ) = @_;
print "$title\n" if $title;
printf $hdr, @ChangeHandler::ACTIONS,
qw(ALGORITHM START END EXIT DATABASE TABLE);
return;
}
sub print_results {
my ( @values ) = @_;
printf $hdr, @values;
return;
}
}
# Sub: get_server_time
# Return HH:MM:SS of SELECT NOW() from the server.
#
# Parameters:
# $dbh - dbh
sub get_server_time {
my ( $dbh ) = @_;
return unless $dbh;
my $now;
eval {
my $sql = "SELECT NOW()";
MKDEBUG && _d($dbh, $sql);
($now) = $dbh->selectrow_array($sql);
MKDEBUG && _d("Server time:", $now);
$now =~ s/^\S+\s+//;
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d("Failed to get server time:", $EVAL_ERROR);
}
return $now
}
sub get_current_user {
my ( $dbh ) = @_;
return unless $dbh;
my $user;
eval {
my $sql = "SELECT CURRENT_USER()";
MKDEBUG && _d($dbh, $sql);
($user) = $dbh->selectrow_array($sql);
};
if ( $EVAL_ERROR ) {
MKDEBUG && _d("Error getting current user:", $EVAL_ERROR);
}
return $user;
}
sub _d {
my ($package, undef, $line) = caller 0;
@_ = map { (my $temp = $_) =~ s/\n/\n# /g; $temp; }
map { defined $_ ? $_ : 'undef' }
@_;
print STDERR "# $package:$line $PID ", join(' ', @_), "\n";
}
# ############################################################################
# Run the program.
# ############################################################################
if ( !caller ) { exit main(@ARGV); }
1; # Because this is a module as well as a script.
# ############################################################################
# Documentation
# ############################################################################
=pod
=head1 NAME
pt-table-sync - Synchronize MySQL table data efficiently.
=head1 SYNOPSIS
Usage: pt-table-sync [OPTION...] DSN [DSN...]
pt-table-sync synchronizes data efficiently between MySQL tables.
This tool changes data, so for maximum safety, you should back up your data
before you use it. When synchronizing a server that is a replication slave with
the --replicate or --sync-to-master methods, it B<always> makes the changes on
the replication master, B<never> the replication slave directly. This is in
general the only safe way to bring a replica back in sync with its master;
changes to the replica are usually the source of the problems in the first
place. However, the changes it makes on the master should be no-op changes that
set the data to their current values, and actually affect only the replica.
Please read the detailed documentation that follows to learn more about this.
Sync db.tbl on host1 to host2:
pt-table-sync --execute h=host1,D=db,t=tbl h=host2
Sync all tables on host1 to host2 and host3:
pt-table-sync --execute host1 host2 host3
Make slave1 have the same data as its replication master:
pt-table-sync --execute --sync-to-master slave1
Resolve differences that L<pt-table-checksum> found on all slaves of master1:
pt-table-sync --execute --replicate test.checksum master1
Same as above but only resolve differences on slave1:
pt-table-sync --execute --replicate test.checksum \
--sync-to-master slave1
Sync master2 in a master-master replication configuration, where master2's copy
of db.tbl is known or suspected to be incorrect:
pt-table-sync --execute --sync-to-master h=master2,D=db,t=tbl
Note that in the master-master configuration, the following will NOT do what you
want, because it will make changes directly on master2, which will then flow
through replication and change master1's data:
# Don't do this in a master-master setup!
pt-table-sync --execute h=master1,D=db,t=tbl master2
=head1 RISKS
The following section is included to inform users about the potential risks,
whether known or unknown, of using this tool. The two main categories of risks
are those created by the nature of the tool (e.g. read-only tools vs. read-write
tools) and those created by bugs.
With great power comes great responsibility! This tool changes data, so it is a
good idea to back up your data. It is also very powerful, which means it is
very complex, so you should run it with the L<"--dry-run"> option to see what it
will do, until you're familiar with its operation. If you want to see which
rows are different, without changing any data, use L<"--print"> instead of
L<"--execute">.
Be careful when using pt-table-sync in any master-master setup. Master-master
replication is inherently tricky, and it's easy to make mistakes. You need to
be sure you're using the tool correctly for master-master replication. See the
L<"SYNOPSIS"> for the overview of the correct usage.
Also be careful with tables that have foreign key constraints with C<ON DELETE>
or C<ON UPDATE> definitions because these might cause unintended changes on the
child tables.
In general, this tool is best suited when your tables have a primary key or
unique index. Although it can synchronize data in tables lacking a primary key
or unique index, it might be best to synchronize that data by another means.
At the time of this release, there is a potential bug using
L<"--lock-and-rename"> with MySQL 5.1, a bug detecting certain differences,
a bug using ROUND() across different platforms, and a bug mixing collations.
The authoritative source for updated information is always the online issue
tracking system. Issues that affect this tool will be marked as such. You can
see a list of such issues at the following URL:
L<http://www.percona.com/bugs/pt-table-sync>.
See also L<"BUGS"> for more information on filing bugs and getting help.
=head1 DESCRIPTION
pt-table-sync does one-way and bidirectional synchronization of table data.
It does B<not> synchronize table structures, indexes, or any other schema
objects. The following describes one-way synchronization.
L<"BIDIRECTIONAL SYNCING"> is described later.
This tool is complex and functions in several different ways. To use it
safely and effectively, you should understand three things: the purpose
of L<"--replicate">, finding differences, and specifying hosts. These
three concepts are closely related and determine how the tool will run.
The following is the abbreviated logic:
if DSN has a t part, sync only that table:
if 1 DSN:
if --sync-to-master:
The DSN is a slave. Connect to its master and sync.
if more than 1 DSN:
The first DSN is the source. Sync each DSN in turn.
else if --replicate:
if --sync-to-master:
The DSN is a slave. Connect to its master, find records
of differences, and fix.
else:
The DSN is the master. Find slaves and connect to each,
find records of differences, and fix.
else:
if only 1 DSN and --sync-to-master:
The DSN is a slave. Connect to its master, find tables and
filter with --databases etc, and sync each table to the master.
else:
find tables, filtering with --databases etc, and sync each
DSN to the first.
pt-table-sync can run in one of two ways: with L<"--replicate"> or without.
The default is to run without L<"--replicate"> which causes pt-table-sync
to automatically find differences efficiently with one of several
algorithms (see L<"ALGORITHMS">). Alternatively, the value of
L<"--replicate">, if specified, causes pt-table-sync to use the differences
already found by having previously ran L<pt-table-checksum> with its own
C<--replicate> option. Strictly speaking, you don't need to use
L<"--replicate"> because pt-table-sync can find differences, but many
people use L<"--replicate"> if, for example, they checksum regularly
using L<pt-table-checksum> then fix differences as needed with pt-table-sync.
If you're unsure, read each tool's documentation carefully and decide for
yourself, or consult with an expert.
Regardless of whether L<"--replicate"> is used or not, you need to specify
which hosts to sync. There are two ways: with L<"--sync-to-master"> or
without. Specifying L<"--sync-to-master"> makes pt-table-sync expect
one and only slave DSN on the command line. The tool will automatically
discover the slave's master and sync it so that its data is the same as
its master. This is accomplished by making changes on the master which
then flow through replication and update the slave to resolve its differences.
B<Be careful though>: although this option specifies and syncs a single
slave, if there are other slaves on the same master, they will receive
via replication the changes intended for the slave that you're trying to
sync.
Alternatively, if you do not specify L<"--sync-to-master">, the first
DSN given on the command line is the source host. There is only ever
one source host. If you do not also specify L<"--replicate">, then you
must specify at least one other DSN as the destination host. There
can be one or more destination hosts. Source and destination hosts
must be independent; they cannot be in the same replication topology.
pt-table-sync will die with an error if it detects that a destination
host is a slave because changes are written directly to destination hosts
(and it's not safe to write directly to slaves). Or, if you specify
L<"--replicate"> (but not L<"--sync-to-master">) then pt-table-sync expects
one and only one master DSN on the command line. The tool will automatically
discover all the master's slaves and sync them to the master. This is
the only way to sync several (all) slaves at once (because
L<"--sync-to-master"> only specifies one slave).
Each host on the command line is specified as a DSN. The first DSN
(or only DSN for cases like L<"--sync-to-master">) provides default values
for other DSNs, whether those other DSNs are specified on the command line
or auto-discovered by the tool. So in this example,
pt-table-sync --execute h=host1,u=msandbox,p=msandbox h=host2
the host2 DSN inherits the C<u> and C<p> DSN parts from the host1 DSN.
Use the L<"--explain-hosts"> option to see how pt-table-sync will interpret
the DSNs given on the command line.
=head1 OUTPUT
If you specify the L<"--verbose"> option, you'll see information about the
differences between the tables. There is one row per table. Each server is
printed separately. For example,
# Syncing h=host1,D=test,t=test1
# DELETE REPLACE INSERT UPDATE ALGORITHM START END EXIT DATABASE.TABLE
# 0 0 3 0 Chunk 13:00:00 13:00:17 2 test.test1
Table test.test1 on host1 required 3 C<INSERT> statements to synchronize
and it used the Chunk algorithm (see L<"ALGORITHMS">). The sync operation
for this table started at 13:00:00 and ended 17 seconds later (times taken
from C<NOW()> on the source host). Because differences were found, its
L<"EXIT STATUS"> was 2.
If you specify the L<"--print"> option, you'll see the actual SQL statements
that the script uses to synchronize the table if L<"--execute"> is also
specified.
If you want to see the SQL statements that pt-table-sync is using to select
chunks, nibbles, rows, etc., then specify L<"--print"> once and L<"--verbose">
twice. Be careful though: this can print a lot of SQL statements.
There are cases where no combination of C<INSERT>, C<UPDATE> or C<DELETE>
statements can resolve differences without violating some unique key. For
example, suppose there's a primary key on column a and a unique key on column b.
Then there is no way to sync these two tables with straightforward UPDATE
statements:
+---+---+ +---+---+
| a | b | | a | b |
+---+---+ +---+---+
| 1 | 2 | | 1 | 1 |
| 2 | 1 | | 2 | 2 |
+---+---+ +---+---+
The tool rewrites queries to C<DELETE> and C<REPLACE> in this case. This is
automatically handled after the first index violation, so you don't have to
worry about it.
=head1 REPLICATION SAFETY
Synchronizing a replication master and slave safely is a non-trivial problem, in
general. There are all sorts of issues to think about, such as other processes
changing data, trying to change data on the slave, whether the destination and
source are a master-master pair, and much more.
In general, the safe way to do it is to change the data on the master, and let
the changes flow through replication to the slave like any other changes.
However, this works only if it's possible to REPLACE into the table on the
master. REPLACE works only if there's a unique index on the table (otherwise it
just acts like an ordinary INSERT).
If your table has unique keys, you should use the L<"--sync-to-master"> and/or
L<"--replicate"> options to sync a slave to its master. This will generally do
the right thing. When there is no unique key on the table, there is no choice
but to change the data on the slave, and pt-table-sync will detect that you're
trying to do so. It will complain and die unless you specify
C<--no-check-slave> (see L<"--[no]check-slave">).
If you're syncing a table without a primary or unique key on a master-master
pair, you must change the data on the destination server. Therefore, you need
to specify C<--no-bin-log> for safety (see L<"--[no]bin-log">). If you don't,
the changes you make on the destination server will replicate back to the
source server and change the data there!
The generally safe thing to do on a master-master pair is to use the
L<"--sync-to-master"> option so you don't change the data on the destination
server. You will also need to specify C<--no-check-slave> to keep
pt-table-sync from complaining that it is changing data on a slave.
=head1 ALGORITHMS
pt-table-sync has a generic data-syncing framework which uses different
algorithms to find differences. The tool automatically chooses the best
algorithm for each table based on indexes, column types, and the algorithm
preferences specified by L<"--algorithms">. The following algorithms are
available, listed in their default order of preference:
=over
=item Chunk
Finds an index whose first column is numeric (including date and time types),
and divides the column's range of values into chunks of approximately
L<"--chunk-size"> rows. Syncs a chunk at a time by checksumming the entire
chunk. If the chunk differs on the source and destination, checksums each
chunk's rows individually to find the rows that differ.
It is efficient when the column has sufficient cardinality to make the chunks
end up about the right size.
The initial per-chunk checksum is quite small and results in minimal network
traffic and memory consumption. If a chunk's rows must be examined, only the
primary key columns and a checksum are sent over the network, not the entire
row. If a row is found to be different, the entire row will be fetched, but not
before.
=item Nibble
Finds an index and ascends the index in fixed-size nibbles of L<"--chunk-size">
rows, using a non-backtracking algorithm (see L<pt-archiver> for more on this
algorithm). It is very similar to L<"Chunk">, but instead of pre-calculating
the boundaries of each piece of the table based on index cardinality, it uses
C<LIMIT> to define each nibble's upper limit, and the previous nibble's upper
limit to define the lower limit.
It works in steps: one query finds the row that will define the next nibble's
upper boundary, and the next query checksums the entire nibble. If the nibble
differs between the source and destination, it examines the nibble row-by-row,
just as L<"Chunk"> does.
=item GroupBy
Selects the entire table grouped by all columns, with a COUNT(*) column added.
Compares all columns, and if they're the same, compares the COUNT(*) column's
value to determine how many rows to insert or delete into the destination.
Works on tables with no primary key or unique index.
=item Stream
Selects the entire table in one big stream and compares all columns. Selects
all columns. Much less efficient than the other algorithms, but works when
there is no suitable index for them to use.
=item Future Plans
Possibilities for future algorithms are TempTable (what I originally called
bottom-up in earlier versions of this tool), DrillDown (what I originally
called top-down), and GroupByPrefix (similar to how SqlYOG Job Agent works).
Each algorithm has strengths and weaknesses. If you'd like to implement your
favorite technique for finding differences between two sources of data on
possibly different servers, I'm willing to help. The algorithms adhere to a
simple interface that makes it pretty easy to write your own.
=back
=head1 BIDIRECTIONAL SYNCING
Bidirectional syncing is a new, experimental feature. To make it work
reliably there are a number of strict limitations:
* only works when syncing one server to other independent servers
* does not work in any way with replication
* requires that the table(s) are chunkable with the Chunk algorithm
* is not N-way, only bidirectional between two servers at a time
* does not handle DELETE changes
For example, suppose we have three servers: c1, r1, r2. c1 is the central
server, a pseudo-master to the other servers (viz. r1 and r2 are not slaves
to c1). r1 and r2 are remote servers. Rows in table foo are updated and
inserted on all three servers and we want to synchronize all the changes
between all the servers. Table foo has columns:
id int PRIMARY KEY
ts timestamp auto updated
name varchar
Auto-increment offsets are used so that new rows from any server do not
create conflicting primary key (id) values. In general, newer rows, as
determined by the ts column, take precedence when a same but differing row
is found during the bidirectional sync. "Same but differing" means that
two rows have the same primary key (id) value but different values for some
other column, like the name column in this example. Same but differing
conflicts are resolved by a "conflict". A conflict compares some column of
the competing rows to determine a "winner". The winning row becomes the
source and its values are used to update the other row.
There are subtle differences between three columns used to achieve
bidirectional syncing that you should be familiar with: chunk column
(L<"--chunk-column">), comparison column(s) (L<"--columns">), and conflict
column (L<"--conflict-column">). The chunk column is only used to chunk the
table; e.g. "WHERE id >= 5 AND id < 10". Chunks are checksummed and when
chunk checksums reveal a difference, the tool selects the rows in that
chunk and checksums the L<"--columns"> for each row. If a column checksum
differs, the rows have one or more conflicting column values. In a
traditional unidirectional sync, the conflict is a moot point because it can
be resolved simply by updating the entire destination row with the source
row's values. In a bidirectional sync, however, the L<"--conflict-column">
(in accordance with other C<--conflict-*> options list below) is compared
to determine which row is "correct" or "authoritative"; this row becomes
the "source".
To sync all three servers completely, two runs of pt-table-sync are required.
The first run syncs c1 and r1, then syncs c1 and r2 including any changes
from r1. At this point c1 and r2 are completely in sync, but r1 is missing
any changes from r2 because c1 didn't have these changes when it and r1
were synced. So a second run is needed which syncs the servers in the same
order, but this time when c1 and r1 are synced r1 gets r2's changes.
The tool does not sync N-ways, only bidirectionally between the first DSN
given on the command line and each subsequent DSN in turn. So the tool in
this example would be ran twice like:
pt-table-sync --bidirectional h=c1 h=r1 h=r2
The L<"--bidirectional"> option enables this feature and causes various
sanity checks to be performed. You must specify other options that tell
pt-table-sync how to resolve conflicts for same but differing rows.
These options are:
* --conflict-column
* --conflict-comparison
* --conflict-value
* --conflict-threshold
* --conflict-error"> (optional)
Use L<"--print"> to test this option before L<"--execute">. The printed
SQL statements will have comments saying on which host the statement
would be executed if you used L<"--execute">.
Technical side note: the first DSN is always the "left" server and the other
DSNs are always the "right" server. Since either server can become the source
or destination it's confusing to think of them as "src" and "dst". Therefore,
they're generically referred to as left and right. It's easy to remember
this because the first DSN is always to the left of the other server DSNs on
the command line.
=head1 EXIT STATUS
The following are the exit statuses (also called return values, or return codes)
when pt-table-sync finishes and exits.
STATUS MEANING
====== =======================================================
0 Success.
1 Internal error.
2 At least one table differed on the destination.
3 Combination of 1 and 2.
=head1 OPTIONS
Specify at least one of L<"--print">, L<"--execute">, or L<"--dry-run">.
L<"--where"> and L<"--replicate"> are mutually exclusive.
This tool accepts additional command-line arguments. Refer to the
L<"SYNOPSIS"> and usage information for details.
=over
=item --algorithms
type: string; default: Chunk,Nibble,GroupBy,Stream
Algorithm to use when comparing the tables, in order of preference.
For each table, pt-table-sync will check if the table can be synced with
the given algorithms in the order that they're given. The first algorithm
that can sync the table is used. See L<"ALGORITHMS">.
=item --ask-pass
Prompt for a password when connecting to MySQL.
=item --bidirectional
Enable bidirectional sync between first and subsequent hosts.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --[no]bin-log
default: yes
Log to the binary log (C<SET SQL_LOG_BIN=1>).
Specifying C<--no-bin-log> will C<SET SQL_LOG_BIN=0>.
=item --buffer-in-mysql
Instruct MySQL to buffer queries in its memory.
This option adds the C<SQL_BUFFER_RESULT> option to the comparison queries.
This causes MySQL to execute the queries and place them in a temporary table
internally before sending the results back to pt-table-sync. The advantage of
this strategy is that pt-table-sync can fetch rows as desired without using a
lot of memory inside the Perl process, while releasing locks on the MySQL table
(to reduce contention with other queries). The disadvantage is that it uses
more memory on the MySQL server instead.
You probably want to leave L<"--[no]buffer-to-client"> enabled too, because
buffering into a temp table and then fetching it all into Perl's memory is
probably a silly thing to do. This option is most useful for the GroupBy and
Stream algorithms, which may fetch a lot of data from the server.
=item --[no]buffer-to-client
default: yes
Fetch rows one-by-one from MySQL while comparing.
This option enables C<mysql_use_result> which causes MySQL to hold the selected
rows on the server until the tool fetches them. This allows the tool to use
less memory but may keep the rows locked on the server longer.
If this option is disabled by specifying C<--no-buffer-to-client> then
C<mysql_store_result> is used which causes MySQL to send all selected rows to
the tool at once. This may result in the results "cursor" being held open for
a shorter time on the server, but if the tables are large, it could take a long
time anyway, and use all your memory.
For most non-trivial data sizes, you want to leave this option enabled.
This option is disabled when L<"--bidirectional"> is used.
=item --charset
short form: -A; type: string
Default character set. If the value is utf8, sets Perl's binmode on
STDOUT to utf8, passes the mysql_enable_utf8 option to DBD::mysql, and
runs SET NAMES UTF8 after connecting to MySQL. Any other value sets
binmode on STDOUT without the utf8 layer, and runs SET NAMES after
connecting to MySQL.
=item --[no]check-master
default: yes
With L<"--sync-to-master">, try to verify that the detected
master is the real master.
=item --[no]check-privileges
default: yes
Check that user has all necessary privileges on source and destination table.
=item --[no]check-slave
default: yes
Check whether the destination server is a slave.
If the destination server is a slave, it's generally unsafe to make changes on
it. However, sometimes you have to; L<"--replace"> won't work unless there's a
unique index, for example, so you can't make changes on the master in that
scenario. By default pt-table-sync will complain if you try to change data on
a slave. Specify C<--no-check-slave> to disable this check. Use it at your own
risk.
=item --[no]check-triggers
default: yes
Check that no triggers are defined on the destination table.
Triggers were introduced in MySQL v5.0.2, so for older versions this option
has no effect because triggers will not be checked.
=item --chunk-column
type: string
Chunk the table on this column.
=item --chunk-index
type: string
Chunk the table using this index.
=item --chunk-size
type: string; default: 1000
Number of rows or data size per chunk.
The size of each chunk of rows for the L<"Chunk"> and L<"Nibble"> algorithms.
The size can be either a number of rows, or a data size. Data sizes are
specified with a suffix of k=kibibytes, M=mebibytes, G=gibibytes. Data sizes
are converted to a number of rows by dividing by the average row length.
=item --columns
short form: -c; type: array
Compare this comma-separated list of columns.
=item --config
type: Array
Read this comma-separated list of config files; if specified, this must be the
first option on the command line.
=item --conflict-column
type: string
Compare this column when rows conflict during a L<"--bidirectional"> sync.
When a same but differing row is found the value of this column from each
row is compared according to L<"--conflict-comparison">, L<"--conflict-value">
and L<"--conflict-threshold"> to determine which row has the correct data and
becomes the source. The column can be any type for which there is an
appropriate L<"--conflict-comparison"> (this is almost all types except, for
example, blobs).
This option only works with L<"--bidirectional">.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --conflict-comparison
type: string
Choose the L<"--conflict-column"> with this property as the source.
The option affects how the L<"--conflict-column"> values from the conflicting
rows are compared. Possible comparisons are one of these MAGIC_comparisons:
newest|oldest|greatest|least|equals|matches
COMPARISON CHOOSES ROW WITH
========== =========================================================
newest Newest temporal --conflict-column value
oldest Oldest temporal --conflict-column value
greatest Greatest numerical "--conflict-column value
least Least numerical --conflict-column value
equals --conflict-column value equal to --conflict-value
matches --conflict-column value matching Perl regex pattern
--conflict-value
This option only works with L<"--bidirectional">.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --conflict-error
type: string; default: warn
How to report unresolvable conflicts and conflict errors
This option changes how the user is notified when a conflict cannot be
resolved or causes some kind of error. Possible values are:
* warn: Print a warning to STDERR about the unresolvable conflict
* die: Die, stop syncing, and print a warning to STDERR
This option only works with L<"--bidirectional">.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --conflict-threshold
type: string
Amount by which one L<"--conflict-column"> must exceed the other.
The L<"--conflict-threshold"> prevents a conflict from being resolved if
the absolute difference between the two L<"--conflict-column"> values is
less than this amount. For example, if two L<"--conflict-column"> have
timestamp values "2009-12-01 12:00:00" and "2009-12-01 12:05:00" the difference
is 5 minutes. If L<"--conflict-threshold"> is set to "5m" the conflict will
be resolved, but if L<"--conflict-threshold"> is set to "6m" the conflict
will fail to resolve because the difference is not greater than or equal
to 6 minutes. In this latter case, L<"--conflict-error"> will report
the failure.
This option only works with L<"--bidirectional">.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --conflict-value
type: string
Use this value for certain L<"--conflict-comparison">.
This option gives the value for C<equals> and C<matches>
L<"--conflict-comparison">.
This option only works with L<"--bidirectional">.
See L<"BIDIRECTIONAL SYNCING"> for more information.
=item --databases
short form: -d; type: hash
Sync only this comma-separated list of databases.
A common request is to sync tables from one database with tables from another
database on the same or different server. This is not yet possible.
L<"--databases"> will not do it, and you can't do it with the D part of the DSN
either because in the absence of a table name it assumes the whole server
should be synced and the D part controls only the connection's default database.
=item --defaults-file
short form: -F; type: string
Only read mysql options from the given file. You must give an absolute pathname.
=item --dry-run
Analyze, decide the sync algorithm to use, print and exit.
Implies L<"--verbose"> so you can see the results. The results are in the same
output format that you'll see from actually running the tool, but there will be
zeros for rows affected. This is because the tool actually executes, but stops
before it compares any data and just returns zeros. The zeros do not mean there
are no changes to be made.
=item --engines
short form: -e; type: hash
Sync only this comma-separated list of storage engines.
=item --execute
Execute queries to make the tables have identical data.
This option makes pt-table-sync actually sync table data by executing all
the queries that it created to resolve table differences. Therefore, B<the
tables will be changed!> And unless you also specify L<"--verbose">, the
changes will be made silently. If this is not what you want, see
L<"--print"> or L<"--dry-run">.
=item --explain-hosts
Print connection information and exit.
Print out a list of hosts to which pt-table-sync will connect, with all
the various connection options, and exit.
=item --float-precision
type: int
Precision for C<FLOAT> and C<DOUBLE> number-to-string conversion. Causes FLOAT
and DOUBLE values to be rounded to the specified number of digits after the
decimal point, with the ROUND() function in MySQL. This can help avoid
checksum mismatches due to different floating-point representations of the same
values on different MySQL versions and hardware. The default is no rounding;
the values are converted to strings by the CONCAT() function, and MySQL chooses
the string representation. If you specify a value of 2, for example, then the
values 1.008 and 1.009 will be rounded to 1.01, and will checksum as equal.
=item --[no]foreign-key-checks
default: yes
Enable foreign key checks (C<SET FOREIGN_KEY_CHECKS=1>).
Specifying C<--no-foreign-key-checks> will C<SET FOREIGN_KEY_CHECKS=0>.
=item --function
type: string
Which hash function you'd like to use for checksums.
The default is C<CRC32>. Other good choices include C<MD5> and C<SHA1>. If you
have installed the C<FNV_64> user-defined function, C<pt-table-sync> will detect
it and prefer to use it, because it is much faster than the built-ins. You can
also use MURMUR_HASH if you've installed that user-defined function. Both of
these are distributed with Maatkit. See L<pt-table-checksum> for more
information and benchmarks.
=item --help
Show help and exit.
=item --[no]hex-blob
default: yes
C<HEX()> C<BLOB>, C<TEXT> and C<BINARY> columns.
When row data from the source is fetched to create queries to sync the
data (i.e. the queries seen with L<"--print"> and executed by L<"--execute">),
binary columns are wrapped in HEX() so the binary data does not produce
an invalid SQL statement. You can disable this option but you probably
shouldn't.
=item --host
short form: -h; type: string
Connect to host.
=item --ignore-columns
type: Hash
Ignore this comma-separated list of column names in comparisons.
This option causes columns not to be compared. However, if a row is determined
to differ between tables, all columns in that row will be synced, regardless.
(It is not currently possible to exclude columns from the sync process itself,
only from the comparison.)
=item --ignore-databases
type: Hash
Ignore this comma-separated list of databases.
=item --ignore-engines
type: Hash; default: FEDERATED,MRG_MyISAM
Ignore this comma-separated list of storage engines.
=item --ignore-tables
type: Hash
Ignore this comma-separated list of tables.
Table names may be qualified with the database name.
=item --[no]index-hint
default: yes
Add FORCE/USE INDEX hints to the chunk and row queries.
By default C<pt-table-sync> adds a FORCE/USE INDEX hint to each SQL statement
to coerce MySQL into using the index chosen by the sync algorithm or specified
by L<"--chunk-index">. This is usually a good thing, but in rare cases the
index may not be the best for the query so you can suppress the index hint
by specifying C<--no-index-hint> and let MySQL choose the index.
This does not affect the queries printed by L<"--print">; it only affects the
chunk and row queries that C<pt-table-sync> uses to select and compare rows.
=item --lock
type: int
Lock tables: 0=none, 1=per sync cycle, 2=per table, or 3=globally.
This uses C<LOCK TABLES>. This can help prevent tables being changed while
you're examining them. The possible values are as follows:
VALUE MEANING
===== =======================================================
0 Never lock tables.
1 Lock and unlock one time per sync cycle (as implemented
by the syncing algorithm). This is the most granular
level of locking available. For example, the Chunk
algorithm will lock each chunk of C<N> rows, and then
unlock them if they are the same on the source and the
destination, before moving on to the next chunk.
2 Lock and unlock before and after each table.
3 Lock and unlock once for every server (DSN) synced, with
C<FLUSH TABLES WITH READ LOCK>.
A replication slave is never locked if L<"--replicate"> or L<"--sync-to-master">
is specified, since in theory locking the table on the master should prevent any
changes from taking place. (You are not changing data on your slave, right?)
If L<"--wait"> is given, the master (source) is locked and then the tool waits
for the slave to catch up to the master before continuing.
If C<--transaction> is specified, C<LOCK TABLES> is not used. Instead, lock
and unlock are implemented by beginning and committing transactions.
The exception is if L<"--lock"> is 3.
If C<--no-transaction> is specified, then C<LOCK TABLES> is used for any
value of L<"--lock">. See L<"--[no]transaction">.
=item --lock-and-rename
Lock the source and destination table, sync, then swap names. This is useful as
a less-blocking ALTER TABLE, once the tables are reasonably in sync with each
other (which you may choose to accomplish via any number of means, including
dump and reload or even something like L<pt-archiver>). It requires exactly two
DSNs and assumes they are on the same server, so it does no waiting for
replication or the like. Tables are locked with LOCK TABLES.
=item --password
short form: -p; type: string
Password to use when connecting.
=item --pid
type: string
Create the given PID file. The file contains the process ID of the script.
The PID file is removed when the script exits. Before starting, the script
checks if the PID file already exists. If it does not, then the script creates
and writes its own PID to it. If it does, then the script checks the following:
if the file contains a PID and a process is running with that PID, then
the script dies; or, if there is no process running with that PID, then the
script overwrites the file with its own PID and starts; else, if the file
contains no PID, then the script dies.
=item --port
short form: -P; type: int
Port number to use for connection.
=item --print
Print queries that will resolve differences.
If you don't trust C<pt-table-sync>, or just want to see what it will do, this
is a good way to be safe. These queries are valid SQL and you can run them
yourself if you want to sync the tables manually.
=item --recursion-method
type: string
Preferred recursion method used to find slaves.
Possible methods are:
METHOD USES
=========== ================
processlist SHOW PROCESSLIST
hosts SHOW SLAVE HOSTS
The processlist method is preferred because SHOW SLAVE HOSTS is not reliable.
However, the hosts method is required if the server uses a non-standard
port (not 3306). Usually pt-table-sync does the right thing and finds
the slaves, but you may give a preferred method and it will be used first.
If it doesn't find any slaves, the other methods will be tried.
=item --replace
Write all C<INSERT> and C<UPDATE> statements as C<REPLACE>.
This is automatically switched on as needed when there are unique index
violations.
=item --replicate
type: string
Sync tables listed as different in this table.
Specifies that C<pt-table-sync> should examine the specified table to find data
that differs. The table is exactly the same as the argument of the same name to
L<pt-table-checksum>. That is, it contains records of which tables (and ranges
of values) differ between the master and slave.
For each table and range of values that shows differences between the master and
slave, C<pt-table-checksum> will sync that table, with the appropriate C<WHERE>
clause, to its master.
This automatically sets L<"--wait"> to 60 and causes changes to be made on the
master instead of the slave.
If L<"--sync-to-master"> is specified, the tool will assume the server you
specified is the slave, and connect to the master as usual to sync.
Otherwise, it will try to use C<SHOW PROCESSLIST> to find slaves of the server
you specified. If it is unable to find any slaves via C<SHOW PROCESSLIST>, it
will inspect C<SHOW SLAVE HOSTS> instead. You must configure each slave's
C<report-host>, C<report-port> and other options for this to work right. After
finding slaves, it will inspect the specified table on each slave to find data
that needs to be synced, and sync it.
The tool examines the master's copy of the table first, assuming that the master
is potentially a slave as well. Any table that shows differences there will
B<NOT> be synced on the slave(s). For example, suppose your replication is set
up as A->B, B->C, B->D. Suppose you use this argument and specify server B.
The tool will examine server B's copy of the table. If it looks like server B's
data in table C<test.tbl1> is different from server A's copy, the tool will not
sync that table on servers C and D.
=item --set-vars
type: string; default: wait_timeout=10000
Set these MySQL variables. Immediately after connecting to MySQL, this
string will be appended to SET and executed.
=item --socket
short form: -S; type: string
Socket file to use for connection.
=item --sync-to-master
Treat the DSN as a slave and sync it to its master.
Treat the server you specified as a slave. Inspect C<SHOW SLAVE STATUS>,
connect to the server's master, and treat the master as the source and the slave
as the destination. Causes changes to be made on the master. Sets L<"--wait">
to 60 by default, sets L<"--lock"> to 1 by default, and disables
L<"--[no]transaction"> by default. See also L<"--replicate">, which changes
this option's behavior.
=item --tables
short form: -t; type: hash
Sync only this comma-separated list of tables.
Table names may be qualified with the database name.
=item --timeout-ok
Keep going if L<"--wait"> fails.
If you specify L<"--wait"> and the slave doesn't catch up to the master's
position before the wait times out, the default behavior is to abort. This
option makes the tool keep going anyway. B<Warning>: if you are trying to get a
consistent comparison between the two servers, you probably don't want to keep
going after a timeout.
=item --[no]transaction
Use transactions instead of C<LOCK TABLES>.
The granularity of beginning and committing transactions is controlled by
L<"--lock">. This is enabled by default, but since L<"--lock"> is disabled by
default, it has no effect.
Most options that enable locking also disable transactions by default, so if
you want to use transactional locking (via C<LOCK IN SHARE MODE> and C<FOR
UPDATE>, you must specify C<--transaction> explicitly.
If you don't specify C<--transaction> explicitly C<pt-table-sync> will decide on
a per-table basis whether to use transactions or table locks. It currently
uses transactions on InnoDB tables, and table locks on all others.
If C<--no-transaction> is specified, then C<pt-table-sync> will not use
transactions at all (not even for InnoDB tables) and locking is controlled
by L<"--lock">.
When enabled, either explicitly or implicitly, the transaction isolation level
is set C<REPEATABLE READ> and transactions are started C<WITH CONSISTENT
SNAPSHOT>.
=item --trim
C<TRIM()> C<VARCHAR> columns in C<BIT_XOR> and C<ACCUM> modes. Helps when
comparing MySQL 4.1 to >= 5.0.
This is useful when you don't care about the trailing space differences between
MySQL versions which vary in their handling of trailing spaces. MySQL 5.0 and
later all retain trailing spaces in C<VARCHAR>, while previous versions would
remove them.
=item --[no]unique-checks
default: yes
Enable unique key checks (C<SET UNIQUE_CHECKS=1>).
Specifying C<--no-unique-checks> will C<SET UNIQUE_CHECKS=0>.
=item --user
short form: -u; type: string
User for login if not current user.
=item --verbose
short form: -v; cumulative: yes
Print results of sync operations.
See L<"OUTPUT"> for more details about the output.
=item --version
Show version and exit.
=item --wait
short form: -w; type: time
How long to wait for slaves to catch up to their master.
Make the master wait for the slave to catch up in replication before comparing
the tables. The value is the number of seconds to wait before timing out (see
also L<"--timeout-ok">). Sets L<"--lock"> to 1 and L<"--[no]transaction"> to 0
by default. If you see an error such as the following,
MASTER_POS_WAIT returned -1
It means the timeout was exceeded and you need to increase it.
The default value of this option is influenced by other options. To see what
value is in effect, run with L<"--help">.
To disable waiting entirely (except for locks), specify L<"--wait"> 0. This
helps when the slave is lagging on tables that are not being synced.
=item --where
type: string
C<WHERE> clause to restrict syncing to part of the table.
=item --[no]zero-chunk
default: yes
Add a chunk for rows with zero or zero-equivalent values. The only has an
effect when L<"--chunk-size"> is specified. The purpose of the zero chunk
is to capture a potentially large number of zero values that would imbalance
the size of the first chunk. For example, if a lot of negative numbers were
inserted into an unsigned integer column causing them to be stored as zeros,
then these zero values are captured by the zero chunk instead of the first
chunk and all its non-zero values.
=back
=head1 DSN OPTIONS
These DSN options are used to create a DSN. Each option is given like
C<option=value>. The options are case-sensitive, so P and p are not the
same option. There cannot be whitespace before or after the C<=> and
if the value contains whitespace it must be quoted. DSN options are
comma-separated. See the L<percona-toolkit> manpage for full details.
=over
=item * A
dsn: charset; copy: yes
Default character set.
=item * D
dsn: database; copy: yes
Database containing the table to be synced.
=item * F
dsn: mysql_read_default_file; copy: yes
Only read default options from the given file
=item * h
dsn: host; copy: yes
Connect to host.
=item * p
dsn: password; copy: yes
Password to use when connecting.
=item * P
dsn: port; copy: yes
Port number to use for connection.
=item * S
dsn: mysql_socket; copy: yes
Socket file to use for connection.
=item * t
copy: yes
Table to be synced.
=item * u
dsn: user; copy: yes
User for login if not current user.
=back
=head1 ENVIRONMENT
The environment variable C<PTDEBUG> enables verbose debugging output to STDERR.
To enable debugging and capture all output to a file, run the tool like:
PTDEBUG=1 pt-table-sync ... > FILE 2>&1
Be careful: debugging output is voluminous and can generate several megabytes
of output.
=head1 SYSTEM REQUIREMENTS
You need Perl, DBI, DBD::mysql, and some core packages that ought to be
installed in any reasonably new version of Perl.
=head1 BUGS
For a list of known bugs, see L<http://www.percona.com/bugs/pt-table-sync>.
Please report bugs at L<https://bugs.launchpad.net/percona-toolkit>.
Include the following information in your bug report:
=over
=item * Complete command-line used to run the tool
=item * Tool L<"--version">
=item * MySQL version of all servers involved
=item * Output from the tool including STDERR
=item * Input files (log/dump/config files, etc.)
=back
If possible, include debugging output by running the tool with C<PTDEBUG>;
see L<"ENVIRONMENT">.
=head1 DOWNLOADING
Visit L<http://www.percona.com/software/percona-toolkit/> to download the
latest release of Percona Toolkit. Or, get the latest release from the
command line:
wget percona.com/get/percona-toolkit.tar.gz
wget percona.com/get/percona-toolkit.rpm
wget percona.com/get/percona-toolkit.deb
You can also get individual tools from the latest release:
wget percona.com/get/TOOL
Replace C<TOOL> with the name of any tool.
=head1 AUTHORS
Baron Schwartz
=head1 ACKNOWLEDGMENTS
My work is based in part on Giuseppe Maxia's work on distributed databases,
L<http://www.sysadminmag.com/articles/2004/0408/> and code derived from that
article. There is more explanation, and a link to the code, at
L<http://www.perlmonks.org/?node_id=381053>.
Another programmer extended Maxia's work even further. Fabien Coelho changed
and generalized Maxia's technique, introducing symmetry and avoiding some
problems that might have caused too-frequent checksum collisions. This work
grew into pg_comparator, L<http://www.coelho.net/pg_comparator/>. Coelho also
explained the technique further in a paper titled "Remote Comparison of Database
Tables" (L<http://cri.ensmp.fr/classement/doc/A-375.pdf>).
This existing literature mostly addressed how to find the differences between
the tables, not how to resolve them once found. I needed a tool that would not
only find them efficiently, but would then resolve them. I first began thinking
about how to improve the technique further with my article
L<http://tinyurl.com/mysql-data-diff-algorithm>,
where I discussed a number of problems with the Maxia/Coelho "bottom-up"
algorithm. After writing that article, I began to write this tool. I wanted to
actually implement their algorithm with some improvements so I was sure I
understood it completely. I discovered it is not what I thought it was, and is
considerably more complex than it appeared to me at first. Fabien Coelho was
kind enough to address some questions over email.
The first versions of this tool implemented a version of the Coelho/Maxia
algorithm, which I called "bottom-up", and my own, which I called "top-down."
Those algorithms are considerably more complex than the current algorithms and
I have removed them from this tool, and may add them back later. The
improvements to the bottom-up algorithm are my original work, as is the
top-down algorithm. The techniques to actually resolve the differences are
also my own work.
Another tool that can synchronize tables is the SQLyog Job Agent from webyog.
Thanks to Rohit Nadhani, SJA's author, for the conversations about the general
techniques. There is a comparison of pt-table-sync and SJA at
L<http://tinyurl.com/maatkit-vs-sqlyog>
Thanks to the following people and organizations for helping in many ways:
The Rimm-Kaufman Group L<http://www.rimmkaufman.com/>,
MySQL AB L<http://www.mysql.com/>,
Blue Ridge InternetWorks L<http://www.briworks.com/>,
Percona L<http://www.percona.com/>,
Fabien Coelho,
Giuseppe Maxia and others at MySQL AB,
Kristian Koehntopp (MySQL AB),
Rohit Nadhani (WebYog),
The helpful monks at Perlmonks,
And others too numerous to mention.
=head1 ABOUT PERCONA TOOLKIT
This tool is part of Percona Toolkit, a collection of advanced command-line
tools developed by Percona for MySQL support and consulting. Percona Toolkit
was forked from two projects in June, 2011: Maatkit and Aspersa. Those
projects were created by Baron Schwartz and developed primarily by him and
Daniel Nichter, both of whom are employed by Percona. Visit
L<http://www.percona.com/software/> for more software developed by Percona.
=head1 COPYRIGHT, LICENSE, AND WARRANTY
This program is copyright 2007-2011 Baron Schwartz, 2011 Percona Inc.
Feedback and improvements are welcome.
THIS PROGRAM IS PROVIDED "AS IS" AND WITHOUT ANY EXPRESS OR IMPLIED
WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE IMPLIED WARRANTIES OF
MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation, version 2; OR the Perl Artistic License. On UNIX and similar
systems, you can issue `man perlgpl' or `man perlartistic' to read these
licenses.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc., 59 Temple
Place, Suite 330, Boston, MA 02111-1307 USA.
=head1 VERSION
pt-table-sync 1.0.1
=cut
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment