--- trunk/bin/BackupPC_updatedb 2005/09/11 12:39:24 116 +++ trunk/bin/BackupPC_updatedb 2005/10/10 13:39:11 160 @@ -11,6 +11,7 @@ use Time::HiRes qw/time/; use File::Pid; use POSIX qw/strftime/; +use BackupPC::SearchLib; use constant BPC_FTYPE_DIR => 5; use constant EST_CHUNK => 100000; @@ -43,24 +44,13 @@ my $user = $Conf{SearchUser} || ''; my $use_hest = $Conf{HyperEstraierIndex}; -my ($index_path, $index_node_url); -if ($use_hest) { - use HyperEstraier; - if ($use_hest =~ m#^http://#) { - $index_node_url = $use_hest; - } else { - $index_path = $TopDir . '/' . $index_path; - $index_path =~ s#//#/#g; - } -} -print "-- $use_hest : $index_path OR $index_node_url --\n"; - +my ($index_path, $index_node_url) = BackupPC::SearchLib::getHyperEstraier_url($use_hest); my $dbh = DBI->connect($dsn, $user, "", { RaiseError => 1, AutoCommit => 0 }); my %opt; -if ( !getopts("cdm:v:i", \%opt ) ) { +if ( !getopts("cdm:v:ij", \%opt ) ) { print STDERR <new(); - $hest_db->open($index_path, $HyperEstraier::Database::DBWRITER | $HyperEstraier::Database::DBCREAT); + $hest_db->open($TopDir . $index_path, $HyperEstraier::Database::DBWRITER | $HyperEstraier::Database::DBCREAT); print " directly"; } elsif ($index_node_url) { $hest_node ||= HyperEstraier::Node->new($index_node_url); @@ -200,8 +197,10 @@ my $fid = $row->{'fid'} || die "no fid?"; my $uri = 'file:///' . $fid; - my $id = ($hest_db || $hest_node)->uri_to_id($uri); - next unless ($id == -1); + unless ($skip_check) { + my $id = ($hest_db || $hest_node)->uri_to_id($uri); + next unless ($id == -1); + } # create a document object my $doc = HyperEstraier::Document->new; @@ -257,11 +256,12 @@ ## update index ## -if (($opt{i} || ($index_path && ! -e $index_path)) && !$opt{c}) { +if (($opt{i} || $opt{j} || ($index_path && ! -e $index_path)) && !$opt{c}) { # update all print "force update of HyperEstraier index "; print "importing existing data" unless (-e $index_path); print "by -i flag" if ($opt{i}); + print "by -j flag" if ($opt{j}); print "\n"; hest_update(); } @@ -270,9 +270,10 @@ if ($opt{c}) { sub do_index { my $index = shift || return; - my ($table,$col,$unique) = split(/_/, $index); + my ($table,$col,$unique) = split(/:/, $index); $unique ||= ''; - $index =~ s/,/_/g; + $index =~ s/\W+/_/g; + print "$index on $table($col)" . ( $unique ? "u" : "" ) . " "; $dbh->do(qq{ create $unique index $index on $table($col) }); } @@ -291,65 +292,97 @@ ID SERIAL PRIMARY KEY, hostID INTEGER NOT NULL references hosts(id), name VARCHAR(30) NOT NULL, - share VARCHAR(200) NOT NULL, - localpath VARCHAR(200) + share VARCHAR(200) NOT NULL ); }); + + $dbh->do(qq{ + create table dvds ( + ID SERIAL PRIMARY KEY, + num INTEGER NOT NULL, + name VARCHAR(255) NOT NULL, + mjesto VARCHAR(255) + ); + }); $dbh->do(qq{ create table backups ( + id serial, hostID INTEGER NOT NULL references hosts(id), num INTEGER NOT NULL, date integer NOT NULL, type CHAR(4) not null, shareID integer not null references shares(id), - size integer not null, - PRIMARY KEY(hostID, num, shareID) + size bigint not null, + inc_size bigint not null default -1, + inc_deleted boolean default false, + PRIMARY KEY(id) ); }); - #do_index('backups_hostid,num_unique'); - - $dbh->do(qq{ - create table dvds ( - ID SERIAL PRIMARY KEY, - num INTEGER NOT NULL, - name VARCHAR(255) NOT NULL, - mjesto VARCHAR(255) + $dbh->do(qq{ + create table files ( + ID SERIAL, + shareID INTEGER NOT NULL references shares(id), + backupNum INTEGER NOT NULL, + name VARCHAR(255) NOT NULL, + path VARCHAR(255) NOT NULL, + date integer NOT NULL, + type INTEGER NOT NULL, + size bigint NOT NULL, + primary key(id) ); }); - $dbh->do(qq{ - create table files ( - ID SERIAL PRIMARY KEY, - shareID INTEGER NOT NULL references shares(id), - backupNum INTEGER NOT NULL, - name VARCHAR(255) NOT NULL, - path VARCHAR(255) NOT NULL, - date integer NOT NULL, - type INTEGER NOT NULL, - size INTEGER NOT NULL, - dvdid INTEGER references dvds(id) + + $dbh->do( qq{ + create table archive ( + id serial, + dvd_nr int not null, + total_size bigint default -1, + note text, + username varchar(20) not null, + date timestamp default now(), + primary key(id) + ); + } + ); + + $dbh->do( qq{ + create table archive_backup + ( + archive_id int not null references archive(id) on delete cascade, + backup_id int not null references backups(id), + primary key(archive_id, backup_id) ); }); - print "creating indexes:"; + print "creating indexes: "; foreach my $index (qw( - hosts_name - backups_hostID - backups_num - shares_hostID - shares_name - files_shareID - files_path - files_name - files_date - files_size + hosts:name + backups:hostID + backups:num + backups:shareID + shares:hostID + shares:name + files:shareID + files:path + files:name + files:date + files:size + archive:dvd_nr )) { - print " $index"; do_index($index); } + + print " creating sequence: "; + foreach my $seq (qw/dvd_nr/) { + print "$seq "; + $dbh->do( qq{ CREATE SEQUENCE $seq } ); + } + + print "...\n"; $dbh->commit; @@ -465,7 +498,7 @@ $hostID, $backupNum, $backup->{'endTime'}, - $backup->{'type'}, + substr($backup->{'type'},0,4), $shareID, $size, ); @@ -510,14 +543,14 @@ $sth->{insert_share} ||= $dbh->prepare(qq{ INSERT INTO shares - (hostID,name,share,localpath) - VALUES (?,?,?,?) + (hostID,name,share) + VALUES (?,?,?) }); my $drop_down = $hostname . '/' . $share; $drop_down =~ s#//+#/#g; - $sth->{insert_share}->execute($hostID,$share, $drop_down ,undef); + $sth->{insert_share}->execute($hostID,$share, $drop_down); return $dbh->last_insert_id(undef,undef,'shares',undef); }