4 |
|
|
5 |
use Cwd qw/abs_path/; |
use Cwd qw/abs_path/; |
6 |
use File::Temp qw/tempdir/; |
use File::Temp qw/tempdir/; |
|
use Data::Dumper; |
|
7 |
use lib './lib'; |
use lib './lib'; |
8 |
|
|
9 |
use WebPAC::Lookup; |
use WebPAC::Common 0.02; |
10 |
use WebPAC::Input::ISIS; |
use WebPAC::Parser 0.08; |
11 |
use WebPAC::Store 0.03; |
use WebPAC::Input 0.14; |
12 |
use WebPAC::Normalize::XML; |
use WebPAC::Store 0.14; |
13 |
|
use WebPAC::Normalize 0.22; |
14 |
use WebPAC::Output::TT; |
use WebPAC::Output::TT; |
15 |
use WebPAC::Output::Estraier 0.02; |
use WebPAC::Validate 0.06; |
16 |
use YAML qw/LoadFile/; |
use WebPAC::Output::MARC; |
17 |
use LWP::Simple; |
use WebPAC::Config; |
18 |
|
use Getopt::Long; |
19 |
|
use File::Path; |
20 |
|
use Time::HiRes qw/time/; |
21 |
|
use File::Slurp; |
22 |
|
use Data::Dump qw/dump/; |
23 |
|
use Storable qw/dclone/; |
24 |
|
|
25 |
my $limit = shift @ARGV; |
use Proc::Queue size => 1; |
26 |
|
use POSIX ":sys_wait_h"; # imports WNOHANG |
27 |
|
|
28 |
my $config = LoadFile('conf/config.yml'); |
=head1 NAME |
29 |
|
|
30 |
print "config = ",Dumper($config); |
run.pl - start WebPAC indexing |
31 |
|
|
32 |
die "no databases in config file!\n" unless ($config->{databases}); |
B<this command will probably go away. Don't get used to it!> |
33 |
|
|
34 |
|
Options: |
35 |
|
|
36 |
|
=over 4 |
37 |
|
|
38 |
|
=item --offset 42 |
39 |
|
|
40 |
|
start loading (all) databases at offset 42 |
41 |
|
|
42 |
|
=item --limit 100 |
43 |
|
|
44 |
|
limit loading to 100 records |
45 |
|
|
46 |
|
=item --clean |
47 |
|
|
48 |
|
remove database and Hyper Estraier index before indexing |
49 |
|
|
50 |
|
=item --only=database_name/input_filter |
51 |
|
|
52 |
|
reindex just single database (legacy name is --one) |
53 |
|
|
54 |
|
C</input_filter> is optional part which can be C<name> |
55 |
|
or C<type> from input |
56 |
|
|
57 |
|
=item --config conf/config.yml |
58 |
|
|
59 |
|
path to YAML configuration file |
60 |
|
|
61 |
|
=item --stats |
62 |
|
|
63 |
|
disable indexing, modify_* in configuration and dump statistics about field |
64 |
|
and subfield usage for each input |
65 |
|
|
66 |
|
=item --validate path/to/validation_file |
67 |
|
|
68 |
|
turn on extra validation of imput records, see L<WebPAC::Validation> |
69 |
|
|
70 |
|
=item --marc-lint |
71 |
|
|
72 |
|
By default turned on if normalisation file has C<marc*> directives. You can disable lint |
73 |
|
messages with C<--no-marc-lint>. |
74 |
|
|
75 |
|
=item --marc-dump |
76 |
|
|
77 |
|
Force dump or input and marc record for debugging. |
78 |
|
|
79 |
|
=item --parallel 4 |
80 |
|
|
81 |
|
Run databases in parallel (aproximatly same as number of processors in |
82 |
|
machine if you want to use full load) |
83 |
|
|
84 |
|
=item --only-links |
85 |
|
|
86 |
|
Create just links |
87 |
|
|
88 |
|
=item --merge |
89 |
|
|
90 |
|
Create merged index of databases which have links |
91 |
|
|
92 |
|
=back |
93 |
|
|
94 |
|
=cut |
95 |
|
|
96 |
|
my $offset; |
97 |
|
my $limit; |
98 |
|
|
99 |
|
my $clean = 0; |
100 |
|
my $config_path; |
101 |
|
my $debug = 0; |
102 |
|
my $only_filter; |
103 |
|
my $stats = 0; |
104 |
|
my $validate_path; |
105 |
|
my $marc_lint = 1; |
106 |
|
my $marc_dump = 0; |
107 |
|
my $parallel = 0; |
108 |
|
my $only_links = 0; |
109 |
|
my $merge = 0; |
110 |
|
|
111 |
|
my $log = _new WebPAC::Common()->_get_logger(); |
112 |
|
|
113 |
|
GetOptions( |
114 |
|
"limit=i" => \$limit, |
115 |
|
"offset=i" => \$offset, |
116 |
|
"clean" => \$clean, |
117 |
|
"one=s" => \$only_filter, |
118 |
|
"only=s" => \$only_filter, |
119 |
|
"config" => \$config_path, |
120 |
|
"debug+" => \$debug, |
121 |
|
"stats" => \$stats, |
122 |
|
"validate=s" => \$validate_path, |
123 |
|
"marc-lint!" => \$marc_lint, |
124 |
|
"marc-dump!" => \$marc_dump, |
125 |
|
"parallel=i" => \$parallel, |
126 |
|
"only-links!" => \$only_links, |
127 |
|
"merge" => \$merge, |
128 |
|
); |
129 |
|
|
130 |
|
my $config = new WebPAC::Config( path => $config_path ); |
131 |
|
|
132 |
|
#print "config = ",dump($config) if ($debug); |
133 |
|
|
134 |
|
die "no databases in config file!\n" unless ($config->databases); |
135 |
|
|
136 |
|
$log->info( "-" x 79 ); |
137 |
|
|
138 |
|
|
139 |
|
my $estcmd_fh; |
140 |
|
my $estcmd_path = './estcmd-merge.sh'; |
141 |
|
if ($merge) { |
142 |
|
open($estcmd_fh, '>', $estcmd_path) || $log->logdie("can't open $estcmd_path: $!"); |
143 |
|
print $estcmd_fh 'cd /data/estraier/_node/ || exit 1',$/; |
144 |
|
print $estcmd_fh 'sudo /etc/init.d/hyperestraier stop',$/; |
145 |
|
$log->info("created merge batch file $estcmd_path"); |
146 |
|
} |
147 |
|
|
148 |
|
|
149 |
|
my $validate; |
150 |
|
$validate = new WebPAC::Validate( |
151 |
|
path => $validate_path, |
152 |
|
) if ($validate_path); |
153 |
|
|
154 |
|
|
155 |
|
my $use_indexer = $config->use_indexer; |
156 |
|
if ($stats) { |
157 |
|
$log->debug("option --stats disables update of indexing engine..."); |
158 |
|
$use_indexer = undef; |
159 |
|
} else { |
160 |
|
$log->info("using $use_indexer indexing engine..."); |
161 |
|
} |
162 |
|
|
163 |
|
# parse normalize files and create source files for lookup and normalization |
164 |
|
|
165 |
|
my $parser = new WebPAC::Parser( config => $config ); |
166 |
|
|
167 |
my $total_rows = 0; |
my $total_rows = 0; |
168 |
|
my $start_t = time(); |
169 |
|
|
170 |
while (my ($database, $db_config) = each %{ $config->{databases} }) { |
my @links; |
171 |
|
|
172 |
my $type = lc($db_config->{input}->{type}); |
if ($parallel) { |
173 |
|
$log->info("Using $parallel processes for speedup"); |
174 |
|
Proc::Queue::size($parallel); |
175 |
|
} |
176 |
|
|
177 |
die "I know only how to handle input type isis, not '$type'!\n" unless ($type eq 'isis'); |
sub create_ds_config { |
178 |
|
my ($db_config, $database, $input, $mfn) = @_; |
179 |
|
my $c = dclone( $db_config ); |
180 |
|
$c->{_} = $database || $log->logconfess("need database"); |
181 |
|
$c->{_mfn} = $mfn || $log->logconfess("need mfn"); |
182 |
|
$c->{input} = $input || $log->logconfess("need input"); |
183 |
|
return $c; |
184 |
|
} |
185 |
|
|
186 |
my $abs_path = abs_path($0); |
while (my ($database, $db_config) = each %{ $config->databases }) { |
|
$abs_path =~ s#/[^/]*$#/#; |
|
187 |
|
|
188 |
my $lookup = new WebPAC::Lookup( |
my ($only_database,$only_input) = split(m#/#, $only_filter) if ($only_filter); |
189 |
lookup_file => $db_config->{input}->{lookup}, |
next if ($only_database && $database !~ m/$only_database/i); |
|
); |
|
190 |
|
|
191 |
my $db_path = $config->{webpac}->{db_path} . '/' . $database; |
if ($parallel) { |
192 |
|
my $f=fork; |
193 |
|
if(defined ($f) and $f==0) { |
194 |
|
$log->info("Created processes $$ for speedup"); |
195 |
|
} else { |
196 |
|
next; |
197 |
|
} |
198 |
|
} |
199 |
|
|
200 |
|
my $indexer; |
201 |
|
if ($use_indexer && $parser->have_rules( 'search', $database )) { |
202 |
|
|
203 |
my $log = $lookup->_get_logger; |
my $cfg_name = $use_indexer; |
204 |
$log->info("working on $database in $db_path"); |
$cfg_name =~ s/\-.*$//; |
205 |
|
|
206 |
my $path = './db/'; |
my $indexer_config = $config->get( $cfg_name ) || $log->logdie("can't find '$cfg_name' part in confguration"); |
207 |
|
$indexer_config->{database} = $database; |
208 |
|
$indexer_config->{clean} = $clean; |
209 |
|
$indexer_config->{label} = $db_config->{name}; |
210 |
|
|
211 |
my $db = new WebPAC::Store( |
# force clean if database has links |
212 |
path => $db_path, |
$indexer_config->{clean} = 1 if ($db_config->{links}); |
213 |
); |
|
214 |
|
if ($use_indexer eq 'hyperestraier') { |
215 |
|
|
216 |
|
# open Hyper Estraier database |
217 |
|
use WebPAC::Output::Estraier '0.10'; |
218 |
|
$indexer = new WebPAC::Output::Estraier( %{ $indexer_config } ); |
219 |
|
|
220 |
|
} elsif ($use_indexer eq 'hyperestraier-native') { |
221 |
|
|
222 |
|
# open Hyper Estraier database |
223 |
|
use WebPAC::Output::EstraierNative; |
224 |
|
$indexer = new WebPAC::Output::EstraierNative( %{ $indexer_config } ); |
225 |
|
|
226 |
|
} elsif ($use_indexer eq 'kinosearch') { |
227 |
|
|
228 |
|
# open KinoSearch |
229 |
|
use WebPAC::Output::KinoSearch; |
230 |
|
$indexer_config->{clean} = 1 unless (-e $indexer_config->{index_path}); |
231 |
|
$indexer = new WebPAC::Output::KinoSearch( %{ $indexer_config } ); |
232 |
|
|
233 |
|
} else { |
234 |
|
$log->logdie("unknown use_indexer: $use_indexer"); |
235 |
|
} |
236 |
|
|
237 |
|
$log->logide("can't continue without valid indexer") unless ($indexer); |
238 |
|
} |
239 |
|
|
240 |
|
|
241 |
|
# |
242 |
|
# store Hyper Estraier links to other databases |
243 |
|
# |
244 |
|
if (ref($db_config->{links}) eq 'ARRAY' && $use_indexer) { |
245 |
|
foreach my $link (@{ $db_config->{links} }) { |
246 |
|
if ($use_indexer eq 'hyperestraier') { |
247 |
|
if ($merge) { |
248 |
|
print $estcmd_fh 'sudo -u www-data estcmd merge ' . $database . ' ' . $link->{to},$/; |
249 |
|
} else { |
250 |
|
$log->info("saving link $database -> $link->{to} [$link->{credit}]"); |
251 |
|
push @links, sub { |
252 |
|
$log->info("adding link $database -> $link->{to} [$link->{credit}]"); |
253 |
|
$indexer->add_link( |
254 |
|
from => $database, |
255 |
|
to => $link->{to}, |
256 |
|
credit => $link->{credit}, |
257 |
|
); |
258 |
|
}; |
259 |
|
} |
260 |
|
} else { |
261 |
|
$log->warn("NOT IMPLEMENTED WITH $use_indexer: adding link $database -> $link->{to} [$link->{credit}]"); |
262 |
|
} |
263 |
|
} |
264 |
|
} |
265 |
|
next if ($only_links); |
266 |
|
|
267 |
|
|
268 |
|
# |
269 |
|
# now WebPAC::Store |
270 |
|
# |
271 |
|
my $abs_path = abs_path($0); |
272 |
|
$abs_path =~ s#/[^/]*$#/#; |
273 |
|
|
274 |
my $est_config = $config->{hyperestraier} || $log->logdie("can't find 'hyperestraier' part in confguration"); |
my $db_path = $config->webpac('db_path'); |
|
$est_config->{database} = $database; |
|
275 |
|
|
276 |
$log->info("using HyperEstraier URL $est_config->{masterurl}"); |
if ($clean) { |
277 |
|
$log->info("creating new database '$database' in $db_path"); |
278 |
|
rmtree( $db_path ) || $log->warn("can't remove $db_path: $!"); |
279 |
|
} else { |
280 |
|
$log->info("working on database '$database' in $db_path"); |
281 |
|
} |
282 |
|
|
283 |
my $est = new WebPAC::Output::Estraier( |
my $store = new WebPAC::Store( |
284 |
%{ $est_config }, |
path => $db_path, |
285 |
|
debug => $debug, |
286 |
); |
); |
287 |
|
|
288 |
|
|
289 |
# |
# |
290 |
# now, iterate through input formats |
# now, iterate through input formats |
291 |
# |
# |
293 |
my @inputs; |
my @inputs; |
294 |
if (ref($db_config->{input}) eq 'ARRAY') { |
if (ref($db_config->{input}) eq 'ARRAY') { |
295 |
@inputs = @{ $db_config->{input} }; |
@inputs = @{ $db_config->{input} }; |
296 |
} else { |
} elsif ($db_config->{input}) { |
297 |
push @inputs, $db_config->{input}; |
push @inputs, $db_config->{input}; |
298 |
|
} else { |
299 |
|
$log->info("database $database doesn't have inputs defined"); |
300 |
} |
} |
301 |
|
|
302 |
foreach my $input (@inputs) { |
foreach my $input (@inputs) { |
|
$log->info("working on input $input->{path} [$input->{type}]"); |
|
303 |
|
|
304 |
my $isis = new WebPAC::Input::ISIS( |
my $input_name = $input->{name} || $log->logdie("input without a name isn't valid: ",dump($input)); |
|
code_page => $config->{webpac}->{webpac_encoding}, |
|
|
limit_mfn => $input->{limit}, |
|
|
); |
|
305 |
|
|
306 |
my $maxmfn = $isis->open( |
next if ($only_input && ($input_name !~ m#$only_input#i && $input->{type} !~ m#$only_input#i)); |
307 |
filename => $input->{path}, |
|
308 |
code_page => $input->{encoding}, # database encoding |
my $type = lc($input->{type}); |
309 |
|
|
310 |
|
die "I know only how to handle input types ", join(",", $config->webpac('inputs') ), " not '$type'!\n" unless (grep(/$type/, $config->webpac('inputs'))); |
311 |
|
|
312 |
|
my $input_module = $config->webpac('inputs')->{$type}; |
313 |
|
|
314 |
|
my @lookups = $parser->have_lookup_create($database, $input); |
315 |
|
|
316 |
|
$log->info("working on input '$input_name' in $input->{path} [type: $input->{type}] using $input_module", |
317 |
|
@lookups ? " creating lookups: ".join(", ", @lookups) : "" |
318 |
); |
); |
319 |
|
|
320 |
my $n = new WebPAC::Normalize::XML( |
if ($stats) { |
321 |
# filter => { 'foo' => sub { shift } }, |
# disable modification of records if --stats is in use |
322 |
db => $db, |
delete($input->{modify_records}); |
323 |
lookup_regex => $lookup->regex, |
delete($input->{modify_file}); |
324 |
lookup => $lookup, |
} |
325 |
|
|
326 |
|
my $input_db = new WebPAC::Input( |
327 |
|
module => $input_module, |
328 |
|
encoding => $config->webpac('webpac_encoding'), |
329 |
|
limit => $limit || $input->{limit}, |
330 |
|
offset => $offset, |
331 |
|
recode => $input->{recode}, |
332 |
|
stats => $stats, |
333 |
|
modify_records => $input->{modify_records}, |
334 |
|
modify_file => $input->{modify_file}, |
335 |
); |
); |
336 |
|
$log->logdie("can't create input using $input_module") unless ($input); |
337 |
|
|
338 |
|
if (defined( $input->{lookup} )) { |
339 |
|
$log->warn("$database/$input_name has depriciated lookup definition, removing it..."); |
340 |
|
delete( $input->{lookup} ); |
341 |
|
} |
342 |
|
|
343 |
|
my $lookup_coderef; |
344 |
|
|
345 |
|
if (@lookups) { |
346 |
|
|
347 |
|
my $rules = $parser->lookup_create_rules($database, $input) || $log->logdie("no rules found for $database/$input"); |
348 |
|
|
349 |
|
$lookup_coderef = sub { |
350 |
|
my $rec = shift || die "need rec!"; |
351 |
|
my $mfn = $rec->{'000'}->[0] || die "need mfn in 000"; |
352 |
|
|
353 |
|
WebPAC::Normalize::data_structure( |
354 |
|
row => $rec, |
355 |
|
rules => $rules, |
356 |
|
config => create_ds_config( $db_config, $database, $input, $mfn ), |
357 |
|
); |
358 |
|
|
359 |
|
#warn "current lookup: ", dump(WebPAC::Normalize::_get_lookup()); |
360 |
|
}; |
361 |
|
|
362 |
|
WebPAC::Normalize::_set_lookup( undef ); |
363 |
|
|
364 |
|
$log->debug("created lookup_coderef using:\n$rules"); |
365 |
|
|
366 |
|
}; |
367 |
|
|
368 |
|
my $lookup_jar; |
369 |
|
|
370 |
|
my $maxmfn = $input_db->open( |
371 |
|
path => $input->{path}, |
372 |
|
code_page => $input->{encoding}, # database encoding |
373 |
|
lookup_coderef => $lookup_coderef, |
374 |
|
lookup => $lookup_jar, |
375 |
|
%{ $input }, |
376 |
|
load_row => sub { |
377 |
|
my $a = shift; |
378 |
|
return $store->load_row( |
379 |
|
database => $database, |
380 |
|
input => $input_name, |
381 |
|
id => $a->{id}, |
382 |
|
); |
383 |
|
}, |
384 |
|
save_row => sub { |
385 |
|
my $a = shift; |
386 |
|
return $store->save_row( |
387 |
|
database => $database, |
388 |
|
input => $input_name, |
389 |
|
id => $a->{id}, |
390 |
|
row => $a->{row}, |
391 |
|
); |
392 |
|
}, |
393 |
|
|
|
$n->open( |
|
|
tag => $input->{normalize}->{tag}, |
|
|
xml_file => $input->{normalize}->{path}, |
|
394 |
); |
); |
395 |
|
|
396 |
for ( 0 ... $isis->size ) { |
my $lookup_data = WebPAC::Normalize::_get_lookup(); |
397 |
|
|
398 |
my $row = $isis->fetch || next; |
if (defined( $lookup_data->{$database}->{$input_name} )) { |
399 |
|
$log->debug("created following lookups: ", dump( $lookup_data )); |
400 |
|
|
401 |
my $mfn = $row->{'000'}->[0] || die "can't find MFN"; |
foreach my $key (keys %{ $lookup_data->{$database}->{$input_name} }) { |
402 |
|
$store->save_lookup( |
403 |
|
database => $database, |
404 |
|
input => $input_name, |
405 |
|
key => $key, |
406 |
|
data => $lookup_data->{$database}->{$input_name}->{$key}, |
407 |
|
); |
408 |
|
} |
409 |
|
} |
410 |
|
|
411 |
my $ds = $n->data_structure($row); |
my $report_fh; |
412 |
|
if ($stats || $validate) { |
413 |
|
my $path = "out/report/${database}-${input_name}.txt"; |
414 |
|
open($report_fh, '>', $path) || $log->logdie("can't open $path: $!"); |
415 |
|
|
416 |
|
print $report_fh "Report for database '$database' input '$input_name' records ", |
417 |
|
$offset || 1, "-", $limit || $input->{limit} || $maxmfn, "\n\n"; |
418 |
|
$log->info("Generating report file $path"); |
419 |
|
} |
420 |
|
|
421 |
$est->add( |
my $marc; |
422 |
id => $input->{name} . "#" . $mfn, |
if ($parser->have_rules( 'marc', $database, $input_name )) { |
423 |
ds => $ds, |
$marc = new WebPAC::Output::MARC( |
424 |
type => $config->{hyperestraier}->{type}, |
path => "out/marc/${database}-${input_name}.marc", |
425 |
|
lint => $marc_lint, |
426 |
|
dump => $marc_dump, |
427 |
); |
); |
428 |
|
} |
429 |
|
|
430 |
|
my $rules = $parser->normalize_rules($database,$input_name) || $log->logdie("no normalize rules found for $database/$input_name"); |
431 |
|
$log->debug("parsed normalize rules:\n$rules"); |
432 |
|
|
433 |
|
# reset position in database |
434 |
|
$input_db->seek(1); |
435 |
|
|
436 |
|
# generate name of config key for indexer (strip everything after -) |
437 |
|
my $indexer_config = $use_indexer; |
438 |
|
$indexer_config =~ s/^(\w+)-?.*$/$1/g if ($indexer_config); |
439 |
|
|
440 |
|
my $lookup_hash; |
441 |
|
my $depends = $parser->depends($database,$input_name); |
442 |
|
|
443 |
|
if ($depends) { |
444 |
|
$log->debug("$database/$input_name depends on: ", dump($depends)) if ($depends); |
445 |
|
$log->logdie("parser->depends didn't return HASH") unless (ref($depends) eq 'HASH'); |
446 |
|
|
447 |
|
foreach my $db (keys %$depends) { |
448 |
|
foreach my $i (keys %{$depends->{$db}}) { |
449 |
|
foreach my $k (keys %{$depends->{$db}->{$i}}) { |
450 |
|
my $t = time(); |
451 |
|
$log->debug("loading lookup $db/$i"); |
452 |
|
$lookup_hash->{$db}->{$i}->{$k} = $store->load_lookup( |
453 |
|
database => $db, |
454 |
|
input => $i, |
455 |
|
key => $k, |
456 |
|
); |
457 |
|
$log->debug(sprintf("lookup $db/$i took %.2fs", time() - $t)); |
458 |
|
} |
459 |
|
} |
460 |
|
} |
461 |
|
|
462 |
|
$log->debug("lookup_hash = ", sub { dump( $lookup_hash ) }); |
463 |
|
} |
464 |
|
|
465 |
|
|
466 |
|
foreach my $pos ( 0 ... $input_db->size ) { |
467 |
|
|
468 |
|
my $row = $input_db->fetch || next; |
469 |
|
|
470 |
|
my $mfn = $row->{'000'}->[0]; |
471 |
|
|
472 |
|
if (! $mfn || $mfn !~ m#^\d+$#) { |
473 |
|
$log->warn("record $pos doesn't have valid MFN but '$mfn', using $pos"); |
474 |
|
$mfn = $pos; |
475 |
|
push @{ $row->{'000'} }, $pos; |
476 |
|
} |
477 |
|
|
478 |
|
|
479 |
|
if ($validate) { |
480 |
|
if ( my $errors = $validate->validate_errors( $row, $input_db->dump ) ) { |
481 |
|
$log->error( "MFN $mfn validation error:\n", |
482 |
|
$validate->report_error( $errors ) |
483 |
|
); |
484 |
|
} |
485 |
|
} |
486 |
|
|
487 |
|
my $ds = WebPAC::Normalize::data_structure( |
488 |
|
row => $row, |
489 |
|
rules => $rules, |
490 |
|
lookup => $lookup_hash, |
491 |
|
config => create_ds_config( $db_config, $database, $input, $mfn ), |
492 |
|
marc_encoding => 'utf-8', |
493 |
|
load_row_coderef => sub { |
494 |
|
my ($database,$input,$mfn) = @_; |
495 |
|
return $store->load_row( |
496 |
|
database => $database, |
497 |
|
input => $input, |
498 |
|
id => $mfn, |
499 |
|
); |
500 |
|
}, |
501 |
|
); |
502 |
|
|
503 |
|
$log->debug("ds = ",dump($ds)) if ($ds); |
504 |
|
|
505 |
|
$store->save_ds( |
506 |
|
database => $database, |
507 |
|
input => $input_name, |
508 |
|
id => $mfn, |
509 |
|
ds => $ds, |
510 |
|
) if ($ds && !$stats); |
511 |
|
|
512 |
|
$indexer->add( |
513 |
|
id => "${input_name}/${mfn}", |
514 |
|
ds => $ds, |
515 |
|
type => $config->get($indexer_config)->{type}, |
516 |
|
) if ($indexer && $ds); |
517 |
|
|
518 |
|
if ($marc) { |
519 |
|
my $i = 0; |
520 |
|
|
521 |
|
while (my $fields = WebPAC::Normalize::_get_marc_fields( fetch_next => 1 ) ) { |
522 |
|
$marc->add( |
523 |
|
id => $mfn . ( $i ? "/$i" : '' ), |
524 |
|
fields => $fields, |
525 |
|
leader => WebPAC::Normalize::marc_leader(), |
526 |
|
row => $row, |
527 |
|
); |
528 |
|
$i++; |
529 |
|
} |
530 |
|
|
531 |
|
$log->info("Created $i instances of MFN $mfn\n") if ($i > 1); |
532 |
|
} |
533 |
|
|
534 |
$total_rows++; |
$total_rows++; |
535 |
} |
} |
536 |
|
|
537 |
}; |
if ($validate) { |
538 |
|
my $errors = $validate->report; |
539 |
|
if ($errors) { |
540 |
|
$log->info("validation errors:\n$errors\n" ); |
541 |
|
print $report_fh "$errors\n" if ($report_fh); |
542 |
|
} |
543 |
|
} |
544 |
|
|
545 |
|
if ($stats) { |
546 |
|
my $s = $input_db->stats; |
547 |
|
$log->info("statistics of fields usage:\n$s"); |
548 |
|
print $report_fh "Statistics of fields usage:\n$s" if ($report_fh); |
549 |
|
} |
550 |
|
|
551 |
|
# close MARC file |
552 |
|
$marc->finish if ($marc); |
553 |
|
|
554 |
|
# close report |
555 |
|
close($report_fh) if ($report_fh) |
556 |
|
|
557 |
|
} |
558 |
|
|
559 |
|
eval { $indexer->finish } if ($indexer && $indexer->can('finish')); |
560 |
|
|
561 |
|
my $dt = time() - $start_t; |
562 |
|
$log->info("$total_rows records ", $indexer ? "indexed " : "", |
563 |
|
sprintf("in %.2f sec [%.2f rec/sec]", |
564 |
|
$dt, ($total_rows / $dt) |
565 |
|
) |
566 |
|
); |
567 |
|
|
568 |
|
|
569 |
|
# end forked process |
570 |
|
if ($parallel) { |
571 |
|
$log->info("parallel process $$ finished"); |
572 |
|
exit(0); |
573 |
|
} |
574 |
|
|
|
$log->info("$total_rows records indexed"); |
|
575 |
} |
} |
576 |
|
|
577 |
|
if ($parallel) { |
578 |
|
# wait all children to finish |
579 |
|
sleep(1) while wait != -1; |
580 |
|
$log->info("all parallel processes finished"); |
581 |
|
} |
582 |
|
|
583 |
|
# |
584 |
|
# handle links or merge after indexing |
585 |
|
# |
586 |
|
|
587 |
|
if ($merge) { |
588 |
|
print $estcmd_fh 'sudo /etc/init.d/hyperestraier start',$/; |
589 |
|
close($estcmd_fh); |
590 |
|
chmod 0700, $estcmd_path || $log->warn("can't chmod 0700 $estcmd_path: $!"); |
591 |
|
system $estcmd_path; |
592 |
|
} else { |
593 |
|
foreach my $link (@links) { |
594 |
|
$log->logdie("coderef in link ", Dumper($link), " is ", ref($link), " and not CODE") unless (ref($link) eq 'CODE'); |
595 |
|
$link->(); |
596 |
|
} |
597 |
|
} |