bd141a442e
Adds the "Attributes" field to z3950 servers. The feature here is not quite de same. In the old patches, the attributes were applied to individual query parts if the part already contains "@attr" and the additionnal attribute is not already in the query part. Here, the content of the new field is prepended to all PQF queries sent to the server. This new way of doing is simpler and works for the sponsor. Test plan: I) Apply the patch II) Run updatedatabase.pl 1) Add a new z3950 server with the following parameters: Hostname : catalogue.banq.qc.ca Port : 210 Database : IRIS Syntax : Marc21 2) Perform a z3950 search on that server. Keyword (Any) : egypt 2.1) Nothing Found. 3) Add attributes on the server administration page @attr 4=1 4) Perform the same z3950 search 4.1) A lot of results Signed-off-by: Katrin Fischer <katrin.fischer.83@web.de> Signed-off-by: Martin Renvoize <martin.renvoize@ptfs-europe.com> Signed-off-by: David Nind <david@davidnind.com> Signed-off-by: Katrin Fischer <katrin.fischer.83@web.de> Signed-off-by: Martin Renvoize <martin.renvoize@ptfs-europe.com>
666 lines
22 KiB
Perl
666 lines
22 KiB
Perl
package C4::Breeding;
|
|
|
|
# Copyright 2000-2002 Katipo Communications
|
|
# Parts Copyright 2013 Prosentient Systems
|
|
#
|
|
# This file is part of Koha.
|
|
#
|
|
# Koha is free software; you can redistribute it and/or modify it
|
|
# under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# Koha is distributed in the hope that it will be useful, but
|
|
# WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with Koha; if not, see <http://www.gnu.org/licenses>.
|
|
|
|
use strict;
|
|
use warnings;
|
|
|
|
use C4::Biblio;
|
|
use C4::Koha;
|
|
use C4::Charset;
|
|
use MARC::File::USMARC;
|
|
use MARC::Field;
|
|
use C4::ImportBatch;
|
|
use C4::AuthoritiesMarc; #GuessAuthTypeCode, FindDuplicateAuthority
|
|
use C4::Languages;
|
|
use Koha::Database;
|
|
use Koha::XSLT_Handler;
|
|
|
|
use vars qw(@ISA @EXPORT @EXPORT_OK %EXPORT_TAGS);
|
|
|
|
BEGIN {
|
|
require Exporter;
|
|
@ISA = qw(Exporter);
|
|
@EXPORT = qw(&BreedingSearch &Z3950Search &Z3950SearchAuth);
|
|
}
|
|
|
|
=head1 NAME
|
|
|
|
C4::Breeding : module to add biblios to import_records via
|
|
the breeding/reservoir API.
|
|
|
|
=head1 SYNOPSIS
|
|
|
|
Z3950Search($pars, $template);
|
|
($count, @results) = &BreedingSearch($title,$isbn);
|
|
|
|
=head1 DESCRIPTION
|
|
|
|
This module contains routines related to Koha's Z39.50 search into
|
|
cataloguing reservoir features.
|
|
|
|
=head2 BreedingSearch
|
|
|
|
($count, @results) = &BreedingSearch($title,$isbn);
|
|
C<$title> contains the title,
|
|
C<$isbn> contains isbn or issn,
|
|
|
|
C<$count> is the number of items in C<@results>. C<@results> is an
|
|
array of references-to-hash; the keys are the items from the C<import_records> and
|
|
C<import_biblios> tables of the Koha database.
|
|
|
|
=cut
|
|
|
|
sub BreedingSearch {
|
|
my ($search,$isbn) = @_;
|
|
my $dbh = C4::Context->dbh;
|
|
my $count = 0;
|
|
my ($query,@bind);
|
|
my $sth;
|
|
my @results;
|
|
|
|
# normalise ISBN like at import
|
|
$isbn = C4::Koha::GetNormalizedISBN($isbn);
|
|
|
|
$query = "SELECT import_record_id, file_name, isbn, title, author
|
|
FROM import_biblios
|
|
JOIN import_records USING (import_record_id)
|
|
JOIN import_batches USING (import_batch_id)
|
|
WHERE ";
|
|
@bind=();
|
|
if (defined($search) && length($search)>0) {
|
|
$search =~ s/(\s+)/\%/g;
|
|
$query .= "title like ? OR author like ?";
|
|
push(@bind,"%$search%", "%$search%");
|
|
}
|
|
if ($#bind!=-1 && defined($isbn) && length($isbn)>0) {
|
|
$query .= " and ";
|
|
}
|
|
if (defined($isbn) && length($isbn)>0) {
|
|
$query .= "isbn like ?";
|
|
push(@bind,"$isbn%");
|
|
}
|
|
$sth = $dbh->prepare($query);
|
|
$sth->execute(@bind);
|
|
while (my $data = $sth->fetchrow_hashref) {
|
|
$results[$count] = $data;
|
|
# FIXME - hack to reflect difference in name
|
|
# of columns in old marc_breeding and import_records
|
|
# There needs to be more separation between column names and
|
|
# field names used in the templates </soapbox>
|
|
$data->{'file'} = $data->{'file_name'};
|
|
$data->{'id'} = $data->{'import_record_id'};
|
|
$count++;
|
|
} # while
|
|
|
|
$sth->finish;
|
|
return($count, @results);
|
|
} # sub breedingsearch
|
|
|
|
|
|
=head2 Z3950Search
|
|
|
|
Z3950Search($pars, $template);
|
|
|
|
Parameters for Z3950 search are all passed via the $pars hash. It may contain isbn, title, author, dewey, subject, lccall, controlnumber, stdid, srchany.
|
|
Also it should contain an arrayref id that points to a list of id's of the z3950 targets to be queried (see z3950servers table).
|
|
This code is used in acqui/z3950_search and cataloging/z3950_search.
|
|
The second parameter $template is a Template object. The routine uses this parameter to store the found values into the template.
|
|
|
|
=cut
|
|
|
|
sub Z3950Search {
|
|
my ($pars, $template)= @_;
|
|
|
|
my @id= @{$pars->{id}};
|
|
my $page= $pars->{page};
|
|
my $biblionumber= $pars->{biblionumber};
|
|
|
|
my $show_next = 0;
|
|
my $total_pages = 0;
|
|
my @results;
|
|
my @breeding_loop = ();
|
|
my @oConnection;
|
|
my @oResult;
|
|
my @errconn;
|
|
my $s = 0;
|
|
my $imported=0;
|
|
|
|
my ( $zquery, $squery ) = _bib_build_query( $pars );
|
|
|
|
my $schema = Koha::Database->new()->schema();
|
|
my $rs = $schema->resultset('Z3950server')->search(
|
|
{ id => [ @id ] },
|
|
{ result_class => 'DBIx::Class::ResultClass::HashRefInflator' },
|
|
);
|
|
my @servers = $rs->all;
|
|
foreach my $server ( @servers ) {
|
|
my $server_zquery = $zquery;
|
|
if(my $attributes = $server->{attributes}){
|
|
$server_zquery = "$attributes $zquery";
|
|
}
|
|
$oConnection[$s] = _create_connection( $server );
|
|
$oResult[$s] =
|
|
$server->{servertype} eq 'zed'?
|
|
$oConnection[$s]->search_pqf( $server_zquery ):
|
|
$oConnection[$s]->search(new ZOOM::Query::CQL(
|
|
_translate_query( $server, $squery )));
|
|
$s++;
|
|
}
|
|
my $xslh = Koha::XSLT_Handler->new;
|
|
|
|
my $nremaining = $s;
|
|
while ( $nremaining-- ) {
|
|
my $k;
|
|
my $event;
|
|
while ( ( $k = ZOOM::event( \@oConnection ) ) != 0 ) {
|
|
$event = $oConnection[ $k - 1 ]->last_event();
|
|
last if $event == ZOOM::Event::ZEND;
|
|
}
|
|
|
|
if ( $k != 0 ) {
|
|
$k--;
|
|
my ($error)= $oConnection[$k]->error_x(); #ignores errmsg, addinfo, diagset
|
|
if ($error) {
|
|
if ($error =~ m/^(10000|10007)$/ ) {
|
|
push(@errconn, { server => $servers[$k]->{host}, error => $error } );
|
|
}
|
|
}
|
|
else {
|
|
my $numresults = $oResult[$k]->size();
|
|
my $i;
|
|
my $res;
|
|
if ( $numresults > 0 and $numresults >= (($page-1)*20)) {
|
|
$show_next = 1 if $numresults >= ($page*20);
|
|
$total_pages = int($numresults/20)+1 if $total_pages < ($numresults/20);
|
|
for ($i = ($page-1)*20; $i < (($numresults < ($page*20)) ? $numresults : ($page*20)); $i++) {
|
|
if ( $oResult[$k]->record($i) ) {
|
|
undef $error;
|
|
( $res, $error ) = _handle_one_result( $oResult[$k]->record($i), $servers[$k], ++$imported, $biblionumber, $xslh ); #ignores error in sequence numbering
|
|
push @breeding_loop, $res if $res;
|
|
push @errconn, { server => $servers[$k]->{servername}, error => $error, seq => $i+1 } if $error;
|
|
}
|
|
else {
|
|
push @errconn, { 'server' => $servers[$k]->{servername}, error => ( ( $oConnection[$k]->error_x() )[0] ), seq => $i+1 };
|
|
}
|
|
}
|
|
} #if $numresults
|
|
}
|
|
} # if $k !=0
|
|
|
|
$template->param(
|
|
numberpending => $nremaining,
|
|
current_page => $page,
|
|
total_pages => $total_pages,
|
|
show_nextbutton => $show_next?1:0,
|
|
show_prevbutton => $page!=1,
|
|
);
|
|
} # while nremaining
|
|
|
|
#close result sets and connections
|
|
foreach(0..$s-1) {
|
|
$oResult[$_]->destroy();
|
|
$oConnection[$_]->destroy();
|
|
}
|
|
|
|
$template->param(
|
|
breeding_loop => \@breeding_loop,
|
|
servers => \@servers,
|
|
errconn => \@errconn
|
|
);
|
|
}
|
|
|
|
sub _auth_build_query {
|
|
my ( $pars ) = @_;
|
|
|
|
my $qry_build = {
|
|
nameany => '@attr 1=1002 "#term" ',
|
|
authorany => '@attr 1=1003 "#term" ',
|
|
authorcorp => '@attr 1=2 "#term" ',
|
|
authorpersonal => '@attr 1=1 "#term" ',
|
|
authormeetingcon => '@attr 1=3 "#term" ',
|
|
subject => '@attr 1=21 "#term" ',
|
|
subjectsubdiv => '@attr 1=47 "#term" ',
|
|
title => '@attr 1=4 "#term" ',
|
|
uniformtitle => '@attr 1=6 "#term" ',
|
|
srchany => '@attr 1=1016 "#term" ',
|
|
controlnumber => '@attr 1=12 "#term" ',
|
|
};
|
|
|
|
return _build_query( $pars, $qry_build );
|
|
}
|
|
|
|
sub _bib_build_query {
|
|
|
|
my ( $pars ) = @_;
|
|
|
|
my $qry_build = {
|
|
isbn => '@attr 1=7 @attr 5=1 "#term" ',
|
|
issn => '@attr 1=8 @attr 5=1 "#term" ',
|
|
title => '@attr 1=4 "#term" ',
|
|
author => '@attr 1=1003 "#term" ',
|
|
dewey => '@attr 1=16 "#term" ',
|
|
subject => '@attr 1=21 "#term" ',
|
|
lccall => '@attr 1=16 @attr 2=3 @attr 3=1 @attr 4=1 @attr 5=1 '.
|
|
'@attr 6=1 "#term" ',
|
|
controlnumber => '@attr 1=12 "#term" ',
|
|
srchany => '@attr 1=1016 "#term" ',
|
|
stdid => '@attr 1=1007 "#term" ',
|
|
publicationyear => '@attr 1=31 "#term" '
|
|
};
|
|
|
|
return _build_query( $pars, $qry_build );
|
|
}
|
|
|
|
sub _build_query {
|
|
|
|
my ( $pars, $qry_build ) = @_;
|
|
|
|
my $zquery='';
|
|
my $squery='';
|
|
my $nterms=0;
|
|
foreach my $k ( sort keys %$pars ) {
|
|
#note that the sort keys forces an identical result under Perl 5.18
|
|
#one of the unit tests is based on that assumption
|
|
if( ( my $val=$pars->{$k} ) && $qry_build->{$k} ) {
|
|
$qry_build->{$k} =~ s/#term/$val/g;
|
|
$zquery .= $qry_build->{$k};
|
|
$squery .= "[$k]=\"$val\" and ";
|
|
$nterms++;
|
|
}
|
|
}
|
|
$zquery = "\@and " . $zquery for 2..$nterms;
|
|
$squery =~ s/ and $//;
|
|
return ( $zquery, $squery );
|
|
}
|
|
|
|
sub _handle_one_result {
|
|
my ( $zoomrec, $servhref, $seq, $bib, $xslh )= @_;
|
|
|
|
my $raw= $zoomrec->raw();
|
|
my $marcrecord;
|
|
if( $servhref->{servertype} eq 'sru' ) {
|
|
$marcrecord= MARC::Record->new_from_xml( $raw, 'UTF-8',
|
|
$servhref->{syntax} );
|
|
} else {
|
|
($marcrecord) = MarcToUTF8Record($raw, C4::Context->preference('marcflavour'), $servhref->{encoding} // "iso-5426" ); #ignores charset return values
|
|
}
|
|
SetUTF8Flag($marcrecord);
|
|
my $error;
|
|
( $marcrecord, $error ) = _do_xslt_proc($marcrecord, $servhref, $xslh);
|
|
|
|
my $batch_id = GetZ3950BatchId($servhref->{servername});
|
|
my $breedingid = AddBiblioToBatch($batch_id, $seq, $marcrecord, 'UTF-8', 0);
|
|
#Last zero indicates: no update for batch record counts
|
|
|
|
|
|
#call to TransformMarcToKoha replaced by next call
|
|
#we only need six fields from the marc record
|
|
my $row;
|
|
$row = _add_rowdata(
|
|
{
|
|
biblionumber => $bib,
|
|
server => $servhref->{servername},
|
|
breedingid => $breedingid,
|
|
}, $marcrecord) if $breedingid;
|
|
return ( $row, $error );
|
|
}
|
|
|
|
sub _do_xslt_proc {
|
|
my ( $marc, $server, $xslh ) = @_;
|
|
return $marc if !$server->{add_xslt};
|
|
|
|
my $htdocs = C4::Context->config('intrahtdocs');
|
|
my $theme = C4::Context->preference("template"); #staff
|
|
my $lang = C4::Languages::getlanguage() || 'en';
|
|
|
|
my @files= split ',', $server->{add_xslt};
|
|
my $xml = $marc->as_xml;
|
|
foreach my $f ( @files ) {
|
|
$f =~ s/^\s+//; $f =~ s/\s+$//; next if !$f;
|
|
$f = C4::XSLT::_get_best_default_xslt_filename(
|
|
$htdocs, $theme, $lang, $f ) unless $f =~ /^\//;
|
|
$xml = $xslh->transform( $xml, $f );
|
|
last if $xslh->err; #skip other files
|
|
}
|
|
if( !$xslh->err ) {
|
|
return MARC::Record->new_from_xml($xml, 'UTF-8');
|
|
} else {
|
|
return ( $marc, $xslh->err ); #original record in case of errors
|
|
}
|
|
}
|
|
|
|
sub _add_rowdata {
|
|
my ($row, $record)=@_;
|
|
my %fetch= (
|
|
title => 'biblio.title',
|
|
author => 'biblio.author',
|
|
isbn =>'biblioitems.isbn',
|
|
lccn =>'biblioitems.lccn', #LC control number (not call number)
|
|
edition =>'biblioitems.editionstatement'
|
|
);
|
|
$fetch{date} = C4::Context->preference('marcflavour') eq "MARC21" ? 'biblio.copyrightdate' : 'biblioitems.publicationyear';
|
|
|
|
foreach my $k (keys %fetch) {
|
|
$row->{$k} = C4::Biblio::TransformMarcToKohaOneField( $fetch{$k}, $record );
|
|
}
|
|
$row->{date}//= $row->{date2};
|
|
$row->{isbn}=_isbn_replace($row->{isbn});
|
|
|
|
$row = _add_custom_field_rowdata($row, $record);
|
|
|
|
return $row;
|
|
}
|
|
|
|
sub _add_custom_field_rowdata
|
|
{
|
|
my ( $row, $record ) = @_;
|
|
my $pref_newtags = C4::Context->preference('AdditionalFieldsInZ3950ResultSearch');
|
|
my $pref_flavour = C4::Context->preference('MarcFlavour');
|
|
|
|
$pref_newtags =~ s/^\s+|\s+$//g;
|
|
$pref_newtags =~ s/\h+/ /g;
|
|
|
|
my @addnumberfields;
|
|
|
|
foreach my $field (split /\,/, $pref_newtags) {
|
|
$field =~ s/^\s+|\s+$//g ; # trim whitespace
|
|
my ($tag, $subtags) = split(/\$/, $field);
|
|
|
|
if ( $record->field($tag) ) {
|
|
my @content = ();
|
|
|
|
for my $marcfield ($record->field($tag)) {
|
|
if ( $subtags ) {
|
|
my $str = '';
|
|
for my $code (split //, $subtags) {
|
|
if ( $marcfield->subfield($code) ) {
|
|
$str .= $marcfield->subfield($code) . ' ';
|
|
}
|
|
}
|
|
if ( not $str eq '') {
|
|
push @content, $str;
|
|
}
|
|
} elsif ( $tag == 10 ) {
|
|
push @content, ( $pref_flavour eq "MARC21" ? $marcfield->data : $marcfield->as_string );
|
|
} elsif ( $tag < 10 ) {
|
|
push @content, $marcfield->data();
|
|
} else {
|
|
push @content, $marcfield->as_string();
|
|
}
|
|
}
|
|
|
|
if ( @content ) {
|
|
$row->{$field} = \@content;
|
|
push( @addnumberfields, $field );
|
|
}
|
|
}
|
|
}
|
|
|
|
$row->{'addnumberfields'} = \@addnumberfields;
|
|
|
|
return $row;
|
|
}
|
|
|
|
sub _isbn_replace {
|
|
my ($isbn) = @_;
|
|
return unless defined $isbn;
|
|
$isbn =~ s/ |-|\.//g;
|
|
$isbn =~ s/\|/ \| /g;
|
|
$isbn =~ s/\(/ \(/g;
|
|
return $isbn;
|
|
}
|
|
|
|
sub _create_connection {
|
|
my ( $server ) = @_;
|
|
my $option1= new ZOOM::Options();
|
|
$option1->option( 'async' => 1 );
|
|
$option1->option( 'elementSetName', 'F' );
|
|
$option1->option( 'preferredRecordSyntax', $server->{syntax} );
|
|
$option1->option( 'timeout', $server->{timeout} ) if $server->{timeout};
|
|
|
|
if( $server->{servertype} eq 'sru' ) {
|
|
foreach( split ',', $server->{sru_options}//'' ) {
|
|
#first remove surrounding spaces at comma and equals-sign
|
|
s/^\s+|\s+$//g;
|
|
my @temp= split '=', $_, 2;
|
|
@temp= map { my $c=$_; $c=~s/^\s+|\s+$//g; $c; } @temp;
|
|
$option1->option( $temp[0] => $temp[1] ) if @temp;
|
|
}
|
|
} elsif( $server->{servertype} eq 'zed' ) {
|
|
$option1->option( 'databaseName', $server->{db} );
|
|
$option1->option( 'user', $server->{userid} ) if $server->{userid};
|
|
$option1->option( 'password', $server->{password} ) if $server->{password};
|
|
}
|
|
my $obj= ZOOM::Connection->create($option1);
|
|
if( $server->{servertype} eq 'sru' ) {
|
|
my $host= $server->{host};
|
|
if( $host !~ /^https?:\/\// ) {
|
|
#Normally, host will not be prefixed by protocol.
|
|
#In that case we can (safely) assume http.
|
|
#In case someone prefixed with https, give it a try..
|
|
$host = 'http://' . $host;
|
|
}
|
|
$obj->connect( $host.':'.$server->{port}.'/'.$server->{db} );
|
|
} else {
|
|
$obj->connect( $server->{host}, $server->{port} );
|
|
}
|
|
return $obj;
|
|
}
|
|
|
|
sub _translate_query { #SRU query adjusted per server cf. srufields column
|
|
my ($server, $query) = @_;
|
|
|
|
#sru_fields is in format title=field,isbn=field,...
|
|
#if a field doesn't exist, try anywhere or remove [field]=
|
|
my @parts= split(',', $server->{sru_fields} );
|
|
my %trans= map { if( /=/ ) { ( $`,$' ) } else { () } } @parts;
|
|
my $any= $trans{srchany}?$trans{srchany}.'=':'';
|
|
|
|
my $q=$query;
|
|
foreach my $key (keys %trans) {
|
|
my $f=$trans{$key};
|
|
if( $f ) {
|
|
$q=~s/\[$key\]/$f/g;
|
|
} else {
|
|
$q=~s/\[$key\]=/$any/g;
|
|
}
|
|
}
|
|
$q=~s/\[\w+\]=/$any/g; # remove remaining fields (not found in field list)
|
|
return $q;
|
|
}
|
|
|
|
=head2 ImportBreedingAuth
|
|
|
|
ImportBreedingAuth( $marcrecord, $filename, $encoding, $heading );
|
|
|
|
ImportBreedingAuth imports MARC records in the reservoir (import_records table) or returns their id if they already exist.
|
|
|
|
=cut
|
|
|
|
sub ImportBreedingAuth {
|
|
my ( $marcrecord, $filename, $encoding, $heading ) = @_;
|
|
my $dbh = C4::Context->dbh;
|
|
|
|
my $batch_id = GetZ3950BatchId($filename);
|
|
my $searchbreeding = $dbh->prepare("select import_record_id from import_auths where control_number=? and authorized_heading=?");
|
|
|
|
my $controlnumber = $marcrecord->field('001')->data;
|
|
|
|
# Normalize the record so it doesn't have separated diacritics
|
|
SetUTF8Flag($marcrecord);
|
|
|
|
$searchbreeding->execute($controlnumber,$heading);
|
|
my ($breedingid) = $searchbreeding->fetchrow;
|
|
|
|
return $breedingid if $breedingid;
|
|
$breedingid = AddAuthToBatch($batch_id, 0, $marcrecord, $encoding);
|
|
return $breedingid;
|
|
}
|
|
|
|
=head2 Z3950SearchAuth
|
|
|
|
Z3950SearchAuth($pars, $template);
|
|
|
|
Parameters for Z3950 search are all passed via the $pars hash. It may contain nameany, namepersonal, namecorp, namemeetingcon,
|
|
title, uniform title, subject, subjectsubdiv, srchany.
|
|
Also it should contain an arrayref id that points to a list of IDs of the z3950 targets to be queried (see z3950servers table).
|
|
This code is used in cataloging/z3950_auth_search.
|
|
The second parameter $template is a Template object. The routine uses this parameter to store the found values into the template.
|
|
|
|
=cut
|
|
|
|
sub Z3950SearchAuth {
|
|
my ($pars, $template)= @_;
|
|
|
|
my $dbh = C4::Context->dbh;
|
|
my @id= @{$pars->{id}};
|
|
my $page= $pars->{page};
|
|
|
|
|
|
my $show_next = 0;
|
|
my $total_pages = 0;
|
|
my @encoding;
|
|
my @results;
|
|
my @serverhost;
|
|
my @breeding_loop = ();
|
|
my @oConnection;
|
|
my @oResult;
|
|
my @errconn;
|
|
my @servers;
|
|
my $s = 0;
|
|
my $query;
|
|
my $nterms=0;
|
|
|
|
my $marcflavour = C4::Context->preference('marcflavour');
|
|
my $marc_type = $marcflavour eq 'UNIMARC' ? 'UNIMARCAUTH' : $marcflavour;
|
|
my $authid= $pars->{authid};
|
|
my ( $zquery, $squery ) = _auth_build_query( $pars );
|
|
foreach my $servid (@id) {
|
|
my $sth = $dbh->prepare("select * from z3950servers where id=?");
|
|
$sth->execute($servid);
|
|
while ( my $server = $sth->fetchrow_hashref ) {
|
|
$oConnection[$s] = _create_connection( $server );
|
|
|
|
if ( $server->{servertype} eq 'zed' ) {
|
|
my $server_zquery = $zquery;
|
|
if ( my $attributes = $server->{attributes} ) {
|
|
$server_zquery = "$attributes $zquery";
|
|
}
|
|
$oResult[$s] = $oConnection[$s]->search_pqf( $server_zquery );
|
|
}
|
|
else {
|
|
$oResult[$s] = $oConnection[$s]->search(
|
|
new ZOOM::Query::CQL(_translate_query( $server, $squery ))
|
|
);
|
|
}
|
|
$encoding[$s] = ($server->{encoding}?$server->{encoding}:"iso-5426");
|
|
$servers[$s] = $server;
|
|
$s++;
|
|
} ## while fetch
|
|
} # foreach
|
|
my $nremaining = $s;
|
|
|
|
while ( $nremaining-- ) {
|
|
my $k;
|
|
my $event;
|
|
while ( ( $k = ZOOM::event( \@oConnection ) ) != 0 ) {
|
|
$event = $oConnection[ $k - 1 ]->last_event();
|
|
last if $event == ZOOM::Event::ZEND;
|
|
}
|
|
|
|
if ( $k != 0 ) {
|
|
$k--;
|
|
my ($error )= $oConnection[$k]->error_x(); #ignores errmsg, addinfo, diagset
|
|
if ($error) {
|
|
if ($error =~ m/^(10000|10007)$/ ) {
|
|
push(@errconn, {'server' => $serverhost[$k]});
|
|
}
|
|
}
|
|
else {
|
|
my $numresults = $oResult[$k]->size();
|
|
my $i;
|
|
my $result = '';
|
|
if ( $numresults > 0 and $numresults >= (($page-1)*20)) {
|
|
$show_next = 1 if $numresults >= ($page*20);
|
|
$total_pages = int($numresults/20)+1 if $total_pages < ($numresults/20);
|
|
for ($i = ($page-1)*20; $i < (($numresults < ($page*20)) ? $numresults : ($page*20)); $i++) {
|
|
my $rec = $oResult[$k]->record($i);
|
|
if ($rec) {
|
|
my $marcrecord;
|
|
my $marcdata;
|
|
$marcdata = $rec->raw();
|
|
|
|
my ($charset_result, $charset_errors);
|
|
if( $servers[$k]->{servertype} eq 'sru' ) {
|
|
$marcrecord = MARC::Record->new_from_xml( $marcdata, 'UTF-8', $servers[$k]->{syntax} );
|
|
} else {
|
|
( $marcrecord, $charset_result, $charset_errors ) = MarcToUTF8Record( $marcdata, $marc_type, $encoding[$k] );
|
|
}
|
|
my $heading;
|
|
my $heading_authtype_code;
|
|
$heading_authtype_code = GuessAuthTypeCode($marcrecord);
|
|
$heading = C4::AuthoritiesMarc::GetAuthorizedHeading({ record => $marcrecord });
|
|
|
|
my $breedingid = ImportBreedingAuth( $marcrecord, $serverhost[$k], $encoding[$k], $heading );
|
|
my %row_data;
|
|
$row_data{server} = $servers[$k]->{'servername'};
|
|
$row_data{breedingid} = $breedingid;
|
|
$row_data{heading} = $heading;
|
|
$row_data{authid} = $authid;
|
|
$row_data{heading_code} = $heading_authtype_code;
|
|
push( @breeding_loop, \%row_data );
|
|
}
|
|
else {
|
|
push(@breeding_loop,{'server'=>$servers[$k]->{'servername'},'title'=>join(': ',$oConnection[$k]->error_x()),'breedingid'=>-1,'authid'=>-1});
|
|
}
|
|
}
|
|
} #if $numresults
|
|
}
|
|
} # if $k !=0
|
|
|
|
$template->param(
|
|
numberpending => $nremaining,
|
|
current_page => $page,
|
|
total_pages => $total_pages,
|
|
show_nextbutton => $show_next?1:0,
|
|
show_prevbutton => $page!=1,
|
|
);
|
|
} # while nremaining
|
|
|
|
#close result sets and connections
|
|
foreach(0..$s-1) {
|
|
$oResult[$_]->destroy();
|
|
$oConnection[$_]->destroy();
|
|
}
|
|
|
|
@servers = ();
|
|
foreach my $id (@id) {
|
|
push @servers, {id => $id};
|
|
}
|
|
$template->param(
|
|
breeding_loop => \@breeding_loop,
|
|
servers => \@servers,
|
|
errconn => \@errconn
|
|
);
|
|
}
|
|
|
|
1;
|
|
__END__
|
|
|