3e85c9e97f
* adding 3 subs in Biblio.pm - GetNoZebraIndexes, that get the index structure in a new systempreference (added with this commit) - _DelBiblioNoZebra, that retrieve all index entries for a biblio and remove in a variable the biblio reference - _AddBiblioNoZebra, that add index entries for a biblio. Note that the 2 _Add and _Del subs work only in a hash variable, to speed up things in case of a modif (ie : delete+add). The effective SQL update is done in the ModZebra sub (that existed before, and dealed with zebra index). I think the code has to be more deeply tested, but it works at least partially.
1495 lines
58 KiB
Perl
Executable file
1495 lines
58 KiB
Perl
Executable file
package C4::Search;
|
|
|
|
# This file is part of Koha.
|
|
#
|
|
# Koha is free software; you can redistribute it and/or modify it under the
|
|
# terms of the GNU General Public License as published by the Free Software
|
|
# Foundation; either version 2 of the License, or (at your option) any later
|
|
# version.
|
|
#
|
|
# Koha is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
# WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR
|
|
# A PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License along with
|
|
# Koha; if not, write to the Free Software Foundation, Inc., 59 Temple Place,
|
|
# Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
use strict;
|
|
require Exporter;
|
|
use C4::Context;
|
|
use C4::Biblio; # GetMarcFromKohaField
|
|
use C4::Koha; # getFacets
|
|
use Lingua::Stem;
|
|
|
|
use vars qw($VERSION @ISA @EXPORT @EXPORT_OK %EXPORT_TAGS);
|
|
|
|
# set the version for version checking
|
|
$VERSION = do { my @v = '$Revision$' =~ /\d+/g;
|
|
shift(@v) . "." . join( "_", map { sprintf "%03d", $_ } @v );
|
|
};
|
|
|
|
=head1 NAME
|
|
|
|
C4::Search - Functions for searching the Koha catalog.
|
|
|
|
=head1 SYNOPSIS
|
|
|
|
see opac/opac-search.pl or catalogue/search.pl for example of usage
|
|
|
|
=head1 DESCRIPTION
|
|
|
|
This module provides the searching facilities for the Koha into a zebra catalog.
|
|
|
|
=head1 FUNCTIONS
|
|
|
|
=cut
|
|
|
|
@ISA = qw(Exporter);
|
|
@EXPORT = qw(
|
|
&SimpleSearch
|
|
&findseealso
|
|
&FindDuplicate
|
|
&searchResults
|
|
&getRecords
|
|
&buildQuery
|
|
&NZgetRecords
|
|
&EditBiblios
|
|
);
|
|
|
|
# make all your functions, whether exported or not;
|
|
|
|
=head2 findseealso($dbh,$fields);
|
|
|
|
C<$dbh> is a link to the DB handler.
|
|
|
|
use C4::Context;
|
|
my $dbh =C4::Context->dbh;
|
|
|
|
C<$fields> is a reference to the fields array
|
|
|
|
This function modify the @$fields array and add related fields to search on.
|
|
|
|
=cut
|
|
|
|
sub findseealso {
|
|
my ( $dbh, $fields ) = @_;
|
|
my $tagslib = GetMarcStructure( $dbh, 1 );
|
|
for ( my $i = 0 ; $i <= $#{$fields} ; $i++ ) {
|
|
my ($tag) = substr( @$fields[$i], 1, 3 );
|
|
my ($subfield) = substr( @$fields[$i], 4, 1 );
|
|
@$fields[$i] .= ',' . $tagslib->{$tag}->{$subfield}->{seealso}
|
|
if ( $tagslib->{$tag}->{$subfield}->{seealso} );
|
|
}
|
|
}
|
|
|
|
=head2 FindDuplicate
|
|
|
|
($biblionumber,$biblionumber,$title) = FindDuplicate($record);
|
|
|
|
=cut
|
|
|
|
sub FindDuplicate {
|
|
my ($record) = @_;
|
|
return;
|
|
my $dbh = C4::Context->dbh;
|
|
my $result = TransformMarcToKoha( $dbh, $record, '' );
|
|
my $sth;
|
|
my $query;
|
|
my $search;
|
|
my $type;
|
|
my ( $biblionumber, $title );
|
|
|
|
# search duplicate on ISBN, easy and fast..
|
|
#$search->{'avoidquerylog'}=1;
|
|
if ( $result->{isbn} ) {
|
|
$query = "isbn=$result->{isbn}";
|
|
}
|
|
else {
|
|
$result->{title} =~ s /\\//g;
|
|
$result->{title} =~ s /\"//g;
|
|
$result->{title} =~ s /\(//g;
|
|
$result->{title} =~ s /\)//g;
|
|
$query = "ti,ext=$result->{title}";
|
|
}
|
|
my ($possible_duplicate_record) =
|
|
C4::Biblio::getRecord( "biblioserver", $query, "usmarc" ); # FIXME :: hardcoded !
|
|
if ($possible_duplicate_record) {
|
|
my $marcrecord =
|
|
MARC::Record->new_from_usmarc($possible_duplicate_record);
|
|
my $result = TransformMarcToKoha( $dbh, $marcrecord, '' );
|
|
|
|
# FIXME :: why 2 $biblionumber ?
|
|
return $result->{'biblionumber'}, $result->{'biblionumber'},
|
|
$result->{'title'}
|
|
if $result;
|
|
}
|
|
}
|
|
|
|
=head2 SimpleSearch
|
|
|
|
($error,$results) = SimpleSearch($query,@servers);
|
|
|
|
this function performs a simple search on the catalog using zoom.
|
|
|
|
=over 2
|
|
|
|
=item C<input arg:>
|
|
|
|
* $query could be a simple keyword or a complete CCL query wich is depending on your ccl file.
|
|
* @servers is optionnal. default one is read on koha.xml
|
|
|
|
=item C<Output arg:>
|
|
* $error is a string which containt the description error if there is one. Else it's empty.
|
|
* \@results is an array of marc record.
|
|
|
|
=item C<usage in the script:>
|
|
|
|
=back
|
|
|
|
my ($error, $marcresults) = SimpleSearch($query);
|
|
|
|
if (defined $error) {
|
|
$template->param(query_error => $error);
|
|
warn "error: ".$error;
|
|
output_html_with_http_headers $input, $cookie, $template->output;
|
|
exit;
|
|
}
|
|
|
|
my $hits = scalar @$marcresults;
|
|
my @results;
|
|
|
|
for(my $i=0;$i<$hits;$i++) {
|
|
my %resultsloop;
|
|
my $marcrecord = MARC::File::USMARC::decode($marcresults->[$i]);
|
|
my $biblio = TransformMarcToKoha(C4::Context->dbh,$marcrecord,'');
|
|
|
|
#build the hash for the template.
|
|
$resultsloop{highlight} = ($i % 2)?(1):(0);
|
|
$resultsloop{title} = $biblio->{'title'};
|
|
$resultsloop{subtitle} = $biblio->{'subtitle'};
|
|
$resultsloop{biblionumber} = $biblio->{'biblionumber'};
|
|
$resultsloop{author} = $biblio->{'author'};
|
|
$resultsloop{publishercode} = $biblio->{'publishercode'};
|
|
$resultsloop{publicationyear} = $biblio->{'publicationyear'};
|
|
|
|
push @results, \%resultsloop;
|
|
}
|
|
$template->param(result=>\@results);
|
|
|
|
=cut
|
|
|
|
sub SimpleSearch {
|
|
my $query = shift;
|
|
my @servers = @_;
|
|
my @results;
|
|
my @tmpresults;
|
|
my @zconns;
|
|
return ( "No query entered", undef ) unless $query;
|
|
|
|
#@servers = (C4::Context->config("biblioserver")) unless @servers;
|
|
@servers =
|
|
("biblioserver") unless @servers
|
|
; # FIXME hardcoded value. See catalog/search.pl & opac-search.pl too.
|
|
|
|
# Connect & Search
|
|
for ( my $i = 0 ; $i < @servers ; $i++ ) {
|
|
$zconns[$i] = C4::Context->Zconn( $servers[$i], 1 );
|
|
$tmpresults[$i] =
|
|
$zconns[$i]
|
|
->search( new ZOOM::Query::CCL2RPN( $query, $zconns[$i] ) );
|
|
|
|
# getting error message if one occured.
|
|
my $error =
|
|
$zconns[$i]->errmsg() . " ("
|
|
. $zconns[$i]->errcode() . ") "
|
|
. $zconns[$i]->addinfo() . " "
|
|
. $zconns[$i]->diagset();
|
|
|
|
return ( $error, undef ) if $zconns[$i]->errcode();
|
|
}
|
|
my $hits;
|
|
my $ev;
|
|
while ( ( my $i = ZOOM::event( \@zconns ) ) != 0 ) {
|
|
$ev = $zconns[ $i - 1 ]->last_event();
|
|
if ( $ev == ZOOM::Event::ZEND ) {
|
|
$hits = $tmpresults[ $i - 1 ]->size();
|
|
}
|
|
if ( $hits > 0 ) {
|
|
for ( my $j = 0 ; $j < $hits ; $j++ ) {
|
|
my $record = $tmpresults[ $i - 1 ]->record($j)->raw();
|
|
push @results, $record;
|
|
}
|
|
}
|
|
}
|
|
return ( undef, \@results );
|
|
}
|
|
|
|
# performs the search
|
|
sub getRecords {
|
|
my (
|
|
$koha_query, $federated_query, $sort_by_ref,
|
|
$servers_ref, $results_per_page, $offset,
|
|
$expanded_facet, $branches, $query_type,
|
|
$scan
|
|
) = @_;
|
|
|
|
my @servers = @$servers_ref;
|
|
my @sort_by = @$sort_by_ref;
|
|
|
|
# create the zoom connection and query object
|
|
my $zconn;
|
|
my @zconns;
|
|
my @results;
|
|
my $results_hashref = ();
|
|
|
|
### FACETED RESULTS
|
|
my $facets_counter = ();
|
|
my $facets_info = ();
|
|
my $facets = getFacets();
|
|
|
|
#### INITIALIZE SOME VARS USED CREATE THE FACETED RESULTS
|
|
my @facets_loop; # stores the ref to array of hashes for template
|
|
for ( my $i = 0 ; $i < @servers ; $i++ ) {
|
|
$zconns[$i] = C4::Context->Zconn( $servers[$i], 1 );
|
|
|
|
# perform the search, create the results objects
|
|
# if this is a local search, use the $koha-query, if it's a federated one, use the federated-query
|
|
my $query_to_use;
|
|
if ( $servers[$i] =~ /biblioserver/ ) {
|
|
$query_to_use = $koha_query;
|
|
}
|
|
else {
|
|
$query_to_use = $federated_query;
|
|
}
|
|
|
|
# warn "HERE : $query_type => $query_to_use";
|
|
# check if we've got a query_type defined
|
|
eval {
|
|
if ($query_type)
|
|
{
|
|
if ( $query_type =~ /^ccl/ ) {
|
|
$query_to_use =~
|
|
s/\:/\=/g; # change : to = last minute (FIXME)
|
|
|
|
# warn "CCL : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
elsif ( $query_type =~ /^cql/ ) {
|
|
|
|
# warn "CQL : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CQL( $query_to_use, $zconns[$i] ) );
|
|
}
|
|
elsif ( $query_type =~ /^pqf/ ) {
|
|
|
|
# warn "PQF : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::PQF( $query_to_use, $zconns[$i] ) );
|
|
}
|
|
}
|
|
else {
|
|
if ($scan) {
|
|
|
|
# warn "preparing to scan";
|
|
$results[$i] =
|
|
$zconns[$i]->scan(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
else {
|
|
|
|
# warn "LAST : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
}
|
|
};
|
|
if ($@) {
|
|
warn "prob with query toto $query_to_use " . $@;
|
|
}
|
|
|
|
# concatenate the sort_by limits and pass them to the results object
|
|
my $sort_by;
|
|
foreach my $sort (@sort_by) {
|
|
$sort_by .= $sort . " "; # used to be $sort,
|
|
}
|
|
$results[$i]->sort( "yaz", $sort_by ) if $sort_by;
|
|
}
|
|
while ( ( my $i = ZOOM::event( \@zconns ) ) != 0 ) {
|
|
my $ev = $zconns[ $i - 1 ]->last_event();
|
|
if ( $ev == ZOOM::Event::ZEND ) {
|
|
my $size = $results[ $i - 1 ]->size();
|
|
if ( $size > 0 ) {
|
|
my $results_hash;
|
|
#$results_hash->{'server'} = $servers[$i-1];
|
|
# loop through the results
|
|
$results_hash->{'hits'} = $size;
|
|
my $times;
|
|
if ( $offset + $results_per_page <= $size ) {
|
|
$times = $offset + $results_per_page;
|
|
}
|
|
else {
|
|
$times = $size;
|
|
}
|
|
for ( my $j = $offset ; $j < $times ; $j++ )
|
|
{ #(($offset+$count<=$size) ? ($offset+$count):$size) ; $j++){
|
|
my $records_hash;
|
|
my $record;
|
|
my $facet_record;
|
|
## This is just an index scan
|
|
if ($scan) {
|
|
my ( $term, $occ ) = $results[ $i - 1 ]->term($j);
|
|
|
|
# here we create a minimal MARC record and hand it off to the
|
|
# template just like a normal result ... perhaps not ideal, but
|
|
# it works for now
|
|
my $tmprecord = MARC::Record->new();
|
|
$tmprecord->encoding('UTF-8');
|
|
my $tmptitle;
|
|
|
|
# srote the minimal record in author/title (depending on MARC flavour)
|
|
if ( C4::Context->preference("marcflavour") eq
|
|
"UNIMARC" )
|
|
{
|
|
$tmptitle = MARC::Field->new(
|
|
'200', ' ', ' ',
|
|
a => $term,
|
|
f => $occ
|
|
);
|
|
}
|
|
else {
|
|
$tmptitle = MARC::Field->new(
|
|
'245', ' ', ' ',
|
|
a => $term,
|
|
b => $occ
|
|
);
|
|
}
|
|
$tmprecord->append_fields($tmptitle);
|
|
$results_hash->{'RECORDS'}[$j] =
|
|
$tmprecord->as_usmarc();
|
|
}
|
|
else {
|
|
$record = $results[ $i - 1 ]->record($j)->raw();
|
|
|
|
#warn "RECORD $j:".$record;
|
|
$results_hash->{'RECORDS'}[$j] =
|
|
$record; # making a reference to a hash
|
|
# Fill the facets while we're looping
|
|
$facet_record = MARC::Record->new_from_usmarc($record);
|
|
|
|
#warn $servers[$i-1].$facet_record->title();
|
|
for ( my $k = 0 ; $k <= @$facets ; $k++ ) {
|
|
if ( $facets->[$k] ) {
|
|
my @fields;
|
|
for my $tag ( @{ $facets->[$k]->{'tags'} } ) {
|
|
push @fields, $facet_record->field($tag);
|
|
}
|
|
for my $field (@fields) {
|
|
my @subfields = $field->subfields();
|
|
for my $subfield (@subfields) {
|
|
my ( $code, $data ) = @$subfield;
|
|
if ( $code eq
|
|
$facets->[$k]->{'subfield'} )
|
|
{
|
|
$facets_counter->{ $facets->[$k]
|
|
->{'link_value'} }->{$data}++;
|
|
}
|
|
}
|
|
}
|
|
$facets_info->{ $facets->[$k]->{'link_value'} }
|
|
->{'label_value'} =
|
|
$facets->[$k]->{'label_value'};
|
|
$facets_info->{ $facets->[$k]->{'link_value'} }
|
|
->{'expanded'} = $facets->[$k]->{'expanded'};
|
|
}
|
|
}
|
|
}
|
|
}
|
|
$results_hashref->{ $servers[ $i - 1 ] } = $results_hash;
|
|
}
|
|
|
|
#print "connection ", $i-1, ": $size hits";
|
|
#print $results[$i-1]->record(0)->render() if $size > 0;
|
|
# BUILD FACETS
|
|
for my $link_value (
|
|
sort { $facets_counter->{$b} <=> $facets_counter->{$a} }
|
|
keys %$facets_counter
|
|
)
|
|
{
|
|
my $expandable;
|
|
my $number_of_facets;
|
|
my @this_facets_array;
|
|
for my $one_facet (
|
|
sort {
|
|
$facets_counter->{$link_value}
|
|
->{$b} <=> $facets_counter->{$link_value}->{$a}
|
|
} keys %{ $facets_counter->{$link_value} }
|
|
)
|
|
{
|
|
$number_of_facets++;
|
|
if ( ( $number_of_facets < 6 )
|
|
|| ( $expanded_facet eq $link_value )
|
|
|| ( $facets_info->{$link_value}->{'expanded'} ) )
|
|
{
|
|
|
|
# sanitize the link value ), ( will cause errors with CCL
|
|
my $facet_link_value = $one_facet;
|
|
$facet_link_value =~ s/(\(|\))/ /g;
|
|
|
|
# fix the length that will display in the label
|
|
my $facet_label_value = $one_facet;
|
|
$facet_label_value = substr( $one_facet, 0, 20 ) . "..."
|
|
unless length($facet_label_value) <= 20;
|
|
|
|
# well, if it's a branch, label by the name, not the code
|
|
if ( $link_value =~ /branch/ ) {
|
|
$facet_label_value =
|
|
$branches->{$one_facet}->{'branchname'};
|
|
}
|
|
|
|
# but we're down with the whole label being in the link's title
|
|
my $facet_title_value = $one_facet;
|
|
|
|
push @this_facets_array,
|
|
(
|
|
{
|
|
facet_count =>
|
|
$facets_counter->{$link_value}->{$one_facet},
|
|
facet_label_value => $facet_label_value,
|
|
facet_title_value => $facet_title_value,
|
|
facet_link_value => $facet_link_value,
|
|
type_link_value => $link_value,
|
|
},
|
|
);
|
|
}
|
|
}
|
|
unless ( $facets_info->{$link_value}->{'expanded'} ) {
|
|
$expandable = 1
|
|
if ( ( $number_of_facets > 6 )
|
|
&& ( $expanded_facet ne $link_value ) );
|
|
}
|
|
push @facets_loop,
|
|
(
|
|
{
|
|
type_link_value => $link_value,
|
|
type_id => $link_value . "_id",
|
|
type_label =>
|
|
$facets_info->{$link_value}->{'label_value'},
|
|
facets => \@this_facets_array,
|
|
expandable => $expandable,
|
|
expand => $link_value,
|
|
}
|
|
);
|
|
}
|
|
}
|
|
}
|
|
use Data::Dumper;
|
|
warn Dumper($results_hashref);
|
|
return ( undef, $results_hashref, \@facets_loop );
|
|
}
|
|
|
|
# build the query itself
|
|
sub buildQuery {
|
|
my ( $query, $operators, $operands, $indexes, $limits, $sort_by ) = @_;
|
|
|
|
my @operators = @$operators if $operators;
|
|
my @indexes = @$indexes if $indexes;
|
|
my @operands = @$operands if $operands;
|
|
my @limits = @$limits if $limits;
|
|
my @sort_by = @$sort_by if $sort_by;
|
|
|
|
my $human_search_desc; # a human-readable query
|
|
my $machine_search_desc; #a machine-readable query
|
|
# FIXME: the locale should be set based on the syspref
|
|
my $stemmer = Lingua::Stem->new( -locale => 'EN-US' );
|
|
|
|
# FIXME: these should be stored in the db so the librarian can modify the behavior
|
|
$stemmer->add_exceptions(
|
|
{
|
|
'and' => 'and',
|
|
'or' => 'or',
|
|
'not' => 'not',
|
|
}
|
|
);
|
|
|
|
# STEP I: determine if this is a form-based / simple query or if it's complex (if complex,
|
|
# we can't handle field weighting, stemming until a formal query parser is written
|
|
# I'll work on this soon -- JF
|
|
#if (!$query) { # form-based
|
|
# check if this is a known query language query, if it is, return immediately:
|
|
if ( $query =~ /^ccl=/ ) {
|
|
return ( undef, $', $', $', 'ccl' );
|
|
}
|
|
if ( $query =~ /^cql=/ ) {
|
|
return ( undef, $', $', $', 'cql' );
|
|
}
|
|
if ( $query =~ /^pqf=/ ) {
|
|
return ( undef, $', $', $', 'pqf' );
|
|
}
|
|
if ( $query =~ /(\(|\))/ ) { # sorry, too complex
|
|
return ( undef, $query, $query, $query, 'ccl' );
|
|
}
|
|
|
|
# form-based queries are limited to non-nested a specific depth, so we can easily
|
|
# modify the incoming query operands and indexes to do stemming and field weighting
|
|
# Once we do so, we'll end up with a value in $query, just like if we had an
|
|
# incoming $query from the user
|
|
else {
|
|
$query = ""
|
|
; # clear it out so we can populate properly with field-weighted stemmed query
|
|
my $previous_operand
|
|
; # a flag used to keep track if there was a previous query
|
|
# if there was, we can apply the current operator
|
|
for ( my $i = 0 ; $i <= @operands ; $i++ ) {
|
|
my $operand = $operands[$i];
|
|
my $index = $indexes[$i];
|
|
my $stemmed_operand;
|
|
my $stemming = C4::Context->parameters("Stemming") || 0;
|
|
my $weight_fields = C4::Context->parameters("WeightFields") || 0;
|
|
|
|
if ( $operands[$i] ) {
|
|
|
|
# STEMMING FIXME: need to refine the field weighting so stemmed operands don't disrupt the query ranking
|
|
if ($stemming) {
|
|
my @words = split( / /, $operands[$i] );
|
|
my $stems = $stemmer->stem(@words);
|
|
foreach my $stem (@$stems) {
|
|
$stemmed_operand .= "$stem";
|
|
$stemmed_operand .= "?"
|
|
unless ( $stem =~ /(and$|or$|not$)/ )
|
|
|| ( length($stem) < 3 );
|
|
$stemmed_operand .= " ";
|
|
|
|
#warn "STEM: $stemmed_operand";
|
|
}
|
|
|
|
#$operand = $stemmed_operand;
|
|
}
|
|
|
|
# FIELD WEIGHTING - This is largely experimental stuff. What I'm committing works
|
|
# pretty well but will work much better when we have an actual query parser
|
|
my $weighted_query;
|
|
if ($weight_fields) {
|
|
$weighted_query .=
|
|
" rk=("; # Specifies that we're applying rank
|
|
# keyword has different weight properties
|
|
if ( ( $index =~ /kw/ ) || ( !$index ) )
|
|
{ # FIXME: do I need to add right-truncation in the case of stemming?
|
|
# a simple way to find out if this query uses an index
|
|
if ( $operand =~ /(\=|\:)/ ) {
|
|
$weighted_query .= " $operand";
|
|
}
|
|
else {
|
|
$weighted_query .=
|
|
" Title-cover,ext,r1=\"$operand\""
|
|
; # index label as exact
|
|
$weighted_query .=
|
|
" or ti,ext,r2=$operand"; # index as exact
|
|
#$weighted_query .= " or ti,phr,r3=$operand"; # index as phrase
|
|
#$weighted_query .= " or any,ext,r4=$operand"; # index as exact
|
|
$weighted_query .=
|
|
" or kw,wrdl,r5=$operand"; # index as exact
|
|
$weighted_query .= " or wrd,fuzzy,r9=$operand";
|
|
$weighted_query .= " or wrd=$stemmed_operand"
|
|
if $stemming;
|
|
}
|
|
}
|
|
elsif ( $index =~ /au/ ) {
|
|
$weighted_query .=
|
|
" $index,ext,r1=$operand"; # index label as exact
|
|
#$weighted_query .= " or (title-sort-az=0 or $index,startswithnt,st-word,r3=$operand #)";
|
|
$weighted_query .=
|
|
" or $index,phr,r3=$operand"; # index as phrase
|
|
$weighted_query .= " or $index,rt,wrd,r3=$operand";
|
|
}
|
|
elsif ( $index =~ /ti/ ) {
|
|
$weighted_query .=
|
|
" Title-cover,ext,r1=$operand"; # index label as exact
|
|
$weighted_query .= " or Title-series,ext,r2=$operand";
|
|
|
|
#$weighted_query .= " or ti,ext,r2=$operand";
|
|
#$weighted_query .= " or ti,phr,r3=$operand";
|
|
#$weighted_query .= " or ti,wrd,r3=$operand";
|
|
$weighted_query .=
|
|
" or (title-sort-az=0 or Title-cover,startswithnt,st-word,r3=$operand #)";
|
|
$weighted_query .=
|
|
" or (title-sort-az=0 or Title-cover,phr,r6=$operand)";
|
|
|
|
#$weighted_query .= " or Title-cover,wrd,r5=$operand";
|
|
#$weighted_query .= " or ti,ext,r6=$operand";
|
|
#$weighted_query .= " or ti,startswith,phr,r7=$operand";
|
|
#$weighted_query .= " or ti,phr,r8=$operand";
|
|
#$weighted_query .= " or ti,wrd,r9=$operand";
|
|
|
|
#$weighted_query .= " or ti,ext,r2=$operand"; # index as exact
|
|
#$weighted_query .= " or ti,phr,r3=$operand"; # index as phrase
|
|
#$weighted_query .= " or any,ext,r4=$operand"; # index as exact
|
|
#$weighted_query .= " or kw,wrd,r5=$operand"; # index as exact
|
|
}
|
|
else {
|
|
$weighted_query .=
|
|
" $index,ext,r1=$operand"; # index label as exact
|
|
#$weighted_query .= " or $index,ext,r2=$operand"; # index as exact
|
|
$weighted_query .=
|
|
" or $index,phr,r3=$operand"; # index as phrase
|
|
$weighted_query .= " or $index,rt,wrd,r3=$operand";
|
|
$weighted_query .=
|
|
" or $index,wrd,r5=$operand"
|
|
; # index as word right-truncated
|
|
$weighted_query .= " or $index,wrd,fuzzy,r8=$operand";
|
|
}
|
|
$weighted_query .= ")"; # close rank specification
|
|
$operand = $weighted_query;
|
|
}
|
|
|
|
# only add an operator if there is a previous operand
|
|
if ($previous_operand) {
|
|
if ( $operators[ $i - 1 ] ) {
|
|
$query .= " $operators[$i-1] $index: $operand";
|
|
if ( !$index ) {
|
|
$human_search_desc .=
|
|
" $operators[$i-1] $operands[$i]";
|
|
}
|
|
else {
|
|
$human_search_desc .=
|
|
" $operators[$i-1] $index: $operands[$i]";
|
|
}
|
|
}
|
|
|
|
# the default operator is and
|
|
else {
|
|
$query .= " and $index: $operand";
|
|
$human_search_desc .= " and $index: $operands[$i]";
|
|
}
|
|
}
|
|
else {
|
|
if ( !$index ) {
|
|
$query .= " $operand";
|
|
$human_search_desc .= " $operands[$i]";
|
|
}
|
|
else {
|
|
$query .= " $index: $operand";
|
|
$human_search_desc .= " $index: $operands[$i]";
|
|
}
|
|
$previous_operand = 1;
|
|
}
|
|
} #/if $operands
|
|
} # /for
|
|
}
|
|
|
|
# add limits
|
|
my $limit_query;
|
|
my $limit_search_desc;
|
|
foreach my $limit (@limits) {
|
|
|
|
# FIXME: not quite right yet ... will work on this soon -- JF
|
|
my $type = $1 if $limit =~ m/([^:]+):([^:]*)/;
|
|
if ( $limit =~ /available/ ) {
|
|
$limit_query .=
|
|
" (($query and datedue=0000-00-00) or ($query and datedue=0000-00-00 not lost=1) or ($query and datedue=0000-00-00 not lost=2))";
|
|
|
|
#$limit_search_desc.=" and available";
|
|
}
|
|
elsif ( ($limit_query) && ( index( $limit_query, $type, 0 ) > 0 ) ) {
|
|
if ( $limit_query !~ /\(/ ) {
|
|
$limit_query =
|
|
substr( $limit_query, 0, index( $limit_query, $type, 0 ) )
|
|
. "("
|
|
. substr( $limit_query, index( $limit_query, $type, 0 ) )
|
|
. " or $limit )"
|
|
if $limit;
|
|
$limit_search_desc =
|
|
substr( $limit_search_desc, 0,
|
|
index( $limit_search_desc, $type, 0 ) )
|
|
. "("
|
|
. substr( $limit_search_desc,
|
|
index( $limit_search_desc, $type, 0 ) )
|
|
. " or $limit )"
|
|
if $limit;
|
|
}
|
|
else {
|
|
chop $limit_query;
|
|
chop $limit_search_desc;
|
|
$limit_query .= " or $limit )" if $limit;
|
|
$limit_search_desc .= " or $limit )" if $limit;
|
|
}
|
|
}
|
|
elsif ( ($limit_query) && ( $limit =~ /mc/ ) ) {
|
|
$limit_query .= " or $limit" if $limit;
|
|
$limit_search_desc .= " or $limit" if $limit;
|
|
}
|
|
|
|
# these are treated as AND
|
|
elsif ($limit_query) {
|
|
if ($limit =~ /branch/){
|
|
$limit_query .= " ) and ( $limit" if $limit;
|
|
$limit_search_desc .= " ) and ( $limit" if $limit;
|
|
}else{
|
|
$limit_query .= " or $limit" if $limit;
|
|
$limit_search_desc .= " or $limit" if $limit;
|
|
}
|
|
}
|
|
|
|
# otherwise, there is nothing but the limit
|
|
else {
|
|
$limit_query .= "$limit" if $limit;
|
|
$limit_search_desc .= "$limit" if $limit;
|
|
}
|
|
}
|
|
|
|
# if there's also a query, we need to AND the limits to it
|
|
if ( ($limit_query) && ($query) ) {
|
|
$limit_query = " and (" . $limit_query . ")";
|
|
$limit_search_desc = " and ($limit_search_desc)" if $limit_search_desc;
|
|
|
|
}
|
|
$query .= $limit_query;
|
|
$human_search_desc .= $limit_search_desc;
|
|
|
|
# now normalize the strings
|
|
$query =~ s/ / /g; # remove extra spaces
|
|
$query =~ s/^ //g; # remove any beginning spaces
|
|
$query =~ s/:/=/g; # causes probs for server
|
|
$query =~ s/==/=/g; # remove double == from query
|
|
|
|
my $federated_query = $human_search_desc;
|
|
$federated_query =~ s/ / /g;
|
|
$federated_query =~ s/^ //g;
|
|
$federated_query =~ s/:/=/g;
|
|
my $federated_query_opensearch = $federated_query;
|
|
|
|
# my $federated_query_RPN = new ZOOM::Query::CCL2RPN( $query , C4::Context->ZConn('biblioserver'));
|
|
|
|
$human_search_desc =~ s/ / /g;
|
|
$human_search_desc =~ s/^ //g;
|
|
my $koha_query = $query;
|
|
|
|
#warn "QUERY:".$koha_query;
|
|
#warn "SEARCHDESC:".$human_search_desc;
|
|
#warn "FEDERATED QUERY:".$federated_query;
|
|
return ( undef, $human_search_desc, $koha_query, $federated_query );
|
|
}
|
|
|
|
# IMO this subroutine is pretty messy still -- it's responsible for
|
|
# building the HTML output for the template
|
|
sub searchResults {
|
|
my ( $searchdesc, $hits, $results_per_page, $offset, @marcresults ) = @_;
|
|
|
|
my $dbh = C4::Context->dbh;
|
|
my $toggle;
|
|
my $even = 1;
|
|
my @newresults;
|
|
my $span_terms_hashref;
|
|
for my $span_term ( split( / /, $searchdesc ) ) {
|
|
$span_term =~ s/(.*=|\)|\(|\+|\.)//g;
|
|
$span_terms_hashref->{$span_term}++;
|
|
}
|
|
|
|
#Build brancnames hash
|
|
#find branchname
|
|
#get branch information.....
|
|
my %branches;
|
|
my $bsth =
|
|
$dbh->prepare("SELECT branchcode,branchname FROM branches")
|
|
; # FIXME : use C4::Koha::GetBranches
|
|
$bsth->execute();
|
|
while ( my $bdata = $bsth->fetchrow_hashref ) {
|
|
$branches{ $bdata->{'branchcode'} } = $bdata->{'branchname'};
|
|
}
|
|
|
|
#Build itemtype hash
|
|
#find itemtype & itemtype image
|
|
my %itemtypes;
|
|
$bsth =
|
|
$dbh->prepare("SELECT itemtype,description,imageurl,summary FROM itemtypes");
|
|
$bsth->execute();
|
|
while ( my $bdata = $bsth->fetchrow_hashref ) {
|
|
$itemtypes{ $bdata->{'itemtype'} }->{description} =
|
|
$bdata->{'description'};
|
|
$itemtypes{ $bdata->{'itemtype'} }->{imageurl} = $bdata->{'imageurl'};
|
|
$itemtypes{ $bdata->{'itemtype'} }->{summary} = $bdata->{'summary'};
|
|
}
|
|
|
|
#search item field code
|
|
my $sth =
|
|
$dbh->prepare(
|
|
"select tagfield from marc_subfield_structure where kohafield like 'items.itemnumber'"
|
|
);
|
|
$sth->execute;
|
|
my ($itemtag) = $sth->fetchrow;
|
|
|
|
## find column names of items related to MARC
|
|
my $sth2 = $dbh->prepare("SHOW COLUMNS from items");
|
|
$sth2->execute;
|
|
my %subfieldstosearch;
|
|
while ( ( my $column ) = $sth2->fetchrow ) {
|
|
my ( $tagfield, $tagsubfield ) =
|
|
&GetMarcFromKohaField( "items." . $column, "" );
|
|
$subfieldstosearch{$column} = $tagsubfield;
|
|
}
|
|
my $times;
|
|
|
|
if ( $hits && $offset + $results_per_page <= $hits ) {
|
|
$times = $offset + $results_per_page;
|
|
}
|
|
else {
|
|
$times = $hits;
|
|
}
|
|
|
|
for ( my $i = $offset ; $i <= $times - 1 ; $i++ ) {
|
|
my $marcrecord;
|
|
$marcrecord = MARC::File::USMARC::decode( $marcresults[$i] );
|
|
|
|
my $oldbiblio = TransformMarcToKoha( $dbh, $marcrecord, '' );
|
|
|
|
# add image url if there is one
|
|
if ( $itemtypes{ $oldbiblio->{itemtype} }->{imageurl} =~ /^http:/ ) {
|
|
$oldbiblio->{imageurl} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{imageurl};
|
|
$oldbiblio->{description} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{description};
|
|
}
|
|
else {
|
|
$oldbiblio->{imageurl} =
|
|
getitemtypeimagesrc() . "/"
|
|
. $itemtypes{ $oldbiblio->{itemtype} }->{imageurl}
|
|
if ( $itemtypes{ $oldbiblio->{itemtype} }->{imageurl} );
|
|
$oldbiblio->{description} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{description};
|
|
}
|
|
#
|
|
# build summary if there is one (the summary is defined in itemtypes table
|
|
#
|
|
if ($itemtypes{ $oldbiblio->{itemtype} }->{summary}) {
|
|
my $summary = $itemtypes{ $oldbiblio->{itemtype} }->{summary};
|
|
my @fields = $marcrecord->fields();
|
|
foreach my $field (@fields) {
|
|
my $tag = $field->tag();
|
|
my $tagvalue = $field->as_string();
|
|
$summary =~ s/\[(.?.?.?.?)$tag\*(.*?)]/$1$tagvalue$2\[$1$tag$2]/g;
|
|
unless ($tag<10) {
|
|
my @subf = $field->subfields;
|
|
for my $i (0..$#subf) {
|
|
my $subfieldcode = $subf[$i][0];
|
|
my $subfieldvalue = $subf[$i][1];
|
|
my $tagsubf = $tag.$subfieldcode;
|
|
$summary =~ s/\[(.?.?.?.?)$tagsubf(.*?)]/$1$subfieldvalue$2\[$1$tagsubf$2]/g;
|
|
}
|
|
}
|
|
}
|
|
$summary =~ s/\[(.*?)]//g;
|
|
$summary =~ s/\n/<br>/g;
|
|
$oldbiblio->{summary} = $summary;
|
|
}
|
|
# add spans to search term in results
|
|
foreach my $term ( keys %$span_terms_hashref ) {
|
|
|
|
#warn "term: $term";
|
|
my $old_term = $term;
|
|
if ( length($term) > 3 ) {
|
|
$term =~ s/(.*=|\)|\(|\+|\.|\?)//g;
|
|
|
|
#FIXME: is there a better way to do this?
|
|
$oldbiblio->{'title'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'subtitle'} =~
|
|
s/$term/<span class=term>$&<\/span>/gi;
|
|
|
|
$oldbiblio->{'author'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'publishercode'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'place'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'pages'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'notes'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
$oldbiblio->{'size'} =~ s/$term/<span class=term>$&<\/span>/gi;
|
|
}
|
|
}
|
|
|
|
if ( $i % 2 ) {
|
|
$toggle = "#ffffcc";
|
|
}
|
|
else {
|
|
$toggle = "white";
|
|
}
|
|
$oldbiblio->{'toggle'} = $toggle;
|
|
my @fields = $marcrecord->field($itemtag);
|
|
my @items_loop;
|
|
my $items;
|
|
my $ordered_count = 0;
|
|
my $onloan_count = 0;
|
|
my $wthdrawn_count = 0;
|
|
my $itemlost_count = 0;
|
|
my $itembinding_count = 0;
|
|
my $norequests = 1;
|
|
|
|
foreach my $field (@fields) {
|
|
my $item;
|
|
foreach my $code ( keys %subfieldstosearch ) {
|
|
$item->{$code} = $field->subfield( $subfieldstosearch{$code} );
|
|
}
|
|
if ( $item->{wthdrawn} ) {
|
|
$wthdrawn_count++;
|
|
}
|
|
elsif ( $item->{notforloan} == -1 ) {
|
|
$ordered_count++;
|
|
$norequests = 0;
|
|
}
|
|
elsif ( $item->{itemlost} ) {
|
|
$itemlost_count++;
|
|
}
|
|
elsif ( $item->{binding} ) {
|
|
$itembinding_count++;
|
|
}
|
|
elsif ( ( $item->{onloan} ) && ( $item->{onloan} != '0000-00-00' ) )
|
|
{
|
|
$onloan_count++;
|
|
$norequests = 0;
|
|
}
|
|
else {
|
|
$norequests = 0;
|
|
if ( $item->{'homebranch'} ) {
|
|
$items->{ $item->{'homebranch'} }->{count}++;
|
|
}
|
|
|
|
# Last resort
|
|
elsif ( $item->{'holdingbranch'} ) {
|
|
$items->{ $item->{'homebranch'} }->{count}++;
|
|
}
|
|
$items->{ $item->{homebranch} }->{itemcallnumber} =
|
|
$item->{itemcallnumber};
|
|
$items->{ $item->{homebranch} }->{location} =
|
|
$item->{location};
|
|
}
|
|
} # notforloan, item level and biblioitem level
|
|
for my $key ( keys %$items ) {
|
|
|
|
#warn "key: $key";
|
|
my $this_item = {
|
|
branchname => $branches{$key},
|
|
branchcode => $key,
|
|
count => $items->{$key}->{count},
|
|
itemcallnumber => $items->{$key}->{itemcallnumber},
|
|
location => $items->{$key}->{location},
|
|
};
|
|
push @items_loop, $this_item;
|
|
}
|
|
$oldbiblio->{norequests} = $norequests;
|
|
$oldbiblio->{items_loop} = \@items_loop;
|
|
$oldbiblio->{onloancount} = $onloan_count;
|
|
$oldbiblio->{wthdrawncount} = $wthdrawn_count;
|
|
$oldbiblio->{itemlostcount} = $itemlost_count;
|
|
$oldbiblio->{bindingcount} = $itembinding_count;
|
|
$oldbiblio->{orderedcount} = $ordered_count;
|
|
|
|
# FIXME
|
|
# Ugh ... this is ugly, I'll re-write it better above then delete it
|
|
# my $norequests = 1;
|
|
# my $noitems = 1;
|
|
# if (@items) {
|
|
# $noitems = 0;
|
|
# foreach my $itm (@items) {
|
|
# $norequests = 0 unless $itm->{'itemnotforloan'};
|
|
# }
|
|
# }
|
|
# $oldbiblio->{'noitems'} = $noitems;
|
|
# $oldbiblio->{'norequests'} = $norequests;
|
|
# $oldbiblio->{'even'} = $even = not $even;
|
|
# $oldbiblio->{'itemcount'} = $counts{'total'};
|
|
# my $totalitemcounts = 0;
|
|
# foreach my $key (keys %counts){
|
|
# if ($key ne 'total'){
|
|
# $totalitemcounts+= $counts{$key};
|
|
# $oldbiblio->{'locationhash'}->{$key}=$counts{$key};
|
|
# }
|
|
# }
|
|
# my ($locationtext, $locationtextonly, $notavailabletext) = ('','','');
|
|
# foreach (sort keys %{$oldbiblio->{'locationhash'}}) {
|
|
# if ($_ eq 'notavailable') {
|
|
# $notavailabletext="Not available";
|
|
# my $c=$oldbiblio->{'locationhash'}->{$_};
|
|
# $oldbiblio->{'not-available-p'}=$c;
|
|
# } else {
|
|
# $locationtext.="$_";
|
|
# my $c=$oldbiblio->{'locationhash'}->{$_};
|
|
# if ($_ eq 'Item Lost') {
|
|
# $oldbiblio->{'lost-p'} = $c;
|
|
# } elsif ($_ eq 'Withdrawn') {
|
|
# $oldbiblio->{'withdrawn-p'} = $c;
|
|
# } elsif ($_ eq 'On Loan') {
|
|
# $oldbiblio->{'on-loan-p'} = $c;
|
|
# } else {
|
|
# $locationtextonly.= $_;
|
|
# $locationtextonly.= " ($c)<br/> " if $totalitemcounts > 1;
|
|
# }
|
|
# if ($totalitemcounts>1) {
|
|
# $locationtext.=" ($c)<br/> ";
|
|
# }
|
|
# }
|
|
# }
|
|
# if ($notavailabletext) {
|
|
# $locationtext.= $notavailabletext;
|
|
# } else {
|
|
# $locationtext=~s/, $//;
|
|
# }
|
|
# $oldbiblio->{'location'} = $locationtext;
|
|
# $oldbiblio->{'location-only'} = $locationtextonly;
|
|
# $oldbiblio->{'use-location-flags-p'} = 1;
|
|
|
|
push( @newresults, $oldbiblio );
|
|
}
|
|
return @newresults;
|
|
}
|
|
|
|
|
|
=head2 EditBiblios
|
|
|
|
($countchanged,$listunchanged) = EditBiblios($listbiblios, $tagsubfield,$initvalue,$targetvalue,$test);
|
|
|
|
this function changes all the values $initvalue in subfield $tag$subfield in any record in $listbiblios
|
|
test parameter if set donot perform change to records in database.
|
|
|
|
=over 2
|
|
|
|
=item C<input arg:>
|
|
|
|
* $listbiblios is an array ref to marcrecords to be changed
|
|
* $tagsubfield is the reference of the subfield to change.
|
|
* $initvalue is the value to search the record for
|
|
* $targetvalue is the value to set the subfield to
|
|
* $test is to be set only not to perform changes in database.
|
|
|
|
=item C<Output arg:>
|
|
* $countchanged counts all the changes performed.
|
|
* $listunchanged contains the list of all the biblionumbers of records unchanged.
|
|
|
|
=item C<usage in the script:>
|
|
|
|
=back
|
|
|
|
my ($countchanged, $listunchanged) = EditBiblios($results->{RECORD}, $tagsubfield,$initvalue,$targetvalue);;
|
|
#If one wants to display unchanged records, you should get biblios foreach @$listunchanged
|
|
$template->param(countchanged => $countchanged, loopunchanged=>$listunchanged);
|
|
|
|
=cut
|
|
sub EditBiblios{
|
|
my ($listbiblios,$tagsubfield,$initvalue,$targetvalue,$test)=@_;
|
|
my $countmatched;
|
|
my @unmatched;
|
|
my ($tag,$subfield)=($1,$2) if ($tagsubfield=~/^(\d{1,3})(.)$/);
|
|
my ($bntag,$bnsubf) = GetMarcFromKohaField('biblio.biblionumber');
|
|
|
|
foreach my $usmarc (@$listbiblios){
|
|
my $record=MARC::Record->new_from_usmarc($usmarc);
|
|
my $biblionumber;
|
|
if ($bntag>10){
|
|
$biblionumber = $record->subfield($bntag,$bnsubf);
|
|
}else {
|
|
$biblionumber=$record->field($bntag)->data;
|
|
}
|
|
#GetBiblionumber is to be written.
|
|
#Could be replaced by TransformMarcToKoha (But Would be longer)
|
|
if ($record->field($tag)){
|
|
foreach my $field ($record->field($tag)){
|
|
if ($field->delete_subfield('code' =>$subfield,'match'=>qr($initvalue))){
|
|
$countmatched++;
|
|
$field->update($subfield,$targetvalue) if ($targetvalue);
|
|
}
|
|
}
|
|
# warn $record->as_formatted;
|
|
ModBiblio($record,$biblionumber,GetFrameworkCode($biblionumber)) unless ($test);
|
|
} else {
|
|
push @unmatched, $biblionumber;
|
|
}
|
|
}
|
|
return ($countmatched,\@unmatched);
|
|
}
|
|
|
|
#----------------------------------------------------------------------
|
|
#
|
|
# Non-Zebra GetRecords#
|
|
#----------------------------------------------------------------------
|
|
|
|
=item
|
|
NZgetRecords has the same API as zera getRecords, even if some parameters are not managed
|
|
=cut
|
|
|
|
sub NZgetRecords {
|
|
my (
|
|
$koha_query, $federated_query, $sort_by_ref,
|
|
$servers_ref, $results_per_page, $offset,
|
|
$expanded_facet, $branches, $query_type,
|
|
$scan
|
|
) = @_;
|
|
my $result = NZanalyse($koha_query);
|
|
# use Data::Dumper;
|
|
# warn "==========".@$sort_by_ref[0];
|
|
return (undef,NZorder($result,@$sort_by_ref[0],$results_per_page,$offset),undef);
|
|
}
|
|
|
|
=item
|
|
|
|
NZanalyse : get a CQL string as parameter, and returns a list of biblionumber;title,biblionumber;title,...
|
|
the list is builded from inverted index in nozebra SQL table
|
|
note that title is here only for convenience : the sorting will be very fast when requested on title
|
|
if the sorting is requested on something else, we will have to reread all results, and that may be longer.
|
|
|
|
=cut
|
|
|
|
sub NZanalyse {
|
|
my ($string) = @_;
|
|
# if we have a ", replace the content to discard temporarily any and/or/not inside
|
|
my $commacontent;
|
|
if ($string =~/"/) {
|
|
$string =~ s/"(.*?)"/__X__/;
|
|
$commacontent = $1;
|
|
# print "commacontent : $commacontent\n";
|
|
}
|
|
# split the query string in 3 parts : X AND Y means : $left="X", $operand="AND" and $right="Y"
|
|
# then, call again NZanalyse with $left and $right
|
|
# (recursive until we find a leaf (=> something without and/or/not)
|
|
$string =~ /(.*)( and | or | not )(.*)/;
|
|
my $left = $1;
|
|
my $right = $3;
|
|
my $operand = $2;
|
|
# it's not a leaf, we have a and/or/not
|
|
if ($operand) {
|
|
# reintroduce comma content if needed
|
|
$right =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
$left =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
# print "noeud : $left / $operand / $right\n";
|
|
my $leftresult = NZanalyse($left);
|
|
my $rightresult = NZanalyse($right);
|
|
# OK, we have the results for right and left part of the query
|
|
# depending of operand, intersect, union or exclude both lists
|
|
# to get a result list
|
|
if ($operand eq ' and ') {
|
|
my @leftresult = split /,/, $leftresult;
|
|
# my @rightresult = split /,/,$leftresult;
|
|
my $finalresult;
|
|
# parse the left results, and if the biblionumber exist in the right result, save it in finalresult
|
|
# the result is stored twice, to have the same weight for AND than OR.
|
|
# example : TWO : 61,61,64,121 (two is twice in the biblio #61) / TOWER : 61,64,130
|
|
# result : 61,61,61,61,64,64 for two AND tower : 61 has more weight than 64
|
|
foreach (@leftresult) {
|
|
if ($rightresult =~ "$_,") {
|
|
$finalresult .= "$_,$_,";
|
|
}
|
|
}
|
|
return $finalresult;
|
|
} elsif ($operand eq ' or ') {
|
|
# just merge the 2 strings
|
|
return $leftresult.$rightresult;
|
|
} elsif ($operand eq ' not ') {
|
|
my @leftresult = split /,/, $leftresult;
|
|
# my @rightresult = split /,/,$leftresult;
|
|
my $finalresult;
|
|
foreach (@leftresult) {
|
|
unless ($rightresult =~ "$_,") {
|
|
$finalresult .= "$_,";
|
|
}
|
|
}
|
|
return $finalresult;
|
|
} else {
|
|
# this error is impossible, because of the regexp that isolate the operand, but just in case...
|
|
die "error : operand unknown : $operand for $string";
|
|
}
|
|
# it's a leaf, do the real SQL query and return the result
|
|
} else {
|
|
$string =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
$string =~ s/-|\.|\?|,|;|!|'|\(|\)|\[|\]|{|}|"|<|>|&|\+|\*|\// /g;
|
|
# print "feuille : $string\n";
|
|
# parse the string in in operator/operand/value again
|
|
$string =~ /(.*)(=|>|>=|<|<=)(.*)/;
|
|
my $left = $1;
|
|
my $operator = $2;
|
|
my $right = $3;
|
|
my $results;
|
|
if ($operator) {
|
|
#do a specific search
|
|
my $dbh = C4::Context->dbh;
|
|
$operator='LIKE' if $operator eq '=' and $right=~ /%/;
|
|
my $sth = $dbh->prepare("SELECT biblionumbers FROM nozebra WHERE indexname=? AND value $operator ?");
|
|
# print "$left / $operator / $right\n";
|
|
# split each word, query the DB and build the biblionumbers result
|
|
foreach (split / /,$right) {
|
|
my $biblionumbers;
|
|
$sth->execute($left,$_);
|
|
while (my $line = $sth->fetchrow) {
|
|
$biblionumbers .= $line;
|
|
}
|
|
# do a AND with existing list if there is one, otherwise, use the biblionumbers list as 1st result list
|
|
if ($results) {
|
|
my @leftresult = split /;/, $biblionumbers;
|
|
my $temp;
|
|
foreach (@leftresult) {
|
|
if ($results =~ "$_;") {
|
|
$temp .= "$_;$_;";
|
|
}
|
|
}
|
|
$results = $temp;
|
|
} else {
|
|
$results = $biblionumbers;
|
|
}
|
|
}
|
|
} else {
|
|
#do a complete search (all indexes)
|
|
my $dbh = C4::Context->dbh;
|
|
my $sth = $dbh->prepare("SELECT biblionumbers FROM nozebra WHERE value LIKE ?");
|
|
# split each word, query the DB and build the biblionumbers result
|
|
foreach (split / /,$string) {
|
|
my $biblionumbers;
|
|
$sth->execute($_);
|
|
while (my $line = $sth->fetchrow) {
|
|
$biblionumbers .= $line;
|
|
}
|
|
# do a AND with existing list if there is one, otherwise, use the biblionumbers list as 1st result list
|
|
if ($results) {
|
|
my @leftresult = split /,/, $biblionumbers;
|
|
my $temp;
|
|
foreach (@leftresult) {
|
|
if ($results =~ "$_;") {
|
|
$temp .= "$_;$_;";
|
|
}
|
|
}
|
|
$results = $temp;
|
|
} else {
|
|
$results = $biblionumbers;
|
|
}
|
|
}
|
|
}
|
|
return $results;
|
|
}
|
|
}
|
|
|
|
sub NZorder {
|
|
my ($biblionumbers, $ordering,$results_per_page,$offset) = @_;
|
|
# order title asc by default
|
|
# $ordering = '1=36 <i' unless $ordering;
|
|
$results_per_page=20 unless $results_per_page;
|
|
$offset = 0 unless $offset;
|
|
my $dbh = C4::Context->dbh;
|
|
#
|
|
# order by POPULARITY
|
|
#
|
|
if ($ordering =~ /1=9523/) {
|
|
my %result;
|
|
my %popularity;
|
|
# popularity is not in MARC record, it's builded from a specific query
|
|
my $sth = $dbh->prepare("select sum(issues) from items where biblionumber=?");
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
$result{$biblionumber}=GetMarcBiblio($biblionumber);
|
|
$sth->execute($biblionumber);
|
|
my $popularity= $sth->fetchrow ||0;
|
|
# hint : the key is popularity.title because we can have
|
|
# many results with the same popularity. In this cas, sub-ordering is done by title
|
|
# we also have biblionumber to avoid bug for 2 biblios with the same title & popularity
|
|
# (un-frequent, I agree, but we won't forget anything that way ;-)
|
|
$popularity{sprintf("%10d",$popularity).$title.$biblionumber} = $biblionumber;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq '1=9523 >i') { # sort popularity DESC
|
|
foreach my $key (sort {$b <=> $a} (keys %popularity)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$popularity{$key}}->as_usmarc();
|
|
}
|
|
} else { # sort popularity ASC
|
|
foreach my $key (sort (keys %popularity)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$popularity{$key}}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY author
|
|
#
|
|
} elsif ($ordering eq '1=1003 <i'){
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my $author;
|
|
if (C4::Context->preference('marcflavour') eq 'UNIMARC') {
|
|
$author=$record->subfield('200','f');
|
|
$author=$record->subfield('700','a') unless $author;
|
|
} else {
|
|
$author=$record->subfield('100','a');
|
|
}
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$author.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq '1=1003 <i') { # sort by title desc
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $a <=> $b } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY callnumber
|
|
#
|
|
} elsif ($ordering eq '1=20 <i'){
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my $callnumber;
|
|
my ($callnumber_tag,$callnumber_subfield)=GetMarcFromKohaField($dbh,'items.itemcallnumber');
|
|
($callnumber_tag,$callnumber_subfield)= GetMarcFromKohaField('biblioitems.callnumber') unless $callnumber_tag;
|
|
if (C4::Context->preference('marcflavour') eq 'UNIMARC') {
|
|
$callnumber=$record->subfield('200','f');
|
|
} else {
|
|
$callnumber=$record->subfield('100','a');
|
|
}
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$callnumber.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq '1=1003 <i') { # sort by title desc
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $a <=> $b } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
} elsif ($ordering =~ /1=31/){ #pub year
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my ($publicationyear_tag,$publicationyear_subfield)=GetMarcFromKohaField($dbh,'biblioitems.publicationyear');
|
|
my $publicationyear=$record->subfield($publicationyear_tag,$publicationyear_subfield);
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$publicationyear.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq '1=31 <i') { # sort by title desc
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $a <=> $b } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY title
|
|
#
|
|
} elsif ($ordering =~ /1=36/) {
|
|
# the title is in the biblionumbers string, so we just need to build a hash, sort it and return
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
# hint & speed improvement : we can order without reading the record
|
|
# so order, and read records only for the requested page !
|
|
$result{$title.$biblionumber}=$biblionumber;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq '1=36 <i') { # sort by title desc
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $a <=> $b } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
}
|
|
# for the requested page, replace biblionumber by the complete record
|
|
# speed improvement : avoid reading too much things
|
|
for (my $counter=$offset;$counter<=$offset+$results_per_page;$counter++) {
|
|
$result_hash->{'RECORDS'}[$counter] = GetMarcBiblio($result_hash->{'RECORDS'}[$counter])->as_usmarc;
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
} else {
|
|
#
|
|
# order by ranking
|
|
#
|
|
# we need 2 hashes to order by ranking : the 1st one to count the ranking, the 2nd to order by ranking
|
|
my %result;
|
|
my %count_ranking;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
$title =~ /(.*)-(\d)/;
|
|
# get weight
|
|
my $ranking =$2;
|
|
# note that we + the ranking because ranking is calculated on weight of EACH term requested.
|
|
# if we ask for "two towers", and "two" has weight 2 in biblio N, and "towers" has weight 4 in biblio N
|
|
# biblio N has ranking = 6
|
|
$count_ranking{$biblionumber} =+ $ranking;
|
|
}
|
|
# build the result by "inverting" the count_ranking hash
|
|
# hing : as usual, we don't order by ranking only, to avoid having only 1 result for each rank. We build an hash on concat(ranking,biblionumber) instead
|
|
# warn "counting";
|
|
foreach (keys %count_ranking) {
|
|
$result{sprintf("%10d",$count_ranking{$_}).'-'.$_} = $_;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
foreach my $key (sort {$b <=> $a} (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
# for the requested page, replace biblionumber by the complete record
|
|
# speed improvement : avoid reading too much things
|
|
for (my $counter=$offset;$counter<=$offset+$results_per_page;$counter++) {
|
|
$result_hash->{'RECORDS'}[$counter] = GetMarcBiblio($result_hash->{'RECORDS'}[$counter])->as_usmarc;
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
}
|
|
}
|
|
|
|
END { } # module clean-up code here (global destructor)
|
|
|
|
1;
|
|
__END__
|
|
|
|
=head1 AUTHOR
|
|
|
|
Koha Developement team <info@koha.org>
|
|
|
|
=cut
|