c698f23940
Signed-off-by: Chris Cormack <crc@liblime.com> Signed-off-by: Joshua Ferraro <jmf@liblime.com>
1643 lines
64 KiB
Perl
1643 lines
64 KiB
Perl
package C4::Search;
|
|
|
|
# This file is part of Koha.
|
|
#
|
|
# Koha is free software; you can redistribute it and/or modify it under the
|
|
# terms of the GNU General Public License as published by the Free Software
|
|
# Foundation; either version 2 of the License, or (at your option) any later
|
|
# version.
|
|
#
|
|
# Koha is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
# WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR
|
|
# A PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License along with
|
|
# Koha; if not, write to the Free Software Foundation, Inc., 59 Temple Place,
|
|
# Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
use strict;
|
|
require Exporter;
|
|
use C4::Context;
|
|
use C4::Biblio; # GetMarcFromKohaField
|
|
use C4::Koha; # getFacets
|
|
use Lingua::Stem;
|
|
use C4::Date;
|
|
|
|
use vars qw($VERSION @ISA @EXPORT @EXPORT_OK %EXPORT_TAGS);
|
|
|
|
# set the version for version checking
|
|
$VERSION = 3.00;
|
|
$DEBUG=1;
|
|
|
|
=head1 NAME
|
|
|
|
C4::Search - Functions for searching the Koha catalog.
|
|
|
|
=head1 SYNOPSIS
|
|
|
|
see opac/opac-search.pl or catalogue/search.pl for example of usage
|
|
|
|
=head1 DESCRIPTION
|
|
|
|
This module provides the searching facilities for the Koha into a zebra catalog.
|
|
|
|
=head1 FUNCTIONS
|
|
|
|
=cut
|
|
|
|
@ISA = qw(Exporter);
|
|
@EXPORT = qw(
|
|
&SimpleSearch
|
|
&findseealso
|
|
&FindDuplicate
|
|
&searchResults
|
|
&getRecords
|
|
&buildQuery
|
|
&NZgetRecords
|
|
&ModBiblios
|
|
);
|
|
|
|
# make all your functions, whether exported or not;
|
|
|
|
=head2 findseealso($dbh,$fields);
|
|
|
|
C<$dbh> is a link to the DB handler.
|
|
|
|
use C4::Context;
|
|
my $dbh =C4::Context->dbh;
|
|
|
|
C<$fields> is a reference to the fields array
|
|
|
|
This function modify the @$fields array and add related fields to search on.
|
|
|
|
=cut
|
|
|
|
sub findseealso {
|
|
my ( $dbh, $fields ) = @_;
|
|
my $tagslib = GetMarcStructure( 1 );
|
|
for ( my $i = 0 ; $i <= $#{$fields} ; $i++ ) {
|
|
my ($tag) = substr( @$fields[$i], 1, 3 );
|
|
my ($subfield) = substr( @$fields[$i], 4, 1 );
|
|
@$fields[$i] .= ',' . $tagslib->{$tag}->{$subfield}->{seealso}
|
|
if ( $tagslib->{$tag}->{$subfield}->{seealso} );
|
|
}
|
|
}
|
|
|
|
=head2 FindDuplicate
|
|
|
|
($biblionumber,$biblionumber,$title) = FindDuplicate($record);
|
|
|
|
=cut
|
|
|
|
sub FindDuplicate {
|
|
my ($record) = @_;
|
|
my $dbh = C4::Context->dbh;
|
|
my $result = TransformMarcToKoha( $dbh, $record, '' );
|
|
my $sth;
|
|
my $query;
|
|
my $search;
|
|
my $type;
|
|
my ( $biblionumber, $title );
|
|
|
|
# search duplicate on ISBN, easy and fast..
|
|
# ... normalize first
|
|
if ( $result->{isbn} ) {
|
|
$result->{isbn} =~ s/\(.*$//;
|
|
$result->{isbn} =~ s/\s+$//;
|
|
}
|
|
#$search->{'avoidquerylog'}=1;
|
|
if ( $result->{isbn} ) {
|
|
$query = "isbn=$result->{isbn}";
|
|
}
|
|
else {
|
|
$result->{title} =~ s /\\//g;
|
|
$result->{title} =~ s /\"//g;
|
|
$result->{title} =~ s /\(//g;
|
|
$result->{title} =~ s /\)//g;
|
|
# remove valid operators
|
|
$result->{title} =~ s/(and|or|not)//g;
|
|
$query = "ti,ext=$result->{title}";
|
|
$query .= " and mt=$result->{itemtype}" if ($result->{itemtype});
|
|
if ($result->{author}){
|
|
$result->{author} =~ s /\\//g;
|
|
$result->{author} =~ s /\"//g;
|
|
$result->{author} =~ s /\(//g;
|
|
$result->{author} =~ s /\)//g;
|
|
# remove valid operators
|
|
$result->{author} =~ s/(and|or|not)//g;
|
|
$query .= " and au,ext=$result->{author}";
|
|
}
|
|
}
|
|
my ($error,$searchresults) =
|
|
SimpleSearch($query); # FIXME :: hardcoded !
|
|
my @results;
|
|
foreach my $possible_duplicate_record (@$searchresults) {
|
|
my $marcrecord =
|
|
MARC::Record->new_from_usmarc($possible_duplicate_record);
|
|
my $result = TransformMarcToKoha( $dbh, $marcrecord, '' );
|
|
|
|
# FIXME :: why 2 $biblionumber ?
|
|
if ($result){
|
|
push @results, $result->{'biblionumber'};
|
|
push @results, $result->{'title'};
|
|
}
|
|
}
|
|
return @results;
|
|
}
|
|
|
|
=head2 SimpleSearch
|
|
|
|
($error,$results) = SimpleSearch($query,@servers);
|
|
|
|
this function performs a simple search on the catalog using zoom.
|
|
|
|
=over 2
|
|
|
|
=item C<input arg:>
|
|
|
|
* $query could be a simple keyword or a complete CCL query wich is depending on your ccl file.
|
|
* @servers is optionnal. default one is read on koha.xml
|
|
|
|
=item C<Output arg:>
|
|
* $error is a string which containt the description error if there is one. Else it's empty.
|
|
* \@results is an array of marc record.
|
|
|
|
=item C<usage in the script:>
|
|
|
|
=back
|
|
|
|
my ($error, $marcresults) = SimpleSearch($query);
|
|
|
|
if (defined $error) {
|
|
$template->param(query_error => $error);
|
|
warn "error: ".$error;
|
|
output_html_with_http_headers $input, $cookie, $template->output;
|
|
exit;
|
|
}
|
|
|
|
my $hits = scalar @$marcresults;
|
|
my @results;
|
|
|
|
for(my $i=0;$i<$hits;$i++) {
|
|
my %resultsloop;
|
|
my $marcrecord = MARC::File::USMARC::decode($marcresults->[$i]);
|
|
my $biblio = TransformMarcToKoha(C4::Context->dbh,$marcrecord,'');
|
|
|
|
#build the hash for the template.
|
|
$resultsloop{highlight} = ($i % 2)?(1):(0);
|
|
$resultsloop{title} = $biblio->{'title'};
|
|
$resultsloop{subtitle} = $biblio->{'subtitle'};
|
|
$resultsloop{biblionumber} = $biblio->{'biblionumber'};
|
|
$resultsloop{author} = $biblio->{'author'};
|
|
$resultsloop{publishercode} = $biblio->{'publishercode'};
|
|
$resultsloop{publicationyear} = $biblio->{'publicationyear'};
|
|
|
|
push @results, \%resultsloop;
|
|
}
|
|
$template->param(result=>\@results);
|
|
|
|
=cut
|
|
|
|
sub SimpleSearch {
|
|
my $query = shift;
|
|
if (C4::Context->preference('NoZebra')) {
|
|
my $result = NZorder(NZanalyse($query))->{'biblioserver'}->{'RECORDS'};
|
|
return (undef,$result);
|
|
} else {
|
|
my @servers = @_;
|
|
my @results;
|
|
my @tmpresults;
|
|
my @zconns;
|
|
return ( "No query entered", undef ) unless $query;
|
|
|
|
#@servers = (C4::Context->config("biblioserver")) unless @servers;
|
|
@servers =
|
|
("biblioserver") unless @servers
|
|
; # FIXME hardcoded value. See catalog/search.pl & opac-search.pl too.
|
|
|
|
# Connect & Search
|
|
for ( my $i = 0 ; $i < @servers ; $i++ ) {
|
|
$zconns[$i] = C4::Context->Zconn( $servers[$i], 1 );
|
|
$tmpresults[$i] =
|
|
$zconns[$i]
|
|
->search( new ZOOM::Query::CCL2RPN( $query, $zconns[$i] ) );
|
|
|
|
# getting error message if one occured.
|
|
my $error =
|
|
$zconns[$i]->errmsg() . " ("
|
|
. $zconns[$i]->errcode() . ") "
|
|
. $zconns[$i]->addinfo() . " "
|
|
. $zconns[$i]->diagset();
|
|
|
|
return ( $error, undef ) if $zconns[$i]->errcode();
|
|
}
|
|
my $hits;
|
|
my $ev;
|
|
while ( ( my $i = ZOOM::event( \@zconns ) ) != 0 ) {
|
|
$ev = $zconns[ $i - 1 ]->last_event();
|
|
if ( $ev == ZOOM::Event::ZEND ) {
|
|
$hits = $tmpresults[ $i - 1 ]->size();
|
|
}
|
|
if ( $hits > 0 ) {
|
|
for ( my $j = 0 ; $j < $hits ; $j++ ) {
|
|
my $record = $tmpresults[ $i - 1 ]->record($j)->raw();
|
|
push @results, $record;
|
|
}
|
|
}
|
|
}
|
|
return ( undef, \@results );
|
|
}
|
|
}
|
|
|
|
# performs the search
|
|
sub getRecords {
|
|
my (
|
|
$koha_query, $simple_query, $sort_by_ref,
|
|
$servers_ref, $results_per_page, $offset,
|
|
$expanded_facet, $branches, $query_type,
|
|
$scan
|
|
) = @_;
|
|
# warn "Query : $koha_query";
|
|
my @servers = @$servers_ref;
|
|
my @sort_by = @$sort_by_ref;
|
|
|
|
# create the zoom connection and query object
|
|
my $zconn;
|
|
my @zconns;
|
|
my @results;
|
|
my $results_hashref = ();
|
|
|
|
### FACETED RESULTS
|
|
my $facets_counter = ();
|
|
my $facets_info = ();
|
|
my $facets = getFacets();
|
|
|
|
#### INITIALIZE SOME VARS USED CREATE THE FACETED RESULTS
|
|
my @facets_loop; # stores the ref to array of hashes for template
|
|
for ( my $i = 0 ; $i < @servers ; $i++ ) {
|
|
$zconns[$i] = C4::Context->Zconn( $servers[$i], 1 );
|
|
|
|
# perform the search, create the results objects
|
|
# if this is a local search, use the $koha-query, if it's a federated one, use the federated-query
|
|
my $query_to_use;
|
|
if ( $servers[$i] =~ /biblioserver/ ) {
|
|
$query_to_use = $koha_query;
|
|
}
|
|
else {
|
|
$query_to_use = $simple_query;
|
|
}
|
|
|
|
$query_to_use = $simple_query if $scan;
|
|
|
|
# check if we've got a query_type defined
|
|
eval {
|
|
if ($query_type)
|
|
{
|
|
if ( $query_type =~ /^ccl/ ) {
|
|
$query_to_use =~
|
|
s/\:/\=/g; # change : to = last minute (FIXME)
|
|
|
|
# warn "CCL : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
elsif ( $query_type =~ /^cql/ ) {
|
|
|
|
# warn "CQL : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CQL( $query_to_use, $zconns[$i] ) );
|
|
}
|
|
elsif ( $query_type =~ /^pqf/ ) {
|
|
|
|
# warn "PQF : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::PQF( $query_to_use, $zconns[$i] ) );
|
|
}
|
|
}
|
|
else {
|
|
if ($scan) {
|
|
warn "preparing to scan:$query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->scan(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
else {
|
|
# warn "LAST : $query_to_use";
|
|
$results[$i] =
|
|
$zconns[$i]->search(
|
|
new ZOOM::Query::CCL2RPN( $query_to_use, $zconns[$i] )
|
|
);
|
|
}
|
|
}
|
|
};
|
|
if ($@) {
|
|
warn "WARNING: query problem with $query_to_use " . $@;
|
|
}
|
|
|
|
# concatenate the sort_by limits and pass them to the results object
|
|
my $sort_by;
|
|
foreach my $sort (@sort_by) {
|
|
if ($sort eq "author_az") {
|
|
$sort_by.="1=1003 <i ";
|
|
}
|
|
elsif ($sort eq "author_za") {
|
|
$sort_by.="1=1003 >i ";
|
|
}
|
|
elsif ($sort eq "popularity_asc") {
|
|
$sort_by.="1=9003 <i ";
|
|
}
|
|
elsif ($sort eq "popularity_dsc") {
|
|
$sort_by.="1=9003 >i ";
|
|
}
|
|
elsif ($sort eq "call_number_asc") {
|
|
$sort_by.="1=20 <i ";
|
|
}
|
|
elsif ($sort eq "call_number_dsc") {
|
|
$sort_by.="1=20 >i ";
|
|
}
|
|
elsif ($sort eq "pubdate_asc") {
|
|
$sort_by.="1=31 <i ";
|
|
}
|
|
elsif ($sort eq "pubdate_dsc") {
|
|
$sort_by.="1=31 >i ";
|
|
}
|
|
elsif ($sort eq "acqdate_asc") {
|
|
$sort_by.="1=32 <i ";
|
|
}
|
|
elsif ($sort eq "acqdate_dsc") {
|
|
$sort_by.="1=32 >i ";
|
|
}
|
|
elsif ($sort eq "title_az") {
|
|
$sort_by.="1=4 <i ";
|
|
}
|
|
elsif ($sort eq "title_za") {
|
|
$sort_by.="1=4 >i ";
|
|
}
|
|
}
|
|
if ($sort_by) {
|
|
if ( $results[$i]->sort( "yaz", $sort_by ) < 0) {
|
|
warn "WARNING sort $sort_by failed";
|
|
}
|
|
}
|
|
}
|
|
while ( ( my $i = ZOOM::event( \@zconns ) ) != 0 ) {
|
|
my $ev = $zconns[ $i - 1 ]->last_event();
|
|
if ( $ev == ZOOM::Event::ZEND ) {
|
|
my $size = $results[ $i - 1 ]->size();
|
|
if ( $size > 0 ) {
|
|
my $results_hash;
|
|
#$results_hash->{'server'} = $servers[$i-1];
|
|
# loop through the results
|
|
$results_hash->{'hits'} = $size;
|
|
my $times;
|
|
if ( $offset + $results_per_page <= $size ) {
|
|
$times = $offset + $results_per_page;
|
|
}
|
|
else {
|
|
$times = $size;
|
|
}
|
|
for ( my $j = $offset ; $j < $times ; $j++ )
|
|
{ #(($offset+$count<=$size) ? ($offset+$count):$size) ; $j++){
|
|
my $records_hash;
|
|
my $record;
|
|
my $facet_record;
|
|
## This is just an index scan
|
|
if ($scan) {
|
|
my ( $term, $occ ) = $results[ $i - 1 ]->term($j);
|
|
# here we create a minimal MARC record and hand it off to the
|
|
# template just like a normal result ... perhaps not ideal, but
|
|
# it works for now
|
|
my $tmprecord = MARC::Record->new();
|
|
$tmprecord->encoding('UTF-8');
|
|
my $tmptitle;
|
|
|
|
# srote the minimal record in author/title (depending on MARC flavour)
|
|
if ( C4::Context->preference("marcflavour") eq
|
|
"UNIMARC" )
|
|
{
|
|
$tmptitle = MARC::Field->new(
|
|
'200', ' ', ' ',
|
|
a => $term,
|
|
f => $occ
|
|
);
|
|
}
|
|
else {
|
|
$tmptitle = MARC::Field->new(
|
|
'245', ' ', ' ',
|
|
a => $term,
|
|
b => $occ
|
|
);
|
|
}
|
|
$tmprecord->append_fields($tmptitle);
|
|
$results_hash->{'RECORDS'}[$j] =
|
|
$tmprecord->as_usmarc();
|
|
}
|
|
else {
|
|
$record = $results[ $i - 1 ]->record($j)->raw();
|
|
|
|
#warn "RECORD $j:".$record;
|
|
$results_hash->{'RECORDS'}[$j] =
|
|
$record; # making a reference to a hash
|
|
# Fill the facets while we're looping
|
|
$facet_record = MARC::Record->new_from_usmarc($record);
|
|
|
|
#warn $servers[$i-1].$facet_record->title();
|
|
for ( my $k = 0 ; $k <= @$facets ; $k++ ) {
|
|
if ( $facets->[$k] ) {
|
|
my @fields;
|
|
for my $tag ( @{ $facets->[$k]->{'tags'} } ) {
|
|
push @fields, $facet_record->field($tag);
|
|
}
|
|
for my $field (@fields) {
|
|
my @subfields = $field->subfields();
|
|
for my $subfield (@subfields) {
|
|
my ( $code, $data ) = @$subfield;
|
|
if ( $code eq
|
|
$facets->[$k]->{'subfield'} )
|
|
{
|
|
$facets_counter->{ $facets->[$k]
|
|
->{'link_value'} }->{$data}++;
|
|
}
|
|
}
|
|
}
|
|
$facets_info->{ $facets->[$k]->{'link_value'} }
|
|
->{'label_value'} =
|
|
$facets->[$k]->{'label_value'};
|
|
$facets_info->{ $facets->[$k]->{'link_value'} }
|
|
->{'expanded'} = $facets->[$k]->{'expanded'};
|
|
}
|
|
}
|
|
}
|
|
}
|
|
$results_hashref->{ $servers[ $i - 1 ] } = $results_hash;
|
|
}
|
|
|
|
#print "connection ", $i-1, ": $size hits";
|
|
#print $results[$i-1]->record(0)->render() if $size > 0;
|
|
# BUILD FACETS
|
|
for my $link_value (
|
|
sort { $facets_counter->{$b} <=> $facets_counter->{$a} }
|
|
keys %$facets_counter
|
|
)
|
|
{
|
|
my $expandable;
|
|
my $number_of_facets;
|
|
my @this_facets_array;
|
|
for my $one_facet (
|
|
sort {
|
|
$facets_counter->{$link_value}
|
|
->{$b} <=> $facets_counter->{$link_value}->{$a}
|
|
} keys %{ $facets_counter->{$link_value} }
|
|
)
|
|
{
|
|
$number_of_facets++;
|
|
if ( ( $number_of_facets < 6 )
|
|
|| ( $expanded_facet eq $link_value )
|
|
|| ( $facets_info->{$link_value}->{'expanded'} ) )
|
|
{
|
|
|
|
# sanitize the link value ), ( will cause errors with CCL
|
|
my $facet_link_value = $one_facet;
|
|
$facet_link_value =~ s/(\(|\))/ /g;
|
|
|
|
# fix the length that will display in the label
|
|
my $facet_label_value = $one_facet;
|
|
$facet_label_value = substr( $one_facet, 0, 20 ) . "..."
|
|
unless length($facet_label_value) <= 20;
|
|
|
|
# well, if it's a branch, label by the name, not the code
|
|
if ( $link_value =~ /branch/ ) {
|
|
$facet_label_value =
|
|
$branches->{$one_facet}->{'branchname'};
|
|
}
|
|
|
|
# but we're down with the whole label being in the link's title
|
|
my $facet_title_value = $one_facet;
|
|
|
|
push @this_facets_array,
|
|
(
|
|
{
|
|
facet_count =>
|
|
$facets_counter->{$link_value}->{$one_facet},
|
|
facet_label_value => $facet_label_value,
|
|
facet_title_value => $facet_title_value,
|
|
facet_link_value => $facet_link_value,
|
|
type_link_value => $link_value,
|
|
},
|
|
);
|
|
}
|
|
}
|
|
unless ( $facets_info->{$link_value}->{'expanded'} ) {
|
|
$expandable = 1
|
|
if ( ( $number_of_facets > 6 )
|
|
&& ( $expanded_facet ne $link_value ) );
|
|
}
|
|
push @facets_loop,
|
|
(
|
|
{
|
|
type_link_value => $link_value,
|
|
type_id => $link_value . "_id",
|
|
type_label =>
|
|
$facets_info->{$link_value}->{'label_value'},
|
|
facets => \@this_facets_array,
|
|
expandable => $expandable,
|
|
expand => $link_value,
|
|
}
|
|
);
|
|
}
|
|
}
|
|
}
|
|
return ( undef, $results_hashref, \@facets_loop );
|
|
}
|
|
|
|
# STOPWORDS
|
|
sub _remove_stopwords {
|
|
my ($operand,$index) = @_;
|
|
# phrase and exact-qualified indexes shoudln't have stopwords removed
|
|
if ($index!~m/phr|ext/){
|
|
# remove stopwords from operand : parse all stopwords & remove them (case insensitive)
|
|
# we use IsAlpha unicode definition, to deal correctly with diacritics.
|
|
# otherwise, a french word like "leçon" woudl be split into "le" "çon", le
|
|
# is an empty word, we get "çon" and wouldn't find anything...
|
|
foreach (keys %{C4::Context->stopwords}) {
|
|
next if ($_ =~/(and|or|not)/); # don't remove operators
|
|
$operand=~ s/\P{IsAlpha}$_\P{IsAlpha}/ /i;
|
|
$operand=~ s/^$_\P{IsAlpha}/ /i;
|
|
$operand=~ s/\P{IsAlpha}$_$/ /i;
|
|
}
|
|
}
|
|
return $operand;
|
|
}
|
|
|
|
# TRUNCATION
|
|
sub _detect_truncation {
|
|
my ($operand,$index) = @_;
|
|
my (@nontruncated,@righttruncated,@lefttruncated,@rightlefttruncated,@regexpr);
|
|
$operand =~s/^ //g;
|
|
my @wordlist= split (/\s/,$operand);
|
|
foreach my $word (@wordlist){
|
|
if ($word=~s/^\*([^\*]+)\*$/$1/){
|
|
push @rightlefttruncated,$word;
|
|
}
|
|
elsif($word=~s/^\*([^\*]+)$/$1/){
|
|
push @lefttruncated,$word;
|
|
}
|
|
elsif ($word=~s/^([^\*]+)\*$/$1/){
|
|
push @righttruncated,$word;
|
|
}
|
|
elsif (index($word,"*")<0){
|
|
push @nontruncated,$word;
|
|
}
|
|
else {
|
|
push @regexpr,$word;
|
|
}
|
|
}
|
|
return (\@nontruncated,\@righttruncated,\@lefttruncated,\@rightlefttruncated,\@regexpr);
|
|
}
|
|
|
|
sub _build_stemmed_operand {
|
|
my ($operand) = @_;
|
|
my $stemmed_operand;
|
|
# FIXME: the locale should be set based on the user's language and/or search choice
|
|
my $stemmer = Lingua::Stem->new( -locale => 'EN-US' );
|
|
# FIXME: these should be stored in the db so the librarian can modify the behavior
|
|
$stemmer->add_exceptions(
|
|
{
|
|
'and' => 'and',
|
|
'or' => 'or',
|
|
'not' => 'not',
|
|
}
|
|
|
|
);
|
|
my @words = split( / /, $operand );
|
|
my $stems = $stemmer->stem(@words);
|
|
for my $stem (@$stems) {
|
|
$stemmed_operand .= "$stem";
|
|
$stemmed_operand .= "?" unless ( $stem =~ /(and$|or$|not$)/ ) || ( length($stem) < 3 );
|
|
$stemmed_operand .= " ";
|
|
}
|
|
#warn "STEMMED OPERAND: $stemmed_operand";
|
|
return $stemmed_operand;
|
|
}
|
|
|
|
sub _build_weighted_query {
|
|
# FIELD WEIGHTING - This is largely experimental stuff. What I'm committing works
|
|
# pretty well but will work much better when we have an actual query parser
|
|
my ($operand,$stemmed_operand,$index) = @_;
|
|
my $stemming = C4::Context->preference("QueryStemming") || 0;
|
|
my $weight_fields = C4::Context->preference("QueryWeightFields") || 0;
|
|
my $fuzzy_enabled = C4::Context->preference("QueryFuzzy") || 0;
|
|
|
|
my $weighted_query .= "(rk=("; # Specifies that we're applying rank
|
|
|
|
# Keyword, or, no index specified
|
|
if ( ( $index eq 'kw' ) || ( !$index ) ) {
|
|
$weighted_query .= "Title-cover,ext,r1=\"$operand\""; # exact title-cover
|
|
$weighted_query .= " or ti,ext,r2=\"$operand\""; # exact title
|
|
$weighted_query .= " or ti,phr,r3=\"$operand\""; # phrase title
|
|
#$weighted_query .= " or any,ext,r4=$operand"; # exact any
|
|
#$weighted_query .=" or kw,wrdl,r5=\"$operand\""; # word list any
|
|
$weighted_query .= " or wrd,fuzzy,r8=\"$operand\"" if $fuzzy_enabled; # add fuzzy, word list
|
|
$weighted_query .= " or wrd,right-Truncation,r9=\"$stemmed_operand\"" if ($stemming and $stemmed_operand); # add stemming, right truncation
|
|
# embedded sorting: 0 a-z; 1 z-a
|
|
# $weighted_query .= ") or (sort1,aut=1";
|
|
}
|
|
# if the index already has more than one qualifier, just wrap the operand
|
|
# in quotes and pass it back
|
|
elsif ($index =~ ',') {
|
|
$weighted_query .=" $index=\"$operand\"";
|
|
}
|
|
#TODO: build better cases based on specific search indexes
|
|
else {
|
|
$weighted_query .= " $index,ext,r1=\"$operand\""; # exact index
|
|
#$weighted_query .= " or (title-sort-az=0 or $index,startswithnt,st-word,r3=$operand #)";
|
|
$weighted_query .= " or $index,phr,r3=\"$operand\""; # phrase index
|
|
$weighted_query .= " or $index,rt,wrd,r3=\"$operand\""; # word list index
|
|
}
|
|
$weighted_query .= "))"; # close rank specification
|
|
return $weighted_query;
|
|
}
|
|
|
|
# build the query itself
|
|
sub buildQuery {
|
|
my ( $operators, $operands, $indexes, $limits, $sort_by ) = @_;
|
|
|
|
my @operators = @$operators if $operators;
|
|
my @indexes = @$indexes if $indexes;
|
|
my @operands = @$operands if $operands;
|
|
my @limits = @$limits if $limits;
|
|
my @sort_by = @$sort_by if $sort_by;
|
|
|
|
my $stemming = C4::Context->preference("QueryStemming") || 0;
|
|
my $weight_fields = C4::Context->preference("QueryWeightFields") || 0;
|
|
my $fuzzy_enabled = C4::Context->preference("QueryFuzzy") || 0;
|
|
|
|
my $query = $operands[0];
|
|
my $simple_query = $operands[0];
|
|
my $query_cgi;
|
|
my $query_search_desc;
|
|
|
|
my $limit;
|
|
my $limit_cgi;
|
|
my $limit_desc;
|
|
|
|
# STEP I: determine if this is a form-based / simple query or if it's complex (if complex,
|
|
# pass it off to zebra directly)
|
|
|
|
# check if this is a known query language query, if it is, return immediately,
|
|
# the user is responsible for constructing valid syntax:
|
|
if ( $query =~ /^ccl=/ ) {
|
|
return ( undef, $', $', $', '', '', '', 'ccl' );
|
|
}
|
|
if ( $query =~ /^cql=/ ) {
|
|
return ( undef, $', $', $', '', '', '', 'cql' );
|
|
}
|
|
if ( $query =~ /^pqf=/ ) {
|
|
return ( undef, $', $', $', '', '', '', 'pqf' );
|
|
}
|
|
|
|
# FIXME: this is bound to be broken now
|
|
if ( $query =~ /(\(|\))/ ) { # sorry, too complex, assume CCL
|
|
return ( undef, $query, $query_cgi, $query_search_desc, $limit, $limit_cgi, $limit_desc, 'ccl' );
|
|
}
|
|
|
|
# form-based queries are limited to non-nested at a specific depth, so we can easily
|
|
# modify the incoming query operands and indexes to do stemming and field weighting
|
|
# Once we do so, we'll end up with a value in $query, just like if we had an
|
|
# incoming $query from the user
|
|
else {
|
|
$query = ""; # clear it out so we can populate properly with field-weighted stemmed query
|
|
my $previous_operand; # a flag used to keep track if there was a previous query
|
|
# if there was, we can apply the current operator
|
|
# for every operand
|
|
for ( my $i = 0 ; $i <= @operands ; $i++ ) {
|
|
|
|
# COMBINE OPERANDS, INDEXES AND OPERATORS
|
|
if ( $operands[$i] ) {
|
|
|
|
$weight_fields = 0 if $operands[$i] =~ /(:|=)/;
|
|
my $operand = $operands[$i];
|
|
my $index = $indexes[$i];
|
|
|
|
# if there's no index, don't use one, it will throw a CCL error
|
|
my $index_plus = "$index:" if $index;
|
|
my $index_plus_comma="$index," if $index;
|
|
|
|
# Remove Stopwords
|
|
$operand = _remove_stopwords($operand,$index);
|
|
warn "OPERAND w/out STOPWORDS: >$operand<" if $DEBUG;
|
|
|
|
my $indexes_set;
|
|
|
|
# Detect Truncation
|
|
my ($nontruncated,$righttruncated,$lefttruncated,$rightlefttruncated,$regexpr);
|
|
my $truncated_operand;
|
|
($nontruncated,$righttruncated,$lefttruncated,$rightlefttruncated,$regexpr) = _detect_truncation($operand,$index);
|
|
warn "TRUNCATION: NON:>@$nontruncated< RIGHT:>@$righttruncated< LEFT:>@$lefttruncated< RIGHTLEFT:>@$rightlefttruncated< REGEX:>@$regexpr<" if $DEBUG;
|
|
# Apply Truncation
|
|
# Problem is when build_weights gets ahold if this is wraps in quotes which breaks the truncation :/
|
|
if (scalar(@$righttruncated)+scalar(@$lefttruncated)+scalar(@$rightlefttruncated)>0){
|
|
$indexes_set = 1;
|
|
undef $weight_fields;
|
|
my $previous_truncation_operand;
|
|
if (scalar(@$nontruncated)>0) {
|
|
$truncated_operand.= "$index_plus @$nontruncated ";
|
|
$previous_truncation_operand = 1;
|
|
}
|
|
if (scalar(@$righttruncated)>0){
|
|
$truncated_operand .= "and " if $previous_truncation_operand;
|
|
$truncated_operand .= "$index_plus_comma"."rtrn:@$righttruncated ";
|
|
$previous_truncation_operand = 1;
|
|
}
|
|
if (scalar(@$lefttruncated)>0){
|
|
$truncated_operand .= "and " if $previous_truncation_operand;
|
|
$truncated_operand .= "$index_plus_comma"."ltrn:@$lefttruncated ";
|
|
$previous_truncation_operand = 1;
|
|
}
|
|
if (scalar(@$rightlefttruncated)>0){
|
|
$truncated_operand .= "and " if $previous_truncation_operand;
|
|
$truncated_operand .= "$index_plus_comma"."rltrn:@$rightlefttruncated ";
|
|
$previous_truncation_operand = 1;
|
|
}
|
|
}
|
|
$operand = $truncated_operand if $truncated_operand;
|
|
warn "TRUNCATED OPERAND: >$truncated_operand<" if $DEBUG;
|
|
|
|
# Handle Stemming
|
|
my $stemmed_operand;
|
|
$stemmed_operand = _build_stemmed_operand($operand) if $stemming;
|
|
warn "STEMMED OPERAND: >$stemmed_operand<" if $DEBUG;
|
|
|
|
# Handle Field Weighting
|
|
my $weighted_operand;
|
|
$weighted_operand = _build_weighted_query($operand,$stemmed_operand,$index) if $weight_fields;
|
|
warn "FIELD WEIGHTED OPERAND: >$weighted_operand<" if $DEBUG;
|
|
$operand = $weighted_operand if $weight_fields;
|
|
$indexes_set = 1 if $weight_fields;
|
|
|
|
# If there's a previous operand, we need to add an operator
|
|
if ($previous_operand) {
|
|
|
|
# user-specified operator
|
|
if ( $operators[$i-1] ) {
|
|
$query .= " $operators[$i-1] ";
|
|
$query .= " $index_plus " unless $indexes_set;
|
|
$query .= " $operand";
|
|
$query_cgi .="&op=$operators[$i-1]";
|
|
$query_cgi .="&idx=$index" if $index;
|
|
$query_cgi .="&q=$operands[$i]" if $operands[$i];
|
|
$query_search_desc .=" $operators[$i-1] $index_plus $operands[$i]";
|
|
}
|
|
|
|
# the default operator is and
|
|
else {
|
|
$query .= " and ";
|
|
$query .= "$index_plus " unless $indexes_set;
|
|
$query .= "$operand";
|
|
$query_cgi .="&op=and&idx=$index" if $index;
|
|
$query_cgi .="&q=$operands[$i]" if $operands[$i];
|
|
$query_search_desc .= " and $index_plus $operands[$i]";
|
|
}
|
|
}
|
|
|
|
else {
|
|
# field-weighted queries already have indexes set
|
|
$query .=" $index_plus " unless $indexes_set;
|
|
$query .= $operand;
|
|
$query_search_desc .= " $index_plus $operands[$i]";
|
|
$query_cgi.="&idx=$index" if $index;
|
|
$query_cgi.="&q=$operands[$i]" if $operands[$i];
|
|
|
|
$previous_operand = 1;
|
|
}
|
|
} #/if $operands
|
|
} # /for
|
|
}
|
|
warn "QUERY BEFORE LIMITS: >$query<" if $DEBUG;
|
|
|
|
# add limits
|
|
my $group_OR_limits;
|
|
foreach my $this_limit (@limits) {
|
|
if ( $this_limit =~ /available/ ) {
|
|
# FIXME: switch to zebra search for null values
|
|
$limit .= " (($query and datedue=0000-00-00) or ($query and datedue=0000-00-00 not lost=1) or ($query and datedue=0000-00-00 not lost=2))";
|
|
$limit_cgi .= "&limit=available";
|
|
$limit_desc .="";
|
|
}
|
|
|
|
# these are treated as OR
|
|
elsif ( $this_limit =~ /mc/ ) {
|
|
$group_OR_limits .= " or " if $group_OR_limits;
|
|
$group_OR_limits .= "$this_limit";
|
|
$limit_cgi .="&limit=$this_limit";
|
|
$limit_desc .= " or $this_limit";
|
|
}
|
|
|
|
# regular old limits
|
|
else {
|
|
$limit .= " and " if $limit || $query;
|
|
$limit .= "$this_limit";
|
|
$limit_cgi .="&limit=$this_limit";
|
|
$limit_desc .=" and $this_limit";
|
|
}
|
|
}
|
|
if ($group_OR_limits) {
|
|
$limit.=" and " if ($query || $limit );
|
|
$limit.="($group_OR_limits)";
|
|
}
|
|
# normalize the strings
|
|
for ($query, $query_search_desc, $limit, $limit_desc) {
|
|
$_ =~ s/ / /g; # remove extra spaces
|
|
$_ =~ s/^ //g; # remove any beginning spaces
|
|
$_ =~ s/ $//g; # remove any beginning spaces
|
|
$_ =~ s/:/=/g; # causes probs for server
|
|
$_ =~ s/==/=/g; # remove double == from query
|
|
|
|
}
|
|
|
|
$query_cgi =~ s/^&//;
|
|
|
|
# append the limit to the query
|
|
$query .= $limit;
|
|
|
|
warn "QUERY:".$query if $DEBUG;
|
|
warn "QUERY CGI:".$query_cgi if $DEBUG;
|
|
warn "QUERY DESC:".$query_search_desc if $DEBUG;
|
|
warn "LIMIT:".$limit if $DEBUG;
|
|
warn "LIMIT CGI:".$limit_cgi if $DEBUG;
|
|
warn "LIMIT DESC:".$limit_desc if $DEBUG;
|
|
|
|
return ( undef, $query,$simple_query,$query_cgi,$query_search_desc,$limit,$limit_cgi,$limit_desc );
|
|
}
|
|
|
|
# IMO this subroutine is pretty messy still -- it's responsible for
|
|
# building the HTML output for the template
|
|
sub searchResults {
|
|
my ( $searchdesc, $hits, $results_per_page, $offset, @marcresults ) = @_;
|
|
|
|
my $dbh = C4::Context->dbh;
|
|
my $toggle;
|
|
my $even = 1;
|
|
my @newresults;
|
|
my $span_terms_hashref;
|
|
for my $span_term ( split( / /, $searchdesc ) ) {
|
|
$span_term =~ s/(.*=|\)|\(|\+|\.)//g;
|
|
$span_terms_hashref->{$span_term}++;
|
|
}
|
|
|
|
#Build brancnames hash
|
|
#find branchname
|
|
#get branch information.....
|
|
my %branches;
|
|
my $bsth =
|
|
$dbh->prepare("SELECT branchcode,branchname FROM branches")
|
|
; # FIXME : use C4::Koha::GetBranches
|
|
$bsth->execute();
|
|
while ( my $bdata = $bsth->fetchrow_hashref ) {
|
|
$branches{ $bdata->{'branchcode'} } = $bdata->{'branchname'};
|
|
}
|
|
|
|
#Build itemtype hash
|
|
#find itemtype & itemtype image
|
|
my %itemtypes;
|
|
$bsth =
|
|
$dbh->prepare("SELECT itemtype,description,imageurl,summary,notforloan FROM itemtypes");
|
|
$bsth->execute();
|
|
while ( my $bdata = $bsth->fetchrow_hashref ) {
|
|
$itemtypes{ $bdata->{'itemtype'} }->{description} =
|
|
$bdata->{'description'};
|
|
$itemtypes{ $bdata->{'itemtype'} }->{imageurl} = $bdata->{'imageurl'};
|
|
$itemtypes{ $bdata->{'itemtype'} }->{summary} = $bdata->{'summary'};
|
|
$itemtypes{ $bdata->{'itemtype'} }->{notforloan} = $bdata->{'notforloan'};
|
|
}
|
|
|
|
#search item field code
|
|
my $sth =
|
|
$dbh->prepare(
|
|
"select tagfield from marc_subfield_structure where kohafield like 'items.itemnumber'"
|
|
);
|
|
$sth->execute;
|
|
my ($itemtag) = $sth->fetchrow;
|
|
|
|
## find column names of items related to MARC
|
|
my $sth2 = $dbh->prepare("SHOW COLUMNS from items");
|
|
$sth2->execute;
|
|
my %subfieldstosearch;
|
|
while ( ( my $column ) = $sth2->fetchrow ) {
|
|
my ( $tagfield, $tagsubfield ) =
|
|
&GetMarcFromKohaField( "items." . $column, "" );
|
|
$subfieldstosearch{$column} = $tagsubfield;
|
|
}
|
|
my $times;
|
|
|
|
if ( $hits && $offset + $results_per_page <= $hits ) {
|
|
$times = $offset + $results_per_page;
|
|
}
|
|
else {
|
|
$times = $hits;
|
|
}
|
|
|
|
for ( my $i = $offset ; $i <= $times - 1 ; $i++ ) {
|
|
my $marcrecord;
|
|
$marcrecord = MARC::File::USMARC::decode( $marcresults[$i] );
|
|
my $oldbiblio = TransformMarcToKoha( $dbh, $marcrecord, '' );
|
|
# add image url if there is one
|
|
if ( $itemtypes{ $oldbiblio->{itemtype} }->{imageurl} =~ /^http:/ ) {
|
|
$oldbiblio->{imageurl} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{imageurl};
|
|
$oldbiblio->{description} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{description};
|
|
}
|
|
else {
|
|
$oldbiblio->{imageurl} =
|
|
getitemtypeimagesrc() . "/"
|
|
. $itemtypes{ $oldbiblio->{itemtype} }->{imageurl}
|
|
if ( $itemtypes{ $oldbiblio->{itemtype} }->{imageurl} );
|
|
$oldbiblio->{description} =
|
|
$itemtypes{ $oldbiblio->{itemtype} }->{description};
|
|
}
|
|
#
|
|
# build summary if there is one (the summary is defined in itemtypes table
|
|
#
|
|
if ($itemtypes{ $oldbiblio->{itemtype} }->{summary}) {
|
|
my $summary = $itemtypes{ $oldbiblio->{itemtype} }->{summary};
|
|
my @fields = $marcrecord->fields();
|
|
foreach my $field (@fields) {
|
|
my $tag = $field->tag();
|
|
my $tagvalue = $field->as_string();
|
|
$summary =~ s/\[(.?.?.?.?)$tag\*(.*?)]/$1$tagvalue$2\[$1$tag$2]/g;
|
|
unless ($tag<10) {
|
|
my @subf = $field->subfields;
|
|
for my $i (0..$#subf) {
|
|
my $subfieldcode = $subf[$i][0];
|
|
my $subfieldvalue = $subf[$i][1];
|
|
my $tagsubf = $tag.$subfieldcode;
|
|
$summary =~ s/\[(.?.?.?.?)$tagsubf(.*?)]/$1$subfieldvalue$2\[$1$tagsubf$2]/g;
|
|
}
|
|
}
|
|
}
|
|
$summary =~ s/\[(.*?)]//g;
|
|
$summary =~ s/\n/<br>/g;
|
|
$oldbiblio->{summary} = $summary;
|
|
}
|
|
# add spans to search term in results for search term highlighting
|
|
foreach my $term ( keys %$span_terms_hashref ) {
|
|
my $old_term = $term;
|
|
if ( length($term) > 3 ) {
|
|
$term =~ s/(.*=|\)|\(|\+|\.|\?|\[|\])//g;
|
|
$term =~ s/\\//g;
|
|
$term =~ s/\*//g;
|
|
|
|
#FIXME: is there a better way to do this?
|
|
$oldbiblio->{'title'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'subtitle'} =~
|
|
s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
|
|
$oldbiblio->{'author'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'publishercode'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'place'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'pages'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'notes'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
$oldbiblio->{'size'} =~ s/$term/<span class=\"term\">$&<\/span>/gi;
|
|
}
|
|
}
|
|
|
|
if ( $i % 2 ) {
|
|
$toggle = "#ffffcc";
|
|
}
|
|
else {
|
|
$toggle = "white";
|
|
}
|
|
$oldbiblio->{'toggle'} = $toggle;
|
|
my @fields = $marcrecord->field($itemtag);
|
|
my @items_loop;
|
|
my $items;
|
|
my $ordered_count = 0;
|
|
my $onloan_count = 0;
|
|
my $wthdrawn_count = 0;
|
|
my $itemlost_count = 0;
|
|
my $norequests = 1;
|
|
|
|
#
|
|
# check the loan status of the item :
|
|
# it is not stored in the MARC record, for pref (zebra reindexing)
|
|
# reason. Thus, we have to get the status from a specific SQL query
|
|
#
|
|
my $sth_issue = $dbh->prepare("
|
|
SELECT date_due,returndate
|
|
FROM issues
|
|
WHERE itemnumber=? AND returndate IS NULL");
|
|
my $items_count=scalar(@fields);
|
|
foreach my $field (@fields) {
|
|
my $item;
|
|
foreach my $code ( keys %subfieldstosearch ) {
|
|
$item->{$code} = $field->subfield( $subfieldstosearch{$code} );
|
|
}
|
|
$sth_issue->execute($item->{itemnumber});
|
|
$item->{due_date} = format_date($sth_issue->fetchrow);
|
|
$item->{onloan} = 1 if $item->{due_date};
|
|
# at least one item can be reserved : suppose no
|
|
$norequests = 1;
|
|
if ( $item->{wthdrawn} ) {
|
|
$wthdrawn_count++;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{unavailable}=1;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{wthdrawn}=1;
|
|
}
|
|
elsif ( $item->{itemlost} ) {
|
|
$itemlost_count++;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{unavailable}=1;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{itemlost}=1;
|
|
}
|
|
unless ( $item->{notforloan}) {
|
|
# OK, this one can be issued, so at least one can be reserved
|
|
$norequests = 0;
|
|
}
|
|
if ( ( $item->{onloan} ) && ( $item->{onloan} != '0000-00-00' ) )
|
|
{
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{unavailable}=1;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{onloancount} = 1;
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{due_date} = $item->{due_date};
|
|
$onloan_count++;
|
|
}
|
|
if ( $item->{'homebranch'} ) {
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{count}++;
|
|
}
|
|
|
|
# Last resort
|
|
elsif ( $item->{'holdingbranch'} ) {
|
|
$items->{ $item->{'holdingbranch'} }->{count}++;
|
|
}
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{itemcallnumber} = $item->{itemcallnumber};
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{location} = $item->{location};
|
|
$items->{ $item->{'homebranch'}.'--'.$item->{'itemcallnumber'} }->{branchcode} = $item->{homebranch};
|
|
} # notforloan, item level and biblioitem level
|
|
|
|
# last check for norequest : if itemtype is notforloan, it can't be reserved either, whatever the items
|
|
$norequests = 1 if $itemtypes{$oldbiblio->{itemtype}}->{notforloan};
|
|
|
|
for my $key ( sort keys %$items ) {
|
|
my $this_item = {
|
|
branchname => $branches{$items->{$key}->{branchcode}},
|
|
branchcode => $items->{$key}->{branchcode},
|
|
count => $items->{$key}->{count},
|
|
itemcallnumber => $items->{$key}->{itemcallnumber},
|
|
location => $items->{$key}->{location},
|
|
onloancount => $items->{$key}->{onloancount},
|
|
due_date => $items->{$key}->{due_date},
|
|
wthdrawn => $items->{$key}->{wthdrawn},
|
|
lost => $items->{$key}->{itemlost},
|
|
};
|
|
push @items_loop, $this_item;
|
|
}
|
|
$oldbiblio->{norequests} = $norequests;
|
|
$oldbiblio->{items_count} = $items_count;
|
|
$oldbiblio->{items_loop} = \@items_loop;
|
|
$oldbiblio->{onloancount} = $onloan_count;
|
|
$oldbiblio->{wthdrawncount} = $wthdrawn_count;
|
|
$oldbiblio->{itemlostcount} = $itemlost_count;
|
|
$oldbiblio->{orderedcount} = $ordered_count;
|
|
$oldbiblio->{isbn} =~ s/-//g; # deleting - in isbn to enable amazon content
|
|
push( @newresults, $oldbiblio );
|
|
}
|
|
return @newresults;
|
|
}
|
|
|
|
|
|
|
|
#----------------------------------------------------------------------
|
|
#
|
|
# Non-Zebra GetRecords#
|
|
#----------------------------------------------------------------------
|
|
|
|
=head2 NZgetRecords
|
|
|
|
NZgetRecords has the same API as zera getRecords, even if some parameters are not managed
|
|
|
|
=cut
|
|
|
|
sub NZgetRecords {
|
|
my (
|
|
$koha_query, $simple_query, $sort_by_ref,
|
|
$servers_ref, $results_per_page, $offset,
|
|
$expanded_facet, $branches, $query_type,
|
|
$scan
|
|
) = @_;
|
|
my $result = NZanalyse($koha_query);
|
|
return (undef,NZorder($result,@$sort_by_ref[0],$results_per_page,$offset),undef);
|
|
}
|
|
|
|
=head2 NZanalyse
|
|
|
|
NZanalyse : get a CQL string as parameter, and returns a list of biblionumber;title,biblionumber;title,...
|
|
the list is builded from inverted index in nozebra SQL table
|
|
note that title is here only for convenience : the sorting will be very fast when requested on title
|
|
if the sorting is requested on something else, we will have to reread all results, and that may be longer.
|
|
|
|
=cut
|
|
|
|
sub NZanalyse {
|
|
my ($string,$server) = @_;
|
|
# $server contains biblioserver or authorities, depending on what we search on.
|
|
#warn "querying : $string on $server";
|
|
$server='biblioserver' unless $server;
|
|
# if we have a ", replace the content to discard temporarily any and/or/not inside
|
|
my $commacontent;
|
|
if ($string =~/"/) {
|
|
$string =~ s/"(.*?)"/__X__/;
|
|
$commacontent = $1;
|
|
# print "commacontent : $commacontent\n";
|
|
}
|
|
# split the query string in 3 parts : X AND Y means : $left="X", $operand="AND" and $right="Y"
|
|
# then, call again NZanalyse with $left and $right
|
|
# (recursive until we find a leaf (=> something without and/or/not)
|
|
$string =~ /(.*)( and | or | not | AND | OR | NOT )(.*)/;
|
|
my $left = $1;
|
|
my $right = $3;
|
|
my $operand = lc($2);
|
|
# it's not a leaf, we have a and/or/not
|
|
if ($operand) {
|
|
# reintroduce comma content if needed
|
|
$right =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
$left =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
# warn "node : $left / $operand / $right\n";
|
|
my $leftresult = NZanalyse($left,$server);
|
|
my $rightresult = NZanalyse($right,$server);
|
|
# OK, we have the results for right and left part of the query
|
|
# depending of operand, intersect, union or exclude both lists
|
|
# to get a result list
|
|
if ($operand eq ' and ') {
|
|
my @leftresult = split /;/, $leftresult;
|
|
# my @rightresult = split /;/,$leftresult;
|
|
my $finalresult;
|
|
# parse the left results, and if the biblionumber exist in the right result, save it in finalresult
|
|
# the result is stored twice, to have the same weight for AND than OR.
|
|
# example : TWO : 61,61,64,121 (two is twice in the biblio #61) / TOWER : 61,64,130
|
|
# result : 61,61,61,61,64,64 for two AND tower : 61 has more weight than 64
|
|
foreach (@leftresult) {
|
|
if ($rightresult =~ "$_;") {
|
|
$finalresult .= "$_;$_;";
|
|
}
|
|
}
|
|
return $finalresult;
|
|
} elsif ($operand eq ' or ') {
|
|
# just merge the 2 strings
|
|
return $leftresult.$rightresult;
|
|
} elsif ($operand eq ' not ') {
|
|
my @leftresult = split /;/, $leftresult;
|
|
# my @rightresult = split /;/,$leftresult;
|
|
my $finalresult;
|
|
foreach (@leftresult) {
|
|
unless ($rightresult =~ "$_;") {
|
|
$finalresult .= "$_;";
|
|
}
|
|
}
|
|
return $finalresult;
|
|
} else {
|
|
# this error is impossible, because of the regexp that isolate the operand, but just in case...
|
|
die "error : operand unknown : $operand for $string";
|
|
}
|
|
# it's a leaf, do the real SQL query and return the result
|
|
} else {
|
|
$string =~ s/__X__/"$commacontent"/ if $commacontent;
|
|
$string =~ s/-|\.|\?|,|;|!|'|\(|\)|\[|\]|{|}|"|&|\+|\*|\// /g;
|
|
# warn "leaf : $string\n";
|
|
# parse the string in in operator/operand/value again
|
|
$string =~ /(.*)(>=|<=)(.*)/;
|
|
my $left = $1;
|
|
my $operator = $2;
|
|
my $right = $3;
|
|
unless ($operator) {
|
|
$string =~ /(.*)(>|<|=)(.*)/;
|
|
$left = $1;
|
|
$operator = $2;
|
|
$right = $3;
|
|
}
|
|
my $results;
|
|
# automatic replace for short operators
|
|
$left='title' if $left =~ '^ti';
|
|
$left='author' if $left =~ '^au';
|
|
$left='publisher' if $left =~ '^pb';
|
|
$left='subject' if $left =~ '^su';
|
|
$left='koha-Auth-Number' if $left =~ '^an';
|
|
$left='keyword' if $left =~ '^kw';
|
|
if ($operator) {
|
|
#do a specific search
|
|
my $dbh = C4::Context->dbh;
|
|
$operator='LIKE' if $operator eq '=' and $right=~ /%/;
|
|
my $sth = $dbh->prepare("SELECT biblionumbers,value FROM nozebra WHERE server=? AND indexname=? AND value $operator ?");
|
|
warn "$left / $operator / $right\n";
|
|
# split each word, query the DB and build the biblionumbers result
|
|
foreach (split / /,$right) {
|
|
my ($biblionumbers,$value);
|
|
next unless $_;
|
|
warn "EXECUTE : $server, $left, $_";
|
|
$sth->execute($server, $left, $_);
|
|
while (my ($line,$value) = $sth->fetchrow) {
|
|
# if we are dealing with a numeric value, use only numeric results (in case of >=, <=, > or <)
|
|
# otherwise, fill the result
|
|
$biblionumbers .= $line unless ($right =~ /\d/ && $value =~ /\D/);
|
|
warn "result : $value ". ($right =~ /\d/) . "==".(!$value =~ /\d/) ;#= $line";
|
|
}
|
|
# do a AND with existing list if there is one, otherwise, use the biblionumbers list as 1st result list
|
|
if ($results) {
|
|
my @leftresult = split /;/, $biblionumbers;
|
|
my $temp;
|
|
foreach my $entry (@leftresult) { # $_ contains biblionumber,title-weight
|
|
# remove weight at the end
|
|
my $cleaned = $entry;
|
|
$cleaned =~ s/-\d*$//;
|
|
# if the entry already in the hash, take it & increase weight
|
|
# warn "===== $cleaned =====";
|
|
if ($results =~ "$cleaned") {
|
|
$temp .= "$entry;$entry;";
|
|
# warn "INCLUDING $entry";
|
|
}
|
|
}
|
|
$results = $temp;
|
|
} else {
|
|
$results = $biblionumbers;
|
|
}
|
|
}
|
|
} else {
|
|
#do a complete search (all indexes)
|
|
my $dbh = C4::Context->dbh;
|
|
my $sth = $dbh->prepare("SELECT biblionumbers FROM nozebra WHERE server=? AND value LIKE ?");
|
|
# split each word, query the DB and build the biblionumbers result
|
|
foreach (split / /,$string) {
|
|
next if C4::Context->stopwords->{uc($_)}; # skip if stopword
|
|
#warn "search on all indexes on $_";
|
|
my $biblionumbers;
|
|
next unless $_;
|
|
$sth->execute($server, $_);
|
|
while (my $line = $sth->fetchrow) {
|
|
$biblionumbers .= $line;
|
|
}
|
|
# do a AND with existing list if there is one, otherwise, use the biblionumbers list as 1st result list
|
|
if ($results) {
|
|
# warn "RES for $_ = $biblionumbers";
|
|
my @leftresult = split /;/, $biblionumbers;
|
|
my $temp;
|
|
foreach my $entry (@leftresult) { # $_ contains biblionumber,title-weight
|
|
# remove weight at the end
|
|
my $cleaned = $entry;
|
|
$cleaned =~ s/-\d*$//;
|
|
# if the entry already in the hash, take it & increase weight
|
|
# warn "===== $cleaned =====";
|
|
if ($results =~ "$cleaned") {
|
|
$temp .= "$entry;$entry;";
|
|
# warn "INCLUDING $entry";
|
|
}
|
|
}
|
|
$results = $temp;
|
|
} else {
|
|
# warn "NEW RES for $_ = $biblionumbers";
|
|
$results = $biblionumbers;
|
|
}
|
|
}
|
|
}
|
|
# warn "return : $results for LEAF : $string";
|
|
return $results;
|
|
}
|
|
}
|
|
|
|
=head2 NZorder
|
|
|
|
$finalresult = NZorder($biblionumbers, $ordering,$results_per_page,$offset);
|
|
|
|
TODO :: Description
|
|
|
|
=cut
|
|
|
|
|
|
sub NZorder {
|
|
my ($biblionumbers, $ordering,$results_per_page,$offset) = @_;
|
|
# order title asc by default
|
|
# $ordering = '1=36 <i' unless $ordering;
|
|
$results_per_page=20 unless $results_per_page;
|
|
$offset = 0 unless $offset;
|
|
my $dbh = C4::Context->dbh;
|
|
#
|
|
# order by POPULARITY
|
|
#
|
|
if ($ordering =~ /popularity/) {
|
|
my %result;
|
|
my %popularity;
|
|
# popularity is not in MARC record, it's builded from a specific query
|
|
my $sth = $dbh->prepare("select sum(issues) from items where biblionumber=?");
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
$result{$biblionumber}=GetMarcBiblio($biblionumber);
|
|
$sth->execute($biblionumber);
|
|
my $popularity= $sth->fetchrow ||0;
|
|
# hint : the key is popularity.title because we can have
|
|
# many results with the same popularity. In this cas, sub-ordering is done by title
|
|
# we also have biblionumber to avoid bug for 2 biblios with the same title & popularity
|
|
# (un-frequent, I agree, but we won't forget anything that way ;-)
|
|
$popularity{sprintf("%10d",$popularity).$title.$biblionumber} = $biblionumber;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq 'popularity_dsc') { # sort popularity DESC
|
|
foreach my $key (sort {$b cmp $a} (keys %popularity)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$popularity{$key}}->as_usmarc();
|
|
}
|
|
} else { # sort popularity ASC
|
|
foreach my $key (sort (keys %popularity)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$popularity{$key}}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY author
|
|
#
|
|
} elsif ($ordering =~/author/){
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my $author;
|
|
if (C4::Context->preference('marcflavour') eq 'UNIMARC') {
|
|
$author=$record->subfield('200','f');
|
|
$author=$record->subfield('700','a') unless $author;
|
|
} else {
|
|
$author=$record->subfield('100','a');
|
|
}
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$author.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq 'author_za') { # sort by author desc
|
|
foreach my $key (sort { $b cmp $a } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by author ASC
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY callnumber
|
|
#
|
|
} elsif ($ordering =~/callnumber/){
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my $callnumber;
|
|
my ($callnumber_tag,$callnumber_subfield)=GetMarcFromKohaField($dbh,'items.itemcallnumber');
|
|
($callnumber_tag,$callnumber_subfield)= GetMarcFromKohaField('biblioitems.callnumber') unless $callnumber_tag;
|
|
if (C4::Context->preference('marcflavour') eq 'UNIMARC') {
|
|
$callnumber=$record->subfield('200','f');
|
|
} else {
|
|
$callnumber=$record->subfield('100','a');
|
|
}
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$callnumber.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq 'call_number_dsc') { # sort by title desc
|
|
foreach my $key (sort { $b cmp $a } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $a cmp $b } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
} elsif ($ordering =~ /pubdate/){ #pub year
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
my $record=GetMarcBiblio($biblionumber);
|
|
my ($publicationyear_tag,$publicationyear_subfield)=GetMarcFromKohaField('biblioitems.publicationyear','');
|
|
my $publicationyear=$record->subfield($publicationyear_tag,$publicationyear_subfield);
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
$result{$publicationyear.$biblionumber}=$record;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq 'pubdate_dsc') { # sort by pubyear desc
|
|
foreach my $key (sort { $b cmp $a } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
} else { # sort by pub year ASC
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key}->as_usmarc();
|
|
}
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
#
|
|
# ORDER BY title
|
|
#
|
|
} elsif ($ordering =~ /title/) {
|
|
# the title is in the biblionumbers string, so we just need to build a hash, sort it and return
|
|
my %result;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
# hint : the result is sorted by title.biblionumber because we can have X biblios with the same title
|
|
# and we don't want to get only 1 result for each of them !!!
|
|
# hint & speed improvement : we can order without reading the record
|
|
# so order, and read records only for the requested page !
|
|
$result{$title.$biblionumber}=$biblionumber;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
if ($ordering eq 'title_az') { # sort by title desc
|
|
foreach my $key (sort (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
} else { # sort by title ASC
|
|
foreach my $key (sort { $b cmp $a } (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
}
|
|
# limit the $results_per_page to result size if it's more
|
|
$results_per_page = $numbers-1 if $numbers < $results_per_page;
|
|
# for the requested page, replace biblionumber by the complete record
|
|
# speed improvement : avoid reading too much things
|
|
for (my $counter=$offset;$counter<=$offset+$results_per_page;$counter++) {
|
|
$result_hash->{'RECORDS'}[$counter] = GetMarcBiblio($result_hash->{'RECORDS'}[$counter])->as_usmarc;
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
} else {
|
|
#
|
|
# order by ranking
|
|
#
|
|
# we need 2 hashes to order by ranking : the 1st one to count the ranking, the 2nd to order by ranking
|
|
my %result;
|
|
my %count_ranking;
|
|
foreach (split /;/,$biblionumbers) {
|
|
my ($biblionumber,$title) = split /,/,$_;
|
|
$title =~ /(.*)-(\d)/;
|
|
# get weight
|
|
my $ranking =$2;
|
|
# note that we + the ranking because ranking is calculated on weight of EACH term requested.
|
|
# if we ask for "two towers", and "two" has weight 2 in biblio N, and "towers" has weight 4 in biblio N
|
|
# biblio N has ranking = 6
|
|
$count_ranking{$biblionumber} += $ranking;
|
|
}
|
|
# build the result by "inverting" the count_ranking hash
|
|
# hing : as usual, we don't order by ranking only, to avoid having only 1 result for each rank. We build an hash on concat(ranking,biblionumber) instead
|
|
# warn "counting";
|
|
foreach (keys %count_ranking) {
|
|
$result{sprintf("%10d",$count_ranking{$_}).'-'.$_} = $_;
|
|
}
|
|
# sort the hash and return the same structure as GetRecords (Zebra querying)
|
|
my $result_hash;
|
|
my $numbers=0;
|
|
foreach my $key (sort {$b cmp $a} (keys %result)) {
|
|
$result_hash->{'RECORDS'}[$numbers++] = $result{$key};
|
|
}
|
|
# limit the $results_per_page to result size if it's more
|
|
$results_per_page = $numbers-1 if $numbers < $results_per_page;
|
|
# for the requested page, replace biblionumber by the complete record
|
|
# speed improvement : avoid reading too much things
|
|
for (my $counter=$offset;$counter<=$offset+$results_per_page;$counter++) {
|
|
$result_hash->{'RECORDS'}[$counter] = GetMarcBiblio($result_hash->{'RECORDS'}[$counter])->as_usmarc if $result_hash->{'RECORDS'}[$counter];
|
|
}
|
|
my $finalresult=();
|
|
$result_hash->{'hits'} = $numbers;
|
|
$finalresult->{'biblioserver'} = $result_hash;
|
|
return $finalresult;
|
|
}
|
|
}
|
|
=head2 ModBiblios
|
|
|
|
($countchanged,$listunchanged) = ModBiblios($listbiblios, $tagsubfield,$initvalue,$targetvalue,$test);
|
|
|
|
this function changes all the values $initvalue in subfield $tag$subfield in any record in $listbiblios
|
|
test parameter if set donot perform change to records in database.
|
|
|
|
=over 2
|
|
|
|
=item C<input arg:>
|
|
|
|
* $listbiblios is an array ref to marcrecords to be changed
|
|
* $tagsubfield is the reference of the subfield to change.
|
|
* $initvalue is the value to search the record for
|
|
* $targetvalue is the value to set the subfield to
|
|
* $test is to be set only not to perform changes in database.
|
|
|
|
=item C<Output arg:>
|
|
* $countchanged counts all the changes performed.
|
|
* $listunchanged contains the list of all the biblionumbers of records unchanged.
|
|
|
|
=item C<usage in the script:>
|
|
|
|
=back
|
|
|
|
my ($countchanged, $listunchanged) = EditBiblios($results->{RECORD}, $tagsubfield,$initvalue,$targetvalue);;
|
|
#If one wants to display unchanged records, you should get biblios foreach @$listunchanged
|
|
$template->param(countchanged => $countchanged, loopunchanged=>$listunchanged);
|
|
|
|
=cut
|
|
|
|
sub ModBiblios{
|
|
my ($listbiblios,$tagsubfield,$initvalue,$targetvalue,$test)=@_;
|
|
my $countmatched;
|
|
my @unmatched;
|
|
my ($tag,$subfield)=($1,$2) if ($tagsubfield=~/^(\d{1,3})([a-z0-9A-Z@])?$/);
|
|
if ((length($tag)<3)&& $subfield=~/0-9/){
|
|
$tag=$tag.$subfield;
|
|
undef $subfield;
|
|
}
|
|
my ($bntag,$bnsubf) = GetMarcFromKohaField('biblio.biblionumber');
|
|
my ($itemtag,$itemsubf) = GetMarcFromKohaField('items.itemnumber');
|
|
foreach my $usmarc (@$listbiblios){
|
|
my $record;
|
|
$record=eval{MARC::Record->new_from_usmarc($usmarc)};
|
|
my $biblionumber;
|
|
if ($@){
|
|
# usmarc is not a valid usmarc May be a biblionumber
|
|
if ($tag eq $itemtag){
|
|
my $bib=GetBiblioFromItemNumber($usmarc);
|
|
$record=GetMarcItem($bib->{'biblionumber'},$usmarc) ;
|
|
$biblionumber=$bib->{'biblionumber'};
|
|
} else {
|
|
$record=GetMarcBiblio($usmarc);
|
|
$biblionumber=$usmarc;
|
|
}
|
|
} else {
|
|
if ($bntag >= 010){
|
|
$biblionumber = $record->subfield($bntag,$bnsubf);
|
|
}else {
|
|
$biblionumber=$record->field($bntag)->data;
|
|
}
|
|
}
|
|
#GetBiblionumber is to be written.
|
|
#Could be replaced by TransformMarcToKoha (But Would be longer)
|
|
if ($record->field($tag)){
|
|
my $modify=0;
|
|
foreach my $field ($record->field($tag)){
|
|
if ($subfield){
|
|
if ($field->delete_subfield('code' =>$subfield,'match'=>qr($initvalue))){
|
|
$countmatched++;
|
|
$modify=1;
|
|
$field->update($subfield,$targetvalue) if ($targetvalue);
|
|
}
|
|
} else {
|
|
if ($tag >= 010){
|
|
if ($field->delete_field($field)){
|
|
$countmatched++;
|
|
$modify=1;
|
|
}
|
|
} else {
|
|
$field->data=$targetvalue if ($field->data=~qr($initvalue));
|
|
}
|
|
}
|
|
}
|
|
# warn $record->as_formatted;
|
|
if ($modify){
|
|
ModBiblio($record,$biblionumber,GetFrameworkCode($biblionumber)) unless ($test);
|
|
} else {
|
|
push @unmatched, $biblionumber;
|
|
}
|
|
} else {
|
|
push @unmatched, $biblionumber;
|
|
}
|
|
}
|
|
return ($countmatched,\@unmatched);
|
|
}
|
|
|
|
END { } # module clean-up code here (global destructor)
|
|
|
|
1;
|
|
__END__
|
|
|
|
=head1 AUTHOR
|
|
|
|
Koha Developement team <info@koha.org>
|
|
|
|
=cut
|