Tomas Cohen Arazi
b54c854d24
This patch adds logging of unhandled exceptions that could occur. This is happening on busy production sites right now. This is also useful for plugin jobs that might not be 100% following the guidelines and would benefit from this. But as the [DO NOT PUSH] patch highlights, this is something we really want to have on our current codebase, as a database connection drop might make us reach that `catch` block we are adding logging to on this patch. To test: 1. Apply the [DO NOT PUSH] patch 2. Run: $ ktd --shell k$ restart_all ; tail -f /var/log/koha/kohadev/worker*.log 3. Pick a valid barcode on the staff UI 4. Use the 'Batch delete items' tool in the cataloguing section 5. Start the job for deleting the item => FAIL: The item got deleted, but the job marked as failed and no logs about the reasons 6. Apply this patch and repeat 2-5 => SUCCESS: Same scenario but there's a log with the error message 7. Sign off :-D Signed-off-by: David Nind <david@davidnind.com> Signed-off-by: Martin Renvoize <martin.renvoize@ptfs-europe.com> Signed-off-by: Katrin Fischer <katrin.fischer@bsz-bw.de>
211 lines
6.4 KiB
Perl
Executable file
211 lines
6.4 KiB
Perl
Executable file
#!/usr/bin/perl
|
|
|
|
# This file is part of Koha.
|
|
#
|
|
# Koha is free software; you can redistribute it and/or modify it
|
|
# under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# Koha is distributed in the hope that it will be useful, but
|
|
# WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with Koha; if not, see <http://www.gnu.org/licenses>.
|
|
|
|
=head1 NAME
|
|
|
|
background_jobs_worker.pl - Worker script that will process background jobs
|
|
|
|
=head1 SYNOPSIS
|
|
|
|
./background_jobs_worker.pl [--queue QUEUE] [-m|--max-processes MAX_PROCESSES]
|
|
|
|
=head1 DESCRIPTION
|
|
|
|
This script will connect to the Stomp server (RabbitMQ) and subscribe to the queues passed in parameter (or the 'default' queue),
|
|
or if a Stomp server is not active it will poll the database every 10s for new jobs in the passed queue.
|
|
|
|
You can specify some queues only (using --queue, which is repeatable) if you want to run several workers that will handle their own jobs.
|
|
|
|
--m --max-processes specifies how many jobs to process simultaneously
|
|
|
|
Max processes will be set from the command line option, the environment variable MAX_PROCESSES, or the koha-conf file, in that order of precedence.
|
|
By default the script will only run one job at a time.
|
|
|
|
=head1 OPTIONS
|
|
|
|
=over
|
|
|
|
=item B<--queue>
|
|
|
|
Repeatable. Give the job queues this worker will process.
|
|
|
|
The different values available are:
|
|
|
|
default
|
|
long_tasks
|
|
elastic_index
|
|
|
|
=back
|
|
|
|
=cut
|
|
|
|
use Modern::Perl;
|
|
use JSON qw( decode_json );
|
|
use Try::Tiny;
|
|
use Pod::Usage;
|
|
use Getopt::Long;
|
|
use Parallel::ForkManager;
|
|
use Time::HiRes;
|
|
|
|
use C4::Context;
|
|
use Koha::Logger;
|
|
use Koha::BackgroundJobs;
|
|
use C4::Context;
|
|
|
|
$SIG{'PIPE'} = 'IGNORE'; # See BZ 35111; added to ignore PIPE error when connection lost on Ubuntu.
|
|
|
|
my ( $help, @queues );
|
|
|
|
my $max_processes = $ENV{MAX_PROCESSES};
|
|
$max_processes ||= C4::Context->config('background_jobs_worker')->{max_processes} if C4::Context->config('background_jobs_worker');
|
|
$max_processes ||= 1;
|
|
my $mq_timeout = $ENV{MQ_TIMEOUT} // 10;
|
|
|
|
my $not_found_retries = {};
|
|
my $max_retries = $ENV{MAX_RETRIES} || 10;
|
|
|
|
GetOptions(
|
|
'm|max-processes=i' => \$max_processes,
|
|
'h|help' => \$help,
|
|
'queue=s' => \@queues,
|
|
) || pod2usage(1);
|
|
|
|
|
|
pod2usage(0) if $help;
|
|
|
|
unless (@queues) {
|
|
push @queues, 'default';
|
|
}
|
|
|
|
my $conn;
|
|
try {
|
|
$conn = Koha::BackgroundJob->connect;
|
|
} catch {
|
|
warn sprintf "Cannot connect to the message broker, the jobs will be processed anyway (%s)", $_;
|
|
};
|
|
|
|
my $pm = Parallel::ForkManager->new($max_processes);
|
|
|
|
if ( $conn ) {
|
|
# FIXME cf note in Koha::BackgroundJob about $namespace
|
|
my $namespace = C4::Context->config('memcached_namespace');
|
|
for my $queue (@queues) {
|
|
$conn->subscribe(
|
|
{
|
|
destination => sprintf( "/queue/%s-%s", $namespace, $queue ),
|
|
ack => 'client',
|
|
'prefetch-count' => 1,
|
|
}
|
|
);
|
|
}
|
|
}
|
|
while (1) {
|
|
if ( $conn ) {
|
|
my $frame = $conn->receive_frame( { timeout => $mq_timeout } );
|
|
if ( !defined $frame ) {
|
|
# timeout or connection issue?
|
|
$pm->reap_finished_children;
|
|
next; # will reconnect automatically
|
|
}
|
|
|
|
my $args = try {
|
|
my $body = $frame->body;
|
|
decode_json($body); # TODO Should this be from_json? Check utf8 flag.
|
|
} catch {
|
|
Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Frame not processed - %s", $_);
|
|
return;
|
|
};
|
|
|
|
unless ( $args ) {
|
|
Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Frame does not have correct args, ignoring it");
|
|
$conn->nack( { frame => $frame, requeue => 'false' } );
|
|
next;
|
|
}
|
|
|
|
my $job = Koha::BackgroundJobs->find( $args->{job_id} );
|
|
|
|
if ( $job && $job->status ne 'new' ) {
|
|
Koha::Logger->get( { interface => 'worker' } )
|
|
->warn( sprintf "Job %s has wrong status %s", $args->{job_id}, $job->status );
|
|
|
|
# nack without requeue, we do not want to process this frame again
|
|
$conn->nack( { frame => $frame, requeue => 'false' } );
|
|
next;
|
|
}
|
|
|
|
unless ($job) {
|
|
$not_found_retries->{ $args->{job_id} } //= 0;
|
|
if ( ++$not_found_retries->{ $args->{job_id} } >= $max_retries ) {
|
|
Koha::Logger->get( { interface => 'worker' } )
|
|
->warn( sprintf "Job %s not found, no more retry", $args->{job_id} );
|
|
|
|
# nack without requeue, we do not want to process this frame again
|
|
$conn->nack( { frame => $frame, requeue => 'false' } );
|
|
next;
|
|
}
|
|
|
|
Koha::Logger->get( { interface => 'worker' } )
|
|
->debug( sprintf "Job %s not found, will retry later", $args->{job_id} );
|
|
|
|
# nack to force requeue
|
|
$conn->nack( { frame => $frame, requeue => 'true' } );
|
|
Time::HiRes::sleep(0.5);
|
|
next;
|
|
}
|
|
$conn->ack( { frame => $frame } );
|
|
|
|
$pm->start and next;
|
|
srand(); # ensure each child process begins with a new seed
|
|
process_job( $job, $args );
|
|
$pm->finish;
|
|
|
|
} else {
|
|
my $jobs = Koha::BackgroundJobs->search({ status => 'new', queue => \@queues });
|
|
while ( my $job = $jobs->next ) {
|
|
my $args = try {
|
|
$job->json->decode($job->data);
|
|
} catch {
|
|
Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Cannot decode data for job id=%s", $job->id);
|
|
$job->status('failed')->store;
|
|
return;
|
|
};
|
|
|
|
next unless $args;
|
|
|
|
$pm->start and next;
|
|
srand(); # ensure each child process begins with a new seed
|
|
process_job( $job, { job_id => $job->id, %$args } );
|
|
$pm->finish;
|
|
|
|
}
|
|
$pm->reap_finished_children;
|
|
sleep 10;
|
|
}
|
|
}
|
|
$conn->disconnect;
|
|
$pm->wait_all_children;
|
|
|
|
sub process_job {
|
|
my ( $job, $args ) = @_;
|
|
try {
|
|
$job->process($args);
|
|
} catch {
|
|
Koha::Logger->get( { interface => 'worker' } )
|
|
->warn( sprintf "Uncaught exception processing job id=%s: %s", $job->id, $_ );
|
|
$job->status('failed')->store;
|
|
};
|
|
}
|