#!/usr/bin/perl # This file is part of Koha. # # Koha is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # Koha is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Koha; if not, see . =head1 NAME background_jobs_worker.pl - Worker script that will process background jobs =head1 SYNOPSIS ./background_jobs_worker.pl [--queue QUEUE] [-m|--max-processes MAX_PROCESSES] =head1 DESCRIPTION This script will connect to the Stomp server (RabbitMQ) and subscribe to the queues passed in parameter (or the 'default' queue), or if a Stomp server is not active it will poll the database every 10s for new jobs in the passed queue. You can specify some queues only (using --queue, which is repeatable) if you want to run several workers that will handle their own jobs. --m --max-processes specifies how many jobs to process simultaneously Max processes will be set from the command line option, the environment variable MAX_PROCESSES, or the koha-conf file, in that order of precedence. By default the script will only run one job at a time. =head1 OPTIONS =over =item B<--queue> Repeatable. Give the job queues this worker will process. The different values available are: default long_tasks elastic_index =back =cut use Modern::Perl; use JSON qw( decode_json ); use Try::Tiny; use Pod::Usage; use Getopt::Long; use Parallel::ForkManager; use Time::HiRes; use C4::Context; use Koha::Logger; use Koha::BackgroundJobs; use C4::Context; $SIG{'PIPE'} = 'IGNORE'; # See BZ 35111; added to ignore PIPE error when connection lost on Ubuntu. my ( $help, @queues ); my $max_processes = $ENV{MAX_PROCESSES}; $max_processes ||= C4::Context->config('background_jobs_worker')->{max_processes} if C4::Context->config('background_jobs_worker'); $max_processes ||= 1; my $mq_timeout = $ENV{MQ_TIMEOUT} // 10; my $not_found_retries = {}; my $max_retries = $ENV{MAX_RETRIES} || 10; GetOptions( 'm|max-processes=i' => \$max_processes, 'h|help' => \$help, 'queue=s' => \@queues, ) || pod2usage(1); pod2usage(0) if $help; unless (@queues) { push @queues, 'default'; } my $conn; try { $conn = Koha::BackgroundJob->connect; } catch { warn sprintf "Cannot connect to the message broker, the jobs will be processed anyway (%s)", $_; }; my $pm = Parallel::ForkManager->new($max_processes); if ( $conn ) { # FIXME cf note in Koha::BackgroundJob about $namespace my $namespace = C4::Context->config('memcached_namespace'); for my $queue (@queues) { $conn->subscribe( { destination => sprintf( "/queue/%s-%s", $namespace, $queue ), ack => 'client', 'prefetch-count' => 1, } ); } } while (1) { if ( $conn ) { my $frame = $conn->receive_frame( { timeout => $mq_timeout } ); if ( !defined $frame ) { # timeout or connection issue? $pm->reap_finished_children; next; # will reconnect automatically } my $args = try { my $body = $frame->body; decode_json($body); # TODO Should this be from_json? Check utf8 flag. } catch { Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Frame not processed - %s", $_); return; }; unless ( $args ) { Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Frame does not have correct args, ignoring it"); $conn->nack( { frame => $frame, requeue => 'false' } ); next; } my $job = Koha::BackgroundJobs->find( $args->{job_id} ); if ( $job && $job->status ne 'new' ) { Koha::Logger->get( { interface => 'worker' } ) ->warn( sprintf "Job %s has wrong status %s", $args->{job_id}, $job->status ); # nack without requeue, we do not want to process this frame again $conn->nack( { frame => $frame, requeue => 'false' } ); next; } unless ($job) { $not_found_retries->{ $args->{job_id} } //= 0; if ( ++$not_found_retries->{ $args->{job_id} } >= $max_retries ) { Koha::Logger->get( { interface => 'worker' } ) ->warn( sprintf "Job %s not found, no more retry", $args->{job_id} ); # nack without requeue, we do not want to process this frame again $conn->nack( { frame => $frame, requeue => 'false' } ); next; } Koha::Logger->get( { interface => 'worker' } ) ->debug( sprintf "Job %s not found, will retry later", $args->{job_id} ); # nack to force requeue $conn->nack( { frame => $frame, requeue => 'true' } ); Time::HiRes::sleep(0.5); next; } $conn->ack( { frame => $frame } ); $pm->start and next; srand(); # ensure each child process begins with a new seed process_job( $job, $args ); $pm->finish; } else { my $jobs = Koha::BackgroundJobs->search({ status => 'new', queue => \@queues }); while ( my $job = $jobs->next ) { my $args = try { $job->json->decode($job->data); } catch { Koha::Logger->get({ interface => 'worker' })->warn(sprintf "Cannot decode data for job id=%s", $job->id); $job->status('failed')->store; return; }; next unless $args; $pm->start and next; srand(); # ensure each child process begins with a new seed process_job( $job, { job_id => $job->id, %$args } ); $pm->finish; } $pm->reap_finished_children; sleep 10; } } $conn->disconnect; $pm->wait_all_children; sub process_job { my ( $job, $args ) = @_; try { $job->process($args); } catch { Koha::Logger->get( { interface => 'worker' } ) ->warn( sprintf "Uncaught exception processing job id=%s: %s", $job->id, $_ ); $job->status('failed')->store; }; }