Difference between revisions of "ISFDB:Configure Nightly Processing"

From ISFDB
Jump to navigation Jump to search
(Page creation)
 
(→‎Monthly Processing: Clarified that the monthly process is currently commented out on the live server and that the algorithm doesn't scale well)
 
(2 intermediate revisions by the same user not shown)
Line 1: Line 1:
Configure crontab to run nightly_update.py in the "nightly" subdirectory under INSTALL_HTML. This module should be scheduled to run two times. The first run should be scheduled to run every night with the parameter "nightly", e.g.:
+
Note: the following three files reside in the "nightly" subdirectory under INSTALL_HTML.
  
00    01    *    *    *    /var/www/html/nightly/nightly_update.py nightly > /dev/null 2>&1
+
== Weekly Processing ==
  
When it runs, this task will regenerate database statistics and rerun the nightly cleanup reports.
+
Configure crontab to run weekly_job.py once a week, e.g.:
  
The second run of this cron job should be scheduled to run once a week with the parameter "weekly", e.g.:
+
00    01    *    *    7    /var/www/html/nightly/weekly_job.py > /dev/null 2>&1
  
00    02    *    *    2    /var/www/html/nightly/nightly_update.py weekly > /dev/null 2>&1
+
When it runs, this task will regenerate database statistics AND rerun the cleanup reports.
  
When it runs, this task will recreate the "Suspected Duplicate Authors" cleanup report. This process takes a long time and puts a lot of stress on the server, so it should not be run nightly on a live system.
+
== Nightly Processing ==
 +
 
 +
Configure crontab to run nightly_job.py once a day EXCEPT when weekly processing runs, e.g.:
 +
 
 +
00    01    *    *    1-6    /var/www/html/nightly/nightly_job.py > /dev/null 2>&1
 +
 
 +
When it runs, this task will rerun the cleanup reports.
 +
 
 +
== Monthly Processing ==
 +
 
 +
Configure crontab to run monthly_job.py once a month, e.g.:
 +
 
 +
00    02    7    *    *    /var/www/html/nightly/monthly_job.py > /dev/null 2>&1
 +
 
 +
When it runs, this task will recreate the "Suspected Duplicate Authors" cleanup report. Note that the algorithm doesn't scale well with the number of author records in the database. It takes a long time and puts a lot of stress on the server, so it should only be run infrequently. At the moment it's commented out on the live server due to performance issues.
  
 
[[Category:Installation instructions]]
 
[[Category:Installation instructions]]

Latest revision as of 13:49, 30 November 2023

Note: the following three files reside in the "nightly" subdirectory under INSTALL_HTML.

Weekly Processing

Configure crontab to run weekly_job.py once a week, e.g.:

00    01    *    *    7    /var/www/html/nightly/weekly_job.py > /dev/null 2>&1

When it runs, this task will regenerate database statistics AND rerun the cleanup reports.

Nightly Processing

Configure crontab to run nightly_job.py once a day EXCEPT when weekly processing runs, e.g.:

00    01    *    *    1-6    /var/www/html/nightly/nightly_job.py > /dev/null 2>&1

When it runs, this task will rerun the cleanup reports.

Monthly Processing

Configure crontab to run monthly_job.py once a month, e.g.:

00    02    7    *    *    /var/www/html/nightly/monthly_job.py > /dev/null 2>&1

When it runs, this task will recreate the "Suspected Duplicate Authors" cleanup report. Note that the algorithm doesn't scale well with the number of author records in the database. It takes a long time and puts a lot of stress on the server, so it should only be run infrequently. At the moment it's commented out on the live server due to performance issues.