Difference between revisions of "HeavyJobs"
(new task) |
|||
Line 1: | Line 1: | ||
− | |||
<noinclude><big>[[OurWork]] < [[DevelopmentTeam]] < [[DevelopmentTeamPriorities|Priorities]] < </noinclude>[[HeavyJobs]] {{JustTinyEditIcon|HeavyJobs}}<noinclude></big> | <noinclude><big>[[OurWork]] < [[DevelopmentTeam]] < [[DevelopmentTeamPriorities|Priorities]] < </noinclude>[[HeavyJobs]] {{JustTinyEditIcon|HeavyJobs}}<noinclude></big> | ||
__NOTOC__ | __NOTOC__ | ||
Line 15: | Line 14: | ||
** aggregators | ** aggregators | ||
* we have startup scripts that will resume proper job processing after a machine reboot | * we have startup scripts that will resume proper job processing after a machine reboot | ||
− | * we can monitor overall health of all heavy job processing with zabbix, including | + | * we can monitor overall health of all heavy job processing with zabbix, including system administrator alerts |
== Bugs and Todos == | == Bugs and Todos == |
Revision as of 00:40, 11 April 2008
What (summary)
Manage long-running jobs on available compute resources (servers) using db tables to keep track of work, and inter-process communication to keep track of workers.
Why this is important
We will use this infrastructure to manage our algorithmic data collection. This is a strategic direction for the company.
DoneDone
We will be satisfied with this infrastructure when:
- we can launch, balance, and diagnose all steps of our pilot whois refresh path.
- fetchers
- parsers
- aggregators
- we have startup scripts that will resume proper job processing after a machine reboot
- we can monitor overall health of all heavy job processing with zabbix, including system administrator alerts
Bugs and Todos
(non-prioritized at the moment)
- Workers should do partially completed chunks before starting new chunks.
- A worker should terminate when a manager has no more work to do.