euscanwww: Fixed tasks to work with processing.scan and processing.misc

Fixed tasks to work with processing.* changes.
Moved periodic tasks config to settings.py using CELERYBEAT_SCHEDULE

Signed-off-by: volpino <fox91@anche.no>
This commit is contained in:
volpino 2012-07-07 15:13:59 +02:00
parent 80e8157627
commit a6b215cec9
10 changed files with 65 additions and 58 deletions

View File

@ -23,16 +23,12 @@ Running Celery
You'll need: You'll need:
* celeryd (celery daemon for running tasks):: * celeryd (celery daemon for running tasks)::
python manage.py celeryd -E -l INFO python manage.py celeryd -B -E -l INFO
* celerycam (for monitoring celery and see the results in the django admin page):: * celerycam (for monitoring celery and see the results in the django admin page)::
python manage.py celerycam python manage.py celerycam
* celerybeat (for running periodic tasks)::
python manage.py celerybeat -l INFO
TODO TODO
==== ====

View File

@ -1,4 +1,3 @@
class FakeLogger(object): class FakeLogger(object):
def __getattr__(self, key): def __getattr__(self, key):
return lambda *x, **y: None return lambda *x, **y: None
@ -29,4 +28,3 @@ def set_verbosity_level(logger, verbosity):
logger.setLevel(levels[verbosity]) logger.setLevel(levels[verbosity])
return logger return logger

View File

@ -3,6 +3,7 @@ from djeuscan import charts
from djeuscan.processing import FakeLogger from djeuscan.processing import FakeLogger
def regen_rrds(logger=None): def regen_rrds(logger=None):
""" """
Regenerates the rrd database Regenerates the rrd database

View File

@ -2,6 +2,7 @@ import os
from django.conf import settings from django.conf import settings
def _launch_command(cmd, logger=None): def _launch_command(cmd, logger=None):
""" """
Helper for launching shell commands inside tasks Helper for launching shell commands inside tasks
@ -43,6 +44,7 @@ def _launch_command(cmd, logger=None):
fp.wait() fp.wait()
def emerge_sync(logger): def emerge_sync(logger):
""" """
Launches an emerge --sync Launches an emerge --sync
@ -81,10 +83,12 @@ def layman_sync(logger, cache=True):
for overlay in installed_overlays: for overlay in installed_overlays:
logger.info('Generating cache for overlay %s...' % overlay) logger.info('Generating cache for overlay %s...' % overlay)
overlay_path = os.path.join(l.config['storage'], overlay) overlay_path = os.path.join(l.config['storage'], overlay)
if not os.path.exists(os.path.join(overlay_path, 'profiles/repo_name')): repo_path = os.path.join(overlay_path, 'profiles/repo_name')
if not os.path.exists(repo_path):
continue continue
_launch_command(cmd + ['--repo', overlay], logger) _launch_command(cmd + ['--repo', overlay], logger)
def eix_update(logger): def eix_update(logger):
""" """
Launches eix-update Launches eix-update

View File

@ -1,7 +1,4 @@
import os.path
from gentoolkit.query import Query from gentoolkit.query import Query
from gentoolkit.errors import GentoolkitFatalError
from django.db.transaction import commit_on_success from django.db.transaction import commit_on_success
from django.core.management.color import color_style from django.core.management.color import color_style
@ -10,6 +7,7 @@ from django.core.exceptions import ValidationError
from djeuscan.models import Package, Herd, Maintainer from djeuscan.models import Package, Herd, Maintainer
from djeuscan.processing import FakeLogger from djeuscan.processing import FakeLogger
class ScanMetadata(object): class ScanMetadata(object):
def __init__(self, logger=None): def __init__(self, logger=None):
self.style = color_style() self.style = color_style()
@ -61,8 +59,10 @@ class ScanMetadata(object):
old_herds = set(existing_herds).difference(herds.keys()) old_herds = set(existing_herds).difference(herds.keys())
existing_maintainers = [m.email for m in obj.maintainers.all()] existing_maintainers = [m.email for m in obj.maintainers.all()]
new_maintainers = set(maintainers.keys()).difference(existing_maintainers) new_maintainers = set(maintainers.keys()).\
old_maintainers = set(existing_maintainers).difference(maintainers.keys()) difference(existing_maintainers)
old_maintainers = set(existing_maintainers).\
difference(maintainers.keys())
for herd in obj.herds.all(): for herd in obj.herds.all():
if herd.herd in old_herds: if herd.herd in old_herds:
@ -125,6 +125,7 @@ class ScanMetadata(object):
) )
return maintainer return maintainer
@commit_on_success @commit_on_success
def scan_metadata(packages=None, category=None, logger=None): def scan_metadata(packages=None, category=None, logger=None):
scan_handler = ScanMetadata(logger=logger) scan_handler = ScanMetadata(logger=logger)
@ -139,4 +140,3 @@ def scan_metadata(packages=None, category=None, logger=None):
scan_handler.scan('%s/%s' % (pkg.category, pkg.name), pkg) scan_handler.scan('%s/%s' % (pkg.category, pkg.name), pkg)
else: else:
scan_handler.scan(pkg) scan_handler.scan(pkg)

View File

@ -83,7 +83,9 @@ class ScanPortage(object):
package['homepage'] = ' '.join(p.homepages) package['homepage'] = ' '.join(p.homepages)
package['description'] = p.description package['description'] = p.description
package['versions'] = [] package['versions'] = []
package['versions'].append((p._cpv, p.slot, p.repository or 'gentoo')) package['versions'].append(
(p._cpv, p.slot, p.repository or 'gentoo')
)
if package_name: if package_name:
yield package yield package
@ -123,8 +125,11 @@ class ScanPortage(object):
package[elem.tag] = elem.text or "" package[elem.tag] = elem.text or ""
elif elem.tag == "version": elif elem.tag == "version":
# append version data to versions # append version data to versions
cpv = "%s/%s-%s" % \ cpv = "%s/%s-%s" % (
(package["category"], package["package"], elem.attrib["id"]) package["category"],
package["package"],
elem.attrib["id"]
)
slot = elem.attrib.get("slot", "0") slot = elem.attrib.get("slot", "0")
overlay = elem.attrib.get("repository", "gentoo") overlay = elem.attrib.get("repository", "gentoo")
package["versions"].append((cpv, slot, overlay)) package["versions"].append((cpv, slot, overlay))
@ -177,7 +182,9 @@ class ScanPortage(object):
for data in self.scan_eix_xml(query, category): for data in self.scan_eix_xml(query, category):
#for data in self.scan_gentoopm(query, category): #for data in self.scan_gentoopm(query, category):
cat, pkg = data['category'], data['package'] cat, pkg = data['category'], data['package']
package = self.store_package(cat, pkg, data['homepage'], data['description']) package = self.store_package(
cat, pkg, data['homepage'], data['description']
)
packages_alive.add("%s/%s" % (cat, pkg)) packages_alive.add("%s/%s" % (cat, pkg))
for cpv, slot, overlay in data['versions']: for cpv, slot, overlay in data['versions']:
self.store_version(package, cpv, slot, overlay) self.store_version(package, cpv, slot, overlay)
@ -300,8 +307,9 @@ class ScanPortage(object):
@commit_on_success @commit_on_success
def scan_portage(packages=None, category=None, no_log=False, purge_packages=False, def scan_portage(packages=None, category=None, no_log=False,
purge_versions=False, prefetch=False, logger=None): purge_packages=False, purge_versions=False, prefetch=False,
logger=None):
logger = logger or FakeLogger() logger = logger or FakeLogger()
@ -335,4 +343,3 @@ def scan_portage(packages=None, category=None, no_log=False, purge_packages=Fals
scan_handler.scan(pkg) scan_handler.scan(pkg)
logger.info('Done.') logger.info('Done.')

View File

@ -69,7 +69,9 @@ class ScanUpstream(object):
# Set all versions dead, then set found versions alive and # Set all versions dead, then set found versions alive and
# delete old versions # delete old versions
if self.purge_versions: if self.purge_versions:
Version.objects.filter(package=obj, packaged=False).update(alive=False) Version.objects.filter(
package=obj, packaged=False
).update(alive=False)
return obj return obj
@ -109,7 +111,6 @@ class ScanUpstream(object):
package.n_versions += 1 package.n_versions += 1
package.save() package.save()
def purge_old_versions(self): def purge_old_versions(self):
if not self.purge_versions: if not self.purge_versions:
return return
@ -132,6 +133,7 @@ class ScanUpstream(object):
versions.delete() versions.delete()
@commit_on_success @commit_on_success
def scan_upstream(packages=None, purge_versions=False, def scan_upstream(packages=None, purge_versions=False,
logger=None): logger=None):
@ -144,8 +146,6 @@ def scan_upstream(packages=None, purge_versions=False,
if not packages: if not packages:
packages = Package.objects.all() packages = Package.objects.all()
result = True
for pkg in packages: for pkg in packages:
if isinstance(pkg, Package): if isinstance(pkg, Package):
scan_handler.scan('%s/%s' % (pkg.category, pkg.name)) scan_handler.scan('%s/%s' % (pkg.category, pkg.name))

View File

@ -4,19 +4,15 @@ Celery tasks for djeuscan
from itertools import islice from itertools import islice
from celery.task import task, periodic_task from celery.task import task
from celery.task.schedules import crontab
from celery.task.sets import TaskSet from celery.task.sets import TaskSet
from django.conf import settings from django.conf import settings
from djeuscan.models import Package, RefreshPackageQuery from djeuscan.models import Package, RefreshPackageQuery
from djeuscan.processing.regen_rrds import regen_rrds from djeuscan.processing.misc import regen_rrds, update_counters, \
from djeuscan.processing.update_counters import update_counters update_portage_trees
from djeuscan.processing.scan_metadata import scan_metadata from djeuscan.processing.scan import scan_metadata, scan_portage, scan_upstream
from djeuscan.processing.scan_portage import scan_portage
from djeuscan.processing.scan_upstream import scan_upstream
from djeuscan.processing.update_portage_trees import update_portage_trees
class TaskFailedException(Exception): class TaskFailedException(Exception):
@ -87,13 +83,10 @@ def _scan_metadata_task(packages):
logger.info("Starting metadata scanning subtask for %d packages...", logger.info("Starting metadata scanning subtask for %d packages...",
len(packages)) len(packages))
result = scan_metadata( scan_metadata(
packages=packages, packages=packages,
logger=logger, logger=logger,
) )
if not result:
raise TaskFailedException
return result
@task @task
@ -126,7 +119,7 @@ def _scan_portage_task(packages, no_log=False, purge_packages=False,
else: else:
logger.info("Starting portage scanning for all packages...") logger.info("Starting portage scanning for all packages...")
result = scan_portage( scan_portage(
packages=packages, packages=packages,
no_log=no_log, no_log=no_log,
purge_packages=purge_packages, purge_packages=purge_packages,
@ -134,9 +127,6 @@ def _scan_portage_task(packages, no_log=False, purge_packages=False,
prefetch=prefetch, prefetch=prefetch,
logger=logger, logger=logger,
) )
if not result:
raise TaskFailedException
return result
@task @task
@ -182,8 +172,9 @@ def _scan_upstream_task(packages, purge_versions=False):
purge_versions=purge_versions, purge_versions=purge_versions,
logger=logger, logger=logger,
) )
if not result: # TODO: implement some kind of error raising in case of failure
raise TaskFailedException #if not result:
# raise TaskFailedException
return result return result
@ -258,7 +249,9 @@ def scan_package_task(package):
_scan_upstream_task([package]) _scan_upstream_task([package])
@periodic_task(run_every=crontab(minute="*/1")) # Periodic tasks
@task
def consume_refresh_package_request(): def consume_refresh_package_request():
""" """
Satisfies user requests for package refreshing, runs every minute Satisfies user requests for package refreshing, runs every minute
@ -273,14 +266,6 @@ def consume_refresh_package_request():
return result return result
@periodic_task(run_every=crontab(hour=03, minute=00, day_of_week=1))
def update_periodic_task():
"""
Runs a whole update once a week
"""
update_task()
admin_tasks = [ admin_tasks = [
regen_rrds_task, regen_rrds_task,
update_counters_task, update_counters_task,

View File

@ -1,6 +1,7 @@
# Django settings for euscanwww project. # Django settings for euscanwww project.
import os.path import os.path
from datetime import timedelta
DEBUG = True DEBUG = True
TEMPLATE_DEBUG = DEBUG TEMPLATE_DEBUG = DEBUG
@ -238,6 +239,22 @@ CELERYD_CONCURRENCY = 4
TASKS_CONCURRENTLY = 4 TASKS_CONCURRENTLY = 4
TASKS_SUBTASK_PACKAGES = 32 TASKS_SUBTASK_PACKAGES = 32
CELERYBEAT_SCHEDULE = {
"refresh_package_query": {
"task": "djeuscan.tasks.consume_refresh_package_request",
"schedule": timedelta(minutes=1),
},
"weekly_update": {
"task": "djeuscan.tasks.update_task",
"schedule": timedelta(days=7),
},
"daily_update": {
"task": "djeuscan.tasks.update_task",
"schedule": timedelta(days=1),
"kwargs": {"scan_upstream": False}
},
}
# LDAP authentication # LDAP authentication
# TODO: Test data - change me! # TODO: Test data - change me!
AUTH_LDAP_SERVER_URI = "ldap://localhost" AUTH_LDAP_SERVER_URI = "ldap://localhost"

View File

@ -1,4 +1,3 @@
python manage.py celeryd -E -l INFO & python manage.py celeryd -B -E -l INFO &
python manage.py celerybeat -l INFO &
python manage.py celerycam & python manage.py celerycam &
python manage.py runserver python manage.py runserver