dont use celery to index - seems it gets blocked on index file access
[wolnelektury.git] / apps / catalogue / utils.py
index 29a1857..a48ec03 100644 (file)
@@ -8,7 +8,10 @@ import random
 import time
 from base64 import urlsafe_b64encode
 
+from django.http import HttpResponse, HttpResponseRedirect, Http404, HttpResponsePermanentRedirect
 from django.core.files.uploadedfile import UploadedFile
+from django.core.files.base import File
+from django.core.files.storage import DefaultStorage
 from django.utils.hashcompat import sha_constructor
 from django.conf import settings
 from celery.task import task
@@ -17,8 +20,9 @@ from errno import EEXIST, ENOENT
 from fcntl import flock, LOCK_EX
 from zipfile import ZipFile
 
-from librarian import DocProvider
-
+from reporting.utils import read_chunks
+from celery.task import task
+import catalogue.models
 
 # Use the system (hardware-based) random number generator if it exists.
 if hasattr(random, 'SystemRandom'):
@@ -55,19 +59,6 @@ class ExistingFile(UploadedFile):
         pass
 
 
-class ORMDocProvider(DocProvider):
-    """Used for getting books' children."""
-
-    def __init__(self, book):
-        self.book = book
-
-    def by_slug(self, slug):
-        if slug == self.book.slug:
-            return self.book.xml_file
-        else:
-            return type(self.book).objects.get(slug=slug).xml_file
-
-
 class LockFile(object):
     """
     A file lock monitor class; createas an ${objname}.lock
@@ -111,8 +102,10 @@ def create_zip(paths, zip_slug):
         if not path.exists(path.join(zip_path, zip_filename)):
             zipf = ZipFile(path.join(zip_path, zip_filename), 'w')
             try:
-                for p in paths:
-                    zipf.write(p, path.basename(p))
+                for arcname, p in paths:
+                    if arcname is None:
+                        arcname = path.basename(p)
+                    zipf.write(p, arcname)
             finally:
                 zipf.close()
 
@@ -129,3 +122,64 @@ def remove_zip(zip_slug):
     except OSError as oe:
         if oe.errno != ENOENT:
             raise oe
+
+
+class AttachmentHttpResponse(HttpResponse):
+    """Response serving a file to be downloaded.
+    """
+    def __init__ (self, file_path, file_name, mimetype):
+        super(AttachmentHttpResponse, self).__init__(mimetype=mimetype)
+        self['Content-Disposition'] = 'attachment; filename=%s' % file_name
+        self.file_path = file_path
+        self.file_name = file_name
+
+        with open(DefaultStorage().path(self.file_path)) as f:
+            for chunk in read_chunks(f):
+                self.write(chunk)
+
+@task
+def async_build_pdf(book_id, customizations, file_name):
+    """
+    A celery task to generate pdf files.
+    Accepts the same args as Book.build_pdf, but with book id as first parameter
+    instead of Book instance
+    """
+    book = catalogue.models.Book.objects.get(id=book_id)
+    print "will gen %s" % DefaultStorage().path(file_name)
+    if not DefaultStorage().exists(file_name):
+        book.build_pdf(customizations=customizations, file_name=file_name)
+    print "done."
+
+
+class MultiQuerySet(object):
+    def __init__(self, *args, **kwargs):
+        self.querysets = args
+        self._count = None
+    
+    def count(self):
+        if not self._count:
+            self._count = sum(len(qs) for qs in self.querysets)
+        return self._count
+    
+    def __len__(self):
+        return self.count()
+        
+    def __getitem__(self, item):
+        try:
+            indices = (offset, stop, step) = item.indices(self.count())
+        except AttributeError:
+            # it's not a slice - make it one
+            return self[item : item + 1][0]
+        items = []
+        total_len = stop - offset
+        for qs in self.querysets:
+            if len(qs) < offset:
+                offset -= len(qs)
+            else:
+                items += list(qs[offset:stop])
+                if len(items) >= total_len:
+                    return items
+                else:
+                    offset = 0
+                    stop = total_len - len(items)
+                    continue
\ No newline at end of file