diff piecrust/environment.py @ 411:e7b865f8f335

bake: Enable multiprocess baking. Baking is now done by running a worker per CPU, and sending jobs to them. This changes several things across the codebase: * Ability to not cache things related to pages other than the 'main' page (i.e. the page at the bottom of the execution stack). * Decouple the baking process from the bake records, so only the main process keeps track (and modifies) the bake record. * Remove the need for 'batch page getters' and loading a page directly from the page factories. There are various smaller changes too included here, including support for scope performance timers that are saved with the bake record and can be printed out to the console. Yes I got carried away. For testing, the in-memory 'mock' file-system doesn't work anymore, since we're spawning processes, so this is replaced by a 'tmpfs' file-system which is saved in temporary files on disk and deleted after tests have run.
author Ludovic Chabant <ludovic@chabant.com>
date Fri, 12 Jun 2015 17:09:19 -0700
parents c2ca72fb7f0b
children a1567766c83c
line wrap: on
line diff
--- a/piecrust/environment.py	Sat May 30 15:41:52 2015 -0700
+++ b/piecrust/environment.py	Fri Jun 12 17:09:19 2015 -0700
@@ -1,9 +1,7 @@
-import re
 import time
 import json
 import logging
 import hashlib
-import threading
 import contextlib
 import collections
 import repoze.lru
@@ -12,9 +10,6 @@
 logger = logging.getLogger(__name__)
 
 
-re_fs_cache_key = re.compile(r'[^\d\w\-\._]+')
-
-
 def _make_fs_cache_key(key):
     return hashlib.md5(key.encode('utf8')).hexdigest()
 
@@ -26,57 +21,58 @@
     def __init__(self, size=2048):
         self.cache = repoze.lru.LRUCache(size)
         self.fs_cache = None
+        self._last_access_hit = None
         self._invalidated_fs_items = set()
-        self._lock = threading.RLock()
 
-    @contextlib.contextmanager
-    def startBatchGet(self):
-        logger.debug("Starting batch cache operation.")
-        with self._lock:
-            yield
-        logger.debug("Ending batch cache operation.")
+    @property
+    def last_access_hit(self):
+        return self._last_access_hit
 
     def invalidate(self, key):
-        with self._lock:
-            logger.debug("Invalidating cache item '%s'." % key)
-            self.cache.invalidate(key)
-            if self.fs_cache:
-                logger.debug("Invalidating FS cache item '%s'." % key)
-                fs_key = _make_fs_cache_key(key)
-                self._invalidated_fs_items.add(fs_key)
+        logger.debug("Invalidating cache item '%s'." % key)
+        self.cache.invalidate(key)
+        if self.fs_cache:
+            logger.debug("Invalidating FS cache item '%s'." % key)
+            fs_key = _make_fs_cache_key(key)
+            self._invalidated_fs_items.add(fs_key)
 
-    def get(self, key, item_maker, fs_cache_time=None):
+    def put(self, key, item, save_to_fs=True):
+        self.cache.put(key, item)
+        if self.fs_cache and save_to_fs:
+            fs_key = _make_fs_cache_key(key)
+            item_raw = json.dumps(item)
+            self.fs_cache.write(fs_key, item_raw)
+
+    def get(self, key, item_maker, fs_cache_time=None, save_to_fs=True):
+        self._last_access_hit = True
         item = self.cache.get(key)
         if item is None:
-            logger.debug("Acquiring lock for: %s" % key)
-            with self._lock:
-                item = self.cache.get(key)
-                if item is None:
-                    if (self.fs_cache is not None and
-                            fs_cache_time is not None):
-                        # Try first from the file-system cache.
-                        fs_key = _make_fs_cache_key(key)
-                        if (fs_key not in self._invalidated_fs_items and
-                                self.fs_cache.isValid(fs_key, fs_cache_time)):
-                            logger.debug("'%s' found in file-system cache." %
-                                         key)
-                            item_raw = self.fs_cache.read(fs_key)
-                            item = json.loads(
-                                    item_raw,
-                                    object_pairs_hook=collections.OrderedDict)
-                            self.cache.put(key, item)
-                            return item
+            if (self.fs_cache is not None and
+                    fs_cache_time is not None):
+                # Try first from the file-system cache.
+                fs_key = _make_fs_cache_key(key)
+                if (fs_key not in self._invalidated_fs_items and
+                        self.fs_cache.isValid(fs_key, fs_cache_time)):
+                    logger.debug("'%s' found in file-system cache." %
+                                 key)
+                    item_raw = self.fs_cache.read(fs_key)
+                    item = json.loads(
+                            item_raw,
+                            object_pairs_hook=collections.OrderedDict)
+                    self.cache.put(key, item)
+                    return item
 
-                    # Look into the mem-cache.
-                    logger.debug("'%s' not found in cache, must build." % key)
-                    item = item_maker()
-                    self.cache.put(key, item)
+            # Look into the mem-cache.
+            logger.debug("'%s' not found in cache, must build." % key)
+            item = item_maker()
+            self.cache.put(key, item)
+            self._last_access_hit = False
 
-                    # Save to the file-system if needed.
-                    if (self.fs_cache is not None and
-                            fs_cache_time is not None):
-                        item_raw = json.dumps(item)
-                        self.fs_cache.write(fs_key, item_raw)
+            # Save to the file-system if needed.
+            if self.fs_cache is not None and save_to_fs:
+                item_raw = json.dumps(item)
+                self.fs_cache.write(fs_key, item_raw)
+
         return item
 
 
@@ -88,7 +84,7 @@
         self.start_time = time.clock()
 
 
-class ExecutionInfoStack(threading.local):
+class ExecutionInfoStack(object):
     def __init__(self):
         self._page_stack = []
 
@@ -131,14 +127,28 @@
         self.rendered_segments_repository = MemCache()
         self.fs_caches = {
                 'renders': self.rendered_segments_repository}
+        self.fs_cache_only_for_main_page = False
+        self._timers = {}
 
     def initialize(self, app):
-        self.start_time = time.clock()
+        self.start_time = time.perf_counter()
         self.exec_info_stack.clear()
         self.was_cache_cleaned = False
         self.base_asset_url_format = '%uri%'
         self._onSubCacheDirChanged(app)
 
+    def registerTimer(self, category):
+        self._timers[category] = 0
+
+    @contextlib.contextmanager
+    def timerScope(self, category):
+        start = time.perf_counter()
+        yield
+        self._timers[category] += time.perf_counter() - start
+
+    def stepTimer(self, category, value):
+        self._timers[category] += value
+
     def _onSubCacheDirChanged(self, app):
         for name, repo in self.fs_caches.items():
             cache = app.cache.getCache(name)