# HG changeset patch # User Ludovic Chabant # Date 1422752882 28800 # Node ID a47580a0955b27fc34e52dfbde18166a69d5bcf2 # Parent 09e350db7f8f1805651ade65a3def2a77a520635 bake: Better error handling for the processing pipeline. Pipeline jobs now keep track of whether they've seen any errors. This is aggregated into an overall "success" flag for the processing record. Also, jobs keep going as long as there's no critical (i.e. internal) failure happening. Errors raised by processors are also better tracked: the actual processor that failed, along with the input file, are tracks in the processing record. The `bake` command returns a failure exit code if processing saw any error. diff -r 09e350db7f8f -r a47580a0955b piecrust/commands/builtin/baking.py --- a/piecrust/commands/builtin/baking.py Sat Jan 31 16:28:16 2015 -0800 +++ b/piecrust/commands/builtin/baking.py Sat Jan 31 17:08:02 2015 -0800 @@ -40,19 +40,20 @@ out_dir = (ctx.args.output or os.path.join(ctx.app.root_dir, '_counter')) + success = True start_time = time.clock() try: # Bake the site sources. - self._bakeSources(ctx, out_dir) + success = success & self._bakeSources(ctx, out_dir) # Bake the assets. if not ctx.args.no_assets: - self._bakeAssets(ctx, out_dir) + success = success & self._bakeAssets(ctx, out_dir) # All done. logger.info('-------------------------') logger.info(format_timed(start_time, 'done baking')) - return 0 + return 0 if success else 1 except Exception as ex: if ctx.app.debug: logger.exception(ex) @@ -65,12 +66,14 @@ ctx.app, out_dir, force=ctx.args.force) baker.bake() + return True def _bakeAssets(self, ctx, out_dir): proc = ProcessorPipeline( ctx.app, out_dir, force=ctx.args.force) - proc.run() + record = proc.run() + return record.success class ShowRecordCommand(ChefCommand): diff -r 09e350db7f8f -r a47580a0955b piecrust/processing/base.py --- a/piecrust/processing/base.py Sat Jan 31 16:28:16 2015 -0800 +++ b/piecrust/processing/base.py Sat Jan 31 17:08:02 2015 -0800 @@ -11,7 +11,8 @@ ProcessorPipelineRecordEntry, TransitionalProcessorPipelineRecord, FLAG_PROCESSED, FLAG_OVERRIDEN, FLAG_BYPASSED_STRUCTURED_PROCESSING) from piecrust.processing.tree import ( - ProcessingTreeBuilder, ProcessingTreeRunner, ProcessingTreeError, + ProcessingTreeBuilder, ProcessingTreeRunner, + ProcessingTreeError, ProcessorError, STATE_DIRTY, print_node, get_node_name_tree) @@ -239,8 +240,10 @@ self.processDirectory(ctx, path) # Wait on all workers. + record.current.success = True for w in pool: w.join() + record.current.success &= w.success if abort.is_set(): raise Exception("Worker pool was aborted.") @@ -310,6 +313,7 @@ super(ProcessingWorker, self).__init__() self.wid = wid self.ctx = ctx + self.success = True def run(self): while(not self.ctx.abort_event.is_set()): @@ -320,11 +324,13 @@ break try: - self._unsafeRun(job) + success = self._unsafeRun(job) logger.debug("[%d] Done with file." % self.wid) self.ctx.work_queue.task_done() + self.success &= success except Exception as ex: self.ctx.abort_event.set() + self.success = False logger.error("[%d] Critical error, aborting." % self.wid) logger.exception(ex) break @@ -347,7 +353,7 @@ record_entry.flags |= FLAG_OVERRIDEN logger.info(format_timed(start_time, '%s [not baked, overridden]' % rel_path)) - return + return True processors = pipeline.getFilteredProcessors( job.mount_info['processors']) @@ -355,9 +361,12 @@ builder = ProcessingTreeBuilder(processors) tree_root = builder.build(rel_path) except ProcessingTreeError as ex: - record_entry.errors.append(str(ex)) - logger.error("Error processing %s: %s" % (rel_path, ex)) - return + msg = str(ex) + logger.error("Error processing %s: %s" % (rel_path, msg)) + while ex: + record_entry.errors.append(str(ex)) + ex = ex.__cause__ + return False print_node(tree_root, recursive=True) leaves = tree_root.getLeaves() @@ -380,10 +389,18 @@ pipeline.out_dir, self.ctx.pipeline_lock) if runner.processSubTree(tree_root): record_entry.flags |= FLAG_PROCESSED - logger.info(format_timed(start_time, "[%d] %s" % (self.wid, rel_path))) + logger.info(format_timed( + start_time, "[%d] %s" % (self.wid, rel_path))) + return True except ProcessingTreeError as ex: - record_entry.errors.append(str(ex)) - logger.error("Error processing %s: %s" % (rel_path, ex)) + msg = str(ex) + if isinstance(ex, ProcessorError): + msg = str(ex.__cause__) + logger.error("Error processing %s: %s" % (rel_path, msg)) + while ex: + record_entry.errors.append(str(ex)) + ex = ex.__cause__ + return False def make_mount_infos(mounts, root_dir): diff -r 09e350db7f8f -r a47580a0955b piecrust/processing/records.py --- a/piecrust/processing/records.py Sat Jan 31 16:28:16 2015 -0800 +++ b/piecrust/processing/records.py Sat Jan 31 17:08:02 2015 -0800 @@ -9,6 +9,7 @@ super(ProcessorPipelineRecord, self).__init__() self.out_dir = None self.process_time = None + self.success = False def hasOverrideEntry(self, rel_path): return self.findEntry(rel_path) is not None diff -r 09e350db7f8f -r a47580a0955b piecrust/processing/tree.py --- a/piecrust/processing/tree.py Sat Jan 31 16:28:16 2015 -0800 +++ b/piecrust/processing/tree.py Sat Jan 31 17:08:02 2015 -0800 @@ -24,6 +24,16 @@ pass +class ProcessorError(ProcessingTreeError): + def __init__(self, proc_name, in_path, *args): + super(ProcessorError, self).__init__(*args) + self.proc_name = proc_name + self.in_path = in_path + + def __str__(self): + return "Processor %s failed on: %s" % (self.proc_name, self.in_path) + + class ProcessingTreeNode(object): def __init__(self, path, available_procs, level=0): self.path = path @@ -154,8 +164,7 @@ colored=False)) return True except Exception as e: - raise ProcessingTreeError("Error processing: %s" % - node.path) from e + raise ProcessorError(proc.PROCESSOR_NAME, full_path) from e # All outputs of a node must go to the same directory, so we can get # the output directory off of the first output. @@ -183,7 +192,7 @@ print_node(node, "-> %s [clean]" % out_dir) return False except Exception as e: - raise Exception("Error processing: %s" % node.path) from e + raise ProcessorError(proc.PROCESSOR_NAME, full_path) from e def _computeNodeState(self, node): if node.state != STATE_UNKNOWN: