From c7c65d0dc02695eb31ad8f77324656497c046da4 Mon Sep 17 00:00:00 2001 From: florian <f.spreckelsen@inidscale.com> Date: Tue, 4 Oct 2022 14:17:36 +0200 Subject: [PATCH] STY: autopep'd --- src/caoscrawler/crawl.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/src/caoscrawler/crawl.py b/src/caoscrawler/crawl.py index 93196413..1694ffa1 100644 --- a/src/caoscrawler/crawl.py +++ b/src/caoscrawler/crawl.py @@ -239,7 +239,8 @@ class Crawler(object): elif len(crawler_definitions) == 2: crawler_definition = crawler_definitions[1] else: - raise RuntimeError("Crawler definition must not contain more than two documents.") + raise RuntimeError( + "Crawler definition must not contain more than two documents.") # TODO: at this point this function can already load the cfood schema extensions # from the crawler definition and add them to the yaml schema that will be @@ -443,7 +444,8 @@ class Crawler(object): items = [items] self.run_id = uuid.uuid1() - local_converters = Crawler.initialize_converters(crawler_definition, converter_registry) + local_converters = Crawler.initialize_converters( + crawler_definition, converter_registry) # This recursive crawling procedure generates the update list: self.target_data: List[db.Record] = [] self._crawl(items, @@ -525,7 +527,7 @@ class Crawler(object): if (isinstance(p.value, list)): for el in p.value: if (isinstance(el, db.Entity) and el.id is None - and self.get_identified_record_from_local_cache(el) is None): + and self.get_identified_record_from_local_cache(el) is None): return False if (isinstance(p.value, db.Entity) and p.value.id is None and self.get_identified_record_from_local_cache(p.value) is None): -- GitLab