diff --git a/src/caoscrawler/crawl.py b/src/caoscrawler/crawl.py
index f016df6ab0ad0ab848c6cf9d02f938fd5878a9cf..a26b0da7f24fa74d912c9a60bb717f6edde8d9dc 100644
--- a/src/caoscrawler/crawl.py
+++ b/src/caoscrawler/crawl.py
@@ -336,7 +336,7 @@ class Crawler(object):
             #                 )
 
             raise RuntimeError(
-                f"Could not finish split_into_inserts_and_updates. "
+                "Could not finish split_into_inserts_and_updates. "
                 "It might be due to a circular dependency")
 
         return st.export_record_lists()
@@ -692,14 +692,14 @@ class Crawler(object):
             cache_clear()
             self.execute_parent_updates_in_list(to_be_updated, securityMode=self.securityMode,
                                                 run_id=self.run_id, unique_names=unique_names)
-            logger.info(f"Added parent RecordTypes where necessary.")
+            logger.info("Added parent RecordTypes where necessary.")
             self.execute_inserts_in_list(
                 to_be_inserted, self.securityMode, self.run_id, unique_names=unique_names)
-            logger.info(f"Executed inserts:\n"
+            logger.info("Executed inserts:\n"
                         + self.create_entity_summary(to_be_inserted))
             self.execute_updates_in_list(
                 to_be_updated, self.securityMode, self.run_id, unique_names=unique_names)
-            logger.info(f"Executed updates:\n"
+            logger.info("Executed updates:\n"
                         + self.create_entity_summary(to_be_updated))
 
         update_cache = UpdateCache()
@@ -832,13 +832,13 @@ def _update_status_record(run_id, n_inserts, n_updates, status):
         cr_rec.get_property('status').value = status
         (cr_rec
             .add_property(db.execute_query(
-                f"FIND Property with name='number_of_inserted_entities'", unique=True).id,
+                "FIND Property with name='number_of_inserted_entities'", unique=True).id,
                 n_inserts)
             .add_property(
-                db.execute_query(f"FIND Property with name='number_of_updated_entities'",
+                db.execute_query("FIND Property with name='number_of_updated_entities'",
                                  unique=True).id, n_updates)
             .add_property(
-                db.execute_query(f"FIND Property with name='finished'",
+                db.execute_query("FIND Property with name='finished'",
                                  unique=True).id, datetime.now().isoformat()))
         cr_rec.update()
 
diff --git a/src/caoscrawler/sync_graph.py b/src/caoscrawler/sync_graph.py
index b858327773ebb8903a4fbb0d4e9a16f69126d300..1108bb174a67723fc3dd62bb0c231e9066deb259 100644
--- a/src/caoscrawler/sync_graph.py
+++ b/src/caoscrawler/sync_graph.py
@@ -215,7 +215,7 @@ class SyncGraph:
         """exports the SyncGraph in form of db.Entities
 
         All nodes are converted to db.Entity objects and reference values that are SyncNodes are
-        replaced by their corresponding (newly created)  db.Entity objects.
+        replaced by their corresponding (newly created) db.Entity objects.
 
         Since the result is returned in form of two lists, one with Entities that have a valid ID
         one with those that haven't, an error is raised if there are any SyncNodes without an