diff --git a/src/caoscrawler/conv_impl/spss.py b/src/caoscrawler/conv_impl/spss.py
index 5bd980728b61cfc65818ffb833018ce036fc5a9d..a3201ef0bdcb82280c16f2ab4a5a2e4a47f0c1be 100644
--- a/src/caoscrawler/conv_impl/spss.py
+++ b/src/caoscrawler/conv_impl/spss.py
@@ -44,11 +44,6 @@ ORIGINAL_TYPES = {
 class SPSSConverter(converters.TableConverter):
     """Converter for SAV files (stored by SPSS)."""
 
-    # def __init__(self, *args, **kwargs):
-    #     super().__init__(*args, **kwargs)
-    #     from IPython import embed
-    #     embed()
-
     def create_children(self, values: GeneralStore, element: StructureElement) -> list:
         assert isinstance(element, File)
         # The default dtype backend "numpy_nullable" does not handle dates well.
@@ -56,12 +51,6 @@ class SPSSConverter(converters.TableConverter):
         df = pd.io.spss.read_spss(element.path, dtype_backend="pyarrow")
         dtypes = read_column_types(element.path)
 
-        # if element.path.endswith(".sav"):
-        #     sav_df, meta = pyreadstat.read_sav(element.path, metadataonly=True)
-        # from IPython import embed
-        # embed()
-        df.drop(range(15, len(df.index)), inplace=True)
-
         # Fix datetime columns
         for name, dtype in dtypes.items():
             if dtype != "DATETIME":
@@ -70,9 +59,6 @@ class SPSSConverter(converters.TableConverter):
             col.fillna(np.nan, inplace=True)
             col.replace([np.nan], [None], inplace=True)
 
-        # from IPython import embed
-        # embed()
-
         return self._children_from_dataframe(df)
 
 
@@ -216,21 +202,6 @@ directory: # corresponds to the directory given to the crawler
 # python3 -m caosadvancedtools.models.parser datamodel.yaml --sync
 
 """
-# #
-# # Code for creating enum records:
-# #
-# """
-#     for name, values in enums.items():
-#         for line in f"""
-# cont = db.Container()
-# for value in {repr(values)}:
-#     rec = db.Record(name=value).add_parent(name="{name}")
-#     cont.append(rec)
-# cont.insert()
-#         """.splitlines(keepends=True):
-#             if line.strip():
-#                 output += f"# {line}"
-#     output += "#\n"
 
     # Actual datamodel
     output += """