@@ -395,23 +395,23 @@ def dataset_create(
395395
396396 package .add_counts ()
397397
398- # Repeat for parquet
399- # Set up cache directory to store parquet files. The sqlite files created from this will be saved in the dataset
400- if not os .path .exists (cache_dir ):
401- os .makedirs (cache_dir )
402-
403- pqpackage = DatasetParquetPackage (
404- dataset ,
405- path = output_path ,
406- input_paths = input_paths ,
407- specification_dir = None , # TBD: package should use this specification object
408- )
409- pqpackage .create_temp_table (input_paths )
410- pqpackage .load_facts (input_paths , cache_dir )
411- pqpackage .load_fact_resource (input_paths , cache_dir )
412- pqpackage .load_entities (input_paths , cache_dir , organisation_path )
413- pqpackage .pq_to_sqlite (output_path , cache_dir )
414- pqpackage .close_conn ()
398+ # # Repeat for parquet
399+ # # Set up cache directory to store parquet files. The sqlite files created from this will be saved in the dataset
400+ # if not os.path.exists(cache_dir):
401+ # os.makedirs(cache_dir)
402+ #
403+ # pqpackage = DatasetParquetPackage(
404+ # dataset,
405+ # path=output_path,
406+ # input_paths=input_paths,
407+ # specification_dir=None, # TBD: package should use this specification object
408+ # )
409+ # pqpackage.create_temp_table(input_paths)
410+ # pqpackage.load_facts(input_paths, cache_dir)
411+ # pqpackage.load_fact_resource(input_paths, cache_dir)
412+ # pqpackage.load_entities(input_paths, cache_dir, organisation_path)
413+ # pqpackage.pq_to_sqlite(output_path, cache_dir)
414+ # pqpackage.close_conn()
415415
416416
417417def dataset_dump (input_path , output_path ):
0 commit comments