add: 1900 csv files in CSV_DIR for aggregation
This commit is contained in:
parent
3610421c7f
commit
35943d2e0c
|
@ -960,8 +960,8 @@ def create_batch_of_csv_files():
|
|||
shutil.copy(first_csv_file, logging_dir)
|
||||
manage_csv_files(logging_dir)
|
||||
|
||||
# keep at most 100 files at CSV_DIR for logging
|
||||
manage_csv_files(CSV_DIR, 100)
|
||||
# keep at most 1900 files at CSV_DIR for logging and aggregation
|
||||
manage_csv_files(CSV_DIR, 1900)
|
||||
|
||||
# prepare for compression
|
||||
csv_data = read_csv_as_string(first_csv_file)
|
||||
|
|
|
@ -632,8 +632,8 @@ def create_batch_of_csv_files():
|
|||
shutil.copy(first_csv_file, logging_dir)
|
||||
manage_csv_files(logging_dir)
|
||||
|
||||
# keep at most 100 files at CSV_DIR for logging
|
||||
manage_csv_files(CSV_DIR, 100)
|
||||
# keep at most 1900 files at CSV_DIR for logging and aggregation
|
||||
manage_csv_files(CSV_DIR, 1900)
|
||||
|
||||
# print("The batch csv file is: {}".format(recent_csv_files[-1]))
|
||||
|
||||
|
|
Loading…
Reference in New Issue