update script
Browse files- generate_files.py +16 -16
generate_files.py
CHANGED
|
@@ -5,13 +5,13 @@ import pyarrow.parquet as pq
|
|
| 5 |
from huggingface_hub import hf_hub_download, login
|
| 6 |
|
| 7 |
# Update this if needed
|
| 8 |
-
DATASET = "severo/
|
| 9 |
|
| 10 |
|
| 11 |
def write_table(table, filename):
|
| 12 |
-
for row_group_size in ["128k"
|
| 13 |
-
for use_content_defined_chunking in ["cdc"
|
| 14 |
-
for compression in ["none"
|
| 15 |
path = f"hf://datasets/{DATASET}/{row_group_size}/{use_content_defined_chunking}/{compression}/{filename}"
|
| 16 |
print(f"\nTrying to write to {path}")
|
| 17 |
pq.write_table(
|
|
@@ -34,18 +34,18 @@ def main():
|
|
| 34 |
|
| 35 |
for filename in [
|
| 36 |
"2023_05.parquet",
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
]:
|
| 50 |
print(f"\n\nProcessing {filename}")
|
| 51 |
|
|
|
|
| 5 |
from huggingface_hub import hf_hub_download, login
|
| 6 |
|
| 7 |
# Update this if needed
|
| 8 |
+
DATASET = "severo/notebooks_on_the_hub"
|
| 9 |
|
| 10 |
|
| 11 |
def write_table(table, filename):
|
| 12 |
+
for row_group_size in ["128k", "1m"]:
|
| 13 |
+
for use_content_defined_chunking in ["cdc", "no_cdc"]:
|
| 14 |
+
for compression in ["none", "snappy"]:
|
| 15 |
path = f"hf://datasets/{DATASET}/{row_group_size}/{use_content_defined_chunking}/{compression}/{filename}"
|
| 16 |
print(f"\nTrying to write to {path}")
|
| 17 |
pq.write_table(
|
|
|
|
| 34 |
|
| 35 |
for filename in [
|
| 36 |
"2023_05.parquet",
|
| 37 |
+
"2023_06.parquet",
|
| 38 |
+
"2023_07.parquet",
|
| 39 |
+
"2023_08.parquet",
|
| 40 |
+
"2023_09.parquet",
|
| 41 |
+
"2023_10.parquet",
|
| 42 |
+
"2023_11.parquet",
|
| 43 |
+
"2023_12.parquet",
|
| 44 |
+
"2024_01.parquet",
|
| 45 |
+
"2024_02.parquet",
|
| 46 |
+
"2024_03.parquet",
|
| 47 |
+
"2024_04.parquet",
|
| 48 |
+
"2024_05.parquet",
|
| 49 |
]:
|
| 50 |
print(f"\n\nProcessing {filename}")
|
| 51 |
|