File size: 3,249 Bytes
2eae90c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 |
import os
import timeit
from os.path import join
from typing import Callable, Union
import dask.dataframe as dd
import pandas as pd
from linetimer import CodeTimer, linetimer
from common_utils import (
ANSWERS_BASE_DIR,
DATASET_BASE_DIR,
FILE_TYPE,
INCLUDE_IO,
LOG_TIMINGS,
SHOW_RESULTS,
append_row,
on_second_call,
)
def read_ds(path: str) -> Union:
if INCLUDE_IO:
return dd.read_parquet(path)
if FILE_TYPE == "feather":
raise ValueError("file type feather not supported for dask queries")
return dd.from_pandas(pd.read_parquet(path), npartitions=os.cpu_count())
def get_query_answer(query: int, base_dir: str = ANSWERS_BASE_DIR) -> dd.DataFrame:
answer_df = pd.read_csv(
join(base_dir, f"q{query}.out"),
sep="|",
parse_dates=True,
infer_datetime_format=True,
)
return answer_df.rename(columns=lambda x: x.strip())
def test_results(q_num: int, result_df: pd.DataFrame):
with CodeTimer(name=f"Testing result of dask Query {q_num}", unit="s"):
answer = get_query_answer(q_num)
for c, t in answer.dtypes.items():
s1 = result_df[c]
s2 = answer[c]
if t.name == "object":
s1 = s1.astype("string").apply(lambda x: x.strip())
s2 = s2.astype("string").apply(lambda x: x.strip())
pd.testing.assert_series_equal(left=s1, right=s2, check_index=False)
@on_second_call
def get_line_item_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "lineitem.parquet"))
@on_second_call
def get_orders_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "orders.parquet"))
@on_second_call
def get_customer_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "customer.parquet"))
@on_second_call
def get_region_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "region.parquet"))
@on_second_call
def get_nation_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "nation.parquet"))
@on_second_call
def get_supplier_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "supplier.parquet"))
@on_second_call
def get_part_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "part.parquet"))
@on_second_call
def get_part_supp_ds(base_dir: str = DATASET_BASE_DIR) -> dd.DataFrame:
return read_ds(join(base_dir, "partsupp.parquet"))
def run_query(q_num: str, query: Callable):
@linetimer(name=f"Overall execution of dask Query {q_num}", unit="s")
def run():
import dask
dask.config.set(scheduler="threads")
with CodeTimer(name=f"Get result of dask Query {q_num}", unit="s"):
t0 = timeit.default_timer()
result = query()
secs = timeit.default_timer() - t0
if LOG_TIMINGS:
append_row(
solution="dask", version=dask.__version__, q=f"q{q_num}", secs=secs
)
else:
test_results(q_num, result)
if SHOW_RESULTS:
print(result)
run()
|