@@ -99,7 +99,9 @@ def _normalize_haskell_rows(tf: tarfile.TarFile) -> list[dict[str, str]]:
9999 table_member = f"param-eq-haskell/results/{ dataset } _table_counts.csv"
100100 raw_rows = _read_csv_text (_read_text (tf , table_member ))
101101 exprs_by_algorithm = {
102- path_algorithm : _load_expression_lines (tf , f"param-eq-haskell/results/exprs/{ path_algorithm } _exprs_{ dataset } " )
102+ path_algorithm : _load_expression_lines (
103+ tf , f"param-eq-haskell/results/exprs/{ path_algorithm } _exprs_{ dataset } "
104+ )
103105 for path_algorithm in ("Bingo" , "EPLEX" , "FEAT" , "GOMEA" , "Operon" , "SBP" , "SRjl" )
104106 }
105107 sympy_by_algorithm = {
@@ -150,17 +152,17 @@ def _normalize_runtime_rows(tf: tarfile.TarFile) -> list[dict[str, str]]:
150152 node_count = int (benchmark_name .split ("/" )[- 1 ])
151153 _ , numeric , unit , * _ = time_line .split ()
152154 runtime_ms = float (numeric ) * _to_runtime_multiplier (unit )
153- rows .append (
154- {
155- "benchmark_name" : benchmark_name ,
156- "node_count" : str (node_count ),
157- "runtime_ms" : f"{ runtime_ms :.9f} " ,
158- }
159- )
155+ rows .append ({
156+ "benchmark_name" : benchmark_name ,
157+ "node_count" : str (node_count ),
158+ "runtime_ms" : f"{ runtime_ms :.9f} " ,
159+ })
160160 return rows
161161
162162
163- def _normalize_manifest (bundle : ArchiveBundle , haskell_rows : Iterable [dict [str , str ]], runtime_rows : Iterable [dict [str , str ]]) -> dict [str , Any ]:
163+ def _normalize_manifest (
164+ bundle : ArchiveBundle , haskell_rows : Iterable [dict [str , str ]], runtime_rows : Iterable [dict [str , str ]]
165+ ) -> dict [str , Any ]:
164166 haskell_rows = list (haskell_rows )
165167 runtime_rows = list (runtime_rows )
166168 return {
@@ -169,7 +171,9 @@ def _normalize_manifest(bundle: ArchiveBundle, haskell_rows: Iterable[dict[str,
169171 "haskell_row_count" : len (haskell_rows ),
170172 "paper_row_count" : sum (row ["is_paper_row" ] == "1" for row in haskell_rows ),
171173 "pagie_paper_rows" : sum (row ["dataset" ] == "pagie" and row ["is_paper_row" ] == "1" for row in haskell_rows ),
172- "kotanchek_paper_rows" : sum (row ["dataset" ] == "kotanchek" and row ["is_paper_row" ] == "1" for row in haskell_rows ),
174+ "kotanchek_paper_rows" : sum (
175+ row ["dataset" ] == "kotanchek" and row ["is_paper_row" ] == "1" for row in haskell_rows
176+ ),
173177 "runtime_row_count" : len (runtime_rows ),
174178 "matcher_height_limit" : 8 ,
175179 }
0 commit comments