@@ -1352,7 +1352,7 @@ def fast_eval( # noqa: C901
13521352 res_eval = blosc2 .uninit (shape , dtype , chunks = chunks , blocks = blocks , cparams = cparams , ** kwargs )
13531353 try :
13541354 res_eval ._set_pref_expr (expression , operands , fp_accuracy = fp_accuracy )
1355- print ("expr->miniexpr:" , expression , fp_accuracy )
1355+ # print("expr->miniexpr:", expression, fp_accuracy)
13561356 # Data to compress is fetched from operands, so it can be uninitialized here
13571357 data = np .empty (res_eval .schunk .chunksize , dtype = np .uint8 )
13581358 # Exercise prefilter for each chunk
@@ -2104,7 +2104,7 @@ def reduce_slices( # noqa: C901
21042104 else :
21052105 expression_miniexpr = f"{ reduce_op_str } ({ expression } )"
21062106 res_eval ._set_pref_expr (expression_miniexpr , operands , fp_accuracy , aux_reduc )
2107- print ("expr->miniexpr:" , expression , reduce_op , fp_accuracy )
2107+ # print("expr->miniexpr:", expression, reduce_op, fp_accuracy)
21082108 # Data won't even try to be compressed, so buffers can be unitialized and reused
21092109 data = np .empty (res_eval .schunk .chunksize , dtype = np .uint8 )
21102110 chunk_data = np .empty (res_eval .schunk .chunksize + blosc2 .MAX_OVERHEAD , dtype = np .uint8 )
0 commit comments