@@ -319,8 +319,8 @@ def train_once(
319319 metrics_logger = None
320320 if RANK == 0 :
321321 metrics_logger = logger_utils .set_up_loggers (log_dir ,
322- flags .FLAGS ,
323- hyperparameters )
322+ flags .FLAGS ,
323+ hyperparameters )
324324 workload .attach_metrics_logger (metrics_logger )
325325
326326 global_start_time = get_time ()
@@ -470,13 +470,13 @@ def train_once(
470470 metrics = {'eval_results' : eval_results , 'global_step' : global_step }
471471
472472 if log_dir is not None and RANK == 0 :
473- metrics_logger .append_scalar_metrics (
474- {'score' : train_state ['accumulated_submission_time' ]},
475- global_step = global_step ,
476- preemption_count = preemption_count )
477- metrics_logger .finish ()
478- if save_checkpoints :
479- checkpoint_utils .save_checkpoint (
473+ metrics_logger .append_scalar_metrics (
474+ {'score' : train_state ['accumulated_submission_time' ]},
475+ global_step = global_step ,
476+ preemption_count = preemption_count )
477+ metrics_logger .finish ()
478+ if save_checkpoints :
479+ checkpoint_utils .save_checkpoint (
480480 framework = FLAGS .framework ,
481481 optimizer_state = optimizer_state ,
482482 model_params = model_params ,
0 commit comments