|
|
@@ -999,8 +999,17 @@ def main(unused_argv):
|
|
|
eval_delta = eval_end - eval_start
|
|
|
utils.print_out("eval time for ckpt: %.2f mins (%.2f sent/sec, %.2f tokens/sec)" %
|
|
|
(eval_delta / 60., eval_speed, eval_speed * (eval_src_tokens + eval_output_tokens) / eval_sentences), f=sys.stderr)
|
|
|
+ logging_data = {
|
|
|
+ 'infer_speed_sent': eval_speed,
|
|
|
+ 'infer_speed_toks': eval_speed * (eval_src_tokens + eval_output_tokens) / eval_sentences,
|
|
|
+ }
|
|
|
for lat in sorted(eval_latencies):
|
|
|
utils.print_out("eval latency_%s for ckpt: %.2f ms" % (lat, eval_latencies[lat] * 1000))
|
|
|
+ logging_data['infer_latency_{}'.format(lat)] = eval_latencies[lat] * 1000
|
|
|
+
|
|
|
+ dllogger.log((), logging_data)
|
|
|
+ dllogger.flush()
|
|
|
+
|
|
|
|
|
|
if translate_mode:
|
|
|
detokenize(hparams, hparams.translate_file + ".trans.tok", hparams.translate_file + ".trans")
|
|
|
@@ -1082,8 +1091,8 @@ def main(unused_argv):
|
|
|
(epochs + 1, eval_delta / 60., eval_speed, eval_speed * (eval_src_tokens + eval_output_tokens) / eval_sentences), f=sys.stderr)
|
|
|
logging_data.update({
|
|
|
'bleu': bleu_score,
|
|
|
- 'eval_speed_sent': eval_speed,
|
|
|
- 'eval_speed_toks': eval_speed * (eval_src_tokens + eval_output_tokens) / eval_sentences,
|
|
|
+ 'infer_speed_sent': eval_speed,
|
|
|
+ 'infer_speed_toks': eval_speed * (eval_src_tokens + eval_output_tokens) / eval_sentences,
|
|
|
})
|
|
|
for lat in sorted(eval_latencies):
|
|
|
utils.print_out("eval latency_%s for epoch %d: %.2f ms" % (lat, epochs + 1, eval_latencies[lat] * 1000))
|