Add train loss and lr to loggers (#6732)

Signed-off-by: Glenn Jocher <glenn.jocher@ultralytics.com>
This commit is contained in:
Glenn Jocher 2023-12-01 22:09:20 +01:00 committed by GitHub
parent 4425abce59
commit 9618025416
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
5 changed files with 25 additions and 7 deletions

View file

@ -90,8 +90,10 @@ def on_train_epoch_end(trainer):
if trainer.epoch == 1:
_log_debug_samples(sorted(trainer.save_dir.glob('train_batch*.jpg')), 'Mosaic')
# Report the current training progress
for k, v in trainer.validator.metrics.results_dict.items():
for k, v in trainer.label_loss_items(trainer.tloss, prefix='train').items():
task.get_logger().report_scalar('train', k, v, iteration=trainer.epoch)
for k, v in trainer.lr.items():
task.get_logger().report_scalar('lr', k, v, iteration=trainer.epoch)
def on_fit_epoch_end(trainer):
@ -102,6 +104,8 @@ def on_fit_epoch_end(trainer):
series='Epoch Time',
value=trainer.epoch_time,
iteration=trainer.epoch)
for k, v in trainer.metrics.items():
task.get_logger().report_scalar('val', k, v, iteration=trainer.epoch)
if trainer.epoch == 0:
from ultralytics.utils.torch_utils import model_info_for_loggers
for k, v in model_info_for_loggers(trainer).items():