Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
21 changes: 11 additions & 10 deletions templates/gan/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,10 @@
from config import get_default_parser


PRINT_FREQ = 100
FAKE_IMG_FNAME = "fake_sample_epoch_{:04d}.png"
REAL_IMG_FNAME = "real_sample_epoch_{:04d}.png"
LOGS_FNAME = "logs.tsv"
PLOT_FNAME = "plot.svg"
SAMPLES_FNAME = "samples.svg"
CKPT_PREFIX = "networks"


def run(local_rank: int, config: Any, *args: Any, **kwargs: Any):
Expand Down Expand Up @@ -112,7 +109,10 @@ def run(local_rank: int, config: Any, *args: Any, **kwargs: Any):
lr_scheduler=lr_scheduler,
output_names=["errD", "errG", "D_x", "D_G_z1", "D_G_z2"],
)
logger_handler = get_logger(config=config, train_engine=train_engine, optimizers=optimizers)

# setup ignite logger only on rank 0
if rank == 0:
logger_handler = get_logger(config=config, train_engine=train_engine, optimizers=optimizers)

# -----------------------------------
# resume from the saved checkpoints
Expand Down Expand Up @@ -192,12 +192,13 @@ def create_plots(engine):
# close the logger after the training completed / terminated
# ------------------------------------------------------------

if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()
if rank == 0:
if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()

# -----------------------------------------
# where is my best and last checkpoint ?
Expand Down
18 changes: 11 additions & 7 deletions templates/image_classification/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,10 @@ def run(local_rank: int, config: Any, *args: Any, **kwargs: Any):
lr_scheduler=lr_scheduler,
output_names=None,
)
logger_handler = get_logger(config=config, train_engine=train_engine, eval_engine=eval_engine, optimizers=optimizer)

# setup ignite logger only on rank 0
if rank == 0:
logger_handler = get_logger(config=config, train_engine=train_engine, eval_engine=eval_engine, optimizers=optimizer)

# -----------------------------------
# resume from the saved checkpoints
Expand Down Expand Up @@ -198,12 +201,13 @@ def _():
# close the logger after the training completed / terminated
# ------------------------------------------------------------

if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()
if rank == 0:
if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()

# -----------------------------------------
# where is my best and last checkpoint ?
Expand Down
18 changes: 11 additions & 7 deletions templates/single/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,10 @@ def run(local_rank: int, config: Any, *args: Any, **kwargs: Any):
lr_scheduler=lr_scheduler,
output_names=None,
)
logger_handler = get_logger(config=config, train_engine=train_engine, eval_engine=eval_engine, optimizers=optimizer)

# setup ignite logger only on rank 0
if rank == 0:
logger_handler = get_logger(config=config, train_engine=train_engine, eval_engine=eval_engine, optimizers=optimizer)

# -----------------------------------
# resume from the saved checkpoints
Expand Down Expand Up @@ -175,12 +178,13 @@ def _():
# close the logger after the training completed / terminated
# ------------------------------------------------------------

if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()
if rank == 0:
if isinstance(logger_handler, WandBLogger):
# why handle differently for wandb ?
# See : https://github.com/pytorch/ignite/issues/1894
logger_handler.finish()
elif logger_handler:
logger_handler.close()

# -----------------------------------------
# where is my best and last checkpoint ?
Expand Down