pytorch-ignite 0.6.0.dev20260109__py3-none-any.whl → 0.6.0.dev20260111__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
ignite/__init__.py CHANGED
@@ -6,4 +6,4 @@ import ignite.handlers
6
6
  import ignite.metrics
7
7
  import ignite.utils
8
8
 
9
- __version__ = "0.6.0.dev20260109"
9
+ __version__ = "0.6.0.dev20260111"
@@ -325,7 +325,7 @@ class OutputHandler(BaseOutputHandler):
325
325
  def __init__(
326
326
  self,
327
327
  tag: str,
328
- metric_names: Optional[List[str]] = None,
328
+ metric_names: Optional[Union[List[str], str]] = None,
329
329
  output_transform: Optional[Callable] = None,
330
330
  global_step_transform: Optional[Callable[[Engine, Union[str, Events]], int]] = None,
331
331
  state_attributes: Optional[List[str]] = None,
@@ -344,8 +344,7 @@ class OutputHandler(BaseOutputHandler):
344
344
 
345
345
  if not isinstance(global_step, int):
346
346
  raise TypeError(
347
- f"global_step must be int, got {type(global_step)}."
348
- " Please check the output of global_step_transform."
347
+ f"global_step must be int, got {type(global_step)}. Please check the output of global_step_transform."
349
348
  )
350
349
 
351
350
  for key, value in metrics.items():
@@ -317,7 +317,7 @@ class FastaiLRFinder:
317
317
  lrs = lrs[skip_start:-skip_end]
318
318
  losses = losses[skip_start:-skip_end]
319
319
 
320
- plt.legend(legends)
320
+ plt.legend(legends) # pyrefly: ignore
321
321
  # Plot loss as a function of the learning rate
322
322
  ax.plot(lrs, losses)
323
323
  if log_lr:
@@ -5,7 +5,6 @@ from typing import Any, Callable, List, Optional, Union
5
5
  from torch.optim import Optimizer
6
6
 
7
7
  from ignite.engine import Engine, Events
8
-
9
8
  from ignite.handlers.base_logger import BaseLogger, BaseOptimizerParamsHandler, BaseOutputHandler
10
9
  from ignite.handlers.utils import global_step_from_engine # noqa
11
10
 
@@ -237,7 +236,7 @@ class OutputHandler(BaseOutputHandler):
237
236
  def __init__(
238
237
  self,
239
238
  tag: str,
240
- metric_names: Optional[List[str]] = None,
239
+ metric_names: Optional[Union[List[str], str]] = None,
241
240
  output_transform: Optional[Callable] = None,
242
241
  global_step_transform: Optional[Callable[[Engine, Union[str, Events]], int]] = None,
243
242
  state_attributes: Optional[List[str]] = None,
@@ -256,8 +255,7 @@ class OutputHandler(BaseOutputHandler):
256
255
 
257
256
  if not isinstance(global_step, int):
258
257
  raise TypeError(
259
- f"global_step must be int, got {type(global_step)}."
260
- " Please check the output of global_step_transform."
258
+ f"global_step must be int, got {type(global_step)}. Please check the output of global_step_transform."
261
259
  )
262
260
 
263
261
  metrics.update({"step": global_step})
@@ -5,7 +5,6 @@ from typing import Any, Callable, List, Optional, Union
5
5
  from torch.optim import Optimizer
6
6
 
7
7
  from ignite.engine import Engine, Events
8
-
9
8
  from ignite.handlers.base_logger import (
10
9
  BaseLogger,
11
10
  BaseOptimizerParamsHandler,
@@ -322,7 +321,7 @@ class OutputHandler(BaseOutputHandler):
322
321
  def __init__(
323
322
  self,
324
323
  tag: str,
325
- metric_names: Optional[List[str]] = None,
324
+ metric_names: Optional[Union[List[str], str]] = None,
326
325
  output_transform: Optional[Callable] = None,
327
326
  global_step_transform: Optional[Callable[[Engine, Union[str, Events]], int]] = None,
328
327
  state_attributes: Optional[List[str]] = None,
@@ -340,8 +339,7 @@ class OutputHandler(BaseOutputHandler):
340
339
  global_step = self.global_step_transform(engine, event_name)
341
340
  if not isinstance(global_step, int):
342
341
  raise TypeError(
343
- f"global_step must be int, got {type(global_step)}."
344
- " Please check the output of global_step_transform."
342
+ f"global_step must be int, got {type(global_step)}. Please check the output of global_step_transform."
345
343
  )
346
344
 
347
345
  for key, value in metrics.items():
@@ -1,7 +1,7 @@
1
1
  """Visdom logger and its helper handlers."""
2
2
 
3
3
  import os
4
- from typing import Any, Callable, Dict, List, Optional, Union, TYPE_CHECKING
4
+ from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Union
5
5
 
6
6
  import torch
7
7
  import torch.nn as nn
@@ -14,7 +14,6 @@ from ignite.handlers.base_logger import (
14
14
  BaseOutputHandler,
15
15
  BaseWeightsScalarHandler,
16
16
  )
17
-
18
17
  from ignite.handlers.utils import global_step_from_engine # noqa
19
18
 
20
19
  __all__ = [
@@ -361,7 +360,7 @@ class OutputHandler(BaseOutputHandler, _BaseVisDrawer):
361
360
  def __init__(
362
361
  self,
363
362
  tag: str,
364
- metric_names: Optional[str] = None,
363
+ metric_names: Optional[Union[List[str], str]] = None,
365
364
  output_transform: Optional[Callable] = None,
366
365
  global_step_transform: Optional[Callable[[Engine, Union[str, Events]], int]] = None,
367
366
  show_legend: bool = False,
@@ -382,8 +381,7 @@ class OutputHandler(BaseOutputHandler, _BaseVisDrawer):
382
381
 
383
382
  if not isinstance(global_step, int):
384
383
  raise TypeError(
385
- f"global_step must be int, got {type(global_step)}."
386
- " Please check the output of global_step_transform."
384
+ f"global_step must be int, got {type(global_step)}. Please check the output of global_step_transform."
387
385
  )
388
386
 
389
387
  for key, value in metrics.items():
@@ -6,7 +6,6 @@ from warnings import warn
6
6
  from torch.optim import Optimizer
7
7
 
8
8
  from ignite.engine import Engine, Events
9
-
10
9
  from ignite.handlers.base_logger import BaseLogger, BaseOptimizerParamsHandler, BaseOutputHandler
11
10
  from ignite.handlers.utils import global_step_from_engine # noqa
12
11
 
@@ -277,7 +276,7 @@ class OutputHandler(BaseOutputHandler):
277
276
  def __init__(
278
277
  self,
279
278
  tag: str,
280
- metric_names: Optional[List[str]] = None,
279
+ metric_names: Optional[Union[List[str], str]] = None,
281
280
  output_transform: Optional[Callable] = None,
282
281
  global_step_transform: Optional[Callable[[Engine, Union[str, Events]], int]] = None,
283
282
  sync: Optional[bool] = None,
@@ -294,8 +293,7 @@ class OutputHandler(BaseOutputHandler):
294
293
  global_step = self.global_step_transform(engine, event_name)
295
294
  if not isinstance(global_step, int):
296
295
  raise TypeError(
297
- f"global_step must be int, got {type(global_step)}."
298
- " Please check the output of global_step_transform."
296
+ f"global_step must be int, got {type(global_step)}. Please check the output of global_step_transform."
299
297
  )
300
298
 
301
299
  metrics = self._setup_output_metrics_state_attrs(engine, log_text=True, key_tuple=False)
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: pytorch-ignite
3
- Version: 0.6.0.dev20260109
3
+ Version: 0.6.0.dev20260111
4
4
  Summary: A lightweight library to help with training neural networks in PyTorch.
5
5
  Project-URL: Homepage, https://pytorch-ignite.ai
6
6
  Project-URL: Repository, https://github.com/pytorch/ignite
@@ -1,4 +1,4 @@
1
- ignite/__init__.py,sha256=1gr3jaj8-9mq5MSVFKOJ9C7hDUoUwFudcoLgPq4ZI_Q,194
1
+ ignite/__init__.py,sha256=MiNYi1mCgbI-dp3ZwOIcFyMc9ZoULQr7MFWApr8c9Ow,194
2
2
  ignite/_utils.py,sha256=XDPpUDJ8ykLXWMV2AYTqGSj8XCfApsyzsQ3Vij_OB4M,182
3
3
  ignite/exceptions.py,sha256=5ZWCVLPC9rgoW8t84D-VeEleqz5O7XpAGPpCdU8rKd0,150
4
4
  ignite/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -62,26 +62,26 @@ ignite/engine/utils.py,sha256=QG5mkdg4OipspqgpNQcJuoHTYdr2Sx5LS16kfjOHDdI,1073
62
62
  ignite/handlers/__init__.py,sha256=Qq85YTtHPcii6UAfMOoCPg9RwigH96iqxOJKIlRfDqw,2728
63
63
  ignite/handlers/base_logger.py,sha256=wPiGn9iCh5ung1GaRUf_qAlqe63h1NpUUQ0XK709p2k,13011
64
64
  ignite/handlers/checkpoint.py,sha256=u6cFUDxAoSSBKCBprmDud2LEZGDEYHvyCoLUmtG3Xd4,46309
65
- ignite/handlers/clearml_logger.py,sha256=0-57RYznIz-EgTsKtkKFPdGGFQXJIhq146H_qiE8hVc,37897
65
+ ignite/handlers/clearml_logger.py,sha256=MlqHtFjC04yrtvAmFTzPPgwDkiRrXGrVpi1NRAoMqmk,37890
66
66
  ignite/handlers/early_stopping.py,sha256=UA6TiKho5CbD085R-16H8w3r0BYPQcWQjhEXg8aITSw,4139
67
67
  ignite/handlers/ema_handler.py,sha256=SmUyyWIFPZW3yMvjD_sSk5m_LfnMFl9R-uQdbXNFfY0,11854
68
68
  ignite/handlers/fbresearch_logger.py,sha256=MfQeiBIXBYLEwZoDIld2oCceMeTAsz8rc5cd7fLtpJs,11133
69
- ignite/handlers/lr_finder.py,sha256=EMcQR3NDPOuh2s85a5Zu5Bqt0I4pg1cACJpjSa5cO4A,22100
69
+ ignite/handlers/lr_finder.py,sha256=VEJyUTjEXkJzrSaVzMNIwON5uUIpBGIeGWbNucX_kzk,22119
70
70
  ignite/handlers/mlflow_logger.py,sha256=M5Mggrnr2wMsms8wbEaHqNtTk5L1zNs1MlPWD0ZCpDQ,13894
71
71
  ignite/handlers/neptune_logger.py,sha256=Rv-O_i0zGZC2Ozzeetxv7rtD7iP3IeWEcbY-U28Mkzg,27348
72
72
  ignite/handlers/param_scheduler.py,sha256=Tn4o27YBrp5JsuadHobIrsHfmvB_cR1IrV_oV1Eo7us,68373
73
- ignite/handlers/polyaxon_logger.py,sha256=5b7Zxhksne8Ufg_SBTG-rlf_9CPSjkBQOJR4-ynoZnQ,12354
73
+ ignite/handlers/polyaxon_logger.py,sha256=7nyOhu4rgg4Sj6uTNUobPG9IJqYxjb_fopKiF4fk9vc,12346
74
74
  ignite/handlers/state_param_scheduler.py,sha256=B89YKZyj9DXLXQyr3amDNMslUOWNHZDis2DXIwW0q10,20841
75
75
  ignite/handlers/stores.py,sha256=8XM_Qqsitfu0WtOOE-K2FMtv51vD90r3GgQlCzRABYc,2616
76
- ignite/handlers/tensorboard_logger.py,sha256=q3YxXkbIFayBggI_kcHyl-upttVVjjnqFOLgyjj2cRo,27967
76
+ ignite/handlers/tensorboard_logger.py,sha256=e6gS9b7dkGjm-iR4iTQWPZcYKrDUml7cmrriNXoMBoU,27959
77
77
  ignite/handlers/terminate_on_nan.py,sha256=RFSKd3Oqn9Me2xLCos4lSE-hnY7fYWWjE9blioeMlIs,2103
78
78
  ignite/handlers/time_limit.py,sha256=heTuS-ReBbOUCm1NcNJGhzxI080Hanc4hOLB2Y4GeZk,1567
79
79
  ignite/handlers/time_profilers.py,sha256=8iCcBYPxv0vKFSO_ujFV0ST54a9PD9ezFLvYTIu9lFI,30482
80
80
  ignite/handlers/timing.py,sha256=nHeBHvPwYdPRMAx-jk_8MjZit4a7rmsmIWkUrajAG-s,4705
81
81
  ignite/handlers/tqdm_logger.py,sha256=3kxH39vM0LCDVwIZl9HQRaWM2Pr6bYC_l9oydFJmdM4,13093
82
82
  ignite/handlers/utils.py,sha256=X4LRqo1kqGsbmX0pEuZKYR6K4C8sZudAqxCLriiXtCg,872
83
- ignite/handlers/visdom_logger.py,sha256=RY5ss3NAPad7d3xFFnqczCtuO6RgmWq9ROz-sFf6imI,21862
84
- ignite/handlers/wandb_logger.py,sha256=vGok3gADQmTNkc6KkfFBreYoHAO8EneuU65xjBpT5-Q,14837
83
+ ignite/handlers/visdom_logger.py,sha256=OlnqVDXKYTea6VG7rwcQdzrfYmai0SSxNLh2KqsjreU,21860
84
+ ignite/handlers/wandb_logger.py,sha256=9HbwRMHzWckrZ-m0rkMF5Ug6r9C9J4sdq73yqaAHWYE,14829
85
85
  ignite/metrics/__init__.py,sha256=m-8F8J17r-aEwsO6Ww-8AqDRN59WFfYBwCDKwqGDSmI,3627
86
86
  ignite/metrics/accumulation.py,sha256=xWdsm9u6JfsfODX_GUKzQc_omrdFDJ4yELBR-xXgc4s,12448
87
87
  ignite/metrics/accuracy.py,sha256=W8mO4W11VzryMXKy8G7W_g4A9PH9RYpejW_tQ-T_Txw,10245
@@ -153,7 +153,7 @@ ignite/metrics/regression/spearman_correlation.py,sha256=IzmN4WIe7C4cTUU3BOkBmaw
153
153
  ignite/metrics/regression/wave_hedges_distance.py,sha256=Ji_NRUgnZ3lJgi5fyNFLRjbHO648z4dBmqVDQU9ImKA,2792
154
154
  ignite/metrics/vision/__init__.py,sha256=lPBAEq1idc6Q17poFm1SjttE27irHF1-uNeiwrxnLrU,159
155
155
  ignite/metrics/vision/object_detection_average_precision_recall.py,sha256=4wwiNVd658ynIpIbQlffTA-ehvyJ2EzmJ5pBSBuA8XQ,25091
156
- pytorch_ignite-0.6.0.dev20260109.dist-info/METADATA,sha256=r-rroOFoSErSxxGGo-lndF5NkiaCmmXEf3p5gyX9qfE,27979
157
- pytorch_ignite-0.6.0.dev20260109.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
158
- pytorch_ignite-0.6.0.dev20260109.dist-info/licenses/LICENSE,sha256=SwJvaRmy1ql-k9_nL4WnER4_ODTMF9fWoP9HXkoicgw,1527
159
- pytorch_ignite-0.6.0.dev20260109.dist-info/RECORD,,
156
+ pytorch_ignite-0.6.0.dev20260111.dist-info/METADATA,sha256=5xtdAYoZ8V5bblHaFYKlRQ3u415PHC74htcAKkojBQk,27979
157
+ pytorch_ignite-0.6.0.dev20260111.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
158
+ pytorch_ignite-0.6.0.dev20260111.dist-info/licenses/LICENSE,sha256=SwJvaRmy1ql-k9_nL4WnER4_ODTMF9fWoP9HXkoicgw,1527
159
+ pytorch_ignite-0.6.0.dev20260111.dist-info/RECORD,,