nshtrainer 1.0.0b24__py3-none-any.whl → 1.0.0b26__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (71) hide show
  1. nshtrainer/.nshconfig.generated.json +6 -0
  2. nshtrainer/_checkpoint/metadata.py +1 -1
  3. nshtrainer/callbacks/__init__.py +3 -0
  4. nshtrainer/callbacks/checkpoint/__init__.py +4 -0
  5. nshtrainer/callbacks/checkpoint/best_checkpoint.py +1 -2
  6. nshtrainer/callbacks/checkpoint/last_checkpoint.py +1 -2
  7. nshtrainer/callbacks/checkpoint/time_checkpoint.py +114 -0
  8. nshtrainer/callbacks/print_table.py +2 -2
  9. nshtrainer/configs/__init__.py +95 -10
  10. nshtrainer/configs/_checkpoint/__init__.py +6 -0
  11. nshtrainer/configs/_checkpoint/metadata/__init__.py +5 -0
  12. nshtrainer/configs/_directory/__init__.py +5 -1
  13. nshtrainer/configs/_hf_hub/__init__.py +6 -0
  14. nshtrainer/configs/callbacks/__init__.py +44 -1
  15. nshtrainer/configs/callbacks/actsave/__init__.py +5 -0
  16. nshtrainer/configs/callbacks/base/__init__.py +4 -0
  17. nshtrainer/configs/callbacks/checkpoint/__init__.py +14 -0
  18. nshtrainer/configs/callbacks/checkpoint/_base/__init__.py +6 -0
  19. nshtrainer/configs/callbacks/checkpoint/best_checkpoint/__init__.py +7 -0
  20. nshtrainer/configs/callbacks/checkpoint/last_checkpoint/__init__.py +6 -0
  21. nshtrainer/configs/callbacks/checkpoint/on_exception_checkpoint/__init__.py +5 -0
  22. nshtrainer/configs/callbacks/debug_flag/__init__.py +5 -0
  23. nshtrainer/configs/callbacks/directory_setup/__init__.py +5 -0
  24. nshtrainer/configs/callbacks/early_stopping/__init__.py +6 -0
  25. nshtrainer/configs/callbacks/ema/__init__.py +5 -0
  26. nshtrainer/configs/callbacks/finite_checks/__init__.py +5 -0
  27. nshtrainer/configs/callbacks/gradient_skipping/__init__.py +5 -0
  28. nshtrainer/configs/callbacks/log_epoch/__init__.py +5 -0
  29. nshtrainer/configs/callbacks/lr_monitor/__init__.py +5 -0
  30. nshtrainer/configs/callbacks/norm_logging/__init__.py +5 -0
  31. nshtrainer/configs/callbacks/print_table/__init__.py +5 -0
  32. nshtrainer/configs/callbacks/rlp_sanity_checks/__init__.py +5 -0
  33. nshtrainer/configs/callbacks/shared_parameters/__init__.py +5 -0
  34. nshtrainer/configs/callbacks/timer/__init__.py +5 -0
  35. nshtrainer/configs/callbacks/wandb_upload_code/__init__.py +5 -0
  36. nshtrainer/configs/callbacks/wandb_watch/__init__.py +5 -0
  37. nshtrainer/configs/loggers/__init__.py +16 -1
  38. nshtrainer/configs/loggers/_base/__init__.py +4 -0
  39. nshtrainer/configs/loggers/actsave/__init__.py +5 -0
  40. nshtrainer/configs/loggers/csv/__init__.py +5 -0
  41. nshtrainer/configs/loggers/tensorboard/__init__.py +5 -0
  42. nshtrainer/configs/loggers/wandb/__init__.py +8 -0
  43. nshtrainer/configs/lr_scheduler/__init__.py +10 -4
  44. nshtrainer/configs/lr_scheduler/_base/__init__.py +4 -0
  45. nshtrainer/configs/lr_scheduler/linear_warmup_cosine/__init__.py +5 -3
  46. nshtrainer/configs/lr_scheduler/reduce_lr_on_plateau/__init__.py +6 -0
  47. nshtrainer/configs/metrics/__init__.py +5 -0
  48. nshtrainer/configs/metrics/_config/__init__.py +4 -0
  49. nshtrainer/configs/nn/__init__.py +21 -1
  50. nshtrainer/configs/nn/mlp/__init__.py +5 -1
  51. nshtrainer/configs/nn/nonlinearity/__init__.py +18 -1
  52. nshtrainer/configs/optimizer/__init__.py +5 -1
  53. nshtrainer/configs/profiler/__init__.py +11 -1
  54. nshtrainer/configs/profiler/_base/__init__.py +4 -0
  55. nshtrainer/configs/profiler/advanced/__init__.py +5 -0
  56. nshtrainer/configs/profiler/pytorch/__init__.py +5 -0
  57. nshtrainer/configs/profiler/simple/__init__.py +5 -0
  58. nshtrainer/configs/trainer/__init__.py +35 -6
  59. nshtrainer/configs/trainer/_config/__init__.py +33 -6
  60. nshtrainer/configs/trainer/trainer/__init__.py +9 -0
  61. nshtrainer/configs/util/__init__.py +19 -1
  62. nshtrainer/configs/util/_environment_info/__init__.py +14 -0
  63. nshtrainer/configs/util/config/__init__.py +8 -1
  64. nshtrainer/configs/util/config/dtype/__init__.py +4 -0
  65. nshtrainer/configs/util/config/duration/__init__.py +5 -1
  66. nshtrainer/model/mixins/logger.py +30 -12
  67. nshtrainer/trainer/_config.py +40 -21
  68. nshtrainer/trainer/trainer.py +4 -4
  69. {nshtrainer-1.0.0b24.dist-info → nshtrainer-1.0.0b26.dist-info}/METADATA +2 -1
  70. {nshtrainer-1.0.0b24.dist-info → nshtrainer-1.0.0b26.dist-info}/RECORD +71 -69
  71. {nshtrainer-1.0.0b24.dist-info → nshtrainer-1.0.0b26.dist-info}/WHEEL +1 -1
@@ -6,3 +6,8 @@ from nshtrainer.callbacks.timer import CallbackConfigBase as CallbackConfigBase
6
6
  from nshtrainer.callbacks.timer import (
7
7
  EpochTimerCallbackConfig as EpochTimerCallbackConfig,
8
8
  )
9
+
10
+ __all__ = [
11
+ "CallbackConfigBase",
12
+ "EpochTimerCallbackConfig",
13
+ ]
@@ -8,3 +8,8 @@ from nshtrainer.callbacks.wandb_upload_code import (
8
8
  from nshtrainer.callbacks.wandb_upload_code import (
9
9
  WandbUploadCodeCallbackConfig as WandbUploadCodeCallbackConfig,
10
10
  )
11
+
12
+ __all__ = [
13
+ "CallbackConfigBase",
14
+ "WandbUploadCodeCallbackConfig",
15
+ ]
@@ -6,3 +6,8 @@ from nshtrainer.callbacks.wandb_watch import CallbackConfigBase as CallbackConfi
6
6
  from nshtrainer.callbacks.wandb_watch import (
7
7
  WandbWatchCallbackConfig as WandbWatchCallbackConfig,
8
8
  )
9
+
10
+ __all__ = [
11
+ "CallbackConfigBase",
12
+ "WandbWatchCallbackConfig",
13
+ ]
@@ -5,7 +5,6 @@ __codegen__ = True
5
5
  from nshtrainer.loggers import ActSaveLoggerConfig as ActSaveLoggerConfig
6
6
  from nshtrainer.loggers import BaseLoggerConfig as BaseLoggerConfig
7
7
  from nshtrainer.loggers import CSVLoggerConfig as CSVLoggerConfig
8
- from nshtrainer.loggers import LoggerConfig as LoggerConfig
9
8
  from nshtrainer.loggers import TensorboardLoggerConfig as TensorboardLoggerConfig
10
9
  from nshtrainer.loggers import WandbLoggerConfig as WandbLoggerConfig
11
10
  from nshtrainer.loggers.wandb import CallbackConfigBase as CallbackConfigBase
@@ -21,3 +20,19 @@ from . import actsave as actsave
21
20
  from . import csv as csv
22
21
  from . import tensorboard as tensorboard
23
22
  from . import wandb as wandb
23
+
24
+ __all__ = [
25
+ "ActSaveLoggerConfig",
26
+ "BaseLoggerConfig",
27
+ "CSVLoggerConfig",
28
+ "CallbackConfigBase",
29
+ "TensorboardLoggerConfig",
30
+ "WandbLoggerConfig",
31
+ "WandbUploadCodeCallbackConfig",
32
+ "WandbWatchCallbackConfig",
33
+ "_base",
34
+ "actsave",
35
+ "csv",
36
+ "tensorboard",
37
+ "wandb",
38
+ ]
@@ -3,3 +3,7 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.loggers._base import BaseLoggerConfig as BaseLoggerConfig
6
+
7
+ __all__ = [
8
+ "BaseLoggerConfig",
9
+ ]
@@ -4,3 +4,8 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.loggers.actsave import ActSaveLoggerConfig as ActSaveLoggerConfig
6
6
  from nshtrainer.loggers.actsave import BaseLoggerConfig as BaseLoggerConfig
7
+
8
+ __all__ = [
9
+ "ActSaveLoggerConfig",
10
+ "BaseLoggerConfig",
11
+ ]
@@ -4,3 +4,8 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.loggers.csv import BaseLoggerConfig as BaseLoggerConfig
6
6
  from nshtrainer.loggers.csv import CSVLoggerConfig as CSVLoggerConfig
7
+
8
+ __all__ = [
9
+ "BaseLoggerConfig",
10
+ "CSVLoggerConfig",
11
+ ]
@@ -6,3 +6,8 @@ from nshtrainer.loggers.tensorboard import BaseLoggerConfig as BaseLoggerConfig
6
6
  from nshtrainer.loggers.tensorboard import (
7
7
  TensorboardLoggerConfig as TensorboardLoggerConfig,
8
8
  )
9
+
10
+ __all__ = [
11
+ "BaseLoggerConfig",
12
+ "TensorboardLoggerConfig",
13
+ ]
@@ -11,3 +11,11 @@ from nshtrainer.loggers.wandb import (
11
11
  from nshtrainer.loggers.wandb import (
12
12
  WandbWatchCallbackConfig as WandbWatchCallbackConfig,
13
13
  )
14
+
15
+ __all__ = [
16
+ "BaseLoggerConfig",
17
+ "CallbackConfigBase",
18
+ "WandbLoggerConfig",
19
+ "WandbUploadCodeCallbackConfig",
20
+ "WandbWatchCallbackConfig",
21
+ ]
@@ -5,14 +5,20 @@ __codegen__ = True
5
5
  from nshtrainer.lr_scheduler import (
6
6
  LinearWarmupCosineDecayLRSchedulerConfig as LinearWarmupCosineDecayLRSchedulerConfig,
7
7
  )
8
- from nshtrainer.lr_scheduler import LRSchedulerConfig as LRSchedulerConfig
9
8
  from nshtrainer.lr_scheduler import LRSchedulerConfigBase as LRSchedulerConfigBase
10
9
  from nshtrainer.lr_scheduler import ReduceLROnPlateauConfig as ReduceLROnPlateauConfig
11
- from nshtrainer.lr_scheduler.linear_warmup_cosine import (
12
- DurationConfig as DurationConfig,
13
- )
14
10
  from nshtrainer.lr_scheduler.reduce_lr_on_plateau import MetricConfig as MetricConfig
15
11
 
16
12
  from . import _base as _base
17
13
  from . import linear_warmup_cosine as linear_warmup_cosine
18
14
  from . import reduce_lr_on_plateau as reduce_lr_on_plateau
15
+
16
+ __all__ = [
17
+ "LRSchedulerConfigBase",
18
+ "LinearWarmupCosineDecayLRSchedulerConfig",
19
+ "MetricConfig",
20
+ "ReduceLROnPlateauConfig",
21
+ "_base",
22
+ "linear_warmup_cosine",
23
+ "reduce_lr_on_plateau",
24
+ ]
@@ -3,3 +3,7 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.lr_scheduler._base import LRSchedulerConfigBase as LRSchedulerConfigBase
6
+
7
+ __all__ = [
8
+ "LRSchedulerConfigBase",
9
+ ]
@@ -2,12 +2,14 @@ from __future__ import annotations
2
2
 
3
3
  __codegen__ = True
4
4
 
5
- from nshtrainer.lr_scheduler.linear_warmup_cosine import (
6
- DurationConfig as DurationConfig,
7
- )
8
5
  from nshtrainer.lr_scheduler.linear_warmup_cosine import (
9
6
  LinearWarmupCosineDecayLRSchedulerConfig as LinearWarmupCosineDecayLRSchedulerConfig,
10
7
  )
11
8
  from nshtrainer.lr_scheduler.linear_warmup_cosine import (
12
9
  LRSchedulerConfigBase as LRSchedulerConfigBase,
13
10
  )
11
+
12
+ __all__ = [
13
+ "LRSchedulerConfigBase",
14
+ "LinearWarmupCosineDecayLRSchedulerConfig",
15
+ ]
@@ -9,3 +9,9 @@ from nshtrainer.lr_scheduler.reduce_lr_on_plateau import MetricConfig as MetricC
9
9
  from nshtrainer.lr_scheduler.reduce_lr_on_plateau import (
10
10
  ReduceLROnPlateauConfig as ReduceLROnPlateauConfig,
11
11
  )
12
+
13
+ __all__ = [
14
+ "LRSchedulerConfigBase",
15
+ "MetricConfig",
16
+ "ReduceLROnPlateauConfig",
17
+ ]
@@ -5,3 +5,8 @@ __codegen__ = True
5
5
  from nshtrainer.metrics import MetricConfig as MetricConfig
6
6
 
7
7
  from . import _config as _config
8
+
9
+ __all__ = [
10
+ "MetricConfig",
11
+ "_config",
12
+ ]
@@ -3,3 +3,7 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.metrics._config import MetricConfig as MetricConfig
6
+
7
+ __all__ = [
8
+ "MetricConfig",
9
+ ]
@@ -8,7 +8,6 @@ from nshtrainer.nn import GELUNonlinearityConfig as GELUNonlinearityConfig
8
8
  from nshtrainer.nn import LeakyReLUNonlinearityConfig as LeakyReLUNonlinearityConfig
9
9
  from nshtrainer.nn import MishNonlinearityConfig as MishNonlinearityConfig
10
10
  from nshtrainer.nn import MLPConfig as MLPConfig
11
- from nshtrainer.nn import NonlinearityConfig as NonlinearityConfig
12
11
  from nshtrainer.nn import PReLUConfig as PReLUConfig
13
12
  from nshtrainer.nn import ReLUNonlinearityConfig as ReLUNonlinearityConfig
14
13
  from nshtrainer.nn import SigmoidNonlinearityConfig as SigmoidNonlinearityConfig
@@ -24,3 +23,24 @@ from nshtrainer.nn.nonlinearity import (
24
23
 
25
24
  from . import mlp as mlp
26
25
  from . import nonlinearity as nonlinearity
26
+
27
+ __all__ = [
28
+ "BaseNonlinearityConfig",
29
+ "ELUNonlinearityConfig",
30
+ "GELUNonlinearityConfig",
31
+ "LeakyReLUNonlinearityConfig",
32
+ "MLPConfig",
33
+ "MishNonlinearityConfig",
34
+ "PReLUConfig",
35
+ "ReLUNonlinearityConfig",
36
+ "SiLUNonlinearityConfig",
37
+ "SigmoidNonlinearityConfig",
38
+ "SoftmaxNonlinearityConfig",
39
+ "SoftplusNonlinearityConfig",
40
+ "SoftsignNonlinearityConfig",
41
+ "SwiGLUNonlinearityConfig",
42
+ "SwishNonlinearityConfig",
43
+ "TanhNonlinearityConfig",
44
+ "mlp",
45
+ "nonlinearity",
46
+ ]
@@ -4,4 +4,8 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.nn.mlp import BaseNonlinearityConfig as BaseNonlinearityConfig
6
6
  from nshtrainer.nn.mlp import MLPConfig as MLPConfig
7
- from nshtrainer.nn.mlp import NonlinearityConfig as NonlinearityConfig
7
+
8
+ __all__ = [
9
+ "BaseNonlinearityConfig",
10
+ "MLPConfig",
11
+ ]
@@ -9,7 +9,6 @@ from nshtrainer.nn.nonlinearity import (
9
9
  LeakyReLUNonlinearityConfig as LeakyReLUNonlinearityConfig,
10
10
  )
11
11
  from nshtrainer.nn.nonlinearity import MishNonlinearityConfig as MishNonlinearityConfig
12
- from nshtrainer.nn.nonlinearity import NonlinearityConfig as NonlinearityConfig
13
12
  from nshtrainer.nn.nonlinearity import PReLUConfig as PReLUConfig
14
13
  from nshtrainer.nn.nonlinearity import ReLUNonlinearityConfig as ReLUNonlinearityConfig
15
14
  from nshtrainer.nn.nonlinearity import (
@@ -32,3 +31,21 @@ from nshtrainer.nn.nonlinearity import (
32
31
  SwishNonlinearityConfig as SwishNonlinearityConfig,
33
32
  )
34
33
  from nshtrainer.nn.nonlinearity import TanhNonlinearityConfig as TanhNonlinearityConfig
34
+
35
+ __all__ = [
36
+ "BaseNonlinearityConfig",
37
+ "ELUNonlinearityConfig",
38
+ "GELUNonlinearityConfig",
39
+ "LeakyReLUNonlinearityConfig",
40
+ "MishNonlinearityConfig",
41
+ "PReLUConfig",
42
+ "ReLUNonlinearityConfig",
43
+ "SiLUNonlinearityConfig",
44
+ "SigmoidNonlinearityConfig",
45
+ "SoftmaxNonlinearityConfig",
46
+ "SoftplusNonlinearityConfig",
47
+ "SoftsignNonlinearityConfig",
48
+ "SwiGLUNonlinearityConfig",
49
+ "SwishNonlinearityConfig",
50
+ "TanhNonlinearityConfig",
51
+ ]
@@ -3,5 +3,9 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.optimizer import AdamWConfig as AdamWConfig
6
- from nshtrainer.optimizer import OptimizerConfig as OptimizerConfig
7
6
  from nshtrainer.optimizer import OptimizerConfigBase as OptimizerConfigBase
7
+
8
+ __all__ = [
9
+ "AdamWConfig",
10
+ "OptimizerConfigBase",
11
+ ]
@@ -4,7 +4,6 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.profiler import AdvancedProfilerConfig as AdvancedProfilerConfig
6
6
  from nshtrainer.profiler import BaseProfilerConfig as BaseProfilerConfig
7
- from nshtrainer.profiler import ProfilerConfig as ProfilerConfig
8
7
  from nshtrainer.profiler import PyTorchProfilerConfig as PyTorchProfilerConfig
9
8
  from nshtrainer.profiler import SimpleProfilerConfig as SimpleProfilerConfig
10
9
 
@@ -12,3 +11,14 @@ from . import _base as _base
12
11
  from . import advanced as advanced
13
12
  from . import pytorch as pytorch
14
13
  from . import simple as simple
14
+
15
+ __all__ = [
16
+ "AdvancedProfilerConfig",
17
+ "BaseProfilerConfig",
18
+ "PyTorchProfilerConfig",
19
+ "SimpleProfilerConfig",
20
+ "_base",
21
+ "advanced",
22
+ "pytorch",
23
+ "simple",
24
+ ]
@@ -3,3 +3,7 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.profiler._base import BaseProfilerConfig as BaseProfilerConfig
6
+
7
+ __all__ = [
8
+ "BaseProfilerConfig",
9
+ ]
@@ -6,3 +6,8 @@ from nshtrainer.profiler.advanced import (
6
6
  AdvancedProfilerConfig as AdvancedProfilerConfig,
7
7
  )
8
8
  from nshtrainer.profiler.advanced import BaseProfilerConfig as BaseProfilerConfig
9
+
10
+ __all__ = [
11
+ "AdvancedProfilerConfig",
12
+ "BaseProfilerConfig",
13
+ ]
@@ -4,3 +4,8 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.profiler.pytorch import BaseProfilerConfig as BaseProfilerConfig
6
6
  from nshtrainer.profiler.pytorch import PyTorchProfilerConfig as PyTorchProfilerConfig
7
+
8
+ __all__ = [
9
+ "BaseProfilerConfig",
10
+ "PyTorchProfilerConfig",
11
+ ]
@@ -4,3 +4,8 @@ __codegen__ = True
4
4
 
5
5
  from nshtrainer.profiler.simple import BaseProfilerConfig as BaseProfilerConfig
6
6
  from nshtrainer.profiler.simple import SimpleProfilerConfig as SimpleProfilerConfig
7
+
8
+ __all__ = [
9
+ "BaseProfilerConfig",
10
+ "SimpleProfilerConfig",
11
+ ]
@@ -3,16 +3,13 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.trainer import TrainerConfig as TrainerConfig
6
+ from nshtrainer.trainer._config import AcceleratorConfigBase as AcceleratorConfigBase
6
7
  from nshtrainer.trainer._config import ActSaveLoggerConfig as ActSaveLoggerConfig
7
8
  from nshtrainer.trainer._config import BaseLoggerConfig as BaseLoggerConfig
8
9
  from nshtrainer.trainer._config import (
9
10
  BestCheckpointCallbackConfig as BestCheckpointCallbackConfig,
10
11
  )
11
- from nshtrainer.trainer._config import CallbackConfig as CallbackConfig
12
12
  from nshtrainer.trainer._config import CallbackConfigBase as CallbackConfigBase
13
- from nshtrainer.trainer._config import (
14
- CheckpointCallbackConfig as CheckpointCallbackConfig,
15
- )
16
13
  from nshtrainer.trainer._config import CheckpointSavingConfig as CheckpointSavingConfig
17
14
  from nshtrainer.trainer._config import CSVLoggerConfig as CSVLoggerConfig
18
15
  from nshtrainer.trainer._config import (
@@ -32,7 +29,6 @@ from nshtrainer.trainer._config import (
32
29
  LearningRateMonitorConfig as LearningRateMonitorConfig,
33
30
  )
34
31
  from nshtrainer.trainer._config import LogEpochCallbackConfig as LogEpochCallbackConfig
35
- from nshtrainer.trainer._config import LoggerConfig as LoggerConfig
36
32
  from nshtrainer.trainer._config import MetricConfig as MetricConfig
37
33
  from nshtrainer.trainer._config import (
38
34
  NormLoggingCallbackConfig as NormLoggingCallbackConfig,
@@ -40,7 +36,7 @@ from nshtrainer.trainer._config import (
40
36
  from nshtrainer.trainer._config import (
41
37
  OnExceptionCheckpointCallbackConfig as OnExceptionCheckpointCallbackConfig,
42
38
  )
43
- from nshtrainer.trainer._config import ProfilerConfig as ProfilerConfig
39
+ from nshtrainer.trainer._config import PluginConfigBase as PluginConfigBase
44
40
  from nshtrainer.trainer._config import (
45
41
  RLPSanityChecksCallbackConfig as RLPSanityChecksCallbackConfig,
46
42
  )
@@ -48,6 +44,7 @@ from nshtrainer.trainer._config import SanityCheckingConfig as SanityCheckingCon
48
44
  from nshtrainer.trainer._config import (
49
45
  SharedParametersCallbackConfig as SharedParametersCallbackConfig,
50
46
  )
47
+ from nshtrainer.trainer._config import StrategyConfigBase as StrategyConfigBase
51
48
  from nshtrainer.trainer._config import (
52
49
  TensorboardLoggerConfig as TensorboardLoggerConfig,
53
50
  )
@@ -55,3 +52,35 @@ from nshtrainer.trainer._config import WandbLoggerConfig as WandbLoggerConfig
55
52
 
56
53
  from . import _config as _config
57
54
  from . import trainer as trainer
55
+
56
+ __all__ = [
57
+ "AcceleratorConfigBase",
58
+ "ActSaveLoggerConfig",
59
+ "BaseLoggerConfig",
60
+ "BestCheckpointCallbackConfig",
61
+ "CSVLoggerConfig",
62
+ "CallbackConfigBase",
63
+ "CheckpointSavingConfig",
64
+ "DebugFlagCallbackConfig",
65
+ "DirectoryConfig",
66
+ "EarlyStoppingCallbackConfig",
67
+ "EnvironmentConfig",
68
+ "GradientClippingConfig",
69
+ "HuggingFaceHubConfig",
70
+ "LastCheckpointCallbackConfig",
71
+ "LearningRateMonitorConfig",
72
+ "LogEpochCallbackConfig",
73
+ "MetricConfig",
74
+ "NormLoggingCallbackConfig",
75
+ "OnExceptionCheckpointCallbackConfig",
76
+ "PluginConfigBase",
77
+ "RLPSanityChecksCallbackConfig",
78
+ "SanityCheckingConfig",
79
+ "SharedParametersCallbackConfig",
80
+ "StrategyConfigBase",
81
+ "TensorboardLoggerConfig",
82
+ "TrainerConfig",
83
+ "WandbLoggerConfig",
84
+ "_config",
85
+ "trainer",
86
+ ]
@@ -2,16 +2,13 @@ from __future__ import annotations
2
2
 
3
3
  __codegen__ = True
4
4
 
5
+ from nshtrainer.trainer._config import AcceleratorConfigBase as AcceleratorConfigBase
5
6
  from nshtrainer.trainer._config import ActSaveLoggerConfig as ActSaveLoggerConfig
6
7
  from nshtrainer.trainer._config import BaseLoggerConfig as BaseLoggerConfig
7
8
  from nshtrainer.trainer._config import (
8
9
  BestCheckpointCallbackConfig as BestCheckpointCallbackConfig,
9
10
  )
10
- from nshtrainer.trainer._config import CallbackConfig as CallbackConfig
11
11
  from nshtrainer.trainer._config import CallbackConfigBase as CallbackConfigBase
12
- from nshtrainer.trainer._config import (
13
- CheckpointCallbackConfig as CheckpointCallbackConfig,
14
- )
15
12
  from nshtrainer.trainer._config import CheckpointSavingConfig as CheckpointSavingConfig
16
13
  from nshtrainer.trainer._config import CSVLoggerConfig as CSVLoggerConfig
17
14
  from nshtrainer.trainer._config import (
@@ -31,7 +28,6 @@ from nshtrainer.trainer._config import (
31
28
  LearningRateMonitorConfig as LearningRateMonitorConfig,
32
29
  )
33
30
  from nshtrainer.trainer._config import LogEpochCallbackConfig as LogEpochCallbackConfig
34
- from nshtrainer.trainer._config import LoggerConfig as LoggerConfig
35
31
  from nshtrainer.trainer._config import MetricConfig as MetricConfig
36
32
  from nshtrainer.trainer._config import (
37
33
  NormLoggingCallbackConfig as NormLoggingCallbackConfig,
@@ -39,7 +35,7 @@ from nshtrainer.trainer._config import (
39
35
  from nshtrainer.trainer._config import (
40
36
  OnExceptionCheckpointCallbackConfig as OnExceptionCheckpointCallbackConfig,
41
37
  )
42
- from nshtrainer.trainer._config import ProfilerConfig as ProfilerConfig
38
+ from nshtrainer.trainer._config import PluginConfigBase as PluginConfigBase
43
39
  from nshtrainer.trainer._config import (
44
40
  RLPSanityChecksCallbackConfig as RLPSanityChecksCallbackConfig,
45
41
  )
@@ -47,8 +43,39 @@ from nshtrainer.trainer._config import SanityCheckingConfig as SanityCheckingCon
47
43
  from nshtrainer.trainer._config import (
48
44
  SharedParametersCallbackConfig as SharedParametersCallbackConfig,
49
45
  )
46
+ from nshtrainer.trainer._config import StrategyConfigBase as StrategyConfigBase
50
47
  from nshtrainer.trainer._config import (
51
48
  TensorboardLoggerConfig as TensorboardLoggerConfig,
52
49
  )
53
50
  from nshtrainer.trainer._config import TrainerConfig as TrainerConfig
54
51
  from nshtrainer.trainer._config import WandbLoggerConfig as WandbLoggerConfig
52
+
53
+ __all__ = [
54
+ "AcceleratorConfigBase",
55
+ "ActSaveLoggerConfig",
56
+ "BaseLoggerConfig",
57
+ "BestCheckpointCallbackConfig",
58
+ "CSVLoggerConfig",
59
+ "CallbackConfigBase",
60
+ "CheckpointSavingConfig",
61
+ "DebugFlagCallbackConfig",
62
+ "DirectoryConfig",
63
+ "EarlyStoppingCallbackConfig",
64
+ "EnvironmentConfig",
65
+ "GradientClippingConfig",
66
+ "HuggingFaceHubConfig",
67
+ "LastCheckpointCallbackConfig",
68
+ "LearningRateMonitorConfig",
69
+ "LogEpochCallbackConfig",
70
+ "MetricConfig",
71
+ "NormLoggingCallbackConfig",
72
+ "OnExceptionCheckpointCallbackConfig",
73
+ "PluginConfigBase",
74
+ "RLPSanityChecksCallbackConfig",
75
+ "SanityCheckingConfig",
76
+ "SharedParametersCallbackConfig",
77
+ "StrategyConfigBase",
78
+ "TensorboardLoggerConfig",
79
+ "TrainerConfig",
80
+ "WandbLoggerConfig",
81
+ ]
@@ -2,5 +2,14 @@ from __future__ import annotations
2
2
 
3
3
  __codegen__ = True
4
4
 
5
+ from nshtrainer.trainer.trainer import AcceleratorConfigBase as AcceleratorConfigBase
5
6
  from nshtrainer.trainer.trainer import EnvironmentConfig as EnvironmentConfig
7
+ from nshtrainer.trainer.trainer import StrategyConfigBase as StrategyConfigBase
6
8
  from nshtrainer.trainer.trainer import TrainerConfig as TrainerConfig
9
+
10
+ __all__ = [
11
+ "AcceleratorConfigBase",
12
+ "EnvironmentConfig",
13
+ "StrategyConfigBase",
14
+ "TrainerConfig",
15
+ ]
@@ -32,9 +32,27 @@ from nshtrainer.util._environment_info import (
32
32
  )
33
33
  from nshtrainer.util._environment_info import GitRepositoryConfig as GitRepositoryConfig
34
34
  from nshtrainer.util.config import DTypeConfig as DTypeConfig
35
- from nshtrainer.util.config import DurationConfig as DurationConfig
36
35
  from nshtrainer.util.config import EpochsConfig as EpochsConfig
37
36
  from nshtrainer.util.config import StepsConfig as StepsConfig
38
37
 
39
38
  from . import _environment_info as _environment_info
40
39
  from . import config as config
40
+
41
+ __all__ = [
42
+ "DTypeConfig",
43
+ "EnvironmentCUDAConfig",
44
+ "EnvironmentClassInformationConfig",
45
+ "EnvironmentConfig",
46
+ "EnvironmentGPUConfig",
47
+ "EnvironmentHardwareConfig",
48
+ "EnvironmentLSFInformationConfig",
49
+ "EnvironmentLinuxEnvironmentConfig",
50
+ "EnvironmentPackageConfig",
51
+ "EnvironmentSLURMInformationConfig",
52
+ "EnvironmentSnapshotConfig",
53
+ "EpochsConfig",
54
+ "GitRepositoryConfig",
55
+ "StepsConfig",
56
+ "_environment_info",
57
+ "config",
58
+ ]
@@ -31,3 +31,17 @@ from nshtrainer.util._environment_info import (
31
31
  EnvironmentSnapshotConfig as EnvironmentSnapshotConfig,
32
32
  )
33
33
  from nshtrainer.util._environment_info import GitRepositoryConfig as GitRepositoryConfig
34
+
35
+ __all__ = [
36
+ "EnvironmentCUDAConfig",
37
+ "EnvironmentClassInformationConfig",
38
+ "EnvironmentConfig",
39
+ "EnvironmentGPUConfig",
40
+ "EnvironmentHardwareConfig",
41
+ "EnvironmentLSFInformationConfig",
42
+ "EnvironmentLinuxEnvironmentConfig",
43
+ "EnvironmentPackageConfig",
44
+ "EnvironmentSLURMInformationConfig",
45
+ "EnvironmentSnapshotConfig",
46
+ "GitRepositoryConfig",
47
+ ]
@@ -3,9 +3,16 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.util.config import DTypeConfig as DTypeConfig
6
- from nshtrainer.util.config import DurationConfig as DurationConfig
7
6
  from nshtrainer.util.config import EpochsConfig as EpochsConfig
8
7
  from nshtrainer.util.config import StepsConfig as StepsConfig
9
8
 
10
9
  from . import dtype as dtype
11
10
  from . import duration as duration
11
+
12
+ __all__ = [
13
+ "DTypeConfig",
14
+ "EpochsConfig",
15
+ "StepsConfig",
16
+ "dtype",
17
+ "duration",
18
+ ]
@@ -3,3 +3,7 @@ from __future__ import annotations
3
3
  __codegen__ = True
4
4
 
5
5
  from nshtrainer.util.config.dtype import DTypeConfig as DTypeConfig
6
+
7
+ __all__ = [
8
+ "DTypeConfig",
9
+ ]
@@ -2,6 +2,10 @@ from __future__ import annotations
2
2
 
3
3
  __codegen__ = True
4
4
 
5
- from nshtrainer.util.config.duration import DurationConfig as DurationConfig
6
5
  from nshtrainer.util.config.duration import EpochsConfig as EpochsConfig
7
6
  from nshtrainer.util.config.duration import StepsConfig as StepsConfig
7
+
8
+ __all__ = [
9
+ "EpochsConfig",
10
+ "StepsConfig",
11
+ ]
@@ -64,6 +64,32 @@ class LoggerLightningModuleMixin(mixin_base_type(LightningModule)):
64
64
 
65
65
  self._logger_prefix_stack = deque[_LogContextKwargs]()
66
66
 
67
+ @property
68
+ def logging_enabled(self) -> bool:
69
+ # Logging is disabled in barebones mode.
70
+ if (trainer := self._trainer) is not None and trainer.barebones:
71
+ # Warn the user once that logging is disabled in barebones mode.
72
+ if not hasattr(self, "_barebones_logging_warned"):
73
+ rank_zero_warn(
74
+ "Logging is disabled in barebones mode. "
75
+ "This is to reduce the overhead of logging in barebones mode. "
76
+ "If you want to enable logging, set `barebones=False` in the Trainer.",
77
+ )
78
+ self._barebones_logging_warned = True
79
+ return False
80
+
81
+ # If no loggers are registered, then logging is disabled.
82
+ if not self.logger:
83
+ return False
84
+
85
+ # Check if the topmost non-null context is disabled
86
+ for context in reversed(self._logger_prefix_stack):
87
+ if context.disabled is not None:
88
+ return not context.disabled
89
+
90
+ # Otherwise, logging is enabled.
91
+ return True
92
+
67
93
  @contextmanager
68
94
  def log_context(
69
95
  self,
@@ -121,22 +147,14 @@ class LoggerLightningModuleMixin(mixin_base_type(LightningModule)):
121
147
  metric_attribute: str | None = None,
122
148
  rank_zero_only: bool = False,
123
149
  ) -> None:
150
+ # If logging is disabled, then do nothing.
151
+ if not self.logging_enabled:
152
+ return
153
+
124
154
  # join all prefixes
125
155
  prefix = "".join(c.prefix for c in self._logger_prefix_stack if c.prefix)
126
156
  name = f"{prefix}{name}"
127
157
 
128
- # check for disabled context:
129
- # if the topmost non-null context is disabled, then we don't log
130
- for c in reversed(self._logger_prefix_stack):
131
- if c.disabled is not None:
132
- if c.disabled:
133
- rank_zero_warn(
134
- f"Skipping logging of {name} due to disabled context"
135
- )
136
- return
137
- else:
138
- break
139
-
140
158
  fn_kwargs = _LogContextKwargs()
141
159
  for c in self._logger_prefix_stack:
142
160
  fn_kwargs = fn_kwargs.copy_from(c)