Skip to content

Commit

Permalink
core renaming (#835)
Browse files Browse the repository at this point in the history
  • Loading branch information
Scitator authored Jun 4, 2020
1 parent 0c08738 commit a485633
Show file tree
Hide file tree
Showing 45 changed files with 286 additions and 286 deletions.
8 changes: 4 additions & 4 deletions catalyst/contrib/dl/callbacks/alchemy_logger.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
CallbackOrder,
CallbackScope,
)
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner


class AlchemyLogger(Callback):
Expand Down Expand Up @@ -101,7 +101,7 @@ def _log_metrics(
name=metric_name, value=metric_value, step=step,
)

def on_batch_end(self, runner: _Runner):
def on_batch_end(self, runner: IRunner):
"""Translate batch metrics to Alchemy."""
if self.log_on_batch_end:
mode = runner.loader_name
Expand All @@ -113,7 +113,7 @@ def on_batch_end(self, runner: _Runner):
suffix=self.batch_log_suffix,
)

def on_loader_end(self, runner: _Runner):
def on_loader_end(self, runner: IRunner):
"""Translate loader metrics to Alchemy."""
if self.log_on_epoch_end:
mode = runner.loader_name
Expand All @@ -125,7 +125,7 @@ def on_loader_end(self, runner: _Runner):
suffix=self.epoch_log_suffix,
)

def on_epoch_end(self, runner: _Runner):
def on_epoch_end(self, runner: IRunner):
"""Translate epoch metrics to Alchemy."""
extra_mode = "_base"
splitted_epoch_metrics = utils.split_dict_to_subdicts(
Expand Down
14 changes: 7 additions & 7 deletions catalyst/contrib/dl/callbacks/cutmix_callback.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
import torch

from catalyst.core.callbacks import CriterionCallback
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner


class CutmixCallback(CriterionCallback):
Expand Down Expand Up @@ -52,14 +52,14 @@ def __init__(
self.index = None
self.is_needed = True

def _compute_loss(self, runner: _Runner, criterion):
def _compute_loss(self, runner: IRunner, criterion):
"""Computes loss.
If self.is_needed is ``False`` then calls ``_compute_loss``
from ``CriterionCallback``, otherwise computes loss value.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
criterion: that is used to compute loss
"""
if not self.is_needed:
Expand Down Expand Up @@ -101,19 +101,19 @@ def _rand_bbox(self, size, lam):

return bbx1, bby1, bbx2, bby2

def on_loader_start(self, runner: _Runner) -> None:
def on_loader_start(self, runner: IRunner) -> None:
"""Checks if it is needed for the loader.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
self.is_needed = not self.on_train_only or runner.is_train_loader

def on_batch_start(self, runner: _Runner) -> None:
def on_batch_start(self, runner: IRunner) -> None:
"""Mixes data according to Cutmix algorithm.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
if not self.is_needed:
return
Expand Down
6 changes: 3 additions & 3 deletions catalyst/contrib/dl/callbacks/gradnorm_logger.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
from torch.nn.parallel import DistributedDataParallel

from catalyst.core.callback import Callback, CallbackNode, CallbackOrder
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner
from catalyst.tools.typing import Model


Expand Down Expand Up @@ -65,11 +65,11 @@ def grad_norm(*, model: Model, prefix: str, norm_type: int,) -> Dict:

return grad_norm

def on_batch_end(self, runner: _Runner) -> None:
def on_batch_end(self, runner: IRunner) -> None:
"""On batch end event
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
if not runner.is_train_loader:
return
Expand Down
14 changes: 7 additions & 7 deletions catalyst/contrib/dl/callbacks/knn_metric.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
import torch

from catalyst.core.callback import Callback, CallbackOrder
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner


class KNNMetricCallback(Callback):
Expand Down Expand Up @@ -165,11 +165,11 @@ def _knn(self, train_set, test_set=None):

return result

def on_batch_end(self, runner: _Runner) -> None:
def on_batch_end(self, runner: IRunner) -> None:
"""Batch end hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
features: torch.Tensor = runner.output[
self.features_key
Expand All @@ -181,11 +181,11 @@ def on_batch_end(self, runner: _Runner) -> None:
self.features.extend(features)
self.targets.extend(targets)

def on_loader_end(self, runner: _Runner) -> None:
def on_loader_end(self, runner: IRunner) -> None:
"""Loader end hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
self.features = np.stack(self.features)
self.targets = np.stack(self.targets)
Expand Down Expand Up @@ -216,11 +216,11 @@ def on_loader_end(self, runner: _Runner) -> None:

self._reset_cache()

def on_epoch_end(self, runner: _Runner) -> None:
def on_epoch_end(self, runner: IRunner) -> None:
"""Epoch end hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
if self.cv_loader_names is not None:
for k, vs in self.cv_loader_names.items():
Expand Down
14 changes: 7 additions & 7 deletions catalyst/contrib/dl/callbacks/mask_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
import torch.nn.functional as F

from catalyst.core.callback import Callback, CallbackOrder
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner
from catalyst.dl import utils


Expand Down Expand Up @@ -48,11 +48,11 @@ def __init__(
self.counter = 0
self._keys_from_runner = ["out_dir", "out_prefix"]

def on_stage_start(self, runner: _Runner):
def on_stage_start(self, runner: IRunner):
"""Stage start hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
for key in self._keys_from_runner:
value = getattr(runner, key, None)
Expand All @@ -66,20 +66,20 @@ def on_stage_start(self, runner: _Runner):
self.out_prefix = str(self.out_dir) + "/" + str(self.out_prefix)
os.makedirs(os.path.dirname(self.out_prefix), exist_ok=True)

def on_loader_start(self, runner: _Runner):
def on_loader_start(self, runner: IRunner):
"""Loader start hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
lm = runner.loader_name
os.makedirs(f"{self.out_prefix}/{lm}/", exist_ok=True)

def on_batch_end(self, runner: _Runner):
def on_batch_end(self, runner: IRunner):
"""Batch end hook.
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
lm = runner.loader_name
names = runner.input.get(self.name_key, [])
Expand Down
6 changes: 3 additions & 3 deletions catalyst/contrib/dl/callbacks/neptune_logger.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
CallbackOrder,
CallbackScope,
)
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner


class NeptuneLogger(Callback):
Expand Down Expand Up @@ -135,7 +135,7 @@ def _log_metrics(
metric_value = metrics[name]
self.experiment.log_metric(metric_name, y=metric_value, x=step)

def on_batch_end(self, runner: _Runner):
def on_batch_end(self, runner: IRunner):
"""Log batch metrics to Neptune."""
if self.log_on_batch_end:
mode = runner.loader_name
Expand All @@ -147,7 +147,7 @@ def on_batch_end(self, runner: _Runner):
suffix=self.batch_log_suffix,
)

def on_loader_end(self, runner: _Runner):
def on_loader_end(self, runner: IRunner):
"""Translate epoch metrics to Neptune."""
if self.log_on_epoch_end:
mode = runner.loader_name
Expand Down
14 changes: 7 additions & 7 deletions catalyst/contrib/dl/callbacks/periodic_loader_callback.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
from torch.utils.data import DataLoader

from catalyst.core.callback import Callback, CallbackOrder
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner


class PeriodicLoaderCallback(Callback):
Expand Down Expand Up @@ -41,11 +41,11 @@ def __init__(self, **kwargs):
)
self.loader_periods[loader] = int(period)

def on_stage_start(self, runner: _Runner) -> None:
def on_stage_start(self, runner: IRunner) -> None:
"""Collect information about loaders.
Arguments:
runner (_Runner): current runner
runner (IRunner): current runner
"""
# store pointers to data loader objects
for name, loader in runner.loaders.items():
Expand Down Expand Up @@ -75,7 +75,7 @@ def on_stage_start(self, runner: _Runner) -> None:
f"There will be no loaders in epoch {epoch_with_err}!"
)

def on_epoch_start(self, runner: _Runner) -> None:
def on_epoch_start(self, runner: IRunner) -> None:
"""Set loaders for current epoch.
If validation is not required then the first loader
from loaders used in current epoch will be used
Expand All @@ -85,7 +85,7 @@ def on_epoch_start(self, runner: _Runner) -> None:
in the epochs where this loader is missing.
Arguments:
runner (_Runner): current runner
runner (IRunner): current runner
"""
epoch_num = runner.epoch
# loaders to use in current epoch
Expand All @@ -105,12 +105,12 @@ def on_epoch_start(self, runner: _Runner) -> None:
)
runner.loaders = epoch_loaders

def on_epoch_end(self, runner: _Runner) -> None:
def on_epoch_end(self, runner: IRunner) -> None:
"""Store validation metrics and use latest validation score
when validation loader is not required.
Arguments:
runner (_Runner): current runner
runner (IRunner): current runner
"""
if self.valid_loader in runner.loaders:
self.valid_metrics = {
Expand Down
12 changes: 6 additions & 6 deletions catalyst/contrib/dl/callbacks/telegram_logger.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

from catalyst import utils
from catalyst.core.callback import Callback, CallbackNode, CallbackOrder
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner
from catalyst.tools import settings


Expand Down Expand Up @@ -69,14 +69,14 @@ def _send_text(self, text: str):
except Exception as e:
logging.getLogger(__name__).warning(f"telegram.send.error:{e}")

def on_stage_start(self, runner: _Runner):
def on_stage_start(self, runner: IRunner):
"""Notify about starting a new stage."""
if self.log_on_stage_start:
text = f"{runner.stage_name} stage was started"

self._send_text(text)

def on_loader_start(self, runner: _Runner):
def on_loader_start(self, runner: IRunner):
"""Notify about starting running the new loader."""
if self.log_on_loader_start:
text = (
Expand All @@ -85,7 +85,7 @@ def on_loader_start(self, runner: _Runner):

self._send_text(text)

def on_loader_end(self, runner: _Runner):
def on_loader_end(self, runner: IRunner):
"""Translate ``runner.metric_manager`` to telegram channel."""
if self.log_on_loader_end:
metrics = runner.loader_metrics
Expand All @@ -108,14 +108,14 @@ def on_loader_end(self, runner: _Runner):

self._send_text(text)

def on_stage_end(self, runner: _Runner):
def on_stage_end(self, runner: IRunner):
"""Notify about finishing a stage."""
if self.log_on_stage_end:
text = f"{runner.stage_name} stage was finished"

self._send_text(text)

def on_exception(self, runner: _Runner):
def on_exception(self, runner: IRunner):
"""Notify about raised ``Exception``."""
if self.log_on_exception:
exception = runner.exception
Expand Down
4 changes: 2 additions & 2 deletions catalyst/contrib/dl/callbacks/tests/test_gradnorm_logger.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
from catalyst.contrib.dl.callbacks.gradnorm_logger import GradNormLogger
from catalyst.core.callback import Callback, CallbackOrder
from catalyst.core.callbacks import CriterionCallback, OptimizerCallback
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner
from catalyst.dl import SupervisedRunner


Expand Down Expand Up @@ -103,7 +103,7 @@ def __init__(self, prefix: str):
super().__init__(CallbackOrder.External)
self.prefix = prefix

def on_batch_end(self, runner: _Runner):
def on_batch_end(self, runner: IRunner):
if not runner.is_train_loader:
return

Expand Down
6 changes: 3 additions & 3 deletions catalyst/contrib/dl/callbacks/tests/test_tracer_callback.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
from catalyst.contrib.dl.callbacks.tracer_callback import TracerCallback
from catalyst.core.callback import Callback, CallbackOrder
from catalyst.core.callbacks import CriterionCallback, OptimizerCallback
from catalyst.core.runner import _Runner
from catalyst.core.runner import IRunner
from catalyst.dl import SupervisedRunner
from catalyst.dl.utils import get_device, get_trace_name

Expand Down Expand Up @@ -144,10 +144,10 @@ def __init__(self, path: Union[str, Path], inputs: torch.Tensor):
self.inputs: torch.Tensor = inputs
self.device = get_device()

def on_stage_end(self, runner: _Runner):
def on_stage_end(self, runner: IRunner):
"""
Args:
runner (_Runner): current runner
runner (IRunner): current runner
"""
assert self.path.exists(), "Traced model was not found"

Expand Down
Loading

0 comments on commit a485633

Please sign in to comment.