Skip to content

classifier

Classifier(module, loss_fn='binary_cross_entropy_with_logits', optimizer_fn='sgd', lr=0.001, output_is_logit=True, is_class_incremental=False, device='cpu', seed=42, **kwargs)

Bases: DeepEstimator, MiniBatchClassifier

Wrapper for PyTorch classification models that automatically handles increases in the number of classes by adding output neurons in case the number of observed classes exceeds the current number of output neurons.

PARAMETER DESCRIPTION
module

Torch Module that builds the autoencoder to be wrapped. The Module should accept parameter n_features so that the returned model's input shape can be determined based on the number of features in the initial training example.

TYPE: Type[Module]

loss_fn

Loss function to be used for training the wrapped model. Can be a loss function provided by torch.nn.functional or one of the following: 'mse', 'l1', 'cross_entropy', 'binary_cross_entropy_with_logits', 'binary_crossentropy', 'smooth_l1', 'kl_div'.

TYPE: Union[str, Callable] DEFAULT: 'binary_cross_entropy_with_logits'

optimizer_fn

Optimizer to be used for training the wrapped model. Can be an optimizer class provided by torch.optim or one of the following: "adam", "adam_w", "sgd", "rmsprop", "lbfgs".

TYPE: Union[str, Callable] DEFAULT: 'sgd'

lr

Learning rate of the optimizer.

TYPE: float DEFAULT: 0.001

output_is_logit

Whether the module produces logits as output. If true, either softmax or sigmoid is applied to the outputs when predicting.

TYPE: bool DEFAULT: True

is_class_incremental

Whether the classifier should adapt to the appearance of previously unobserved classes by adding an unit to the output layer of the network. This works only if the last trainable layer is an nn.Linear layer. Note also, that output activation functions can not be adapted, meaning that a binary classifier with a sigmoid output can not be altered to perform multi-class predictions.

TYPE: bool DEFAULT: False

device

Device to run the wrapped model on. Can be "cpu" or "cuda".

TYPE: str DEFAULT: 'cpu'

seed

Random seed to be used for training the wrapped model.

TYPE: int DEFAULT: 42

**kwargs

Parameters to be passed to the build_fn function aside from n_features.

DEFAULT: {}

Examples:

>>> from river import metrics, preprocessing, compose, datasets
>>> from deep_river import classification
>>> from torch import nn
>>> from torch import manual_seed
>>> _ = manual_seed(42)
>>> class MyModule(nn.Module):
...     def __init__(self, n_features):
...         super(MyModule, self).__init__()
...         self.dense0 = nn.Linear(n_features,5)
...         self.nonlin = nn.ReLU()
...         self.dense1 = nn.Linear(5, 2)
...         self.softmax = nn.Softmax(dim=-1)
...
...     def forward(self, X, **kwargs):
...         X = self.nonlin(self.dense0(X))
...         X = self.nonlin(self.dense1(X))
...         X = self.softmax(X)
...         return X
>>> model_pipeline = compose.Pipeline(
...     preprocessing.StandardScaler,
...     Classifier(module=MyModule,
...                loss_fn="binary_cross_entropy",
...                optimizer_fn='adam')
... )
>>> dataset = datasets.Phishing()
>>> metric = metrics.Accuracy()
>>> for x, y in dataset:
...     y_pred = model_pipeline.predict_one(x)  # make a prediction
...     metric = metric.update(y, y_pred)  # update the metric
...     model_pipeline = model_pipeline.learn_one(x,y)
>>> print(f'Accuracy: {metric.get()}')
Accuracy: 0.6728

learn_many(X, y)

Performs one step of training with a batch of examples.

PARAMETER DESCRIPTION
X

Input examples.

TYPE: DataFrame

y

Target values.

TYPE: Series

RETURNS DESCRIPTION
Classifier

The classifier itself.

learn_one(x, y, **kwargs)

Performs one step of training with a single example.

PARAMETER DESCRIPTION
x

Input example.

TYPE: dict

y

Target value.

TYPE: ClfTarget

RETURNS DESCRIPTION
Classifier

The classifier itself.

predict_proba_many(X)

Predict the probability of each label given the input.

PARAMETER DESCRIPTION
X

Input examples.

TYPE: DataFrame

RETURNS DESCRIPTION
DataFrame

DataFrame of probabilities for each label.

predict_proba_one(x)

Predict the probability of each label given the input.

PARAMETER DESCRIPTION
x

Input example.

TYPE: dict

RETURNS DESCRIPTION
Dict[ClfTarget, float]

Dictionary of probabilities for each label.