Source code for FCNClassifier_module

from typing import Any, Dict, List

import numpy as np
import keras
from aeon.classification.deep_learning import FCNClassifier
from skopt.space import Categorical
from sklearn.pipeline import Pipeline
from sklearn.impute import SimpleImputer
from sklearn.preprocessing import StandardScaler
from sklearn.base import BaseEstimator, TransformerMixin

from ml_grid.pipeline.data import pipe
from ml_grid.util.param_space import ParamSpace


[docs] class TimeSeriesStandardScaler(BaseEstimator, TransformerMixin): def __init__(self, epsilon=1e-6):
[docs] self.scaler = StandardScaler()
[docs] self.imputer = SimpleImputer(strategy="median")
[docs] self.epsilon = epsilon
[docs] def fit(self, X, y=None): n_samples, n_dims, n_timesteps = X.shape # Reshape to (n_samples * n_timesteps, n_dims) to scale each dimension X_reshaped = X.transpose(0, 2, 1).reshape(-1, n_dims) # Handle infs/nans before anything else X_reshaped = np.where(np.isinf(X_reshaped), np.nan, X_reshaped) # Impute missing values before scaling X_imputed = self.imputer.fit_transform(X_reshaped) self.scaler.fit(X_imputed) # If scale is near zero (constant feature), use 1.0 to avoid division by zero and value explosion self.scaler.scale_ = np.where( self.scaler.scale_ < self.epsilon, 1.0, self.scaler.scale_ ) return self
[docs] def transform(self, X): n_samples, n_dims, n_timesteps = X.shape X_reshaped = X.transpose(0, 2, 1).reshape(-1, n_dims) # Handle infs/nans X_reshaped = np.where(np.isinf(X_reshaped), np.nan, X_reshaped) X_imputed = self.imputer.transform(X_reshaped) X_scaled_reshaped = self.scaler.transform(X_imputed) # Clip to prevent extreme values (e.g. > 20 sigma) which cause gradient explosion X_scaled_reshaped = np.clip(X_scaled_reshaped, -20, 20) # Reshape back to original 3D format X_scaled = X_scaled_reshaped.reshape(n_samples, n_timesteps, n_dims).transpose( 0, 2, 1 ) return X_scaled
[docs] class FCNClassifierWrapper(FCNClassifier):
[docs] def fit(self, X, y, **kwargs): if isinstance(self.kernel_size, tuple): self.kernel_size = list(self.kernel_size) if isinstance(self.n_filters, tuple): self.n_filters = list(self.n_filters) return super().fit(X, y, **kwargs)
def _fit(self, X, y): if self.metrics is None: self._metrics = [] elif isinstance(self.metrics, str): self._metrics = [self.metrics] else: self._metrics = self.metrics # Clone optimizer to ensure a fresh instance for each fit/fold if hasattr(self, "optimizer") and isinstance( self.optimizer, keras.optimizers.Optimizer ): self.optimizer = self.optimizer.from_config(self.optimizer.get_config()) return super()._fit(X, y) def _predict_proba(self, X, **kwargs): # Safety net: intercept NaNs in probability outputs proba = super()._predict_proba(X, **kwargs) if np.isnan(proba).any(): # Replace NaNs with uniform probability (1/n_classes) n_classes = len(self.classes_) uniform_prob = 1.0 / n_classes proba = np.where(np.isnan(proba), uniform_prob, proba) # Normalize rows to sum to 1 row_sums = proba.sum(axis=1) # Avoid division by zero if a row is all zeros row_sums[row_sums == 0] = 1 proba = proba / row_sums[:, np.newaxis] return proba def _predict(self, X, **kwargs): # Use our safe predict_proba to generate predictions probs = self._predict_proba(X, **kwargs) return np.array([self.classes_[np.argmax(prob)] for prob in probs])
[docs] class FCNClassifier_class: """A wrapper for the aeon FCNClassifier time-series classifier. This class provides a consistent interface for the FCNClassifier, including defining a hyperparameter search space. Attributes: algorithm_implementation: An instance of the aeon FCNClassifier. method_name (str): The name of the classifier method. parameter_space (Dict[str, List[Any]]): The hyperparameter search space for the classifier. """
[docs] algorithm_implementation: FCNClassifier
[docs] method_name: str
[docs] parameter_space: Dict[str, List[Any]]
def __init__(self, ml_grid_object: pipe): """Initializes the FCNClassifier_class. Args: ml_grid_object (pipe): An instance of the main data pipeline object. """ random_state_val = ml_grid_object.global_params.random_state_val verbose_param = ml_grid_object.verbose param_space = ParamSpace( ml_grid_object.local_param_dict.get("param_space_size") ) log_epoch = param_space.param_dict.get("log_epoch") if isinstance(log_epoch, list): log_epoch = log_epoch[0] fcn_model = FCNClassifierWrapper() self.algorithm_implementation = Pipeline( [("scaler", TimeSeriesStandardScaler()), ("model", fcn_model)] ) self.method_name = "FCNClassifier" if getattr(ml_grid_object.global_params, "test_mode", False): self.parameter_space = { "model__n_epochs": [1], "model__n_layers": [1], "model__n_filters": [(32,)], "model__kernel_size": [(3,)], "model__verbose": [0], "model__optimizer": [keras.optimizers.Adam(learning_rate=0.001)], } return if ml_grid_object.global_params.bayessearch: base_params = { "n_layers": [3], "n_filters": Categorical( [(128, 128, 128), (256, 256, 256), (128, 256, 128)] ), "kernel_size": Categorical( [(8, 8, 8), (5, 5, 5), (3, 3, 3), (8, 5, 3)] ), "dilation_rate": Categorical([1]), "strides": Categorical([1]), "padding": Categorical(["same"]), "activation": Categorical(["relu"]), "use_bias": Categorical([True]), "n_epochs": log_epoch, "batch_size": Categorical([16]), "use_mini_batch_size": Categorical([False]), "random_state": Categorical([random_state_val]), "verbose": Categorical([verbose_param]), "loss": Categorical(["categorical_crossentropy"]), "metrics": Categorical([None]), "optimizer": Categorical( [ keras.optimizers.Adam(learning_rate=0.00001, clipnorm=1.0), keras.optimizers.SGD(learning_rate=0.00001, clipnorm=1.0), ] ), } self.parameter_space = { f"model__{key}": value for key, value in base_params.items() } else: base_params = { "n_layers": [3], "n_filters": [(128, 256, 128)], "kernel_size": [(8, 5, 3)], "dilation_rate": [1], "strides": [1], "padding": ["same"], "activation": ["relu"], "use_bias": [True], "n_epochs": [100], "batch_size": [16], "use_mini_batch_size": [False], "random_state": [random_state_val], "verbose": [verbose_param], "loss": ["categorical_crossentropy"], "metrics": [None], "optimizer": [ keras.optimizers.Adam(learning_rate=0.00001, clipnorm=1.0), keras.optimizers.SGD(learning_rate=0.00001, clipnorm=1.0), ], } self.parameter_space = { f"model__{key}": value for key, value in base_params.items() }