Time Series Classification

Perform time series classification in Amazon SageMaker

SageMaker Algorithm

CPU Training

GPU Training

Multi-GPU Training

Incremental Training

LSTM-FCN

InceptionTime

CNN-SVC

LSTM-FCN SageMaker Algorithm

The LSTM-FCN SageMaker Algorithm performs time series classification with the Long Short-Term Memory Fully Convolutional Network (LSTM-FCN). The LSTM-FCN model consists of two blocks: a recurrent block and a convolutional block. The two blocks process the input time series in parallel. After that their output representations are concatenated and passed to a final output layer with softmax activation. For additional information, see the algorithm's AWS Marketplace listing page and GitHub repository.

InceptionTime SageMaker Algorithm

The InceptionTime SageMaker Algorithm performs time series classification with the InceptionTime Network. The InceptionTime network consists of a stack of Inception blocks linked by residual connections. Each block contains three convolutional layers and a max pooling layer. The four layers process the block input in parallel. After that their output representations are concatenated and passed to a batch normalization layer followed by a dense layer with ReLU activation. The output of the last block is passed to an average pooling layer, and then to a final output layer with softmax activation. The algorithm trains an ensemble of InceptionTime networks and generates the final predicted class labels by averaging the class probabilities predicted by the different networks in the ensemble. For additional information, see the algorithm's AWS Marketplace listing page and GitHub repository.

CNN-SVC SageMaker Algorithm

The CNN-SVC SageMaker Algorithm performs time series classification with an unsupervised convolutional neural network (CNN) followed by a support vector classifier (SVC). The CNN network encodes the input time series into a number of time-independent features, which are then used as inputs by the SVC algorithm. The CNN network consists of a stack of exponentially dilated causal convolutional blocks with residual connections and is trained in an unsupervised manner using a contrastive learning procedure that minimizes the triplet loss. The algorithm can be used for time series with different lengths and with missing values. The algorithm also supports missing class labels. For additional information, see the algorithm's AWS Marketplace listing page and GitHub repository.