Making Time Series Embeddings More Interpretable in Deep Learning
Extracting Higher-Level Features via Symbolic Approximation Representations
DOI:
https://doi.org/10.32473/flairs.36.133107Keywords:
Deep Learning, Symbolic Time Series Analysis, Explainable Embeddings, Symbolic Representation, Transformer, High-Level Feature ExtractionAbstract
With the success of language models in deep learning, multiple new time series embeddings have been proposed. However, the interpretability of those representations is often still lacking compared to word embeddings. This paper tackles this issue, aiming to present some criteria for making time series embeddings applied in deep learning models more interpretable using higher-level features in symbolic form. For that, we investigate two different approaches for extracting symbolic approximation representations regarding the frequency and the trend information, i.e. the Symbolic Fourier Approximation (SFA) and the Symbolic Aggregate approXimation (SAX). In particular, we analyze and discuss the impact of applying the different representation approaches. Furthermore, in our experimentation, we apply a state-of-the-art Transformer model to demonstrate the efficacy of the proposed approach regarding explainability in a comprehensive evaluation using a large set of time series datasets.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2023 Leonid Schwenke, Martin Atzmueller
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.