Making Time Series Embeddings More Interpretable in Deep Learning

Extracting Higher-Level Features via Symbolic Approximation Representations

Authors

DOI:

https://doi.org/10.32473/flairs.36.133107

Keywords:

Deep Learning, Symbolic Time Series Analysis, Explainable Embeddings, Symbolic Representation, Transformer, High-Level Feature Extraction

Abstract

With the success of language models in deep learning, multiple new time series embeddings have been proposed. However, the interpretability of those representations is often still lacking compared to word embeddings. This paper tackles this issue, aiming to present some criteria for making time series embeddings applied in deep learning models more interpretable using higher-level features in symbolic form. For that, we investigate two different approaches for extracting symbolic approximation representations regarding the frequency and the trend information, i.e. the Symbolic Fourier Approximation (SFA) and the Symbolic Aggregate approXimation (SAX). In particular, we analyze and discuss the impact of applying the different representation approaches. Furthermore, in our experimentation, we apply a state-of-the-art Transformer model to demonstrate the efficacy of the proposed approach regarding explainability in a comprehensive evaluation using a large set of time series datasets.

Downloads

Published

08-05-2023

How to Cite

Schwenke, L., & Atzmueller, M. (2023). Making Time Series Embeddings More Interpretable in Deep Learning: Extracting Higher-Level Features via Symbolic Approximation Representations. The International FLAIRS Conference Proceedings, 36(1). https://doi.org/10.32473/flairs.36.133107

Issue

Section

Special Track: Neural Networks and Data Mining