TY - GEN
T1 - ON MINI-BATCH TRAINING WITH VARYING LENGTH TIME SERIES
AU - Iwana, Brian Kenji
N1 - Funding Information:
This work was partially supported by MEXT-Japan (Grant No. J21K17808) and R3QR Program (Qdai-jump Research Program) 01252.
Publisher Copyright:
© 2022 IEEE
PY - 2022
Y1 - 2022
N2 - In real-world time series recognition applications, it is possible to have data with varying length patterns. However, when using artificial neural networks (ANN), it is standard practice to use fixed-sized mini-batches. To do this, time series data with varying lengths are typically normalized so that all the patterns are the same length. Normally, this is done using zero padding or truncation without much consideration. We propose a novel method of normalizing the lengths of the time series in a dataset by exploiting the dynamic matching ability of Dynamic Time Warping (DTW). In this way, the time series lengths in a dataset can be set to a fixed size while maintaining features typical to the dataset. In the experiments, all 11 datasets with varying length time series from the 2018 UCR Time Series Archive are used. We evaluate the proposed method by comparing it with 18 other length normalization methods on a Convolutional Neural Network (CNN), a Long-Short Term Memory network (LSTM), and a Bidirectional LSTM (BLSTM). The code is publicly available at https://github.com/uchidalab/vary length time series.
AB - In real-world time series recognition applications, it is possible to have data with varying length patterns. However, when using artificial neural networks (ANN), it is standard practice to use fixed-sized mini-batches. To do this, time series data with varying lengths are typically normalized so that all the patterns are the same length. Normally, this is done using zero padding or truncation without much consideration. We propose a novel method of normalizing the lengths of the time series in a dataset by exploiting the dynamic matching ability of Dynamic Time Warping (DTW). In this way, the time series lengths in a dataset can be set to a fixed size while maintaining features typical to the dataset. In the experiments, all 11 datasets with varying length time series from the 2018 UCR Time Series Archive are used. We evaluate the proposed method by comparing it with 18 other length normalization methods on a Convolutional Neural Network (CNN), a Long-Short Term Memory network (LSTM), and a Bidirectional LSTM (BLSTM). The code is publicly available at https://github.com/uchidalab/vary length time series.
UR - http://www.scopus.com/inward/record.url?scp=85131236700&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85131236700&partnerID=8YFLogxK
U2 - 10.1109/ICASSP43922.2022.9746542
DO - 10.1109/ICASSP43922.2022.9746542
M3 - Conference contribution
AN - SCOPUS:85131236700
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
SP - 4483
EP - 4487
BT - 2022 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 47th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2022
Y2 - 23 May 2022 through 27 May 2022
ER -