Betting on Decomposition-Based Time Series Augmentation

In many machine learning applications, especially those involving time series data, obtaining a large, diverse, and high-quality labeled dataset is one of the biggest hurdles. Data collection is often constrained by availability, privacy, annotation costs, and time. 

To address this, data augmentation — the process of generating new data from existing samples — has become a key strategy. However, traditional augmentation methods like adding noise, scaling, or using generative models often fail to preserve the underlying structure of time series data.


Our Solution: Structure-Preserving Augmentation

Time series data is often composed of three main components:

Our method decomposes a time series into these components, applies transformations to each component individually, and then reconstructs a new series. This ensures the augmented data retains the structural integrity of the original.


Layer 1: Transforming Components from the Same Series

In the first layer, we take a single time series and decompose it using techniques like Seasonal-Trend Decomposition using Loess (STL). We then apply transformations such as:

Finally, we recombine the transformed components to synthesize a new time series.

This approach allows us to generate realistic variations while preserving the original structure.


Layer 2: Mix-Matching Components from Different Series

If we have multiple time series from similar sources (e.g., different machines of the same origin), we can further expand our augmentation by mixing components across series. For example:

This creates a new series that is structurally valid but more diverse than any individual source.

This second layer significantly increases the variety of the augmented dataset while maintaining coherence.

Smart Automation and Simple Control 

To make this solution both powerful and easy to use, we introduce two key features: 

  1. Automatic Decomposition Strategy Depending on the characteristics of the time series, the system automatically chooses between: 

a. STL decomposition for series with strong seasonality 

b. Wavelet decomposition for series without clear seasonal patterns 

This decision is based on the energy in the seasonal component, compared against an empirically determined threshold. 

  1. Single “Divergence” Parameter Instead of manually tuning multiple augmentation parameters, users can adjust a single “divergence” setting. Internally, this controls: 

a. Amplitude and time scaling 

b. Whether to apply slicing or remixing 

c. Degree of transformation per component 

This makes the solution accessible to non-experts while still offering the flexibility and power needed by advanced users. 

Why It Matters 

This decomposition-based augmentation strategy — developed and delivered by Edge Impulse — offers a scalable, structure-aware, and user-friendly way to generate realistic synthetic data , improving model robustness and reducing costly data collection.  

Whether you're working with sensor data, industrial signals, or any other time series, this approach helps you get more value from the data you already have. 

Best Practice

Example Project

IMU Exercise Recognition 

Train Accuracy

(20% of all subjects)

Test Accuracy

(80% of all subjects)

With Augmentation 

(x2, divergence = 0.1)

85%

74%

Without Augmentation

82%

66%

Comments

Subscribe

Are you interested in bringing machine learning intelligence to your devices? We're happy to help.

Subscribe to our newsletter