Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add window size selector #1237

Merged
merged 23 commits into from
Jan 4, 2024
Merged
Show file tree
Hide file tree
Changes from 5 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion fedot/core/composer/gp_composer/specific_operators.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ def parameter_change_mutation(pipeline: Pipeline, requirements, graph_gen_params
node_mutation_probability = get_mutation_prob(mut_id=parameters.mutation_strength,
node=pipeline.root_node)
for node in pipeline.nodes:
if random() < node_mutation_probability:
if node.operation.metadata.id != 'lagged' and random() < node_mutation_probability:
operation_name = node.operation.operation_type
current_params = node.parameters

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,8 @@

import numpy as np
import pandas as pd

from fedot.utilities.window_size_selector import WindowSizeSelector
from golem.core.log import default_log
from scipy.ndimage import gaussian_filter
from sklearn.decomposition import TruncatedSVD
Expand Down Expand Up @@ -103,7 +105,7 @@ def transform_for_fit(self, input_data: InputData) -> OutputData:
self._update_column_types(output_data)
return output_data

def _check_and_correct_window_size(self, time_series: np.array, forecast_length: int):
def _check_and_correct_window_size(self, time_series: np.ndarray, forecast_length: int):
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Я про то, это этот метод вызывается в transform и transform_for_fit

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Окно подстраивается только если оно по дефолту не задано. Если задать его руками или изменить в процессе композиции/тюнинга, то оно не будет переопределяться.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

То есть сейчас это работает так:

  1. Дефолтное значение длины окна 0.
  2. Пусть у нас 3 фолда валидации
  3. Заходим в первый фолд, подбираем длину окна
  4. На следующих фолдах и при финальном обучении на всем ряду размер окна не подбираем?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Добавил тесты что при вычислении графа на фолдах и внутри тюнера окно не перенастраивается.

""" Method check if the length of the time series is not enough for
lagged transformation - clip it

Expand All @@ -115,6 +117,18 @@ def _check_and_correct_window_size(self, time_series: np.array, forecast_length:

"""

if self.window_size == 0:
def get_window(ts: np.ndarray):
return int(WindowSizeSelector(method='hac', window_range=(5, 60))
.get_window_size(ts) * len(ts) / 100)

if time_series.ndim > 1:
new = np.mean([get_window(time_series[:, i].ravel()) for i in range(time_series.shape[1])])
else:
new = get_window(time_series)

self.params.update(window_size=new)

# Maximum threshold
if self.window_size + forecast_length > len(time_series):
raise ValueError(f"Window size is to high ({self.window_size}) for provided data len {len(time_series)}")
Expand Down
2 changes: 1 addition & 1 deletion fedot/core/repository/data/default_operation_params.json
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@
"verbose": -1
},
"lagged": {
"window_size": 10
"window_size": 0
},
"diff_filter": {
"window_size": 3,
Expand Down
234 changes: 234 additions & 0 deletions fedot/utilities/window_size_selector.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,234 @@
import math
from typing import Union

import numpy as np
import pandas as pd
from scipy.signal import find_peaks
from statsmodels.tsa.stattools import acf


class WindowSizeSelector:
"""Class to select appropriate window size to catch periodicity for time series analysis.
There are two group of algorithms implemented:
Whole-Series-Based (WSB):
1. 'hac' - highest_autocorrelation
2. 'dff' - dominant_fourier_frequency
Subsequence-based (SB):
1. 'mwf' - multi_window_finder
2. 'sss' - summary_statistics_subsequence
Args:
method: by ``default``, it is 'dff'.
You can choose between: 'hac', 'dff', 'sss' or 'mwf'.
window_range: % of time series length, by ``default`` it is (5, 50).
Attributes:
length_ts(int): length of the time_series.
window_max(int): maximum window size in real values.
window_min(int): minimum window size in real values.
dict_methods(dict): dictionary with all implemented methods.
Example:
To find window size for single time series::
ts = np.random.rand(1000)
ws_selector = WindowSizeSelector(method='hac')
window_size = ws_selector.get_window_size(time_series=ts)
To find window size for multiple time series::
ts = np.random.rand(1000, 10)
ws_selector = WindowSizeSelector(method='hac')
window_size = ws_selector.apply(time_series=ts, average='median')
Reference:
(c) "Windows Size Selection in Unsupervised Time Series Analytics: A Review and Benchmark. Arik Ermshaus,
Patrick Schafer, and Ulf Leser. 2022"
"""

def __init__(self, method: str = 'dff', window_range: tuple = (5, 50)):

assert window_range[0] < window_range[1], 'Upper bound of window range should be bigger than lower bound'

self.dict_methods = {'hac': self.autocorrelation,
'dff': self.dominant_fourier_frequency,
'mwf': self.mwf,
'sss': self.summary_statistics_subsequence}
self.wss_algorithm = method
self.window_range = window_range
self.window_max = None
self.window_min = None
self.length_ts = None

def apply(self, time_series: Union[pd.DataFrame, np.array], average: str = 'median') -> int:
"""Method to run WSS class over selected time series in parallel mode via joblib
Args:
time_series: time series to study
average: 'mean' or 'median' to average window size over all time series
Returns:
window_size_selected: value which has been chosen as appropriate window size
"""
methods = {'mean': np.mean, 'median': np.median}
assert average in methods.keys(), 'Hyperparameters error: `average` should be mean or median'

if isinstance(time_series, pd.DataFrame):
time_series = time_series.values

window_list = [self.get_window_size(ts) for ts in time_series]
return round(methods[average](window_list))

def get_window_size(self, time_series: np.array) -> int:
"""Main function to run WSS class over selected time series
Note:
One of the reason of ValueError is that time series size can be equal or smaller than 50.
In case of it try to initially set window_size min and max.
Returns:
window_size_selected: value which has been chosen as appropriate window size
"""
if time_series.shape[0] == 1: # If time series is a part of multivariate one
time_series = np.array(time_series[0])
self.length_ts = len(time_series)

self.window_max = int(round(self.length_ts * self.window_range[1] / 100)) # in real values
self.window_min = int(round(self.length_ts * self.window_range[0] / 100)) # in real values

window_size_selected = self.dict_methods[self.wss_algorithm](time_series=time_series)
return round(window_size_selected * 100 / self.length_ts) # in %

def dominant_fourier_frequency(self, time_series: np.array) -> int:
"""
Method to find dominant fourier frequency in time series and return appropriate window size. It is based on
the assumption that the dominant frequency is the one with the highest magnitude in the Fourier transform. The
window size is then the inverse of the dominant frequency.
"""
fourier = np.fft.fft(time_series)
freq = np.fft.fftfreq(time_series.shape[0], 1)

magnitudes, window_sizes = [], []

for coef, freq in zip(fourier, freq):
if coef and freq > 0:
window_size = int(1 / freq)
mag = math.sqrt(coef.real * coef.real + coef.imag * coef.imag)

if self.window_min <= window_size < self.window_max:
window_sizes.append(window_size)
magnitudes.append(mag)

return window_sizes[np.argmax(magnitudes)]

def autocorrelation(self, time_series: np.array) -> int:
"""Method to find the highest autocorrelation in time series and return appropriate window size. It is based on
the assumption that the lag of highest autocorrelation coefficient corresponds to the window size that best
captures the periodicity of the time series.
"""
ts_len = time_series.shape[0]
acf_values = acf(time_series, fft=True, nlags=int(ts_len / 2))

peaks, _ = find_peaks(acf_values)
peaks = peaks[np.logical_and(peaks >= self.window_min, peaks < self.window_max)]
corrs = acf_values[peaks]

if peaks.shape[0] == 0: # if there is no peaks in range (window_min, window_max) return window_min
return self.window_min
return peaks[np.argmax(corrs)]

def mwf(self, time_series: np.array) -> int:
""" Method to find the window size that minimizes the moving average residual. It is based on the assumption
that the window size that best captures the periodicity of the time series is the one that minimizes the
difference between the moving average and the time series.
"""

all_averages, window_sizes = [], []

for w in range(self.window_min, self.window_max, 1):
movingAvg = np.array(self.movmean(time_series, w))
all_averages.append(movingAvg)
window_sizes.append(w)

movingAvgResiduals = []

for i, w in enumerate(window_sizes):
moving_avg = all_averages[i][:len(all_averages[-1])]
movingAvgResidual = np.log(abs(moving_avg - (moving_avg).mean()).sum())
movingAvgResiduals.append(movingAvgResidual)

b = (np.diff(np.sign(np.diff(movingAvgResiduals))) > 0).nonzero()[0] + 1 # local min

if len(b) == 0:
return self.window_min
if len(b) < 3:
return window_sizes[b[0]]

reswin = np.array([window_sizes[b[i]] / (i + 1) for i in range(3)])
w = np.mean(reswin)

return int(w)

def movmean(self, ts, w):
"""Fast moving average function"""
moving_avg = np.cumsum(ts, dtype=float)
moving_avg[w:] = moving_avg[w:] - moving_avg[:-w]
return moving_avg[w - 1:] / w

def summary_statistics_subsequence(self, time_series: np.array, threshold=.89) -> int:
"""Method to find the window size that maximizes the subsequence unsupervised similarity score (SUSS). It is
based on the assumption that the window size that best captures the periodicity of the time series is the one
that maximizes the similarity between subsequences of the time series.
"""
# lbound = self.window_min
time_series = (time_series - time_series.min()) / (time_series.max() - time_series.min())

ts_mean = np.mean(time_series)
ts_std = np.std(time_series)
ts_min_max = np.max(time_series) - np.min(time_series)

stats = (ts_mean, ts_std, ts_min_max)

max_score = self.suss_score(time_series=time_series, window_size=1, stats=stats)
min_score = self.suss_score(time_series=time_series, window_size=time_series.shape[0] - 1, stats=stats)

exp = 0

# exponential search (to find window size interval)
while True:
window_size = 2 ** exp

if window_size < self.window_min:
exp += 1
continue

score = 1 - (self.suss_score(time_series, window_size, stats) - min_score) / (max_score - min_score)

if score > threshold:
break

exp += 1

lbound, ubound = max(self.window_min, 2 ** (exp - 1)), 2 ** exp + 1

# binary search (to find window size in interval)
while lbound <= ubound:
window_size = int((lbound + ubound) / 2)
score = 1 - (self.suss_score(time_series, window_size, stats) - min_score) / (max_score - min_score)

if score < threshold:
lbound = window_size + 1
elif score > threshold:
ubound = window_size - 1
else:
break

return 2 * lbound

def suss_score(self, time_series, window_size, stats):
roll = pd.Series(time_series).rolling(window_size)
ts_mean, ts_std, ts_min_max = stats

roll_mean = roll.mean().to_numpy()[window_size:]
roll_std = roll.std(ddof=0).to_numpy()[window_size:]
roll_min = roll.min().to_numpy()[window_size:]
roll_max = roll.max().to_numpy()[window_size:]

X = np.array([
roll_mean - ts_mean,
roll_std - ts_std,
(roll_max - roll_min) - ts_min_max
])

X = np.sqrt(np.sum(np.square(X), axis=0)) / np.sqrt(window_size)

return np.mean(X)
8 changes: 5 additions & 3 deletions test/unit/optimizer/gp_operators/test_mutation.py
Original file line number Diff line number Diff line change
Expand Up @@ -131,8 +131,6 @@ def test_boosting_mutation_for_non_lagged_ts_model():
"""

graph = PipelineAdapter().restore(get_ts_forecasting_graph())

boosting_graph = get_ts_forecasting_graph_with_boosting()
requirements = PipelineComposerRequirements(primary=['ridge'],
secondary=['ridge'])
pipeline = boosting_mutation(graph,
Expand All @@ -143,7 +141,11 @@ def test_boosting_mutation_for_non_lagged_ts_model():
data_train, data_test = get_ts_data()
pipeline.fit(data_train)
result = pipeline.predict(data_test)
assert boosting_graph.descriptive_id == pipeline.descriptive_id

boosting_pipeline = PipelineAdapter().restore(get_ts_forecasting_graph_with_boosting())
boosting_pipeline.fit(data_train)

assert boosting_pipeline.descriptive_id == pipeline.descriptive_id
assert result is not None


Expand Down
Loading