2017-01-14 03:42:00 +04:00
|
|
|
#!/usr/bin/python
|
|
|
|
# -*- coding: utf8 -*-
|
|
|
|
|
2017-05-05 22:01:45 +04:00
|
|
|
"""Benchmarks to FTS methods"""
|
|
|
|
|
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
import datetime
|
2017-05-08 20:12:08 +04:00
|
|
|
import time
|
|
|
|
from copy import deepcopy
|
|
|
|
|
2016-09-08 16:03:32 +04:00
|
|
|
import matplotlib as plt
|
2017-03-03 15:53:55 +04:00
|
|
|
import matplotlib.cm as cmx
|
2017-05-08 20:12:08 +04:00
|
|
|
import matplotlib.colors as pltcolors
|
2016-09-08 01:51:00 +04:00
|
|
|
import matplotlib.pyplot as plt
|
2017-05-08 20:12:08 +04:00
|
|
|
import numpy as np
|
|
|
|
import pandas as pd
|
2016-09-08 01:51:00 +04:00
|
|
|
from mpl_toolkits.mplot3d import Axes3D
|
2017-05-08 20:12:08 +04:00
|
|
|
|
|
|
|
from probabilistic import ProbabilityDistribution
|
|
|
|
from pyFTS import song, chen, yu, ismailefendi, sadaei, hofts, pwfts, ifts, cheng, ensemble, hwang
|
2017-05-10 02:04:51 +04:00
|
|
|
from pyFTS.benchmarks import Measures, naive, arima, ResidualAnalysis, quantreg
|
|
|
|
from pyFTS.benchmarks import Util as bUtil
|
2017-05-08 20:12:08 +04:00
|
|
|
from pyFTS.common import Transformations, Util
|
2017-01-23 00:41:42 +04:00
|
|
|
# from sklearn.cross_validation import KFold
|
2017-05-08 20:12:08 +04:00
|
|
|
from pyFTS.partitioners import Grid
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
colors = ['grey', 'rosybrown', 'maroon', 'red','orange', 'yellow', 'olive', 'green',
|
|
|
|
'cyan', 'blue', 'darkblue', 'purple', 'darkviolet']
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
ncol = len(colors)
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
styles = ['-','--','-.',':','.']
|
|
|
|
|
|
|
|
nsty = len(styles)
|
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def get_benchmark_point_methods():
|
2017-05-07 04:19:04 +04:00
|
|
|
"""Return all non FTS methods for point forecasting"""
|
2017-05-07 00:04:37 +04:00
|
|
|
return [naive.Naive, arima.ARIMA, quantreg.QuantileRegression]
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def get_point_methods():
|
2017-05-07 04:19:04 +04:00
|
|
|
"""Return all FTS methods for point forecasting"""
|
|
|
|
return [song.ConventionalFTS, chen.ConventionalFTS, yu.WeightedFTS, ismailefendi.ImprovedWeightedFTS,
|
2017-05-07 18:41:31 +04:00
|
|
|
cheng.TrendWeightedFTS, sadaei.ExponentialyWeightedFTS, hofts.HighOrderFTS, hwang.HighOrderFTS,
|
2017-05-07 04:19:04 +04:00
|
|
|
pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-05 22:01:45 +04:00
|
|
|
def get_benchmark_interval_methods():
|
2017-05-07 04:19:04 +04:00
|
|
|
"""Return all non FTS methods for interval forecasting"""
|
2017-05-05 22:01:45 +04:00
|
|
|
return [quantreg.QuantileRegression]
|
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def get_interval_methods():
|
2017-05-07 04:19:04 +04:00
|
|
|
"""Return all FTS methods for interval forecasting"""
|
2017-03-03 15:53:55 +04:00
|
|
|
return [ifts.IntervalFTS, pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
def get_probabilistic_methods():
|
|
|
|
"""Return all FTS methods for probabilistic forecasting"""
|
|
|
|
return [quantreg.QuantileRegression, ensemble.EnsembleFTS, pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
def run_point(mfts, partitioner, train_data, test_data, window_key=None, transformation=None, indexer=None):
|
|
|
|
"""
|
|
|
|
Point forecast benchmark function to be executed on sliding window
|
|
|
|
:param mfts: FTS model
|
|
|
|
:param partitioner: Universe of Discourse partitioner
|
|
|
|
:param train_data: data used to train the model
|
|
|
|
:param test_data: ata used to test the model
|
|
|
|
:param window_key: id of the sliding window
|
|
|
|
:param transformation: data transformation
|
|
|
|
:param indexer: seasonal indexer
|
|
|
|
:return: a dictionary with the benchmark results
|
|
|
|
"""
|
|
|
|
|
|
|
|
if mfts.benchmark_only:
|
|
|
|
_key = mfts.shortname + str(mfts.order if mfts.order is not None else "")
|
|
|
|
else:
|
|
|
|
pttr = str(partitioner.__module__).split('.')[-1]
|
|
|
|
_key = mfts.shortname + " n = " + str(mfts.order) + " " + pttr + " q = " + str(partitioner.partitions)
|
|
|
|
mfts.partitioner = partitioner
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
|
|
|
_start = time.time()
|
|
|
|
mfts.train(train_data, partitioner.sets, order=mfts.order)
|
|
|
|
_end = time.time()
|
|
|
|
times = _end - _start
|
|
|
|
|
|
|
|
_start = time.time()
|
|
|
|
_rmse, _smape, _u = Measures.get_point_statistics(test_data, mfts, indexer)
|
|
|
|
_end = time.time()
|
|
|
|
times += _end - _start
|
|
|
|
|
|
|
|
ret = {'key': _key, 'obj': mfts, 'rmse': _rmse, 'smape': _smape, 'u': _u, 'time': times, 'window': window_key}
|
|
|
|
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
|
|
|
def point_sliding_window(data, windowsize, train=0.8, models=None, partitioners=[Grid.GridPartitioner],
|
|
|
|
partitions=[10], max_order=3, transformation=None, indexer=None, dump=False,
|
2017-05-08 21:49:45 +04:00
|
|
|
benchmark_models=None, benchmark_models_parameters = None,
|
2017-05-10 02:04:51 +04:00
|
|
|
save=False, file=None, sintetic=False):
|
2017-05-05 22:01:45 +04:00
|
|
|
"""
|
|
|
|
Sliding window benchmarks for FTS point forecasters
|
|
|
|
:param data:
|
|
|
|
:param windowsize: size of sliding window
|
|
|
|
:param train: percentual of sliding window data used to train the models
|
|
|
|
:param models: FTS point forecasters
|
|
|
|
:param partitioners: Universe of Discourse partitioner
|
|
|
|
:param partitions: the max number of partitions on the Universe of Discourse
|
|
|
|
:param max_order: the max order of the models (for high order models)
|
|
|
|
:param transformation: data transformation
|
|
|
|
:param indexer: seasonal indexer
|
|
|
|
:param dump:
|
2017-05-10 02:04:51 +04:00
|
|
|
:param benchmark_models: Non FTS models to benchmark
|
|
|
|
:param benchmark_models_parameters: Non FTS models parameters
|
2017-05-05 22:01:45 +04:00
|
|
|
:param save: save results
|
|
|
|
:param file: file path to save the results
|
|
|
|
:param sintetic: if true only the average and standard deviation of the results
|
|
|
|
:return: DataFrame with the results
|
|
|
|
"""
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
if benchmark_models is None: # and models is None:
|
|
|
|
benchmark_models = [naive.Naive, arima.ARIMA, arima.ARIMA, arima.ARIMA, arima.ARIMA,
|
|
|
|
quantreg.QuantileRegression, quantreg.QuantileRegression]
|
|
|
|
|
|
|
|
if benchmark_models_parameters is None:
|
|
|
|
benchmark_models_parameters = [1, (1, 0, 0), (1, 0, 1), (2, 0, 1), (2, 0, 2), 1, 2]
|
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
_process_start = time.time()
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
print("Process Start: {0: %H:%M:%S}".format(datetime.datetime.now()))
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
pool = []
|
|
|
|
jobs = []
|
2017-03-03 15:53:55 +04:00
|
|
|
objs = {}
|
|
|
|
rmse = {}
|
|
|
|
smape = {}
|
|
|
|
u = {}
|
|
|
|
times = {}
|
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
if models is None:
|
|
|
|
models = get_point_methods()
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
for model in models:
|
|
|
|
mfts = model("")
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
if mfts.is_high_order:
|
|
|
|
for order in np.arange(1, max_order + 1):
|
|
|
|
if order >= mfts.min_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts = model("")
|
2017-05-10 02:04:51 +04:00
|
|
|
mfts.order = order
|
|
|
|
pool.append(mfts)
|
|
|
|
else:
|
|
|
|
mfts.order = 1
|
|
|
|
pool.append(mfts)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
if benchmark_models is not None:
|
|
|
|
for count, model in enumerate(benchmark_models, start=0):
|
|
|
|
par = benchmark_models_parameters[count]
|
|
|
|
mfts = model(str(par if par is not None else ""))
|
|
|
|
mfts.order = par
|
|
|
|
pool.append(mfts)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
experiments = 0
|
|
|
|
for ct, train, test in Util.sliding_window(data, windowsize, train):
|
|
|
|
experiments += 1
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
benchmarks_only = {}
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
if dump: print('\nWindow: {0}\n'.format(ct))
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
for partition in partitions:
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
for partitioner in partitioners:
|
2017-03-22 06:17:06 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
data_train_fs = partitioner(train, partition, transformation=transformation)
|
2017-03-22 06:17:06 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
for _id, m in enumerate(pool,start=0):
|
|
|
|
if m.benchmark_only and m.shortname in benchmarks_only:
|
|
|
|
continue
|
2017-03-03 15:53:55 +04:00
|
|
|
else:
|
2017-05-10 02:04:51 +04:00
|
|
|
benchmarks_only[m.shortname] = m
|
|
|
|
|
|
|
|
tmp = run_point(deepcopy(m), data_train_fs, train, test, ct, transformation)
|
|
|
|
|
|
|
|
if tmp['key'] not in objs:
|
|
|
|
objs[tmp['key']] = tmp['obj']
|
|
|
|
rmse[tmp['key']] = []
|
|
|
|
smape[tmp['key']] = []
|
|
|
|
u[tmp['key']] = []
|
|
|
|
times[tmp['key']] = []
|
|
|
|
rmse[tmp['key']].append(tmp['rmse'])
|
|
|
|
smape[tmp['key']].append(tmp['smape'])
|
|
|
|
u[tmp['key']].append(tmp['u'])
|
|
|
|
times[tmp['key']].append(tmp['time'])
|
|
|
|
print(tmp['key'], tmp['window'])
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
_process_end = time.time()
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
print("Process End: {0: %H:%M:%S}".format(datetime.datetime.now()))
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
print("Process Duration: {0}".format(_process_end - _process_start))
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
return bUtil.save_dataframe_point(experiments, file, objs, rmse, save, sintetic, smape, times, u)
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
|
|
|
|
def all_point_forecasters(data_train, data_test, partitions, max_order=3, statistics=True, residuals=True,
|
|
|
|
series=True, save=False, file=None, tam=[20, 5], models=None, transformation=None,
|
|
|
|
distributions=False):
|
2017-05-05 22:01:45 +04:00
|
|
|
"""
|
|
|
|
Fixed data benchmark for FTS point forecasters
|
|
|
|
:param data_train: data used to train the models
|
|
|
|
:param data_test: data used to test the models
|
|
|
|
:param partitions: the max number of partitions on the Universe of Discourse
|
|
|
|
:param max_order: the max order of the models (for high order models)
|
|
|
|
:param statistics: print statistics
|
|
|
|
:param residuals: print and plot residuals
|
|
|
|
:param series: plot time series
|
|
|
|
:param save: save results
|
|
|
|
:param file: file path to save the results
|
|
|
|
:param tam: figure dimensions to plot the graphs
|
|
|
|
:param models: list of models to benchmark
|
|
|
|
:param transformation: data transformation
|
|
|
|
:param distributions: plot distributions
|
|
|
|
:return:
|
|
|
|
"""
|
2017-03-03 15:53:55 +04:00
|
|
|
if models is None:
|
|
|
|
models = get_point_methods()
|
|
|
|
|
|
|
|
objs = []
|
2017-02-27 22:53:29 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
data_train_fs = Grid.GridPartitioner(data_train, partitions, transformation=transformation)
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-24 16:40:48 +04:00
|
|
|
count = 1
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors = []
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
for count, model in enumerate(models, start=0):
|
2017-01-24 16:40:48 +04:00
|
|
|
#print(model)
|
|
|
|
mfts = model("")
|
2017-05-03 00:16:49 +04:00
|
|
|
if not mfts.is_high_order:
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts.train(data_train, data_train_fs.sets)
|
2017-01-24 16:40:48 +04:00
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append( colors[count % ncol] )
|
2017-01-23 17:00:27 +04:00
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
2017-05-03 00:16:49 +04:00
|
|
|
if order >= mfts.min_order:
|
2017-01-25 18:17:07 +04:00
|
|
|
mfts = model(" n = " + str(order))
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts.train(data_train, data_train_fs.sets, order=order)
|
2017-01-25 18:17:07 +04:00
|
|
|
objs.append(mfts)
|
2017-03-03 15:53:55 +04:00
|
|
|
lcolors.append(colors[(count + order) % ncol])
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if statistics:
|
2017-03-03 15:53:55 +04:00
|
|
|
print_point_statistics(data_test, objs)
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if residuals:
|
|
|
|
print(ResidualAnalysis.compareResiduals(data_test, objs))
|
2017-03-03 15:53:55 +04:00
|
|
|
ResidualAnalysis.plot_residuals(data_test, objs, save=save, file=file, tam=tam)
|
2017-01-25 18:17:07 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if series:
|
2017-03-03 15:53:55 +04:00
|
|
|
plot_compared_series(data_test, objs, lcolors, typeonlegend=False, save=save, file=file, tam=tam,
|
|
|
|
intervals=False)
|
2017-01-25 18:17:07 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
if distributions:
|
|
|
|
lcolors.insert(0,'black')
|
|
|
|
pmfs = []
|
|
|
|
pmfs.append(
|
|
|
|
ProbabilityDistribution.ProbabilityDistribution("Original", 100, [min(data_test), max(data_test)], data=data_test) )
|
2017-02-24 20:29:55 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
for m in objs:
|
|
|
|
forecasts = m.forecast(data_test)
|
|
|
|
pmfs.append(
|
|
|
|
ProbabilityDistribution.ProbabilityDistribution(m.shortname, 100, [min(data_test), max(data_test)],
|
|
|
|
data=forecasts))
|
|
|
|
print(getProbabilityDistributionStatistics(pmfs,data_test))
|
|
|
|
|
|
|
|
plot_probability_distributions(pmfs, lcolors, tam=tam)
|
|
|
|
|
|
|
|
|
|
|
|
def print_point_statistics(data, models, externalmodels = None, externalforecasts = None, indexers=None):
|
|
|
|
ret = "Model & Order & RMSE & SMAPE & Theil's U \\\\ \n"
|
|
|
|
for count,model in enumerate(models,start=0):
|
2017-05-02 03:56:47 +04:00
|
|
|
_rmse, _smape, _u = Measures.get_point_statistics(data, model, indexers)
|
2017-03-03 15:53:55 +04:00
|
|
|
ret += model.shortname + " & "
|
|
|
|
ret += str(model.order) + " & "
|
|
|
|
ret += str(_rmse) + " & "
|
|
|
|
ret += str(_smape)+ " & "
|
|
|
|
ret += str(_u)
|
2017-01-27 14:26:47 +04:00
|
|
|
#ret += str(round(Measures.TheilsInequality(np.array(data[fts.order:]), np.array(forecasts[:-1])), 4))
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += " \\\\ \n"
|
|
|
|
if externalmodels is not None:
|
|
|
|
l = len(externalmodels)
|
|
|
|
for k in np.arange(0,l):
|
|
|
|
ret += externalmodels[k] + " & "
|
|
|
|
ret += " 1 & "
|
2017-03-03 15:53:55 +04:00
|
|
|
ret += str(round(Measures.rmse(data, externalforecasts[k][:-1]), 2)) + " & "
|
|
|
|
ret += str(round(Measures.smape(data, externalforecasts[k][:-1]), 2))+ " & "
|
|
|
|
ret += str(round(Measures.UStatistic(data, externalforecasts[k][:-1]), 2))
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += " \\\\ \n"
|
2017-03-03 15:53:55 +04:00
|
|
|
print(ret)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def getProbabilityDistributionStatistics(pmfs, data):
|
|
|
|
ret = "Model & Entropy & Empirical Likelihood & Pseudo Likelihood \\\\ \n"
|
|
|
|
for k in pmfs:
|
|
|
|
ret += k.name + " & "
|
|
|
|
ret += str(k.entropy()) + " & "
|
|
|
|
ret += str(k.empiricalloglikelihood())+ " & "
|
|
|
|
ret += str(k.pseudologlikelihood(data))
|
|
|
|
ret += " \\\\ \n"
|
2017-01-23 17:00:27 +04:00
|
|
|
return ret
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
def interval_sliding_window(data, windowsize, train=0.8, models=None, partitioners=[Grid.GridPartitioner],
|
|
|
|
partitions=[10], max_order=3, transformation=None, indexer=None, dump=False,
|
|
|
|
save=False, file=None, synthetic=True):
|
2017-03-03 15:53:55 +04:00
|
|
|
if models is None:
|
|
|
|
models = get_interval_methods()
|
|
|
|
|
|
|
|
objs = {}
|
|
|
|
lcolors = {}
|
|
|
|
sharpness = {}
|
|
|
|
resolution = {}
|
|
|
|
coverage = {}
|
2017-03-22 06:17:06 +04:00
|
|
|
times = {}
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
experiments = 0
|
2017-03-22 06:17:06 +04:00
|
|
|
for ct, training,test in Util.sliding_window(data, windowsize, train):
|
2017-04-05 01:39:15 +04:00
|
|
|
experiments += 1
|
2017-03-03 15:53:55 +04:00
|
|
|
for partition in partitions:
|
|
|
|
for partitioner in partitioners:
|
|
|
|
pttr = str(partitioner.__module__).split('.')[-1]
|
2017-03-22 06:17:06 +04:00
|
|
|
data_train_fs = partitioner(training, partition, transformation=transformation)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
for count, model in enumerate(models, start=0):
|
|
|
|
|
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " " + pttr+ " q = " +str(partition)
|
|
|
|
|
|
|
|
mfts.partitioner = data_train_fs
|
2017-05-03 00:16:49 +04:00
|
|
|
if not mfts.is_high_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if dump: print(ct,_key)
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
sharpness[_key] = []
|
|
|
|
resolution[_key] = []
|
|
|
|
coverage[_key] = []
|
2017-03-22 06:17:06 +04:00
|
|
|
times[_key] = []
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
|
|
|
mfts.train(training, data_train_fs.sets)
|
|
|
|
_end = time.time()
|
|
|
|
_tdiff = _end - _start
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
2017-04-07 02:23:29 +04:00
|
|
|
_sharp, _res, _cov = Measures.get_interval_statistics(test, mfts)
|
2017-03-22 06:17:06 +04:00
|
|
|
_end = time.time()
|
|
|
|
_tdiff += _end - _start
|
2017-03-03 15:53:55 +04:00
|
|
|
sharpness[_key].append(_sharp)
|
|
|
|
resolution[_key].append(_res)
|
|
|
|
coverage[_key].append(_cov)
|
2017-03-22 06:17:06 +04:00
|
|
|
times[_key].append(_tdiff)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
else:
|
|
|
|
for order in np.arange(1, max_order + 1):
|
2017-05-03 00:16:49 +04:00
|
|
|
if order >= mfts.min_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " n = " + str(order) + " " + pttr + " q = " + str(partition)
|
|
|
|
mfts.partitioner = data_train_fs
|
|
|
|
|
|
|
|
if dump: print(ct,_key)
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
sharpness[_key] = []
|
|
|
|
resolution[_key] = []
|
|
|
|
coverage[_key] = []
|
2017-03-22 06:17:06 +04:00
|
|
|
times[_key] = []
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
|
|
|
mfts.train(training, data_train_fs.sets, order=order)
|
|
|
|
_end = time.time()
|
|
|
|
|
|
|
|
_tdiff = _end - _start
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
2017-04-07 02:23:29 +04:00
|
|
|
_sharp, _res, _cov = Measures.get_interval_statistics(test, mfts)
|
2017-03-22 06:17:06 +04:00
|
|
|
_end = time.time()
|
|
|
|
_tdiff += _end - _start
|
2017-03-03 15:53:55 +04:00
|
|
|
sharpness[_key].append(_sharp)
|
|
|
|
resolution[_key].append(_res)
|
|
|
|
coverage[_key].append(_cov)
|
2017-03-22 06:17:06 +04:00
|
|
|
times[_key].append(_tdiff)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
return bUtil.save_dataframe_interval(coverage, experiments, file, objs, resolution, save, sharpness, synthetic, times)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
|
|
|
|
def all_interval_forecasters(data_train, data_test, partitions, max_order=3,save=False, file=None, tam=[20, 5],
|
2017-01-27 14:26:47 +04:00
|
|
|
models=None, transformation=None):
|
|
|
|
if models is None:
|
2017-03-03 15:53:55 +04:00
|
|
|
models = get_interval_methods()
|
2017-01-24 16:40:48 +04:00
|
|
|
|
|
|
|
objs = []
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
data_train_fs = Grid.GridPartitioner(data_train,partitions, transformation=transformation).sets
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors = []
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
for count, model in Util.enumerate2(models, start=0, step=2):
|
2017-01-24 16:40:48 +04:00
|
|
|
mfts = model("")
|
2017-05-03 00:16:49 +04:00
|
|
|
if not mfts.is_high_order:
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-24 16:40:48 +04:00
|
|
|
mfts.train(data_train, data_train_fs)
|
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append( colors[count % ncol] )
|
2017-01-24 16:40:48 +04:00
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
2017-05-03 00:16:49 +04:00
|
|
|
if order >= mfts.min_order:
|
2017-01-25 18:17:07 +04:00
|
|
|
mfts = model(" n = " + str(order))
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-25 18:17:07 +04:00
|
|
|
mfts.train(data_train, data_train_fs, order=order)
|
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append(colors[count % ncol])
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
print_interval_statistics(data_test, objs)
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
plot_compared_series(data_test, objs, lcolors, typeonlegend=False, save=save, file=file, tam=tam, intervals=True)
|
2017-01-24 16:40:48 +04:00
|
|
|
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def print_interval_statistics(original, models):
|
2017-01-30 03:59:50 +04:00
|
|
|
ret = "Model & Order & Sharpness & Resolution & Coverage \\\\ \n"
|
2017-01-11 00:05:51 +04:00
|
|
|
for fts in models:
|
2017-04-07 02:23:29 +04:00
|
|
|
_sharp, _res, _cov = Measures.get_interval_statistics(original, fts)
|
2017-01-23 17:00:27 +04:00
|
|
|
ret += fts.shortname + " & "
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += str(fts.order) + " & "
|
2017-03-03 15:53:55 +04:00
|
|
|
ret += str(_sharp) + " & "
|
|
|
|
ret += str(_res) + " & "
|
|
|
|
ret += str(_cov) + " \\\\ \n"
|
|
|
|
print(ret)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def plot_distribution(dist):
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in dist.index:
|
|
|
|
alpha = np.array([dist[x][k] for x in dist]) * 100
|
|
|
|
x = [k for x in np.arange(0, len(alpha))]
|
|
|
|
y = dist.columns
|
|
|
|
plt.scatter(x, y, c=alpha, marker='s', linewidths=0, cmap='Oranges', norm=pltcolors.Normalize(vmin=0, vmax=1),
|
|
|
|
vmin=0, vmax=1, edgecolors=None)
|
|
|
|
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def plot_compared_series(original, models, colors, typeonlegend=False, save=False, file=None, tam=[20, 5],
|
|
|
|
points=True, intervals=True, linewidth=1.5):
|
2017-01-13 01:25:10 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-11 00:05:51 +04:00
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
legends = []
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
ax.plot(original, color='black', label="Original", linewidth=linewidth*1.5)
|
2017-01-27 14:26:47 +04:00
|
|
|
|
|
|
|
for count, fts in enumerate(models, start=0):
|
2017-05-03 00:16:49 +04:00
|
|
|
if fts.has_point_forecasting and points:
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted = fts.forecast(original)
|
2017-01-23 00:41:42 +04:00
|
|
|
mi.append(min(forecasted) * 0.95)
|
|
|
|
ma.append(max(forecasted) * 1.05)
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in np.arange(0, fts.order):
|
|
|
|
forecasted.insert(0, None)
|
2017-01-13 01:25:10 +04:00
|
|
|
lbl = fts.shortname
|
|
|
|
if typeonlegend: lbl += " (Point)"
|
2017-03-03 15:53:55 +04:00
|
|
|
ax.plot(forecasted, color=colors[count], label=lbl, ls="-",linewidth=linewidth)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-05-03 00:16:49 +04:00
|
|
|
if fts.has_interval_forecasting and intervals:
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted = fts.forecastInterval(original)
|
|
|
|
lower = [kk[0] for kk in forecasted]
|
|
|
|
upper = [kk[1] for kk in forecasted]
|
2017-01-23 00:41:42 +04:00
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in np.arange(0, fts.order):
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
2017-01-13 01:25:10 +04:00
|
|
|
lbl = fts.shortname
|
|
|
|
if typeonlegend: lbl += " (Interval)"
|
2017-03-03 15:53:55 +04:00
|
|
|
ax.plot(lower, color=colors[count], label=lbl, ls="--",linewidth=linewidth)
|
|
|
|
ax.plot(upper, color=colors[count], ls="--",linewidth=linewidth)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
2017-01-27 14:26:47 +04:00
|
|
|
lgd = ax.legend(handles0, labels0, loc=2, bbox_to_anchor=(1, 1))
|
|
|
|
legends.append(lgd)
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax.set_title(fts.name)
|
|
|
|
ax.set_ylim([min(mi), max(ma)])
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
Util.showAndSaveImage(fig, file, save, lgd=legends)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-02-16 00:16:13 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def plot_probability_distributions(pmfs, lcolors, tam=[15, 7]):
|
|
|
|
fig = plt.figure(figsize=tam)
|
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
|
|
|
for k,m in enumerate(pmfs,start=0):
|
|
|
|
m.plot(ax, color=lcolors[k])
|
|
|
|
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0)
|
|
|
|
|
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
def ahead_sliding_window(data, windowsize, train, steps, models=None, resolution = None, partitioners=[Grid.GridPartitioner],
|
2017-05-10 02:04:51 +04:00
|
|
|
partitions=[10], max_order=3, transformation=None, indexer=None, dump=False,
|
|
|
|
save=False, file=None, synthetic=False):
|
2017-03-03 15:53:55 +04:00
|
|
|
if models is None:
|
|
|
|
models = [pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
|
|
|
objs = {}
|
|
|
|
lcolors = {}
|
|
|
|
crps_interval = {}
|
|
|
|
crps_distr = {}
|
2017-03-22 06:17:06 +04:00
|
|
|
times1 = {}
|
|
|
|
times2 = {}
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-04-05 01:39:15 +04:00
|
|
|
experiments = 0
|
2017-03-03 15:53:55 +04:00
|
|
|
for ct, train,test in Util.sliding_window(data, windowsize, train):
|
2017-04-05 01:39:15 +04:00
|
|
|
experiments += 1
|
2017-03-03 15:53:55 +04:00
|
|
|
for partition in partitions:
|
|
|
|
for partitioner in partitioners:
|
|
|
|
pttr = str(partitioner.__module__).split('.')[-1]
|
|
|
|
data_train_fs = partitioner(train, partition, transformation=transformation)
|
|
|
|
|
|
|
|
for count, model in enumerate(models, start=0):
|
|
|
|
|
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " " + pttr+ " q = " +str(partition)
|
|
|
|
|
|
|
|
mfts.partitioner = data_train_fs
|
2017-05-03 00:16:49 +04:00
|
|
|
if not mfts.is_high_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if dump: print(ct,_key)
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
crps_interval[_key] = []
|
|
|
|
crps_distr[_key] = []
|
2017-03-22 06:17:06 +04:00
|
|
|
times1[_key] = []
|
|
|
|
times2[_key] = []
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts.train(train, data_train_fs.sets)
|
2017-03-22 06:17:06 +04:00
|
|
|
_end = time.time()
|
|
|
|
|
|
|
|
_tdiff = _end - _start
|
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
_crps1, _crps2, _t1, _t2 = Measures.get_distribution_statistics(test,mfts,steps=steps,resolution=resolution)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
crps_interval[_key].append(_crps1)
|
|
|
|
crps_distr[_key].append(_crps2)
|
2017-03-22 06:17:06 +04:00
|
|
|
times1[_key] = _tdiff + _t1
|
|
|
|
times2[_key] = _tdiff + _t2
|
|
|
|
|
|
|
|
if dump: print(_crps1, _crps2, _tdiff, _t1, _t2)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
else:
|
|
|
|
for order in np.arange(1, max_order + 1):
|
2017-05-03 00:16:49 +04:00
|
|
|
if order >= mfts.min_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " n = " + str(order) + " " + pttr + " q = " + str(partition)
|
|
|
|
mfts.partitioner = data_train_fs
|
|
|
|
|
|
|
|
if dump: print(ct,_key)
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
crps_interval[_key] = []
|
|
|
|
crps_distr[_key] = []
|
2017-03-22 06:17:06 +04:00
|
|
|
times1[_key] = []
|
|
|
|
times2[_key] = []
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
_start = time.time()
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts.train(train, data_train_fs.sets, order=order)
|
2017-03-22 06:17:06 +04:00
|
|
|
_end = time.time()
|
|
|
|
|
|
|
|
_tdiff = _end - _start
|
|
|
|
|
2017-05-07 04:19:04 +04:00
|
|
|
_crps1, _crps2, _t1, _t2 = Measures.get_distribution_statistics(test, mfts, steps=steps,
|
2017-03-22 06:17:06 +04:00
|
|
|
resolution=resolution)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
crps_interval[_key].append(_crps1)
|
|
|
|
crps_distr[_key].append(_crps2)
|
2017-03-22 06:17:06 +04:00
|
|
|
times1[_key] = _tdiff + _t1
|
|
|
|
times2[_key] = _tdiff + _t2
|
|
|
|
|
|
|
|
if dump: print(_crps1, _crps2, _tdiff, _t1, _t2)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-05-10 02:04:51 +04:00
|
|
|
return bUtil.save_dataframe_ahead(experiments, file, objs, crps_interval, crps_distr, times1, times2, save, synthetic)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
|
|
|
|
def all_ahead_forecasters(data_train, data_test, partitions, start, steps, resolution = None, max_order=3,save=False, file=None, tam=[20, 5],
|
|
|
|
models=None, transformation=None, option=2):
|
|
|
|
if models is None:
|
|
|
|
models = [pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
|
|
|
if resolution is None: resolution = (max(data_train) - min(data_train)) / 100
|
|
|
|
|
|
|
|
objs = []
|
|
|
|
|
|
|
|
data_train_fs = Grid.GridPartitioner(data_train, partitions, transformation=transformation).sets
|
|
|
|
lcolors = []
|
|
|
|
|
|
|
|
for count, model in Util.enumerate2(models, start=0, step=2):
|
|
|
|
mfts = model("")
|
2017-05-03 00:16:49 +04:00
|
|
|
if not mfts.is_high_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
mfts.train(data_train, data_train_fs)
|
|
|
|
objs.append(mfts)
|
|
|
|
lcolors.append( colors[count % ncol] )
|
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
2017-05-03 00:16:49 +04:00
|
|
|
if order >= mfts.min_order:
|
2017-03-03 15:53:55 +04:00
|
|
|
mfts = model(" n = " + str(order))
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
mfts.train(data_train, data_train_fs, order=order)
|
|
|
|
objs.append(mfts)
|
|
|
|
lcolors.append(colors[count % ncol])
|
|
|
|
|
|
|
|
distributions = [False for k in objs]
|
|
|
|
|
|
|
|
distributions[0] = True
|
|
|
|
|
|
|
|
print_distribution_statistics(data_test[start:], objs, steps, resolution)
|
|
|
|
|
2017-05-02 03:56:47 +04:00
|
|
|
plot_compared_intervals_ahead(data_test, objs, lcolors, distributions=distributions, time_from=start, time_to=steps,
|
2017-03-22 06:17:06 +04:00
|
|
|
interpol=False, save=save, file=file, tam=tam, resolution=resolution, option=option)
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
def print_distribution_statistics(original, models, steps, resolution):
|
|
|
|
ret = "Model & Order & Interval & Distribution \\\\ \n"
|
|
|
|
for fts in models:
|
2017-05-07 04:19:04 +04:00
|
|
|
_crps1, _crps2, _t1, _t2 = Measures.get_distribution_statistics(original, fts, steps, resolution)
|
2017-03-03 15:53:55 +04:00
|
|
|
ret += fts.shortname + " & "
|
|
|
|
ret += str(fts.order) + " & "
|
|
|
|
ret += str(_crps1) + " & "
|
|
|
|
ret += str(_crps2) + " \\\\ \n"
|
|
|
|
print(ret)
|
|
|
|
|
|
|
|
|
2017-05-02 03:56:47 +04:00
|
|
|
def plot_compared_intervals_ahead(original, models, colors, distributions, time_from, time_to,
|
2017-03-03 15:53:55 +04:00
|
|
|
interpol=False, save=False, file=None, tam=[20, 5], resolution=None,
|
|
|
|
cmap='Blues',option=2):
|
2017-01-13 01:25:10 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-11 00:05:51 +04:00
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
cm = plt.get_cmap(cmap)
|
|
|
|
cNorm = pltcolors.Normalize(vmin=0, vmax=1)
|
|
|
|
scalarMap = cmx.ScalarMappable(norm=cNorm, cmap=cm)
|
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
if resolution is None: resolution = (max(original) - min(original)) / 100
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
for count, fts in enumerate(models, start=0):
|
2017-05-03 00:16:49 +04:00
|
|
|
if fts.has_probability_forecasting and distributions[count]:
|
2017-04-15 02:57:39 +04:00
|
|
|
density = fts.forecastAheadDistribution(original[time_from - fts.order:time_from], time_to,
|
2017-05-02 03:56:47 +04:00
|
|
|
resolution=resolution, method=option)
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
Y = []
|
|
|
|
X = []
|
|
|
|
C = []
|
|
|
|
S = []
|
2017-01-13 01:25:10 +04:00
|
|
|
y = density.columns
|
|
|
|
t = len(y)
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
ss = time_to ** 2
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in density.index:
|
2017-03-03 15:53:55 +04:00
|
|
|
#alpha = [scalarMap.to_rgba(density[col][k]) for col in density.columns]
|
|
|
|
col = [density[col][k]*5 for col in density.columns]
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
x = [time_from + k for x in np.arange(0, t)]
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
s = [ss for x in np.arange(0, t)]
|
|
|
|
|
|
|
|
ic = resolution/10
|
|
|
|
|
|
|
|
for cc in np.arange(0, resolution, ic):
|
|
|
|
Y.append(y + cc)
|
|
|
|
X.append(x)
|
|
|
|
C.append(col)
|
|
|
|
S.append(s)
|
|
|
|
|
|
|
|
Y = np.hstack(Y)
|
|
|
|
X = np.hstack(X)
|
|
|
|
C = np.hstack(C)
|
|
|
|
S = np.hstack(S)
|
|
|
|
|
|
|
|
s = ax.scatter(X, Y, c=C, marker='s',s=S, linewidths=0, edgecolors=None, cmap=cmap)
|
|
|
|
s.set_clim([0, 1])
|
|
|
|
cb = fig.colorbar(s)
|
|
|
|
|
|
|
|
cb.set_label('Density')
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-05-03 00:16:49 +04:00
|
|
|
if fts.has_interval_forecasting:
|
2017-01-13 01:25:10 +04:00
|
|
|
forecasts = fts.forecastAheadInterval(original[time_from - fts.order:time_from], time_to)
|
2017-01-11 00:05:51 +04:00
|
|
|
lower = [kk[0] for kk in forecasts]
|
|
|
|
upper = [kk[1] for kk in forecasts]
|
|
|
|
mi.append(min(lower))
|
|
|
|
ma.append(max(upper))
|
2017-01-23 00:41:42 +04:00
|
|
|
for k in np.arange(0, time_from - fts.order):
|
2017-01-11 00:05:51 +04:00
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
ax.plot(lower, color=colors[count], label=fts.shortname)
|
|
|
|
ax.plot(upper, color=colors[count])
|
|
|
|
|
|
|
|
else:
|
|
|
|
forecasts = fts.forecast(original)
|
|
|
|
mi.append(min(forecasts))
|
|
|
|
ma.append(max(forecasts))
|
|
|
|
for k in np.arange(0, time_from):
|
|
|
|
forecasts.insert(0, None)
|
|
|
|
ax.plot(forecasts, color=colors[count], label=fts.shortname)
|
|
|
|
|
|
|
|
ax.plot(original, color='black', label="Original")
|
2017-01-13 01:25:10 +04:00
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0, loc=2)
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax.set_title(fts.name)
|
2017-01-30 03:59:50 +04:00
|
|
|
_mi = min(mi)
|
|
|
|
if _mi < 0:
|
|
|
|
_mi *= 1.1
|
|
|
|
else:
|
|
|
|
_mi *= 0.9
|
|
|
|
_ma = max(ma)
|
|
|
|
if _ma < 0:
|
|
|
|
_ma *= 0.9
|
|
|
|
else:
|
|
|
|
_ma *= 1.1
|
|
|
|
|
|
|
|
ax.set_ylim([_mi, _ma])
|
2017-01-11 00:05:51 +04:00
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
#plt.colorbar()
|
|
|
|
|
2017-01-20 19:51:20 +04:00
|
|
|
Util.showAndSaveImage(fig, file, save)
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
def plotCompared(original, forecasts, labels, title):
|
|
|
|
fig = plt.figure(figsize=[13, 6])
|
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
ax.plot(original, color='k', label="Original")
|
|
|
|
for c in range(0, len(forecasts)):
|
|
|
|
ax.plot(forecasts[c], label=labels[c])
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0)
|
|
|
|
ax.set_title(title)
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
ax.set_ylim([min(original), max(original)])
|
|
|
|
|
2017-05-08 20:12:08 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
def SelecaoSimples_MenorRMSE(original, parameters, modelo):
|
|
|
|
ret = []
|
|
|
|
errors = []
|
|
|
|
forecasted_best = []
|
|
|
|
print("Série Original")
|
|
|
|
fig = plt.figure(figsize=[20, 12])
|
|
|
|
fig.suptitle("Comparação de modelos ")
|
|
|
|
ax0 = fig.add_axes([0, 0.5, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax0.set_xlim([0, len(original)])
|
|
|
|
ax0.set_ylim([min(original), max(original)])
|
|
|
|
ax0.set_title('Série Temporal')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
ax0.plot(original, label="Original")
|
|
|
|
min_rmse = 100000.0
|
|
|
|
best = None
|
|
|
|
for p in parameters:
|
2017-05-03 00:16:49 +04:00
|
|
|
sets = Grid.GridPartitioner(original, p).sets
|
2017-01-11 00:05:51 +04:00
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(original, sets)
|
|
|
|
# print(original)
|
|
|
|
forecasted = fts.forecast(original)
|
|
|
|
forecasted.insert(0, original[0])
|
|
|
|
# print(forecasted)
|
|
|
|
ax0.plot(forecasted, label=fts.name)
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(original))
|
|
|
|
print(p, error)
|
|
|
|
errors.append(error)
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best = fts
|
|
|
|
forecasted_best = forecasted
|
|
|
|
handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
ax0.legend(handles0, labels0)
|
|
|
|
ax1 = fig.add_axes([0.7, 0.5, 0.3, 0.45]) # left, bottom, width, height
|
|
|
|
ax1.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax1.set_ylabel('RMSE')
|
|
|
|
ax1.set_xlabel('Quantidade de Partições')
|
|
|
|
ax1.set_xlim([min(parameters), max(parameters)])
|
|
|
|
ax1.plot(parameters, errors)
|
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2016-09-02 22:55:55 +04:00
|
|
|
# Modelo diferencial
|
2017-01-11 00:05:51 +04:00
|
|
|
print("\nSérie Diferencial")
|
|
|
|
difffts = Transformations.differential(original)
|
|
|
|
errors = []
|
|
|
|
forecastedd_best = []
|
|
|
|
ax2 = fig.add_axes([0, 0, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax2.set_xlim([0, len(difffts)])
|
|
|
|
ax2.set_ylim([min(difffts), max(difffts)])
|
|
|
|
ax2.set_title('Série Temporal')
|
|
|
|
ax2.set_ylabel('F(T)')
|
|
|
|
ax2.set_xlabel('T')
|
|
|
|
ax2.plot(difffts, label="Original")
|
|
|
|
min_rmse = 100000.0
|
|
|
|
bestd = None
|
|
|
|
for p in parameters:
|
|
|
|
sets = Grid.GridPartitionerTrimf(difffts, p)
|
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(difffts, sets)
|
|
|
|
forecasted = fts.forecast(difffts)
|
|
|
|
forecasted.insert(0, difffts[0])
|
|
|
|
ax2.plot(forecasted, label=fts.name)
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(difffts))
|
|
|
|
print(p, error)
|
|
|
|
errors.append(error)
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
bestd = fts
|
|
|
|
forecastedd_best = forecasted
|
|
|
|
handles0, labels0 = ax2.get_legend_handles_labels()
|
|
|
|
ax2.legend(handles0, labels0)
|
|
|
|
ax3 = fig.add_axes([0.7, 0, 0.3, 0.45]) # left, bottom, width, height
|
|
|
|
ax3.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax3.set_ylabel('RMSE')
|
|
|
|
ax3.set_xlabel('Quantidade de Partições')
|
|
|
|
ax3.set_xlim([min(parameters), max(parameters)])
|
|
|
|
ax3.plot(parameters, errors)
|
|
|
|
ret.append(bestd)
|
|
|
|
ret.append(forecastedd_best)
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
|
|
|
def compareModelsPlot(original, models_fo, models_ho):
|
|
|
|
fig = plt.figure(figsize=[13, 6])
|
2016-09-02 22:55:55 +04:00
|
|
|
fig.suptitle("Comparação de modelos ")
|
2017-01-11 00:05:51 +04:00
|
|
|
ax0 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height
|
2016-09-02 22:55:55 +04:00
|
|
|
rows = []
|
|
|
|
for model in models_fo:
|
|
|
|
fts = model["model"]
|
2016-10-18 21:45:07 +04:00
|
|
|
ax0.plot(model["forecasted"], label=model["name"])
|
2016-09-02 22:55:55 +04:00
|
|
|
for model in models_ho:
|
|
|
|
fts = model["model"]
|
2016-10-18 21:45:07 +04:00
|
|
|
ax0.plot(model["forecasted"], label=model["name"])
|
2016-09-02 22:55:55 +04:00
|
|
|
handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
ax0.legend(handles0, labels0)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
|
|
|
|
def compareModelsTable(original, models_fo, models_ho):
|
|
|
|
fig = plt.figure(figsize=[12, 4])
|
2016-09-02 22:55:55 +04:00
|
|
|
fig.suptitle("Comparação de modelos ")
|
2017-01-11 00:05:51 +04:00
|
|
|
columns = ['Modelo', 'Ordem', 'Partições', 'RMSE', 'MAPE (%)']
|
2016-09-02 22:55:55 +04:00
|
|
|
rows = []
|
|
|
|
for model in models_fo:
|
|
|
|
fts = model["model"]
|
2017-01-11 00:05:51 +04:00
|
|
|
error_r = Measures.rmse(model["forecasted"], original)
|
|
|
|
error_m = round(Measures.mape(model["forecasted"], original) * 100, 2)
|
|
|
|
rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m])
|
2016-09-02 22:55:55 +04:00
|
|
|
for model in models_ho:
|
|
|
|
fts = model["model"]
|
2017-01-11 00:05:51 +04:00
|
|
|
error_r = Measures.rmse(model["forecasted"][fts.order:], original[fts.order:])
|
|
|
|
error_m = round(Measures.mape(model["forecasted"][fts.order:], original[fts.order:]) * 100, 2)
|
|
|
|
rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m])
|
|
|
|
ax1 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height
|
2016-09-02 22:55:55 +04:00
|
|
|
ax1.set_xticks([])
|
|
|
|
ax1.set_yticks([])
|
|
|
|
ax1.table(cellText=rows,
|
2017-01-11 00:05:51 +04:00
|
|
|
colLabels=columns,
|
|
|
|
cellLoc='center',
|
|
|
|
bbox=[0, 0, 1, 1])
|
2016-09-02 22:55:55 +04:00
|
|
|
sup = "\\begin{tabular}{"
|
|
|
|
header = ""
|
|
|
|
body = ""
|
|
|
|
footer = ""
|
|
|
|
|
|
|
|
for c in columns:
|
|
|
|
sup = sup + "|c"
|
|
|
|
if len(header) > 0:
|
|
|
|
header = header + " & "
|
|
|
|
header = header + "\\textbf{" + c + "} "
|
|
|
|
sup = sup + "|} \\hline\n"
|
2017-01-11 00:05:51 +04:00
|
|
|
header = header + "\\\\ \\hline \n"
|
|
|
|
|
2016-09-02 22:55:55 +04:00
|
|
|
for r in rows:
|
|
|
|
lin = ""
|
|
|
|
for c in r:
|
|
|
|
if len(lin) > 0:
|
|
|
|
lin = lin + " & "
|
|
|
|
lin = lin + str(c)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
body = body + lin + "\\\\ \\hline \n"
|
|
|
|
|
2016-09-02 22:55:55 +04:00
|
|
|
return sup + header + body + "\\end{tabular}"
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
def simpleSearch_RMSE(train, test, model, partitions, orders, save=False, file=None, tam=[10, 15],
|
2017-03-22 06:17:06 +04:00
|
|
|
plotforecasts=False, elev=30, azim=144, intervals=False,parameters=None,
|
|
|
|
partitioner=Grid.GridPartitioner,transformation=None,indexer=None):
|
2017-03-03 15:53:55 +04:00
|
|
|
_3d = len(orders) > 1
|
2017-01-11 00:05:51 +04:00
|
|
|
ret = []
|
2017-01-21 06:38:32 +04:00
|
|
|
errors = np.array([[0 for k in range(len(partitions))] for kk in range(len(orders))])
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted_best = []
|
2017-01-21 06:38:32 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-23 00:41:42 +04:00
|
|
|
# fig.suptitle("Comparação de modelos ")
|
2017-01-21 06:38:32 +04:00
|
|
|
if plotforecasts:
|
2017-01-23 00:41:42 +04:00
|
|
|
ax0 = fig.add_axes([0, 0.4, 0.9, 0.5]) # left, bottom, width, height
|
2017-01-27 14:26:47 +04:00
|
|
|
ax0.set_xlim([0, len(train)])
|
|
|
|
ax0.set_ylim([min(train) * 0.9, max(train) * 1.1])
|
2017-01-21 06:38:32 +04:00
|
|
|
ax0.set_title('Forecasts')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
min_rmse = 1000000.0
|
2017-01-11 00:05:51 +04:00
|
|
|
best = None
|
2017-01-30 03:59:50 +04:00
|
|
|
|
|
|
|
for pc, p in enumerate(partitions, start=0):
|
|
|
|
|
2017-03-22 06:17:06 +04:00
|
|
|
sets = partitioner(train, p, transformation=transformation).sets
|
2017-01-30 03:59:50 +04:00
|
|
|
for oc, o in enumerate(orders, start=0):
|
2017-01-21 06:38:32 +04:00
|
|
|
fts = model("q = " + str(p) + " n = " + str(o))
|
2017-03-22 06:17:06 +04:00
|
|
|
fts.appendTransformation(transformation)
|
|
|
|
fts.train(train, sets, o, parameters=parameters)
|
2017-01-30 03:59:50 +04:00
|
|
|
if not intervals:
|
|
|
|
forecasted = fts.forecast(test)
|
2017-05-03 00:16:49 +04:00
|
|
|
if not fts.has_seasonality:
|
2017-03-03 15:53:55 +04:00
|
|
|
error = Measures.rmse(np.array(test[o:]), np.array(forecasted[:-1]))
|
|
|
|
else:
|
|
|
|
error = Measures.rmse(np.array(test[o:]), np.array(forecasted))
|
2017-01-30 03:59:50 +04:00
|
|
|
for kk in range(o):
|
|
|
|
forecasted.insert(0, None)
|
2017-03-03 15:53:55 +04:00
|
|
|
if plotforecasts: ax0.plot(forecasted, label=fts.name)
|
2017-01-30 03:59:50 +04:00
|
|
|
else:
|
|
|
|
forecasted = fts.forecastInterval(test)
|
|
|
|
error = 1.0 - Measures.rmse_interval(np.array(test[o:]), np.array(forecasted[:-1]))
|
2017-01-11 00:05:51 +04:00
|
|
|
errors[oc, pc] = error
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best = fts
|
|
|
|
forecasted_best = forecasted
|
2017-01-30 03:59:50 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
# print(min_rmse)
|
2017-01-21 06:38:32 +04:00
|
|
|
if plotforecasts:
|
2017-01-23 00:41:42 +04:00
|
|
|
# handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
# ax0.legend(handles0, labels0)
|
2017-01-27 14:26:47 +04:00
|
|
|
ax0.plot(test, label="Original", linewidth=3.0, color="black")
|
2017-03-03 15:53:55 +04:00
|
|
|
if _3d: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
2017-01-21 06:38:32 +04:00
|
|
|
if not plotforecasts: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax1 = fig.add_axes([0.6, 0.5, 0.45, 0.45], projection='3d')
|
2017-03-03 15:53:55 +04:00
|
|
|
if _3d:
|
|
|
|
ax1.set_title('Error Surface')
|
|
|
|
ax1.set_ylabel('Model order')
|
|
|
|
ax1.set_xlabel('Number of partitions')
|
|
|
|
ax1.set_zlabel('RMSE')
|
|
|
|
X, Y = np.meshgrid(partitions, orders)
|
|
|
|
surf = ax1.plot_surface(X, Y, errors, rstride=1, cstride=1, antialiased=True)
|
|
|
|
else:
|
|
|
|
ax1 = fig.add_axes([0, 1, 0.9, 0.9])
|
|
|
|
ax1.set_title('Error Curve')
|
|
|
|
ax1.set_ylabel('Number of partitions')
|
|
|
|
ax1.set_xlabel('RMSE')
|
|
|
|
ax0.plot(errors,partitions)
|
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2017-03-22 06:17:06 +04:00
|
|
|
ret.append(min_rmse)
|
2017-03-03 15:53:55 +04:00
|
|
|
|
|
|
|
# plt.tight_layout()
|
|
|
|
|
|
|
|
Util.showAndSaveImage(fig, file, save)
|
|
|
|
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
|
|
|
def sliding_window_simple_search(data, windowsize, model, partitions, orders, save=False, file=None, tam=[10, 15],
|
|
|
|
plotforecasts=False, elev=30, azim=144, intervals=False, parameters=None):
|
|
|
|
_3d = len(orders) > 1
|
|
|
|
ret = []
|
|
|
|
errors = np.array([[0 for k in range(len(partitions))] for kk in range(len(orders))])
|
|
|
|
forecasted_best = []
|
|
|
|
fig = plt.figure(figsize=tam)
|
|
|
|
# fig.suptitle("Comparação de modelos ")
|
|
|
|
if plotforecasts:
|
|
|
|
ax0 = fig.add_axes([0, 0.4, 0.9, 0.5]) # left, bottom, width, height
|
|
|
|
ax0.set_xlim([0, len(data)])
|
|
|
|
ax0.set_ylim([min(data) * 0.9, max(data) * 1.1])
|
|
|
|
ax0.set_title('Forecasts')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
min_rmse = 1000000.0
|
|
|
|
best = None
|
|
|
|
|
|
|
|
for pc, p in enumerate(partitions, start=0):
|
|
|
|
|
|
|
|
sets = Grid.GridPartitioner(data, p).sets
|
|
|
|
for oc, o in enumerate(orders, start=0):
|
|
|
|
_error = []
|
|
|
|
for ct, train, test in Util.sliding_window(data, windowsize, 0.8):
|
|
|
|
fts = model("q = " + str(p) + " n = " + str(o))
|
|
|
|
fts.train(data, sets, o, parameters=parameters)
|
|
|
|
if not intervals:
|
|
|
|
forecasted = fts.forecast(test)
|
2017-05-03 00:16:49 +04:00
|
|
|
if not fts.has_seasonality:
|
2017-03-03 15:53:55 +04:00
|
|
|
_error.append( Measures.rmse(np.array(test[o:]), np.array(forecasted[:-1])) )
|
|
|
|
else:
|
|
|
|
_error.append( Measures.rmse(np.array(test[o:]), np.array(forecasted)) )
|
|
|
|
for kk in range(o):
|
|
|
|
forecasted.insert(0, None)
|
|
|
|
if plotforecasts: ax0.plot(forecasted, label=fts.name)
|
|
|
|
else:
|
|
|
|
forecasted = fts.forecastInterval(test)
|
|
|
|
_error.append( 1.0 - Measures.rmse_interval(np.array(test[o:]), np.array(forecasted[:-1])) )
|
|
|
|
error = np.nanmean(_error)
|
|
|
|
errors[oc, pc] = error
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best = fts
|
|
|
|
forecasted_best = forecasted
|
|
|
|
|
|
|
|
# print(min_rmse)
|
|
|
|
if plotforecasts:
|
|
|
|
# handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
# ax0.legend(handles0, labels0)
|
|
|
|
ax0.plot(test, label="Original", linewidth=3.0, color="black")
|
|
|
|
if _3d: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
|
|
|
if not plotforecasts: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
|
|
|
# ax1 = fig.add_axes([0.6, 0.5, 0.45, 0.45], projection='3d')
|
|
|
|
if _3d:
|
|
|
|
ax1.set_title('Error Surface')
|
|
|
|
ax1.set_ylabel('Model order')
|
|
|
|
ax1.set_xlabel('Number of partitions')
|
|
|
|
ax1.set_zlabel('RMSE')
|
|
|
|
X, Y = np.meshgrid(partitions, orders)
|
|
|
|
surf = ax1.plot_surface(X, Y, errors, rstride=1, cstride=1, antialiased=True)
|
|
|
|
else:
|
|
|
|
ax1 = fig.add_axes([0, 1, 0.9, 0.9])
|
|
|
|
ax1.set_title('Error Curve')
|
|
|
|
ax1.set_ylabel('Number of partitions')
|
|
|
|
ax1.set_xlabel('RMSE')
|
|
|
|
ax0.plot(errors,partitions)
|
2017-01-11 00:05:51 +04:00
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2016-09-02 22:55:55 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
# plt.tight_layout()
|
|
|
|
|
|
|
|
Util.showAndSaveImage(fig, file, save)
|
2017-01-21 06:38:32 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
return ret
|
2017-01-23 00:41:42 +04:00
|
|
|
|
|
|
|
|
|
|
|
def pftsExploreOrderAndPartitions(data,save=False, file=None):
|
|
|
|
fig, axes = plt.subplots(nrows=4, ncols=1, figsize=[6, 8])
|
2017-05-03 00:16:49 +04:00
|
|
|
data_fs1 = Grid.GridPartitioner(data, 10).sets
|
2017-01-23 00:41:42 +04:00
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
|
|
|
axes[0].set_title('Point Forecasts by Order')
|
|
|
|
axes[2].set_title('Interval Forecasts by Order')
|
|
|
|
|
|
|
|
for order in np.arange(1, 6):
|
2017-03-03 15:53:55 +04:00
|
|
|
fts = pwfts.ProbabilisticWeightedFTS("")
|
2017-01-23 00:41:42 +04:00
|
|
|
fts.shortname = "n = " + str(order)
|
|
|
|
fts.train(data, data_fs1, order=order)
|
|
|
|
point_forecasts = fts.forecast(data)
|
|
|
|
interval_forecasts = fts.forecastInterval(data)
|
|
|
|
lower = [kk[0] for kk in interval_forecasts]
|
|
|
|
upper = [kk[1] for kk in interval_forecasts]
|
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
|
|
|
for k in np.arange(0, order):
|
|
|
|
point_forecasts.insert(0, None)
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
axes[0].plot(point_forecasts, label=fts.shortname)
|
|
|
|
axes[2].plot(lower, label=fts.shortname)
|
|
|
|
axes[2].plot(upper)
|
|
|
|
|
|
|
|
axes[1].set_title('Point Forecasts by Number of Partitions')
|
|
|
|
axes[3].set_title('Interval Forecasts by Number of Partitions')
|
|
|
|
|
|
|
|
for partitions in np.arange(5, 11):
|
2017-05-03 00:16:49 +04:00
|
|
|
data_fs = Grid.GridPartitioner(data, partitions).sets
|
2017-03-03 15:53:55 +04:00
|
|
|
fts = pwfts.ProbabilisticWeightedFTS("")
|
2017-01-23 00:41:42 +04:00
|
|
|
fts.shortname = "q = " + str(partitions)
|
|
|
|
fts.train(data, data_fs, 1)
|
|
|
|
point_forecasts = fts.forecast(data)
|
|
|
|
interval_forecasts = fts.forecastInterval(data)
|
|
|
|
lower = [kk[0] for kk in interval_forecasts]
|
|
|
|
upper = [kk[1] for kk in interval_forecasts]
|
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
|
|
|
point_forecasts.insert(0, None)
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
axes[1].plot(point_forecasts, label=fts.shortname)
|
|
|
|
axes[3].plot(lower, label=fts.shortname)
|
|
|
|
axes[3].plot(upper)
|
|
|
|
|
|
|
|
for ax in axes:
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.plot(data, label="Original", color="black", linewidth=1.5)
|
|
|
|
handles, labels = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles, labels, loc=2, bbox_to_anchor=(1, 1))
|
|
|
|
ax.set_ylim([min(mi), max(ma)])
|
|
|
|
ax.set_xlim([0, len(data)])
|
|
|
|
|
|
|
|
plt.tight_layout()
|
|
|
|
|
2017-03-03 15:53:55 +04:00
|
|
|
Util.showAndSaveImage(fig, file, save)
|
|
|
|
|