2017-01-14 03:42:00 +04:00
|
|
|
#!/usr/bin/python
|
|
|
|
# -*- coding: utf8 -*-
|
|
|
|
|
2016-09-08 01:51:00 +04:00
|
|
|
import numpy as np
|
|
|
|
import pandas as pd
|
2016-09-08 16:03:32 +04:00
|
|
|
import matplotlib as plt
|
2016-11-08 20:08:06 +04:00
|
|
|
import matplotlib.colors as pltcolors
|
2017-02-01 20:54:37 +04:00
|
|
|
import matplotlib.cm as cmx
|
2016-09-08 01:51:00 +04:00
|
|
|
import matplotlib.pyplot as plt
|
|
|
|
from mpl_toolkits.mplot3d import Axes3D
|
2017-01-23 00:41:42 +04:00
|
|
|
# from sklearn.cross_validation import KFold
|
2017-02-16 00:16:13 +04:00
|
|
|
from pyFTS.benchmarks import Measures, naive, ResidualAnalysis, ProbabilityDistribution
|
2016-12-22 17:04:33 +04:00
|
|
|
from pyFTS.partitioners import Grid
|
2017-01-20 19:51:20 +04:00
|
|
|
from pyFTS.common import Membership, FuzzySet, FLR, Transformations, Util
|
2017-02-22 23:18:15 +04:00
|
|
|
from pyFTS import fts, chen, yu, ismailefendi, sadaei, hofts, hwang, pwfts, ifts
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
colors = ['grey', 'rosybrown', 'maroon', 'red','orange', 'yellow', 'olive', 'green',
|
|
|
|
'cyan', 'blue', 'darkblue', 'purple', 'darkviolet']
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
ncol = len(colors)
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
styles = ['-','--','-.',':','.']
|
|
|
|
|
|
|
|
nsty = len(styles)
|
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
def get_point_methods():
|
|
|
|
return [naive.Naive, chen.ConventionalFTS, yu.WeightedFTS, ismailefendi.ImprovedWeightedFTS,
|
|
|
|
sadaei.ExponentialyWeightedFTS, hofts.HighOrderFTS, pwfts.ProbabilisticWeightedFTS]
|
|
|
|
|
|
|
|
|
|
|
|
def sliding_window(data, windowsize, train=0.8,models=None,partitioners=[Grid.GridPartitioner],
|
|
|
|
partitions=[10], max_order=3,transformation=None,indexer=None):
|
|
|
|
if models is None:
|
|
|
|
models = get_point_methods()
|
|
|
|
|
|
|
|
objs = {}
|
|
|
|
lcolors = {}
|
|
|
|
rmse = {}
|
|
|
|
smape = {}
|
|
|
|
u = {}
|
|
|
|
|
|
|
|
for train,test in Util.sliding_window(data, windowsize, train):
|
|
|
|
for partition in partitions:
|
|
|
|
for partitioner in partitioners:
|
|
|
|
pttr = str(partitioner.__module__).split('.')[-1]
|
|
|
|
if transformation is not None:
|
|
|
|
data_train_fs = partitioner(transformation.apply(train), partition)
|
|
|
|
else:
|
|
|
|
data_train_fs = partitioner(train, partition)
|
|
|
|
|
|
|
|
for count, model in enumerate(models, start=0):
|
|
|
|
|
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " " + pttr+ " q = " +str(partition)
|
|
|
|
|
|
|
|
mfts.partitioner = data_train_fs
|
|
|
|
if not mfts.isHighOrder:
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
rmse[_key] = []
|
|
|
|
smape[_key] = []
|
|
|
|
u[_key] = []
|
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
|
|
|
mfts.train(train, data_train_fs.sets)
|
|
|
|
|
|
|
|
_rmse, _smape, _u = get_point_statistics(test, mfts, indexer)
|
|
|
|
rmse[_key].append(_rmse)
|
|
|
|
smape[_key].append(_smape)
|
|
|
|
u[_key].append(_u)
|
|
|
|
|
|
|
|
else:
|
|
|
|
for order in np.arange(1, max_order + 1):
|
|
|
|
if order >= mfts.minOrder:
|
|
|
|
mfts = model("")
|
|
|
|
_key = mfts.shortname + " n = " + str(order) + " " + pttr + " q = " + str(partition)
|
|
|
|
mfts.partitioner = data_train_fs
|
|
|
|
|
|
|
|
if _key not in objs:
|
|
|
|
objs[_key] = mfts
|
|
|
|
lcolors[_key] = colors[count % ncol]
|
|
|
|
rmse[_key] = []
|
|
|
|
smape[_key] = []
|
|
|
|
u[_key] = []
|
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
|
|
|
|
mfts.train(train, data_train_fs.sets, order=order)
|
|
|
|
|
|
|
|
_rmse, _smape, _u = get_point_statistics(test, mfts, indexer)
|
|
|
|
rmse[_key].append(_rmse)
|
|
|
|
smape[_key].append(_smape)
|
|
|
|
u[_key].append(_u)
|
|
|
|
|
|
|
|
ret = "Model\t&Order\t&Scheme\t&Partitions\t&RMSE AVG\t&RMSE STD\t& SMAPE AVG\t& SMAPE STD\t& U AVG\t& U STD \\\\ \n"
|
|
|
|
for k in sorted(objs.keys()):
|
|
|
|
mfts = objs[k]
|
|
|
|
ret += mfts.shortname + "\t &"
|
|
|
|
ret += str( mfts.order ) + "\t &"
|
|
|
|
ret += mfts.partitioner.name + "\t &"
|
|
|
|
ret += str(mfts.partitioner.partitions) + "\t &"
|
|
|
|
ret += str(round(np.mean(rmse[k]),2)) + "\t &"
|
|
|
|
ret += str(round(np.std(rmse[k]), 2)) + "\t &"
|
|
|
|
ret += str(round(np.mean(smape[k]), 2)) + "\t &"
|
|
|
|
ret += str(round(np.std(smape[k]), 2)) + "\t &"
|
|
|
|
ret += str(round(np.mean(u[k]), 2)) + "\t &"
|
|
|
|
ret += str(round(np.std(u[k]), 2)) + "\t &"
|
|
|
|
ret += str(len(mfts)) + "\\\\ \n"
|
|
|
|
|
|
|
|
print(ret)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
def allPointForecasters(data_train, data_test, partitions, max_order=3, statistics=True, residuals=True,
|
2017-02-16 00:16:13 +04:00
|
|
|
series=True, save=False, file=None, tam=[20, 5], models=None, transformation=None,
|
|
|
|
distributions=False):
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
if models is None:
|
2017-02-24 20:29:55 +04:00
|
|
|
models = get_point_methods()
|
2017-01-27 14:26:47 +04:00
|
|
|
|
|
|
|
objs = []
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(transformation.apply(data_train),partitions)
|
|
|
|
else:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(data_train, partitions)
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-24 16:40:48 +04:00
|
|
|
count = 1
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors = []
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
for count, model in enumerate(models, start=0):
|
2017-01-24 16:40:48 +04:00
|
|
|
#print(model)
|
|
|
|
mfts = model("")
|
|
|
|
if not mfts.isHighOrder:
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-24 16:40:48 +04:00
|
|
|
mfts.train(data_train, data_train_fs)
|
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append( colors[count % ncol] )
|
2017-01-23 17:00:27 +04:00
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
2017-01-25 18:17:07 +04:00
|
|
|
if order >= mfts.minOrder:
|
|
|
|
mfts = model(" n = " + str(order))
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-25 18:17:07 +04:00
|
|
|
mfts.train(data_train, data_train_fs, order=order)
|
|
|
|
objs.append(mfts)
|
2017-02-16 00:16:13 +04:00
|
|
|
lcolors.append(colors[(count + order) % ncol])
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if statistics:
|
2017-02-24 20:29:55 +04:00
|
|
|
print(print_point_statistics(data_test, objs))
|
2017-01-23 17:00:27 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if residuals:
|
|
|
|
print(ResidualAnalysis.compareResiduals(data_test, objs))
|
2017-01-27 14:26:47 +04:00
|
|
|
ResidualAnalysis.plotResiduals2(data_test, objs, save=save, file=file, tam=tam)
|
2017-01-25 18:17:07 +04:00
|
|
|
|
2017-01-26 16:19:34 +04:00
|
|
|
if series:
|
2017-01-27 14:26:47 +04:00
|
|
|
plotComparedSeries(data_test, objs, lcolors, typeonlegend=False, save=save, file=file, tam=tam,
|
|
|
|
intervals=False)
|
2017-01-25 18:17:07 +04:00
|
|
|
|
2017-02-16 00:16:13 +04:00
|
|
|
if distributions:
|
|
|
|
lcolors.insert(0,'black')
|
|
|
|
pmfs = []
|
|
|
|
pmfs.append(
|
2017-02-22 23:18:15 +04:00
|
|
|
ProbabilityDistribution.ProbabilityDistribution("Original", 100, [min(data_test), max(data_test)], data=data_test) )
|
2017-02-16 00:16:13 +04:00
|
|
|
|
|
|
|
for m in objs:
|
2017-02-22 23:18:15 +04:00
|
|
|
forecasts = m.forecast(data_test)
|
2017-02-16 00:16:13 +04:00
|
|
|
pmfs.append(
|
2017-02-22 23:18:15 +04:00
|
|
|
ProbabilityDistribution.ProbabilityDistribution(m.shortname, 100, [min(data_test), max(data_test)],
|
2017-02-16 00:16:13 +04:00
|
|
|
data=forecasts))
|
2017-02-22 23:18:15 +04:00
|
|
|
print(getProbabilityDistributionStatistics(pmfs,data_test))
|
2017-02-16 00:16:13 +04:00
|
|
|
|
2017-02-22 23:18:15 +04:00
|
|
|
plotProbabilityDistributions(pmfs, lcolors,tam=tam)
|
2017-02-16 00:16:13 +04:00
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
def get_point_statistics(data, model, indexer=None):
|
|
|
|
if indexer is not None:
|
|
|
|
ndata = np.array(indexer.get_data(data[model.order:]))
|
|
|
|
else:
|
|
|
|
ndata = np.array(data[model.order:])
|
|
|
|
|
|
|
|
if model.isMultivariate or indexer is None:
|
|
|
|
forecasts = model.forecast(data)
|
|
|
|
elif not model.isMultivariate and indexer is not None:
|
|
|
|
forecasts = model.forecast(indexer.get_data(data))
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
if model.hasSeasonality:
|
|
|
|
nforecasts = np.array(forecasts)
|
|
|
|
else:
|
|
|
|
nforecasts = np.array(forecasts[:-1])
|
|
|
|
ret = list()
|
|
|
|
ret.append(np.round(Measures.rmse(ndata, nforecasts), 2))
|
|
|
|
ret.append(np.round(Measures.smape(ndata, nforecasts), 2))
|
|
|
|
ret.append(np.round(Measures.UStatistic(ndata, nforecasts), 2))
|
2017-02-10 17:09:59 +04:00
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
return ret
|
2017-02-10 17:09:59 +04:00
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
|
|
|
|
def print_point_statistics(data, models, externalmodels = None, externalforecasts = None, indexers=None):
|
|
|
|
ret = "Model & Order & RMSE & SMAPE & Theil's U \\\\ \n"
|
|
|
|
for count,model in enumerate(models,start=0):
|
|
|
|
_rmse, _smape, _u = get_point_statistics(data, model, indexers[count])
|
2017-02-09 17:04:48 +04:00
|
|
|
ret += model.shortname + " & "
|
|
|
|
ret += str(model.order) + " & "
|
2017-02-24 20:29:55 +04:00
|
|
|
ret += str(_rmse) + " & "
|
|
|
|
ret += str(_smape)+ " & "
|
|
|
|
ret += str(_u)
|
2017-01-27 14:26:47 +04:00
|
|
|
#ret += str(round(Measures.TheilsInequality(np.array(data[fts.order:]), np.array(forecasts[:-1])), 4))
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += " \\\\ \n"
|
|
|
|
if externalmodels is not None:
|
|
|
|
l = len(externalmodels)
|
|
|
|
for k in np.arange(0,l):
|
|
|
|
ret += externalmodels[k] + " & "
|
|
|
|
ret += " 1 & "
|
2017-02-09 17:04:48 +04:00
|
|
|
ret += str(round(Measures.rmse(data, externalforecasts[k][:-1]), 2)) + " & "
|
|
|
|
ret += str(round(Measures.smape(data, externalforecasts[k][:-1]), 2))+ " & "
|
2017-02-19 08:02:59 +04:00
|
|
|
ret += str(round(Measures.UStatistic(data, externalforecasts[k][:-1]), 2))
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += " \\\\ \n"
|
2017-01-23 17:00:27 +04:00
|
|
|
return ret
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
|
2017-02-16 00:16:13 +04:00
|
|
|
def getProbabilityDistributionStatistics(pmfs, data):
|
|
|
|
ret = "Model & Entropy & Empirical Likelihood & Pseudo Likelihood \\\\ \n"
|
|
|
|
for k in pmfs:
|
|
|
|
ret += k.name + " & "
|
|
|
|
ret += str(k.entropy()) + " & "
|
|
|
|
ret += str(k.empiricalloglikelihood())+ " & "
|
|
|
|
ret += str(k.pseudologlikelihood(data))
|
|
|
|
ret += " \\\\ \n"
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
def allIntervalForecasters(data_train, data_test, partitions, max_order=3,save=False, file=None, tam=[20, 5],
|
|
|
|
models=None, transformation=None):
|
|
|
|
if models is None:
|
2017-02-22 23:18:15 +04:00
|
|
|
models = [ifts.IntervalFTS, pwfts.ProbabilisticWeightedFTS]
|
2017-01-24 16:40:48 +04:00
|
|
|
|
|
|
|
objs = []
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(transformation.apply(data_train),partitions)
|
|
|
|
else:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(data_train, partitions)
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors = []
|
2017-01-24 16:40:48 +04:00
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
for count, model in Util.enumerate2(models, start=0, step=2):
|
2017-01-24 16:40:48 +04:00
|
|
|
mfts = model("")
|
|
|
|
if not mfts.isHighOrder:
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-24 16:40:48 +04:00
|
|
|
mfts.train(data_train, data_train_fs)
|
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append( colors[count % ncol] )
|
2017-01-24 16:40:48 +04:00
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
2017-01-25 18:17:07 +04:00
|
|
|
if order >= mfts.minOrder:
|
|
|
|
mfts = model(" n = " + str(order))
|
2017-01-27 14:26:47 +04:00
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
2017-01-25 18:17:07 +04:00
|
|
|
mfts.train(data_train, data_train_fs, order=order)
|
|
|
|
objs.append(mfts)
|
2017-01-27 14:26:47 +04:00
|
|
|
lcolors.append(colors[count % ncol])
|
|
|
|
|
2017-01-24 16:40:48 +04:00
|
|
|
print(getIntervalStatistics(data_test, objs))
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
plotComparedSeries(data_test, objs, lcolors, typeonlegend=False, save=save, file=file, tam=tam, intervals=True)
|
2017-01-24 16:40:48 +04:00
|
|
|
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
def getIntervalStatistics(original, models):
|
2017-01-30 03:59:50 +04:00
|
|
|
ret = "Model & Order & Sharpness & Resolution & Coverage \\\\ \n"
|
2017-01-11 00:05:51 +04:00
|
|
|
for fts in models:
|
2017-01-23 17:00:27 +04:00
|
|
|
forecasts = fts.forecastInterval(original)
|
|
|
|
ret += fts.shortname + " & "
|
2017-01-24 16:40:48 +04:00
|
|
|
ret += str(fts.order) + " & "
|
2017-01-23 17:00:27 +04:00
|
|
|
ret += str(round(Measures.sharpness(forecasts), 2)) + " & "
|
|
|
|
ret += str(round(Measures.resolution(forecasts), 2)) + " & "
|
2017-01-30 03:59:50 +04:00
|
|
|
ret += str(round(Measures.coverage(original[fts.order:], forecasts[:-1]), 2)) + " \\\\ \n"
|
2017-01-11 00:05:51 +04:00
|
|
|
return ret
|
|
|
|
|
|
|
|
|
2016-11-06 03:24:36 +04:00
|
|
|
def plotDistribution(dist):
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in dist.index:
|
|
|
|
alpha = np.array([dist[x][k] for x in dist]) * 100
|
|
|
|
x = [k for x in np.arange(0, len(alpha))]
|
|
|
|
y = dist.columns
|
|
|
|
plt.scatter(x, y, c=alpha, marker='s', linewidths=0, cmap='Oranges', norm=pltcolors.Normalize(vmin=0, vmax=1),
|
|
|
|
vmin=0, vmax=1, edgecolors=None)
|
|
|
|
|
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
def plotComparedSeries(original, models, colors, typeonlegend=False, save=False, file=None, tam=[20, 5],
|
2017-02-24 20:29:55 +04:00
|
|
|
points=True, intervals=True, linewidth=1.5):
|
2017-01-13 01:25:10 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-11 00:05:51 +04:00
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
legends = []
|
|
|
|
|
2017-02-24 20:29:55 +04:00
|
|
|
ax.plot(original, color='black', label="Original", linewidth=linewidth*1.5)
|
2017-01-27 14:26:47 +04:00
|
|
|
|
|
|
|
for count, fts in enumerate(models, start=0):
|
2017-02-24 20:29:55 +04:00
|
|
|
if fts.hasPointForecasting and points:
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted = fts.forecast(original)
|
2017-01-23 00:41:42 +04:00
|
|
|
mi.append(min(forecasted) * 0.95)
|
|
|
|
ma.append(max(forecasted) * 1.05)
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in np.arange(0, fts.order):
|
|
|
|
forecasted.insert(0, None)
|
2017-01-13 01:25:10 +04:00
|
|
|
lbl = fts.shortname
|
|
|
|
if typeonlegend: lbl += " (Point)"
|
2017-02-24 20:29:55 +04:00
|
|
|
ax.plot(forecasted, color=colors[count], label=lbl, ls="-",linewidth=linewidth)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-01-21 06:38:32 +04:00
|
|
|
if fts.hasIntervalForecasting and intervals:
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted = fts.forecastInterval(original)
|
|
|
|
lower = [kk[0] for kk in forecasted]
|
|
|
|
upper = [kk[1] for kk in forecasted]
|
2017-01-23 00:41:42 +04:00
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in np.arange(0, fts.order):
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
2017-01-13 01:25:10 +04:00
|
|
|
lbl = fts.shortname
|
|
|
|
if typeonlegend: lbl += " (Interval)"
|
2017-02-24 20:29:55 +04:00
|
|
|
ax.plot(lower, color=colors[count], label=lbl, ls="--",linewidth=linewidth)
|
|
|
|
ax.plot(upper, color=colors[count], ls="--",linewidth=linewidth)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
2017-01-27 14:26:47 +04:00
|
|
|
lgd = ax.legend(handles0, labels0, loc=2, bbox_to_anchor=(1, 1))
|
|
|
|
legends.append(lgd)
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax.set_title(fts.name)
|
|
|
|
ax.set_ylim([min(mi), max(ma)])
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
|
2017-01-27 14:26:47 +04:00
|
|
|
Util.showAndSaveImage(fig, file, save, lgd=legends)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-02-16 00:16:13 +04:00
|
|
|
|
2017-02-22 23:18:15 +04:00
|
|
|
def plotProbabilityDistributions(pmfs,lcolors,tam=[15, 7]):
|
|
|
|
fig = plt.figure(figsize=tam)
|
2017-02-16 00:16:13 +04:00
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
|
|
|
for k,m in enumerate(pmfs,start=0):
|
|
|
|
m.plot(ax, color=lcolors[k])
|
|
|
|
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0)
|
|
|
|
|
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
def allAheadForecasters(data_train, data_test, partitions, start, steps, resolution = None, max_order=3,save=False, file=None, tam=[20, 5],
|
|
|
|
models=None, transformation=None, option=2):
|
|
|
|
if models is None:
|
2017-02-22 23:18:15 +04:00
|
|
|
models = [pwfts.ProbabilisticWeightedFTS]
|
2017-02-01 20:54:37 +04:00
|
|
|
|
|
|
|
if resolution is None: resolution = (max(data_train) - min(data_train)) / 100
|
|
|
|
|
|
|
|
objs = []
|
|
|
|
|
|
|
|
if transformation is not None:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(transformation.apply(data_train),partitions)
|
|
|
|
else:
|
|
|
|
data_train_fs = Grid.GridPartitionerTrimf(data_train, partitions)
|
|
|
|
|
|
|
|
lcolors = []
|
|
|
|
|
|
|
|
for count, model in Util.enumerate2(models, start=0, step=2):
|
|
|
|
mfts = model("")
|
|
|
|
if not mfts.isHighOrder:
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
mfts.train(data_train, data_train_fs)
|
|
|
|
objs.append(mfts)
|
|
|
|
lcolors.append( colors[count % ncol] )
|
|
|
|
else:
|
|
|
|
for order in np.arange(1,max_order+1):
|
|
|
|
if order >= mfts.minOrder:
|
|
|
|
mfts = model(" n = " + str(order))
|
|
|
|
if transformation is not None:
|
|
|
|
mfts.appendTransformation(transformation)
|
|
|
|
mfts.train(data_train, data_train_fs, order=order)
|
|
|
|
objs.append(mfts)
|
|
|
|
lcolors.append(colors[count % ncol])
|
|
|
|
|
|
|
|
distributions = [False for k in objs]
|
|
|
|
|
|
|
|
distributions[0] = True
|
|
|
|
|
|
|
|
print(getDistributionStatistics(data_test[start:], objs, steps, resolution))
|
|
|
|
|
|
|
|
#plotComparedIntervalsAhead(data_test, objs, lcolors, distributions=, save=save, file=file, tam=tam, intervals=True)
|
|
|
|
|
|
|
|
|
|
|
|
def getDistributionStatistics(original, models, steps, resolution):
|
|
|
|
ret = "Model & Order & Interval & Distribution \\\\ \n"
|
|
|
|
for fts in models:
|
|
|
|
densities1 = fts.forecastAheadDistribution(original,steps,resolution, parameters=3)
|
|
|
|
densities2 = fts.forecastAheadDistribution(original, steps, resolution, parameters=2)
|
|
|
|
ret += fts.shortname + " & "
|
|
|
|
ret += str(fts.order) + " & "
|
|
|
|
ret += str(round(Measures.crps(original, densities1), 3)) + " & "
|
|
|
|
ret += str(round(Measures.crps(original, densities2), 3)) + " \\\\ \n"
|
|
|
|
return ret
|
|
|
|
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-14 03:42:00 +04:00
|
|
|
def plotComparedIntervalsAhead(original, models, colors, distributions, time_from, time_to,
|
2017-02-01 20:54:37 +04:00
|
|
|
interpol=False, save=False, file=None, tam=[20, 5], resolution=None,
|
|
|
|
cmap='Blues',option=2):
|
2017-01-13 01:25:10 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-11 00:05:51 +04:00
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
cm = plt.get_cmap(cmap)
|
|
|
|
cNorm = pltcolors.Normalize(vmin=0, vmax=1)
|
|
|
|
scalarMap = cmx.ScalarMappable(norm=cNorm, cmap=cm)
|
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
if resolution is None: resolution = (max(original) - min(original)) / 100
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
for count, fts in enumerate(models, start=0):
|
2017-01-11 00:05:51 +04:00
|
|
|
if fts.hasDistributionForecasting and distributions[count]:
|
2017-01-30 03:59:50 +04:00
|
|
|
density = fts.forecastAheadDistribution(original[time_from - fts.order:time_from],
|
2017-02-01 20:54:37 +04:00
|
|
|
time_to, resolution, parameters=option)
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
Y = []
|
|
|
|
X = []
|
|
|
|
C = []
|
|
|
|
S = []
|
2017-01-13 01:25:10 +04:00
|
|
|
y = density.columns
|
|
|
|
t = len(y)
|
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
ss = time_to ** 2
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
for k in density.index:
|
2017-02-01 20:54:37 +04:00
|
|
|
#alpha = [scalarMap.to_rgba(density[col][k]) for col in density.columns]
|
|
|
|
col = [density[col][k]*5 for col in density.columns]
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
x = [time_from + k for x in np.arange(0, t)]
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
s = [ss for x in np.arange(0, t)]
|
|
|
|
|
|
|
|
ic = resolution/10
|
|
|
|
|
|
|
|
for cc in np.arange(0, resolution, ic):
|
|
|
|
Y.append(y + cc)
|
|
|
|
X.append(x)
|
|
|
|
C.append(col)
|
|
|
|
S.append(s)
|
|
|
|
|
|
|
|
Y = np.hstack(Y)
|
|
|
|
X = np.hstack(X)
|
|
|
|
C = np.hstack(C)
|
|
|
|
S = np.hstack(S)
|
|
|
|
|
|
|
|
s = ax.scatter(X, Y, c=C, marker='s',s=S, linewidths=0, edgecolors=None, cmap=cmap)
|
|
|
|
s.set_clim([0, 1])
|
|
|
|
cb = fig.colorbar(s)
|
|
|
|
|
|
|
|
cb.set_label('Density')
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
if fts.hasIntervalForecasting:
|
2017-01-13 01:25:10 +04:00
|
|
|
forecasts = fts.forecastAheadInterval(original[time_from - fts.order:time_from], time_to)
|
2017-01-11 00:05:51 +04:00
|
|
|
lower = [kk[0] for kk in forecasts]
|
|
|
|
upper = [kk[1] for kk in forecasts]
|
|
|
|
mi.append(min(lower))
|
|
|
|
ma.append(max(upper))
|
2017-01-23 00:41:42 +04:00
|
|
|
for k in np.arange(0, time_from - fts.order):
|
2017-01-11 00:05:51 +04:00
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
ax.plot(lower, color=colors[count], label=fts.shortname)
|
|
|
|
ax.plot(upper, color=colors[count])
|
|
|
|
|
|
|
|
else:
|
|
|
|
forecasts = fts.forecast(original)
|
|
|
|
mi.append(min(forecasts))
|
|
|
|
ma.append(max(forecasts))
|
|
|
|
for k in np.arange(0, time_from):
|
|
|
|
forecasts.insert(0, None)
|
|
|
|
ax.plot(forecasts, color=colors[count], label=fts.shortname)
|
|
|
|
|
|
|
|
ax.plot(original, color='black', label="Original")
|
2017-01-13 01:25:10 +04:00
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0, loc=2)
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax.set_title(fts.name)
|
2017-01-30 03:59:50 +04:00
|
|
|
_mi = min(mi)
|
|
|
|
if _mi < 0:
|
|
|
|
_mi *= 1.1
|
|
|
|
else:
|
|
|
|
_mi *= 0.9
|
|
|
|
_ma = max(ma)
|
|
|
|
if _ma < 0:
|
|
|
|
_ma *= 0.9
|
|
|
|
else:
|
|
|
|
_ma *= 1.1
|
|
|
|
|
|
|
|
ax.set_ylim([_mi, _ma])
|
2017-01-11 00:05:51 +04:00
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
|
2017-02-01 20:54:37 +04:00
|
|
|
#plt.colorbar()
|
|
|
|
|
2017-01-20 19:51:20 +04:00
|
|
|
Util.showAndSaveImage(fig, file, save)
|
2017-01-13 01:25:10 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
def plotCompared(original, forecasts, labels, title):
|
|
|
|
fig = plt.figure(figsize=[13, 6])
|
|
|
|
ax = fig.add_subplot(111)
|
|
|
|
ax.plot(original, color='k', label="Original")
|
|
|
|
for c in range(0, len(forecasts)):
|
|
|
|
ax.plot(forecasts[c], label=labels[c])
|
|
|
|
handles0, labels0 = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles0, labels0)
|
|
|
|
ax.set_title(title)
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.set_xlim([0, len(original)])
|
|
|
|
ax.set_ylim([min(original), max(original)])
|
|
|
|
|
|
|
|
|
|
|
|
def SelecaoKFold_MenorRMSE(original, parameters, modelo):
|
|
|
|
nfolds = 5
|
|
|
|
ret = []
|
|
|
|
errors = np.array([[0 for k in parameters] for z in np.arange(0, nfolds)])
|
|
|
|
forecasted_best = []
|
|
|
|
print("Série Original")
|
|
|
|
fig = plt.figure(figsize=[18, 10])
|
|
|
|
fig.suptitle("Comparação de modelos ")
|
|
|
|
ax0 = fig.add_axes([0, 0.5, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax0.set_xlim([0, len(original)])
|
|
|
|
ax0.set_ylim([min(original), max(original)])
|
|
|
|
ax0.set_title('Série Temporal')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
ax0.plot(original, label="Original")
|
|
|
|
min_rmse_fold = 100000.0
|
|
|
|
best = None
|
|
|
|
fc = 0 # Fold count
|
|
|
|
kf = KFold(len(original), n_folds=nfolds)
|
|
|
|
for train_ix, test_ix in kf:
|
|
|
|
train = original[train_ix]
|
|
|
|
test = original[test_ix]
|
|
|
|
min_rmse = 100000.0
|
|
|
|
best_fold = None
|
|
|
|
forecasted_best_fold = []
|
|
|
|
errors_fold = []
|
|
|
|
pc = 0 # Parameter count
|
|
|
|
for p in parameters:
|
|
|
|
sets = Grid.GridPartitionerTrimf(train, p)
|
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(train, sets)
|
|
|
|
forecasted = [fts.forecast(xx) for xx in test]
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(test))
|
|
|
|
errors_fold.append(error)
|
|
|
|
print(fc, p, error)
|
|
|
|
errors[fc, pc] = error
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best_fold = fts
|
|
|
|
forecasted_best_fold = forecasted
|
|
|
|
pc = pc + 1
|
|
|
|
forecasted_best_fold = [best_fold.forecast(xx) for xx in original]
|
|
|
|
ax0.plot(forecasted_best_fold, label=best_fold.name)
|
|
|
|
if np.mean(errors_fold) < min_rmse_fold:
|
|
|
|
min_rmse_fold = np.mean(errors)
|
|
|
|
best = best_fold
|
|
|
|
forecasted_best = forecasted_best_fold
|
|
|
|
fc = fc + 1
|
|
|
|
handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
ax0.legend(handles0, labels0)
|
|
|
|
ax1 = Axes3D(fig, rect=[0.7, 0.5, 0.3, 0.45], elev=30, azim=144)
|
|
|
|
# ax1 = fig.add_axes([0.6, 0.0, 0.45, 0.45], projection='3d')
|
|
|
|
ax1.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax1.set_zlabel('RMSE')
|
|
|
|
ax1.set_xlabel('K-fold')
|
|
|
|
ax1.set_ylabel('Partições')
|
|
|
|
X, Y = np.meshgrid(np.arange(0, nfolds), parameters)
|
|
|
|
surf = ax1.plot_surface(X, Y, errors.T, rstride=1, cstride=1, antialiased=True)
|
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2016-09-02 22:55:55 +04:00
|
|
|
|
|
|
|
# Modelo diferencial
|
2017-01-11 00:05:51 +04:00
|
|
|
print("\nSérie Diferencial")
|
|
|
|
errors = np.array([[0 for k in parameters] for z in np.arange(0, nfolds)])
|
|
|
|
forecastedd_best = []
|
|
|
|
ax2 = fig.add_axes([0, 0, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax2.set_xlim([0, len(original)])
|
|
|
|
ax2.set_ylim([min(original), max(original)])
|
|
|
|
ax2.set_title('Série Temporal')
|
|
|
|
ax2.set_ylabel('F(T)')
|
|
|
|
ax2.set_xlabel('T')
|
|
|
|
ax2.plot(original, label="Original")
|
|
|
|
min_rmse = 100000.0
|
|
|
|
min_rmse_fold = 100000.0
|
|
|
|
bestd = None
|
|
|
|
fc = 0
|
|
|
|
diff = Transformations.differential(original)
|
|
|
|
kf = KFold(len(original), n_folds=nfolds)
|
|
|
|
for train_ix, test_ix in kf:
|
|
|
|
train = diff[train_ix]
|
|
|
|
test = diff[test_ix]
|
|
|
|
min_rmse = 100000.0
|
|
|
|
best_fold = None
|
|
|
|
forecasted_best_fold = []
|
|
|
|
errors_fold = []
|
|
|
|
pc = 0
|
|
|
|
for p in parameters:
|
|
|
|
sets = Grid.GridPartitionerTrimf(train, p)
|
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(train, sets)
|
|
|
|
forecasted = [fts.forecastDiff(test, xx) for xx in np.arange(len(test))]
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(test))
|
|
|
|
print(fc, p, error)
|
|
|
|
errors[fc, pc] = error
|
|
|
|
errors_fold.append(error)
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best_fold = fts
|
|
|
|
pc = pc + 1
|
|
|
|
forecasted_best_fold = [best_fold.forecastDiff(original, xx) for xx in np.arange(len(original))]
|
|
|
|
ax2.plot(forecasted_best_fold, label=best_fold.name)
|
|
|
|
if np.mean(errors_fold) < min_rmse_fold:
|
|
|
|
min_rmse_fold = np.mean(errors)
|
|
|
|
best = best_fold
|
|
|
|
forecasted_best = forecasted_best_fold
|
|
|
|
fc = fc + 1
|
|
|
|
handles0, labels0 = ax2.get_legend_handles_labels()
|
|
|
|
ax2.legend(handles0, labels0)
|
|
|
|
ax3 = Axes3D(fig, rect=[0.7, 0, 0.3, 0.45], elev=30, azim=144)
|
|
|
|
# ax1 = fig.add_axes([0.6, 0.0, 0.45, 0.45], projection='3d')
|
|
|
|
ax3.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax3.set_zlabel('RMSE')
|
|
|
|
ax3.set_xlabel('K-fold')
|
|
|
|
ax3.set_ylabel('Partições')
|
|
|
|
X, Y = np.meshgrid(np.arange(0, nfolds), parameters)
|
|
|
|
surf = ax3.plot_surface(X, Y, errors.T, rstride=1, cstride=1, antialiased=True)
|
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
|
|
|
def SelecaoSimples_MenorRMSE(original, parameters, modelo):
|
|
|
|
ret = []
|
|
|
|
errors = []
|
|
|
|
forecasted_best = []
|
|
|
|
print("Série Original")
|
|
|
|
fig = plt.figure(figsize=[20, 12])
|
|
|
|
fig.suptitle("Comparação de modelos ")
|
|
|
|
ax0 = fig.add_axes([0, 0.5, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax0.set_xlim([0, len(original)])
|
|
|
|
ax0.set_ylim([min(original), max(original)])
|
|
|
|
ax0.set_title('Série Temporal')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
ax0.plot(original, label="Original")
|
|
|
|
min_rmse = 100000.0
|
|
|
|
best = None
|
|
|
|
for p in parameters:
|
|
|
|
sets = Grid.GridPartitionerTrimf(original, p)
|
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(original, sets)
|
|
|
|
# print(original)
|
|
|
|
forecasted = fts.forecast(original)
|
|
|
|
forecasted.insert(0, original[0])
|
|
|
|
# print(forecasted)
|
|
|
|
ax0.plot(forecasted, label=fts.name)
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(original))
|
|
|
|
print(p, error)
|
|
|
|
errors.append(error)
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best = fts
|
|
|
|
forecasted_best = forecasted
|
|
|
|
handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
ax0.legend(handles0, labels0)
|
|
|
|
ax1 = fig.add_axes([0.7, 0.5, 0.3, 0.45]) # left, bottom, width, height
|
|
|
|
ax1.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax1.set_ylabel('RMSE')
|
|
|
|
ax1.set_xlabel('Quantidade de Partições')
|
|
|
|
ax1.set_xlim([min(parameters), max(parameters)])
|
|
|
|
ax1.plot(parameters, errors)
|
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2016-09-02 22:55:55 +04:00
|
|
|
# Modelo diferencial
|
2017-01-11 00:05:51 +04:00
|
|
|
print("\nSérie Diferencial")
|
|
|
|
difffts = Transformations.differential(original)
|
|
|
|
errors = []
|
|
|
|
forecastedd_best = []
|
|
|
|
ax2 = fig.add_axes([0, 0, 0.65, 0.45]) # left, bottom, width, height
|
|
|
|
ax2.set_xlim([0, len(difffts)])
|
|
|
|
ax2.set_ylim([min(difffts), max(difffts)])
|
|
|
|
ax2.set_title('Série Temporal')
|
|
|
|
ax2.set_ylabel('F(T)')
|
|
|
|
ax2.set_xlabel('T')
|
|
|
|
ax2.plot(difffts, label="Original")
|
|
|
|
min_rmse = 100000.0
|
|
|
|
bestd = None
|
|
|
|
for p in parameters:
|
|
|
|
sets = Grid.GridPartitionerTrimf(difffts, p)
|
|
|
|
fts = modelo(str(p) + " particoes")
|
|
|
|
fts.train(difffts, sets)
|
|
|
|
forecasted = fts.forecast(difffts)
|
|
|
|
forecasted.insert(0, difffts[0])
|
|
|
|
ax2.plot(forecasted, label=fts.name)
|
|
|
|
error = Measures.rmse(np.array(forecasted), np.array(difffts))
|
|
|
|
print(p, error)
|
|
|
|
errors.append(error)
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
bestd = fts
|
|
|
|
forecastedd_best = forecasted
|
|
|
|
handles0, labels0 = ax2.get_legend_handles_labels()
|
|
|
|
ax2.legend(handles0, labels0)
|
|
|
|
ax3 = fig.add_axes([0.7, 0, 0.3, 0.45]) # left, bottom, width, height
|
|
|
|
ax3.set_title('Comparação dos Erros Quadráticos Médios')
|
|
|
|
ax3.set_ylabel('RMSE')
|
|
|
|
ax3.set_xlabel('Quantidade de Partições')
|
|
|
|
ax3.set_xlim([min(parameters), max(parameters)])
|
|
|
|
ax3.plot(parameters, errors)
|
|
|
|
ret.append(bestd)
|
|
|
|
ret.append(forecastedd_best)
|
|
|
|
return ret
|
|
|
|
|
|
|
|
|
|
|
|
def compareModelsPlot(original, models_fo, models_ho):
|
|
|
|
fig = plt.figure(figsize=[13, 6])
|
2016-09-02 22:55:55 +04:00
|
|
|
fig.suptitle("Comparação de modelos ")
|
2017-01-11 00:05:51 +04:00
|
|
|
ax0 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height
|
2016-09-02 22:55:55 +04:00
|
|
|
rows = []
|
|
|
|
for model in models_fo:
|
|
|
|
fts = model["model"]
|
2016-10-18 21:45:07 +04:00
|
|
|
ax0.plot(model["forecasted"], label=model["name"])
|
2016-09-02 22:55:55 +04:00
|
|
|
for model in models_ho:
|
|
|
|
fts = model["model"]
|
2016-10-18 21:45:07 +04:00
|
|
|
ax0.plot(model["forecasted"], label=model["name"])
|
2016-09-02 22:55:55 +04:00
|
|
|
handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
ax0.legend(handles0, labels0)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
|
|
|
|
def compareModelsTable(original, models_fo, models_ho):
|
|
|
|
fig = plt.figure(figsize=[12, 4])
|
2016-09-02 22:55:55 +04:00
|
|
|
fig.suptitle("Comparação de modelos ")
|
2017-01-11 00:05:51 +04:00
|
|
|
columns = ['Modelo', 'Ordem', 'Partições', 'RMSE', 'MAPE (%)']
|
2016-09-02 22:55:55 +04:00
|
|
|
rows = []
|
|
|
|
for model in models_fo:
|
|
|
|
fts = model["model"]
|
2017-01-11 00:05:51 +04:00
|
|
|
error_r = Measures.rmse(model["forecasted"], original)
|
|
|
|
error_m = round(Measures.mape(model["forecasted"], original) * 100, 2)
|
|
|
|
rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m])
|
2016-09-02 22:55:55 +04:00
|
|
|
for model in models_ho:
|
|
|
|
fts = model["model"]
|
2017-01-11 00:05:51 +04:00
|
|
|
error_r = Measures.rmse(model["forecasted"][fts.order:], original[fts.order:])
|
|
|
|
error_m = round(Measures.mape(model["forecasted"][fts.order:], original[fts.order:]) * 100, 2)
|
|
|
|
rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m])
|
|
|
|
ax1 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height
|
2016-09-02 22:55:55 +04:00
|
|
|
ax1.set_xticks([])
|
|
|
|
ax1.set_yticks([])
|
|
|
|
ax1.table(cellText=rows,
|
2017-01-11 00:05:51 +04:00
|
|
|
colLabels=columns,
|
|
|
|
cellLoc='center',
|
|
|
|
bbox=[0, 0, 1, 1])
|
2016-09-02 22:55:55 +04:00
|
|
|
sup = "\\begin{tabular}{"
|
|
|
|
header = ""
|
|
|
|
body = ""
|
|
|
|
footer = ""
|
|
|
|
|
|
|
|
for c in columns:
|
|
|
|
sup = sup + "|c"
|
|
|
|
if len(header) > 0:
|
|
|
|
header = header + " & "
|
|
|
|
header = header + "\\textbf{" + c + "} "
|
|
|
|
sup = sup + "|} \\hline\n"
|
2017-01-11 00:05:51 +04:00
|
|
|
header = header + "\\\\ \\hline \n"
|
|
|
|
|
2016-09-02 22:55:55 +04:00
|
|
|
for r in rows:
|
|
|
|
lin = ""
|
|
|
|
for c in r:
|
|
|
|
if len(lin) > 0:
|
|
|
|
lin = lin + " & "
|
|
|
|
lin = lin + str(c)
|
2017-01-11 00:05:51 +04:00
|
|
|
|
|
|
|
body = body + lin + "\\\\ \\hline \n"
|
|
|
|
|
2016-09-02 22:55:55 +04:00
|
|
|
return sup + header + body + "\\end{tabular}"
|
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
|
2017-01-30 03:59:50 +04:00
|
|
|
def simpleSearch_RMSE(train, test, model, partitions, orders, save=False, file=None, tam=[10, 15],
|
2017-02-05 02:40:27 +04:00
|
|
|
plotforecasts=False, elev=30, azim=144, intervals=False,parameters=None):
|
|
|
|
_3d = len(orders) > 1
|
2017-01-11 00:05:51 +04:00
|
|
|
ret = []
|
2017-01-21 06:38:32 +04:00
|
|
|
errors = np.array([[0 for k in range(len(partitions))] for kk in range(len(orders))])
|
2017-01-11 00:05:51 +04:00
|
|
|
forecasted_best = []
|
2017-01-21 06:38:32 +04:00
|
|
|
fig = plt.figure(figsize=tam)
|
2017-01-23 00:41:42 +04:00
|
|
|
# fig.suptitle("Comparação de modelos ")
|
2017-01-21 06:38:32 +04:00
|
|
|
if plotforecasts:
|
2017-01-23 00:41:42 +04:00
|
|
|
ax0 = fig.add_axes([0, 0.4, 0.9, 0.5]) # left, bottom, width, height
|
2017-01-27 14:26:47 +04:00
|
|
|
ax0.set_xlim([0, len(train)])
|
|
|
|
ax0.set_ylim([min(train) * 0.9, max(train) * 1.1])
|
2017-01-21 06:38:32 +04:00
|
|
|
ax0.set_title('Forecasts')
|
|
|
|
ax0.set_ylabel('F(T)')
|
|
|
|
ax0.set_xlabel('T')
|
|
|
|
min_rmse = 1000000.0
|
2017-01-11 00:05:51 +04:00
|
|
|
best = None
|
2017-01-30 03:59:50 +04:00
|
|
|
|
|
|
|
for pc, p in enumerate(partitions, start=0):
|
|
|
|
|
|
|
|
sets = Grid.GridPartitionerTrimf(train, p)
|
|
|
|
for oc, o in enumerate(orders, start=0):
|
2017-01-21 06:38:32 +04:00
|
|
|
fts = model("q = " + str(p) + " n = " + str(o))
|
2017-02-05 02:40:27 +04:00
|
|
|
fts.train(train, sets, o,parameters=parameters)
|
2017-01-30 03:59:50 +04:00
|
|
|
if not intervals:
|
|
|
|
forecasted = fts.forecast(test)
|
2017-02-05 02:40:27 +04:00
|
|
|
if not fts.hasSeasonality:
|
|
|
|
error = Measures.rmse(np.array(test[o:]), np.array(forecasted[:-1]))
|
|
|
|
else:
|
|
|
|
error = Measures.rmse(np.array(test[o:]), np.array(forecasted))
|
2017-01-30 03:59:50 +04:00
|
|
|
for kk in range(o):
|
|
|
|
forecasted.insert(0, None)
|
2017-02-05 02:40:27 +04:00
|
|
|
if plotforecasts: ax0.plot(forecasted, label=fts.name)
|
2017-01-30 03:59:50 +04:00
|
|
|
else:
|
|
|
|
forecasted = fts.forecastInterval(test)
|
|
|
|
error = 1.0 - Measures.rmse_interval(np.array(test[o:]), np.array(forecasted[:-1]))
|
2017-01-11 00:05:51 +04:00
|
|
|
errors[oc, pc] = error
|
|
|
|
if error < min_rmse:
|
|
|
|
min_rmse = error
|
|
|
|
best = fts
|
|
|
|
forecasted_best = forecasted
|
2017-01-30 03:59:50 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
# print(min_rmse)
|
2017-01-21 06:38:32 +04:00
|
|
|
if plotforecasts:
|
2017-01-23 00:41:42 +04:00
|
|
|
# handles0, labels0 = ax0.get_legend_handles_labels()
|
|
|
|
# ax0.legend(handles0, labels0)
|
2017-01-27 14:26:47 +04:00
|
|
|
ax0.plot(test, label="Original", linewidth=3.0, color="black")
|
2017-02-05 02:40:27 +04:00
|
|
|
if _3d: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
2017-01-21 06:38:32 +04:00
|
|
|
if not plotforecasts: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim)
|
2017-01-11 00:05:51 +04:00
|
|
|
# ax1 = fig.add_axes([0.6, 0.5, 0.45, 0.45], projection='3d')
|
2017-02-05 02:40:27 +04:00
|
|
|
if _3d:
|
|
|
|
ax1.set_title('Error Surface')
|
|
|
|
ax1.set_ylabel('Model order')
|
|
|
|
ax1.set_xlabel('Number of partitions')
|
|
|
|
ax1.set_zlabel('RMSE')
|
|
|
|
X, Y = np.meshgrid(partitions, orders)
|
|
|
|
surf = ax1.plot_surface(X, Y, errors, rstride=1, cstride=1, antialiased=True)
|
|
|
|
else:
|
|
|
|
ax1 = fig.add_axes([0, 1, 0.9, 0.9])
|
|
|
|
ax1.set_title('Error Curve')
|
|
|
|
ax1.set_ylabel('Number of partitions')
|
|
|
|
ax1.set_xlabel('RMSE')
|
|
|
|
ax0.plot(errors,partitions)
|
2017-01-11 00:05:51 +04:00
|
|
|
ret.append(best)
|
|
|
|
ret.append(forecasted_best)
|
2016-09-02 22:55:55 +04:00
|
|
|
|
2017-01-23 00:41:42 +04:00
|
|
|
# plt.tight_layout()
|
|
|
|
|
|
|
|
Util.showAndSaveImage(fig, file, save)
|
2017-01-21 06:38:32 +04:00
|
|
|
|
2017-01-11 00:05:51 +04:00
|
|
|
return ret
|
2017-01-23 00:41:42 +04:00
|
|
|
|
|
|
|
|
|
|
|
def pftsExploreOrderAndPartitions(data,save=False, file=None):
|
|
|
|
fig, axes = plt.subplots(nrows=4, ncols=1, figsize=[6, 8])
|
|
|
|
data_fs1 = Grid.GridPartitionerTrimf(data, 10)
|
|
|
|
mi = []
|
|
|
|
ma = []
|
|
|
|
|
|
|
|
axes[0].set_title('Point Forecasts by Order')
|
|
|
|
axes[2].set_title('Interval Forecasts by Order')
|
|
|
|
|
|
|
|
for order in np.arange(1, 6):
|
2017-02-22 23:18:15 +04:00
|
|
|
fts = pwfts.ProbabilisticWeightedFTS("")
|
2017-01-23 00:41:42 +04:00
|
|
|
fts.shortname = "n = " + str(order)
|
|
|
|
fts.train(data, data_fs1, order=order)
|
|
|
|
point_forecasts = fts.forecast(data)
|
|
|
|
interval_forecasts = fts.forecastInterval(data)
|
|
|
|
lower = [kk[0] for kk in interval_forecasts]
|
|
|
|
upper = [kk[1] for kk in interval_forecasts]
|
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
|
|
|
for k in np.arange(0, order):
|
|
|
|
point_forecasts.insert(0, None)
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
axes[0].plot(point_forecasts, label=fts.shortname)
|
|
|
|
axes[2].plot(lower, label=fts.shortname)
|
|
|
|
axes[2].plot(upper)
|
|
|
|
|
|
|
|
axes[1].set_title('Point Forecasts by Number of Partitions')
|
|
|
|
axes[3].set_title('Interval Forecasts by Number of Partitions')
|
|
|
|
|
|
|
|
for partitions in np.arange(5, 11):
|
|
|
|
data_fs = Grid.GridPartitionerTrimf(data, partitions)
|
2017-02-22 23:18:15 +04:00
|
|
|
fts = pwfts.ProbabilisticWeightedFTS("")
|
2017-01-23 00:41:42 +04:00
|
|
|
fts.shortname = "q = " + str(partitions)
|
|
|
|
fts.train(data, data_fs, 1)
|
|
|
|
point_forecasts = fts.forecast(data)
|
|
|
|
interval_forecasts = fts.forecastInterval(data)
|
|
|
|
lower = [kk[0] for kk in interval_forecasts]
|
|
|
|
upper = [kk[1] for kk in interval_forecasts]
|
|
|
|
mi.append(min(lower) * 0.95)
|
|
|
|
ma.append(max(upper) * 1.05)
|
|
|
|
point_forecasts.insert(0, None)
|
|
|
|
lower.insert(0, None)
|
|
|
|
upper.insert(0, None)
|
|
|
|
axes[1].plot(point_forecasts, label=fts.shortname)
|
|
|
|
axes[3].plot(lower, label=fts.shortname)
|
|
|
|
axes[3].plot(upper)
|
|
|
|
|
|
|
|
for ax in axes:
|
|
|
|
ax.set_ylabel('F(T)')
|
|
|
|
ax.set_xlabel('T')
|
|
|
|
ax.plot(data, label="Original", color="black", linewidth=1.5)
|
|
|
|
handles, labels = ax.get_legend_handles_labels()
|
|
|
|
ax.legend(handles, labels, loc=2, bbox_to_anchor=(1, 1))
|
|
|
|
ax.set_ylim([min(mi), max(ma)])
|
|
|
|
ax.set_xlim([0, len(data)])
|
|
|
|
|
|
|
|
plt.tight_layout()
|
|
|
|
|
|
|
|
Util.showAndSaveImage(fig, file, save)
|
2017-02-16 00:16:13 +04:00
|
|
|
|