Source code for pyFTS.benchmarks.benchmarks

#!/usr/bin/python
# -*- coding: utf8 -*-

"""Benchmarks methods for FTS methods"""


import datetime
import time
from copy import deepcopy
import traceback

import matplotlib as plt

import matplotlib.pyplot as plt
import numpy as np
from mpl_toolkits.mplot3d import Axes3D
from itertools import product

from pyFTS.common.transformations.differential import Differential
from pyFTS.models import song, chen, yu, ismailefendi, sadaei, hofts, pwfts, ifts, cheng, hwang
from pyFTS.models.multivariate import mvfts, wmvfts, cmvfts
from pyFTS.models.ensemble import ensemble
from pyFTS.benchmarks import Measures, naive, arima, ResidualAnalysis, quantreg, knn
from pyFTS.benchmarks import Util as bUtil
from pyFTS.common import Util as cUtil
# from sklearn.cross_validation import KFold
from pyFTS.partitioners import Grid

colors = ['grey', 'darkgrey', 'rosybrown', 'maroon', 'red','orange', 'gold', 'yellow', 'olive', 'green',
          'darkgreen', 'cyan', 'lightblue','blue', 'darkblue', 'purple', 'darkviolet' ]

ncol = len(colors)

styles = ['-','--','-.',':','.']

nsty = len(styles)


def __pop(key, default, kwargs):
    if key in kwargs:
        return kwargs.pop(key)
    else:
        return default


[docs]def get_benchmark_point_methods(): """Return all non FTS methods for point forecasting""" return [naive.Naive, arima.ARIMA, quantreg.QuantileRegression]
[docs]def get_point_methods(): """Return all FTS methods for point forecasting""" return [song.ConventionalFTS, chen.ConventionalFTS, yu.WeightedFTS, ismailefendi.ImprovedWeightedFTS, cheng.TrendWeightedFTS, sadaei.ExponentialyWeightedFTS, hofts.HighOrderFTS, hofts.WeightedHighOrderFTS, hwang.HighOrderFTS, pwfts.ProbabilisticWeightedFTS]
[docs]def get_point_multivariate_methods(): """Return all multivariate FTS methods por point forecasting""" return [mvfts.MVFTS, wmvfts.WeightedMVFTS, cmvfts.ClusteredMVFTS]
[docs]def get_benchmark_interval_methods(): """Return all non FTS methods for point_to_interval forecasting""" return [ arima.ARIMA, quantreg.QuantileRegression]
[docs]def get_interval_methods(): """Return all FTS methods for point_to_interval forecasting""" return [ifts.IntervalFTS, ifts.WeightedIntervalFTS, pwfts.ProbabilisticWeightedFTS]
[docs]def get_probabilistic_methods(): """Return all FTS methods for probabilistic forecasting""" return [ensemble.AllMethodEnsembleFTS, pwfts.ProbabilisticWeightedFTS]
[docs]def get_benchmark_probabilistic_methods(): """Return all FTS methods for probabilistic forecasting""" return [arima.ARIMA, quantreg.QuantileRegression, knn.KNearestNeighbors]
[docs]def multivariate_sliding_window_benchmarks2(data, windowsize, train=0.8, **kwargs): from pyFTS.models.multivariate import common, variable, mvfts tag = __pop('tag', None, kwargs) dataset = __pop('dataset', None, kwargs) distributed = __pop('distributed', False, kwargs) variables = __pop('variables', {}, kwargs) target_variable = __pop('target_variable', '', kwargs) type = kwargs.get("type", 'point') steps_ahead = __pop('steps_ahead', [1], kwargs) steps_ahead = [k for k in steps_ahead] fts_methods = __pop('methods', [], kwargs) if fts_methods is not None: methods_parameters = __pop('methods_parameters', None, kwargs) if type == 'point': experiment_method = mv_run_point2 synthesis_method = process_point_jobs2 elif type == 'interval': experiment_method = mv_run_interval2 synthesis_method = process_interval_jobs2 elif type == 'distribution': experiment_method = mv_run_probabilistic2 synthesis_method = process_probabilistic_jobs2 else: raise ValueError("Type parameter has a unkown value!") if distributed: import pyFTS.distributed.dispy as dispy nodes = kwargs.get("nodes", ['127.0.0.1']) cluster, http_server = dispy.start_dispy_cluster(experiment_method, nodes) inc = __pop("inc", 0.1, kwargs) file = kwargs.get('file', "benchmarks.db") conn = bUtil.open_benchmark_db(file) jobs = [] for ct, train, test in cUtil.sliding_window(data, windowsize, train, inc=inc, **kwargs): for id, fts_method in enumerate(fts_methods): kwargs['steps_ahead'] = max(steps_ahead) parameters = {} if methods_parameters is not None: parameters = methods_parameters[id] vars = [] tvar = None for key, value in variables.items(): var = variable.Variable(key, data=train, **value) vars.append(var) if key == target_variable: tvar = var model = fts_method(explanatory_variables=vars, target_variable=tvar, **parameters) if not distributed: try: job = experiment_method(model, train, test, ct, **kwargs) synthesis_method(dataset, tag, job, conn) except Exception as ex: print('EXCEPTION! ', fts_method) traceback.print_exc() else: job = cluster.submit(model, train, test, ct, **kwargs) job.id = id jobs.append(job) if distributed: for job in jobs: job() if job.status == dispy.dispy.DispyJob.Finished and job is not None: tmp = job.result synthesis_method(dataset, tag, tmp, conn) else: print("status", job.status) print("result", job.result) print("stdout", job.stdout) print("stderr", job.exception) cluster.wait() # wait for all jobs to finish dispy.stop_dispy_cluster(cluster, http_server) conn.close()
[docs]def sliding_window_benchmarks2(data, windowsize, train=0.8, **kwargs): tag = __pop('tag', None, kwargs) dataset = __pop('dataset', None, kwargs) distributed = __pop('distributed', False, kwargs) transformations = kwargs.get('transformations', [None]) type = kwargs.get("type", 'point') orders = __pop("orders", [1, 2, 3], kwargs) partitioners_methods = __pop("partitioners_methods", [Grid.GridPartitioner], kwargs) partitions = __pop("partitions", [10], kwargs) partitions = [k for k in partitions] steps_ahead = __pop('steps_ahead', [1], kwargs) steps_ahead = [k for k in steps_ahead] fts_methods = __pop('methods', [], kwargs) if fts_methods is not None: methods_parameters = __pop('methods_parameters', None, kwargs) ix_methods = [k for k in np.arange(len(fts_methods))] benchmark_models = __pop("benchmark_models", False, kwargs) benchmark_methods = __pop("benchmark_methods", None, kwargs) benchmark_methods_parameters = __pop("benchmark_methods_parameters", None, kwargs) if type == 'point': experiment_method = run_point2 synthesis_method = process_point_jobs2 elif type == 'interval': experiment_method = run_interval2 synthesis_method = process_interval_jobs2 elif type == 'distribution': experiment_method = run_probabilistic2 synthesis_method = process_probabilistic_jobs2 else: raise ValueError("Type parameter has a unkown value!") if distributed: import pyFTS.distributed.dispy as dispy nodes = kwargs.get("nodes", ['127.0.0.1']) cluster, http_server = dispy.start_dispy_cluster(experiment_method, nodes) inc = __pop("inc", 0.1, kwargs) file = kwargs.get('file', "benchmarks.db") conn = bUtil.open_benchmark_db(file) jobs = [] for ct, train, test in cUtil.sliding_window(data, windowsize, train, inc=inc, **kwargs): if benchmark_models: for bm, method in enumerate(benchmark_methods): kwargs['steps_ahead'] = max(steps_ahead) kwargs['parameters'] = benchmark_methods_parameters[bm] if not distributed: try: job = experiment_method(method, None, None, None, None, train, test, ct, **kwargs) synthesis_method(dataset, tag, job, conn) except Exception as ex: print('EXCEPTION! ', method, benchmark_methods_parameters[bm]) traceback.print_exc() else: job = cluster.submit(method, None, None, None, None, train, test, ct, **kwargs) jobs.append(job) if fts_methods is not None: params = [ix_methods, orders, partitioners_methods, partitions, transformations] for id, instance in enumerate(product(*params)): fts_method = fts_methods[instance[0]] kwargs['steps_ahead'] = max(steps_ahead) if methods_parameters is not None: kwargs['parameters'] = methods_parameters[instance[0]] if not distributed: try: job = experiment_method(fts_method, instance[1], instance[2], instance[3], instance[4], train, test, ct, **kwargs) synthesis_method(dataset, tag, job, conn) except Exception as ex: print('EXCEPTION! ', instance) traceback.print_exc() else: job = cluster.submit(fts_method, instance[1], instance[2], instance[3], instance[4], train, test, ct, **kwargs) job.id = id jobs.append(job) if distributed: for job in jobs: job() if job.status == dispy.dispy.DispyJob.Finished and job is not None: tmp = job.result synthesis_method(dataset, tag, tmp, conn) else: print("status", job.status) print("result", job.result) print("stdout", job.stdout) print("stderr", job.exception) cluster.wait() # wait for all jobs to finish dispy.stop_dispy_cluster(cluster, http_server) conn.close()
[docs]def sliding_window_benchmarks(data, windowsize, train=0.8, **kwargs): """ Sliding window benchmarks for FTS forecasters. For each data window, a train and test datasets will be splitted. For each train split, number of partitions and partitioning method will be created a partitioner model. And for each partitioner, order, steps ahead and FTS method a foreasting model will be trained. Then all trained models are benchmarked on the test data and the metrics are stored on a sqlite3 database (identified by the 'file' parameter) for posterior analysis. All these process can be distributed on a dispy cluster, setting the atributed 'distributed' to true and informing the list of dispy nodes on 'nodes' parameter. The number of experiments is determined by 'windowsize' and 'inc' parameters. :param data: test data :param windowsize: size of sliding window :param train: percentual of sliding window data used to train the models :param kwargs: dict, optional arguments :keyword benchmark_methods: a list with Non FTS models to benchmark. The default is None. :keyword benchmark_methods_parameters: a list with Non FTS models parameters. The default is None. :keyword benchmark_models: A boolean value indicating if external FTS methods will be used on benchmark. The default is False. :keyword build_methods: A boolean value indicating if the default FTS methods will be used on benchmark. The default is True. :keyword dataset: the dataset name to identify the current set of benchmarks results on database. :keyword distributed: A boolean value indicating if the forecasting procedure will be distributed in a dispy cluster. . The default is False :keyword file: file path to save the results. The default is benchmarks.db. :keyword inc: a float on interval [0,1] indicating the percentage of the windowsize to move the window :keyword methods: a list with FTS class names. The default depends on the forecasting type and contains the list of all FTS methods. :keyword models: a list with prebuilt FTS objects. The default is None. :keyword nodes: a list with the dispy cluster nodes addresses. The default is [127.0.0.1]. :keyword orders: a list with orders of the models (for high order models). The default is [1,2,3]. :keyword partitions: a list with the numbers of partitions on the Universe of Discourse. The default is [10]. :keyword partitioners_models: a list with prebuilt Universe of Discourse partitioners objects. The default is None. :keyword partitioners_methods: a list with Universe of Discourse partitioners class names. The default is [partitioners.Grid.GridPartitioner]. :keyword progress: If true a progress bar will be displayed during the benchmarks. The default is False. :keyword start: in the multi step forecasting, the index of the data where to start forecasting. The default is 0. :keyword steps_ahead: a list with the forecasting horizons, i. e., the number of steps ahead to forecast. The default is 1. :keyword tag: a name to identify the current set of benchmarks results on database. :keyword type: the forecasting type, one of these values: point(default), interval or distribution. The default is point. :keyword transformations: a list with data transformations do apply . The default is [None]. """ tag = __pop('tag', None, kwargs) dataset = __pop('dataset', None, kwargs) distributed = __pop('distributed', False, kwargs) transformations = kwargs.get('transformations', [None]) progress = kwargs.get('progress', None) type = kwargs.get("type", 'point') orders = __pop("orders", [1,2,3], kwargs) partitioners_models = __pop("partitioners_models", None, kwargs) partitioners_methods = __pop("partitioners_methods", [Grid.GridPartitioner], kwargs) partitions = __pop("partitions", [10], kwargs) steps_ahead = __pop('steps_ahead', [1], kwargs) methods = __pop('methods', None, kwargs) models = __pop('models', None, kwargs) pool = [] if models is None else models if methods is None: if type == 'point': methods = get_point_methods() elif type == 'interval': methods = get_interval_methods() elif type == 'distribution': methods = get_probabilistic_methods() build_methods = __pop("build_methods", True, kwargs) if build_methods: for method in methods: mfts = method() if mfts.is_high_order: for order in orders: if order >= mfts.min_order: mfts = method() mfts.order = order pool.append(mfts) else: mfts.order = 1 pool.append(mfts) benchmark_models = __pop("benchmark_models", False, kwargs) if benchmark_models != False: benchmark_methods = __pop("benchmark_methods", None, kwargs) benchmark_methods_parameters = __pop("benchmark_methods_parameters", None, kwargs) benchmark_pool = [] if ( benchmark_models is None or not isinstance(benchmark_models, list)) \ else benchmark_models if benchmark_models is None and benchmark_methods is None: if type == 'point'or type == 'partition': benchmark_methods = get_benchmark_point_methods() elif type == 'interval': benchmark_methods = get_benchmark_interval_methods() elif type == 'distribution': benchmark_methods = get_benchmark_probabilistic_methods() if benchmark_methods is not None: for transformation in transformations: for count, model in enumerate(benchmark_methods, start=0): par = benchmark_methods_parameters[count] mfts = model(**par) mfts.append_transformation(transformation) benchmark_pool.append(mfts) if type == 'point': experiment_method = run_point synthesis_method = process_point_jobs elif type == 'interval': experiment_method = run_interval synthesis_method = process_interval_jobs elif type == 'distribution': experiment_method = run_probabilistic synthesis_method = process_probabilistic_jobs else: raise ValueError("Type parameter has a unkown value!") if distributed: import pyFTS.distributed.dispy as dispy nodes = kwargs.get("nodes", ['127.0.0.1']) cluster, http_server = dispy.start_dispy_cluster(experiment_method, nodes) jobs = [] inc = __pop("inc", 0.1, kwargs) if progress: from tqdm import tqdm _tdata = len(data) / (windowsize * inc) _tasks = (len(partitioners_models) * len(orders) * len(partitions) * len(transformations) * len(steps_ahead)) _tbcmk = len(benchmark_pool)*len(steps_ahead) progressbar = tqdm(total=_tdata*_tasks + _tdata*_tbcmk, desc="Benchmarks:") file = kwargs.get('file', "benchmarks.db") conn = bUtil.open_benchmark_db(file) for ct, train, test in cUtil.sliding_window(data, windowsize, train, inc=inc, **kwargs): if benchmark_models != False: for model in benchmark_pool: for step in steps_ahead: kwargs['steps_ahead'] = step if not distributed: if progress: progressbar.update(1) try: job = experiment_method(deepcopy(model), None, train, test, **kwargs) synthesis_method(dataset, tag, job, conn) except Exception as ex: print('EXCEPTION! ', model.shortname, model.order) traceback.print_exc() else: job = cluster.submit(deepcopy(model), None, train, test, **kwargs) jobs.append(job) partitioners_pool = [] if partitioners_models is None: for transformation in transformations: for partition in partitions: for partitioner in partitioners_methods: data_train_fs = partitioner(data=train, npart=partition, transformation=transformation) partitioners_pool.append(data_train_fs) else: partitioners_pool = partitioners_models for step in steps_ahead: for partitioner in partitioners_pool: for _id, model in enumerate(pool,start=0): kwargs['steps_ahead'] = step if not distributed: if progress: progressbar.update(1) try: job = experiment_method(deepcopy(model), deepcopy(partitioner), train, test, **kwargs) synthesis_method(dataset, tag, job, conn) except Exception as ex: print('EXCEPTION! ',model.shortname, model.order, partitioner.name, partitioner.partitions, str(partitioner.transformation)) traceback.print_exc() else: job = cluster.submit(deepcopy(model), deepcopy(partitioner), train, test, **kwargs) job.id = id # associate an ID to identify jobs (if needed later) jobs.append(job) if progress: progressbar.close() if distributed: for job in jobs: if progress: progressbar.update(1) job() if job.status == dispy.dispy.DispyJob.Finished and job is not None: tmp = job.result synthesis_method(dataset, tag, tmp, conn) else: print("status",job.status) print("result",job.result) print("stdout",job.stdout) print("stderr",job.exception) cluster.wait() # wait for all jobs to finish dispy.stop_dispy_cluster(cluster, http_server) conn.close()
[docs]def run_point(mfts, partitioner, train_data, test_data, window_key=None, **kwargs): """ Run the point forecasting benchmarks :param mfts: FTS model :param partitioner: Universe of Discourse partitioner :param train_data: data used to train the model :param test_data: ata used to test the model :param window_key: id of the sliding window :param transformation: data transformation :param indexer: seasonal indexer :return: a dictionary with the benchmark results """ import time from pyFTS.models import yu, chen, hofts, pwfts,ismailefendi,sadaei, song, cheng, hwang from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, naive, arima, quantreg from pyFTS.common import Transformations tmp = [song.ConventionalFTS, chen.ConventionalFTS, yu.WeightedFTS, ismailefendi.ImprovedWeightedFTS, cheng.TrendWeightedFTS, sadaei.ExponentialyWeightedFTS, hofts.HighOrderFTS, hwang.HighOrderFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp4 = [naive.Naive, arima.ARIMA, quantreg.QuantileRegression] tmp3 = [Measures.get_point_statistics] tmp5 = [Differential] indexer = kwargs.get('indexer', None) steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) if mfts.benchmark_only: _key = mfts.shortname + str(mfts.order if mfts.order is not None else "") else: pttr = str(partitioner.__module__).split('.')[-1] _key = mfts.shortname + " n = " + str(mfts.order) + " " + pttr + " q = " + str(partitioner.partitions) mfts.partitioner = partitioner mfts.append_transformation(partitioner.transformation) _key += str(steps_ahead) _key += str(method) if method is not None else "" _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start _start = time.time() _rmse, _smape, _u = Measures.get_point_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'key': _key, 'obj': mfts, 'rmse': _rmse, 'smape': _smape, 'u': _u, 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method} return ret
[docs]def run_interval(mfts, partitioner, train_data, test_data, window_key=None, **kwargs): """ Run the interval forecasting benchmarks :param mfts: FTS model :param partitioner: Universe of Discourse partitioner :param train_data: data used to train the model :param test_data: ata used to test the model :param window_key: id of the sliding window :param transformation: data transformation :param indexer: seasonal indexer :return: a dictionary with the benchmark results """ import time from pyFTS.models import hofts,ifts,pwfts from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp4 = [arima.ARIMA, quantreg.QuantileRegression] tmp3 = [Measures.get_interval_statistics] steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) if mfts.benchmark_only: _key = mfts.shortname + str(mfts.order if mfts.order is not None else "") + str(mfts.alpha) else: pttr = str(partitioner.__module__).split('.')[-1] _key = mfts.shortname + " n = " + str(mfts.order) + " " + pttr + " q = " + str(partitioner.partitions) mfts.partitioner = partitioner mfts.append_transformation(partitioner.transformation) _key += str(steps_ahead) _key += str(method) if method is not None else "" _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start _start = time.time() #_sharp, _res, _cov, _q05, _q25, _q75, _q95, _w05, _w25 metrics = Measures.get_interval_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'key': _key, 'obj': mfts, 'sharpness': metrics[0], 'resolution': metrics[1], 'coverage': metrics[2], 'time': times,'Q05': metrics[3], 'Q25': metrics[4], 'Q75': metrics[5], 'Q95': metrics[6], 'winkler05': metrics[7], 'winkler25': metrics[8], 'window': window_key,'steps': steps_ahead, 'method': method} return ret
[docs]def run_probabilistic(mfts, partitioner, train_data, test_data, window_key=None, **kwargs): """ Run the probabilistic forecasting benchmarks :param mfts: FTS model :param partitioner: Universe of Discourse partitioner :param train_data: data used to train the model :param test_data: ata used to test the model :param steps: :param resolution: :param window_key: id of the sliding window :param transformation: data transformation :param indexer: seasonal indexer :return: a dictionary with the benchmark results """ import time import numpy as np from pyFTS.models import hofts, ifts, pwfts from pyFTS.models.ensemble import ensemble from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, knn from pyFTS.models.seasonal import SeasonalIndexer tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, pwfts.ProbabilisticWeightedFTS, arima.ARIMA, ensemble.AllMethodEnsembleFTS, knn.KNearestNeighbors] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp3 = [Measures.get_distribution_statistics, SeasonalIndexer.SeasonalIndexer, SeasonalIndexer.LinearSeasonalIndexer] indexer = kwargs.get('indexer', None) steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) if mfts.benchmark_only: _key = mfts.shortname + str(mfts.order if mfts.order is not None else "") + str(mfts.alpha) else: pttr = str(partitioner.__module__).split('.')[-1] _key = mfts.shortname + " n = " + str(mfts.order) + " " + pttr + " q = " + str(partitioner.partitions) mfts.partitioner = partitioner mfts.append_transformation(partitioner.transformation) _key += str(steps_ahead) _key += str(method) if method is not None else "" if mfts.has_seasonality: mfts.indexer = indexer _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start _crps1, _t1, _brier = Measures.get_distribution_statistics(test_data, mfts, **kwargs) _t1 += times ret = {'key': _key, 'obj': mfts, 'crps': _crps1, 'time': _t1, 'brier': _brier, 'window': window_key, 'steps': steps_ahead, 'method': method} return ret
def __build_model(fts_method, order, parameters, partitioner_method, partitions, train_data, transformation): mfts = fts_method(**parameters) if mfts.benchmark_only or mfts.is_wrapper: pttr = '' else: fs = partitioner_method(npart=partitions, data=train_data, transformation=transformation) pttr = str(fs.__module__).split('.')[-1] if order > 1: mfts = fts_method(partitioner=fs, order=order, **parameters) else: mfts.partitioner = fs if transformation is not None: mfts.append_transformation(transformation) return mfts, pttr
[docs]def mv_run_point2(mfts, train_data, test_data, window_key=None, **kwargs): import time from pyFTS.models import hofts, ifts, pwfts from pyFTS.models.multivariate import mvfts, wmvfts, granular from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, BSTS, benchmarks tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, ifts.WeightedIntervalFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, mvfts.MVFTS, wmvfts.WeightedMVFTS, granular.GranularWMVFTS] tmp4 = [arima.ARIMA, quantreg.QuantileRegression, BSTS.ARIMA] tmp3 = [Measures.get_interval_statistics] steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _start = time.time() _rmse, _smape, _u = Measures.get_point_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'rmse': _rmse, 'smape': _smape, 'u': _u, 'window': window_key, 'steps': steps_ahead, 'method': method} else: _start = time.time() forecasts = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start tmp_test = test_data[mfts.target_variable.data_label].values[mfts.order:mfts.order + steps_ahead] eval = Measures.get_point_ahead_statistics(tmp_test, forecasts) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def run_point2(fts_method, order, partitioner_method, partitions, transformation, train_data, test_data, window_key=None, **kwargs): import time from pyFTS.models import yu, chen, hofts, pwfts,ismailefendi,sadaei, song, cheng, hwang from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, naive, arima, quantreg, benchmarks from pyFTS.common import Transformations tmp = [song.ConventionalFTS, chen.ConventionalFTS, yu.WeightedFTS, ismailefendi.ImprovedWeightedFTS, cheng.TrendWeightedFTS, sadaei.ExponentialyWeightedFTS, hofts.HighOrderFTS, hwang.HighOrderFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp4 = [naive.Naive, arima.ARIMA, quantreg.QuantileRegression] tmp3 = [Measures.get_point_statistics] tmp5 = [Differential] indexer = kwargs.get('indexer', None) steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) parameters = kwargs.get('parameters', {}) mfts, pttr = benchmarks.__build_model(fts_method, order, parameters, partitioner_method, partitions, train_data, transformation) _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _start = time.time() _rmse, _smape, _u = Measures.get_point_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'rmse': _rmse, 'smape': _smape, 'u': _u, 'window': window_key, 'steps': steps_ahead, 'method': method} else: _start = time.time() forecasts = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start eval = Measures.get_point_ahead_statistics(test_data[mfts.order:mfts.order+steps_ahead], forecasts) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def mv_run_interval2(mfts,train_data, test_data, window_key=None, **kwargs): import time from pyFTS.models import hofts,ifts,pwfts from pyFTS.models.multivariate import mvfts, wmvfts, granular from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, BSTS, benchmarks tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, ifts.WeightedIntervalFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, mvfts.MVFTS, wmvfts.WeightedMVFTS, granular.GranularWMVFTS ] tmp4 = [arima.ARIMA, quantreg.QuantileRegression, BSTS.ARIMA] tmp3 = [Measures.get_interval_statistics] steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) parameters = kwargs.get('parameters',{}) _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _start = time.time() metrics = Measures.get_interval_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'sharpness': metrics[0], 'resolution': metrics[1], 'coverage': metrics[2], 'time': times,'pinball05': metrics[3], 'pinball25': metrics[4], 'pinball75': metrics[5], 'pinball95': metrics[6], 'winkler05': metrics[7], 'winkler25': metrics[8], 'window': window_key,'steps': steps_ahead, 'method': method} else: _start = time.time() intervals = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start tmp_test = test_data[mfts.target_variable.data_label].values[mfts.order:mfts.order + steps_ahead] eval = Measures.get_interval_ahead_statistics(tmp_test, intervals) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def run_interval2(fts_method, order, partitioner_method, partitions, transformation, train_data, test_data, window_key=None, **kwargs): import time from pyFTS.models import hofts,ifts,pwfts from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, BSTS, benchmarks tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, ifts.WeightedIntervalFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp4 = [arima.ARIMA, quantreg.QuantileRegression, BSTS.ARIMA] tmp3 = [Measures.get_interval_statistics] steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) parameters = kwargs.get('parameters',{}) mfts, pttr = benchmarks.__build_model(fts_method, order, parameters, partitioner_method, partitions, train_data, transformation) _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _start = time.time() metrics = Measures.get_interval_statistics(test_data, mfts, **kwargs) _end = time.time() times += _end - _start ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'sharpness': metrics[0], 'resolution': metrics[1], 'coverage': metrics[2], 'time': times,'pinball05': metrics[3], 'pinball25': metrics[4], 'pinball75': metrics[5], 'pinball95': metrics[6], 'winkler05': metrics[7], 'winkler25': metrics[8], 'window': window_key,'steps': steps_ahead, 'method': method} else: _start = time.time() intervals = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start eval = Measures.get_interval_ahead_statistics(test_data[mfts.order:mfts.order+steps_ahead], intervals) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def mv_run_probabilistic2(mfts, train_data, test_data, window_key=None, **kwargs): import time from pyFTS.models import hofts, ifts, pwfts from pyFTS.models.multivariate import mvfts, wmvfts, granular from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, BSTS, benchmarks tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, ifts.WeightedIntervalFTS, pwfts.ProbabilisticWeightedFTS] tmp2 = [Grid.GridPartitioner, mvfts.MVFTS, wmvfts.WeightedMVFTS, granular.GranularWMVFTS] tmp4 = [arima.ARIMA, quantreg.QuantileRegression, BSTS.ARIMA] tmp3 = [Measures.get_interval_statistics] steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) parameters = kwargs.get('parameters', {}) _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _crps1, _t1, _brier = Measures.get_distribution_statistics(test_data, mfts, **kwargs) times += _t1 ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'crps': _crps1, 'brier': _brier, 'window': window_key, 'steps': steps_ahead, 'method': method} else: _start = time.time() distributions = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start tmp_test = test_data[mfts.target_variable.data_label].values[mfts.order:mfts.order + steps_ahead] eval = Measures.get_distribution_ahead_statistics(tmp_test, distributions) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': None, 'order': mfts.order, 'partitions': None, 'transformation': None, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def run_probabilistic2(fts_method, order, partitioner_method, partitions, transformation, train_data, test_data, window_key=None, **kwargs): import time import numpy as np from pyFTS.models import hofts, ifts, pwfts from pyFTS.models.ensemble import ensemble from pyFTS.partitioners import Grid, Entropy, FCM from pyFTS.benchmarks import Measures, arima, quantreg, knn, benchmarks from pyFTS.models.seasonal import SeasonalIndexer tmp = [hofts.HighOrderFTS, ifts.IntervalFTS, pwfts.ProbabilisticWeightedFTS, arima.ARIMA, ensemble.AllMethodEnsembleFTS, knn.KNearestNeighbors] tmp2 = [Grid.GridPartitioner, Entropy.EntropyPartitioner, FCM.FCMPartitioner] tmp3 = [Measures.get_distribution_statistics, SeasonalIndexer.SeasonalIndexer, SeasonalIndexer.LinearSeasonalIndexer] indexer = kwargs.get('indexer', None) steps_ahead = kwargs.get('steps_ahead', 1) method = kwargs.get('method', None) parameters = kwargs.get('parameters', {}) mfts, pttr = benchmarks.__build_model(fts_method, order, parameters, partitioner_method, partitions, train_data, transformation) if mfts.has_seasonality: mfts.indexer = indexer _start = time.time() mfts.fit(train_data, **kwargs) _end = time.time() times = _end - _start if steps_ahead == 1: _crps1, _t1, _brier = Measures.get_distribution_statistics(test_data, mfts, **kwargs) times += _t1 ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'crps': _crps1, 'brier': _brier, 'window': window_key, 'steps': steps_ahead, 'method': method} else: _start = time.time() distributions = mfts.predict(test_data, **kwargs) _end = time.time() times += _end - _start eval = Measures.get_distribution_ahead_statistics(test_data[mfts.order:mfts.order+steps_ahead], distributions) for key in eval.keys(): eval[key]["time"] = times eval[key]["method"] = method ret = {'model': mfts.shortname, 'partitioner': pttr, 'order': order, 'partitions': partitions, 'transformation': '' if transformation is None else transformation.name, 'size': len(mfts), 'time': times, 'window': window_key, 'steps': steps_ahead, 'method': method, 'ahead_results': eval } return ret
[docs]def common_process_time_jobs(conn, data, job): dta = deepcopy(data) dta.append(job['steps']) dta.append(job['method']) for key in ["time"]: if key in job: data2 = deepcopy(dta) data2.extend([key, job[key]]) bUtil.insert_benchmark(data2, conn)
[docs]def common_process_point_jobs(conn, data, job): dta = deepcopy(data) dta.append(job['steps']) dta.append(job['method']) for key in ["rmse", "mape", "u", "time"]: if key in job: data2 = deepcopy(dta) data2.extend([key, job[key]]) bUtil.insert_benchmark(data2, conn)
[docs]def process_point_jobs(dataset, tag, job, conn): """ Extract information from a dictionary with point benchmark results and save it on a database :param dataset: the benchmark dataset name :param tag: alias for the benchmark group being executed :param job: a dictionary with the benchmark results :param conn: a connection to a Sqlite database :return: """ data = bUtil.process_common_data(dataset, tag, 'point',job) common_process_point_jobs(conn, data, job)
[docs]def process_point_jobs2(dataset, tag, job, conn): """ Extract information from a dictionary with point benchmark results and save it on a database :param dataset: the benchmark dataset name :param tag: alias for the benchmark group being executed :param job: a dictionary with the benchmark results :param conn: a connection to a Sqlite database :return: """ data = bUtil.process_common_data2(dataset, tag, 'point',job) if job['steps'] == 1: common_process_point_jobs(conn, data, job) else: for k in range(job['steps']): j2 = job['ahead_results'][k] common_process_point_jobs(conn, data, j2)
[docs]def common_process_interval_jobs(conn, data, job): dta = deepcopy(data) dta.append(job['steps']) dta.append(job['method']) for key in ["sharpness","resolution","coverage","time","pinball05", "pinball25","pinball75","pinball95", "winkler05", "winkler25"]: if key in job: data2 = deepcopy(dta) data2.extend([key, job[key]]) bUtil.insert_benchmark(data2, conn)
[docs]def process_interval_jobs(dataset, tag, job, conn): """ Extract information from an dictionary with interval benchmark results and save it on a database :param dataset: the benchmark dataset name :param tag: alias for the benchmark group being executed :param job: a dictionary with the benchmark results :param conn: a connection to a Sqlite database :return: """ data = bUtil.process_common_data(dataset, tag, 'interval', job) common_process_interval_jobs(conn, data, job)
[docs]def process_interval_jobs2(dataset, tag, job, conn): data = bUtil.process_common_data2(dataset, tag, 'interval', job) if job['steps'] == 1: common_process_interval_jobs(conn, data, job) else: for k in range(job['steps']): j2 = job['ahead_results'][k] common_process_interval_jobs(conn, data, j2)
[docs]def common_process_probabilistic_jobs(conn, data, job): dta = deepcopy(data) dta.append(job['steps']) dta.append(job['method']) for key in ["crps","time","brier"]: if key in job: data2 = deepcopy(dta) data2.extend([key, job[key]]) bUtil.insert_benchmark(data2, conn)
[docs]def process_probabilistic_jobs(dataset, tag, job, conn): """ Extract information from an dictionary with probabilistic benchmark results and save it on a database :param dataset: the benchmark dataset name :param tag: alias for the benchmark group being executed :param job: a dictionary with the benchmark results :param conn: a connection to a Sqlite database :return: """ data = bUtil.process_common_data(dataset, tag, 'density', job) common_process_probabilistic_jobs(conn, data, job)
[docs]def process_probabilistic_jobs2(dataset, tag, job, conn): """ Extract information from an dictionary with probabilistic benchmark results and save it on a database :param dataset: the benchmark dataset name :param tag: alias for the benchmark group being executed :param job: a dictionary with the benchmark results :param conn: a connection to a Sqlite database :return: """ data = bUtil.process_common_data2(dataset, tag, 'density', job) if job['steps'] == 1: common_process_probabilistic_jobs(conn,data,job) else: for k in range(job['steps']): j2 = job['ahead_results'][k] common_process_probabilistic_jobs(conn, data, j2)
[docs]def plot_point(axis, points, order, label, color='red', ls='-', linewidth=1): mi = min(points) * 0.95 ma = max(points) * 1.05 for k in np.arange(0, order): points.insert(0, None) axis.plot(points, color=color, label=label, ls=ls,linewidth=linewidth) return [mi, ma]
[docs]def plot_compared_series(original, models, colors, typeonlegend=False, save=False, file=None, tam=[20, 5], points=True, intervals=True, linewidth=1.5): """ Plot the forecasts of several one step ahead models, by point or by interval :param original: Original time series data (list) :param models: List of models to compare :param colors: List of models colors :param typeonlegend: Add the type of forecast (point / interval) on legend :param save: Save the picture on file :param file: Filename to save the picture :param tam: Size of the picture :param points: True to plot the point forecasts, False otherwise :param intervals: True to plot the interval forecasts, False otherwise :param linewidth: :return: """ fig = plt.figure(figsize=tam) ax = fig.add_subplot(111) mi = [] ma = [] legends = [] ax.plot(original, color='black', label="Original", linewidth=linewidth*1.5) for count, fts in enumerate(models, start=0): try: if fts.has_point_forecasting and points: forecasts = fts.forecast(original) if isinstance(forecasts, np.ndarray): forecasts = forecasts.tolist() mi.append(min(forecasts) * 0.95) ma.append(max(forecasts) * 1.05) for k in np.arange(0, fts.order): forecasts.insert(0, None) lbl = fts.shortname + str(fts.order if fts.is_high_order and not fts.benchmark_only else "") if typeonlegend: lbl += " (Point)" ax.plot(forecasts, color=colors[count], label=lbl, ls="-",linewidth=linewidth) if fts.has_interval_forecasting and intervals: forecasts = fts.forecast_interval(original) lbl = fts.shortname + " " + str(fts.order if fts.is_high_order and not fts.benchmark_only else "") if not points and intervals: ls = "-" else: ls = "--" tmpmi, tmpma = cUtil.plot_interval(ax, forecasts, fts.order, label=lbl, typeonlegend=typeonlegend, color=colors[count], ls=ls, linewidth=linewidth) mi.append(tmpmi) ma.append(tmpma) except ValueError as ex: print(fts.shortname) handles0, labels0 = ax.get_legend_handles_labels() lgd = ax.legend(handles0, labels0, loc=2, bbox_to_anchor=(1, 1)) legends.append(lgd) # ax.set_title(fts.name) ax.set_ylim([min(mi), max(ma)]) ax.set_ylabel('F(T)') ax.set_xlabel('T') ax.set_xlim([0, len(original)])
#Util.show_and_save_image(fig, file, save, lgd=legends)
[docs]def plotCompared(original, forecasts, labels, title): fig = plt.figure(figsize=[13, 6]) ax = fig.add_subplot(111) ax.plot(original, color='k', label="Original") for c in range(0, len(forecasts)): ax.plot(forecasts[c], label=labels[c]) handles0, labels0 = ax.get_legend_handles_labels() ax.legend(handles0, labels0) ax.set_title(title) ax.set_ylabel('F(T)') ax.set_xlabel('T') ax.set_xlim([0, len(original)]) ax.set_ylim([min(original), max(original)])
[docs]def compareModelsPlot(original, models_fo, models_ho): fig = plt.figure(figsize=[13, 6]) fig.suptitle("Comparação de modelos ") ax0 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height rows = [] for model in models_fo: fts = model["model"] ax0.plot(model["forecasted"], label=model["name"]) for model in models_ho: fts = model["model"] ax0.plot(model["forecasted"], label=model["name"]) handles0, labels0 = ax0.get_legend_handles_labels() ax0.legend(handles0, labels0)
[docs]def compareModelsTable(original, models_fo, models_ho): fig = plt.figure(figsize=[12, 4]) fig.suptitle("Comparação de modelos ") columns = ['Modelo', 'Ordem', 'Partições', 'RMSE', 'MAPE (%)'] rows = [] for model in models_fo: fts = model["model"] error_r = Measures.rmse(model["forecasted"], original) error_m = round(Measures.mape(model["forecasted"], original) * 100, 2) rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m]) for model in models_ho: fts = model["model"] error_r = Measures.rmse(model["forecasted"][fts.order:], original[fts.order:]) error_m = round(Measures.mape(model["forecasted"][fts.order:], original[fts.order:]) * 100, 2) rows.append([model["name"], fts.order, len(fts.sets), error_r, error_m]) ax1 = fig.add_axes([0, 0, 1, 1]) # left, bottom, width, height ax1.set_xticks([]) ax1.set_yticks([]) ax1.table(cellText=rows, colLabels=columns, cellLoc='center', bbox=[0, 0, 1, 1]) sup = "\\begin{tabular}{" header = "" body = "" footer = "" for c in columns: sup = sup + "|c" if len(header) > 0: header = header + " & " header = header + "\\textbf{" + c + "} " sup = sup + "|} \\hline\n" header = header + "\\\\ \\hline \n" for r in rows: lin = "" for c in r: if len(lin) > 0: lin = lin + " & " lin = lin + str(c) body = body + lin + "\\\\ \\hline \n" return sup + header + body + "\\end{tabular}"
[docs]def simpleSearch_RMSE(train, test, model, partitions, orders, save=False, file=None, tam=[10, 15], plotforecasts=False, elev=30, azim=144, intervals=False,parameters=None, partitioner=Grid.GridPartitioner,transformation=None,indexer=None): _3d = len(orders) > 1 ret = [] if _3d: errors = np.array([[0 for k in range(len(partitions))] for kk in range(len(orders))]) else: errors = [] forecasted_best = [] fig = plt.figure(figsize=tam) # fig.suptitle("Comparação de modelos ") if plotforecasts: ax0 = fig.add_axes([0, 0.4, 0.9, 0.5]) # left, bottom, width, height ax0.set_xlim([0, len(train)]) ax0.set_ylim([min(train) * 0.9, max(train) * 1.1]) ax0.set_title('Forecasts') ax0.set_ylabel('F(T)') ax0.set_xlabel('T') min_rmse = 1000000.0 best = None for pc, p in enumerate(partitions, start=0): sets = partitioner(data=train, npart=p, transformation=transformation).sets for oc, o in enumerate(orders, start=0): fts = model("q = " + str(p) + " n = " + str(o)) fts.append_transformation(transformation) fts.train(train, sets=sets, order=o, parameters=parameters) if not intervals: forecasted = fts.forecast(test) if not fts.has_seasonality: error = Measures.rmse(np.array(test[o:]), np.array(forecasted[:-1])) else: error = Measures.rmse(np.array(test[o:]), np.array(forecasted)) for kk in range(o): forecasted.insert(0, None) if plotforecasts: ax0.plot(forecasted, label=fts.name) else: forecasted = fts.forecast_interval(test) error = 1.0 - Measures.rmse_interval(np.array(test[o:]), np.array(forecasted[:-1])) if _3d: errors[oc, pc] = error else: errors.append( error ) if error < min_rmse: min_rmse = error best = fts forecasted_best = forecasted # print(min_rmse) if plotforecasts: # handles0, labels0 = ax0.get_legend_handles_labels() # ax0.legend(handles0, labels0) ax0.plot(test, label="Original", linewidth=3.0, color="black") if _3d: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim) if _3d and not plotforecasts: ax1 = Axes3D(fig, rect=[0, 1, 0.9, 0.9], elev=elev, azim=azim) ax1.set_title('Error Surface') ax1.set_ylabel('Model order') ax1.set_xlabel('Number of partitions') ax1.set_zlabel('RMSE') X, Y = np.meshgrid(partitions, orders) surf = ax1.plot_surface(X, Y, errors, rstride=1, cstride=1, antialiased=True) else: ax1 = fig.add_axes([0, 1, 0.9, 0.9]) ax1.set_title('Error Curve') ax1.set_xlabel('Number of partitions') ax1.set_ylabel('RMSE') ax1.plot(partitions, errors) ret.append(best) ret.append(forecasted_best) ret.append(min_rmse) # plt.tight_layout() cUtil.show_and_save_image(fig, file, save) return ret
[docs]def pftsExploreOrderAndPartitions(data,save=False, file=None): fig, axes = plt.subplots(nrows=4, ncols=1, figsize=[6, 8]) data_fs1 = Grid.GridPartitioner(data=data, npart=10).sets mi = [] ma = [] axes[0].set_title('Point Forecasts by Order') axes[2].set_title('Interval Forecasts by Order') for order in np.arange(1, 6): fts = pwfts.ProbabilisticWeightedFTS("") fts.shortname = "n = " + str(order) fts.train(data, sets=data_fs1.sets, order=order) point_forecasts = fts.forecast(data) interval_forecasts = fts.forecast_interval(data) lower = [kk[0] for kk in interval_forecasts] upper = [kk[1] for kk in interval_forecasts] mi.append(min(lower) * 0.95) ma.append(max(upper) * 1.05) for k in np.arange(0, order): point_forecasts.insert(0, None) lower.insert(0, None) upper.insert(0, None) axes[0].plot(point_forecasts, label=fts.shortname) axes[2].plot(lower, label=fts.shortname) axes[2].plot(upper) axes[1].set_title('Point Forecasts by Number of Partitions') axes[3].set_title('Interval Forecasts by Number of Partitions') for partitions in np.arange(5, 11): data_fs = Grid.GridPartitioner(data=data, npart=partitions).sets fts = pwfts.ProbabilisticWeightedFTS("") fts.shortname = "q = " + str(partitions) fts.train(data, sets=data_fs.sets, order=1) point_forecasts = fts.forecast(data) interval_forecasts = fts.forecast_interval(data) lower = [kk[0] for kk in interval_forecasts] upper = [kk[1] for kk in interval_forecasts] mi.append(min(lower) * 0.95) ma.append(max(upper) * 1.05) point_forecasts.insert(0, None) lower.insert(0, None) upper.insert(0, None) axes[1].plot(point_forecasts, label=fts.shortname) axes[3].plot(lower, label=fts.shortname) axes[3].plot(upper) for ax in axes: ax.set_ylabel('F(T)') ax.set_xlabel('T') ax.plot(data, label="Original", color="black", linewidth=1.5) handles, labels = ax.get_legend_handles_labels() ax.legend(handles, labels, loc=2, bbox_to_anchor=(1, 1)) ax.set_ylim([min(mi), max(ma)]) ax.set_xlim([0, len(data)]) plt.tight_layout() cUtil.show_and_save_image(fig, file, save)
[docs]def train_test_time(data, windowsize, train=0.8, **kwargs): import time tag = __pop('tag', None, kwargs) steps = __pop('steps', 0, kwargs) dataset = __pop('dataset', None, kwargs) partitions = __pop('partitions', 10, kwargs) fts_methods = __pop('methods', [], kwargs) file = kwargs.get('file', "benchmarks.db") inc = __pop("inc", 0.1, kwargs) conn = bUtil.open_benchmark_db(file) for ct, train, test in cUtil.sliding_window(data, windowsize, train, inc=inc, **kwargs): partitioner = Grid.GridPartitioner(data=train, npart=partitions) for id, fts_method in enumerate(fts_methods): print(dataset, fts_method, ct) times = [] model = fts_method(partitioner = partitioner, **kwargs) _start = time.time() model.fit(train, **kwargs) _end = time.time() times.append( _end - _start ) _start = time.time() model.predict(train, **kwargs) _end = time.time() times.append(_end - _start) for ct, method in enumerate(['train','test']): job = { 'steps': steps, 'method': method, 'time': times[ct], 'model': model.shortname, 'transformation': None, 'order': model.order, 'partitioner': partitioner.name, 'partitions': partitions, 'size': len(model) } data = bUtil.process_common_data2(dataset, tag, 'train', job) common_process_time_jobs(conn, data, job) conn.close()
[docs]def distributed_model_train_test_time(models, data, windowsize, train=0.8, **kwargs): """ Assess the train and test times for a given list of configured models and save the results on a database. :param models: A list of FTS models already configured, but not yet trained, :param data: time series data, including train and test data :param windowsize: Train/test data windows :param train: Percent of data window that will be used to train the models :param kwargs: :return: """ import time tag = __pop('tag', None, kwargs) num_batches = kwargs.get('num_batches', 1) dataset = __pop('dataset', None, kwargs) file = kwargs.get('file', "benchmarks.db") inc = __pop("inc", 0.5, kwargs) conn = bUtil.open_benchmark_db(file) for ct, train, test in cUtil.sliding_window(data, windowsize, train, inc=inc, **kwargs): for id, model in enumerate(models): print(dataset, model, ct) model.fit(train, **kwargs) for time in model.__dict__['training_time']: job = { 'steps': num_batches, 'method': 'train', 'time': time, 'model': model.shortname, 'transformation': None, 'order': model.order, 'partitioner': None, 'partitions': None, 'size': len(model) } data = bUtil.process_common_data2(dataset, tag, 'train', job) common_process_time_jobs(conn, data, job) model.predict(train, **kwargs) for time in model.__dict__['forecasting_time']: job = { 'steps': num_batches, 'method': 'test', 'time': time, 'model': model.shortname, 'transformation': None, 'order': model.order, 'partitioner': None, 'partitions': None, 'size': len(model) } data = bUtil.process_common_data2(dataset, tag, 'test', job) common_process_time_jobs(conn, data, job) conn.close()