Refine
Has Fulltext
- yes (16)
Keywords
- Optimierung (11)
- Modellierung (9)
- Soft Computing (7)
- Simulation (4)
- Computational Intelligence (3)
- Evolutionärer Algorithmus (3)
- Genetic Programming (3)
- Globale Optimierung (3)
- Mehrkriterielle Optimierung (3)
- Optimization (3)
- Sequentielle Parameter Optimierung (3)
- Versuchsplanung (3)
- Co-Kriging (2)
- Evolutionary Algorithms (2)
- Evolutionäre Algorithmen (2)
- Genetisches Programmieren (2)
- Kriging (2)
- Metaheuristik (2)
- Multi-Criteria Optimization (2)
- Multiobjective Optimization (2)
- Optimierungsproblem (2)
- Prognose (2)
- Sequential Parameter Optimization (2)
- Surrogat-Modellierung (2)
- Surrogate Modeling (2)
- Zeitreihenanalyse (2)
- Adaptive Schrittweite (1)
- Biogas (1)
- Biogas Plant (1)
- Board Games (1)
- Cyclone Dust Separator (1)
- Datenanalyse (1)
- Design of Experiments (1)
- Ensemble Methods (1)
- Entstauber (1)
- Event Detection (1)
- Evolution Strategies (1)
- Evolutionsstrategie (1)
- Evolutionsstrategien (1)
- Expected Improvement (1)
- Fehlende Daten (1)
- Finanzwirtschaft (1)
- Genetic Algorithms (1)
- Genetic programming (1)
- Genetische Algorithmen (1)
- Genetische Programmierung (1)
- Imputation (1)
- Klassifikation (1)
- Maschinelles Lernen (1)
- Metamodel (1)
- Missing Data (1)
- Mixed Models (1)
- Mixed-Effects Models (1)
- Modeling (1)
- Modelling (1)
- Modelllernen (1)
- Multi-criteria Optimization (1)
- Multi-fidelity (1)
- N-tuple Systems (1)
- Online Adaptation (1)
- Parametertuning (1)
- R (1)
- Regression (1)
- Reinforcement Learning (1)
- Simulated annealing (1)
- Simulationsmodell (1)
- Software (1)
- Statistics (1)
- Statistische Versuchsplanung (1)
- Surrogate Models (1)
- Surrogate Optimization (1)
- Surrogate-Model-Based Optimization (1)
- Surrogate-model-based Optimization (1)
- Surrogatmodellbasierte Optimierung (1)
- System Identification (1)
- Systemidentifikation (1)
- Temporal Difference Learning (TDL) (1)
- Time-series (1)
- Unsicherheit (1)
- Varianzanalyse (1)
- Vorverarbeitung (1)
- Wasserwirtschaft (1)
- Water Quality Monitoring (1)
- Zeitreihe (1)
- Zylon Enstauber (1)
- classification (1)
- model-assisted optimization (1)
- modellbasierte Optimierung (1)
- regression (1)
Institute
- Fakultät 10 / Institut für Informatik (16) (remove)
Evolutionary algorithm (EA) is an umbrella term used to describe population-based stochastic direct search algorithms that in some sense mimic natural evolution. Prominent representatives of such algorithms are genetic algorithms, evolution strategies, evolutionary programming, and genetic programming. On the basis of the evolutionary cycle, similarities and differences between these algorithms are described. We briefly discuss how EAs can be adapted to work well in case of multiple objectives, and dynamic or noisy optimization problems. We look at the tuning of algorithms and present some recent developments coming from theory. Finally, typical applications of EAs to real-world problems are shown, with special emphasis on data-mining applications
Computational intelligence methods have gained importance in several real-world domains such as process optimization, system identification, data mining, or statistical quality control. Tools are missing, which determine the applicability of computational intelligence methods in these application domains in an objective manner. Statistics provide methods for comparing algorithms on certain data sets. In the past, several test suites were presented and considered as state of the art. However, there are several drawbacks of these test suites, namely: (i) problem instances are somehow artificial and have no direct link to real-world settings; (ii) since there is a fixed number of test instances, algorithms can be fitted or tuned to this specific and very limited set of test functions; (iii) statistical tools for comparisons of several algorithms on several test problem instances are relatively complex and not easily to analyze. We propose amethodology to overcome these dificulties. It is based on standard ideas from statistics: analysis of variance and its extension to mixed models. This work combines essential ideas from two approaches: problem generation and statistical analysis of computer experiments.
Learning board games by self-play has a long tradition in computational intelligence for games. Based on Tesauro’s seminal success with TD-Gammon in 1994, many successful agents use temporal difference learning today. But in order to be successful with temporal difference learning on game tasks, often a careful selection of features and a large number of training games is necessary. Even for board games of moderate complexity like Connect-4, we found in previous work that a very rich initial feature set and several millions of game plays are required. In this work we investigate different approaches of online-adaptable learning rates like Incremental Delta Bar Delta (IDBD) or Temporal Coherence Learning (TCL) whether they have the potential to speed up learning for such a complex task. We propose a new variant of TCL with geometric step size changes. We compare those algorithms with several other state-of-the-art learning rate adaptation algorithms and perform a case study on the sensitivity with respect to their meta parameters. We show that in this set of learning algorithms those with geometric step size changes outperform those other algorithms with constant step size changes. Algorithms with nonlinear output functions are slightly better than linear ones. Algorithms with geometric step size changes learn faster by a factor of 4 as compared to previously published results on the task Connect-4.
An essential task for operation and planning of biogas plants is the optimization of substrate feed mixtures. Optimizing the monetary gain requires the determination of the exact amounts of maize, manure, grass silage, and other substrates. Accurate simulation models are mandatory for this optimization, because the underlying chemical processes are very slow. The simulation models themselves may be time-consuming to evaluate, hence we show how to use surrogate-model-based approaches to optimize biogas plants efficiently. In detail, a Kriging surrogate is employed. To improve model quality of this surrogate, we integrate cheaply available data into the optimization process. Doing so, Multi-fidelity modeling methods like Co-Kriging are employed. Furthermore, a two-layered modeling approach is employed to avoid deterioration of model quality due to discontinuities in the search space. At the same time, the cheaply available data is shown to be very useful for initialization of the employed optimization algorithms. Overall, we show how biogas plants can be efficiently modeled using data-driven methods, avoiding discontinuities as well as including cheaply available data. The application of the derived surrogate models to an optimization process is shown to be very difficult, yet successful for a lower problem dimension.
SOMA - Systematische Optimierung von Modellen in IT- und Automatisierungstechnik (Schlussbericht)
(2013)
Das im Rahmen der Förderlinie IngenieurNachwuchs geförderte Forschungsvorhaben Systematische Optimierung von Modellen für Informations- und Automatisierungs-technik (kurz: SOMA) startete im August 2009. Eine wesentliche Zielsetzung war die Entwicklung und Optimierung von Modellen zur Prognose von Zielgrößen. Ein wichtiges Merkmal ist dabei die effiziente Optimierung dieser Modelle, welche es ermöglichen soll, mit einer streng limitierten Anzahl an Auswertungen gute Parametereinstellungen zu bestimmen. Mithilfe dieser genaueren Parametrierungen der unterliegenden Modelle können unter Einbeziehung neuer merkmalserzeugender Verfahren insbesondere für kleine und mittelständische Unternehmen verbesserte Lösungen erzielt werden. Als direkter Gewinn derartiger Verbesserungen konnte für KMUs ein geeignetes Framework für Modellierungs- und Prognoseaufgaben be- reitgestellt werden, sodass mit geringem technischem und personellen Aufwand performante und nahezu optimale Lösungen erzielt werden können. Dieser Schluss-bericht beschreibt die im Projekt durchgeführten Maßnahmen und Ergebnisse.
This paper introduces UniFIeD, a new data preprocessing method for time series. UniFIeD can cope with large intervals of missing data. A scalable test function generator, which allows the simulation of time series with different gap sizes, is presented additionally. An experimental study demonstrates that (i) UniFIeD shows a significant better performance than simple imputation methods and (ii) UniFIeD is able to handle situations, where advanced imputation methods fail. The results are independent from the underlying error measurements.
Dieser Schlussbericht beschreibt die im Projekt „Methoden der Computational Intelligence für Vorhersagemodelle in der Finanzund Wasserwirtschaft“ (FIWA) im Zeitraum von Juni 2009 bis einschließlich November 2012 erzielten Ergebnisse. In der Praxis werden für diese Vorhersagemodelle Verfahren der linearen und nichtlinearen Regression, NN, Support Vector Machines (SVM) und viele weitere Verfahren eingesetzt. Das Projekt FIWA befasste sich mit der Entwicklung modularer Systeme zur Analyse und Prognose von Daten aus der Finanz- und Wasserwirtschaft mittels Verfahren der Computational Intelligence (CI) mit methodischem Fokus auf dem CI-Unterbereich Genetic Programming (GP). Ein zentrales Ergebnis der wissenschaftlichtechnischen Arbeit im Projekt FIWA ist die Entwicklung der Open-Source Software RGP. Dabei handelt es sich um ein Software- Framework für GP, welches auf die automatische Erstellung von Vorhersagemodellen spezialisiert ist. Für die Finanzwirtschaft stand ein Handelssimulator zu Verfügung, der auf Basis von echten Finanzdaten die Qualität verschiedener Strategien testen kann. Dieser wurde im Projekt weiterentwickelt. GP wurde genutzt, um auf Basis der Simulationen genaue Vorhersagen und damit verbesserte Handelsstrategien zu entwerfen. Auch für die Wasserwirtschaft wurden Prognoseverfahren mit GP entwickelt. Der Schwerpunkt lag dabei auf der Füllstandprognose für Regenüberlaufbecken. Hier konnten moderne Verfahren mit GP oder SVM klassische Methoden deutlich schlagen oder verbessern. Auch der Einsatz von Sequentieller Parameter Optimierung zeigte signifikante Verbesserungen für die Prognosegenauigkeit. Dabei war die Kombination von klassischen Methoden und GP besonders erfolgreich. GP ist nach wie vor ein sehr aktives Forschungsgebiet und erlaubt auch für die Folgezeit zahlreiche Kooperationen mit den Partnern der Fachhochschule Köln. Sowohl für technische Anwendungen als auch zur Lösung von Forschungsfragen bieten sich zahlreiche Möglichkeiten an.
Cyclone Dust Separators are devices often used to filter solid particles from flue gas. Such cyclones are supposed to filter as much solid particles from the carrying gas as possible. At the same time, they should only introduce a minimal pressure loss to the system. Hence, collection efficiency has to be maximized and pressure loss minimized. Both the collection efficiency and pressure loss are heavily influenced by the cyclones geometry. In this paper, we optimize seven geometrical parameters of an analytical cyclone model. Furthermore, noise variables are introduced to the model, representing the non-deterministic structure of the real-world problem. This is used to investigate robustness and sensitivity of solutions. Both the deterministic as well as the stochastic model are optimized with an SMS-EMOA. The SMS-EMOA is compared to a single objective optimization algorithm. For the harder, stochastic optimization problem, a surrogate-model-supported SMS-EMOA is compared against the model-free SMS-EMOA. The model supported approach yields better solutions with the same run-time budget.
RGP is genetic programming system based on, as well as fully integrated into, the R environment. The system implements classical tree-based genetic programming as well as other variants including, for example, strongly typed genetic programming and Pareto genetic programming. It strives for high modularity through a consistent architecture that allows the customization and replacement of every algorithm component, while maintaining accessibility for new users by adhering to the "convention over configuration" principle.
Multi-criteria optimization has gained increasing attention during the last decades. This article exemplifies multi-criteria features, which are implemented in the statistical software package SPOT. It describes related software packages such as mco and emoa and gives a comprehensive introduction to simple multi criteria optimization tasks. Several hands-on examples are used for illustration. The article is well-suited as a starting point for performing multi-criteria optimization tasks with SPOT.