Refine
Year of publication
Document Type
- Report (22)
- Working Paper (13)
- Article (7)
- Preprint (5)
- Conference Proceeding (1)
Has Fulltext
- yes (48)
Keywords
- Optimierung (17)
- Optimization (12)
- Soft Computing (9)
- Modellierung (7)
- Evolutionärer Algorithmus (5)
- Globale Optimierung (5)
- Simulation (5)
- Benchmarking (4)
- Metaheuristik (4)
- Computational Intelligence (3)
- Genetisches Programmieren (3)
- Kriging (3)
- Maschinelles Lernen (3)
- Mehrkriterielle Optimierung (3)
- Modeling (3)
- Sequential Parameter Optimization (3)
- Sequentielle Parameter Optimierung (3)
- Surrogate (3)
- Surrogate Models (3)
- Versuchsplanung (3)
- 3D Printing (2)
- Bayesian Optimization (2)
- Co-Kriging (2)
- Combined simulation (2)
- Design of Experiments (2)
- Ensemble Methods (2)
- Event Detection (2)
- Evolutionary Computation (2)
- Evolutionsstrategie (2)
- Genetic Programming (2)
- Genetische Algorithmen (2)
- Imputation (2)
- Künstliche Intelligenz (2)
- Machine Learning (2)
- Metamodel (2)
- Multi-Criteria Optimization (2)
- Multiobjective Optimization (2)
- Optimierungsproblem (2)
- Parallelization (2)
- Prognose (2)
- R (2)
- Surrogat-Modellierung (2)
- Surrogate Modeling (2)
- Surrogate-based (2)
- Taxonomie (2)
- Taxonomy (2)
- 3D-Druck (1)
- Algorithm Tuning (1)
- Algorithmus (1)
- Angewandte Mathematik (1)
- Artificial intelligence (1)
- Automated Learning (1)
- BBOB (1)
- Bayesian Learning (1)
- Bayesian Regression (1)
- Big Data (1)
- Big data platform (1)
- Biogas (1)
- Biogas Plant (1)
- Ccomputational fluid dynamics (1)
- Cognition (1)
- Computational fluid dynamics (1)
- Conditional inference tree (1)
- Cyber-physische Produktionssysteme (1)
- Cyclone Dust Separator (1)
- Data Analysis (1)
- Data Mining (1)
- Data Modelling (1)
- Datenanalyse (1)
- Decision tree (1)
- Discrete Optimization (1)
- Electrostatic Precipitator (1)
- Ensemble based modeling (1)
- Entdeckendes Lernen (1)
- Entstauber (1)
- Erfahrungsbericht (1)
- Evolution Strategies (1)
- Evolutionary Algorithms (1)
- Evolutionary Robotics (1)
- Evolutionsstrategien (1)
- Evolutionäre Algorithmen (1)
- Expected Improvement (1)
- Expensive Optimization (1)
- Experiment (1)
- Experimental Algorithmics (1)
- Faserverbundwerkstoffe (1)
- Feature selection (1)
- Fehlende Daten (1)
- Finanzwirtschaft (1)
- Flowcurve (1)
- Flushing (1)
- Forschendes Lernen (1)
- Function Approximation (1)
- Funktionstest (1)
- Gaussian Process (1)
- Genetic Algorithms (1)
- Genetic programming (1)
- Health condition monitoring (1)
- Heuristics (1)
- Hot rolling (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Knowledge extraction (1)
- Kognition (1)
- Kognitive Referenzarchitektur (1)
- Lineare Regression (1)
- Machine learning (1)
- Massive Online Analysis (1)
- Meta-model (1)
- Metaheuristics (1)
- Metal (1)
- Metamodell (1)
- Metamodels (1)
- Missing Data (1)
- Mixed Models (1)
- Mixed-Effects Models (1)
- Model Selection (1)
- Modelierung (1)
- Multi-criteria Optimization (1)
- Multi-fidelity (1)
- Neural Networks (1)
- Neural and Evolutionary Computing (1)
- Numerische Strömungssimulation (1)
- On-line Algorithm (1)
- Parallelisierung (1)
- Parameter Tuning (1)
- Parametertuning (1)
- Performance (1)
- Predictive Analytics (1)
- Promotion (1)
- Regression (1)
- Robotics (1)
- SPOT (1)
- Sensor placement (1)
- Sensortechnik (1)
- Signalanalyse (1)
- Simulated annealing (1)
- Simulation-based Optimization (1)
- Simulationsmodell (1)
- Social Learning (1)
- Spülen (1)
- Stacked Generalization (1)
- Stacking (1)
- Statistics (1)
- Statistische Versuchsplanung (1)
- Structural Health Monitoring (1)
- Surrogate Mod (1)
- Surrogate Model (1)
- Surrogate Optimization (1)
- Surrogate model (1)
- Surrogate model based optimization (1)
- Surrogate-Model-Based Optimization (1)
- Surrogate-model-based Optimization (1)
- Surrogates (1)
- Surrogatmodellbasierte Optimierung (1)
- Test Function (1)
- Test function generator (1)
- Time Series (1)
- Time-series (1)
- Trinkwasser (1)
- Trinkwasserversorgung (1)
- Univariate Data (1)
- Unsicherheit (1)
- Variable reduction (1)
- Varianzanalyse (1)
- Verunreinigung (1)
- Vorgehensmodell (1)
- Vorverarbeitung (1)
- Wasserverteilung (1)
- Wasserwirtschaft (1)
- Water Distribution Systems (1)
- Water Quality Monitoring (1)
- Water distribution systems (1)
- Zeitreihe (1)
- Zeitreihenanalyse (1)
- Zylon Enstauber (1)
Multi-criteria optimization has gained increasing attention during the last decades. This article exemplifies multi-criteria features, which are implemented in the statistical software package SPOT. It describes related software packages such as mco and emoa and gives a comprehensive introduction to simple multi criteria optimization tasks. Several hands-on examples are used for illustration. The article is well-suited as a starting point for performing multi-criteria optimization tasks with SPOT.
There is a strong need for sound statistical analysis of simulation and optimization algorithms. Based on this analysis, improved parameter settings can be determined. This will be referred to as tuning. Model-based investigations are common approaches in simulation and optimization. The sequential parameter optimization toolbox (SPOT), which is implemented as a package for the statistical programming language R, provides sophisticated means for tuning and understanding simulation and optimization algorithms. The toolbox includes methods for tuning based on classical regression and analysis of variance techniques; tree-based models such as classification and regressions trees (CART) and random forest; Gaussian process models (Kriging), and combinations of different meta-modeling approaches. This article exemplifies how an existing optimization algorithm, namely simulated annealing, can be tuned using the SPOT framework.
Surrogate-based optimization and nature-inspired metaheuristics have become the state of the art in solving real-world optimization problems. Still, it is difficult for beginners and even experts to get an overview that explains their advantages in comparison to the large number of available methods in the scope of continuous optimization. Available taxonomies lack the integration of surrogate-based approaches and thus their embedding in the larger context of this broad field.
This article presents a taxonomy of the field, which further matches the idea of nature-inspired algorithms, as it is based on the human behavior in path finding. Intuitive analogies make it easy to conceive the most basic principles of the search algorithms, even for beginners and non-experts in this area of research. However, this scheme does not oversimplify the high complexity of the different algorithms, as the class identifier only defines a descriptive meta-level of the algorithm search strategies. The taxonomy was established by exploring and matching algorithm schemes, extracting similarities and differences, and creating a set of classification indicators to distinguish between five distinct classes. In practice, this taxonomy allows recommendations for the applicability of the corresponding algorithms and helps developers trying to create or improve their own algorithms.
This survey compiles ideas and recommendations from more than a dozen researchers with different backgrounds and from different institutes around the world. Promoting best practice in benchmarking is its main goal. The article discusses eight essential topics in benchmarking: clearly stated goals, well- specified problems, suitable algorithms, adequate performance measures, thoughtful analysis, effective and efficient designs, comprehensible presentations, and guaranteed reproducibility. The final goal is to provide well-accepted guidelines (rules) that might be useful for authors and reviewers. As benchmarking in optimization is an active and evolving field of research this manuscript is meant to co-evolve over time by means of periodic updates.
Computational intelligence methods have gained importance in several real-world domains such as process optimization, system identification, data mining, or statistical quality control. Tools are missing, which determine the applicability of computational intelligence methods in these application domains in an objective manner. Statistics provide methods for comparing algorithms on certain data sets. In the past, several test suites were presented and considered as state of the art. However, there are several drawbacks of these test suites, namely: (i) problem instances are somehow artificial and have no direct link to real-world settings; (ii) since there is a fixed number of test instances, algorithms can be fitted or tuned to this specific and very limited set of test functions; (iii) statistical tools for comparisons of several algorithms on several test problem instances are relatively complex and not easily to analyze. We propose a methodology to overcome these difficulties. It is based on standard ideas from statistics: analysis of variance and its extension to mixed models. This paper combines essential ideas from two approaches: problem generation and statistical analysis of computer experiments.
When using machine learning techniques for learning a function approximation from given data it is often a difficult task to select the right modeling technique.
In many real-world settings is no preliminary knowledge about the objective function available. Then it might be beneficial if the algorithm could learn all models by itself and select the model that suits best to the problem.
This approach is known as automated model selection. In this work we propose a
generalization of this approach.
It combines the predictions of several into one more accurate ensemble surrogate model. This approach is studied in a fundamental way, by first evaluating minimalistic ensembles of only two surrogate models in detail and then proceeding to ensembles with three and more surrogate models.
The results show to what extent combinations of models can perform better than single surrogate models and provides insights into the scalability and robustness of the approach. The study focuses on multi-modal functions topologies, which are important in surrogate-assisted global optimization.
This paper introduces CAAI, a novel cognitive architecture for artificial intelligence in cyber-physical production systems. The goal of the architecture is to reduce the implementation effort for the usage of artificial intelligence algorithms. The core of the CAAI is a cognitive module that processes declarative goals of the user, selects suitable models and algorithms, and creates a configuration for the execution of a processing pipeline on a big data platform. Constant observation and evaluation against performance criteria assess the performance of pipelines for many and varying use cases. Based on these evaluations, the pipelines are automatically adapted if necessary. The modular design with well-defined interfaces enables the reusability and extensibility of pipeline components. A big data platform implements this modular design supported by technologies such as Docker, Kubernetes, and Kafka for virtualization and orchestration of the individual components and their communication. The implementation of the architecture is evaluated using a real-world use case.
Dieser Schlussbericht beschreibt die im Projekt „CI-basierte mehrkriterielle Optimierungsverfahren für Anwendungen in der Industrie“ (CIMO) im Zeitraum von November 2011 bis einschließlich Oktober 2014 erzielten Ergebnisse. Für aufwändige Optimierungsprobleme aus der Industrie wurden geeignete Lösungsverfahren entwickelt. Der Schwerpunkt lag hierbei auf Methoden aus den Bereichen Computational Intelligence (CI) und Surrogatmodellierung. Diese bieten die Möglichkeit, wichtige Herausforderung von aufwändigen, komplexen Optimierungsproblemen zu lösen. Die entwickelten Methoden können verschiedene konfliktäre Zielgrößen berücksichtigen, verschiedene Hierarchieebenen des Problems in die Optimierung integrieren, Nebenbedingungen beachten, vektorielle aber auch strukturierte Daten verarbeiten (kombinatorische Optimierung) sowie die Notwendigkeit teurer/zeitaufwändiger Zielfunktionsberechnungen reduzieren. Die entwickelten Methoden wurden schwerpunktmäßig auf einer Problemstellung aus der Kraftwerkstechnik angewendet, nämlich der Optimierung der Geometrie eines Fliehkraftabscheiders (auch: Zyklon), der Staubanteile aus Abgasen filtert. Das Optimierungsproblem, das diese FIiehkraftabscheider aufwerfen, führt zu konfliktären Zielsetzungen (z.B. Druckverlust, Abscheidegrad). Zyklone können unter anderem über aufwändige Computational Fluid Dynamics (CFD) Simulationen berechnet werden, es stehen aber auch einfache analytische Gleichungen als Schätzung zu Verfügung. Die Verknüpfung von beidem zeigt hier beispielhaft wie Hierarchieebenen eines Optimierungsproblems mit den Methoden des Projektes verbunden werden können. Neben dieser Schwerpunktanwendung konnte auch gezeigt werden, dass die Methoden in vielen weiteren Bereichen Erfolgreich zur Anwendung kommen können: Biogaserzeugung, Wasserwirtschaft, Stahlindustrie. Die besondere Herausforderung der behandelten Probleme und Methoden bietet viele wichtige Forschungsmöglichkeiten für zukünftige Projekte, die derzeit durch die Projektpartner vorbereitet werden.
The availability of several CPU cores on current computers enables
parallelization and increases the computational power significantly.
Optimization algorithms have to be adapted to exploit these highly
parallelized systems and evaluate multiple candidate solutions in
each iteration. This issue is especially challenging for expensive
optimization problems, where surrogate models are employed to
reduce the load of objective function evaluations.
This paper compares different approaches for surrogate modelbased
optimization in parallel environments. Additionally, an easy
to use method, which was developed for an industrial project, is
proposed. All described algorithms are tested with a variety of
standard benchmark functions. Furthermore, they are applied to
a real-world engineering problem, the electrostatic precipitator
problem. Expensive computational fluid dynamics simulations are
required to estimate the performance of the precipitator. The task
is to optimize a gas-distribution system so that a desired velocity
distribution is achieved for the gas flow throughout the precipitator.
The vast amount of possible configurations leads to a complex
discrete valued optimization problem. The experiments indicate
that a hybrid approach works best, which proposes candidate solutions
based on different surrogate model-based infill criteria and
evolutionary operators.
As the amount of data gathered by monitoring systems increases, using computational tools to analyze it becomes a necessity.
Machine learning algorithms can be used in both regression and classification problems, providing useful insights while avoiding the bias and proneness to errors of humans. In this paper, a specific kind of decision tree algorithm, called conditional inference tree, is used to extract relevant knowledge from data that pertains to electrical motors. The model is chosen due to its flexibility, strong statistical foundation, as well as great capabilities to generalize and cope with problems in the data. The obtained knowledge is organized in a structured way and then analyzed in the context of health condition monitoring. The final
results illustrate how the approach can be used to gain insight into the system and present the results in an understandable, user-friendly manner