Refine
Year of publication
Document Type
- Report (17)
- Working Paper (16)
- Article (8)
- Preprint (5)
- Book (4)
- Conference Proceeding (1)
- Doctoral Thesis (1)
Language
- English (52) (remove)
Has Fulltext
- yes (52)
Keywords
- Optimierung (15)
- Optimization (13)
- Benchmarking (5)
- Modellierung (5)
- Simulation (5)
- Globale Optimierung (4)
- Soft Computing (4)
- Disaster Risk Reduction (3)
- Evolutionärer Algorithmus (3)
- Kriging (3)
- Maschinelles Lernen (3)
- Mehrkriterielle Optimierung (3)
- Metaheuristik (3)
- Modeling (3)
- Resilienz (3)
- Risikomanagement (3)
- Sequential Parameter Optimization (3)
- Surrogate (3)
- Surrogate Models (3)
- Verwundbarkeit (3)
- 3D Printing (2)
- Anpassung (2)
- Bayesian Optimization (2)
- Climate Change Adaptation (2)
- Co-Kriging (2)
- Combined simulation (2)
- Ensemble Methods (2)
- Evolutionary Computation (2)
- Extreme Events (2)
- Gaussian Process (2)
- Genetic Programming (2)
- Imputation (2)
- Klimawandelanpassung (2)
- Klimaänderung (2)
- Metamodel (2)
- Multi-Criteria Optimization (2)
- Multiobjective Optimization (2)
- Natural Hazards (2)
- Naturgefahr (2)
- Parallelization (2)
- R (2)
- Resilience (2)
- Risk Management (2)
- Sequentielle Parameter Optimierung (2)
- Surrogat-Modellierung (2)
- Surrogate Modeling (2)
- Surrogate-based (2)
- Taxonomie (2)
- Taxonomy (2)
- Test Function (2)
- Vulnerability (2)
- 3D-Druck (1)
- Adaptive Schrittweite (1)
- Algorithm Tuning (1)
- Algorithmus (1)
- Artificial intelligence (1)
- Automated Learning (1)
- Automation (1)
- BBOB (1)
- Bacteria (1)
- Bakterien (1)
- Bayesian Learning (1)
- Bayesian Regression (1)
- Big Data (1)
- Big data platform (1)
- Biogas (1)
- Biogas Plant (1)
- Biotechnologie (1)
- Biotechnology (1)
- Board Games (1)
- Brazil (1)
- Business Intelligence (1)
- Ccomputational fluid dynamics (1)
- Cognition (1)
- Composite Materials (1)
- Computational Intelligence (1)
- Computational fluid dynamics (1)
- Conditional inference tree (1)
- Continuous Optimization (1)
- Cyclone Dust Separator (1)
- Data Analysis (1)
- Data Mining (1)
- Data Modelling (1)
- Data-Warehouse-Konzept (1)
- Datenanalyse (1)
- Decision tree (1)
- Design (1)
- Design of Experiments (1)
- Discrete Optimization (1)
- Dye (1)
- Electrostatic Precipitator (1)
- Ensemble based modeling (1)
- Entstauber (1)
- Event Detection (1)
- Evolutionary Algorithms (1)
- Evolutionary Robotics (1)
- Evolutionäre Algorithmen (1)
- Expected Improvement (1)
- Expensive Optimization (1)
- Experiment (1)
- Experimental Algorithmics (1)
- Feature selection (1)
- Fehlende Daten (1)
- Flowcurve (1)
- Flushing (1)
- Function Approximation (1)
- Funktionstest (1)
- Färben (1)
- Gauß-Prozess (1)
- Genetische Programmierung (1)
- Geografie (1)
- Health condition monitoring (1)
- Heuristics (1)
- Hot rolling (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Kapitalmarkt (1)
- Knowledge extraction (1)
- Kognition (1)
- Krisenmanagement (1)
- Künstliche Intelligenz (1)
- Landnutzungsänderungen (1)
- Lessons learned (1)
- Lifetime Prediction (1)
- Lineare Regression (1)
- Machine Learning (1)
- Machine learning (1)
- Massive Online Analysis (1)
- Material (1)
- Meta-model (1)
- Metaheuristics (1)
- Metal (1)
- Metamodell (1)
- Metamodels (1)
- Missing Data (1)
- Mixed Models (1)
- Mixed-Effects Models (1)
- Mode (1)
- Model Selection (1)
- Modelierung (1)
- Modelling (1)
- Modelllernen (1)
- Multi-criteria Optimization (1)
- Multi-fidelity (1)
- Multiaxial Fatigue (1)
- N-tuple Systems (1)
- Nachhaltigkeit (1)
- Neural Networks (1)
- Neural and Evolutionary Computing (1)
- Numerische Strömungssimulation (1)
- On-line Algorithm (1)
- Online Adaptation (1)
- Optimierungsproblem (1)
- Pandemie (1)
- Parallelisierung (1)
- Parameter Tuning (1)
- Performance (1)
- Plywise Analysis (1)
- Prognose (1)
- Referenzmodell (1)
- Regression (1)
- Reinforcement Learning (1)
- Robotics (1)
- SAP (1)
- SPOT (1)
- Sensor placement (1)
- Sensortechnik (1)
- Simulation-based Optimization (1)
- Simulationsmodell (1)
- Social Learning (1)
- Software (1)
- Spülen (1)
- Stacked Generalization (1)
- Stacking (1)
- Standardisierung (1)
- Statistics (1)
- Statistische Versuchsplanung (1)
- Surrogate Mod (1)
- Surrogate Model (1)
- Surrogate Optimization (1)
- Surrogate model (1)
- Surrogate model based optimization (1)
- Surrogate-Model-Based Optimization (1)
- Surrogates (1)
- Sustainability (1)
- System Identification (1)
- Systemidentifikation (1)
- Temporal Difference Learning (TDL) (1)
- Test function generator (1)
- Testgröße (1)
- Textiles (1)
- Time Series (1)
- Time-series (1)
- Trinkwasserversorgung (1)
- Univariate Data (1)
- Unsicherheit (1)
- Variable reduction (1)
- Varianzanalyse (1)
- Versuchsplanung (1)
- Verunreinigung (1)
- Vorverarbeitung (1)
- Wasseraufbereitungskosten (1)
- Wasserressourcen (1)
- Wasserverteilung (1)
- Water Distribution Systems (1)
- Water Quality Monitoring (1)
- Water distribution systems (1)
- Zeitreihe (1)
- Zeitreihenanalyse (1)
- Zylon Enstauber (1)
- economic assessment (1)
- ecosystem services (1)
- opportunity costs (1)
- water resources (1)
- watershed protection (1)
- Ökosystem-Dienstleistungen (1)
Institute
- Fakultät für Informatik und Ingenieurwissenschaften (F10) (26)
- Fakultät 10 / Institut für Informatik (11)
- Fakultät 09 / Institut für Rettungsingenieurwesen und Gefahrenabwehr (4)
- Fakultät 02 / Köln International School of Design (2)
- Fakultät 04 / Institut für Versicherungswesen (2)
- Fakultät 10 / Institut für Data Science, Engineering, and Analytics (2)
- Fakultät 08 / Institut für Fahrzeugtechnik (1)
- Institut für Technologie und Ressourcenmanagement in den Tropen und Subtropen (ITT) (1)
In the present paper a calculation tool for the lifetime prediction of composite materials with focus on local multiaxial
stress states and different local stress ratios within each lamina is developed. The approach is based on repetitiv, progressive in-plane stress calculations using classical laminate theory with subsequent analysis of the material stressing effort and use of appropriate material degradation models. Therefore experimentally data of S-N curves are
used to generate anistropic constant life diagrams for a closer examination of critical fracture planes under any given combination of local stress ratios. The model is verified against various balanced angle plies and multi-directional
laminates with arbitrary stacking sequences and varying stress ratios throughout the analysis. Different sections of the
model, such as residual strength and residual stiffness, are examined and verified over a wide range of load cycles. The obtained results agree very well with the analyzed experimental data.
RGP is genetic programming system based on, as well as fully integrated into, the R environment. The system implements classical tree-based genetic programming as well as other variants including, for example, strongly typed genetic programming and Pareto genetic programming. It strives for high modularity through a consistent architecture that allows the customization and replacement of every algorithm component, while maintaining accessibility for new users by adhering to the "convention over configuration" principle.
Multi-criteria optimization has gained increasing attention during the last decades. This article exemplifies multi-criteria features, which are implemented in the statistical software package SPOT. It describes related software packages such as mco and emoa and gives a comprehensive introduction to simple multi criteria optimization tasks. Several hands-on examples are used for illustration. The article is well-suited as a starting point for performing multi-criteria optimization tasks with SPOT.
Surrogate-based optimization and nature-inspired metaheuristics have become the state of the art in solving real-world optimization problems. Still, it is difficult for beginners and even experts to get an overview that explains their advantages in comparison to the large number of available methods in the scope of continuous optimization. Available taxonomies lack the integration of surrogate-based approaches and thus their embedding in the larger context of this broad field.
This article presents a taxonomy of the field, which further matches the idea of nature-inspired algorithms, as it is based on the human behavior in path finding. Intuitive analogies make it easy to conceive the most basic principles of the search algorithms, even for beginners and non-experts in this area of research. However, this scheme does not oversimplify the high complexity of the different algorithms, as the class identifier only defines a descriptive meta-level of the algorithm search strategies. The taxonomy was established by exploring and matching algorithm schemes, extracting similarities and differences, and creating a set of classification indicators to distinguish between five distinct classes. In practice, this taxonomy allows recommendations for the applicability of the corresponding algorithms and helps developers trying to create or improve their own algorithms.
This survey compiles ideas and recommendations from more than a dozen researchers with different backgrounds and from different institutes around the world. Promoting best practice in benchmarking is its main goal. The article discusses eight essential topics in benchmarking: clearly stated goals, well- specified problems, suitable algorithms, adequate performance measures, thoughtful analysis, effective and efficient designs, comprehensible presentations, and guaranteed reproducibility. The final goal is to provide well-accepted guidelines (rules) that might be useful for authors and reviewers. As benchmarking in optimization is an active and evolving field of research this manuscript is meant to co-evolve over time by means of periodic updates.
Computational intelligence methods have gained importance in several real-world domains such as process optimization, system identification, data mining, or statistical quality control. Tools are missing, which determine the applicability of computational intelligence methods in these application domains in an objective manner. Statistics provide methods for comparing algorithms on certain data sets. In the past, several test suites were presented and considered as state of the art. However, there are several drawbacks of these test suites, namely: (i) problem instances are somehow artificial and have no direct link to real-world settings; (ii) since there is a fixed number of test instances, algorithms can be fitted or tuned to this specific and very limited set of test functions; (iii) statistical tools for comparisons of several algorithms on several test problem instances are relatively complex and not easily to analyze. We propose a methodology to overcome these difficulties. It is based on standard ideas from statistics: analysis of variance and its extension to mixed models. This paper combines essential ideas from two approaches: problem generation and statistical analysis of computer experiments.
When using machine learning techniques for learning a function approximation from given data it is often a difficult task to select the right modeling technique.
In many real-world settings is no preliminary knowledge about the objective function available. Then it might be beneficial if the algorithm could learn all models by itself and select the model that suits best to the problem.
This approach is known as automated model selection. In this work we propose a
generalization of this approach.
It combines the predictions of several into one more accurate ensemble surrogate model. This approach is studied in a fundamental way, by first evaluating minimalistic ensembles of only two surrogate models in detail and then proceeding to ensembles with three and more surrogate models.
The results show to what extent combinations of models can perform better than single surrogate models and provides insights into the scalability and robustness of the approach. The study focuses on multi-modal functions topologies, which are important in surrogate-assisted global optimization.
This paper introduces CAAI, a novel cognitive architecture for artificial intelligence in cyber-physical production systems. The goal of the architecture is to reduce the implementation effort for the usage of artificial intelligence algorithms. The core of the CAAI is a cognitive module that processes declarative goals of the user, selects suitable models and algorithms, and creates a configuration for the execution of a processing pipeline on a big data platform. Constant observation and evaluation against performance criteria assess the performance of pipelines for many and varying use cases. Based on these evaluations, the pipelines are automatically adapted if necessary. The modular design with well-defined interfaces enables the reusability and extensibility of pipeline components. A big data platform implements this modular design supported by technologies such as Docker, Kubernetes, and Kafka for virtualization and orchestration of the individual components and their communication. The implementation of the architecture is evaluated using a real-world use case.
Collective Defined Contribution Plans – Backtesting Based on German Capital Market Data 1950 - 2022
(2022)
Using historical capital market data for Germany (1950-2022) we analyze and compare (individual) defined contribution (IDC-) and collective defined contribution (CDC) pension plans. To this end we define simple asset liability management rules that govern a CDC pension plan and compare these to IDC-plans with the same asset allovation. Our main result is, that the CDC pension plans allow for a significant improvement of the risk return profile compared to individual pension plans. Hereby we consider different risk measures. This empirical study affirms the theoretical results based on stochastic CDC-models.
The availability of several CPU cores on current computers enables
parallelization and increases the computational power significantly.
Optimization algorithms have to be adapted to exploit these highly
parallelized systems and evaluate multiple candidate solutions in
each iteration. This issue is especially challenging for expensive
optimization problems, where surrogate models are employed to
reduce the load of objective function evaluations.
This paper compares different approaches for surrogate modelbased
optimization in parallel environments. Additionally, an easy
to use method, which was developed for an industrial project, is
proposed. All described algorithms are tested with a variety of
standard benchmark functions. Furthermore, they are applied to
a real-world engineering problem, the electrostatic precipitator
problem. Expensive computational fluid dynamics simulations are
required to estimate the performance of the precipitator. The task
is to optimize a gas-distribution system so that a desired velocity
distribution is achieved for the gas flow throughout the precipitator.
The vast amount of possible configurations leads to a complex
discrete valued optimization problem. The experiments indicate
that a hybrid approach works best, which proposes candidate solutions
based on different surrogate model-based infill criteria and
evolutionary operators.