Refine
Year of publication
Document Type
- Report (17)
- Working Paper (16)
- Article (8)
- Preprint (5)
- Book (4)
- Conference Proceeding (1)
- Doctoral Thesis (1)
Language
- English (52) (remove)
Has Fulltext
- yes (52)
Keywords
- Optimierung (15)
- Optimization (13)
- Benchmarking (5)
- Modellierung (5)
- Simulation (5)
- Globale Optimierung (4)
- Soft Computing (4)
- Disaster Risk Reduction (3)
- Evolutionärer Algorithmus (3)
- Kriging (3)
- Maschinelles Lernen (3)
- Mehrkriterielle Optimierung (3)
- Metaheuristik (3)
- Modeling (3)
- Resilienz (3)
- Risikomanagement (3)
- Sequential Parameter Optimization (3)
- Surrogate (3)
- Surrogate Models (3)
- Verwundbarkeit (3)
- 3D Printing (2)
- Anpassung (2)
- Bayesian Optimization (2)
- Climate Change Adaptation (2)
- Co-Kriging (2)
- Combined simulation (2)
- Ensemble Methods (2)
- Evolutionary Computation (2)
- Extreme Events (2)
- Gaussian Process (2)
- Genetic Programming (2)
- Imputation (2)
- Klimawandelanpassung (2)
- Klimaänderung (2)
- Metamodel (2)
- Multi-Criteria Optimization (2)
- Multiobjective Optimization (2)
- Natural Hazards (2)
- Naturgefahr (2)
- Parallelization (2)
- R (2)
- Resilience (2)
- Risk Management (2)
- Sequentielle Parameter Optimierung (2)
- Surrogat-Modellierung (2)
- Surrogate Modeling (2)
- Surrogate-based (2)
- Taxonomie (2)
- Taxonomy (2)
- Test Function (2)
- Vulnerability (2)
- 3D-Druck (1)
- Adaptive Schrittweite (1)
- Algorithm Tuning (1)
- Algorithmus (1)
- Artificial intelligence (1)
- Automated Learning (1)
- Automation (1)
- BBOB (1)
- Bacteria (1)
- Bakterien (1)
- Bayesian Learning (1)
- Bayesian Regression (1)
- Big Data (1)
- Big data platform (1)
- Biogas (1)
- Biogas Plant (1)
- Biotechnologie (1)
- Biotechnology (1)
- Board Games (1)
- Brazil (1)
- Business Intelligence (1)
- Ccomputational fluid dynamics (1)
- Cognition (1)
- Composite Materials (1)
- Computational Intelligence (1)
- Computational fluid dynamics (1)
- Conditional inference tree (1)
- Continuous Optimization (1)
- Cyclone Dust Separator (1)
- Data Analysis (1)
- Data Mining (1)
- Data Modelling (1)
- Data-Warehouse-Konzept (1)
- Datenanalyse (1)
- Decision tree (1)
- Design (1)
- Design of Experiments (1)
- Discrete Optimization (1)
- Dye (1)
- Electrostatic Precipitator (1)
- Ensemble based modeling (1)
- Entstauber (1)
- Event Detection (1)
- Evolutionary Algorithms (1)
- Evolutionary Robotics (1)
- Evolutionäre Algorithmen (1)
- Expected Improvement (1)
- Expensive Optimization (1)
- Experiment (1)
- Experimental Algorithmics (1)
- Feature selection (1)
- Fehlende Daten (1)
- Flowcurve (1)
- Flushing (1)
- Function Approximation (1)
- Funktionstest (1)
- Färben (1)
- Gauß-Prozess (1)
- Genetische Programmierung (1)
- Geografie (1)
- Health condition monitoring (1)
- Heuristics (1)
- Hot rolling (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Kapitalmarkt (1)
- Knowledge extraction (1)
- Kognition (1)
- Krisenmanagement (1)
- Künstliche Intelligenz (1)
- Landnutzungsänderungen (1)
- Lessons learned (1)
- Lifetime Prediction (1)
- Lineare Regression (1)
- Machine Learning (1)
- Machine learning (1)
- Massive Online Analysis (1)
- Material (1)
- Meta-model (1)
- Metaheuristics (1)
- Metal (1)
- Metamodell (1)
- Metamodels (1)
- Missing Data (1)
- Mixed Models (1)
- Mixed-Effects Models (1)
- Mode (1)
- Model Selection (1)
- Modelierung (1)
- Modelling (1)
- Modelllernen (1)
- Multi-criteria Optimization (1)
- Multi-fidelity (1)
- Multiaxial Fatigue (1)
- N-tuple Systems (1)
- Nachhaltigkeit (1)
- Neural Networks (1)
- Neural and Evolutionary Computing (1)
- Numerische Strömungssimulation (1)
- On-line Algorithm (1)
- Online Adaptation (1)
- Optimierungsproblem (1)
- Pandemie (1)
- Parallelisierung (1)
- Parameter Tuning (1)
- Performance (1)
- Plywise Analysis (1)
- Prognose (1)
- Referenzmodell (1)
- Regression (1)
- Reinforcement Learning (1)
- Robotics (1)
- SAP (1)
- SPOT (1)
- Sensor placement (1)
- Sensortechnik (1)
- Simulation-based Optimization (1)
- Simulationsmodell (1)
- Social Learning (1)
- Software (1)
- Spülen (1)
- Stacked Generalization (1)
- Stacking (1)
- Standardisierung (1)
- Statistics (1)
- Statistische Versuchsplanung (1)
- Surrogate Mod (1)
- Surrogate Model (1)
- Surrogate Optimization (1)
- Surrogate model (1)
- Surrogate model based optimization (1)
- Surrogate-Model-Based Optimization (1)
- Surrogates (1)
- Sustainability (1)
- System Identification (1)
- Systemidentifikation (1)
- Temporal Difference Learning (TDL) (1)
- Test function generator (1)
- Testgröße (1)
- Textiles (1)
- Time Series (1)
- Time-series (1)
- Trinkwasserversorgung (1)
- Univariate Data (1)
- Unsicherheit (1)
- Variable reduction (1)
- Varianzanalyse (1)
- Versuchsplanung (1)
- Verunreinigung (1)
- Vorverarbeitung (1)
- Wasseraufbereitungskosten (1)
- Wasserressourcen (1)
- Wasserverteilung (1)
- Water Distribution Systems (1)
- Water Quality Monitoring (1)
- Water distribution systems (1)
- Zeitreihe (1)
- Zeitreihenanalyse (1)
- Zylon Enstauber (1)
- economic assessment (1)
- ecosystem services (1)
- opportunity costs (1)
- water resources (1)
- watershed protection (1)
- Ökosystem-Dienstleistungen (1)
Institute
- Fakultät für Informatik und Ingenieurwissenschaften (F10) (26)
- Fakultät 10 / Institut für Informatik (11)
- Fakultät 09 / Institut für Rettungsingenieurwesen und Gefahrenabwehr (4)
- Fakultät 02 / Köln International School of Design (2)
- Fakultät 04 / Institut für Versicherungswesen (2)
- Fakultät 10 / Institut für Data Science, Engineering, and Analytics (2)
- Fakultät 08 / Institut für Fahrzeugtechnik (1)
- Institut für Technologie und Ressourcenmanagement in den Tropen und Subtropen (ITT) (1)
Cyclone separators are popular devices used to filter dust from the emitted flue gases. They are applied as pre-filters in many industrial processes including energy production and grain processing facilities.
Increasing computational power and the availability of 3D printers provide new tools for the combination of modeling and experimentation, which necessary for constructing efficient cyclones. Several simulation tools can be run in parallel, e.g., long running CFD simulations can be accompanied by experiments with 3D printers. Furthermore, results from analytical and data-driven models can be incorporated. There are fundamental differences between these modeling approaches: some models, e.g., analytical models, use domain knowledge, whereas data-driven models do not require any information about the underlying processes.
At the same time, data-driven models require input and output data, whereas analytical models do not. Combining results from models with different input-output structure is of great interest. This combination inspired the development of a new methodology. An optimization via multimodel simulation approach, which combines results from different models, is introduced.
Using cyclonic dust separators (cyclones) as a real-world simulation problem, the feasibility of this approach is demonstrated. Pros and cons of this approach are discussed and experiences from the experiments are presented.
Furthermore, technical problems, which are related to 3D-printing approaches, are discussed.
The use of surrogate models is a standard method to deal with complex, realworld
optimization problems. The first surrogate models were applied to continuous
optimization problems. In recent years, surrogate models gained importance
for discrete optimization problems. This article, which consists of three
parts, takes care of this development. The first part presents a survey of modelbased
methods, focusing on continuous optimization. It introduces a taxonomy,
which is useful as a guideline for selecting adequate model-based optimization
tools. The second part provides details for the case of discrete optimization
problems. Here, six strategies for dealing with discrete data structures are introduced.
A new approach for combining surrogate information via stacking
is proposed in the third part. The implementation of this approach will be
available in the open source R package SPOT2. The article concludes with a
discussion of recent developments and challenges in both application domains.
Data pre-processing is a key research topic in data mining because it plays a
crucial role in improving the accuracy of any data mining algorithm. In most
real world cases, a significant amount of the recorded data is found missing
due to most diverse errors. This loss of data is nearly always unavoidable.
Recovery of missing data plays a vital role in avoiding inaccurate data
mining decisions. Most multivariate imputation methods are not compatible
to univariate datasets and the traditional univariate imputation techniques
become highly biased as the missing data gap increases. With the current
technological advancements abundant data is being captured every second.
Hence, we intend to develop a new algorithm that enables maximum
utilization of the available big datasets for imputation. In this paper, we
present a Seasonal and Trend decomposition using Loess (STL) based
Seasonal Moving Window Algorithm, which is capable of handling patterns
with trend as well as cyclic characteristics. We show that the algorithm is
highly suitable for pre-processing of large datasets.
When researchers and practitioners in the field of
computational intelligence are confronted with real-world
problems, the question arises which method is the best to
apply. Nowadays, there are several, well established test
suites and well known artificial benchmark functions
available.
However, relevance and applicability of these methods to
real-world problems remains an open question in many
situations. Furthermore, the generalizability of these
methods cannot be taken for granted.
This paper describes a data-driven approach for the
generation of test instances, which is based on
real-world data. The test instance generation uses
data-preprocessing, feature extraction, modeling, and
parameterization. We apply this methodology on a classical
design of experiment real-world project and generate test
instances for benchmarking, e.g. design methods, surrogate
techniques, and optimization algorithms. While most
available results of methods applied on real-world
problems lack availability of the data for comparison,
our future goal is to create a toolbox covering multiple
data sets of real-world projects to provide a test
function generator to the research community.
This report presents a practical approach to stacked generalization in surrogate model based optimization. It exemplifies the integration of stacking methods into the surrogate model building process. First, a brief overview of the current state in surrogate model based opti- mization is presented. Stacked generalization is introduced as a promising ensemble surrogate modeling approach. Then two examples (the first is based on a real world application and the second on a set of artificial test functions) are presented. These examples clearly illustrate two properties of stacked generalization: (i) combining information from two poor performing models can result in a good performing model and (ii) even if the ensemble contains a good performing model, combining its information with information from poor performing models results in a relatively small performance decrease only.
When using machine learning techniques for learning a function approximation from given data it is often a difficult task to select the right modeling technique.
In many real-world settings is no preliminary knowledge about the objective function available. Then it might be beneficial if the algorithm could learn all models by itself and select the model that suits best to the problem.
This approach is known as automated model selection. In this work we propose a
generalization of this approach.
It combines the predictions of several into one more accurate ensemble surrogate model. This approach is studied in a fundamental way, by first evaluating minimalistic ensembles of only two surrogate models in detail and then proceeding to ensembles with three and more surrogate models.
The results show to what extent combinations of models can perform better than single surrogate models and provides insights into the scalability and robustness of the approach. The study focuses on multi-modal functions topologies, which are important in surrogate-assisted global optimization.
In this paper we present a comparison of different data driven modeling methods. The first instance of a data driven linear Bayesian model is compared with several linear regression models, a Kriging model and a genetic programming model.
The models are build on industrial data for the development of a robust gas sensor.
The data contain limited amount of samples and a high variance.
The mean square error of the models implemented in a test dataset is used as the comparison strategy.
The results indicate that standard linear regression approaches as well as Kriging and GP show good results,
whereas the Bayesian approach, despite the fact that it requires additional resources, does not lead to improved results.
This paper proposes an experimental methodology for on-line machine learning algorithms, i.e., for algorithms that work on data that are available in a sequential order.
It is demonstrated how established tools from experimental algorithmics (EA) can be applied in the on-line or streaming data setting.
The massive on-line analysis (MOA) framework is used to perform the experiments.
Benefits of a well-defined report structure are discussed.
The application of methods from the EA community to on-line or streaming data is referred to as experimental algorithmics for streaming data (EADS).
Land-use intensification and urbanisation processes are degrading ecosystem services in the Guapiaçu-Macacu watershed in the state of Rio de Janeiro, Brazil. Paying farmers to forgo agricultural production activities in order to restore natural watershed services might be a viable means of securing water resources over the long term for the approximately 2.5 million urban water users in the region. This study quantified the costs of changing current land-use patterns to enhance watershed services. These costs are compared to estimates of the avoided water treatment costs for the public potable water supply as a proxy of willingness-to-pay for watershed services. Farm-household data was used to estimate the opportunity costs of abandoning current land uses in order to allow natural vegetation succession; a process that is very likely to improve water quality in terms of reducing erosion and subsequently water turbidity. Opportunity cost estimates were extrapolated to the watershed scale based on land-use classifications and a vulnerability analysis for identifying priority areas for watershed management interventions. Water quality and treatment cost data from the primary local water treatment plant (principal water user in the study area) were analysed to assess the potential demand for watershed services. The conversion of agricultural land uses for the benefit of watershed service provision was found to entail high opportunity costs in the study area, which is near the city of Rio de Janeiro. Alternative, relatively low-cost practices that support watershed conservation do exist for the livestock production systems. Other options include: implementing soil conservation techniques, permanent protection of areas that are vulnerable to erosion, protecting and restoring riparian and headwater areas, and applying more sustainable agricultural practices. These measures have the potential to directly reduce the amount of sediment and nutrients reaching water bodies and, in turn, decrease the costs of treatment required for providing the potable water supply. Based on treatment costs, the state water utility company’s willingness-to-pay for watershed services alone will not be sufficient to compensate farmers for forgoing agricultural production activities in order to improve the provision of additional watershed services. The results suggest that the opportunity costs of land-cover changes at the scale needed to improve water quality will likely exceed the cost of additional investments in water treatment. Monetary incentives conditioned on specific adjustments to existing production systems could offer a complementary role for improving watershed services. The willingness-to-pay analysis, however, only focused on chemical treatment costs and one of a potentially wide range of ecosystem services provided by the natural vegetation in the Guapiaçu-Macacu watershed (water quality maintenance for potable water provision). Other ecosystem services provided by forest cover include carbon sequestration and storage, moderation of extreme weather events, regulation of water flows, landscape aesthetics, and biodiversity protection. Factoring these additional ecosystem services into the willingness-to-pay equation is likely to change the conclusions of the assessment in favour of additional conservation action, either through payments for ecosystem services (PES) or other policy instruments. This effort contributes to the growing body of related scientific literature by offering additional knowledge on how to combine spatially explicit economic and environmental information to provide valuable insights into the feasibility of implementing PES schemes at the scale of entire watersheds. This is relevant to helping inform decision-making processes with respect to the economic scope of incentive-based watershed management in the context of the Guapiaçu-Macacu watershed. Furthermore, the findings of this research can serve long-term watershed conservation initiatives and public policy in other watersheds of the Atlantic Forest biome by facilitating the targeting of conservation incentives for a cost-effective watershed management.
Sequential Parameter Optimization is a model-based optimization methodology, which includes several techniques for handling uncertainty. Simple approaches such as sharp- ening and more sophisticated approaches such as optimal computing budget allocation are available. For many real world engineering problems, the objective function can be evaluated at different levels of fidelity. For instance, a CFD simulation might provide a very time consuming but accurate way to estimate the quality of a solution.The same solution could be evaluated based on simplified mathematical equations, leading to a cheaper but less accurate estimate. Combining these different levels of fidelity in a model-based optimization process is referred to as multi-fidelity optimization. This chapter describes uncertainty-handling techniques for meta-model based search heuristics in combination with multi-fidelity optimization. Co-Kriging is one power- ful method to correlate multiple sets of data from different levels of fidelity. For the first time, Sequential Parameter Optimization with co-Kriging is applied to noisy test functions. This study will introduce these techniques and discuss how they can be applied to real-world examples.
We propose to apply typed Genetic Programming (GP) to the problem of finding surrogate-model ensembles for global optimization on compute-intensive target functions. In a model ensemble, base-models such as linear models, random forest models, or Kriging models, as well as pre- and post-processing methods, are combined. In theory, an optimal ensemble will join the strengths of its comprising base-models while avoiding their weaknesses, offering higher prediction accuracy and robustness. This study defines a grammar of model ensemble expressions and searches the set for optimal ensembles via GP. We performed an extensive experimental study based on 10 different objective functions and 2 sets of base-models. We arrive at promising results, as on unseen test data, our ensembles perform not significantly worse than the best base-model.
Computational intelligence methods have gained importance in several real-world domains such as process optimization, system identification, data mining, or statistical quality control. Tools are missing, which determine the applicability of computational intelligence methods in these application domains in an objective manner. Statistics provide methods for comparing algorithms on certain data sets. In the past, several test suites were presented and considered as state of the art. However, there are several drawbacks of these test suites, namely: (i) problem instances are somehow artificial and have no direct link to real-world settings; (ii) since there is a fixed number of test instances, algorithms can be fitted or tuned to this specific and very limited set of test functions; (iii) statistical tools for comparisons of several algorithms on several test problem instances are relatively complex and not easily to analyze. We propose amethodology to overcome these dificulties. It is based on standard ideas from statistics: analysis of variance and its extension to mixed models. This work combines essential ideas from two approaches: problem generation and statistical analysis of computer experiments.
This volume addresses the topics of Disaster Risk Reduction (DRR) and Climate Change Adaption (CCA) from the perspective of participants of the DAAD Alumni Summer School 2013. Contributions from 16 countries are gathered in this volume, thereby sharing specific knowledge on climate extremes, disasters, adaptation and prevention measures as well as current strategies in a range of different national contexts. The DAAD Alumni Summer School opened up a forum for integrative thinking and learning across cultures, disciplines and institutions. This volume is directly linked to the first volume of the series and presents a further outcome of the Summer School 2013. It invites the reader to look beyond common perspectives of DRR and CCA and relates climate change and natural disasters with interdisciplinary and bottom-up policy making. The outcomes presented in the two volumes are a starting point for further international & transdisciplinary knowledge exchange activities planned for the upcoming years.
Learning board games by self-play has a long tradition in computational intelligence for games. Based on Tesauro’s seminal success with TD-Gammon in 1994, many successful agents use temporal difference learning today. But in order to be successful with temporal difference learning on game tasks, often a careful selection of features and a large number of training games is necessary. Even for board games of moderate complexity like Connect-4, we found in previous work that a very rich initial feature set and several millions of game plays are required. In this work we investigate different approaches of online-adaptable learning rates like Incremental Delta Bar Delta (IDBD) or Temporal Coherence Learning (TCL) whether they have the potential to speed up learning for such a complex task. We propose a new variant of TCL with geometric step size changes. We compare those algorithms with several other state-of-the-art learning rate adaptation algorithms and perform a case study on the sensitivity with respect to their meta parameters. We show that in this set of learning algorithms those with geometric step size changes outperform those other algorithms with constant step size changes. Algorithms with nonlinear output functions are slightly better than linear ones. Algorithms with geometric step size changes learn faster by a factor of 4 as compared to previously published results on the task Connect-4.
In facing recent natural and man-made disasters Disaster Risk Reduction (DRR) and Climate Change Adaption (CCA) calls for integrative thinking and learning across cultures, disciplines and institutions. In times of increasing complexity, insecurity and uncertainty thinking outside the box becomes essential. This first volume of “Integrative Risk and Security Research” presents related research contributions developed in the context of the 2013 DAAD Alumni Summer School "Coping with Disasters and Climate Extremes - Challenges & Cooperation Potential". It invites the reader to look beyond common perspectives of DRR and CCA and relates climate change and natural disasters with interdisciplinary and bottom-up policy making.
An essential task for operation and planning of biogas plants is the optimization of substrate feed mixtures. Optimizing the monetary gain requires the determination of the exact amounts of maize, manure, grass silage, and other substrates. Accurate simulation models are mandatory for this optimization, because the underlying chemical processes are very slow. The simulation models themselves may be time-consuming to evaluate, hence we show how to use surrogate-model-based approaches to optimize biogas plants efficiently. In detail, a Kriging surrogate is employed. To improve model quality of this surrogate, we integrate cheaply available data into the optimization process. Doing so, Multi-fidelity modeling methods like Co-Kriging are employed. Furthermore, a two-layered modeling approach is employed to avoid deterioration of model quality due to discontinuities in the search space. At the same time, the cheaply available data is shown to be very useful for initialization of the employed optimization algorithms. Overall, we show how biogas plants can be efficiently modeled using data-driven methods, avoiding discontinuities as well as including cheaply available data. The application of the derived surrogate models to an optimization process is shown to be very difficult, yet successful for a lower problem dimension.
This paper introduces UniFIeD, a new data preprocessing method for time series. UniFIeD can cope with large intervals of missing data. A scalable test function generator, which allows the simulation of time series with different gap sizes, is presented additionally. An experimental study demonstrates that (i) UniFIeD shows a significant better performance than simple imputation methods and (ii) UniFIeD is able to handle situations, where advanced imputation methods fail. The results are independent from the underlying error measurements.
Cyclone Dust Separators are devices often used to filter solid particles from flue gas. Such cyclones are supposed to filter as much solid particles from the carrying gas as possible. At the same time, they should only introduce a minimal pressure loss to the system. Hence, collection efficiency has to be maximized and pressure loss minimized. Both the collection efficiency and pressure loss are heavily influenced by the cyclones geometry. In this paper, we optimize seven geometrical parameters of an analytical cyclone model. Furthermore, noise variables are introduced to the model, representing the non-deterministic structure of the real-world problem. This is used to investigate robustness and sensitivity of solutions. Both the deterministic as well as the stochastic model are optimized with an SMS-EMOA. The SMS-EMOA is compared to a single objective optimization algorithm. For the harder, stochastic optimization problem, a surrogate-model-supported SMS-EMOA is compared against the model-free SMS-EMOA. The model supported approach yields better solutions with the same run-time budget.
RGP is genetic programming system based on, as well as fully integrated into, the R environment. The system implements classical tree-based genetic programming as well as other variants including, for example, strongly typed genetic programming and Pareto genetic programming. It strives for high modularity through a consistent architecture that allows the customization and replacement of every algorithm component, while maintaining accessibility for new users by adhering to the "convention over configuration" principle.
Multi-criteria optimization has gained increasing attention during the last decades. This article exemplifies multi-criteria features, which are implemented in the statistical software package SPOT. It describes related software packages such as mco and emoa and gives a comprehensive introduction to simple multi criteria optimization tasks. Several hands-on examples are used for illustration. The article is well-suited as a starting point for performing multi-criteria optimization tasks with SPOT.