Refine
Document Type
- Working Paper (13) (remove)
Has Fulltext
- yes (13)
Keywords
- Optimization (4)
- Benchmarking (3)
- Modeling (3)
- Optimierung (3)
- Künstliche Intelligenz (2)
- Parallelization (2)
- Artificial intelligence (1)
- Bayesian Optimization (1)
- Big Data (1)
- Big data platform (1)
- Cognition (1)
- Conditional inference tree (1)
- Decision tree (1)
- Design of Experiments (1)
- Discrete Optimization (1)
- Electrostatic Precipitator (1)
- Evolutionary Computation (1)
- Expensive Optimization (1)
- Faserverbundwerkstoffe (1)
- Flowcurve (1)
- Funktionstest (1)
- Health condition monitoring (1)
- Hot rolling (1)
- Imputation (1)
- Industrie 4.0 (1)
- Industry 4.0 (1)
- Knowledge extraction (1)
- Kognition (1)
- Kriging (1)
- Machine learning (1)
- Meta-model (1)
- Metal (1)
- Metamodels (1)
- Modelierung (1)
- Neural and Evolutionary Computing (1)
- Parallelisierung (1)
- Performance (1)
- Robotics (1)
- Signalanalyse (1)
- Simulation (1)
- Structural Health Monitoring (1)
- Surrogate (1)
- Surrogate model based optimization (1)
- Surrogates (1)
- Test Function (1)
- Test function generator (1)
- Time Series (1)
- Univariate Data (1)
- Vorgehensmodell (1)
Data pre-processing is a key research topic in data mining because it plays a
crucial role in improving the accuracy of any data mining algorithm. In most
real world cases, a significant amount of the recorded data is found missing
due to most diverse errors. This loss of data is nearly always unavoidable.
Recovery of missing data plays a vital role in avoiding inaccurate data
mining decisions. Most multivariate imputation methods are not compatible
to univariate datasets and the traditional univariate imputation techniques
become highly biased as the missing data gap increases. With the current
technological advancements abundant data is being captured every second.
Hence, we intend to develop a new algorithm that enables maximum
utilization of the available big datasets for imputation. In this paper, we
present a Seasonal and Trend decomposition using Loess (STL) based
Seasonal Moving Window Algorithm, which is capable of handling patterns
with trend as well as cyclic characteristics. We show that the algorithm is
highly suitable for pre-processing of large datasets.
The use of surrogate models is a standard method to deal with complex, realworld
optimization problems. The first surrogate models were applied to continuous
optimization problems. In recent years, surrogate models gained importance
for discrete optimization problems. This article, which consists of three
parts, takes care of this development. The first part presents a survey of modelbased
methods, focusing on continuous optimization. It introduces a taxonomy,
which is useful as a guideline for selecting adequate model-based optimization
tools. The second part provides details for the case of discrete optimization
problems. Here, six strategies for dealing with discrete data structures are introduced.
A new approach for combining surrogate information via stacking
is proposed in the third part. The implementation of this approach will be
available in the open source R package SPOT2. The article concludes with a
discussion of recent developments and challenges in both application domains.
To maximize the throughput of a hot rolling mill,
the number of passes has to be reduced. This can be achieved by maximizing the thickness reduction in each pass. For this purpose, exact predictions of roll force and torque are required. Hence, the predictive models that describe the physical behavior of the product have to be accurate and cover a wide range of different materials.
Due to market requirements a lot of new materials are tested and rolled. If these materials are chosen to be rolled more often, a suitable flow curve has to be established. It is not reasonable to determine those flow curves in laboratory, because of costs and time. A strong demand for quick parameter determination and the optimization of flow curve parameter with minimum costs is the logical consequence. Therefore parameter estimation and the optimization with real data, which were collected during previous runs, is a promising idea. Producers benefit from this data-driven approach and receive a huge gain in flexibility when rolling new
materials, optimizing current production, and increasing quality. This concept would also allow to optimize flow curve parameters, which have already been treated by standard methods. In this article, a new data-driven approach for predicting the physical behavior of the product and setting important parameters is presented.
We demonstrate how the prediction quality of the roll force and roll torque can be optimized sustainably. This offers the opportunity to continuously increase the workload in each pass to the theoretical maximum while product quality and process stability can also be improved.