The high-level radioactive waste repository must ensure its performance for a long period of time enough to sufficiently reduce the potential risk of the waste, and for this purpose, multibarrier systems consisting of engineered and natural barrier systems are applied. If waste nuclides leak, the dominating mechanisms facilitating their movement toward human habitats include advection, dispersion and diffusion along groundwater flows. Therefore, it is of great importance to accurately assess the hydrogeological and geochemical characteristics of the host rock because it acts as a flow medium. Normally, borehole investigations were used to evaluate the characteristics and the use of multi-packer system is more efficient and economical compared to standpipes, as it divides a single borehole into multiple sections by installing multiple packers. For effective analyses and groundwater sampling, the entire system is designed by preselecting sections where groundwater flow is clearly remarkable. The selection is based on the analyses of various borehole and rock core logging data. Generally, sections with a high frequency of joints and evident water flow are chosen. Analyzing the logging data, which can be considered continuous, gives several local points where the results exhibit significant local changes. These clear deviations can be considered outliers within the data set, and machine learning algorithms have been frequently applied to classify them. The algorithms applied in this study include DBSCAN (density based spatial clustering of application with noise), OCSVM (one class support vector method), KNN (K nearest neighbor), and isolation forest, of which are widely used in many applications. This paper aims to evaluate the applicability of the aforementioned four algorithms to the design of multi-packer system. The data used for this evaluation were obtained from DB-2 borehole logging data, which is a deep borehole locates near KURT.
This paper proposes an outlier detection model based on machine learning that can diagnose the presence or absence of major engine parts through unsupervised learning analysis of main engine big data of a ship. Engine big data of the ship was collected for more than seven months, and expert knowledge and correlation analysis were performed to select features that are closely related to the operation of the main engine. For unsupervised learning analysis, ensemble model wherein many predictive models are strategically combined to increase the model performance, is used for anomaly detection. As a result, the proposed model successfully detected the anomalous engine status from the normal status. To validate our approach, clustering analysis was conducted to find out the different patterns of anomalies the anomalous point. By examining distribution of each cluster, we could successfully find the patterns of anomalies.
High-performance concrete (HPC) is a new terminology used in concrete construction industry. Several studies have shown that concrete strength development is determined not only by the water-to-cement ratio but also influenced by the content of other conc
High-performance concrete(HPC) is a new terminology used in concrete construction industry. Several studies have shown that concrete strength development is determined not only by the water-to-cement ratio but also influenced by the content of other concrete ingredients. HPC is a highly complex material, which makes modeling its behavior a very difficult task. This paper aimed at demonstrating the possibilities of adapting artificial neural network (ANN) to predict the comprresive strength of HPC. Mahalanobis Distance(MD) outlier detection method used for the purpose increase prediction ability of ANN. The detailed procedure of calculating Mahalanobis Distance (MD) is described. The effects of outlier compared with before and after artificial neural network training. MD outlier detection method successfully removed existence of outlier and improved the neural network training and prediction perfomance.
This study investigates the problem of outlier detection based on discrete wavelet transform in the context of time series data where the identification and treatment of outliers constitute an important component. An outlier is defined as a data point that deviates so much from the rest of observations within a data sample. In this work we focus on the application of the traditional method suggested by Tukey (1977) for detecting outliers in the closed price series of the Saudi Arabia stock market (Tadawul) between Oct. 2011 and Dec. 2019. The method is applied to the details obtained from the MODWT (Maximal-Overlap Discrete Wavelet Transform) of the original series. The result show that the suggested methodology was successful in detecting all of the outliers in the series. The findings of this study suggest that we can model and forecast the volatility of returns from the reconstructed series without outliers using GARCH models. The estimated GARCH volatility model was compared to other asymmetric GARCH models using standard forecast error metrics. It is found that the performance of the standard GARCH model were as good as that of the gjrGARCH model over the out-of-sample forecasts for returns among other GARCH specifications.