Deep learning-based computer vision anomaly detection algorithms are widely utilized in various fields. Especially in the manufacturing industry, the difficulty in collecting abnormal data compared to normal data, and the challenge of defining all potential abnormalities in advance, have led to an increasing demand for unsupervised learning methods that rely on normal data. In this study, we conducted a comparative analysis of deep learning-based unsupervised learning algorithms that define and detect abnormalities that can occur when transparent contact lenses are immersed in liquid solution. We validated and applied the unsupervised learning algorithms used in this study to the existing anomaly detection benchmark dataset, MvTecAD. The existing anomaly detection benchmark dataset primarily consists of solid objects, whereas in our study, we compared unsupervised learning-based algorithms in experiments judging the shape and presence of lenses submerged in liquid. Among the algorithms analyzed, EfficientAD showed an AUROC and F1-score of 0.97 in image-level tests. However, the F1-score decreased to 0.18 in pixel-level tests, making it challenging to determine the locations where abnormalities occurred. Despite this, EfficientAD demonstrated excellent performance in image-level tests classifying normal and abnormal instances, suggesting that with the collection and training of large-scale data in real industrial settings, it is expected to exhibit even better performance.
With the recent surge in YouTube usage, there has been a proliferation of user-generated videos where individuals evaluate cosmetics. Consequently, many companies are increasingly utilizing evaluation videos for their product marketing and market research. However, a notable drawback is the manual classification of these product review videos incurring significant costs and time. Therefore, this paper proposes a deep learning-based cosmetics search algorithm to automate this task. The algorithm consists of two networks: One for detecting candidates in images using shape features such as circles, rectangles, etc and Another for filtering and categorizing these candidates. The reason for choosing a Two-Stage architecture over One-Stage is that, in videos containing background scenes, it is more robust to first detect cosmetic candidates before classifying them as specific objects. Although Two-Stage structures are generally known to outperform One-Stage structures in terms of model architecture, this study opts for Two-Stage to address issues related to the acquisition of training and validation data that arise when using One-Stage. Acquiring data for the algorithm that detects cosmetic candidates based on shape and the algorithm that classifies candidates into specific objects is cost-effective, ensuring the overall robustness of the algorithm.
Nowadays, artificial intelligence model approaches such as machine and deep learning have been widely used to predict variations of water quality in various freshwater bodies. In particular, many researchers have tried to predict the occurrence of cyanobacterial blooms in inland water, which pose a threat to human health and aquatic ecosystems. Therefore, the objective of this study were to: 1) review studies on the application of machine learning models for predicting the occurrence of cyanobacterial blooms and its metabolites and 2) prospect for future study on the prediction of cyanobacteria by machine learning models including deep learning. In this study, a systematic literature search and review were conducted using SCOPUS, which is Elsevier’s abstract and citation database. The key results showed that deep learning models were usually used to predict cyanobacterial cells, while machine learning models focused on predicting cyanobacterial metabolites such as concentrations of microcystin, geosmin, and 2-methylisoborneol (2-MIB) in reservoirs. There was a distinct difference in the use of input variables to predict cyanobacterial cells and metabolites. The application of deep learning models through the construction of big data may be encouraged to build accurate models to predict cyanobacterial metabolites.
The color image of the brand comes first and is an important visual element that leads consumers to the consumption of the product. To express more effectively what the brand wants to convey through design, the printing market is striving to print accurate colors that match the intention. In ‘offset printing’ mainly used in printing, colors are often printed in CMYK (Cyan, Magenta, Yellow, Key) colors. However, it is possible to print more accurate colors by making ink of the desired color instead of dotting CMYK colors. The resulting ink is called ‘spot color’ ink. Spot color ink is manufactured by repeating the process of mixing the existing inks. In this repetition of trial and error, the manufacturing cost of ink increases, resulting in economic loss, and environmental pollution is caused by wasted inks. In this study, a deep learning algorithm to predict printed spot colors was designed to solve this problem. The algorithm uses a single DNN (Deep Neural Network) model to predict printed spot colors based on the information of the paper and the proportions of inks to mix. More than 8,000 spot color ink data were used for learning, and all color was quantified by dividing the visible light wavelength range into 31 sections and the reflectance for each section. The proposed algorithm predicted more than 80% of spot color inks as very similar colors. The average value of the calculated difference between the actual color and the predicted color through ‘Delta E’ provided by CIE is 5.29. It is known that when Delta E is less than 10, it is difficult to distinguish the difference in printed color with the naked eye. The algorithm of this study has a more accurate prediction ability than previous studies, and it can be added flexibly even when new inks are added. This can be usefully used in real industrial sites, and it will reduce the attempts of the operator by checking the color of ink in a virtual environment. This will reduce the manufacturing cost of spot color inks and lead to improved working conditions for workers. In addition, it is expected to contribute to solving the environmental pollution problem by reducing unnecessarily wasted ink.
본 연구는 화재진압 및 피난활동을 지원하는 딥러닝 기반의 알고리즘 개발에 관한 기초 연구로 선박 화재 시 연기감지기가 작동하기 전에 검출된 연기 데이터를 분석 및 활용하여 원격지까지 연기가 확산 되기 전에 연기 확산거리를 예측하는 것이 목적이다. 다음과 같은 절차에 따라 제안 알고리즘을 검토하였다. 첫 번째 단계로, 딥러닝 기반 객체 검출 알고리즘인 YOLO(You Only Look Once)모델에 화재시뮬레이션을 통하여 얻은 연기 영상을 적용하여 학습을 진행하였다. 학습된 YOLO모델의 mAP(mean Average Precision)은 98.71%로 측정되었으며, 9 FPS(Frames Per Second)의 처리 속도로 연기를 검출하였다. 두 번째 단계로 YOLO로부터 연기 형상이 추출된 경계 상자의 좌표값을 통해 연기 확산거리를 추정하였으며 이를 시계열 예측 알고리즘인 LSTM(Long Short-Term Memory)에 적용하여 학습을 진행하였다. 그 결과, 화재시뮬레이션으로부터 얻은 Fast 화재의 연기영상에서 경계 상자의 좌표값으로부터 추정한 화재발생~30초까지의 연기 확산거리 데이터를 LSTM 학습모델에 입력하여 31초~90초까지의 연기 확산거리 데이터를 예측하였다. 그리고 추정한 연기 확산거리와 예측한 연기 확산거리의 평균제곱근 오차는 2.74로 나타났다.
딥러닝 모델은 주어진 학습용 데이터에서 탐지하고자 하는 물체의 특징을 추출하기 때문에, 딥러닝 모델 학습을 위한 학습용 데이터 구축은 매우 중요하다. 본 연구에서는 균열을 탐지하는 딥러닝 모델의 성능을 향상시키기 위해, 실제 콘크리트 구조물이나 아스팔트 도로 표면에서 자주 발견될 수 있는 나뭇가지, 거미줄, 전선 등을 학습 데이터에 자동으로 포함시키고, negative 영역으로 분류하는 알고리즘을 개발하였다. 제안된 알고리즘을 사용하여 학습된 딥러닝 모델을 실제 도로 표면에 발생한 균열 탐지에 적용하여 실제 균열 탐지에 사용될 수 있음을 보였다.
Carbonation of reinforced concrete is a major factor in the deterioration of reinforced concrete, and prediction of the resistance to carbonation is important in determining the durability life of reinforced concrete structures. In this study, basic research on the prediction of carbonation penetration depth of concrete using Deep Learning algorithm among artificial neural network theory was carried out. The data used in the experiment were analyzed by deep running algorithm by setting W/B, cement and blast furnace slag, fly ash content, relative humidity of the carbonated laboratory, temperature, CO2 concentration, Deep learning algorithms were used to study 60,000 times, and the analysis of the number of hidden layers was compared.
Reinforcement learning has been applied to various problems in robotics. However, it was still hard to train complex robotic manipulation tasks since there is a few models which can be applicable to general tasks. Such general models require a lot of training episodes. In these reasons, deep neural networks which have shown to be good function approximators have not been actively used for robot manipulation task. Recently, some of these challenges are solved by a set of methods, such as Guided Policy Search, which guide or limit search directions while training of a deep neural network based policy model. These frameworks are already applied to a humanoid robot, PR2. However, in robotics, it is not trivial to adjust existing algorithms designed for one robot to another robot. In this paper, we present our implementation of Guided Policy Search to the robotic arms of the Baxter Research Robot. To meet the goals and needs of the project, we build on an existing implementation of Baxter Agent class for the Guided Policy Search algorithm code using the built-in Python interface. This work is expected to play an important role in popularizing robot manipulation reinforcement learning methods on cost-effective robot platforms.
As the importance of maintenance of reinforced concrete structures spreads, interest in the durability of structures is increasing. Among them, carbonation of concrete is one of the main deterioration factors of reinforced concrete structures. For quantitative evaluation of carbonation, many researchers are predicting carbonation considering water-cement ratio and environmental requirements. In this study, we studied the parameters based on the concrete made of ordinary Portland cement in the existing experimental data. The depth of carbonation deduced from the learning is applied to the carbonation by applying the deep learning.