• Title/Summary/Keyword: DNN (Deep Neural Network)

Search Result 254, Processing Time 0.032 seconds

Watermarking for Digital Hologram by a Deep Neural Network and its Training Considering the Hologram Data Characteristics (딥 뉴럴 네트워크에 의한 디지털 홀로그램의 워터마킹 및 홀로그램 데이터 특성을 고려한 학습)

  • Lee, Juwon;Lee, Jae-Eun;Seo, Young-Ho;Kim, Dong-Wook
    • Journal of Broadcast Engineering
    • /
    • v.26 no.3
    • /
    • pp.296-307
    • /
    • 2021
  • A digital hologram (DH) is an ultra-high value-added video content that includes 3D information in 2D data. Therefore, its intellectual property rights must be protected for its distribution. For this, this paper proposes a watermarking method of DH using a deep neural network. This method is a watermark (WM) invisibility, attack robustness, and blind watermarking method that does not use host information in WM extraction. The proposed network consists of four sub-networks: pre-processing for each of the host and WM, WM embedding watermark, and WM extracting watermark. This network expand the WM data to the host instead of shrinking host data to WM and concatenate it to the host to insert the WM by considering the characteristics of a DH having a strong high frequency component. In addition, in the training of this network, the difference in performance according to the data distribution property of DH is identified, and a method of selecting a training data set with the best performance in all types of DH is presented. The proposed method is tested for various types and strengths of attacks to show its performance. It also shows that this method has high practicality as it operates independently of the resolution of the host DH and WM data.

Initialization by using truncated distributions in artificial neural network (절단된 분포를 이용한 인공신경망에서의 초기값 설정방법)

  • Kim, MinJong;Cho, Sungchul;Jeong, Hyerin;Lee, YungSeop;Lim, Changwon
    • The Korean Journal of Applied Statistics
    • /
    • v.32 no.5
    • /
    • pp.693-702
    • /
    • 2019
  • Deep learning has gained popularity for the classification and prediction task. Neural network layers become deeper as more data becomes available. Saturation is the phenomenon that the gradient of an activation function gets closer to 0 and can happen when the value of weight is too big. Increased importance has been placed on the issue of saturation which limits the ability of weight to learn. To resolve this problem, Glorot and Bengio (Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, 249-256, 2010) claimed that efficient neural network training is possible when data flows variously between layers. They argued that variance over the output of each layer and variance over input of each layer are equal. They proposed a method of initialization that the variance of the output of each layer and the variance of the input should be the same. In this paper, we propose a new method of establishing initialization by adopting truncated normal distribution and truncated cauchy distribution. We decide where to truncate the distribution while adapting the initialization method by Glorot and Bengio (2010). Variances are made over output and input equal that are then accomplished by setting variances equal to the variance of truncated distribution. It manipulates the distribution so that the initial values of weights would not grow so large and with values that simultaneously get close to zero. To compare the performance of our proposed method with existing methods, we conducted experiments on MNIST and CIFAR-10 data using DNN and CNN. Our proposed method outperformed existing methods in terms of accuracy.

Arc Detection using Logistic Regression (로지스틱 회기를 이용한 아크 검출)

  • Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.26 no.5
    • /
    • pp.566-574
    • /
    • 2021
  • The arc is one of factors causing electrical fires. Over past decades, various researches have been carried out to detect arc occurrences. Even though frequency analysis, wavelet and statistical features have been used, arc detection performance is degraded due to diverse arc waveforms. On the contray, Deep neural network (DNN) direcly utilizes raw data without feature extraction, based on end-to-end learning. However, a disadvantage of the DNN is processing complexity, posing the difficulty of being migrated into a termnial device. To solve this, this paper proposes an arc detection method using a logistic regression that is one of simple machine learning methods.

Assessment of wall convergence for tunnels using machine learning techniques

  • Mahmoodzadeh, Arsalan;Nejati, Hamid Reza;Mohammadi, Mokhtar;Ibrahim, Hawkar Hashim;Mohammed, Adil Hussein;Rashidi, Shima
    • Geomechanics and Engineering
    • /
    • v.31 no.3
    • /
    • pp.265-279
    • /
    • 2022
  • Tunnel convergence prediction is essential for the safe construction and design of tunnels. This study proposes five machine learning models of deep neural network (DNN), K-nearest neighbors (KNN), Gaussian process regression (GPR), support vector regression (SVR), and decision trees (DT) to predict the convergence phenomenon during or shortly after the excavation of tunnels. In this respect, a database including 650 datasets (440 for training, 110 for validation, and 100 for test) was gathered from the previously constructed tunnels. In the database, 12 effective parameters on the tunnel convergence and a target of tunnel wall convergence were considered. Both 5-fold and hold-out cross validation methods were used to analyze the predicted outcomes in the ML models. Finally, the DNN method was proposed as the most robust model. Also, to assess each parameter's contribution to the prediction problem, the backward selection method was used. The results showed that the highest and lowest impact parameters for tunnel convergence are tunnel depth and tunnel width, respectively.

Speaker Adaptation Using i-Vector Based Clustering

  • Kim, Minsoo;Jang, Gil-Jin;Kim, Ji-Hwan;Lee, Minho
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.7
    • /
    • pp.2785-2799
    • /
    • 2020
  • We propose a novel speaker adaptation method using acoustic model clustering. The similarity of different speakers is defined by the cosine distance between their i-vectors (intermediate vectors), and various efficient clustering algorithms are applied to obtain a number of speaker subsets with different characteristics. The speaker-independent model is then retrained with the training data of the individual speaker subsets grouped by the clustering results, and an unknown speech is recognized by the retrained model of the closest cluster. The proposed method is applied to a large-scale speech recognition system implemented by a hybrid hidden Markov model and deep neural network framework. An experiment was conducted to evaluate the word error rates using Resource Management database. When the proposed speaker adaptation method using i-vector based clustering was applied, the performance, as compared to that of the conventional speaker-independent speech recognition model, was improved relatively by as much as 12.2% for the conventional fully neural network, and by as much as 10.5% for the bidirectional long short-term memory.

Deep learning-based anomaly detection in acceleration data of long-span cable-stayed bridges

  • Seungjun Lee;Jaebeom Lee;Minsun Kim;Sangmok Lee;Young-Joo Lee
    • Smart Structures and Systems
    • /
    • v.33 no.2
    • /
    • pp.93-103
    • /
    • 2024
  • Despite the rapid development of sensors, structural health monitoring (SHM) still faces challenges in monitoring due to the degradation of devices and harsh environmental loads. These challenges can lead to measurement errors, missing data, or outliers, which can affect the accuracy and reliability of SHM systems. To address this problem, this study proposes a classification method that detects anomaly patterns in sensor data. The proposed classification method involves several steps. First, data scaling is conducted to adjust the scale of the raw data, which may have different magnitudes and ranges. This step ensures that the data is on the same scale, facilitating the comparison of data across different sensors. Next, informative features in the time and frequency domains are extracted and used as input for a deep neural network model. The model can effectively detect the most probable anomaly pattern, allowing for the timely identification of potential issues. To demonstrate the effectiveness of the proposed method, it was applied to actual data obtained from a long-span cable-stayed bridge in China. The results of the study have successfully verified the proposed method's applicability to practical SHM systems for civil infrastructures. The method has the potential to significantly enhance the safety and reliability of civil infrastructures by detecting potential issues and anomalies at an early stage.

Efficient AIOT Information Link Processing in Cloud Edge Environment Using Blockchain-Based Time Series Information (블록체인 기반의 시계열 정보를 이용한 클라우드 엣지 환경의 효율적인 AIoT 정보 연계 처리 기법)

  • Jeong, Yoon-Su
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.3
    • /
    • pp.9-15
    • /
    • 2021
  • With the recent development of 5G and artificial intelligence technologies, it is interested in AIOT technology to collect, process, and analyze information in cloud edge environments. AIIoT technology is being applied to various smart environments, but research is needed to perform fast response processing through accurate analysis of collected information. In this paper, we propose a technique to minimize bandwidth and processing time by blocking the connection processing between AIOT information through fast processing and accurate analysis/forecasting of information collected in the smart environment. The proposed technique generates seeds for data indexes on AIOT devices by multipointing information collected by blockchain, and blocks them along with collection information to deliver them to the data center. At this time, we deploy Deep Neural Network (DNN) models between cloud and AIOT devices to reduce network overhead. Furthermore, server/data centers have improved the accuracy of inaccurate AIIoT information through the analysis and predicted results delivered to minimize latency. Furthermore, the proposed technique minimizes data latency by allowing it to be partitioned into a layered multilayer network because it groups it into blockchain by applying weights to AIOT information.

Efficient Tire Wear and Defect Detection Algorithm Based on Deep Learning (심층학습 기법을 활용한 효과적인 타이어 마모도 분류 및 손상 부위 검출 알고리즘)

  • Park, Hye-Jin;Lee, Young-Woon;Kim, Byung-Gyu
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.8
    • /
    • pp.1026-1034
    • /
    • 2021
  • Tire wear and defect are important factors for safe driving condition. These defects are generally inspected by some specialized experts or very expensive equipments such as stereo depth camera and depth gauge. In this paper, we propose tire safety vision inspector based on deep neural network (DNN). The status of tire wear is categorized into three: 'safety', 'warning', and 'danger' based on depth of tire tread. We propose an attention mechanism for emphasizing the feature of tread area. The attention-based feature is concatenated to output feature maps of the last convolution layer of ResNet-101 to extract more robust feature. Through experiments, the proposed tire wear classification model improves 1.8% of accuracy compared to the existing ResNet-101 model. For detecting the tire defections, the developed tire defect detection model shows up-to 91% of accuracy using the Mask R-CNN model. From these results, we can see that the suggested models are useful for checking on the safety condition of working tire in real environment.

Deep-Learning-Based Water Shield Automation System by Predicting River Overflow and Vehicle Flooding Possibility (하천 범람 및 차량 침수 가능성 예측을 통한 딥러닝 기반 차수막 자동화 시스템)

  • Seung-Jae Ham;Min-Su Kang;Seong-Woo Jeong;Joonhyuk Yoo
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.3
    • /
    • pp.133-139
    • /
    • 2023
  • This paper proposes a two-stage Water Shield Automation System (WSAS) to predict the possibility of river overflow and vehicle flooding due to sudden rainfall. The WSAS uses a two-stage Deep Neural Network (DNN) model. First, a river overflow prediction module is designed with LSTM to decide whether the river is flooded by predicting the river's water level rise. Second, a vehicle flooding prediction module predicts flooding of underground parking lots by detecting flooded tires with YOLOv5 from CCTV images. Finally, the WSAS automatically installs the water barrier whenever the river overflow and vehicle flooding events happen in the underground parking lots. The only constraint to implementing is that collecting training data for flooded vehicle tires is challenging. This paper exploits the Image C&S data augmentation technique to synthesize flooded tire images. Experimental results validate the superiority of WSAS by showing that the river overflow prediction module can reduce RMSE by three times compared with the previous method, and the vehicle flooding detection module can increase mAP by 20% compared with the naive detection method, respectively.

Predicting Program Code Changes Using a CNN Model (CNN 모델을 이용한 프로그램 코드 변경 예측)

  • Kim, Dong Kwan
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.9
    • /
    • pp.11-19
    • /
    • 2021
  • A software system is required to change during its life cycle due to various requirements such as adding functionalities, fixing bugs, and adjusting to new computing environments. Such program code modification should be considered as carefully as a new system development becase unexpected software errors could be introduced. In addition, when reusing open source programs, we can expect higher quality software if code changes of the open source program are predicted in advance. This paper proposes a Convolutional Neural Network (CNN)-based deep learning model to predict source code changes. In this paper, the prediction of code changes is considered as a kind of a binary classification problem in deep learning and labeled datasets are used for supervised learning. Java projects and code change logs are collected from GitHub for training and testing datasets. Software metrics are computed from the collected Java source code and they are used as input data for the proposed model to detect code changes. The performance of the proposed model has been measured by using evaluation metrics such as precision, recall, F1-score, and accuracy. The experimental results show the proposed CNN model has achieved 95% in terms of F1-Score and outperformed the multilayer percept-based DNN model whose F1-Score is 92%.