• Title/Summary/Keyword: deep Learning

Search Result 5,795, Processing Time 0.033 seconds

Image Reconstruction Method for Photonic Integrated Interferometric Imaging Based on Deep Learning

  • Qianchen Xu;Weijie Chang;Feng Huang;Wang Zhang
    • Current Optics and Photonics
    • /
    • v.8 no.4
    • /
    • pp.391-398
    • /
    • 2024
  • An image reconstruction algorithm is vital for the image quality of a photonic integrated interferometric imaging (PIII) system. However, image reconstruction algorithms have limitations that always lead to degraded image reconstruction. In this paper, a novel image reconstruction algorithm based on deep learning is proposed. Firstly, the principle of optical signal transmission through the PIII system is investigated. A dataset suitable for image reconstruction of the PIII system is constructed. Key aspects such as model and loss functions are compared and constructed to solve the problem of image blurring and noise influence. By comparing it with other algorithms, the proposed algorithm is verified to have good reconstruction results not only qualitatively but also quantitatively.

Use of artificial intelligence in the management of T1 colorectal cancer: a new tool in the arsenal or is deep learning out of its depth?

  • James Weiquan Li;Lai Mun Wang;Katsuro Ichimasa;Kenneth Weicong Lin;James Chi-Yong Ngu;Tiing Leong Ang
    • Clinical Endoscopy
    • /
    • v.57 no.1
    • /
    • pp.24-35
    • /
    • 2024
  • The field of artificial intelligence is rapidly evolving, and there has been an interest in its use to predict the risk of lymph node metastasis in T1 colorectal cancer. Accurately predicting lymph node invasion may result in fewer patients undergoing unnecessary surgeries; conversely, inadequate assessments will result in suboptimal oncological outcomes. This narrative review aims to summarize the current literature on deep learning for predicting the probability of lymph node metastasis in T1 colorectal cancer, highlighting areas of potential application and barriers that may limit its generalizability and clinical utility.

Multi-label Lane Detection Algorithm for Autonomous Vehicle Using Deep Learning (자율주행 차량을 위한 멀티 레이블 차선 검출 딥러닝 알고리즘)

  • Chae Song Park;Kyong Su Yi
    • Journal of Auto-vehicle Safety Association
    • /
    • v.16 no.1
    • /
    • pp.29-34
    • /
    • 2024
  • This paper presents a multi-label lane detection method for autonomous vehicles based on deep learning. The proposed algorithm can detect two types of lanes: center lane and normal lane. The algorithm uses a convolution neural network with an encoder-decoder architecture to extract features from input images and produce a multi-label heatmap for predicting lane's label. This architecture has the potential to detect more diverse types of lanes in that it can add the number of labels by extending the heatmap's dimension. The proposed algorithm was tested on an OpenLane dataset and achieved 85 Frames Per Second (FPS) in end to-end inference time. The results demonstrate the usability and computational efficiency of the proposed algorithm for the lane detection in autonomous vehicles.

Methodology of Applying Randomness for Boosting Image Classification Performance (이미지 분류 성능 향상을 위한 무작위성 적용 방법론)

  • Juyong Park;Yuri Jeon;Miyeong Kim;Jeongmin Lee;Yoonsuk Hyun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.5
    • /
    • pp.251-257
    • /
    • 2024
  • Securing various types of training data in image Classification is important for improving performance. However, increasing the amount of original data is cost-limited, so data diversity can be secured by transforming images through data augmentation. Recently, a new deep learning approach using Generative AI models like GAN or Diffusion Based models has emerged in the Data Augmentation task, and reinforcement learning-based methods such as AutoAugment and Deep AutoAugment using existing basic Augmentation techniques are also showing good performance. However, this method has the disadvantage of having a complicated optimization procedure and high computational cost. This paper conducted various experiments with existing methods Cutmix, Mixup, RandAugment. By combining these techniques appropriately, we obtained higher performance than existing method without much effort. Additionally, our ablation experiment shows that additional hyper-parameter adjustments are needed for the basic augmentation types when we use RandAugment. Our code is available at https://github.com/lliee1/Randomness_Analysis.

Deep Learning-Based Face Recognition through Low-Light Enhancement (딥러닝 기반 저조도 향상 기술을 활용한 얼굴 인식 성능 개선)

  • Changwoo Baek;Kyeongbo Kong
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.5
    • /
    • pp.243-250
    • /
    • 2024
  • This study explores enhancing facial recognition performance in low-light environments using deep learning-based low-light enhancement techniques. Facial recognition technology is widely used in edge devices like smartphones, smart home devices, and security systems, but low-light conditions reduce accuracy due to degraded image quality and increased noise. We reviewed the latest techniques, including Zero-DCE, Zero-DCE++, and SCI (Self-Calibrated Illumination), and applied them as preprocessing steps in facial recognition on edge devices. Using the K-face dataset, experiments on the Qualcomm QRB5165 platform showed significant improvements in F1 SCORE from 0.57 to 0.833 with SCI. Processing times were 0.15ms for SCI, 0.4ms for Zero-DCE, and 0.7ms for Zero-DCE++, all much shorter than the facial recognition model MobileFaceNet's 5ms. These results indicate that these techniques can be effectively used in resource-limited edge devices, enhancing facial recognition in low-light conditions for various applications.

Human Cardiac Abnormality Detection Using Deep Learning with Heart Sound in Newborn Children

  • Eashita Wazed;Hieyong Jeong
    • Annual Conference of KIPS
    • /
    • 2024.10a
    • /
    • pp.461-462
    • /
    • 2024
  • In pediatric healthcare, early detection of cardiovascular diseases in newborns is crucial. Analyzing heart sounds using stethoscopes can be subjective and reliant on physician expertise, potentially leading to delayed diagnosis. There is a need for a simple method that can help even inexperienced doctors detect heart abnormalities without an electrocardiogram or ultrasound. Automated heart sound diagnosis systems can aid clinicians by providing accurate and early detection of abnormal heartbeats. To address this, we developed an intelligent deep-learning model incorporating CNN and LSTM to detect heart abnormalities based on artificial intelligence using heart sound data from stethoscope recordings. Our research achieved a high accuracy rate of 97.8%. Using audio data to introduce advanced models for cardiac abnormalities in children is essential for enhancing early detection and intervention in pediatric cardiovascular healthcare.

Denoising Laplace-domain Seismic Wavefields using Deep Learning

  • Lydie Uwibambe;Jun Hyeon Jo;Wansoo Ha
    • Economic and Environmental Geology
    • /
    • v.57 no.5
    • /
    • pp.499-512
    • /
    • 2024
  • Random noise in seismic data can significantly impair hydrocarbon exploration by degrading the quality of subsurface imaging. We propose a deep learning approach to attenuate random noise in Laplace-domain seismic wavefields. Our method employs a modified U-Net architecture, trained on diverse synthetic P-wave velocity models simulating the Gulf of Mexico subsurface. We rigorously evaluated the network's denoising performance using both the synthetic Pluto velocity model and real Gulf of Mexico field data. We assessed the effectiveness of our approach through Laplace-domain full waveform inversion. The results consistently show that our U-Net approach outperforms traditional singular value decomposition methods in noise attenuation across various scenarios. Numerical examples demonstrate that our method effectively attenuates random noise and significantly enhances the accuracy of subsequent seismic imaging processes.

A Study on Transferring Cloud Dataset for Smoke Extraction Based on Deep Learning (딥러닝 기반 연기추출을 위한 구름 데이터셋의 전이학습에 대한 연구)

  • Kim, Jiyong;Kwak, Taehong;Kim, Yongil
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_2
    • /
    • pp.695-706
    • /
    • 2022
  • Medium and high-resolution optical satellites have proven their effectiveness in detecting wildfire areas. However, smoke plumes generated by wildfire scatter visible light incidents on the surface, thereby interrupting accurate monitoring of the area where wildfire occurs. Therefore, a technology to extract smoke in advance is required. Deep learning technology is expected to improve the accuracy of smoke extraction, but the lack of training datasets limits the application. However, for clouds, which have a similar property of scattering visible light, a large amount of training datasets has been accumulated. The purpose of this study is to develop a smoke extraction technique using deep learning, and the limits due to the lack of datasets were overcome by using a cloud dataset on transfer learning. To check the effectiveness of transfer learning, a small-scale smoke extraction training set was made, and the smoke extraction performance was compared before and after applying transfer learning using a public cloud dataset. As a result, not only the performance in the visible light wavelength band was enhanced but also in the near infrared (NIR) and short-wave infrared (SWIR). Through the results of this study, it is expected that the lack of datasets, which is a critical limit for using deep learning on smoke extraction, can be solved, and therefore, through the advancement of smoke extraction technology, it will be possible to present an advantage in monitoring wildfires.

Detection Ability of Occlusion Object in Deep Learning Algorithm depending on Image Qualities (영상품질별 학습기반 알고리즘 폐색영역 객체 검출 능력 분석)

  • LEE, Jeong-Min;HAM, Geon-Woo;BAE, Kyoung-Ho;PARK, Hong-Ki
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.22 no.3
    • /
    • pp.82-98
    • /
    • 2019
  • The importance of spatial information is rapidly rising. In particular, 3D spatial information construction and modeling for Real World Objects, such as smart cities and digital twins, has become an important core technology. The constructed 3D spatial information is used in various fields such as land management, landscape analysis, environment and welfare service. Three-dimensional modeling with image has the hig visibility and reality of objects by generating texturing. However, some texturing might have occlusion area inevitably generated due to physical deposits such as roadside trees, adjacent objects, vehicles, banners, etc. at the time of acquiring image Such occlusion area is a major cause of the deterioration of reality and accuracy of the constructed 3D modeling. Various studies have been conducted to solve the occlusion area. Recently the researches of deep learning algorithm have been conducted for detecting and resolving the occlusion area. For deep learning algorithm, sufficient training data is required, and the collected training data quality directly affects the performance and the result of the deep learning. Therefore, this study analyzed the ability of detecting the occlusion area of the image using various image quality to verify the performance and the result of deep learning according to the quality of the learning data. An image containing an object that causes occlusion is generated for each artificial and quantified image quality and applied to the implemented deep learning algorithm. The study found that the image quality for adjusting brightness was lower at 0.56 detection ratio for brighter images and that the image quality for pixel size and artificial noise control decreased rapidly from images adjusted from the main image to the middle level. In the F-measure performance evaluation method, the change in noise-controlled image resolution was the highest at 0.53 points. The ability to detect occlusion zones by image quality will be used as a valuable criterion for actual application of deep learning in the future. In the acquiring image, it is expected to contribute a lot to the practical application of deep learning by providing a certain level of image acquisition.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.