• Title/Summary/Keyword: deep learning models

Search Result 1,392, Processing Time 0.05 seconds

Development of a Deep Learning Model for Detecting Fake Reviews Using Author Linguistic Features (작성자 언어적 특성 기반 가짜 리뷰 탐지 딥러닝 모델 개발)

  • Shin, Dong Hoon;Shin, Woo Sik;Kim, Hee Woong
    • The Journal of Information Systems
    • /
    • v.31 no.4
    • /
    • pp.01-23
    • /
    • 2022
  • Purpose This study aims to propose a deep learning-based fake review detection model by combining authors' linguistic features and semantic information of reviews. Design/methodology/approach This study used 358,071 review data of Yelp to develop fake review detection model. We employed linguistic inquiry and word count (LIWC) to extract 24 linguistic features of authors. Then we used deep learning architectures such as multilayer perceptron(MLP), long short-term memory(LSTM) and transformer to learn linguistic features and semantic features for fake review detection. Findings The results of our study show that detection models using both linguistic and semantic features outperformed other models using single type of features. In addition, this study confirmed that differences in linguistic features between fake reviewer and authentic reviewer are significant. That is, we found that linguistic features complement semantic information of reviews and further enhance predictive power of fake detection model.

Selection of Three (E)UV Channels for Solar Satellite Missions by Deep Learning

  • Lim, Daye;Moon, Yong-Jae;Park, Eunsu;Lee, Jin-Yi
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.46 no.1
    • /
    • pp.42.2-43
    • /
    • 2021
  • We address a question of what are three main channels that can best translate other channels in ultraviolet (UV) and extreme UV (EUV) observations. For this, we compare the image translations among the nine channels of the Atmospheric Imaging Assembly on the Solar Dynamics Observatory using a deep learning model based on conditional generative adversarial networks. In this study, we develop 170 deep learning models: 72 models for single-channel input, 56 models for double-channel input, and 42 models for triple-channel input. All models have a single-channel output. Then we evaluate the model results by pixel-to-pixel correlation coefficients (CCs) within the solar disk. Major results from this study are as follows. First, the model with 131 Å shows the best performance (average CC = 0.84) among single-channel models. Second, the model with 131 and 1600 Å shows the best translation (average CC = 0.95) among double-channel models. Third, among the triple-channel models with the highest average CC (0.97), the model with 131, 1600, and 304 Å is suggested in that the minimum CC (0.96) is the highest. Interestingly they are representative coronal, photospheric, and chromospheric lines, respectively. Our results may be used as a secondary perspective in addition to primary scientific purposes in selecting a few channels of an UV/EUV imaging instrument for future solar satellite missions.

  • PDF

Super-Resolution Transmission Electron Microscope Image of Nanomaterials Using Deep Learning (딥러닝을 이용한 나노소재 투과전자 현미경의 초해상 이미지 획득)

  • Nam, Chunghee
    • Korean Journal of Materials Research
    • /
    • v.32 no.8
    • /
    • pp.345-353
    • /
    • 2022
  • In this study, using deep learning, super-resolution images of transmission electron microscope (TEM) images were generated for nanomaterial analysis. 1169 paired images with 256 × 256 pixels (high resolution: HR) from TEM measurements and 32 × 32 pixels (low resolution: LR) produced using the python module openCV were trained with deep learning models. The TEM images were related to DyVO4 nanomaterials synthesized by hydrothermal methods. Mean-absolute-error (MAE), peak-signal-to-noise-ratio (PSNR), and structural similarity (SSIM) were used as metrics to evaluate the performance of the models. First, a super-resolution image (SR) was obtained using the traditional interpolation method used in computer vision. In the SR image at low magnification, the shape of the nanomaterial improved. However, the SR images at medium and high magnification failed to show the characteristics of the lattice of the nanomaterials. Second, to obtain a SR image, the deep learning model includes a residual network which reduces the loss of spatial information in the convolutional process of obtaining a feature map. In the process of optimizing the deep learning model, it was confirmed that the performance of the model improved as the number of data increased. In addition, by optimizing the deep learning model using the loss function, including MAE and SSIM at the same time, improved results of the nanomaterial lattice in SR images were achieved at medium and high magnifications. The final proposed deep learning model used four residual blocks to obtain the characteristic map of the low-resolution image, and the super-resolution image was completed using Upsampling2D and the residual block three times.

An Explainable Deep Learning-Based Classification Method for Facial Image Quality Assessment

  • Kuldeep Gurjar;Surjeet Kumar;Arnav Bhavsar;Kotiba Hamad;Yang-Sae Moon;Dae Ho Yoon
    • Journal of Information Processing Systems
    • /
    • v.20 no.4
    • /
    • pp.558-573
    • /
    • 2024
  • Considering factors such as illumination, camera quality variations, and background-specific variations, identifying a face using a smartphone-based facial image capture application is challenging. Face Image Quality Assessment refers to the process of taking a face image as input and producing some form of "quality" estimate as an output. Typically, quality assessment techniques use deep learning methods to categorize images. The models used in deep learning are shown as black boxes. This raises the question of the trustworthiness of the models. Several explainability techniques have gained importance in building this trust. Explainability techniques provide visual evidence of the active regions within an image on which the deep learning model makes a prediction. Here, we developed a technique for reliable prediction of facial images before medical analysis and security operations. A combination of gradient-weighted class activation mapping and local interpretable model-agnostic explanations were used to explain the model. This approach has been implemented in the preselection of facial images for skin feature extraction, which is important in critical medical science applications. We demonstrate that the use of combined explanations provides better visual explanations for the model, where both the saliency map and perturbation-based explainability techniques verify predictions.

A Hierarchical Deep Convolutional Neural Network for Crop Species and Diseases Classification (Deep Convolutional Neural Network(DCNN)을 이용한 계층적 농작물의 종류와 질병 분류 기법)

  • Borin, Min;Rah, HyungChul;Yoo, Kwan-Hee
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.11
    • /
    • pp.1653-1671
    • /
    • 2022
  • Crop diseases affect crop production, more than 30 billion USD globally. We proposed a classification study of crop species and diseases using deep learning algorithms for corn, cucumber, pepper, and strawberry. Our study has three steps of species classification, disease detection, and disease classification, which is noteworthy for using captured images without additional processes. We designed deep learning approach of deep learning convolutional neural networks based on Mask R-CNN model to classify crop species. Inception and Resnet models were presented for disease detection and classification sequentially. For classification, we trained Mask R-CNN network and achieved loss value of 0.72 for crop species classification and segmentation. For disease detection, InceptionV3 and ResNet101-V2 models were trained for nodes of crop species on 1,500 images of normal and diseased labels, resulting in the accuracies of 0.984, 0.969, 0.956, and 0.962 for corn, cucumber, pepper, and strawberry by InceptionV3 model with higher accuracy and AUC. For disease classification, InceptionV3 and ResNet 101-V2 models were trained for nodes of crop species on 1,500 images of diseased label, resulting in the accuracies of 0.995 and 0.992 for corn and cucumber by ResNet101 with higher accuracy and AUC whereas 0.940 and 0.988 for pepper and strawberry by Inception.

A Comparative Analysis of Deep Learning Frameworks for Image Learning (이미지 학습을 위한 딥러닝 프레임워크 비교분석)

  • jong-min Kim;Dong-Hwi Lee
    • Convergence Security Journal
    • /
    • v.22 no.4
    • /
    • pp.129-133
    • /
    • 2022
  • Deep learning frameworks are still evolving, and there are various frameworks. Typical deep learning frameworks include TensorFlow, PyTorch, and Keras. The Deepram framework utilizes optimization models in image classification through image learning. In this paper, we use the TensorFlow and PyTorch frameworks, which are most widely used in the deep learning image recognition field, to proceed with image learning, and compare and analyze the results derived in this process to know the optimized framework. was made.

Deep Meta Learning Based Classification Problem Learning Method for Skeletal Maturity Indication (골 성숙도 판별을 위한 심층 메타 학습 기반의 분류 문제 학습 방법)

  • Min, Jeong Won;Kang, Dong Joong
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.2
    • /
    • pp.98-107
    • /
    • 2018
  • In this paper, we propose a method to classify the skeletal maturity with a small amount of hand wrist X-ray image using deep learning-based meta-learning. General deep-learning techniques require large amounts of data, but in many cases, these data sets are not available for practical application. Lack of learning data is usually solved through transfer learning using pre-trained models with large data sets. However, transfer learning performance may be degraded due to over fitting for unknown new task with small data, which results in poor generalization capability. In addition, medical images require high cost resources such as a professional manpower and mcuh time to obtain labeled data. Therefore, in this paper, we use meta-learning that can classify using only a small amount of new data by pre-trained models trained with various learning tasks. First, we train the meta-model by using a separate data set composed of various learning tasks. The network learns to classify the bone maturity using the bone maturity data composed of the radiographs of the wrist. Then, we compare the results of the classification using the conventional learning algorithm with the results of the meta learning by the same number of learning data sets.

Transfer Learning-Based Feature Fusion Model for Classification of Maneuver Weapon Systems

  • Jinyong Hwang;You-Rak Choi;Tae-Jin Park;Ji-Hoon Bae
    • Journal of Information Processing Systems
    • /
    • v.19 no.5
    • /
    • pp.673-687
    • /
    • 2023
  • Convolutional neural network-based deep learning technology is the most commonly used in image identification, but it requires large-scale data for training. Therefore, application in specific fields in which data acquisition is limited, such as in the military, may be challenging. In particular, the identification of ground weapon systems is a very important mission, and high identification accuracy is required. Accordingly, various studies have been conducted to achieve high performance using small-scale data. Among them, the ensemble method, which achieves excellent performance through the prediction average of the pre-trained models, is the most representative method; however, it requires considerable time and effort to find the optimal combination of ensemble models. In addition, there is a performance limitation in the prediction results obtained by using an ensemble method. Furthermore, it is difficult to obtain the ensemble effect using models with imbalanced classification accuracies. In this paper, we propose a transfer learning-based feature fusion technique for heterogeneous models that extracts and fuses features of pre-trained heterogeneous models and finally, fine-tunes hyperparameters of the fully connected layer to improve the classification accuracy. The experimental results of this study indicate that it is possible to overcome the limitations of the existing ensemble methods by improving the classification accuracy through feature fusion between heterogeneous models based on transfer learning.

A Model of Strawberry Pest Recognition using Artificial Intelligence Learning

  • Guangzhi Zhao
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.2
    • /
    • pp.133-143
    • /
    • 2023
  • In this study, we propose a big data set of strawberry pests collected directly for diagnosis model learning and an automatic pest diagnosis model architecture based on deep learning. First, a big data set related to strawberry pests, which did not exist anywhere before, was directly collected from the web. A total of more than 12,000 image data was directly collected and classified, and this data was used to train a deep learning model. Second, the deep-learning-based automatic pest diagnosis module is a module that classifies what kind of pest or disease corresponds to when a user inputs a desired picture. In particular, we propose a model architecture that can optimally classify pests based on a convolutional neural network among deep learning models. Through this, farmers can easily identify diseases and pests without professional knowledge, and can respond quickly accordingly.

White Blood Cell Types Classification Using Deep Learning Models

  • Bagido, Rufaidah Ali;Alzahrani, Manar;Arif, Muhammad
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.9
    • /
    • pp.223-229
    • /
    • 2021
  • Classification of different blood cell types is an essential task for human's medical treatment. The white blood cells have different types of cells. Counting total White Blood Cells (WBC) and differential of the WBC types are required by the physicians to diagnose the disease correctly. This paper used transfer learning methods to the pre-trained deep learning models to classify different WBCs. The best pre-trained model was Inception ResNetV2 with Adam optimizer that produced classification accuracy of 98.4% for the dataset comprising four types of WBCs.