• Title/Summary/Keyword: artificial intelligence-based model

Search Result 1,231, Processing Time 0.028 seconds

Weather Recognition Based on 3C-CNN

  • Tan, Ling;Xuan, Dawei;Xia, Jingming;Wang, Chao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.8
    • /
    • pp.3567-3582
    • /
    • 2020
  • Human activities are often affected by weather conditions. Automatic weather recognition is meaningful to traffic alerting, driving assistance, and intelligent traffic. With the boost of deep learning and AI, deep convolutional neural networks (CNN) are utilized to identify weather situations. In this paper, a three-channel convolutional neural network (3C-CNN) model is proposed on the basis of ResNet50.The model extracts global weather features from the whole image through the ResNet50 branch, and extracts the sky and ground features from the top and bottom regions by two CNN5 branches. Then the global features and the local features are merged by the Concat function. Finally, the weather image is classified by Softmax classifier and the identification result is output. In addition, a medium-scale dataset containing 6,185 outdoor weather images named WeatherDataset-6 is established. 3C-CNN is used to train and test both on the Two-class Weather Images and WeatherDataset-6. The experimental results show that 3C-CNN achieves best on both datasets, with the average recognition accuracy up to 94.35% and 95.81% respectively, which is superior to other classic convolutional neural networks such as AlexNet, VGG16, and ResNet50. It is prospected that our method can also work well for images taken at night with further improvement.

A study on the standardization strategy for building of learning data set for machine learning applications (기계학습 활용을 위한 학습 데이터세트 구축 표준화 방안에 관한 연구)

  • Choi, JungYul
    • Journal of Digital Convergence
    • /
    • v.16 no.10
    • /
    • pp.205-212
    • /
    • 2018
  • With the development of high performance CPU / GPU, artificial intelligence algorithms such as deep neural networks, and a large amount of data, machine learning has been extended to various applications. In particular, a large amount of data collected from the Internet of Things, social network services, web pages, and public data is accelerating the use of machine learning. Learning data sets for machine learning exist in various formats according to application fields and data types, and thus it is difficult to effectively process data and apply them to machine learning. Therefore, this paper studied a method for building a learning data set for machine learning in accordance with standardized procedures. This paper first analyzes the requirement of learning data set according to problem types and data types. Based on the analysis, this paper presents the reference model to build learning data set for machine learning applications. This paper presents the target standardization organization and a standard development strategy for building learning data set.

A study on classification of textile design and extraction of regions of interest (텍스타일 디자인 분류 및 관심 영역 도출에 대한 연구)

  • Chae, Seung Wan;Lee, Woo Chang;Lee, Byoung Woo;Lee, Choong Kwon
    • Smart Media Journal
    • /
    • v.10 no.2
    • /
    • pp.70-75
    • /
    • 2021
  • Grouping and classifying similar designs in design increase efficiency in terms of management and provide convenience in terms of use. Using artificial intelligence algorithms, this study attempted to classify textile designs into four categories: dots, flower patterns, stripes, and geometry. In particular, we explored whether it is possible to find and explain the regions of interest underlying classification from the perspective of artificial intelligence. We randomly extracted a total of 4,536 designs at a ratio of 8:2, comprising 3,629 for training and 907 for testing. The models used in the classification were VGG-16 and ResNet-34, both of which showed excellent classification performance with precision on flower pattern designs of 0.79%, 0.89% and recall of 0.95% and 0.38%. Analysis using the Local Interpretable Model-agnostic Explanation (LIME) technique has shown that geometry and flower-patterned designs derived shapes and petals from the region of interest on which classification was based.

A Study on the Liver and Tumor Segmentation and Hologram Visualization of CT Images Using Deep Learning (딥러닝을 이용한 CT 영상의 간과 종양 분할과 홀로그램 시각화 기법 연구)

  • Kim, Dae Jin;Kim, Young Jae;Jeon, Youngbae;Hwang, Tae-sik;Choi, Seok Won;Baek, Jeong-Heum;Kim, Kwang Gi
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.5
    • /
    • pp.757-768
    • /
    • 2022
  • In this paper, we proposed a system that visualizes a hologram device in 3D by utilizing the CT image segmentation function based on artificial intelligence deep learning. The input axial CT medical image is converted into Sagittal and Coronal, and the input image and the converted image are divided into 3D volumes using ResUNet, a deep learning model. In addition, the volume is created by segmenting the tumor region in the segmented liver image. Each result is integrated into one 3D volume, displayed in a medical image viewer, and converted into a video. When the converted video is transmitted to the hologram device and output from the device, a 3D image with a sense of space can be checked. As for the performance of the deep learning model, in Axial, the basic input image, DSC showed 95.0% performance in liver region segmentation and 67.5% in liver tumor region segmentation. If the system is applied to a real-world care environment, additional physical contact is not required, making it safer for patients to explain changes before and after surgery more easily. In addition, it will provide medical staff with information on liver and liver tumors necessary for treatment or surgery in a three-dimensional manner, and help patients manage them after surgery by comparing and observing the liver before and after liver resection.

VS3-NET: Neural variational inference model for machine-reading comprehension

  • Park, Cheoneum;Lee, Changki;Song, Heejun
    • ETRI Journal
    • /
    • v.41 no.6
    • /
    • pp.771-781
    • /
    • 2019
  • We propose the VS3-NET model to solve the task of question answering questions with machine-reading comprehension that searches for an appropriate answer in a given context. VS3-NET is a model that trains latent variables for each question using variational inferences based on a model of a simple recurrent unit-based sentences and self-matching networks. The types of questions vary, and the answers depend on the type of question. To perform efficient inference and learning, we introduce neural question-type models to approximate the prior and posterior distributions of the latent variables, and we use these approximated distributions to optimize a reparameterized variational lower bound. The context given in machine-reading comprehension usually comprises several sentences, leading to performance degradation caused by context length. Therefore, we model a hierarchical structure using sentence encoding, in which as the context becomes longer, the performance degrades. Experimental results show that the proposed VS3-NET model has an exact-match score of 76.8% and an F1 score of 84.5% on the SQuAD test set.

Multimodal audiovisual speech recognition architecture using a three-feature multi-fusion method for noise-robust systems

  • Sanghun Jeon;Jieun Lee;Dohyeon Yeo;Yong-Ju Lee;SeungJun Kim
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.22-34
    • /
    • 2024
  • Exposure to varied noisy environments impairs the recognition performance of artificial intelligence-based speech recognition technologies. Degraded-performance services can be utilized as limited systems that assure good performance in certain environments, but impair the general quality of speech recognition services. This study introduces an audiovisual speech recognition (AVSR) model robust to various noise settings, mimicking human dialogue recognition elements. The model converts word embeddings and log-Mel spectrograms into feature vectors for audio recognition. A dense spatial-temporal convolutional neural network model extracts features from log-Mel spectrograms, transformed for visual-based recognition. This approach exhibits improved aural and visual recognition capabilities. We assess the signal-to-noise ratio in nine synthesized noise environments, with the proposed model exhibiting lower average error rates. The error rate for the AVSR model using a three-feature multi-fusion method is 1.711%, compared to the general 3.939% rate. This model is applicable in noise-affected environments owing to its enhanced stability and recognition rate.

Artificial Intelligence for Assistance of Facial Expression Practice Using Emotion Classification (감정 분류를 이용한 표정 연습 보조 인공지능)

  • Dong-Kyu, Kim;So Hwa, Lee;Jae Hwan, Bong
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.6
    • /
    • pp.1137-1144
    • /
    • 2022
  • In this study, an artificial intelligence(AI) was developed to help with facial expression practice in order to express emotions. The developed AI used multimodal inputs consisting of sentences and facial images for deep neural networks (DNNs). The DNNs calculated similarities between the emotions predicted by the sentences and the emotions predicted by facial images. The user practiced facial expressions based on the situation given by sentences, and the AI provided the user with numerical feedback based on the similarity between the emotion predicted by sentence and the emotion predicted by facial expression. ResNet34 structure was trained on FER2013 public data to predict emotions from facial images. To predict emotions in sentences, KoBERT model was trained in transfer learning manner using the conversational speech dataset for emotion classification opened to the public by AIHub. The DNN that predicts emotions from the facial images demonstrated 65% accuracy, which is comparable to human emotional classification ability. The DNN that predicts emotions from the sentences achieved 90% accuracy. The performance of the developed AI was evaluated through experiments with changing facial expressions in which an ordinary person was participated.

Breast Cytology Diagnosis using a Hybrid Case-based Reasoning and Genetic Algorithms Approach

  • Ahn, Hyun-Chul;Kim, Kyoung-Jae
    • Proceedings of the Korea Inteligent Information System Society Conference
    • /
    • 2007.05a
    • /
    • pp.389-398
    • /
    • 2007
  • Case-based reasoning (CBR) is one of the most popular prediction techniques for medical diagnosis because it is easy to apply, has no possibility of overfitting, and provides a good explanation for the output. However, it has a critical limitation - its prediction performance is generally lower than other artificial intelligence techniques like artificial neural networks (ANNs). In order to obtain accurate results from CBR, effective retrieval and matching of useful prior cases for the problem is essential, but it is still a controversial issue to design a good matching and retrieval mechanism for CBR systems. In this study, we propose a novel approach to enhance the prediction performance of CBR. Our suggestion is the simultaneous optimization of feature weights, instance selection, and the number of neighbors that combine using genetic algorithms (GAs). Our model improves the prediction performance in three ways - (1) measuring similarity between cases more accurately by considering relative importance of each feature, (2) eliminating redundant or erroneous reference cases, and (3) combining several similar cases represent significant patterns. To validate the usefulness of our model, this study applied it to a real-world case for evaluating cytological features derived directly from a digital scan of breast fine needle aspirate (FNA) slides. Experimental results showed that the prediction accuracy of conventional CBR may be improved significantly by using our model. We also found that our proposed model outperformed all the other optimized models for CBR using GA.

  • PDF

Development of Artificial Neural Network Model for Estimation of Cable Tension of Cable-Stayed Bridge (사장교 케이블의 장력 추정을 위한 인공신경망 모델 개발)

  • Kim, Ki-Jung;Park, Yoo-Sin;Park, Sung-Woo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.3
    • /
    • pp.414-419
    • /
    • 2020
  • An artificial intelligence-based cable tension estimation model was developed to expand the utilization of data obtained from cable accelerometers of cable-stayed bridges. The model was based on an algorithm for selecting the natural frequency in the tension estimation process based on the vibration method and an applied artificial neural network (ANN). The training data of the ANN was composed after converting the cable acceleration data into the frequency, and machine learning was carried out using the characteristics with a pattern on the natural frequency. When developing the training data, the frequencies with various amplitudes can be used to represent the frequencies of multiple shapes to improve the selection performance for natural frequencies. The performance of the model was estimated by comparing it with the control criteria of the tension estimated by an expert. As a result of the verification using 139 frequencies obtained from the cable accelerometer as the input, the natural frequency was determined to be similar to the real criteria and the estimated tension of the cable by the natural frequency was 96.4% of the criteria.

CCTV-Based Multi-Factor Authentication System

  • Kwon, Byoung-Wook;Sharma, Pradip Kumar;Park, Jong-Hyuk
    • Journal of Information Processing Systems
    • /
    • v.15 no.4
    • /
    • pp.904-919
    • /
    • 2019
  • Many security systems rely solely on solutions based on Artificial Intelligence, which are weak in nature. These security solutions can be easily manipulated by malicious users who can gain unlawful access. Some security systems suggest using fingerprint-based solutions, but they can be easily deceived by copying fingerprints with clay. Image-based security is undoubtedly easy to manipulate, but it is also a solution that does not require any special training on the part of the user. In this paper, we propose a multi-factor security framework that operates in a three-step process to authenticate the user. The motivation of the research lies in utilizing commonly available and inexpensive devices such as onsite CCTV cameras and smartphone camera and providing fully secure user authentication. We have used technologies such as Argon2 for hashing image features and physically unclonable identification for secure device-server communication. We also discuss the methodological workflow of the proposed multi-factor authentication framework. In addition, we present the service scenario of the proposed model. Finally, we analyze qualitatively the proposed model and compare it with state-of-the-art methods to evaluate the usability of the model in real-world applications.