• Title/Summary/Keyword: artificial dataset

Search Result 448, Processing Time 0.027 seconds

Real-world multimodal lifelog dataset for human behavior study

  • Chung, Seungeun;Jeong, Chi Yoon;Lim, Jeong Mook;Lim, Jiyoun;Noh, Kyoung Ju;Kim, Gague;Jeong, Hyuntae
    • ETRI Journal
    • /
    • v.44 no.3
    • /
    • pp.426-437
    • /
    • 2022
  • To understand the multilateral characteristics of human behavior and physiological markers related to physical, emotional, and environmental states, extensive lifelog data collection in a real-world environment is essential. Here, we propose a data collection method using multimodal mobile sensing and present a long-term dataset from 22 subjects and 616 days of experimental sessions. The dataset contains over 10 000 hours of data, including physiological, data such as photoplethysmography, electrodermal activity, and skin temperature in addition to the multivariate behavioral data. Furthermore, it consists of 10 372 user labels with emotional states and 590 days of sleep quality data. To demonstrate feasibility, human activity recognition was applied on the sensor data using a convolutional neural network-based deep learning model with 92.78% recognition accuracy. From the activity recognition result, we extracted the daily behavior pattern and discovered five representative models by applying spectral clustering. This demonstrates that the dataset contributed toward understanding human behavior using multimodal data accumulated throughout daily lives under natural conditions.

Study of Posture Evaluation Method in Chest PA Examination based on Artificial Intelligence (인공지능 기반 흉부 후전방향 검사에서 자세 평가 방법에 관한 연구)

  • Ho Seong Hwang;Yong Seok Choi;Dae Won Lee;Dong Hyun Kim;Ho Chul Kim
    • Journal of Biomedical Engineering Research
    • /
    • v.44 no.3
    • /
    • pp.167-175
    • /
    • 2023
  • Chest PA is the basic examination of radiographic imaging. Moreover, Chest PA's demands are constantly increasing because of the Increase in respiratory diseases. However, it is not meeting the demand due to problems such as a shortage of radiological technologist, sexual shame caused by patient contact, and the spread of infectious diseases. There have been many cases of using artificial intelligence to solve this problem. Therefore, the purpose of this research is to build an artificial intelligence dataset of Chest PA and to find a posture evaluation method. To construct the posture dataset, the posture image is acquired during actual and simulated examination and classified correct and incorrect posture of the patient. And to evaluate the artificial intelligence posture method, a posture estimation algorithm is used to preprocess the dataset and an artificial intelligence classification algorithm is applied. As a result, Chest PA posture dataset is validated with in over 95% accuracy in all artificial intelligence classification and the accuracy is improved through the Top-Down posture estimation algorithm AlphaPose and the classification InceptionV3 algorithm. Based on this, it will be possible to build a non-face-to-face automatic Chest PA examination system using artificial intelligence.

A Virtual Battlefield Situation Dataset Generation for Battlefield Analysis based on Artificial Intelligence

  • Cho, Eunji;Jin, Soyeon;Shin, Yukyung;Lee, Woosin
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.6
    • /
    • pp.33-42
    • /
    • 2022
  • In the existing intelligent command control system study, the analysis results of the commander's battlefield situation questions are provided from knowledge-based situation data. Analysis reporters write these results in various expressions of natural language. However, it is important to analyze situations about information and intelligence according to context. Analyzing the battlefield situation using artificial intelligence is necessary. We propose a virtual dataset generation method based on battlefield simulation scenarios in order to provide a dataset necessary for the battlefield situation analysis based on artificial intelligence. Dataset is generated after identifying battlefield knowledge elements in scenarios. When a candidate hypothesis is created, a unit hypothesis is automatically created. By combining unit hypotheses, similar identification hypothesis combinations are generated. An aggregation hypothesis is generated by grouping candidate hypotheses. Dataset generator SW implementation demonstrates that the proposed method can be generated the virtual battlefield situation dataset.

COVID-19: Improving the accuracy using data augmentation and pre-trained DCNN Models

  • Saif Hassan;Abdul Ghafoor;Zahid Hussain Khand;Zafar Ali;Ghulam Mujtaba;Sajid Khan
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.7
    • /
    • pp.170-176
    • /
    • 2024
  • Since the World Health Organization (WHO) has declared COVID-19 as pandemic, many researchers have started working on developing vaccine and developing AI systems to detect COVID-19 patient using Chest X-ray images. The purpose of this work is to improve the performance of pre-trained Deep convolution neural nets (DCNNs) on Chest X-ray images dataset specially COVID-19 which is developed by collecting from different sources such as GitHub, Kaggle. To improve the performance of Deep CNNs, data augmentation is used in this study. The COVID-19 dataset collected from GitHub was containing 257 images while the other two classes normal and pneumonia were having more than 500 images each class. There were two issues whike training DCNN model on this dataset, one is unbalanced and second is the data is very less. In order to handle these both issues, we performed data augmentation such as rotation, flipping to increase and balance the dataset. After data augmentation each class contains 510 images. Results show that augmentation on Chest X-ray images helps in improving accuracy. The accuracy before and after augmentation produced by our proposed architecture is 96.8% and 98.4% respectively.

ETLi: Efficiently annotated traffic LiDAR dataset using incremental and suggestive annotation

  • Kang, Jungyu;Han, Seung-Jun;Kim, Nahyeon;Min, Kyoung-Wook
    • ETRI Journal
    • /
    • v.43 no.4
    • /
    • pp.630-639
    • /
    • 2021
  • Autonomous driving requires a computerized perception of the environment for safety and machine-learning evaluation. Recognizing semantic information is difficult, as the objective is to instantly recognize and distinguish items in the environment. Training a model with real-time semantic capability and high reliability requires extensive and specialized datasets. However, generalized datasets are unavailable and are typically difficult to construct for specific tasks. Hence, a light detection and ranging semantic dataset suitable for semantic simultaneous localization and mapping and specialized for autonomous driving is proposed. This dataset is provided in a form that can be easily used by users familiar with existing two-dimensional image datasets, and it contains various weather and light conditions collected from a complex and diverse practical setting. An incremental and suggestive annotation routine is proposed to improve annotation efficiency. A model is trained to simultaneously predict segmentation labels and suggest class-representative frames. Experimental results demonstrate that the proposed algorithm yields a more efficient dataset than uniformly sampled datasets.

Handwritten Hangul Graphemes Classification Using Three Artificial Neural Networks

  • Aaron Daniel Snowberger;Choong Ho Lee
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.2
    • /
    • pp.167-173
    • /
    • 2023
  • Hangul is unique compared to other Asian languages because of its simple letter forms that combine to create syllabic shapes. There are 24 basic letters that can be combined to form 27 additional complex letters. This produces 51 graphemes. Hangul optical character recognition has been a research topic for some time; however, handwritten Hangul recognition continues to be challenging owing to the various writing styles, slants, and cursive-like nature of the handwriting. In this study, a dataset containing thousands of samples of 51 Hangul graphemes was gathered from 110 freshmen university students to create a robust dataset with high variance for training an artificial neural network. The collected dataset included 2200 samples for each consonant grapheme and 1100 samples for each vowel grapheme. The dataset was normalized to the MNIST digits dataset, trained in three neural networks, and the obtained results were compared.

Recommendations for the Construction of a Quslity-Controlled Stress Measurement Dataset (품질이 관리된 스트레스 측정용 테이터셋 구축을 위한 제언)

  • Tai Hoon KIM;In Seop NA
    • Smart Media Journal
    • /
    • v.13 no.2
    • /
    • pp.44-51
    • /
    • 2024
  • The construction of a stress measurement detaset plays a curcial role in various modern applications. In particular, for the efficient training of artificial intelligence models for stress measurement, it is essential to compare various biases and construct a quality-controlled dataset. In this paper, we propose the construction of a stress measurement dataset with quality management through the comparison of various biases. To achieve this, we introduce strss definitions and measurement tools, the process of building an artificial intelligence stress dataset, strategies to overcome biases for quality improvement, and considerations for stress data collection. Specifically, to manage dataset quality, we discuss various biases such as selection bias, measurement bias, causal bias, confirmation bias, and artificial intelligence bias that may arise during stress data collection. Through this paper, we aim to systematically understand considerations for stress data collection and various biases that may occur during the construction of a stress dataset, contributing to the construction of a dataset with guaranteed quality by overcoming these biases.

A Study on the Generation of Datasets for Applied AI to OLED Life Prediction

  • CHUNG, Myung-Ae;HAN, Dong Hun;AHN, Seongdeok;KANG, Min Soo
    • Korean Journal of Artificial Intelligence
    • /
    • v.10 no.2
    • /
    • pp.7-11
    • /
    • 2022
  • OLED displays cannot be used permanently due to burn-in or generation of dark spots due to degradation. Therefore, the time when the display can operate normally is very important. It is close to impossible to physically measure the time when the display operates normally. Therefore, the time that works normally should be predicted in a way other than a physical way. Therefore, if you do computer simulations based on artificial intelligence, you can increase the accuracy of prediction by saving time and continuous learning. Therefore, if we do computer simulations based on artificial intelligence, we can increase the accuracy of prediction by saving time and continuous learning. In this paper, a dataset in the form of development from generation to diffusion of dark spots, which is one of the causes related to the life of OLED, was generated by applying the finite element method. The dark spots were generated in nine conditions, such as 0.1 to 2.0 ㎛ with the size of pinholes, the number was 10 to 100, and 50% with water content. The learning data created in this way may be a criterion for generating an artificial intelligence-based dataset.

Multi-type Image Noise Classification by Using Deep Learning

  • Waqar Ahmed;Zahid Hussain Khand;Sajid Khan;Ghulam Mujtaba;Muhammad Asif Khan;Ahmad Waqas
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.7
    • /
    • pp.143-147
    • /
    • 2024
  • Image noise classification is a classical problem in the field of image processing, machine learning, deep learning and computer vision. In this paper, image noise classification is performed using deep learning. Keras deep learning library of TensorFlow is used for this purpose. 6900 images images are selected from the Kaggle database for the classification purpose. Dataset for labeled noisy images of multiple type was generated with the help of Matlab from a dataset of non-noisy images. Labeled dataset comprised of Salt & Pepper, Gaussian and Sinusoidal noise. Different training and tests sets were partitioned to train and test the model for image classification. In deep neural networks CNN (Convolutional Neural Network) is used due to its in-depth and hidden patterns and features learning in the images to be classified. This deep learning of features and patterns in images make CNN outperform the other classical methods in many classification problems.

A Study on Synthetic Dataset Generation Method for Maritime Traffic Situation Awareness (해상교통 상황인지 향상을 위한 합성 데이터셋 구축방안 연구)

  • Youngchae Lee;Sekil Park
    • Journal of Information Technology Applications and Management
    • /
    • v.30 no.6
    • /
    • pp.69-80
    • /
    • 2023
  • Ship collision accidents not only cause loss of life and property damage, but also cause marine pollution and can become national disasters, so prevention is very important. Most of these ship collision accidents are caused by human factors due to the navigation officer's lack of vigilance and carelessness, and in many cases, they can be prevented through the support of a system that helps with situation awareness. Recently, artificial intelligence has been used to develop systems that help navigators recognize the situation, but the sea is very wide and deep, so it is difficult to secure maritime traffic datasets, which also makes it difficult to develop artificial intelligence models. In this paper, to solve these difficulties, we propose a method to build a dataset with characteristics similar to actual maritime traffic datasets. The proposed method uses segmentation and inpainting technologies to build a foreground and background dataset, and then applies compositing technology to create a synthetic dataset. Through prototype implementation and result analysis of the proposed method, it was confirmed that the proposed method is effective in overcoming the difficulties of dataset construction and complementing various scenes similar to reality.