• Title/Summary/Keyword: Deep Learning Dataset

Search Result 815, Processing Time 0.026 seconds

Gesture-Based Emotion Recognition by 3D-CNN and LSTM with Keyframes Selection

  • Ly, Son Thai;Lee, Guee-Sang;Kim, Soo-Hyung;Yang, Hyung-Jeong
    • International Journal of Contents
    • /
    • v.15 no.4
    • /
    • pp.59-64
    • /
    • 2019
  • In recent years, emotion recognition has been an interesting and challenging topic. Compared to facial expressions and speech modality, gesture-based emotion recognition has not received much attention with only a few efforts using traditional hand-crafted methods. These approaches require major computational costs and do not offer many opportunities for improvement as most of the science community is conducting their research based on the deep learning technique. In this paper, we propose an end-to-end deep learning approach for classifying emotions based on bodily gestures. In particular, the informative keyframes are first extracted from raw videos as input for the 3D-CNN deep network. The 3D-CNN exploits the short-term spatiotemporal information of gesture features from selected keyframes, and the convolutional LSTM networks learn the long-term feature from the features results of 3D-CNN. The experimental results on the FABO dataset exceed most of the traditional methods results and achieve state-of-the-art results for the deep learning-based technique for gesture-based emotion recognition.

Validation Data Augmentation for Improving the Grading Accuracy of Diabetic Macular Edema using Deep Learning (딥러닝을 이용한 당뇨성황반부종 등급 분류의 정확도 개선을 위한 검증 데이터 증강 기법)

  • Lee, Tae Soo
    • Journal of Biomedical Engineering Research
    • /
    • v.40 no.2
    • /
    • pp.48-54
    • /
    • 2019
  • This paper proposed a method of validation data augmentation for improving the grading accuracy of diabetic macular edema (DME) using deep learning. The data augmentation technique is basically applied in order to secure diversity of data by transforming one image to several images through random translation, rotation, scaling and reflection in preparation of input data of the deep neural network (DNN). In this paper, we apply this technique in the validation process of the trained DNN, and improve the grading accuracy by combining the classification results of the augmented images. To verify the effectiveness, 1,200 retinal images of Messidor dataset was divided into training and validation data at the ratio 7:3. By applying random augmentation to 359 validation data, $1.61{\pm}0.55%$ accuracy improvement was achieved in the case of six times augmentation (N=6). This simple method has shown that the accuracy can be improved in the N range from 2 to 6 with the correlation coefficient of 0.5667. Therefore, it is expected to help improve the diagnostic accuracy of DME with the grading information provided by the proposed DNN.

Region of Interest Localization for Bone Age Estimation Using Whole-Body Bone Scintigraphy

  • Do, Thanh-Cong;Yang, Hyung Jeong;Kim, Soo Hyung;Lee, Guee Sang;Kang, Sae Ryung;Min, Jung Joon
    • Smart Media Journal
    • /
    • v.10 no.2
    • /
    • pp.22-29
    • /
    • 2021
  • In the past decade, deep learning has been applied to various medical image analysis tasks. Skeletal bone age estimation is clinically important as it can help prevent age-related illness and pave the way for new anti-aging therapies. Recent research has applied deep learning techniques to the task of bone age assessment and achieved positive results. In this paper, we propose a bone age prediction method using a deep convolutional neural network. Specifically, we first train a classification model that automatically localizes the most discriminative region of an image and crops it from the original image. The regions of interest are then used as input for a regression model to estimate the age of the patient. The experiments are conducted on a whole-body scintigraphy dataset that was collected by Chonnam National University Hwasun Hospital. The experimental results illustrate the potential of our proposed method, which has a mean absolute error of 3.35 years. Our proposed framework can be used as a robust supporting tool for clinicians to prevent age-related diseases.

Estimation of tomato maturity as a continuous index using deep neural networks

  • Taehyeong Kim;Dae-Hyun Lee;Seung-Woo Kang;Soo-Hyun Cho;Kyoung-Chul Kim
    • Korean Journal of Agricultural Science
    • /
    • v.49 no.4
    • /
    • pp.837-845
    • /
    • 2022
  • In this study, tomato maturity was estimated based on deep learning for a harvesting robot. Tomato images were obtained using a RGB camera installed on a monitoring robot, which was developed previously, and the samples were cropped to 128 × 128 size images to generate a dataset for training the classification model. The classification model was constructed based on convolutional neural networks, and the mean-variance loss was used to learn implicitly the distribution of the data features by class. In the test stage, the tomato maturity was estimated as a continuous index, which has a range of 0 to 1, by calculating the expected class value. The results show that the F1-score of the classification was approximately 0.94, and the performance was similar to that of a deep learning-based classification task in the agriculture field. In addition, it was possible to estimate the distribution in each maturity stage. From the results, it was found that our approach can not only classify the discrete maturation stages of the tomatoes but also can estimate the continuous maturity.

Forecasting the Precipitation of the Next Day Using Deep Learning (딥러닝 기법을 이용한 내일강수 예측)

  • Ha, Ji-Hun;Lee, Yong Hee;Kim, Yong-Hyuk
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.26 no.2
    • /
    • pp.93-98
    • /
    • 2016
  • For accurate precipitation forecasts the choice of weather factors and prediction method is very important. Recently, machine learning has been widely used for forecasting precipitation, and artificial neural network, one of machine learning techniques, showed good performance. In this paper, we suggest a new method for forecasting precipitation using DBN, one of deep learning techniques. DBN has an advantage that initial weights are set by unsupervised learning, so this compensates for the defects of artificial neural networks. We used past precipitation, temperature, and the parameters of the sun and moon's motion as features for forecasting precipitation. The dataset consists of observation data which had been measured for 40 years from AWS in Seoul. Experiments were based on 8-fold cross validation. As a result of estimation, we got probabilities of test dataset, so threshold was used for the decision of precipitation. CSI and Bias were used for indicating the precision of precipitation. Our experimental results showed that DBN performed better than MLP.

A Deep Learning Performance Comparison of R and Tensorflow (R과 텐서플로우 딥러닝 성능 비교)

  • Sung-Bong Jang
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.487-494
    • /
    • 2023
  • In this study, performance comparison was performed on R and TensorFlow, which are free deep learning tools. In the experiment, six types of deep neural networks were built using each tool, and the neural networks were trained using the 10-year Korean temperature dataset. The number of nodes in the input layer of the constructed neural network was set to 10, the number of output layers was set to 5, and the hidden layer was set to 5, 10, and 20 to conduct experiments. The dataset includes 3600 temperature data collected from Gangnam-gu, Seoul from March 1, 2013 to March 29, 2023. For performance comparison, the future temperature was predicted for 5 days using the trained neural network, and the root mean square error (RMSE) value was measured using the predicted value and the actual value. Experiment results shows that when there was one hidden layer, the learning error of R was 0.04731176, and TensorFlow was measured at 0.06677193, and when there were two hidden layers, R was measured at 0.04782134 and TensorFlow was measured at 0.05799060. Overall, R was measured to have better performance. We tried to solve the difficulties in tool selection by providing quantitative performance information on the two tools to users who are new to machine learning.

Multi-scale Attention and Deep Ensemble-Based Animal Skin Lesions Classification (다중 스케일 어텐션과 심층 앙상블 기반 동물 피부 병변 분류 기법)

  • Kwak, Min Ho;Kim, Kyeong Tae;Choi, Jae Young
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.8
    • /
    • pp.1212-1223
    • /
    • 2022
  • Skin lesions are common diseases that range from skin rashes to skin cancer, which can lead to death. Note that early diagnosis of skin diseases can be important because early diagnosis of skin diseases considerably can reduce the course of treatment and the harmful effect of the disease. Recently, the development of computer-aided diagnosis (CAD) systems based on artificial intelligence has been actively made for the early diagnosis of skin diseases. In a typical CAD system, the accurate classification of skin lesion types is of great importance for improving the diagnosis performance. Motivated by this, we propose a novel deep ensemble classification with multi-scale attention networks. The proposed deep ensemble networks are jointly trained using a single loss function in an end-to-end manner. In addition, the proposed deep ensemble network is equipped with a multi-scale attention mechanism and segmentation information of the original skin input image, which improves the classification performance. To demonstrate our method, the publicly available human skin disease dataset (HAM 10000) and the private animal skin lesion dataset were used for the evaluation. Experiment results showed that the proposed methods can achieve 97.8% and 81% accuracy on each HAM10000 and animal skin lesion dataset. This research work would be useful for developing a more reliable CAD system which helps doctors early diagnose skin diseases.

Bi-LSTM model with time distribution for bandwidth prediction in mobile networks

  • Hyeonji Lee;Yoohwa Kang;Minju Gwak;Donghyeok An
    • ETRI Journal
    • /
    • v.46 no.2
    • /
    • pp.205-217
    • /
    • 2024
  • We propose a bandwidth prediction approach based on deep learning. The approach is intended to accurately predict the bandwidth of various types of mobile networks. We first use a machine learning technique, namely, the gradient boosting algorithm, to recognize the connected mobile network. Second, we apply a handover detection algorithm based on network recognition to account for vertical handover that causes the bandwidth variance. Third, as the communication performance offered by 3G, 4G, and 5G networks varies, we suggest a bidirectional long short-term memory model with time distribution for bandwidth prediction per network. To increase the prediction accuracy, pretraining and fine-tuning are applied for each type of network. We use a dataset collected at University College Cork for network recognition, handover detection, and bandwidth prediction. The performance evaluation indicates that the handover detection algorithm achieves 88.5% accuracy, and the bandwidth prediction model achieves a high accuracy, with a root-mean-square error of only 2.12%.

Breast Cancer Classification in Ultrasound Images using Semi-supervised method based on Pseudo-labeling

  • Seokmin Han
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.124-131
    • /
    • 2024
  • Breast cancer classification using ultrasound, while widely employed, faces challenges due to its relatively low predictive value arising from significant overlap in characteristics between benign and malignant lesions, as well as operator-dependency. To alleviate these challenges and reduce dependency on radiologist interpretation, the implementation of automatic breast cancer classification in ultrasound image can be helpful. To deal with this problem, we propose a semi-supervised deep learning framework for breast cancer classification. In the proposed method, we could achieve reasonable performance utilizing less than 50% of the training data for supervised learning in comparison to when we utilized a 100% labeled dataset for training. Though it requires more modification, this methodology may be able to alleviate the time-consuming annotation burden on radiologists by reducing the number of annotation, contributing to a more efficient and effective breast cancer detection process in ultrasound images.

Recognition of Dog Breeds based on Deep Learning using a Random-Label and Web Image Mining (웹 이미지 마이닝과 랜덤 레이블을 이용한 딥러닝 기반 개 품종 인식)

  • Kang, Min-Seok;Hong, Kwang-Seok
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.201-202
    • /
    • 2018
  • In this paper, a dog breed image provided by Dataset of existing ImageNet and Oxford-IIIT Pet Image is combined with a dog breed image obtained through data mining on Internet and a random-label is added. this paper introduces to recognize 122 classes of dog breeds and 1 class that is not dog breeds. The recognition rate of dog breeds using both conventional DB and collection DB was improved 1.5% over Top-1 compared to recognition rate of dog breeds using only existing DB. The image recognition rate about non-dog image, was 93% recognition rate in case of 10000 random DBs.

  • PDF