• Title/Summary/Keyword: LE(Learning Evaluation

Search Result 15, Processing Time 0.027 seconds

Design and Implementation of Teacher Supporting Component on Web (웹 기반 교육 시스템에서 교수지원 컴포넌트의 구현)

  • Haeng-Kon Kim;Jeon-Geun Kang
    • Journal of the Korea Computer Industry Society
    • /
    • v.2 no.9
    • /
    • pp.1139-1146
    • /
    • 2001
  • The value of Web in information societyis increasing in the field of education. Web Based Instruction (WBI) has an unlimited possibility of access the information. It also overcome the constraint of time and space more than the previous class learning method. The previous WBI system informed the results of study to learners after a professor's lecture and test was given. The uniform test didn't make the professor recognize the learner's evaluation according to the level of the learner and it couldn't be a reason to change a teaching method. In this paper, we discuss the Learning Evaluation(LE) component which can support a teaching method to teachers. We suggest LE component for supporting teachers as suggesting visually the test result of pre-test by step and post-test with several ways after analyzing, designing and realizing the evaluation of the test based on CDP(Component Development Process).

  • PDF

Controller Learning Method of Self-driving Bicycle Using State-of-the-art Deep Reinforcement Learning Algorithms

  • Choi, Seung-Yoon;Le, Tuyen Pham;Chung, Tae-Choong
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.10
    • /
    • pp.23-31
    • /
    • 2018
  • Recently, there have been many studies on machine learning. Among them, studies on reinforcement learning are actively worked. In this study, we propose a controller to control bicycle using DDPG (Deep Deterministic Policy Gradient) algorithm which is the latest deep reinforcement learning method. In this paper, we redefine the compensation function of bicycle dynamics and neural network to learn agents. When using the proposed method for data learning and control, it is possible to perform the function of not allowing the bicycle to fall over and reach the further given destination unlike the existing method. For the performance evaluation, we have experimented that the proposed algorithm works in various environments such as fixed speed, random, target point, and not determined. Finally, as a result, it is confirmed that the proposed algorithm shows better performance than the conventional neural network algorithms NAF and PPO.

A STUDY OF USING CKKS HOMOMORPHIC ENCRYPTION OVER THE LAYERS OF A CONVOLUTIONAL NEURAL NETWORK MODEL

  • Castaneda, Sebastian Soler;Nam, Kevin;Joo, Youyeon;Paek, Yunheung
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.161-164
    • /
    • 2022
  • Homomorphic Encryption (HE) schemes have been recently growing as a reliable solution to preserve users' information owe to maintaining and operating the user data in the encrypted state. In addition to that, several Neural Networks models merged with HE schemes have been developed as a prospective tool for privacy-preserving machine learning. Those mentioned works demonstrated that it is possible to match the accuracy of non-encrypted models but there is always a trade-off in the computation time. In this work, we evaluate the implementation of CKKS HE scheme operations over the layers of a LeNet5 convolutional inference model, however, owing to the limitations of the evaluation environment, the scope of this work is not to develop a complete LeNet5 encrypted model. The evaluation was performed using the MNIST dataset with Microsoft SEAL (MSEAL) open-source homomorphic encryption library ported version on Python (PyFhel). The behavior of the encrypted model, the limitations faced and a small description of related and future work is also provided.

Analysis of bias correction performance of satellite-derived precipitation products by deep learning model

  • Le, Xuan-Hien;Nguyen, Giang V.;Jung, Sungho;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.148-148
    • /
    • 2022
  • Spatiotemporal precipitation data is one of the primary quantities in hydrological as well as climatological studies. Despite the fact that the estimation of these data has made considerable progress owing to advances in remote sensing, the discrepancy between satellite-derived precipitation product (SPP) data and observed data is still remarkable. This study aims to propose an effective deep learning model (DLM) for bias correction of SPPs. In which TRMM (The Tropical Rainfall Measuring Mission), CMORPH (CPC Morphing technique), and PERSIANN-CDR (Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks) are three SPPs with a spatial resolution of 0.25o exploited for bias correction, and APHRODITE (Asian Precipitation - Highly-Resolved Observational Data Integration Towards Evaluation) data is used as a benchmark to evaluate the effectiveness of DLM. We selected the Mekong River Basin as a case study area because it is one of the largest watersheds in the world and spans many countries. The adjusted dataset has demonstrated an impressive performance of DLM in bias correction of SPPs in terms of both spatial and temporal evaluation. The findings of this study indicate that DLM can generate reliable estimates for the gridded satellite-based precipitation bias correction.

  • PDF

Evaluation performance of machine learning in merging multiple satellite-based precipitation with gauge observation data

  • Nhuyen, Giang V.;Le, Xuan-hien;Jung, Sungho;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.143-143
    • /
    • 2022
  • Precipitation plays an essential role in water resources management and disaster prevention. Therefore, the understanding related to spatiotemporal characteristics of rainfall is necessary. Nowadays, highly accurate precipitation is mainly obtained from gauge observation systems. However, the density of gauge stations is a sparse and uneven distribution in mountainous areas. With the proliferation of technology, satellite-based precipitation sources are becoming increasingly common and can provide rainfall information in regions with complex topography. Nevertheless, satellite-based data is that it still remains uncertain. To overcome the above limitation, this study aims to take the strengthens of machine learning to generate a new reanalysis of precipitation data by fusion of multiple satellite precipitation products (SPPs) with gauge observation data. Several machine learning algorithms (i.e., Random Forest, Support Vector Regression, and Artificial Neural Network) have been adopted. To investigate the robustness of the new reanalysis product, observed data were collected to evaluate the accuracy of the products through Kling-Gupta efficiency (KGE), probability of detection (POD), false alarm rate (FAR), and critical success index (CSI). As a result, the new precipitation generated through the machine learning model showed higher accuracy than original satellite rainfall products, and its spatiotemporal variability was better reflected than others. Thus, reanalysis of satellite precipitation product based on machine learning can be useful source input data for hydrological simulations in ungauged river basins.

  • PDF

A review of Chinese named entity recognition

  • Cheng, Jieren;Liu, Jingxin;Xu, Xinbin;Xia, Dongwan;Liu, Le;Sheng, Victor S.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.6
    • /
    • pp.2012-2030
    • /
    • 2021
  • Named Entity Recognition (NER) is used to identify entity nouns in the corpus such as Location, Person and Organization, etc. NER is also an important basic of research in various natural language fields. The processing of Chinese NER has some unique difficulties, for example, there is no obvious segmentation boundary between each Chinese character in a Chinese sentence. The Chinese NER task is often combined with Chinese word segmentation, and so on. In response to these problems, we summarize the recognition methods of Chinese NER. In this review, we first introduce the sequence labeling system and evaluation metrics of NER. Then, we divide Chinese NER methods into rule-based methods, statistics-based machine learning methods and deep learning-based methods. Subsequently, we analyze in detail the model framework based on deep learning and the typical Chinese NER methods. Finally, we put forward the current challenges and future research directions of Chinese NER technology.

Detection of Anomaly Lung Sound using Deep Temporal Feature Extraction (깊은 시계열 특성 추출을 이용한 폐 음성 이상 탐지)

  • Kim-Ngoc T. Le;Gyurin Byun;Hyunseung Choo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.11a
    • /
    • pp.605-607
    • /
    • 2023
  • Recent research has highlighted the effectiveness of Deep Learning (DL) techniques in automating the detection of lung sound anomalies. However, the available lung sound datasets often suffer from limitations in both size and balance, prompting DL methods to employ data preprocessing such as augmentation and transfer learning techniques. These strategies, while valuable, contribute to the increased complexity of DL models and necessitate substantial training memory. In this study, we proposed a streamlined and lightweight DL method but effectively detects lung sound anomalies from small and imbalanced dataset. The utilization of 1D dilated convolutional neural networks enhances sensitivity to lung sound anomalies by efficiently capturing deep temporal features and small variations. We conducted a comprehensive evaluation of the ICBHI dataset and achieved a notable improvement over state-of-the-art results, increasing the average score of sensitivity and specificity metrics by 2.7%.

Performance Evaluation of Efficient Vision Transformers on Embedded Edge Platforms (임베디드 엣지 플랫폼에서의 경량 비전 트랜스포머 성능 평가)

  • Minha Lee;Seongjae Lee;Taehyoun Kim
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.3
    • /
    • pp.89-100
    • /
    • 2023
  • Recently, on-device artificial intelligence (AI) solutions using mobile devices and embedded edge devices have emerged in various fields, such as computer vision, to address network traffic burdens, low-energy operations, and security problems. Although vision transformer deep learning models have outperformed conventional convolutional neural network (CNN) models in computer vision, they require more computations and parameters than CNN models. Thus, they are not directly applicable to embedded edge devices with limited hardware resources. Many researchers have proposed various model compression methods or lightweight architectures for vision transformers; however, there are only a few studies evaluating the effects of model compression techniques of vision transformers on performance. Regarding this problem, this paper presents a performance evaluation of vision transformers on embedded platforms. We investigated the behaviors of three vision transformers: DeiT, LeViT, and MobileViT. Each model performance was evaluated by accuracy and inference time on edge devices using the ImageNet dataset. We assessed the effects of the quantization method applied to the models on latency enhancement and accuracy degradation by profiling the proportion of response time occupied by major operations. In addition, we evaluated the performance of each model on GPU and EdgeTPU-based edge devices. In our experimental results, LeViT showed the best performance in CPU-based edge devices, and DeiT-small showed the highest performance improvement in GPU-based edge devices. In addition, only MobileViT models showed performance improvement on EdgeTPU. Summarizing the analysis results through profiling, the degree of performance improvement of each vision transformer model was highly dependent on the proportion of parts that could be optimized in the target edge device. In summary, to apply vision transformers to on-device AI solutions, either proper operation composition and optimizations specific to target edge devices must be considered.

Performance Enhancement and Evaluation of a Deep Learning Framework on Embedded Systems using Unified Memory (통합메모리를 이용한 임베디드 환경에서의 딥러닝 프레임워크 성능 개선과 평가)

  • Lee, Minhak;Kang, Woochul
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.7
    • /
    • pp.417-423
    • /
    • 2017
  • Recently, many embedded devices that have the computing capability required for deep learning have become available; hence, many new applications using these devices are emerging. However, these embedded devices have an architecture different from that of PCs and high-performance servers. In this paper, we propose a method that improves the performance of deep-learning framework by considering the architecture of an embedded device that shares memory between the CPU and the GPU. The proposed method is implemented in Caffe, an open-source deep-learning framework, and is evaluated on an NVIDIA Jetson TK1 embedded device. In the experiment, we investigate the image recognition performance of several state-of-the-art deep-learning networks, including AlexNet, VGGNet, and GoogLeNet. Our results show that the proposed method can achieve significant performance gain. For instance, in AlexNet, we could reduce image recognition latency by about 33% and energy consumption by about 50%.

Evaluation of Recurrent Neural Network Variants for Person Re-identification

  • Le, Cuong Vo;Tuan, Nghia Nguyen;Hong, Quan Nguyen;Lee, Hyuk-Jae
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.6 no.3
    • /
    • pp.193-199
    • /
    • 2017
  • Instead of using only spatial features from a single frame for person re-identification, a combination of spatial and temporal factors boosts the performance of the system. A recurrent neural network (RNN) shows its effectiveness in generating highly discriminative sequence-level human representations. In this work, we implement RNN, three Long Short Term Memory (LSTM) network variants, and Gated Recurrent Unit (GRU) on Caffe deep learning framework, and we then conduct experiments to compare performance in terms of size and accuracy for person re-identification. We propose using GRU for the optimized choice as the experimental results show that the GRU achieves the highest accuracy despite having fewer parameters than the others.