• Title/Summary/Keyword: Information input algorithm

Search Result 2,444, Processing Time 0.028 seconds

A Low Power Resource Allocation and Scheduling Algorithm for High Level Synthesis (상위 레벨 합성을 위한 저 전력 스케줄링 및 자원할당 알고리즘)

  • Sin, Mu-Kyoung;Lin, Chi-Ho
    • The KIPS Transactions:PartA
    • /
    • v.8A no.3
    • /
    • pp.279-286
    • /
    • 2001
  • This paper proposes a low power resource allocation and scheduling algorithm that minimized power consumption such as DSP circuit in high-level synthesis process. In this paper, we have used list-scheduling method for low power design in scheduling step. Also, it increase possibility to reuse input through resource sharing when assign resource. After scheduling, the resources allocation uses the power function in consideration of the result of calculating average hamming distances and switching activity between two input. First, it obtain switching activity about input value after calculate average hamming distances between two operator and find power value make use of bit pattern of the input value. Resource allocation process assign operator to minimize average hamming distance and power dissipation on all occasions which is allocated at each control step according to increase control step. As comparing the existed method, the execution time becomes fast according to number of operator and be most numberous control step. And in case of power that consume, there is decrease effect from 6% to 8% to be small.

  • PDF

Development and application of prediction model of hyperlipidemia using SVM and meta-learning algorithm (SVM과 meta-learning algorithm을 이용한 고지혈증 유병 예측모형 개발과 활용)

  • Lee, Seulki;Shin, Taeksoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.111-124
    • /
    • 2018
  • This study aims to develop a classification model for predicting the occurrence of hyperlipidemia, one of the chronic diseases. Prior studies applying data mining techniques for predicting disease can be classified into a model design study for predicting cardiovascular disease and a study comparing disease prediction research results. In the case of foreign literatures, studies predicting cardiovascular disease were predominant in predicting disease using data mining techniques. Although domestic studies were not much different from those of foreign countries, studies focusing on hypertension and diabetes were mainly conducted. Since hypertension and diabetes as well as chronic diseases, hyperlipidemia, are also of high importance, this study selected hyperlipidemia as the disease to be analyzed. We also developed a model for predicting hyperlipidemia using SVM and meta learning algorithms, which are already known to have excellent predictive power. In order to achieve the purpose of this study, we used data set from Korea Health Panel 2012. The Korean Health Panel produces basic data on the level of health expenditure, health level and health behavior, and has conducted an annual survey since 2008. In this study, 1,088 patients with hyperlipidemia were randomly selected from the hospitalized, outpatient, emergency, and chronic disease data of the Korean Health Panel in 2012, and 1,088 nonpatients were also randomly extracted. A total of 2,176 people were selected for the study. Three methods were used to select input variables for predicting hyperlipidemia. First, stepwise method was performed using logistic regression. Among the 17 variables, the categorical variables(except for length of smoking) are expressed as dummy variables, which are assumed to be separate variables on the basis of the reference group, and these variables were analyzed. Six variables (age, BMI, education level, marital status, smoking status, gender) excluding income level and smoking period were selected based on significance level 0.1. Second, C4.5 as a decision tree algorithm is used. The significant input variables were age, smoking status, and education level. Finally, C4.5 as a decision tree algorithm is used. In SVM, the input variables selected by genetic algorithms consisted of 6 variables such as age, marital status, education level, economic activity, smoking period, and physical activity status, and the input variables selected by genetic algorithms in artificial neural network consist of 3 variables such as age, marital status, and education level. Based on the selected parameters, we compared SVM, meta learning algorithm and other prediction models for hyperlipidemia patients, and compared the classification performances using TP rate and precision. The main results of the analysis are as follows. First, the accuracy of the SVM was 88.4% and the accuracy of the artificial neural network was 86.7%. Second, the accuracy of classification models using the selected input variables through stepwise method was slightly higher than that of classification models using the whole variables. Third, the precision of artificial neural network was higher than that of SVM when only three variables as input variables were selected by decision trees. As a result of classification models based on the input variables selected through the genetic algorithm, classification accuracy of SVM was 88.5% and that of artificial neural network was 87.9%. Finally, this study indicated that stacking as the meta learning algorithm proposed in this study, has the best performance when it uses the predicted outputs of SVM and MLP as input variables of SVM, which is a meta classifier. The purpose of this study was to predict hyperlipidemia, one of the representative chronic diseases. To do this, we used SVM and meta-learning algorithms, which is known to have high accuracy. As a result, the accuracy of classification of hyperlipidemia in the stacking as a meta learner was higher than other meta-learning algorithms. However, the predictive performance of the meta-learning algorithm proposed in this study is the same as that of SVM with the best performance (88.6%) among the single models. The limitations of this study are as follows. First, various variable selection methods were tried, but most variables used in the study were categorical dummy variables. In the case with a large number of categorical variables, the results may be different if continuous variables are used because the model can be better suited to categorical variables such as decision trees than general models such as neural networks. Despite these limitations, this study has significance in predicting hyperlipidemia with hybrid models such as met learning algorithms which have not been studied previously. It can be said that the result of improving the model accuracy by applying various variable selection techniques is meaningful. In addition, it is expected that our proposed model will be effective for the prevention and management of hyperlipidemia.

A FUZZY NEURAL NETWORK-BASED DECISION OF ROAD IMAGE QUALITY FOR THE EXTRACTION OF LANE-RELATED INFORMATION

  • YI U. K.;LEE J. W.;BAEK K. R.
    • International Journal of Automotive Technology
    • /
    • v.6 no.1
    • /
    • pp.53-63
    • /
    • 2005
  • We propose a fuzzy neural network (FNN) theory capable of deciding the quality of a road image prior to extracting lane-related information. The accuracy of lane-related information obtained by image processing depends on the quality of the raw images, which can be classified as good or bad according to how visible the lane marks on the images are. Enhancing the accuracy of the information by an image-processing algorithm is limited due to noise corruption which makes image processing difficult. The FNN, on the other hand, decides whether road images are good or bad with respect to the degree of noise corruption. A cumulative distribution function (CDF), a function of edge histogram, is utilized to extract input parameters from the FNN according to the fact that the shape of the CDF is deeply correlated to the road image quality. A suitability analysis shows that this deep correlation exists between the parameters and the image quality. The input pattern vector of the FNN consists of nine parameters in which eight parameters are from the CDF and one is from the intensity distribution of raw images. Experimental results showed that the proposed FNN system was quite successful. We carried out simulations with real images taken in various lighting and weather conditions, and obtained successful decision-making about $99\%$ of the time.

Discriminant Metric Learning Approach for Face Verification

  • Chen, Ju-Chin;Wu, Pei-Hsun;Lien, Jenn-Jier James
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.2
    • /
    • pp.742-762
    • /
    • 2015
  • In this study, we propose a distance metric learning approach called discriminant metric learning (DML) for face verification, which addresses a binary-class problem for classifying whether or not two input images are of the same subject. The critical issue for solving this problem is determining the method to be used for measuring the distance between two images. Among various methods, the large margin nearest neighbor (LMNN) method is a state-of-the-art algorithm. However, to compensate the LMNN's entangled data distribution due to high levels of appearance variations in unconstrained environments, DML's goal is to penalize violations of the negative pair distance relationship, i.e., the images with different labels, while being integrated with LMNN to model the distance relation between positive pairs, i.e., the images with the same label. The likelihoods of the input images, estimated using DML and LMNN metrics, are then weighted and combined for further analysis. Additionally, rather than using the k-nearest neighbor (k-NN) classification mechanism, we propose a verification mechanism that measures the correlation of the class label distribution of neighbors to reduce the false negative rate of positive pairs. From the experimental results, we see that DML can modify the relation of negative pairs in the original LMNN space and compensate for LMNN's performance on faces with large variances, such as pose and expression.

Multi-Slice Joint Task Offloading and Resource Allocation Scheme for Massive MIMO Enabled Network

  • Yin Ren;Aihuang Guo;Chunlin Song
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.3
    • /
    • pp.794-815
    • /
    • 2023
  • The rapid development of mobile communication not only has made the industry gradually diversified, but also has enhanced the service quality requirements of users. In this regard, it is imperative to consider jointly network slicing and mobile edge computing. The former mainly ensures the requirements of varied vertical services preferably, and the latter solves the conflict between the user's own energy and harsh latency. At present, the integration of the two faces many challenges and need to carry out at different levels. The main target of the paper is to minimize the energy consumption of the system, and introduce a multi-slice joint task offloading and resource allocation scheme for massive multiple input multiple output enabled heterogeneous networks. The problem is formulated by collaborative optimizing offloading ratios, user association, transmission power and resource slicing, while being limited by the dissimilar latency and rate of multi-slice. To solve it, assign the optimal problem to two sub-problems of offloading decision and resource allocation, then solve them separately by exploiting the alternative optimization technique and Karush-Kuhn-Tucker conditions. Finally, a novel slices task offloading and resource allocation algorithm is proposed to get the offloading and resource allocation strategies. Numerous simulation results manifest that the proposed scheme has certain feasibility and effectiveness, and its performance is better than the other baseline scheme.

Privacy-Preserving Deep Learning using Collaborative Learning of Neural Network Model

  • Hye-Kyeong Ko
    • International journal of advanced smart convergence
    • /
    • v.12 no.2
    • /
    • pp.56-66
    • /
    • 2023
  • The goal of deep learning is to extract complex features from multidimensional data use the features to create models that connect input and output. Deep learning is a process of learning nonlinear features and functions from complex data, and the user data that is employed to train deep learning models has become the focus of privacy concerns. Companies that collect user's sensitive personal information, such as users' images and voices, own this data for indefinite period of times. Users cannot delete their personal information, and they cannot limit the purposes for which the data is used. The study has designed a deep learning method that employs privacy protection technology that uses distributed collaborative learning so that multiple participants can use neural network models collaboratively without sharing the input datasets. To prevent direct leaks of personal information, participants are not shown the training datasets during the model training process, unlike traditional deep learning so that the personal information in the data can be protected. The study used a method that can selectively share subsets via an optimization algorithm that is based on modified distributed stochastic gradient descent, and the result showed that it was possible to learn with improved learning accuracy while protecting personal information.

An Adaptive K-best detection algorithm for MIMO systems (다중 송수신 안테나 시스템에서 적응 K-best 검출 알고리즘)

  • Kim, Jong-Wook;Kang, Ji-Won;Lee, Chung-Yong
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.43 no.10 s.352
    • /
    • pp.1-7
    • /
    • 2006
  • Lattice decoding concept has been proposed for the implementation of the Maximum-Likelihood detection which is the optimal receiver from the viewpoint of the BER (Bit Error Rate) performance for MIMO (Multiple Input Multiple Output) systems. Sphere decoding algorithm and K-best decoding algorithm are based on the lattice decoding concept. A K-best decoding algorithm shows a good BER performance with relatively low complexity. However, with small K value, the error propagation effect severely degrades the performance. In this paper, we propose an adaptive K-best decoding algorithm which has lower average complexity and better BER performance than conventional K-best decoding algorithm.

Design of a systolic array for forward-backward propagation of back-propagation algorithm (역전파 알고리즘의 전방향, 역방향 동시 수행을 위한 스스톨릭 배열의 설계)

  • 장명숙;유기영
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.33B no.9
    • /
    • pp.49-61
    • /
    • 1996
  • Back-propagation(BP) algorithm needs a lot of time to train the artificial neural network (ANN) to get high accuracy level in classification tasks. So there have been extensive researches to process back-propagation algorithm on parallel processors. This paper prsents a linear systolic array which calculates forward-backward propagation of BP algorithm at the same time using effective space-time transformation and PE structure. First, we analyze data flow of forwared and backward propagations and then, represent the BP algorithm into data dapendency graph (DG) which shows parallelism inherent in the BP algorithm. Next, apply space-time transformation on the DG of ANN is turn with orthogonal direction projection. By doing so, we can get a snakelike systolic array. Also we calculate the interval of input for parallel processing, calculate the indices to make the right datas be used at the right PE when forward and bvackward propagations are processed in the same PE. And then verify the correctness of output when forward and backward propagations are executed at the same time. By doing so, the proposed system maximizes parallelism of BP algorithm, minimizes th enumber of PEs. And it reduces the execution time by 2 times through making idle PEs participate in forward-backward propagation at the same time.

  • PDF

Behavior Pattern Prediction Algorithm Based on 2D Pose Estimation and LSTM from Videos (비디오 영상에서 2차원 자세 추정과 LSTM 기반의 행동 패턴 예측 알고리즘)

  • Choi, Jiho;Hwang, Gyutae;Lee, Sang Jun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.17 no.4
    • /
    • pp.191-197
    • /
    • 2022
  • This study proposes an image-based Pose Intention Network (PIN) algorithm for rehabilitation via patients' intentions. The purpose of the PIN algorithm is for enabling an active rehabilitation exercise, which is implemented by estimating the patient's motion and classifying the intention. Existing rehabilitation involves the inconvenience of attaching a sensor directly to the patient's skin. In addition, the rehabilitation device moves the patient, which is a passive rehabilitation method. Our algorithm consists of two steps. First, we estimate the user's joint position through the OpenPose algorithm, which is efficient in estimating 2D human pose in an image. Second, an intention classifier is constructed for classifying the motions into three categories, and a sequence of images including joint information is used as input. The intention network also learns correlations between joints and changes in joints over a short period of time, which can be easily used to determine the intention of the motion. To implement the proposed algorithm and conduct real-world experiments, we collected our own dataset, which is composed of videos of three classes. The network is trained using short segment clips of the video. Experimental results demonstrate that the proposed algorithm is effective for classifying intentions based on a short video clip.

Performance Analysis of Correntropy-Based Blind Algorithms Robust to Impulsive Noise (충격성 잡음에 강인한 코렌트로피 기반 블라인드 알고리듬의 성능분석)

  • Kim, Namyong
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.12
    • /
    • pp.2324-2330
    • /
    • 2015
  • In blind signal processing in impulsive noise environment the maximum cross-correntropy (MCC) algorithm shows superior performance compared to MSE-based algorithms. But optimum weight conditions of MCC algorithm and its properties related with robustness to impulsive noise have not been studied sufficiently. In this paper, through the analysis of the behavior of its optimum weight and the relationship with the MSE-based LMS algorithm, it is shown that the optimum weight of MCC and MSE-based LMS have an equal solution. Also the factor that keeps optimum weight of MCC undisturbed and stable under impulsive noise is proven to be the magnitude controlled input through simulation.