• Title/Summary/Keyword: 성능최적화 기법

Search Result 1,330, Processing Time 0.026 seconds

Strawberry Pests and Diseases Detection Technique Optimized for Symptoms Using Deep Learning Algorithm (딥러닝을 이용한 병징에 최적화된 딸기 병충해 검출 기법)

  • Choi, Young-Woo;Kim, Na-eun;Paudel, Bhola;Kim, Hyeon-tae
    • Journal of Bio-Environment Control
    • /
    • v.31 no.3
    • /
    • pp.255-260
    • /
    • 2022
  • This study aimed to develop a service model that uses a deep learning algorithm for detecting diseases and pests in strawberries through image data. In addition, the pest detection performance of deep learning models was further improved by proposing segmented image data sets specialized in disease and pest symptoms. The CNN-based YOLO deep learning model was selected to enhance the existing R-CNN-based model's slow learning speed and inference speed. A general image data set and a proposed segmented image dataset was prepared to train the pest and disease detection model. When the deep learning model was trained with the general training data set, the pest detection rate was 81.35%, and the pest detection reliability was 73.35%. On the other hand, when the deep learning model was trained with the segmented image dataset, the pest detection rate increased to 91.93%, and detection reliability was increased to 83.41%. This study concludes with the possibility of improving the performance of the deep learning model by using a segmented image dataset instead of a general image dataset.

Parallel Processing of Satellite Images using CUDA Library: Focused on NDVI Calculation (CUDA 라이브러리를 이용한 위성영상 병렬처리 : NDVI 연산을 중심으로)

  • LEE, Kang-Hun;JO, Myung-Hee;LEE, Won-Hee
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.19 no.3
    • /
    • pp.29-42
    • /
    • 2016
  • Remote sensing allows acquisition of information across a large area without contacting objects, and has thus been rapidly developed by application to different areas. Thus, with the development of remote sensing, satellites are able to rapidly advance in terms of their image resolution. As a result, satellites that use remote sensing have been applied to conduct research across many areas of the world. However, while research on remote sensing is being implemented across various areas, research on data processing is presently insufficient; that is, as satellite resources are further developed, data processing continues to lag behind. Accordingly, this paper discusses plans to maximize the performance of satellite image processing by utilizing the CUDA(Compute Unified Device Architecture) Library of NVIDIA, a parallel processing technique. The discussion in this paper proceeds as follows. First, standard KOMPSAT(Korea Multi-Purpose Satellite) images of various sizes are subdivided into five types. NDVI(Normalized Difference Vegetation Index) is implemented to the subdivided images. Next, ArcMap and the two techniques, each based on CPU or GPU, are used to implement NDVI. The histograms of each image are then compared after each implementation to analyze the different processing speeds when using CPU and GPU. The results indicate that both the CPU version and GPU version images are equal with the ArcMap images, and after the histogram comparison, the NDVI code was correctly implemented. In terms of the processing speed, GPU showed 5 times faster results than CPU. Accordingly, this research shows that a parallel processing technique using CUDA Library can enhance the data processing speed of satellites images, and that this data processing benefits from multiple advanced remote sensing techniques as compared to a simple pixel computation like NDVI.

Development of deep learning network based low-quality image enhancement techniques for improving foreign object detection performance (이물 객체 탐지 성능 개선을 위한 딥러닝 네트워크 기반 저품질 영상 개선 기법 개발)

  • Ki-Yeol Eom;Byeong-Seok Min
    • Journal of Internet Computing and Services
    • /
    • v.25 no.1
    • /
    • pp.99-107
    • /
    • 2024
  • Along with economic growth and industrial development, there is an increasing demand for various electronic components and device production of semiconductor, SMT component, and electrical battery products. However, these products may contain foreign substances coming from manufacturing process such as iron, aluminum, plastic and so on, which could lead to serious problems or malfunctioning of the product, and fire on the electric vehicle. To solve these problems, it is necessary to determine whether there are foreign materials inside the product, and may tests have been done by means of non-destructive testing methodology such as ultrasound ot X-ray. Nevertheless, there are technical challenges and limitation in acquiring X-ray images and determining the presence of foreign materials. In particular Small-sized or low-density foreign materials may not be visible even when X-ray equipment is used, and noise can also make it difficult to detect foreign objects. Moreover, in order to meet the manufacturing speed requirement, the x-ray acquisition time should be reduced, which can result in the very low signal- to-noise ratio(SNR) lowering the foreign material detection accuracy. Therefore, in this paper, we propose a five-step approach to overcome the limitations of low resolution, which make it challenging to detect foreign substances. Firstly, global contrast of X-ray images are increased through histogram stretching methodology. Second, to strengthen the high frequency signal and local contrast, we applied local contrast enhancement technique. Third, to improve the edge clearness, Unsharp masking is applied to enhance edges, making objects more visible. Forth, the super-resolution method of the Residual Dense Block (RDB) is used for noise reduction and image enhancement. Last, the Yolov5 algorithm is employed to train and detect foreign objects after learning. Using the proposed method in this study, experimental results show an improvement of more than 10% in performance metrics such as precision compared to low-density images.

A 13b 100MS/s 0.70㎟ 45nm CMOS ADC for IF-Domain Signal Processing Systems (IF 대역 신호처리 시스템 응용을 위한 13비트 100MS/s 0.70㎟ 45nm CMOS ADC)

  • Park, Jun-Sang;An, Tai-Ji;Ahn, Gil-Cho;Lee, Mun-Kyo;Go, Min-Ho;Lee, Seung-Hoon
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.3
    • /
    • pp.46-55
    • /
    • 2016
  • This work proposes a 13b 100MS/s 45nm CMOS ADC with a high dynamic performance for IF-domain high-speed signal processing systems based on a four-step pipeline architecture to optimize operating specifications. The SHA employs a wideband high-speed sampling network properly to process high-frequency input signals exceeding a sampling frequency. The SHA and MDACs adopt a two-stage amplifier with a gain-boosting technique to obtain the required high DC gain and the wide signal-swing range, while the amplifier and bias circuits use the same unit-size devices repeatedly to minimize device mismatch. Furthermore, a separate analog power supply voltage for on-chip current and voltage references minimizes performance degradation caused by the undesired noise and interference from adjacent functional blocks during high-speed operation. The proposed ADC occupies an active die area of $0.70mm^2$, based on various process-insensitive layout techniques to minimize the physical process imperfection effects. The prototype ADC in a 45nm CMOS demonstrates a measured DNL and INL within 0.77LSB and 1.57LSB, with a maximum SNDR and SFDR of 64.2dB and 78.4dB at 100MS/s, respectively. The ADC is implemented with long-channel devices rather than minimum channel-length devices available in this CMOS technology to process a wide input range of $2.0V_{PP}$ for the required system and to obtain a high dynamic performance at IF-domain input signal bands. The ADC consumes 425.0mW with a single analog voltage of 2.5V and two digital voltages of 2.5V and 1.1V.

A Recidivism Prediction Model Based on XGBoost Considering Asymmetric Error Costs (비대칭 오류 비용을 고려한 XGBoost 기반 재범 예측 모델)

  • Won, Ha-Ram;Shim, Jae-Seung;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.127-137
    • /
    • 2019
  • Recidivism prediction has been a subject of constant research by experts since the early 1970s. But it has become more important as committed crimes by recidivist steadily increase. Especially, in the 1990s, after the US and Canada adopted the 'Recidivism Risk Assessment Report' as a decisive criterion during trial and parole screening, research on recidivism prediction became more active. And in the same period, empirical studies on 'Recidivism Factors' were started even at Korea. Even though most recidivism prediction studies have so far focused on factors of recidivism or the accuracy of recidivism prediction, it is important to minimize the prediction misclassification cost, because recidivism prediction has an asymmetric error cost structure. In general, the cost of misrecognizing people who do not cause recidivism to cause recidivism is lower than the cost of incorrectly classifying people who would cause recidivism. Because the former increases only the additional monitoring costs, while the latter increases the amount of social, and economic costs. Therefore, in this paper, we propose an XGBoost(eXtream Gradient Boosting; XGB) based recidivism prediction model considering asymmetric error cost. In the first step of the model, XGB, being recognized as high performance ensemble method in the field of data mining, was applied. And the results of XGB were compared with various prediction models such as LOGIT(logistic regression analysis), DT(decision trees), ANN(artificial neural networks), and SVM(support vector machines). In the next step, the threshold is optimized to minimize the total misclassification cost, which is the weighted average of FNE(False Negative Error) and FPE(False Positive Error). To verify the usefulness of the model, the model was applied to a real recidivism prediction dataset. As a result, it was confirmed that the XGB model not only showed better prediction accuracy than other prediction models but also reduced the cost of misclassification most effectively.

Self-optimizing feature selection algorithm for enhancing campaign effectiveness (캠페인 효과 제고를 위한 자기 최적화 변수 선택 알고리즘)

  • Seo, Jeoung-soo;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.173-198
    • /
    • 2020
  • For a long time, many studies have been conducted on predicting the success of campaigns for customers in academia, and prediction models applying various techniques are still being studied. Recently, as campaign channels have been expanded in various ways due to the rapid revitalization of online, various types of campaigns are being carried out by companies at a level that cannot be compared to the past. However, customers tend to perceive it as spam as the fatigue of campaigns due to duplicate exposure increases. Also, from a corporate standpoint, there is a problem that the effectiveness of the campaign itself is decreasing, such as increasing the cost of investing in the campaign, which leads to the low actual campaign success rate. Accordingly, various studies are ongoing to improve the effectiveness of the campaign in practice. This campaign system has the ultimate purpose to increase the success rate of various campaigns by collecting and analyzing various data related to customers and using them for campaigns. In particular, recent attempts to make various predictions related to the response of campaigns using machine learning have been made. It is very important to select appropriate features due to the various features of campaign data. If all of the input data are used in the process of classifying a large amount of data, it takes a lot of learning time as the classification class expands, so the minimum input data set must be extracted and used from the entire data. In addition, when a trained model is generated by using too many features, prediction accuracy may be degraded due to overfitting or correlation between features. Therefore, in order to improve accuracy, a feature selection technique that removes features close to noise should be applied, and feature selection is a necessary process in order to analyze a high-dimensional data set. Among the greedy algorithms, SFS (Sequential Forward Selection), SBS (Sequential Backward Selection), SFFS (Sequential Floating Forward Selection), etc. are widely used as traditional feature selection techniques. It is also true that if there are many risks and many features, there is a limitation in that the performance for classification prediction is poor and it takes a lot of learning time. Therefore, in this study, we propose an improved feature selection algorithm to enhance the effectiveness of the existing campaign. The purpose of this study is to improve the existing SFFS sequential method in the process of searching for feature subsets that are the basis for improving machine learning model performance using statistical characteristics of the data to be processed in the campaign system. Through this, features that have a lot of influence on performance are first derived, features that have a negative effect are removed, and then the sequential method is applied to increase the efficiency for search performance and to apply an improved algorithm to enable generalized prediction. Through this, it was confirmed that the proposed model showed better search and prediction performance than the traditional greed algorithm. Compared with the original data set, greed algorithm, genetic algorithm (GA), and recursive feature elimination (RFE), the campaign success prediction was higher. In addition, when performing campaign success prediction, the improved feature selection algorithm was found to be helpful in analyzing and interpreting the prediction results by providing the importance of the derived features. This is important features such as age, customer rating, and sales, which were previously known statistically. Unlike the previous campaign planners, features such as the combined product name, average 3-month data consumption rate, and the last 3-month wireless data usage were unexpectedly selected as important features for the campaign response, which they rarely used to select campaign targets. It was confirmed that base attributes can also be very important features depending on the type of campaign. Through this, it is possible to analyze and understand the important characteristics of each campaign type.

Optimization of Multiclass Support Vector Machine using Genetic Algorithm: Application to the Prediction of Corporate Credit Rating (유전자 알고리즘을 이용한 다분류 SVM의 최적화: 기업신용등급 예측에의 응용)

  • Ahn, Hyunchul
    • Information Systems Review
    • /
    • v.16 no.3
    • /
    • pp.161-177
    • /
    • 2014
  • Corporate credit rating assessment consists of complicated processes in which various factors describing a company are taken into consideration. Such assessment is known to be very expensive since domain experts should be employed to assess the ratings. As a result, the data-driven corporate credit rating prediction using statistical and artificial intelligence (AI) techniques has received considerable attention from researchers and practitioners. In particular, statistical methods such as multiple discriminant analysis (MDA) and multinomial logistic regression analysis (MLOGIT), and AI methods including case-based reasoning (CBR), artificial neural network (ANN), and multiclass support vector machine (MSVM) have been applied to corporate credit rating.2) Among them, MSVM has recently become popular because of its robustness and high prediction accuracy. In this study, we propose a novel optimized MSVM model, and appy it to corporate credit rating prediction in order to enhance the accuracy. Our model, named 'GAMSVM (Genetic Algorithm-optimized Multiclass Support Vector Machine),' is designed to simultaneously optimize the kernel parameters and the feature subset selection. Prior studies like Lorena and de Carvalho (2008), and Chatterjee (2013) show that proper kernel parameters may improve the performance of MSVMs. Also, the results from the studies such as Shieh and Yang (2008) and Chatterjee (2013) imply that appropriate feature selection may lead to higher prediction accuracy. Based on these prior studies, we propose to apply GAMSVM to corporate credit rating prediction. As a tool for optimizing the kernel parameters and the feature subset selection, we suggest genetic algorithm (GA). GA is known as an efficient and effective search method that attempts to simulate the biological evolution phenomenon. By applying genetic operations such as selection, crossover, and mutation, it is designed to gradually improve the search results. Especially, mutation operator prevents GA from falling into the local optima, thus we can find the globally optimal or near-optimal solution using it. GA has popularly been applied to search optimal parameters or feature subset selections of AI techniques including MSVM. With these reasons, we also adopt GA as an optimization tool. To empirically validate the usefulness of GAMSVM, we applied it to a real-world case of credit rating in Korea. Our application is in bond rating, which is the most frequently studied area of credit rating for specific debt issues or other financial obligations. The experimental dataset was collected from a large credit rating company in South Korea. It contained 39 financial ratios of 1,295 companies in the manufacturing industry, and their credit ratings. Using various statistical methods including the one-way ANOVA and the stepwise MDA, we selected 14 financial ratios as the candidate independent variables. The dependent variable, i.e. credit rating, was labeled as four classes: 1(A1); 2(A2); 3(A3); 4(B and C). 80 percent of total data for each class was used for training, and remaining 20 percent was used for validation. And, to overcome small sample size, we applied five-fold cross validation to our dataset. In order to examine the competitiveness of the proposed model, we also experimented several comparative models including MDA, MLOGIT, CBR, ANN and MSVM. In case of MSVM, we adopted One-Against-One (OAO) and DAGSVM (Directed Acyclic Graph SVM) approaches because they are known to be the most accurate approaches among various MSVM approaches. GAMSVM was implemented using LIBSVM-an open-source software, and Evolver 5.5-a commercial software enables GA. Other comparative models were experimented using various statistical and AI packages such as SPSS for Windows, Neuroshell, and Microsoft Excel VBA (Visual Basic for Applications). Experimental results showed that the proposed model-GAMSVM-outperformed all the competitive models. In addition, the model was found to use less independent variables, but to show higher accuracy. In our experiments, five variables such as X7 (total debt), X9 (sales per employee), X13 (years after founded), X15 (accumulated earning to total asset), and X39 (the index related to the cash flows from operating activity) were found to be the most important factors in predicting the corporate credit ratings. However, the values of the finally selected kernel parameters were found to be almost same among the data subsets. To examine whether the predictive performance of GAMSVM was significantly greater than those of other models, we used the McNemar test. As a result, we found that GAMSVM was better than MDA, MLOGIT, CBR, and ANN at the 1% significance level, and better than OAO and DAGSVM at the 5% significance level.

Optimized Mix Proportioning of Steel and Hybrid Reinforced Concrete Using Harmony Search Algorithm (화음탐색법을 이용한 강섬유 및 하이브리드 섬유보강 콘크리트의 최적배합 설계)

  • Lee, Chi-Hoon;Lee, Joo-Ha;Yoon, Young-Soo
    • Journal of the Korea Concrete Institute
    • /
    • v.18 no.2 s.92
    • /
    • pp.151-159
    • /
    • 2006
  • The guide line of the SFRC mix design was not established, and the convenience of the practical application on the spot is not so good. In this paper, hence, the program which is optimized to result the mix proportion by the flexural strength and toughness, was developed to apply to SFRC on the practical spot. This program could minimize the number of trial mixes and get an economical and appropriate mixture. In addition, the theoretical background on which the program is based, will be the basis of the embodied method to mixing SFRC. Additionally, new algorithm, in this paper, was used to develop the mix proportioning program of SFRC. The new algorithm is the Harmony Search which is the heuristic method mimicking the improvisation of music players, Musical performances seek a best state determined by aesthetic estimation, as the optimization algorithms seek a best state determined by objected function value. And, it was developed the program about single fiber reinforced concrete, beside to the hybrid fiber reinforced concrete that two kinds of steel fibers, which have the different geometry, was reinforced. This will be able to keep the world trend to study, hence, offers the basis of the next research about hybrid fiber reinforced concrete.

Development of Communication Emulate Technique in control system for Automatic Machine. (자동화 기기를 위한 제어 시스템에서의 통신 Emulate 개발)

  • 이범석;정화영
    • Proceedings of the Korea Society for Industrial Systems Conference
    • /
    • 2000.05a
    • /
    • pp.101-106
    • /
    • 2000
  • 자동화 기기 분야에서 컴퓨터의 적용 및 응용은 하드웨어 발달에 따라 매우 빠르고 민감하게 반영되어왔다. 이는 컴퓨터 하드웨어의 빠른 개발 주기에 맞춰 저 가격, 고 효율성, 높은 신뢰성, 호환성 등의 장점을 가진 PC가 현대의 컴퓨터 흐름을 주도하게 되면서 자동화 산업분야 또한 이를 적용하여 왔기 때문이다. 이에 따라, 자동화 기기 분야에서는 고 가격, 긴 개발기간 등을 필요로 했던 과거와 달리 저 가격, 짧은 개발기간, 다양한 개발환경 등을 이룰 수 있었다. 또한, 생산량 증가에만 의존하던 과거와 달리 현대에 이르러서는 시스템의 최적화, 효율의 극대화, 시스템의 안정성, 운용의 편리성, 호환성 등의 개념들이 도입되고 있는 것이다. 자동화 기기를 구성하는 요인으로는 크게 시스템의 틀을 이루는 기계부분과 이를 제어하는 제어 시스템부로 나뉠 수 있다. 제어 시스템에서는 기계부분의 동작을 제어하는 동작 제어부와 이에 관한 정보를 화면에 나타내는 GUI(Graphical User Interface)부분으로 나뉘게된다. 현재에는 이를 통합하여 하나의 하드웨어에서 제어부와 GUI를 모두 담당하는 방법이 연구 진행되고 있으나, 하드웨어를 둘로 나누거나 하나로 하여도 제어부와 GUI 사이의 통신부분은 빼놓을 수 없는 요소가 된다. 따라서, 본 논문에서는 시스템의 안정성을 위하여 두 시스템간에 송·수신되는 데이터를 추적할 수 있도록 하는 Emulate 기법을 구현 및 개발하고자 한다. 이는, 두 시스템간의 통신 데이터를 실시간으로 누적, 저장하여 사용자로 하여금 시스템의 운용상태를 분석할 수 있게 하였으며, 시스템 오류발생 시 Emulate 자료를 근거로 시스템의 운용상태를 파악할 수 있게 하였다.근 제한기능을 제공하며 각 클라이언트와 서버간의 실시간 연결 혹은 지연연결을 지원하는 독립적인 애플리케이션이다. 이러한 처방전달 메시징시스템을 구성하는 각 요소에 대해 정의하고 개념적 모델을 설계하고자 한다.에게 청구되며, 소비자에게 전송 되는 청구서는 사용자DB를 참조하여 사용자가 미리 정의한 원하는 형태로 변환되어 전달되며, 필요시 암호화 과정을 거치는 것이 가능해야 한다. 전송된 청구서는 전자우편의 경우, 암호해독이 가능한 전용 브라우저를 통해 열람 되며, 이는 다시 전용 브라우저를 통해 지불인증이 승인되어 청구 제시서버에게 전송된다. EBPP 시스템의 제어 흐름은 크게 기업이 청구 정보를 소비자에게 제시하는 흐름과 소비자의 지불 승인으로 인해 기업이 은행에 지불을 요구하는 흐름으로 구분할 수 있다. 본 논문에서는 통합 청구서버 및 정구 제시서버의 역할 및 구성 요소들에 대해 서술하고, EBPP 시스템과 연동하여야 하는 메일 서버와의 상호 작용에 대해 서술할 것이다. 본 시스템을 아직 구현이 되지 않은 관계로 시스템의 성능 등의 수치적 결과를 제시할 수 없는 상태다., 취약계층을 위한 일차의료, 의약관리), ${\circled}2$ 보건소 조직 개편 및 민간의료기관과 협력체계 확립, ${\circled}3$ 전문인력 확보 및 인력구성 조정, 그리고 ${\circled}4$ 방문보건사업의 강화 등이다., 대사(代謝)와 관계(關係)있음을 시사(示唆)해 주고 있다.ble nutrient (TDN) was highest in booting stage (59.7%); however no significant difference was foun

  • PDF

Optimal Design of Multi-Plate Clutch Featuring MR Fluid (MR 유체를 적용한 Multi-Plate Clutch의 최적설계)

  • Park, Jin-Young;Kim, Young-Choon;Oh, Jong-Seok;Jeon, Jae-Hoon;Jeong, Jun-Hong
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.5
    • /
    • pp.77-83
    • /
    • 2020
  • 4WD technology is being actively applied to passenger cars. Therefore, dry multi-plate clutches are used for transfer cases. On the other hand, dry clutches have problems related to large vibrations and poor ride quality. To solve this problem, this paper proposes a multi-plate clutch with an MR fluid. When fastening the multi-plate clutch in the transfer case, the proposed MR clutch was applied to reduce the shock and friction, which is a key component in a four-wheel-drive system. MR multi-plate clutch has a fluid coupling mode and a compression mode. A torque model equation was derived for the optimal design. The analysis was performed using Ansys Maxwell to optimize the design parameters of the multi-plate clutch. Electromagnetic field analysis confirmed the strength of the magnetic field when the number of disks and plates were changed, and the maximum strength of the magnetic field was 0.45 Tesla. By applying this to the torque equation, the spacing between the plates was 2 mm, and the inner and outer diameters of the plates were selected to be 45 mm and 55 mm, respectively. Overall, this paper proposes an optimal design technique to maximize the performance of an MR multi-plate clutch.