• Title/Summary/Keyword: Minimize total error

Search Result 76, Processing Time 0.027 seconds

A Recidivism Prediction Model Based on XGBoost Considering Asymmetric Error Costs (비대칭 오류 비용을 고려한 XGBoost 기반 재범 예측 모델)

  • Won, Ha-Ram;Shim, Jae-Seung;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.127-137
    • /
    • 2019
  • Recidivism prediction has been a subject of constant research by experts since the early 1970s. But it has become more important as committed crimes by recidivist steadily increase. Especially, in the 1990s, after the US and Canada adopted the 'Recidivism Risk Assessment Report' as a decisive criterion during trial and parole screening, research on recidivism prediction became more active. And in the same period, empirical studies on 'Recidivism Factors' were started even at Korea. Even though most recidivism prediction studies have so far focused on factors of recidivism or the accuracy of recidivism prediction, it is important to minimize the prediction misclassification cost, because recidivism prediction has an asymmetric error cost structure. In general, the cost of misrecognizing people who do not cause recidivism to cause recidivism is lower than the cost of incorrectly classifying people who would cause recidivism. Because the former increases only the additional monitoring costs, while the latter increases the amount of social, and economic costs. Therefore, in this paper, we propose an XGBoost(eXtream Gradient Boosting; XGB) based recidivism prediction model considering asymmetric error cost. In the first step of the model, XGB, being recognized as high performance ensemble method in the field of data mining, was applied. And the results of XGB were compared with various prediction models such as LOGIT(logistic regression analysis), DT(decision trees), ANN(artificial neural networks), and SVM(support vector machines). In the next step, the threshold is optimized to minimize the total misclassification cost, which is the weighted average of FNE(False Negative Error) and FPE(False Positive Error). To verify the usefulness of the model, the model was applied to a real recidivism prediction dataset. As a result, it was confirmed that the XGB model not only showed better prediction accuracy than other prediction models but also reduced the cost of misclassification most effectively.

Identification of Factors Affecting Errors of Velocity Calculation on Application of MLSPIV and Analysys of its Errors through Labortory Experiment (MLSPIV를 이용한 유속산정시 오차요인 규명 및 실내실험을 통한 유속산정오차 분석)

  • Kim, Young-Sung;Lee, Hyun-Seok
    • Journal of Korea Water Resources Association
    • /
    • v.43 no.2
    • /
    • pp.153-165
    • /
    • 2010
  • Large-Scale Particle Image Velocimetry (LSPIV) is an extension of particle image velocimetry (PIV) for measurement of flows spanning large areas in laboratory or field conditions. LSPIV is composed of six elements - seeding, illumination, recording, image transformation, image processing, postprocessing - based on PIV. Possible error elements at each step of Mobile LSPIV (MLSPIV), which is a mobile version of LSPIV, in field applications are identified and summarized the effect of the errors which were quantified in the previous studies. The total number of elemental errors is 27, and five error sources were evaluated previously, seven elemental errors are not effective to the current MLSPIV system. Among 15 elemental errors, four errors - sampling time, image resolution, tracer, and wind - are investigated through an experiment at a laboratory to figure out how those errors affect to velocity calculation. The analysis to figure out the effect of the number of images used for image processing on the velocity calculation error shows that if over 50 images or more are used, the error due to it goes below 1 %. The effect of the image resolution on velocity calculation was investigated through various image resolution using digital camera. Low resolution image set made 3 % of velocity calculation error comparing with high resolution image set as a reference. For the effect of tracers and wind, the wind effect on tracer is decreasing remarkably with increasing the flume bulk velocity. To minimize the velocity evaluation error due to wind, tracers with high specific gravity is favorable.

Tunnel Reverse Engineering Using Terrestrial LiDAR (지상LiDAR를 이용한 터널의 Reverse Engineering)

  • Cho, Hyung Sig;Sohn, Hong Gyoo;Kim, Jong Suk;Lee, Suk Kun
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.28 no.6D
    • /
    • pp.931-936
    • /
    • 2008
  • Surveying by using terrestrial LiDAR(Light Detection And Ranging) is more rapid than by using total station which enables tunnel section profile surveying to be done in suitable time and minimize centerline error, occurrence of overcut and undercut. Therefore, utilization of terrestrial LiDAR has increased more and more in section profile survey and measurement field Moreover, studies of terrestrial LiDAR for accurate and efficient utilization is now ongoing vigorously. Average end area formula, which was generally used to calculate overcut and undercut, was compared with existing methods such as total station survey and photogrammetry. However, there are no criteria of spacing distance for calculating overcut and undercut through terrestrial LiDAR surveying which can acquire 3D information of whole tunnel. This research performed reverse engineering to decide optimal spacing distance when surveying tunnel section profile by comparing whole tunnel volume and tunnel volume in difference spacing distance. This result was utilized to produce CAD drawing for the test tunnel site where there is no design drawings. In addition to this, efficiency of LiDAR and accuracy of CAD drawing was compared with targetless total station surveying of tunnel section profile. Finally, error analysis of target coordinate's accuracy and incidence angle was done in order to verify the accuracy of terrestrial LiDAR technology.

Development of the Forecasting Model for Parts in an Automobile (자동차 부품 수요의 예측 모형 개발)

  • Hong, Jung-Sik;Ahn, Jae-Kyung;Hong, Suk-Kee
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.27 no.3
    • /
    • pp.233-238
    • /
    • 2001
  • This paper deals with demand forecasting of parts in an automobile model which has been extinct. It is important to estimate how much inventory of each part in the extinct model should be stocked because production lines of some parts may be replaced by new ones although there is still demands for the model. Furthermore, in some countries, there is a strong regulation that the automobile manufacturing company should provide customers with auto parts for several years whenever they are requested. The major characteristic of automobile parts demand forecasting is that there exists a close correlation between the number of running cars and the demand of each part. In this sense, the total demand of each part in a year is determined by two factors, the total number of running cars in that year and the failure rate of the part. The total number of running cars in year k can be estimated sequentially by the amount of shipped cars and proportion of discarded cars in years 1, 2,$\cdots$, i. However, it is very difficult to estimate the failure rate of each part because available inter-failure time data is not complete. The failure rate is, therefore, determined so as to minimize the mean squared error between the estimated demand and the observed demand of a part in years 1, 2,$\cdots$, i. In this paper, data obtained from a Korean automobile manufacturing company are used to illustrate our model.

  • PDF

DEVELOPMENT OF A CRYOGENIC TESTING SYSTEM FOR MID-INFRARED DETECTORS ON SPICA

  • Nishiyama, Miho;Kaneda, Hidehiro;Ishihara, Daisuke;Oseki, Shinji;Takeuchi, Nami;Nagayama, Takahiro;Wada, Takehiko
    • Publications of The Korean Astronomical Society
    • /
    • v.32 no.1
    • /
    • pp.355-357
    • /
    • 2017
  • For future space IR missions, such as SPICA, it is crucial to establish an experimental method for evaluating the performance of mid-IR detectors. In particular, the wavelength dependence of the sensitivity is important but difficult to be measured properly. We are now preparing a testing system for mid-IR Si:As/Si:Sb detectors on SPICA. We have designed a cryogenic optical system in which IR signal light from a pinhole is collimated, passed through an optical filter, and focused onto a detector. With this system, we can measure the photoresponse of the detector for various IR light using optical filters with different wavelength properties. We have fabricated aluminum mirrors which are adopted to minimize thermal distortion effects and evaluated the surface figure errors. The total wavefront error of the optical system is $1.3{\mu}m$ RMS, which is small enough for the target wavelengths ($20-37{\mu}m$) of SPICA. The point spread function measured at a room temperature is consistent with that predicted by the simulation. We report the optical performance of the system at cryogenic temperatures.

Error Forecasting & Optimal Stopping Rule under Decreasing Failure Rate (감소(減少)하는 고장률(故障率)하에서 오류예측 및 테스트 시간(時間)의 최적화(最適化)에 관한 연구(硏究))

  • Choe, Myeong-Ho;Yun, Deok-Gyun
    • Journal of Korean Society for Quality Management
    • /
    • v.17 no.2
    • /
    • pp.17-26
    • /
    • 1989
  • This paper is concerned with forecasting the existing number of errors in the computer software and optimizing the stopping time of the software test based upon the forecasted number of errors. The most commonly used models have assessed software reliability under the assumption that the software failure late is proportional to the current fault content of the software but invariant to time since software faults are independents of others and equally likely to cause a failure during testing. In practice, it has been observed that in many situations, the failure rate decrease. Hence, this paper proposes a mathematical model to describe testing situations where the failure rate of software limearly decreases proportional to testing time. The least square method is used to estimate parameters of the mathematical model. A cost model to optimize the software testing time is also proposed. In this cost mode two cost factors are considered. The first cost is to test execution cost directly proportional to test time and the second cost is the failure cost incurred after delivery of the software to user. The failure cost is assumed to be proportional to the number of errors remained in the software at the test stopping time. The optimal stopping time is determined to minimize the total cost, which is the sum of test execution cast and the failure cost. A numerical example is solved to illustrate the proposed procedure.

  • PDF

Decomposition-based Process Planning far Layered Manufacturing of Functionally Gradient Materials (기능성 경사복합재의 적층조형을 위한 분해기반 공정계획)

  • Shin K.H.;Kim S.H.
    • Korean Journal of Computational Design and Engineering
    • /
    • v.11 no.3
    • /
    • pp.223-233
    • /
    • 2006
  • Layered manufacturing(LM) is emerging as a new technology that enables the fabrication of three dimensional heterogeneous objects such as Multi-materials and Functionally Gradient Materials (FGMs). Among various types of heterogeneous objects, more attention has recently paid on the fabrication of FGMs because of their potentials in engineering applications. The necessary steps for LM fabrication of FGMs include representation and process planning of material information inside an FGM. This paper introduces a new process planning algorithm that takes into account the processing of material information. The detailed tasks are discretization (i.e., decomposition-based approximation of volume fraction), orientation (build direction selection), and adaptive slicing of heterogeneous objects. In particular, this paper focuses on the discretization process that converts all of the material information inside an FGM into material features like geometric features. It is thus possible to choose an optimal build direction among various pre-selected ones by approximately estimating build time. This is because total build time depends on the complexity of features. This discretization process also allows adaptive slicing of heterogeneous objects to minimize surface finish and material composition error. In addition, tool path planning can be simplified into fill pattern generation. Specific examples are shown to illustrate the overall procedure.

A pilot study on remake of dental prosthesis of dental laboratory working (치과보철물의 재제작 실태에 관한 예비조사)

  • Nam, Shin-Eun
    • Journal of Technologic Dentistry
    • /
    • v.40 no.3
    • /
    • pp.173-180
    • /
    • 2018
  • Purpose: The purpose of this study was to verify the remake rate and cause of dental prosthesis and to investigate major factors of remake of dental prosthesis. Methods: This study carried out self-administered questionnaire survey from 126 nationwide dental laboratory CEO and directors, which was conducted from September to October in 2016. Results: Total remake rate of dental prosthesis was 10.1% at the nationwide dental laboratories. It was in order of remake rate of dental prosthesis 11.8% for CAD/CAM, 11.5% for porcelain and 11.0% for implant prosthesis. Error of clinical impression was the highest remake cause(63.7%). Nevertheless, dental laboratory take the responsibility of expense for remake of dental prosthesis, regardless of remake cause(67.4%). There was no relation between dental laboratory characteristics and the remake rate of dental prosthesis(p>.05). Conclusion : The remake rate of dental prostheses should be reduced to minimize the economic loss of dental laboratories and raise dental prosthesis satisfaction. It is required to communicate of dentist, dental technicians, and patients, moreover, undistorted information about oral environment should be provided to the dental technicians.

Swarm Intelligence-based Power Allocation and Relay Selection Algorithm for wireless cooperative network

  • Xing, Yaxin;Chen, Yueyun;Lv, Chen;Gong, Zheng;Xu, Ling
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.3
    • /
    • pp.1111-1130
    • /
    • 2016
  • Cooperative communications can significantly improve the wireless transmission performance with the help of relay nodes. In cooperative communication networks, relay selection and power allocation are two key issues. In this paper, we propose a relay selection and power allocation scheme RS-PA-PSACO (Relay Selection-Power Allocation-Particle Swarm Ant Colony Optimization) based on PSACO (Particle Swarm Ant Colony Optimization) algorithm. This scheme can effectively reduce the computational complexity and select the optimal relay nodes. As one of the swarm intelligence algorithms, PSACO which combined both PSO (Particle Swarm Optimization) and ACO (Ant Colony Optimization) algorithms is effective to solve non-linear optimization problems through a fast global search at a low cost. The proposed RS-PA-PSACO algorithm can simultaneously obtain the optimal solutions of relay selection and power allocation to minimize the SER (Symbol Error Rate) with a fixed total power constraint both in AF (Amplify and Forward) and DF (Decode and Forward) modes. Simulation results show that the proposed scheme improves the system performance significantly both in reliability and power efficiency at a low complexity.

Development of a Design System for Multi-Stage Gear Drives (1st Report : Procposal of Formal Processes for Dimensional Design of Gears) (다단 치차장치 설계 시스템 개발에 관한 연구(제 1보: 정식화된 제원 설계 프로세스의 제안))

  • Jeong, Tae-Hyeong
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.17 no.9
    • /
    • pp.202-209
    • /
    • 2000
  • In recent years the concern of designing multi-stage gear drives increases with the more application of gear drives in high-speed and high-load. until now however research on the gear drive design has been focused on single gear pairs and the design has been depended on experiences and know-how of designers and carried out commonly by trial and error. We propose the automation of the dimensional design of gears and the configuration design for gear arrangement of two-and three-stage cylindrical gear drives. The dimensional design is divided into two types of design processes to determine the dimensions of gears. The first design process(Process I) uses the total volume of gears to determine gear ratio and uses K factor unit load and aspect ratio to determine gear dimensions. The second one(Process II) makes use of Niemann's formula and center distance to calculate gear ratio and dimensions. Process I and II employ material data from AGMA and ISO standards respectively. The configuration design determines the positions of gears to minimize the volume of gearbox by simulated annealing algorithm. Finally the availability of the design algorithm is validated by the design examples of two-and three-stage gear drives.

  • PDF