• Title/Summary/Keyword: software algorithms

Search Result 1,093, Processing Time 0.036 seconds

Echocardiography Core Laboratory Validation of a Novel Vendor-Independent Web-Based Software for the Assessment of Left Ventricular Global Longitudinal Strain

  • Ernest Spitzer;Benjamin Camacho;Blaz Mrevlje;Hans-Jelle Brandendburg;Claire B. Ren
    • Journal of Cardiovascular Imaging
    • /
    • v.31 no.3
    • /
    • pp.135-141
    • /
    • 2023
  • BACKGROUND: Global longitudinal strain (GLS) is an accurate and reproducible parameter of left ventricular (LV) systolic function which has shown meaningful prognostic value. Fast, user-friendly, and accurate tools are required for its widespread implementation. We aim to compare a novel web-based tool with two established algorithms for strain analysis and test its reproducibility. METHODS: Thirty echocardiographic datasets with focused LV acquisitions were analyzed using three different semi-automated endocardial GLS algorithms by two readers. Analyses were repeated by one reader for the purpose of intra-observer variability. CAAS Qardia (Pie Medical Imaging) was compared with 2DCPA and AutoLV (TomTec). RESULTS: Mean GLS values were -15.0 ± 3.5% from Qardia, -15.3 ± 4.0% from 2DCPA, and -15.2 ± 3.8% from AutoLV. Mean GLS between Qardia and 2DCPA were not statistically different (p = 0.359), with a bias of -0.3%, limits of agreement (LOA) of 3.7%, and an intraclass correlation coefficient (ICC) of 0.88. Mean GLS between Qardia and AutoLV were not statistically different (p = 0.637), with a bias of -0.2%, LOA of 3.4%, and an ICC of 0.89. The coefficient of variation (CV) for intra-observer variability was 4.4% for Qardia, 8.4% 2DCPA, and 7.7% AutoLV. The CV for inter-observer variability was 4.5%, 8.1%, and 8.0%, respectively. CONCLUSIONS: In echocardiographic datasets of good image quality analyzed at an independent core laboratory using a standardized annotation method, a novel web-based tool for GLS analysis showed consistent results when compared with two algorithms of an established platform. Moreover, inter- and intra-observer reproducibility results were excellent.

Optimal Selection of Classifier Ensemble Using Genetic Algorithms (유전자 알고리즘을 이용한 분류자 앙상블의 최적 선택)

  • Kim, Myung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.99-112
    • /
    • 2010
  • Ensemble learning is a method for improving the performance of classification and prediction algorithms. It is a method for finding a highly accurateclassifier on the training set by constructing and combining an ensemble of weak classifiers, each of which needs only to be moderately accurate on the training set. Ensemble learning has received considerable attention from machine learning and artificial intelligence fields because of its remarkable performance improvement and flexible integration with the traditional learning algorithms such as decision tree (DT), neural networks (NN), and SVM, etc. In those researches, all of DT ensemble studies have demonstrated impressive improvements in the generalization behavior of DT, while NN and SVM ensemble studies have not shown remarkable performance as shown in DT ensembles. Recently, several works have reported that the performance of ensemble can be degraded where multiple classifiers of an ensemble are highly correlated with, and thereby result in multicollinearity problem, which leads to performance degradation of the ensemble. They have also proposed the differentiated learning strategies to cope with performance degradation problem. Hansen and Salamon (1990) insisted that it is necessary and sufficient for the performance enhancement of an ensemble that the ensemble should contain diverse classifiers. Breiman (1996) explored that ensemble learning can increase the performance of unstable learning algorithms, but does not show remarkable performance improvement on stable learning algorithms. Unstable learning algorithms such as decision tree learners are sensitive to the change of the training data, and thus small changes in the training data can yield large changes in the generated classifiers. Therefore, ensemble with unstable learning algorithms can guarantee some diversity among the classifiers. To the contrary, stable learning algorithms such as NN and SVM generate similar classifiers in spite of small changes of the training data, and thus the correlation among the resulting classifiers is very high. This high correlation results in multicollinearity problem, which leads to performance degradation of the ensemble. Kim,s work (2009) showedthe performance comparison in bankruptcy prediction on Korea firms using tradition prediction algorithms such as NN, DT, and SVM. It reports that stable learning algorithms such as NN and SVM have higher predictability than the unstable DT. Meanwhile, with respect to their ensemble learning, DT ensemble shows the more improved performance than NN and SVM ensemble. Further analysis with variance inflation factor (VIF) analysis empirically proves that performance degradation of ensemble is due to multicollinearity problem. It also proposes that optimization of ensemble is needed to cope with such a problem. This paper proposes a hybrid system for coverage optimization of NN ensemble (CO-NN) in order to improve the performance of NN ensemble. Coverage optimization is a technique of choosing a sub-ensemble from an original ensemble to guarantee the diversity of classifiers in coverage optimization process. CO-NN uses GA which has been widely used for various optimization problems to deal with the coverage optimization problem. The GA chromosomes for the coverage optimization are encoded into binary strings, each bit of which indicates individual classifier. The fitness function is defined as maximization of error reduction and a constraint of variance inflation factor (VIF), which is one of the generally used methods to measure multicollinearity, is added to insure the diversity of classifiers by removing high correlation among the classifiers. We use Microsoft Excel and the GAs software package called Evolver. Experiments on company failure prediction have shown that CO-NN is effectively applied in the stable performance enhancement of NNensembles through the choice of classifiers by considering the correlations of the ensemble. The classifiers which have the potential multicollinearity problem are removed by the coverage optimization process of CO-NN and thereby CO-NN has shown higher performance than a single NN classifier and NN ensemble at 1% significance level, and DT ensemble at 5% significance level. However, there remain further research issues. First, decision optimization process to find optimal combination function should be considered in further research. Secondly, various learning strategies to deal with data noise should be introduced in more advanced further researches in the future.

A Design of Software Receiver for GNSS Signal Processing

  • Choi, Seung-Hyun;Kim, Jae-Hyun;Shin, Cheon-Sig;Lee, Sang-Uk;Kim, Jae-Hoon
    • Journal of Satellite, Information and Communications
    • /
    • v.2 no.2
    • /
    • pp.48-52
    • /
    • 2007
  • Recently, the research of GPS receiver which uses the Software-Defined Radio(SDR) technique is being actively proceeded instead of traditional hardware-based receiver. The software-based GPS receiver indicates that the signal acquisition and tracking treated by the hardware-based platform are processed as the software technique through a microprocessor. In this paper, GPS software receiver is designed by using SDR technique and then the signal acquisition, tracking, and the navigation message decoding parts are verified through the PC-based simulation. Moreover, the efficient algorithms are developed about the signal acquisition and tracking parts in order to obtain the accurate pseudorange. Finally, the pseudorange is calculated through the relative channel delay received through the different satellite of L1 frequency band. GPS software receiver proposed in this paper will be included in the element of GPS/Galileo complex system of development target and will provide not only the method that verifies the performance for Galileo Sensor Station standard but also usability by providing various debugging environments.

  • PDF

A Novel Compressed Sensing Technique for Traffic Matrix Estimation of Software Defined Cloud Networks

  • Qazi, Sameer;Atif, Syed Muhammad;Kadri, Muhammad Bilal
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.10
    • /
    • pp.4678-4702
    • /
    • 2018
  • Traffic Matrix estimation has always caught attention from researchers for better network management and future planning. With the advent of high traffic loads due to Cloud Computing platforms and Software Defined Networking based tunable routing and traffic management algorithms on the Internet, it is more necessary as ever to be able to predict current and future traffic volumes on the network. For large networks such origin-destination traffic prediction problem takes the form of a large under- constrained and under-determined system of equations with a dynamic measurement matrix. Previously, the researchers had relied on the assumption that the measurement (routing) matrix is stationary due to which the schemes are not suitable for modern software defined networks. In this work, we present our Compressed Sensing with Dynamic Model Estimation (CS-DME) architecture suitable for modern software defined networks. Our main contributions are: (1) we formulate an approach in which measurement matrix in the compressed sensing scheme can be accurately and dynamically estimated through a reformulation of the problem based on traffic demands. (2) We show that the problem formulation using a dynamic measurement matrix based on instantaneous traffic demands may be used instead of a stationary binary routing matrix which is more suitable to modern Software Defined Networks that are constantly evolving in terms of routing by inspection of its Eigen Spectrum using two real world datasets. (3) We also show that linking this compressed measurement matrix dynamically with the measured parameters can lead to acceptable estimation of Origin Destination (OD) Traffic flows with marginally poor results with other state-of-art schemes relying on fixed measurement matrices. (4) Furthermore, using this compressed reformulated problem, a new strategy for selection of vantage points for most efficient traffic matrix estimation is also presented through a secondary compression technique based on subset of link measurements. Experimental evaluation of proposed technique using real world datasets Abilene and GEANT shows that the technique is practical to be used in modern software defined networks. Further, the performance of the scheme is compared with recent state of the art techniques proposed in research literature.

A Metrics-Based Approach to the Reorganization of Class Hierarchy Structures (클래스계층구조의 품질평가척도를 기반으로 하는 재구성기법)

  • Hwang, Sun-Hyung;Yang, Hea-Sool;Hwang, Young-Sub
    • The KIPS Transactions:PartD
    • /
    • v.10D no.5
    • /
    • pp.859-872
    • /
    • 2003
  • Class hierarchies often constitute the backbone of object-oriented software. Their quality is therefore quite crucial. Building class hierarchies with good qualify is a very important and common tasks on the object oriented software development, but such hierarchies are not so easy to build. Moreover, the class hierarchy structure under construction is frequently restructured and refined until it becomes suitable for the requirement on the iterative and incremental development lifecycle. Therefore, there has been renewal of interest in all methodologies and tools to assist the object oriented developers in this task. In this paper, we define a set of quantitative metrics which provide a wav of capturing features of a rough estimation of complexity of class hierarchy structure. In addition to, we suggest a set of algorithms that transform a original class hierarchy structure into reorganized one based on the proposed metrics for class hierarchy structure. Furthermore, we also prove that each algorithm is "object-preserving". That is, we prove that the set of objects are never changed before and after applying the algorithm on a class hierarchy. The technique presented in this paper can be used as a guidelines of the construction, restructuring and refinement of class hierarchies. Moreover, the proposed set of algorithms based on metrics can be helpful for developers as an useful instrument for the object-oriented software development.velopment.

Resource Clustering Simulator for Desktop Virtualization Based on Intra Cloud (인트라 클라우드 기반 데스크탑 가상화를 위한 리소스 클러스터링 시뮬레이터)

  • Kim, Hyun-Woo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.1
    • /
    • pp.45-50
    • /
    • 2019
  • With the gradual advancement of IT, passive work processes are automated and the overall quality of life has greatly improved. This is made possible by the formation of an organic topology between a wide variety of real-life smart devices. To serve these diverse smart devices, businesses or users are using the cloud. The services in the cloud are divided into Infrastructure as a Service (IaaS), Platform as a Service (PaaS) and Software as a Service (SaaS). SaaS runs on PaaS, and PaaS runs on IaaS. Since IaaS is the basis of all services, an algorithm is required to operate virtualization resources efficiently. Among them, desktop resource virtualization is used for resource high availability of unused state time of existing desktop PC. Clustering of hierarchical structures is important for high availability of these resources. In addition, it is very important to select a suitable algorithm because many clustering algorithms are mainly used depending on the distribution ratio and environment of the desktop PC. If various attempts are made to find an algorithm suitable for desktop resource virtualization in an operating environment, a great deal of power, time, and manpower will be incurred. Therefore, this paper proposes a resource clustering simulator for cluster selection of desktop virtualization. This provides a clustering simulation to properly select clustering algorithms and apply elements in different environments of desktop PCs.

Hierarchical Particle Swarm Optimization for Multi UAV Waypoints Planning Under Various Threats (다양한 위협 하에서 복수 무인기의 경로점 계획을 위한 계층적 입자 군집 최적화)

  • Chung, Wonmo;Kim, Myunggun;Lee, Sanha;Lee, Sang-Pill;Park, Chun-Shin;Son, Hungsun
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.50 no.6
    • /
    • pp.385-391
    • /
    • 2022
  • This paper presents to develop a path planning algorithm combining gradient descent-based path planning (GBPP) and particle swarm optimization (PSO) for considering prohibited flight areas, terrain information, and characteristics of fixed-wing unmmaned aerial vehicle (UAV) in 3D space. Path can be generated fast using GBPP, but it is often happened that an unsafe path can be generated by converging to a local minimum depending on the initial path. Bio-inspired swarm intelligence algorithms, such as Genetic algorithm (GA) and PSO, can avoid the local minima problem by sampling several paths. However, if the number of optimal variable increases due to an increase in the number of UAVs and waypoints, it requires heavy computation time and efforts due to increasing the number of particles accordingly. To solve the disadvantages of the two algorithms, hierarchical path planning algorithm associated with hierarchical particle swarm optimization (HPSO) is developed by defining the initial path, which is the input of GBPP, as two variables including particles variables. Feasibility of the proposed algorithm is verified by software-in-the-loop simulation (SILS) of flight control computer (FCC) for UAVs.

Compensation of Installation Errors in a Laser Vision System and Dimensional Inspection of Automobile Chassis

  • Barkovski Igor Dunin;Samuel G.L.;Yang Seung-Han
    • Journal of Mechanical Science and Technology
    • /
    • v.20 no.4
    • /
    • pp.437-446
    • /
    • 2006
  • Laser vision inspection systems are becoming popular for automated inspection of manufactured components. The performance of such systems can be enhanced by improving accuracy of the hardware and robustness of the software used in the system. This paper presents a new approach for enhancing the capability of a laser vision system by applying hardware compensation and using efficient analysis software. A 3D geometrical model is developed to study and compensate for possible distortions in installation of gantry robot on which the vision system is mounted. Appropriate compensation is applied to the inspection data obtained from the laser vision system based on the parameters in 3D model. The present laser vision system is used for dimensional inspection of car chassis sub frame and lower arm assembly module. An algorithm based on simplex search techniques is used for analyzing the compensated inspection data. The details of 3D model, parameters used for compensation and the measurement data obtained from the system are presented in this paper. The details of search algorithm used for analyzing the measurement data and the results obtained are also presented in the paper. It is observed from the results that, by applying compensation and using appropriate algorithms for analyzing, the error in evaluation of the inspection data can be significantly minimized, thus reducing the risk of rejecting good parts.

The Performance Evaluation of New Web Caching with Related Content using Colored Petri Net Simulation

  • Abdullaev, Sarvar;Ko, Franz I.S.
    • The Journal of Society for e-Business Studies
    • /
    • v.13 no.3
    • /
    • pp.67-83
    • /
    • 2008
  • Web caching server is one of the important components of any web site, as it makes the access of users to web content much faster while balancing the network and server load. There are many caching approaches proposed in order to solve this problem. Classically, we know several caching algorithms like FIFO, LRU, LRU-min and etc. Most of them have their advantages and disadvantages based on specific context. The purpose of this paper is to introduce the new concept of caching objects along with their related content and test the performance of the proposed caching method using CPN Tools simulator program. It is generally known that the use of Colored Petri Nets(CPNs) for modeling the simulation of new idea is one of the popular methods of system evaluation throughout software development projects. Moreover it could be very handy in describing the overall logic of the system. The CPN Tools is one of the most robust simulator programs which supports all necessary tools and functions to build and run the simulation model for CPN. Therefore in this paper, we will use CPN Tools software in order to build a simulator for our innovative caching approach named as web caching with related content. Then we will analyze the results derived from the simulation of our model and will make corresponding conclusions.

  • PDF

An integrated approach for optimum design of HPC mix proportion using genetic algorithm and artificial neural networks

  • Parichatprecha, Rattapoohm;Nimityongskul, Pichai
    • Computers and Concrete
    • /
    • v.6 no.3
    • /
    • pp.253-268
    • /
    • 2009
  • This study aims to develop a cost-based high-performance concrete (HPC) mix optimization system based on an integrated approach using artificial neural networks (ANNs) and genetic algorithms (GA). ANNs are used to predict the three main properties of HPC, namely workability, strength and durability, which are used to evaluate fitness and constraint violations in the GA process. Multilayer back-propagation neural networks are trained using the results obtained from experiments and previous research. The correlation between concrete components and its properties is established. GA is employed to arrive at an optimal mix proportion of HPC by minimizing its total cost. A system prototype, called High Performance Concrete Mix-Design System using Genetic Algorithm and Neural Networks (HPCGANN), was developed in MATLAB. The architecture of the proposed system consists of three main parts: 1) User interface; 2) ANNs prediction models software; and 3) GA engine software. The validation of the proposed system is carried out by comparing the results obtained from the system with the trial batches. The results indicate that the proposed system can be used to enable the design of HPC mix which corresponds to its required performance. Furthermore, the proposed system takes into account the influence of the fluctuating unit price of materials in order to achieve the lowest cost of concrete, which cannot be easily obtained by traditional methods or trial-and-error techniques.