• Title/Summary/Keyword: adaptive algorithm

Search Result 4,664, Processing Time 0.033 seconds

A Study on the Implementation of an Agile SFFS Based on 5DOF Manipulator (5축 매니퓰레이터를 이용한 쾌속 임의형상제작시스템의 구현에 관한 연구)

  • Kim Seung-Woo;Jung Yong-Rae
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.42 no.1
    • /
    • pp.1-11
    • /
    • 2005
  • Several Solid Freeform Fabrication Systems(SFFS) are commercialized in a few companies for rapid prototyping. However, they have many technical problems including the limitation of applicable materials. A new method of agile prototyping is required for the recent manufacturing environments of multi-item and small quantity production. The objectives of this paper include the development of a novel method of SFFS, the CAFL/sup VM/(Computer Aided Fabrication of Lamination for Various Material), and the manufacture of the various material samples for the certification of the proposed system and the creation of new application areas. For these objectives, the technologies for a highly accurate robot path control, the optimization of support structure, CAD modeling, adaptive slicing was implemented. However, there is an important problem with the conventional 2D lamination method. That is the inaccuracy of 3D model surface, which is caused by the stair-type surface generated in virtue of vertical 2D cutting. In this paper, We design the new control algorithm that guarantees the constant speed, precise positioning and tangential cutting on the 5DOF SFFS. We develop the tangential cutting algorithm to be controlled with constant speed and successfully implemented in the 5DOF CAFL/sup VM/ system developed in this paper. Finally, this paper confirms its high-performance through the experimental results from the application into CAFL/sup VM/ system.

Hardware optimized high quality image signal processor for single-chip CMOS Image Sensor (Single-chip CMOS Image Sensor를 위한 하드웨어 최적화된 고화질 Image Signal Processor 설계)

  • Lee, Won-Jae;Jung, Yun-Ho;Lee, Seong-Joo;Kim, Jae-Seok
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.5
    • /
    • pp.103-111
    • /
    • 2007
  • In this paper, we propose a VLSI architecture of hardware optimized high quality image signal processor for a Single-chip CMOS Image Sensor(CIS). The Single-chip CIS is usually used for mobile applications, so it has to be implemented as small as possible while maintaining the image quality. Several image processing algorithms are used in ISP to improve captured image quality. Among the several image processing blocks, demosaicing and image filter are the core blocks in ISP. These blocks need line memories, but the number of line memories is limited in a low cost Single-chip CIS. In our design, high quality edge-adaptive and cross channel correlation considered demosaicing algorithm is adopted. To minimize the number of required line memories for image filter, we share the line memories using the characteristics of demosaicing algorithm which consider the cross correlation. Based on the proposed method, we can achieve both high quality and low hardware complexity with a small number of line memories. The proposed method was implemented and verified successfully using verilog HDL and FPGA. It was synthesized to gate-level circuits using 0.25um CMOS standard cell library. The total logic gate count is 37K, and seven and half line memories are used.

Real-Time DSP Implementation of IMT-2000 Speech Coding Algorithm (IMT-2000 음성부호화 알고리즘의 실시간 DSP 구현)

  • Seo, Jeong-Uk;Gwon, Hong-Seok;Park, Man-Ho;Bae, Geon-Seong
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.38 no.3
    • /
    • pp.304-315
    • /
    • 2001
  • In this paper, we peformed the real-time implementation of AMR(Adaptive Multi-Rate) speech coding algorithm which is adopted for IMT-2000 service using TMS320C6201, i.e., a Texas Instrument´s fixed-point DSP. With the ANSI C source code released from ETSI, optimization is performed to make it run in real-time with memory as small as possible using the C compiler and assembly language. Implemented AMR speech codec has the size of 32.06 kWords program memory, 9.75 kWords data RAM memory, and 19.89 kWords data ROM memory. And, The time required for processing one frame of 20 ms length speech data is about 4.38 ms, and it is short enough for real-time operation. It is verified that the decoded result of the implemented speech codec on the DSP is identical with the PC simulation result using ANSI C code for test sequences. Also, actual sound input/output test using microphone and speaker demonstrates its proper real-time operation without distortions or delays.

  • PDF

Implementation of a Jitter and Glitch Removing Circuit for UHF RFID System Based on ISO/IEC 18000-6C Standard (UHF대역 RFID 수신단(리더)의 지터(비트동기) 및 글리치 제거회로 설계)

  • Kim, Sang-Hoon;Lee, Yong-Joo;Sim, Jae-Hee;Lee, Yong-Surk
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.32 no.1A
    • /
    • pp.83-90
    • /
    • 2007
  • In this paper, we propose an implementation and an algorithm of 'Jitter and Glitch Removing Circuit' for UHF RFID reader system based on ISO/IEC 18000-6C standard. We analyze the response of TI(Texas Instrument) Gen2 tag with a reader using the proposed algorithm. In ISO/IEC 18000-6C standard, a bit rate accuracy(tolerance) is up to +/-22% during tag-to-interrogator communication and +/-1% during interrogator-to-tag communication. In order to solve tolerance problems, we implement the Jitter and Glitch Removing Circuit using the concept of tolerance and tolerance-accumulation instead of PLL(DPLL, ADPLL). The main clock is 19.2MHz and the LF(Link Frequency) is determined as 40kHz to meet the local radio regulation in korea. As a result of simulations, the error-rate is zero within 15% tolerance of tag responses. And in the case of using the adaptive LF generation circuit, the error-rate varies from 0.000589 to zero between 15% and 22% tolerance of tag responses. In conclusion, the error-rate is zero between 0%-22% tolerance of tag response specified in ISO/IEC 18000-6C standard.

Development of the Algorithm for Traffic Accident Auto-Detection in Signalized Intersection (신호교차로 내 실시간 교통사고 자동검지 알고리즘 개발)

  • O, Ju-Taek;Im, Jae-Geuk;Hwang, Bo-Hui
    • Journal of Korean Society of Transportation
    • /
    • v.27 no.5
    • /
    • pp.97-111
    • /
    • 2009
  • Image-based traffic information collection systems have entered widespread adoption and use in many countries since these systems are not only capable of replacing existing loop-based detectors which have limitations in management and administration, but are also capable of providing and managing a wide variety of traffic related information. In addition, these systems are expanding rapidly in terms of purpose and scope of use. Currently, the utilization of image processing technology in the field of traffic accident management is limited to installing surveillance cameras on locations where traffic accidents are expected to occur and digitalizing of recorded data. Accurately recording the sequence of situations around a traffic accident in a signal intersection and then objectively and clearly analyzing how such accident occurred is more urgent and important than anything else in resolving a traffic accident. Therefore, in this research, we intend to present a technology capable of overcoming problems in which advanced existing technologies exhibited limitations in handling real-time due to large data capacity such as object separation of vehicles and tracking, which pose difficulties due to environmental diversities and changes at a signal intersection with complex traffic situations, as pointed out by many past researches while presenting and implementing an active and environmentally adaptive methodology capable of effectively reducing false detection situations which frequently occur even with the Gaussian complex model analytical method which has been considered the best among well-known environmental obstacle reduction methods. To prove that the technology developed by this research has performance advantage over existing automatic traffic accident recording systems, a test was performed by entering image data from an actually operating crossroad online in real-time. The test results were compared with the performance of other existing technologies.

A Finite Element Simulation of Cancellous Bone Remodeling Based on Volumetric Strain (스폰지 뼈의 Remodeling 예측을 위한 체적 변형률을 이용한 유한요소 알고리즘)

  • Kim, Young;Vanderby, Ray
    • Journal of Biomedical Engineering Research
    • /
    • v.21 no.4
    • /
    • pp.373-384
    • /
    • 2000
  • The goal of this paper is to develop a computational method to predict cancellous bone density distributions based upon continuum levels of volumetric strain. Volumetric strain is defined as the summation of normal strains, excluding shear strains, within an elastic range of loadings. Volumetric strain at a particular location in a cancellous structure changes with changes of the boundary conditions (prescribed displacements, tractions, and pressure). This change in the volumetric strain is postulated to predict the adaptive change in the bone apparent density. This bone remodeling theory based on volumetric strain is then used with the finite element method to compute the apparent density distribution for cancellous bone in both lumbar spine and proximal femur using an iterative algorithm, considering the dead zone of strain stimuli. The apparent density distribution of cancellous bone predicted by this method has the same pattern as experimental data reported in the literature (Wolff 1892, Keller et al. 1989, Cody et al. 1992). The resulting bone apparent density distributions predict Young's modulus and strength distributions throughout cancellous bone in agreement with the literature (Keller et al. 1989, Carter and Hayes 1977). The method was convergent and sensitive to changes in boundary conditions. Therefore, the computational algorithm of the present study appears to be a useful approach to predict the apparent density distribution of cancellous bone (i.e. a numerical approximation for Wolff's Law)

  • PDF

Memory-Free Skin-Detection Algorithm and Implementation of Hardware Design for Small-Sized Display Device (소형 DISPLAY 장치를 위한 비 메모리 피부 검출 알고리즘 및 HARDWARE 구현)

  • Im, Jeong-Uk;Song, Jin-Gun;Ha, Joo-Young;Kang, Bong-Soon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.11 no.8
    • /
    • pp.1456-1464
    • /
    • 2007
  • The research of skin-tone detection has been conducting continuously to enlarge the importance in security, surveillance and administration of the information and 'Password Control System' for using face and skin recognition in airports, harbors and general companies. As well as tile rapid diffusion of the application range in image communications and an electron transaction using wide range of communication network, the importance of the accurate detection of skin color has been augmenting recently. In this paper, it will set up the boundaries of skin colors using the information of Cb and Cr in YCbCr color model of human skin color which is from hundreds compiled portrait images for each race, and suggest a efficient yet simple structure about the skin detection which has been followed by whether the comprehension of the boundaries of skin or not with adaptive skin-range set. With the possibility of the 1D Processes which does not use any memory, it is able to be applied to relatively small-sized hardware and system such as mobile apparatuses. To add the selective mode, it is not only available the improvement of tie skin detection, but also showing the correspondent results about previous face recognition technologies using complicated algorithm.

A Link-Label Based Node-to-Link Optimal Path Algorithm Considering Non Additive Path Cost (비가산성 경로비용을 반영한 링크표지기반 Node-to-Link 최적경로탐색)

  • Lee, Mee Young;Nam, Doohee
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.18 no.5
    • /
    • pp.91-99
    • /
    • 2019
  • Existing node-to-node based optimal path searching is built on the assumption that all destination nodes can be arrived at from an origin node. However, the recent appearance of the adaptive path search algorithm has meant that the optimal path solution cannot be derived in node-to-node path search. In order to reflect transportation data at the links in real-time, the necessity of the node-to-link (or link-to-node; NL) problem is being recognized. This research assumes existence of a network with link-label and non-additive path costs as a solution to the node-to-link optimal path problem. At the intersections in which the link-label has a turn penalty, the network retains its shape. Non-additive path cost requires that M-similar paths be enumerated so that the ideal path can be ascertained. In this, the research proposes direction deletion and turn restriction so that regulation of the loop in the link-label entry-link-based network transformation method will ensure that an optimal solution is derived up until the final link. Using this method on a case study shows that the proposed method derives the optimal solution through learning. The research concludes by bringing to light the necessity of verification in large-scale networks.

Comparison of Multi-angle TerraSAR-X Staring Mode Image Registration Method through Coarse to Fine Step (Coarse to Fine 단계를 통한 TerraSAR-X Staring Mode 다중 관측각 영상 정합기법 비교 분석)

  • Lee, Dongjun;Kim, Sang-Wan
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.3
    • /
    • pp.475-491
    • /
    • 2021
  • With the recent increase in available high-resolution (< ~1 m) satellite SAR images, the demand for precise registration of SAR images is increasing in various fields including change detection. The registration between high-resolution SAR images acquired in different look angle is difficult due to speckle noise and geometric distortion caused by the characteristics of SAR images. In this study, registration is performed in two stages, coarse and fine, using the x-band SAR data imaged at staring spotlight mode of TerraSAR-X. For the coarse registration, a method combining the adaptive sampling method and SAR-SIFT (Scale Invariant Feature Transform) is applied, and three rigid methods (NCC: Normalized Cross Correlation, Phase Congruency-NCC, MI: Mutual Information) and one non-rigid (Gefolki: Geoscience extended Flow Optical Flow Lucas-Kanade Iterative), for the fine registration stage, was performed for performance comparison. The results were compared by using RMSE (Root Mean Square Error) and FSIM (Feature Similarity) index, and all rigid models showed poor results in all image combinations. It is confirmed that the rigid models have a large registration error in the rugged terrain area. As a result of applying the Gefolki algorithm, it was confirmed that the RMSE of Gefolki showed the best result as a 1~3 pixels, and the FSIM index also obtained a higher value than 0.02~0.03 compared to other rigid methods. It was confirmed that the mis-registration due to terrain effect could be sufficiently reduced by the Gefolki algorithm.

The Adaptive Personalization Method According to Users Purchasing Index : Application to Beverage Purchasing Predictions (고객별 구매빈도에 동적으로 적응하는 개인화 시스템 : 음료수 구매 예측에의 적용)

  • Park, Yoon-Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.95-108
    • /
    • 2011
  • TThis is a study of the personalization method that intelligently adapts the level of clustering considering purchasing index of a customer. In the e-biz era, many companies gather customers' demographic and transactional information such as age, gender, purchasing date and product category. They use this information to predict customer's preferences or purchasing patterns so that they can provide more customized services to their customers. The previous Customer-Segmentation method provides customized services for each customer group. This method clusters a whole customer set into different groups based on their similarity and builds predictive models for the resulting groups. Thus, it can manage the number of predictive models and also provide more data for the customers who do not have enough data to build a good predictive model by using the data of other similar customers. However, this method often fails to provide highly personalized services to each customer, which is especially important to VIP customers. Furthermore, it clusters the customers who already have a considerable amount of data as well as the customers who only have small amount of data, which causes to increase computational cost unnecessarily without significant performance improvement. The other conventional method called 1-to-1 method provides more customized services than the Customer-Segmentation method for each individual customer since the predictive model are built using only the data for the individual customer. This method not only provides highly personalized services but also builds a relatively simple and less costly model that satisfies with each customer. However, the 1-to-1 method has a limitation that it does not produce a good predictive model when a customer has only a few numbers of data. In other words, if a customer has insufficient number of transactional data then the performance rate of this method deteriorate. In order to overcome the limitations of these two conventional methods, we suggested the new method called Intelligent Customer Segmentation method that provides adaptive personalized services according to the customer's purchasing index. The suggested method clusters customers according to their purchasing index, so that the prediction for the less purchasing customers are based on the data in more intensively clustered groups, and for the VIP customers, who already have a considerable amount of data, clustered to a much lesser extent or not clustered at all. The main idea of this method is that applying clustering technique when the number of transactional data of the target customer is less than the predefined criterion data size. In order to find this criterion number, we suggest the algorithm called sliding window correlation analysis in this study. The algorithm purposes to find the transactional data size that the performance of the 1-to-1 method is radically decreased due to the data sparity. After finding this criterion data size, we apply the conventional 1-to-1 method for the customers who have more data than the criterion and apply clustering technique who have less than this amount until they can use at least the predefined criterion amount of data for model building processes. We apply the two conventional methods and the newly suggested method to Neilsen's beverage purchasing data to predict the purchasing amounts of the customers and the purchasing categories. We use two data mining techniques (Support Vector Machine and Linear Regression) and two types of performance measures (MAE and RMSE) in order to predict two dependent variables as aforementioned. The results show that the suggested Intelligent Customer Segmentation method can outperform the conventional 1-to-1 method in many cases and produces the same level of performances compare with the Customer-Segmentation method spending much less computational cost.