• Title/Summary/Keyword: Processing Speed

Search Result 4,286, Processing Time 0.111 seconds

An Implementation Method of the Character Recognizer for the Sorting Rate Improvement of an Automatic Postal Envelope Sorting Machine (우편물 자동구분기의 구분율 향상을 위한 문자인식기의 구현 방법)

  • Lim, Kil-Taek;Jeong, Seon-Hwa;Jang, Seung-Ick;Kim, Ho-Yon
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.12 no.4
    • /
    • pp.15-24
    • /
    • 2007
  • The recognition of postal address images is indispensable for the automatic sorting of postal envelopes. The process of the address image recognition is composed of three steps-address image preprocessing, character recognition, address interpretation. The extracted character images from the preprocessing step are forwarded to the character recognition step, in which multiple candidate characters with reliability scores are obtained for each character image extracted. aracters with reliability scores are obtained for each character image extracted. Utilizing those character candidates with scores, we obtain the final valid address for the input envelope image through the address interpretation step. The envelope sorting rate depends on the performance of all three steps, among which character recognition step could be said to be very important. The good character recognizer would be the one which could produce valid candidates with very reliable scores to help the address interpretation step go easy. In this paper, we propose the method of generating character candidates with reliable recognition scores. We utilize the existing MLP(multilayered perceptrons) neural network of the address recognition system in the current automatic postal envelope sorters, as the classifier for the each image from the preprocessing step. The MLP is well known to be one of the best classifiers in terms of processing speed and recognition rate. The false alarm problem, however, might be occurred in recognition results, which made the address interpretation hard. To make address interpretation easy and improve the envelope sorting rate, we propose promising methods to reestimate the recognition score (confidence) of the existing MLP classifier: the generation method of the statistical recognition properties of the classifier and the method of the combination of the MLP and the subspace classifier which roles as a reestimator of the confidence. To confirm the superiority of the proposed method, we have used the character images of the real postal envelopes from the sorters in the post office. The experimental results show that the proposed method produces high reliability in terms of error and rejection for individual characters and non-characters.

  • PDF

A Real-Time Head Tracking Algorithm Using Mean-Shift Color Convergence and Shape Based Refinement (Mean-Shift의 색 수렴성과 모양 기반의 재조정을 이용한 실시간 머리 추적 알고리즘)

  • Jeong Dong-Gil;Kang Dong-Goo;Yang Yu Kyung;Ra Jong Beom
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.6
    • /
    • pp.1-8
    • /
    • 2005
  • In this paper, we propose a two-stage head tracking algorithm adequate for real-time active camera system having pan-tilt-zoom functions. In the color convergence stage, we first assume that the shape of a head is an ellipse and its model color histogram is acquired in advance. Then, the min-shift method is applied to roughly estimate a target position by examining the histogram similarity of the model and a candidate ellipse. To reflect the temporal change of object color and enhance the reliability of mean-shift based tracking, the target histogram obtained in the previous frame is considered to update the model histogram. In the updating process, to alleviate error-accumulation due to outliers in the target ellipse of the previous frame, the target histogram in the previous frame is obtained within an ellipse adaptively shrunken on the basis of the model histogram. In addition, to enhance tracking reliability further, we set the initial position closer to the true position by compensating the global motion, which is rapidly estimated on the basis of two 1-D projection datasets. In the subsequent stage, we refine the position and size of the ellipse obtained in the first stage by using shape information. Here, we define a robust shape-similarity function based on the gradient direction. Extensive experimental results proved that the proposed algorithm performs head hacking well, even when a person moves fast, the head size changes drastically, or the background has many clusters and distracting colors. Also, the propose algorithm can perform tracking with the processing speed of about 30 fps on a standard PC.

Enhanced Production of Carboxymethylcellulase by a Newly Isolated Marine Microorganism Bacillus atrophaeus LBH-18 Using Rice Bran, a Byproduct from the Rice Processing Industry (미강을 이용한 해양미생물 Bacillus atrophaeus LBH-18 유래의 carboxymethylcellulase 생산의 최적화)

  • Kim, Yi-Joon;Cao, Wa;Lee, Yu-Jeong;Lee, Sang-Un;Jeong, Jeong-Han;Lee, Jin-Woo
    • Journal of Life Science
    • /
    • v.22 no.10
    • /
    • pp.1295-1306
    • /
    • 2012
  • A microorganism producing carboxymethylcellulase (CMCase) was isolated from seawater and identified as Bacillus atrophaeus. This species was designated as B. atrophaeus LBH-18 based on its evolutionary distance and the phylogenetic tree resulting from 16S rDNA sequencing and the neighbor-joining method. The optimal conditions for rice bran (68.1 g/l), peptone (9.1 g/l), and initial pH (7.0) of the medium for cell growth was determined by Design Expert Software based on the response surface method; conditions for production of CMCase were 55.2 g/l, 6.6 g/l, and 7.1, respectively. The optimal temperature for cell growth and the production of CMCase by B. atrophaeus LBH-18 was $30^{\circ}C$. The optimal conditions of agitation speed and aeration rate for cell growth in a 7-l bioreactor were 324 rpm and 0.9 vvm, respectively, whereas those for production of CMCase were 343 rpm and 0.6 vvm, respectively. The optimal inner pressure for cell growth and production of CMCase in a 100-l bioreactor was 0.06 MPa. Maximal production of CMCase under optimal conditions in a 100-l bioreactor was 127.5 U/ml, which was 1.32 times higher than that without an inner pressure. In this study, rice bran was developed as a carbon source for industrial scale production of CMCase by B. atrophaeus LBH-18. Reduced time for the production of CMCase from 7 to 10 days to 3 days by using a bacterial strain with submerged fermentation also resulted in increased productivity of CMCase and a decrease in its production cost.

X-tree Diff: An Efficient Change Detection Algorithm for Tree-structured Data (X-tree Diff: 트리 기반 데이터를 위한 효율적인 변화 탐지 알고리즘)

  • Lee, Suk-Kyoon;Kim, Dong-Ah
    • The KIPS Transactions:PartC
    • /
    • v.10C no.6
    • /
    • pp.683-694
    • /
    • 2003
  • We present X-tree Diff, a change detection algorithm for tree-structured data. Our work is motivated by need to monitor massive volume of web documents and detect suspicious changes, called defacement attack on web sites. From this context, our algorithm should be very efficient in speed and use of memory space. X-tree Diff uses a special ordered labeled tree, X-tree, to represent XML/HTML documents. X-tree nodes have a special field, tMD, which stores a 128-bit hash value representing the structure and data of subtrees, so match identical subtrees form the old and new versions. During this process, X-tree Diff uses the Rule of Delaying Ambiguous Matchings, implying that it perform exact matching where a node in the old version has one-to one corrspondence with the corresponding node in the new, by delaying all the others. It drastically reduces the possibility of wrong matchings. X-tree Diff propagates such exact matchings upwards in Step 2, and obtain more matchings downwsards from roots in Step 3. In step 4, nodes to ve inserted or deleted are decided, We aldo show thst X-tree Diff runs on O(n), woere n is the number of noses in X-trees, in worst case as well as in average case, This result is even better than that of BULD Diff algorithm, which is O(n log(n)) in worst case, We experimented X-tree Diff on reat data, which are about 11,000 home pages from about 20 wev sites, instead of synthetic documets manipulated for experimented for ex[erimentation. Currently, X-treeDiff algorithm is being used in a commeercial hacking detection system, called the WIDS(Web-Document Intrusion Detection System), which is to find changes occured in registered websites, and report suspicious changes to users.

Processing and Quality Control of Flux Data at Gwangneung Forest (광릉 산림의 플럭스 자료 처리와 품질 관리)

  • Lim, Hee-Jeong;Lee, Young-Hee
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.10 no.3
    • /
    • pp.82-93
    • /
    • 2008
  • In order to ensure a standardized data analysis of the eddy covariance measurements, Hong and Kim's quality control program has been updated and used to process eddy covariance data measured at two levels on the main flux tower at Gwangneung site from January to May in 2005. The updated program was allowed to remove outliers automatically for $CO_2$ and latent heat fluxes. The flag system consists of four quality groups(G, D, B and M). During the study period, the missing data were about 25% of the total records. About 60% of the good quality data were obtained after the quality control. The number of record in G group was larger at 40m than at 20m. It is due that the level of 20m was within the roughness sublayer where the presence of the canopy influences directly on the character of the turbulence. About 60% of the bad data were due to low wind speed. Energy balance closure at this site was about 40% during the study period. Large imbalance is attributed partly to the combined effects of the neglected heat storage terms, inaccuracy of ground heat flux and advection due to local wind system near the surface. The analysis of wind direction indicates that the frequent occurrence of positive momentum flux was closely associated with mountain valley wind system at this site. The negative $CO_2$ flux at night was examined in terms of averaging time. The results show that when averaging time is larger than 10min, the magnitude of calculated $CO_2$ fluxes increases rapidly, suggesting that the 30min $CO_2$ flux is influenced severely by the mesoscale motion or nonstationarity. A proper choice of averaging time needs to be considered to get accurate turbulent fluxes during nighttime.

Design of MAHA Supercomputing System for Human Genome Analysis (대용량 유전체 분석을 위한 고성능 컴퓨팅 시스템 MAHA)

  • Kim, Young Woo;Kim, Hong-Yeon;Bae, Seungjo;Kim, Hag-Young;Woo, Young-Choon;Park, Soo-Jun;Choi, Wan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.2
    • /
    • pp.81-90
    • /
    • 2013
  • During the past decade, many changes and attempts have been tried and are continued developing new technologies in the computing area. The brick wall in computing area, especially power wall, changes computing paradigm from computing hardwares including processor and system architecture to programming environment and application usage. The high performance computing (HPC) area, especially, has been experienced catastrophic changes, and it is now considered as a key to the national competitiveness. In the late 2000's, many leading countries rushed to develop Exascale supercomputing systems, and as a results tens of PetaFLOPS system are prevalent now. In Korea, ICT is well developed and Korea is considered as a one of leading countries in the world, but not for supercomputing area. In this paper, we describe architecture design of MAHA supercomputing system which is aimed to develop 300 TeraFLOPS system for bio-informatics applications like human genome analysis and protein-protein docking. MAHA supercomputing system is consists of four major parts - computing hardware, file system, system software and bio-applications. MAHA supercomputing system is designed to utilize heterogeneous computing accelerators (co-processors like GPGPUs and MICs) to get more performance/$, performance/area, and performance/power. To provide high speed data movement and large capacity, MAHA file system is designed to have asymmetric cluster architecture, and consists of metadata server, data server, and client file system on top of SSD and MAID storage servers. MAHA system softwares are designed to provide user-friendliness and easy-to-use based on integrated system management component - like Bio Workflow management, Integrated Cluster management and Heterogeneous Resource management. MAHA supercomputing system was first installed in Dec., 2011. The theoretical performance of MAHA system was 50 TeraFLOPS and measured performance of 30.3 TeraFLOPS with 32 computing nodes. MAHA system will be upgraded to have 100 TeraFLOPS performance at Jan., 2013.

A hybrid algorithm for the synthesis of computer-generated holograms

  • Nguyen The Anh;An Jun Won;Choe Jae Gwang;Kim Nam
    • Proceedings of the Optical Society of Korea Conference
    • /
    • 2003.07a
    • /
    • pp.60-61
    • /
    • 2003
  • A new approach to reduce the computation time of genetic algorithm (GA) for making binary phase holograms is described. Synthesized holograms having diffraction efficiency of 75.8% and uniformity of 5.8% are proven in computer simulation and experimentally demonstrated. Recently, computer-generated holograms (CGHs) having high diffraction efficiency and flexibility of design have been widely developed in many applications such as optical information processing, optical computing, optical interconnection, etc. Among proposed optimization methods, GA has become popular due to its capability of reaching nearly global. However, there exits a drawback to consider when we use the genetic algorithm. It is the large amount of computation time to construct desired holograms. One of the major reasons that the GA' s operation may be time intensive results from the expense of computing the cost function that must Fourier transform the parameters encoded on the hologram into the fitness value. In trying to remedy this drawback, Artificial Neural Network (ANN) has been put forward, allowing CGHs to be created easily and quickly (1), but the quality of reconstructed images is not high enough to use in applications of high preciseness. For that, we are in attempt to find a new approach of combiningthe good properties and performance of both the GA and ANN to make CGHs of high diffraction efficiency in a short time. The optimization of CGH using the genetic algorithm is merely a process of iteration, including selection, crossover, and mutation operators [2]. It is worth noting that the evaluation of the cost function with the aim of selecting better holograms plays an important role in the implementation of the GA. However, this evaluation process wastes much time for Fourier transforming the encoded parameters on the hologram into the value to be solved. Depending on the speed of computer, this process can even last up to ten minutes. It will be more effective if instead of merely generating random holograms in the initial process, a set of approximately desired holograms is employed. By doing so, the initial population will contain less trial holograms equivalent to the reduction of the computation time of GA's. Accordingly, a hybrid algorithm that utilizes a trained neural network to initiate the GA's procedure is proposed. Consequently, the initial population contains less random holograms and is compensated by approximately desired holograms. Figure 1 is the flowchart of the hybrid algorithm in comparison with the classical GA. The procedure of synthesizing a hologram on computer is divided into two steps. First the simulation of holograms based on ANN method [1] to acquire approximately desired holograms is carried. With a teaching data set of 9 characters obtained from the classical GA, the number of layer is 3, the number of hidden node is 100, learning rate is 0.3, and momentum is 0.5, the artificial neural network trained enables us to attain the approximately desired holograms, which are fairly good agreement with what we suggested in the theory. The second step, effect of several parameters on the operation of the hybrid algorithm is investigated. In principle, the operation of the hybrid algorithm and GA are the same except the modification of the initial step. Hence, the verified results in Ref [2] of the parameters such as the probability of crossover and mutation, the tournament size, and the crossover block size are remained unchanged, beside of the reduced population size. The reconstructed image of 76.4% diffraction efficiency and 5.4% uniformity is achieved when the population size is 30, the iteration number is 2000, the probability of crossover is 0.75, and the probability of mutation is 0.001. A comparison between the hybrid algorithm and GA in term of diffraction efficiency and computation time is also evaluated as shown in Fig. 2. With a 66.7% reduction in computation time and a 2% increase in diffraction efficiency compared to the GA method, the hybrid algorithm demonstrates its efficient performance. In the optical experiment, the phase holograms were displayed on a programmable phase modulator (model XGA). Figures 3 are pictures of diffracted patterns of the letter "0" from the holograms generated using the hybrid algorithm. Diffraction efficiency of 75.8% and uniformity of 5.8% are measured. We see that the simulation and experiment results are fairly good agreement with each other. In this paper, Genetic Algorithm and Neural Network have been successfully combined in designing CGHs. This method gives a significant reduction in computation time compared to the GA method while still allowing holograms of high diffraction efficiency and uniformity to be achieved. This work was supported by No.mOl-2001-000-00324-0 (2002)) from the Korea Science & Engineering Foundation.

  • PDF

An Outlier Detection Using Autoencoder for Ocean Observation Data (해양 이상 자료 탐지를 위한 오토인코더 활용 기법 최적화 연구)

  • Kim, Hyeon-Jae;Kim, Dong-Hoon;Lim, Chaewook;Shin, Yongtak;Lee, Sang-Chul;Choi, Youngjin;Woo, Seung-Buhm
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.33 no.6
    • /
    • pp.265-274
    • /
    • 2021
  • Outlier detection research in ocean data has traditionally been performed using statistical and distance-based machine learning algorithms. Recently, AI-based methods have received a lot of attention and so-called supervised learning methods that require classification information for data are mainly used. This supervised learning method requires a lot of time and costs because classification information (label) must be manually designated for all data required for learning. In this study, an autoencoder based on unsupervised learning was applied as an outlier detection to overcome this problem. For the experiment, two experiments were designed: one is univariate learning, in which only SST data was used among the observation data of Deokjeok Island and the other is multivariate learning, in which SST, air temperature, wind direction, wind speed, air pressure, and humidity were used. Period of data is 25 years from 1996 to 2020, and a pre-processing considering the characteristics of ocean data was applied to the data. An outlier detection of actual SST data was tried with a learned univariate and multivariate autoencoder. We tried to detect outliers in real SST data using trained univariate and multivariate autoencoders. To compare model performance, various outlier detection methods were applied to synthetic data with artificially inserted errors. As a result of quantitatively evaluating the performance of these methods, the multivariate/univariate accuracy was about 96%/91%, respectively, indicating that the multivariate autoencoder had better outlier detection performance. Outlier detection using an unsupervised learning-based autoencoder is expected to be used in various ways in that it can reduce subjective classification errors and cost and time required for data labeling.

Performance Optimization of Numerical Ocean Modeling on Cloud Systems (클라우드 시스템에서 해양수치모델 성능 최적화)

  • JUNG, KWANGWOOG;CHO, YANG-KI;TAK, YONG-JIN
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.27 no.3
    • /
    • pp.127-143
    • /
    • 2022
  • Recently, many attempts to run numerical ocean models in cloud computing environments have been tried actively. A cloud computing environment can be an effective means to implement numerical ocean models requiring a large-scale resource or quickly preparing modeling environment for global or large-scale grids. Many commercial and private cloud computing systems provide technologies such as virtualization, high-performance CPUs and instances, ether-net based high-performance-networking, and remote direct memory access for High Performance Computing (HPC). These new features facilitate ocean modeling experimentation on commercial cloud computing systems. Many scientists and engineers expect cloud computing to become mainstream in the near future. Analysis of the performance and features of commercial cloud services for numerical modeling is essential in order to select appropriate systems as this can help to minimize execution time and the amount of resources utilized. The effect of cache memory is large in the processing structure of the ocean numerical model, which processes input/output of data in a multidimensional array structure, and the speed of the network is important due to the communication characteristics through which a large amount of data moves. In this study, the performance of the Regional Ocean Modeling System (ROMS), the High Performance Linpack (HPL) benchmarking software package, and STREAM, the memory benchmark were evaluated and compared on commercial cloud systems to provide information for the transition of other ocean models into cloud computing. Through analysis of actual performance data and configuration settings obtained from virtualization-based commercial clouds, we evaluated the efficiency of the computer resources for the various model grid sizes in the virtualization-based cloud systems. We found that cache hierarchy and capacity are crucial in the performance of ROMS using huge memory. The memory latency time is also important in the performance. Increasing the number of cores to reduce the running time for numerical modeling is more effective with large grid sizes than with small grid sizes. Our analysis results will be helpful as a reference for constructing the best computing system in the cloud to minimize time and cost for numerical ocean modeling.

The Trend of Cigarette Design and Tobacco Flavor System Development

  • Wu, Jimmy Z.
    • Journal of the Korean Society of Tobacco Science
    • /
    • v.24 no.1
    • /
    • pp.67-73
    • /
    • 2002
  • In light of addressing consumer health concern, coping with anti-tobacco movement, and promoting new product, tobacco industry is actively pursuing to make a new generation of cigarettes with low tar and nicotine deliveries, and less harmful substances. Low tar and low nicotine cigarettes increases their market shares dramatically world wide, especially in KT&G, multinational tobacco companies, EU countries, even in China regulated by CNTC to set up yearly target to lower tar and nicotine deliveries. On the other hand, to design a new cigarette with reduced harmful substances begins to gain speed. The "modified Hoffmann list" publishes thirty plus substances in tobacco leaf and main smoke stream, which is the prime suspect causing health problems. Various ways and means are developed to reduce such components including new tobacco breeds, new curing method, tobacco leaf treatment before processing, selected filtration system, innovated casing system to reduce free radicals, as well as some non conventional cigarette products. In TSRC held this year, the main topic is related to reduce tobacco specific nitrosamines in tobacco leaf. The new generation of cigarette is in the horizon. It still needs a lot help to produce commercial products with satisfied taste and aroma characters. The flavor industry is not regulated by many governments demanding which ingredients might or might not be for tobacco use. However, most of the cigarette companies self impose a list of ingredients to guide flavor suppliers to design flavors. Unfortunately, the number of ingredients in those lists is getting shorter every year. It is understandable that the health is not the only reason. Some cigarette companies are playing safe to protect the company from potential lawsuit, while others are just copying from their competitors. Moreover, it is obvious that it needs more assistance from casings and flavors to design new generation of cigarettes with missing certain flavor components in tobacco leaf and main smoke stream. These flavor components are either non-existed or at lower level at new form of cured tobacco leaf or filtered in the main smoke stream along with reduced harmful substances. The use of carbon filters and other selected filtration system poses another tough task for flavor system design. Specific flavor components are missing from the smoke analysis data, which brings a notion of "carbon taste" and "dryness" of mouth feel. It is ever more demanded by cigarette industry to flavor suppliers to produce flavors as body enhancer, tobacco notes, salivating agents, harshness reducer, and various of aromatic notes provided they are safe to use. Another trend is that water based flavor or flavor with reduced ethanol as solvent is gaining popularity. It is preferred by some cigarette companies that the flavor is compounded with all natural ingredients or all ingredients should he GMO free. The new generation of cigarettes demands many ways of new thinking process. It is also vital for tobacco industry. It reflects the real needs for the consumers that the cigarette product should be safe to use as well as bearing the taste and aroma characters smokers always enjoyed. An effective tobacco flavor system is definitely a part of the equation. The global trend of tobacco industry is like trends of any other industries lead by consumer needs, benefited with new technology availability, affected by the global economy, and subjected for various rules and regulations. Anti-tobacco organizations and media exceptionally scrutinize cigarette, as a legal commercial product. Cigarette is probably the most studied commercial product for its composition, structure, deliveries, effects, as well as its new developmental trend. Therefore, any new trend of cigarette development would be within these boundaries. This paper is trying to point out what it would be like for tobacco industry in the next few yews and what concerns the tobacco industry. It focuses mostly on the efforts to produce safer cigarettes. It is such a vital task for the tobacco industry and its affiliate industries such as cigarette papers, filters, flavors, and other materials. The facts and knowledge presented in this paper might be well known for the public. Some of the comments and predictions are very much personal opinion for a further discussion.