• Title/Summary/Keyword: Processing Speed

Search Result 4,314, Processing Time 0.049 seconds

Enhanced Production of Carboxymethylcellulase by a Newly Isolated Marine Microorganism Bacillus atrophaeus LBH-18 Using Rice Bran, a Byproduct from the Rice Processing Industry (미강을 이용한 해양미생물 Bacillus atrophaeus LBH-18 유래의 carboxymethylcellulase 생산의 최적화)

  • Kim, Yi-Joon;Cao, Wa;Lee, Yu-Jeong;Lee, Sang-Un;Jeong, Jeong-Han;Lee, Jin-Woo
    • Journal of Life Science
    • /
    • v.22 no.10
    • /
    • pp.1295-1306
    • /
    • 2012
  • A microorganism producing carboxymethylcellulase (CMCase) was isolated from seawater and identified as Bacillus atrophaeus. This species was designated as B. atrophaeus LBH-18 based on its evolutionary distance and the phylogenetic tree resulting from 16S rDNA sequencing and the neighbor-joining method. The optimal conditions for rice bran (68.1 g/l), peptone (9.1 g/l), and initial pH (7.0) of the medium for cell growth was determined by Design Expert Software based on the response surface method; conditions for production of CMCase were 55.2 g/l, 6.6 g/l, and 7.1, respectively. The optimal temperature for cell growth and the production of CMCase by B. atrophaeus LBH-18 was $30^{\circ}C$. The optimal conditions of agitation speed and aeration rate for cell growth in a 7-l bioreactor were 324 rpm and 0.9 vvm, respectively, whereas those for production of CMCase were 343 rpm and 0.6 vvm, respectively. The optimal inner pressure for cell growth and production of CMCase in a 100-l bioreactor was 0.06 MPa. Maximal production of CMCase under optimal conditions in a 100-l bioreactor was 127.5 U/ml, which was 1.32 times higher than that without an inner pressure. In this study, rice bran was developed as a carbon source for industrial scale production of CMCase by B. atrophaeus LBH-18. Reduced time for the production of CMCase from 7 to 10 days to 3 days by using a bacterial strain with submerged fermentation also resulted in increased productivity of CMCase and a decrease in its production cost.

X-tree Diff: An Efficient Change Detection Algorithm for Tree-structured Data (X-tree Diff: 트리 기반 데이터를 위한 효율적인 변화 탐지 알고리즘)

  • Lee, Suk-Kyoon;Kim, Dong-Ah
    • The KIPS Transactions:PartC
    • /
    • v.10C no.6
    • /
    • pp.683-694
    • /
    • 2003
  • We present X-tree Diff, a change detection algorithm for tree-structured data. Our work is motivated by need to monitor massive volume of web documents and detect suspicious changes, called defacement attack on web sites. From this context, our algorithm should be very efficient in speed and use of memory space. X-tree Diff uses a special ordered labeled tree, X-tree, to represent XML/HTML documents. X-tree nodes have a special field, tMD, which stores a 128-bit hash value representing the structure and data of subtrees, so match identical subtrees form the old and new versions. During this process, X-tree Diff uses the Rule of Delaying Ambiguous Matchings, implying that it perform exact matching where a node in the old version has one-to one corrspondence with the corresponding node in the new, by delaying all the others. It drastically reduces the possibility of wrong matchings. X-tree Diff propagates such exact matchings upwards in Step 2, and obtain more matchings downwsards from roots in Step 3. In step 4, nodes to ve inserted or deleted are decided, We aldo show thst X-tree Diff runs on O(n), woere n is the number of noses in X-trees, in worst case as well as in average case, This result is even better than that of BULD Diff algorithm, which is O(n log(n)) in worst case, We experimented X-tree Diff on reat data, which are about 11,000 home pages from about 20 wev sites, instead of synthetic documets manipulated for experimented for ex[erimentation. Currently, X-treeDiff algorithm is being used in a commeercial hacking detection system, called the WIDS(Web-Document Intrusion Detection System), which is to find changes occured in registered websites, and report suspicious changes to users.

Processing and Quality Control of Flux Data at Gwangneung Forest (광릉 산림의 플럭스 자료 처리와 품질 관리)

  • Lim, Hee-Jeong;Lee, Young-Hee
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.10 no.3
    • /
    • pp.82-93
    • /
    • 2008
  • In order to ensure a standardized data analysis of the eddy covariance measurements, Hong and Kim's quality control program has been updated and used to process eddy covariance data measured at two levels on the main flux tower at Gwangneung site from January to May in 2005. The updated program was allowed to remove outliers automatically for $CO_2$ and latent heat fluxes. The flag system consists of four quality groups(G, D, B and M). During the study period, the missing data were about 25% of the total records. About 60% of the good quality data were obtained after the quality control. The number of record in G group was larger at 40m than at 20m. It is due that the level of 20m was within the roughness sublayer where the presence of the canopy influences directly on the character of the turbulence. About 60% of the bad data were due to low wind speed. Energy balance closure at this site was about 40% during the study period. Large imbalance is attributed partly to the combined effects of the neglected heat storage terms, inaccuracy of ground heat flux and advection due to local wind system near the surface. The analysis of wind direction indicates that the frequent occurrence of positive momentum flux was closely associated with mountain valley wind system at this site. The negative $CO_2$ flux at night was examined in terms of averaging time. The results show that when averaging time is larger than 10min, the magnitude of calculated $CO_2$ fluxes increases rapidly, suggesting that the 30min $CO_2$ flux is influenced severely by the mesoscale motion or nonstationarity. A proper choice of averaging time needs to be considered to get accurate turbulent fluxes during nighttime.

Design of MAHA Supercomputing System for Human Genome Analysis (대용량 유전체 분석을 위한 고성능 컴퓨팅 시스템 MAHA)

  • Kim, Young Woo;Kim, Hong-Yeon;Bae, Seungjo;Kim, Hag-Young;Woo, Young-Choon;Park, Soo-Jun;Choi, Wan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.2
    • /
    • pp.81-90
    • /
    • 2013
  • During the past decade, many changes and attempts have been tried and are continued developing new technologies in the computing area. The brick wall in computing area, especially power wall, changes computing paradigm from computing hardwares including processor and system architecture to programming environment and application usage. The high performance computing (HPC) area, especially, has been experienced catastrophic changes, and it is now considered as a key to the national competitiveness. In the late 2000's, many leading countries rushed to develop Exascale supercomputing systems, and as a results tens of PetaFLOPS system are prevalent now. In Korea, ICT is well developed and Korea is considered as a one of leading countries in the world, but not for supercomputing area. In this paper, we describe architecture design of MAHA supercomputing system which is aimed to develop 300 TeraFLOPS system for bio-informatics applications like human genome analysis and protein-protein docking. MAHA supercomputing system is consists of four major parts - computing hardware, file system, system software and bio-applications. MAHA supercomputing system is designed to utilize heterogeneous computing accelerators (co-processors like GPGPUs and MICs) to get more performance/$, performance/area, and performance/power. To provide high speed data movement and large capacity, MAHA file system is designed to have asymmetric cluster architecture, and consists of metadata server, data server, and client file system on top of SSD and MAID storage servers. MAHA system softwares are designed to provide user-friendliness and easy-to-use based on integrated system management component - like Bio Workflow management, Integrated Cluster management and Heterogeneous Resource management. MAHA supercomputing system was first installed in Dec., 2011. The theoretical performance of MAHA system was 50 TeraFLOPS and measured performance of 30.3 TeraFLOPS with 32 computing nodes. MAHA system will be upgraded to have 100 TeraFLOPS performance at Jan., 2013.

A hybrid algorithm for the synthesis of computer-generated holograms

  • Nguyen The Anh;An Jun Won;Choe Jae Gwang;Kim Nam
    • Proceedings of the Optical Society of Korea Conference
    • /
    • 2003.07a
    • /
    • pp.60-61
    • /
    • 2003
  • A new approach to reduce the computation time of genetic algorithm (GA) for making binary phase holograms is described. Synthesized holograms having diffraction efficiency of 75.8% and uniformity of 5.8% are proven in computer simulation and experimentally demonstrated. Recently, computer-generated holograms (CGHs) having high diffraction efficiency and flexibility of design have been widely developed in many applications such as optical information processing, optical computing, optical interconnection, etc. Among proposed optimization methods, GA has become popular due to its capability of reaching nearly global. However, there exits a drawback to consider when we use the genetic algorithm. It is the large amount of computation time to construct desired holograms. One of the major reasons that the GA' s operation may be time intensive results from the expense of computing the cost function that must Fourier transform the parameters encoded on the hologram into the fitness value. In trying to remedy this drawback, Artificial Neural Network (ANN) has been put forward, allowing CGHs to be created easily and quickly (1), but the quality of reconstructed images is not high enough to use in applications of high preciseness. For that, we are in attempt to find a new approach of combiningthe good properties and performance of both the GA and ANN to make CGHs of high diffraction efficiency in a short time. The optimization of CGH using the genetic algorithm is merely a process of iteration, including selection, crossover, and mutation operators [2]. It is worth noting that the evaluation of the cost function with the aim of selecting better holograms plays an important role in the implementation of the GA. However, this evaluation process wastes much time for Fourier transforming the encoded parameters on the hologram into the value to be solved. Depending on the speed of computer, this process can even last up to ten minutes. It will be more effective if instead of merely generating random holograms in the initial process, a set of approximately desired holograms is employed. By doing so, the initial population will contain less trial holograms equivalent to the reduction of the computation time of GA's. Accordingly, a hybrid algorithm that utilizes a trained neural network to initiate the GA's procedure is proposed. Consequently, the initial population contains less random holograms and is compensated by approximately desired holograms. Figure 1 is the flowchart of the hybrid algorithm in comparison with the classical GA. The procedure of synthesizing a hologram on computer is divided into two steps. First the simulation of holograms based on ANN method [1] to acquire approximately desired holograms is carried. With a teaching data set of 9 characters obtained from the classical GA, the number of layer is 3, the number of hidden node is 100, learning rate is 0.3, and momentum is 0.5, the artificial neural network trained enables us to attain the approximately desired holograms, which are fairly good agreement with what we suggested in the theory. The second step, effect of several parameters on the operation of the hybrid algorithm is investigated. In principle, the operation of the hybrid algorithm and GA are the same except the modification of the initial step. Hence, the verified results in Ref [2] of the parameters such as the probability of crossover and mutation, the tournament size, and the crossover block size are remained unchanged, beside of the reduced population size. The reconstructed image of 76.4% diffraction efficiency and 5.4% uniformity is achieved when the population size is 30, the iteration number is 2000, the probability of crossover is 0.75, and the probability of mutation is 0.001. A comparison between the hybrid algorithm and GA in term of diffraction efficiency and computation time is also evaluated as shown in Fig. 2. With a 66.7% reduction in computation time and a 2% increase in diffraction efficiency compared to the GA method, the hybrid algorithm demonstrates its efficient performance. In the optical experiment, the phase holograms were displayed on a programmable phase modulator (model XGA). Figures 3 are pictures of diffracted patterns of the letter "0" from the holograms generated using the hybrid algorithm. Diffraction efficiency of 75.8% and uniformity of 5.8% are measured. We see that the simulation and experiment results are fairly good agreement with each other. In this paper, Genetic Algorithm and Neural Network have been successfully combined in designing CGHs. This method gives a significant reduction in computation time compared to the GA method while still allowing holograms of high diffraction efficiency and uniformity to be achieved. This work was supported by No.mOl-2001-000-00324-0 (2002)) from the Korea Science & Engineering Foundation.

  • PDF

An Outlier Detection Using Autoencoder for Ocean Observation Data (해양 이상 자료 탐지를 위한 오토인코더 활용 기법 최적화 연구)

  • Kim, Hyeon-Jae;Kim, Dong-Hoon;Lim, Chaewook;Shin, Yongtak;Lee, Sang-Chul;Choi, Youngjin;Woo, Seung-Buhm
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.33 no.6
    • /
    • pp.265-274
    • /
    • 2021
  • Outlier detection research in ocean data has traditionally been performed using statistical and distance-based machine learning algorithms. Recently, AI-based methods have received a lot of attention and so-called supervised learning methods that require classification information for data are mainly used. This supervised learning method requires a lot of time and costs because classification information (label) must be manually designated for all data required for learning. In this study, an autoencoder based on unsupervised learning was applied as an outlier detection to overcome this problem. For the experiment, two experiments were designed: one is univariate learning, in which only SST data was used among the observation data of Deokjeok Island and the other is multivariate learning, in which SST, air temperature, wind direction, wind speed, air pressure, and humidity were used. Period of data is 25 years from 1996 to 2020, and a pre-processing considering the characteristics of ocean data was applied to the data. An outlier detection of actual SST data was tried with a learned univariate and multivariate autoencoder. We tried to detect outliers in real SST data using trained univariate and multivariate autoencoders. To compare model performance, various outlier detection methods were applied to synthetic data with artificially inserted errors. As a result of quantitatively evaluating the performance of these methods, the multivariate/univariate accuracy was about 96%/91%, respectively, indicating that the multivariate autoencoder had better outlier detection performance. Outlier detection using an unsupervised learning-based autoencoder is expected to be used in various ways in that it can reduce subjective classification errors and cost and time required for data labeling.

Performance Optimization of Numerical Ocean Modeling on Cloud Systems (클라우드 시스템에서 해양수치모델 성능 최적화)

  • JUNG, KWANGWOOG;CHO, YANG-KI;TAK, YONG-JIN
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.27 no.3
    • /
    • pp.127-143
    • /
    • 2022
  • Recently, many attempts to run numerical ocean models in cloud computing environments have been tried actively. A cloud computing environment can be an effective means to implement numerical ocean models requiring a large-scale resource or quickly preparing modeling environment for global or large-scale grids. Many commercial and private cloud computing systems provide technologies such as virtualization, high-performance CPUs and instances, ether-net based high-performance-networking, and remote direct memory access for High Performance Computing (HPC). These new features facilitate ocean modeling experimentation on commercial cloud computing systems. Many scientists and engineers expect cloud computing to become mainstream in the near future. Analysis of the performance and features of commercial cloud services for numerical modeling is essential in order to select appropriate systems as this can help to minimize execution time and the amount of resources utilized. The effect of cache memory is large in the processing structure of the ocean numerical model, which processes input/output of data in a multidimensional array structure, and the speed of the network is important due to the communication characteristics through which a large amount of data moves. In this study, the performance of the Regional Ocean Modeling System (ROMS), the High Performance Linpack (HPL) benchmarking software package, and STREAM, the memory benchmark were evaluated and compared on commercial cloud systems to provide information for the transition of other ocean models into cloud computing. Through analysis of actual performance data and configuration settings obtained from virtualization-based commercial clouds, we evaluated the efficiency of the computer resources for the various model grid sizes in the virtualization-based cloud systems. We found that cache hierarchy and capacity are crucial in the performance of ROMS using huge memory. The memory latency time is also important in the performance. Increasing the number of cores to reduce the running time for numerical modeling is more effective with large grid sizes than with small grid sizes. Our analysis results will be helpful as a reference for constructing the best computing system in the cloud to minimize time and cost for numerical ocean modeling.

The Trend of Cigarette Design and Tobacco Flavor System Development

  • Wu, Jimmy Z.
    • Journal of the Korean Society of Tobacco Science
    • /
    • v.24 no.1
    • /
    • pp.67-73
    • /
    • 2002
  • In light of addressing consumer health concern, coping with anti-tobacco movement, and promoting new product, tobacco industry is actively pursuing to make a new generation of cigarettes with low tar and nicotine deliveries, and less harmful substances. Low tar and low nicotine cigarettes increases their market shares dramatically world wide, especially in KT&G, multinational tobacco companies, EU countries, even in China regulated by CNTC to set up yearly target to lower tar and nicotine deliveries. On the other hand, to design a new cigarette with reduced harmful substances begins to gain speed. The "modified Hoffmann list" publishes thirty plus substances in tobacco leaf and main smoke stream, which is the prime suspect causing health problems. Various ways and means are developed to reduce such components including new tobacco breeds, new curing method, tobacco leaf treatment before processing, selected filtration system, innovated casing system to reduce free radicals, as well as some non conventional cigarette products. In TSRC held this year, the main topic is related to reduce tobacco specific nitrosamines in tobacco leaf. The new generation of cigarette is in the horizon. It still needs a lot help to produce commercial products with satisfied taste and aroma characters. The flavor industry is not regulated by many governments demanding which ingredients might or might not be for tobacco use. However, most of the cigarette companies self impose a list of ingredients to guide flavor suppliers to design flavors. Unfortunately, the number of ingredients in those lists is getting shorter every year. It is understandable that the health is not the only reason. Some cigarette companies are playing safe to protect the company from potential lawsuit, while others are just copying from their competitors. Moreover, it is obvious that it needs more assistance from casings and flavors to design new generation of cigarettes with missing certain flavor components in tobacco leaf and main smoke stream. These flavor components are either non-existed or at lower level at new form of cured tobacco leaf or filtered in the main smoke stream along with reduced harmful substances. The use of carbon filters and other selected filtration system poses another tough task for flavor system design. Specific flavor components are missing from the smoke analysis data, which brings a notion of "carbon taste" and "dryness" of mouth feel. It is ever more demanded by cigarette industry to flavor suppliers to produce flavors as body enhancer, tobacco notes, salivating agents, harshness reducer, and various of aromatic notes provided they are safe to use. Another trend is that water based flavor or flavor with reduced ethanol as solvent is gaining popularity. It is preferred by some cigarette companies that the flavor is compounded with all natural ingredients or all ingredients should he GMO free. The new generation of cigarettes demands many ways of new thinking process. It is also vital for tobacco industry. It reflects the real needs for the consumers that the cigarette product should be safe to use as well as bearing the taste and aroma characters smokers always enjoyed. An effective tobacco flavor system is definitely a part of the equation. The global trend of tobacco industry is like trends of any other industries lead by consumer needs, benefited with new technology availability, affected by the global economy, and subjected for various rules and regulations. Anti-tobacco organizations and media exceptionally scrutinize cigarette, as a legal commercial product. Cigarette is probably the most studied commercial product for its composition, structure, deliveries, effects, as well as its new developmental trend. Therefore, any new trend of cigarette development would be within these boundaries. This paper is trying to point out what it would be like for tobacco industry in the next few yews and what concerns the tobacco industry. It focuses mostly on the efforts to produce safer cigarettes. It is such a vital task for the tobacco industry and its affiliate industries such as cigarette papers, filters, flavors, and other materials. The facts and knowledge presented in this paper might be well known for the public. Some of the comments and predictions are very much personal opinion for a further discussion.

Application of Remote Sensing Techniques to Survey and Estimate the Standing-Stock of Floating Debris in the Upper Daecheong Lake (원격탐사 기법 적용을 통한 대청호 상류 유입 부유쓰레기 조사 및 현존량 추정 연구)

  • Youngmin Kim;Seon Woong Jang ;Heung-Min Kim;Tak-Young Kim;Suho Bak
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.5_1
    • /
    • pp.589-597
    • /
    • 2023
  • Floating debris in large quantities from land during heavy rainfall has adverse social, economic, and environmental impacts, but the monitoring system for the concentration area and amount is insufficient. In this study, we proposed an efficient monitoring method for floating debris entering the river during heavy rainfall in Daecheong Lake, the largest water supply source in the central region, and applied remote sensing techniques to estimate the standing-stock of floating debris. To investigate the status of floating debris in the upper of Daecheong Lake, we used a tracking buoy equipped with a low-orbit satellite communication terminal to identify the movement route and behavior characteristics, and used a drone to estimate the potential concentration area and standing-stock of floating debris. The location tracking buoys moved rapidly during the period when the cumulative rainfall for 3 days increased by more than 200 to 300 mm. In the case of Hotan Bridge, which showed the longest distance, it moved about 72.8 km for one day, and the maximum moving speed at this time was 5.71 km/h. As a result of calculating the standing-stock of floating debris using a drone after heavy rainfall, it was found to be 658.8 to 9,165.4 tons, with the largest amount occurring in the Seokhori area. In this study, we were able to identify the main concentrations of floating debris by using location-tracking buoys and drones. It is believed that remote sensing-based monitoring methods, which are more mobile and quicker than traditional monitoring methods, can contribute to reducing the cost of collecting and processing large amounts of floating debris that flows in during heavy rain periods in the future.

Effect of Accelerated Storage on the Microstructure and Water Absorption Characteristics of Korean Adzuki Bean (Vigna angularis L.) Cultivar (팥의 가속화 저장에 따른 미세구조 및 수분흡수 특성)

  • Jieun Kwak;Seon-Min Oh;You-Geun Oh;Yu-Chan Choi;Hyun-Jin Park;Suk-Bo Song;Jeong-Heui Lee;Jeom-Sig Lee
    • KOREAN JOURNAL OF CROP SCIENCE
    • /
    • v.68 no.3
    • /
    • pp.167-174
    • /
    • 2023
  • This study investigated the microstructure and water absorption characteristics of the Korean adzuki bean (Vigna angularis L.) cultivar under accelerated storage. The germination rate, acid value, redness (a*), and yellowness (b*) values showed no significant differences after three months of storage compared to pre-storage under low temperatures (4℃). However, a statistically significant difference was observed under accelerated high temperatures (45℃). In particular, after storage for three months, the germination rate and acid value were 0% and 33.63 mg KOH/100g, respectively, under accelerated high temperatures. After storage for three months, the holes, hilum damage, and spaces between the seed coat and cotyledon shortened the time and speed of water absorption under accelerated high temperatures compared to that under low temperatures. Conversely, further research is required to investigate the reason for the low rate of parallel water absorption.