• Title/Summary/Keyword: and parallel processing

Search Result 2,006, Processing Time 0.032 seconds

Benchmark Results of a Monte Carlo Treatment Planning system (몬데카를로 기반 치료계획시스템의 성능평가)

  • Cho, Byung-Chul
    • Progress in Medical Physics
    • /
    • v.13 no.3
    • /
    • pp.149-155
    • /
    • 2002
  • Recent advances in radiation transport algorithms, computer hardware performance, and parallel computing make the clinical use of Monte Carlo based dose calculations possible. To compare the speed and accuracies of dose calculations between different developed codes, a benchmark tests were proposed at the XIIth ICCR (International Conference on the use of Computers in Radiation Therapy, Heidelberg, Germany 2000). A Monte Carlo treatment planning comprised of 28 various Intel Pentium CPUs was implemented for routine clinical use. The purpose of this study was to evaluate the performance of our system using the above benchmark tests. The benchmark procedures are comprised of three parts. a) speed of photon beams dose calculation inside a given phantom of 30.5 cm$\times$39.5 cm $\times$ 30 cm deep and filled with 5 ㎣ voxels within 2% statistical uncertainty. b) speed of electron beams dose calculation inside the same phantom as that of the photon beams. c) accuracy of photon and electron beam calculation inside heterogeneous slab phantom compared with the reference results of EGS4/PRESTA calculation. As results of the speed benchmark tests, it took 5.5 minutes to achieve less than 2% statistical uncertainty for 18 MV photon beams. Though the net calculation for electron beams was an order of faster than the photon beam, the overall calculation time was similar to that of photon beam case due to the overhead time to maintain parallel processing. Since our Monte Carlo code is EGSnrc, which is an improved version of EGS4, the accuracy tests of our system showed, as expected, very good agreement with the reference data. In conclusion, our Monte Carlo treatment planning system shows clinically meaningful results. Though other more efficient codes are developed such like MCDOSE and VMC++, BEAMnrc based on EGSnrc code system may be used for routine clinical Monte Carlo treatment planning in conjunction with clustering technique.

  • PDF

Evaluation of MR-SENSE Reconstruction by Filtering Effect and Spatial Resolution of the Sensitivity Map for the Simulation-Based Linear Coil Array (선형적 위상배열 코일구조의 시뮬레이션을 통한 민감도지도의 공간 해상도 및 필터링 변화에 따른 MR-SENSE 영상재구성 평가)

  • Lee, D.H.;Hong, C.P.;Han, B.S.;Kim, H.J.;Suh, J.J.;Kim, S.H.;Lee, C.H.;Lee, M.W.
    • Journal of Biomedical Engineering Research
    • /
    • v.32 no.3
    • /
    • pp.245-250
    • /
    • 2011
  • Parallel imaging technique can provide several advantages for a multitude of MRI applications. Especially, in SENSE technique, sensitivity maps were always required in order to determine the reconstruction matrix, therefore, a number of difference approaches using sensitivity information from coils have been demonstrated to improve of image quality. Moreover, many filtering methods were proposed such as adaptive matched filter and nonlinear diffusion technique to optimize the suppression of background noise and to improve of image quality. In this study, we performed SENSE reconstruction using computer simulations to confirm the most suitable method for the feasibility of filtering effect and according to changing order of polynomial fit that were applied on variation of spatial resolution of sensitivity map. The image was obtained at 0.32T(Magfinder II, Genpia, Korea) MRI system using spin-echo pulse sequence(TR/TE = 500/20 ms, FOV = 300 mm, matrix = $128{\times}128$, thickness = 8 mm). For the simulation, obtained image was multiplied with four linear-array coil sensitivities which were formed of 2D-gaussian distribution and the image was complex white gaussian noise was added. Image processing was separated to apply two methods which were polynomial fitting and filtering according to spatial resolution of sensitivity map and each coil image was subsampled corresponding to reduction factor(r-factor) of 2 and 4. The results were compared to mean value of geomety factor(g-factor) and artifact power(AP) according to r-factor 2 and 4. Our results were represented while changing of spatial resolution of sensitivity map and r-factor, polynomial fit methods were represented the better results compared with general filtering methods. Although our result had limitation of computer simulation study instead of applying to experiment and coil geometric array such as linear, our method may be useful for determination of optimal sensitivity map in a linear coil array.

Parallel Computation For The Edit Distance Based On The Four-Russians' Algorithm (4-러시안 알고리즘 기반의 편집거리 병렬계산)

  • Kim, Young Ho;Jeong, Ju-Hui;Kang, Dae Woong;Sim, Jeong Seop
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.2 no.2
    • /
    • pp.67-74
    • /
    • 2013
  • Approximate string matching problems have been studied in diverse fields. Recently, fast approximate string matching algorithms are being used to reduce the time and costs for the next generation sequencing. To measure the amounts of errors between two strings, we use a distance function such as the edit distance. Given two strings X(|X| = m) and Y(|Y| = n) over an alphabet ${\Sigma}$, the edit distance between X and Y is the minimum number of edit operations to convert X into Y. The edit distance between X and Y can be computed using the well-known dynamic programming technique in O(mn) time and space. The edit distance also can be computed using the Four-Russians' algorithm whose preprocessing step runs in $O((3{\mid}{\Sigma}{\mid})^{2t}t^2)$ time and $O((3{\mid}{\Sigma}{\mid})^{2t}t)$ space and the computation step runs in O(mn/t) time and O(mn) space where t represents the size of the block. In this paper, we present a parallelized version of the computation step of the Four-Russians' algorithm. Our algorithm computes the edit distance between X and Y in O(m+n) time using m/t threads. Then we implemented both the sequential version and our parallelized version of the Four-Russians' algorithm using CUDA to compare the execution times. When t = 1 and t = 2, our algorithm runs about 10 times and 3 times faster than the sequential algorithm, respectively.

New VLSI Architecture of Parallel Multiplier-Accumulator Based on Radix-2 Modified Booth Algorithm (Radix-2 MBA 기반 병렬 MAC의 VLSI 구조)

  • Seo, Young-Ho;Kim, Dong-Wook
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.45 no.4
    • /
    • pp.94-104
    • /
    • 2008
  • In this paper, we propose a new architecture of multiplier-and-accumulator (MAC) for high speed multiplication and accumulation arithmetic. By combining multiplication with accumulation and devising a hybrid type of carry save adder (CSA), the performance was improved. Since the accumulator which has the largest delay in MAC was removed and its function was included into CSA, the overall performance becomes to be elevated. The proposed CSA tree uses 1's complement-based radix-2 modified booth algorithm (MBA) and has the modified array for the sign extension in order to increase the bit density of operands. The CSA propagates the carries by the least significant bits of the partial products and generates the least significant bits in advance for decreasing the number of the input bits of the final adder. Also, the proposed MAC accumulates the intermediate results in the type of sum and carry bits not the output of the final adder for improving the performance by optimizing the efficiency of pipeline scheme. The proposed architecture was synthesized with $250{\mu}m,\;180{\mu}m,\;130{\mu}m$ and 90nm standard CMOS library after designing it. We analyzed the results such as hardware resource, delay, and pipeline which are based on the theoretical and experimental estimation. We used Sakurai's alpha power low for the delay modeling. The proposed MAC has the superior properties to the standard design in many ways and its performance is twice as much than the previous research in the similar clock frequency.

Optimized Hardware Design of Deblocking Filter for H.264/AVC (H.264/AVC를 위한 디블록킹 필터의 최적화된 하드웨어 설계)

  • Jung, Youn-Jin;Ryoo, Kwang-Ki
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.47 no.1
    • /
    • pp.20-27
    • /
    • 2010
  • This paper describes a design of 5-stage pipelined de-blocking filter with power reduction scheme and proposes a efficient memory architecture and filter order for high performance H.264/AVC Decoder. Generally the de-blocking filter removes block boundary artifacts and enhances image quality. Nevertheless filter has a few disadvantage that it requires a number of memory access and iterated operations because of filter operation for 4 time to one edge. So this paper proposes a optimized filter ordering and efficient hardware architecture for the reduction of memory access and total filter cycles. In proposed filter parallel processing is available because of structured 5-stage pipeline consisted of memory read, threshold decider, pre-calculation, filter operation and write back. Also it can reduce power consumption because it uses a clock gating scheme which disable unnecessary clock switching. Besides total number of filtering cycle is decreased by new filter order. The proposed filter is designed with Verilog-HDL and functionally verified with the whole H.264/AVC decoder using the Modelsim 6.2g simulator. Input vectors are QCIF images generated by JM9.4 standard encoder software. As a result of experiment, it shows that the filter can make about 20% total filter cycles reduction and it requires small transposition buffer size.

Multi-DOF Real-time Hybrid Dynamic Test of a Steel Frame Structure (강 뼈대 구조물의 다자유도 실시간 하이브리드 동적 실험)

  • Kim, Sehoon;Na, Okpin;Kim, Sungil
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.33 no.2
    • /
    • pp.443-453
    • /
    • 2013
  • The hybrid test is one of the most advanced test methods to predict the structural dynamic behavior with the interaction between a physical substructure and a numerical modeling in the hybrid control system. The purpose of this study is to perform the multi-directional dynamic test of a steel frame structure with the real-time hybrid system and to evaluate the validation of the results. In this study, FEAPH, nonlinear finite element analysis program for hybrid only, was developed and the hybrid control system was optimized. The inefficient computational time was improved with a fixed number iteration method and parallel computational techniques used in FEAPH. Furthermore, the previously used data communication method and the interface between a substructure and an analysis program were simplified in the control system. As the results, the total processing time in real-time hybrid test was shortened up to 10 times of actual measured seismic period. In order to verify the accuracy and validation of the hybrid system, the linear and nonlinear dynamic tests with a steel framed structure were carried out so that the trend of displacement responses was almost in accord with the numerical results. However, the maximum displacement responses had somewhat differences due to the analysis errors in material nonlinearities and the occurrence of permanent displacements. Therefore, if the proper material model and numerical algorithms are developed, the real-time hybrid system could be used to evaluate the structural dynamic behavior and would be an effective testing method as a substitute for a shaking table test.

Materials Compatibility and Structure Optimization of Test Department Probe for Quality Test of Fingerprint Sensor (지문인식센서 품질평가를 위한 검사부 프로브의 소재 적합성과 구조 최적화 연구)

  • Son, Eun-Won;Youn, Ji Won;Kim, Dae Up;Lim, Jae-Won;Kim, Kwang-Seok
    • Journal of the Microelectronics and Packaging Society
    • /
    • v.24 no.4
    • /
    • pp.73-77
    • /
    • 2017
  • Recently, fingerprint sensors have widely used for personal information security, and require quality evaluation to reduce an error of their recognition rate. Quality of fingerprint sensors is evaluated by variation of their electrical resistance introducing by contacts between a probe tip and a sensor electrode, Investigation on the materials compatability and structure optimization of probe is required to reduce deformation of sensor electrode for repeatability of quality testing. Nickel, steel(SK4), beryllium copper, and phosphor bronze were considered as probe materials, and beryllium copper was the most appropriate for materials of probe tips, considering indentation and contact resistance while being contacted probe tips on electrodes. Probes of an inspection part were manufactured with the single-unit structure for physical damage prevention and parallel processing capability. Inspection repeatability was evaluated by voltage variation of fingerprint sensors when the specific current was applied. A single-unit inspection part with beryllium copper probe tips showed excellent repeatability within ${\pm}0.003V$ of its voltage variation.

Potential of Onion Peel Extract as a Functional Ingredient for Functional Foods (건강기능성식품 기능성원료로서 양파껍질추출물의 품질특성)

  • Jeon, Seon-Young;Baek, Jeong-Hwa;Jeong, Eun-Jeong;Cha, Yong-Jun
    • Journal of Life Science
    • /
    • v.22 no.9
    • /
    • pp.1207-1213
    • /
    • 2012
  • Onion peels is a natural source of high-value functional ingredients produced in the onion industry without suitable processing. The objective of this study was to evaluate characteristics of onion peel extract (OPE), including its biological activities, obtained from solvent extraction in 3 times pilot scales (Lot A, B, and C). Mineral analysis showed that K was present in the largest amount (13,767.56-15,506.78 ppm), followed by Na and Ca at 8,602.44-9,796.00 ppm and 4,255.78-4,903.33 ppm, respectively. The amounts of total phenol, total flavonoid, and quercetin in the OPE were in the ranges of 598.57~626.73, 211.73~233.64, and 93.78~107.29 mg/g, respectively. The biological activities such as antioxidant and effects of fibrinolysis increased in parallel with the concentration of OPE. The $IC_{50}$ value of DPPH radical scavenging activity was in the range of 517.58~557.32 ppm in the OPE. The $IC_{50}$ value for superoxide dismutase (SOD)-like activity was in the range of 11,900.91~12,690.35 ppm. A clear zone of OPE (20,000 ppm) in fibrinolysis test was three times higher than the plasmin as a reference. In conclusion, OPE could be used as a good source of antioxidants and fibrinolytic activities.

The Relationship between Visual Stress and MBTI Personality Types (시각적 스트레스와 MBTI 성격유형과의 관계)

  • Kim, Sun-Uk;Han, Seung-Jo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.13 no.9
    • /
    • pp.4036-4044
    • /
    • 2012
  • This study is aimed to investigate the association between web-based visual stress and MBTI personality types. The stressor deriving visual stress is built by 14 vowels out of Korean alphabet as a content and parallel striples as a background on the screen, which is given to each subject during 5min. The dependent variable indicating how much human takes visual stress is the reduction rate of flicker fusion frequency, which is evaluated with visual flicker fusion frequency tester. The independent variables are gender and 8 MBTI personality types(E-I, S-N, T-F, and J-P), and hypotheses are based on human information processing model and previous studies. The results address that the reduction rate is not significantly affected by gender, S-N, and J-P, but E-I and T-F have significant influences on it. The reduction rate in I-type is almost 2 times as much as that in E-type and T-type has the rate 2.2 times more than F-type. This study can be applicable to determine the adequate personnel for jobs requiring less sensibility to visual stressors in areas that human error may lead to critical damages to an overall system.

DFT-spread OFDM Communication System for the Power Efficiency and Nonlinear Distortion in Underwater Communication (수중통신에서 비선형 왜곡과 전력효율을 위한 DFT-spread OFDM 통신 시스템)

  • Lee, Woo-Min;Ryn, Heung-Gyoon
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.8A
    • /
    • pp.777-784
    • /
    • 2010
  • Recently, the necessity of underwater communication and demand for transmitting and receiving various data such as voice or high resolution image data are increasing as well. The performance of underwater acoustic communication system is influenced by characteristics of the underwater communication channels. Especially, ISI(inter symbol interference) occurs because of delay spread according to multi-path and communication performance is degraded. In this paper, we study the OFDM technique to overcome the delay spread in underwater channel and by using CP, we compensate for delay spread. But PAPR which OFDM system has problem is very high. Therefore, we use DFT-spread OFDM method to avoid nonlinear distortion by high PAPR and to improve efficiency of amplifier. DFT-spread OFDM technique obtains high PAPR reduction effect because of each parallel data loads to all subcarrier by DFT spread processing before IFFT. In this paper, we show performance about delay spread through OFDM system and verify method that DFT spread OFDM is more suitable than OFDM for underwater communication. And we analyze performance according to two subcarrier mapping methods(Interleaved, Localized). Through the simulation results, performance of DFT spread OFDM is better about 5~6dB at $10^{-4}$ than OFDM. When compared to BER according to subcarrier mapping, Interleaved method is better about 3.5dB at $10^{-4}$ than Localized method.