• Title/Summary/Keyword: 성능최적화 기법

Search Result 1,330, Processing Time 0.026 seconds

An 8b 240 MS/s 1.36 ㎟ 104 mW 0.18 um CMOS ADC for High-Performance Display Applications (고성능 디스플레이 응용을 위한 8b 240 MS/s 1.36 ㎟ 104 mW 0.18 um CMOS ADC)

  • In Kyung-Hoon;Kim Se-Won;Cho Young-Jae;Moon Kyoung-Jun;Jee Yong;Lee Seung-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.42 no.1
    • /
    • pp.47-55
    • /
    • 2005
  • This work describes an 8b 240 MS/s CMOS ADC as one of embedded core cells for high-performance displays requiring low power and small size at high speed. The proposed ADC uses externally connected pins only for analog inputs, digital outputs, and supplies. The ADC employs (1) a two-step pipelined architecture to optimize power and chip size at the target sampling frequency of 240 MHz, (2) advanced bootstrapping techniques to achieve high signal bandwidth in the input SHA, and (3) RC filter-based on-chip I/V references to improve noise performance with a power-off function added for portable applications. The prototype ADC is implemented in a 0.18 um CMOS and simultaneously integrated in a DVD system with dual-mode inputs. The measured DNL and INL are within 0.49 LSB and 0.69 LSB, respectively. The prototype ADC shows the SFDR of 53 dB for a 10 MHz input sinewave at 240 MS/s while maintaining the SNDR exceeding 38 dB and the SFDR exceeding 50 dB for input frequencies up to the Nyquist frequency at 240 MS/s. The ADC consumes, 104 mW at 240 MS/s and the active die area is 1.36 ㎟.

A 14b 150MS/s 140mW $2.0mm^2$ 0.13um CMOS ADC for SDR (Software Defined Radio 시스템을 위한 14비트 150MS/s 140mW $2.0mm^2$ 0.13um CMOS A/D 변환기)

  • Yoo, Pil-Seon;Kim, Cha-Dong;Lee, Seung-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.45 no.4
    • /
    • pp.27-35
    • /
    • 2008
  • This work proposes a 14b 150MS/s 0.13um CMOS ADC for SDR systems requiring simultaneously high resolution, low power, and small size at high speed. The proposed ADC employs a calibration-free four-step pipeline architecture optimizing the scaling factor for the input trans-conductance of amplifiers and the sampling capacitance in each stage to minimize thermal noise effects and power consumption at the target resolution and sampling rate. A signal- insensitive 3-D fully symmetric layout achieves a 14b level resolution by reducing a capacitor mismatch of three MDACs. The proposed supply- and temperature- insensitive current and voltage references with on-chip RC filters minimizing the effect of switching noise are implemented with off-chip C filters. The prototype ADC in a 0.13um 1P8M CMOS technology demonstrates a measured DNL and INL within 0.81LSB and 2.83LSB, at 14b, respectively. The ADC shows a maximum SNDR of 64dB and 61dB and a maximum SFDR of 71dB and 70dB at 120MS/s and 150MS/s, respectively. The ADC with an active die area of $2.0mm^2$ consumes 140mW at 150MS/s and 1.2V.

An Efficient Addressing Scheme Using (x, y) Coordinates in Environments of Smart Grid (스마트 그리드 환경에서 (x, y) 좌표값을 이용한 효율적인 주소 할당 방법)

  • Cho, Yang-Hyun;Lim, Song-Bin;Kim, Gyung-Mok
    • Journal of the Korea Society of Computer and Information
    • /
    • v.17 no.1
    • /
    • pp.61-69
    • /
    • 2012
  • Smart Grid is the next-generation intelligent power grid that maximizes energy efficiency with the convergence of IT technologies and the existing power grid. Smart Grid is created solution for standardization and interoperability. Smart Grid industry enables consumers to check power rates in real time for active power consumption. It also enables suppliers to measure their expected power generation load, which stabilizes the operation of the power system. Smart industy was ecolved actively cause Wireless communication is being considered for AMI system and wireless communication using ZigBee sensor has been applied in various industly. In this paper, we proposed efficient addressing scheme for improving the performance of the routing algorithm using ZigBee in Smart Grid environment. A distributed address allocation scheme used an existing algorithm has wasted address space. Therefore proposing x, y coordinate axes from divide address space of 16 bit to solve this problem. Each node was reduced not only bitwise but also multi hop using the coordinate axes while routing than Cskip algorithm. I compared the performance between the standard and the proposed mechanism through the numerical analysis. Simulation verify performance about decrease averaging multi hop count that compare proposing algorithm and another. The numerical analysis results show that proposed algorithm reduce multi hop than ZigBee distributed address assignment and another.

G-RAID: A Green RAID Mechanism for enhancing Energy-Efficiency in Massive Storage System (G-RAID: 대용량 저장장치에서 에너지 효율향상을 위한 그린 RAID 기법)

  • Kim, Young-Hwan;Suck, Jin-Sun;Park, Chang-Won;Hong, Ji-Man
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.6
    • /
    • pp.21-30
    • /
    • 2011
  • In the global IT market, a lot of issues for responding to various environmental regulations emerged. In case of the data centers, it is consuming huge amounts of energy to maintain. So there have been various technical attempts as Consolidation, Virtualization, Optimization to efficiently manage energy and data storage to fix the problems. In this paper, we propose a new RAID(Redundant Array of Independent Disks) mechanism which is differing the intensity of power consumption and works to provide data protection and disaster recovery(backup, mirroring etc.) to stratify multiple volumes. G-RAID minimize the power consumption and the lower of I/O performance by selecting the volume depending on the frequency of data access while classifying the power consumption between volumes in storage system. Also, it is possible that a filesystem and block map information of G-RAID is processed by basic unit which is group located in a row for the blocks to work efficiently and can minimize the performance degradation of block mapping load by the access frequency in each groups. As a result, we obtained to elevate a little bit of response time caused by block relocation work, but showed the decrease of power consumption by 38%.

Crashworthiness Analysis and Shape Design Optimization of Thin-walled Corrugated Tubes under Axial Impact (축 방향 충격을 받는 박판 파형관의 충돌안전도 해석 및 형상 최적설계)

  • Ahn, Seung Ho;Jung, Hyun Seung;Kim, Jin Sung;Son, Seung Wan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.5
    • /
    • pp.128-135
    • /
    • 2021
  • Thin-walled tubes have been widely used as energy absorbing devices because they are light and have high energy-absorption efficiency. However, the downside is that conventional thin-walled tubes usually exhibit an excessive initial peak crushing force (IPCF) and a large fluctuation in the load-displacement curve, and thus lack stability as energy absorbing devices. Corrugated tubes were introduced to reduce IPCF and to increase the stability of collision energy-absorbing devices. Since the performance of corrugated tubes is highly influence by geometry, design optimization methods can be utilized to optimize the performance of corrugated tubes. In this paper, we utilize shape design optimization based on an adaptive surrogate model for crashworthiness analysis. The amplitude and wavelength of the corrugation, as well as curvature changes in the features, are the design variables. A morphing methodology is adopted to perform shape design parameterization. Through numerical examples, we compare optimal design results based on the adaptive surrogate model, with optimal results based on conventional surrogate models, and we show that direct optimal design methods produce more efficient results.

A Hybrid SVM Classifier for Imbalanced Data Sets (불균형 데이터 집합의 분류를 위한 하이브리드 SVM 모델)

  • Lee, Jae Sik;Kwon, Jong Gu
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.125-140
    • /
    • 2013
  • We call a data set in which the number of records belonging to a certain class far outnumbers the number of records belonging to the other class, 'imbalanced data set'. Most of the classification techniques perform poorly on imbalanced data sets. When we evaluate the performance of a certain classification technique, we need to measure not only 'accuracy' but also 'sensitivity' and 'specificity'. In a customer churn prediction problem, 'retention' records account for the majority class, and 'churn' records account for the minority class. Sensitivity measures the proportion of actual retentions which are correctly identified as such. Specificity measures the proportion of churns which are correctly identified as such. The poor performance of the classification techniques on imbalanced data sets is due to the low value of specificity. Many previous researches on imbalanced data sets employed 'oversampling' technique where members of the minority class are sampled more than those of the majority class in order to make a relatively balanced data set. When a classification model is constructed using this oversampled balanced data set, specificity can be improved but sensitivity will be decreased. In this research, we developed a hybrid model of support vector machine (SVM), artificial neural network (ANN) and decision tree, that improves specificity while maintaining sensitivity. We named this hybrid model 'hybrid SVM model.' The process of construction and prediction of our hybrid SVM model is as follows. By oversampling from the original imbalanced data set, a balanced data set is prepared. SVM_I model and ANN_I model are constructed using the imbalanced data set, and SVM_B model is constructed using the balanced data set. SVM_I model is superior in sensitivity and SVM_B model is superior in specificity. For a record on which both SVM_I model and SVM_B model make the same prediction, that prediction becomes the final solution. If they make different prediction, the final solution is determined by the discrimination rules obtained by ANN and decision tree. For a record on which SVM_I model and SVM_B model make different predictions, a decision tree model is constructed using ANN_I output value as input and actual retention or churn as target. We obtained the following two discrimination rules: 'IF ANN_I output value <0.285, THEN Final Solution = Retention' and 'IF ANN_I output value ${\geq}0.285$, THEN Final Solution = Churn.' The threshold 0.285 is the value optimized for the data used in this research. The result we present in this research is the structure or framework of our hybrid SVM model, not a specific threshold value such as 0.285. Therefore, the threshold value in the above discrimination rules can be changed to any value depending on the data. In order to evaluate the performance of our hybrid SVM model, we used the 'churn data set' in UCI Machine Learning Repository, that consists of 85% retention customers and 15% churn customers. Accuracy of the hybrid SVM model is 91.08% that is better than that of SVM_I model or SVM_B model. The points worth noticing here are its sensitivity, 95.02%, and specificity, 69.24%. The sensitivity of SVM_I model is 94.65%, and the specificity of SVM_B model is 67.00%. Therefore the hybrid SVM model developed in this research improves the specificity of SVM_B model while maintaining the sensitivity of SVM_I model.

RGB Channel Selection Technique for Efficient Image Segmentation (효율적인 이미지 분할을 위한 RGB 채널 선택 기법)

  • 김현종;박영배
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.10
    • /
    • pp.1332-1344
    • /
    • 2004
  • Upon development of information super-highway and multimedia-related technoiogies in recent years, more efficient technologies to transmit, store and retrieve the multimedia data are required. Among such technologies, firstly, it is common that the semantic-based image retrieval is annotated separately in order to give certain meanings to the image data and the low-level property information that include information about color, texture, and shape Despite the fact that the semantic-based information retrieval has been made by utilizing such vocabulary dictionary as the key words that given, however it brings about a problem that has not yet freed from the limit of the existing keyword-based text information retrieval. The second problem is that it reveals a decreased retrieval performance in the content-based image retrieval system, and is difficult to separate the object from the image that has complex background, and also is difficult to extract an area due to excessive division of those regions. Further, it is difficult to separate the objects from the image that possesses multiple objects in complex scene. To solve the problems, in this paper, I established a content-based retrieval system that can be processed in 5 different steps. The most critical process of those 5 steps is that among RGB images, the one that has the largest and the smallest background are to be extracted. Particularly. I propose the method that extracts the subject as well as the background by using an Image, which has the largest background. Also, to solve the second problem, I propose the method in which multiple objects are separated using RGB channel selection techniques having optimized the excessive division of area by utilizing Watermerge's threshold value with the object separation using the method of RGB channels separation. The tests proved that the methods proposed by me were superior to the existing methods in terms of retrieval performances insomuch as to replace those methods that developed for the purpose of retrieving those complex objects that used to be difficult to retrieve up until now.

An Efficient Spatial Index Structure for Main Memory (메인 메모리를 위한 효율적인 공간 인덱스 구조)

  • Lee, Ki-Young;Lim, Myung-Jae;Kang, Jeong-Jin;Kim, Joung-Joon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.9 no.2
    • /
    • pp.13-20
    • /
    • 2009
  • Recently there is growing interest in LBS requiring real-time services and the spatial main memory DBMS for efficient Telematics services. In order to optimize existing disk-based spatial indexes of the spatial main memory DBMS in the main memory, spatial index structures have been proposed, which minimize failures in cache access by reducing the entry size. However, because the reduction of entry size requires compression based on the MBR of the parent node or the removal of redundant MBR, the cost of MBR reconstruction increases in index update and the efficiency of search is lowered in index search. Thus, to reduce the cost of MBR reconstruction, this paper proposed the RSMB (relative-sized MBR)compression technique, which applies the base point of compression differently in case of broad distribution and narrow distribution. In case of broad distribution, compression is made based on the left-bottom point of the extended MBR of the parent node, and in case of narrow distribution, the whole MBR is divided into cells of the same size and compression is made based on the left-bottom point of each cell. In addition, MBR was compressed using a relative coordinate and size to reduce the cost of search in index search. Lastly, we evaluated the performance of the proposed RSMBR compression technique using real data, and proved its superiority.

  • PDF

New Beamforming Schemes with Optimum Receive Combining for Multiuser MIMO Downlink Channels (다중사용자 다중입출력 하향링크 시스템을 위한 최적 수신 결합을 이용한 새로운 빔 형성 기법)

  • Lee, Sang-Rim;Park, Seok-Hwan;Moon, Sung-Hyun;Lee, In-Kyu
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.48 no.8
    • /
    • pp.15-26
    • /
    • 2011
  • In this paper, we present a new beamforming scheme for a downlink of multiuser multiple-input multipleoutput (MIMO) communication systems. Recently, a block-diagonalization (BD) algorithm has been proposed for the multiuser MIMO downlink where both a base station and each user have multiple antennas. However, the BD algorithm is not efficient when the number of supported streams per user is smaller than that of receive antennas. Since the BD method utilizes the space based on the channel matrix without considering the receive combining, the degree of freedom for beamforming cannot be fully exploited at the transmitter. In this paper, we optimize the receive beamforming vector under a zero forcing (ZF) constraint, where all inter-user interference is driven to zero. We propose an efficient algorithm to find the optimum receive vector by an iterative procedure. The proposed algorithm requires two phase values feedforward information for the receive combining vector. Also, we present another algorithm which needs only one phase value by using a decomposition of the complex general unitary matrix. Simulation results show that the proposed beamforming scheme outperforms the conventional BD algorithm in terms of error probability and obtains the diversity enhancement by utilizing the degree of freedom at the base station.

Fine Grained Resource Scaling Approach for Virtualized Environment (가상화 환경에서 세밀한 자원 활용률 적용을 위한 스케일 기법)

  • Lee, Donhyuck;Oh, Sangyoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.7
    • /
    • pp.11-21
    • /
    • 2013
  • Recently operating a large scale computing resource like a data center becomes easier because of the virtualization technology that virtualize servers and enable flexible resource provision. The most of public cloud services provides automatic scaling in the form of scale-in or scale-out and these scaling approaches works well to satisfy the service level agreement (SLA) of users. However, a novel scaling approach is required to operate private clouds that has smaller amount of computing resources than vast resources of public clouds. In this paper, we propose a hybrid server scaling architecture and related algorithms using both scale-in and scale-out to achieve higher resource utilization rate for private clouds. We uses dynamic resource allocation and live migration to run our proposed algorithm. Our propose system aims to provide a fine-grain resource scaling by steps. Thus private cloud systems are able to keep stable service and to reduce server management cost by optimizing server utilization. The experiment results show that our proposed approach performs better in resource utilization than the scale-out approach based on the number of users.