• Title/Summary/Keyword: 계산적 오버헤드

Search Result 121, Processing Time 0.022 seconds

Distributed Key Management Using Regression Model for Hierarchical Mobile Sensor Networks (계층적인 이동 센서 네트워크에서 회귀모델을 이용한 분산 키 관리)

  • Kim Mi-Hui;Chae Ki-Joon
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.43 no.7 s.349
    • /
    • pp.1-13
    • /
    • 2006
  • In this paper, we introduce a novel key management scheme that is based on the key pre-distribution but provides the key re-distribution method, in order to manage keys for message encryption and authentication of lower-layer sensor nodes on hierarchical mobile sensor networks. The characteristics of our key management are as follows: First, the role of key management is distributed to aggregator nodes as well as a sink node, to overcome the weakness of centralized management. Second, a sink node generates keys using regression model, thus it stores only the information for calculating the keys using the key information received from nodes, but does not store the relationship between a node and a key, and the keys themselves. As the disadvantage of existing key pre-distributions, they do not support the key re-distribution after the deployment of nodes, and it is hard to extend the key information in the case that sensor nodes in the network enlarge. Thirdly, our mechanism provides the resilience to node capture(${\lambda}$-security), also provided by the existing key pre-distributions, and fourth offers the key freshness through key re-distribution, key distribution to mobile nodes, and scalability to make up for the weak points in the existing key pre-distributions. Fifth, our mechanism does not fix the relationship between a node and a key, thus supports the anonymity and untraceability of mobile nodes. Lastly, we compare ours with existing mechanisms, and verify our performance through the overhead analysis of communication, computation, and memory.

MAC Scheduling Algorithm for Efficient Management of Wireless Resources in Bluetooth Systems (블루투스 시스템에서의 효율적 무선자원관리를 위한 MAC 스케쥴링 기법)

  • 주양익;권오석;오종수;김용석;이태진;엄두섭;차균현
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.9A
    • /
    • pp.702-709
    • /
    • 2003
  • In this paper, we propose an efficient and QoS-aware MAC scheduling algorithm for Bluetooth, which considers both throughput and delay performance of each Master-Slave pair in scheduling decisions, and thus, attempts to maximize overall performance. The proposed algorithm, MTDPP (Modified Throughput-Delay Priority Policy), makes up for the drawbacks of T-D PP (Throughput-Delay Priority Policy) proposed in [6] and improves the performance. Since Bluetooth employs a master-driven TDD based scheduling algorithm, which is basically operated with the Round Robin policy, many slots may be wasted by POLL or NULL packets when there is no data waiting for transmission in queues. To overcome this link wastage problem, several algorithms have been proposed. Among them, queue state-based priority policy and low power mode-based algorithm can perform with high throughput and reasonable fairness. However, their performances may depend on traffic characteristics, i.e., static or dynamic, and they require additional computational and signaling overheads. In order to tackle such problems, we propose a new scheduling algorithm. Performance of our proposed algorithm is evaluated with respect to throughput and delay. Simulation results show that overall performances can be improved by selecting suitable parameters of our algorithm.

A Data Aggregation Scheme for Enhancing the Efficiency of Data Aggregation and Correctness in Wireless Sensor Networks (무선 센서 네트워크에서 데이터 수집의 효율성 및 정확성 향상을 위한 데이터 병합기법)

  • Kim, Hyun-Tae;Yu, Tae-Young;Jung, Kyu-Su;Jeon, Yeong-Bae;Ra, In-Ho
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.16 no.5
    • /
    • pp.531-536
    • /
    • 2006
  • Recently, many of researchers have been studied in data processing oriented middleware for wireless sensor networks with the rapid advances on sensor and wireless communication technologies. In a wireless sensor network, a middleware should handle the data loss problem at an intermediate sensor node caused by instantaneous data burstness to support efficient processing and fast delivering of the sensing data. To handle this problem, a simple data discarding or data compressing policy for reducing the total amount of data to be transferred is typically used. But, data discarding policy decreases the correctness of a collected data, in other hand, data compressing policy requires additional processing overhead with the high complexity of the given algorithm. In this paper, it proposes a data-average method for enhancing the efficiency of data aggregation and correctness where the sensed data should be delivered only with the limited computing power and energy resource. With the proposed method, unnecessary data transfer of the overlapped data is eliminated and data correctness is enhanced by using the proposed averaging scheme when an instantaneous data burstness is occurred. Finally, with the TOSSTM simulation results on TinyBB, we show that the correctness of the transferred data is enhanced.

Design of Efficient frequency Offset Estimator for MB-OFDM based UWB Systems (MB-OFDM 기반 UWB 시스템을 위한 효율적인 주파수 옵셋 추정기의 설계)

  • Kim, Kil-Hwan;Jung, Yun-Ho;Kim, Jae-Seok
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.3C
    • /
    • pp.311-321
    • /
    • 2009
  • This paper proposes an efficient frequency offset estimation algorithm for MB-OFDM based UWB systems. The time-frequency interleaving in MB-OFDM extends the time-interval between two transmitted OFDM symbols in the same sub-band. The extended time-interval causes not only the degradation of the system performance by reducing frequency offset estimation range, but also the increase of the hardware complexity by requiring the larger number of storing samples. The proposed estimation algorithm expands the estimation range by applying the proposed sign detection scheme. Simulation results show that the estimation range is increased above 30 ppm compared with a conventional auto-correlation based scheme. The estimation is performed on only one sub-band, and the frequency offsets of the others are calculated by relation to center frequency. This way reduced the number of the storing samples by about l/3. The frequency offset estimator with the proposed algorithm was designed into the architecture which minimizes hardware overhead by time-sharing operators and memory units, and which was synthesized to gate-level circuits using $0.13{\mu}m$ CMOS technology, and the total gates were about 47K.

A Utility-Based Hybrid Error Recovery Scheme for Multimedia Transmission over 3G Cellular Broadcast Networks (3G 방송망에서의 효율적인 멀티미디어 전송을 위한 유틸리티 기반 하이브라드 에러 복구기법)

  • Kang Kyung-Tae;Cho Yong-Jin;Cho Yong-Woo;Cho Jin-Sung;Shin Heon-Shik
    • Journal of KIISE:Information Networking
    • /
    • v.33 no.4
    • /
    • pp.333-342
    • /
    • 2006
  • The cdma2000 lxEV - DO mobile communication system provides broadcast and multicast services (BCMCS) to meet an increasing demand from multimedia data services. The servicing of video streams over a BCMCS network must, however, face a challenge from the unreliable and error-prone nature of the radio channel. The BCMCS network uses Reed-Solomon coding integrated with the MAC protocol for error recovery. We analyze this coding technique and show that it is not effective in the case of slowly moving mobiles. To improve the playback quality of an MPEG-4 FGS video stream, we propose the Hybrid error recovery scheme, which combines Reed-Solomon with ARQ, using slots which are saved by reducing the Reed-Solomon coding overhead. The target packets to be retransmitted are prioritized by a utility function to reduce the packet error rate in the application layer within a fixed retransmission budget. This is achieved by considering of the map of the error control block at each mobile node. The proposed Hybrid error recovery scheme also uses the characteristics of MPEG-4 FGS (fine granularity scalability) to improve the video quality even when conditions are adverse: slow-moving nodes and a high error rate in the physical channel.

k-Nearest Neighbor Querv Processing using Approximate Indexing in Road Network Databases (도로 네트워크 데이타베이스에서 근사 색인을 이용한 k-최근접 질의 처리)

  • Lee, Sang-Chul;Kim, Sang-Wook
    • Journal of KIISE:Databases
    • /
    • v.35 no.5
    • /
    • pp.447-458
    • /
    • 2008
  • In this paper, we address an efficient processing scheme for k-nearest neighbor queries to retrieve k static objects in road network databases. Existing methods cannot expect a query processing speed-up by index structures in road network databases, since it is impossible to build an index by the network distance, which cannot meet the triangular inequality requirement, essential for index creation, but only possible in a totally ordered set. Thus, these previous methods suffer from a serious performance degradation in query processing. Another method using pre-computed network distances also suffers from a serious storage overhead to maintain a huge amount of pre-computed network distances. To solve these performance and storage problems at the same time, this paper proposes a novel approach that creates an index for moving objects by approximating their network distances and efficiently processes k-nearest neighbor queries by means of the approximate index. For this approach, we proposed a systematic way of mapping each moving object on a road network into the corresponding absolute position in the m-dimensional space. To meet the triangular inequality this paper proposes a new notion of average network distance, and uses FastMap to map moving objects to their corresponding points in the m-dimensional space. After then, we present an approximate indexing algorithm to build an R*-tree, a multidimensional index, on the m-dimensional points of moving objects. The proposed scheme presents a query processing algorithm capable of efficiently evaluating k-nearest neighbor queries by finding k-nearest points (i.e., k-nearest moving objects) from the m-dimensional index. Finally, a variety of extensive experiments verifies the performance enhancement of the proposed approach by performing especially for the real-life road network databases.

USB Device Authentication Protocol based on OTP (OTP 기반의 USB 디바이스 인증 프로토콜)

  • Jeong, Yoon-Su;Kim, Yong-Tae;Park, Gil-Cheol
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.15 no.8
    • /
    • pp.1735-1742
    • /
    • 2011
  • Now a days, as a mass-storage USB becomes comfortable to carry, function of USB is being developed fast. However, there is a problem that the personal information which is stored in USB could be exposed being used with negative purpose without other certification process. This paper suggests OTP(One-Time Password)-based certification protocol of USB to securely protect personal information stored in USB without additional certification information. The proposed OTP based certification protocol of USB not only demands low calculations but also prevents physical approach of USB of other network and does not allow unnecessary service access of user because it conducts simple action and uses one-way hash function. Therefore, communication overhead and service delay is improved. In the experiment, the proposed protocol compares and evaluates throughput of certification server according to the numbers of USB and delay time of packet certification with a device(USB driver) which simply save device and a device(USB Token) which can calculate by oneself. As a result, it is improved as the number of 12.5% in the certification delay time on average and is improved as the number of 10.8% in the throughput of certification server according to the numbers of USB.

JMP+RAND: Mitigating Memory Sharing-Based Side-Channel Attack by Embedding Random Values in Binaries (JMP+RAND: 바이너리 난수 삽입을 통한 메모리 공유 기반 부채널 공격 방어 기법)

  • Kim, Taehun;Shin, Youngjoo
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.5
    • /
    • pp.101-106
    • /
    • 2020
  • Since computer became available, much effort has been made to achieve information security. Even though memory protection defense mechanisms were studied the most among of them, the problems of existing memory protection defense mechanisms were found due to improved performance of computer and new defense mechanisms were needed due to the advent of the side-channel attacks. In this paper, we propose JMP+RAND that embedding random values of 5 to 8 bytes per page to defend against memory sharing based side-channel attacks and bridging the gap of existing memory protection defense mechanism. Unlike the defense mechanism of the existing side-channel attacks, JMP+RAND uses static binary rewriting and continuous jmp instruction and random values to defend against the side-channel attacks in advance. We numerically calculated the time it takes for a memory sharing-based side-channel attack to binary adopted JMP+RAND technique and verified that the attacks are impossible in a realistic time. Modern architectures have very low overhead for JMP+RAND because of the very fast and accurate branching of jmp instruction using branch prediction. Since random value can be embedded only in specific programs using JMP+RAND, it is expected to be highly efficient when used with memory deduplication technique, especially in a cloud computing environment.

Performance analysis of Frequent Itemset Mining Technique based on Transaction Weight Constraints (트랜잭션 가중치 기반의 빈발 아이템셋 마이닝 기법의 성능분석)

  • Yun, Unil;Pyun, Gwangbum
    • Journal of Internet Computing and Services
    • /
    • v.16 no.1
    • /
    • pp.67-74
    • /
    • 2015
  • In recent years, frequent itemset mining for considering the importance of each item has been intensively studied as one of important issues in the data mining field. According to strategies utilizing the item importance, itemset mining approaches for discovering itemsets based on the item importance are classified as follows: weighted frequent itemset mining, frequent itemset mining using transactional weights, and utility itemset mining. In this paper, we perform empirical analysis with respect to frequent itemset mining algorithms based on transactional weights. The mining algorithms compute transactional weights by utilizing the weight for each item in large databases. In addition, these algorithms discover weighted frequent itemsets on the basis of the item frequency and weight of each transaction. Consequently, we can see the importance of a certain transaction through the database analysis because the weight for the transaction has higher value if it contains many items with high values. We not only analyze the advantages and disadvantages but also compare the performance of the most famous algorithms in the frequent itemset mining field based on the transactional weights. As a representative of the frequent itemset mining using transactional weights, WIS introduces the concept and strategies of transactional weights. In addition, there are various other state-of-the-art algorithms, WIT-FWIs, WIT-FWIs-MODIFY, and WIT-FWIs-DIFF, for extracting itemsets with the weight information. To efficiently conduct processes for mining weighted frequent itemsets, three algorithms use the special Lattice-like data structure, called WIT-tree. The algorithms do not need to an additional database scanning operation after the construction of WIT-tree is finished since each node of WIT-tree has item information such as item and transaction IDs. In particular, the traditional algorithms conduct a number of database scanning operations to mine weighted itemsets, whereas the algorithms based on WIT-tree solve the overhead problem that can occur in the mining processes by reading databases only one time. Additionally, the algorithms use the technique for generating each new itemset of length N+1 on the basis of two different itemsets of length N. To discover new weighted itemsets, WIT-FWIs performs the itemset combination processes by using the information of transactions that contain all the itemsets. WIT-FWIs-MODIFY has a unique feature decreasing operations for calculating the frequency of the new itemset. WIT-FWIs-DIFF utilizes a technique using the difference of two itemsets. To compare and analyze the performance of the algorithms in various environments, we use real datasets of two types (i.e., dense and sparse) in terms of the runtime and maximum memory usage. Moreover, a scalability test is conducted to evaluate the stability for each algorithm when the size of a database is changed. As a result, WIT-FWIs and WIT-FWIs-MODIFY show the best performance in the dense dataset, and in sparse dataset, WIT-FWI-DIFF has mining efficiency better than the other algorithms. Compared to the algorithms using WIT-tree, WIS based on the Apriori technique has the worst efficiency because it requires a large number of computations more than the others on average.

Methods of Incorporating Design for Production Considerations into Concept Design Investigations (개념설계 단계에서 총 건조비를 최소로 하는 생산지향적 설계 적용 방법)

  • H.S.,Bong
    • Bulletin of the Society of Naval Architects of Korea
    • /
    • v.27 no.3
    • /
    • pp.131-136
    • /
    • 1990
  • 여러해 전부터 선박의 생산실적이나 생산성 관련 자료를 기록하고 보완하는 작업을 꾸준히 개선토록 노력해온 결과중 중요한 것 하나는, 선박의 여러 가지 설계 검토과정에서 충분히 활용할 수 있는 함축성 있고 믿을만한 형태의 생산정보를 제공해줄 수 있게 되었다는 것이라고 말 할 수 있겠다. 이러한 자료들은 생산계획상 각 단계(stage)에서의 작업량, 예상재료비와 인건비의 산출등이 포함될 수 있으며, 선박이나 해상구조물의 전반적인 설계방법론(design methodology)을 개선코자 한다면 ''생산지향적 설계(Design for Production)''의 근간이 되는 선박건조전략(build strategy), 구매정책(purchasing policy)과 생산기술(production technology)에 대한 폭넓은 지식이 한데 어우러져야 한다. 최근에는 CIMS의 일부분에서 보는 바와 같은 경영관리, 설계 및 생산지원 시스템의 도입으로 이와 같은 설계 프로세스의 추진을 가능케하고 있다. 이와 병행하여 설계를 지원하기 위한 전산기술, 특히 대화형 화상처리기술(interactive graphics)의 발달은 설계자가 선박의 형상이나 구조 배치를 여러 가지로 변화시켜 가면서 눈으로 즉시 확인할 수 있도록 설계자의 능력을 배가시키는데 크게 기여하고 있다. 여러 가지의 설계안(alternative design arrangement)을 신속히 만들어내고 이를 즉시 검토 평가할 수 있는 능력을 초기설계 단계에서 가질 수 있다면 이는 분명히 큰 장점일 것이며, 더구나 설계초기 단계에 생산관련인자를 설계에서 고려할 수 있다면 이는 더욱 두드러진 발전일 것이다. 생산공법과 관련생산 비용을 정확히 반영한 각 가지의 설계안을 짧은 시간내에 검토하고 생산소요 비용을 산출하여 비교함으로써, 수주계약단계에서 실제적인 생산공법과 신뢰성있는 생산실적자료를 기준으로 하여 총 건조비(total production cost)를 최소로 하는 최적의 설계를 선택할 수 있도록 해 줄 것이다. 이제 이와 같은 새로운 설계도구(design tool)를 제공해 주므로써 초기설계에 각종 생산관련 정보나 지식 및 실적자료가 반영가능토록 발전되었다. 본 논문은 영국의 뉴카슬대학교(Univ. of Newcastle upon Type)에서 위에 언급한 특징들을 반영하여 새로운 선박구조 설계 방법을 개발한 연구결과를 보여주고 있다. 본 선계연구는 5단계로 구분되는데; (1) 컴퓨터 그라픽스를 이용하고 생산정보 데이타베이스와 연결시켜 구조형상(geometry)을 정의하고 구조부재 칫수(scantling) 계산/결정 (2) 블럭 분할(block division) 및 강재 배치(panel arrangement)의 확정을 위해 생산기술 및 건조방식에 대한 정보 제공 (3) 상기 (1) 및 (2)를 활용하여 아래 각 생산 단계에서의 생산작업 분석(work content assessment) a) 생산 준비 단계(Preparation) b) 가공 조립 단계(Fabrication/Assembly) c) 탑재 단계(Erection) (4) 각각의 설계(안)에 대하여 재료비(material cost), 인건비(labour cost) 및 오버헤드 비용(overhead cost)을 산출키 위한 조선소의 생산시설 및 각종 품셈 정보 (5) 총 건조 비용(total production cost)을 산출하여 각각의 설계안을 비교 검토. 본 설계 방식을 산적화물선(Bulk Carrier) 설계에 적용하여 구조배치(structural geometry), 표준화의 정도(levels of standardisation), 구조생산공법(structural topology) 등의 변화에 따른 설계 결과의 민감도를 분석(sensitivity studies)하였다. 전산장비는 설계자의 대화형 접근을 용이하도록 하기 위해 VAX의 화상 처리장치를 이용하여 각 설계안에 대한 구조형상과 작업분석, 건조비 현황 등을 제시할 수 있도록 하였다. 결론적으로 본 연구는 설계초기 단계에서 상세한 건조비 모델(detailed production cost model)을 대화형 화상 처리방법에 접합시켜 이를 이용하여 여러가지 설계안의 도출과 비교검토를 신속히 처리할 수 있도록 함은 물론, 각종 생산 실적정보를 초기설계에 반영하는 최초의 시도라고 믿으며, 생산지향적(Design for Production) 최적설계분야의 발전에 많은 도움이 되기를 기대해 마지 않는다. 참고로 본 시스템의 설계 적용결과를 부록에 요약 소개하며, 상세한 내용은 참고문헌 [4] 또는 [7]을 참조 요망한다.

  • PDF