• Title/Summary/Keyword: network optimization

Search Result 2,239, Processing Time 0.029 seconds

Fast Cooperative Sensing with Low Overhead in Cognitive Radios

  • Dai, Zeyang;Liu, Jian;Li, Yunji;Long, Keping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.1
    • /
    • pp.58-73
    • /
    • 2014
  • As is well known, cooperative sensing can significantly improve the sensing accuracy as compared to local sensing in cognitive radio networks (CRNs). However, a large number of cooperative secondary users (SUs) reporting their local detection results to the fusion center (FC) would cause much overhead, such as sensing delay and energy consumption. In this paper, we propose a fast cooperative sensing scheme, called double threshold fusion (DTF), to reduce the sensing overhead while satisfying a given sensing accuracy requirement. In DTF, FC respectively compares the number of successfully received local decisions and that of failed receptions with two different thresholds to make a final decision in each reporting sub-slot during a sensing process, where cooperative SUs sequentially report their local decisions in a selective fashion to reduce the reporting overhead. By jointly considering sequential detection and selective reporting techniques in DTF, the overhead of cooperative sensing can be significantly reduced. Besides, we study the performance optimization problems with different objectives for DTF and develop three optimum fusion rules accordingly. Simulation results reveal that DTF shows evident performance gains over an existing scheme.

Topology Aggregation for Hierarchical Wireless Tactical Networks

  • Pak, Woo-Guil;Choi, Young-June
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.5 no.2
    • /
    • pp.344-358
    • /
    • 2011
  • Wireless tactical network (WTN) is the most important present-day technology enabling modern network centric warfare. It inherits many features from WMNs, since the WTN is based on existing wireless mesh networks (WMNs). However, it also has distinctive characteristics, such as hierarchical structures and tight QoS (Quality-of-Service) requirements. Little research has been conducted on hierarchical protocols to support various QoS in WMN. We require new protocols specifically optimized for WTNs. Control packets are generally required to find paths and reserve resources for QoS requirements, so data throughput is not degraded due to overhead. The fundamental solution is to adopt topology aggregation, in which a low tier node aggregates and simplifies the topology information and delivers it to a high tier node. The overhead from control packet exchange can be reduced greatly due to decreased information size. Although topology aggregation is effective for low overhead, it also causes the inaccuracy of topology information; thus, incurring low QoS support capability. Therefore, we need a new topology aggregation algorithm to achieve high accuracy. In this paper, we propose a new aggregation algorithm based on star topology. Noting the hierarchical characteristics in military and hierarchical networks, star topology aggregation can be used effectively. Our algorithm uses a limited number of bypasses to increase the exactness of the star topology aggregation. It adjusts topology parameters whenever it adds a bypass. Consequently, the result is highly accurate and has low computational complexity.

A Study on the Activation Technique of Detection nodes for Intrusion Detection in Wireless Sensor Networks (무선 센서네트워크에서 침입탐지를 위한 탐지노드 활성화기법 연구)

  • Seong, Ki-Taek
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.12 no.11
    • /
    • pp.5238-5244
    • /
    • 2011
  • Recently, wireless sensor networks have become increasingly interesting areas over extensive application fields such as military, ecological, and health-related areas. Almost sensor networks have mission-critical tasks that requires very high security. Therefore, extensive work has been done for securing sensor networks from outside attackers, efficient cryptographic systems, secure key management and authorization, but little work has yet been done to protect these networks from inside threats. This paper proposed an method to select which nodes should activate their idle nodes as detectors to be able to watch all packets in the sensor network. Suggested method is modeled as optimization equation, and heuristic Greedy algorithm based simulation results are presented to verify my approach.

CNN-based Fast Split Mode Decision Algorithm for Versatile Video Coding (VVC) Inter Prediction

  • Yeo, Woon-Ha;Kim, Byung-Gyu
    • Journal of Multimedia Information System
    • /
    • v.8 no.3
    • /
    • pp.147-158
    • /
    • 2021
  • Versatile Video Coding (VVC) is the latest video coding standard developed by Joint Video Exploration Team (JVET). In VVC, the quadtree plus multi-type tree (QT+MTT) structure of coding unit (CU) partition is adopted, and its computational complexity is considerably high due to the brute-force search for recursive rate-distortion (RD) optimization. In this paper, we aim to reduce the time complexity of inter-picture prediction mode since the inter prediction accounts for a large portion of the total encoding time. The problem can be defined as classifying the split mode of each CU. To classify the split mode effectively, a novel convolutional neural network (CNN) called multi-level tree (MLT-CNN) architecture is introduced. For boosting classification performance, we utilize additional information including inter-picture information while training the CNN. The overall algorithm including the MLT-CNN inference process is implemented on VVC Test Model (VTM) 11.0. The CUs of size 128×128 can be the inputs of the CNN. The sequences are encoded at the random access (RA) configuration with five QP values {22, 27, 32, 37, 42}. The experimental results show that the proposed algorithm can reduce the computational complexity by 11.53% on average, and 26.14% for the maximum with an average 1.01% of the increase in Bjøntegaard delta bit rate (BDBR). Especially, the proposed method shows higher performance on the sequences of the A and B classes, reducing 9.81%~26.14% of encoding time with 0.95%~3.28% of the BDBR increase.

A Parallel Genetic Algorithm for Solving Deadlock Problem within Multi-Unit Resources Systems

  • Ahmed, Rabie;Saidani, Taoufik;Rababa, Malek
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.12
    • /
    • pp.175-182
    • /
    • 2021
  • Deadlock is a situation in which two or more processes competing for resources are waiting for the others to finish, and neither ever does. There are two different forms of systems, multi-unit and single-unit resource systems. The difference is the number of instances (or units) of each type of resource. Deadlock problem can be modeled as a constrained combinatorial problem that seeks to find a possible scheduling for the processes through which the system can avoid entering a deadlock state. To solve deadlock problem, several algorithms and techniques have been introduced, but the use of metaheuristics is one of the powerful methods to solve it. Genetic algorithms have been effective in solving many optimization issues, including deadlock Problem. In this paper, an improved parallel framework of the genetic algorithm is introduced and adapted effectively and efficiently to deadlock problem. The proposed modified method is implemented in java and tested on a specific dataset. The experiment shows that proposed approach can produce optimal solutions in terms of burst time and the number of feasible solutions in each advanced generation. Further, the proposed approach enables all types of crossovers to work with high performance.

Design and Optimization for Distributed Compress-and-Forward System based on Multi-Relay Network

  • Bao, Junwei;Xu, Dazhuan;Luo, Hao;Zhang, Ruidan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.2949-2963
    • /
    • 2019
  • A novel distributed compress-and-forward (CF) system based on multi-relay network is presented. In this system, as the direct link between the source and destination is invalid due to some reasons, such as the limited power, special working environment, or even economic factors, relays are employed to receive analog signals and carry on distributed compressed encoding. Subsequently, the digital signals are transmitted to the destination via wireless channel. Moreover, a theoretical analysis for the system is provided by utilizing the Chief Executive Officer (CEO) theory and Shannon channel capacity theory, and the rate-distortion function as well as the connection between the transmission rate and the channel capacity are constructed. In addition, an optimal signal-to-noise ratio (SNR) -based power allocation method is proposed to maximize the quantization SNR under the limited total power. Simulation result shows that the proposed CF system outperforms the amplify-and-forward (AF) system versus the SNR performance.

A Reinforcement Learning Framework for Autonomous Cell Activation and Customized Energy-Efficient Resource Allocation in C-RANs

  • Sun, Guolin;Boateng, Gordon Owusu;Huang, Hu;Jiang, Wei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.8
    • /
    • pp.3821-3841
    • /
    • 2019
  • Cloud radio access networks (C-RANs) have been regarded in recent times as a promising concept in future 5G technologies where all DSP processors are moved into a central base band unit (BBU) pool in the cloud, and distributed remote radio heads (RRHs) compress and forward received radio signals from mobile users to the BBUs through radio links. In such dynamic environment, automatic decision-making approaches, such as artificial intelligence based deep reinforcement learning (DRL), become imperative in designing new solutions. In this paper, we propose a generic framework of autonomous cell activation and customized physical resource allocation schemes for energy consumption and QoS optimization in wireless networks. We formulate the problem as fractional power control with bandwidth adaptation and full power control and bandwidth allocation models and set up a Q-learning model to satisfy the QoS requirements of users and to achieve low energy consumption with the minimum number of active RRHs under varying traffic demand and network densities. Extensive simulations are conducted to show the effectiveness of our proposed solution compared to existing schemes.

Hyper Parameter Tuning Method based on Sampling for Optimal LSTM Model

  • Kim, Hyemee;Jeong, Ryeji;Bae, Hyerim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.1
    • /
    • pp.137-143
    • /
    • 2019
  • As the performance of computers increases, the use of deep learning, which has faced technical limitations in the past, is becoming more diverse. In many fields, deep learning has contributed to the creation of added value and used on the bases of more data as the application become more divers. The process for obtaining a better performance model will require a longer time than before, and therefore it will be necessary to find an optimal model that shows the best performance more quickly. In the artificial neural network modeling a tuning process that changes various elements of the neural network model is used to improve the model performance. Except Gride Search and Manual Search, which are widely used as tuning methods, most methodologies have been developed focusing on heuristic algorithms. The heuristic algorithm can get the results in a short time, but the results are likely to be the local optimal solution. Obtaining a global optimal solution eliminates the possibility of a local optimal solution. Although the Brute Force Method is commonly used to find the global optimal solution, it is not applicable because of an infinite number of hyper parameter combinations. In this paper, we use a statistical technique to reduce the number of possible cases, so that we can find the global optimal solution.

A Task Scheduling Strategy in Cloud Computing with Service Differentiation

  • Xue, Yuanzheng;Jin, Shunfu;Wang, Xiushuang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5269-5286
    • /
    • 2018
  • Task scheduling is one of the key issues in improving system performance and optimizing resource management in cloud computing environment. In order to provide appropriate services for heterogeneous users, we propose a novel task scheduling strategy with service differentiation, in which the delay sensitive tasks are assigned to the rapid cloud with high-speed processing, whereas the fault sensitive tasks are assigned to the reliable cloud with service restoration. Considering that a user can receive service from either local SaaS (Software as a Service) servers or public IaaS (Infrastructure as a Service) cloud, we establish a hybrid queueing network based system model. With the assumption of Poisson arriving process, we analyze the system model in steady state. Moreover, we derive the performance measures in terms of average response time of the delay sensitive tasks and utilization of VMs (Virtual Machines) in reliable cloud. We provide experimental results to validate the proposed strategy and the system model. Furthermore, we investigate the Nash equilibrium behavior and the social optimization behavior of the delay sensitive tasks. Finally, we carry out an improved intelligent searching algorithm to obtain the optimal arrival rate of total tasks and present a pricing policy for the delay sensitive tasks.

Wireless Sensor Networks have Applied the Routing History Cache Routing Algorithm (무선센서 네트워크에서 Routing History Cache를 이용한 라우팅 알고리즘)

  • Lee, Doo-Wan;Jang, Kyung-Sik
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2009.10a
    • /
    • pp.1018-1021
    • /
    • 2009
  • Wireless Sensor Network collects a data from the specific area and the control is composed of small sensor nodes. Like this sensors to after that is established at the beginning are operated with the battery, the operational duration until several years must be continued from several months and will be able to apply the resources which is restricted in efficiently there must be. In this paper RHC (rounting history cache) applies in Directed Diffusion which apply a data central concept a reliability and an efficiency in data transfer course set. RHC algorithms which proposes each sensor node updated RHC of oneself with periodic and because storing the optimization course the course and, every event occurrence hour they reset the energy is wasted the fact that a reliability with minimization of duplication message improved.

  • PDF