• Title/Summary/Keyword: Network Latency

Search Result 764, Processing Time 0.02 seconds

A Deadlock Free Router Design for Network-on-Chip Architecture (NOC 구조용 교착상태 없는 라우터 설계)

  • Agarwal, Ankur;Mustafa, Mehmet;Shiuku, Ravi;Pandya, A.S.;Lho, Young-Ugh
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.11 no.4
    • /
    • pp.696-706
    • /
    • 2007
  • Multiprocessor system on chip (MPSoC) platform has set a new innovative trend for the System on Chip (SoC) design. With the rapidly approaching billion transistors era, some of the main problem in deep sub-micron technologies characterized by gate lengths in the range of 60-90 nm will arise from non scalable wire delays, errors in signal integrity and un-synchronized communication. These problems may be addressed by the use of Network on Chip (NOC) architecture for future SoC. Most future SoCs will use network architecture and a packet based communication protocol for on chip communication. This paper presents an adaptive wormhole routing with proactive turn prohibition to guarantee deadlock free on chip communication for NOC architecture. It shows a simple muting architecture with five full-duplex, flit-wide communication channels. We provide simulation results for message latency and compare results with those of dimension ordered techniques operating at the same link rates.

Kalman Filtering-based Traffic Prediction for Software Defined Intra-data Center Networks

  • Mbous, Jacques;Jiang, Tao;Tang, Ming;Fu, Songnian;Liu, Deming
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.2964-2985
    • /
    • 2019
  • Global data center IP traffic is expected to reach 20.6 zettabytes (ZB) by the end of 2021. Intra-data center networks (Intra-DCN) will account for 71.5% of the data center traffic flow and will be the largest portion of the traffic. The understanding of traffic distribution in IntraDCN is still sketchy. It causes significant amount of bandwidth to go unutilized, and creates avoidable choke points. Conventional transport protocols such as Optical Packet Switching (OPS) and Optical Burst Switching (OBS) allow a one-sided view of the traffic flow in the network. This therefore causes disjointed and uncoordinated decision-making at each node. For effective resource planning, there is the need to consider joining the distributed with centralized management which anticipates the system's needs and regulates the entire network. Methods derived from Kalman filters have proved effective in planning road networks. Considering the network available bandwidth as data transport highways, we propose an intelligent enhanced SDN concept applied to OBS architecture. A management plane (MP) is added to conventional control (CP) and data planes (DP). The MP assembles the traffic spatio-temporal parameters from ingress nodes, uses Kalman filtering prediction-based algorithm to estimate traffic demand. Prior to packets arrival at edges nodes, it regularly forwards updates of resources allocation to CPs. Simulations were done on a hybrid scheme (1+1) and on the centralized OBS. The results demonstrated that the proposition decreases the packet loss ratio. It also improves network latency and throughput-up to 84 and 51%, respectively, versus the traditional scheme.

Implementation of High Performance TCP Proxy Logic against TCP Flooding Attack on Network Interface Card (TCP 플러딩 공격 방어를 위한 네트워크 인터페이스용 고성능 TCP 프락시 제어 로직 구현)

  • Kim, Byoung-Koo;Kim, Ik-Kyun;Kim, Dae-Won;Oh, Jin-Tae;Jang, Jong-Soo;Chung, Tai-Myoung
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.21 no.2
    • /
    • pp.119-129
    • /
    • 2011
  • TCP-related Flooding attacks still dominate Distributed Denial of Service Attack. It is a great challenge to accurately detect the TCP flood attack in hish speed network. In this paper, we propose the NIC_Cookie logic implementation, which is a kind of security offload engine against TCP-related DDoS attacks, on network interface card. NIC_Cookie has robustness against DDoS attack itself and it is independent on server OS and external network configuration. It supports not IP-based response method but packet-level response, therefore it can handle attacks of NAT-based user group. We evaluate that the latency time of NIC_Cookie logics is $7{\times}10^{-6}$ seconds and we show 2Gbps wire-speed performance through a benchmark test.

Blockchain (A-PBFT) Based Authentication Method for Secure Lora Network (안전한 Lora 네트워크를 위한 블록체인(A-PBFT) 기반 인증 기법)

  • Kim, Sang-Geun
    • Journal of Industrial Convergence
    • /
    • v.20 no.10
    • /
    • pp.17-24
    • /
    • 2022
  • Lora, a non-band network technology of the long-distance wireless standard LPWAN standard, uses ABP and OTTA methods and AES-128-based encryption algorithm (shared key) for internal terminal authentication and integrity verification. Lora's recent firmware tampering vulnerability and shared-key encryption algorithm structure make it difficult to defend against MITM attacks. In this study, the consensus algorithm(PBFT) is applied to the Lora network to enhance safety. It performs authentication and PBFT block chain creation by searching for node groups using the GPS module. As a result of the performance analysis, we established a new Lora trust network and proved that the latency of the consensus algorithm was improved. This study is a 4th industry convergence study and is intended to help improve the security technology of Lora devices in the future.

Performance Evaluation Using Neural Network Learning of Indoor Autonomous Vehicle Based on LiDAR (라이다 기반 실내 자율주행 차량에서 신경망 학습을 사용한 성능평가 )

  • Yonghun Kwon;Inbum Jung
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.12 no.3
    • /
    • pp.93-102
    • /
    • 2023
  • Data processing through the cloud causes many problems, such as latency and increased communication costs in the communication process. Therefore, many researchers study edge computing in the IoT, and autonomous driving is a representative application. In indoor self-driving, unlike outdoor, GPS and traffic information cannot be used, so the surrounding environment must be recognized using sensors. An efficient autonomous driving system is required because it is a mobile environment with resource constraints. This paper proposes a machine-learning method using neural networks for autonomous driving in an indoor environment. The neural network model predicts the most appropriate driving command for the current location based on the distance data measured by the LiDAR sensor. We designed six learning models to evaluate according to the number of input data of the proposed neural networks. In addition, we made an autonomous vehicle based on Raspberry Pi for driving and learning and an indoor driving track produced for collecting data and evaluation. Finally, we compared six neural network models in terms of accuracy, response time, and battery consumption, and the effect of the number of input data on performance was confirmed.

Mobile Ultra-Broadband, Super Internet-of-Things and Artificial Intelligence for 6G Visions

  • Hamza Ali Alshawabkeh
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.12
    • /
    • pp.235-245
    • /
    • 2023
  • Smart applications based on the Network of Everything also known as Internet of Everything (IoE) are increasing popularity as network connectivity requires rise further. As a result, there will be a greater need for developing 6G technologies for wireless communications in order to overcome the primary limitations of visible 5G networks. Furthermore, implementing neural networks into 6G will bring remedies for the most complex optimizing networks challenges. Future 6G mobile phone networks must handle huge applications that require data and an increasing amount of users. With a ten-year time skyline from thought to the real world, it is presently time for pondering what 6th era (6G) remote correspondence will be just before 5G application. In this article, we talk about 6G dreams to clear the street for the headway of 6G and then some. We start with the conversation of imaginative 5G organizations and afterward underline the need of exploring 6G. Treating proceeding and impending remote organization improvement in a serious way, we expect 6G to contain three critical components: cell phones super broadband, very The Web of Things (or IoT and falsely clever (artificial intelligence). The 6G project is currently in its early phases, and people everywhere must envision and come up with its conceptualization, realization, implementation, and use cases. To that aim, this article presents an environment for Presented Distributed Artificial Intelligence as-a-Services (DAIaaS) supplying in IoE and 6G applications. The case histories and the DAIaaS architecture have been evaluated in terms of from end to end latency and bandwidth consumption, use of energy, and cost savings, with suggestion to improve efficiency.

IoT Edge Architecture Model to Prevent Blockchain-Based Security Threats (블록체인 기반의 보안 위협을 예방할 수 있는 IoT 엣지 아키텍처 모델)

  • Yoon-Su Jeong
    • Journal of Internet of Things and Convergence
    • /
    • v.10 no.2
    • /
    • pp.77-84
    • /
    • 2024
  • Over the past few years, IoT edges have begun to emerge based on new low-latency communication protocols such as 5G. However, IoT edges, despite their enormous advantages, pose new complementary threats, requiring new security solutions to address them. In this paper, we propose a cloud environment-based IoT edge architecture model that complements IoT systems. The proposed model acts on machine learning to prevent security threats in advance with network traffic data extracted from IoT edge devices. In addition, the proposed model ensures load and security in the access network (edge) by allocating some of the security data at the local node. The proposed model further reduces the load on the access network (edge) and secures the vulnerable part by allocating some functions of data processing and management to the local node among IoT edge environments. The proposed model virtualizes various IoT functions as a name service, and deploys hardware functions and sufficient computational resources to local nodes as needed.

Session Key Distribution Scheme in V2I of VANET using Identity-Based Cryptography (VANET의 V2I 환경에서 IBC를 이용한 세션키 분배 기법)

  • Roh, Hyo-Sun;Jung, Sou-Hwan
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.46 no.1
    • /
    • pp.112-120
    • /
    • 2009
  • This paper proposes a session key distribution scheme on non-interactive key distribution algorithm of Identity-based cryptography in V2I of VANET. In the current VANET, IEEE 802.11i is used to provide secure data communication between the vehicle and infrastructure. However, since the 4-way handshake procedure reply when the vehicle handover to another RSU/AP, IEEE 802.11i increases the communication overhead and latency. The proposed scheme using non-interactive key distribution algorithm of Identity-based cryptography provided session key generation and exchange without message exchange and reduced communication overhead and latency than the IEEE 802.11i.

Building Low Delay Application Layer Multicasting Trees for Streaming Services (스트리밍 서비스를 위한 적은 지연의 응용계층 멀티캐스트 트리 구축)

  • Kim, Jong-Gyung
    • The Journal of the Korea Contents Association
    • /
    • v.8 no.10
    • /
    • pp.20-26
    • /
    • 2008
  • The quality of stream remaking is decided the load of a server and Jitter through the traffic of the transmission path between end to end. In order to improve these problems in this paper, I propose tree construction method of low-delay-level-multicast. In this case which the network congestion will be occurred by streaming quality, I also propose the technique which dynamically changes the transmission path. This technique first constructs the overlay structure for relaxing the overload of server. Secondly, in order to decrease Jitter of client, it makes upload bandwidth and low latency balanced. In the evaluation of the performance, this paper showed better enhancement of about $15%{\sim}24%$ than P2CAST[4] in the simulation about node average join count, average bandwidth, service request refusal ratio, RTT measurement of nodes, and node average join count by defect ratio.

A Joint Wakeup Scheduling and MAC Protocol for Energy Efficient Data Forwarding in Wireless Sensor Networks (무선 센서 네트워크에서 에너지 효율적인 데이터 전송을 위한 스케줄링/ MAC 통합 프로토콜)

  • Cho, Jae-Kyu;Kwon, Tae-Kyoung;Choi, Yang-Hee
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.33 no.4B
    • /
    • pp.207-214
    • /
    • 2008
  • Under future internet environment, wireless sensor networks will be used in a wide range of applications. A major problem for designing sensor protocol is developing the most energy efficient technique to monitor an area of interest for a long time since sensors have some constraints such as small and a limited energy level. In addition, data latency is often a critical issue since sensory data is transmitted via multi hop fashion and need to be delivered timely for taking an appropriate action. Our motivation for designing a data forwarding protocol is to minimize energy consumption while keeping data latency bound in wireless sensor networks. In this paper, we propose a data forwarding protocol that consists of wakeup scheduling and MAC protocols, the latter of which is designed to achieve load balancing. Simulation results show that the proposed framework provides more energy-efficient delivery than other protocol.