• Title/Summary/Keyword: network time-delay

Search Result 1,471, Processing Time 0.033 seconds

The MCSTOP Algorithm about the Minimum Cost Spanning Tree and the Optimum Path Generation for the Multicasting Path Assignment (최적 경로 생성 및 최소 비용 신장 트리를 이용한 멀티캐스트 경로 배정 알고리즘 : MCSTOP)

  • Park, Moon-Sung;Kim, Jin-Suk
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.4
    • /
    • pp.1033-1043
    • /
    • 1998
  • In this paper, we present an improved multicasting path assignment algorithm based on the minimum cost spanning tree. In the method presented in this paper, a multicasting path is assigned preferentially when a node to be received is found among the next degree nodes of the searching node in the multicasting path assignment of the constrained steiner tree (CST). If nodes of the legacy group exist between nodes of the new group, a new path among the nodes of new group is assigned as long as the nodes may be excluded from the new multicasting path assignment taking into consideration characteristics of nodes in the legacy group. In assigning the multicasting path additionally, where the source and destination nodes which can be set for the new multicasting path exist in the domain of identical network (local area network) and conditions for degree constraint are satisfied, a method of producing and assigning a new multicasting path is used. The results of comparison of CST with MCSTOP, MCSTOp algorithm enhanced performance capabilities about the communication cost, the propagation delay, and the computation time for the multicasting assignment paths more than CST algorithm. Further to this, research activities need study for the application of the international standard protocol(multicasting path assignment technology in the multipoint communication service (MCS) of the ITU-T T.120).

  • PDF

Patent Analysis on 5G Technology Trends from the Perspective of Smart Factory (특허 분석을 통한 스마트공장 관점의 5G 기술개발 동향 연구)

  • Cho, Eunnuri;Chang, Tai-Woo
    • The Journal of Society for e-Business Studies
    • /
    • v.25 no.3
    • /
    • pp.95-108
    • /
    • 2020
  • The development of 5G technology, which is a next-generation communication technology capable of processing large amounts of data in real-time and solving delays, is drawing attention. Not only in the United States but also Korea, 5G is focused on supporting R&D as a national strategic technology. The strategy for the smart factory, one of the core services of the 5G, aims to increase the flexibility of manufacturing production lines. The existing wired communications devices can be replaced into wireless ones with the ultra-low-delay and ultra-high-speed characteristics of 5G. For the efficient development of 5G technology, it is necessary to keep abreast of the status and trend. In this study, based on the collected data of 1517 Korea patents and 1928 US patents, 5G technologies trend was analyzed and key technologies were identified by network analysis and topic modeling. We expect that it will be used for decision making for policy establishment and technology strategy of related industries to provide the trends of technology development related to the introduction of 5G technology to smart factories.

A Flow Control Scheme based on Queue Priority (큐의 우선순위에 근거한 흐름제어방식)

  • Lee, Gwang-Jun;Son, Ji-Yeon;Son, Chang-Won
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.1
    • /
    • pp.237-245
    • /
    • 1997
  • In this paper, a flow control mechanism is proposed which is based on the priority control between communication path of a node. In this scheme, demanding length of a data queue for any pre-defined, then each node in that path is forced to maintains buffer size under the limit by controlling priority level of the path. The communication path which requires higher bandwidth sets its demanding queue length smaller. By providing relationship between the priority of a path and length of its queue, the high bandwidth requesting path has a better chance to get high bandwidth by defining the smaller demanding queue size. And also, by forcing a path which has high flow rate to maintain small queue size in the path of the communication, the scheme keep the transmission delay of the path small. The size of the demanding queue of a path is regularly adjusted to meet the applications requirement, and the load status of the network during the life time of the communication. The priority control based on the demanding queue size is also provided in the intermediate nodes as well as the end nodes. By that the flow control can provide a quicker result than end to-end flow control, it provides better performance advantage especially for the high speed network.

  • PDF

Efficient Parallel Spatial Join Processing Method in a Shared-Nothing Database Cluster System (비공유 공간 클러스터 환경에서 효율적인 병렬 공간 조인 처리 기법)

  • Chung, Warn-Ill;Lee, Chung-Ho;Bae, Hae-Young
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.591-602
    • /
    • 2003
  • Delay and discontinuance phenomenon of service are cause by sudden increase of the network communication amount and the quantity consumed of resources when Internet users are driven excessively to a conventional single large database sewer. To solve these problems, spatial database cluster consisted of several single nodes on high-speed network to offer high-performance is risen. But, research about spatial join operation that can reduce the performance of whole system in case process at single node is not achieved. So, in this paper, we propose efficient parallel spatial join processing method in a spatial database cluster system that uses data partitions and replications method that considers the characteristics of space data. Since proposed method does not need the creation step and the assignment step of tasks, and does not occur additional message transmission between cluster nodes that appear in existent parallel spatial join method, it shows performance improvement of 23% than the conventional parallel R-tree spatial join for a shared-nothing architecture about expensive spatial join queries. Also, It can minimize the response time to user because it removes redundant refinement operation at each cluster node.

Handover Scheme between WiFi and Mobile WiMax (WiFi와 mobile WiMax간 핸드오버 방안)

  • Park, Seung-Kyun
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.1
    • /
    • pp.34-41
    • /
    • 2011
  • At present wireless internet access service is available through the 3G network, mobile WiMAX and WiFi anytime and anywhere. In this environment where there are various networks, users should be able to select specific networks depending on different situations. And it is necessary to provide mobility support between homogeneous and between heterogenous networks. Given this situation, the many proposals have been presented to link 3G, which has the largest service area among various networks, with mobile WiMAX(IEEE 802.16e), or with WiFi(IEEE 802.11). But, recently, with the increasing volume of wireless internet use and wireless internet data, due to the advents of net-book, e-book and smart phone, the service area of WiFi and mobile WiMAX has rapidly expanded. Especially, the availability of real-time application such as internet phone has led to the relative shrinking of the proportion of 3G mobile communication network giving conventional voice service, and enlargement of those of wireless internet access networks like WiFi and mobile WiMAX. This paper suggests a handover scheme based on PMIPv6, whitch support mobility between WiFi and mobile WiMAX, and minimizes handover delay. In this scheme, the mobile node has a dual stack structure composed of two interfaces-WiFi and mobile WiMAX. Since WiFi dose not support mobility, it is suggested that the mobile node have the capacity to deal with handover signaling between gateway in case of handover between homogeneous networks. This handover scheme, suggested comparing with current handovers between homogeneous networks, has proved, in its analytic evaluation, to be able to reduce handover, transmission, and signaling overhead.

Congestion Control Algorithms Evaluation of TCP Linux Variants in Dumbbell (덤벨 네트워크에서 TCP 리눅스 변종의 혼잡 제어 알고리즘 평가)

  • Mateen, Ahamed;Zaman, Muhanmmad
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.16 no.1
    • /
    • pp.139-145
    • /
    • 2016
  • Dumbbell is the most basic topology that can be used in almost all kind of network experiment within it or just by little expansion. While Transmission Control Protocol TCP is the basic protocol that is used for the connectivity among networks and stations. TCP major and basic goal is to provide path and services to different applications for communication. For that reason TCP has to transfer a lot of data through a communication medium that cause serious congestion problem. To calculate the congestion problem, different kind of pre-cure solutions are developer which are Loss Based Variant and Delay Based Variant. While LBV keep track of the data that is going to be passed through TCP protocol, if the data packets start dropping that means congestion occurrence which notify as a symptom, TCP CUBIC use LBV for notifying the loss. Similarly the DBV work with the acknowledgment procedure that is used in when data ACK get late with respect to its set data rate time, TCP COMPOUND/VAGAS are examples of DBV. Many algorithms have been purposed to control the congestion in different TCP variants but the loss of data packets did not completely controlled. In this paper, the congestion control algorithms are implemented and corresponding results are analyzed in Dumbbell topology, it is typically used to analyze the TCP traffic flows. Fairness of throughput is evaluated for different TCP variants using network simulator (NS-2).

Macroscopic-Microscopic Sequential Traffic Simulation Analysis and Dynamic O/D Estimation for Sub-area (거시-미시 순차적 교통시뮬레이션 방법과 부분상세지구의 동적 O/D추정)

  • Lee, Jin Hak;Kim, Ikki;Kim, Dae Hyun
    • Journal of Korean Society of Transportation
    • /
    • v.32 no.6
    • /
    • pp.567-578
    • /
    • 2014
  • The study suggested a method to improve analysis accuracy such that the interactive effects of transportation changes between outside and inside of sub-areas were sequentially considered in the analysis by linking a macroscopic network analysis and a microscopic traffic simulation. A dynamic O/D estimation method was developed for practical implement of sub-area microscopic simulation analysis by using the results of macroscopic network analysis, the results of selected link analysis at the cordon line of the sub-area, departure time data of household travel survey, timely observed traffic volume data at the cordon. This estimated dynamic O/D for the sub-area made it possible to analyze traffic phenomena in details. Various detailed phenomena such as traffic queues, delay at intersection, and conflicts between vehicles, which is impossible to be grasped through a macroscopic analysis, can be analyzed with the dynamic microscopic traffic analysis. Through implementing an empirical study and validation, the study provided a reference result about accuracy of a microscopic traffic simulation of a sub-area to help its application for real transportation policy analysis.

A Maximally Disjoint Multipath Routing Protocol Based on AODV in Mobile Ad Hoc Networks (모바일 애드혹 네트워크에서의 AODV 기반 치대 비중첩 다중경로 라우팅 프로토콜)

  • Kim Jungtae;Moh Sangman;Chung Ilyong
    • The KIPS Transactions:PartC
    • /
    • v.12C no.3 s.99
    • /
    • pp.429-436
    • /
    • 2005
  • A mobile ad hoc network (MANET) is a collection of mobile nodes without any fixed infrastructure or my form of centralized administration such as access points and base stations. The ad hoc on-demand distance vector routing (AODV) protocol is an on-demand routing protocol for MANETs, which is one of the Internet-Drafts submitted to the Internet engineering task force (IETF) MANET working group. This paper proposes a new multipath routing protocol called maximally disjoint multipath AODV (MDAODV), which exploits maximally node- and link-disjoint paths and outperforms the conventional multipath protocol based on AODV as well as the basic AODV protocol. The key idea is to extend only route request (RREQ) message by adding source routing information and to make the destination node select two paths from multiple RREQs received for a predetermined time period. Compared to the conventional multipath routing protocol, the proposed MDAODV provides more reliable and robust routing paths and higher performance. It also makes the destination node determine the maximally node- and link-disjoint paths, reducing the overhead incurred at intermediate nodes. Our extensive simulation study shows that the proposed MDAODV outperforms the conventional multipath routing protocol based on AODV in terms of packet delivery ratio and average end-to-end delay, and reduces routing overhead.

Distributed Edge Computing for DNA-Based Intelligent Services and Applications: A Review (딥러닝을 사용하는 IoT빅데이터 인프라에 필요한 DNA 기술을 위한 분산 엣지 컴퓨팅기술 리뷰)

  • Alemayehu, Temesgen Seyoum;Cho, We-Duke
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.12
    • /
    • pp.291-306
    • /
    • 2020
  • Nowadays, Data-Network-AI (DNA)-based intelligent services and applications have become a reality to provide a new dimension of services that improve the quality of life and productivity of businesses. Artificial intelligence (AI) can enhance the value of IoT data (data collected by IoT devices). The internet of things (IoT) promotes the learning and intelligence capability of AI. To extract insights from massive volume IoT data in real-time using deep learning, processing capability needs to happen in the IoT end devices where data is generated. However, deep learning requires a significant number of computational resources that may not be available at the IoT end devices. Such problems have been addressed by transporting bulks of data from the IoT end devices to the cloud datacenters for processing. But transferring IoT big data to the cloud incurs prohibitively high transmission delay and privacy issues which are a major concern. Edge computing, where distributed computing nodes are placed close to the IoT end devices, is a viable solution to meet the high computation and low-latency requirements and to preserve the privacy of users. This paper provides a comprehensive review of the current state of leveraging deep learning within edge computing to unleash the potential of IoT big data generated from IoT end devices. We believe that the revision will have a contribution to the development of DNA-based intelligent services and applications. It describes the different distributed training and inference architectures of deep learning models across multiple nodes of the edge computing platform. It also provides the different privacy-preserving approaches of deep learning on the edge computing environment and the various application domains where deep learning on the network edge can be useful. Finally, it discusses open issues and challenges leveraging deep learning within edge computing.

A User Optimer Traffic Assignment Model Reflecting Route Perceived Cost (경로인지비용을 반영한 사용자최적통행배정모형)

  • Lee, Mi-Yeong;Baek, Nam-Cheol;Mun, Byeong-Seop;Gang, Won-Ui
    • Journal of Korean Society of Transportation
    • /
    • v.23 no.2
    • /
    • pp.117-130
    • /
    • 2005
  • In both deteministic user Optimal Traffic Assignment Model (UOTAM) and stochastic UOTAM, travel time, which is a major ccriterion for traffic loading over transportation network, is defined by the sum of link travel time and turn delay at intersections. In this assignment method, drivers actual route perception processes and choice behaviors, which can become main explanatory factors, are not sufficiently considered: therefore may result in biased traffic loading. Even though there have been some efforts in Stochastic UOTAM for reflecting drivers' route perception cost by assuming cumulative distribution function of link travel time, it has not been fundamental fruitions, but some trials based on the unreasonable assumptions of Probit model of truncated travel time distribution function and Logit model of independency of inter-link congestion. The critical reason why deterministic UOTAM have not been able to reflect route perception cost is that the route perception cost has each different value according to each origin, destination, and path connection the origin and destination. Therefore in order to find the optimum route between OD pair, route enumeration problem that all routes connecting an OD pair must be compared is encountered, and it is the critical reason causing computational failure because uncountable number of path may be enumerated as the scale of transportation network become bigger. The purpose of this study is to propose a method to enable UOTAM to reflect route perception cost without route enumeration between an O-D pair. For this purpose, this study defines a link as a least definition of path. Thus since each link can be treated as a path, in two links searching process of the link label based optimum path algorithm, the route enumeration between OD pair can be reduced the scale of finding optimum path to all links. The computational burden of this method is no more than link label based optimum path algorithm. Each different perception cost is embedded as a quantitative value generated by comparing the sub-path from the origin to the searching link and the searched link.