• Title/Summary/Keyword: completion time algorithm

Search Result 120, Processing Time 0.029 seconds

All-port Broadcasting Algorithms on Wormhole Routed Star Graph Networks (웜홀 라우팅을 지원하는 스타그래프 네트워크에서 전 포트 브로드캐스팅 알고리즘)

  • Kim, Cha-Young;Lee, Sang-Kyu;Lee, Ju-Young
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.29 no.2
    • /
    • pp.65-74
    • /
    • 2002
  • Recently star networks are considered as attractive alternatives to the widely used hypercube for interconnection networks in parallel processing systems by many researchers. One of the fundamental communication problems on star graph networks is broadcasing In this paper we consider the broadcasting problems in star graph networks using wormhole routing. In wormhole routed system minimizing link contention is more critical for the system performance than the distance between two communicating nodes. We use Hamiltonian paths in star graph to set up link-disjoint communication paths We present a broadcast algorithm in n-dimensional star graph of N(=n!) nodes such that the total completion time is no larger than $([long_n n!]+1)$ steps where $([long_n n!]+1)$ is the lower bound This result is significant improvement over the previous n-1 step broadcasting algorithm.

QoS-Constrained Resource Reconfiguration Scheme based on Temporal Quorum Status Estimation in Computational Grids (계산 그리드에서의 시변 쿼럼 상태 예측을 기반으로 한 QoS 제약 자원 재구성 방법)

  • Byungsang, Kim;Dong Su, Nam;Bong Hwan, Lee
    • Journal of KIISE:Information Networking
    • /
    • v.31 no.5
    • /
    • pp.460-468
    • /
    • 2004
  • Quality of Service (QoS)-constrained policy has an advantage to guarantee QoS requirements requested by users. Quorum systems can ensure the consistency and availability of replicated data despite the benign failure of data repositories. We propose a Quorum based resource management scheme, which includes a system resource and network resource, both of which can satisfy the requirements of application QoS. We also propose the resource reconfiguration algorithm based on temporal execution time estimation method. Resource reconfiguration means that reshuffling the current available resource set for maintaining the quality level of the resources. We evaluate the effectiveness of resource reconfiguration mechanism with a Heart Hemodynamics analysis. Our approach shows the increase of the stability of execution environment as well as decrease the completion time compare to the method that is not adapted the resource reconfiguration.

Drone Force Deployment Optimization Algorithm For Efficient Military Drone Operations (효율적 군용 드론 작전 운영을 위한 Drone Force Deployment Optimization 알고리즘)

  • Song, Ju-Young;Jang, Hyeon-Deok;Chung, Jong-Moon
    • Journal of Internet Computing and Services
    • /
    • v.21 no.1
    • /
    • pp.211-219
    • /
    • 2020
  • One of the major advancements of the Fourth Industrial Revolution is the use of Internet of Drones (IoD), which combines the Internet of Things (IoT) and drone technology. IoD technology is especially important for efficiently and economically operating C4ISR operations in actual battlefields supporting various combat situations. The purpose of this study is to solve the problems of limited battery capacity of drones and lack of budgeting criteria for military drone transcription, introduction, and operation. If the mission area is defined and corresponding multi-drone hovering check points and mission completion time limits are set, then an energy and time co-optimized scheduling and operation control scheme is needed. Because such a scheme does not exist, in this paper, a Drone Force Deployment Optimization (DFDO) scheme is proposed to help schedule multi-drone operation scheduling and networked based remote multi-drone control.

A Proposal for Partial Automation Preparation System of BIM-based Energy Conservation Plan - Case Study on Automation Process Using BIM Software and Excel VBA - (BIM기반 에너지절약계획서 건축부문 부분자동화 작성 시스템 제안 - BIM 소프트웨어와 EXCEL VBA를 이용한 자동화과정을 중심으로 -)

  • Ryu, Jea-Ho;Hwang, Jong-Min;Kim, Sol-Yee;Seo, Hwa-Yeong;Lee, Ji-Hyun
    • Journal of KIBIM
    • /
    • v.12 no.2
    • /
    • pp.49-59
    • /
    • 2022
  • The main idea of this study is to propose a BIM-based automation system drawing up a report of energy conservation plan in the architecture division. In order to obtain a building permit, an energy conservation plan must be prepared for buildings with a total floor area of 500m2 or more under the current law. Currently, it is adopted as a general method to complete a report by obtaining data and drawings necessary for an energy conservation plan through manual work and input them directly into the verification system. This method takes a lot of effort and time in the design phase which ultimately increases the initial cost of the business, including the services of companies specialized in the environmental field. However, in preparation for mandatory BIM work process in the future, it is necessary to introduce BIM-based automatic creation system that has an advantage for shortening the whole process to enable rapid permission of energy-saving designs for buildings. There may be many methods of automation, but this study introduces how to build an application using Dynamo of Revit, in terms of utilizing BIM, and write an energy conservation plan by automatic completion of report through Dynamo and Excel's VBA algorithm, which can save time and cost in preparing the report of energy conservation plan compared with the manual process. Also we have insisted that the digital transformation of architectural process is a necessary for an efficient use of our automation system in the current energy conservation plan workflow.

Optimal path planning and analysis for the maximization of multi UAVs survivability for missions involving multiple threats and locations (다수의 위협과 복수의 목적지가 존재하는 임무에서 복수 무인기의 생존율 극대화를 위한 최적 경로 계획 및 분석)

  • Jeong, Seongsik;Jang, Dae-Sung;Park, Hyunjin;Seong, Taehyun;Ahn, Jaemyung
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.43 no.6
    • /
    • pp.488-496
    • /
    • 2015
  • This paper proposes a framework to determine the routes of multiple unmanned aerial vehicles (UAVs) to conduct multiple tasks in different locations considering the survivability of the vehicles. The routing problem can be formulated as the vehicle routing problem (VRP) with different cost matrices representing the trade-off between the safety of the UAVs and the mission completion time. The threat level for a UAV at a certain location was modeled considering the detection probability and the shoot-down probability. The minimal-cost path connecting two locations considering the threat level and the flight distance was obtained using the Dijkstra algorithm in hexagonal cells. A case study for determining the optimal routes for a persistent multi-UAVs surveillance and reconnaissance missions given multiple enemy bases was conducted and its results were discussed.

Evaluation of Short and Long-Term Modal Parameters of a Cable-Stayed Bridge Based on Operational Modal Analysis (운용모드해석에 기반한 사장교의 장단기 동특성 평가)

  • Park, Jong-Chil
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.26 no.4
    • /
    • pp.20-29
    • /
    • 2022
  • The operational modal analysis (OMA) technique, which extracts the modal parameters of a structural system using ambient vibrations, has been actively developed as a field of structural health monitoring of cable-supported bridges. In this paper, the short and long-term modal parameters of a cable-stayed bridge were evaluated using the acceleration data obtained from the two ambient vibration tests (AVTs) and three years of continuous measurements. A total of 27 vertical modes and 1 lateral mode in the range 0.1 ~ 2.5 Hz were extracted from the high-resolution AVTs which were conducted in the 6th and 19th years after its completion. Existing OMA methods such as Peak-Picking (PP), Eigensystem Realization Algorithm with Data Correlation (ERADC), Frequency Domain Decomposition (FDD) and Time Domain Decomposition (TDD) were applied for modal parameters extraction, and it was confirmed that there was no significant difference between the applied methods. From the correlation analysis between long-term natural frequencies and environmental factors, it was confirmed that temperature change is the dominant factor influencing natural frequency fluctuations. It was revealed that the decreased natural frequencies of the bridge were not due to changes in structural performance and integrity, but to the environmental effects caused by the temperature difference between the two AVTs. In addition, when the TDD technique is applied, the accuracy of extracted mode shapes is improved by adding a proposed algorithm that normalizes the sequence so that the autocorrelations at zero lag equal 1.

ATM Cell Encipherment Method using Rijndael Algorithm in Physical Layer (Rijndael 알고리즘을 이용한 물리 계층 ATM 셀 보안 기법)

  • Im Sung-Yeal;Chung Ki-Dong
    • The KIPS Transactions:PartC
    • /
    • v.13C no.1 s.104
    • /
    • pp.83-94
    • /
    • 2006
  • This paper describes ATM cell encipherment method using Rijndael Algorithm adopted as an AES(Advanced Encryption Standard) by NIST in 2001. ISO 9160 describes the requirement of physical layer data processing in encryption/decryption. For the description of ATM cell encipherment method, we implemented ATM data encipherment equipment which satisfies the requirements of ISO 9160, and verified the encipherment/decipherment processing at ATM STM-1 rate(155.52Mbps). The DES algorithm can process data in the block size of 64 bits and its key length is 64 bits, but the Rijndael algorithm can process data in the block size of 128 bits and the key length of 128, 192, or 256 bits selectively. So it is more flexible in high bit rate data processing and stronger in encription strength than DES. For tile real time encryption of high bit rate data stream. Rijndael algorithm was implemented in FPGA in this experiment. The boundary of serial UNI cell was detected by the CRC method, and in the case of user data cell the payload of 48 octets (384 bits) is converted in parallel and transferred to 3 Rijndael encipherment module in the block size of 128 bits individually. After completion of encryption, the header stored in buffer is attached to the enciphered payload and retransmitted in the format of cell. At the receiving end, the boundary of ceil is detected by the CRC method and the payload type is decided. n the payload type is the user data cell, the payload of the cell is transferred to the 3-Rijndael decryption module in the block sire of 128 bits for decryption of data. And in the case of maintenance cell, the payload is extracted without decryption processing.

Measuring the Public Service Quality Using Process Mining: Focusing on N City's Building Licensing Complaint Service (프로세스 마이닝을 이용한 공공서비스의 품질 측정: N시의 건축 인허가 민원 서비스를 중심으로)

  • Lee, Jung Seung
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.35-52
    • /
    • 2019
  • As public services are provided in various forms, including e-government, the level of public demand for public service quality is increasing. Although continuous measurement and improvement of the quality of public services is needed to improve the quality of public services, traditional surveys are costly and time-consuming and have limitations. Therefore, there is a need for an analytical technique that can measure the quality of public services quickly and accurately at any time based on the data generated from public services. In this study, we analyzed the quality of public services based on data using process mining techniques for civil licensing services in N city. It is because the N city's building license complaint service can secure data necessary for analysis and can be spread to other institutions through public service quality management. This study conducted process mining on a total of 3678 building license complaint services in N city for two years from January 2014, and identified process maps and departments with high frequency and long processing time. According to the analysis results, there was a case where a department was crowded or relatively few at a certain point in time. In addition, there was a reasonable doubt that the increase in the number of complaints would increase the time required to complete the complaints. According to the analysis results, the time required to complete the complaint was varied from the same day to a year and 146 days. The cumulative frequency of the top four departments of the Sewage Treatment Division, the Waterworks Division, the Urban Design Division, and the Green Growth Division exceeded 50% and the cumulative frequency of the top nine departments exceeded 70%. Higher departments were limited and there was a great deal of unbalanced load among departments. Most complaint services have a variety of different patterns of processes. Research shows that the number of 'complementary' decisions has the greatest impact on the length of a complaint. This is interpreted as a lengthy period until the completion of the entire complaint is required because the 'complement' decision requires a physical period in which the complainant supplements and submits the documents again. In order to solve these problems, it is possible to drastically reduce the overall processing time of the complaints by preparing thoroughly before the filing of the complaints or in the preparation of the complaints, or the 'complementary' decision of other complaints. By clarifying and disclosing the cause and solution of one of the important data in the system, it helps the complainant to prepare in advance and convinces that the documents prepared by the public information will be passed. The transparency of complaints can be sufficiently predictable. Documents prepared by pre-disclosed information are likely to be processed without problems, which not only shortens the processing period but also improves work efficiency by eliminating the need for renegotiation or multiple tasks from the point of view of the processor. The results of this study can be used to find departments with high burdens of civil complaints at certain points of time and to flexibly manage the workforce allocation between departments. In addition, as a result of analyzing the pattern of the departments participating in the consultation by the characteristics of the complaints, it is possible to use it for automation or recommendation when requesting the consultation department. In addition, by using various data generated during the complaint process and using machine learning techniques, the pattern of the complaint process can be found. It can be used for automation / intelligence of civil complaint processing by making this algorithm and applying it to the system. This study is expected to be used to suggest future public service quality improvement through process mining analysis on civil service.

Development of a Stock Trading System Using M & W Wave Patterns and Genetic Algorithms (M&W 파동 패턴과 유전자 알고리즘을 이용한 주식 매매 시스템 개발)

  • Yang, Hoonseok;Kim, Sunwoong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.63-83
    • /
    • 2019
  • Investors prefer to look for trading points based on the graph shown in the chart rather than complex analysis, such as corporate intrinsic value analysis and technical auxiliary index analysis. However, the pattern analysis technique is difficult and computerized less than the needs of users. In recent years, there have been many cases of studying stock price patterns using various machine learning techniques including neural networks in the field of artificial intelligence(AI). In particular, the development of IT technology has made it easier to analyze a huge number of chart data to find patterns that can predict stock prices. Although short-term forecasting power of prices has increased in terms of performance so far, long-term forecasting power is limited and is used in short-term trading rather than long-term investment. Other studies have focused on mechanically and accurately identifying patterns that were not recognized by past technology, but it can be vulnerable in practical areas because it is a separate matter whether the patterns found are suitable for trading. When they find a meaningful pattern, they find a point that matches the pattern. They then measure their performance after n days, assuming that they have bought at that point in time. Since this approach is to calculate virtual revenues, there can be many disparities with reality. The existing research method tries to find a pattern with stock price prediction power, but this study proposes to define the patterns first and to trade when the pattern with high success probability appears. The M & W wave pattern published by Merrill(1980) is simple because we can distinguish it by five turning points. Despite the report that some patterns have price predictability, there were no performance reports used in the actual market. The simplicity of a pattern consisting of five turning points has the advantage of reducing the cost of increasing pattern recognition accuracy. In this study, 16 patterns of up conversion and 16 patterns of down conversion are reclassified into ten groups so that they can be easily implemented by the system. Only one pattern with high success rate per group is selected for trading. Patterns that had a high probability of success in the past are likely to succeed in the future. So we trade when such a pattern occurs. It is a real situation because it is measured assuming that both the buy and sell have been executed. We tested three ways to calculate the turning point. The first method, the minimum change rate zig-zag method, removes price movements below a certain percentage and calculates the vertex. In the second method, high-low line zig-zag, the high price that meets the n-day high price line is calculated at the peak price, and the low price that meets the n-day low price line is calculated at the valley price. In the third method, the swing wave method, the high price in the center higher than n high prices on the left and right is calculated as the peak price. If the central low price is lower than the n low price on the left and right, it is calculated as valley price. The swing wave method was superior to the other methods in the test results. It is interpreted that the transaction after checking the completion of the pattern is more effective than the transaction in the unfinished state of the pattern. Genetic algorithms(GA) were the most suitable solution, although it was virtually impossible to find patterns with high success rates because the number of cases was too large in this simulation. We also performed the simulation using the Walk-forward Analysis(WFA) method, which tests the test section and the application section separately. So we were able to respond appropriately to market changes. In this study, we optimize the stock portfolio because there is a risk of over-optimized if we implement the variable optimality for each individual stock. Therefore, we selected the number of constituent stocks as 20 to increase the effect of diversified investment while avoiding optimization. We tested the KOSPI market by dividing it into six categories. In the results, the portfolio of small cap stock was the most successful and the high vol stock portfolio was the second best. This shows that patterns need to have some price volatility in order for patterns to be shaped, but volatility is not the best.

The Availability of the step optimization in Monaco Planning system (모나코 치료계획 시스템에서 단계적 최적화 조건 실현의 유용성)

  • Kim, Dae Sup
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.26 no.2
    • /
    • pp.207-216
    • /
    • 2014
  • Purpose : We present a method to reduce this gap and complete the treatment plan, to be made by the re-optimization is performed in the same conditions as the initial treatment plan different from Monaco treatment planning system. Materials and Methods : The optimization is carried in two steps when performing the inverse calculation for volumetric modulated radiation therapy or intensity modulated radiation therapy in Monaco treatment planning system. This study was the first plan with a complete optimization in two steps by performing all of the treatment plan, without changing the optimized condition from Step 1 to Step 2, a typical sequential optimization performed. At this time, the experiment was carried out with a pencil beam and Monte Carlo algorithm is applied In step 2. We compared initial plan and re-optimized plan with the same optimized conditions. And then evaluated the planning dose by measurement. When performing a re-optimization for the initial treatment plan, the second plan applied the step optimization. Results : When the common optimization again carried out in the same conditions in the initial treatment plan was completed, the result is not the same. From a comparison of the treatment planning system, similar to the dose-volume the histogram showed a similar trend, but exhibit different values that do not satisfy the conditions best optimized dose, dose homogeneity and dose limits. Also showed more than 20% different in comparison dosimetry. If different dose algorithms, this measure is not the same out. Conclusion : The process of performing a number of trial and error, and you get to the ultimate goal of treatment planning optimization process. If carried out to optimize the completion of the initial trust only the treatment plan, we could be made of another treatment plan. The similar treatment plan could not satisfy to optimization results. When you perform re-optimization process, you will need to apply the step optimized conditions, making sure the dose distribution through the optimization process.