• Title/Summary/Keyword: Complexity analysis

Search Result 2,370, Processing Time 0.033 seconds

The Effects of Emergent Leader on Team Cognitive Complexity and Team Performance

  • Choi, Kyoosang
    • Journal of the Korean Data Analysis Society
    • /
    • v.20 no.6
    • /
    • pp.2781-2792
    • /
    • 2018
  • From a cognitive perspective, this study investigates the role of emergent leaders in developing team cognition and affecting team performance. With application of the cognitive complexity theory, this study hypothesizes that emergent leaders' cognitive complexity will be positively associated with team cognitive complexity, and that team cognitive complexity will be positively associated with team performance. In addition, team cognitive complexity is hypothesized to mediate the effect of the cognitive complexity of emergent leaders on team performance. To test the research hypotheses, data were obtained from 100 teams comprising a total of 339 undergraduate students who participated in a business simulation game. The findings of this study suggests that the cognitive complexity of emergent leaders is a significant predictor of team cognitive complexity, and that team cognitive complexity is positively related to team performance. Moreover, team cognitive complexity significantly mediates the effect of emergent leaders' cognitive complexity and team performance.

Determining the complexity level of proceduralized tasks in a digitalized main control room using the TACOM measure

  • Inseok Jang;Jinkyun Park
    • Nuclear Engineering and Technology
    • /
    • v.54 no.11
    • /
    • pp.4170-4180
    • /
    • 2022
  • The task complexity (TACOM) measure was previously developed to quantify the complexity of proceduralized tasks conducted by nuclear power plant operators. Following the development of the TACOM measure, its appropriateness has been validated by investigating the relationship between TACOM scores and three kinds of human performance data, namely response times, human error probabilities, and subjective workload scores. However, the information reflected in quantified TACOM scores is still insufficient to determine the levels of complexity of proceduralized tasks for human reliability analysis (HRA) applications. In this regard, the objective of this study is to suggest criteria for determining the levels of task complexity based on logistic regression between human error occurrences in digitalized main control rooms and TACOM scores. Analysis results confirmed that the likelihood of human error occurrence according to the TACOM score is secured. This result strongly implies that the TACOM measure can be used to identify the levels of task complexity, which could be applicable to various research domains including HRA.

Case study of the large switching software metrics and their fault analysis (대형 교환 소프트웨어의 복잡성과 고장분석 사례 연구)

  • 이재기;남상식;김창봉;이규대
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.27 no.10C
    • /
    • pp.887-901
    • /
    • 2002
  • Software management model divided into the software project model and design estimation model, software matrices model, reliability growth model, process improvement model(or process maturity model) etc. Among these software management models, software complexity model make an estimated of the product software. For a practice of software managed, need to guideline of the static analysis of software. Especially, Software complexity model introduced for the estimation of software quantity and program complexity. In case of measurement the software matrices, its need for us to analysis of software quality and products. On the other hand, we known that complexity program include many defects and consuming of source cost. So, we apply to complexity model using of the program complexity, control structure and volume matrices, interface metrics, process complexity metrics method. In this paper, we represent that the analysis of fault data detected during the system test. Also, we analysis of program control structure and interface, volume matrices in various aspect of switching software. Others, their results utilized similar of project and system development.

Measurement of Classes Complexity in the Object-Oriented Analysis Phase (객체지향 분석 단계에서의 클래스 복잡도 측정)

  • Kim, Yu-Kyung;Park, Jai-Nyun
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.10
    • /
    • pp.720-731
    • /
    • 2001
  • Complexity metrics have been developed for the structured paradigm of software development are not suitable for use with the object-oriented(OO) paradigm, because they do not support key object-oriented concepts such as inheritance, polymorphism. message passing and encapsulation. There are many researches on OO software metrics such as program complexity or design metrics. But metrics measuring the complexity of classes at the OO analysis phase are needed because they provide earlier feedback to the development project. and earlier feedback means more effective developing and less costly maintenance. In this paper, we propose the new metrics to measure the complexity of analysis classes which draw out in the analysis based on RUP(Rational Unified Process). By the collaboration complexity, is denoted by CC, we mean the maximum number of the collaborations can be achieved with each of the collaborator and determine the potential complexity. And the interface complexity, is denoted by IC, shows the difficulty related to understand the interface of collaborators each other. We verify theoretically the suggested metrics for Weyuker's nine properties. Moreover, we show the computation results for analysis classes of the system which automatically respond to questions of the user using the text mining technique. As a result of the comparison of CC and CBO and WMC suggested by Chidamber and Kemerer, the class that have highly the proposed metric value maintain the high complexity at the design phase too. And the complexity can be represented by CC and IC more than CBO and WMC. We can expect that our metrics may provide us the earlier feedback and hence possible to predict the efforts, costs and time required to remainder processes. As a result, we expect to develop the cost-effective OO software by reviewing the complexity of analysis classes in the first stage of SDLC(Software Development Life Cycle).

  • PDF

The Relationship between English Proficiency and Syntactic Complexity for Korean College Students (한국 대학생의 에세이에 나타난 영어 능력 수준과 통사적 복잡성 간의 관계 탐색)

  • Lee, Young-Ju
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.3
    • /
    • pp.439-444
    • /
    • 2021
  • This study investigates the relationship between syntactic complexity and English proficiency for Korean college students, using the recently developed TAASSC(the Tool for the Automatic Analysis of Syntactic Sophistication and Complexity) program. Essays on the ICNALE(International Corpus Network of Asian Learners of English) corpus were employed and phrasal complexity indices and clausal complexity indices, respectively were used to predict English proficiency level for Korean students. Results of stepwise regression analysis showed that indices of phrasal complexity explained 8% of variance in English proficiency, while indices of clausal complexity accounted for approximately 11%. That is, indices of clausal complexity were slightly better predictors of English proficiency than indices of phrasal complexity, which contradicts Biber et at.(2011)'s claim that phrasal complexity is the hallmark of writing development.

Regression Analysis of the Relationships between Complexity Metrics and Faults on the Telecommunication Program (통신 소프트웨어의 프로그램 결함과 복잡도의 관련성 분석을 위한 회귀분석 모델)

  • Lee, Gyeong-Hwan;Jeong, Chang-Sin;Hwang, Seon-Myeong;Jo, Byeong-Gyu;Park, Ji-Hun;Kim, Gang-Tae
    • Journal of KIISE:Software and Applications
    • /
    • v.26 no.11
    • /
    • pp.1282-1287
    • /
    • 1999
  • 통신 프로그램은 고도의 신뢰성과 기능성, 확장성, 그리고 유지 보수성이 필요하다. 프로그램 테스트의 결과와 McCabe의 Complexity를 측정한 데이타를 가지고 회귀모델을 만들고 그 신뢰성을 분석함으로서 프로그램의 결함과 복잡도의 관련성을 평가한다.본 연구에서 사용한 통신 프로그램은 500개 블록이 59가지 기능을 수행하는 교환 기능 중에서 복잡도가 너무 많아서 통계 처리의 bias가 될 블록을 제외하고 394 블록을 선정하여 SAS에 의해서 통계 분석을 하고 회귀 분석 모델을 설계하였다. t 분포에 의하여 방정식의 유의성 수준을 검증하고 프로그램의 결함수에 가장 큰 영향을 주고 있는 복잡도가 McCabe의 복잡도와 설계 복잡도 임을 밝혀냈다. 이 연구 결과에 의해서 설계 정보 및 유지 보수 정보를 얻을 수 있다. Abstract Switching software requires high reliability, functionality, extendability and maintainability. For doing, software quality model based on MaCabe's complexity measure is investigated. It is experimentally shown using regression analysis the program fault density depends on the complexity and size of the function unit. The software should be verified and tested if it satisfies its requirements with automated analysis tools. In this paper we propose the regression model with the test data.The sample program for the regression model consists of more than 500 blocks, where each block compose of 10 files, which has 59 functions of switching activity.Among them we choose 394 blocks and analyzed for 59 functions by testing tools and SAS package. We developed Regression Analysis Model and evaluated significant of the equation based on McCabe's cyclomatic complexity, block design complexity, design complexity, and integration complexity.The results of our experimental study are that number of fault are under the influence of McCabe's complexity number and design complexity.

Complexity Metrics for Analysis Classes in the Unified Software Development Process (Unified Process의 분석 클래스에 대한 복잡도 척도)

  • 김유경;박재년
    • The KIPS Transactions:PartD
    • /
    • v.8D no.1
    • /
    • pp.71-80
    • /
    • 2001
  • Object-Oriented (OO) methodology to use the concept like encapsulation, inheritance, polymorphism, and message passing demands metrics that are different from structured methodology. There are many studies for OO software metrics such as program complexity or design metrics. But the metrics for the analysis class need to decrease the complexity in the analysis phase so that greatly reduce the effort and the cost of system development. In this paper, we propose new metrics to measure the complexity of analysis classes which draw out in the analysis phase based on Unified Process. By the collaboration complexity, is denoted by CC, we mean the maximum number of the collaborations can be achieved with each of the collaborator and detennine the potential complexity. And the interface complexity, is denoted by IC, shows the difficulty related to understand the interface of collaborators each other. We prove mathematically that the suggested metrics satisfy OO characteristics such as class size and inheritance. And we verify it theoretically for Weyuker' s nine properties. Moreover, we show the computation results for analysis classes of the system which automatically respond to questions of the it's user using the text mining technique. As we compared CC and IC to CBO and WMC, the complexity can be represented by CC and IC more than CBO and WMC. We expect to develop the cost-effective OO software by reviewing the complexity of analysis classes in the first stage of SDLC (Software Development Life Cycle).

  • PDF

A Study on the Correlationship between EEG Complexity by Nonlinear Dynamics Analysis and Impedance Cardiography (비선형 동역학적 방법을 통한 뇌파 복잡도와 임피던스 심장기록법(ICG) 지표와의 상관성 연구)

  • Ryu, Jae-Min;Park, Young-Bae;Park, Young-Jae
    • The Journal of the Society of Korean Medicine Diagnostics
    • /
    • v.11 no.2
    • /
    • pp.128-140
    • /
    • 2007
  • Purpose: We performed this study to examine the correlationship between EEG complexity and impedance cardiography data using correlation analysis. Method: This study performed on 30 healthy subjects(16 males, 14 females). Before and after natural respiration, ICG data were recorded, and EEG raw data were measured by moving windows during 15 minutes. The correlation dimension(D2) was calculated from 15 minutes data. 8 channels EEG data were analysed with 9 index of ICG data by correlation analysis. Result: 1. ACI of impedance cardiography had significant correlationship with ch.4 of EEG complexity(p=0.03). 2. VI of impedance cardiography had significant correlationship with ch.3 of EEG complexity(p=0.034) and ch.4 of EEG complexity(p=0.017). 3. HR, TFC, PEP, LVET, STR of impedance cardiography had no significant correlationship with all of 8 channel EEG complexity. Conclusions: These results suggest that nonlinear analysis of EEG and impedance cardiography have some significant correlationship. And it can make out relationship between brain system and cardiovascular system. In the future, therefore, more study of these fields are necessary.

  • PDF

A Study on the Type and the Facilities in Compositeness of the Domestic Discount Store (국내 대형할인점의 복합화에 따른 유형과 시설에 관한 연구)

  • 문선욱;양정필
    • Korean Institute of Interior Design Journal
    • /
    • no.41
    • /
    • pp.137-145
    • /
    • 2003
  • This research analyzed the space scheme in connection with complexity, one of the new changes in the discount stores, and has a goal of predicting the direction of space scheme in the upcoming complexity era. The research was conducted in the following way. Firstly, this researcher tried to grasp what kinds of changes were required in the overall distribution industry socially and economically. Secondly, the characteristic and situation of discount stores were scrutinized. Thirdly, the domestic stores' complexity status was classified and types of those were elicited. Fourthly, the time-series change and use were analyzed. The result of this analysis reveals that the types of complexity can be divided by location and adjustment to environmental changes. The time-series analysis shows that total operating area, the number of parked cars and the tenant ratio have increased dramatically in 2000 and 2003. And, according to the correlation analysis between factors, the tenant ratio has, a strong correlation with other two factors. Self-complexity takes the basic form of living facilities and complexity with other facilities is combined with other cultural, sales, educational and administrative ones. Mass-complexity is merged with the stadiums, parks or station sites. As you've seen, the concept of complex shopping mall for the realization of one stop shopping and convenience will continue in the days to come. It is desirable that the study on the large-scale shopping spaces will be conducted continually for the preparedness of future life style.

A Study on Complexity Analysis of Extensible Profile Verification Software for Energy Storage System (에너지저장장치용 확장성 프로파일 검증 소프트웨어 복잡도 분석에 관한 연구)

  • Kwon, Hyeokyoung;Ryu, Youngsu;Park, Jaehong;Kwon, Kiwon
    • Journal of Internet Computing and Services
    • /
    • v.17 no.5
    • /
    • pp.59-65
    • /
    • 2016
  • Recently, a study has been progressed about the energy storage system for resolving energy shortage problems in the world. The energy storage system can maximize energy storage system's energy usage by monitoring and controlling about all energy infrastructures on energy network. However, compatibility problems among main components or devices of the energy storage system are obstacles to development of energy storage system products. An extensible profile and extensible profile verification software being able to verify the extensible profile have been required in order to resolve compatibility problems. In this paper, the study on complexity analysis for the extensible profile verification software for the energy storage system is performed. A XML based profile and C language structure based profile are used for analysis of the profile verification software. The complexity of complex verification structure that parses the XML based profile several times and simple verification structure that parses the C language structure based profile are analyzed and compared. Time complexity, space complexity, and cyclomatic complexity are used for complexity analysis. By using these complexity analysis, the study result that compares and analyzes the complexity of XML based and C language structure based profile verification software is presented.