• Title/Summary/Keyword: statistical process

Search Result 3,311, Processing Time 0.028 seconds

A study on the aspect-based sentiment analysis of multilingual customer reviews (다국어 사용자 후기에 대한 속성기반 감성분석 연구)

  • Sungyoung Ji;Siyoon Lee;Daewoo Choi;Kee-Hoon Kang
    • The Korean Journal of Applied Statistics
    • /
    • v.36 no.6
    • /
    • pp.515-528
    • /
    • 2023
  • With the growth of the e-commerce market, consumers increasingly rely on user reviews to make purchasing decisions. Consequently, researchers are actively conducting studies to effectively analyze these reviews. Among the various methods of sentiment analysis, the aspect-based sentiment analysis approach, which examines user reviews from multiple angles rather than solely relying on simple positive or negative sentiments, is gaining widespread attention. Among the various methodologies for aspect-based sentiment analysis, there is an analysis method using a transformer-based model, which is the latest natural language processing technology. In this paper, we conduct an aspect-based sentiment analysis on multilingual user reviews using two real datasets from the latest natural language processing technology model. Specifically, we use restaurant data from the SemEval 2016 public dataset and multilingual user review data from the cosmetic domain. We compare the performance of transformer-based models for aspect-based sentiment analysis and apply various methodologies to improve their performance. Models using multilingual data are expected to be highly useful in that they can analyze multiple languages in one model without building separate models for each language.

CNN Model for Prediction of Tensile Strength based on Pore Distribution Characteristics in Cement Paste (시멘트풀의 공극분포특성에 기반한 인장강도 예측 CNN 모델)

  • Sung-Wook Hong;Tong-Seok Han
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.5
    • /
    • pp.339-346
    • /
    • 2023
  • The uncertainties of microstructural features affect the properties of materials. Numerous pores that are randomly distributed in materials make it difficult to predict the properties of the materials. The distribution of pores in cementitious materials has a great influence on their mechanical properties. Existing studies focus on analyzing the statistical relationship between pore distribution and material responses, and the correlation between them is not yet fully determined. In this study, the mechanical response of cementitious materials is predicted through an image-based data approach using a convolutional neural network (CNN), and the correlation between pore distribution and material response is analyzed. The dataset for machine learning consists of high-resolution micro-CT images and the properties (tensile strength) of cementitious materials. The microstructures are characterized, and the mechanical properties are evaluated through 2D direct tension simulations using the phase-field fracture model. The attributes of input images are analyzed to identify the spot with the greatest influence on the prediction of material response through CNN. The correlation between pore distribution characteristics and material response is analyzed by comparing the active regions during the CNN process and the pore distribution.

Development of SVM-based Construction Project Document Classification Model to Derive Construction Risk (건설 리스크 도출을 위한 SVM 기반의 건설프로젝트 문서 분류 모델 개발)

  • Kang, Donguk;Cho, Mingeon;Cha, Gichun;Park, Seunghee
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.43 no.6
    • /
    • pp.841-849
    • /
    • 2023
  • Construction projects have risks due to various factors such as construction delays and construction accidents. Based on these construction risks, the method of calculating the construction period of the construction project is mainly made by subjective judgment that relies on supervisor experience. In addition, unreasonable shortening construction to meet construction project schedules delayed by construction delays and construction disasters causes negative consequences such as poor construction, and economic losses are caused by the absence of infrastructure due to delayed schedules. Data-based scientific approaches and statistical analysis are needed to solve the risks of such construction projects. Data collected in actual construction projects is stored in unstructured text, so to apply data-based risks, data pre-processing involves a lot of manpower and cost, so basic data through a data classification model using text mining is required. Therefore, in this study, a document-based data generation classification model for risk management was developed through a data classification model based on SVM (Support Vector Machine) by collecting construction project documents and utilizing text mining. Through quantitative analysis through future research results, it is expected that risk management will be possible by being used as efficient and objective basic data for construction project process management.

Examination of Aggregate Quality Using Image Processing Based on Deep-Learning (딥러닝 기반 영상처리를 이용한 골재 품질 검사)

  • Kim, Seong Kyu;Choi, Woo Bin;Lee, Jong Se;Lee, Won Gok;Choi, Gun Oh;Bae, You Suk
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.6
    • /
    • pp.255-266
    • /
    • 2022
  • The quality control of coarse aggregate among aggregates, which are the main ingredients of concrete, is currently carried out by SPC(Statistical Process Control) method through sampling. We construct a smart factory for manufacturing innovation by changing the quality control of coarse aggregates to inspect the coarse aggregates based on this image by acquired images through the camera instead of the current sieve analysis. First, obtained images were preprocessed, and HED(Hollistically-nested Edge Detection) which is the filter learned by deep learning segment each object. After analyzing each aggregate by image processing the segmentation result, fineness modulus and the aggregate shape rate are determined by analyzing result. The quality of aggregate obtained through the video was examined by calculate fineness modulus and aggregate shape rate and the accuracy of the algorithm was more than 90% accurate compared to that of aggregates through the sieve analysis. Furthermore, the aggregate shape rate could not be examined by conventional methods, but the content of this paper also allowed the measurement of the aggregate shape rate. For the aggregate shape rate, it was verified with the length of models, which showed a difference of ±4.5%. In the case of measuring the length of the aggregate, the algorithm result and actual length of the aggregate showed a ±6% difference. Analyzing the actual three-dimensional data in a two-dimensional video made a difference from the actual data, which requires further research.

An Exploratory Study on the Trustworthiness Analysis of Generative AI (생성형 AI의 신뢰도에 대한 탐색적 연구)

  • Soyon Kim;Ji Yeon Cho;Bong Gyou Lee
    • Journal of Internet Computing and Services
    • /
    • v.25 no.1
    • /
    • pp.79-90
    • /
    • 2024
  • This study focused on user trust in ChatGPT, a generative AI technology, and explored the factors that affect usage status and intention to continue using, and whether the influence of trust varies depending on the purpose. For this purpose, the survey was conducted targeting people in their 20s and 30s who use ChatGPT the most. The statistical analysis deploying IBM SPSS 27 and SmartPLS 4.0. A structural equation model was formulated on the foundation of Bhattacherjee's Expectation-Confirmation Model (ECM), employing path analysis and Multi-Group Analysis (MGA) for hypothesis validation. The main findings are as follows: Firstly, ChatGPT is mainly used for specific needs or objectives rather than as a daily tool. The majority of users are cognizant of its hallucination effects; however, this did not hinder its use. Secondly, the hypothesis testing indicated that independent variables such as expectation- confirmation, perceived usefulness, and user satisfaction all exert a positive influence on the dependent variable, the intention for continuance intention. Thirdly, the influence of trust varied depending on the user's purpose in utilizing ChatGPT. trust was significant when ChatGPT is used for information retrieval but not for creative purposes. This study will be used to solve reliability problems in the process of introducing generative AI in society and companies in the future and to establish policies and derive improvement measures for successful employment.

A Foundational Study on Developing a Structural Model for AI-based Sentencing Prediciton Based on Violent Crime Judgment (인공지능기술 적용을 위한 강력범죄 판결문 기반 양형 예측 구조모델 개발 기초 연구)

  • Woongil Park;Eunbi Cho;Jeong-Hyeon Chang;Joo-chang Kim
    • Journal of Internet Computing and Services
    • /
    • v.25 no.1
    • /
    • pp.91-98
    • /
    • 2024
  • With the advancement of ICT (Information and Communication Technology), searching for judgments through the internet has become increasingly convenient. However, predicting sentencing based on judgments remains a challenging task for individuals. This is because sentencing involves a complex process of applying aggravating and mitigating factors within the framework of legal provisions, and it often depends on the subjective judgment of the judge. Therefore, this research aimed to develop a model for predicting sentencing using artificial intelligence by focusing on structuring the data from judgments, making it suitable for AI applications. Through theoretical and statistical analysis of previous studies, we identified variables with high explanatory power for predicting sentencing. Additionally, by analyzing 50 legal judgments related to serious crimes that are publicly available, we presented a framework for extracting essential information from judgments. This framework encompasses basic case information, sentencing details, reasons for sentencing, the reasons for the determination of the sentence, as well as information about offenders, victims, and accomplices evident within the specific content of the judgments. This research is expected to contribute to the development of artificial intelligence technologies in the field of law in the future.

Case Study on Physical Activity Guidance Experience to Maintain Balance in Adults with Cerebellar Ataxia (소뇌성 운동실조증 성인의 균형 유지를 위한 신체활동 지도 경험 사례 연구)

  • Jeonghyeon Kim
    • Journal of Industrial Convergence
    • /
    • v.22 no.3
    • /
    • pp.51-65
    • /
    • 2024
  • This study aims to understand positive changes in balance and gait function and difficulties in the instructor's guidance process through repetitive basic motor skill-based physical activities targeting people with cerebellar ataxia. For this purpose, five adults with cerebellar ataxia were selected as research subjects, and their three instructors participated as research participants. To collect quantitative data, the average and standard deviation were examined through pre-and post-evaluation of the research participants' physical activity classes for 16 weeks. The mean and standard deviation of the collected data were calculated using the Shapiro-Wilk test in the SAS 9.1 statistical program (p<.05). As a qualitative data collection method, the cultural description method of developmental research(DSR) proposed by Spradley(1980) was adopted, and the collected data were analyzed inductively according to the analysis method of Mertens(1990). Through this, 31 concepts, 10 subcategories, and 4 categories were discovered. As a result, the difficulties experienced by the research participants included insufficient guidance environment, dissatisfaction of consumers, difficulty in guidance, and non-cooperation of colleagues. Based on these research results, it was found that institutional, legal, and policy support should be provided not only to public institutions but also to private physical activity institutions that can use vouchers in order to maintain the balance of adults with cerebellar ataxia as well as to guide their physical activities.

An Empirical Analysis of the Determinants of Defense Cost Sharing between Korea and the U.S. (한미 방위비 분담금 결정요인에 대한 실증분석)

  • Yonggi Min;Sunggyun Shin;Yongjoon Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.1
    • /
    • pp.183-192
    • /
    • 2024
  • The purpose of this study is to empirically analyze the determining factors (economy, security, domestic politics, administration, and international politics) that affect the ROK-US defense cost sharing decision. Through this, we will gain a deeper understanding of the defense cost sharing decision process and improve the efficiency of defense cost sharing calculation and execution. The scope of the study is ROK-US defense cost sharing from 1991 to 2021. The data used in the empirical analysis were various secondary data such as Ministry of National Defense, government statistical data, SIPRI, and media reports. As an empirical analysis method, multiple regression analysis using time series was used and the data was analyzed using an autoregressive model. As a result of empirical research through multiple regression analysis, we derived the following results. It was analyzed that the size of Korea's economy, that is, GDP, the previous year's defense cost share, and the number of U.S. troops stationed in Korea had a positive influence on the decision on defense cost sharing. This indicates that Korea's economic growth is a major factor influencing the increase in defense cost sharing, and that the gradual increase in the budget and the negotiation method of the Special Agreement (SMA) for cost sharing of stationing US troops in Korea play an important role. On the other hand, the political tendencies of the ruling party, North Korea's military threats, and China's defense budget were found to have no statistically significant influence on the decision to share defense costs.

Uncertainty Calculation Algorithm for the Estimation of the Radiochronometry of Nuclear Material (핵물질 연대측정을 위한 불확도 추정 알고리즘 연구)

  • JaeChan Park;TaeHoon Jeon;JungHo Song;MinSu Ju;JinYoung Chung;KiNam Kwon;WooChul Choi;JaeHak Cheong
    • Journal of Radiation Industry
    • /
    • v.17 no.4
    • /
    • pp.345-357
    • /
    • 2023
  • Nuclear forensics has been understood as a mendatory component in the international society for nuclear material control and non-proliferation verification. Radiochronometry of nuclear activities for nuclear forensics are decay series characteristics of nuclear materials and the Bateman equation to estimate when nuclear materials were purified and produced. Radiochronometry values have uncertainty of measurement due to the uncertainty factors in the estimation process. These uncertainties should be calculated using appropriate evaluation methods that are representative of the accuracy and reliability. The IAEA, US, and EU have been researched on radiochronometry and uncertainty of measurement, although the uncertainty calculation method using the Bateman equation is limited by the underestimation of the decay constant and the impossibility of estimating the age of more than one generation, so it is necessary to conduct uncertainty calculation research using computer simulation such as Monte Carlo method. This highlights the need for research using computational simulations, such as the Monte Carlo method, to overcome these limitations. In this study, we have analyzed mathematical models and the LHS (Latin Hypercube Sampling) methods to enhance the reliability of radiochronometry which is to develop an uncertainty algorithm for nuclear material radiochronometry using Bateman Equation. We analyzed the LHS method, which can obtain effective statistical results with a small number of samples, and applied it to algorithms that are Monte Carlo methods for uncertainty calculation by computer simulation. This was implemented through the MATLAB computational software. The uncertainty calculation model using mathematical models demonstrated characteristics based on the relationship between sensitivity coefficients and radiative equilibrium. Computational simulation random sampling showed characteristics dependent on random sampling methods, sampling iteration counts, and the probability distribution of uncertainty factors. For validation, we compared models from various international organizations, mathematical models, and the Monte Carlo method. The developed algorithm was found to perform calculations at an equivalent level of accuracy compared to overseas institutions and mathematical model-based methods. To enhance usability, future research and comparisons·validations need to incorporate more complex decay chains and non-homogeneous conditions. The results of this study can serve as foundational technology in the nuclear forensics field, providing tools for the identification of signature nuclides and aiding in the research, development, comparison, and validation of related technologies.

Extraction of Nature Pigment with Antioxidant Properties from Sprout Barley - Optimization Using CCD-RSM (새싹보리로부터 항산화기능성을 갖는 천연색소의 추출 - CCD-RSM을 이용한 최적화)

  • Dong Hwan Kim;Seung Bum Lee
    • Applied Chemistry for Engineering
    • /
    • v.35 no.3
    • /
    • pp.222-229
    • /
    • 2024
  • The use of low-toxic, hypoallergenic, and environmentally friendly natural pigments has increased. With growing interest in health, research on natural extracts containing beneficial substances for the human body is actively underway. In this study, natural pigments were extracted from sprout barley using a solvent extraction method and CCD-RSM was used to optimize the extraction process. The experiment's independent variables included extraction temperature, alcohol/ultra-pure volume ratio, and extraction time. The response variables were set to achieve a target chromaticity (L = 45, a = -35, b = 45), and to maximize DPPH radical scavenging activity evaluating the antioxidant capacity. The statistical significance of the main effect, interaction effect, and effect on the response value was evaluated and analyzed through the F and P values for the regression equation variables calculated using RSM optimization. Additionally, the reliability of the experiment was also confirmed through the P values of the probability plot graph. The extraction conditions for optimizing the four reaction values are 76.1 vol.% alcohol/ultra pure water volume ratio, an extraction temperature of 52.9 ℃ , and an extraction time of 49.6 min. Under these conditions, the theoretical values of the reaction values are L = 45.4, a = -36.8, and b = 45.0 DPPH radical scavenging activity = 30.9%. When the actual experiment was conducted under these optimal extraction conditions and analyzed, the measured values were L = 46.2, a = -36.1, and b = 48.2, and antioxidant capacity = 31.1% with an average error rate of 2.9%.